[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <b5ffed37-5a2a-4bcf-bdc9-532e72aafebc@intel.com>
Date: Wed, 24 Jul 2024 11:37:48 -0700
From: Jacob Keller <jacob.e.keller@...el.com>
To: Larysa Zaremba <larysa.zaremba@...el.com>,
<intel-wired-lan@...ts.osuosl.org>
CC: Tony Nguyen <anthony.l.nguyen@...el.com>, "David S. Miller"
<davem@...emloft.net>, Eric Dumazet <edumazet@...gle.com>, Jakub Kicinski
<kuba@...nel.org>, Paolo Abeni <pabeni@...hat.com>, Alexei Starovoitov
<ast@...nel.org>, Daniel Borkmann <daniel@...earbox.net>, "Jesper Dangaard
Brouer" <hawk@...nel.org>, John Fastabend <john.fastabend@...il.com>, "Maciej
Fijalkowski" <maciej.fijalkowski@...el.com>, <netdev@...r.kernel.org>,
<linux-kernel@...r.kernel.org>, <bpf@...r.kernel.org>,
<magnus.karlsson@...el.com>, Michal Kubiak <michal.kubiak@...el.com>,
Wojciech Drewek <wojciech.drewek@...el.com>, Amritha Nambiar
<amritha.nambiar@...el.com>
Subject: Re: [PATCH iwl-net v2 5/6] ice: remove ICE_CFG_BUSY locking from
AF_XDP code
On 7/24/2024 9:48 AM, Larysa Zaremba wrote:
> Locking used in ice_qp_ena() and ice_qp_dis() does pretty much nothing,
> because ICE_CFG_BUSY is a state flag that is supposed to be set in a PF
> state, not VSI one. Therefore it does not protect the queue pair from
> e.g. reset.
>
Yea, unfortunately a lot of places accidentally use the wrong flags. I
wonder if this is something sparse could help with identifying by having
the flags tagged in some way...
Reviewed-by: Jacob Keller <jacob.e.keller@...el.com>
> Despite being useless, it still can deadlock the unfortunate functions that
> have fell into the same ICE_CFG_BUSY-VSI trap. This happens if ice_qp_ena
> returns an error.
>
This wording makes it sound like other functions have this issue. Is it
only these two left?
Seems like there are a few other places which check this:
> ice_xsk.c
> 176: while (test_and_set_bit(ICE_CFG_BUSY, vsi->state)) {
> 253: clear_bit(ICE_CFG_BUSY, vsi->state);
>
These two are fixed by your patch.
> ice_main.c
> 334: while (test_and_set_bit(ICE_CFG_BUSY, vsi->state))
> 475: clear_bit(ICE_CFG_BUSY, vsi->state);
These two appear to be ice_vsi_sync_fltr.
> 3791: while (test_and_set_bit(ICE_CFG_BUSY, vsi->state))
> 3828: clear_bit(ICE_CFG_BUSY, vsi->state);
These two appear to be ice_vlan_rx_add_vid.
> 3854: while (test_and_set_bit(ICE_CFG_BUSY, vsi->state))
> 3897: clear_bit(ICE_CFG_BUSY, vsi->state);
These two appear to be ice_vlan_rx_kill_vid.
> ice.h
> 299: ICE_CFG_BUSY,
>
This is part of the ice_pf_state enumeration. So yes, we really
shouldn't be checking it in the vsi->state. In the strictest sense this
could be leading to a out-of-bounds read or set, but we happen to luck
into working because the DECLARE_BITMAP uses longs so there is junk data
after the end of the actual state bit size. The bit functions don't get
passed the size so can't have annotations which would catch this.
Obviously not your fault, and don't need to be fixed in this series,
but its at least a semantic bug if not actually trigger-able by
anything. It looks like VLAN functions *are* using this flag
intentionally, if incorrectly. Its unclear what the correct fix is to me
offhand. Perhaps just creating a VSI specific flag for VLANs... or
perhaps replacing the flag with a regular synchronization primitive....
Reviewed-by: Jacob Keller <jacob.e.keller@...el.com>
> Remove ICE_CFG_BUSY locking from ice_qp_dis() and ice_qp_ena().
>
> Fixes: 2d4238f55697 ("ice: Add support for AF_XDP")
> Reviewed-by: Wojciech Drewek <wojciech.drewek@...el.com>
> Signed-off-by: Larysa Zaremba <larysa.zaremba@...el.com>
> ---
> drivers/net/ethernet/intel/ice/ice_xsk.c | 9 ---------
> 1 file changed, 9 deletions(-)
>
> diff --git a/drivers/net/ethernet/intel/ice/ice_xsk.c b/drivers/net/ethernet/intel/ice/ice_xsk.c
> index 5dd50a2866cc..d23fd4ea9129 100644
> --- a/drivers/net/ethernet/intel/ice/ice_xsk.c
> +++ b/drivers/net/ethernet/intel/ice/ice_xsk.c
> @@ -163,7 +163,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx)
> struct ice_q_vector *q_vector;
> struct ice_tx_ring *tx_ring;
> struct ice_rx_ring *rx_ring;
> - int timeout = 50;
> int err;
>
> if (q_idx >= vsi->num_rxq || q_idx >= vsi->num_txq)
> @@ -173,13 +172,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx)
> rx_ring = vsi->rx_rings[q_idx];
> q_vector = rx_ring->q_vector;
>
> - while (test_and_set_bit(ICE_CFG_BUSY, vsi->state)) {
> - timeout--;
> - if (!timeout)
> - return -EBUSY;
> - usleep_range(1000, 2000);
> - }
> -
> ice_qvec_dis_irq(vsi, rx_ring, q_vector);
> ice_qvec_toggle_napi(vsi, q_vector, false);
>
> @@ -250,7 +242,6 @@ static int ice_qp_ena(struct ice_vsi *vsi, u16 q_idx)
> ice_qvec_ena_irq(vsi, q_vector);
>
> netif_tx_start_queue(netdev_get_tx_queue(vsi->netdev, q_idx));
> - clear_bit(ICE_CFG_BUSY, vsi->state);
>
> return 0;
> }
Powered by blists - more mailing lists