[<prev] [next>] [day] [month] [year] [list]
Message-ID: <SJ0PR18MB3882E0E7B9E5C7FCEF5D6787CCDA9@SJ0PR18MB3882.namprd18.prod.outlook.com>
Date: Tue, 14 Sep 2021 15:40:57 +0000
From: Shai Malin <smalin@...vell.com>
To: Leon Romanovsky <leon@...nel.org>
CC: "netdev@...r.kernel.org" <netdev@...r.kernel.org>,
"davem@...emloft.net" <davem@...emloft.net>,
"kuba@...nel.org" <kuba@...nel.org>,
"linux-rdma@...r.kernel.org" <linux-rdma@...r.kernel.org>,
"jgg@...pe.ca" <jgg@...pe.ca>, Ariel Elior <aelior@...vell.com>,
"malin1024@...il.com" <malin1024@...il.com>,
Michal Kalderon <mkalderon@...vell.com>
Subject: RE: [PATCH net] qed: rdma - don't wait for resources under hw error
recovery flow
On Tue, Sep 14, 2021 at 1:01PM +0300, Leon Romanovsky wrote:
> On Tue, Sep 14, 2021 at 06:23:02AM +0000, Shai Malin wrote:
> > On Mon, Sep 13, 2021 at 5:45:00PM +0300, Leon Romanovsky wrote:
> > > On Mon, Sep 13, 2021 at 03:14:42PM +0300, Shai Malin wrote:
> > > > If the HW device is during recovery, the HW resources will never return,
> > > > hence we shouldn't wait for the CID (HW context ID) bitmaps to clear.
> > > > This fix speeds up the error recovery flow.
> > > >
> > > > Fixes: 64515dc899df ("qed: Add infrastructure for error detection and
> > > recovery")
> > > > Signed-off-by: Michal Kalderon <mkalderon@...vell.com>
> > > > Signed-off-by: Ariel Elior <aelior@...vell.com>
> > > > Signed-off-by: Shai Malin <smalin@...vell.com>
> > > > ---
> > > > drivers/net/ethernet/qlogic/qed/qed_iwarp.c | 7 +++++++
> > > > drivers/net/ethernet/qlogic/qed/qed_roce.c | 7 +++++++
> > > > 2 files changed, 14 insertions(+)
> > > >
> > > > diff --git a/drivers/net/ethernet/qlogic/qed/qed_iwarp.c
> > > b/drivers/net/ethernet/qlogic/qed/qed_iwarp.c
> > > > index fc8b3e64f153..4967e383c31a 100644
> > > > --- a/drivers/net/ethernet/qlogic/qed/qed_iwarp.c
> > > > +++ b/drivers/net/ethernet/qlogic/qed/qed_iwarp.c
> > > > @@ -1323,6 +1323,13 @@ static int qed_iwarp_wait_for_all_cids(struct
> > > qed_hwfn *p_hwfn)
> > > > int rc;
> > > > int i;
> > > >
> > > > + /* If the HW device is during recovery, all resources are immediately
> > > > + * reset without receiving a per-cid indication from HW. In this case
> > > > + * we don't expect the cid_map to be cleared.
> > > > + */
> > > > + if (p_hwfn->cdev->recov_in_prog)
> > > > + return 0;
> > >
> > > How do you ensure that this doesn't race with recovery flow?
> >
> > The HW recovery will start with the management FW which will detect and
> report
> > the problem to the driver and it also set "cdev->recov_in_prog = ture" for all
> > the devices on the same HW.
> > The qedr recovery flow is actually the qedr_remove flow but if
> > "cdev->recov_in_prog = true" it will "ignore" the FW/HW resources.
> > The changes introduced with this patch are part of this qedr remove flow.
> > The cdev->recov_in_prog will be set to false only as part of the following
> > probe and after the HW was re-initialized.
>
> I asked how do you make sure that recov_in_prog is not changing to be
> "true" right after your "if ..." check?
>
> Thanks
Thanks Leon - it's a valid point. Moving the "if..." to the while loop
for both RoCE and iWARP will solve it.
I will fix it with V2.
>
> >
> > >
> > > > +
> > > > rc = qed_iwarp_wait_cid_map_cleared(p_hwfn,
> > > > &p_hwfn->p_rdma_info-
> > > >tcp_cid_map);
> > > > if (rc)
> > > > diff --git a/drivers/net/ethernet/qlogic/qed/qed_roce.c
> > > b/drivers/net/ethernet/qlogic/qed/qed_roce.c
> > > > index f16a157bb95a..aff5a2871b8f 100644
> > > > --- a/drivers/net/ethernet/qlogic/qed/qed_roce.c
> > > > +++ b/drivers/net/ethernet/qlogic/qed/qed_roce.c
> > > > @@ -71,6 +71,13 @@ void qed_roce_stop(struct qed_hwfn *p_hwfn)
> > > > struct qed_bmap *rcid_map = &p_hwfn->p_rdma_info->real_cid_map;
> > > > int wait_count = 0;
> > > >
> > > > + /* If the HW device is during recovery, all resources are immediately
> > > > + * reset without receiving a per-cid indication from HW. In this case
> > > > + * we don't expect the cid bitmap to be cleared.
> > > > + */
> > > > + if (p_hwfn->cdev->recov_in_prog)
> > > > + return;
> > > > +
> > > > /* when destroying a_RoCE QP the control is returned to the user after
> > > > * the synchronous part. The asynchronous part may take a little longer.
> > > > * We delay for a short while if an async destroy QP is still expected.
> > > > --
> > > > 2.22.0
> > > >
Powered by blists - more mailing lists