[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <c698d201-0099-47a7-b5b0-d2322da63ea0@grimberg.me>
Date: Sat, 27 Dec 2025 12:14:11 +0200
From: Sagi Grimberg <sagi@...mberg.me>
To: Mohamed Khalfella <mkhalfella@...estorage.com>,
Chaitanya Kulkarni <kch@...dia.com>, Christoph Hellwig <hch@....de>,
Jens Axboe <axboe@...nel.dk>, Keith Busch <kbusch@...nel.org>
Cc: Aaron Dailey <adailey@...estorage.com>,
Randy Jennings <randyj@...estorage.com>, John Meneghini
<jmeneghi@...hat.com>, Hannes Reinecke <hare@...e.de>,
linux-nvme@...ts.infradead.org, linux-kernel@...r.kernel.org
Subject: Re: [RFC PATCH 08/14] nvme: Implement cross-controller reset recovery
On 26/11/2025 4:11, Mohamed Khalfella wrote:
> A host that has more than one path connecting to an nvme subsystem
> typically has an nvme controller associated with every path. This is
> mostly applicable to nvmeof. If one path goes down, inflight IOs on that
> path should not be retried immediately on another path because this
> could lead to data corruption as described in TP4129. TP8028 defines
> cross-controller reset mechanism that can be used by host to terminate
> IOs on the failed path using one of the remaining healthy paths. Only
> after IOs are terminated, or long enough time passes as defined by
> TP4129, inflight IOs should be retried on another path. Implement core
> cross-controller reset shared logic to be used by the transports.
>
> Signed-off-by: Mohamed Khalfella <mkhalfella@...estorage.com>
> ---
> drivers/nvme/host/constants.c | 1 +
> drivers/nvme/host/core.c | 133 ++++++++++++++++++++++++++++++++++
> drivers/nvme/host/nvme.h | 10 +++
> 3 files changed, 144 insertions(+)
>
> diff --git a/drivers/nvme/host/constants.c b/drivers/nvme/host/constants.c
> index dc90df9e13a2..f679efd5110e 100644
> --- a/drivers/nvme/host/constants.c
> +++ b/drivers/nvme/host/constants.c
> @@ -46,6 +46,7 @@ static const char * const nvme_admin_ops[] = {
> [nvme_admin_virtual_mgmt] = "Virtual Management",
> [nvme_admin_nvme_mi_send] = "NVMe Send MI",
> [nvme_admin_nvme_mi_recv] = "NVMe Receive MI",
> + [nvme_admin_cross_ctrl_reset] = "Cross Controller Reset",
> [nvme_admin_dbbuf] = "Doorbell Buffer Config",
> [nvme_admin_format_nvm] = "Format NVM",
> [nvme_admin_security_send] = "Security Send",
> diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c
> index f5b84bc327d3..f38b70ca9cee 100644
> --- a/drivers/nvme/host/core.c
> +++ b/drivers/nvme/host/core.c
> @@ -554,6 +554,138 @@ void nvme_cancel_admin_tagset(struct nvme_ctrl *ctrl)
> }
> EXPORT_SYMBOL_GPL(nvme_cancel_admin_tagset);
>
> +static struct nvme_ctrl *nvme_find_ccr_ctrl(struct nvme_ctrl *ictrl,
> + u32 min_cntlid)
> +{
> + struct nvme_subsystem *subsys = ictrl->subsys;
> + struct nvme_ctrl *sctrl;
> + unsigned long flags;
> +
> + mutex_lock(&nvme_subsystems_lock);
This looks like the wrong lock to take here?
> + list_for_each_entry(sctrl, &subsys->ctrls, subsys_entry) {
> + if (sctrl->cntlid < min_cntlid)
> + continue;
The use of min_cntlid is not clear to me.
> +
> + if (atomic_dec_if_positive(&sctrl->ccr_limit) < 0)
> + continue;
> +
> + spin_lock_irqsave(&sctrl->lock, flags);
> + if (sctrl->state != NVME_CTRL_LIVE) {
> + spin_unlock_irqrestore(&sctrl->lock, flags);
> + atomic_inc(&sctrl->ccr_limit);
> + continue;
> + }
> +
> + /*
> + * We got a good candidate source controller that is locked and
> + * LIVE. However, no guarantee sctrl will not be deleted after
> + * sctrl->lock is released. Get a ref of both sctrl and admin_q
> + * so they do not disappear until we are done with them.
> + */
> + WARN_ON_ONCE(!blk_get_queue(sctrl->admin_q));
> + nvme_get_ctrl(sctrl);
> + spin_unlock_irqrestore(&sctrl->lock, flags);
> + goto found;
> + }
> + sctrl = NULL;
> +found:
> + mutex_unlock(&nvme_subsystems_lock);
> + return sctrl;
> +}
> +
> +static int nvme_issue_wait_ccr(struct nvme_ctrl *sctrl, struct nvme_ctrl *ictrl)
> +{
> + unsigned long flags, tmo, remain;
> + struct nvme_ccr_entry ccr = { };
> + union nvme_result res = { 0 };
> + struct nvme_command c = { };
> + u32 result;
> + int ret = 0;
> +
> + init_completion(&ccr.complete);
> + ccr.ictrl = ictrl;
> +
> + spin_lock_irqsave(&sctrl->lock, flags);
> + list_add_tail(&ccr.list, &sctrl->ccrs);
> + spin_unlock_irqrestore(&sctrl->lock, flags);
> +
> + c.ccr.opcode = nvme_admin_cross_ctrl_reset;
> + c.ccr.ciu = ictrl->ciu;
> + c.ccr.icid = cpu_to_le16(ictrl->cntlid);
> + c.ccr.cirn = cpu_to_le64(ictrl->cirn);
> + ret = __nvme_submit_sync_cmd(sctrl->admin_q, &c, &res,
> + NULL, 0, NVME_QID_ANY, 0);
> + if (ret)
> + goto out;
> +
> + result = le32_to_cpu(res.u32);
> + if (result & 0x01) /* Immediate Reset */
> + goto out;
> +
> + tmo = msecs_to_jiffies(max(ictrl->cqt, ictrl->kato * 1000));
> + remain = wait_for_completion_timeout(&ccr.complete, tmo);
> + if (!remain)
I think remain is redundant here.
> + ret = -EAGAIN;
> +out:
> + spin_lock_irqsave(&sctrl->lock, flags);
> + list_del(&ccr.list);
> + spin_unlock_irqrestore(&sctrl->lock, flags);
> + return ccr.ccrs == 1 ? 0 : ret;
Why would you still return 0 and not EAGAIN? you expired on timeout but
still
return success if you have ccrs=1? btw you have ccrs in the ccr struct
and in the controller
as a list. Lets rename to distinguish the two.
> +}
> +
> +unsigned long nvme_recover_ctrl(struct nvme_ctrl *ictrl)
> +{
I'd call it nvme_fence_controller()
> + unsigned long deadline, now, timeout;
> + struct nvme_ctrl *sctrl;
> + u32 min_cntlid = 0;
> + int ret;
> +
> + timeout = nvme_recovery_timeout_ms(ictrl);
> + dev_info(ictrl->device, "attempting CCR, timeout %lums\n", timeout);
> +
> + now = jiffies;
> + deadline = now + msecs_to_jiffies(timeout);
> + while (time_before(now, deadline)) {
> + sctrl = nvme_find_ccr_ctrl(ictrl, min_cntlid);
> + if (!sctrl) {
> + /* CCR failed, switch to time-based recovery */
> + return deadline - now;
It is not clear what is the return code semantics of this function.
How about making it success/failure and have the caller choose what to do?
> + }
> +
> + ret = nvme_issue_wait_ccr(sctrl, ictrl);
> + atomic_inc(&sctrl->ccr_limit);
inc after you wait for the ccr? shouldn't this be before?
> +
> + if (!ret) {
> + dev_info(ictrl->device, "CCR succeeded using %s\n",
> + dev_name(sctrl->device));
> + blk_put_queue(sctrl->admin_q);
> + nvme_put_ctrl(sctrl);
> + return 0;
> + }
> +
> + /* Try another controller */
> + min_cntlid = sctrl->cntlid + 1;
OK, I see why min_cntlid is used. That is very non-intuitive.
I'm wandering if it will be simpler to take one-shot at ccr and
if it fails fallback to crt. I mean, if the sctrl is alive, and it was
unable
to reset the ictrl in time, how would another ctrl do a better job here?
> + blk_put_queue(sctrl->admin_q);
> + nvme_put_ctrl(sctrl);
> + now = jiffies;
> + }
> +
> + dev_info(ictrl->device, "CCR reached timeout, call it done\n");
> + return 0;
> +}
> +EXPORT_SYMBOL_GPL(nvme_recover_ctrl);
> +
> +void nvme_end_ctrl_recovery(struct nvme_ctrl *ctrl)
> +{
> + unsigned long flags;
> +
> + spin_lock_irqsave(&ctrl->lock, flags);
> + WRITE_ONCE(ctrl->state, NVME_CTRL_RESETTING);
This needs to be a proper state transition.
Powered by blists - more mailing lists