[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <40d69074-31a8-d06a-ade9-90de7712c553@deltatee.com>
Date: Thu, 1 Mar 2018 10:40:11 -0700
From: Logan Gunthorpe <logang@...tatee.com>
To: Sagi Grimberg <sagi@...mberg.me>, linux-kernel@...r.kernel.org,
linux-pci@...r.kernel.org, linux-nvme@...ts.infradead.org,
linux-rdma@...r.kernel.org, linux-nvdimm@...ts.01.org,
linux-block@...r.kernel.org
Cc: Stephen Bates <sbates@...thlin.com>,
Christoph Hellwig <hch@....de>, Jens Axboe <axboe@...nel.dk>,
Keith Busch <keith.busch@...el.com>,
Bjorn Helgaas <bhelgaas@...gle.com>,
Jason Gunthorpe <jgg@...lanox.com>,
Max Gurtovoy <maxg@...lanox.com>,
Dan Williams <dan.j.williams@...el.com>,
Jérôme Glisse <jglisse@...hat.com>,
Benjamin Herrenschmidt <benh@...nel.crashing.org>,
Alex Williamson <alex.williamson@...hat.com>,
Steve Wise <swise@...ngridcomputing.com>
Subject: Re: [PATCH v2 10/10] nvmet: Optionally use PCI P2P memory
On 01/03/18 04:03 AM, Sagi Grimberg wrote:
> Can you describe what would be the plan to have it when these devices
> do come along? I'd say that p2p_dev needs to become a nvmet_ns reference
> and not from nvmet_ctrl. Then, when cmb capable devices come along, the
> ns can prefer to use its own cmb instead of locating a p2p_dev device?
The patchset already supports CMB drives. That's essentially what patch
7 is for. We change the nvme-pci driver to use the p2pmem code to
register and manage the CMB memory. After that, it is simply available
to the nvmet code. We have already been using this with a couple
prototype CMB drives.
>> +static int nvmet_p2pdma_add_client(struct nvmet_ctrl *ctrl,
>> + struct nvmet_ns *ns)
>> +{
>> + int ret;
>> +
>> + if (!blk_queue_pci_p2pdma(ns->bdev->bd_queue)) {
>> + pr_err("peer-to-peer DMA is not supported by %s\n",
>> + ns->device_path);
>> + return -EINVAL;
>
> I'd say that just skip it instead of failing it, in theory one can
> connect nvme devices via p2p mem and expose other devices in the
> same subsystem. The message would be pr_debug to reduce chattiness.
No, it's actually a bit more complicated than you make it. There's a
couple cases:
1) The user adds a namespace but there hasn't been a connection and no
p2pmem has been selected. In this case the add_client function is never
called and the user can add whatever namespace they like.
2) When the first connect happens, nvmet_setup_p2pmem() will call
add_client() for each namespace and rdma device. If any of them fail
then it does not use a P2P device and falls back, as you'd like, to
regular memory.
3) When the user adds a namespace after a port is in use and a
compatible P2P device has been found. In this case, if the user tries to
add a namespace that is not compatible with the P2P device in use then
it fails adding the new namespace. The only alternative here is to tear
everything down, ensure there are no P2P enabled buffers open and start
using regular memory again... That is very difficult.
I also disagree that these messages should be pr_debug. If a user is
trying to use P2P memory and they enable it but the system doesn't
choose to use their memory they must know why that is so they can make
the necessary adjustments. If the system doesn't at least print a dmesg
they are in the dark.
>> + }
>> +
>> + ret = pci_p2pdma_add_client(&ctrl->p2p_clients, nvmet_ns_dev(ns));
>> + if (ret)
>> + pr_err("failed to add peer-to-peer DMA client %s: %d\n",
>> + ns->device_path, ret);
>> +
>> + return ret;
>> +}
>> +
>> int nvmet_ns_enable(struct nvmet_ns *ns)
>> {
>> struct nvmet_subsys *subsys = ns->subsys;
>> @@ -299,6 +319,14 @@ int nvmet_ns_enable(struct nvmet_ns *ns)
>> if (ret)
>> goto out_blkdev_put;
>> + list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) {
>> + if (ctrl->p2p_dev) {
>> + ret = nvmet_p2pdma_add_client(ctrl, ns);
>> + if (ret)
>> + goto out_remove_clients;
>
> Is this really a fatal failure given that we fall-back to main
> memory? Why not continue with main memory (and warn at best)?
See above. It's fatal because we are already using p2p memory and we
can't easily tear that all down when a user adds a new namespace.
>> + ctrl->p2p_dev = pci_p2pmem_find(&ctrl->p2p_clients);
>
> This is the first p2p_dev found right? What happens if I have more than
> a single p2p device? In theory I'd have more p2p memory I can use. Have
> you considered making pci_p2pmem_find return the least used suitable
> device?
Yes, it currently returns the first found. I imagine a bunch of
improvements could be made to it in future work. However, I'd probably
start with finding the nearest p2p device and then falling back to the
least used if that's necessary. At this time though it's not clear how
complicated these systems will get and what's actually needed here.
>> @@ -68,6 +69,7 @@ struct nvmet_rdma_rsp {
>> u8 n_rdma;
>> u32 flags;
>> u32 invalidate_rkey;
>> + struct pci_dev *p2p_dev;
>
> Given that p2p_client is in nvmet_req, I think it make sense
> that the p2p_dev itself would also live there. In theory, nothing
> is preventing FC from using it as well.
Fair point. I can look at moving it in v3.
Thanks,
Logan
Powered by blists - more mailing lists