[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190624072752.GA3954@lst.de>
Date: Mon, 24 Jun 2019 09:27:52 +0200
From: Christoph Hellwig <hch@....de>
To: Logan Gunthorpe <logang@...tatee.com>
Cc: linux-kernel@...r.kernel.org, linux-block@...r.kernel.org,
linux-nvme@...ts.infradead.org, linux-pci@...r.kernel.org,
linux-rdma@...r.kernel.org, Jens Axboe <axboe@...nel.dk>,
Christoph Hellwig <hch@....de>,
Bjorn Helgaas <bhelgaas@...gle.com>,
Dan Williams <dan.j.williams@...el.com>,
Sagi Grimberg <sagi@...mberg.me>,
Keith Busch <kbusch@...nel.org>,
Jason Gunthorpe <jgg@...pe.ca>,
Stephen Bates <sbates@...thlin.com>
Subject: Re: [RFC PATCH 00/28] Removing struct page from P2PDMA
This is not going to fly.
For one passing a dma_addr_t through the block layer is a layering
violation, and one that I think will also bite us in practice.
The host physical to PCIe bus address mapping can have offsets, and
those offsets absolutely can be different for differnet root ports.
So with your caller generated dma_addr_t everything works fine with
a switched setup as the one you are probably testing on, but on a
sufficiently complicated setup with multiple root ports it can break.
Also duplicating the whole block I/O stack, including hooks all over
the fast path is pretty much a no-go.
I've been pondering for a while if we wouldn't be better off just
passing a phys_addr_t + len instead of the page, offset, len tuple
in the bio_vec, though. If you look at the normal I/O path here
is what we normally do:
- we get a page as input, either because we have it at hand (e.g.
from the page cache) or from get_user_pages (which actually caculates
it from a pfn in the page tables)
- once in the bio all the merging decisions are based on the physical
address, so we have to convert it to the physical address there,
potentially multiple times
- then dma mapping all works off the physical address, which it gets
from the page at the start
- then only the dma address is used for the I/O
- on I/O completion we often but not always need the page again. In
the direct I/O case for reference counting and dirty status, in the
file system also for things like marking the page uptodate
So if we move to a phys_addr_t we'd need to go back to the page at least
once. But because of how the merging works we really only need to do
it once per segment, as we can just do pointer arithmerics do get the
following pages. As we generally go at least once from a physical
address to a page in the merging code even a relatively expensive vmem_map
looks shouldn't be too bad. Even more so given that the super hot path
(small blkdev direct I/O) can actually trivially cache the affected pages
as well.
Linus kinda hates the pfn approach, but part of that was really that
it was proposed for file system data, which we all found out really
can't work as-is without pages the hard way. Another part probably
was potential performance issue, but between the few page lookups, and
the fact that using a single phys_addr_t instead of pfn/page + offset
should avoid quite a few calculations performance should not actually
be affected, although we'll have to be careful to actually verify that.
Powered by blists - more mailing lists