[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20240219134113.24396ad9@meshulam.tesarici.cz>
Date: Mon, 19 Feb 2024 13:41:13 +0100
From: Petr Tesařík <petr@...arici.cz>
To: Will Deacon <will@...nel.org>
Cc: linux-kernel@...r.kernel.org, kernel-team@...roid.com,
iommu@...ts.linux.dev, Christoph Hellwig <hch@....de>, Marek Szyprowski
<m.szyprowski@...sung.com>, Robin Murphy <robin.murphy@....com>, Petr
Tesarik <petr.tesarik1@...wei-partners.com>, Dexuan Cui
<decui@...rosoft.com>
Subject: Re: [PATCH v3 2/3] swiotlb: Enforce page alignment in
swiotlb_alloc()
On Mon, 5 Feb 2024 19:01:26 +0000
Will Deacon <will@...nel.org> wrote:
> When allocating pages from a restricted DMA pool in swiotlb_alloc(),
> the buffer address is blindly converted to a 'struct page *' that is
> returned to the caller. In the unlikely event of an allocation bug,
> page-unaligned addresses are not detected and slots can silently be
> double-allocated.
>
> Add a simple check of the buffer alignment in swiotlb_alloc() to make
> debugging a little easier if something has gone wonky.
>
> Cc: Christoph Hellwig <hch@....de>
> Cc: Marek Szyprowski <m.szyprowski@...sung.com>
> Cc: Robin Murphy <robin.murphy@....com>
> Cc: Petr Tesarik <petr.tesarik1@...wei-partners.com>
> Cc: Dexuan Cui <decui@...rosoft.com>
> Signed-off-by: Will Deacon <will@...nel.org>
Reviewed-by: Petr Tesarik <petr.tesarik1@...wei-partners.com>
Petr T
> ---
> kernel/dma/swiotlb.c | 6 ++++++
> 1 file changed, 6 insertions(+)
>
> diff --git a/kernel/dma/swiotlb.c b/kernel/dma/swiotlb.c
> index 9ff909a0039a..adbb3143238b 100644
> --- a/kernel/dma/swiotlb.c
> +++ b/kernel/dma/swiotlb.c
> @@ -1643,6 +1643,12 @@ struct page *swiotlb_alloc(struct device *dev, size_t size)
> return NULL;
>
> tlb_addr = slot_addr(pool->start, index);
> + if (unlikely(!PAGE_ALIGNED(tlb_addr))) {
> + dev_WARN_ONCE(dev, 1, "Cannot allocate pages from non page-aligned swiotlb addr 0x%pa.\n",
> + &tlb_addr);
> + swiotlb_release_slots(dev, tlb_addr);
> + return NULL;
> + }
>
> return pfn_to_page(PFN_DOWN(tlb_addr));
> }
Powered by blists - more mailing lists