[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <6de64ad1fa9c4c82de190d4f71923984979782d8.camel@nvidia.com>
Date: Wed, 24 May 2023 12:03:16 +0000
From: Dragos Tatulea <dtatulea@...dia.com>
To: Tariq Toukan <tariqt@...dia.com>, "ttoukan.linux@...il.com"
<ttoukan.linux@...il.com>, "linyunsheng@...wei.com" <linyunsheng@...wei.com>,
"jbrouer@...hat.com" <jbrouer@...hat.com>, "saeed@...nel.org"
<saeed@...nel.org>, Saeed Mahameed <saeedm@...dia.com>,
"netdev@...r.kernel.org" <netdev@...r.kernel.org>
CC: "maxtram95@...il.com" <maxtram95@...il.com>, "lorenzo@...nel.org"
<lorenzo@...nel.org>, "alexander.duyck@...il.com"
<alexander.duyck@...il.com>, "kheib@...hat.com" <kheib@...hat.com>,
"ilias.apalodimas@...aro.org" <ilias.apalodimas@...aro.org>,
"mkabat@...hat.com" <mkabat@...hat.com>, "brouer@...hat.com"
<brouer@...hat.com>, "atzin@...hat.com" <atzin@...hat.com>,
"fmaurer@...hat.com" <fmaurer@...hat.com>, "bpf@...r.kernel.org"
<bpf@...r.kernel.org>, "jbenc@...hat.com" <jbenc@...hat.com>
Subject: Re: mlx5 XDP redirect leaking memory on kernel 6.3
On Wed, 2023-05-24 at 19:26 +0800, Yunsheng Lin wrote:
> On 2023/5/24 0:35, Dragos Tatulea wrote:
> >
> > On Tue, 2023-05-23 at 17:55 +0200, Jesper Dangaard Brouer wrote:
> > >
> > > When the mlx5 driver runs an XDP program doing XDP_REDIRECT, then memory
> > > is getting leaked. Other XDP actions, like XDP_DROP, XDP_PASS and XDP_TX
> > > works correctly. I tested both redirecting back out same mlx5 device and
> > > cpumap redirect (with XDP_PASS), which both cause leaking.
> > >
> > > After removing the XDP prog, which also cause the page_pool to be
> > > released by mlx5, then the leaks are visible via the page_pool periodic
> > > inflight reports. I have this bpftrace[1] tool that I also use to detect
> > > the problem faster (not waiting 60 sec for a report).
> > >
> > > [1]
> > > https://github.com/xdp-project/xdp-project/blob/master/areas/mem/bpftrace/page_pool_track_shutdown01.bt
> > >
> > > I've been debugging and reading through the code for a couple of days,
> > > but I've not found the root-cause, yet. I would appreciate new ideas
> > > where to look and fresh eyes on the issue.
> > >
> > >
> > > To Lin, it looks like mlx5 uses PP_FLAG_PAGE_FRAG, and my current
> > > suspicion is that mlx5 driver doesn't fully release the bias count (hint
> > > see MLX5E_PAGECNT_BIAS_MAX).
>
> It seems mlx5 is implementing it's own frag allocation scheme, it there a
> reason why the native frag allocation scheme in page pool is not used? To
> avoid the "((page->pp_magic & ~0x3UL) == PP_SIGNATURE)" checking?
mlx5 uses fragmentation of the page from within the driver instead of the pre-
partitioning of the page using page_pool_alloc_frag(). As shown in commit
52cc6ffc0ab2 ("page_pool: Refactor page_pool to enable fragmenting after
allocation")
The exception is however the following optimization:
page_pool_put_defragged_page() can be called for XDP_TX directly to avoid the
overhead of fragment management. That's because mlx5 currently supports only one
packet per page for XDP.
> > >
> >
> > Thanks for the report Jesper. Incidentally I've just picked up this issue
> > today
> > as well.
> >
> > On XDP redirect and tx, the page is set to skip the bias counter release
> > with
> > the expectation that page_pool_put_defragged_page will be called from [1].
> > But,
>
> page_pool_put_defragged_page() can only be called when there is only user
> using
> the page, I am not sure how it can ensure that yet.
>
See previous comment.
> > as I found out now, during XDP redirect only one fragment of the page is
> > released in xdp core [2]. This is where the leak is coming from.
> >
> > We'll provide a fix soon.
> >
> > [1]
> > https://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next.git/tree/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c#n665
> >
> > [2]
> > https://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next.git/tree/net/core/xdp.c#n390
> >
> > Thanks,
> > Dragos
> >
> >
Powered by blists - more mailing lists