lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CABdmKX0Du2F+bko=hjLBqdQO-bJSFcG3y1Bbuu2v6J8aVB39sw@mail.gmail.com>
Date: Mon, 5 Feb 2024 12:26:10 -0800
From: "T.J. Mercier" <tjmercier@...gle.com>
To: Michal Hocko <mhocko@...e.com>
Cc: Johannes Weiner <hannes@...xchg.org>, Roman Gushchin <roman.gushchin@...ux.dev>, 
	Shakeel Butt <shakeelb@...gle.com>, Muchun Song <muchun.song@...ux.dev>, 
	Andrew Morton <akpm@...ux-foundation.org>, Efly Young <yangyifei03@...ishou.com>, 
	android-mm@...gle.com, yuzhao@...gle.com, mkoutny@...e.com, 
	Yosry Ahmed <yosryahmed@...gle.com>, cgroups@...r.kernel.org, linux-mm@...ck.org, 
	linux-kernel@...r.kernel.org
Subject: Re: [PATCH v3] mm: memcg: Use larger batches for proactive reclaim

On Mon, Feb 5, 2024 at 11:40 AM Michal Hocko <mhocko@...e.com> wrote:
>
> On Mon 05-02-24 11:29:49, T.J. Mercier wrote:
> > On Mon, Feb 5, 2024 at 2:40 AM Michal Hocko <mhocko@...e.com> wrote:
> > >
> > > On Fri 02-02-24 23:38:54, T.J. Mercier wrote:
> > > > Before 388536ac291 ("mm:vmscan: fix inaccurate reclaim during proactive
> > > > reclaim") we passed the number of pages for the reclaim request directly
> > > > to try_to_free_mem_cgroup_pages, which could lead to significant
> > > > overreclaim. After 0388536ac291 the number of pages was limited to a
> > > > maximum 32 (SWAP_CLUSTER_MAX) to reduce the amount of overreclaim.
> > > > However such a small batch size caused a regression in reclaim
> > > > performance due to many more reclaim start/stop cycles inside
> > > > memory_reclaim.
> > >
> > > You have mentioned that in one of the previous emails but it is good to
> > > mention what is the source of that overhead for the future reference.
> >
> > I can add a sentence about the restart cost being amortized over more
> > pages with a large batch size. It covers things like repeatedly
> > flushing stats, walking the tree, evaluating protection limits, etc.
> >
> > > > Reclaim tries to balance nr_to_reclaim fidelity with fairness across
> > > > nodes and cgroups over which the pages are spread. As such, the bigger
> > > > the request, the bigger the absolute overreclaim error. Historic
> > > > in-kernel users of reclaim have used fixed, small sized requests to
> > > > approach an appropriate reclaim rate over time. When we reclaim a user
> > > > request of arbitrary size, use decaying batch sizes to manage error while
> > > > maintaining reasonable throughput.
> > >
> > > These numbers are with MGLRU or the default reclaim implementation?
> >
> > These numbers are for both. root uses the memcg LRU (MGLRU was
> > enabled), and /uid_0 does not.
>
> Thanks it would be nice to outline that in the changelog.

Ok, I'll update the table below for each case.

> > > > root - full reclaim       pages/sec   time (sec)
> > > > pre-0388536ac291      :    68047        10.46
> > > > post-0388536ac291     :    13742        inf
> > > > (reclaim-reclaimed)/4 :    67352        10.51
> > > >
> > > > /uid_0 - 1G reclaim       pages/sec   time (sec)  overreclaim (MiB)
> > > > pre-0388536ac291      :    258822       1.12            107.8
> > > > post-0388536ac291     :    105174       2.49            3.5
> > > > (reclaim-reclaimed)/4 :    233396       1.12            -7.4
> > > >
> > > > /uid_0 - full reclaim     pages/sec   time (sec)
> > > > pre-0388536ac291      :    72334        7.09
> > > > post-0388536ac291     :    38105        14.45
> > > > (reclaim-reclaimed)/4 :    72914        6.96
> > > >
> > > > Fixes: 0388536ac291 ("mm:vmscan: fix inaccurate reclaim during proactive reclaim")
> > > > Signed-off-by: T.J. Mercier <tjmercier@...gle.com>
> > > > Reviewed-by: Yosry Ahmed <yosryahmed@...gle.com>
> > > > Acked-by: Johannes Weiner <hannes@...xchg.org>
> > > >
> > > > ---
> > > > v3: Formatting fixes per Yosry Ahmed and Johannes Weiner. No functional
> > > > changes.
> > > > v2: Simplify the request size calculation per Johannes Weiner and Michal Koutný
> > > >
> > > >  mm/memcontrol.c | 6 ++++--
> > > >  1 file changed, 4 insertions(+), 2 deletions(-)
> > > >
> > > > diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> > > > index 46d8d02114cf..f6ab61128869 100644
> > > > --- a/mm/memcontrol.c
> > > > +++ b/mm/memcontrol.c
> > > > @@ -6976,9 +6976,11 @@ static ssize_t memory_reclaim(struct kernfs_open_file *of, char *buf,
> > > >               if (!nr_retries)
> > > >                       lru_add_drain_all();
> > > >
> > > > +             /* Will converge on zero, but reclaim enforces a minimum */
> > > > +             unsigned long batch_size = (nr_to_reclaim - nr_reclaimed) / 4;
> > >
> > > This doesn't fit into the existing coding style. I do not think there is
> > > a strong reason to go against it here.
> >
> > There's been some back and forth here. You'd prefer to move this to
> > the top of the while loop, under the declaration of reclaimed? It's
> > farther from its use there, but it does match the existing style in
> > the file better.
>
> This is not something I deeply care about but generally it is better to
> not mix styles unless that is a clear win. If you want to save one LOC
> you can just move it up - just couple of lines up, or you can keep the
> definition closer and have a separate declaration.

I find it nicer to have to search as little as possible for both the
declaration (type) and definition, but I am not attached to it either
and it's not worth annoying anyone over here. Let's move it up like
Yosry suggested initially.

> > > > +
> > > >               reclaimed = try_to_free_mem_cgroup_pages(memcg,
> > > > -                                     min(nr_to_reclaim - nr_reclaimed, SWAP_CLUSTER_MAX),
> > > > -                                     GFP_KERNEL, reclaim_options);
> > > > +                                     batch_size, GFP_KERNEL, reclaim_options);
> > >
> > > Also with the increased reclaim target do we need something like this?
> > >
> > > diff --git a/mm/vmscan.c b/mm/vmscan.c
> > > index 4f9c854ce6cc..94794cf5ee9f 100644
> > > --- a/mm/vmscan.c
> > > +++ b/mm/vmscan.c
> > > @@ -1889,7 +1889,7 @@ static unsigned long shrink_inactive_list(unsigned long nr_to_scan,
> > >
> > >                 /* We are about to die and free our memory. Return now. */
> > >                 if (fatal_signal_pending(current))
> > > -                       return SWAP_CLUSTER_MAX;
> > > +                       return sc->nr_to_reclaim;
> > >         }
> > >
> > >         lru_add_drain();
> > > >
> > > >               if (!reclaimed && !nr_retries--)
> > > >                       return -EAGAIN;
> > > > --
> >
> > This is interesting, but I don't think it's closely related to this
> > change. This section looks like it was added to delay OOM kills due to
> > apparent lack of reclaim progress when pages are isolated and the
> > direct reclaimer is scheduled out. A couple things:
> >
> > In the context of proactive reclaim, current is not really undergoing
> > reclaim due to memory pressure. It's initiated from userspace. So
> > whether it has a fatal signal pending or not doesn't seem like it
> > should influence the return value of shrink_inactive_list for some
> > probably unrelated process. It seems more straightforward to me to
> > return 0, and add another fatal signal pending check to the caller
> > (shrink_lruvec) to bail out early (dealing with OOM kill avoidance
> > there if necessary) instead of waiting to accumulate fake
> > SWAP_CLUSTER_MAX values from shrink_inactive_list.
>
> The point of this code is to bail out early if the caller has fatal
> signals pending. That could be SIGTERM sent to the process performing
> the reclaim for whatever reason. The bail out is tuned for
> SWAP_CLUSTER_MAX as you can see and your patch is increasing the reclaim
> target which means that bailout wouldn't work properly and you wouldn't
> get any useful work done but not really bail out.

It's increasing to 1/4 of what it was 6 months ago before 88536ac291
("mm:vmscan: fix inaccurate reclaim during proactive reclaim") and
this hasn't changed since then, so if anything the bailout should
happen quicker than originally tuned for.

> > As far as changing the value, SWAP_CLUSTER_MAX puts the final value of
> > sc->nr_reclaimed pretty close to sc->nr_to_reclaim. Since there's a
> > loop for each evictable lru in shrink_lruvec, we could end up with 4 *
> > sc->nr_to_reclaim in sc->nr_reclaimed if we switched to
> > sc->nr_to_reclaim from SWAP_CLUSTER_MAX... an even bigger lie. So I
> > don't think we'd want to do that.
>
> The actual number returned from the reclaim is not really important
> because memory_reclaim would break out of the loop and userspace would
> never see the result.

This makes sense, but it makes me uneasy. I can't point to anywhere
this would cause a problem currently (except maybe super unlikely
overflow of nr_reclaimed), but it feels like a setup for future
unintended consequences.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ