lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Fri, 8 Dec 2023 16:24:15 +0800
From:   Kairui Song <ryncsn@...il.com>
To:     Yu Zhao <yuzhao@...gle.com>
Cc:     Andrew Morton <akpm@...ux-foundation.org>, linux-mm@...ck.org,
        linux-kernel@...r.kernel.org,
        Charan Teja Kalla <quic_charante@...cinc.com>,
        Kalesh Singh <kaleshsingh@...gle.com>, stable@...r.kernel.org
Subject: Re: [PATCH mm-unstable v1 1/4] mm/mglru: fix underprotected page cache

Yu Zhao <yuzhao@...gle.com> 于2023年12月8日周五 14:14写道:
>
> Unmapped folios accessed through file descriptors can be
> underprotected. Those folios are added to the oldest generation based
> on:
> 1. The fact that they are less costly to reclaim (no need to walk the
>    rmap and flush the TLB) and have less impact on performance (don't
>    cause major PFs and can be non-blocking if needed again).
> 2. The observation that they are likely to be single-use. E.g., for
>    client use cases like Android, its apps parse configuration files
>    and store the data in heap (anon); for server use cases like MySQL,
>    it reads from InnoDB files and holds the cached data for tables in
>    buffer pools (anon).
>
> However, the oldest generation can be very short lived, and if so, it
> doesn't provide the PID controller with enough time to respond to a
> surge of refaults. (Note that the PID controller uses weighted
> refaults and those from evicted generations only take a half of the
> whole weight.) In other words, for a short lived generation, the
> moving average smooths out the spike quickly.
>
> To fix the problem:
> 1. For folios that are already on LRU, if they can be beyond the
>    tracking range of tiers, i.e., five accesses through file
>    descriptors, move them to the second oldest generation to give them
>    more time to age. (Note that tiers are used by the PID controller
>    to statistically determine whether folios accessed multiple times
>    through file descriptors are worth protecting.)
> 2. When adding unmapped folios to LRU, adjust the placement of them so
>    that they are not too close to the tail. The effect of this is
>    similar to the above.
>
> On Android, launching 55 apps sequentially:
>                            Before     After      Change
>   workingset_refault_anon  25641024   25598972   0%
>   workingset_refault_file  115016834  106178438  -8%

Hi Yu,

Thanks you for your amazing works on MGLRU.

I believe this is the similar issue I was trying to resolve previously:
https://lwn.net/Articles/945266/
The idea is to use refault distance to decide if the page should be
place in oldest generation or some other gen, which per my test,
worked very well, and we have been using refault distance for MGLRU in
multiple workloads.

There are a few issues left in my previous RFC series, like anon pages
in MGLRU shouldn't be considered, I wanted to collect feedback or test
cases, but unfortunately it seems didn't get too much attention
upstream.

I think both this patch and my previous series are for solving the
file pages underpertected issue, and I did a quick test using this
series, for mongodb test, refault distance seems still a better
solution (I'm not saying these two optimization are mutually exclusive
though, just they do have some conflicts in implementation and solving
similar problem):

Previous result:
==================================================================
Execution Results after 905 seconds
------------------------------------------------------------------
                  Executed        Time (µs)       Rate
  STOCK_LEVEL     2542            27121571486.2   0.09 txn/s
------------------------------------------------------------------
  TOTAL           2542            27121571486.2   0.09 txn/s

This patch:
==================================================================
Execution Results after 900 seconds
------------------------------------------------------------------
                  Executed        Time (µs)       Rate
  STOCK_LEVEL     1594            27061522574.4   0.06 txn/s
------------------------------------------------------------------
  TOTAL           1594            27061522574.4   0.06 txn/s

Unpatched version is always around ~500.

I think there are a few points here:
- Refault distance make use of page shadow so it can better
distinguish evicted pages of different access pattern (re-access
distance).
- Throttled refault distance can help hold part of workingset when
memory is too small to hold the whole workingset.

So maybe part of this patch and the bits of previous series can be
combined to work better on this issue, how do you think?

>
> Fixes: ac35a4902374 ("mm: multi-gen LRU: minimal implementation")
> Signed-off-by: Yu Zhao <yuzhao@...gle.com>
> Reported-by: Charan Teja Kalla <quic_charante@...cinc.com>
> Tested-by: Kalesh Singh <kaleshsingh@...gle.com>
> Cc: stable@...r.kernel.org
> ---
>  include/linux/mm_inline.h | 23 ++++++++++++++---------
>  mm/vmscan.c               |  2 +-
>  mm/workingset.c           |  6 +++---
>  3 files changed, 18 insertions(+), 13 deletions(-)
>
> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
> index 9ae7def16cb2..f4fe593c1400 100644
> --- a/include/linux/mm_inline.h
> +++ b/include/linux/mm_inline.h
> @@ -232,22 +232,27 @@ static inline bool lru_gen_add_folio(struct lruvec *lruvec, struct folio *folio,
>         if (folio_test_unevictable(folio) || !lrugen->enabled)
>                 return false;
>         /*
> -        * There are three common cases for this page:
> -        * 1. If it's hot, e.g., freshly faulted in or previously hot and
> -        *    migrated, add it to the youngest generation.
> -        * 2. If it's cold but can't be evicted immediately, i.e., an anon page
> -        *    not in swapcache or a dirty page pending writeback, add it to the
> -        *    second oldest generation.
> -        * 3. Everything else (clean, cold) is added to the oldest generation.
> +        * There are four common cases for this page:
> +        * 1. If it's hot, i.e., freshly faulted in, add it to the youngest
> +        *    generation, and it's protected over the rest below.
> +        * 2. If it can't be evicted immediately, i.e., a dirty page pending
> +        *    writeback, add it to the second youngest generation.
> +        * 3. If it should be evicted first, e.g., cold and clean from
> +        *    folio_rotate_reclaimable(), add it to the oldest generation.
> +        * 4. Everything else falls between 2 & 3 above and is added to the
> +        *    second oldest generation if it's considered inactive, or the
> +        *    oldest generation otherwise. See lru_gen_is_active().
>          */
>         if (folio_test_active(folio))
>                 seq = lrugen->max_seq;
>         else if ((type == LRU_GEN_ANON && !folio_test_swapcache(folio)) ||
>                  (folio_test_reclaim(folio) &&
>                   (folio_test_dirty(folio) || folio_test_writeback(folio))))
> -               seq = lrugen->min_seq[type] + 1;
> -       else
> +               seq = lrugen->max_seq - 1;
> +       else if (reclaiming || lrugen->min_seq[type] + MIN_NR_GENS >= lrugen->max_seq)
>                 seq = lrugen->min_seq[type];
> +       else
> +               seq = lrugen->min_seq[type] + 1;

For example. maybe still keep the pages on oldest gen by default, but
if the page have a eligible shadow, then put it on min_seq + 1?

>
>         gen = lru_gen_from_seq(seq);
>         flags = (gen + 1UL) << LRU_GEN_PGOFF;
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 4e3b835c6b4a..e67631c60ac0 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -4260,7 +4260,7 @@ static bool sort_folio(struct lruvec *lruvec, struct folio *folio, struct scan_c
>         }
>
>         /* protected */
> -       if (tier > tier_idx) {
> +       if (tier > tier_idx || refs == BIT(LRU_REFS_WIDTH)) {
>                 int hist = lru_hist_from_seq(lrugen->min_seq[type]);
>
>                 gen = folio_inc_gen(lruvec, folio, false);
> diff --git a/mm/workingset.c b/mm/workingset.c
> index 7d3dacab8451..2a2a34234df9 100644
> --- a/mm/workingset.c
> +++ b/mm/workingset.c
> @@ -313,10 +313,10 @@ static void lru_gen_refault(struct folio *folio, void *shadow)
>          * 1. For pages accessed through page tables, hotter pages pushed out
>          *    hot pages which refaulted immediately.
>          * 2. For pages accessed multiple times through file descriptors,
> -        *    numbers of accesses might have been out of the range.
> +        *    they would have been protected by sort_folio().
>          */
> -       if (lru_gen_in_fault() || refs == BIT(LRU_REFS_WIDTH)) {
> -               folio_set_workingset(folio);
> +       if (lru_gen_in_fault() || refs >= BIT(LRU_REFS_WIDTH) - 1) {
> +               set_mask_bits(&folio->flags, 0, LRU_REFS_MASK | BIT(PG_workingset));
>                 mod_lruvec_state(lruvec, WORKINGSET_RESTORE_BASE + type, delta);
>         }

Also this can combine with refault distance check for setting the
reference flag.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ