[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CA+G9fYsdsgRmwLtSKJSzB1eWcUQ1z-_aaU+BNcQpker34XT6_w@mail.gmail.com>
Date: Wed, 17 Jun 2020 19:07:20 +0530
From: Naresh Kamboju <naresh.kamboju@...aro.org>
To: Michal Hocko <mhocko@...nel.org>,
Chris Down <chris@...isdown.name>,
Yafang Shao <laoar.shao@...il.com>
Cc: Anders Roxell <anders.roxell@...aro.org>,
"Linux F2FS DEV, Mailing List"
<linux-f2fs-devel@...ts.sourceforge.net>,
linux-ext4 <linux-ext4@...r.kernel.org>,
linux-block <linux-block@...r.kernel.org>,
Andrew Morton <akpm@...ux-foundation.org>,
open list <linux-kernel@...r.kernel.org>,
Linux-Next Mailing List <linux-next@...r.kernel.org>,
linux-mm <linux-mm@...ck.org>, Arnd Bergmann <arnd@...db.de>,
Andreas Dilger <adilger.kernel@...ger.ca>,
Jaegeuk Kim <jaegeuk@...nel.org>,
"Theodore Ts'o" <tytso@....edu>, Chao Yu <chao@...nel.org>,
Hugh Dickins <hughd@...gle.com>,
Andrea Arcangeli <aarcange@...hat.com>,
Matthew Wilcox <willy@...radead.org>,
Chao Yu <yuchao0@...wei.com>, lkft-triage@...ts.linaro.org,
Johannes Weiner <hannes@...xchg.org>,
Roman Gushchin <guro@...com>, Cgroups <cgroups@...r.kernel.org>
Subject: Re: mm: mkfs.ext4 invoked oom-killer on i386 - pagecache_get_page
On Thu, 21 May 2020 at 22:04, Michal Hocko <mhocko@...nel.org> wrote:
>
> On Thu 21-05-20 11:55:16, Michal Hocko wrote:
> > On Wed 20-05-20 20:09:06, Chris Down wrote:
> > > Hi Naresh,
> > >
> > > Naresh Kamboju writes:
> > > > As a part of investigation on this issue LKFT teammate Anders Roxell
> > > > git bisected the problem and found bad commit(s) which caused this problem.
> > > >
> > > > The following two patches have been reverted on next-20200519 and retested the
> > > > reproducible steps and confirmed the test case mkfs -t ext4 got PASS.
> > > > ( invoked oom-killer is gone now)
> > > >
> > > > Revert "mm, memcg: avoid stale protection values when cgroup is above
> > > > protection"
> > > > This reverts commit 23a53e1c02006120f89383270d46cbd040a70bc6.
> > > >
> > > > Revert "mm, memcg: decouple e{low,min} state mutations from protection
> > > > checks"
> > > > This reverts commit 7b88906ab7399b58bb088c28befe50bcce076d82.
> > >
> > > Thanks Anders and Naresh for tracking this down and reverting.
> > >
> > > I'll take a look tomorrow. I don't see anything immediately obviously wrong
> > > in either of those commits from a (very) cursory glance, but they should
> > > only be taking effect if protections are set.
> >
> > Agreed. If memory.{low,min} is not used then the patch should be
> > effectively a nop.
>
> I was staring into the code and do not see anything. Could you give the
> following debugging patch a try and see whether it triggers?
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index cc555903a332..df2e8df0eb71 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2404,6 +2404,8 @@ static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc,
> * sc->priority further than desirable.
> */
> scan = max(scan, SWAP_CLUSTER_MAX);
> +
> + trace_printk("scan:%lu protection:%lu\n", scan, protection);
> } else {
> scan = lruvec_size;
> }
> @@ -2648,6 +2650,7 @@ static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc)
> mem_cgroup_calculate_protection(target_memcg, memcg);
>
> if (mem_cgroup_below_min(memcg)) {
> + trace_printk("under min:%lu emin:%lu\n", memcg->memory.min, memcg->memory.emin);
> /*
> * Hard protection.
> * If there is no reclaimable memory, OOM.
> @@ -2660,6 +2663,7 @@ static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc)
> * there is an unprotected supply
> * of reclaimable memory from other cgroups.
> */
> + trace_printk("under low:%lu elow:%lu\n", memcg->memory.low, memcg->memory.elow);
> if (!sc->memcg_low_reclaim) {
> sc->memcg_low_skipped = 1;
> continue;
As per your suggestions on debugging this problem,
trace_printk is replaced with printk and applied to your patch on top of the
problematic kernel and here is the test output and link.
mkfs -t ext4 /dev/disk/by-id/ata-TOSHIBA_MG04ACA100N_Y8RQK14KF6XF
mke2fs 1.43.8 (1-Jan-2018)
Creating filesystem with 244190646 4k blocks and 61054976 inodes
Filesystem UUID: 7c380766-0ed8-41ba-a0de-3c08e78f1891
Superblock backups stored on blocks:
32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208,
4096000, 7962624, 11239424, 20480000, 23887872, 71663616, 78675968,
102400000, 214990848
Allocating group tables: 0/7453 done
Writing inode tables: 0/7453 done
Creating journal (262144 blocks): [ 51.544525] under min:0 emin:0
[ 51.845304] under min:0 emin:0
[ 51.848738] under min:0 emin:0
[ 51.858147] under min:0 emin:0
[ 51.861333] under min:0 emin:0
[ 51.862034] under min:0 emin:0
[ 51.862442] under min:0 emin:0
[ 51.862763] under min:0 emin:0
Full test log link,
https://lkft.validation.linaro.org/scheduler/job/1497412#L1451
- Naresh
Powered by blists - more mailing lists