lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 1 Jul 2009 10:26:44 +0800
From:	Wu Fengguang <fengguang.wu@...il.com>
To:	KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>
Cc:	David Woodhouse <dwmw2@...radead.org>,
	David Howells <dhowells@...hat.com>,
	Minchan Kim <minchan.kim@...il.com>,
	Mel Gorman <mel@....ul.ie>,
	Johannes Weiner <hannes@...xchg.org>,
	"riel@...hat.com" <riel@...hat.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>,
	Christoph Lameter <cl@...ux-foundation.org>,
	"peterz@...radead.org" <peterz@...radead.org>,
	"tytso@....edu" <tytso@....edu>,
	"linux-mm@...ck.org" <linux-mm@...ck.org>,
	"elladan@...imo.com" <elladan@...imo.com>,
	"npiggin@...e.de" <npiggin@...e.de>,
	"Barnes, Jesse" <jesse.barnes@...el.com>
Subject: Re: Found the commit that causes the OOMs

On Wed, Jul 01, 2009 at 10:16:45AM +0800, Wu Fengguang wrote:
> On Wed, Jul 01, 2009 at 10:18:03AM +0900, KOSAKI Motohiro wrote:
> > > On Mon, Jun 29, 2009 at 11:56:47PM +0800, David Woodhouse wrote:
> > > > On Mon, 2009-06-29 at 16:54 +0100, David Howells wrote:
> > > > > Wu Fengguang <fengguang.wu@...el.com> wrote:
> > > > >
> > > > > > Yes this time the OOM order/flags are much different from all previous OOMs.
> > > > > >
> > > > > > btw, I found that msgctl11 is pretty good at making a lot of SUnreclaim and
> > > > > > PageTables pages:
> > > > >
> > > > > I got David Woodhouse to run this on one of this boxes, but he doesn't see the
> > > > > problem, I think because he's got 4GB of RAM, and never comes close to running
> > > > > out.
> > > > >
> > > > > I've asked him to reboot with mem=1G to see if that helps reproduce it.
> > > >
> > > > msgctl11 invoked oom-killer: gfp_mask=0xd0, order=1, oom_adj=0
> > > > Pid: 5795, comm: msgctl11 Not tainted 2.6.31-rc1 #147
> > > > Call Trace:
> > > >  [<ffffffff81092c77>] oom_kill_process.clone.0+0xac/0x254
> > > >  [<ffffffff81092b5c>] ? badness+0x24d/0x2bc
> > > >  [<ffffffff81092f5f>] __out_of_memory+0x140/0x157
> > > >  [<ffffffff8109308f>] out_of_memory+0x119/0x150
> > > >  [<ffffffff81095c65>] ? drain_local_pages+0x16/0x18
> > > >  [<ffffffff810967ab>] __alloc_pages_nodemask+0x45a/0x55b
> > > >  [<ffffffff810a32b0>] ? __inc_zone_page_state+0x2e/0x30
> > > >  [<ffffffff810bb6b9>] alloc_pages_current+0xae/0xb6
> > > >  [<ffffffff810a604a>] ? do_wp_page+0x621/0x6c3
> > > >  [<ffffffff81094d7e>] __get_free_pages+0xe/0x4b
> > > >  [<ffffffff810403a7>] copy_process+0xab/0x11a5
> > > >  [<ffffffff810327c8>] ? check_preempt_wakeup+0x11a/0x142
> > > >  [<ffffffff810a7a06>] ? handle_mm_fault+0x678/0x6e9
> > > >  [<ffffffff810415ec>] do_fork+0x14b/0x338
> > > >  [<ffffffff8105b50a>] ? up_read+0xe/0x10
> > > >  [<ffffffff814ee655>] ? do_page_fault+0x2da/0x307
> > > >  [<ffffffff8100a55c>] sys_clone+0x28/0x2a
> > > >  [<ffffffff8100bfc3>] stub_clone+0x13/0x20
> > > >  [<ffffffff8100bcdb>] ? system_call_fastpath+0x16/0x1b
> > > > Mem-Info:
> > > > Node 0 DMA per-cpu:
> > > > CPU    0: hi:    0, btch:   1 usd:   0
> > > > CPU    1: hi:    0, btch:   1 usd:   0
> > > > CPU    2: hi:    0, btch:   1 usd:   0
> > > > CPU    3: hi:    0, btch:   1 usd:   0
> > > > CPU    4: hi:    0, btch:   1 usd:   0
> > > > CPU    5: hi:    0, btch:   1 usd:   0
> > > > CPU    6: hi:    0, btch:   1 usd:   0
> > > > CPU    7: hi:    0, btch:   1 usd:   0
> > > > Node 0 DMA32 per-cpu:
> > > > CPU    0: hi:  186, btch:  31 usd:   0
> > > > CPU    1: hi:  186, btch:  31 usd:  20
> > > > CPU    2: hi:  186, btch:  31 usd:  19
> > > > CPU    3: hi:  186, btch:  31 usd:  20
> > > > CPU    4: hi:  186, btch:  31 usd:  19
> > > > CPU    5: hi:  186, btch:  31 usd:  24
> > > > CPU    6: hi:  186, btch:  31 usd:  41
> > > > CPU    7: hi:  186, btch:  31 usd:  25
> > > > Active_anon:72835 active_file:89 inactive_anon:575
> > > >  inactive_file:103 unevictable:0 dirty:36 writeback:0 unstable:0
> > > >  free:2467 slab:38211 mapped:229 pagetables:66918 bounce:0
> > > > Node 0 DMA free:4036kB min:60kB low:72kB high:88kB active_anon:3228kB inactive_a
> > > > non:256kB active_file:0kB inactive_file:0kB unevictable:0kB present:15356kB page
> > > > s_scanned:0 all_unreclaimable? no
> > > > lowmem_reserve[]: 0 994 994 994
> > > > Node 0 DMA32 free:5832kB min:4000kB low:5000kB high:6000kB active_anon:288112kB
> > > > inactive_anon:2044kB active_file:356kB inactive_file:412kB unevictable:0kB prese
> > > > nt:1018080kB pages_scanned:0 all_unreclaimable? no
> > > > lowmem_reserve[]: 0 0 0 0
> > > > Node 0 DMA: 1*4kB 2*8kB 1*16kB 0*32kB 1*64kB 2*128kB 0*256kB 1*512kB 1*1024kB 1*
> > > > 2048kB 0*4096kB = 3940kB
> > > > Node 0 DMA32: 852*4kB 1*8kB 0*16kB 1*32kB 1*64kB 0*128kB 1*256kB 1*512kB 1*1024k
> > > > B 0*2048kB 0*4096kB = 5304kB
> > > > 437 total pagecache pages
> > > > 0 pages in swap cache
> > > > Swap cache stats: add 0, delete 0, find 0/0
> > > > Free swap  = 0kB
> > > > Total swap = 0kB
> > > > 262144 pages RAM
> > > > 6503 pages reserved
> > > > 205864 pages shared
> > > > 226536 pages non-shared
> > > > Out of memory: kill process 3855 (msgctl11) score 179248 or a child
> > > > Killed process 4222 (msgctl11)
> > > 
> > > More data: I boot 2.6.30-rc1 with mem=1G and enabled 1GB swap and run msgctl11.
> > > 
> > > It goes OOM at the 2nd run. They are very interesting numbers: memory leaked?
> > > 
> > >         [ 2259.825958] msgctl11 invoked oom-killer: gfp_mask=0x84d0, order=0, oom_adj=0
> > >         [ 2259.828092] Pid: 29657, comm: msgctl11 Not tainted 2.6.31-rc1 #22
> > >         [ 2259.830505] Call Trace:
> > >         [ 2259.832010]  [<ffffffff8156f366>] ? _spin_unlock+0x26/0x30
> > >         [ 2259.834219]  [<ffffffff810c8b26>] oom_kill_process+0x176/0x270
> > >         [ 2259.837603]  [<ffffffff810c8def>] ? badness+0x18f/0x300
> > >         [ 2259.839906]  [<ffffffff810c9095>] __out_of_memory+0x135/0x170
> > >         [ 2259.842035]  [<ffffffff810c91c5>] out_of_memory+0xf5/0x180
> > >         [ 2259.844270]  [<ffffffff810cd86c>] __alloc_pages_nodemask+0x6ac/0x6c0
> > >         [ 2259.846743]  [<ffffffff810f8fa8>] alloc_pages_current+0x78/0x100
> > >         [ 2259.849083]  [<ffffffff81033515>] pte_alloc_one+0x15/0x50
> > >         [ 2259.851282]  [<ffffffff810e0eda>] __pte_alloc+0x2a/0xf0
> > >         [ 2259.853454]  [<ffffffff810e16e2>] handle_mm_fault+0x742/0x830
> > >         [ 2259.855793]  [<ffffffff815725cb>] do_page_fault+0x1cb/0x330
> > >         [ 2259.858033]  [<ffffffff8156fdf5>] page_fault+0x25/0x30
> > >         [ 2259.860301] Mem-Info:
> > >         [ 2259.861706] Node 0 DMA per-cpu:
> > >         [ 2259.862523] CPU    0: hi:    0, btch:   1 usd:   0
> > >         [ 2259.864454] CPU    1: hi:    0, btch:   1 usd:   0
> > >         [ 2259.866608] Node 0 DMA32 per-cpu:
> > >         [ 2259.867404] CPU    0: hi:  186, btch:  31 usd: 197
> > >         [ 2259.869283] CPU    1: hi:  186, btch:  31 usd: 175
> > >         [ 2259.870511] Active_anon:0 active_file:11 inactive_anon:0
> > > 
> > > zero anon pages!
> > >
> > >        [ 2259.870512]  inactive_file:0 unevictable:0 dirty:0 writeback:0 unstable:0
> > >        [ 2259.870513]  free:1986 slab:42170 mapped:96 pagetables:59427 bounce:0
> > 
> > I bet this is NOT zero. it only hidden. 
> 
> Yes, very likely! I noticed that it's all about direct scans:
> 
> pgscan_kswapd_dma 0
> pgscan_kswapd_dma32 0
> pgscan_kswapd_normal 0
> pgscan_kswapd_movable 0
> pgscan_direct_dma 0
> pgscan_direct_dma32 7295
> pgscan_direct_normal 143810
> pgscan_direct_movable 0
> zone_reclaim_failed 0
> 
> > I guess this system's memory usage is,
> >    pagetables: 60k pages
> >    kernel stack: 60k pages
> >    anon (hidden): 60k pages
> >    slab: 40k pages
> >    other: 30k pages
> >    ===================
> >    total: 250k pages = 1GB
> > 
> > What is "hidden" anon pages?
> > each shrink_{in}active_list isolate 32 pages from lru. it mean anon or file lru
> > accounting decrease temporary.
> > 
> > if system have plenty thread or process, heavy memory pressure makes 
> > #-of-thread x 32pages isolation.
> > 
> > msgctl11 makes >10K processes.
> 
> More exactly, ~16K processes:
> 
>         msgctl11    0  INFO  :  Using upto 16298 pids
> 
> So the maximum number of isolated pages is 16K * 32 = 512K, or 2GiB.
> 
> > I have debugging patch for this case.
> > Wu, Can you please try this patch?
> 
> OK. But the OOM is not quite reproducible. Sometimes it produces these
> messages:

This time I got the OOM: there are 69817 isolated pages (just as expected)!

[ 1521.979074] msgctl11 invoked oom-killer: gfp_mask=0x200da, order=0, oom_adj=0
[ 1521.980996] Pid: 16405, comm: msgctl11 Not tainted 2.6.31-rc1 #27
[ 1521.983271] Call Trace:
[ 1521.983936]  [<ffffffff8158dc1b>] ? _spin_unlock+0x2b/0x40
[ 1521.985195]  [<ffffffff810d3526>] oom_kill_process+0x176/0x270
[ 1521.987384]  [<ffffffff810d37f7>] ? badness+0x197/0x310
[ 1521.989019]  [<ffffffff810d3ab5>] __out_of_memory+0x145/0x180
[ 1521.990981]  [<ffffffff810d3bed>] out_of_memory+0xfd/0x190
[ 1521.993199]  [<ffffffff810d83bc>] __alloc_pages_nodemask+0x6bc/0x6d0
[ 1521.995770]  [<ffffffff81012e69>] ? sched_clock+0x9/0x10
[ 1521.997880]  [<ffffffff8110485e>] alloc_page_vma+0x8e/0x1c0
[ 1522.000091]  [<ffffffff810ea5aa>] do_wp_page+0x23a/0x840
[ 1522.002246]  [<ffffffff810ec7b6>] handle_mm_fault+0x656/0x840
[ 1522.003476]  [<ffffffff81590ecb>] do_page_fault+0x1cb/0x330
[ 1522.004995]  [<ffffffff8158e6e5>] page_fault+0x25/0x30
[ 1522.007006] Mem-Info:
[ 1522.007535] Node 0 DMA per-cpu:
[ 1522.009342] CPU    0: hi:    0, btch:   1 usd:   0
[ 1522.011277] CPU    1: hi:    0, btch:   1 usd:   0
[ 1522.013401] Node 0 DMA32 per-cpu:
[ 1522.015291] CPU    0: hi:  186, btch:  31 usd: 176
[ 1522.017232] CPU    1: hi:  186, btch:  31 usd: 155
[ 1522.019259] Active_anon:11 active_file:6 inactive_anon:0
[ 1522.019260]  inactive_file:0 unevictable:0 dirty:0 writeback:0 unstable:0
[ 1522.019261]  free:1985 slab:44399 mapped:132 pagetables:61830 bounce:0
[ 1522.019262]  isolate:69817
[ 1522.025145] Node 0 DMA free:3964kB min:56kB low:68kB high:84kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB present:15164kB pages_scanned:655 all_unreclaimable? yes
[ 1522.030180] lowmem_reserve[]: 0 982 982 982
[ 1522.031506] Node 0 DMA32 free:3976kB min:3980kB low:4972kB high:5968kB active_anon:44kB inactive_anon:0kB active_file:24kB inactive_file:0kB unevictable:0kB present:1005984kB pages_scanned:249 all_unreclaimable? no
[ 1522.037463] lowmem_reserve[]: 0 0 0 0
[ 1522.039637] Node 0 DMA: 3*4kB 0*8kB 1*16kB 1*32kB 1*64kB 0*128kB 1*256kB 1*512kB 1*1024kB 1*2048kB 0*4096kB = 3964kB
[ 1522.043998] Node 0 DMA32: 102*4kB 6*8kB 0*16kB 0*32kB 1*64kB 1*128kB 1*256kB 2*512kB 2*1024kB 0*2048kB 0*4096kB = 3976kB
[ 1522.049241] 1312 total pagecache pages
[ 1522.050996] 1112 pages in swap cache
[ 1522.051759] Swap cache stats: add 218714, delete 217602, find 97535/130636
[ 1522.055428] Free swap  = 1037356kB
[ 1522.057113] Total swap = 1048568kB

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ