lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Tue, 1 Dec 2015 14:09:39 +0800
From:	"Chenjie (K)" <chenjie6@...wei.com>
To:	David Rientjes <rientjes@...gle.com>
CC:	Andrew Morton <akpm@...ux-foundation.org>, <linux-mm@...ck.org>,
	<linux-kernel@...r.kernel.org>, <David.Woodhouse@...el.com>,
	<zhihui.gao@...wei.com>, <lizefan@...wei.com>,
	<stable@...r.kernel.org>
Subject: Re: [PATCH] bugfix oom kill init lead panic


Please ignore the last email.

This is log , kill pid 9134:

Mem-info:
Normal per-cpu:
CPU    0: hi:   90, btch:  15 usd:   0
CPU    1: hi:   90, btch:  15 usd:   0
active_anon:997 inactive_anon:36468 isolated_anon:0
  active_file:1 inactive_file:0 isolated_file:0
  unevictable:10743 dirty:0 writeback:0 unstable:0
  free:451 slab_reclaimable:1345 slab_unreclaimable:2902
  mapped:588 shmem:37100 pagetables:91 bounce:0
  free_cma:0
Normal free:1804kB min:1812kB low:2264kB high:2716kB active_anon:3988kB 
inactive_anon:145872kB active_file:4kB inactive_file:0kB 
unevictable:42972kB isolated(anon):0kB isolated(file):0kB 
present:307200kB managed:205416kB mlocked:0kB dirty:0kB writeback:0kB 
mapped:2352kB shmem:148400kB slab_reclaimable:5380kB 
slab_unreclaimable:11608kB kernel_stack:520kB pagetables:364kB 
unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:22 
all_unreclaimable? yes
lowmem_reserve[]: 0 0 0
Normal: 139*4kB (UEMR) 67*8kB (UEMR) 5*16kB (UMR) 1*32kB (R) 2*64kB (R) 
4*128kB (R) 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 1844kB
47844 total pagecache pages
76800 pages of RAM
594 free pages
22676 reserved pages
2796 slab pages
529874 pages shared
0 pages swap cached
[ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name
[ 1021]     0  1021      493      139       5        0             0 RATC
[ 1023]     0  1023      493      123       5        0         -1000 RATC
[ 1112]     0  1112     1302      511       4        0         -1000 sshd
[ 1505]     0  1505      490      130       5        0             0 
take_cpu_rate
[ 1506]     0  1506      490       32       5        0             0 
take_cpu_rate
[ 1508]     0  1508      490       32       5        0             0 
take_cpu_rate
[ 8553]     0  8553      462       45       4        0             0 
lockf.test
[ 8554]     0  8554      462       86       4        0             0 
lockf.test
[ 8555]     0  8555      462       86       4        0             0 
lockf.test
[ 8557]     0  8557      462       86       4        0             0 
lockf.test
[10732]     0 10732      462       16       4        0             0 
lockf.test
[10068]     0 10068      730       92       5        0             0 
runtest.sh
[10072]     0 10072      697       99       4        0             0 sleep
[ 8606]     0  8606      730       92       5        0             0 
runtest.sh
[ 8610]     0  8610      697       99       4        0             0 sleep
[ 8611]     0  8611      732       99       5        0             0 tr
[ 8612]     0  8612      730       99       5        0             0 dd
[ 9132]     0  9132      947      129       5        0             0 
syslog-ng
[ 9133]     0  9133      384       32       3        0             0 getty
[ 9134]     0  9134      467      164       4        0             0 init
*****************Start oom extend info.*****************
Vmallocinfo Start >>>>>>>>>>>>>>>>>>>>
0xbf000000-0xbf006000   24576 module_alloc_update_bounds+0xc/0x5c 
pages=5 vmalloc
0xbf00a000-0xbf00e000   16384 module_alloc_update_bounds+0xc/0x5c 
pages=3 vmalloc
0xbf011000-0xbf013000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf015000-0xbf017000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf019000-0xbf01b000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf01d000-0xbf021000   16384 module_alloc_update_bounds+0xc/0x5c 
pages=3 vmalloc
0xbf024000-0xbf032000   57344 module_alloc_update_bounds+0xc/0x5c 
pages=13 vmalloc
0xbf039000-0xbf03c000   12288 module_alloc_update_bounds+0xc/0x5c 
pages=2 vmalloc
0xbf03f000-0xbf042000   12288 module_alloc_update_bounds+0xc/0x5c 
pages=2 vmalloc
0xbf044000-0xbf046000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf048000-0xbf04a000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf04c000-0xbf04e000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf050000-0xbf054000   16384 module_alloc_update_bounds+0xc/0x5c 
pages=3 vmalloc
0xbf056000-0xbf059000   12288 module_alloc_update_bounds+0xc/0x5c 
pages=2 vmalloc
0xbf05b000-0xbf069000   57344 module_alloc_update_bounds+0xc/0x5c 
pages=13 vmalloc
0xbf06d000-0xbf08e000  135168 module_alloc_update_bounds+0xc/0x5c 
pages=32 vmalloc
0xbf096000-0xbf0ce000  229376 module_alloc_update_bounds+0xc/0x5c 
pages=55 vmalloc
0xbf0e8000-0xbf0ea000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xbf0ec000-0xbf0fe000   73728 module_alloc_update_bounds+0xc/0x5c 
pages=17 vmalloc
0xbf105000-0xbf12a000  151552 module_alloc_update_bounds+0xc/0x5c 
pages=36 vmalloc
0xbf13d000-0xbf145000   32768 module_alloc_update_bounds+0xc/0x5c 
pages=7 vmalloc
0xbf149000-0xbf168000  126976 module_alloc_update_bounds+0xc/0x5c 
pages=30 vmalloc
0xbf173000-0xbf179000   24576 module_alloc_update_bounds+0xc/0x5c 
pages=5 vmalloc
0xbf17c000-0xbf180000   16384 module_alloc_update_bounds+0xc/0x5c 
pages=3 vmalloc
0xbf182000-0xbf186000   16384 module_alloc_update_bounds+0xc/0x5c 
pages=3 vmalloc
0xbf192000-0xbf1a5000   77824 module_alloc_update_bounds+0xc/0x5c 
pages=18 vmalloc
0xbf1ab000-0xbf1ad000    8192 module_alloc_update_bounds+0xc/0x5c 
pages=1 vmalloc
0xd3000000-0xd3021000  135168 ekbox_reinit+0x3c/0xcc phys=9fa00000 ioremap
0xd3022000-0xd3024000    8192 of_iomap+0x30/0x3c phys=1a001000 ioremap
0xd3024000-0xd3027000   12288 of_iomap+0x30/0x3c phys=1a000000 ioremap
0xd3028000-0xd302a000    8192 of_iomap+0x30/0x3c phys=20000000 ioremap
0xd302a000-0xd302c000    8192 of_iomap+0x30/0x3c phys=20011000 ioremap
0xd302c000-0xd302f000   12288 of_iomap+0x30/0x3c phys=1a000000 ioremap
0xd3030000-0xd3032000    8192 of_iomap+0x30/0x3c phys=20013000 ioremap
0xd3032000-0xd3034000    8192 bsp_init_led+0x320/0x5e8 phys=f000b000 ioremap
0xd3034000-0xd3036000    8192 bsp_init_led+0x394/0x5e8 phys=f000b000 ioremap
0xd303a000-0xd307b000  266240 atomic_pool_init+0x0/0x11c phys=8d500000 user
0xd3080000-0xd3082000    8192 l2cache_init+0xb8/0x3b8 phys=16800000 ioremap
0xd3916000-0xd3959000  274432 0xbf08e26c pages=66 vmalloc
0xd3959000-0xd3965000   49152 0xbf08e280 pages=11 vmalloc
0xd3965000-0xd396a000   20480 0xbf08e2e0 pages=4 vmalloc
0xd396a000-0xd396d000   12288 0xbf08e2ec pages=2 vmalloc
0xd3a3f000-0xd3a42000   12288 pcpu_extend_area_map+0x18/0xa0 pages=2 vmalloc
0xd3ad8000-0xd3ada000    8192 0xbf1a54ac phys=9fffe000 ioremap
0xd3adc000-0xd3ade000    8192 0xbf1a517c phys=30000000 ioremap
0xd3ade000-0xd3ae0000    8192 0xbf1a51b0 phys=20000000 ioremap
0xd3af2000-0xd3af4000    8192 0xbf1a517c phys=30001000 ioremap
0xd3af4000-0xd3af6000    8192 0xbf1a51b0 phys=20000000 ioremap
0xd3af6000-0xd3af8000    8192 0xbf1a517c phys=30002000 ioremap
0xd3af8000-0xd3afa000    8192 0xbf1a51b0 phys=20000000 ioremap
0xd3afa000-0xd3afc000    8192 0xbf1a517c phys=30003000 ioremap
0xd3afc000-0xd3afe000    8192 0xbf1a51b0 phys=20000000 ioremap
0xd3c00000-0xd3e01000 2101248 kbox_proc_mem_write+0x104/0x1cc 
phys=9f800000 ioremap
0xd4000000-0xd8001000 67112960 devm_ioremap+0x38/0x70 phys=40000000 ioremap
0xfe001000-0xfe002000    4096 iotable_init+0x0/0xb4 phys=20001000 ioremap
0xfe200000-0xfe201000    4096 iotable_init+0x0/0xb4 phys=1a000000 ioremap
0xfee00000-0xff000000 2097152 pci_reserve_io+0x0/0x30 ioremap
Vmallocinfo End <<<<<<<<<<<<<<<<<<<<

[SLUB]Slabinfo Start >>>>>>>>>>>>>>>>>>>>
# name            <active_objs> <num_objs> <objsize> <objperslab> 
<pagesperslab> : slabdata <active_slabs> <num_slabs>
nfs_direct_cache       0      0    120   34    1 : slabdata      0      0
nfs_commit_data       18     18    448   18    2 : slabdata      1      1
nfs_read_data          0      0    576   14    2 : slabdata      0      0
nfs_inode_cache        0      0    816   20    4 : slabdata      0      0
rpc_inode_cache        0      0    512   16    2 : slabdata      0      0
jffs2_refblock       432    432    248   16    1 : slabdata     27     27
jffs2_i              274    546    552   14    2 : slabdata     39     39
bsg_cmd                0      0    288   14    1 : slabdata      0      0
mqueue_inode_cache     23     23    704   23    4 : slabdata      1      1
squashfs_inode_cache      0      0    576   14    2 : slabdata      0      0
ext2_inode_cache       0      0    640   12    2 : slabdata      0      0
pid_namespace        102    102     80   51    1 : slabdata      2      2
user_namespace         0      0    224   18    1 : slabdata      0      0
posix_timers_cache      0      0    160   25    1 : slabdata      0      0
UDP-Lite               0      0    640   12    2 : slabdata      0      0
UDP                   24     24    640   12    2 : slabdata      2      2
tw_sock_TCP           64     64    128   32    1 : slabdata      2      2
TCP                   24     24   1344   12    4 : slabdata      2      2
eventpoll_pwq        204    204     40  102    1 : slabdata      2      2
sgpool-128            12     12   2560   12    8 : slabdata      1      1
sgpool-64             12     12   1280   12    4 : slabdata      1      1
sgpool-16             12     12    320   12    1 : slabdata      1      1
blkdev_queue          50     50   1272   25    8 : slabdata      2      2
blkdev_requests       54     54    216   18    1 : slabdata      3      3
fsnotify_event_holder   7696   8960     16  256    1 : slabdata     35 
    35
fsnotify_event       112    112     72   56    1 : slabdata      2      2
biovec-256            10     10   3072   10    8 : slabdata      1      1
biovec-128             0      0   1536   21    8 : slabdata      0      0
biovec-64              0      0    768   21    4 : slabdata      0      0
sock_inode_cache      64     64    512   16    2 : slabdata      4      4
skbuff_fclone_cache     63     63    384   21    2 : slabdata      3      3
file_lock_cache     1904   1904    120   34    1 : slabdata     56     56
net_namespace          0      0   2240   14    8 : slabdata      0      0
shmem_inode_cache  10195  10890    528   15    2 : slabdata    726    726
proc_inode_cache     162    256    504   16    2 : slabdata     16     16
sigqueue             169    308    144   28    1 : slabdata     11     11
bdev_cache            23     23    704   23    4 : slabdata      1      1
inode_cache         2919   3315    472   17    2 : slabdata    195    195
dentry             13824  20412    144   28    1 : slabdata    729    729
buffer_head            0      0     64   64    1 : slabdata      0      0
vm_area_struct       615   1150     88   46    1 : slabdata     25     25
signal_cache         165    240    640   12    2 : slabdata     20     20
sighand_cache        108    144   1344   12    4 : slabdata     12     12
task_struct          127    182   2304   14    8 : slabdata     13     13
anon_vma_chain      2269   2816     32  128    1 : slabdata     22     22
anon_vma            2409   2409     56   73    1 : slabdata     33     33
debug_objects_cache   3593   8500     24  170    1 : slabdata     50     50
radix_tree_node     1259   1352    304   13    1 : slabdata    104    104
idr_layer_cache      105    105   1080   15    4 : slabdata      7      7
kmalloc-8192          20     20   8192    4    8 : slabdata      5      5
kmalloc-4096          48     48   4096    8    8 : slabdata      6      6
kmalloc-2048          96     96   2048   16    8 : slabdata      6      6
kmalloc-1024         330    400   1024   16    4 : slabdata     25     25
kmalloc-512          902    960    512   16    2 : slabdata     60     60
kmalloc-256          211    240    256   16    1 : slabdata     15     15
kmalloc-192         1221   1323    192   21    1 : slabdata     63     63
kmalloc-128         2318   2624    128   32    1 : slabdata     82     82
kmalloc-64         23503  24192     64   64    1 : slabdata    378    378
kmem_cache_node      128    128     64   64    1 : slabdata      2      2
kmem_cache            96     96    128   32    1 : slabdata      3      3
Slabinfo End <<<<<<<<<<<<<<<<<<<<


Filesystem            1K-blocks    Used   Available Use(%)   Mounted on
tmpfs                   524288        4   524284        0%   /tmp
none                     10240    10240        0      100%   /var
tmpfs                   108248       16   108232        0%   /dev
tmpfs                   108248        0   108248        0%   /dev/shm
tmpfs                   173192   138152    35040       79%   /tmp
*****smap info of all task:*****
   smaps info of task-sshd[1112], rss:2044 kB:
   smaps info of task-init[9134], rss:656 kB:
   smaps info of task-init[1], rss:656 kB:
   smaps info of task-RATC[1021], rss:556 kB:
   smaps info of task-take_cpu_rate[1505], rss:520 kB:
********    mem info     *****	Total:                 216496 kB
	Total free:              1804 kB
	User space:            192836 kB
	Mlock:                      0 kB
	Kernel space:           21856 kB
	Bootmem reserved:       90704 kB
	kernel_image_info:
	    Kernel code:   0x80008000-0x8043369f
	    Kernel data:   0x80496000-0x8056559b
	module info:
	    physmap 3kB         Live 0xbf1ab000 (O)
			refrence count: 1
                                -
	    Drv_Gmac_K 71kB         Live 0xbf192000 (O)
			refrence count: 0
                                -
	    rtos_snapshot 10kB         Live 0xbf182000 (O)
			refrence count: 0
                                -
	    rtos_kbox_panic 10kB         Live 0xbf17c000 (O)
			refrence count: 0
                                -
	    rsm 16kB         Live 0xbf173000 (O)
			refrence count: 0
                                -
	    nfsv4 119kB         Live 0xbf149000
			refrence count: 0
                                -
	    nfsv3 24kB         Live 0xbf13d000
			refrence count: 0
                                -
	    nfs 141kB         Live 0xbf105000
			refrence count: 2
			nfsv4,
			nfsv3,
	    lockd 64kB         Live 0xbf0ec000
			refrence count: 2
			nfsv3,
			nfs,
	    nfs_acl 2kB         Live 0xbf0e8000
			refrence count: 1
			nfsv3,
	    sunrpc 216kB         Live 0xbf096000
			refrence count: 5
			nfsv4,
			nfsv3,
			nfs,
			lockd,
			nfs_acl,
	    jffs2 125kB         Live 0xbf06d000
			refrence count: 1
                                -
	    cfi_cmdset_0002 51kB         Live 0xbf05b000
			refrence count: 1
                                -
	    cfi_probe 5kB         Live 0xbf056000
			refrence count: 0
                                -
	    cfi_util 11kB         Live 0xbf050000
			refrence count: 2
			cfi_cmdset_0002,
			cfi_probe,
	    gen_probe 2kB         Live 0xbf04c000
			refrence count: 1
			cfi_probe,
	    cmdlinepart 2kB         Live 0xbf048000
			refrence count: 0
                                -
	    chipreg 2kB         Live 0xbf044000
			refrence count: 2
			physmap,
			cfi_probe,
	    mtdblock 4kB         Live 0xbf03f000
			refrence count: 0
                                -
	    mtd_blkdevs 7kB         Live 0xbf039000
			refrence count: 1
			mtdblock,
	    mtd 48kB         Live 0xbf024000
			refrence count: 17
			physmap,
			jffs2,
			cfi_cmdset_0002,
			cmdlinepart,
			mtdblock,
			mtd_blkdevs,
	    uio 8kB         Live 0xbf01d000
			refrence count: 0
                                -
	    xt_tcpudp 2kB         Live 0xbf019000
			refrence count: 0
                                -
	    ipt_REJECT 2kB         Live 0xbf015000
			refrence count: 0
                                -
	    iptable_filter 1kB         Live 0xbf011000
			refrence count: 0
                                -
	    ip_tables 11kB         Live 0xbf00a000
			refrence count: 1
			iptable_filter,
	    x_tables 16kB         Live 0xbf000000
			refrence count: 4
			xt_tcpudp,
			ipt_REJECT,
			iptable_filter,
			ip_tables,
******     pagecache_info:     ******
   /rel_mem_filecache_tc11/0 : nrpages = 24051.
   /volatile/log/auth.log : nrpages = 1947.
   /usr/bin/gdb : nrpages = 845.
   /lib/libcrypto.so.1.0.0 : nrpages = 418.
   /usr/lib/libgio-2.0.so.0.3600.4 : nrpages = 391.
   /lib/libc-2.18.so : nrpages = 370.
   /usr/lib/libperl.so.5.14.3 : nrpages = 348.
   /usr/lib/libglib-2.0.so.0.3600.4 : nrpages = 314.
   /usr/lib/libstdc++.so.6.0.17 : nrpages = 277.
   /fs_stress_t/fs_stress_t/testcase/bin/fsback/unit : nrpages = 256.
   /fs_stress_t/fs_stress_t_src/testcase/fsback/unit : nrpages = 256.
   /usr/bin/makedumpfile : nrpages = 238.
   /usr/bin/perf : nrpages = 228.
   /usr/sbin/sshd : nrpages = 226.
   /usr/bin/ssh : nrpages = 202.
   /usr/lib/libbfd-2.23.2.so : nrpages = 195.
   /volatile/log/syslog : nrpages = 192.
   /volatile/log/kern.log : nrpages = 186.
   /lib/libm-2.18.so : nrpages = 173.
   /volatile/log/error : nrpages = 162.
*****************End oom extend info.*****************
Out of memory: Kill process 9134 (init) score 3 or sacrifice child
Killed process 9134 (init) total-vm:1868kB, anon-rss:84kB, file-rss:572kB
Kill process 1 (init) sharing same memory
[RSM][SIG]Kernel:tr(pid:8611|tid:8611) send SIG[9] to init(pid:1|tid:1).
CPU: 0 PID: 8611 Comm: tr Tainted: G           O 3.10.53-HULK2 #1
[<c0018e68>] (unwind_backtrace+0x0/0x11c) from [<c0014548>] 
(show_stack+0x10/0x14)
[<c0014548>] (show_stack+0x10/0x14) from [<bf173640>] 
(send_signal_entry+0xd4/0x144 [rsm])
[<bf173640>] (send_signal_entry+0xd4/0x144 [rsm]) from [<c0037028>] 
(__send_signal+0x2bc/0x310)
[<c0037028>] (__send_signal+0x2bc/0x310) from [<c003710c>] 
(send_signal+0x90/0x94)
[<c003710c>] (send_signal+0x90/0x94) from [<c0037b88>] 
(do_send_sig_info+0x3c/0x64)
[<c0037b88>] (do_send_sig_info+0x3c/0x64) from [<c00c7718>] 
(oom_kill_process+0x34c/0x3d8)
[<c00c7718>] (oom_kill_process+0x34c/0x3d8) from [<c00c7bf4>] 
(out_of_memory+0x26c/0x2b0)
[<c00c7bf4>] (out_of_memory+0x26c/0x2b0) from [<c00cad8c>] 
(__alloc_pages_nodemask+0x558/0x6f8)
[<c00cad8c>] (__alloc_pages_nodemask+0x558/0x6f8) from [<c0106e2c>] 
(pipe_write+0x238/0x494)
[<c0106e2c>] (pipe_write+0x238/0x494) from [<c00ff5ec>] 
(do_sync_write+0x74/0x98)
[<c00ff5ec>] (do_sync_write+0x74/0x98) from [<c00fff80>] 
(vfs_write+0xcc/0x1a8)
[<c00fff80>] (vfs_write+0xcc/0x1a8) from [<c0100374>] (SyS_write+0x38/0x64)
[<c0100374>] (SyS_write+0x38/0x64) from [<c0010960>] 
(ret_fast_syscall+0x0/0x60)
[RSM][SIG]Kernel:tr(pid:8611|tid:8611) send SIG[9] to 
init(pid:9134|tid:9134).
CPU: 0 PID: 8611 Comm: tr Tainted: G           O 3.10.53-HULK2 #1
[<c0018e68>] (unwind_backtrace+0x0/0x11c) from [<c0014548>] 
(show_stack+0x10/0x14)
[<c0014548>] (show_stack+0x10/0x14) from [<bf173640>] 
(send_signal_entry+0xd4/0x144 [rsm])
[<bf173640>] (send_signal_entry+0xd4/0x144 [rsm]) from [<c0037028>] 
(__send_signal+0x2bc/0x310)
[<c0037028>] (__send_signal+0x2bc/0x310) from [<c003710c>] 
(send_signal+0x90/0x94)
[<c003710c>] (send_signal+0x90/0x94) from [<c0037b88>] 
(do_send_sig_info+0x3c/0x64)
[<c0037b88>] (do_send_sig_info+0x3c/0x64) from [<c00c7750>] 
(oom_kill_process+0x384/0x3d8)
[<c00c7750>] (oom_kill_process+0x384/0x3d8) from [<c00c7bf4>] 
(out_of_memory+0x26c/0x2b0)
[<c00c7bf4>] (out_of_memory+0x26c/0x2b0) from [<c00cad8c>] 
(__alloc_pages_nodemask+0x558/0x6f8)
[<c00cad8c>] (__alloc_pages_nodemask+0x558/0x6f8) from [<c0106e2c>] 
(pipe_write+0x238/0x494)
[<c0106e2c>] (pipe_write+0x238/0x494) from [<c00ff5ec>] 
(do_sync_write+0x74/0x98)
[<c00ff5ec>] (do_sync_write+0x74/0x98) from [<c00fff80>] 
(vfs_write+0xcc/0x1a8)
[<c00fff80>] (vfs_write+0xcc/0x1a8) from [<c0100374>] (SyS_write+0x38/0x64)
[<c0100374>] (SyS_write+0x38/0x64) from [<c0010960>] 
(ret_fast_syscall+0x0/0x60)
[RSM][SIG]init(pid:9134|tid:9134) deliver SIG[9].
[RSM][SIG]init(pid:1|tid:1) deliver SIG[9].
Kernel panic - not syncing: Attempted to kill init! exitcode=0x00000009









On 2015/12/1 12:15, Chenjie (K) wrote:
>
> Thank you reply, we run a test case.
>
> A new log:
>
> Out of memory: Kill process 8520 (sshd) score 11 or sacrifice child
> Killed process 8520 (sshd) total-vm:5812kB, anon-rss:404kB, file-rss:2132kB
> [RSM][SIG]Kernel:dd(pid:8612|tid:8612) send SIG[9] to
> sshd(pid:8520|tid:8520).
> CPU: 0 PID: 8612 Comm: dd Tainted: G           O 3.10.53-HULK2 #1
> [<c0018e68>] (unwind_backtrace+0x0/0x11c) from [<c0014548>]
> (show_stack+0x10/0x14)
> [<c0014548>] (show_stack+0x10/0x14) from [<bf173640>]
> (send_signal_entry+0xd4/0x144 [rsm])
> [<bf173640>] (send_signal_entry+0xd4/0x144 [rsm]) from [<c0037028>]
> (__send_signal+0x2bc/0x310)
> [<c0037028>] (__send_signal+0x2bc/0x310) from [<c003710c>]
> (send_signal+0x90/0x94)
> [<c003710c>] (send_signal+0x90/0x94) from [<c0037b88>]
> (do_send_sig_info+0x3c/0x64)
> [<c0037b88>] (do_send_sig_info+0x3c/0x64) from [<c00c7750>]
> (oom_kill_process+0x384/0x3d8)
> [<c00c7750>] (oom_kill_process+0x384/0x3d8) from [<c00c7bf4>]
> (out_of_memory+0x26c/0x2b0)
> [<c00c7bf4>] (out_of_memory+0x26c/0x2b0) from [<c00cad8c>]
> (__alloc_pages_nodemask+0x558/0x6f8)
> [<c00cad8c>] (__alloc_pages_nodemask+0x558/0x6f8) from [<c00d647c>]
> (shmem_getpage_gfp+0x1bc/0x5e0)
> [<c00d647c>] (shmem_getpage_gfp+0x1bc/0x5e0) from [<c00c4a24>]
> (generic_file_buffered_write+0xdc/0x23c)
> [<c00c4a24>] (generic_file_buffered_write+0xdc/0x23c) from [<c00c5b7c>]
> (__generic_file_aio_write+0x33c/0x3a8)
> [<c00c5b7c>] (__generic_file_aio_write+0x33c/0x3a8) from [<c00c5c3c>]
> (generic_file_aio_write+0x54/0xb0)
> [<c00c5c3c>] (generic_file_aio_write+0x54/0xb0) from [<c00ff5ec>]
> (do_sync_write+0x74/0x98)
> [<c00ff5ec>] (do_sync_write+0x74/0x98) from [<c00fff80>]
> (vfs_write+0xcc/0x1a8)
> [<c00fff80>] (vfs_write+0xcc/0x1a8) from [<c0100374>] (SyS_write+0x38/0x64)
> [<c0100374>] (SyS_write+0x38/0x64) from [<c0010960>]
> (ret_fast_syscall+0x0/0x60)
> [RSM][SIG]sshd(pid:8520|tid:8520) deliver SIG[9].
> [RSM][SIG]tr(pid:9088|tid:9088) deliver SIG[9].
> dd invoked oom-killer: gfp_mask=0x200da, order=0, oom_score_adj=0
> dd cpuset=/ mems_allowed=0
> CPU: 0 PID: 8612 Comm: dd Tainted: G           O 3.10.53-HULK2 #1
> [<c0018e68>] (unwind_backtrace+0x0/0x11c) from [<c0014548>]
> (show_stack+0x10/0x14)
> [<c0014548>] (show_stack+0x10/0x14) from [<c02e22b4>]
> (dump_header.isra.12+0x90/0x1c0)
> [<c02e22b4>] (dump_header.isra.12+0x90/0x1c0) from [<c00c7428>]
> (oom_kill_process+0x5c/0x3d8)
> [<c00c7428>] (oom_kill_process+0x5c/0x3d8) from [<c00c7bf4>]
> (out_of_memory+0x26c/0x2b0)
> [<c00c7bf4>] (out_of_memory+0x26c/0x2b0) from [<c00cad8c>]
> (__alloc_pages_nodemask+0x558/0x6f8)
> [<c00cad8c>] (__alloc_pages_nodemask+0x558/0x6f8) from [<c00d647c>]
> (shmem_getpage_gfp+0x1bc/0x5e0)
> [<c00d647c>] (shmem_getpage_gfp+0x1bc/0x5e0) from [<c00c4a24>]
> (generic_file_buffered_write+0xdc/0x23c)
> [<c00c4a24>] (generic_file_buffered_write+0xdc/0x23c) from [<c00c5b7c>]
> (__generic_file_aio_write+0x33c/0x3a8)
> [<c00c5b7c>] (__generic_file_aio_write+0x33c/0x3a8) from [<c00c5c3c>]
> (generic_file_aio_write+0x54/0xb0)
> [<c00c5c3c>] (generic_file_aio_write+0x54/0xb0) from [<c00ff5ec>]
> (do_sync_write+0x74/0x98)
> [<c00ff5ec>] (do_sync_write+0x74/0x98) from [<c00fff80>]
> (vfs_write+0xcc/0x1a8)
> [<c00fff80>] (vfs_write+0xcc/0x1a8) from [<c0100374>] (SyS_write+0x38/0x64)
> [<c0100374>] (SyS_write+0x38/0x64) from [<c0010960>]
> (ret_fast_syscall+0x0/0x60)
> Mem-info:
> Normal per-cpu:
> CPU    0: hi:   90, btch:  15 usd:  15
> CPU    1: hi:   90, btch:  15 usd:  19
> active_anon:1505 inactive_anon:35705 isolated_anon:0
>   active_file:0 inactive_file:1 isolated_file:0
>   unevictable:10743 dirty:0 writeback:0 unstable:0
>   free:452 slab_reclaimable:1377 slab_unreclaimable:2922
>   mapped:833 shmem:36300 pagetables:161 bounce:0
>   free_cma:0
> Normal free:1808kB min:1812kB low:2264kB high:2716kB active_anon:6020kB
> inactive_anon:142820kB active_file:0kB inactive_file:4kB
> unevictable:42972kB isolated(anon):0kB isolated(file):0kB
> present:307200kB managed:205416kB mlocked:0kB dirty:0kB writeback:0kB
> mapped:3332kB shmem:145200kB slab_reclaimable:5508kB
> slab_unreclaimable:11688kB kernel_stack:672kB pagetables:644kB
> unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:23
> all_unreclaimable? yes
> lowmem_reserve[]: 0 0 0
> Normal: 24*4kB (UEMR) 21*8kB (UER) 5*16kB (UEM) 2*32kB (MR) 2*64kB (M)
> 2*128kB (M) 2*256kB (M) 1*512kB (M) 0*1024kB 0*2048kB 0*4096kB = 1816kB
> 47049 total pagecache pages
> 76800 pages of RAM
> 679 free pages
> 22676 reserved pages
> 2830 slab pages
> 532708 pages shared
> 0 pages swap cached
> [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name
> [ 1021]     0  1021      493      139       5        0             0 RATC
> [ 1023]     0  1023      493      123       5        0         -1000 RATC
> [ 1112]     0  1112     1302      511       4        0         -1000 sshd
> [ 1130]     0  1130      824      278       5        0             0 crond
> [ 1505]     0  1505      490      130       5        0             0
> take_cpu_rate
> [ 1506]     0  1506      490       32       5        0             0
> take_cpu_rate
> [ 1508]     0  1508      490       32       5        0             0
> take_cpu_rate
> [ 2315]     0  2315      731      165       5        0             0 getty
> [11839]     0 11839      731      191       5        0             0
> debug_runtest.s
> [11848]     0 11848      731      165       5        0             0
> istress.sh
> [11852]     0 11852      730      156       5        0             0
> spacectl.sh
> [12109]     0 12109      730      156       5        0             0
> bsd_stress.sh
> [ 8552]     0  8552      462      130       4        0             0
> lockf.test
> [ 8553]     0  8553      462       45       4        0             0
> lockf.test
> [ 8554]     0  8554      462       86       4        0             0
> lockf.test
> [ 8555]     0  8555      462       86       4        0             0
> lockf.test
> [ 8557]     0  8557      462       86       4        0             0
> lockf.test
> [10732]     0 10732      462       16       4        0             0
> lockf.test
> [10042]     0 10042      730      132       5        0             0 bash
> [10043]     0 10043      730      159       5        0             0
> runtest.sh
> [10068]     0 10068      730       92       5        0             0
> runtest.sh
> [10069]     0 10069      730      159       5        0             0
> rel_mem_inodeca
> [10072]     0 10072      697       99       4        0             0 sleep
> [ 8403]     0  8403      697       98       4        0             0 cp
> [ 8569]     0  8569      730      159       5        0             0
> runtest.sh
> [ 8606]     0  8606      730       92       5        0             0
> runtest.sh
> [ 8607]     0  8607      730      155       5        0             0
> rel_mem_filecac
> [ 8610]     0  8610      697       99       4        0             0 sleep
> [ 8611]     0  8611      732       99       5        0             0 tr
> [ 8612]     0  8612      730       99       5        0             0 dd
> [ 9073]     0  9073     1454      593       6        0             0 sshd
> [ 9083]   502  9083     1302      283       5        0             0 sshd
> [ 9086]     0  9086     1463      542       7        0             0
> syslog-ng
> [ 9090]     0  9090      730       19       4        0             0
> rel_mem_inodeca
>
> more info about it
> *****************Start oom extend info.*****************
> Vmallocinfo Start >>>>>>>>>>>>>>>>>>>>
> 0xbf000000-0xbf006000   24576 module_alloc_update_bounds+0xc/0x5c
> pages=5 vmalloc
> 0xbf00a000-0xbf00e000   16384 module_alloc_update_bounds+0xc/0x5c
> pages=3 vmalloc
> 0xbf011000-0xbf013000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf015000-0xbf017000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf019000-0xbf01b000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf01d000-0xbf021000   16384 module_alloc_update_bounds+0xc/0x5c
> pages=3 vmalloc
> 0xbf024000-0xbf032000   57344 module_alloc_update_bounds+0xc/0x5c
> pages=13 vmalloc
> 0xbf039000-0xbf03c000   12288 module_alloc_update_bounds+0xc/0x5c
> pages=2 vmalloc
> 0xbf03f000-0xbf042000   12288 module_alloc_update_bounds+0xc/0x5c
> pages=2 vmalloc
> 0xbf044000-0xbf046000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf048000-0xbf04a000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf04c000-0xbf04e000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf050000-0xbf054000   16384 module_alloc_update_bounds+0xc/0x5c
> pages=3 vmalloc
> 0xbf056000-0xbf059000   12288 module_alloc_update_bounds+0xc/0x5c
> pages=2 vmalloc
> 0xbf05b000-0xbf069000   57344 module_alloc_update_bounds+0xc/0x5c
> pages=13 vmalloc
> 0xbf06d000-0xbf08e000  135168 module_alloc_update_bounds+0xc/0x5c
> pages=32 vmalloc
> 0xbf096000-0xbf0ce000  229376 module_alloc_update_bounds+0xc/0x5c
> pages=55 vmalloc
> 0xbf0e8000-0xbf0ea000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xbf0ec000-0xbf0fe000   73728 module_alloc_update_bounds+0xc/0x5c
> pages=17 vmalloc
> 0xbf105000-0xbf12a000  151552 module_alloc_update_bounds+0xc/0x5c
> pages=36 vmalloc
> 0xbf13d000-0xbf145000   32768 module_alloc_update_bounds+0xc/0x5c
> pages=7 vmalloc
> 0xbf149000-0xbf168000  126976 module_alloc_update_bounds+0xc/0x5c
> pages=30 vmalloc
> 0xbf173000-0xbf179000   24576 module_alloc_update_bounds+0xc/0x5c
> pages=5 vmalloc
> 0xbf17c000-0xbf180000   16384 module_alloc_update_bounds+0xc/0x5c
> pages=3 vmalloc
> 0xbf182000-0xbf186000   16384 module_alloc_update_bounds+0xc/0x5c
> pages=3 vmalloc
> 0xbf192000-0xbf1a5000   77824 module_alloc_update_bounds+0xc/0x5c
> pages=18 vmalloc
> 0xbf1ab000-0xbf1ad000    8192 module_alloc_update_bounds+0xc/0x5c
> pages=1 vmalloc
> 0xd3000000-0xd3021000  135168 ekbox_reinit+0x3c/0xcc phys=9fa00000 ioremap
> 0xd3022000-0xd3024000    8192 of_iomap+0x30/0x3c phys=1a001000 ioremap
> 0xd3024000-0xd3027000   12288 of_iomap+0x30/0x3c phys=1a000000 ioremap
> 0xd3028000-0xd302a000    8192 of_iomap+0x30/0x3c phys=20000000 ioremap
> 0xd302a000-0xd302c000    8192 of_iomap+0x30/0x3c phys=20011000 ioremap
> 0xd302c000-0xd302f000   12288 of_iomap+0x30/0x3c phys=1a000000 ioremap
> 0xd3030000-0xd3032000    8192 of_iomap+0x30/0x3c phys=20013000 ioremap
> 0xd3032000-0xd3034000    8192 bsp_init_led+0x320/0x5e8 phys=f000b000
> ioremap
> 0xd3034000-0xd3036000    8192 bsp_init_led+0x394/0x5e8 phys=f000b000
> ioremap
> 0xd303a000-0xd307b000  266240 atomic_pool_init+0x0/0x11c phys=8d500000 user
> 0xd3080000-0xd3082000    8192 l2cache_init+0xb8/0x3b8 phys=16800000 ioremap
> 0xd3916000-0xd3959000  274432 0xbf08e26c pages=66 vmalloc
> 0xd3959000-0xd3965000   49152 0xbf08e280 pages=11 vmalloc
> 0xd3965000-0xd396a000   20480 0xbf08e2e0 pages=4 vmalloc
> 0xd396a000-0xd396d000   12288 0xbf08e2ec pages=2 vmalloc
> 0xd3a3f000-0xd3a42000   12288 pcpu_extend_area_map+0x18/0xa0 pages=2
> vmalloc
> 0xd3ad8000-0xd3ada000    8192 0xbf1a54ac phys=9fffe000 ioremap
> 0xd3adc000-0xd3ade000    8192 0xbf1a517c phys=30000000 ioremap
> 0xd3ade000-0xd3ae0000    8192 0xbf1a51b0 phys=20000000 ioremap
> 0xd3af2000-0xd3af4000    8192 0xbf1a517c phys=30001000 ioremap
> 0xd3af4000-0xd3af6000    8192 0xbf1a51b0 phys=20000000 ioremap
> 0xd3af6000-0xd3af8000    8192 0xbf1a517c phys=30002000 ioremap
> 0xd3af8000-0xd3afa000    8192 0xbf1a51b0 phys=20000000 ioremap
> 0xd3afa000-0xd3afc000    8192 0xbf1a517c phys=30003000 ioremap
> 0xd3afc000-0xd3afe000    8192 0xbf1a51b0 phys=20000000 ioremap
> 0xd3c00000-0xd3e01000 2101248 kbox_proc_mem_write+0x104/0x1cc
> phys=9f800000 ioremap
> 0xd4000000-0xd8001000 67112960 devm_ioremap+0x38/0x70 phys=40000000 ioremap
> 0xfe001000-0xfe002000    4096 iotable_init+0x0/0xb4 phys=20001000 ioremap
> 0xfe200000-0xfe201000    4096 iotable_init+0x0/0xb4 phys=1a000000 ioremap
> 0xfee00000-0xff000000 2097152 pci_reserve_io+0x0/0x30 ioremap
> Vmallocinfo End <<<<<<<<<<<<<<<<<<<<
>
> [SLUB]Slabinfo Start >>>>>>>>>>>>>>>>>>>>
> # name            <active_objs> <num_objs> <objsize> <objperslab>
> <pagesperslab> : slabdata <active_slabs> <num_slabs>
> nfs_direct_cache       0      0    120   34    1 : slabdata      0      0
> nfs_commit_data       18     18    448   18    2 : slabdata      1      1
> nfs_read_data          0      0    576   14    2 : slabdata      0      0
> nfs_inode_cache        0      0    816   20    4 : slabdata      0      0
> rpc_inode_cache        0      0    512   16    2 : slabdata      0      0
> jffs2_refblock       432    432    248   16    1 : slabdata     27     27
> jffs2_i              274    546    552   14    2 : slabdata     39     39
> bsg_cmd                0      0    288   14    1 : slabdata      0      0
> mqueue_inode_cache     23     23    704   23    4 : slabdata      1      1
> squashfs_inode_cache      0      0    576   14    2 : slabdata
> 0      0
> ext2_inode_cache       0      0    640   12    2 : slabdata      0      0
> pid_namespace        102    102     80   51    1 : slabdata      2      2
> user_namespace         0      0    224   18    1 : slabdata      0      0
> posix_timers_cache      0      0    160   25    1 : slabdata      0      0
> UDP-Lite               0      0    640   12    2 : slabdata      0      0
> UDP                   24     24    640   12    2 : slabdata      2      2
> tw_sock_TCP           64     64    128   32    1 : slabdata      2      2
> TCP                   24     24   1344   12    4 : slabdata      2      2
> eventpoll_pwq        204    204     40  102    1 : slabdata      2      2
> sgpool-128            12     12   2560   12    8 : slabdata      1      1
> sgpool-64             12     12   1280   12    4 : slabdata      1      1
> sgpool-16             12     12    320   12    1 : slabdata      1      1
> blkdev_queue          50     50   1272   25    8 : slabdata      2      2
> blkdev_requests       54     54    216   18    1 : slabdata      3      3
> fsnotify_event_holder   7696   8960     16  256    1 : slabdata     35
>     35
> fsnotify_event       112    112     72   56    1 : slabdata      2      2
> biovec-256            10     10   3072   10    8 : slabdata      1      1
> biovec-128             0      0   1536   21    8 : slabdata      0      0
> biovec-64              0      0    768   21    4 : slabdata      0      0
> sock_inode_cache      64     64    512   16    2 : slabdata      4      4
> skbuff_fclone_cache     63     63    384   21    2 : slabdata      3      3
> file_lock_cache     1904   1904    120   34    1 : slabdata     56     56
> net_namespace          0      0   2240   14    8 : slabdata      0      0
> shmem_inode_cache  10184  10890    528   15    2 : slabdata    726    726
> proc_inode_cache     391    496    504   16    2 : slabdata     31     31
> sigqueue             169    308    144   28    1 : slabdata     11     11
> bdev_cache            23     23    704   23    4 : slabdata      1      1
> inode_cache         2920   3315    472   17    2 : slabdata    195    195
> dentry             13777  20412    144   28    1 : slabdata    729    729
> buffer_head            0      0     64   64    1 : slabdata      0      0
> vm_area_struct      1702   1748     88   46    1 : slabdata     38     38
> signal_cache         171    240    640   12    2 : slabdata     20     20
> sighand_cache        110    144   1344   12    4 : slabdata     12     12
> task_struct          134    182   2304   14    8 : slabdata     13     13
> anon_vma_chain      2364   3328     32  128    1 : slabdata     26     26
> anon_vma            2409   2409     56   73    1 : slabdata     33     33
> debug_objects_cache   3594   8500     24  170    1 : slabdata     50     50
> radix_tree_node     1259   1352    304   13    1 : slabdata    104    104
> idr_layer_cache      105    105   1080   15    4 : slabdata      7      7
> kmalloc-8192          20     20   8192    4    8 : slabdata      5      5
> kmalloc-4096          48     48   4096    8    8 : slabdata      6      6
> kmalloc-2048          96     96   2048   16    8 : slabdata      6      6
> kmalloc-1024         330    400   1024   16    4 : slabdata     25     25
> kmalloc-512          911    976    512   16    2 : slabdata     61     61
> kmalloc-256          211    240    256   16    1 : slabdata     15     15
> kmalloc-192         1228   1344    192   21    1 : slabdata     64     64
> kmalloc-128         2324   2624    128   32    1 : slabdata     82     82
> kmalloc-64         23749  24192     64   64    1 : slabdata    378    378
> kmem_cache_node      128    128     64   64    1 : slabdata      2      2
> kmem_cache            96     96    128   32    1 : slabdata      3      3
> Slabinfo End <<<<<<<<<<<<<<<<<<<<
>
>
> Filesystem            1K-blocks    Used   Available Use(%)   Mounted on
> tmpfs                   524288        4   524284        0%   /tmp
> none                     10240    10240        0      100%   /var
> tmpfs                   108248       16   108232        0%   /dev
> tmpfs                   108248        0   108248        0%   /dev/shm
> tmpfs                   173192   134960    38232       77%   /tmp
> *****smap info of all task:*****
>    smaps info of task-sshd[9073], rss:2372 kB:
>    smaps info of task-syslog-ng[9086], rss:2168 kB:
>    smaps info of task-sshd[1112], rss:2044 kB:
>    smaps info of task-sshd[9083], rss:1132 kB:
>    smaps info of task-crond[1130], rss:1112 kB:
> ********    mem info     *****    Total:                 216496 kB
>      Total free:              1808 kB
>      User space:            191816 kB
>      Mlock:                      0 kB
>      Kernel space:           22872 kB
>      Bootmem reserved:       90704 kB
>      kernel_image_info:
>          Kernel code:   0x80008000-0x8043369f
>          Kernel data:   0x80496000-0x8056559b
>      module info:
>          physmap 3kB         Live 0xbf1ab000 (O)
>              refrence count: 1
>                                 -
>          Drv_Gmac_K 71kB         Live 0xbf192000 (O)
>              refrence count: 0
>                                 -
>          rtos_snapshot 10kB         Live 0xbf182000 (O)
>              refrence count: 0
>                                 -
>          rtos_kbox_panic 10kB         Live 0xbf17c000 (O)
>              refrence count: 0
>                                 -
>          rsm 16kB         Live 0xbf173000 (O)
>              refrence count: 0
>                                 -
>          nfsv4 119kB         Live 0xbf149000
>              refrence count: 0
>                                 -
>          nfsv3 24kB         Live 0xbf13d000
>              refrence count: 0
>                                 -
>          nfs 141kB         Live 0xbf105000
>              refrence count: 2
>              nfsv4,
>              nfsv3,
>          lockd 64kB         Live 0xbf0ec000
>              refrence count: 2
>              nfsv3,
>              nfs,
>          nfs_acl 2kB         Live 0xbf0e8000
>              refrence count: 1
>              nfsv3,
>          sunrpc 216kB         Live 0xbf096000
>              refrence count: 5
>              nfsv4,
>              nfsv3,
>              nfs,
>              lockd,
>              nfs_acl,
>          jffs2 125kB         Live 0xbf06d000
>              refrence count: 1
>                                 -
>          cfi_cmdset_0002 51kB         Live 0xbf05b000
>              refrence count: 1
>                                 -
>          cfi_probe 5kB         Live 0xbf056000
>              refrence count: 0
>                                 -
>          cfi_util 11kB         Live 0xbf050000
>              refrence count: 2
>              cfi_cmdset_0002,
>              cfi_probe,
>          gen_probe 2kB         Live 0xbf04c000
>              refrence count: 1
>              cfi_probe,
>          cmdlinepart 2kB         Live 0xbf048000
>              refrence count: 0
>                                 -
>          chipreg 2kB         Live 0xbf044000
>              refrence count: 2
>              physmap,
>              cfi_probe,
>          mtdblock 4kB         Live 0xbf03f000
>              refrence count: 0
>                                 -
>          mtd_blkdevs 7kB         Live 0xbf039000
>              refrence count: 1
>              mtdblock,
>          mtd 48kB         Live 0xbf024000
>              refrence count: 17
>              physmap,
>              jffs2,
>              cfi_cmdset_0002,
>              cmdlinepart,
>              mtdblock,
>              mtd_blkdevs,
>          uio 8kB         Live 0xbf01d000
>              refrence count: 0
>                                 -
>          xt_tcpudp 2kB         Live 0xbf019000
>              refrence count: 0
>                                 -
>          ipt_REJECT 2kB         Live 0xbf015000
>              refrence count: 0
>                                 -
>          iptable_filter 1kB         Live 0xbf011000
>              refrence count: 0
>                                 -
>          ip_tables 11kB         Live 0xbf00a000
>              refrence count: 1
>              iptable_filter,
>          x_tables 16kB         Live 0xbf000000
>              refrence count: 4
>              xt_tcpudp,
>              ipt_REJECT,
>              iptable_filter,
>              ip_tables,
> ******     pagecache_info:     ******
>    /rel_mem_filecache_tc11/0 : nrpages = 23260.
>    /volatile/log/auth.log : nrpages = 1947.
>    /usr/bin/gdb : nrpages = 845.
>    /lib/libcrypto.so.1.0.0 : nrpages = 418.
>    /usr/lib/libgio-2.0.so.0.3600.4 : nrpages = 391.
>    /lib/libc-2.18.so : nrpages = 370.
>    /usr/lib/libperl.so.5.14.3 : nrpages = 348.
>    /usr/lib/libglib-2.0.so.0.3600.4 : nrpages = 314.
>    /usr/lib/libstdc++.so.6.0.17 : nrpages = 277.
>    /fs_stress_t/fs_stress_t/testcase/bin/fsback/unit : nrpages = 256.
>    /fs_stress_t/fs_stress_t_src/testcase/fsback/unit : nrpages = 256.
>    /usr/bin/makedumpfile : nrpages = 238.
>    /usr/bin/perf : nrpages = 228.
>    /usr/sbin/sshd : nrpages = 226.
>    /usr/bin/ssh : nrpages = 202.
>    /usr/lib/libbfd-2.23.2.so : nrpages = 195.
>    /volatile/log/syslog : nrpages = 192.
>    /volatile/log/kern.log : nrpages = 186.
>    /lib/libm-2.18.so : nrpages = 173.
>    /volatile/log/error : nrpages = 162.
> *****************End oom extend info.*****************
>
>
>
>
> On 2015/12/1 6:08, David Rientjes wrote:
>> On Mon, 30 Nov 2015, Chenjie (K) wrote:
>>
>>> My kernel version is 3.10 ,but the 4.3 is the same
>>> and the newest code is
>>>
>>>     for_each_process(p) {
>>>         if (!process_shares_mm(p, mm))
>>>             continue;
>>>         if (same_thread_group(p, victim))
>>>             continue;
>>>         if (unlikely(p->flags & PF_KTHREAD))
>>>             continue;
>>>         if (p->signal->oom_score_adj == OOM_SCORE_ADJ_MIN)
>>>             continue;
>>>
>>> so this not add the !is_global_init also.
>>>
>>> when we vfork (CLONE_VM) a process,the copy_mm
>>>     if (clone_flags & CLONE_VM) {
>>>         atomic_inc(&oldmm->mm_users);
>>>         mm = oldmm;
>>>         goto good_mm;
>>>     }
>>> use the parent mm.
>>>
>>
>> I think it might be a legitimate fix, but if the oom killer is killing
>> pid
>> 9134 in your log then I assume the next call to the oom killer will panic
>> the system anyway unless there is actually a process using less memory
>> that can be killed.  Would you mind enabling vm.oom_dump_tasks (it should
>> default to enabled) and post the entire oom killer log?
>>

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ