[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20211128150658.GC5295@xsang-OptiPlex-9020>
Date: Sun, 28 Nov 2021 23:06:58 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Mel Gorman <mgorman@...hsingularity.net>
Cc: 0day robot <lkp@...el.com>, LKML <linux-kernel@...r.kernel.org>,
lkp@...ts.01.org, ying.huang@...el.com, feng.tang@...el.com,
zhengjun.xing@...ux.intel.com, fengwei.yin@...el.com,
guobing.chen@...el.com, ming.a.chen@...el.com, frank.du@...el.com,
Shuhua.Fan@...el.com, wangyang.guo@...el.com,
Wenhuan.Huang@...el.com, jessica.ji@...el.com, shan.kang@...el.com,
guangli.li@...el.com, tiejun.li@...el.com, yu.ma@...el.com,
dapeng1.mi@...el.com, jiebin.sun@...el.com, gengxin.xie@...el.com,
fan.zhao@...el.com, Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...nel.org>,
Vincent Guittot <vincent.guittot@...aro.org>,
Valentin Schneider <valentin.schneider@....com>,
Aubrey Li <aubrey.li@...ux.intel.com>,
Barry Song <song.bao.hua@...ilicon.com>,
Mike Galbraith <efault@....de>,
Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
Mel Gorman <mgorman@...hsingularity.net>, yu.c.chen@...el.com
Subject: [sched/fair] b4d95a034c:
phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s -26.3% regression
Greeting,
FYI, we noticed a -26.3% regression of phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s due to commit:
commit: b4d95a034cffb1e4424874645549d3cac2de5c02 ("[PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs")
url: https://github.com/0day-ci/linux/commits/Mel-Gorman/Adjust-NUMA-imbalance-for-multiple-LLCs/20211125-232336
base: https://git.kernel.org/cgit/linux/kernel/git/tip/tip.git 8c92606ab81086db00cbb73347d124b4eb169b7e
in testcase: phoronix-test-suite
on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 128G memory
with following parameters:
test: tiobench-1.3.1
option_a: Random Write
option_b: 64MB
option_c: 8
cpufreq_governor: performance
ucode: 0x5003006
test-description: The Phoronix Test Suite is the most comprehensive testing and benchmarking platform available that provides an extensible framework for which new tests can be easily added.
test-url: http://www.phoronix-test-suite.com/
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@...el.com>
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
=========================================================================================
compiler/cpufreq_governor/kconfig/option_a/option_b/option_c/rootfs/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/Random Write/64MB/8/debian-x86_64-phoronix/lkp-csl-2sp8/tiobench-1.3.1/phoronix-test-suite/0x5003006
commit:
fee45dc486 ("sched/fair: Use weight of SD_NUMA domain in find_busiest_group")
b4d95a034c ("sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs")
fee45dc486dd343a b4d95a034cffb1e442487464554
---------------- ---------------------------
%stddev %change %stddev
\ | \
190841 ± 4% -26.3% 140600 ± 3% phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s
5.17 ±128% +1.3e+05% 6530 ± 64% proc-vmstat.numa_hint_faults
76503 ± 40% -25.3% 57153 ± 4% interrupts.CAL:Function_call_interrupts
4574 ± 50% -82.7% 791.14 ± 42% interrupts.CPU1.CAL:Function_call_interrupts
3.32 ± 41% +882.9% 32.65 ± 7% perf-stat.i.cpu-migrations
51246 ± 10% +104.4% 104748 ± 3% perf-stat.i.node-store-misses
1465 ± 21% -24.6% 1105 ± 13% numa-vmstat.node0.nr_active_anon
82443 ± 2% -47.6% 43196 ± 14% numa-vmstat.node0.nr_anon_pages
10866 ± 4% -8.3% 9965 ± 4% numa-vmstat.node0.nr_kernel_stack
14846 ± 15% -50.1% 7413 ± 43% numa-vmstat.node0.nr_mapped
1033 ± 2% -31.7% 706.14 ± 15% numa-vmstat.node0.nr_page_table_pages
1465 ± 21% -24.6% 1105 ± 13% numa-vmstat.node0.nr_zone_active_anon
8909 ± 26% +47.1% 13103 ± 20% numa-vmstat.node1.nr_active_file
8603 ± 15% +458.9% 48088 ± 11% numa-vmstat.node1.nr_anon_pages
8949 ± 5% +9.9% 9834 ± 4% numa-vmstat.node1.nr_kernel_stack
416.00 ± 7% +79.4% 746.14 ± 14% numa-vmstat.node1.nr_page_table_pages
8909 ± 26% +47.1% 13103 ± 20% numa-vmstat.node1.nr_zone_active_file
5844 ± 22% -24.3% 4426 ± 13% numa-meminfo.node0.Active(anon)
121357 ± 13% -45.1% 66683 ± 26% numa-meminfo.node0.AnonHugePages
329764 ± 2% -47.6% 172811 ± 14% numa-meminfo.node0.AnonPages
346450 -47.6% 181374 ± 14% numa-meminfo.node0.AnonPages.max
2050555 ± 13% -29.7% 1441806 ± 36% numa-meminfo.node0.Inactive
10866 ± 4% -8.3% 9966 ± 4% numa-meminfo.node0.KernelStack
59355 ± 15% -50.0% 29668 ± 43% numa-meminfo.node0.Mapped
2872827 ± 12% -20.3% 2288843 ± 24% numa-meminfo.node0.MemUsed
4133 ± 3% -31.6% 2829 ± 15% numa-meminfo.node0.PageTables
37735 ± 26% +47.9% 55814 ± 18% numa-meminfo.node1.Active
35639 ± 26% +47.1% 52416 ± 20% numa-meminfo.node1.Active(file)
5616 ± 27% +912.0% 56834 ± 44% numa-meminfo.node1.AnonHugePages
34408 ± 15% +459.0% 192349 ± 11% numa-meminfo.node1.AnonPages
39089 ± 19% +418.8% 202789 ± 12% numa-meminfo.node1.AnonPages.max
8950 ± 5% +9.9% 9833 ± 4% numa-meminfo.node1.KernelStack
1666 ± 6% +79.0% 2983 ± 14% numa-meminfo.node1.PageTables
4925 ± 8% -14.0% 4237 ± 8% slabinfo.kmalloc-cg-16.active_objs
4925 ± 8% -14.0% 4237 ± 8% slabinfo.kmalloc-cg-16.num_objs
3328 +11.4% 3709 ± 3% slabinfo.kmalloc-cg-192.active_objs
3328 +11.4% 3709 ± 3% slabinfo.kmalloc-cg-192.num_objs
2545 ± 3% +11.8% 2845 ± 3% slabinfo.kmalloc-cg-1k.active_objs
2545 ± 3% +11.8% 2845 ± 3% slabinfo.kmalloc-cg-1k.num_objs
1054 ± 6% +24.3% 1310 ± 3% slabinfo.kmalloc-cg-2k.active_objs
1054 ± 6% +24.3% 1310 ± 3% slabinfo.kmalloc-cg-2k.num_objs
4376 ± 5% +22.2% 5347 ± 2% slabinfo.kmalloc-cg-64.active_objs
4376 ± 5% +22.2% 5347 ± 2% slabinfo.kmalloc-cg-64.num_objs
2663 ± 7% +27.0% 3382 ± 3% slabinfo.kmalloc-cg-96.active_objs
2663 ± 7% +27.0% 3382 ± 3% slabinfo.kmalloc-cg-96.num_objs
1446 ± 9% -21.6% 1133 ± 7% slabinfo.task_group.active_objs
1446 ± 9% -21.6% 1133 ± 7% slabinfo.task_group.num_objs
14208 ± 5% -13.5% 12296 ± 3% slabinfo.vmap_area.active_objs
14213 ± 5% -13.5% 12297 ± 3% slabinfo.vmap_area.num_objs
8.25 ±110% -6.1 2.14 ±159% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
8.25 ±110% -6.1 2.14 ±159% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
8.25 ±110% -6.1 2.14 ±159% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
8.25 ±110% -6.1 2.14 ±159% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
8.25 ±110% -6.1 2.14 ±159% perf-profile.calltrace.cycles-pp.read
7.96 ±124% -5.5 2.49 ±158% perf-profile.calltrace.cycles-pp.zap_pte_range.unmap_page_range.unmap_vmas.exit_mmap.mmput
6.44 ±111% -5.3 1.19 ±244% perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.unmap_page_range.unmap_vmas.exit_mmap
6.40 ±108% -4.3 2.14 ±159% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.40 ±108% -4.3 2.14 ±159% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64
6.40 ±108% -4.3 2.14 ±159% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read
6.40 ±108% -4.3 2.14 ±159% perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read
5.41 ±105% -4.2 1.19 ±244% perf-profile.calltrace.cycles-pp.release_task.wait_task_zombie.do_wait.kernel_waitid.__do_sys_waitid
4.22 ±101% -4.2 0.00 perf-profile.calltrace.cycles-pp.__dentry_kill.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
4.22 ±101% -4.2 0.00 perf-profile.calltrace.cycles-pp.d_invalidate.proc_invalidate_siblings_dcache.release_task.wait_task_zombie.do_wait
4.22 ±101% -4.2 0.00 perf-profile.calltrace.cycles-pp.proc_invalidate_siblings_dcache.release_task.wait_task_zombie.do_wait.kernel_waitid
4.22 ±101% -4.2 0.00 perf-profile.calltrace.cycles-pp.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache.release_task.wait_task_zombie
4.22 ±101% -4.2 0.00 perf-profile.calltrace.cycles-pp.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache.release_task
8.36 ±154% -4.0 4.36 ±179% perf-profile.calltrace.cycles-pp.mmput.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve
8.36 ±154% -4.0 4.36 ±179% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.begin_new_exec.load_elf_binary.exec_binprm
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.kernel_waitid.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.do_wait.kernel_waitid.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe
5.41 ±105% -3.6 1.79 ±169% perf-profile.calltrace.cycles-pp.wait_task_zombie.do_wait.kernel_waitid.__do_sys_waitid.do_syscall_64
8.36 ±154% +0.1 8.49 ±177% perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve.do_execveat_common
9.47 ±137% -7.0 2.49 ±158% perf-profile.children.cycles-pp.unmap_vmas
8.25 ±110% -6.1 2.14 ±159% perf-profile.children.cycles-pp.ksys_read
8.25 ±110% -6.1 2.14 ±159% perf-profile.children.cycles-pp.vfs_read
8.25 ±110% -6.1 2.14 ±159% perf-profile.children.cycles-pp.seq_read_iter
8.25 ±110% -6.1 2.14 ±159% perf-profile.children.cycles-pp.read
7.96 ±124% -5.5 2.49 ±158% perf-profile.children.cycles-pp.zap_pte_range
7.96 ±124% -5.5 2.49 ±158% perf-profile.children.cycles-pp.unmap_page_range
6.44 ±111% -5.3 1.19 ±244% perf-profile.children.cycles-pp.page_remove_rmap
6.40 ±108% -4.3 2.14 ±159% perf-profile.children.cycles-pp.new_sync_read
6.40 ±108% -4.3 2.14 ±159% perf-profile.children.cycles-pp.proc_reg_read_iter
6.40 ±108% -4.3 2.14 ±159% perf-profile.children.cycles-pp.show_interrupts
5.41 ±105% -4.2 1.19 ±244% perf-profile.children.cycles-pp.release_task
4.22 ±101% -4.2 0.00 perf-profile.children.cycles-pp.__dentry_kill
4.22 ±101% -4.2 0.00 perf-profile.children.cycles-pp.d_invalidate
4.22 ±101% -4.2 0.00 perf-profile.children.cycles-pp.proc_invalidate_siblings_dcache
4.22 ±101% -4.2 0.00 perf-profile.children.cycles-pp.shrink_dcache_parent
4.22 ±101% -4.2 0.00 perf-profile.children.cycles-pp.shrink_dentry_list
5.41 ±105% -3.6 1.79 ±169% perf-profile.children.cycles-pp.__do_sys_waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.children.cycles-pp.waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.children.cycles-pp.kernel_waitid
5.41 ±105% -3.6 1.79 ±169% perf-profile.children.cycles-pp.do_wait
5.41 ±105% -3.6 1.79 ±169% perf-profile.children.cycles-pp.wait_task_zombie
8.36 ±154% +0.1 8.49 ±177% perf-profile.children.cycles-pp.begin_new_exec
3.82 ±101% -2.6 1.19 ±244% perf-profile.self.cycles-pp.page_remove_rmap
phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s
220000 +------------------------------------------------------------------+
| + |
210000 |-+ :: |
200000 |-+ +.++ : +. .+ + + + + +.+ |
| + : + .+ : ++ +. + .+ : : + :+ +.+ + :: : : ++ |
190000 |.+ +: ++ + + + +.+ :: +.+ +: + + : :: :+ +|
180000 |-+ + + + + +.+ + + |
| |
170000 |-+ |
160000 |-+ |
| O |
150000 |-+ OO O O OO O O O O |
140000 |-O O O O OO OO O O |
| O O O O O O O |
130000 +------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
---
0DAY/LKP+ Test Infrastructure Open Source Technology Center
https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation
Thanks,
Oliver Sang
View attachment "config-5.16.0-rc1-00010-gb4d95a034cff" of type "text/plain" (173486 bytes)
View attachment "job-script" of type "text/plain" (7595 bytes)
View attachment "job.yaml" of type "text/plain" (4837 bytes)
View attachment "reproduce" of type "text/plain" (300 bytes)
Powered by blists - more mailing lists