lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Sun, 28 Nov 2021 23:06:58 +0800
From:   kernel test robot <oliver.sang@...el.com>
To:     Mel Gorman <mgorman@...hsingularity.net>
Cc:     0day robot <lkp@...el.com>, LKML <linux-kernel@...r.kernel.org>,
        lkp@...ts.01.org, ying.huang@...el.com, feng.tang@...el.com,
        zhengjun.xing@...ux.intel.com, fengwei.yin@...el.com,
        guobing.chen@...el.com, ming.a.chen@...el.com, frank.du@...el.com,
        Shuhua.Fan@...el.com, wangyang.guo@...el.com,
        Wenhuan.Huang@...el.com, jessica.ji@...el.com, shan.kang@...el.com,
        guangli.li@...el.com, tiejun.li@...el.com, yu.ma@...el.com,
        dapeng1.mi@...el.com, jiebin.sun@...el.com, gengxin.xie@...el.com,
        fan.zhao@...el.com, Peter Zijlstra <peterz@...radead.org>,
        Ingo Molnar <mingo@...nel.org>,
        Vincent Guittot <vincent.guittot@...aro.org>,
        Valentin Schneider <valentin.schneider@....com>,
        Aubrey Li <aubrey.li@...ux.intel.com>,
        Barry Song <song.bao.hua@...ilicon.com>,
        Mike Galbraith <efault@....de>,
        Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
        Mel Gorman <mgorman@...hsingularity.net>, yu.c.chen@...el.com
Subject: [sched/fair]  b4d95a034c:
 phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s -26.3% regression



Greeting,

FYI, we noticed a -26.3% regression of phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s due to commit:


commit: b4d95a034cffb1e4424874645549d3cac2de5c02 ("[PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs")
url: https://github.com/0day-ci/linux/commits/Mel-Gorman/Adjust-NUMA-imbalance-for-multiple-LLCs/20211125-232336
base: https://git.kernel.org/cgit/linux/kernel/git/tip/tip.git 8c92606ab81086db00cbb73347d124b4eb169b7e

in testcase: phoronix-test-suite
on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 128G memory
with following parameters:

	test: tiobench-1.3.1
	option_a: Random Write
	option_b: 64MB
	option_c: 8
	cpufreq_governor: performance
	ucode: 0x5003006

test-description: The Phoronix Test Suite is the most comprehensive testing and benchmarking platform available that provides an extensible framework for which new tests can be easily added.
test-url: http://www.phoronix-test-suite.com/



If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@...el.com>


Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        sudo bin/lkp install job.yaml           # job file is attached in this email
        bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
        sudo bin/lkp run generated-yaml-file

        # if come across any failure that blocks the test,
        # please remove ~/.lkp and /lkp dir to run from a clean state.

=========================================================================================
compiler/cpufreq_governor/kconfig/option_a/option_b/option_c/rootfs/tbox_group/test/testcase/ucode:
  gcc-9/performance/x86_64-rhel-8.3/Random Write/64MB/8/debian-x86_64-phoronix/lkp-csl-2sp8/tiobench-1.3.1/phoronix-test-suite/0x5003006

commit: 
  fee45dc486 ("sched/fair: Use weight of SD_NUMA domain in find_busiest_group")
  b4d95a034c ("sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs")

fee45dc486dd343a b4d95a034cffb1e442487464554 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    190841 ±  4%     -26.3%     140600 ±  3%  phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s
      5.17 ±128%  +1.3e+05%       6530 ± 64%  proc-vmstat.numa_hint_faults
     76503 ± 40%     -25.3%      57153 ±  4%  interrupts.CAL:Function_call_interrupts
      4574 ± 50%     -82.7%     791.14 ± 42%  interrupts.CPU1.CAL:Function_call_interrupts
      3.32 ± 41%    +882.9%      32.65 ±  7%  perf-stat.i.cpu-migrations
     51246 ± 10%    +104.4%     104748 ±  3%  perf-stat.i.node-store-misses
      1465 ± 21%     -24.6%       1105 ± 13%  numa-vmstat.node0.nr_active_anon
     82443 ±  2%     -47.6%      43196 ± 14%  numa-vmstat.node0.nr_anon_pages
     10866 ±  4%      -8.3%       9965 ±  4%  numa-vmstat.node0.nr_kernel_stack
     14846 ± 15%     -50.1%       7413 ± 43%  numa-vmstat.node0.nr_mapped
      1033 ±  2%     -31.7%     706.14 ± 15%  numa-vmstat.node0.nr_page_table_pages
      1465 ± 21%     -24.6%       1105 ± 13%  numa-vmstat.node0.nr_zone_active_anon
      8909 ± 26%     +47.1%      13103 ± 20%  numa-vmstat.node1.nr_active_file
      8603 ± 15%    +458.9%      48088 ± 11%  numa-vmstat.node1.nr_anon_pages
      8949 ±  5%      +9.9%       9834 ±  4%  numa-vmstat.node1.nr_kernel_stack
    416.00 ±  7%     +79.4%     746.14 ± 14%  numa-vmstat.node1.nr_page_table_pages
      8909 ± 26%     +47.1%      13103 ± 20%  numa-vmstat.node1.nr_zone_active_file
      5844 ± 22%     -24.3%       4426 ± 13%  numa-meminfo.node0.Active(anon)
    121357 ± 13%     -45.1%      66683 ± 26%  numa-meminfo.node0.AnonHugePages
    329764 ±  2%     -47.6%     172811 ± 14%  numa-meminfo.node0.AnonPages
    346450           -47.6%     181374 ± 14%  numa-meminfo.node0.AnonPages.max
   2050555 ± 13%     -29.7%    1441806 ± 36%  numa-meminfo.node0.Inactive
     10866 ±  4%      -8.3%       9966 ±  4%  numa-meminfo.node0.KernelStack
     59355 ± 15%     -50.0%      29668 ± 43%  numa-meminfo.node0.Mapped
   2872827 ± 12%     -20.3%    2288843 ± 24%  numa-meminfo.node0.MemUsed
      4133 ±  3%     -31.6%       2829 ± 15%  numa-meminfo.node0.PageTables
     37735 ± 26%     +47.9%      55814 ± 18%  numa-meminfo.node1.Active
     35639 ± 26%     +47.1%      52416 ± 20%  numa-meminfo.node1.Active(file)
      5616 ± 27%    +912.0%      56834 ± 44%  numa-meminfo.node1.AnonHugePages
     34408 ± 15%    +459.0%     192349 ± 11%  numa-meminfo.node1.AnonPages
     39089 ± 19%    +418.8%     202789 ± 12%  numa-meminfo.node1.AnonPages.max
      8950 ±  5%      +9.9%       9833 ±  4%  numa-meminfo.node1.KernelStack
      1666 ±  6%     +79.0%       2983 ± 14%  numa-meminfo.node1.PageTables
      4925 ±  8%     -14.0%       4237 ±  8%  slabinfo.kmalloc-cg-16.active_objs
      4925 ±  8%     -14.0%       4237 ±  8%  slabinfo.kmalloc-cg-16.num_objs
      3328           +11.4%       3709 ±  3%  slabinfo.kmalloc-cg-192.active_objs
      3328           +11.4%       3709 ±  3%  slabinfo.kmalloc-cg-192.num_objs
      2545 ±  3%     +11.8%       2845 ±  3%  slabinfo.kmalloc-cg-1k.active_objs
      2545 ±  3%     +11.8%       2845 ±  3%  slabinfo.kmalloc-cg-1k.num_objs
      1054 ±  6%     +24.3%       1310 ±  3%  slabinfo.kmalloc-cg-2k.active_objs
      1054 ±  6%     +24.3%       1310 ±  3%  slabinfo.kmalloc-cg-2k.num_objs
      4376 ±  5%     +22.2%       5347 ±  2%  slabinfo.kmalloc-cg-64.active_objs
      4376 ±  5%     +22.2%       5347 ±  2%  slabinfo.kmalloc-cg-64.num_objs
      2663 ±  7%     +27.0%       3382 ±  3%  slabinfo.kmalloc-cg-96.active_objs
      2663 ±  7%     +27.0%       3382 ±  3%  slabinfo.kmalloc-cg-96.num_objs
      1446 ±  9%     -21.6%       1133 ±  7%  slabinfo.task_group.active_objs
      1446 ±  9%     -21.6%       1133 ±  7%  slabinfo.task_group.num_objs
     14208 ±  5%     -13.5%      12296 ±  3%  slabinfo.vmap_area.active_objs
     14213 ±  5%     -13.5%      12297 ±  3%  slabinfo.vmap_area.num_objs
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.calltrace.cycles-pp.read
      7.96 ±124%      -5.5        2.49 ±158%  perf-profile.calltrace.cycles-pp.zap_pte_range.unmap_page_range.unmap_vmas.exit_mmap.mmput
      6.44 ±111%      -5.3        1.19 ±244%  perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.unmap_page_range.unmap_vmas.exit_mmap
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.calltrace.cycles-pp.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read
      5.41 ±105%      -4.2        1.19 ±244%  perf-profile.calltrace.cycles-pp.release_task.wait_task_zombie.do_wait.kernel_waitid.__do_sys_waitid
      4.22 ±101%      -4.2        0.00        perf-profile.calltrace.cycles-pp.__dentry_kill.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      4.22 ±101%      -4.2        0.00        perf-profile.calltrace.cycles-pp.d_invalidate.proc_invalidate_siblings_dcache.release_task.wait_task_zombie.do_wait
      4.22 ±101%      -4.2        0.00        perf-profile.calltrace.cycles-pp.proc_invalidate_siblings_dcache.release_task.wait_task_zombie.do_wait.kernel_waitid
      4.22 ±101%      -4.2        0.00        perf-profile.calltrace.cycles-pp.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache.release_task.wait_task_zombie
      4.22 ±101%      -4.2        0.00        perf-profile.calltrace.cycles-pp.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache.release_task
      8.36 ±154%      -4.0        4.36 ±179%  perf-profile.calltrace.cycles-pp.mmput.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve
      8.36 ±154%      -4.0        4.36 ±179%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.begin_new_exec.load_elf_binary.exec_binprm
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.kernel_waitid.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.do_wait.kernel_waitid.__do_sys_waitid.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.calltrace.cycles-pp.wait_task_zombie.do_wait.kernel_waitid.__do_sys_waitid.do_syscall_64
      8.36 ±154%      +0.1        8.49 ±177%  perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve.do_execveat_common
      9.47 ±137%      -7.0        2.49 ±158%  perf-profile.children.cycles-pp.unmap_vmas
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.children.cycles-pp.ksys_read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.children.cycles-pp.vfs_read
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.children.cycles-pp.seq_read_iter
      8.25 ±110%      -6.1        2.14 ±159%  perf-profile.children.cycles-pp.read
      7.96 ±124%      -5.5        2.49 ±158%  perf-profile.children.cycles-pp.zap_pte_range
      7.96 ±124%      -5.5        2.49 ±158%  perf-profile.children.cycles-pp.unmap_page_range
      6.44 ±111%      -5.3        1.19 ±244%  perf-profile.children.cycles-pp.page_remove_rmap
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.children.cycles-pp.new_sync_read
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.children.cycles-pp.proc_reg_read_iter
      6.40 ±108%      -4.3        2.14 ±159%  perf-profile.children.cycles-pp.show_interrupts
      5.41 ±105%      -4.2        1.19 ±244%  perf-profile.children.cycles-pp.release_task
      4.22 ±101%      -4.2        0.00        perf-profile.children.cycles-pp.__dentry_kill
      4.22 ±101%      -4.2        0.00        perf-profile.children.cycles-pp.d_invalidate
      4.22 ±101%      -4.2        0.00        perf-profile.children.cycles-pp.proc_invalidate_siblings_dcache
      4.22 ±101%      -4.2        0.00        perf-profile.children.cycles-pp.shrink_dcache_parent
      4.22 ±101%      -4.2        0.00        perf-profile.children.cycles-pp.shrink_dentry_list
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.children.cycles-pp.__do_sys_waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.children.cycles-pp.waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.children.cycles-pp.kernel_waitid
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.children.cycles-pp.do_wait
      5.41 ±105%      -3.6        1.79 ±169%  perf-profile.children.cycles-pp.wait_task_zombie
      8.36 ±154%      +0.1        8.49 ±177%  perf-profile.children.cycles-pp.begin_new_exec
      3.82 ±101%      -2.6        1.19 ±244%  perf-profile.self.cycles-pp.page_remove_rmap


                                                                                
                phoronix-test-suite.tiobench.RandomWrite.64MB.8.mb_s            
                                                                                
  220000 +------------------------------------------------------------------+   
         |                +                                                 |   
  210000 |-+              ::                                                |   
  200000 |-+   +.++      : +.  .+             +      +     +     +  +.+     |   
         |  +  :   +  .+ :   ++  +.   +   .+ : :  +  :+ +.+ +    :: : :  ++ |   
  190000 |.+ +:     ++  +          + + +.+  :: +.+ +:  +     +  : ::   :+  +|   
  180000 |-+  +                     +       +       +         +.+  +   +    |   
         |                                                                  |   
  170000 |-+                                                                |   
  160000 |-+                                                                |   
         |  O                                                               |   
  150000 |-+  OO  O  O    OO       O  O       O  O                          |   
  140000 |-O     O  O   O    OO OO       O          O                       |   
         |             O            O  O    O  O  O    O                    |   
  130000 +------------------------------------------------------------------+   
                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


---
0DAY/LKP+ Test Infrastructure                   Open Source Technology Center
https://lists.01.org/hyperkitty/list/lkp@lists.01.org       Intel Corporation

Thanks,
Oliver Sang


View attachment "config-5.16.0-rc1-00010-gb4d95a034cff" of type "text/plain" (173486 bytes)

View attachment "job-script" of type "text/plain" (7595 bytes)

View attachment "job.yaml" of type "text/plain" (4837 bytes)

View attachment "reproduce" of type "text/plain" (300 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ