lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 21 Nov 2022 11:03:08 +0800 From: kernel test robot <yujie.liu@...el.com> To: David Hildenbrand <david@...hat.com> CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, Andrew Morton <akpm@...ux-foundation.org>, Jason Gunthorpe <jgg@...dia.com>, John Hubbard <jhubbard@...dia.com>, Andrea Arcangeli <aarcange@...hat.com>, Hugh Dickins <hughd@...gle.com>, Peter Xu <peterx@...hat.com>, Alistair Popple <apopple@...dia.com>, Nadav Amit <namit@...are.com>, Yang Shi <shy828301@...il.com>, Vlastimil Babka <vbabka@...e.cz>, Michal Hocko <mhocko@...nel.org>, Mike Kravetz <mike.kravetz@...cle.com>, Andrea Parri <parri.andrea@...il.com>, Will Deacon <will@...nel.org>, Peter Zijlstra <peterz@...radead.org>, "Paul E. McKenney" <paulmck@...nel.org>, "Christoph von Recklinghausen" <crecklin@...hat.com>, Don Dutile <ddutile@...hat.com>, <linux-kernel@...r.kernel.org>, <linux-mm@...ck.org>, <ying.huang@...el.com>, <feng.tang@...el.com>, <zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com> Subject: [linus:master] [mm] 088b8aa537: vm-scalability.throughput -6.5% regression Greeting, FYI, we noticed a -6.5% regression of vm-scalability.throughput due to commit: commit: 088b8aa537c2c767765f1c19b555f21ffe555786 ("mm: fix PageAnonExclusive clearing racing with concurrent RCU GUP-fast") https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master in testcase: vm-scalability on test machine: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz (Cascade Lake) with 128G memory with following parameters: thp_enabled: never thp_defrag: never nr_task: 1 nr_pmem: 2 priority: 1 test: swap-w-seq cpufreq_governor: performance test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us. test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/ Details are as below: ========================================================================================= compiler/cpufreq_governor/kconfig/nr_pmem/nr_task/priority/rootfs/tbox_group/test/testcase/thp_defrag/thp_enabled: gcc-11/performance/x86_64-rhel-8.3/2/1/1/debian-11.1-x86_64-20220510.cgz/lkp-csl-2sp9/swap-w-seq/vm-scalability/never/never commit: e7b72c48d6 ("mm/mremap_pages: save a few cycles in get_dev_pagemap()") 088b8aa537 ("mm: fix PageAnonExclusive clearing racing with concurrent RCU GUP-fast") e7b72c48d677c244 088b8aa537c2c767765f1c19b55 ---------------- --------------------------- %stddev %change %stddev \ | \ 1061123 -6.5% 991719 ± 3% vm-scalability.median 1061123 -6.5% 991719 ± 3% vm-scalability.throughput 106.50 +6.7% 113.64 ± 3% vm-scalability.time.elapsed_time 106.50 +6.7% 113.64 ± 3% vm-scalability.time.elapsed_time.max 95.00 -3.9% 91.33 vm-scalability.time.percent_of_cpu_this_job_got 464.83 ± 45% +77.2% 823.50 ± 32% numa-vmstat.node1.workingset_refault_anon 41828947 ± 3% +29.1% 54019629 ± 2% turbostat.IRQ 46348 ± 3% +36.6% 63322 ± 41% turbostat.POLL 593.33 -3.5% 572.83 ± 2% vmstat.swap.si 626370 -5.7% 590394 ± 3% vmstat.swap.so 272382 ± 4% +14.9% 313058 ± 2% vmstat.system.in 6487093 ± 55% -69.9% 1954992 ± 36% proc-vmstat.compact_migrate_scanned 1184 ± 5% -24.7% 892.17 ± 20% proc-vmstat.kswapd_low_wmark_hit_quickly 96594 -5.3% 91472 ± 2% proc-vmstat.nr_dirty_background_threshold 193425 -5.3% 183168 ± 2% proc-vmstat.nr_dirty_threshold 1004718 -5.1% 953385 proc-vmstat.nr_free_pages 283.33 ± 19% -28.5% 202.67 ± 24% proc-vmstat.nr_inactive_file 282.17 ± 19% -28.4% 202.17 ± 24% proc-vmstat.nr_zone_inactive_file 1504 ± 2% -16.9% 1250 ± 10% proc-vmstat.pageoutrun 3589 ± 13% +115.7% 7743 ± 12% proc-vmstat.pgactivate 22448232 +3.3% 23184909 proc-vmstat.pgalloc_normal 9440 ± 22% +69.2% 15974 proc-vmstat.pgmajfault 19906 ± 3% +4.9% 20882 ± 2% proc-vmstat.pgreuse 9569 ± 22% +69.7% 16238 proc-vmstat.pswpin 1089 ± 5% +38.4% 1507 ± 23% proc-vmstat.workingset_refault_anon 1.77e+09 -4.6% 1.689e+09 ± 2% perf-stat.i.branch-instructions 190.04 +1.4% 192.69 perf-stat.i.cpu-migrations 25.17 ± 10% +14.9 40.02 ± 9% perf-stat.i.iTLB-load-miss-rate% 690930 ± 4% +57.4% 1087333 ± 4% perf-stat.i.iTLB-load-misses 2149984 ± 8% -24.2% 1630299 ± 13% perf-stat.i.iTLB-loads 6.979e+09 -4.4% 6.675e+09 ± 2% perf-stat.i.instructions 10160 ± 4% -36.6% 6443 ± 5% perf-stat.i.instructions-per-iTLB-miss 69.47 ± 31% +78.3% 123.87 ± 16% perf-stat.i.major-faults 223174 -6.0% 209787 ± 3% perf-stat.i.minor-faults 223243 -6.0% 209911 ± 3% perf-stat.i.page-faults 24.44 ± 8% +15.8 40.28 ± 9% perf-stat.overall.iTLB-load-miss-rate% 10119 ± 4% -39.2% 6151 ± 4% perf-stat.overall.instructions-per-iTLB-miss 0.91 ± 5% -9.1% 0.82 ± 4% perf-stat.overall.ipc 57.75 ± 6% +8.2 65.97 ± 5% perf-stat.overall.node-load-miss-rate% 1.753e+09 -4.5% 1.674e+09 ± 2% perf-stat.ps.branch-instructions 188.25 +1.5% 191.00 perf-stat.ps.cpu-migrations 684471 ± 4% +57.5% 1077795 ± 4% perf-stat.ps.iTLB-load-misses 2129593 ± 8% -24.1% 1616077 ± 13% perf-stat.ps.iTLB-loads 6.914e+09 -4.3% 6.617e+09 ± 2% perf-stat.ps.instructions 68.83 ± 31% +78.5% 122.87 ± 16% perf-stat.ps.major-faults 221092 -5.9% 207957 ± 3% perf-stat.ps.minor-faults 221161 -5.9% 208080 ± 3% perf-stat.ps.page-faults 2.62 ± 37% -2.0 0.62 ± 79% perf-profile.calltrace.cycles-pp.try_to_unmap_one.rmap_walk_anon.try_to_unmap.shrink_page_list.shrink_inactive_list 2.82 ± 36% -1.9 0.88 ± 60% perf-profile.calltrace.cycles-pp.rmap_walk_anon.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_lruvec 2.86 ± 35% -1.9 0.92 ± 60% perf-profile.calltrace.cycles-pp.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_lruvec.shrink_node_memcgs 0.00 +2.4 2.36 ± 44% perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list 0.00 +2.4 2.37 ± 44% perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list 0.00 +2.4 2.41 ± 44% perf-profile.calltrace.cycles-pp.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list.shrink_lruvec 0.00 +2.4 2.41 ± 44% perf-profile.calltrace.cycles-pp.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list.shrink_lruvec.shrink_node_memcgs 2.87 ± 35% -1.9 0.97 ± 49% perf-profile.children.cycles-pp.try_to_unmap 2.63 ± 37% -1.9 0.75 ± 47% perf-profile.children.cycles-pp.try_to_unmap_one 0.23 ± 29% +0.1 0.35 ± 29% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm 0.39 ± 18% +0.1 0.52 ± 10% perf-profile.children.cycles-pp.sync_regs 0.07 ± 20% +0.1 0.20 ± 23% perf-profile.children.cycles-pp.llist_reverse_order 0.53 ± 17% +0.1 0.67 ± 12% perf-profile.children.cycles-pp.error_entry 0.20 ± 31% +0.2 0.35 ± 13% perf-profile.children.cycles-pp.flush_tlb_func 0.00 +0.2 0.19 ± 17% perf-profile.children.cycles-pp.native_flush_tlb_local 0.35 ± 25% +0.5 0.88 ± 19% perf-profile.children.cycles-pp.__flush_smp_call_function_queue 0.35 ± 26% +0.5 0.88 ± 19% perf-profile.children.cycles-pp.__sysvec_call_function_single 0.43 ± 25% +0.6 0.99 ± 17% perf-profile.children.cycles-pp.sysvec_call_function_single 0.73 ± 23% +0.8 1.54 ± 14% perf-profile.children.cycles-pp.asm_sysvec_call_function_single 0.00 +2.4 2.41 ± 44% perf-profile.children.cycles-pp.arch_tlbbatch_flush 0.00 +2.4 2.41 ± 44% perf-profile.children.cycles-pp.try_to_unmap_flush_dirty 0.95 ± 11% -0.2 0.75 ± 10% perf-profile.self.cycles-pp.shrink_page_list 0.23 ± 13% -0.0 0.18 ± 17% perf-profile.self.cycles-pp.cpuidle_idle_call 0.10 ± 29% +0.1 0.16 ± 15% perf-profile.self.cycles-pp.__handle_mm_fault 0.38 ± 18% +0.1 0.51 ± 10% perf-profile.self.cycles-pp.sync_regs 0.07 ± 20% +0.1 0.20 ± 23% perf-profile.self.cycles-pp.llist_reverse_order 0.00 +0.2 0.19 ± 18% perf-profile.self.cycles-pp.native_flush_tlb_local 0.09 ± 31% +0.2 0.33 ± 36% perf-profile.self.cycles-pp.__flush_smp_call_function_queue If you fix the issue, kindly add following tag | Reported-by: kernel test robot <yujie.liu@...el.com> | Link: https://lore.kernel.org/oe-lkp/202211211037.2b2e5e1f-yujie.liu@intel.com To reproduce: git clone https://github.com/intel/lkp-tests.git cd lkp-tests sudo bin/lkp install job.yaml # job file is attached in this email bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run sudo bin/lkp run generated-yaml-file # if come across any failure that blocks the test, # please remove ~/.lkp and /lkp dir to run from a clean state. Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://01.org/lkp View attachment "config-6.0.0-rc3-00139-g088b8aa537c2" of type "text/plain" (164374 bytes) View attachment "job-script" of type "text/plain" (8453 bytes) View attachment "job.yaml" of type "text/plain" (5883 bytes) View attachment "reproduce" of type "text/plain" (999 bytes)
Powered by blists - more mailing lists