[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202211211037.2b2e5e1f-yujie.liu@intel.com>
Date: Mon, 21 Nov 2022 11:03:08 +0800
From: kernel test robot <yujie.liu@...el.com>
To: David Hildenbrand <david@...hat.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Jason Gunthorpe <jgg@...dia.com>,
John Hubbard <jhubbard@...dia.com>,
Andrea Arcangeli <aarcange@...hat.com>,
Hugh Dickins <hughd@...gle.com>, Peter Xu <peterx@...hat.com>,
Alistair Popple <apopple@...dia.com>,
Nadav Amit <namit@...are.com>, Yang Shi <shy828301@...il.com>,
Vlastimil Babka <vbabka@...e.cz>,
Michal Hocko <mhocko@...nel.org>,
Mike Kravetz <mike.kravetz@...cle.com>,
Andrea Parri <parri.andrea@...il.com>,
Will Deacon <will@...nel.org>,
Peter Zijlstra <peterz@...radead.org>,
"Paul E. McKenney" <paulmck@...nel.org>,
"Christoph von Recklinghausen" <crecklin@...hat.com>,
Don Dutile <ddutile@...hat.com>,
<linux-kernel@...r.kernel.org>, <linux-mm@...ck.org>,
<ying.huang@...el.com>, <feng.tang@...el.com>,
<zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>
Subject: [linus:master] [mm] 088b8aa537: vm-scalability.throughput -6.5%
regression
Greeting,
FYI, we noticed a -6.5% regression of vm-scalability.throughput due to commit:
commit: 088b8aa537c2c767765f1c19b555f21ffe555786 ("mm: fix PageAnonExclusive clearing racing with concurrent RCU GUP-fast")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
in testcase: vm-scalability
on test machine: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz (Cascade Lake) with 128G memory
with following parameters:
thp_enabled: never
thp_defrag: never
nr_task: 1
nr_pmem: 2
priority: 1
test: swap-w-seq
cpufreq_governor: performance
test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
Details are as below:
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_pmem/nr_task/priority/rootfs/tbox_group/test/testcase/thp_defrag/thp_enabled:
gcc-11/performance/x86_64-rhel-8.3/2/1/1/debian-11.1-x86_64-20220510.cgz/lkp-csl-2sp9/swap-w-seq/vm-scalability/never/never
commit:
e7b72c48d6 ("mm/mremap_pages: save a few cycles in get_dev_pagemap()")
088b8aa537 ("mm: fix PageAnonExclusive clearing racing with concurrent RCU GUP-fast")
e7b72c48d677c244 088b8aa537c2c767765f1c19b55
---------------- ---------------------------
%stddev %change %stddev
\ | \
1061123 -6.5% 991719 ± 3% vm-scalability.median
1061123 -6.5% 991719 ± 3% vm-scalability.throughput
106.50 +6.7% 113.64 ± 3% vm-scalability.time.elapsed_time
106.50 +6.7% 113.64 ± 3% vm-scalability.time.elapsed_time.max
95.00 -3.9% 91.33 vm-scalability.time.percent_of_cpu_this_job_got
464.83 ± 45% +77.2% 823.50 ± 32% numa-vmstat.node1.workingset_refault_anon
41828947 ± 3% +29.1% 54019629 ± 2% turbostat.IRQ
46348 ± 3% +36.6% 63322 ± 41% turbostat.POLL
593.33 -3.5% 572.83 ± 2% vmstat.swap.si
626370 -5.7% 590394 ± 3% vmstat.swap.so
272382 ± 4% +14.9% 313058 ± 2% vmstat.system.in
6487093 ± 55% -69.9% 1954992 ± 36% proc-vmstat.compact_migrate_scanned
1184 ± 5% -24.7% 892.17 ± 20% proc-vmstat.kswapd_low_wmark_hit_quickly
96594 -5.3% 91472 ± 2% proc-vmstat.nr_dirty_background_threshold
193425 -5.3% 183168 ± 2% proc-vmstat.nr_dirty_threshold
1004718 -5.1% 953385 proc-vmstat.nr_free_pages
283.33 ± 19% -28.5% 202.67 ± 24% proc-vmstat.nr_inactive_file
282.17 ± 19% -28.4% 202.17 ± 24% proc-vmstat.nr_zone_inactive_file
1504 ± 2% -16.9% 1250 ± 10% proc-vmstat.pageoutrun
3589 ± 13% +115.7% 7743 ± 12% proc-vmstat.pgactivate
22448232 +3.3% 23184909 proc-vmstat.pgalloc_normal
9440 ± 22% +69.2% 15974 proc-vmstat.pgmajfault
19906 ± 3% +4.9% 20882 ± 2% proc-vmstat.pgreuse
9569 ± 22% +69.7% 16238 proc-vmstat.pswpin
1089 ± 5% +38.4% 1507 ± 23% proc-vmstat.workingset_refault_anon
1.77e+09 -4.6% 1.689e+09 ± 2% perf-stat.i.branch-instructions
190.04 +1.4% 192.69 perf-stat.i.cpu-migrations
25.17 ± 10% +14.9 40.02 ± 9% perf-stat.i.iTLB-load-miss-rate%
690930 ± 4% +57.4% 1087333 ± 4% perf-stat.i.iTLB-load-misses
2149984 ± 8% -24.2% 1630299 ± 13% perf-stat.i.iTLB-loads
6.979e+09 -4.4% 6.675e+09 ± 2% perf-stat.i.instructions
10160 ± 4% -36.6% 6443 ± 5% perf-stat.i.instructions-per-iTLB-miss
69.47 ± 31% +78.3% 123.87 ± 16% perf-stat.i.major-faults
223174 -6.0% 209787 ± 3% perf-stat.i.minor-faults
223243 -6.0% 209911 ± 3% perf-stat.i.page-faults
24.44 ± 8% +15.8 40.28 ± 9% perf-stat.overall.iTLB-load-miss-rate%
10119 ± 4% -39.2% 6151 ± 4% perf-stat.overall.instructions-per-iTLB-miss
0.91 ± 5% -9.1% 0.82 ± 4% perf-stat.overall.ipc
57.75 ± 6% +8.2 65.97 ± 5% perf-stat.overall.node-load-miss-rate%
1.753e+09 -4.5% 1.674e+09 ± 2% perf-stat.ps.branch-instructions
188.25 +1.5% 191.00 perf-stat.ps.cpu-migrations
684471 ± 4% +57.5% 1077795 ± 4% perf-stat.ps.iTLB-load-misses
2129593 ± 8% -24.1% 1616077 ± 13% perf-stat.ps.iTLB-loads
6.914e+09 -4.3% 6.617e+09 ± 2% perf-stat.ps.instructions
68.83 ± 31% +78.5% 122.87 ± 16% perf-stat.ps.major-faults
221092 -5.9% 207957 ± 3% perf-stat.ps.minor-faults
221161 -5.9% 208080 ± 3% perf-stat.ps.page-faults
2.62 ± 37% -2.0 0.62 ± 79% perf-profile.calltrace.cycles-pp.try_to_unmap_one.rmap_walk_anon.try_to_unmap.shrink_page_list.shrink_inactive_list
2.82 ± 36% -1.9 0.88 ± 60% perf-profile.calltrace.cycles-pp.rmap_walk_anon.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_lruvec
2.86 ± 35% -1.9 0.92 ± 60% perf-profile.calltrace.cycles-pp.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_lruvec.shrink_node_memcgs
0.00 +2.4 2.36 ± 44% perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list
0.00 +2.4 2.37 ± 44% perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list
0.00 +2.4 2.41 ± 44% perf-profile.calltrace.cycles-pp.arch_tlbbatch_flush.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list.shrink_lruvec
0.00 +2.4 2.41 ± 44% perf-profile.calltrace.cycles-pp.try_to_unmap_flush_dirty.shrink_page_list.shrink_inactive_list.shrink_lruvec.shrink_node_memcgs
2.87 ± 35% -1.9 0.97 ± 49% perf-profile.children.cycles-pp.try_to_unmap
2.63 ± 37% -1.9 0.75 ± 47% perf-profile.children.cycles-pp.try_to_unmap_one
0.23 ± 29% +0.1 0.35 ± 29% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.39 ± 18% +0.1 0.52 ± 10% perf-profile.children.cycles-pp.sync_regs
0.07 ± 20% +0.1 0.20 ± 23% perf-profile.children.cycles-pp.llist_reverse_order
0.53 ± 17% +0.1 0.67 ± 12% perf-profile.children.cycles-pp.error_entry
0.20 ± 31% +0.2 0.35 ± 13% perf-profile.children.cycles-pp.flush_tlb_func
0.00 +0.2 0.19 ± 17% perf-profile.children.cycles-pp.native_flush_tlb_local
0.35 ± 25% +0.5 0.88 ± 19% perf-profile.children.cycles-pp.__flush_smp_call_function_queue
0.35 ± 26% +0.5 0.88 ± 19% perf-profile.children.cycles-pp.__sysvec_call_function_single
0.43 ± 25% +0.6 0.99 ± 17% perf-profile.children.cycles-pp.sysvec_call_function_single
0.73 ± 23% +0.8 1.54 ± 14% perf-profile.children.cycles-pp.asm_sysvec_call_function_single
0.00 +2.4 2.41 ± 44% perf-profile.children.cycles-pp.arch_tlbbatch_flush
0.00 +2.4 2.41 ± 44% perf-profile.children.cycles-pp.try_to_unmap_flush_dirty
0.95 ± 11% -0.2 0.75 ± 10% perf-profile.self.cycles-pp.shrink_page_list
0.23 ± 13% -0.0 0.18 ± 17% perf-profile.self.cycles-pp.cpuidle_idle_call
0.10 ± 29% +0.1 0.16 ± 15% perf-profile.self.cycles-pp.__handle_mm_fault
0.38 ± 18% +0.1 0.51 ± 10% perf-profile.self.cycles-pp.sync_regs
0.07 ± 20% +0.1 0.20 ± 23% perf-profile.self.cycles-pp.llist_reverse_order
0.00 +0.2 0.19 ± 18% perf-profile.self.cycles-pp.native_flush_tlb_local
0.09 ± 31% +0.2 0.33 ± 36% perf-profile.self.cycles-pp.__flush_smp_call_function_queue
If you fix the issue, kindly add following tag
| Reported-by: kernel test robot <yujie.liu@...el.com>
| Link: https://lore.kernel.org/oe-lkp/202211211037.2b2e5e1f-yujie.liu@intel.com
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://01.org/lkp
View attachment "config-6.0.0-rc3-00139-g088b8aa537c2" of type "text/plain" (164374 bytes)
View attachment "job-script" of type "text/plain" (8453 bytes)
View attachment "job.yaml" of type "text/plain" (5883 bytes)
View attachment "reproduce" of type "text/plain" (999 bytes)
Powered by blists - more mailing lists