[<prev] [next>] [day] [month] [year] [list]
Message-ID: <20211007124350.GA5608@xsang-OptiPlex-9020>
Date: Thu, 7 Oct 2021 20:43:50 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Christoph Hellwig <hch@....de>
Cc: Jens Axboe <axboe@...nel.dk>,
Johannes Thumshirn <johannes.thumshirn@....com>,
LKML <linux-kernel@...r.kernel.org>,
Linux Memory Management List <linux-mm@...ck.org>,
lkp@...ts.01.org, lkp@...el.com, ying.huang@...el.com,
feng.tang@...el.com, zhengjun.xing@...ux.intel.com
Subject: [mm] f11826446d: reaim.jobs_per_min 3.9% improvement
Greeting,
FYI, we noticed a 3.9% improvement of reaim.jobs_per_min due to commit:
commit: f11826446de4c3a65d56ad7ad250e4e17fb79d7c ("mm: don't include <linux/blkdev.h> in <linux/backing-dev.h>")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: reaim
on test machine: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
with following parameters:
runtime: 300s
nr_task: 100%
test: new_dbase
cpufreq_governor: performance
ucode: 0x700001e
test-description: REAIM is an updated and improved version of AIM 7 benchmark.
test-url: https://sourceforge.net/projects/re-aim-7/
In addition to that, the commit also has significant impact on the following tests:
+------------------+-------------------------------------------------------------------------------------+
| testcase: change | reaim: reaim.jobs_per_min 1.0% improvement |
| test machine | 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory |
| test parameters | cpufreq_governor=performance |
| | nr_task=100% |
| | runtime=300s |
| | test=dbase |
| | ucode=0x5003006 |
+------------------+-------------------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/100%/debian-10.4-x86_64-20200603.cgz/300s/lkp-cpl-4sp1/new_dbase/reaim/0x700001e
commit:
03acadf5a8 ("mm: don't include <linux/blk-cgroup.h> in <linux/backing-dev.h>")
f11826446d ("mm: don't include <linux/blkdev.h> in <linux/backing-dev.h>")
03acadf5a80307b3 f11826446de4c3a65d56ad7ad25
---------------- ---------------------------
%stddev %change %stddev
\ | \
11.36 ± 2% -21.4% 8.92 reaim.child_systime
900112 +3.9% 935573 reaim.jobs_per_min
6250 +3.9% 6497 reaim.jobs_per_min_child
89.04 +1.2% 90.07 reaim.jti
950116 +2.2% 970847 reaim.max_jobs_per_min
0.99 -3.8% 0.95 reaim.parent_time
10.45 -9.7% 9.44 reaim.std_dev_percent
0.09 -12.9% 0.08 reaim.std_dev_time
2.511e+08 +1.5% 2.549e+08 reaim.time.minor_page_faults
1876 -3.0% 1820 reaim.time.percent_of_cpu_this_job_got
1137 ± 2% -20.2% 907.79 reaim.time.system_time
4514 +1.6% 4587 reaim.time.user_time
1589912 +2.1% 1623661 reaim.time.voluntary_context_switches
1440000 +1.5% 1461600 reaim.workload
5312 ± 5% -20.0% 4250 ± 6% interrupts.CPU63.CAL:Function_call_interrupts
14208 +2.4% 14552 vmstat.system.cs
0.10 -0.0 0.09 mpstat.cpu.all.soft%
6.49 ± 2% -1.1 5.34 mpstat.cpu.all.sys%
229.33 ± 5% +32.6% 304.00 ± 15% slabinfo.biovec-64.active_objs
229.33 ± 5% +32.6% 304.00 ± 15% slabinfo.biovec-64.num_objs
93258 ± 6% +13.0% 105408 ± 6% softirqs.CPU1.RCU
92573 ± 3% +9.1% 100979 ± 5% softirqs.CPU54.RCU
94252 ± 4% +10.0% 103668 ± 6% softirqs.CPU63.RCU
49636 ± 2% +4.1% 51671 proc-vmstat.nr_active_anon
61572 ± 2% +4.0% 64027 proc-vmstat.nr_shmem
49636 ± 2% +4.1% 51671 proc-vmstat.nr_zone_active_anon
2.342e+08 +1.5% 2.378e+08 proc-vmstat.numa_hit
2.345e+08 +1.5% 2.38e+08 proc-vmstat.numa_local
361365 +1.9% 368277 proc-vmstat.pgactivate
2.341e+08 +1.5% 2.376e+08 proc-vmstat.pgalloc_normal
2.528e+08 +1.5% 2.566e+08 proc-vmstat.pgfault
2.34e+08 +1.5% 2.375e+08 proc-vmstat.pgfree
16463378 +1.4% 16693131 proc-vmstat.pgreuse
1.72 ± 5% +17.9% 2.02 ± 7% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.io_schedule.__lock_page_killable.filemap_fault
1.37 ± 15% -38.6% 0.84 ± 31% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.__anon_vma_prepare
150.33 ± 33% -52.8% 70.90 ± 30% perf-sched.sch_delay.max.ms.__sched_text_start.__sched_text_start.do_wait.kernel_wait4.__do_sys_wait4
164.50 ± 32% -54.1% 75.46 ± 16% perf-sched.sch_delay.max.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
25.97 ± 28% -44.2% 14.49 ± 44% perf-sched.sch_delay.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.__anon_vma_prepare
129289 ± 7% +13.4% 146667 perf-sched.total_wait_and_delay.count.ms
5.36 ± 23% +33.1% 7.14 ± 12% perf-sched.wait_and_delay.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
25.79 ± 6% -12.7% 22.52 perf-sched.wait_and_delay.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
102.35 ± 5% -9.3% 92.88 perf-sched.wait_and_delay.avg.ms.__sched_text_start.__sched_text_start.smpboot_thread_fn.kthread.ret_from_fork
29134 ± 8% +11.3% 32428 ± 2% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.do_task_dead.do_exit.do_group_exit
31263 ± 8% +11.7% 34919 ± 2% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.do_wait.kernel_wait4.__do_sys_wait4
11902 ± 9% +18.2% 14064 perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
1755 ± 8% +19.3% 2094 ± 4% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
8019 ± 9% +19.2% 9560 perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
4385 ± 9% +16.3% 5098 ± 2% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.stop_one_cpu
82.67 ± 6% +24.6% 103.00 ± 7% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.schedule_hrtimeout_range_clock.do_sys_poll.__x64_sys_poll
11573 ± 8% +14.7% 13274 perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.smpboot_thread_fn.kthread.ret_from_fork
4.73 ± 26% +36.1% 6.43 ± 12% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
25.49 ± 6% -12.6% 22.27 perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
0.57 ± 31% +1060.6% 6.58 ±180% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.unmap_page_range
13.71 ± 31% -65.7% 4.71 ± 93% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.wp_page_copy
102.35 ± 5% -9.3% 92.87 perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.smpboot_thread_fn.kthread.ret_from_fork
0.43 ± 85% +1876.3% 8.47 ± 58% perf-sched.wait_time.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.__vmalloc_area_node
10.56 ± 41% +1650.8% 184.79 ±199% perf-sched.wait_time.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.unmap_page_range
1.627e+10 +0.9% 1.642e+10 perf-stat.i.branch-instructions
32973172 -8.9% 30048601 perf-stat.i.cache-misses
14218 +2.0% 14501 perf-stat.i.context-switches
6.409e+10 -3.2% 6.205e+10 perf-stat.i.cpu-cycles
2017 +1.0% 2038 perf-stat.i.cpu-migrations
8.326e+09 +0.9% 8.4e+09 perf-stat.i.dTLB-loads
0.06 ± 2% -0.0 0.05 perf-stat.i.dTLB-store-miss-rate%
49.14 +1.8 50.95 perf-stat.i.iTLB-load-miss-rate%
2912410 +1.8% 2965368 perf-stat.i.iTLB-loads
8.246e+10 +0.9% 8.321e+10 perf-stat.i.instructions
0.90 ± 2% +5.4% 0.95 perf-stat.i.ipc
1.00 -2.9% 0.97 perf-stat.i.metric.GHz
419.48 ± 2% -4.1% 402.22 perf-stat.i.metric.K/sec
481.26 +1.2% 486.91 perf-stat.i.metric.M/sec
829642 +0.9% 837144 perf-stat.i.minor-faults
8315387 -7.7% 7674382 perf-stat.i.node-load-misses
72.73 -1.4 71.30 perf-stat.i.node-store-miss-rate%
3045408 -3.6% 2936099 perf-stat.i.node-store-misses
830068 +0.9% 837567 perf-stat.i.page-faults
4.83 -1.4% 4.76 perf-stat.overall.MPKI
8.30 -0.7 7.61 perf-stat.overall.cache-miss-rate%
0.78 -4.1% 0.75 perf-stat.overall.cpi
1938 +6.2% 2058 perf-stat.overall.cycles-between-cache-misses
1.29 +4.3% 1.34 perf-stat.overall.ipc
1.625e+10 +1.2% 1.645e+10 perf-stat.ps.branch-instructions
1.723e+08 +1.3% 1.745e+08 perf-stat.ps.branch-misses
33029764 -8.6% 30195990 perf-stat.ps.cache-misses
14220 +2.3% 14541 perf-stat.ps.context-switches
6.4e+10 -2.9% 6.214e+10 perf-stat.ps.cpu-cycles
2018 +1.3% 2045 perf-stat.ps.cpu-migrations
8.319e+09 +1.2% 8.418e+09 perf-stat.ps.dTLB-loads
3968649 +1.1% 4010723 perf-stat.ps.dTLB-store-misses
5.894e+09 +1.2% 5.968e+09 perf-stat.ps.dTLB-stores
2909823 +2.0% 2969141 perf-stat.ps.iTLB-loads
8.232e+10 +1.2% 8.332e+10 perf-stat.ps.instructions
829765 +1.2% 840028 perf-stat.ps.minor-faults
8327631 -7.4% 7710627 perf-stat.ps.node-load-misses
3049795 -3.3% 2950459 perf-stat.ps.node-store-misses
830191 +1.2% 840454 perf-stat.ps.page-faults
2.486e+13 +1.4% 2.521e+13 perf-stat.total.instructions
4.42 ± 6% -0.9 3.48 ± 8% perf-profile.calltrace.cycles-pp.page_test
3.40 ± 7% -0.9 2.50 ± 10% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.page_test
3.57 ± 7% -0.9 2.68 ± 10% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.page_test
3.58 ± 7% -0.9 2.70 ± 10% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.page_test
3.60 ± 7% -0.9 2.72 ± 10% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.page_test
3.95 ± 3% -0.6 3.39 ± 5% perf-profile.calltrace.cycles-pp.brk
3.13 ± 5% -0.6 2.58 ± 6% perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
3.19 ± 5% -0.6 2.64 ± 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
3.16 ± 5% -0.6 2.61 ± 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
1.59 ± 13% -0.5 1.04 ± 13% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
2.43 ± 6% -0.5 1.92 ± 9% perf-profile.calltrace.cycles-pp.__do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
4.48 ± 6% -0.5 3.97 ± 7% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
2.34 ± 6% -0.5 1.83 ± 9% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.50 ± 7% -0.3 1.15 ± 9% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap
1.78 ± 6% -0.3 1.44 ± 8% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__x64_sys_brk.do_syscall_64
1.77 ± 6% -0.3 1.43 ± 8% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap.__x64_sys_brk
0.76 ± 10% -0.3 0.47 ± 45% perf-profile.calltrace.cycles-pp.charge_memcg.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.99 ± 7% -0.2 0.77 ± 8% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region
0.76 ± 4% -0.1 0.64 ± 6% perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_flush_mmu.tlb_finish_mmu
0.64 ± 4% -0.1 0.57 ± 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_flush_mmu
0.56 ± 2% -0.0 0.53 ± 2% perf-profile.calltrace.cycles-pp.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
0.64 +0.0 0.68 ± 6% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.54 +0.1 0.59 ± 6% perf-profile.calltrace.cycles-pp.filemap_map_pages.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
1.70 ± 2% +0.2 1.95 ± 9% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
1.84 ± 2% +0.3 2.13 ± 8% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
1.85 ± 2% +0.3 2.14 ± 8% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
1.90 ± 2% +0.3 2.20 ± 8% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
5.01 ± 5% -0.9 4.07 ± 7% perf-profile.children.cycles-pp.page_test
6.14 ± 5% -0.7 5.47 ± 7% perf-profile.children.cycles-pp.handle_mm_fault
6.37 ± 4% -0.6 5.77 ± 6% perf-profile.children.cycles-pp.do_user_addr_fault
6.40 ± 4% -0.6 5.81 ± 6% perf-profile.children.cycles-pp.exc_page_fault
4.00 ± 3% -0.6 3.43 ± 5% perf-profile.children.cycles-pp.brk
3.47 ± 8% -0.6 2.90 ± 9% perf-profile.children.cycles-pp.do_anonymous_page
5.41 ± 5% -0.5 4.90 ± 7% perf-profile.children.cycles-pp.__handle_mm_fault
2.02 ± 12% -0.5 1.51 ± 11% perf-profile.children.cycles-pp.__mem_cgroup_charge
1.48 ± 7% -0.4 1.04 ± 13% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
1.39 ± 8% -0.4 0.96 ± 14% perf-profile.children.cycles-pp.__mod_memcg_state
0.98 ± 7% -0.3 0.69 ± 14% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.71 ± 8% -0.2 0.49 ± 14% perf-profile.children.cycles-pp.__count_memcg_events
0.67 ± 13% -0.2 0.48 ± 12% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.97 ± 9% -0.2 0.80 ± 8% perf-profile.children.cycles-pp.charge_memcg
0.47 ± 8% -0.1 0.34 ± 13% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.33 ± 21% -0.1 0.20 ± 19% perf-profile.children.cycles-pp.uncharge_page
0.60 ± 6% -0.1 0.48 ± 8% perf-profile.children.cycles-pp.mod_objcg_state
0.91 ± 4% -0.1 0.82 ± 6% perf-profile.children.cycles-pp.kmem_cache_alloc
0.72 ± 2% -0.1 0.64 ± 7% perf-profile.children.cycles-pp.page_remove_rmap
0.26 ± 10% -0.1 0.20 ± 13% perf-profile.children.cycles-pp.get_mem_cgroup_from_objcg
0.36 ± 3% -0.0 0.33 ± 5% perf-profile.children.cycles-pp.wp_page_copy
0.07 ± 10% -0.0 0.06 ± 13% perf-profile.children.cycles-pp.free_pgd_range
0.07 ± 12% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.free_p4d_range
0.09 ± 7% +0.0 0.11 ± 6% perf-profile.children.cycles-pp.pick_link
0.08 ± 10% +0.0 0.09 ± 13% perf-profile.children.cycles-pp.d_add
0.11 ± 6% +0.0 0.13 ± 4% perf-profile.children.cycles-pp.free_swap_cache
0.20 ± 6% +0.0 0.22 ± 6% perf-profile.children.cycles-pp.__legitimize_path
0.29 ± 5% +0.0 0.32 ± 3% perf-profile.children.cycles-pp._dl_addr
0.22 ± 8% +0.0 0.26 ± 8% perf-profile.children.cycles-pp.up_write
0.95 ± 2% +0.1 1.03 ± 7% perf-profile.children.cycles-pp.filemap_map_pages
0.76 +0.1 0.86 ± 5% perf-profile.children.cycles-pp._raw_spin_lock
1.38 ± 8% -0.4 0.96 ± 14% perf-profile.self.cycles-pp.__mod_memcg_state
0.71 ± 8% -0.2 0.48 ± 14% perf-profile.self.cycles-pp.__count_memcg_events
0.66 ± 13% -0.2 0.47 ± 12% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
0.48 ± 15% -0.2 0.31 ± 13% perf-profile.self.cycles-pp.charge_memcg
0.38 ± 15% -0.1 0.24 ± 19% perf-profile.self.cycles-pp.__mem_cgroup_charge
0.29 ± 22% -0.1 0.18 ± 19% perf-profile.self.cycles-pp.uncharge_page
0.26 ± 10% -0.1 0.18 ± 10% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.26 ± 10% -0.1 0.19 ± 13% perf-profile.self.cycles-pp.get_mem_cgroup_from_objcg
0.15 ± 15% -0.0 0.12 ± 8% perf-profile.self.cycles-pp.obj_cgroup_charge_pages
0.21 ± 5% -0.0 0.18 ± 6% perf-profile.self.cycles-pp.mod_objcg_state
0.09 ± 11% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.uncharge_batch
0.08 ± 8% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.get_obj_cgroup_from_current
0.18 ± 5% +0.0 0.20 ± 6% perf-profile.self.cycles-pp.__handle_mm_fault
0.28 ± 6% +0.0 0.31 ± 5% perf-profile.self.cycles-pp._dl_addr
0.21 ± 9% +0.0 0.25 ± 7% perf-profile.self.cycles-pp.up_write
0.29 ± 2% +0.0 0.33 ± 4% perf-profile.self.cycles-pp.page_remove_rmap
0.42 ± 3% +0.0 0.46 ± 4% perf-profile.self.cycles-pp._raw_spin_lock
reaim.jobs_per_min
950000 +------------------------------------------------------------------+
| O O O O O O O O O O O O OO O O O O O O O O O O O O |
900000 |.+.+.+.+.+.+.+.+.+.+.+.+.+.+.+.+ +.+. .+. .+ |
| : : + + |
| : : |
850000 |-+ : : |
| : : |
800000 |-+ : : |
| : : |
750000 |-+ : : |
| : : |
| : .+ |
700000 |-+ ++ |
| |
650000 +------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
lkp-csl-2ap3: 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/100%/debian-10.4-x86_64-20200603.cgz/300s/lkp-csl-2ap3/dbase/reaim/0x5003006
commit:
03acadf5a8 ("mm: don't include <linux/blk-cgroup.h> in <linux/backing-dev.h>")
f11826446d ("mm: don't include <linux/blkdev.h> in <linux/backing-dev.h>")
03acadf5a80307b3 f11826446de4c3a65d56ad7ad25
---------------- ---------------------------
%stddev %change %stddev
\ | \
11.09 -9.4% 10.04 reaim.child_systime
855151 +1.0% 864077 reaim.jobs_per_min
4453 +1.0% 4500 reaim.jobs_per_min_child
1.33 -1.0% 1.32 reaim.parent_time
11.58 -2.9% 11.24 reaim.std_dev_percent
0.13 -3.8% 0.12 reaim.std_dev_time
150999 -0.9% 149574 reaim.time.major_page_faults
999.98 -9.4% 905.73 reaim.time.system_time
29446 ± 31% -53.7% 13623 ± 48% softirqs.NET_RX
6.62 -0.3 6.31 perf-stat.overall.cache-miss-rate%
0.76 -1.4% 0.75 perf-stat.overall.cpi
2378 +4.5% 2485 perf-stat.overall.cycles-between-cache-misses
1.32 +1.4% 1.33 perf-stat.overall.ipc
0.16 ± 67% +172.0% 0.43 ± 17% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
0.27 ±131% +1018.6% 2.97 ± 65% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.copy_page_to_iter
0.72 ± 34% -44.0% 0.40 ± 50% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.copy_pte_range
0.44 ± 9% +34.0% 0.59 ± 14% perf-sched.sch_delay.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.remove_vma
6.64 ± 31% +142.2% 16.10 ± 42% perf-sched.sch_delay.max.ms.__sched_text_start.__sched_text_start.pipe_read.new_sync_read.vfs_read
1.68 ±132% +466.5% 9.51 ± 40% perf-sched.sch_delay.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.copy_page_to_iter
2.09 ± 41% +199.6% 6.26 ± 35% perf-sched.wait_and_delay.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
1703 ± 2% -12.7% 1487 ± 2% perf-sched.wait_and_delay.count.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
1.93 ± 40% +201.9% 5.83 ± 38% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
2.37 ± 82% +264.4% 8.62 ± 50% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.io_schedule.__lock_page_killable.filemap_fault
0.24 ± 30% +99.5% 0.47 ± 22% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.down_read
43.01 ±107% -81.1% 8.14 ± 53% perf-sched.wait_time.avg.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.shmem_undo_range
12.87 ± 36% +111.7% 27.25 ± 25% perf-sched.wait_time.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.down_read
368.26 ±123% -92.9% 25.96 ± 57% perf-sched.wait_time.max.ms.__sched_text_start.__sched_text_start.preempt_schedule_common.__cond_resched.shmem_undo_range
3.41 ± 5% -0.6 2.78 ± 4% perf-profile.calltrace.cycles-pp.page_test
2.63 ± 6% -0.6 2.02 ± 5% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.page_test
2.60 ± 6% -0.6 2.00 ± 5% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.page_test
2.64 ± 6% -0.6 2.04 ± 5% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.page_test
2.42 ± 7% -0.6 1.82 ± 5% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.page_test
2.54 ± 4% -0.4 2.11 ± 5% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
2.52 ± 4% -0.4 2.09 ± 5% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
2.48 ± 4% -0.4 2.05 ± 5% perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
3.30 ± 3% -0.4 2.88 ± 5% perf-profile.calltrace.cycles-pp.brk
1.78 ± 5% -0.4 1.40 ± 6% perf-profile.calltrace.cycles-pp.__do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
0.95 ± 11% -0.3 0.61 ± 7% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
1.26 ± 4% -0.3 1.00 ± 6% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap.__x64_sys_brk
1.02 ± 5% -0.3 0.76 ± 5% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap
1.27 ± 4% -0.3 1.00 ± 6% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__x64_sys_brk.do_syscall_64
0.57 ± 2% -0.1 0.44 ± 44% perf-profile.calltrace.cycles-pp.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
1.15 +0.2 1.39 ± 15% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
1.20 +0.3 1.45 ± 15% perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
1.60 +0.3 1.88 ± 16% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
1.60 +0.3 1.88 ± 16% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
1.60 +0.3 1.88 ± 16% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
1.63 +0.3 1.91 ± 15% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
4.00 ± 4% -0.6 3.39 ± 5% perf-profile.children.cycles-pp.page_test
3.34 ± 3% -0.4 2.92 ± 5% perf-profile.children.cycles-pp.brk
0.79 ± 11% -0.3 0.48 ± 14% perf-profile.children.cycles-pp.__mod_memcg_state
0.87 ± 10% -0.3 0.56 ± 14% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
1.22 ± 10% -0.3 0.92 ± 14% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.62 ± 9% -0.2 0.44 ± 14% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.39 ± 16% -0.1 0.25 ± 14% perf-profile.children.cycles-pp.lock_page_lruvec_irqsave
0.42 ± 9% -0.1 0.29 ± 20% perf-profile.children.cycles-pp.__count_memcg_events
0.39 ± 11% -0.1 0.27 ± 18% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.42 ± 15% -0.1 0.30 ± 12% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.58 ± 11% -0.1 0.47 ± 11% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
0.17 ± 20% -0.1 0.09 ± 16% perf-profile.children.cycles-pp.uncharge_page
0.36 ± 7% -0.1 0.28 ± 15% perf-profile.children.cycles-pp.mod_objcg_state
0.29 ± 9% -0.1 0.22 ± 15% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.26 ± 7% -0.1 0.19 ± 11% perf-profile.children.cycles-pp.lru_add_drain_cpu
0.25 ± 9% -0.1 0.19 ± 10% perf-profile.children.cycles-pp.lru_add_drain
0.16 ± 7% -0.1 0.10 ± 17% perf-profile.children.cycles-pp.get_mem_cgroup_from_objcg
0.07 ± 6% +0.0 0.09 ± 12% perf-profile.children.cycles-pp.simple_lookup
0.10 ± 9% +0.0 0.12 ± 14% perf-profile.children.cycles-pp.free_swap_cache
0.16 ± 4% +0.0 0.19 ± 11% perf-profile.children.cycles-pp.__d_lookup_rcu
0.30 ± 3% +0.0 0.35 ± 13% perf-profile.children.cycles-pp._dl_addr
0.32 ± 4% +0.1 0.38 ± 12% perf-profile.children.cycles-pp.vma_interval_tree_insert
0.30 ± 4% +0.1 0.35 ± 10% perf-profile.children.cycles-pp.lookup_fast
0.20 ± 5% +0.1 0.26 ± 10% perf-profile.children.cycles-pp.clockevents_program_event
0.22 ± 9% +0.1 0.29 ± 10% perf-profile.children.cycles-pp.ktime_get
0.36 ± 4% +0.1 0.42 ± 11% perf-profile.children.cycles-pp.dput
0.48 ± 3% +0.1 0.56 ± 11% perf-profile.children.cycles-pp.next_uptodate_page
0.46 ± 3% +0.1 0.55 ± 18% perf-profile.children.cycles-pp.intel_idle
0.62 ± 2% +0.1 0.73 ± 13% perf-profile.children.cycles-pp.filename_lookup
0.62 ± 5% +0.1 0.73 ± 13% perf-profile.children.cycles-pp._raw_spin_lock
0.62 ± 2% +0.1 0.72 ± 13% perf-profile.children.cycles-pp.path_lookupat
0.65 +0.1 0.76 ± 12% perf-profile.children.cycles-pp.vfs_statx
0.67 ± 2% +0.1 0.78 ± 13% perf-profile.children.cycles-pp.user_path_at_empty
0.66 ± 2% +0.1 0.77 ± 12% perf-profile.children.cycles-pp.__do_sys_newstat
0.84 ± 2% +0.1 0.96 ± 12% perf-profile.children.cycles-pp.filemap_map_pages
1.01 ± 4% +0.1 1.14 ± 10% perf-profile.children.cycles-pp.link_path_walk
1.07 ± 3% +0.2 1.23 ± 12% perf-profile.children.cycles-pp.walk_component
1.22 +0.3 1.48 ± 14% perf-profile.children.cycles-pp.cpuidle_enter
1.22 +0.3 1.48 ± 14% perf-profile.children.cycles-pp.cpuidle_enter_state
1.60 +0.3 1.88 ± 16% perf-profile.children.cycles-pp.start_secondary
1.63 +0.3 1.91 ± 15% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
1.63 +0.3 1.91 ± 15% perf-profile.children.cycles-pp.cpu_startup_entry
1.63 +0.3 1.91 ± 15% perf-profile.children.cycles-pp.do_idle
0.79 ± 11% -0.3 0.48 ± 14% perf-profile.self.cycles-pp.__mod_memcg_state
0.42 ± 9% -0.1 0.28 ± 19% perf-profile.self.cycles-pp.__count_memcg_events
0.38 ± 12% -0.1 0.26 ± 18% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
0.58 ± 11% -0.1 0.47 ± 11% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.25 ± 17% -0.1 0.15 ± 11% perf-profile.self.cycles-pp.charge_memcg
0.19 ± 16% -0.1 0.10 ± 10% perf-profile.self.cycles-pp.__mem_cgroup_charge
0.15 ± 21% -0.1 0.08 ± 16% perf-profile.self.cycles-pp.uncharge_page
0.16 ± 9% -0.1 0.10 ± 18% perf-profile.self.cycles-pp.get_mem_cgroup_from_objcg
0.09 ± 12% -0.0 0.06 ± 15% perf-profile.self.cycles-pp.obj_cgroup_charge_pages
0.09 ± 6% +0.0 0.11 ± 8% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.07 +0.0 0.09 ± 15% perf-profile.self.cycles-pp.cpuidle_enter_state
0.16 ± 4% +0.0 0.19 ± 11% perf-profile.self.cycles-pp.__d_lookup_rcu
0.02 ±141% +0.0 0.06 ± 13% perf-profile.self.cycles-pp.lockref_put_or_lock
0.30 ± 2% +0.0 0.35 ± 14% perf-profile.self.cycles-pp._dl_addr
0.20 ± 10% +0.1 0.26 ± 11% perf-profile.self.cycles-pp.ktime_get
0.37 ± 4% +0.1 0.45 ± 16% perf-profile.self.cycles-pp._raw_spin_lock
0.46 ± 3% +0.1 0.55 ± 18% perf-profile.self.cycles-pp.intel_idle
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
---
0DAY/LKP+ Test Infrastructure Open Source Technology Center
https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation
Thanks,
Oliver Sang
View attachment "config-5.15.0-rc3-00003-gf11826446de4" of type "text/plain" (169007 bytes)
View attachment "job-script" of type "text/plain" (8045 bytes)
View attachment "job.yaml" of type "text/plain" (5409 bytes)
View attachment "reproduce" of type "text/plain" (10952 bytes)
Powered by blists - more mailing lists