lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <099cfe59-1601-a0fc-3849-13eee8754130@linux.intel.com>
Date:   Wed, 10 Jun 2020 11:21:13 +0800
From:   Xing Zhengjun <zhengjun.xing@...ux.intel.com>
To:     kernel test robot <rong.a.chen@...el.com>,
        Mimi Zohar <zohar@...ux.ibm.com>
Cc:     Lakshmi Ramasubramanian <nramas@...ux.microsoft.com>,
        LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: Re: [LKP] [ima] 8eb613c0b8: stress-ng.icache.ops_per_sec -84.2%
 regression

Hi Mimi,

     Do you have time to take a look at this? we noticed a 3.7% 
regression of boot-time.dhcp and a 84.2% regression of 
stress-ng.icache.ops_per_sec. Thanks.

On 6/3/2020 5:11 PM, kernel test robot wrote:
> Greeting,
> 
> FYI, we noticed a 3.7% regression of boot-time.dhcp due to commit:
> 
> 
> commit: 8eb613c0b8f19627ba1846dcf78bb2c85edbe8dd ("ima: verify mprotect change is consistent with mmap policy")
> https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> 
> in testcase: stress-ng
> on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
> with following parameters:
> 
> 	nr_threads: 100%
> 	disk: 1HDD
> 	testtime: 30s
> 	class: cpu-cache
> 	cpufreq_governor: performance
> 	ucode: 0x500002c
> 
> 
> 
> 
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <rong.a.chen@...el.com>
> 
> 
> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> 
> To reproduce:
> 
>          git clone https://github.com/intel/lkp-tests.git
>          cd lkp-tests
>          bin/lkp install job.yaml  # job file is attached in this email
>          bin/lkp run     job.yaml
> 
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
>    cpu-cache/gcc-9/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/30s/0x500002c
> 
> commit:
>    0c4395fb2a ("evm: Fix possible memory leak in evm_calc_hmac_or_hash()")
>    8eb613c0b8 ("ima: verify mprotect change is consistent with mmap policy")
> 
> 0c4395fb2aa77341 8eb613c0b8f19627ba1846dcf78
> ---------------- ---------------------------
>         fail:runs  %reproduction    fail:runs
>             |             |             |
>             :4           25%           1:4     dmesg.WARNING:at#for_ip_interrupt_entry/0x
>            0:4            3%           0:4     perf-profile.children.cycles-pp.error_entry
>           %stddev     %change         %stddev
>               \          |                \
>     1245570           -84.2%     197151        stress-ng.icache.ops
>       41517           -84.2%       6570        stress-ng.icache.ops_per_sec
>   1.306e+09           -82.1%  2.338e+08        stress-ng.time.minor_page_faults
>        2985           +13.5%       3387        stress-ng.time.system_time
>        4.28           +13.1%       4.85        iostat.cpu.system
>        4.18            +0.6        4.73        mpstat.cpu.all.sys%
>       10121            +9.6%      11096 ±  3%  softirqs.CPU67.SCHED
>      203299            -4.2%     194854 ±  5%  vmstat.system.in
>       26.91            +2.8%      27.67 ±  3%  boot-time.boot
>       16.34            +3.7%      16.94 ±  2%  boot-time.dhcp
>        2183 ±  3%      +3.7%       2263        boot-time.idle
>     1042938 ± 80%   +8208.2%   86649242 ±156%  cpuidle.C1.time
>       48428 ±114%   +1842.4%     940677 ±151%  cpuidle.C1.usage
>       15748 ± 28%    +301.0%      63144 ± 79%  cpuidle.POLL.usage
>       61300 ±  4%     +82.8%     112033 ± 11%  numa-vmstat.node1.nr_active_anon
>       47060 ±  3%    +106.8%      97323 ± 12%  numa-vmstat.node1.nr_anon_pages
>       42.67 ±  2%    +217.0%     135.25 ± 14%  numa-vmstat.node1.nr_anon_transparent_hugepages
>       61301 ±  4%     +82.8%     112032 ± 11%  numa-vmstat.node1.nr_zone_active_anon
>        3816 ±  2%      +3.0%       3931        proc-vmstat.nr_page_table_pages
>    35216541            +2.9%   36244047        proc-vmstat.pgalloc_normal
>   1.308e+09           -82.0%  2.356e+08        proc-vmstat.pgfault
>    35173363            +2.8%   36173843        proc-vmstat.pgfree
>      248171 ±  5%     +82.5%     452893 ± 11%  numa-meminfo.node1.Active
>      244812 ±  4%     +83.5%     449116 ± 11%  numa-meminfo.node1.Active(anon)
>       88290 ±  3%    +214.4%     277591 ± 15%  numa-meminfo.node1.AnonHugePages
>      187940 ±  3%    +107.8%     390486 ± 12%  numa-meminfo.node1.AnonPages
>     1366813 ±  3%     +12.0%    1530428 ±  6%  numa-meminfo.node1.MemUsed
>      571.00 ±  8%     +10.4%     630.50 ±  8%  slabinfo.UDP.active_objs
>      571.00 ±  8%     +10.4%     630.50 ±  8%  slabinfo.UDP.num_objs
>      300.00 ±  5%     +20.0%     360.00 ± 10%  slabinfo.kmem_cache.active_objs
>      300.00 ±  5%     +20.0%     360.00 ± 10%  slabinfo.kmem_cache.num_objs
>      606.33 ±  4%     +17.6%     713.00 ±  8%  slabinfo.kmem_cache_node.active_objs
>      661.33 ±  4%     +16.1%     768.00 ±  8%  slabinfo.kmem_cache_node.num_objs
>      114561 ± 23%     -34.3%      75239 ±  7%  sched_debug.cfs_rq:/.load.max
>       14869 ± 22%     -36.6%       9424 ±  8%  sched_debug.cfs_rq:/.load.stddev
>     4040842 ±  5%     +18.0%    4767515 ± 13%  sched_debug.cpu.avg_idle.max
>     2019061 ±  8%     +25.5%    2534134 ± 14%  sched_debug.cpu.max_idle_balance_cost.max
>      378044 ±  3%     +22.5%     463135 ±  8%  sched_debug.cpu.max_idle_balance_cost.stddev
>       41605           +12.6%      46852 ±  2%  sched_debug.cpu.nr_switches.max
>       44445 ±  2%     +11.8%      49668 ±  3%  sched_debug.cpu.sched_count.max
>        7704 ±  8%     +26.6%       9753 ± 14%  sched_debug.cpu.sched_goidle.max
>      584.16 ±  4%    +135.0%       1372 ± 57%  sched_debug.cpu.sched_goidle.stddev
>        1416 ±  3%     +32.2%       1873 ± 19%  sched_debug.cpu.ttwu_local.stddev
>        0.33 ±141%  +18200.0%      61.00 ±152%  interrupts.40:PCI-MSI.31981573-edge.i40e-eth0-TxRx-4
>       79.67 ±124%     -99.1%       0.75 ±110%  interrupts.89:PCI-MSI.31981622-edge.i40e-eth0-TxRx-53
>        6318           +44.6%       9134 ±  6%  interrupts.CPU1.RES:Rescheduling_interrupts
>        5245 ± 35%     +50.1%       7871        interrupts.CPU11.NMI:Non-maskable_interrupts
>        5245 ± 35%     +50.1%       7871        interrupts.CPU11.PMI:Performance_monitoring_interrupts
>       67.00 ± 58%     -79.9%      13.50 ± 76%  interrupts.CPU16.TLB:TLB_shootdowns
>      861.67           +39.1%       1198 ± 31%  interrupts.CPU23.RES:Rescheduling_interrupts
>      796.33 ±  2%      +9.4%     871.50 ±  2%  interrupts.CPU34.RES:Rescheduling_interrupts
>        0.33 ±141%  +18125.0%      60.75 ±153%  interrupts.CPU4.40:PCI-MSI.31981573-edge.i40e-eth0-TxRx-4
>        1271 ± 43%     +47.2%       1872 ± 26%  interrupts.CPU51.RES:Rescheduling_interrupts
>       79.33 ±125%     -99.4%       0.50 ±173%  interrupts.CPU53.89:PCI-MSI.31981622-edge.i40e-eth0-TxRx-53
>      844.00 ±  2%     +63.7%       1381 ± 34%  interrupts.CPU6.RES:Rescheduling_interrupts
>      786.00 ±  4%     +21.2%     953.00 ± 13%  interrupts.CPU69.RES:Rescheduling_interrupts
>      855.67 ± 11%     +96.3%       1679 ± 60%  interrupts.CPU8.RES:Rescheduling_interrupts
>      178.33 ± 43%     -81.2%      33.50 ± 85%  interrupts.CPU8.TLB:TLB_shootdowns
>      820.67 ±  9%     -18.8%     666.25 ±  9%  interrupts.CPU82.RES:Rescheduling_interrupts
>   4.464e+10            -2.9%  4.336e+10        perf-stat.i.branch-instructions
>        2.65            +0.1        2.73        perf-stat.i.branch-miss-rate%
>        6.32 ±  3%      +5.3%       6.65 ±  2%  perf-stat.i.cpi
>    19226980            -8.5%   17587852        perf-stat.i.dTLB-load-misses
>    5.17e+10            -3.5%  4.989e+10        perf-stat.i.dTLB-loads
>     5874100           -52.1%    2811682        perf-stat.i.dTLB-store-misses
>   1.968e+10            -5.4%  1.861e+10        perf-stat.i.dTLB-stores
>     6448559 ±  3%     -45.3%    3527950 ±  9%  perf-stat.i.iTLB-load-misses
>    31409753 ±  4%     -53.3%   14680095 ±  3%  perf-stat.i.iTLB-loads
>   2.267e+11            -2.9%  2.201e+11        perf-stat.i.instructions
>        0.93            -2.7%       0.91        perf-stat.i.ipc
>        1155            -3.9%       1110        perf-stat.i.metric.M/sec
>     1819587           -82.9%     311898        perf-stat.i.minor-faults
>     1819587           -82.9%     311898        perf-stat.i.page-faults
>        3.49            +3.3%       3.61        perf-stat.overall.MPKI
>        2.26            +0.1        2.32        perf-stat.overall.branch-miss-rate%
>        1.08            +3.5%       1.12        perf-stat.overall.cpi
>        0.04            -0.0        0.04        perf-stat.overall.dTLB-load-miss-rate%
>        0.03            -0.0        0.02        perf-stat.overall.dTLB-store-miss-rate%
>       17.10            +2.4       19.49 ±  6%  perf-stat.overall.iTLB-load-miss-rate%
>       35144 ±  2%     +75.7%      61761 ±  8%  perf-stat.overall.instructions-per-iTLB-miss
>        0.92            -3.3%       0.89        perf-stat.overall.ipc
>   4.286e+10            -3.0%  4.156e+10        perf-stat.ps.branch-instructions
>    18403186            -8.6%   16825390        perf-stat.ps.dTLB-load-misses
>   4.958e+10            -3.6%  4.778e+10        perf-stat.ps.dTLB-loads
>     5647388           -51.3%    2750721        perf-stat.ps.dTLB-store-misses
>   1.886e+10            -5.5%  1.782e+10        perf-stat.ps.dTLB-stores
>     6190595 ±  2%     -44.4%    3440326 ±  9%  perf-stat.ps.iTLB-load-misses
>    30030226 ±  4%     -52.8%   14185074 ±  3%  perf-stat.ps.iTLB-loads
>   2.174e+11            -3.1%  2.107e+11        perf-stat.ps.instructions
>     1737379           -82.0%     312180        perf-stat.ps.minor-faults
>     1737379           -82.0%     312180        perf-stat.ps.page-faults
>   1.635e+14            -3.1%  1.585e+14        perf-stat.total.instructions
>        0.97 ±  3%      -0.3        0.70 ±  9%  perf-profile.calltrace.cycles-pp.__generic_file_write_iter.generic_file_write_iter.new_sync_write.vfs_write.ksys_write
>        0.97 ±  3%      -0.3        0.70 ±  9%  perf-profile.calltrace.cycles-pp.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.new_sync_write.vfs_write
>        0.98 ±  4%      -0.3        0.71 ± 10%  perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>        0.98 ±  4%      -0.3        0.71 ± 10%  perf-profile.calltrace.cycles-pp.generic_file_write_iter.new_sync_write.vfs_write.ksys_write.do_syscall_64
>        0.99 ±  3%      -0.2        0.74 ± 11%  perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.__GI___libc_write
>        1.02 ±  3%      -0.2        0.77 ±  9%  perf-profile.calltrace.cycles-pp.perf_mmap__push.cmd_record.main.__libc_start_main
>        1.00 ±  4%      -0.2        0.76 ± 10%  perf-profile.calltrace.cycles-pp.__GI___libc_write.perf_mmap__push.cmd_record.main.__libc_start_main
>        0.99 ±  3%      -0.2        0.75 ±  9%  perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.__GI___libc_write.perf_mmap__push
>        0.99 ±  3%      -0.2        0.76 ± 10%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__GI___libc_write.perf_mmap__push.cmd_record.main
>        0.99 ±  3%      -0.2        0.76 ± 10%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__GI___libc_write.perf_mmap__push.cmd_record
>        2.99 ±  5%      +0.4        3.37 ±  5%  perf-profile.calltrace.cycles-pp.lapic_next_deadline.clockevents_program_event.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
>        2.06 ±  9%      +0.4        2.45 ± 11%  perf-profile.calltrace.cycles-pp.update_curr.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle
>        2.43 ±  7%      +0.4        2.84 ±  7%  perf-profile.calltrace.cycles-pp.native_write_msr.lapic_next_deadline.clockevents_program_event.hrtimer_interrupt.smp_apic_timer_interrupt
>        0.17 ±141%      +0.4        0.60 ± 13%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.rcu_core.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt
>        0.19 ±141%      +0.5        0.68 ± 14%  perf-profile.calltrace.cycles-pp.reweight_entity.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle
>        1.50 ± 15%      +0.5        2.03 ±  9%  perf-profile.calltrace.cycles-pp.native_write_msr.__intel_pmu_enable_all.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
>        1.57 ± 16%      +0.6        2.12 ±  8%  perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt
>        3.58 ± 15%      +1.0        4.63 ± 12%  perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
>        3.94 ±  3%      +1.7        5.69 ± 28%  perf-profile.calltrace.cycles-pp.hrtimer_active.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle
>       34.96 ±  5%      +3.2       38.16 ±  5%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
>       41.63 ±  5%      +4.8       46.42 ±  6%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
>       19.35 ±  3%      +5.8       25.16 ± 23%  perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
>        0.41 ± 55%      -0.3        0.09 ± 55%  perf-profile.children.cycles-pp.alloc_pages_vma
>        1.22 ±  2%      -0.3        0.91 ± 15%  perf-profile.children.cycles-pp.new_sync_write
>        1.26            -0.3        0.96 ± 13%  perf-profile.children.cycles-pp.ksys_write
>        1.23            -0.3        0.94 ± 15%  perf-profile.children.cycles-pp.vfs_write
>        0.98 ±  4%      -0.3        0.71 ±  8%  perf-profile.children.cycles-pp.__generic_file_write_iter
>        0.98 ±  4%      -0.3        0.71 ±  8%  perf-profile.children.cycles-pp.generic_perform_write
>        0.99 ±  5%      -0.3        0.72 ± 10%  perf-profile.children.cycles-pp.generic_file_write_iter
>        1.02 ±  3%      -0.2        0.77 ±  9%  perf-profile.children.cycles-pp.perf_mmap__push
>        1.00 ±  4%      -0.2        0.76 ± 10%  perf-profile.children.cycles-pp.__GI___libc_write
>        0.16 ± 10%      -0.1        0.07 ± 60%  perf-profile.children.cycles-pp.hrtimer_run_queues
>        0.15 ± 22%      -0.1        0.09 ± 20%  perf-profile.children.cycles-pp.run_posix_cpu_timers
>        0.33 ±  8%      -0.1        0.27 ±  4%  perf-profile.children.cycles-pp.rcu_irq_enter
>        0.04 ± 71%      +0.0        0.07 ±  5%  perf-profile.children.cycles-pp.path_lookupat
>        0.04 ± 71%      +0.0        0.08 ± 10%  perf-profile.children.cycles-pp.filename_lookup
>        0.20 ± 22%      +0.0        0.25 ± 15%  perf-profile.children.cycles-pp.tick_program_event
>        0.20 ± 23%      +0.1        0.26 ± 15%  perf-profile.children.cycles-pp.cmd_stat
>        0.02 ±141%      +0.1        0.08 ± 15%  perf-profile.children.cycles-pp.vsnprintf
>        0.00            +0.1        0.06 ± 14%  perf-profile.children.cycles-pp.lookup_fast
>        0.12 ± 26%      +0.1        0.18 ± 15%  perf-profile.children.cycles-pp.enqueue_entity
>        0.03 ±141%      +0.1        0.09 ± 39%  perf-profile.children.cycles-pp.pagevec_lru_move_fn
>        0.02 ±141%      +0.1        0.09 ±  7%  perf-profile.children.cycles-pp.__x64_sys_mprotect
>        0.02 ±141%      +0.1        0.09 ±  7%  perf-profile.children.cycles-pp.do_mprotect_pkey
>        0.19 ± 38%      +0.1        0.26 ±  7%  perf-profile.children.cycles-pp.update_min_vruntime
>        0.16 ± 29%      +0.1        0.23 ± 16%  perf-profile.children.cycles-pp.iov_iter_fault_in_readable
>        0.19 ±  5%      +0.1        0.27 ± 27%  perf-profile.children.cycles-pp.copy_process
>        0.25 ± 14%      +0.1        0.33 ± 29%  perf-profile.children.cycles-pp.__do_sys_clone
>        0.25 ± 14%      +0.1        0.33 ± 29%  perf-profile.children.cycles-pp._do_fork
>        0.12 ± 13%      +0.1        0.21 ± 17%  perf-profile.children.cycles-pp.activate_task
>        0.14 ± 18%      +0.1        0.23 ± 23%  perf-profile.children.cycles-pp.enqueue_task_fair
>        0.11 ± 29%      +0.1        0.20 ± 41%  perf-profile.children.cycles-pp.rebalance_domains
>        0.04 ± 71%      +0.1        0.13 ± 43%  perf-profile.children.cycles-pp.cpuacct_account_field
>        0.12 ± 22%      +0.1        0.21 ± 16%  perf-profile.children.cycles-pp.ttwu_do_activate
>        0.29 ± 10%      +0.1        0.40 ± 24%  perf-profile.children.cycles-pp.__libc_fork
>        0.00            +0.1        0.11 ± 48%  perf-profile.children.cycles-pp.i40e_handle_lldp_event
>        0.00            +0.1        0.11 ± 48%  perf-profile.children.cycles-pp.i40e_asq_send_command
>        0.33 ±  6%      +0.1        0.44 ± 19%  perf-profile.children.cycles-pp.forkshell
>        0.00            +0.1        0.11 ± 42%  perf-profile.children.cycles-pp.i40e_clean_adminq_subtask
>        0.63            +0.1        0.75 ± 14%  perf-profile.children.cycles-pp._raw_spin_lock_irq
>        0.02 ±141%      +0.1        0.14 ± 62%  perf-profile.children.cycles-pp.__x64_sys_poll
>        0.09 ± 13%      +0.1        0.21 ± 51%  perf-profile.children.cycles-pp.change_prot_numa
>        0.00            +0.1        0.12 ± 43%  perf-profile.children.cycles-pp.i40e_service_task
>        0.57 ± 14%      +0.1        0.70 ± 17%  perf-profile.children.cycles-pp.handle_mm_fault
>        0.02 ±141%      +0.1        0.15 ± 55%  perf-profile.children.cycles-pp.poll
>        0.85            +0.2        1.07 ± 13%  perf-profile.children.cycles-pp.interrupt_entry
>        1.55 ±  5%      +0.3        1.80 ± 11%  perf-profile.children.cycles-pp.read_tsc
>        0.49 ± 24%      +0.3        0.76 ± 15%  perf-profile.children.cycles-pp.reweight_entity
>        0.62 ± 10%      +0.3        0.90 ± 29%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
>        0.27 ± 28%      +0.4        0.63 ± 33%  perf-profile.children.cycles-pp.account_user_time
>        3.00 ±  5%      +0.4        3.38 ±  5%  perf-profile.children.cycles-pp.lapic_next_deadline
>        2.20 ±  8%      +0.4        2.62 ± 10%  perf-profile.children.cycles-pp.update_curr
>        1.58 ± 15%      +0.6        2.13 ±  8%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
>        0.31 ±141%      +0.6        0.90 ± 39%  perf-profile.children.cycles-pp.wake_up_klogd_work_func
>        4.11 ±  9%      +1.0        5.07 ±  7%  perf-profile.children.cycles-pp.native_write_msr
>        3.62 ± 15%      +1.1        4.68 ± 12%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
>        3.95 ±  3%      +1.7        5.70 ± 29%  perf-profile.children.cycles-pp.hrtimer_active
>       34.98 ±  5%      +3.2       38.18 ±  5%  perf-profile.children.cycles-pp.tick_sched_timer
>       15.44            +3.7       19.12 ± 18%  perf-profile.children.cycles-pp.scheduler_tick
>       41.67 ±  5%      +4.8       46.49 ±  6%  perf-profile.children.cycles-pp.__hrtimer_run_queues
>       19.36 ±  3%      +5.8       25.18 ± 23%  perf-profile.children.cycles-pp.update_process_times
>        0.33 ±  8%      -0.1        0.27 ±  7%  perf-profile.self.cycles-pp.rcu_irq_enter
>        0.15 ± 23%      -0.1        0.09 ± 20%  perf-profile.self.cycles-pp.run_posix_cpu_timers
>        0.20 ± 22%      +0.0        0.24 ± 13%  perf-profile.self.cycles-pp.tick_program_event
>        0.10 ± 12%      +0.1        0.17 ± 23%  perf-profile.self.cycles-pp.unmap_page_range
>        0.07 ± 30%      +0.1        0.13 ± 31%  perf-profile.self.cycles-pp.rcu_irq_exit
>        0.00            +0.1        0.06 ± 20%  perf-profile.self.cycles-pp.select_idle_sibling
>        0.18 ± 35%      +0.1        0.26 ±  7%  perf-profile.self.cycles-pp.update_min_vruntime
>        0.04 ± 71%      +0.1        0.13 ± 43%  perf-profile.self.cycles-pp.cpuacct_account_field
>        0.05 ± 78%      +0.1        0.14 ± 49%  perf-profile.self.cycles-pp.change_pte_range
>        0.20 ± 21%      +0.1        0.32 ± 19%  perf-profile.self.cycles-pp.irq_exit
>        0.13 ±  3%      +0.1        0.27 ± 29%  perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
>        0.37 ± 31%      +0.2        0.57 ± 21%  perf-profile.self.cycles-pp.reweight_entity
>        0.84 ±  2%      +0.2        1.07 ± 13%  perf-profile.self.cycles-pp.interrupt_entry
>        0.62 ±  8%      +0.3        0.87 ± 29%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
>        1.49 ±  6%      +0.3        1.75 ±  9%  perf-profile.self.cycles-pp.read_tsc
>        0.22 ± 31%      +0.3        0.50 ± 33%  perf-profile.self.cycles-pp.account_user_time
>        1.47            +0.4        1.91 ± 13%  perf-profile.self.cycles-pp.update_curr
>        4.10 ± 10%      +1.0        5.05 ±  7%  perf-profile.self.cycles-pp.native_write_msr
>        3.93 ±  3%      +1.8        5.70 ± 29%  perf-profile.self.cycles-pp.hrtimer_active
> 
> 
>                                                                                  
>                              stress-ng.time.system_time
>                                                                                  
>    3500 +--------------------------------------------------------------------+
>         |  O  O  O  O O  O  O  O  O     O  O O  O  O  O             O        |
>    3000 |..+..+..+..+.+..+..+..+..+..+..+..+.+..+..+..+..+..+     +.+..+     |
>         |                                                   :     :          |
>    2500 |-+                                                  :    :          |
>         |                                                    :   :           |
>    2000 |-+                                                  :   :           |
>         |                                                    :   :           |
>    1500 |-+                                                   :  :           |
>         |                                                     : :            |
>    1000 |-+                                                   : :            |
>         |                                                     : :            |
>     500 |-+                                                    ::            |
>         |                                                      :             |
>       0 +--------------------------------------------------------------------+
>                                                                                  
>                                                                                                                                                                  
>                             stress-ng.time.minor_page_faults
>                                                                                  
>    1.4e+09 +-----------------------------------------------------------------+
>            |..+..+.+..+..+..+.+..+..+..+.+..+..+..+.+..+..+..+    +..+..+    |
>    1.2e+09 |-+                                               :    :          |
>            |                                                 :    :          |
>      1e+09 |-+                                               :   :           |
>            |                                                  :  :           |
>      8e+08 |-+                                                :  :           |
>            |                                                  :  :           |
>      6e+08 |-+                                                : :            |
>            |                                                  : :            |
>      4e+08 |-+                                                : :            |
>            |                                                   ::            |
>      2e+08 |-+O  O O  O  O  O O  O  O    O  O  O  O O  O  O  O :  O  O  O O  |
>            |                                                   :             |
>          0 +-----------------------------------------------------------------+
>                                                                                  
>                                                                                                                                                                  
>                                   stress-ng.icache.ops
>                                                                                  
>    1.4e+06 +-----------------------------------------------------------------+
>            |                 .+..                                            |
>    1.2e+06 |..+..+.+..+..+..+    +..+..+.+..+..+..+.+..+..+..+    +..+..+    |
>            |                                                 :    :          |
>      1e+06 |-+                                               :    :          |
>            |                                                  :  :           |
>     800000 |-+                                                :  :           |
>            |                                                  :  :           |
>     600000 |-+                                                :  :           |
>            |                                                  : :            |
>     400000 |-+                                                : :            |
>            |                                                   ::            |
>     200000 |-+O  O O  O  O  O O  O  O    O  O  O  O O  O  O  O :: O  O  O O  |
>            |                                                   :             |
>          0 +-----------------------------------------------------------------+
>                                                                                  
>                                                                                                                                                                  
>                              stress-ng.icache.ops_per_sec
>                                                                                  
>    45000 +-------------------------------------------------------------------+
>          |..+..+..+.+..+..+..+..+..+.+..+..+..+..+..+.+..+..+     +..+.+     |
>    40000 |-+                                                :     :          |
>    35000 |-+                                                :     :          |
>          |                                                   :   :           |
>    30000 |-+                                                 :   :           |
>    25000 |-+                                                 :   :           |
>          |                                                   :   :           |
>    20000 |-+                                                  : :            |
>    15000 |-+                                                  : :            |
>          |                                                    : :            |
>    10000 |-+                                                  : :            |
>     5000 |-+O  O  O O  O  O  O  O  O    O  O  O  O  O O  O  O  :  O  O O  O  |
>          |                                                     :             |
>        0 +-------------------------------------------------------------------+
>                                                                                  
>                                                                                  
> [*] bisect-good sample
> [O] bisect-bad  sample
> 
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Rong Chen
> 
> 
> _______________________________________________
> LKP mailing list -- lkp@...ts.01.org
> To unsubscribe send an email to lkp-leave@...ts.01.org
> 

-- 
Zhengjun Xing

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ