lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Fri, 12 Apr 2019 22:43:57 +0800 From: kernel test robot <rong.a.chen@...el.com> To: Eric Dumazet <edumazet@...gle.com> Cc: "David S . Miller" <davem@...emloft.net>, netdev <netdev@...r.kernel.org>, Eric Dumazet <edumazet@...gle.com>, Eric Dumazet <eric.dumazet@...il.com>, Soheil Hassas Yeganeh <soheil@...gle.com>, Willem de Bruijn <willemb@...gle.com>, lkp@...org Subject: [tcp] 01b4c2aab8: lmbench3.TCP.socket.bandwidth.10MB.MB/sec -20.2% regression Greeting, FYI, we noticed a -20.2% regression of lmbench3.TCP.socket.bandwidth.10MB.MB/sec due to commit: commit: 01b4c2aab841d7ed9c5457371785070b2e0b53b1 ("[PATCH v3 net-next 3/3] tcp: add one skb cache for rx") url: https://github.com/0day-ci/linux/commits/Eric-Dumazet/tcp-add-rx-tx-cache-to-reduce-lock-contention/20190323-091742 in testcase: lmbench3 on test machine: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory with following parameters: test_memory_size: 50% nr_threads: 100% mode: development test: TCP cpufreq_governor: performance ucode: 0xb00002e test-url: http://www.bitmover.com/lmbench/ Details are as below: --------------------------------------------------------------------------------------------------> To reproduce: git clone https://github.com/intel/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_threads/rootfs/tbox_group/test/test_memory_size/testcase/ucode: gcc-7/performance/x86_64-rhel-7.6/development/100%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep4/TCP/50%/lmbench3/0xb00002e commit: af0b648e98 ("tcp: add one skb cache for tx") 01b4c2aab8 ("tcp: add one skb cache for rx") af0b648e98a72a54 01b4c2aab841d7ed9c545737178 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | :4 25% 1:4 dmesg.WARNING:at_ip__netif_receive_skb_core/0x :4 25% 1:4 dmesg.WARNING:at_ip_do_select/0x 1:4 -25% :4 dmesg.WARNING:at_ip_ip_finish_output2/0x %stddev %change %stddev \ | \ 30.40 ± 5% -14.1% 26.11 ± 5% lmbench3.TCP.localhost.latency 99117 -20.2% 79133 lmbench3.TCP.socket.bandwidth.10MB.MB/sec 2537 -2.2% 2481 lmbench3.TCP.socket.bandwidth.64B.MB/sec 157430 -1.7% 154819 lmbench3.time.minor_page_faults 3593 +3.1% 3705 lmbench3.time.percent_of_cpu_this_job_got 22.28 ± 5% -3.0 19.29 ± 2% mpstat.cpu.all.idle% 6.19 ± 2% +1.2 7.40 ± 2% mpstat.cpu.all.soft% 508795 ± 2% +22.0% 620794 ± 2% numa-meminfo.node0.Unevictable 516356 ± 2% +18.1% 609977 ± 2% numa-meminfo.node1.Unevictable 1137258 +14.6% 1303104 ± 2% meminfo.Cached 6680579 ± 6% -12.9% 5821954 meminfo.DirectMap2M 298947 ± 3% +24.1% 371141 ± 5% meminfo.DirectMap4k 1025152 +20.1% 1230771 meminfo.Unevictable 379964 ± 75% -141.4% -157480 sched_debug.cfs_rq:/.spread0.avg -392190 +179.0% -1094014 sched_debug.cfs_rq:/.spread0.min 4051 ± 35% +45.1% 5879 ± 14% sched_debug.cpu.load.min 9511408 ± 8% -18.7% 7730274 ± 3% sched_debug.cpu.nr_switches.max 541.00 ± 5% +17.7% 637.00 ± 8% slabinfo.kmem_cache_node.active_objs 592.00 ± 4% +16.2% 688.00 ± 7% slabinfo.kmem_cache_node.num_objs 5185 ± 3% +38.9% 7201 ± 6% slabinfo.skbuff_fclone_cache.active_objs 5187 ± 3% +38.9% 7206 ± 6% slabinfo.skbuff_fclone_cache.num_objs 21.75 ± 5% -12.6% 19.00 ± 3% vmstat.cpu.id 68.75 +4.0% 71.50 vmstat.cpu.sy 1206841 +13.7% 1372779 ± 2% vmstat.memory.cache 1846679 ± 2% -6.3% 1731096 ± 5% vmstat.system.cs 1.512e+08 ± 2% +13.2% 1.711e+08 ± 3% numa-numastat.node0.local_node 1.512e+08 ± 2% +13.2% 1.711e+08 ± 3% numa-numastat.node0.numa_hit 7091 ±173% +201.1% 21353 ± 57% numa-numastat.node0.other_node 1.48e+08 ± 2% +13.2% 1.674e+08 ± 3% numa-numastat.node1.local_node 1.48e+08 ± 2% +13.1% 1.674e+08 ± 3% numa-numastat.node1.numa_hit 2087 +3.6% 2163 turbostat.Avg_MHz 89315160 ± 13% -50.0% 44625119 ± 19% turbostat.C1 0.94 ± 13% -0.5 0.46 ± 18% turbostat.C1% 7386178 ± 49% -44.9% 4072935 ± 5% turbostat.C1E 0.39 ±146% -0.3 0.04 ± 10% turbostat.C1E% 5.109e+08 ± 15% -50.4% 2.534e+08 ± 22% cpuidle.C1.time 89317656 ± 13% -50.0% 44626550 ± 19% cpuidle.C1.usage 2.092e+08 ±145% -88.4% 24217866 ± 3% cpuidle.C1E.time 7389747 ± 49% -44.9% 4075034 ± 5% cpuidle.C1E.usage 75351513 ± 5% -30.3% 52528863 ± 16% cpuidle.POLL.time 29827485 ± 4% -27.6% 21596769 ± 16% cpuidle.POLL.usage 127198 ± 2% +22.0% 155198 ± 2% numa-vmstat.node0.nr_unevictable 127198 ± 2% +22.0% 155198 ± 2% numa-vmstat.node0.nr_zone_unevictable 25297173 ± 3% +15.2% 29129943 ± 3% numa-vmstat.node0.numa_hit 147405 +17.2% 172790 numa-vmstat.node0.numa_interleave 25289758 ± 2% +15.1% 29108245 ± 3% numa-vmstat.node0.numa_local 129088 ± 2% +18.1% 152494 ± 2% numa-vmstat.node1.nr_unevictable 129088 ± 2% +18.1% 152494 ± 2% numa-vmstat.node1.nr_zone_unevictable 24817577 ± 2% +14.5% 28404348 ± 3% numa-vmstat.node1.numa_hit 147161 +17.7% 173180 numa-vmstat.node1.numa_interleave 24646043 ± 2% +14.5% 28221056 ± 3% numa-vmstat.node1.numa_local 152712 +6.6% 162773 proc-vmstat.nr_anon_pages 241.25 +5.9% 255.50 proc-vmstat.nr_anon_transparent_hugepages 284312 +14.6% 325770 ± 2% proc-vmstat.nr_file_pages 6218 -2.9% 6039 proc-vmstat.nr_mapped 2354 +3.9% 2447 proc-vmstat.nr_page_table_pages 256287 +20.1% 307692 proc-vmstat.nr_unevictable 256287 +20.1% 307692 proc-vmstat.nr_zone_unevictable 2.988e+08 ± 2% +12.8% 3.369e+08 ± 3% proc-vmstat.numa_hit 2.988e+08 ± 2% +12.8% 3.369e+08 ± 3% proc-vmstat.numa_local 2.393e+09 ± 2% +13.0% 2.704e+09 ± 3% proc-vmstat.pgalloc_normal 2.393e+09 ± 2% +13.0% 2.703e+09 ± 3% proc-vmstat.pgfree 36.21 ± 2% -10.1% 32.54 ± 3% perf-stat.i.MPKI 2.203e+10 +3.2% 2.274e+10 perf-stat.i.branch-instructions 1.234e+08 ± 2% -6.8% 1.149e+08 perf-stat.i.cache-misses 1853039 ± 2% -6.3% 1736720 ± 5% perf-stat.i.context-switches 1.831e+11 +3.6% 1.896e+11 perf-stat.i.cpu-cycles 177350 ± 8% -39.9% 106565 ± 12% perf-stat.i.cpu-migrations 52299 ± 9% -15.0% 44447 ± 5% perf-stat.i.cycles-between-cache-misses 0.13 ± 10% -0.0 0.10 ± 6% perf-stat.i.dTLB-load-miss-rate% 24601513 ± 2% -28.8% 17506777 ± 7% perf-stat.i.dTLB-load-misses 3.49e+10 +2.5% 3.578e+10 perf-stat.i.dTLB-loads 0.05 ± 8% +0.0 0.06 ± 6% perf-stat.i.dTLB-store-miss-rate% 54414059 +11.6% 60703047 ± 5% perf-stat.i.iTLB-load-misses 13295971 ± 3% -11.2% 11805882 ± 4% perf-stat.i.iTLB-loads 1.125e+11 +3.0% 1.159e+11 perf-stat.i.instructions 3843 ± 4% +58.0% 6073 perf-stat.i.instructions-per-iTLB-miss 82.77 -2.9 79.87 perf-stat.i.node-load-miss-rate% 61535201 ± 2% -16.8% 51169722 perf-stat.i.node-loads 52.20 ± 3% -4.9 47.30 ± 4% perf-stat.i.node-store-miss-rate% 1348075 ± 6% -24.9% 1012041 ± 3% perf-stat.i.node-store-misses 1002686 ± 6% +26.7% 1269969 perf-stat.i.node-stores 14.36 -3.1% 13.93 perf-stat.overall.MPKI 2.15 -0.0 2.10 perf-stat.overall.branch-miss-rate% 7.65 ± 2% -0.5 7.14 perf-stat.overall.cache-miss-rate% 1481 ± 2% +11.2% 1646 ± 2% perf-stat.overall.cycles-between-cache-misses 0.07 ± 4% -0.0 0.05 ± 7% perf-stat.overall.dTLB-load-miss-rate% 80.36 +3.3 83.69 perf-stat.overall.iTLB-load-miss-rate% 2067 -7.5% 1913 ± 4% perf-stat.overall.instructions-per-iTLB-miss 57.31 ± 4% -13.0 44.31 perf-stat.overall.node-store-miss-rate% 2.198e+10 +3.2% 2.269e+10 perf-stat.ps.branch-instructions 1.234e+08 ± 2% -6.8% 1.15e+08 perf-stat.ps.cache-misses 1849729 ± 2% -6.3% 1733566 ± 5% perf-stat.ps.context-switches 1.827e+11 +3.6% 1.893e+11 perf-stat.ps.cpu-cycles 176984 ± 8% -39.9% 106335 ± 12% perf-stat.ps.cpu-migrations 24552688 ± 2% -28.8% 17472051 ± 7% perf-stat.ps.dTLB-load-misses 3.482e+10 +2.5% 3.57e+10 perf-stat.ps.dTLB-loads 54282300 +11.6% 60558515 ± 5% perf-stat.ps.iTLB-load-misses 13269959 ± 3% -11.2% 11782445 ± 4% perf-stat.ps.iTLB-loads 1.123e+11 +3.0% 1.156e+11 perf-stat.ps.instructions 61567371 ± 2% -16.8% 51205094 perf-stat.ps.node-loads 1345363 ± 6% -24.9% 1010000 ± 3% perf-stat.ps.node-store-misses 1001013 ± 6% +26.8% 1268788 perf-stat.ps.node-stores 31.55 ± 4% -15.5 16.09 ± 76% perf-profile.calltrace.cycles-pp.ip_finish_output2.ip_output.__ip_queue_xmit.__tcp_transmit_skb.tcp_write_xmit 28.16 ± 6% -14.1 14.10 ± 76% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.ip_output.__ip_queue_xmit.__tcp_transmit_skb 28.00 ± 6% -14.0 14.02 ± 76% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.ip_output.__ip_queue_xmit 27.74 ± 6% -13.9 13.86 ± 76% perf-profile.calltrace.cycles-pp.do_softirq_own_stack.do_softirq.__local_bh_enable_ip.ip_finish_output2.ip_output 27.62 ± 6% -13.8 13.79 ± 76% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq_own_stack.do_softirq.__local_bh_enable_ip.ip_finish_output2 27.24 ± 6% -13.7 13.53 ± 76% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq_own_stack.do_softirq.__local_bh_enable_ip 26.70 ± 7% -13.5 13.25 ± 76% perf-profile.calltrace.cycles-pp.process_backlog.net_rx_action.__softirqentry_text_start.do_softirq_own_stack.do_softirq 26.05 ± 7% -13.1 12.93 ± 76% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.net_rx_action.__softirqentry_text_start.do_softirq_own_stack 25.54 ± 8% -12.9 12.61 ± 76% perf-profile.calltrace.cycles-pp.ip_rcv.__netif_receive_skb_one_core.process_backlog.net_rx_action.__softirqentry_text_start 24.81 ± 8% -12.7 12.14 ± 76% perf-profile.calltrace.cycles-pp.ip_local_deliver.ip_rcv.__netif_receive_skb_one_core.process_backlog.net_rx_action 24.69 ± 9% -12.6 12.07 ± 76% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.ip_local_deliver.ip_rcv.__netif_receive_skb_one_core.process_backlog 24.63 ± 9% -12.6 12.04 ± 76% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.ip_local_deliver.ip_rcv.__netif_receive_skb_one_core 24.34 ± 9% -12.4 11.93 ± 76% perf-profile.calltrace.cycles-pp.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.ip_local_deliver.ip_rcv 21.18 ± 12% -11.8 9.35 ± 76% perf-profile.calltrace.cycles-pp.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.ip_local_deliver 20.79 ± 12% -11.6 9.19 ± 76% perf-profile.calltrace.cycles-pp.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish 15.49 ± 20% -9.4 6.12 ± 77% perf-profile.calltrace.cycles-pp.sock_def_readable.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu 15.14 ± 20% -9.2 5.94 ± 77% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sock_def_readable.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv 14.75 ± 20% -9.0 5.77 ± 77% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sock_def_readable.tcp_rcv_established.tcp_v4_do_rcv 14.40 ± 21% -8.8 5.61 ± 77% perf-profile.calltrace.cycles-pp.try_to_wake_up.__wake_up_common.__wake_up_common_lock.sock_def_readable.tcp_rcv_established 18.83 ± 6% -7.6 11.23 ± 31% perf-profile.calltrace.cycles-pp.tcp_recvmsg.inet_recvmsg.sock_read_iter.new_sync_read.vfs_read 19.13 ± 6% -7.5 11.60 ± 30% perf-profile.calltrace.cycles-pp.inet_recvmsg.sock_read_iter.new_sync_read.vfs_read.ksys_read 22.17 ± 4% -7.5 14.67 ± 38% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 22.53 ± 4% -7.4 15.13 ± 38% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 20.30 ± 5% -7.3 13.03 ± 28% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 19.89 ± 5% -7.3 12.63 ± 28% perf-profile.calltrace.cycles-pp.sock_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64 12.67 ± 11% -7.1 5.60 ± 78% perf-profile.calltrace.cycles-pp.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_read_iter.new_sync_read 11.16 ± 12% -6.3 4.90 ± 78% perf-profile.calltrace.cycles-pp.wait_woken.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_read_iter 10.69 ± 12% -6.0 4.70 ± 78% perf-profile.calltrace.cycles-pp.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg.inet_recvmsg 10.55 ± 12% -5.9 4.64 ± 78% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg 10.35 ± 12% -5.8 4.53 ± 78% perf-profile.calltrace.cycles-pp.__sched_text_start.schedule.schedule_timeout.wait_woken.sk_wait_data 5.66 ± 21% -3.4 2.22 ± 79% perf-profile.calltrace.cycles-pp.ttwu_do_activate.try_to_wake_up.__wake_up_common.__wake_up_common_lock.sock_def_readable 5.51 ± 22% -3.4 2.15 ± 79% perf-profile.calltrace.cycles-pp.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.__wake_up_common.__wake_up_common_lock 5.05 ± 21% -3.0 2.03 ± 78% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__sched_text_start.schedule.schedule_timeout.wait_woken 3.04 ± 34% -2.3 0.74 ± 75% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 2.99 ± 34% -2.3 0.72 ± 75% perf-profile.calltrace.cycles-pp.__sched_text_start.schedule_idle.do_idle.cpu_startup_entry.start_secondary 4.14 ± 13% -2.1 2.00 ± 75% perf-profile.calltrace.cycles-pp.select_task_rq_fair.try_to_wake_up.__wake_up_common.__wake_up_common_lock.sock_def_readable 3.13 ± 32% -2.1 1.01 ± 76% perf-profile.calltrace.cycles-pp.enqueue_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.__wake_up_common 2.93 ± 32% -2.0 0.94 ± 77% perf-profile.calltrace.cycles-pp.dequeue_entity.dequeue_task_fair.__sched_text_start.schedule.schedule_timeout 3.48 ± 13% -1.9 1.59 ± 75% perf-profile.calltrace.cycles-pp.select_idle_sibling.select_task_rq_fair.try_to_wake_up.__wake_up_common.__wake_up_common_lock 1.58 ± 16% -1.1 0.52 ±105% perf-profile.calltrace.cycles-pp.available_idle_cpu.select_idle_sibling.select_task_rq_fair.try_to_wake_up.__wake_up_common 1.14 ± 7% -0.7 0.47 ±104% perf-profile.calltrace.cycles-pp.switch_mm_irqs_off.__sched_text_start.schedule.schedule_timeout.wait_woken 0.16 ±173% +0.8 0.95 ± 61% perf-profile.calltrace.cycles-pp.do_select.core_sys_select.kern_select.__x64_sys_select.do_syscall_64 0.14 ±173% +1.0 1.19 ± 58% perf-profile.calltrace.cycles-pp.__x64_sys_rt_sigaction.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.40 ±104% +1.3 1.72 ± 79% perf-profile.calltrace.cycles-pp.task_sched_runtime.thread_group_cputime.thread_group_cputime_adjusted.getrusage.__do_sys_getrusage 0.52 ±103% +2.0 2.48 ± 64% perf-profile.calltrace.cycles-pp.core_sys_select.kern_select.__x64_sys_select.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.77 ±104% +2.2 2.99 ± 68% perf-profile.calltrace.cycles-pp.thread_group_cputime_adjusted.getrusage.__do_sys_getrusage.do_syscall_64.entry_SYSCALL_64_after_hwframe 399.25 ± 22% +48.8% 594.00 ± 27% interrupts.47:PCI-MSI.1572878-edge.eth0-TxRx-14 399.25 ± 22% +48.8% 594.00 ± 27% interrupts.CPU14.47:PCI-MSI.1572878-edge.eth0-TxRx-14 28894 ± 27% -36.6% 18322 ± 11% interrupts.CPU15.RES:Rescheduling_interrupts 6661 ± 24% -41.3% 3911 ± 53% interrupts.CPU21.NMI:Non-maskable_interrupts 6661 ± 24% -41.3% 3911 ± 53% interrupts.CPU21.PMI:Performance_monitoring_interrupts 6395 ± 24% -40.9% 3776 ± 54% interrupts.CPU22.NMI:Non-maskable_interrupts 6395 ± 24% -40.9% 3776 ± 54% interrupts.CPU22.PMI:Performance_monitoring_interrupts 25874 ± 18% -29.3% 18288 ± 20% interrupts.CPU29.RES:Rescheduling_interrupts 26229 ± 11% -28.0% 18896 ± 13% interrupts.CPU32.RES:Rescheduling_interrupts 28466 ± 13% -35.1% 18482 ± 28% interrupts.CPU34.RES:Rescheduling_interrupts 28021 ± 26% -35.4% 18102 ± 25% interrupts.CPU36.RES:Rescheduling_interrupts 7321 -52.3% 3491 ± 67% interrupts.CPU37.NMI:Non-maskable_interrupts 7321 -52.3% 3491 ± 67% interrupts.CPU37.PMI:Performance_monitoring_interrupts 7351 -48.8% 3766 ± 55% interrupts.CPU38.NMI:Non-maskable_interrupts 7351 -48.8% 3766 ± 55% interrupts.CPU38.PMI:Performance_monitoring_interrupts 7333 -47.6% 3841 ± 52% interrupts.CPU39.NMI:Non-maskable_interrupts 7333 -47.6% 3841 ± 52% interrupts.CPU39.PMI:Performance_monitoring_interrupts 6408 ± 23% -41.3% 3759 ± 55% interrupts.CPU40.NMI:Non-maskable_interrupts 6408 ± 23% -41.3% 3759 ± 55% interrupts.CPU40.PMI:Performance_monitoring_interrupts 6416 ± 23% -35.2% 4160 ± 43% interrupts.CPU41.NMI:Non-maskable_interrupts 6416 ± 23% -35.2% 4160 ± 43% interrupts.CPU41.PMI:Performance_monitoring_interrupts 6371 ± 24% -39.8% 3838 ± 52% interrupts.CPU42.NMI:Non-maskable_interrupts 6371 ± 24% -39.8% 3838 ± 52% interrupts.CPU42.PMI:Performance_monitoring_interrupts 6383 ± 24% -37.4% 3993 ± 48% interrupts.CPU43.NMI:Non-maskable_interrupts 6383 ± 24% -37.4% 3993 ± 48% interrupts.CPU43.PMI:Performance_monitoring_interrupts 6636 ± 24% -41.7% 3868 ± 55% interrupts.CPU45.NMI:Non-maskable_interrupts 6636 ± 24% -41.7% 3868 ± 55% interrupts.CPU45.PMI:Performance_monitoring_interrupts 5684 ± 32% -31.4% 3900 ± 55% interrupts.CPU46.NMI:Non-maskable_interrupts 5684 ± 32% -31.4% 3900 ± 55% interrupts.CPU46.PMI:Performance_monitoring_interrupts 22917 ± 10% -19.3% 18485 ± 12% interrupts.CPU51.RES:Rescheduling_interrupts 24437 ± 16% -18.6% 19881 ± 21% interrupts.CPU53.RES:Rescheduling_interrupts 26653 ± 16% -30.7% 18474 ± 22% interrupts.CPU61.RES:Rescheduling_interrupts 26867 ± 9% -39.1% 16369 ± 12% interrupts.CPU62.RES:Rescheduling_interrupts 28101 ± 24% -29.3% 19868 ± 22% interrupts.CPU68.RES:Rescheduling_interrupts 30098 ± 34% -39.4% 18243 ± 18% interrupts.CPU73.RES:Rescheduling_interrupts 6384 ± 24% -59.0% 2616 ± 41% interrupts.CPU74.NMI:Non-maskable_interrupts 6384 ± 24% -59.0% 2616 ± 41% interrupts.CPU74.PMI:Performance_monitoring_interrupts 6371 ± 24% -58.7% 2633 ± 40% interrupts.CPU75.NMI:Non-maskable_interrupts 6371 ± 24% -58.7% 2633 ± 40% interrupts.CPU75.PMI:Performance_monitoring_interrupts 6410 ± 23% -59.5% 2595 ± 42% interrupts.CPU76.NMI:Non-maskable_interrupts 6410 ± 23% -59.5% 2595 ± 42% interrupts.CPU76.PMI:Performance_monitoring_interrupts 27943 ± 16% -39.8% 16824 ± 18% interrupts.CPU76.RES:Rescheduling_interrupts 5478 ± 32% -52.7% 2593 ± 41% interrupts.CPU77.NMI:Non-maskable_interrupts 5478 ± 32% -52.7% 2593 ± 41% interrupts.CPU77.PMI:Performance_monitoring_interrupts 6393 ± 24% -58.8% 2631 ± 41% interrupts.CPU78.NMI:Non-maskable_interrupts 6393 ± 24% -58.8% 2631 ± 41% interrupts.CPU78.PMI:Performance_monitoring_interrupts 6401 ± 24% -59.3% 2603 ± 41% interrupts.CPU79.NMI:Non-maskable_interrupts 6401 ± 24% -59.3% 2603 ± 41% interrupts.CPU79.PMI:Performance_monitoring_interrupts 6370 ± 24% -59.1% 2608 ± 41% interrupts.CPU80.NMI:Non-maskable_interrupts 6370 ± 24% -59.1% 2608 ± 41% interrupts.CPU80.PMI:Performance_monitoring_interrupts 7316 -64.1% 2623 ± 40% interrupts.CPU81.NMI:Non-maskable_interrupts 7316 -64.1% 2623 ± 40% interrupts.CPU81.PMI:Performance_monitoring_interrupts 7292 -64.1% 2616 ± 40% interrupts.CPU82.NMI:Non-maskable_interrupts 7292 -64.1% 2616 ± 40% interrupts.CPU82.PMI:Performance_monitoring_interrupts 7307 -63.5% 2668 ± 37% interrupts.CPU83.NMI:Non-maskable_interrupts 7307 -63.5% 2668 ± 37% interrupts.CPU83.PMI:Performance_monitoring_interrupts 7334 -64.4% 2610 ± 40% interrupts.CPU84.NMI:Non-maskable_interrupts 7334 -64.4% 2610 ± 40% interrupts.CPU84.PMI:Performance_monitoring_interrupts 7350 -54.1% 3374 ± 15% interrupts.CPU85.NMI:Non-maskable_interrupts 7350 -54.1% 3374 ± 15% interrupts.CPU85.PMI:Performance_monitoring_interrupts 7312 -59.9% 2933 ± 22% interrupts.CPU86.NMI:Non-maskable_interrupts 7312 -59.9% 2933 ± 22% interrupts.CPU86.PMI:Performance_monitoring_interrupts 7338 -43.4% 4154 ± 45% interrupts.CPU87.NMI:Non-maskable_interrupts 7338 -43.4% 4154 ± 45% interrupts.CPU87.PMI:Performance_monitoring_interrupts 42075 ± 5% -16.4% 35154 ± 6% softirqs.CPU0.SCHED 39762 ± 7% -16.8% 33092 ± 8% softirqs.CPU1.SCHED 39607 ± 7% -19.2% 32002 ± 6% softirqs.CPU10.SCHED 39624 ± 10% -19.8% 31770 ± 6% softirqs.CPU11.SCHED 38851 ± 6% -18.1% 31832 ± 7% softirqs.CPU12.SCHED 39142 ± 6% -15.6% 33029 ± 7% softirqs.CPU13.SCHED 38809 ± 7% -17.5% 32025 ± 6% softirqs.CPU14.SCHED 40786 ± 12% -21.6% 31993 ± 5% softirqs.CPU15.SCHED 38640 ± 6% -17.4% 31903 ± 7% softirqs.CPU16.SCHED 38433 ± 6% -17.1% 31877 ± 6% softirqs.CPU17.SCHED 40026 ± 10% -16.7% 33343 ± 11% softirqs.CPU18.SCHED 41468 ± 9% -25.2% 31023 ± 6% softirqs.CPU19.SCHED 38731 ± 6% -15.0% 32938 ± 4% softirqs.CPU2.SCHED 5406654 ± 5% -8.8% 4931685 ± 3% softirqs.CPU20.NET_RX 39048 ± 6% -18.1% 31963 ± 7% softirqs.CPU20.SCHED 39819 ± 8% -20.6% 31596 ± 7% softirqs.CPU21.SCHED 39843 ± 8% -20.1% 31827 ± 11% softirqs.CPU22.SCHED 39542 ± 10% -19.9% 31668 ± 11% softirqs.CPU24.SCHED 39722 ± 9% -19.2% 32088 ± 9% softirqs.CPU25.SCHED 39669 ± 10% -19.1% 32078 ± 9% softirqs.CPU26.SCHED 39649 ± 11% -19.6% 31887 ± 10% softirqs.CPU27.SCHED 39862 ± 12% -20.6% 31651 ± 10% softirqs.CPU28.SCHED 39196 ± 11% -19.5% 31533 ± 9% softirqs.CPU29.SCHED 40023 ± 9% -19.0% 32406 ± 3% softirqs.CPU3.SCHED 39523 ± 11% -19.1% 31960 ± 10% softirqs.CPU30.SCHED 39695 ± 11% -19.8% 31838 ± 8% softirqs.CPU31.SCHED 38974 ± 11% -19.5% 31383 ± 10% softirqs.CPU32.SCHED 39283 ± 11% -19.2% 31742 ± 10% softirqs.CPU33.SCHED 39383 ± 12% -19.5% 31712 ± 10% softirqs.CPU34.SCHED 40608 ± 12% -21.7% 31797 ± 12% softirqs.CPU36.SCHED 39687 ± 11% -19.6% 31927 ± 11% softirqs.CPU37.SCHED 39664 ± 10% -19.1% 32099 ± 10% softirqs.CPU38.SCHED 39632 ± 10% -19.9% 31762 ± 10% softirqs.CPU39.SCHED 38780 ± 5% -16.3% 32468 ± 7% softirqs.CPU4.SCHED 39578 ± 11% -20.5% 31466 ± 11% softirqs.CPU40.SCHED 39321 ± 10% -18.7% 31961 ± 10% softirqs.CPU41.SCHED 39030 ± 11% -16.7% 32507 ± 9% softirqs.CPU42.SCHED 38372 ± 8% -16.9% 31898 ± 10% softirqs.CPU43.SCHED 38653 ± 7% -17.3% 31949 ± 6% softirqs.CPU44.SCHED 38385 ± 6% -18.4% 31334 ± 6% softirqs.CPU45.SCHED 38379 ± 6% -16.2% 32151 ± 5% softirqs.CPU47.SCHED 38798 ± 5% -15.5% 32780 ± 6% softirqs.CPU48.SCHED 38516 ± 6% -16.0% 32351 ± 6% softirqs.CPU49.SCHED 38685 ± 7% -12.3% 33942 ± 10% softirqs.CPU5.SCHED 38451 ± 6% -16.6% 32063 ± 6% softirqs.CPU50.SCHED 38345 ± 7% -17.3% 31722 ± 6% softirqs.CPU51.SCHED 39662 ± 6% -20.0% 31733 ± 6% softirqs.CPU52.SCHED 38634 ± 7% -17.6% 31849 ± 6% softirqs.CPU53.SCHED 39172 ± 6% -19.7% 31467 ± 7% softirqs.CPU54.SCHED 5379816 ± 4% -8.8% 4905481 ± 2% softirqs.CPU55.NET_RX 38658 ± 7% -18.5% 31505 ± 6% softirqs.CPU55.SCHED 38386 ± 7% -19.0% 31095 ± 4% softirqs.CPU56.SCHED 38370 ± 5% -15.8% 32305 ± 6% softirqs.CPU57.SCHED 38724 ± 7% -18.0% 31762 ± 4% softirqs.CPU58.SCHED 38540 ± 7% -18.5% 31398 ± 5% softirqs.CPU59.SCHED 38642 ± 6% -15.0% 32829 ± 9% softirqs.CPU6.SCHED 38745 ± 7% -18.0% 31789 ± 6% softirqs.CPU60.SCHED 38501 ± 6% -17.6% 31709 ± 6% softirqs.CPU61.SCHED 38697 ± 5% -18.2% 31659 ± 6% softirqs.CPU62.SCHED 39577 ± 7% -20.4% 31503 ± 6% softirqs.CPU63.SCHED 5678631 ± 13% -16.8% 4722162 ± 2% softirqs.CPU64.NET_RX 39312 ± 7% -19.5% 31647 ± 6% softirqs.CPU64.SCHED 38828 ± 5% -18.6% 31606 ± 7% softirqs.CPU65.SCHED 39268 ± 10% -17.9% 32233 ± 11% softirqs.CPU66.SCHED 39052 ± 10% -20.2% 31151 ± 12% softirqs.CPU67.SCHED 38862 ± 12% -20.0% 31094 ± 12% softirqs.CPU68.SCHED 38819 ± 12% -19.8% 31147 ± 11% softirqs.CPU69.SCHED 39013 ± 5% -13.2% 33865 ± 4% softirqs.CPU7.SCHED 39271 ± 10% -22.9% 30261 ± 13% softirqs.CPU70.SCHED 39586 ± 10% -21.1% 31234 ± 12% softirqs.CPU71.SCHED 39970 ± 12% -22.1% 31155 ± 11% softirqs.CPU72.SCHED 39372 ± 10% -20.1% 31441 ± 11% softirqs.CPU73.SCHED 39419 ± 11% -22.7% 30488 ± 10% softirqs.CPU74.SCHED 39041 ± 11% -19.2% 31532 ± 9% softirqs.CPU75.SCHED 39211 ± 11% -20.3% 31254 ± 10% softirqs.CPU76.SCHED 38997 ± 10% -20.4% 31024 ± 11% softirqs.CPU77.SCHED 38902 ± 11% -18.3% 31788 ± 13% softirqs.CPU78.SCHED 39187 ± 11% -19.6% 31506 ± 13% softirqs.CPU79.SCHED 38878 ± 7% -17.1% 32222 ± 6% softirqs.CPU8.SCHED 38849 ± 10% -18.3% 31748 ± 13% softirqs.CPU80.SCHED 39415 ± 10% -18.2% 32243 ± 12% softirqs.CPU81.SCHED 38612 ± 11% -19.0% 31271 ± 11% softirqs.CPU82.SCHED 39255 ± 11% -18.9% 31818 ± 11% softirqs.CPU83.SCHED 38819 ± 10% -19.1% 31403 ± 10% softirqs.CPU84.SCHED 39017 ± 10% -17.7% 32115 ± 12% softirqs.CPU85.SCHED 39175 ± 10% -17.7% 32232 ± 11% softirqs.CPU86.SCHED 38852 ± 10% -17.0% 32244 ± 11% softirqs.CPU87.SCHED 41916 ± 6% -23.5% 32046 ± 8% softirqs.CPU9.SCHED 3452696 ± 8% -18.5% 2812299 ± 6% softirqs.SCHED lmbench3.TCP.socket.bandwidth.10MB.MB_sec 120000 +-+----------------------------------------------------------------+ | | 100000 +-+++ +.+ .+ +. ++.++.+++. + .+ + .++.+| | : : + + ++.+++.++.+ ++ + + ++.++.+ +.++.+++ | | : : + | 80000 OO+OO:OOO OO OOO OO OOO OO | | : : | 60000 +-+ : : | | : : | 40000 +-+ : : | | : : | | :: | 20000 +-+ :: | | :: | 0 +-+----------------------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Rong Chen View attachment "config-5.0.0-11696-g01b4c2a" of type "text/plain" (192819 bytes) View attachment "job-script" of type "text/plain" (7233 bytes) View attachment "job.yaml" of type "text/plain" (4919 bytes) View attachment "reproduce" of type "text/plain" (566 bytes)
Powered by blists - more mailing lists