lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Wed, 08 Jul 2015 09:57:52 +0800 From: Huang Ying <ying.huang@...el.com> To: "Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com> Cc: LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org> Subject: [lkp] [thp] 79553da293d: +1.8% fileio.time.file_system_inputs FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master commit 79553da293d38d63097278de13e28a3b371f43c1 ("thp: cleanup khugepaged startup") ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/disk/iosched/fs/nr_threads: bay/dd-write/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1HDD/cfq/xfs/10dd commit: e39155ea11eac6da056b04669d7c9fc612e2065a 79553da293d38d63097278de13e28a3b371f43c1 e39155ea11eac6da 79553da293d38d63097278de13 ---------------- -------------------------- %stddev %change %stddev \ | \ 30460 ± 0% +34.3% 40920 ± 0% softirqs.BLOCK 231.27 ± 4% -17.3% 191.30 ± 3% uptime.idle 765.75 ± 7% +19.7% 916.50 ± 4% slabinfo.kmalloc-512.active_objs 972.25 ± 8% +17.4% 1141 ± 5% slabinfo.kmalloc-512.num_objs 74.00 ± 0% +15.9% 85.75 ± 10% vmstat.memory.buff 91092 ± 0% -62.3% 34370 ± 1% vmstat.memory.free 22460 ± 1% +29.2% 29026 ± 1% ftrace.global_dirty_state.dirty 35516 ± 1% -11.0% 31615 ± 0% ftrace.global_dirty_state.writeback 2.00 ± 0% +50.0% 3.00 ± 0% ftrace.writeback_single_inode.sda.age 4913 ± 1% +35.7% 6666 ± 1% ftrace.writeback_single_inode.sda.wrote 36958 ± 4% -67.3% 12083 ± 28% meminfo.AnonHugePages 89634 ± 0% +29.2% 115815 ± 1% meminfo.Dirty 1315242 ± 0% +10.4% 1451507 ± 0% meminfo.MemAvailable 87870 ± 0% -63.5% 32046 ± 2% meminfo.MemFree 142291 ± 1% -11.2% 126331 ± 0% meminfo.Writeback 2.76 ± 14% +22.8% 3.39 ± 8% perf-profile.cpu-cycles.__clear_user.iov_iter_zero.read_iter_zero.new_sync_read.__vfs_read 4.94 ± 16% +31.2% 6.48 ± 6% perf-profile.cpu-cycles.__memset.xfs_vm_write_begin.generic_perform_write.xfs_file_buffered_aio_write.xfs_file_write_iter 1.31 ± 34% -50.4% 0.65 ± 36% perf-profile.cpu-cycles.end_page_writeback.end_buffer_async_write.xfs_destroy_ioend.xfs_end_io.process_one_work 0.86 ± 24% -44.1% 0.48 ± 42% perf-profile.cpu-cycles.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault 1.15 ± 27% -50.4% 0.57 ± 39% perf-profile.cpu-cycles.test_clear_page_writeback.end_page_writeback.end_buffer_async_write.xfs_destroy_ioend.xfs_end_io 0.76 ± 39% +65.6% 1.26 ± 23% perf-profile.cpu-cycles.try_to_free_buffers.xfs_vm_releasepage.try_to_release_page.shrink_page_list.shrink_inactive_list 0.00 ± -1% +Inf% 9581 ± 43% latency_stats.avg.do_truncate.do_sys_ftruncate.SyS_ftruncate.system_call_fastpath 0.00 ± -1% +Inf% 500330 ± 58% latency_stats.avg.xfs_file_buffered_aio_write.xfs_file_write_iter.new_sync_write.vfs_write.SyS_write.system_call_fastpath 163371 ± 3% +26.6% 206775 ± 1% latency_stats.hits.ring_buffer_wait.wait_on_pipe.tracing_wait_pipe.tracing_read_pipe.__vfs_read.vfs_read.SyS_read.system_call_fastpath 0.00 ± -1% +Inf% 32695 ± 10% latency_stats.max.do_truncate.do_sys_ftruncate.SyS_ftruncate.system_call_fastpath 0.00 ± -1% +Inf% 669208 ± 18% latency_stats.max.xfs_file_buffered_aio_write.xfs_file_write_iter.new_sync_write.vfs_write.SyS_write.system_call_fastpath 0.00 ± -1% +Inf% 53331 ± 28% latency_stats.sum.do_truncate.do_sys_ftruncate.SyS_ftruncate.system_call_fastpath 0.00 ± -1% +Inf% 706140 ± 12% latency_stats.sum.xfs_file_buffered_aio_write.xfs_file_write_iter.new_sync_write.vfs_write.SyS_write.system_call_fastpath 394.50 ± 24% -53.2% 184.50 ± 37% sched_debug.cfs_rq[3]:/.load 37.00 ± 34% +51.4% 56.00 ± 17% sched_debug.cpu#1.cpu_load[1] 84.75 ± 19% -52.8% 40.00 ± 68% sched_debug.cpu#3.cpu_load[0] 69.25 ± 20% -50.5% 34.25 ± 34% sched_debug.cpu#3.cpu_load[1] 50.25 ± 11% -45.8% 27.25 ± 23% sched_debug.cpu#3.cpu_load[2] 36.25 ± 5% -37.9% 22.50 ± 22% sched_debug.cpu#3.cpu_load[3] 394.50 ± 24% -53.2% 184.50 ± 37% sched_debug.cpu#3.load 2.138e+11 ± 0% -1.4% 2.108e+11 ± 0% perf-stat.L1-dcache-loads 2.483e+08 ± 0% -4.3% 2.376e+08 ± 0% perf-stat.L1-dcache-prefetches 4.605e+09 ± 0% +7.8% 4.962e+09 ± 0% perf-stat.L1-icache-load-misses 5.684e+11 ± 0% +2.0% 5.797e+11 ± 0% perf-stat.L1-icache-loads 1.637e+08 ± 0% +15.8% 1.895e+08 ± 1% perf-stat.LLC-load-misses 1.089e+11 ± 0% +1.3% 1.103e+11 ± 0% perf-stat.branch-loads 8.163e+10 ± 0% +2.9% 8.396e+10 ± 1% perf-stat.bus-cycles 3.116e+08 ± 1% +9.6% 3.415e+08 ± 0% perf-stat.cache-misses 9306030 ± 1% -1.4% 9171655 ± 0% perf-stat.context-switches 151544 ± 2% +8.3% 164060 ± 1% perf-stat.cpu-migrations 2.299e+08 ± 12% +25.5% 2.886e+08 ± 10% perf-stat.dTLB-load-misses 8302411 ± 3% +17.4% 9747172 ± 2% perf-stat.iTLB-load-misses 7.348e+11 ± 0% +2.3% 7.515e+11 ± 0% perf-stat.ref-cycles 52222 ± 6% -72.6% 14324 ± 28% proc-vmstat.compact_isolated 42280 ± 10% -71.6% 11991 ± 35% proc-vmstat.compact_migrate_scanned 5320 ± 1% -99.9% 6.50 ± 35% proc-vmstat.kswapd_high_wmark_hit_quickly 3377 ± 13% +602.5% 23725 ± 1% proc-vmstat.kswapd_low_wmark_hit_quickly 1446 ± 0% -86.9% 190.00 ± 4% proc-vmstat.nr_alloc_batch 22517 ± 0% +28.9% 29013 ± 1% proc-vmstat.nr_dirty 21984 ± 0% -63.5% 8021 ± 2% proc-vmstat.nr_free_pages 1142 ± 15% -24.3% 864.75 ± 4% proc-vmstat.nr_vmscan_immediate_reclaim 35464 ± 1% -11.0% 31557 ± 0% proc-vmstat.nr_writeback 8844 ± 4% +169.2% 23813 ± 1% proc-vmstat.pageoutrun 17473 ± 0% -91.7% 1447 ± 33% proc-vmstat.pgalloc_dma 24252 ± 5% -74.5% 6188 ± 25% proc-vmstat.pgmigrate_success 16476 ± 0% -94.4% 925.00 ± 48% proc-vmstat.pgrefill_dma 8267 ± 9% +816.4% 75755 ± 8% proc-vmstat.pgscan_direct_dma32 16728 ± 0% -94.5% 927.50 ± 38% proc-vmstat.pgscan_kswapd_dma 8267 ± 9% +128.8% 18917 ± 8% proc-vmstat.pgsteal_direct_dma32 15954 ± 0% -97.0% 483.75 ± 59% proc-vmstat.pgsteal_kswapd_dma ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/disk/iosched/fs/nr_threads: bay/dd-write/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1HDD/cfq/xfs/1dd commit: e39155ea11eac6da056b04669d7c9fc612e2065a 79553da293d38d63097278de13e28a3b371f43c1 e39155ea11eac6da 79553da293d38d63097278de13 ---------------- -------------------------- %stddev %change %stddev \ | \ 3.50 ± 14% +1750.0% 64.75 ± 15% ftrace.writeback_single_inode.sda.age 471.00 ± 2% +28.9% 607.00 ± 5% slabinfo.blkdev_requests.active_objs 476.25 ± 1% +29.6% 617.25 ± 4% slabinfo.blkdev_requests.num_objs 10396 ± 1% +45.5% 15129 ± 0% softirqs.BLOCK 37406 ± 3% +18.5% 44317 ± 4% softirqs.RCU 25121 ± 5% -59.5% 10170 ± 12% meminfo.AnonHugePages 1332448 ± 0% +10.1% 1466476 ± 0% meminfo.MemAvailable 89924 ± 0% -63.0% 33289 ± 0% meminfo.MemFree 69.25 ± 2% +32.9% 92.00 ± 11% vmstat.memory.buff 90542 ± 1% -62.6% 33856 ± 0% vmstat.memory.free 17946 ± 2% -8.6% 16406 ± 4% vmstat.system.cs 6455037 ±100% -99.7% 19975 ±173% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 108228 ± 3% +34.7% 145793 ± 5% latency_stats.hits.pipe_wait.pipe_read.new_sync_read.__vfs_read.vfs_read.SyS_read.system_call_fastpath 52101 ± 1% +140.2% 125153 ± 2% latency_stats.hits.ring_buffer_wait.wait_on_pipe.tracing_wait_pipe.tracing_read_pipe.__vfs_read.vfs_read.SyS_read.system_call_fastpath 6737920 ±100% -99.7% 19975 ±173% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 572233 ± 87% -41.2% 336545 ± 3% latency_stats.max.wait_on_page_bit.truncate_inode_pages_range.truncate_pagecache.truncate_setsize.xfs_setattr_size.xfs_vn_setattr.notify_change.do_truncate.do_sys_ftruncate.SyS_ftruncate.system_call_fastpath 9804520 ±112% -99.8% 19975 ±173% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 5007 ± 0% -99.8% 8.50 ± 35% proc-vmstat.kswapd_high_wmark_hit_quickly 4012 ± 5% +584.2% 27456 ± 2% proc-vmstat.kswapd_low_wmark_hit_quickly 1441 ± 0% -87.2% 184.75 ± 2% proc-vmstat.nr_alloc_batch 22489 ± 0% -63.0% 8329 ± 0% proc-vmstat.nr_free_pages 853.75 ± 6% +31.1% 1119 ± 8% proc-vmstat.nr_vmscan_immediate_reclaim 9282 ± 2% +196.4% 27510 ± 2% proc-vmstat.pageoutrun 15361 ± 12% -95.3% 717.00 ± 37% proc-vmstat.pgalloc_dma 14328 ± 12% -98.3% 249.75 ± 79% proc-vmstat.pgrefill_dma 14670 ± 12% -97.8% 321.50 ± 2% proc-vmstat.pgscan_kswapd_dma 13886 ± 13% -100.0% 0.00 ± -1% proc-vmstat.pgsteal_kswapd_dma 108.75 ± 29% -43.2% 61.75 ± 31% sched_debug.cfs_rq[2]:/.runnable_load_avg 396.25 ± 20% -45.9% 214.50 ± 18% sched_debug.cfs_rq[2]:/.utilization_load_avg 664.00 ± 34% +144.8% 1625 ± 59% sched_debug.cfs_rq[3]:/.blocked_load_avg 738.75 ± 31% +132.6% 1718 ± 57% sched_debug.cfs_rq[3]:/.tg_load_contrib 34.75 ± 36% +86.3% 64.75 ± 22% sched_debug.cpu#0.cpu_load[2] 29.00 ± 23% +92.2% 55.75 ± 28% sched_debug.cpu#0.cpu_load[3] 69144 ± 1% +17.7% 81411 ± 13% sched_debug.cpu#1.nr_load_updates 107.75 ± 22% -39.9% 64.75 ± 23% sched_debug.cpu#2.cpu_load[0] 2359 ± 26% -57.2% 1009 ± 34% sched_debug.cpu#2.curr->pid 414.50 ± 10% -32.8% 278.75 ± 23% sched_debug.cpu#2.load 184007 ± 12% +299.5% 735051 ± 61% sched_debug.cpu#3.sched_goidle 1.55 ± 12% +27.2% 1.97 ± 9% perf-profile.cpu-cycles.__xfs_get_blocks.xfs_get_blocks.__block_write_begin.xfs_vm_write_begin.generic_perform_write 1.34 ± 22% +46.9% 1.97 ± 24% perf-profile.cpu-cycles.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel 1.30 ± 20% -34.7% 0.85 ± 8% perf-profile.cpu-cycles.down_write.xfs_ilock.xfs_file_buffered_aio_write.xfs_file_write_iter.new_sync_write 0.30 ± 16% +109.0% 0.62 ± 50% perf-profile.cpu-cycles.fsnotify.vfs_read.sys_read.system_call_fastpath 12.11 ± 9% +19.3% 14.44 ± 9% perf-profile.cpu-cycles.kthread.ret_from_fork 1.34 ± 22% +46.9% 1.97 ± 24% perf-profile.cpu-cycles.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel 12.13 ± 9% +19.0% 14.44 ± 9% perf-profile.cpu-cycles.ret_from_fork 1.36 ± 21% -49.5% 0.69 ± 11% perf-profile.cpu-cycles.rw_verify_area.vfs_write.sys_write.system_call_fastpath 0.93 ± 27% -57.1% 0.40 ± 30% perf-profile.cpu-cycles.security_file_permission.rw_verify_area.vfs_write.sys_write.system_call_fastpath 1.34 ± 22% +46.9% 1.97 ± 24% perf-profile.cpu-cycles.start_kernel.x86_64_start_reservations.x86_64_start_kernel 0.80 ± 31% +76.3% 1.41 ± 15% perf-profile.cpu-cycles.tracing_read_pipe.__vfs_read.vfs_read.sys_read.system_call_fastpath 1.34 ± 22% +46.9% 1.97 ± 24% perf-profile.cpu-cycles.x86_64_start_kernel 1.34 ± 22% +46.9% 1.97 ± 24% perf-profile.cpu-cycles.x86_64_start_reservations.x86_64_start_kernel 2.39e+11 ± 1% -2.5% 2.33e+11 ± 1% perf-stat.L1-dcache-loads 2.559e+08 ± 2% +4.3% 2.669e+08 ± 1% perf-stat.L1-dcache-prefetches 4.836e+09 ± 2% +7.5% 5.197e+09 ± 0% perf-stat.L1-icache-load-misses 6.191e+11 ± 1% +3.2% 6.387e+11 ± 1% perf-stat.L1-icache-loads 1.187e+08 ± 1% +29.7% 1.54e+08 ± 2% perf-stat.LLC-load-misses 1.355e+09 ± 4% +9.5% 1.484e+09 ± 4% perf-stat.LLC-loads 1.104e+08 ± 1% +12.9% 1.246e+08 ± 1% perf-stat.LLC-store-misses 1.244e+09 ± 3% -6.7% 1.161e+09 ± 1% perf-stat.branch-misses 2.697e+08 ± 0% +25.1% 3.373e+08 ± 2% perf-stat.cache-misses 10825819 ± 2% -8.6% 9891678 ± 4% perf-stat.context-switches 63968 ± 1% +54.6% 98882 ± 2% perf-stat.cpu-migrations 2.385e+11 ± 1% -3.0% 2.314e+11 ± 1% perf-stat.dTLB-loads 5457543 ± 7% +22.7% 6698136 ± 5% perf-stat.iTLB-load-misses ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/period/nr_threads/disk/fs/size/filenum/rwmode/iomode: lkp-ws02/fileio/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/600s/100%/1HDD/btrfs/64G/1024f/rndrd/sync commit: e39155ea11eac6da056b04669d7c9fc612e2065a 79553da293d38d63097278de13e28a3b371f43c1 e39155ea11eac6da 79553da293d38d63097278de13 ---------------- -------------------------- %stddev %change %stddev \ | \ 6708548 ± 0% +1.8% 6832358 ± 0% fileio.time.file_system_inputs 27192978 ± 10% +24.2% 33781232 ± 9% cpuidle.C1-NHM.time 291264 ± 3% -19.2% 235461 ± 2% cpuidle.C1-NHM.usage 256.25 ± 5% -8.5% 234.50 ± 4% cpuidle.POLL.usage 6927467 ±122% -77.4% 1563952 ±100% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 126552 ± 73% -100.0% 0.00 ± -1% latency_stats.avg.wait_on_page_bit.extent_write_cache_pages.[btrfs].extent_writepages.[btrfs].btrfs_writepages.[btrfs].do_writepages.__filemap_fdatawrite_range.filemap_fdatawrite_range.btrfs_fdatawrite_range.[btrfs].start_ordered_ops.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync 304943 ± 3% -18.4% 248685 ± 2% latency_stats.hits.pipe_wait.pipe_read.new_sync_read.__vfs_read.vfs_read.SyS_read.system_call_fastpath 31145 ±133% -24.8% 23416 ± 0% latency_stats.max.blk_execute_rq.scsi_execute.scsi_execute_req_flags.ses_recv_diag.[ses].ses_enclosure_data_process.[ses].ses_match_to_enclosure.[ses].ses_intf_add.[ses].class_interface_register.scsi_register_interface.0xffffffffa0006013.do_one_initcall.do_init_module 242127 ± 74% -100.0% 0.00 ± -1% latency_stats.max.wait_on_page_bit.extent_write_cache_pages.[btrfs].extent_writepages.[btrfs].btrfs_writepages.[btrfs].do_writepages.__filemap_fdatawrite_range.filemap_fdatawrite_range.btrfs_fdatawrite_range.[btrfs].start_ordered_ops.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync 253104 ± 73% -100.0% 0.00 ± -1% latency_stats.sum.wait_on_page_bit.extent_write_cache_pages.[btrfs].extent_writepages.[btrfs].btrfs_writepages.[btrfs].do_writepages.__filemap_fdatawrite_range.filemap_fdatawrite_range.btrfs_fdatawrite_range.[btrfs].start_ordered_ops.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync 1452770 ± 9% +519.6% 9001196 ± 47% numa-numastat.node0.local_node 166456 ±143% +3090.2% 5310291 ± 56% numa-numastat.node0.numa_foreign 1452773 ± 9% +519.6% 9001200 ± 47% numa-numastat.node0.numa_hit 7153013 ± 3% -75.4% 1761824 ±164% numa-numastat.node0.numa_miss 7153016 ± 3% -75.4% 1761828 ±164% numa-numastat.node0.other_node 7153013 ± 3% -75.4% 1761824 ±164% numa-numastat.node1.numa_foreign 166456 ±143% +3090.2% 5310291 ± 56% numa-numastat.node1.numa_miss 166473 ±143% +3089.9% 5310296 ± 56% numa-numastat.node1.other_node 1128 ± 2% -39.1% 687.00 ± 35% slabinfo.bio-1.active_objs 1128 ± 2% -36.2% 719.75 ± 30% slabinfo.bio-1.num_objs 206.75 ± 4% -43.4% 117.00 ± 34% slabinfo.btrfs_transaction.active_objs 206.75 ± 4% -43.4% 117.00 ± 34% slabinfo.btrfs_transaction.num_objs 614.00 ± 5% -18.0% 503.25 ± 11% slabinfo.ext4_free_data.active_objs 614.00 ± 5% -18.0% 503.25 ± 11% slabinfo.ext4_free_data.num_objs 1241 ± 5% -13.2% 1076 ± 10% slabinfo.file_lock_ctx.active_objs 1241 ± 5% -13.2% 1076 ± 10% slabinfo.file_lock_ctx.num_objs 3644 ± 2% -9.0% 3317 ± 5% slabinfo.kmalloc-1024.active_objs 1505 ± 1% -59.8% 605.75 ± 5% proc-vmstat.kswapd_high_wmark_hit_quickly 53.00 ± 15% +19436.8% 10354 ± 1% proc-vmstat.kswapd_low_wmark_hit_quickly 3246 ± 2% -82.2% 577.50 ± 0% proc-vmstat.nr_alloc_batch 905.25 ± 54% -72.2% 251.25 ± 38% proc-vmstat.numa_hint_faults 836.50 ± 63% -72.8% 227.50 ± 28% proc-vmstat.numa_hint_faults_local 1003 ± 42% -72.3% 278.25 ± 15% proc-vmstat.numa_pte_updates 3357 ± 0% +296.0% 13296 ± 1% proc-vmstat.pageoutrun 412.00 ± 40% +666.0% 3155 ± 18% proc-vmstat.pgscan_direct_dma32 1192 ± 24% +1926.9% 24174 ± 17% proc-vmstat.pgscan_direct_normal 1673 ± 3% +13.6% 1900 ± 6% proc-vmstat.workingset_activate 186101 ± 1% +39.2% 259051 ± 15% numa-meminfo.node0.Active 175144 ± 0% +43.7% 251597 ± 16% numa-meminfo.node0.Active(file) 8061 ± 34% -34.6% 5270 ± 0% numa-meminfo.node0.Mapped 874427 ± 0% -46.8% 464786 ± 46% numa-meminfo.node0.MemFree 88387 ± 0% +20.4% 106421 ± 8% numa-meminfo.node0.Slab 286764 ± 0% -25.1% 214834 ± 18% numa-meminfo.node1.Active 275866 ± 0% -27.5% 200006 ± 21% numa-meminfo.node1.Active(file) 1072 ± 8% -32.2% 727.50 ± 24% numa-meminfo.node1.Dirty 87375 ± 1% -18.7% 71059 ± 13% numa-meminfo.node1.SReclaimable 21629 ± 5% -11.0% 19250 ± 2% numa-meminfo.node1.SUnreclaim 109004 ± 0% -17.1% 90310 ± 11% numa-meminfo.node1.Slab 43785 ± 0% +43.7% 62898 ± 16% numa-vmstat.node0.nr_active_file 1837 ± 1% -85.3% 270.25 ± 0% numa-vmstat.node0.nr_alloc_batch 218616 ± 0% -46.8% 116208 ± 46% numa-vmstat.node0.nr_free_pages 2015 ± 34% -34.6% 1317 ± 0% numa-vmstat.node0.nr_mapped 34180 ±144% +10851.6% 3743275 ± 57% numa-vmstat.node0.numa_foreign 767794 ± 3% +689.8% 6064164 ± 50% numa-vmstat.node0.numa_hit 707070 ± 3% +749.1% 6003501 ± 50% numa-vmstat.node0.numa_local 5022883 ± 1% -75.4% 1236569 ±170% numa-vmstat.node0.numa_miss 5083607 ± 1% -74.5% 1297231 ±162% numa-vmstat.node0.numa_other 605.25 ± 3% +78.0% 1077 ± 24% numa-vmstat.node0.workingset_activate 68966 ± 0% -27.5% 50001 ± 21% numa-vmstat.node1.nr_active_file 1433 ± 2% -81.5% 265.75 ± 1% numa-vmstat.node1.nr_alloc_batch 21843 ± 1% -18.7% 17764 ± 13% numa-vmstat.node1.nr_slab_reclaimable 5406 ± 5% -11.0% 4812 ± 2% numa-vmstat.node1.nr_slab_unreclaimable 5022889 ± 1% -75.4% 1236573 ±170% numa-vmstat.node1.numa_foreign 34182 ±144% +10851.0% 3743283 ± 57% numa-vmstat.node1.numa_miss 38048 ±129% +9748.4% 3747130 ± 57% numa-vmstat.node1.numa_other 1696 ± 30% -70.6% 498.25 ± 18% sched_debug.cfs_rq[10]:/.avg->runnable_avg_sum 10699 ± 18% -48.9% 5471 ± 57% sched_debug.cfs_rq[10]:/.exec_clock 31648 ± 36% -58.8% 13052 ± 59% sched_debug.cfs_rq[10]:/.min_vruntime 493.00 ± 16% -29.9% 345.75 ± 6% sched_debug.cfs_rq[10]:/.tg->runnable_avg 36.00 ± 32% -72.9% 9.75 ± 16% sched_debug.cfs_rq[10]:/.tg_runnable_contrib 166.75 ± 32% -60.0% 66.75 ± 47% sched_debug.cfs_rq[11]:/.blocked_load_avg 9322 ± 21% -53.0% 4380 ± 41% sched_debug.cfs_rq[11]:/.exec_clock 23106 ± 28% -57.7% 9765 ± 34% sched_debug.cfs_rq[11]:/.min_vruntime 11778 ± 54% -791.7% -81472 ±-61% sched_debug.cfs_rq[11]:/.spread0 493.50 ± 14% -29.7% 346.75 ± 5% sched_debug.cfs_rq[11]:/.tg->runnable_avg 168.25 ± 32% -59.9% 67.50 ± 47% sched_debug.cfs_rq[11]:/.tg_load_contrib 498.50 ± 15% -29.6% 351.00 ± 4% sched_debug.cfs_rq[12]:/.tg->runnable_avg 1718 ± 13% +28.9% 2214 ± 9% sched_debug.cfs_rq[13]:/.exec_clock 501.25 ± 14% -28.7% 357.25 ± 5% sched_debug.cfs_rq[13]:/.tg->runnable_avg 506.50 ± 15% -30.6% 351.75 ± 4% sched_debug.cfs_rq[14]:/.tg->runnable_avg 1210 ± 9% +42.4% 1723 ± 10% sched_debug.cfs_rq[15]:/.exec_clock 6566 ± 9% +17.6% 7721 ± 10% sched_debug.cfs_rq[15]:/.min_vruntime 511.25 ± 14% -30.6% 355.00 ± 3% sched_debug.cfs_rq[15]:/.tg->runnable_avg 515.50 ± 14% -30.4% 358.75 ± 3% sched_debug.cfs_rq[16]:/.tg->runnable_avg -3588 ±-25% +2233.9% -83744 ±-54% sched_debug.cfs_rq[17]:/.spread0 520.25 ± 13% -30.7% 360.75 ± 3% sched_debug.cfs_rq[17]:/.tg->runnable_avg 2859 ± 73% -51.2% 1394 ± 14% sched_debug.cfs_rq[18]:/.exec_clock 11966 ± 67% -45.0% 6585 ± 9% sched_debug.cfs_rq[18]:/.min_vruntime 524.75 ± 13% -30.7% 363.50 ± 3% sched_debug.cfs_rq[18]:/.tg->runnable_avg 523.50 ± 12% -29.7% 368.00 ± 4% sched_debug.cfs_rq[19]:/.tg->runnable_avg 2926 ± 5% +48.9% 4356 ± 18% sched_debug.cfs_rq[1]:/.exec_clock 8401 ± 6% +30.1% 10932 ± 6% sched_debug.cfs_rq[1]:/.min_vruntime 527.50 ± 11% -29.5% 372.00 ± 3% sched_debug.cfs_rq[20]:/.tg->runnable_avg 2098 ± 46% -35.1% 1362 ± 3% sched_debug.cfs_rq[21]:/.exec_clock 532.00 ± 11% -28.9% 378.00 ± 3% sched_debug.cfs_rq[21]:/.tg->runnable_avg 537.25 ± 12% -28.7% 383.00 ± 3% sched_debug.cfs_rq[22]:/.tg->runnable_avg 3184 ±291% -2728.3% -83710 ±-57% sched_debug.cfs_rq[23]:/.spread0 541.50 ± 11% -28.9% 385.25 ± 3% sched_debug.cfs_rq[23]:/.tg->runnable_avg 3163 ± 3% +19.4% 3776 ± 8% sched_debug.cfs_rq[2]:/.exec_clock 0.75 ±110% +400.0% 3.75 ± 47% sched_debug.cfs_rq[2]:/.runnable_load_avg 555.50 ± 18% -20.2% 443.25 ± 14% sched_debug.cfs_rq[2]:/.tg->runnable_avg 8390 ± 2% +23.8% 10384 ± 8% sched_debug.cfs_rq[3]:/.min_vruntime 549.75 ± 18% -19.5% 442.75 ± 14% sched_debug.cfs_rq[3]:/.tg->runnable_avg 68.00 ± 40% +171.3% 184.50 ± 26% sched_debug.cfs_rq[4]:/.blocked_load_avg 2824 ± 14% +30.4% 3682 ± 9% sched_debug.cfs_rq[4]:/.exec_clock 545.50 ± 16% -22.6% 422.25 ± 11% sched_debug.cfs_rq[4]:/.tg->runnable_avg 69.75 ± 39% +171.0% 189.00 ± 25% sched_debug.cfs_rq[4]:/.tg_load_contrib 7824 ± 17% +24.4% 9733 ± 7% sched_debug.cfs_rq[5]:/.min_vruntime 518.25 ± 16% -22.7% 400.50 ± 11% sched_debug.cfs_rq[5]:/.tg->runnable_avg 1599 ± 51% -62.1% 605.50 ± 25% sched_debug.cfs_rq[7]:/.avg->runnable_avg_sum 15577 ± 38% -69.7% 4721 ± 67% sched_debug.cfs_rq[7]:/.exec_clock 37096 ± 35% -68.1% 11832 ± 56% sched_debug.cfs_rq[7]:/.min_vruntime 25768 ± 49% -408.1% -79404 ±-67% sched_debug.cfs_rq[7]:/.spread0 491.50 ± 16% -27.1% 358.50 ± 10% sched_debug.cfs_rq[7]:/.tg->runnable_avg 34.00 ± 52% -64.0% 12.25 ± 27% sched_debug.cfs_rq[7]:/.tg_runnable_contrib 1949 ± 23% -55.6% 865.75 ± 70% sched_debug.cfs_rq[8]:/.avg->runnable_avg_sum 25183 ± 51% -54.8% 11375 ± 44% sched_debug.cfs_rq[8]:/.min_vruntime 13855 ± 94% -676.4% -79862 ±-64% sched_debug.cfs_rq[8]:/.spread0 492.25 ± 16% -28.9% 349.75 ± 6% sched_debug.cfs_rq[8]:/.tg->runnable_avg 41.25 ± 24% -57.0% 17.75 ± 73% sched_debug.cfs_rq[8]:/.tg_runnable_contrib 494.00 ± 16% -28.6% 352.50 ± 6% sched_debug.cfs_rq[9]:/.tg->runnable_avg 46021 ± 28% +155.5% 117598 ± 31% sched_debug.cpu#0.nr_load_updates -52.75 ± -8% +26.5% -66.75 ± -6% sched_debug.cpu#0.nr_uninterruptible 77978 ± 37% +70.5% 132942 ± 22% sched_debug.cpu#0.ttwu_count 33164 ± 42% +141.3% 80030 ± 30% sched_debug.cpu#0.ttwu_local 29131 ± 2% +18.7% 34578 ± 4% sched_debug.cpu#1.nr_load_updates 34945 ± 55% +82.2% 63667 ± 22% sched_debug.cpu#1.ttwu_count 53444 ± 11% -32.0% 36339 ± 18% sched_debug.cpu#10.nr_load_updates 254928 ± 19% -37.0% 160497 ± 36% sched_debug.cpu#10.nr_switches 255022 ± 19% -37.0% 160586 ± 36% sched_debug.cpu#10.sched_count 124863 ± 20% -37.9% 77546 ± 37% sched_debug.cpu#10.sched_goidle 62350 ± 6% -35.1% 40434 ± 25% sched_debug.cpu#10.ttwu_count 24043 ± 13% -39.6% 14532 ± 27% sched_debug.cpu#10.ttwu_local 48349 ± 12% -25.1% 36207 ± 17% sched_debug.cpu#11.nr_load_updates 220073 ± 20% -43.2% 125091 ± 24% sched_debug.cpu#11.nr_switches 220176 ± 20% -43.1% 125175 ± 24% sched_debug.cpu#11.sched_count 107249 ± 20% -43.9% 60128 ± 25% sched_debug.cpu#11.sched_goidle 54997 ± 15% -39.9% 33028 ± 15% sched_debug.cpu#11.ttwu_count 11715 ± 2% +14.5% 13414 ± 3% sched_debug.cpu#14.nr_load_updates 11624 ± 2% +14.1% 13262 ± 6% sched_debug.cpu#15.nr_load_updates 26743 ± 35% +214.0% 83986 ± 33% sched_debug.cpu#15.ttwu_count 11460 ± 3% +13.2% 12974 ± 4% sched_debug.cpu#16.nr_load_updates 11441 ± 5% +10.9% 12687 ± 3% sched_debug.cpu#17.nr_load_updates 3016 ± 3% -7.4% 2793 ± 1% sched_debug.cpu#17.ttwu_local 29953 ± 6% +22.4% 36676 ± 2% sched_debug.cpu#2.nr_load_updates 7674 ± 8% +32.4% 10159 ± 4% sched_debug.cpu#2.ttwu_local 12920 ± 13% +206.5% 39603 ± 17% sched_debug.cpu#20.nr_switches 12984 ± 13% +205.5% 39667 ± 17% sched_debug.cpu#20.sched_count 5279 ± 16% +249.0% 18423 ± 19% sched_debug.cpu#20.sched_goidle 80631 ± 35% -54.4% 36729 ± 33% sched_debug.cpu#21.ttwu_count 11221 ± 2% +10.5% 12400 ± 1% sched_debug.cpu#22.nr_load_updates 9.00 ± 78% -61.1% 3.50 ±173% sched_debug.cpu#23.cpu_load[2] 7.50 ± 86% -73.3% 2.00 ±173% sched_debug.cpu#23.cpu_load[3] 29370 ± 4% +29.7% 38100 ± 6% sched_debug.cpu#3.nr_load_updates 7498 ± 7% +37.2% 10286 ± 7% sched_debug.cpu#3.ttwu_local 29911 ± 12% +24.2% 37158 ± 4% sched_debug.cpu#4.nr_load_updates 140120 ± 24% +59.9% 224002 ± 18% sched_debug.cpu#4.nr_switches 140205 ± 24% +59.8% 224099 ± 18% sched_debug.cpu#4.sched_count 68431 ± 24% +60.7% 109965 ± 18% sched_debug.cpu#4.sched_goidle 7544 ± 7% +31.1% 9887 ± 4% sched_debug.cpu#4.ttwu_local 29295 ± 7% +21.4% 35577 ± 6% sched_debug.cpu#5.nr_load_updates 23076 ± 21% +43.1% 33020 ± 25% sched_debug.cpu#5.ttwu_count 7228 ± 4% +34.2% 9700 ± 6% sched_debug.cpu#5.ttwu_local 6.00 ± 83% -75.0% 1.50 ±173% sched_debug.cpu#6.cpu_load[2] 60164 ± 14% -40.5% 35793 ± 19% sched_debug.cpu#7.nr_load_updates 3.75 ±175% -240.0% -5.25 ±-69% sched_debug.cpu#7.nr_uninterruptible 53640 ± 18% -21.4% 42139 ± 16% sched_debug.cpu#7.ttwu_count 28043 ± 19% -51.0% 13744 ± 29% sched_debug.cpu#7.ttwu_local 6.75 ± 58% -77.8% 1.50 ±173% sched_debug.cpu#8.cpu_load[4] 51965 ± 15% -31.7% 35508 ± 9% sched_debug.cpu#8.nr_load_updates 22902 ± 25% -40.6% 13615 ± 12% sched_debug.cpu#8.ttwu_local 51893 ± 13% -29.9% 36371 ± 23% sched_debug.cpu#9.nr_load_updates 56687 ± 12% -40.8% 33540 ± 43% sched_debug.cpu#9.ttwu_count 22884 ± 18% -39.3% 13901 ± 39% sched_debug.cpu#9.ttwu_local bay: Pentium D Memory: 2G lkp-ws02: Westmere-EP Memory: 16G To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Ying Huang View attachment "job.yaml" of type "text/plain" (3421 bytes) View attachment "reproduce" of type "text/plain" (405 bytes)
Powered by blists - more mailing lists