[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202210271720.1af1ecc3-yujie.liu@intel.com>
Date: Thu, 27 Oct 2022 17:35:39 +0800
From: kernel test robot <yujie.liu@...el.com>
To: Jan Kara <jack@...e.cz>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
Theodore Ts'o <tytso@....edu>,
Stefan Wahren <stefan.wahren@...e.com>,
Ojaswin Mujoo <ojaswin@...ux.ibm.com>,
<linux-kernel@...r.kernel.org>, <linux-ext4@...r.kernel.org>,
<ying.huang@...el.com>, <feng.tang@...el.com>,
<zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>
Subject: [linus:master] [ext4] 613c5a8589:
fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec 36.0% improvement
Greeting,
FYI, we noticed a 36.0% improvement of fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec due to commit:
commit: 613c5a85898d1cd44e68f28d65eccf64a8ace9cf ("ext4: make directory inode spreading reflect flexbg size")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
in testcase: fxmark
on test machine: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
with following parameters:
disk: 1SSD
media: ssd
test: dbench_client
fstype: ext4_no_jnl
directio: directio
cpufreq_governor: performance
test-description: FxMark is a filesystem benchmark that test multicore scalability.
test-url: https://github.com/sslab-gatech/fxmark
Details are as below:
=========================================================================================
compiler/cpufreq_governor/directio/disk/fstype/kconfig/media/rootfs/tbox_group/test/testcase:
gcc-11/performance/directio/1SSD/ext4_no_jnl/x86_64-rhel-8.3/ssd/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp5/dbench_client/fxmark
commit:
1940265ede ("ext4: avoid unnecessary spreading of allocations among groups")
613c5a8589 ("ext4: make directory inode spreading reflect flexbg size")
1940265ede6683f6 613c5a85898d1cd44e68f28d65e
---------------- ---------------------------
fail:runs %reproduction fail:runs
| | |
:6 100% 6:6 stderr.Couldn't_synthesize_bpf_events
%stddev %change %stddev
\ | \
323.96 ± 12% +28.7% 417.02 ± 7% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.idle_sec
7.50 ± 12% +28.9% 9.66 ± 7% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.idle_util
1583 ± 3% +30.8% 2072 ± 2% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.iowait_sec
36.66 ± 3% +31.0% 48.02 ± 2% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.iowait_util
5.60 ± 3% +27.3% 7.14 fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.softirq_sec
0.13 ± 3% +27.5% 0.17 fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.softirq_util
2348 ± 3% -25.4% 1752 ± 4% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.sys_sec
54.36 ± 3% -25.3% 40.61 ± 4% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.sys_util
27.58 ± 4% +21.0% 33.36 ± 3% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.user_sec
0.64 ± 4% +21.1% 0.77 ± 3% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.user_util
1685 ± 4% +36.0% 2292 ± 3% fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec
3.054e+08 +3.3% 3.154e+08 fxmark.time.file_system_outputs
1032 ± 2% -8.7% 942.00 fxmark.time.percent_of_cpu_this_job_got
6151 ± 2% -9.2% 5584 fxmark.time.system_time
154.57 +3.8% 160.38 fxmark.time.user_time
8290165 +5.5% 8747536 fxmark.time.voluntary_context_switches
6242070 ± 2% +9.6% 6844007 ± 2% numa-numastat.node1.numa_hit
33.42 +3.5% 34.59 iostat.cpu.iowait
36.41 -3.1% 35.27 iostat.cpu.system
4347817 -24.5% 3281458 vmstat.memory.cache
45757 +4.7% 47892 vmstat.system.cs
38.17 -9.1% 34.69 boot-time.boot
23.34 -15.2% 19.79 boot-time.dhcp
4374 -10.1% 3934 boot-time.idle
0.29 ± 10% +16.3% 0.33 ± 6% sched_debug.cfs_rq:/.h_nr_running.stddev
3361937 ± 4% -14.6% 2870581 ± 3% sched_debug.cfs_rq:/.min_vruntime.avg
3453968 ± 4% -13.8% 2976035 ± 2% sched_debug.cfs_rq:/.min_vruntime.max
3074025 ± 4% -16.7% 2559836 ± 6% sched_debug.cfs_rq:/.min_vruntime.min
4234791 -25.1% 3170271 meminfo.Cached
13280609 ± 8% -23.2% 10193251 ± 9% meminfo.DirectMap2M
5623315 -19.1% 4546975 meminfo.Memused
3825966 -27.8% 2760633 meminfo.Unevictable
7683659 -13.9% 6614519 meminfo.max_used_kB
45394 +4.8% 47577 perf-stat.i.context-switches
878.12 ± 2% +7.8% 946.23 ± 3% perf-stat.i.cpu-migrations
0.00 ±116% +0.0 0.00 ± 66% perf-stat.i.dTLB-load-miss-rate%
45558 +4.8% 47762 perf-stat.ps.context-switches
886.77 ± 2% +7.5% 953.70 ± 3% perf-stat.ps.cpu-migrations
0.49 ± 19% -0.3 0.18 ± 52% perf-profile.children.cycles-pp.get_cpu_device
0.34 ± 24% -0.1 0.20 ± 20% perf-profile.children.cycles-pp.s_next
0.06 ±114% +0.1 0.14 ± 34% perf-profile.children.cycles-pp.account_process_tick
0.48 ± 20% -0.3 0.16 ± 54% perf-profile.self.cycles-pp.get_cpu_device
0.05 ±141% +0.1 0.14 ± 41% perf-profile.self.cycles-pp.account_process_tick
0.40 ± 9% +0.1 0.50 ± 13% perf-profile.self.cycles-pp.tick_nohz_next_event
109190 ± 30% -60.3% 43316 ± 57% numa-meminfo.node0.AnonHugePages
154521 ± 26% -43.2% 87783 ± 27% numa-meminfo.node0.AnonPages
437517 ± 8% -16.8% 364128 ± 6% numa-meminfo.node0.Inactive
160530 ± 25% -41.4% 94026 ± 25% numa-meminfo.node0.Inactive(anon)
87985 ± 39% +73.3% 152466 ± 16% numa-meminfo.node1.AnonHugePages
157511 ± 26% +41.4% 222774 ± 10% numa-meminfo.node1.AnonPages
207719 ± 17% +34.9% 280314 ± 9% numa-meminfo.node1.Inactive
161988 ± 25% +40.3% 227216 ± 10% numa-meminfo.node1.Inactive(anon)
38635 ± 26% -43.2% 21947 ± 27% numa-vmstat.node0.nr_anon_pages
40136 ± 25% -41.4% 23505 ± 25% numa-vmstat.node0.nr_inactive_anon
40136 ± 25% -41.4% 23505 ± 25% numa-vmstat.node0.nr_zone_inactive_anon
39376 ± 26% +41.5% 55699 ± 10% numa-vmstat.node1.nr_anon_pages
6131790 ± 2% +10.0% 6744942 ± 2% numa-vmstat.node1.nr_dirtied
40495 ± 25% +40.3% 56809 ± 10% numa-vmstat.node1.nr_inactive_anon
40495 ± 25% +40.3% 56809 ± 10% numa-vmstat.node1.nr_zone_inactive_anon
6241898 ± 2% +9.6% 6843576 ± 2% numa-vmstat.node1.numa_hit
38173261 +3.3% 39428621 proc-vmstat.nr_dirtied
1059388 -25.1% 793340 proc-vmstat.nr_file_pages
53799 -1.9% 52794 proc-vmstat.nr_slab_unreclaimable
956491 -27.8% 690158 proc-vmstat.nr_unevictable
9850651 +3.5% 10194926 proc-vmstat.nr_written
956491 -27.8% 690158 proc-vmstat.nr_zone_unevictable
38481863 +3.2% 39707334 proc-vmstat.numa_hit
38362225 +3.2% 39589589 proc-vmstat.numa_local
1333657 +2.1% 1361550 proc-vmstat.pgactivate
38472842 +3.2% 39705274 proc-vmstat.pgalloc_normal
38505715 +3.2% 39737694 proc-vmstat.pgfree
66251245 +2.3% 67801130 proc-vmstat.pgpgout
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://01.org/lkp
View attachment "config-6.0.0-rc4-00003-g613c5a85898d" of type "text/plain" (163956 bytes)
View attachment "job-script" of type "text/plain" (8237 bytes)
View attachment "job.yaml" of type "text/plain" (5672 bytes)
View attachment "reproduce" of type "text/plain" (254 bytes)
Powered by blists - more mailing lists