[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <202211220854.48dad6fd-oliver.sang@intel.com>
Date: Tue, 22 Nov 2022 09:09:34 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Gao Xiang <hsiangkao@...ux.alibaba.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
Zirong Lang <zlang@...hat.com>, <linux-xfs@...r.kernel.org>,
<ying.huang@...el.com>, <feng.tang@...el.com>,
<zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>,
"Darrick J. Wong" <djwong@...nel.org>,
Dave Chinner <dchinner@...hat.com>,
Brian Foster <bfoster@...hat.com>,
LKML <linux-kernel@...r.kernel.org>,
Gao Xiang <hsiangkao@...ux.alibaba.com>
Subject: Re: [PATCH] xfs: account extra freespace btree splits for multiple
allocations
please be noted we noticed Gao Xiang and Dave Chinner have already had lots of
discussion around this patch, which seems there is maybe new version later.
we just sent out this report FYI the possible performance impact of this patch.
Greeting,
FYI, we noticed a -15.1% regression of fxmark.ssd_xfs_MWCM_72_directio.works/sec due to commit:
commit: ada76f94c5b32c1828955d70f692b111a9ec2d33 ("[PATCH] xfs: account extra freespace btree splits for multiple allocations")
url: https://github.com/intel-lab-lkp/linux/commits/Gao-Xiang/xfs-account-extra-freespace-btree-splits-for-multiple-allocations/20221109-114953
base: https://git.kernel.org/cgit/fs/xfs/xfs-linux.git for-next
patch link: https://lore.kernel.org/all/20221109034802.40322-1-hsiangkao@linux.alibaba.com/
patch subject: [PATCH] xfs: account extra freespace btree splits for multiple allocations
in testcase: fxmark
on test machine: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
with following parameters:
disk: 1SSD
media: ssd
test: MWCM
fstype: xfs
directio: directio
cpufreq_governor: performance
test-description: FxMark is a filesystem benchmark that test multicore scalability.
test-url: https://github.com/sslab-gatech/fxmark
If you fix the issue, kindly add following tag
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Link: https://lore.kernel.org/oe-lkp/202211220854.48dad6fd-oliver.sang@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
=========================================================================================
compiler/cpufreq_governor/directio/disk/fstype/kconfig/media/rootfs/tbox_group/test/testcase:
gcc-11/performance/directio/1SSD/xfs/x86_64-rhel-8.3/ssd/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp5/MWCM/fxmark
commit:
4eb559dd15 ("Merge tag 'refcount-cow-domain-6.1_2022-10-31' of git://git.kernel.org/pub/scm/linux/kernel/git/djwong/xfs-linux into xfs-6.1-fixesA")
ada76f94c5 ("xfs: account extra freespace btree splits for multiple allocations")
4eb559dd15671cca ada76f94c5b32c1828955d70f69
---------------- ---------------------------
%stddev %change %stddev
\ | \
260.92 +90.8% 497.93 ± 13% fxmark.ssd_xfs_MWCM_18_directio.idle_sec
29.10 +91.5% 55.74 ± 14% fxmark.ssd_xfs_MWCM_18_directio.idle_util
9.11 ± 2% -17.6% 7.51 ± 8% fxmark.ssd_xfs_MWCM_18_directio.iowait_sec
1.02 ± 2% -17.3% 0.84 ± 8% fxmark.ssd_xfs_MWCM_18_directio.iowait_util
617.96 -38.8% 378.15 ± 19% fxmark.ssd_xfs_MWCM_18_directio.sys_sec
68.93 -38.6% 42.31 ± 18% fxmark.ssd_xfs_MWCM_18_directio.sys_util
1.80 ± 4% -18.5% 1.47 ± 5% fxmark.ssd_xfs_MWCM_18_directio.user_sec
0.20 ± 4% -18.2% 0.16 ± 4% fxmark.ssd_xfs_MWCM_18_directio.user_util
3363535 -29.6% 2369575 ± 7% fxmark.ssd_xfs_MWCM_18_directio.works
67262 -29.6% 47384 ± 7% fxmark.ssd_xfs_MWCM_18_directio.works/sec
0.01 ±111% +51925.0% 3.47 ± 9% fxmark.ssd_xfs_MWCM_1_directio.idle_sec
0.01 ±111% +52382.0% 6.99 ± 9% fxmark.ssd_xfs_MWCM_1_directio.idle_util
0.23 ± 14% +1481.9% 3.64 ± 14% fxmark.ssd_xfs_MWCM_1_directio.iowait_sec
0.46 ± 14% +1496.0% 7.33 ± 14% fxmark.ssd_xfs_MWCM_1_directio.iowait_util
0.58 +20.9% 0.70 ± 2% fxmark.ssd_xfs_MWCM_1_directio.irq_sec
1.15 +21.9% 1.40 ± 2% fxmark.ssd_xfs_MWCM_1_directio.irq_util
0.37 ± 2% +18.9% 0.44 ± 5% fxmark.ssd_xfs_MWCM_1_directio.softirq_sec
0.74 ± 2% +20.0% 0.89 ± 5% fxmark.ssd_xfs_MWCM_1_directio.softirq_util
47.70 -15.4% 40.33 ± 2% fxmark.ssd_xfs_MWCM_1_directio.sys_sec
95.24 -14.7% 81.24 ± 2% fxmark.ssd_xfs_MWCM_1_directio.sys_util
1.20 ± 5% -11.0% 1.07 ± 4% fxmark.ssd_xfs_MWCM_1_directio.user_sec
2.40 ± 6% -10.2% 2.15 ± 4% fxmark.ssd_xfs_MWCM_1_directio.user_util
2639345 -15.6% 2228369 ± 4% fxmark.ssd_xfs_MWCM_1_directio.works
52786 -15.6% 44567 ± 4% fxmark.ssd_xfs_MWCM_1_directio.works/sec
4.11 ± 3% +254.5% 14.56 ± 18% fxmark.ssd_xfs_MWCM_2_directio.idle_sec
4.10 ± 3% +255.9% 14.60 ± 18% fxmark.ssd_xfs_MWCM_2_directio.idle_util
1.42 ± 6% +435.2% 7.57 ± 6% fxmark.ssd_xfs_MWCM_2_directio.iowait_sec
1.41 ± 6% +437.2% 7.59 ± 6% fxmark.ssd_xfs_MWCM_2_directio.iowait_util
0.82 +20.2% 0.99 ± 4% fxmark.ssd_xfs_MWCM_2_directio.irq_sec
0.82 +20.7% 0.99 ± 4% fxmark.ssd_xfs_MWCM_2_directio.irq_util
0.52 ± 4% +30.0% 0.67 ± 4% fxmark.ssd_xfs_MWCM_2_directio.softirq_sec
0.52 ± 4% +30.5% 0.67 ± 4% fxmark.ssd_xfs_MWCM_2_directio.softirq_util
91.86 -18.6% 74.82 ± 4% fxmark.ssd_xfs_MWCM_2_directio.sys_sec
91.73 -18.3% 74.99 ± 4% fxmark.ssd_xfs_MWCM_2_directio.sys_util
1.42 ± 4% -18.7% 1.15 ± 4% fxmark.ssd_xfs_MWCM_2_directio.user_sec
1.42 ± 4% -18.4% 1.16 ± 4% fxmark.ssd_xfs_MWCM_2_directio.user_util
3159498 -23.2% 2425056 ± 5% fxmark.ssd_xfs_MWCM_2_directio.works
63189 -23.3% 48486 ± 5% fxmark.ssd_xfs_MWCM_2_directio.works/sec
413.60 ± 4% +103.5% 841.63 ± 20% fxmark.ssd_xfs_MWCM_36_directio.idle_sec
22.94 ± 4% +103.9% 46.78 ± 20% fxmark.ssd_xfs_MWCM_36_directio.idle_util
1364 -31.6% 933.33 ± 19% fxmark.ssd_xfs_MWCM_36_directio.sys_sec
75.70 -31.5% 51.85 ± 18% fxmark.ssd_xfs_MWCM_36_directio.sys_util
1.82 ± 4% -11.7% 1.60 ± 5% fxmark.ssd_xfs_MWCM_36_directio.user_sec
0.10 ± 4% -11.5% 0.09 ± 5% fxmark.ssd_xfs_MWCM_36_directio.user_util
2850816 -22.9% 2197955 ± 7% fxmark.ssd_xfs_MWCM_36_directio.works
57006 -22.9% 43943 ± 7% fxmark.ssd_xfs_MWCM_36_directio.works/sec
54.22 +63.5% 88.65 ± 2% fxmark.ssd_xfs_MWCM_4_directio.idle_sec
27.33 +64.7% 45.02 ± 2% fxmark.ssd_xfs_MWCM_4_directio.idle_util
6.53 +36.1% 8.89 ± 8% fxmark.ssd_xfs_MWCM_4_directio.iowait_sec
3.29 +37.1% 4.51 ± 8% fxmark.ssd_xfs_MWCM_4_directio.iowait_util
1.37 +21.7% 1.67 ± 3% fxmark.ssd_xfs_MWCM_4_directio.irq_sec
0.69 +22.6% 0.85 ± 3% fxmark.ssd_xfs_MWCM_4_directio.irq_util
0.64 +20.3% 0.77 ± 3% fxmark.ssd_xfs_MWCM_4_directio.softirq_sec
0.32 +21.2% 0.39 ± 3% fxmark.ssd_xfs_MWCM_4_directio.softirq_util
134.06 -28.6% 95.67 ± 2% fxmark.ssd_xfs_MWCM_4_directio.sys_sec
67.58 -28.1% 48.59 ± 2% fxmark.ssd_xfs_MWCM_4_directio.sys_util
1.55 ± 2% -18.4% 1.26 ± 5% fxmark.ssd_xfs_MWCM_4_directio.user_sec
0.78 ± 2% -17.8% 0.64 ± 5% fxmark.ssd_xfs_MWCM_4_directio.user_util
3262962 -26.9% 2383952 fxmark.ssd_xfs_MWCM_4_directio.works
65256 -26.9% 47670 fxmark.ssd_xfs_MWCM_4_directio.works/sec
489.98 ± 3% +160.6% 1277 ± 24% fxmark.ssd_xfs_MWCM_54_directio.idle_sec
18.11 ± 3% +161.0% 47.27 ± 24% fxmark.ssd_xfs_MWCM_54_directio.idle_util
2185 -36.3% 1391 ± 22% fxmark.ssd_xfs_MWCM_54_directio.sys_sec
80.77 -36.2% 51.49 ± 22% fxmark.ssd_xfs_MWCM_54_directio.sys_util
2448443 -24.0% 1859708 ± 9% fxmark.ssd_xfs_MWCM_54_directio.works
48963 -24.1% 37175 ± 9% fxmark.ssd_xfs_MWCM_54_directio.works/sec
614.60 ± 2% +88.5% 1158 ± 9% fxmark.ssd_xfs_MWCM_72_directio.idle_sec
17.04 ± 2% +88.6% 32.13 ± 9% fxmark.ssd_xfs_MWCM_72_directio.idle_util
6.44 ± 4% -14.8% 5.48 ± 9% fxmark.ssd_xfs_MWCM_72_directio.iowait_sec
0.18 ± 4% -14.8% 0.15 ± 10% fxmark.ssd_xfs_MWCM_72_directio.iowait_util
2953 -18.5% 2407 ± 4% fxmark.ssd_xfs_MWCM_72_directio.sys_sec
81.89 -18.5% 66.77 ± 4% fxmark.ssd_xfs_MWCM_72_directio.sys_util
2279692 -15.1% 1935171 ± 2% fxmark.ssd_xfs_MWCM_72_directio.works
45589 -15.1% 38687 ± 2% fxmark.ssd_xfs_MWCM_72_directio.works/sec
934.32 -10.3% 838.20 fxmark.time.elapsed_time
934.32 -10.3% 838.20 fxmark.time.elapsed_time.max
43490089 -20.8% 34446960 ± 3% fxmark.time.file_system_outputs
128601 ± 3% -31.6% 87922 ± 4% fxmark.time.involuntary_context_switches
31.83 -17.8% 26.17 ± 5% fxmark.time.percent_of_cpu_this_job_got
301.24 -26.5% 221.43 ± 5% fxmark.time.system_time
841264 ± 3% -11.2% 747133 ± 5% fxmark.time.voluntary_context_switches
989.01 -9.6% 893.57 uptime.boot
54.77 +10.7% 60.61 iostat.cpu.idle
5.31 ± 4% +21.4% 6.45 ± 4% iostat.cpu.iowait
38.42 -18.0% 31.51 ± 2% iostat.cpu.system
55.66 +5.8 61.45 mpstat.cpu.all.idle%
5.31 ± 5% +1.2 6.46 ± 4% mpstat.cpu.all.iowait%
0.60 ± 2% -0.1 0.53 mpstat.cpu.all.soft%
36.50 -6.6 29.87 ± 3% mpstat.cpu.all.sys%
26496967 -20.9% 20947493 ± 2% numa-numastat.node0.local_node
26548700 -20.8% 21022864 ± 2% numa-numastat.node0.numa_hit
3332141 ± 6% -22.1% 2595871 ± 4% numa-numastat.node1.local_node
3400117 ± 5% -22.4% 2638708 ± 5% numa-numastat.node1.numa_hit
2211 +1.0% 2233 perf-stat.i.minor-faults
0.39 ± 9% +0.1 0.44 ± 3% perf-stat.i.node-load-miss-rate%
2211 +1.0% 2233 perf-stat.i.page-faults
2209 +1.0% 2230 perf-stat.ps.minor-faults
2209 +1.0% 2230 perf-stat.ps.page-faults
53.83 +11.5% 60.00 vmstat.cpu.id
5430 ± 2% -14.0% 4668 ± 3% vmstat.io.bi
123556 -8.3% 113294 ± 2% vmstat.io.bo
11181852 -16.7% 9318758 ± 2% vmstat.memory.cache
7.67 ± 6% -19.6% 6.17 ± 11% vmstat.procs.r
9247281 -15.1% 7847381 meminfo.Cached
3299044 ± 2% -21.5% 2591082 ± 2% meminfo.Dirty
5715841 ± 2% -24.4% 4320653 ± 3% meminfo.Inactive
5389224 ± 2% -25.9% 3991807 ± 3% meminfo.Inactive(file)
1914390 -24.3% 1449759 ± 3% meminfo.KReclaimable
31406 ± 13% -22.0% 24510 ± 7% meminfo.KernelStack
14111361 -16.6% 11775228 ± 2% meminfo.Memused
1914390 -24.3% 1449759 ± 3% meminfo.SReclaimable
1427015 -25.6% 1061962 ± 4% meminfo.SUnreclaim
18397 ± 2% -8.4% 16846 meminfo.Shmem
3341406 -24.8% 2511722 ± 3% meminfo.Slab
24035030 -18.6% 19557858 ± 2% meminfo.max_used_kB
1.23 ± 22% +0.4 1.58 ± 18% perf-profile.calltrace.cycles-pp.xlog_cil_push_work.process_one_work.worker_thread.kthread.ret_from_fork
0.76 ± 44% -0.4 0.38 ± 74% perf-profile.children.cycles-pp.rcu_sched_clock_irq
0.09 ± 10% -0.0 0.05 ± 47% perf-profile.children.cycles-pp.mod_find
0.10 ± 47% +0.1 0.18 ± 26% perf-profile.children.cycles-pp.xfs_iext_lookup_extent
0.17 ± 26% +0.1 0.25 ± 16% perf-profile.children.cycles-pp.s_show
0.05 ± 82% +0.1 0.13 ± 46% perf-profile.children.cycles-pp.timekeeping_advance
0.05 ± 82% +0.1 0.13 ± 46% perf-profile.children.cycles-pp.update_wall_time
0.21 ± 21% +0.1 0.31 ± 20% perf-profile.children.cycles-pp.vsnprintf
0.20 ± 26% +0.1 0.30 ± 17% perf-profile.children.cycles-pp.seq_printf
1.23 ± 22% +0.4 1.58 ± 18% perf-profile.children.cycles-pp.xlog_cil_push_work
0.08 ± 14% -0.0 0.04 ± 72% perf-profile.self.cycles-pp.mod_find
0.14 ± 26% +0.1 0.22 ± 16% perf-profile.self.cycles-pp.xlog_cil_push_work
0.13 ± 37% +0.1 0.21 ± 22% perf-profile.self.cycles-pp.__folio_end_writeback
1339 ± 6% -12.1% 1177 ± 6% numa-meminfo.node0.Active
1317 ± 4% -10.6% 1177 ± 6% numa-meminfo.node0.Active(anon)
2938921 ± 2% -22.3% 2284083 ± 2% numa-meminfo.node0.Dirty
4950896 ± 2% -25.3% 3699254 ± 2% numa-meminfo.node0.Inactive
4806022 ± 2% -27.0% 3510366 ± 3% numa-meminfo.node0.Inactive(file)
1663428 ± 2% -25.0% 1246833 ± 3% numa-meminfo.node0.KReclaimable
15585 ± 4% -12.4% 13655 ± 4% numa-meminfo.node0.KernelStack
1663428 ± 2% -25.0% 1246833 ± 3% numa-meminfo.node0.SReclaimable
1224137 ± 2% -25.5% 912586 ± 4% numa-meminfo.node0.SUnreclaim
15785 ± 5% -12.5% 13811 ± 4% numa-meminfo.node0.Shmem
2887566 -25.2% 2159421 ± 3% numa-meminfo.node0.Slab
358435 ± 3% -14.3% 307335 ± 6% numa-meminfo.node1.Dirty
764404 ± 6% -18.6% 622071 ± 10% numa-meminfo.node1.Inactive
582651 ± 2% -17.3% 482089 ± 9% numa-meminfo.node1.Inactive(file)
250653 ± 10% -18.9% 203397 ± 13% numa-meminfo.node1.KReclaimable
3410163 ± 54% -26.4% 2508459 ± 68% numa-meminfo.node1.MemUsed
250653 ± 10% -18.9% 203397 ± 13% numa-meminfo.node1.SReclaimable
202812 ± 17% -26.2% 149763 ± 5% numa-meminfo.node1.SUnreclaim
453467 ± 9% -22.1% 353161 ± 8% numa-meminfo.node1.Slab
0.71 ± 14% -21.5% 0.56 ± 14% sched_debug.cfs_rq:/.h_nr_running.avg
1405 ± 55% -68.3% 445.08 ± 27% sched_debug.cfs_rq:/.load_avg.avg
5023746 -33.8% 3324248 ± 13% sched_debug.cfs_rq:/.min_vruntime.avg
5080937 -33.2% 3395209 ± 12% sched_debug.cfs_rq:/.min_vruntime.max
4683128 -31.3% 3217873 ± 13% sched_debug.cfs_rq:/.min_vruntime.min
0.59 ± 9% -17.4% 0.49 ± 18% sched_debug.cfs_rq:/.nr_running.avg
650.51 ± 13% -32.3% 440.38 ± 9% sched_debug.cfs_rq:/.runnable_avg.avg
1007 ± 8% -18.5% 821.27 ± 6% sched_debug.cfs_rq:/.runnable_avg.max
476.07 ± 17% -38.7% 291.80 ± 16% sched_debug.cfs_rq:/.runnable_avg.min
517.96 ± 5% -20.3% 412.57 ± 9% sched_debug.cfs_rq:/.util_avg.avg
352.69 ± 5% -26.1% 260.66 ± 14% sched_debug.cfs_rq:/.util_avg.min
113.13 ± 9% -26.8% 82.85 ± 15% sched_debug.cfs_rq:/.util_est_enqueued.stddev
18547 ± 4% -9.5% 16791 ± 3% sched_debug.cpu.curr->pid.max
1549 ± 11% +1106.2% 18692 ±123% sched_debug.cpu.max_idle_balance_cost.stddev
0.70 ± 16% -25.1% 0.52 ± 18% sched_debug.cpu.nr_running.avg
862357 ± 4% -16.3% 722142 ± 3% sched_debug.cpu.nr_switches.avg
1440368 ± 6% -22.2% 1120203 ± 6% sched_debug.cpu.nr_switches.max
756963 ± 4% -14.6% 646359 ± 5% sched_debug.cpu.nr_switches.min
150642 ± 10% -33.8% 99651 ± 18% sched_debug.cpu.nr_switches.stddev
17687949 -23.3% 13563229 ± 2% numa-vmstat.node0.nr_dirtied
734675 ± 2% -22.3% 570784 ± 2% numa-vmstat.node0.nr_dirty
1201580 ± 2% -27.0% 877113 ± 3% numa-vmstat.node0.nr_inactive_file
15586 ± 4% -12.4% 13655 ± 4% numa-vmstat.node0.nr_kernel_stack
3945 ± 5% -12.5% 3453 ± 4% numa-vmstat.node0.nr_shmem
415866 ± 2% -25.1% 311497 ± 3% numa-vmstat.node0.nr_slab_reclaimable
306030 ± 2% -25.5% 228011 ± 4% numa-vmstat.node0.nr_slab_unreclaimable
17687949 -23.3% 13563229 ± 2% numa-vmstat.node0.nr_written
1201580 ± 2% -27.0% 877113 ± 3% numa-vmstat.node0.nr_zone_inactive_file
734697 ± 2% -22.3% 570821 ± 2% numa-vmstat.node0.nr_zone_write_pending
26548703 -20.8% 21023488 ± 2% numa-vmstat.node0.numa_hit
26496970 -20.9% 20948116 ± 2% numa-vmstat.node0.numa_local
2316618 -20.7% 1836559 ± 5% numa-vmstat.node1.nr_dirtied
89623 ± 3% -14.3% 76800 ± 6% numa-vmstat.node1.nr_dirty
145701 ± 2% -17.3% 120486 ± 9% numa-vmstat.node1.nr_inactive_file
62668 ± 10% -18.9% 50828 ± 13% numa-vmstat.node1.nr_slab_reclaimable
50710 ± 17% -26.2% 37434 ± 5% numa-vmstat.node1.nr_slab_unreclaimable
2316618 -20.7% 1836559 ± 5% numa-vmstat.node1.nr_written
145701 ± 2% -17.3% 120486 ± 9% numa-vmstat.node1.nr_zone_inactive_file
89584 ± 3% -14.3% 76792 ± 6% numa-vmstat.node1.nr_zone_write_pending
3400024 ± 5% -22.4% 2639087 ± 5% numa-vmstat.node1.numa_hit
3332048 ± 6% -22.1% 2596249 ± 4% numa-vmstat.node1.numa_local
77508 +1.2% 78432 proc-vmstat.nr_anon_pages
20004568 -23.0% 15399788 ± 2% proc-vmstat.nr_dirtied
825597 ± 2% -21.4% 648654 ± 2% proc-vmstat.nr_dirty
2311462 -15.1% 1961741 proc-vmstat.nr_file_pages
29390035 +2.0% 29973527 proc-vmstat.nr_free_pages
1346949 ± 2% -25.9% 997847 ± 3% proc-vmstat.nr_inactive_file
31402 ± 13% -21.9% 24514 ± 7% proc-vmstat.nr_kernel_stack
4599 ± 2% -8.4% 4211 proc-vmstat.nr_shmem
478540 -24.3% 362476 ± 3% proc-vmstat.nr_slab_reclaimable
356646 -25.6% 265413 ± 4% proc-vmstat.nr_slab_unreclaimable
20004568 -23.0% 15399788 ± 2% proc-vmstat.nr_written
1346949 ± 2% -25.9% 997847 ± 3% proc-vmstat.nr_zone_inactive_file
825584 ± 2% -21.4% 648681 ± 2% proc-vmstat.nr_zone_write_pending
29951284 -21.0% 23664558 ± 2% proc-vmstat.numa_hit
29831574 -21.1% 23546349 ± 2% proc-vmstat.numa_local
51777954 -21.0% 40906350 ± 2% proc-vmstat.pgalloc_normal
5920 ± 2% -7.6% 5472 ± 2% proc-vmstat.pgdeactivate
2342027 -8.5% 2142047 proc-vmstat.pgfault
51819442 -21.0% 40950599 ± 2% proc-vmstat.pgfree
5071045 ± 3% -22.9% 3908046 ± 4% proc-vmstat.pgpgin
1.153e+08 -17.9% 94738236 proc-vmstat.pgpgout
141611 -6.3% 132656 proc-vmstat.pgreuse
10386182 ± 3% -24.2% 7877205 ± 4% proc-vmstat.pgrotated
36803772 -24.1% 27916664 ± 2% proc-vmstat.slabs_scanned
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://01.org/lkp
View attachment "config-6.1.0-rc1-00031-gada76f94c5b3" of type "text/plain" (166032 bytes)
View attachment "job-script" of type "text/plain" (8315 bytes)
View attachment "job.yaml" of type "text/plain" (5576 bytes)
View attachment "reproduce" of type "text/plain" (254 bytes)
Powered by blists - more mailing lists