lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Tue, 22 Nov 2022 18:03:03 +0800
From:   Oliver Sang <oliver.sang@...el.com>
To:     Gao Xiang <hsiangkao@...ux.alibaba.com>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        Zirong Lang <zlang@...hat.com>, <linux-xfs@...r.kernel.org>,
        <ying.huang@...el.com>, <feng.tang@...el.com>,
        <zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>,
        "Darrick J. Wong" <djwong@...nel.org>,
        Dave Chinner <dchinner@...hat.com>,
        Brian Foster <bfoster@...hat.com>,
        LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] xfs: account extra freespace btree splits for multiple
 allocations

hi Gao Xiang,

On Tue, Nov 22, 2022 at 09:33:38AM +0800, Gao Xiang wrote:
> On Tue, Nov 22, 2022 at 09:09:34AM +0800, kernel test robot wrote:
> > 
> > please be noted we noticed Gao Xiang and Dave Chinner have already had lots of
> > discussion around this patch, which seems there is maybe new version later.
> > we just sent out this report FYI the possible performance impact of this patch.
> > 
> > 
> > Greeting,
> > 
> > FYI, we noticed a -15.1% regression of fxmark.ssd_xfs_MWCM_72_directio.works/sec due to commit:
> 
> Thanks for your report!
> 
> At a glance, I have no idea why this commit can have performance
> impacts.  Is the result stable?

in our tests, the result is quite stable.
     45589           -15.1%      38687 ±  2%  fxmark.ssd_xfs_MWCM_72_directio.works/sec

and detail data is as below:
for this commit:
  "fxmark.ssd_xfs_MWCM_72_directio.works/sec": [
    39192.224368,
    39665.690567,
    38980.680601,
    37298.99538,
    37483.256377,
    39504.606569
  ],

for parent:
  "fxmark.ssd_xfs_MWCM_72_directio.works/sec": [
    45381.458009,
    45314.376204,
    45724.688965,
    45751.955937,
    45614.323267,
    45747.216475
  ],


if you still have concern, we could rerun tests. Thanks!


> 
> Thanks,
> Gao Xiang
> 
> > 
> > 
> > commit: ada76f94c5b32c1828955d70f692b111a9ec2d33 ("[PATCH] xfs: account extra freespace btree splits for multiple allocations")
> > url: https://github.com/intel-lab-lkp/linux/commits/Gao-Xiang/xfs-account-extra-freespace-btree-splits-for-multiple-allocations/20221109-114953
> > base: https://git.kernel.org/cgit/fs/xfs/xfs-linux.git for-next
> > patch link: https://lore.kernel.org/all/20221109034802.40322-1-hsiangkao@linux.alibaba.com/
> > patch subject: [PATCH] xfs: account extra freespace btree splits for multiple allocations
> > 
> > in testcase: fxmark
> > on test machine: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
> > with following parameters:
> > 
> > 	disk: 1SSD
> > 	media: ssd
> > 	test: MWCM
> > 	fstype: xfs
> > 	directio: directio
> > 	cpufreq_governor: performance
> > 
> > test-description: FxMark is a filesystem benchmark that test multicore scalability.
> > test-url: https://github.com/sslab-gatech/fxmark
> > 
> > 
> > 
> > If you fix the issue, kindly add following tag
> > | Reported-by: kernel test robot <oliver.sang@...el.com>
> > | Link: https://lore.kernel.org/oe-lkp/202211220854.48dad6fd-oliver.sang@intel.com
> > 
> > 
> > Details are as below:
> > -------------------------------------------------------------------------------------------------->
> > 
> > 
> > To reproduce:
> > 
> >         git clone https://github.com/intel/lkp-tests.git
> >         cd lkp-tests
> >         sudo bin/lkp install job.yaml           # job file is attached in this email
> >         bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
> >         sudo bin/lkp run generated-yaml-file
> > 
> >         # if come across any failure that blocks the test,
> >         # please remove ~/.lkp and /lkp dir to run from a clean state.
> > 
> > =========================================================================================
> > compiler/cpufreq_governor/directio/disk/fstype/kconfig/media/rootfs/tbox_group/test/testcase:
> >   gcc-11/performance/directio/1SSD/xfs/x86_64-rhel-8.3/ssd/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp5/MWCM/fxmark
> > 
> > commit: 
> >   4eb559dd15 ("Merge tag 'refcount-cow-domain-6.1_2022-10-31' of git://git.kernel.org/pub/scm/linux/kernel/git/djwong/xfs-linux into xfs-6.1-fixesA")
> >   ada76f94c5 ("xfs: account extra freespace btree splits for multiple allocations")
> > 
> > 4eb559dd15671cca ada76f94c5b32c1828955d70f69 
> > ---------------- --------------------------- 
> >          %stddev     %change         %stddev
> >              \          |                \  
> >     260.92           +90.8%     497.93 ± 13%  fxmark.ssd_xfs_MWCM_18_directio.idle_sec
> >      29.10           +91.5%      55.74 ± 14%  fxmark.ssd_xfs_MWCM_18_directio.idle_util
> >       9.11 ±  2%     -17.6%       7.51 ±  8%  fxmark.ssd_xfs_MWCM_18_directio.iowait_sec
> >       1.02 ±  2%     -17.3%       0.84 ±  8%  fxmark.ssd_xfs_MWCM_18_directio.iowait_util
> >     617.96           -38.8%     378.15 ± 19%  fxmark.ssd_xfs_MWCM_18_directio.sys_sec
> >      68.93           -38.6%      42.31 ± 18%  fxmark.ssd_xfs_MWCM_18_directio.sys_util
> >       1.80 ±  4%     -18.5%       1.47 ±  5%  fxmark.ssd_xfs_MWCM_18_directio.user_sec
> >       0.20 ±  4%     -18.2%       0.16 ±  4%  fxmark.ssd_xfs_MWCM_18_directio.user_util
> >    3363535           -29.6%    2369575 ±  7%  fxmark.ssd_xfs_MWCM_18_directio.works
> >      67262           -29.6%      47384 ±  7%  fxmark.ssd_xfs_MWCM_18_directio.works/sec
> >       0.01 ±111%  +51925.0%       3.47 ±  9%  fxmark.ssd_xfs_MWCM_1_directio.idle_sec
> >       0.01 ±111%  +52382.0%       6.99 ±  9%  fxmark.ssd_xfs_MWCM_1_directio.idle_util
> >       0.23 ± 14%   +1481.9%       3.64 ± 14%  fxmark.ssd_xfs_MWCM_1_directio.iowait_sec
> >       0.46 ± 14%   +1496.0%       7.33 ± 14%  fxmark.ssd_xfs_MWCM_1_directio.iowait_util
> >       0.58           +20.9%       0.70 ±  2%  fxmark.ssd_xfs_MWCM_1_directio.irq_sec
> >       1.15           +21.9%       1.40 ±  2%  fxmark.ssd_xfs_MWCM_1_directio.irq_util
> >       0.37 ±  2%     +18.9%       0.44 ±  5%  fxmark.ssd_xfs_MWCM_1_directio.softirq_sec
> >       0.74 ±  2%     +20.0%       0.89 ±  5%  fxmark.ssd_xfs_MWCM_1_directio.softirq_util
> >      47.70           -15.4%      40.33 ±  2%  fxmark.ssd_xfs_MWCM_1_directio.sys_sec
> >      95.24           -14.7%      81.24 ±  2%  fxmark.ssd_xfs_MWCM_1_directio.sys_util
> >       1.20 ±  5%     -11.0%       1.07 ±  4%  fxmark.ssd_xfs_MWCM_1_directio.user_sec
> >       2.40 ±  6%     -10.2%       2.15 ±  4%  fxmark.ssd_xfs_MWCM_1_directio.user_util
> >    2639345           -15.6%    2228369 ±  4%  fxmark.ssd_xfs_MWCM_1_directio.works
> >      52786           -15.6%      44567 ±  4%  fxmark.ssd_xfs_MWCM_1_directio.works/sec
> >       4.11 ±  3%    +254.5%      14.56 ± 18%  fxmark.ssd_xfs_MWCM_2_directio.idle_sec
> >       4.10 ±  3%    +255.9%      14.60 ± 18%  fxmark.ssd_xfs_MWCM_2_directio.idle_util
> >       1.42 ±  6%    +435.2%       7.57 ±  6%  fxmark.ssd_xfs_MWCM_2_directio.iowait_sec
> >       1.41 ±  6%    +437.2%       7.59 ±  6%  fxmark.ssd_xfs_MWCM_2_directio.iowait_util
> >       0.82           +20.2%       0.99 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.irq_sec
> >       0.82           +20.7%       0.99 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.irq_util
> >       0.52 ±  4%     +30.0%       0.67 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.softirq_sec
> >       0.52 ±  4%     +30.5%       0.67 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.softirq_util
> >      91.86           -18.6%      74.82 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.sys_sec
> >      91.73           -18.3%      74.99 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.sys_util
> >       1.42 ±  4%     -18.7%       1.15 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.user_sec
> >       1.42 ±  4%     -18.4%       1.16 ±  4%  fxmark.ssd_xfs_MWCM_2_directio.user_util
> >    3159498           -23.2%    2425056 ±  5%  fxmark.ssd_xfs_MWCM_2_directio.works
> >      63189           -23.3%      48486 ±  5%  fxmark.ssd_xfs_MWCM_2_directio.works/sec
> >     413.60 ±  4%    +103.5%     841.63 ± 20%  fxmark.ssd_xfs_MWCM_36_directio.idle_sec
> >      22.94 ±  4%    +103.9%      46.78 ± 20%  fxmark.ssd_xfs_MWCM_36_directio.idle_util
> >       1364           -31.6%     933.33 ± 19%  fxmark.ssd_xfs_MWCM_36_directio.sys_sec
> >      75.70           -31.5%      51.85 ± 18%  fxmark.ssd_xfs_MWCM_36_directio.sys_util
> >       1.82 ±  4%     -11.7%       1.60 ±  5%  fxmark.ssd_xfs_MWCM_36_directio.user_sec
> >       0.10 ±  4%     -11.5%       0.09 ±  5%  fxmark.ssd_xfs_MWCM_36_directio.user_util
> >    2850816           -22.9%    2197955 ±  7%  fxmark.ssd_xfs_MWCM_36_directio.works
> >      57006           -22.9%      43943 ±  7%  fxmark.ssd_xfs_MWCM_36_directio.works/sec
> >      54.22           +63.5%      88.65 ±  2%  fxmark.ssd_xfs_MWCM_4_directio.idle_sec
> >      27.33           +64.7%      45.02 ±  2%  fxmark.ssd_xfs_MWCM_4_directio.idle_util
> >       6.53           +36.1%       8.89 ±  8%  fxmark.ssd_xfs_MWCM_4_directio.iowait_sec
> >       3.29           +37.1%       4.51 ±  8%  fxmark.ssd_xfs_MWCM_4_directio.iowait_util
> >       1.37           +21.7%       1.67 ±  3%  fxmark.ssd_xfs_MWCM_4_directio.irq_sec
> >       0.69           +22.6%       0.85 ±  3%  fxmark.ssd_xfs_MWCM_4_directio.irq_util
> >       0.64           +20.3%       0.77 ±  3%  fxmark.ssd_xfs_MWCM_4_directio.softirq_sec
> >       0.32           +21.2%       0.39 ±  3%  fxmark.ssd_xfs_MWCM_4_directio.softirq_util
> >     134.06           -28.6%      95.67 ±  2%  fxmark.ssd_xfs_MWCM_4_directio.sys_sec
> >      67.58           -28.1%      48.59 ±  2%  fxmark.ssd_xfs_MWCM_4_directio.sys_util
> >       1.55 ±  2%     -18.4%       1.26 ±  5%  fxmark.ssd_xfs_MWCM_4_directio.user_sec
> >       0.78 ±  2%     -17.8%       0.64 ±  5%  fxmark.ssd_xfs_MWCM_4_directio.user_util
> >    3262962           -26.9%    2383952        fxmark.ssd_xfs_MWCM_4_directio.works
> >      65256           -26.9%      47670        fxmark.ssd_xfs_MWCM_4_directio.works/sec
> >     489.98 ±  3%    +160.6%       1277 ± 24%  fxmark.ssd_xfs_MWCM_54_directio.idle_sec
> >      18.11 ±  3%    +161.0%      47.27 ± 24%  fxmark.ssd_xfs_MWCM_54_directio.idle_util
> >       2185           -36.3%       1391 ± 22%  fxmark.ssd_xfs_MWCM_54_directio.sys_sec
> >      80.77           -36.2%      51.49 ± 22%  fxmark.ssd_xfs_MWCM_54_directio.sys_util
> >    2448443           -24.0%    1859708 ±  9%  fxmark.ssd_xfs_MWCM_54_directio.works
> >      48963           -24.1%      37175 ±  9%  fxmark.ssd_xfs_MWCM_54_directio.works/sec
> >     614.60 ±  2%     +88.5%       1158 ±  9%  fxmark.ssd_xfs_MWCM_72_directio.idle_sec
> >      17.04 ±  2%     +88.6%      32.13 ±  9%  fxmark.ssd_xfs_MWCM_72_directio.idle_util
> >       6.44 ±  4%     -14.8%       5.48 ±  9%  fxmark.ssd_xfs_MWCM_72_directio.iowait_sec
> >       0.18 ±  4%     -14.8%       0.15 ± 10%  fxmark.ssd_xfs_MWCM_72_directio.iowait_util
> >       2953           -18.5%       2407 ±  4%  fxmark.ssd_xfs_MWCM_72_directio.sys_sec
> >      81.89           -18.5%      66.77 ±  4%  fxmark.ssd_xfs_MWCM_72_directio.sys_util
> >    2279692           -15.1%    1935171 ±  2%  fxmark.ssd_xfs_MWCM_72_directio.works
> >      45589           -15.1%      38687 ±  2%  fxmark.ssd_xfs_MWCM_72_directio.works/sec
> >     934.32           -10.3%     838.20        fxmark.time.elapsed_time
> >     934.32           -10.3%     838.20        fxmark.time.elapsed_time.max
> >   43490089           -20.8%   34446960 ±  3%  fxmark.time.file_system_outputs
> >     128601 ±  3%     -31.6%      87922 ±  4%  fxmark.time.involuntary_context_switches
> >      31.83           -17.8%      26.17 ±  5%  fxmark.time.percent_of_cpu_this_job_got
> >     301.24           -26.5%     221.43 ±  5%  fxmark.time.system_time
> >     841264 ±  3%     -11.2%     747133 ±  5%  fxmark.time.voluntary_context_switches
> >     989.01            -9.6%     893.57        uptime.boot
> >      54.77           +10.7%      60.61        iostat.cpu.idle
> >       5.31 ±  4%     +21.4%       6.45 ±  4%  iostat.cpu.iowait
> >      38.42           -18.0%      31.51 ±  2%  iostat.cpu.system
> >      55.66            +5.8       61.45        mpstat.cpu.all.idle%
> >       5.31 ±  5%      +1.2        6.46 ±  4%  mpstat.cpu.all.iowait%
> >       0.60 ±  2%      -0.1        0.53        mpstat.cpu.all.soft%
> >      36.50            -6.6       29.87 ±  3%  mpstat.cpu.all.sys%
> >   26496967           -20.9%   20947493 ±  2%  numa-numastat.node0.local_node
> >   26548700           -20.8%   21022864 ±  2%  numa-numastat.node0.numa_hit
> >    3332141 ±  6%     -22.1%    2595871 ±  4%  numa-numastat.node1.local_node
> >    3400117 ±  5%     -22.4%    2638708 ±  5%  numa-numastat.node1.numa_hit
> >       2211            +1.0%       2233        perf-stat.i.minor-faults
> >       0.39 ±  9%      +0.1        0.44 ±  3%  perf-stat.i.node-load-miss-rate%
> >       2211            +1.0%       2233        perf-stat.i.page-faults
> >       2209            +1.0%       2230        perf-stat.ps.minor-faults
> >       2209            +1.0%       2230        perf-stat.ps.page-faults
> >      53.83           +11.5%      60.00        vmstat.cpu.id
> >       5430 ±  2%     -14.0%       4668 ±  3%  vmstat.io.bi
> >     123556            -8.3%     113294 ±  2%  vmstat.io.bo
> >   11181852           -16.7%    9318758 ±  2%  vmstat.memory.cache
> >       7.67 ±  6%     -19.6%       6.17 ± 11%  vmstat.procs.r
> >    9247281           -15.1%    7847381        meminfo.Cached
> >    3299044 ±  2%     -21.5%    2591082 ±  2%  meminfo.Dirty
> >    5715841 ±  2%     -24.4%    4320653 ±  3%  meminfo.Inactive
> >    5389224 ±  2%     -25.9%    3991807 ±  3%  meminfo.Inactive(file)
> >    1914390           -24.3%    1449759 ±  3%  meminfo.KReclaimable
> >      31406 ± 13%     -22.0%      24510 ±  7%  meminfo.KernelStack
> >   14111361           -16.6%   11775228 ±  2%  meminfo.Memused
> >    1914390           -24.3%    1449759 ±  3%  meminfo.SReclaimable
> >    1427015           -25.6%    1061962 ±  4%  meminfo.SUnreclaim
> >      18397 ±  2%      -8.4%      16846        meminfo.Shmem
> >    3341406           -24.8%    2511722 ±  3%  meminfo.Slab
> >   24035030           -18.6%   19557858 ±  2%  meminfo.max_used_kB
> >       1.23 ± 22%      +0.4        1.58 ± 18%  perf-profile.calltrace.cycles-pp.xlog_cil_push_work.process_one_work.worker_thread.kthread.ret_from_fork
> >       0.76 ± 44%      -0.4        0.38 ± 74%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
> >       0.09 ± 10%      -0.0        0.05 ± 47%  perf-profile.children.cycles-pp.mod_find
> >       0.10 ± 47%      +0.1        0.18 ± 26%  perf-profile.children.cycles-pp.xfs_iext_lookup_extent
> >       0.17 ± 26%      +0.1        0.25 ± 16%  perf-profile.children.cycles-pp.s_show
> >       0.05 ± 82%      +0.1        0.13 ± 46%  perf-profile.children.cycles-pp.timekeeping_advance
> >       0.05 ± 82%      +0.1        0.13 ± 46%  perf-profile.children.cycles-pp.update_wall_time
> >       0.21 ± 21%      +0.1        0.31 ± 20%  perf-profile.children.cycles-pp.vsnprintf
> >       0.20 ± 26%      +0.1        0.30 ± 17%  perf-profile.children.cycles-pp.seq_printf
> >       1.23 ± 22%      +0.4        1.58 ± 18%  perf-profile.children.cycles-pp.xlog_cil_push_work
> >       0.08 ± 14%      -0.0        0.04 ± 72%  perf-profile.self.cycles-pp.mod_find
> >       0.14 ± 26%      +0.1        0.22 ± 16%  perf-profile.self.cycles-pp.xlog_cil_push_work
> >       0.13 ± 37%      +0.1        0.21 ± 22%  perf-profile.self.cycles-pp.__folio_end_writeback
> >       1339 ±  6%     -12.1%       1177 ±  6%  numa-meminfo.node0.Active
> >       1317 ±  4%     -10.6%       1177 ±  6%  numa-meminfo.node0.Active(anon)
> >    2938921 ±  2%     -22.3%    2284083 ±  2%  numa-meminfo.node0.Dirty
> >    4950896 ±  2%     -25.3%    3699254 ±  2%  numa-meminfo.node0.Inactive
> >    4806022 ±  2%     -27.0%    3510366 ±  3%  numa-meminfo.node0.Inactive(file)
> >    1663428 ±  2%     -25.0%    1246833 ±  3%  numa-meminfo.node0.KReclaimable
> >      15585 ±  4%     -12.4%      13655 ±  4%  numa-meminfo.node0.KernelStack
> >    1663428 ±  2%     -25.0%    1246833 ±  3%  numa-meminfo.node0.SReclaimable
> >    1224137 ±  2%     -25.5%     912586 ±  4%  numa-meminfo.node0.SUnreclaim
> >      15785 ±  5%     -12.5%      13811 ±  4%  numa-meminfo.node0.Shmem
> >    2887566           -25.2%    2159421 ±  3%  numa-meminfo.node0.Slab
> >     358435 ±  3%     -14.3%     307335 ±  6%  numa-meminfo.node1.Dirty
> >     764404 ±  6%     -18.6%     622071 ± 10%  numa-meminfo.node1.Inactive
> >     582651 ±  2%     -17.3%     482089 ±  9%  numa-meminfo.node1.Inactive(file)
> >     250653 ± 10%     -18.9%     203397 ± 13%  numa-meminfo.node1.KReclaimable
> >    3410163 ± 54%     -26.4%    2508459 ± 68%  numa-meminfo.node1.MemUsed
> >     250653 ± 10%     -18.9%     203397 ± 13%  numa-meminfo.node1.SReclaimable
> >     202812 ± 17%     -26.2%     149763 ±  5%  numa-meminfo.node1.SUnreclaim
> >     453467 ±  9%     -22.1%     353161 ±  8%  numa-meminfo.node1.Slab
> >       0.71 ± 14%     -21.5%       0.56 ± 14%  sched_debug.cfs_rq:/.h_nr_running.avg
> >       1405 ± 55%     -68.3%     445.08 ± 27%  sched_debug.cfs_rq:/.load_avg.avg
> >    5023746           -33.8%    3324248 ± 13%  sched_debug.cfs_rq:/.min_vruntime.avg
> >    5080937           -33.2%    3395209 ± 12%  sched_debug.cfs_rq:/.min_vruntime.max
> >    4683128           -31.3%    3217873 ± 13%  sched_debug.cfs_rq:/.min_vruntime.min
> >       0.59 ±  9%     -17.4%       0.49 ± 18%  sched_debug.cfs_rq:/.nr_running.avg
> >     650.51 ± 13%     -32.3%     440.38 ±  9%  sched_debug.cfs_rq:/.runnable_avg.avg
> >       1007 ±  8%     -18.5%     821.27 ±  6%  sched_debug.cfs_rq:/.runnable_avg.max
> >     476.07 ± 17%     -38.7%     291.80 ± 16%  sched_debug.cfs_rq:/.runnable_avg.min
> >     517.96 ±  5%     -20.3%     412.57 ±  9%  sched_debug.cfs_rq:/.util_avg.avg
> >     352.69 ±  5%     -26.1%     260.66 ± 14%  sched_debug.cfs_rq:/.util_avg.min
> >     113.13 ±  9%     -26.8%      82.85 ± 15%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
> >      18547 ±  4%      -9.5%      16791 ±  3%  sched_debug.cpu.curr->pid.max
> >       1549 ± 11%   +1106.2%      18692 ±123%  sched_debug.cpu.max_idle_balance_cost.stddev
> >       0.70 ± 16%     -25.1%       0.52 ± 18%  sched_debug.cpu.nr_running.avg
> >     862357 ±  4%     -16.3%     722142 ±  3%  sched_debug.cpu.nr_switches.avg
> >    1440368 ±  6%     -22.2%    1120203 ±  6%  sched_debug.cpu.nr_switches.max
> >     756963 ±  4%     -14.6%     646359 ±  5%  sched_debug.cpu.nr_switches.min
> >     150642 ± 10%     -33.8%      99651 ± 18%  sched_debug.cpu.nr_switches.stddev
> >   17687949           -23.3%   13563229 ±  2%  numa-vmstat.node0.nr_dirtied
> >     734675 ±  2%     -22.3%     570784 ±  2%  numa-vmstat.node0.nr_dirty
> >    1201580 ±  2%     -27.0%     877113 ±  3%  numa-vmstat.node0.nr_inactive_file
> >      15586 ±  4%     -12.4%      13655 ±  4%  numa-vmstat.node0.nr_kernel_stack
> >       3945 ±  5%     -12.5%       3453 ±  4%  numa-vmstat.node0.nr_shmem
> >     415866 ±  2%     -25.1%     311497 ±  3%  numa-vmstat.node0.nr_slab_reclaimable
> >     306030 ±  2%     -25.5%     228011 ±  4%  numa-vmstat.node0.nr_slab_unreclaimable
> >   17687949           -23.3%   13563229 ±  2%  numa-vmstat.node0.nr_written
> >    1201580 ±  2%     -27.0%     877113 ±  3%  numa-vmstat.node0.nr_zone_inactive_file
> >     734697 ±  2%     -22.3%     570821 ±  2%  numa-vmstat.node0.nr_zone_write_pending
> >   26548703           -20.8%   21023488 ±  2%  numa-vmstat.node0.numa_hit
> >   26496970           -20.9%   20948116 ±  2%  numa-vmstat.node0.numa_local
> >    2316618           -20.7%    1836559 ±  5%  numa-vmstat.node1.nr_dirtied
> >      89623 ±  3%     -14.3%      76800 ±  6%  numa-vmstat.node1.nr_dirty
> >     145701 ±  2%     -17.3%     120486 ±  9%  numa-vmstat.node1.nr_inactive_file
> >      62668 ± 10%     -18.9%      50828 ± 13%  numa-vmstat.node1.nr_slab_reclaimable
> >      50710 ± 17%     -26.2%      37434 ±  5%  numa-vmstat.node1.nr_slab_unreclaimable
> >    2316618           -20.7%    1836559 ±  5%  numa-vmstat.node1.nr_written
> >     145701 ±  2%     -17.3%     120486 ±  9%  numa-vmstat.node1.nr_zone_inactive_file
> >      89584 ±  3%     -14.3%      76792 ±  6%  numa-vmstat.node1.nr_zone_write_pending
> >    3400024 ±  5%     -22.4%    2639087 ±  5%  numa-vmstat.node1.numa_hit
> >    3332048 ±  6%     -22.1%    2596249 ±  4%  numa-vmstat.node1.numa_local
> >      77508            +1.2%      78432        proc-vmstat.nr_anon_pages
> >   20004568           -23.0%   15399788 ±  2%  proc-vmstat.nr_dirtied
> >     825597 ±  2%     -21.4%     648654 ±  2%  proc-vmstat.nr_dirty
> >    2311462           -15.1%    1961741        proc-vmstat.nr_file_pages
> >   29390035            +2.0%   29973527        proc-vmstat.nr_free_pages
> >    1346949 ±  2%     -25.9%     997847 ±  3%  proc-vmstat.nr_inactive_file
> >      31402 ± 13%     -21.9%      24514 ±  7%  proc-vmstat.nr_kernel_stack
> >       4599 ±  2%      -8.4%       4211        proc-vmstat.nr_shmem
> >     478540           -24.3%     362476 ±  3%  proc-vmstat.nr_slab_reclaimable
> >     356646           -25.6%     265413 ±  4%  proc-vmstat.nr_slab_unreclaimable
> >   20004568           -23.0%   15399788 ±  2%  proc-vmstat.nr_written
> >    1346949 ±  2%     -25.9%     997847 ±  3%  proc-vmstat.nr_zone_inactive_file
> >     825584 ±  2%     -21.4%     648681 ±  2%  proc-vmstat.nr_zone_write_pending
> >   29951284           -21.0%   23664558 ±  2%  proc-vmstat.numa_hit
> >   29831574           -21.1%   23546349 ±  2%  proc-vmstat.numa_local
> >   51777954           -21.0%   40906350 ±  2%  proc-vmstat.pgalloc_normal
> >       5920 ±  2%      -7.6%       5472 ±  2%  proc-vmstat.pgdeactivate
> >    2342027            -8.5%    2142047        proc-vmstat.pgfault
> >   51819442           -21.0%   40950599 ±  2%  proc-vmstat.pgfree
> >    5071045 ±  3%     -22.9%    3908046 ±  4%  proc-vmstat.pgpgin
> >  1.153e+08           -17.9%   94738236        proc-vmstat.pgpgout
> >     141611            -6.3%     132656        proc-vmstat.pgreuse
> >   10386182 ±  3%     -24.2%    7877205 ±  4%  proc-vmstat.pgrotated
> >   36803772           -24.1%   27916664 ±  2%  proc-vmstat.slabs_scanned
> > 
> > 
> > 
> > 
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> > 
> > 
> > -- 
> > 0-DAY CI Kernel Test Service
> > https://01.org/lkp
> > 
> > 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ