lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202210271720.1af1ecc3-yujie.liu@intel.com>
Date:   Thu, 27 Oct 2022 17:35:39 +0800
From:   kernel test robot <yujie.liu@...el.com>
To:     Jan Kara <jack@...e.cz>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        Theodore Ts'o <tytso@....edu>,
        Stefan Wahren <stefan.wahren@...e.com>,
        Ojaswin Mujoo <ojaswin@...ux.ibm.com>,
        <linux-kernel@...r.kernel.org>, <linux-ext4@...r.kernel.org>,
        <ying.huang@...el.com>, <feng.tang@...el.com>,
        <zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>
Subject: [linus:master] [ext4] 613c5a8589:
 fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec 36.0% improvement

Greeting,

FYI, we noticed a 36.0% improvement of fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec due to commit:

commit: 613c5a85898d1cd44e68f28d65eccf64a8ace9cf ("ext4: make directory inode spreading reflect flexbg size")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: fxmark
on test machine: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
with following parameters:

	disk: 1SSD
	media: ssd
	test: dbench_client
	fstype: ext4_no_jnl
	directio: directio
	cpufreq_governor: performance

test-description: FxMark is a filesystem benchmark that test multicore scalability.
test-url: https://github.com/sslab-gatech/fxmark


Details are as below:

=========================================================================================
compiler/cpufreq_governor/directio/disk/fstype/kconfig/media/rootfs/tbox_group/test/testcase:
  gcc-11/performance/directio/1SSD/ext4_no_jnl/x86_64-rhel-8.3/ssd/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp5/dbench_client/fxmark

commit: 
  1940265ede ("ext4: avoid unnecessary spreading of allocations among groups")
  613c5a8589 ("ext4: make directory inode spreading reflect flexbg size")

1940265ede6683f6 613c5a85898d1cd44e68f28d65e 
---------------- --------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :6          100%           6:6     stderr.Couldn't_synthesize_bpf_events
         %stddev     %change         %stddev
             \          |                \  
    323.96 ± 12%     +28.7%     417.02 ±  7%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.idle_sec
      7.50 ± 12%     +28.9%       9.66 ±  7%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.idle_util
      1583 ±  3%     +30.8%       2072 ±  2%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.iowait_sec
     36.66 ±  3%     +31.0%      48.02 ±  2%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.iowait_util
      5.60 ±  3%     +27.3%       7.14        fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.softirq_sec
      0.13 ±  3%     +27.5%       0.17        fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.softirq_util
      2348 ±  3%     -25.4%       1752 ±  4%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.sys_sec
     54.36 ±  3%     -25.3%      40.61 ±  4%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.sys_util
     27.58 ±  4%     +21.0%      33.36 ±  3%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.user_sec
      0.64 ±  4%     +21.1%       0.77 ±  3%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.user_util
      1685 ±  4%     +36.0%       2292 ±  3%  fxmark.ssd_ext4_no_jnl_dbench_client_72_directio.works/sec
 3.054e+08            +3.3%  3.154e+08        fxmark.time.file_system_outputs
      1032 ±  2%      -8.7%     942.00        fxmark.time.percent_of_cpu_this_job_got
      6151 ±  2%      -9.2%       5584        fxmark.time.system_time
    154.57            +3.8%     160.38        fxmark.time.user_time
   8290165            +5.5%    8747536        fxmark.time.voluntary_context_switches
   6242070 ±  2%      +9.6%    6844007 ±  2%  numa-numastat.node1.numa_hit
     33.42            +3.5%      34.59        iostat.cpu.iowait
     36.41            -3.1%      35.27        iostat.cpu.system
   4347817           -24.5%    3281458        vmstat.memory.cache
     45757            +4.7%      47892        vmstat.system.cs
     38.17            -9.1%      34.69        boot-time.boot
     23.34           -15.2%      19.79        boot-time.dhcp
      4374           -10.1%       3934        boot-time.idle
      0.29 ± 10%     +16.3%       0.33 ±  6%  sched_debug.cfs_rq:/.h_nr_running.stddev
   3361937 ±  4%     -14.6%    2870581 ±  3%  sched_debug.cfs_rq:/.min_vruntime.avg
   3453968 ±  4%     -13.8%    2976035 ±  2%  sched_debug.cfs_rq:/.min_vruntime.max
   3074025 ±  4%     -16.7%    2559836 ±  6%  sched_debug.cfs_rq:/.min_vruntime.min
   4234791           -25.1%    3170271        meminfo.Cached
  13280609 ±  8%     -23.2%   10193251 ±  9%  meminfo.DirectMap2M
   5623315           -19.1%    4546975        meminfo.Memused
   3825966           -27.8%    2760633        meminfo.Unevictable
   7683659           -13.9%    6614519        meminfo.max_used_kB
     45394            +4.8%      47577        perf-stat.i.context-switches
    878.12 ±  2%      +7.8%     946.23 ±  3%  perf-stat.i.cpu-migrations
      0.00 ±116%      +0.0        0.00 ± 66%  perf-stat.i.dTLB-load-miss-rate%
     45558            +4.8%      47762        perf-stat.ps.context-switches
    886.77 ±  2%      +7.5%     953.70 ±  3%  perf-stat.ps.cpu-migrations
      0.49 ± 19%      -0.3        0.18 ± 52%  perf-profile.children.cycles-pp.get_cpu_device
      0.34 ± 24%      -0.1        0.20 ± 20%  perf-profile.children.cycles-pp.s_next
      0.06 ±114%      +0.1        0.14 ± 34%  perf-profile.children.cycles-pp.account_process_tick
      0.48 ± 20%      -0.3        0.16 ± 54%  perf-profile.self.cycles-pp.get_cpu_device
      0.05 ±141%      +0.1        0.14 ± 41%  perf-profile.self.cycles-pp.account_process_tick
      0.40 ±  9%      +0.1        0.50 ± 13%  perf-profile.self.cycles-pp.tick_nohz_next_event
    109190 ± 30%     -60.3%      43316 ± 57%  numa-meminfo.node0.AnonHugePages
    154521 ± 26%     -43.2%      87783 ± 27%  numa-meminfo.node0.AnonPages
    437517 ±  8%     -16.8%     364128 ±  6%  numa-meminfo.node0.Inactive
    160530 ± 25%     -41.4%      94026 ± 25%  numa-meminfo.node0.Inactive(anon)
     87985 ± 39%     +73.3%     152466 ± 16%  numa-meminfo.node1.AnonHugePages
    157511 ± 26%     +41.4%     222774 ± 10%  numa-meminfo.node1.AnonPages
    207719 ± 17%     +34.9%     280314 ±  9%  numa-meminfo.node1.Inactive
    161988 ± 25%     +40.3%     227216 ± 10%  numa-meminfo.node1.Inactive(anon)
     38635 ± 26%     -43.2%      21947 ± 27%  numa-vmstat.node0.nr_anon_pages
     40136 ± 25%     -41.4%      23505 ± 25%  numa-vmstat.node0.nr_inactive_anon
     40136 ± 25%     -41.4%      23505 ± 25%  numa-vmstat.node0.nr_zone_inactive_anon
     39376 ± 26%     +41.5%      55699 ± 10%  numa-vmstat.node1.nr_anon_pages
   6131790 ±  2%     +10.0%    6744942 ±  2%  numa-vmstat.node1.nr_dirtied
     40495 ± 25%     +40.3%      56809 ± 10%  numa-vmstat.node1.nr_inactive_anon
     40495 ± 25%     +40.3%      56809 ± 10%  numa-vmstat.node1.nr_zone_inactive_anon
   6241898 ±  2%      +9.6%    6843576 ±  2%  numa-vmstat.node1.numa_hit
  38173261            +3.3%   39428621        proc-vmstat.nr_dirtied
   1059388           -25.1%     793340        proc-vmstat.nr_file_pages
     53799            -1.9%      52794        proc-vmstat.nr_slab_unreclaimable
    956491           -27.8%     690158        proc-vmstat.nr_unevictable
   9850651            +3.5%   10194926        proc-vmstat.nr_written
    956491           -27.8%     690158        proc-vmstat.nr_zone_unevictable
  38481863            +3.2%   39707334        proc-vmstat.numa_hit
  38362225            +3.2%   39589589        proc-vmstat.numa_local
   1333657            +2.1%    1361550        proc-vmstat.pgactivate
  38472842            +3.2%   39705274        proc-vmstat.pgalloc_normal
  38505715            +3.2%   39737694        proc-vmstat.pgfree
  66251245            +2.3%   67801130        proc-vmstat.pgpgout



To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        sudo bin/lkp install job.yaml           # job file is attached in this email
        bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
        sudo bin/lkp run generated-yaml-file

        # if come across any failure that blocks the test,
        # please remove ~/.lkp and /lkp dir to run from a clean state.


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://01.org/lkp

View attachment "config-6.0.0-rc4-00003-g613c5a85898d" of type "text/plain" (163956 bytes)

View attachment "job-script" of type "text/plain" (8237 bytes)

View attachment "job.yaml" of type "text/plain" (5672 bytes)

View attachment "reproduce" of type "text/plain" (254 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ