lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202511071447.b5bd3ddf-lkp@intel.com>
Date: Fri, 7 Nov 2025 15:21:54 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Dave Chinner <dchinner@...hat.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
	Carlos Maiolino <cem@...nel.org>, Jan Kara <jack@...e.cz>, Christoph Hellwig
	<hch@....de>, <linux-xfs@...r.kernel.org>, <oliver.sang@...el.com>
Subject: [linus:master] [xfs]  c91d38b57f:  stress-ng.chmod.ops_per_sec
 3807.5% improvement



Hello,

kernel test robot noticed a 3807.5% improvement of stress-ng.chmod.ops_per_sec on:


commit: c91d38b57f2c4784d885c874b2a1234a01361afd ("xfs: rework datasync tracking and execution")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master


testcase: stress-ng
config: x86_64-rhel-9.4
compiler: gcc-14
test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
parameters:

	nr_threads: 100%
	disk: 1HDD
	testtime: 60s
	fs: xfs
	test: chmod
	cpufreq_governor: performance


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20251107/202511071447.b5bd3ddf-lkp@intel.com

=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  gcc-14/performance/1HDD/xfs/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-icl-2sp8/chmod/stress-ng/60s

commit: 
  bc7d684fea ("xfs: rearrange code in xfs_inode_item_precommit")
  c91d38b57f ("xfs: rework datasync tracking and execution")

bc7d684fea18cc48 c91d38b57f2c4784d885c874b2a 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      6943 ±  2%     -31.3%       4767        uptime.idle
 3.968e+09           -19.6%  3.191e+09        cpuidle..time
    267038 ±  6%    +228.0%     875840 ±  6%  cpuidle..usage
    202471 ± 87%    +110.4%     425943 ± 53%  numa-meminfo.node1.Active
    202464 ± 87%    +110.4%     425935 ± 53%  numa-meminfo.node1.Active(anon)
     11019 ± 11%    +614.5%      78736 ± 19%  numa-meminfo.node1.Shmem
     50689 ± 87%    +110.2%     106536 ± 53%  numa-vmstat.node1.nr_active_anon
      2736 ± 11%    +620.7%      19718 ± 19%  numa-vmstat.node1.nr_shmem
     50689 ± 87%    +110.2%     106536 ± 53%  numa-vmstat.node1.nr_zone_active_anon
     98.29           -47.8%      51.30 ±  2%  iostat.cpu.idle
      0.01 ± 34%  +4.4e+05%      29.80        iostat.cpu.iowait
      0.86 ±  8%   +1983.4%      17.82 ±  6%  iostat.cpu.system
      0.84 ± 13%     +27.9%       1.08 ±  2%  iostat.cpu.user
    678151           +10.8%     751607        meminfo.Active
    678134           +10.8%     751591        meminfo.Active(anon)
     74229 ±  4%     -10.1%      66725 ±  3%  meminfo.AnonHugePages
     16590 ±  5%    +432.3%      88307 ± 14%  meminfo.Shmem
     15.83 ± 23%   +1103.2%     190.50 ±  8%  perf-c2c.DRAM.local
    358.00 ± 30%   +3312.1%      12215 ±  7%  perf-c2c.DRAM.remote
    321.17 ± 32%   +2900.4%       9636 ±  7%  perf-c2c.HITM.local
    241.00 ± 30%   +3726.9%       9222 ±  7%  perf-c2c.HITM.remote
    562.17 ± 31%   +3254.7%      18859 ±  7%  perf-c2c.HITM.total
      1225 ± 14%   +3762.4%      47314 ±  4%  stress-ng.chmod.ops
     20.17 ± 15%   +3807.5%     788.15 ±  4%  stress-ng.chmod.ops_per_sec
    146.50 ± 11%    +752.3%       1248 ±  4%  stress-ng.time.involuntary_context_switches
     11984           +20.8%      14481        stress-ng.time.minor_page_faults
     34.17 ± 16%   +3293.2%       1159 ±  6%  stress-ng.time.percent_of_cpu_this_job_got
     21.27 ± 15%   +3173.1%     696.14 ±  6%  stress-ng.time.system_time
     62731 ± 16%   +1072.8%     735741 ±  9%  stress-ng.time.voluntary_context_switches
     98.30           -48.5       49.84 ±  2%  mpstat.cpu.all.idle%
      0.00 ±105%     +30.7       30.72        mpstat.cpu.all.iowait%
      0.02 ±  8%      +0.1        0.09 ±  6%  mpstat.cpu.all.irq%
      0.01 ± 18%      +0.0        0.02 ±  5%  mpstat.cpu.all.soft%
      0.83 ±  9%     +17.4       18.24 ±  6%  mpstat.cpu.all.sys%
      0.85 ± 14%      +0.2        1.09 ±  2%  mpstat.cpu.all.usr%
      1.00         +4033.3%      41.33 ± 33%  mpstat.max_utilization.seconds
      6.13 ±  6%    +333.0%      26.54 ±  4%  mpstat.max_utilization_pct
     65.17 ± 11%    +969.1%     696.67 ±  6%  turbostat.Avg_MHz
      1.81 ± 11%     +17.6       19.39 ±  6%  turbostat.Busy%
     98.22           -17.5       80.69        turbostat.C1%
     96.82           -25.6%      72.07        turbostat.CPU%c1
      0.72 ±  4%     -55.4%       0.32 ±  2%  turbostat.IPC
    372011 ±  8%    +706.5%    3000169 ±  5%  turbostat.IRQ
     23024 ± 30%   +2629.0%     628341 ±  6%  turbostat.NMI
    243.27            +9.4%     266.04        turbostat.PkgWatt
     15.14            +2.9%      15.57        turbostat.RAMWatt
    225.26 ± 17%     -92.7%      16.48 ±  9%  perf-sched.total_wait_and_delay.average.ms
      4552 ± 19%   +1482.3%      72026 ± 10%  perf-sched.total_wait_and_delay.count.ms
      4996           -22.0%       3898 ± 10%  perf-sched.total_wait_and_delay.max.ms
    225.25 ± 17%     -92.7%      16.46 ±  9%  perf-sched.total_wait_time.average.ms
      4996           -22.0%       3898 ± 10%  perf-sched.total_wait_time.max.ms
    225.26 ± 17%     -92.7%      16.48 ±  9%  perf-sched.wait_and_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      4552 ± 19%   +1482.3%      72026 ± 10%  perf-sched.wait_and_delay.count.[unknown].[unknown].[unknown].[unknown].[unknown]
      4996           -22.0%       3898 ± 10%  perf-sched.wait_and_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    225.25 ± 17%     -92.7%      16.46 ±  9%  perf-sched.wait_time.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      4996           -22.0%       3898 ± 10%  perf-sched.wait_time.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    169578           +10.8%     187927        proc-vmstat.nr_active_anon
    935816            +1.9%     953408        proc-vmstat.nr_file_pages
     20878            +4.0%      21711        proc-vmstat.nr_mapped
      4142 ±  5%    +433.8%      22109 ± 14%  proc-vmstat.nr_shmem
     41118            +1.0%      41531        proc-vmstat.nr_slab_unreclaimable
    169578           +10.8%     187927        proc-vmstat.nr_zone_active_anon
    615.17 ±218%    +464.4%       3472 ± 14%  proc-vmstat.numa_hint_faults_local
    345828 ±  4%     +10.1%     380782 ±  2%  proc-vmstat.numa_hit
    279652 ±  5%     +12.5%     314600 ±  3%  proc-vmstat.numa_local
     60782 ± 60%     +78.2%     108287 ± 14%  proc-vmstat.numa_pte_updates
    382421 ±  4%      +9.6%     419096 ±  2%  proc-vmstat.pgalloc_normal
      5338 ± 20%    +193.0%      15639 ±  6%  proc-vmstat.pgpgout
      7731 ±  7%   +1078.0%      91080 ± 10%  sched_debug.cfs_rq:/.avg_vruntime.avg
     61046 ± 13%    +183.6%     173149 ± 14%  sched_debug.cfs_rq:/.avg_vruntime.max
    747.60 ± 23%   +3826.2%      29352 ± 13%  sched_debug.cfs_rq:/.avg_vruntime.min
     10401 ±  7%     +72.6%      17954 ±  9%  sched_debug.cfs_rq:/.avg_vruntime.stddev
    564880 ± 44%     -71.2%     162663 ±113%  sched_debug.cfs_rq:/.load.max
     79058 ± 37%     -64.4%      28177 ± 90%  sched_debug.cfs_rq:/.load.stddev
      7731 ±  7%   +1078.0%      91080 ± 10%  sched_debug.cfs_rq:/.min_vruntime.avg
     61046 ± 13%    +183.6%     173149 ± 14%  sched_debug.cfs_rq:/.min_vruntime.max
    747.60 ± 23%   +3826.2%      29352 ± 13%  sched_debug.cfs_rq:/.min_vruntime.min
     10401 ±  7%     +72.6%      17954 ±  9%  sched_debug.cfs_rq:/.min_vruntime.stddev
    254.61 ±  6%     +23.3%     313.88 ±  6%  sched_debug.cfs_rq:/.runnable_avg.avg
    254.13 ±  6%     +23.4%     313.61 ±  6%  sched_debug.cfs_rq:/.util_avg.avg
   1388025 ±  3%      -8.9%    1264183 ±  4%  sched_debug.cpu.avg_idle.avg
      4207 ±  7%    +257.6%      15044 ±  6%  sched_debug.cpu.nr_switches.avg
     13511 ± 20%    +276.6%      50888 ± 62%  sched_debug.cpu.nr_switches.max
      1440 ±  9%    +592.5%       9975 ±  4%  sched_debug.cpu.nr_switches.min
      2591 ± 14%    +129.9%       5958 ± 61%  sched_debug.cpu.nr_switches.stddev
     19.42 ± 31%    +137.8%      46.17 ± 17%  sched_debug.cpu.nr_uninterruptible.max
    -14.67          +891.5%    -145.42        sched_debug.cpu.nr_uninterruptible.min
      4.79 ± 10%    +353.6%      21.71 ± 15%  sched_debug.cpu.nr_uninterruptible.stddev
      0.94 ±  5%     -30.7%       0.65 ±  2%  perf-stat.i.MPKI
 6.179e+08 ± 12%    +373.6%  2.927e+09 ±  4%  perf-stat.i.branch-instructions
      3.29 ± 11%      -1.8        1.47 ±  8%  perf-stat.i.branch-miss-rate%
  43227294 ± 15%     +20.8%   52203685 ±  3%  perf-stat.i.branch-misses
     17.76 ±  6%     +19.1       36.85        perf-stat.i.cache-miss-rate%
   1306832 ± 11%    +601.7%    9169533 ±  4%  perf-stat.i.cache-misses
   8673239 ± 10%    +206.9%   26621069 ±  3%  perf-stat.i.cache-references
      3772 ± 10%    +618.3%      27098 ±  8%  perf-stat.i.context-switches
      2.43 ±  4%     +32.5%       3.23        perf-stat.i.cpi
 4.243e+09 ± 11%    +964.2%  4.515e+10 ±  6%  perf-stat.i.cpu-cycles
    116.32 ±  2%    +203.4%     352.94 ±  3%  perf-stat.i.cpu-migrations
      3133 ±  3%     +59.5%       4996 ±  2%  perf-stat.i.cycles-between-cache-misses
 3.037e+09 ± 12%    +376.7%  1.448e+10 ±  4%  perf-stat.i.instructions
      0.55 ±  4%     -40.3%       0.33 ±  2%  perf-stat.i.ipc
      0.43 ±  4%     +46.9%       0.63 ±  2%  perf-stat.overall.MPKI
      6.97 ±  3%      -5.2        1.79 ±  7%  perf-stat.overall.branch-miss-rate%
     15.06 ±  4%     +19.3       34.39        perf-stat.overall.cache-miss-rate%
      1.40 ±  3%    +122.3%       3.12        perf-stat.overall.cpi
      3252 ±  2%     +51.3%       4921 ±  3%  perf-stat.overall.cycles-between-cache-misses
      0.71 ±  3%     -55.1%       0.32        perf-stat.overall.ipc
 6.093e+08 ± 12%    +372.7%   2.88e+09 ±  4%  perf-stat.ps.branch-instructions
  42634564 ± 15%     +20.7%   51439404 ±  3%  perf-stat.ps.branch-misses
   1287683 ± 11%    +600.7%    9022871 ±  4%  perf-stat.ps.cache-misses
   8556164 ± 10%    +206.5%   26227629 ±  3%  perf-stat.ps.cache-references
      3710 ± 10%    +618.7%      26664 ±  8%  perf-stat.ps.context-switches
 4.185e+09 ± 11%    +961.5%  4.442e+10 ±  6%  perf-stat.ps.cpu-cycles
    114.49 ±  2%    +203.8%     347.78 ±  3%  perf-stat.ps.cpu-migrations
 2.995e+09 ± 12%    +375.7%  1.425e+10 ±  4%  perf-stat.ps.instructions
 1.865e+11 ± 12%    +369.7%  8.759e+11 ±  4%  perf-stat.total.instructions




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ