lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <202509241629.3d135124-lkp@intel.com>
Date: Wed, 24 Sep 2025 16:37:05 +0800
From: kernel test robot <oliver.sang@...el.com>
To: <hariconscious@...il.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <netdev@...r.kernel.org>,
	<davem@...emloft.net>, <edumazet@...gle.com>, <kuba@...nel.org>,
	<pabeni@...hat.com>, <horms@...nel.org>, <shuah@...nel.org>,
	<stable@...r.kernel.org>, <linux-kernel@...r.kernel.org>,
	<syzbot+9a4fbb77c9d4aacd3388@...kaller.appspotmail.com>, HariKrishna Sagala
	<hariconscious@...il.com>, <oliver.sang@...el.com>
Subject: Re: [PATCH net RESEND] net/core : fix KMSAN: uninit value in tipc_rcv



Hello,

kernel test robot noticed a 33.9% regression of netperf.Throughput_Mbps on:


commit: 5cde54f8220b582bda9c34ef86e04ec00be4ce4a ("[PATCH net RESEND] net/core : fix KMSAN: uninit value in tipc_rcv")
url: https://github.com/intel-lab-lkp/linux/commits/hariconscious-gmail-com/net-core-fix-KMSAN-uninit-value-in-tipc_rcv/20250920-023232
base: https://git.kernel.org/cgit/linux/kernel/git/davem/net.git cbf658dd09419f1ef9de11b9604e950bdd5c170b
patch link: https://lore.kernel.org/all/20250919183146.4933-1-hariconscious@gmail.com/
patch subject: [PATCH net RESEND] net/core : fix KMSAN: uninit value in tipc_rcv

testcase: netperf
config: x86_64-rhel-9.4
compiler: gcc-14
test machine: 192 threads 2 sockets Intel(R) Xeon(R) 6740E  CPU @ 2.4GHz (Sierra Forest) with 256G memory
parameters:

	ip: ipv4
	runtime: 300s
	nr_threads: 100%
	cluster: cs-localhost
	test: SCTP_STREAM
	cpufreq_governor: performance




If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202509241629.3d135124-lkp@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250924/202509241629.3d135124-lkp@intel.com

=========================================================================================
cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/tbox_group/test/testcase:
  cs-localhost/gcc-14/performance/ipv4/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/300s/lkp-srf-2sp3/SCTP_STREAM/netperf

commit: 
  cbf658dd09 ("Merge tag 'net-6.17-rc7' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net")
  5cde54f822 ("net/core : fix KMSAN: uninit value in tipc_rcv")

cbf658dd09419f1e 5cde54f8220b582bda9c34ef86e 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
 5.164e+09           -14.4%  4.419e+09        cpuidle..time
   1504152           -12.9%    1310310 ±  4%  cpuidle..usage
    125501 ±  2%     +19.6%     150094 ±  3%  meminfo.Mapped
    401239 ±  4%     +18.3%     474475        meminfo.Shmem
      8.63            -1.3        7.31 ±  2%  mpstat.cpu.all.idle%
      0.42            +0.1        0.52        mpstat.cpu.all.irq%
  1.96e+08           -34.2%   1.29e+08        numa-numastat.node0.local_node
 1.961e+08           -34.2%  1.291e+08        numa-numastat.node0.numa_hit
 1.943e+08           -33.3%  1.297e+08        numa-numastat.node1.local_node
 1.944e+08           -33.2%  1.298e+08        numa-numastat.node1.numa_hit
 1.961e+08           -34.2%  1.291e+08        numa-vmstat.node0.numa_hit
  1.96e+08           -34.2%   1.29e+08        numa-vmstat.node0.numa_local
 1.944e+08           -33.2%  1.298e+08        numa-vmstat.node1.numa_hit
 1.943e+08           -33.3%  1.297e+08        numa-vmstat.node1.numa_local
      1386           -33.9%     916.08        netperf.ThroughputBoth_Mbps
    266242           -33.9%     175886        netperf.ThroughputBoth_total_Mbps
      1386           -33.9%     916.08        netperf.Throughput_Mbps
    266242           -33.9%     175886        netperf.Throughput_total_Mbps
    476071           -18.8%     386450 ±  2%  netperf.time.involuntary_context_switches
      6916           +26.0%       8715        netperf.time.percent_of_cpu_this_job_got
     20872           +26.3%      26369        netperf.time.system_time
    117319           -33.4%      78105        netperf.time.voluntary_context_switches
    115788           -33.9%      76498        netperf.workload
   1005775            +1.8%    1024121        proc-vmstat.nr_file_pages
     31705 ±  2%     +18.6%      37602 ±  4%  proc-vmstat.nr_mapped
    100398 ±  4%     +18.3%     118736        proc-vmstat.nr_shmem
   9211977            -4.0%    8843488        proc-vmstat.nr_slab_unreclaimable
 3.905e+08           -33.7%  2.589e+08        proc-vmstat.numa_hit
    226148           +11.7%     252687        proc-vmstat.numa_huge_pte_updates
 3.903e+08           -33.7%  2.587e+08        proc-vmstat.numa_local
 1.164e+08           +11.6%  1.299e+08        proc-vmstat.numa_pte_updates
 1.243e+10           -33.8%  8.227e+09        proc-vmstat.pgalloc_normal
 1.243e+10           -33.8%  8.226e+09        proc-vmstat.pgfree
      4.20 ± 13%     -50.0%       2.10 ± 11%  perf-sched.sch_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      4.20 ± 13%     -50.0%       2.10 ± 11%  perf-sched.total_sch_delay.average.ms
    111.28 ±  3%     +10.7%     123.18 ±  5%  perf-sched.total_wait_and_delay.average.ms
     32792 ±  4%     -16.3%      27444 ±  6%  perf-sched.total_wait_and_delay.count.ms
      3631 ± 10%     +24.8%       4531 ±  5%  perf-sched.total_wait_and_delay.max.ms
    107.08 ±  3%     +13.1%     121.09 ±  5%  perf-sched.total_wait_time.average.ms
      3631 ± 10%     +24.8%       4531 ±  5%  perf-sched.total_wait_time.max.ms
    111.28 ±  3%     +10.7%     123.18 ±  5%  perf-sched.wait_and_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
     32792 ±  4%     -16.3%      27444 ±  6%  perf-sched.wait_and_delay.count.[unknown].[unknown].[unknown].[unknown].[unknown]
      3631 ± 10%     +24.8%       4531 ±  5%  perf-sched.wait_and_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    107.08 ±  3%     +13.1%     121.09 ±  5%  perf-sched.wait_time.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      3631 ± 10%     +24.8%       4531 ±  5%  perf-sched.wait_time.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    201981 ±  5%     +34.9%     272400 ± 13%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      2.36 ±  6%     -14.1%       2.03 ±  5%  sched_debug.cfs_rq:/.h_nr_queued.max
      0.40 ±  4%     -23.5%       0.30 ±  4%  sched_debug.cfs_rq:/.h_nr_queued.stddev
      2.25 ±  3%     -11.1%       2.00 ±  4%  sched_debug.cfs_rq:/.h_nr_runnable.max
      0.38 ±  5%     -24.0%       0.29 ±  4%  sched_debug.cfs_rq:/.h_nr_runnable.stddev
     20700 ± 27%     -42.3%      11950 ± 12%  sched_debug.cfs_rq:/.load.avg
    201981 ±  5%     +34.9%     272400 ± 13%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.28 ±  7%     -18.1%       0.23 ±  5%  sched_debug.cfs_rq:/.nr_queued.stddev
    350.87 ±  3%     -12.8%     306.09 ±  4%  sched_debug.cfs_rq:/.runnable_avg.stddev
    335.58 ±  3%     -20.3%     267.63 ±  5%  sched_debug.cfs_rq:/.util_est.stddev
     21.08 ±  8%     +47.3%      31.05 ±  3%  sched_debug.cpu.clock.stddev
      2128 ±  7%     -15.5%       1799 ±  6%  sched_debug.cpu.curr->pid.stddev
      0.43 ±  3%     -27.4%       0.32 ±  5%  sched_debug.cpu.nr_running.stddev
      7223 ±  2%     -11.7%       6379 ±  3%  sched_debug.cpu.nr_switches.avg
      4597           -17.4%       3798 ±  2%  sched_debug.cpu.nr_switches.min
    129.49           +86.1%     241.00        perf-stat.i.MPKI
 8.424e+09           -42.4%  4.852e+09        perf-stat.i.branch-instructions
      0.29            +0.2        0.45        perf-stat.i.branch-miss-rate%
  17058270           -23.6%   13026512 ±  2%  perf-stat.i.branch-misses
     88.13            +2.9       91.01        perf-stat.i.cache-miss-rate%
 3.228e+09            -6.3%  3.023e+09        perf-stat.i.cache-misses
 3.654e+09            -9.3%  3.315e+09        perf-stat.i.cache-references
      6871 ±  2%     -16.7%       5721 ±  4%  perf-stat.i.context-switches
     22.00          +103.7%      44.81        perf-stat.i.cpi
 5.596e+11            +1.8%  5.697e+11        perf-stat.i.cpu-cycles
      1398           -20.9%       1105        perf-stat.i.cpu-migrations
    188.57            +5.6%     199.20        perf-stat.i.cycles-between-cache-misses
 3.807e+10           -40.8%  2.255e+10        perf-stat.i.instructions
      0.08           -39.9%       0.05        perf-stat.i.ipc
      0.07 ± 43%    +152.2%       0.17 ± 19%  perf-stat.i.major-faults
      8762 ±  6%     -11.8%       7728 ±  4%  perf-stat.i.minor-faults
      8762 ±  6%     -11.8%       7728 ±  4%  perf-stat.i.page-faults
    103.32           +51.1%     156.15        perf-stat.overall.MPKI
      0.24            +0.1        0.31 ±  2%  perf-stat.overall.branch-miss-rate%
     88.33            +2.9       91.19        perf-stat.overall.cache-miss-rate%
     17.83           +64.8%      29.38        perf-stat.overall.cpi
    172.58            +9.0%     188.12        perf-stat.overall.cycles-between-cache-misses
      0.06           -39.3%       0.03        perf-stat.overall.ipc
  83258790            -6.9%   77496361        perf-stat.overall.path-length
 6.993e+09           -40.6%  4.156e+09        perf-stat.ps.branch-instructions
  16501261           -22.8%   12743483        perf-stat.ps.branch-misses
  3.25e+09            -7.0%  3.022e+09        perf-stat.ps.cache-misses
 3.679e+09            -9.9%  3.314e+09        perf-stat.ps.cache-references
      6755           -16.5%       5642 ±  3%  perf-stat.ps.context-switches
 5.608e+11            +1.4%  5.684e+11        perf-stat.ps.cpu-cycles
      1399           -21.5%       1099        perf-stat.ps.cpu-migrations
 3.145e+10           -38.5%  1.935e+10        perf-stat.ps.instructions
      0.06 ± 34%    +195.2%       0.16 ± 20%  perf-stat.ps.major-faults
  9.64e+12           -38.5%  5.928e+12        perf-stat.total.instructions




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ