lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 19 Feb 2014 21:12:35 +0800
From:	Fengguang Wu <fengguang.wu@...el.com>
To:	"Eric W. Biederman" <ebiederm@...ssion.com>
Cc:	LKML <linux-kernel@...r.kernel.org>
Subject: [vfs] 73c12e50a: +13349.5% softirqs.RCU

Hi Eric,

FYI, we noticed the below changes on commit 73c12e50af3a756137071c1e1c1e99cffcbaa835
("vfs: Lazily remove mounts on unlinked files and directories."):
    
test case: dbench

5c39777e797e7ba  73c12e50af3a756137071c1e1  
---------------  -------------------------  
      6.22 ~ 0%    +195.3%      18.36 ~18%  TOTAL dbench.max_latency

Note: the "~ XX%" numbers are stddev percent.

test case: will-it-scale/unlink2

5c39777e797e7ba  73c12e50af3a756137071c1e1  
---------------  -------------------------  
     57178 ~ 2%     -83.9%       9188 ~ 3%  TOTAL numa-vmstat.node1.nr_slab_unreclaimable
     97777 ~ 1%     -85.3%      14413 ~ 6%  TOTAL numa-vmstat.node0.nr_slab_unreclaimable
    227271 ~ 2%     -83.9%      36567 ~ 4%  TOTAL numa-meminfo.node1.SUnreclaim
     78384 ~ 2%  +13349.5%   10542303 ~11%  TOTAL softirqs.RCU
    202419 ~15%  +48738.0%   98857773 ~ 4%  TOTAL cpuidle.C1-SNB.time
    955612 ~31%   +2561.5%   25433795 ~10%  TOTAL cpuidle.C1E-SNB.time
      1278 ~20%   +8648.3%     111872 ~11%  TOTAL cpuidle.C1E-SNB.usage
    364333 ~11%    +786.3%    3229011 ~17%  TOTAL cpuidle.C3-SNB.time
       793 ~ 5%    +568.6%       5305 ~12%  TOTAL cpuidle.C3-SNB.usage
     10882 ~15%   +1943.2%     222341 ~18%  TOTAL cpuidle.C6-SNB.time
         8 ~27%   +4702.4%        403 ~17%  TOTAL cpuidle.C6-SNB.usage
    389688 ~ 2%     -85.3%      57320 ~ 5%  TOTAL numa-meminfo.node0.SUnreclaim
    485651 ~ 2%     -82.5%      84770 ~ 5%  TOTAL numa-meminfo.node0.Slab
     19822 ~ 4%     -91.7%       1649 ~ 7%  TOTAL slabinfo.kmalloc-256.num_slabs
     19822 ~ 4%     -91.7%       1649 ~ 7%  TOTAL slabinfo.kmalloc-256.active_slabs
    634311 ~ 4%     -91.7%      52792 ~ 7%  TOTAL slabinfo.kmalloc-256.num_objs
    768494 ~ 3%     -81.0%     146268 ~ 2%  TOTAL meminfo.Slab
    634179 ~ 4%     -91.7%      52517 ~ 7%  TOTAL slabinfo.kmalloc-256.active_objs
     16094 ~ 3%     -85.4%       2343 ~ 3%  TOTAL slabinfo.dentry.num_slabs
     16094 ~ 3%     -85.4%       2343 ~ 3%  TOTAL slabinfo.dentry.active_slabs
    675979 ~ 3%     -85.4%      98427 ~ 3%  TOTAL slabinfo.dentry.num_objs
    675739 ~ 3%     -85.5%      98065 ~ 3%  TOTAL slabinfo.dentry.active_objs
     12925 ~ 4%     -91.6%       1081 ~ 7%  TOTAL slabinfo.shmem_inode_cache.num_slabs
     12925 ~ 4%     -91.6%       1081 ~ 7%  TOTAL slabinfo.shmem_inode_cache.active_slabs
    153617 ~ 3%     -84.5%      23851 ~ 3%  TOTAL proc-vmstat.nr_slab_unreclaimable
    633368 ~ 4%     -91.6%      53034 ~ 7%  TOTAL slabinfo.shmem_inode_cache.num_objs
    633182 ~ 4%     -91.7%      52584 ~ 7%  TOTAL slabinfo.shmem_inode_cache.active_objs
    611608 ~ 3%     -84.4%      95315 ~ 3%  TOTAL meminfo.SUnreclaim
     56626 ~ 0%    +977.2%     609963 ~ 4%  TOTAL interrupts.RES
    289471 ~ 1%     -79.4%      59729 ~ 4%  TOTAL numa-meminfo.node1.Slab
     95962 ~ 3%     -71.4%      27449 ~ 7%  TOTAL numa-meminfo.node0.SReclaimable
     24059 ~ 3%     -71.4%       6870 ~ 7%  TOTAL numa-vmstat.node0.nr_slab_reclaimable
     39360 ~ 2%     -67.6%      12741 ~ 1%  TOTAL proc-vmstat.nr_slab_reclaimable
    156885 ~ 2%     -67.5%      50952 ~ 1%  TOTAL meminfo.SReclaimable
     62199 ~ 1%     -62.8%      23161 ~ 7%  TOTAL numa-meminfo.node1.SReclaimable
     15622 ~ 1%     -62.9%       5795 ~ 7%  TOTAL numa-vmstat.node1.nr_slab_reclaimable
     40507 ~39%    +153.7%     102749 ~19%  TOTAL cpuidle.C1-NHM.time
     23068 ~ 2%    +135.2%      54252 ~ 9%  TOTAL interrupts.IWI
       357 ~27%    +116.4%        774 ~15%  TOTAL cpuidle.C3-NHM.usage
   4834003 ~ 3%     -47.6%    2534625 ~ 4%  TOTAL numa-numastat.node1.local_node
   4834003 ~ 3%     -47.6%    2534625 ~ 4%  TOTAL numa-numastat.node1.numa_hit
      2167 ~23%     -28.0%       1560 ~ 6%  TOTAL proc-vmstat.nr_tlb_remote_flush_received
  58685570 ~ 2%     -38.5%   36079286 ~ 2%  TOTAL proc-vmstat.pgalloc_normal
  68325712 ~ 1%     -37.7%   42551512 ~ 2%  TOTAL proc-vmstat.pgfree
   1079369 ~ 0%     -36.9%     681261 ~ 0%  TOTAL numa-meminfo.node0.MemUsed
    140248 ~ 2%     -35.2%      90832 ~ 5%  TOTAL softirqs.SCHED
  19460075 ~ 1%     -36.5%   12353959 ~ 2%  TOTAL proc-vmstat.numa_local
  19460075 ~ 1%     -36.5%   12353960 ~ 2%  TOTAL proc-vmstat.numa_hit
     53926 ~24%     -33.4%      35902 ~ 2%  TOTAL meminfo.DirectMap4k
        85 ~31%     +79.6%        153 ~28%  TOTAL cpuidle.C1-NHM.usage
   9675753 ~ 1%     -33.1%    6474766 ~ 1%  TOTAL proc-vmstat.pgalloc_dma32
  14654234 ~ 1%     -32.9%    9829336 ~ 1%  TOTAL numa-numastat.node0.local_node
  14654234 ~ 1%     -32.9%    9829336 ~ 1%  TOTAL numa-numastat.node0.numa_hit
   1878635 ~ 1%     -31.6%    1284774 ~ 2%  TOTAL numa-vmstat.node1.numa_local
   1886571 ~ 1%     -31.5%    1292725 ~ 2%  TOTAL numa-vmstat.node1.numa_hit
      1425 ~ 7%     -20.0%       1140 ~10%  TOTAL numa-meminfo.node1.Mlocked
       356 ~ 7%     -20.0%        285 ~10%  TOTAL numa-vmstat.node1.nr_mlock
       287 ~ 9%     +24.8%        358 ~ 8%  TOTAL numa-vmstat.node0.nr_mlock
      1150 ~ 9%     +24.8%       1435 ~ 8%  TOTAL numa-meminfo.node0.Mlocked
      1441 ~ 7%     -19.8%       1156 ~10%  TOTAL numa-meminfo.node1.Unevictable
       360 ~ 7%     -19.8%        289 ~10%  TOTAL numa-vmstat.node1.nr_unevictable
      1166 ~ 9%     +24.4%       1451 ~ 8%  TOTAL numa-meminfo.node0.Unevictable
       291 ~ 9%     +24.4%        362 ~ 8%  TOTAL numa-vmstat.node0.nr_unevictable
    871951 ~ 0%     -25.9%     645735 ~ 0%  TOTAL numa-meminfo.node1.MemUsed
   8508242 ~ 0%     -24.4%    6432405 ~ 1%  TOTAL numa-vmstat.node0.numa_local
    225490 ~ 2%     -24.1%     171244 ~ 2%  TOTAL cpuidle.C7-SNB.usage
   8623332 ~ 0%     -24.1%    6547495 ~ 1%  TOTAL numa-vmstat.node0.numa_hit
       313 ~28%     -30.7%        217 ~21%  TOTAL cpuidle.C1E-NHM.usage
      1943 ~ 0%   +7380.4%     145404 ~11%  TOTAL time.voluntary_context_switches
      1001 ~ 0%   +1015.5%      11173 ~ 4%  TOTAL vmstat.system.cs
      4815 ~ 0%     +34.0%       6452 ~ 1%  TOTAL vmstat.system.in
      1480 ~ 0%      -1.7%       1454 ~ 0%  TOTAL time.system_time
       485 ~ 0%      -1.6%        477 ~ 0%  TOTAL time.percent_of_cpu_this_job_got
     50.66 ~ 0%      -1.4%      49.97 ~ 0%  TOTAL turbostat.%c0
      7.69 ~ 0%      +1.1%       7.78 ~ 0%  TOTAL boottime.dhcp
       106 ~ 0%      -0.5%        106 ~ 0%  TOTAL turbostat.Cor_W
       134 ~ 0%      -0.4%        133 ~ 0%  TOTAL turbostat.Pkg_W

                           time.voluntary_context_switches

   20000 ++-----------------------------------------------------------------+
   18000 ++                             O    O          O                   |
         |                                                       O          |
   16000 ++                                                O                |
   14000 ++ O                        O                                      |
         O     O                  O        O    O                   O       |
   12000 ++                                                   O             |
   10000 ++                                        O  O                     |
    8000 ++                                                                 |
         |       O  O  O  O  O O                                            |
    6000 ++                                                                 |
    4000 ++                                                                 |
         |                                                                  |
    2000 *+.*..*.*..*..*..*..*.*..*..*..*..*.*..*..*..*.*..*..*..*..*.*..*..*
       0 ++-----------------------------------------------------------------+

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ