lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <bbc24579-b6ee-37cb-4bbf-10e3476537e0@intel.com>
Date:   Tue, 7 Dec 2021 15:14:38 -0800
From:   Dave Hansen <dave.hansen@...el.com>
To:     kernel test robot <oliver.sang@...el.com>,
        Thomas Gleixner <tglx@...utronix.de>
Cc:     Borislav Petkov <bp@...e.de>,
        "Chang S. Bae" <chang.seok.bae@...el.com>,
        LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org,
        lkp@...el.com, ying.huang@...el.com, feng.tang@...el.com,
        zhengjun.xing@...ux.intel.com, fengwei.yin@...el.com
Subject: Re: [x86/signal] 3aac3ebea0: will-it-scale.per_thread_ops -11.9%
 regression

On 12/6/21 5:21 PM, kernel test robot wrote:
> 
> 1bdda24c4af64cd2 3aac3ebea08f2d342364f827c89 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>     980404 ±  3%     -10.2%     880436 ±  2%  will-it-scale.16.threads
>      61274 ±  3%     -10.2%      55027 ±  2%  will-it-scale.per_thread_ops
>     980404 ±  3%     -10.2%     880436 ±  2%  will-it-scale.workload
>    9745749 ± 18%     +26.8%   12356608 ±  4%  meminfo.DirectMap2M

Something else funky is going on here.  Why would there all of a sudden
be so many more 2M pages in the direct map?  I also see gunk like
interrupts on the network card going up.  I can certainly see that
happening if something else on the network was messing around.

Granted, this was seen across several systems, but it's really odd.  I
guess I'll go try to dig up one of the actual ones where this was seen.

I tried on a smaller Skylake system and I don't see any regression at
all or any interesting delta in a perf profile.

Oliver or Chang, could you try to reproduce this by hand on one of the
suspect systems?  Build:

  1bdda24c4a ("signal: Add an optional check for altstack size")

then run will-it-scale by hand.  Then build:

  3aac3ebea0 ("x86/signal: Implement sigaltstack size validation")

and run it again.  Also, do we see any higher core-count regressions?
These all seem to happen with:

	mode=thread
	nr_task=16

That's really odd to see that for these systems with probably ~50 cores
each.  I'd expect to see it get worse at higher core counts.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ