lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <0de7c397-68f0-b217-3890-0c42805f9923@arm.com>
Date:   Thu, 25 Jun 2020 16:21:35 +0100
From:   Lukasz Luba <lukasz.luba@....com>
To:     Qais Yousef <qais.yousef@....com>
Cc:     Ingo Molnar <mingo@...hat.com>,
        Peter Zijlstra <peterz@...radead.org>,
        Juri Lelli <juri.lelli@...hat.com>,
        Vincent Guittot <vincent.guittot@...aro.org>,
        Dietmar Eggemann <dietmar.eggemann@....com>,
        Steven Rostedt <rostedt@...dmis.org>,
        Ben Segall <bsegall@...gle.com>, Mel Gorman <mgorman@...e.de>,
        Patrick Bellasi <patrick.bellasi@...bug.net>,
        Chris Redpath <chris.redpath@....com>,
        linux-kernel@...r.kernel.org
Subject: Re: [PATCH v3 0/2] sched: Optionally skip uclamp logic in fast path

Hi Qais,

On 6/24/20 6:26 PM, Qais Yousef wrote:
> This series attempts to address the report that uclamp logic could be expensive
> sometimes and shows a regression in netperf UDP_STREAM under certain
> conditions.
> 
> The first patch is a fix for how struct uclamp_rq is initialized which is
> required by the 2nd patch which contains the real 'fix'.
> 
> Worth noting that the root cause of the overhead is believed to be system
> specific or related to potential certain code/data layout issues, leading to
> worse I/D $ performance.
> 
> Different systems exhibited different behaviors and the regression did
> disappear in certain kernel version while attempting to reporoduce.
> 
> More info can be found here:
> 
> https://lore.kernel.org/lkml/20200616110824.dgkkbyapn3io6wik@e107158-lin/
> 
> Having the static key seemed the best thing to do to ensure the effect of
> uclamp is minimized for kernels that compile it in but don't have a userspace
> that uses it, which will allow distros to distribute uclamp capable kernels by
> default without having to compromise on performance for some systems that could
> be affected.
> 
> Changes in v3:
> 	* Avoid double negatives and rename the static key to uclamp_used
> 	* Unconditionally enable the static key through any of the paths where
> 	  the user can modify the default uclamp value.
> 	* Use C99 named struct initializer for struct uclamp_rq which is easier
> 	  to read than the memset().
> 
> Changes in v2:
> 	* Add more info in the commit message about the result of perf diff to
> 	  demonstrate that the activate/deactivate_task pressure is reduced in
> 	  the fast path.
> 
> 	* Fix sparse warning reported by the test robot.
> 
> 	* Add an extra commit about using static_branch_likely() instead of
> 	  static_branc_unlikely().
> 
> Thanks
> 
> --
> Qais Yousef
> 
> Cc: Juri Lelli <juri.lelli@...hat.com>
> Cc: Vincent Guittot <vincent.guittot@...aro.org>
> Cc: Dietmar Eggemann <dietmar.eggemann@....com>
> Cc: Steven Rostedt <rostedt@...dmis.org>
> Cc: Ben Segall <bsegall@...gle.com>
> Cc: Mel Gorman <mgorman@...e.de>
> CC: Patrick Bellasi <patrick.bellasi@...bug.net>
> Cc: Chris Redpath <chris.redpath@....com>
> Cc: Lukasz Luba <lukasz.luba@....com>
> Cc: linux-kernel@...r.kernel.org
> 
> Qais Yousef (2):
>    sched/uclamp: Fix initialization of strut uclamp_rq
>    sched/uclamp: Protect uclamp fast path code with static key
> 
>   kernel/sched/core.c | 75 +++++++++++++++++++++++++++++++++++++++------
>   1 file changed, 66 insertions(+), 9 deletions(-)
> 


The results for this v3 series from mmtest netperf-udp (30x each UDP
size) are good.

                        v5.7-rc7-base-noucl    v5.7-rc7-ucl-tsk-nofix 
v5.7-rc7-ucl-tsk-grp-fix_v3
Hmean     send-64          62.15 (   0.00%)       59.65 *  -4.02%* 
65.83 *   5.93%*
Hmean     send-128        122.88 (   0.00%)      119.37 *  -2.85%* 
133.20 *   8.40%*
Hmean     send-256        244.85 (   0.00%)      234.26 *  -4.32%* 
264.01 *   7.83%*
Hmean     send-1024       919.24 (   0.00%)      880.67 *  -4.20%* 
1005.54 *   9.39%*
Hmean     send-2048      1689.45 (   0.00%)     1647.54 *  -2.48%* 
1845.64 *   9.25%*
Hmean     send-3312      2542.36 (   0.00%)     2485.23 *  -2.25%* 
2729.11 *   7.35%*
Hmean     send-4096      2935.69 (   0.00%)     2861.09 *  -2.54%* 
3161.16 *   7.68%*
Hmean     send-8192      4800.35 (   0.00%)     4680.09 *  -2.51%* 
5090.38 *   6.04%*
Hmean     send-16384     7473.66 (   0.00%)     7349.60 *  -1.66%* 
7786.42 *   4.18%*
Hmean     recv-64          62.15 (   0.00%)       59.65 *  -4.03%* 
65.82 *   5.91%*
Hmean     recv-128        122.88 (   0.00%)      119.37 *  -2.85%* 
133.20 *   8.40%*
Hmean     recv-256        244.84 (   0.00%)      234.26 *  -4.32%* 
264.01 *   7.83%*
Hmean     recv-1024       919.24 (   0.00%)      880.67 *  -4.20%* 
1005.54 *   9.39%*
Hmean     recv-2048      1689.44 (   0.00%)     1647.54 *  -2.48%* 
1845.06 *   9.21%*
Hmean     recv-3312      2542.36 (   0.00%)     2485.23 *  -2.25%* 
2728.74 *   7.33%*
Hmean     recv-4096      2935.69 (   0.00%)     2861.09 *  -2.54%* 
3160.74 *   7.67%*
Hmean     recv-8192      4800.35 (   0.00%)     4678.15 *  -2.55%* 
5090.36 *   6.04%*
Hmean     recv-16384     7473.63 (   0.00%)     7349.52 *  -1.66%* 
7786.25 *   4.18%*

I am happy to re-run v4 if there will be, but for now:

Tested-by: Lukasz Luba <lukasz.luba@....com>

Regards,
Lukasz

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ