[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAKfTPtDETjTEno8XURNVqzkqOONuAYRnhvNXmHzC7Hc_crwA_Q@mail.gmail.com>
Date: Mon, 27 Sep 2021 16:19:00 +0200
From: Vincent Guittot <vincent.guittot@...aro.org>
To: Mel Gorman <mgorman@...hsingularity.net>
Cc: Mike Galbraith <efault@....de>,
Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...nel.org>,
Valentin Schneider <valentin.schneider@....com>,
Aubrey Li <aubrey.li@...ux.intel.com>,
Barry Song <song.bao.hua@...ilicon.com>,
Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 2/2] sched/fair: Scale wakeup granularity relative to nr_running
On Mon, 27 Sept 2021 at 13:17, Mel Gorman <mgorman@...hsingularity.net> wrote:
>
> On Thu, Sep 23, 2021 at 02:41:06PM +0200, Vincent Guittot wrote:
> > On Thu, 23 Sept 2021 at 11:22, Mike Galbraith <efault@....de> wrote:
> > >
> > > On Thu, 2021-09-23 at 10:40 +0200, Vincent Guittot wrote:
> > > >
> > > > a 100us value should even be enough to fix Mel's problem without
> > > > impacting common wakeup preemption cases.
> > >
> > > It'd be nice if it turn out to be something that simple, but color me
> > > skeptical. I've tried various preemption throttling schemes, and while
> >
> > Let's see what the results will show. I tend to agree that this will
> > not be enough to cover all use cases and I don't see any other way to
> > cover all cases than getting some inputs from the threads about their
> > latency fairness which bring us back to some kind of latency niceness
> > value
> >
>
> Unfortunately, I didn't get a complete set of results but enough to work
> with. The missing tests have been requeued. The figures below are based
> on a single-socket Skylake machine with 8 CPUs as it had the most set of
> results and is the basic case.
>
> The reported kernels are
>
> vanilla: vanilla 5.15-rc1
> sched-scalewakegran-v2r4: My patch
> sched-moveforward-v1r1: Vincent's patch
I imagine that this is the results for the 1st version which scales
with the number of CPUs
>
>
>
> hackbench-process-pipes
> 5.15.0-rc1 5.15.0-rc1 5.15.0-rc1
> vanilla sched-scalewakegran-v2r4 sched-moveforward-v1r1
> Amean 1 0.3253 ( 0.00%) 0.3330 ( -2.36%) 0.3257 ( -0.10%)
> Amean 4 0.8300 ( 0.00%) 0.7570 ( 8.80%) 0.7560 ( 8.92%)
> Amean 7 1.1003 ( 0.00%) 1.1457 * -4.12%* 1.1163 ( -1.45%)
> Amean 12 1.7263 ( 0.00%) 1.6393 * 5.04%* 1.5963 * 7.53%*
> Amean 21 3.0063 ( 0.00%) 2.6590 * 11.55%* 2.4487 * 18.55%*
> Amean 30 4.2323 ( 0.00%) 3.5657 * 15.75%* 3.3410 * 21.06%*
> Amean 48 6.5657 ( 0.00%) 5.4180 * 17.48%* 5.0857 * 22.54%*
> Amean 79 10.4867 ( 0.00%) 8.4357 * 19.56%* 7.9563 * 24.13%*
> Amean 110 14.8880 ( 0.00%) 11.0423 * 25.83%* 10.7407 * 27.86%*
> Amean 141 19.2083 ( 0.00%) 14.0820 * 26.69%* 13.3780 * 30.35%*
> Amean 172 23.4847 ( 0.00%) 16.9880 * 27.66%* 16.4293 * 30.04%*
> Amean 203 27.3763 ( 0.00%) 20.2480 * 26.04%* 19.6430 * 28.25%*
> Amean 234 31.3707 ( 0.00%) 23.2477 * 25.89%* 22.8287 * 27.23%*
> Amean 265 35.4663 ( 0.00%) 26.2483 * 25.99%* 25.8683 * 27.06%*
> Amean 296 39.2380 ( 0.00%) 29.4237 * 25.01%* 28.8727 * 26.42%*
>
> For hackbench, either Vincent or my patch has a similar impact.
>
> tbench4
> 5.15.0-rc1 5.15.0-rc1 5.15.0-rc1
> vanillasched-scalewakegran-v2r4 sched-moveforward-v1r1
> Hmean 1 598.71 ( 0.00%) 608.31 * 1.60%* 586.05 * -2.11%*
> Hmean 2 1096.74 ( 0.00%) 1110.07 * 1.22%* 1106.70 * 0.91%*
> Hmean 4 1529.35 ( 0.00%) 1531.20 * 0.12%* 1551.11 * 1.42%*
> Hmean 8 2824.32 ( 0.00%) 2847.96 * 0.84%* 2684.21 * -4.96%*
> Hmean 16 2573.30 ( 0.00%) 2591.77 * 0.72%* 2445.41 * -4.97%*
> Hmean 32 2518.77 ( 0.00%) 2532.70 * 0.55%* 2409.30 * -4.35%*
>
> For tbench, it's ok for lower thread counts for 8 threads (machine
> overloaded), Vincent's patch regresses slightly. With these test runs,
> I don't have detailed information as to why but the most likely solution
> is that preemption gets disabled prematurely.
>
> specjbb
> 5.15.0-rc1 5.15.0-rc1 5.15.0-rc1
> vanillasched-scalewakegran-v2r4 sched-moveforward-v1r1
> Hmean tput-1 71199.00 ( 0.00%) 69492.00 * -2.40%* 71126.00 * -0.10%*
> Hmean tput-2 154478.00 ( 0.00%) 146060.00 * -5.45%* 153073.00 * -0.91%*
> Hmean tput-3 211889.00 ( 0.00%) 209386.00 * -1.18%* 219434.00 * 3.56%*
> Hmean tput-4 257842.00 ( 0.00%) 248012.00 * -3.81%* 262903.00 * 1.96%*
> Hmean tput-5 253506.00 ( 0.00%) 242511.00 * -4.34%* 250828.00 * -1.06%*
> Hmean tput-6 246202.00 ( 0.00%) 236480.00 * -3.95%* 244236.00 * -0.80%*
> Hmean tput-7 241133.00 ( 0.00%) 230905.00 * -4.24%* 237619.00 * -1.46%*
> Hmean tput-8 237983.00 ( 0.00%) 230010.00 * -3.35%* 235275.00 * -1.14%*
>
> For specjbb, it's different again, Vincent's patch is better for the
> overloaded case but both patches show light regressions.
>
> --
> Mel Gorman
> SUSE Labs
Powered by blists - more mailing lists