lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <520f0cf10908210512n7e426d1es4b5b0e0d953af7c5@mail.gmail.com>
Date:	Fri, 21 Aug 2009 14:12:08 +0200
From:	John Kacur <jkacur@...il.com>
To:	Peter Zijlstra <peterz@...radead.org>
Cc:	Thomas Gleixner <tglx@...utronix.de>, Ingo Molnar <mingo@...e.hu>,
	linux-kernel <linux-kernel@...r.kernel.org>, dinogun@...il.com
Subject: Re: [PATCH -rt] timer: delay waking softirqs from the jiffy tick

On Fri, Aug 21, 2009 at 11:56 AM, Peter Zijlstra<peterz@...radead.org> wrote:
> Hi,
>
> most people were complaining about broken balancing with the recent -rt
> series.
>
> A look at /proc/sched_debug yielded:
>
> cpu#0, 2393.874 MHz
>  .nr_running                    : 0
>  .load                          : 0
>  .cpu_load[0]                   : 177522
>  .cpu_load[1]                   : 177522
>  .cpu_load[2]                   : 177522
>  .cpu_load[3]                   : 177522
>  .cpu_load[4]                   : 177522
> cpu#1, 2393.874 MHz
>  .nr_running                    : 4
>  .load                          : 4096
>  .cpu_load[0]                   : 181618
>  .cpu_load[1]                   : 180850
>  .cpu_load[2]                   : 180274
>  .cpu_load[3]                   : 179938
>  .cpu_load[4]                   : 179758
>
>
> Which indicated the cpu_load computation was hosed, the 177522 value
> indicates that there is one RT task runnable. Initially I thought the
> old problem of calculating the cpu_load from a softirq had re-surfaced,
> however looking at the code shows its being done from scheduler_tick().
>
> [ we really should fix this RT/cfs interaction some day... ]
>
> A few trace_printk()s later:
>
>    sirq-timer/1-19    [001]   174.289744:     19: 50:S ==> [001]     0:140:R <idle>
>          <idle>-0     [001]   174.290724: enqueue_task_rt: adding task: 19/sirq-timer/1 with load: 177522
>          <idle>-0     [001]   174.290725:      0:140:R   + [001]    19: 50:S sirq-timer/1
>          <idle>-0     [001]   174.290730: scheduler_tick: current load: 177522
>          <idle>-0     [001]   174.290732: scheduler_tick: current: 0/swapper
>          <idle>-0     [001]   174.290736:      0:140:R ==> [001]    19: 50:R sirq-timer/1
>    sirq-timer/1-19    [001]   174.290741: dequeue_task_rt: removing task: 19/sirq-timer/1 with load: 177522
>    sirq-timer/1-19    [001]   174.290743:     19: 50:S ==> [001]     0:140:R <idle>
>
> We see that we always raise the timer softirq before doing the load
> calculation. Avoid this by re-ordering the scheduler_tick() call in
> update_process_times() to occur before we deal with timers.
>
> This lowers the load back to sanity and restores regular load-balancing
> behaviour.
>
> Signed-off-by: Peter Zijlstra <a.p.zijlstra@...llo.nl>
> ---
>  kernel/timer.c |    2 +-
>  1 files changed, 1 insertions(+), 1 deletions(-)
>
> diff --git a/kernel/timer.c b/kernel/timer.c
> index 8137cce..96ac1b4 100644
> --- a/kernel/timer.c
> +++ b/kernel/timer.c
> @@ -1221,10 +1221,10 @@ void update_process_times(int user_tick)
>
>        /* Note: this timer irq context must be accounted for as well. */
>        account_process_tick(p, user_tick);
> +       scheduler_tick();
>        run_local_timers();
>        if (rcu_pending(cpu))
>                rcu_check_callbacks(cpu, user_tick);
> -       scheduler_tick();
>        run_posix_cpu_timers(p);
>  }
>
>
> --

Cool! I applied this to the v2.6.31-rc6-rt5 tree and with the
following results from /proc/sched_debug

Before applying the patch:

cpu#0, 2792.838 MHz
  .cpu_load[0]                   : 180594
  .cpu_load[1]                   : 192061
  .cpu_load[2]                   : 205170
  .cpu_load[3]                   : 204449
  .cpu_load[4]                   : 199281

cpu#1, 2792.838 MHz
  .cpu_load[0]                   : 177522
  .cpu_load[1]                   : 178378
  .cpu_load[2]                   : 178932
  .cpu_load[3]                   : 178808

After applying the patch:

cpu#0, 2792.847 MHz
  .cpu_load[0]                   : 1024
  .cpu_load[1]                   : 960
  .cpu_load[2]                   : 700
  .cpu_load[3]                   : 430
  .cpu_load[4]                   : 260

cpu#1, 2792.847 MHz
  .cpu_load[0]                   : 1024
  .cpu_load[1]                   : 1280
  .cpu_load[2]                   : 1393
  .cpu_load[3]                   : 1390
  .cpu_load[4]                   : 1871
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ