[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <151240c939d02df4979651b84bb99356a938c44d.camel@intel.com>
Date: Wed, 11 Oct 2023 19:34:33 +0000
From: "Pandruvada, Srinivas" <srinivas.pandruvada@...el.com>
To: "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"anna-maria@...utronix.de" <anna-maria@...utronix.de>
CC: "jstultz@...gle.com" <jstultz@...gle.com>,
"riel@...riel.com" <riel@...riel.com>,
"lukasz.luba@....com" <lukasz.luba@....com>,
"kprateek.nayak@....com" <kprateek.nayak@....com>,
"peterz@...radead.org" <peterz@...radead.org>,
"rostedt@...dmis.org" <rostedt@...dmis.org>,
"frederic@...nel.org" <frederic@...nel.org>,
"tglx@...utronix.de" <tglx@...utronix.de>,
"Wysocki, Rafael J" <rafael.j.wysocki@...el.com>,
"paulmck@...nel.org" <paulmck@...nel.org>,
"arjan@...radead.org" <arjan@...radead.org>,
"edumazet@...gle.com" <edumazet@...gle.com>,
"bigeasy@...utronix.de" <bigeasy@...utronix.de>,
"ggherdovich@...e.cz" <ggherdovich@...e.cz>,
"gautham.shenoy@....com" <gautham.shenoy@....com>
Subject: Re: [PATCH v8 00/25] timer: Move from a push remote at enqueue to a
pull at expiry model
Hi Maria,
On Wed, 2023-10-04 at 14:34 +0200, Anna-Maria Behnsen wrote:
> Hi,
>
>
[...]
>
> The proper solution to this problem is to always queue the timers on
> the
> local CPU and allow the non pinned timers to be pulled onto a busy
> CPU at
> expiry time.
Thanks for these patches. I am looking for saving power during video
playback with our low power daemon. I use cgroup v2 isolation to keep
some CPUs idle (CPU 0-11) and video is played on a single module (CPU
12-15).
I have some kernelshark pictures at below link. The traces are
collected with sched, timer and irq. With 6.6-rc5, you can see some
timers still expires on CPUs which I want to keep idle. With timer
patches added, they are mostly pulled to busy CPU.
https://imgur.com/a/8nF5OoP
I can share the .dat files, but they are too big to attach here.
Thanks,
Srinivas
>
> Therefore split the timer storage into local pinned and global
> timers:
> Local pinned timers are always expired on the CPU on which they have
> been
> queued. Global timers can be expired on any CPU.
>
> As long as a CPU is busy it expires both local and global timers.
> When a
> CPU goes idle it arms for the first expiring local timer. If the
> first
> expiring pinned (local) timer is before the first expiring movable
> timer,
> then no action is required because the CPU will wake up before the
> first
> movable timer expires. If the first expiring movable timer is before
> the
> first expiring pinned (local) timer, then this timer is queued into a
> idle
> timerqueue and eventually expired by some other active CPU.
>
> To avoid global locking the timerqueues are implemented as a
> hierarchy. The
> lowest level of the hierarchy holds the CPUs. The CPUs are associated
> to
> groups of 8, which are separated per node. If more than one CPU group
> exist, then a second level in the hierarchy collects the groups.
> Depending
> on the size of the system more than 2 levels are required. Each group
> has a
> "migrator" which checks the timerqueue during the tick for remote
> timers to
> be expired.
>
> If the last CPU in a group goes idle it reports the first expiring
> event in
> the group up to the next group(s) in the hierarchy. If the last CPU
> goes
> idle it arms its timer for the first system wide expiring timer to
> ensure
> that no timer event is missed.
>
>
> Testing
> ~~~~~~~
>
> Enqueue
> ^^^^^^^
>
> The impact of wasting cycles during enqueue by using the heuristic in
> contrast to always queuing the timer on the local CPU was measured
> with a
> micro benchmark. Therefore a timer is enqueued and dequeued in a loop
> with
> 1000 repetitions on a isolated CPU. The time the loop takes is
> measured. A
> quarter of the remaining CPUs was kept busy. This measurement was
> repeated
> several times. With the patch queue the average duration was reduced
> by
> approximately 25%.
>
> 145ns plain v6
> 109ns v6 with patch queue
>
>
> Furthermore the impact of residence in deep idle states of an idle
> system
> was investigated. The patch queue doesn't downgrade this behavior.
>
> dbench test
> ^^^^^^^^^^^
>
> A dbench test starting X pairs of client servers are used to create
> load on
> the system. The measurable value is the throughput. The tests were
> executed
> on a zen3 machine. The base is the tip tree branch timers/core which
> is
> based on a v6.6-rc1.
>
> governor menu
>
> X pairs timers/core pull-model impact
> ----------------------------------------------
> 1 353.19 (0.19) 353.45 (0.30) 0.07%
> 2 700.10 (0.96) 687.00 (0.20) -1.87%
> 4 1329.37 (0.63) 1282.91 (0.64) -3.49%
> 8 2561.16 (1.28) 2493.56 (1.76) -2.64%
> 16 4959.96 (0.80) 4914.59 (0.64) -0.91%
> 32 9741.92 (3.44) 8979.83 (1.13) -7.82%
> 64 16535.40 (2.84) 16388.47 (4.02) -0.89%
> 128 22136.83 (2.42) 23174.50 (1.43) 4.69%
> 256 39256.77 (4.48) 38994.00 (0.39) -0.67%
> 512 36799.03 (1.83) 38091.10 (0.63) 3.51%
> 1024 32903.03 (0.86) 35370.70 (0.89) 7.50%
>
>
> governor teo
>
> X pairs timers/core pull-model impact
> ----------------------------------------------
> 1 350.83 (1.27) 352.45 (0.96) 0.46%
> 2 699.52 (0.85) 690.10 (0.54) -1.35%
> 4 1339.53 (1.99) 1294.71 (2.71) -3.35%
> 8 2574.10 (0.76) 2495.46 (1.97) -3.06%
> 16 4898.50 (1.74) 4783.06 (1.64) -2.36%
> 32 9115.50 (4.63) 9037.83 (1.58) -0.85%
> 64 16663.90 (3.80) 16042.00 (1.72) -3.73%
> 128 25044.93 (1.11) 23250.03 (1.08) -7.17%
> 256 38059.53 (1.70) 39658.57 (2.98) 4.20%
> 512 36369.30 (0.39) 38890.13 (0.36) 6.93%
> 1024 33956.83 (1.14) 35514.83 (0.29) 4.59%
>
>
>
> Ping Pong Oberservation
> ^^^^^^^^^^^^^^^^^^^^^^^
>
> During testing on a mostly idle machine a ping pong game could be
> observed:
> a process_timeout timer is expired remotely on a non idle CPU. Then
> the CPU
> where the schedule_timeout() was executed to enqueue the timer comes
> out of
> idle and restarts the timer using schedule_timeout() and goes back to
> idle
> again. This is due to the fair scheduler which tries to keep the task
> on
> the CPU which it previously executed on.
>
>
>
>
> Possible Next Steps
> ~~~~~~~~~~~~~~~~~~~
>
> Simple deferrable timers are no longer required as they can be
> converted to
> global timers. If a CPU goes idle, a formerly deferrable timer will
> not
> prevent the CPU to sleep as long as possible. Only the last migrator
> CPU
> has to take care of them. Deferrable timers with timer pinned flags
> needs
> to be expired on the specified CPU but must not prevent CPU from
> going
> idle. They require their own timer base which is never taken into
> account
> when calculating the next expiry time. This conversation and required
> cleanup will be done in a follow up series.
>
>
> v7..v8:
> https://lore.kernel.org/r/20230524070629.6377-1-anna-maria@linutronix.de
> - Address review feedback
> - Move marking timer base idle into tick_nohz_stop_tick()
> - Look ahead function to determine possible sleep lenght
>
>
> v6..v7:
> - Address review feedback of Frederic and bigeasy
> - Change lock, unlock fetch next timer interrupt logic after remote
> expiry
> - Move timer_expire_remote() into tick-internal.h
> - Add documentation section about "Required event and timerqueue
> update
> after remote expiry"
> - Fix fallout of kernel test robot
>
>
> v5..v6:
>
> - Address review of Frederic Weisbecker and Peter Zijlstra
> (spelling,
> locking, race in tmigr_handle_remote_cpu())
>
> - unconditionally set TIMER_PINNED flag in add_timer_on();
> introduce
> add_timer() variants which set/unset TIMER_PINNED flag; drop
> fixing
> add_timer_on() call sites, as TIMER_PINNED flag is set
> implicitly;
> Fixing workqueue to use add_timer_global() instead of simply
> add_timer() for unbound work.
>
> - Drop support for siblings to end up in the same level 0 group
> (could be
> added again in a better way as an improvement later on)
>
> - Do not send IPI for new first deferrable timers
>
> v4..v5:
> - address review feedback of Frederic Weisbecker
> - fix issue with group timer update after remote expiry
>
> v3..v4:
> - address review feedback of Frederic Weisbecker
> - address kernel test robot fallout
> - Move patch 16 "add_timer_on(): Make sure callers have
> TIMER_PINNED
> flag" at the begin of the queue to prevent timers to end up in
> global
> timer base when they were queued using add_timer_on()
> - Fix some comments and typos
>
> v2..v3:
> https://lore.kernel.org/r/20170418111102.490432548@linutronix.de/
> - Minimize usage of locks by storing data using atomic_cmpxchg()
> for
> migrator information and information about active cpus.
>
>
> Thanks,
>
> Anna-Maria
>
>
>
> Anna-Maria Behnsen (22):
> tick/sched: Cleanup confusing variables
> tick-sched: Warn when next tick seems to be in the past
> timer: Do not IPI for deferrable timers
> timer: Move store of next event into __next_timer_interrupt()
> timers: Clarify check in forward_timer_base()
> timers: Split out forward timer base functionality
> timers: Use already existing function for forwarding timer base
> timer: Split out get next timer functionality
> timers: Move marking timer bases idle into tick_nohz_stop_tick()
> timers: Introduce add_timer() variants which modify timer flags
> workqueue: Use global variant for add_timer()
> timer: add_timer_on(): Make sure TIMER_PINNED flag is set
> timers: Ease code in run_local_timers()
> timer: Split next timer interrupt logic
> timer: Keep the pinned timers separate from the others
> timer: Retrieve next expiry of pinned/non-pinned timers separately
> timer: Split out "get next timer interrupt" functionality
> timer: Add get next timer interrupt functionality for remote CPUs
> timer: Check if timers base is handled already
> timer: Implement the hierarchical pull model
> timer_migration: Add tracepoints
> timer: Always queue timers on the local CPU
>
> Richard Cochran (linutronix GmbH) (2):
> timer: Restructure internal locking
> tick/sched: Split out jiffies update helper function
>
> Thomas Gleixner (1):
> timer: Rework idle logic
>
> include/linux/cpuhotplug.h | 1 +
> include/linux/timer.h | 16 +-
> include/trace/events/timer_migration.h | 283 ++++
> kernel/time/Makefile | 3 +
> kernel/time/tick-internal.h | 13 +
> kernel/time/tick-sched.c | 69 +-
> kernel/time/timer.c | 514 ++++++--
> kernel/time/timer_migration.c | 1636
> ++++++++++++++++++++++++
> kernel/time/timer_migration.h | 144 +++
> kernel/workqueue.c | 2 +-
> 10 files changed, 2549 insertions(+), 132 deletions(-)
> create mode 100644 include/trace/events/timer_migration.h
> create mode 100644 kernel/time/timer_migration.c
> create mode 100644 kernel/time/timer_migration.h
>
Powered by blists - more mailing lists