[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <363efc75-7ebf-408d-be2d-e6c579674f44@paulmck-laptop>
Date: Mon, 20 May 2024 13:03:18 -0700
From: "Paul E. McKenney" <paulmck@...nel.org>
To: Lai Jiangshan <jiangshanlai@...il.com>
Cc: linux-kernel@...r.kernel.org, rcu@...r.kernel.org, x86@...nel.org,
Lai Jiangshan <jiangshan.ljs@...group.com>,
Peter Zijlstra <peterz@...radead.org>,
Frederic Weisbecker <frederic@...nel.org>
Subject: Re: [PATCH V2 00/11] rcu/x86: Use per-cpu rcu preempt count
On Sun, Apr 07, 2024 at 05:05:47PM +0800, Lai Jiangshan wrote:
> From: Lai Jiangshan <jiangshan.ljs@...group.com>
>
>
> Changed from v1:
> Merge thunk_64.S and thunk_32.S into thunk.S
> Add missing #ifdef in arch/x86/kernel/cpu/common.c
>
> X86 can access percpu data in a single instruction.
>
> Use per-cpu rcu preempt count and make it able to be inlined.
>
> patch 1-8: prepare
> patch 9-11: implement PCPU_RCU_PREEMPT_COUNT
>
> Cc: "Paul E. McKenney" <paulmck@...nel.org>
> Cc: Peter Zijlstra <peterz@...radead.org>
> Cc: Frederic Weisbecker <frederic@...nel.org>
Hello, Lai!
This is really cool stuff, thank you!!!
Two big questions remain: (1) What system-level net performance benefit
is there, taking the increased context-switch overhead into account and
(2) Are the scheduler maintainers on board with these changes?
On #1, I do well recall your ca. 2019 points about the improved code
generated, but I have seen cases where improved code actually ran
more slowly. My guess is that you have the best chance of seeing
system-level benefits on low-end x86 platforms, perhaps the old Atom
or Celeron systems. The rcuref module provides a good way of doing
microbenchmarks, which would be a good start. Other metrics that
might help include overall kernel code size.
On #2, good data for #1 would help greatly.
Thoughts?
Thanx, Paul
> Lai Jiangshan (11):
> lib: Use rcu_preempt_depth() to replace current->rcu_read_lock_nesting
> rcu: Move rcu_preempt_depth_set() to rcupdate.h
> rcu: Reorder tree_exp.h after tree_plugin.h
> rcu: Add macros set_rcu_preempt_special() and
> clear_rcu_preempt_special()
> rcu: Make rcu_read_unlock_special() global
> rcu: Rename marco __LINUX_RCU_H to __KERNEL_RCU_H
> sched/core: Add rcu_preempt_switch()
> x86/entry: Merge thunk_64.S and thunk_32.S into thunk.S
> rcu: Implement PCPU_RCU_PREEMPT_COUNT framework
> x86/rcu: Add rcu_preempt_count
> x86/rcu: Add THUNK rcu_read_unlock_special_thunk
>
> arch/x86/Kconfig | 1 +
> arch/x86/entry/Makefile | 2 +-
> arch/x86/entry/{thunk_64.S => thunk.S} | 5 ++
> arch/x86/entry/thunk_32.S | 18 ----
> arch/x86/include/asm/current.h | 3 +
> arch/x86/include/asm/rcu_preempt.h | 109 +++++++++++++++++++++++++
> arch/x86/kernel/cpu/common.c | 4 +
> include/linux/rcupdate.h | 36 ++++++++
> kernel/rcu/Kconfig | 8 ++
> kernel/rcu/rcu.h | 15 +++-
> kernel/rcu/tree.c | 2 +-
> kernel/rcu/tree_exp.h | 2 +-
> kernel/rcu/tree_plugin.h | 41 ++++++----
> kernel/sched/core.c | 2 +
> lib/locking-selftest.c | 6 +-
> 15 files changed, 212 insertions(+), 42 deletions(-)
> rename arch/x86/entry/{thunk_64.S => thunk.S} (72%)
> delete mode 100644 arch/x86/entry/thunk_32.S
> create mode 100644 arch/x86/include/asm/rcu_preempt.h
>
>
> base-commit: f2f80ac809875855ac843f9e5e7480604b5cbff5
> --
> 2.19.1.6.gb485710b
>
Powered by blists - more mailing lists