[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAEXW_YTQpAKLRp4Nv-zM7NJ7JAnuPKSaLicwzL6w9eqMXQFQYQ@mail.gmail.com>
Date: Mon, 27 Jan 2025 19:09:34 -0500
From: Joel Fernandes <joel@...lfernandes.org>
To: linux-kernel@...r.kernel.org, "Paul E. McKenney" <paulmck@...nel.org>,
Frederic Weisbecker <frederic@...nel.org>, Neeraj Upadhyay <neeraj.upadhyay@...nel.org>,
Joel Fernandes <joel@...lfernandes.org>, Josh Triplett <josh@...htriplett.org>,
Boqun Feng <boqun.feng@...il.com>, Uladzislau Rezki <urezki@...il.com>,
Steven Rostedt <rostedt@...dmis.org>, Mathieu Desnoyers <mathieu.desnoyers@...icios.com>,
Lai Jiangshan <jiangshanlai@...il.com>, Zqiang <qiang.zhang1211@...il.com>
Cc: rcu@...r.kernel.org
Subject: Re: [PATCH] rcu: Merge rcu_seq_done_exact() logic into rcu_seq_done()
On Mon, Jan 27, 2025 at 7:07 PM Joel Fernandes (Google)
<joel@...lfernandes.org> wrote:
>
> The rcu_seq_done() API has a large "false-negative" windows of size
> ULONG_MAX/2, where after wrap around, it is possible that it will think
> that a GP has not completed if a wrap around happens and the delta is
> large.
>
> rcu_seq_done_exact() is more accurate avoiding this wrap around issue,
> by making the window of false-negativity by only 3 GPs. Use this logic
> for rcu_seq_done() which is a nice negative code delta and could
> potentially avoid issues in the future where rcu_seq_done() was
> reporting false-negatives for too long.
>
> rcutorture runs of all scenarios for 15 minutes passed. Code inspection
> was done of all users to convince the change would work.
>
> Signed-off-by: Joel Fernandes (Google) <joel@...lfernandes.org>
I am leaving a 60 minute overnight run of all scenarios on my personal
server for further testing.
thanks,
- Joel
> ---
> kernel/rcu/rcu.h | 13 ++-----------
> kernel/rcu/tree.c | 6 +++---
> 2 files changed, 5 insertions(+), 14 deletions(-)
>
> diff --git a/kernel/rcu/rcu.h b/kernel/rcu/rcu.h
> index eed2951a4962..c2ca196907cb 100644
> --- a/kernel/rcu/rcu.h
> +++ b/kernel/rcu/rcu.h
> @@ -146,19 +146,10 @@ static inline bool rcu_seq_started(unsigned long *sp, unsigned long s)
>
> /*
> * Given a snapshot from rcu_seq_snap(), determine whether or not a
> - * full update-side operation has occurred.
> + * full update-side operation has occurred while also handling
> + * wraparounds that exceed the (ULONG_MAX / 2) safety-factor/guard-band.
> */
> static inline bool rcu_seq_done(unsigned long *sp, unsigned long s)
> -{
> - return ULONG_CMP_GE(READ_ONCE(*sp), s);
> -}
> -
> -/*
> - * Given a snapshot from rcu_seq_snap(), determine whether or not a
> - * full update-side operation has occurred, but do not allow the
> - * (ULONG_MAX / 2) safety-factor/guard-band.
> - */
> -static inline bool rcu_seq_done_exact(unsigned long *sp, unsigned long s)
> {
> unsigned long cur_s = READ_ONCE(*sp);
>
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index b77ccc55557b..835600cec9ba 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -4300,7 +4300,7 @@ EXPORT_SYMBOL_GPL(start_poll_synchronize_rcu_full);
> bool poll_state_synchronize_rcu(unsigned long oldstate)
> {
> if (oldstate == RCU_GET_STATE_COMPLETED ||
> - rcu_seq_done_exact(&rcu_state.gp_seq_polled, oldstate)) {
> + rcu_seq_done(&rcu_state.gp_seq_polled, oldstate)) {
> smp_mb(); /* Ensure GP ends before subsequent accesses. */
> return true;
> }
> @@ -4347,9 +4347,9 @@ bool poll_state_synchronize_rcu_full(struct rcu_gp_oldstate *rgosp)
>
> smp_mb(); // Order against root rcu_node structure grace-period cleanup.
> if (rgosp->rgos_norm == RCU_GET_STATE_COMPLETED ||
> - rcu_seq_done_exact(&rnp->gp_seq, rgosp->rgos_norm) ||
> + rcu_seq_done(&rnp->gp_seq, rgosp->rgos_norm) ||
> rgosp->rgos_exp == RCU_GET_STATE_COMPLETED ||
> - rcu_seq_done_exact(&rcu_state.expedited_sequence, rgosp->rgos_exp)) {
> + rcu_seq_done(&rcu_state.expedited_sequence, rgosp->rgos_exp)) {
> smp_mb(); /* Ensure GP ends before subsequent accesses. */
> return true;
> }
> --
> 2.34.1
>
Powered by blists - more mailing lists