[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20210714085507.GC2725@worktop.programming.kicks-ass.net>
Date: Wed, 14 Jul 2021 10:55:07 +0200
From: Peter Zijlstra <peterz@...radead.org>
To: Thomas Gleixner <tglx@...utronix.de>
Cc: LKML <linux-kernel@...r.kernel.org>,
Ingo Molnar <mingo@...nel.org>,
Juri Lelli <juri.lelli@...hat.com>,
Steven Rostedt <rostedt@...dmis.org>,
Daniel Bristot de Oliveira <bristot@...hat.com>,
Will Deacon <will@...nel.org>,
Waiman Long <longman@...hat.com>,
Boqun Feng <boqun.feng@...il.com>,
Sebastian Andrzej Siewior <bigeasy@...utronix.de>,
Davidlohr Bueso <dave@...olabs.net>
Subject: Re: [patch 17/50] locking/rtmutex: Prepare RT rt_mutex_wake_q for RT
locks
On Tue, Jul 13, 2021 at 05:11:11PM +0200, Thomas Gleixner wrote:
> From: Thomas Gleixner <tglx@...utronix.de>
>
> Add a rtlock_task pointer to rt_mutex_wake_q which allows to handle the RT
> specific wakeup for spin/rwlock waiters. The pointer is just consuming 4/8
> bytes on stack so it is provided unconditionaly to avoid #ifdeffery all
> over the place.
>
> No functional change for non-RT enabled kernels.
>
> Signed-off-by: Thomas Gleixner <tglx@...utronix.de>
> ---
> kernel/locking/rtmutex.c | 16 ++++++++++++++--
> kernel/locking/rtmutex_common.h | 3 +++
> 2 files changed, 17 insertions(+), 2 deletions(-)
> ---
> --- a/kernel/locking/rtmutex.c
> +++ b/kernel/locking/rtmutex.c
> @@ -351,12 +351,24 @@ static __always_inline void rt_mutex_adj
> static __always_inline void rt_mutex_wake_q_add(struct rt_mutex_wake_q_head *wqh,
> struct rt_mutex_waiter *w)
> {
> - wake_q_add(&wqh->head, w->task);
> + if (IS_ENABLED(CONFIG_PREEMPT_RT) && w->wake_state != TASK_NORMAL) {
> + get_task_struct(w->task);
> + wqh->rtlock_task = w->task;
> + } else {
> + wake_q_add(&wqh->head, w->task);
> + }
> }
>
> static __always_inline void rt_mutex_wake_up_q(struct rt_mutex_wake_q_head *wqh)
> {
> - wake_up_q(&wqh->head);
> + if (IS_ENABLED(CONFIG_PREEMPT_RT) && wqh->rtlock_task) {
> + wake_up_state(wqh->rtlock_task, TASK_RTLOCK_WAIT);
> + put_task_struct(wqh->rtlock_task);
> + wqh->rtlock_task = NULL;
> + }
> +
> + if (!wake_q_empty(&wqh->head))
> + wake_up_q(&wqh->head);
>
> /* Pairs with preempt_disable() in mark_wakeup_next_waiter() */
> preempt_enable();
> --- a/kernel/locking/rtmutex_common.h
> +++ b/kernel/locking/rtmutex_common.h
> @@ -43,14 +43,17 @@ struct rt_mutex_waiter {
> * rt_mutex_wake_q_head - Wrapper around regular wake_q_head to support
> * "sleeping" spinlocks on RT
> * @head: The regular wake_q_head for sleeping lock variants
> + * @rtlock_task: Task pointer for RT lock (spin/rwlock) wakeups
> */
> struct rt_mutex_wake_q_head {
> struct wake_q_head head;
> + struct task_struct *rtlock_task;
> };
>
> #define DEFINE_RT_MUTEX_WAKE_Q_HEAD(name) \
> struct rt_mutex_wake_q_head name = { \
> .head = WAKE_Q_HEAD_INITIALIZER(name.head), \
> + .rtlock_task = NULL, \
> }
This is a bit asymmetric, something like the below perhaps?
---
--- a/include/linux/sched/wake_q.h
+++ b/include/linux/sched/wake_q.h
@@ -61,6 +61,11 @@ static inline bool wake_q_empty(struct w
extern void wake_q_add(struct wake_q_head *head, struct task_struct *task);
extern void wake_q_add_safe(struct wake_q_head *head, struct task_struct *task);
-extern void wake_up_q(struct wake_q_head *head);
+extern void __wake_up_q(struct wake_q_head *head, unsigned int state);
+
+static inline void wake_up_q(struct wake_q_head *head)
+{
+ __wake_up_q(head, TASK_NORMAL);
+}
#endif /* _LINUX_SCHED_WAKE_Q_H */
--- a/kernel/locking/rtmutex.c
+++ b/kernel/locking/rtmutex.c
@@ -351,12 +351,20 @@ static __always_inline void rt_mutex_adj
static __always_inline void rt_mutex_wake_q_add(struct rt_mutex_wake_q_head *wqh,
struct rt_mutex_waiter *w)
{
- wake_q_add(&wqh->head, w->task);
+ if (IS_ENABLED(CONFIG_PREEMPT_RT) && w->wake_state != TASK_NORMAL) {
+ wake_q_add(&wqh->rt_head, w->task);
+ } else {
+ wake_q_add(&wqh->head, w->task);
+ }
}
static __always_inline void rt_mutex_wake_up_q(struct rt_mutex_wake_q_head *wqh)
{
- wake_up_q(&wqh->head);
+ if (IS_ENABLED(CONFIG_PREEMPT_RT) && !wake_q_empty(&wqh->rt_head))
+ __wake_up_q(&wqh->rt_head, TASK_RTLOCK_WAIT);
+
+ if (!wake_q_empty(&wqh->head))
+ wake_up_q(&wqh->head);
/* Pairs with preempt_disable() in mark_wakeup_next_waiter() */
preempt_enable();
--- a/kernel/locking/rtmutex_common.h
+++ b/kernel/locking/rtmutex_common.h
@@ -43,14 +43,17 @@ struct rt_mutex_waiter {
* rt_mutex_wake_q_head - Wrapper around regular wake_q_head to support
* "sleeping" spinlocks on RT
* @head: The regular wake_q_head for sleeping lock variants
+ * @rt_head: The wake_q_head for RT lock (spin/rwlock) variants
*/
struct rt_mutex_wake_q_head {
struct wake_q_head head;
+ struct wake_q_head rt_head;
};
#define DEFINE_RT_MUTEX_WAKE_Q_HEAD(name) \
struct rt_mutex_wake_q_head name = { \
.head = WAKE_Q_HEAD_INITIALIZER(name.head), \
+ .rt_head = WAKE_Q_HEAD_INITIALIZER(name.rt_head),\
}
/*
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -916,7 +916,7 @@ void wake_q_add_safe(struct wake_q_head
put_task_struct(task);
}
-void wake_up_q(struct wake_q_head *head)
+void __wake_up_q(struct wake_q_head *head, unsigned int state)
{
struct wake_q_node *node = head->first;
@@ -932,7 +932,7 @@ void wake_up_q(struct wake_q_head *head)
* wake_up_process() executes a full barrier, which pairs with
* the queueing in wake_q_add() so as not to miss wakeups.
*/
- wake_up_process(task);
+ wake_up_state(task, state);
put_task_struct(task);
}
}
Powered by blists - more mailing lists