[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20070103014423.GA16872@linux.vnet.ibm.com>
Date: Tue, 2 Jan 2007 17:44:24 -0800
From: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To: mingo@...e.hu, tglx@...uxtronix.de
Cc: linux-kernel@...r.kernel.org, tytso@...ibm.com, dvhltc@...ibm.com,
dipankar@...ibm.com, oleg@...sign.ru, twoerner.k@...il.com,
josh@...edesktop.org
Subject: [RFC PATCH -rt] Almost-ready-for-prime-time RCU priority boosting
Hello!
An update to the long-standing need for priority boosting of RCU readers
in -rt kernels. This patch passes moderate testing, considerably more
severe than that faced by previous versions. Known shortcomings:
o This patch has not yet been subjected to enterprise-level
stress testing. It therefore likely still contains a few bugs.
My next step is to write some enterprise-level tests, probably
as extensions to the RCU torture tests.
o No tie-in to the OOM system. Note that the RCU priority booster,
unlike other subsystems that respond to OOM, takes action over
a timeframe. Boosting the priority of long-blocked RCU readers
does not immediately complete the grace period, so the RCU priority
booster needs to know the duration of the OOM event rather than
just being told to free up memory immediately. This likely also
means that the RCU priority booster should be given early warning
of impending OOM, so that it has the time it needs to react.
I have not worried much about this yet, since my belief is that
the current approach will get the RCU callbacks processed in
a timely manner in almost all cases. However, the tie-in to
OOM might be needed for small-memory systems.
o Although the RCU priority booster's own priority is easily adjusted
in a running kernel, it currently boosts blocked RCU readers
to a fixed priority just below that of IRQ handler threads.
This is straightforward and important, but I need to get all
the bugs shaken out before worrying about ease of use.
o A design document is needed. This is on my list!
A couple of questions:
o Currently, the rcu_boost_prio() and rcu_unboost_prio() functions
are in kernel/rcupreempt.c, because this allows them to be
declared as static. But if I ignore the desire to declare them
as static, I would instead put them into kernel/rt_mutex.c
with the other priority-inheritance code. Should I move them
to kernel/rt_mutex.c?
o It appears to me that one must always hold ->pi_lock when calling
rt_mutex_setprio(). Is this really the case? (If so, I will
add words to this effect to its comment header. And the longevity
of my test kernels seemed to increase dramatically when I added
this locking, for whatever that is worth.)
Anyway, here is the patch. Any and all comments greatly appreciated.
Signed-off-by: Paul E. McKenney <paulmck@...ux.vnet.ibm.com>
---
include/linux/init_task.h | 13
include/linux/rcupdate.h | 14
include/linux/rcupreempt.h | 30 +
include/linux/sched.h | 17
init/main.c | 1
kernel/Kconfig.preempt | 32 +
kernel/exit.c | 1
kernel/fork.c | 7
kernel/rcupreempt.c | 826 +++++++++++++++++++++++++++++++++++++++++++++
kernel/rtmutex.c | 9
kernel/sched.c | 5
11 files changed, 952 insertions(+), 3 deletions(-)
diff -urpNa -X dontdiff linux-2.6.19-rt12/include/linux/init_task.h linux-2.6.19-rt12-rcubpl/include/linux/init_task.h
--- linux-2.6.19-rt12/include/linux/init_task.h 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/include/linux/init_task.h 2006-12-24 16:20:08.000000000 -0800
@@ -86,6 +86,18 @@ extern struct nsproxy init_nsproxy;
.siglock = __SPIN_LOCK_UNLOCKED(sighand.siglock), \
}
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+#define INIT_RCU_BOOST_PRIO .rcu_prio = MAX_PRIO,
+#define INIT_PREEMPT_RCU_BOOST(tsk) \
+ .rcub_rbdp = NULL, \
+ .rcub_state = RCU_BOOST_IDLE, \
+ .rcub_entry = LIST_HEAD_INIT(tsk.rcub_entry), \
+ .rcub_rbdp_wq = NULL,
+#else /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
+#define INIT_RCU_BOOST_PRIO
+#define INIT_PREEMPT_RCU_BOOST(tsk)
+#endif /* #else #ifdef CONFIG_PREEMPT_RCU_BOOST */
+
extern struct group_info init_groups;
/*
@@ -142,6 +154,7 @@ extern struct group_info init_groups;
.pi_lock = RAW_SPIN_LOCK_UNLOCKED(tsk.pi_lock), \
INIT_TRACE_IRQFLAGS \
INIT_LOCKDEP \
+ INIT_PREEMPT_RCU_BOOST(tsk) \
}
diff -urpNa -X dontdiff linux-2.6.19-rt12/include/linux/rcupdate.h linux-2.6.19-rt12-rcubpl/include/linux/rcupdate.h
--- linux-2.6.19-rt12/include/linux/rcupdate.h 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/include/linux/rcupdate.h 2006-12-24 23:27:52.000000000 -0800
@@ -227,6 +227,20 @@ extern void rcu_barrier(void);
extern void rcu_init(void);
extern void rcu_advance_callbacks(int cpu, int user);
extern void rcu_check_callbacks(int cpu, int user);
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+extern void init_rcu_boost_late(void);
+extern void rcu_exit_wait(void);
+extern void __rcu_preempt_boost(void);
+#define rcu_preempt_boost() \
+ do { \
+ if (unlikely(current->rcu_read_lock_nesting > 0)) \
+ __rcu_preempt_boost(); \
+ } while (0)
+#else /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
+#define init_rcu_boost_late()
+#define rcu_exit_wait()
+#define rcu_preempt_boost()
+#endif /* #else #ifdef CONFIG_PREEMPT_RCU_BOOST */
#endif /* __KERNEL__ */
#endif /* __LINUX_RCUPDATE_H */
diff -urpNa -X dontdiff linux-2.6.19-rt12/include/linux/rcupreempt.h linux-2.6.19-rt12-rcubpl/include/linux/rcupreempt.h
--- linux-2.6.19-rt12/include/linux/rcupreempt.h 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/include/linux/rcupreempt.h 2006-12-24 23:46:50.000000000 -0800
@@ -42,6 +42,36 @@
#include <linux/cpumask.h>
#include <linux/seqlock.h>
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+/*
+ * Task state with respect to being RCU-boosted. This state is changed
+ * by the task itself in response to the following three events:
+ * 1. Preemption (or block on lock) while in RCU read-side critical section.
+ * 2. Outermost rcu_read_unlock() for blocked RCU read-side critical section.
+ * 3. exit() processing.
+ *
+ * The RCU-boost task also updates the state on the following events:
+ * 1. Starting to boost this task's priority.
+ * 2. Finishing boosting this task's priority.
+ * 3. Unboosting this task's priority (due to race with rcu_read_unlock()).
+ *
+ * The following values are chosen to make the update macros work.
+ */
+enum rcu_boost_state {
+ RCU_BOOST_IDLE = 0, /* Not yet blocked if in RCU read-side. */
+ RCU_BOOST_BLOCKED = 1, /* Blocked from RCU read-side. */
+ RCU_BOOSTED = 2, /* Boosting complete. */
+ RCU_EXIT_OK = 3, /* Can complete exit(). */
+ RCU_UNBOOST_IDLE = 4, /* Waiting for unboost. */
+ RCU_UNBOOST_BLOCKED = 5, /* Blocked while waiting for unlock. */
+ RCU_BOOSTING = 6, /* Boost started, not yet complete. */
+ RCU_UNBOOST_EXITING = 7 /* Waiting for unboost during exit(). */
+};
+
+#define N_RCU_BOOST_STATE (RCU_UNBOOST_EXITING + 1)
+
+#endif /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
+
#define rcu_qsctr_inc(cpu)
#define rcu_bh_qsctr_inc(cpu)
#define call_rcu_bh(head, rcu) call_rcu(head, rcu)
diff -urpNa -X dontdiff linux-2.6.19-rt12/include/linux/sched.h linux-2.6.19-rt12-rcubpl/include/linux/sched.h
--- linux-2.6.19-rt12/include/linux/sched.h 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/include/linux/sched.h 2006-12-24 23:31:23.000000000 -0800
@@ -668,6 +668,14 @@ struct signal_struct {
#define is_rt_policy(p) ((p) != SCHED_NORMAL && (p) != SCHED_BATCH)
#define has_rt_policy(p) unlikely(is_rt_policy((p)->policy))
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+#define set_rcu_prio(p, prio) ((p)->rcu_prio = (prio))
+#define get_rcu_prio(p) ((p)->rcu_prio)
+#else /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
+#define set_rcu_prio(p, prio) do { } while (0)
+#define get_rcu_prio(p) MAX_PRIO
+#endif /* #else #ifdef CONFIG_PREEMPT_RCU_BOOST */
+
/*
* Some day this will be a full-fledged user tracking system..
*/
@@ -950,6 +958,9 @@ struct task_struct {
#endif
int load_weight; /* for niceness load balancing purposes */
int prio, static_prio, normal_prio;
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+ int rcu_prio;
+#endif /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
struct list_head run_list;
struct prio_array *array;
@@ -971,6 +982,12 @@ struct task_struct {
atomic_t *rcu_flipctr1;
atomic_t *rcu_flipctr2;
#endif
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+ struct rcu_boost_dat *rcub_rbdp;
+ enum rcu_boost_state rcub_state;
+ struct list_head rcub_entry;
+ struct rcu_boost_dat *rcub_rbdp_wq;
+#endif /* #ifdef CONFIG_PREEMPT_RCU_BOOST */
#if defined(CONFIG_SCHEDSTATS) || defined(CONFIG_TASK_DELAY_ACCT)
struct sched_info sched_info;
diff -urpNa -X dontdiff linux-2.6.19-rt12/init/main.c linux-2.6.19-rt12-rcubpl/init/main.c
--- linux-2.6.19-rt12/init/main.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/init/main.c 2006-12-22 21:24:14.000000000 -0800
@@ -692,6 +692,7 @@ static void __init do_basic_setup(void)
init_workqueues();
usermodehelper_init();
driver_init();
+ init_rcu_boost_late();
#ifdef CONFIG_SYSCTL
sysctl_init();
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/exit.c linux-2.6.19-rt12-rcubpl/kernel/exit.c
--- linux-2.6.19-rt12/kernel/exit.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/exit.c 2006-12-24 17:46:57.000000000 -0800
@@ -955,6 +955,7 @@ fastcall NORET_TYPE void do_exit(long co
exit_pi_state_list(tsk);
if (unlikely(current->pi_state_cache))
kfree(current->pi_state_cache);
+ rcu_exit_wait();
/*
* Make sure we are holding no locks:
*/
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/fork.c linux-2.6.19-rt12-rcubpl/kernel/fork.c
--- linux-2.6.19-rt12/kernel/fork.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/fork.c 2006-12-24 23:44:03.000000000 -0800
@@ -1124,6 +1124,13 @@ static struct task_struct *copy_process(
p->hardirq_context = 0;
p->softirq_context = 0;
#endif
+#ifdef CONFIG_PREEMPT_RCU_BOOST
+ p->rcu_prio = MAX_PRIO;
+ p->rcub_rbdp = NULL;
+ p->rcub_state = RCU_BOOST_IDLE;
+ INIT_LIST_HEAD(&p->rcub_entry);
+ p->rcub_rbdp_wq = NULL;
+#endif
#ifdef CONFIG_LOCKDEP
p->lockdep_depth = 0; /* no locks held yet */
p->curr_chain_key = 0;
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/Kconfig.preempt linux-2.6.19-rt12-rcubpl/kernel/Kconfig.preempt
--- linux-2.6.19-rt12/kernel/Kconfig.preempt 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/Kconfig.preempt 2006-12-22 21:24:14.000000000 -0800
@@ -176,3 +176,35 @@ config RCU_TRACE
Say Y here if you want to enable RCU tracing
Say N if you are unsure.
+
+config PREEMPT_RCU_BOOST
+ bool "Enable priority boosting of RCU read-side critical sections"
+ depends on PREEMPT_RCU
+ default n
+ help
+ This option permits priority boosting of RCU read-side critical
+ sections that have been preempted in order to prevent indefinite
+ delay of grace periods in face of runaway non-realtime processes.
+
+ Say N if you are unsure.
+
+config PREEMPT_RCU_BOOST_STATS
+ bool "Enable RCU priority-boosting statistic printing"
+ depends on PREEMPT_RCU_BOOST
+ default n
+ help
+ This option enables debug printk()s of RCU boost statistics,
+ which are normally only used to debug RCU priority boost
+ implementations.
+
+ Say N if you are unsure.
+
+config PREEMPT_RCU_BOOST_STATS_INTERVAL
+ int "RCU priority-boosting statistic printing interval (seconds)"
+ depends on PREEMPT_RCU_BOOST_STATS
+ default 100
+ range 10 86400
+ help
+ This option controls the timing of debug printk()s of RCU boost
+ statistics, which are normally only used to debug RCU priority
+ boost implementations.
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/rcupreempt.c linux-2.6.19-rt12-rcubpl/kernel/rcupreempt.c
--- linux-2.6.19-rt12/kernel/rcupreempt.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/rcupreempt.c 2007-01-01 16:22:10.000000000 -0800
@@ -49,6 +49,7 @@
#include <linux/byteorder/swabb.h>
#include <linux/cpumask.h>
#include <linux/rcupreempt_trace.h>
+#include <linux/kthread.h>
/*
* PREEMPT_RCU data structures.
@@ -80,6 +81,824 @@ static struct rcu_ctrlblk rcu_ctrlblk =
static DEFINE_PER_CPU(atomic_t [2], rcu_flipctr) =
{ ATOMIC_INIT(0), ATOMIC_INIT(0) };
+#ifndef CONFIG_PREEMPT_RCU_BOOST
+static inline void init_rcu_boost_early(void) { }
+static inline void rcu_read_unlock_unboost(void) { }
+#else /* #ifndef CONFIG_PREEMPT_RCU_BOOST */
+
+/* Macros operating on enum rcu_boost_state to handle state transitions. */
+
+#define RCU_BOOST_STATE_BLOCKED(s) ((s) + 1)
+#define RCU_BOOST_STATE_RCU_READ_UNLOCK(s) ((s) & ~0x3)
+#define RCU_BOOST_STATE_EXIT(s) ((s) + 3)
+#define RCU_BOOST_STATE_UNBOOST(s) ((s) - 4)
+
+/* Defines possible event indices for ->rbs_stats[] (first index). */
+
+#define RCU_BOOST_DAT_BLOCK 0
+#define RCU_BOOST_DAT_BOOST 1
+#define RCU_BOOST_DAT_UNBOOST 2
+#define RCU_BOOST_DAT_UNLOCK 3
+#define RCU_BOOST_DAT_EXIT 4
+#define N_RCU_BOOST_DAT_EVENTS 5
+
+/* RCU-boost per-CPU array element. */
+
+struct rcu_boost_dat {
+ raw_spinlock_t rbs_mutex;
+ struct list_head rbs_toboost;
+ struct list_head rbs_boosted;
+ wait_queue_head_t rbs_target_wq;
+ wait_queue_head_t rbs_booster_wq;
+ int rbs_exit_done;
+ long rbs_blocked;
+ long rbs_boost_attempt;
+ long rbs_boost_wrongstate;
+ long rbs_boost_cmpxchgfail;
+ long rbs_boost_start;
+ long rbs_boost_end;
+ long rbs_unlock;
+ long rbs_unboosted;
+#ifdef CONFIG_PREEMPT_RCU_BOOST_STATS
+ long rbs_stats[N_RCU_BOOST_DAT_EVENTS][N_RCU_BOOST_STATE + 1];
+#endif /* #ifdef CONFIG_PREEMPT_RCU_BOOST_STATS */
+};
+#define RCU_BOOST_ELEMENTS 4
+
+int rcu_boost_idx = -1; /* invalid value in case someone uses RCU early. */
+DEFINE_PER_CPU(struct rcu_boost_dat, rcu_boost_dat[RCU_BOOST_ELEMENTS]);
+static struct task_struct *rcu_boost_task = NULL;
+
+#ifdef CONFIG_PREEMPT_RCU_BOOST_STATS
+
+/*
+ * Function to increment indicated ->rbs_stats[] element.
+ */
+static inline void rcu_boost_dat_stat(struct rcu_boost_dat *rbdp,
+ int event,
+ enum rcu_boost_state oldstate)
+{
+ if (oldstate >= RCU_BOOST_IDLE &&
+ oldstate <= RCU_UNBOOST_EXITING) {
+ rbdp->rbs_stats[event][oldstate]++;
+ } else {
+ rbdp->rbs_stats[event][N_RCU_BOOST_STATE]++;
+ }
+}
+
+#define rcu_boost_dat_stat_block(rbdp, oldstate) \
+ rcu_boost_dat_stat(rbdp, RCU_BOOST_DAT_BLOCK, oldstate)
+#define rcu_boost_dat_stat_boost(rbdp, oldstate) \
+ rcu_boost_dat_stat(rbdp, RCU_BOOST_DAT_BOOST, oldstate)
+#define rcu_boost_dat_stat_unboost(rbdp, oldstate) \
+ rcu_boost_dat_stat(rbdp, RCU_BOOST_DAT_UNBOOST, oldstate)
+#define rcu_boost_dat_stat_unlock(rbdp, oldstate) \
+ rcu_boost_dat_stat(rbdp, RCU_BOOST_DAT_UNLOCK, oldstate)
+#define rcu_boost_dat_stat_exit(rbdp, oldstate) \
+ do { \
+ if (rbdp != NULL) \
+ rcu_boost_dat_stat(rbdp, RCU_BOOST_DAT_EXIT, oldstate); \
+ } while (0)
+
+/*
+ * Prefix for kprint() strings for periodic statistics messages.
+ */
+static char *rcu_boost_state_event[] = {
+ "block:",
+ "boost: ",
+ "unboost:",
+ "unlock: ",
+ "exit: ",
+};
+
+/*
+ * Indicators for numbers in kprint() strings. "!" indicates a state-event
+ * pair that should not happen, while "?" indicates a state that should
+ * not happen.
+ */
+static char *rcu_boost_state_error[] = {
+ /*ibDEIB^e*/
+ " ! !!?", /* block */
+ "! !!!!! ?", /* boost */
+ "!!!! ?", /* unboost */
+ " ! !?", /* unlock */
+ " !!! !!!?", /* exit */
+};
+
+/*
+ * Print out RCU booster task statistics at the specified interval.
+ */
+static void rcu_boost_dat_stat_print(void)
+{
+ char buf[N_RCU_BOOST_STATE * (sizeof(long) * 3 + 2) + 2];
+ int cpu;
+ int event;
+ int i;
+ static time_t lastprint = 0;
+ struct rcu_boost_dat *rbdp;
+ int state;
+ struct rcu_boost_dat sum;
+
+ /* Wait a graceful interval between printk spamming. */
+
+ if (xtime.tv_sec - lastprint <
+ CONFIG_PREEMPT_RCU_BOOST_STATS_INTERVAL)
+ return;
+
+ /* Sum up the state/event-independent counters. */
+
+ sum.rbs_blocked = 0;
+ sum.rbs_boost_attempt = 0;
+ sum.rbs_boost_wrongstate = 0;
+ sum.rbs_boost_cmpxchgfail = 0;
+ sum.rbs_boost_start = 0;
+ sum.rbs_boost_end = 0;
+ sum.rbs_unlock = 0;
+ sum.rbs_unboosted = 0;
+ for_each_possible_cpu(cpu)
+ for (i = 0; i < RCU_BOOST_ELEMENTS; i++) {
+ rbdp = per_cpu(rcu_boost_dat, cpu);
+ sum.rbs_blocked += rbdp[i].rbs_blocked;
+ sum.rbs_boost_attempt += rbdp[i].rbs_boost_attempt;
+ sum.rbs_boost_wrongstate +=
+ rbdp[i].rbs_boost_wrongstate;
+ sum.rbs_boost_cmpxchgfail +=
+ rbdp[i].rbs_boost_cmpxchgfail;
+ sum.rbs_boost_start += rbdp[i].rbs_boost_start;
+ sum.rbs_boost_end += rbdp[i].rbs_boost_end;
+ sum.rbs_unlock += rbdp[i].rbs_unlock;
+ sum.rbs_unboosted += rbdp[i].rbs_unboosted;
+ }
+
+ /* Sum up the state/event-dependent counters. */
+
+ for (event = 0; event < N_RCU_BOOST_DAT_EVENTS; event++)
+ for (state = 0; state < N_RCU_BOOST_STATE; state++) {
+ sum.rbs_stats[event][state] = 0;
+ for_each_possible_cpu(cpu) {
+ for (i = 0; i < RCU_BOOST_ELEMENTS; i++) {
+ sum.rbs_stats[event][state]
+ += per_cpu(rcu_boost_dat,
+ cpu)[i].rbs_stats[event][state];
+ }
+ }
+ }
+
+ /* Print them out! */
+
+ printk(KERN_ALERT
+ "rcu_boost_dat: idx=%d "
+ "b=%ld ul=%ld ub=%ld "
+ "boost: a=%ld ws=%ld cf=%ld s=%ld e=%ld\n",
+ rcu_boost_idx,
+ sum.rbs_blocked, sum.rbs_unlock, sum.rbs_unboosted,
+ sum.rbs_boost_attempt, sum.rbs_boost_wrongstate,
+ sum.rbs_boost_cmpxchgfail, sum.rbs_boost_start,
+ sum.rbs_boost_end);
+ for (event = 0; event < N_RCU_BOOST_DAT_EVENTS; event++) {
+ i = 0;
+ for (state = 0; state <= N_RCU_BOOST_STATE; state++) {
+ i += sprintf(&buf[i], " %ld%c",
+ sum.rbs_stats[event][state],
+ rcu_boost_state_error[event][state]);
+ }
+ printk(KERN_ALERT "rcu_boost_dat %s %s\n",
+ rcu_boost_state_event[event], buf);
+ }
+
+ /* Go away and don't come back for awhile. */
+
+ lastprint = xtime.tv_sec;
+}
+
+#else /* #ifdef CONFIG_PREEMPT_RCU_BOOST_STATS */
+
+#define rcu_boost_dat_stat_block(rbdp, oldstate)
+#define rcu_boost_dat_stat_boost(rbdp, oldstate)
+#define rcu_boost_dat_stat_unboost(rbdp, oldstate)
+#define rcu_boost_dat_stat_unlock(rbdp, oldstate)
+#define rcu_boost_dat_stat_exit(rbdp, oldstate)
+#define rcu_boost_dat_stat_print()
+
+#endif /* #else #ifdef CONFIG_PREEMPT_RCU_BOOST_STATS */
+
+/*
+ * Initialize RCU-boost state. This happens early in the boot process,
+ * when the scheduler does not yet exist. So don't try to use it.
+ */
+static void init_rcu_boost_early(void)
+{
+ struct rcu_boost_dat *rbdp;
+ int cpu;
+ int i;
+
+ for_each_possible_cpu(cpu) {
+ rbdp = per_cpu(rcu_boost_dat, cpu);
+ for (i = 0; i < RCU_BOOST_ELEMENTS; i++) {
+ rbdp[i].rbs_mutex =
+ RAW_SPIN_LOCK_UNLOCKED(rbdp[i].rbs_mutex);
+ INIT_LIST_HEAD(&rbdp[i].rbs_toboost);
+ INIT_LIST_HEAD(&rbdp[i].rbs_boosted);
+ init_waitqueue_head(&rbdp[i].rbs_target_wq);
+ init_waitqueue_head(&rbdp[i].rbs_booster_wq);
+ rbdp[i].rbs_exit_done = 1;
+ rbdp[i].rbs_blocked = 0;
+ rbdp[i].rbs_boost_attempt = 0;
+ rbdp[i].rbs_boost_wrongstate = 0;
+ rbdp[i].rbs_boost_cmpxchgfail = 0;
+ rbdp[i].rbs_boost_start = 0;
+ rbdp[i].rbs_boost_end = 0;
+ rbdp[i].rbs_unlock = 0;
+ rbdp[i].rbs_unboosted = 0;
+#ifdef CONFIG_PREEMPT_RCU_BOOST_STATS
+ {
+ int j, k;
+
+ for (j = 0; j < N_RCU_BOOST_DAT_EVENTS; j++)
+ for (k = 0; k <= N_RCU_BOOST_STATE; k++)
+ rbdp[i].rbs_stats[j][k] = 0;
+ }
+#endif /* #ifdef CONFIG_PREEMPT_RCU_BOOST_STATS */
+ }
+ smp_wmb();
+ rcu_boost_idx = 0;
+ }
+}
+
+/*
+ * Return the current boost index for adding target tasks.
+ * Will be -1 if too early during boot.
+ */
+static inline int rcu_boost_idx_new(void)
+{
+ int idx = rcu_boost_idx;
+
+ smp_read_barrier_depends(); barrier();
+ return idx;
+}
+
+/*
+ * Return the current boost index for boosting target tasks.
+ * May only be invoked by the booster task, so guaranteed to
+ * already be initialized.
+ */
+static inline int rcu_boost_idx_boosting(void)
+{
+ return (rcu_boost_idx + 1) & (RCU_BOOST_ELEMENTS - 1);
+}
+
+/*
+ * Return the list on which the calling task should add itself, or
+ * NULL if too early during initialization.
+ */
+static inline struct rcu_boost_dat *rcu_rbd_new(void)
+{
+ int cpu = raw_smp_processor_id(); /* locks used, so preemption OK. */
+ int idx = rcu_boost_idx_new();
+
+ if (unlikely(idx < 0))
+ return (NULL);
+ return &per_cpu(rcu_boost_dat, cpu)[idx];
+}
+
+/*
+ * Return the list from which to boost target tasks.
+ * May only be invoked by the booster task, so guaranteed to
+ * already be initialized.
+ */
+static inline struct rcu_boost_dat *rcu_rbd_boosting(int cpu)
+{
+ return &per_cpu(rcu_boost_dat, cpu)[rcu_boost_idx_boosting()];
+}
+
+#define PREEMPT_RCU_BOOSTER_PRIO 49 /* Match curr_irq_prio manually for now. */
+#define PREEMPT_RCU_BOOST_PRIO 50 /* Don't allow RCU read-side critical */
+ /* sections to block irq handlers. */
+
+/*
+ * Boost the specified task from an RCU viewpoint.
+ * These two functions might be better in kernel/rtmutex.c?
+ */
+static void rcu_boost_prio(struct task_struct *taskp)
+{
+ unsigned long oldirq;
+
+ spin_lock_irqsave(&taskp->pi_lock, oldirq);
+ taskp->rcu_prio = PREEMPT_RCU_BOOST_PRIO;
+ if (taskp->rcu_prio < taskp->prio)
+ rt_mutex_setprio(taskp, taskp->rcu_prio);
+ spin_unlock_irqrestore(&taskp->pi_lock, oldirq);
+}
+
+/*
+ * Unboost the specified task from an RCU viewpoint.
+ */
+static void rcu_unboost_prio(struct task_struct *taskp)
+{
+ int nprio;
+ unsigned long oldirq;
+
+ spin_lock_irqsave(&taskp->pi_lock, oldirq);
+ taskp->rcu_prio = MAX_PRIO;
+ nprio = rt_mutex_getprio(taskp);
+ if (nprio > taskp->prio)
+ rt_mutex_setprio(taskp, nprio);
+ spin_unlock_irqrestore(&taskp->pi_lock, oldirq);
+}
+
+/*
+ * Boost all of the RCU-reader tasks on the specified list.
+ */
+static void rcu_boost_one_reader_list(struct rcu_boost_dat *rbdp)
+{
+ LIST_HEAD(list);
+ unsigned long oldirq;
+ enum rcu_boost_state oldstate;
+ enum rcu_boost_state newstate;
+ struct task_struct *taskp;
+
+ /*
+ * Splice the toboost list onto a local list. We will still
+ * need to hold the lock when manipulating the local list
+ * because tasks can remove themselves at any time.
+ */
+
+ spin_lock_irqsave(&rbdp->rbs_mutex, oldirq);
+ list_splice_init(&rbdp->rbs_toboost, &list);
+ while (!list_empty(&list)) {
+
+ /*
+ * Pause for a bit before boosting each task.
+ * @@@FIXME: reduce/eliminate pausing in case of OOM.
+ */
+
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+ schedule_timeout_uninterruptible(1);
+ spin_lock_irqsave(&rbdp->rbs_mutex, oldirq);
+
+ /* Interlock with prior target task if raced with exit(). */
+
+ if (rbdp->rbs_exit_done != 1) {
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+ wait_event(rbdp->rbs_booster_wq,
+ rbdp->rbs_exit_done == 1);
+ spin_lock_irqsave(&rbdp->rbs_mutex, oldirq);
+ }
+
+ /*
+ * All tasks might have removed themselves while
+ * we were waiting. Recheck list emptiness.
+ */
+
+ if (list_empty(&list))
+ break;
+
+ /* Remove first task in local list, count the attempt. */
+
+ taskp = list_entry(list.next, typeof(*taskp), rcub_entry);
+ list_del_init(&taskp->rcub_entry);
+ rbdp->rbs_boost_attempt++;
+
+ /* Ignore tasks in unexpected states. */
+
+ if ((oldstate = taskp->rcub_state) != RCU_BOOST_BLOCKED) {
+ list_add_tail(&taskp->rcub_entry, &rbdp->rbs_toboost);
+ rcu_boost_dat_stat_boost(rbdp, oldstate);
+ continue;
+ }
+ taskp->rcub_rbdp_wq = rbdp;
+
+ /*
+ * This cmpxchg should always succeed, since we hold the
+ * lock. Count and ignore the task if it nonetheless fails.
+ */
+
+ if (cmpxchg(&taskp->rcub_state,
+ RCU_BOOST_BLOCKED,
+ RCU_BOOSTING) != RCU_BOOST_BLOCKED) {
+ list_add_tail(&taskp->rcub_entry, &rbdp->rbs_toboost);
+ rcu_boost_dat_stat_boost(rbdp, RCU_BOOST_BLOCKED);
+ taskp->rcub_rbdp_wq = NULL;
+ rbdp->rbs_boost_cmpxchgfail++;
+ continue;
+ }
+ rcu_boost_dat_stat_boost(rbdp, RCU_BOOST_BLOCKED);
+
+ /*
+ * Count the task, add to boosted list, and set up
+ * for the potential exit() race.
+ */
+
+ rbdp->rbs_boost_start++;
+ list_add_tail(&taskp->rcub_entry, &rbdp->rbs_boosted);
+ rbdp->rbs_exit_done = 0;
+
+ /* Boost task's priority. */
+
+ rcu_boost_prio(taskp);
+
+ /* Update state to indicate that boost is complete. */
+
+ newstate = RCU_BOOSTED;
+ if (cmpxchg(&taskp->rcub_state,
+ RCU_BOOSTING, RCU_BOOSTED) == RCU_BOOSTING) {
+ rbdp->rbs_boost_end++;
+ rcu_boost_dat_stat_unboost(rbdp, RCU_BOOSTING);
+ } else {
+
+ /*
+ * The task changed state before we could boost
+ * it. We must therefore unboost it. Note that
+ * the task may well be on some other list, so
+ * we cannot reasonably leverage locking.
+ */
+
+ rcu_unboost_prio(taskp);
+
+ /*
+ * Now transition the task's state to allow for
+ * the unboosting. In theory, we could go through
+ * this loop many times. In practice, this would
+ * require that the target task be blocking and
+ * unblocking ridiculously often.
+ *
+ * If latency becomes a problem, it should be
+ * OK to drop the lock and re-enable irqs across
+ * this loop.
+ */
+
+ do {
+ oldstate = taskp->rcub_state;
+ if (oldstate < RCU_UNBOOST_IDLE)
+ break; /* logged below. */
+ newstate = RCU_BOOST_STATE_UNBOOST(oldstate);
+ } while (cmpxchg(&taskp->rcub_state,
+ oldstate, newstate) != oldstate);
+ rcu_boost_dat_stat_unboost(rbdp, oldstate);
+ }
+
+ /* Do exit dance if needed. */
+
+ if (newstate == RCU_EXIT_OK) {
+ wake_up(&rbdp->rbs_target_wq); /* drop lock??? @@@ */
+ } else {
+ /* At this point, we don't own the task boost state. */
+ rbdp->rbs_exit_done = 1;
+ }
+ }
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+}
+
+/*
+ * Priority-boost tasks stuck in RCU read-side critical sections as
+ * needed (presumably rarely).
+ */
+static int rcu_booster(void *arg)
+{
+ int cpu;
+ struct sched_param sp;
+
+ sp.sched_priority = PREEMPT_RCU_BOOSTER_PRIO;
+ sched_setscheduler(current, SCHED_RR, &sp);
+ current->flags |= PF_NOFREEZE;
+
+ do {
+
+ /* Advance the lists of tasks. */
+
+ rcu_boost_idx = (rcu_boost_idx + 1) % RCU_BOOST_ELEMENTS;
+ for_each_possible_cpu(cpu) {
+
+ /*
+ * Boost all sufficiently aged readers.
+ * Readers must first be preempted or block
+ * on a mutex in an RCU read-side critical section,
+ * then remain in that critical section for
+ * RCU_BOOST_ELEMENTS-1 time intervals.
+ * So most of the time we should end up doing
+ * nothing.
+ */
+
+ rcu_boost_one_reader_list(rcu_rbd_boosting(cpu));
+
+ /*
+ * Large SMP systems may need to sleep sometimes
+ * in this loop. Or have multiple RCU-boost tasks.
+ */
+ }
+
+ /*
+ * Sleep to allow any unstalled RCU read-side critical
+ * sections to age out of the list. @@@ FIXME: reduce,
+ * adjust, or eliminate in case of OOM.
+ */
+
+ schedule_timeout_uninterruptible(HZ / 100);
+
+ /* Print stats if enough time has passed. */
+
+ rcu_boost_dat_stat_print();
+
+ } while (!kthread_should_stop());
+
+ return 0;
+}
+
+/*
+ * Perform the portions of RCU-boost initialization that require the
+ * scheduler to be up and running.
+ */
+void init_rcu_boost_late(void)
+{
+ int i;
+
+ /* Spawn RCU-boost task. */
+
+ printk(KERN_ALERT "Starting RCU priority booster\n");
+ rcu_boost_task = kthread_run(rcu_booster, NULL, "RCU Prio Booster");
+ if (IS_ERR(rcu_boost_task)) {
+ i = PTR_ERR(rcu_boost_task);
+ printk(KERN_ALERT
+ "Unable to create RCU Priority Booster, errno %d\n", -i);
+
+ /*
+ * Continue running, but tasks permanently blocked
+ * in RCU read-side critical sections will be able
+ * to stall grace-period processing, potentially
+ * OOMing the machine.
+ */
+
+ rcu_boost_task = NULL;
+ }
+}
+
+/*
+ * Update task's RCU-boost state to reflect blocking in RCU read-side
+ * critical section, so that the RCU-boost task can find it in case it
+ * later needs its priority boosted.
+ */
+void __rcu_preempt_boost(void)
+{
+ struct rcu_boost_dat *rbdp;
+ unsigned long oldirq;
+ enum rcu_boost_state oldstate;
+ enum rcu_boost_state newstate;
+
+ /* Identify list to place task on for possible later boosting. */
+
+ local_irq_save(oldirq);
+ rbdp = rcu_rbd_new();
+ if (rbdp == NULL) {
+ local_irq_restore(oldirq);
+ printk("Preempted RCU read-side critical section too early.\n");
+ return;
+ }
+ spin_lock(&rbdp->rbs_mutex);
+ rbdp->rbs_blocked++;
+
+ /* Update state. There can be at most two passes through the
+ * following loop: (1) cmpxchg fails due to booster concurrently
+ * changing RCU_UNBOOST_IDLE to RCU_BOOST_IDLE and (2) cmpxchg
+ * succeeds changing RCU_BOOST_IDLE to RCU_BOOST_BLOCKED.
+ */
+
+ do {
+ oldstate = current->rcub_state;
+ switch (oldstate) {
+ case RCU_BOOST_IDLE:
+ case RCU_UNBOOST_IDLE:
+ newstate = RCU_BOOST_STATE_BLOCKED(oldstate);
+ break;
+ default: /* Error. Count, but don't touch state/lists. */
+ case RCU_BOOST_BLOCKED:
+ case RCU_UNBOOST_BLOCKED:
+ /* Been here in same RCU read-side critical section. */
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+ rcu_boost_dat_stat_block(rbdp, oldstate);
+ return;
+ }
+ } while (cmpxchg(¤t->rcub_state, oldstate, newstate) != oldstate);
+ rcu_boost_dat_stat_block(rbdp, oldstate);
+
+ /* Now add ourselves to the list so that the booster can find use. */
+
+ list_add_tail(¤t->rcub_entry, &rbdp->rbs_toboost);
+ current->rcub_rbdp = rbdp;
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+}
+
+/*
+ * Do the list-removal and priority-unboosting "heavy lifting" when
+ * required.
+ */
+static void __rcu_read_unlock_unboost(void)
+{
+ unsigned long oldirq;
+ enum rcu_boost_state oldstate;
+ enum rcu_boost_state newstate;
+ struct rcu_boost_dat *rbdp;
+
+ /*
+ * Acquire the lock -- this prevents some, but not all, races
+ * with the RCU-boost task. (We might have entered another
+ * RCU read-side critical section on some other CPU, thus be
+ * using a different lock than the RCU-boost task, which might
+ * well still be trying to clean up after boosting one of our
+ * earlier RCU read-side critical sections.)
+ */
+
+ rbdp = current->rcub_rbdp;
+ spin_lock_irqsave(&rbdp->rbs_mutex, oldirq);
+
+ /* Remove task from the list it was on. */
+
+ list_del_init(¤t->rcub_entry);
+ rbdp->rbs_unlock++;
+ current->rcub_rbdp = NULL;
+
+ /*
+ * Update state. There can be at most two passes through the
+ * following loop, via two different scenarios:
+ *
+ * (1) cmpxchg fails due to race with booster changing
+ * RCU_BOOSTING to RCU_BOOSTED.
+ * (2) cmpxchg succeeds changing RCU_BOOSTED to RCU_BOOST_IDLE.
+ * The RCU-boost task is not permitted to change the
+ * state of a task in RCU_BOOST_IDLE.
+ *
+ * (1) cmpxchg fails due to race with booster changing
+ * RCU_UNBOOST_BLOCKED to RCU_BOOST_BLOCKED,
+ * (2) cmpxchg succeeds changing RCU_BOOST_BLOCKED to
+ * RCU_BOOST_IDLE. Although the RCU-boost task is
+ * permitted to change the state while in RCU_BOOST_IDLE,
+ * it holds the lock when doing so, so cannot until we
+ * release the lock.
+ */
+
+ do {
+ oldstate = current->rcub_state;
+ switch (oldstate) {
+ case RCU_BOOST_BLOCKED:
+ case RCU_BOOSTED:
+ case RCU_UNBOOST_BLOCKED:
+ case RCU_BOOSTING:
+ newstate = RCU_BOOST_STATE_RCU_READ_UNLOCK(oldstate);
+ break;
+ case RCU_BOOST_IDLE: /* Do-nothing case. */
+ case RCU_UNBOOST_IDLE: /* Do-nothing case. */
+ case RCU_EXIT_OK: /* Error case: still do nothing. */
+ case RCU_UNBOOST_EXITING: /* Error case: still do nothing. */
+ default: /* Error case: still do nothing. */
+ rcu_boost_dat_stat_unlock(rbdp, oldstate);
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+ return;
+ }
+ } while (cmpxchg(¤t->rcub_state, oldstate, newstate) != oldstate);
+
+ rcu_boost_dat_stat_unlock(rbdp, oldstate);
+ spin_unlock_irqrestore(&rbdp->rbs_mutex, oldirq);
+
+ /*
+ * Now unboost ourselves, if warranted and safe. Note that if
+ * the booster is still trying to boost us, then he must do any
+ * unboosting that might be necessary. The state we set above
+ * will prompt him to do so.
+ */
+
+ if (newstate == RCU_BOOST_IDLE) {
+
+ /* The RCU-boost task is done with use, disassociate. */
+
+ current->rcub_rbdp_wq = NULL;
+
+ /* Unboost if we were in fact ever boosted. */
+
+ if (unlikely(current->rcu_prio != MAX_PRIO)) {
+ rcu_unboost_prio(current);
+ rbdp->rbs_unboosted++;
+ }
+ }
+}
+
+/*
+ * Do any state changes and unboosting needed for rcu_read_unlock().
+ * Pass any complex work on to __rcu_read_unlock_unboost().
+ * The vast majority of the time, no work will be needed, as preemption
+ * and blocking within RCU read-side critical sections is comparatively
+ * rare.
+ */
+static inline void rcu_read_unlock_unboost(void)
+{
+
+ if (unlikely(current->rcub_state != RCU_BOOST_IDLE))
+ __rcu_read_unlock_unboost();
+}
+
+/*
+ * Wait, if needed, for the RCU-booster task to finish manipulating this
+ * task's priority.
+ */
+void rcu_exit_wait(void)
+{
+ enum rcu_boost_state oldstate;
+ enum rcu_boost_state newstate;
+
+ if (current->rcub_state == RCU_BOOST_IDLE)
+ return;
+
+/*&&&&*/printk(KERN_ALERT "rcu_boost_exit: pid=%d rcub_state=%d rcub_rbdp_wq=%p\n",
+/*&&&&*/ current->pid, current->rcub_state, current->rcub_rbdp_wq);
+
+ oldstate = current->rcub_state;
+ switch (oldstate) {
+ case RCU_BOOST_BLOCKED:
+ case RCU_BOOSTING:
+ case RCU_BOOSTED:
+ case RCU_UNBOOST_BLOCKED:
+ default:
+
+ /*
+ * Either state-machine or usage error. Unwind out of
+ * any remaining RCU read-side critical sections, and
+ * see if that gets us somewhere useful.
+ */
+
+ rcu_boost_dat_stat_exit(current->rcub_rbdp_wq, oldstate);
+ if (current->rcu_read_lock_nesting == 0)
+ printk(KERN_ALERT "rcu_exit_wait pid %d: bad state\n",
+ current->pid);
+ else
+ printk(KERN_ALERT "rcu_exit_wait pid %d exiting "
+ "with rcu_read_lock() held\n",
+ current->pid);
+ while (current->rcu_read_lock_nesting > 0)
+ rcu_read_unlock();
+ if ((current->rcub_state != RCU_BOOST_IDLE) &&
+ (current->rcub_state != RCU_UNBOOST_IDLE)) {
+ struct sched_param sp;
+
+ /*
+ * No joy. Stall at low priority: a memory leak
+ * is better than strange corruption.
+ */
+
+ printk(KERN_ALERT "rcu_exit_wait() pid %d waiting "
+ "forever due to invalid state\n",
+ current->pid);
+ sp.sched_priority = MAX_PRIO;
+ sched_setscheduler(current, SCHED_NORMAL, &sp);
+ current->flags |= PF_NOFREEZE;
+ for (;;)
+ schedule_timeout_interruptible(3600 * HZ);
+ }
+
+ /* Fall into RCU_BOOST_IDLE and RCU_UNBOOST_IDLE cases. */
+
+ case RCU_BOOST_IDLE:
+ case RCU_UNBOOST_IDLE:
+
+ /*
+ * Update state. There can be at most two passes through
+ * the following loop: (1) cmpxchg fails due to booster
+ * changing RCU_UNBOOST_IDLE to RCU_BOOST_IDLE, and then
+ * (2) cmpxchg succeeds changing RCU_BOOST_IDLE to
+ * RCU_EXIT_OK.
+ */
+
+ do {
+ oldstate = current->rcub_state;
+ newstate = RCU_BOOST_STATE_EXIT(oldstate);
+ } while (cmpxchg(¤t->rcub_state,
+ oldstate, newstate) != oldstate);
+ rcu_boost_dat_stat_exit(current->rcub_rbdp_wq, oldstate);
+ break;
+
+ case RCU_EXIT_OK:
+ case RCU_UNBOOST_EXITING:
+
+ /* This should not happen, but... */
+
+ rcu_boost_dat_stat_exit(current->rcub_rbdp_wq, oldstate);
+ newstate = oldstate;
+ break;
+ }
+
+ /* Wait for the booster to get done with us, if needed. */
+
+ if (newstate == RCU_UNBOOST_EXITING) {
+ wait_event(current->rcub_rbdp_wq->rbs_target_wq,
+ current->rcub_state == RCU_EXIT_OK);
+
+ /* Tell the booster that it is OK to reuse the waitqueue. */
+
+ current->rcub_rbdp_wq->rbs_exit_done = 1;
+ wake_up(¤t->rcub_rbdp_wq->rbs_booster_wq);
+ }
+ current->rcub_rbdp_wq = NULL;
+}
+
+#endif /* #else #ifndef CONFIG_PREEMPT_RCU_BOOST */
+
/*
* Return the number of RCU batches processed thus far. Useful
* for debug and statistics.
@@ -155,6 +974,7 @@ void __rcu_read_unlock(void)
atomic_dec(current->rcu_flipctr2);
current->rcu_flipctr2 = NULL;
}
+ rcu_read_unlock_unboost();
}
local_irq_restore(oldirq);
@@ -345,6 +1165,11 @@ int notrace rcu_pending(int cpu)
rcu_data.nextlist != NULL);
}
+/*
+ * Initialize RCU. This is called very early in boot, so is restricted
+ * to very simple operations. Don't even think about messing with anything
+ * that involves the scheduler, as it doesn't exist yet.
+ */
void __init __rcu_init(void)
{
/*&&&&*/printk("WARNING: experimental RCU implementation.\n");
@@ -356,6 +1181,7 @@ void __init __rcu_init(void)
rcu_data.waittail = &rcu_data.waitlist;
rcu_data.donelist = NULL;
rcu_data.donetail = &rcu_data.donelist;
+ init_rcu_boost_early();
tasklet_init(&rcu_data.rcu_tasklet, rcu_process_callbacks, 0UL);
}
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/rtmutex.c linux-2.6.19-rt12-rcubpl/kernel/rtmutex.c
--- linux-2.6.19-rt12/kernel/rtmutex.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/rtmutex.c 2006-12-24 16:25:44.000000000 -0800
@@ -128,11 +128,14 @@ static inline void init_lists(struct rt_
*/
int rt_mutex_getprio(struct task_struct *task)
{
+ int prio = task->normal_prio;
+
+ if (get_rcu_prio(task) < prio)
+ prio = get_rcu_prio(task);
if (likely(!task_has_pi_waiters(task)))
- return task->normal_prio;
+ return prio;
- return min(task_top_pi_waiter(task)->pi_list_entry.prio,
- task->normal_prio);
+ return min(task_top_pi_waiter(task)->pi_list_entry.prio, prio);
}
/*
diff -urpNa -X dontdiff linux-2.6.19-rt12/kernel/sched.c linux-2.6.19-rt12-rcubpl/kernel/sched.c
--- linux-2.6.19-rt12/kernel/sched.c 2006-12-22 21:21:42.000000000 -0800
+++ linux-2.6.19-rt12-rcubpl/kernel/sched.c 2006-12-24 16:28:13.000000000 -0800
@@ -1949,6 +1949,7 @@ void fastcall sched_fork(struct task_str
* Make sure we do not leak PI boosting priority to the child:
*/
p->prio = current->normal_prio;
+ set_rcu_prio(p, MAX_PRIO);
INIT_LIST_HEAD(&p->run_list);
p->array = NULL;
@@ -2031,6 +2032,7 @@ void fastcall wake_up_new_task(struct ta
else {
p->prio = current->prio;
p->normal_prio = current->normal_prio;
+ set_rcu_prio(p, MAX_PRIO);
__activate_task_after(p, current, rq);
}
set_need_resched();
@@ -3800,6 +3802,8 @@ void __sched __schedule(void)
}
profile_hit(SCHED_PROFILING, __builtin_return_address(0));
+ rcu_preempt_boost();
+
preempt_disable(); // FIXME: disable irqs here
prev = current;
release_kernel_lock(prev);
@@ -5512,6 +5516,7 @@ void __cpuinit init_idle(struct task_str
idle->sleep_avg = 0;
idle->array = NULL;
idle->prio = idle->normal_prio = MAX_PRIO;
+ set_rcu_prio(idle, MAX_PRIO);
idle->state = TASK_RUNNING;
idle->cpus_allowed = cpumask_of_cpu(cpu);
set_task_cpu(idle, cpu);
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists