[<prev] [next>] [day] [month] [year] [list]
Message-ID: <1293205676.3368.80.camel@hermosa.morreale.net>
Date: Fri, 24 Dec 2010 08:47:56 -0700
From: "Peter W. Morreale" <pmorreale@...ell.com>
To: Gregory Haskins <ghaskins@...ell.com>
Cc: Steven Rostedt <rostedt@...dmis.org>, linux-kernel@...r.kernel.org,
Lai Jiangshan <laijs@...fujitsu.com>,
Ingo Molnar <mingo@...e.hu>,
Peter Zijlstra <peterz@...radead.org>,
Thomas Gleixner <tglx@...utronix.de>
Subject: Re: [RFC][RT][PATCH 3/4] rtmutex: Revert Optimize rt lock wakeup
On Thu, 2010-12-23 at 21:45 -0700, Gregory Haskins wrote:
> Hey Steve,
>
> >>> On 12/23/2010 at 05:47 PM, in message <20101223225116.729981172@...dmis.org>,
> Steven Rostedt <rostedt@...dmis.org> wrote:
> > From: Steven Rostedt <srostedt@...hat.com>
> >
> > The commit: rtmutex: Optimize rt lock wakeup
> >
> > Does not do what it was suppose to do.
> > This is because the adaptive waiter sets its state to TASK_(UN)INTERRUPTIBLE
> > before going into the loop. Thus, the test in wakeup_next_waiter()
> > will always fail on an adaptive waiter, as it only tests to see if
> > the pending waiter never has its state set ot TASK_RUNNING unless
> > something else had woke it up.
> >
> > The smp_mb() added to make this test work is just as expensive as
> > just calling wakeup. And since we we fail to wake up anyway, we are
> > doing both a smp_mb() and wakeup as well.
> >
> > I tested this with dbench and we run faster without this patch.
> > I also tried a variant that instead fixed the loop, to change the state
> > only if the spinner was to go to sleep, and that still did not show
> > any improvement.
>
> Just a quick note to say I am a bit skeptical of this patch. I know you are offline next week, so lets plan on hashing it out after the new year before I ack it.
>
We shouldn't be too quick to merely rip this out without a little
thinking. Clearly this is broken, however the intent was clear.
That being that if a waiter is spinning, we don't need to wake it up.
The wake up path is substantially more than a barrier; it includes a
barrier as well as grabbing the task_rq_lock only to find that the task
is oncpu. Then various accounting is updated, etc.
We know definitively that a waiter can only spin if the owner is oncpu,
by definition of adaptive spinning. We also know that only the owner
can release the lock to a waiter (spinning or not). So it seems clear
that avoiding unnecessary contention on the rq lock would be a Good
Thing(tm).
Perhaps this cannot be done safely, but if you saw an improvement in
dbench by merely removing a barrier, imagine the improvement by removing
the contention on the lock.
Happy Holidays to all!
-PWM
> Happy holidays!
> -Greg
>
> >
> > Cc: Gregory Haskins <ghaskins@...ell.com>
> > Cc: Peter Morreale <pmorreale@...ell.com>
> > Signed-off-by: Steven Rostedt <rostedt@...dmis.org>
> > ---
> > kernel/rtmutex.c | 29 ++---------------------------
> > 1 files changed, 2 insertions(+), 27 deletions(-)
> >
> > diff --git a/kernel/rtmutex.c b/kernel/rtmutex.c
> > index 318d7ed..e218873 100644
> > --- a/kernel/rtmutex.c
> > +++ b/kernel/rtmutex.c
> > @@ -554,33 +554,8 @@ static void wakeup_next_waiter(struct rt_mutex *lock,
> > int savestate)
> > */
> > if (!savestate)
> > wake_up_process(pendowner);
> > - else {
> > - /*
> > - * We can skip the actual (expensive) wakeup if the
> > - * waiter is already running, but we have to be careful
> > - * of race conditions because they may be about to sleep.
> > - *
> > - * The waiter-side protocol has the following pattern:
> > - * 1: Set state != RUNNING
> > - * 2: Conditionally sleep if waiter->task != NULL;
> > - *
> > - * And the owner-side has the following:
> > - * A: Set waiter->task = NULL
> > - * B: Conditionally wake if the state != RUNNING
> > - *
> > - * As long as we ensure 1->2 order, and A->B order, we
> > - * will never miss a wakeup.
> > - *
> > - * Therefore, this barrier ensures that waiter->task = NULL
> > - * is visible before we test the pendowner->state. The
> > - * corresponding barrier is in the sleep logic.
> > - */
> > - smp_mb();
> > -
> > - /* If !RUNNING && !RUNNING_MUTEX */
> > - if (pendowner->state & ~TASK_RUNNING_MUTEX)
> > - wake_up_process_mutex(pendowner);
> > - }
> > + else
> > + wake_up_process_mutex(pendowner);
> >
> > rt_mutex_set_owner(lock, pendowner, RT_MUTEX_OWNER_PENDING);
> >
>
>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists