[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20130629071245.GA5084@gmail.com>
Date: Sat, 29 Jun 2013 09:12:45 +0200
From: Ingo Molnar <mingo@...nel.org>
To: Tim Chen <tim.c.chen@...ux.intel.com>
Cc: Ingo Molnar <mingo@...e.hu>,
Andrea Arcangeli <aarcange@...hat.com>,
Mel Gorman <mgorman@...e.de>, "Shi, Alex" <alex.shi@...el.com>,
Andi Kleen <andi@...stfloor.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Michel Lespinasse <walken@...gle.com>,
Davidlohr Bueso <davidlohr.bueso@...com>,
"Wilcox, Matthew R" <matthew.r.wilcox@...el.com>,
Dave Hansen <dave.hansen@...el.com>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Rik van Riel <riel@...hat.com>, linux-kernel@...r.kernel.org,
linux-mm <linux-mm@...ck.org>
Subject: Re: Performance regression from switching lock to rw-sem for
anon-vma tree
* Tim Chen <tim.c.chen@...ux.intel.com> wrote:
> > If my analysis is correct so far then it might be useful to add two
> > more stats: did rwsem_spin_on_owner() fail because lock->owner == NULL
> > [owner released the rwsem], or because owner_running() failed [owner
> > went to sleep]?
>
> Ingo,
>
> I tabulated the cases where rwsem_spin_on_owner returns false and causes
> us to stop spinning.
>
> 97.12% was due to lock's owner switching to another writer
> 0.01% was due to the owner of the lock sleeping
> 2.87% was due to need_resched()
>
> I made a change to allow us to continue to spin even when lock's owner
> switch to another writer. I did get the lock to be acquired now mostly
> (98%) via optimistic spin and lock stealing, but my benchmark's
> throughput actually got reduced by 30% (too many cycles spent on useless
> spinning?).
Hm, I'm running out of quick ideas :-/ The writer-ends-spinning sequence
is pretty similar in the rwsem and in the mutex case. I'd have a look at
one more detail: is the wakeup of another writer in the rwsem case
singular, is only a single writer woken? I suspect the answer is yes ...
A quick glance suggests that the ordering of wakeups of waiters is the
same for mutexes and rwsems: FIFO, single waiter woken on slowpath-unlock.
So that shouldn't make a big difference.
If all last-ditch efforts to analyze it via counters fail then the way I'd
approach it next is brute-force instrumentation:
- First I'd create a workload 'steady state' that can be traced and
examined without worrying that that it ends or switches to some other
workload.
- Then I'd create a relatively lightweight trace (maybe trace_printk() is
lightweight enough), and capture key mutex and rwsem events.
- I'd capture a 1-10 seconds trace in steady state, both with rwsems and
mutexes. I'd have a good look at which tasks take locks and schedule
how and why. I'd try to eliminate any assymetries in behavior, i.e.
make rwsems behave like mutexes.
The risk and difficulty is that tracing can easily skew locking patterns,
so I'd first check whether with such new tracepoints enabled the assymetry
in behavior and regression is still present.
Thanks,
Ingo
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists