[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190606094154.GB6795@fuggles.cambridge.arm.com>
Date: Thu, 6 Jun 2019 10:41:54 +0100
From: Will Deacon <will.deacon@....com>
To: Jan Glauber <jglauber@...vell.com>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>,
Jan Glauber <jglauber@...ium.com>,
Linux List Kernel Mailing <linux-kernel@...r.kernel.org>,
Catalin Marinas <catalin.marinas@....com>,
Linux ARM <linux-arm-kernel@...ts.infradead.org>,
Jayachandran Chandrasekharan Nair <jnair@...vell.com>
Subject: Re: [PATCH] lockref: Limit number of cmpxchg loop retries
On Thu, Jun 06, 2019 at 08:03:27AM +0000, Jan Glauber wrote:
> On Wed, Jun 05, 2019 at 01:16:46PM -0700, Linus Torvalds wrote:
> > On Wed, Jun 5, 2019 at 6:49 AM Jan Glauber <jglauber@...ium.com> wrote:
> > >
> > > Add an upper bound to the loop to force the fallback to spinlocks
> > > after some time. A retry value of 100 should not impact any hardware
> > > that does not have this issue.
> > >
> > > With the retry limit the performance of an open-close testcase
> > > improved between 60-70% on ThunderX2.
> >
> > Btw, did you do any kind of performance analysis across different
> > retry limit values?
>
> I tried 15/50/100/200/500, results were largely identical up to 100.
> For SMT=4 a higher retry value might be better, but unless we can add a
> sysctl value 100 looked like a good compromise to me.
Perhaps I'm just getting confused pre-morning-coffee, but I thought the
original complaint (and the reason for this patch even existing) was that
when many CPUs were hammering the lockref then performance tanked? In which
case, increasing the threshold as the number of CPUs increases seems
counter-intuitive to me because it suggests that the larger the system,
the harder we should try to make the cmpxchg work.
Will
Powered by blists - more mailing lists