[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZKNB88I41b9DLw1t@slm.duckdns.org>
Date: Mon, 3 Jul 2023 11:47:31 -1000
From: Tejun Heo <tj@...nel.org>
To: Pin-yen Lin <treapking@...omium.org>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>,
Brian Norris <briannorris@...omium.org>,
jiangshanlai@...il.com, peterz@...radead.org,
linux-kernel@...r.kernel.org, kernel-team@...a.com,
joshdon@...gle.com, brho@...gle.com, nhuck@...gle.com,
agk@...hat.com, snitzer@...nel.org, void@...ifault.com
Subject: Re: [PATCHSET v1 wq/for-6.5] workqueue: Improve unbound workqueue
execution locality
Hello, Pin-yen.
On Thu, Jun 29, 2023 at 05:49:27PM +0800, Pin-yen Lin wrote:
> > > I find that perplexing given that switching to a per-cpu workqueue remedies
> > > the situation quite a bit, which is how this patchset came to be. #3 is the
> > > same as per-cpu workqueue, so if you're seeing noticeably different
> > > performance numbers between #3 and per-cpu workqueue, there's something
> > > wrong with either the code or test setup.
> >
> In our case, per-cpu workqueue (removing WQ_UNBOUND) doesn't bring us
> better results. But given that pinning tasks to a single CPU core
> helps, we thought that the regression is related to the behavior of
> WQ_UNBOUND. Our findings are listed in [1].
I see.
> We already use WQ_SYSFS and the sysfs interface to pin the tasks, but
> thanks for the suggestion.
Yeah, I have no idea why there's such a drastic regression and the only way
to alleviate that is pinning the execution to a single CPU, which is also
different from other reports. It seems plausible that there are some
scheduling behavior changes and that's interacting negatively with power
saving but I'm sure that's the first thing you guys looked into.
>From workqueue's POV, I'm afraid using WQ_SYSFS and pinning it as needed
seems like about what can be done for your specific case.
Thanks.
--
tejun
Powered by blists - more mailing lists