[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20210412124909.olui5hfhceakt4u4@beryllium.lan>
Date: Mon, 12 Apr 2021 14:49:09 +0200
From: Daniel Wagner <dwagner@...e.de>
To: Jason Gunthorpe <jgg@...pe.ca>
Cc: linux-nvme@...ts.infradead.org, linux-kernel@...r.kernel.org,
Keith Busch <kbusch@...nel.org>, Jens Axboe <axboe@...com>,
Christoph Hellwig <hch@....de>,
Sagi Grimberg <sagi@...mberg.me>,
Steve Wise <swise@...ngridcomputing.com>,
Leon Romanovsky <leon@...nel.org>,
Potnuri Bharat Teja <bharat@...lsio.com>
Subject: Re: [PATCH] nvme: Drop WQ_MEM_RECLAIM flag from core workqueues
Hi Jason,
On Mon, Apr 12, 2021 at 09:31:49AM -0300, Jason Gunthorpe wrote:
> What does early init have to do with WQ_MEM_RECLAIM?
40c17f75dfa9 ("workqueue: allow WQ_MEM_RECLAIM on early init workqueues")
Workqueues can be created early during boot before workqueue subsystem
in fully online - work items are queued waiting for later full
initialization. However, early init wasn't supported for
WQ_MEM_RECLAIM workqueues causing unnecessary annoyances for a subset
of users. Expand early init support to include WQ_MEM_RECLAIM
workqueues.
That's the connection between WQ_MEM_RECLAIM and early init.
> WQ_MEM_RECLIAM is required when any thread in a reclaim context goes
> to sleep waiting for a WQ to complete. For instance by calling
> flush_workqueue() or many other things.
>
> The sleeping reclaim context must be guarenteed that the work can be
> completed without the work, work queue machinery, or anything the work
> has become interconnected with, recursing back into a reclaim.
>
> IIRC the issue here was some destroy or flush work in some error
> condition that happened to be under a reclaim context?
I understand what you are saying and I would totally agree with you but
where is the code for this?
I've grepped through the code and didn't find anything which supports
the guarantee claim. Neither mm nor schedule seems to care about this
flag nor workqueue.c (except the early init bits). Or I must miss
something.
Thanks,
Daniel
Powered by blists - more mailing lists