[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZNVcfdH8+N5Q83J/@casper.infradead.org>
Date: Thu, 10 Aug 2023 22:54:05 +0100
From: Matthew Wilcox <willy@...radead.org>
To: Peter Xu <peterx@...hat.com>
Cc: David Hildenbrand <david@...hat.com>, linux-kernel@...r.kernel.org,
linux-mm@...ck.org, linux-doc@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>,
Jonathan Corbet <corbet@....net>,
Mike Kravetz <mike.kravetz@...cle.com>,
Hugh Dickins <hughd@...gle.com>,
Ryan Roberts <ryan.roberts@....com>,
Yin Fengwei <fengwei.yin@...el.com>,
Yang Shi <shy828301@...il.com>, Zi Yan <ziy@...dia.com>
Subject: Re: [PATCH mm-unstable v1] mm: add a total mapcount for large folios
On Thu, Aug 10, 2023 at 05:48:19PM -0400, Peter Xu wrote:
> > Yes, that comment from Hugh primarily discusses how we could possibly
> > optimize the loop, and if relying on folio_nr_pages_mapped() to reduce the
> > iterations would be racy. As far as I can see, there are cases where "it
> > would be certainly a bad idea" :)
>
> Is the race described about mapcount being changed right after it's read?
> Are you aware of anything specific that will be broken, and will be fixed
> with this patch?
The problem is that people check the mapcount while holding no locks;
not the PTL, not the page lock. So it's an unfixable race.
> Having a total mapcount does sound helpful if partial folio is common
> indeed.
>
> I'm curious whether that'll be so common after the large anon folio work -
> isn't it be sad if partial folio will be a norm? It sounds to me that's
> the case when small page sizes should be used.. and it's prone to waste?
The problem is that entire_mapcount isn't really entire_mapcount.
It's pmd_mapcount. I have had thoughts about using it as entire_mapcount,
but it gets gnarly when people do partial unmaps. So the _usual_ case
ends up touching every struct page. Which sucks. Also it's one of the
things which stands in the way of shrinking struct page.
But it's kind of annoying to explain all of this to you individually.
There have been hundreds of emails about it over the last months on
this mailing list. It would be nice if you could catch up instead of
jumping in.
Powered by blists - more mailing lists