lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87r3qpyciy.fsf@x220.int.ebiederm.org>
Date:	Sat, 09 May 2015 10:51:49 -0500
From:	ebiederm@...ssion.com (Eric W. Biederman)
To:	Ingo Molnar <mingo@...nel.org>
Cc:	Rik van Riel <riel@...hat.com>,
	Linus Torvalds <torvalds@...ux-foundation.org>,
	John Stoffel <john@...ffel.org>,
	Dave Hansen <dave.hansen@...ux.intel.com>,
	Dan Williams <dan.j.williams@...el.com>,
	Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
	Boaz Harrosh <boaz@...xistor.com>, Jan Kara <jack@...e.cz>,
	Mike Snitzer <snitzer@...hat.com>, Neil Brown <neilb@...e.de>,
	Benjamin Herrenschmidt <benh@...nel.crashing.org>,
	Heiko Carstens <heiko.carstens@...ibm.com>,
	Chris Mason <clm@...com>, Paul Mackerras <paulus@...ba.org>,
	"H. Peter Anvin" <hpa@...or.com>, Christoph Hellwig <hch@....de>,
	Alasdair Kergon <agk@...hat.com>,
	"linux-nvdimm\@lists.01.org" <linux-nvdimm@...1.01.org>,
	Mel Gorman <mgorman@...e.de>,
	Matthew Wilcox <willy@...ux.intel.com>,
	Ross Zwisler <ross.zwisler@...ux.intel.com>,
	Martin Schwidefsky <schwidefsky@...ibm.com>,
	Jens Axboe <axboe@...nel.dk>, Theodore Ts'o <tytso@....edu>,
	"Martin K. Petersen" <martin.petersen@...cle.com>,
	Julia Lawall <Julia.Lawall@...6.fr>, Tejun Heo <tj@...nel.org>,
	linux-fsdevel <linux-fsdevel@...r.kernel.org>,
	Andrew Morton <akpm@...ux-foundation.org>
Subject: Re: "Directly mapped persistent memory page cache"

Ingo Molnar <mingo@...nel.org> writes:

> * Rik van Riel <riel@...hat.com> wrote:
>
>> On 05/08/2015 11:54 AM, Linus Torvalds wrote:
>> > On Fri, May 8, 2015 at 7:40 AM, John Stoffel <john@...ffel.org> wrote:
>> >>
>> >> Now go and look at your /home or /data/ or /work areas, where the
>> >> endusers are actually keeping their day to day work.  Photos, mp3,
>> >> design files, source code, object code littered around, etc.
>> > 
>> > However, the big files in that list are almost immaterial from a
>> > caching standpoint.
>> 
>> > The big files in your home directory? Let me make an educated guess.
>> > Very few to *none* of them are actually in your page cache right now.
>> > And you'd never even care if they ever made it into your page cache
>> > *at*all*. Much less whether you could ever cache them using large
>> > pages using some very fancy cache.
>> 
>> However, for persistent memory, all of the files will be "in 
>> memory".
>> 
>> Not instantiating the 4kB struct pages for 2MB areas that are not 
>> currently being accessed with small files may make a difference.
>>
>> For dynamically allocated 4kB page structs, we need some way to 
>> discover where they are. It may make sense, from a simplicity point 
>> of view, to have one mechanism that works both for pmem and for 
>> normal system memory.
>
> I don't think we need to or want to allocate page structs dynamically, 
> which makes the model really simple and robust.
>
> If we 'think big', we can create something very exciting IMHO, that 
> also gets rid of most of the complications with DIO, DAX, etc:
>
> "Directly mapped pmem integrated into the page cache":
> ------------------------------------------------------
>
>   - The pmem filesystem is mapped directly in all cases, it has device 
>     side struct page arrays, and its struct pages are directly in the 
>     page cache, write-through cached. (See further below about how we 
>     can do this.)
>
>     Note that this is radically different from the current approach 
>     that tries to use DIO and DAX to provide specialized "direct
>     access" APIs.
>
>     With the 'directly mapped' approach we have numerous advantages:
>
>        - no double buffering to main RAM: the device pages represent 
>          file content.
>
>        - no bdflush, no VM pressure, no writeback pressure, no
>          swapping: this is a very simple VM model where the device is
>          RAM and we don't have much dirty state. The primary kernel 
>          cache is the dcache and the directly mapped page cache, which 
>          is not a writeback cache in this case but essentially a 
>          logical->physical index cache of filesystem indexing 
>          metadata.
>
>        - every binary mmap()ed would be XIP mapped in essence
>
>        - every read() would be equivalent a DIO read, without the
>          complexity of DIO.
>
>        - every read() or write() done into a data mmap() area would
>          allow device-to-device zero copy DMA.
>
>        - main RAM caching would still be avilable and would work in 
>          many cases by default: as most apps use file processing 
>          buffers in anonymous memory into which they read() data.
>
> We can achieve this by statically allocating all page structs on the 
> device, in the following way:
>
>   - For every 128MB of pmem data we allocate 2MB of struct-page
>     descriptors, 64 bytes each, that describes that 128MB data range 
>     in a 4K granular way. We never have to allocate page structs as 
>     they are always there.
>
>   - Filesystems don't directly see the preallocated page arrays, they
>     still get a 'logical block space' presented that to them looks
>     like a continuous block device (which is 1.5% smaller than the 
>     true size of the device): this allows arbitrary filesystems to be 
>     put into such pmem devices, fsck will just work, etc.
>
>     I.e. no special pmem filesystem: the full range of existing block 
>     device based Linux filesystems can be used.
>
>   - These page structs are initialized in three layers:
>
>        - a single bit at 128MB data granularity: the first struct page
>          of the 2MB large array (32,768 struct page array members) 
>          represents the initialization state of all of them.
>
>        - a single bit at 2MB data granularity: the first struct page
>          of every 32K array within the 2MB array represents the whole
>          2MB data area. There are 64 such bits per 2MB array.
>
>        - a single bit at 4K data granularity: the whole page array.
>
>     A page marked uninitialized at a higher layer means all lower
>     layer struct pages are in their initial state.
>
>     This is a variant of your suggestion: one that keeps everything
>     2MB aligned, so that a single kernel side 2MB TLB covers a
>     continuous chunk of the page array. This allows us to create a
>     linear VMAP physical memory model to simplify index mapping.
>
>   - Looking up such a struct page (from a pfn) involves two simple,
>     easily computable indirections. With locality of access
>     present, 'hot' struct pages will be in the CPU cache. Them being
>     64 bytes each will help this. The on-device format is so simple
>     and so temporary that no fsck is needed for it.
>
>   - 2MB mappings, where desired, are 'natural' in such a layout:
>     everything's 2MB aligned both for kernel and user space use, while
>     4K granularity is still a first class citizen as well.
>
>   - For TB range storage we could make it 1GB granular: We'd allocate 
>     a 1GB array for every 64 GB of data. This would also allow gbpage 
>     TLBs to be taken advantage of: especially on the kernel side 
>     (vmapping the 1GB page array) this might be useful, even if all 
>     actual file usage is 4KB granular. The last block would be allowed
>     to be smaller than 64GB, but size would still be rounded to 1GB to
>     keep the mapping simple.
>
> What do you think?

The tricky bit is what happens when you reboot and run a different
version of the kernel, especially a kernel with things debugging
features like kmemcheck that increase the size of struct page.

I think we could reserve space for struct page entries in the persistent
memory and 64bytes appears to be a reasonable size.  But it would have
to be something that we initialize on mount or initialize on demand.

I don't think we could have persistent struct page entries, as the exact
contents of the struct page entries is too volatile and too different
between architectures.  Especially architecture changes that a pmem
store is likely to see such as switching between a 32bit and a 64bit
kernel.

Further I think where in the persistent memory the struct page arrays
live is something we could leave up to the filesystem.  We could have
some reasonable constraints to make it fast but I think whoever decides
where things live on the persistent memory can make that choice.

For small persistent memories it probably make sense to allocate the
struct page array describing them out of ordinary ram.  For small
memories I don't think we are talking enough memory to worry about.
For TB+ persistent memories where you need 16GiB per TiB it makes sense
to allocate a one or several regions to store your struct page arrays,
as you can't count on ordinary ram having enough capacity, and you may
not even be talking about a system that actually has ordinary ram at
that point.  

Eric
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ