lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Sat, 12 Mar 2016 11:09:36 +0800 From: xuyiping <xuyiping@...ilicon.com> To: Minchan Kim <minchan@...nel.org>, Andrew Morton <akpm@...ux-foundation.org> CC: <linux-mm@...ck.org>, <linux-kernel@...r.kernel.org>, <jlayton@...chiereds.net>, <bfields@...ldses.org>, Vlastimil Babka <vbabka@...e.cz>, Joonsoo Kim <iamjoonsoo.kim@....com>, <koct9i@...il.com>, <aquini@...hat.com>, <virtualization@...ts.linux-foundation.org>, Mel Gorman <mgorman@...e.de>, Hugh Dickins <hughd@...gle.com>, Sergey Senozhatsky <sergey.senozhatsky@...il.com>, <rknize@...orola.com>, Rik van Riel <riel@...hat.com>, Gioh Kim <gurugio@...mail.net> Subject: Re: [PATCH v1 13/19] zsmalloc: factor page chain functionality out On 2016/3/11 15:30, Minchan Kim wrote: > For migration, we need to create sub-page chain of zspage > dynamically so this patch factors it out from alloc_zspage. > > As a minor refactoring, it makes OBJ_ALLOCATED_TAG assign > more clear in obj_malloc(it could be another patch but it's > trivial so I want to put together in this patch). > > Signed-off-by: Minchan Kim <minchan@...nel.org> > --- > mm/zsmalloc.c | 78 ++++++++++++++++++++++++++++++++++------------------------- > 1 file changed, 45 insertions(+), 33 deletions(-) > > diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c > index bfc6a048afac..f86f8aaeb902 100644 > --- a/mm/zsmalloc.c > +++ b/mm/zsmalloc.c > @@ -977,7 +977,9 @@ static void init_zspage(struct size_class *class, struct page *first_page) > unsigned long off = 0; > struct page *page = first_page; > > - VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); > + first_page->freelist = NULL; > + INIT_LIST_HEAD(&first_page->lru); > + set_zspage_inuse(first_page, 0); > > while (page) { > struct page *next_page; > @@ -1022,13 +1024,44 @@ static void init_zspage(struct size_class *class, struct page *first_page) > set_freeobj(first_page, 0); > } > > +static void create_page_chain(struct page *pages[], int nr_pages) > +{ > + int i; > + struct page *page; > + struct page *prev_page = NULL; > + struct page *first_page = NULL; > + > + for (i = 0; i < nr_pages; i++) { > + page = pages[i]; > + > + INIT_LIST_HEAD(&page->lru); > + if (i == 0) { > + SetPagePrivate(page); > + set_page_private(page, 0); > + first_page = page; > + } > + > + if (i == 1) > + set_page_private(first_page, (unsigned long)page); > + if (i >= 1) > + set_page_private(page, (unsigned long)first_page); > + if (i >= 2) > + list_add(&page->lru, &prev_page->lru); > + if (i == nr_pages - 1) > + SetPagePrivate2(page); > + > + prev_page = page; > + } > +} > + > /* > * Allocate a zspage for the given size class > */ > static struct page *alloc_zspage(struct size_class *class, gfp_t flags) > { > - int i, error; > + int i; > struct page *first_page = NULL, *uninitialized_var(prev_page); > + struct page *pages[ZS_MAX_PAGES_PER_ZSPAGE]; > > /* > * Allocate individual pages and link them together as: > @@ -1041,43 +1074,23 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags) *uninitialized_var(prev_page) in alloc_zspage is not in use more. > * (i.e. no other sub-page has this flag set) and PG_private_2 to > * identify the last page. > */ > - error = -ENOMEM; > for (i = 0; i < class->pages_per_zspage; i++) { > struct page *page; > > page = alloc_page(flags); > - if (!page) > - goto cleanup; > - > - INIT_LIST_HEAD(&page->lru); > - if (i == 0) { /* first page */ > - page->freelist = NULL; > - SetPagePrivate(page); > - set_page_private(page, 0); > - first_page = page; > - set_zspage_inuse(page, 0); > + if (!page) { > + while (--i >= 0) > + __free_page(pages[i]); > + return NULL; > } > - if (i == 1) > - set_page_private(first_page, (unsigned long)page); > - if (i >= 1) > - set_page_private(page, (unsigned long)first_page); > - if (i >= 2) > - list_add(&page->lru, &prev_page->lru); > - if (i == class->pages_per_zspage - 1) /* last page */ > - SetPagePrivate2(page); > - prev_page = page; > + > + pages[i] = page; > } > > + create_page_chain(pages, class->pages_per_zspage); > + first_page = pages[0]; > init_zspage(class, first_page); > > - error = 0; /* Success */ > - > -cleanup: > - if (unlikely(error) && first_page) { > - free_zspage(first_page); > - first_page = NULL; > - } > - > return first_page; > } > > @@ -1419,7 +1432,6 @@ static unsigned long obj_malloc(struct size_class *class, > unsigned long m_offset; > void *vaddr; > > - handle |= OBJ_ALLOCATED_TAG; > obj = get_freeobj(first_page); > objidx_to_page_and_ofs(class, first_page, obj, > &m_page, &m_offset); > @@ -1429,10 +1441,10 @@ static unsigned long obj_malloc(struct size_class *class, > set_freeobj(first_page, link->next >> OBJ_ALLOCATED_TAG); > if (!class->huge) > /* record handle in the header of allocated chunk */ > - link->handle = handle; > + link->handle = handle | OBJ_ALLOCATED_TAG; > else > /* record handle in first_page->private */ > - set_page_private(first_page, handle); > + set_page_private(first_page, handle | OBJ_ALLOCATED_TAG); > kunmap_atomic(vaddr); > mod_zspage_inuse(first_page, 1); > zs_stat_inc(class, OBJ_USED, 1); >
Powered by blists - more mailing lists