[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <fcde09be-ae34-4f09-a324-825fb2d4fac2@default>
Date: Wed, 25 Apr 2012 08:40:02 -0700 (PDT)
From: Dan Magenheimer <dan.magenheimer@...cle.com>
To: Nitin Gupta <ngupta@...are.org>, Minchan Kim <minchan@...nel.org>
Cc: Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
Seth Jennings <sjenning@...ux.vnet.ibm.com>,
Andrew Morton <akpm@...ux-foundation.org>,
linux-kernel@...r.kernel.org, linux-mm@...ck.org
Subject: RE: [PATCH 6/6] zsmalloc: make zsmalloc portable
> From: Nitin Gupta [mailto:ngupta@...are.org]
> Subject: Re: [PATCH 6/6] zsmalloc: make zsmalloc portable
>
> On 04/25/2012 02:23 AM, Minchan Kim wrote:
>
> > The zsmalloc uses __flush_tlb_one and set_pte.
> > It's very lower functions so that it makes arhcitecture dependency
> > so currently zsmalloc is used by only x86.
> > This patch changes them with map_vm_area and unmap_kernel_range so
> > it should work all architecture.
> >
> > Signed-off-by: Minchan Kim <minchan@...nel.org>
> > ---
> > drivers/staging/zsmalloc/Kconfig | 4 ----
> > drivers/staging/zsmalloc/zsmalloc-main.c | 27 +++++++++++++++++----------
> > drivers/staging/zsmalloc/zsmalloc_int.h | 1 -
> > 3 files changed, 17 insertions(+), 15 deletions(-)
> >
> > diff --git a/drivers/staging/zsmalloc/Kconfig b/drivers/staging/zsmalloc/Kconfig
> > index a5ab720..9084565 100644
> > --- a/drivers/staging/zsmalloc/Kconfig
> > +++ b/drivers/staging/zsmalloc/Kconfig
> > @@ -1,9 +1,5 @@
> > config ZSMALLOC
> > tristate "Memory allocator for compressed pages"
> > - # X86 dependency is because of the use of __flush_tlb_one and set_pte
> > - # in zsmalloc-main.c.
> > - # TODO: convert these to portable functions
> > - depends on X86
> > default n
> > help
> > zsmalloc is a slab-based memory allocator designed to store
> > diff --git a/drivers/staging/zsmalloc/zsmalloc-main.c b/drivers/staging/zsmalloc/zsmalloc-main.c
> > index ff089f8..cc017b1 100644
> > --- a/drivers/staging/zsmalloc/zsmalloc-main.c
> > +++ b/drivers/staging/zsmalloc/zsmalloc-main.c
> > @@ -442,7 +442,7 @@ static int zs_cpu_notifier(struct notifier_block *nb, unsigned long action,
> > area = &per_cpu(zs_map_area, cpu);
> > if (area->vm)
> > break;
> > - area->vm = alloc_vm_area(2 * PAGE_SIZE, area->vm_ptes);
> > + area->vm = alloc_vm_area(2 * PAGE_SIZE, NULL);
> > if (!area->vm)
> > return notifier_from_errno(-ENOMEM);
> > break;
> > @@ -696,13 +696,22 @@ void *zs_map_object(struct zs_pool *pool, void *handle)
> > } else {
> > /* this object spans two pages */
> > struct page *nextp;
> > + struct page *pages[2];
> > + struct page **page_array = &pages[0];
> > + int err;
> >
> > nextp = get_next_page(page);
> > BUG_ON(!nextp);
> >
> > + page_array[0] = page;
> > + page_array[1] = nextp;
> >
> > - set_pte(area->vm_ptes[0], mk_pte(page, PAGE_KERNEL));
> > - set_pte(area->vm_ptes[1], mk_pte(nextp, PAGE_KERNEL));
> > + /*
> > + * map_vm_area never fail because we already allocated
> > + * pages for page table in alloc_vm_area.
> > + */
> > + err = map_vm_area(area->vm, PAGE_KERNEL, &page_array);
> > + BUG_ON(err);
> >
> > /* We pre-allocated VM area so mapping can never fail */
> > area->vm_addr = area->vm->addr;
> > @@ -730,14 +739,12 @@ void zs_unmap_object(struct zs_pool *pool, void *handle)
> > off = obj_idx_to_offset(page, obj_idx, class->size);
> >
> > area = &__get_cpu_var(zs_map_area);
> > - if (off + class->size <= PAGE_SIZE) {
> > + if (off + class->size <= PAGE_SIZE)
> > kunmap_atomic(area->vm_addr);
> > - } else {
> > - set_pte(area->vm_ptes[0], __pte(0));
> > - set_pte(area->vm_ptes[1], __pte(0));
> > - __flush_tlb_one((unsigned long)area->vm_addr);
> > - __flush_tlb_one((unsigned long)area->vm_addr + PAGE_SIZE);
> > - }
> > + else
> > + unmap_kernel_range((unsigned long)area->vm->addr,
> > + PAGE_SIZE * 2);
> > +
>
>
>
> This would certainly work but would incur unncessary cost. All we need
> to do is to flush the local TLB entry correpsonding to these two pages.
> However, unmap_kernel_range --> flush_tlb_kernel_range woule cause TLB
> flush on all CPUs. Additionally, implementation of this function
> (flush_tlb_kernel_range) on architecutures like x86 seems naive since it
> flushes the entire TLB on all the CPUs.
>
> Even with all this penalty, I'm inclined on keeping this change to
> remove x86 only dependency, keeping improvements as future work.
>
> I think Seth was working on this improvement but not sure about the
> current status. Seth?
I wouldn't normally advocate an architecture-specific ifdef, but the
penalty for portability here seems high enough that it could make
sense here, perhaps hidden away in zsmalloc.h? Perhaps eventually
in a mm header file as "unmap_kernel_page_pair_local()"?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists