[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAGsJ_4z_TR_UKhjxg-rzATodKJoNn2R-17KkqbeC-fLh3dK3sQ@mail.gmail.com>
Date: Wed, 29 Feb 2012 17:48:49 +0800
From: Barry Song <21cnbao@...il.com>
To: Marek Szyprowski <m.szyprowski@...sung.com>
Cc: linux-kernel@...r.kernel.org, linux-arm-kernel@...ts.infradead.org,
linux-media@...r.kernel.org, linux-mm@...ck.org,
linaro-mm-sig@...ts.linaro.org, Ohad Ben-Cohen <ohad@...ery.com>,
Daniel Walker <dwalker@...eaurora.org>,
Russell King <linux@....linux.org.uk>,
Arnd Bergmann <arnd@...db.de>,
Jonathan Corbet <corbet@....net>, Mel Gorman <mel@....ul.ie>,
Michal Nazarewicz <mina86@...a86.com>,
Dave Hansen <dave@...ux.vnet.ibm.com>,
Jesse Barker <jesse.barker@...aro.org>,
Kyungmin Park <kyungmin.park@...sung.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Rob Clark <rob.clark@...aro.org>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
Subject: Re: [Linaro-mm-sig] [PATCHv23 12/16] mm: trigger page reclaim in
alloc_contig_range() to stabilise watermarks
2012/2/23 Marek Szyprowski <m.szyprowski@...sung.com>:
> alloc_contig_range() performs memory allocation so it also should keep
> track on keeping the correct level of memory watermarks. This commit adds
> a call to *_slowpath style reclaim to grab enough pages to make sure that
> the final collection of contiguous pages from freelists will not starve
> the system.
>
> Signed-off-by: Marek Szyprowski <m.szyprowski@...sung.com>
> Signed-off-by: Kyungmin Park <kyungmin.park@...sung.com>
> CC: Michal Nazarewicz <mina86@...a86.com>
> Tested-by: Rob Clark <rob.clark@...aro.org>
> Tested-by: Ohad Ben-Cohen <ohad@...ery.com>
> Tested-by: Benjamin Gaignard <benjamin.gaignard@...aro.org>
> Tested-by: Robert Nelson <robertcnelson@...il.com>
> ---
> include/linux/mmzone.h | 9 +++++++
> mm/page_alloc.c | 62 ++++++++++++++++++++++++++++++++++++++++++++++++
> 2 files changed, 71 insertions(+), 0 deletions(-)
>
> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> index 4781f30..77db8c0 100644
> --- a/include/linux/mmzone.h
> +++ b/include/linux/mmzone.h
> @@ -63,8 +63,10 @@ enum {
>
> #ifdef CONFIG_CMA
> # define is_migrate_cma(migratetype) unlikely((migratetype) == MIGRATE_CMA)
> +# define cma_wmark_pages(zone) zone->min_cma_pages
> #else
> # define is_migrate_cma(migratetype) false
> +# define cma_wmark_pages(zone) 0
> #endif
>
> #define for_each_migratetype_order(order, type) \
> @@ -371,6 +373,13 @@ struct zone {
> /* see spanned/present_pages for more description */
> seqlock_t span_seqlock;
> #endif
> +#ifdef CONFIG_CMA
> + /*
> + * CMA needs to increase watermark levels during the allocation
> + * process to make sure that the system is not starved.
> + */
> + unsigned long min_cma_pages;
> +#endif
> struct free_area free_area[MAX_ORDER];
>
> #ifndef CONFIG_SPARSEMEM
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 7a0d286..39cd74f 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -5092,6 +5092,11 @@ static void __setup_per_zone_wmarks(void)
> low + (min >> 2);
> zone->watermark[WMARK_HIGH] = min_wmark_pages(zone) +
> low + (min >> 1);
> +
> + zone->watermark[WMARK_MIN] += cma_wmark_pages(zone);
> + zone->watermark[WMARK_LOW] += cma_wmark_pages(zone);
> + zone->watermark[WMARK_HIGH] += cma_wmark_pages(zone);
> +
> setup_zone_migrate_reserve(zone);
> spin_unlock_irqrestore(&zone->lock, flags);
> }
> @@ -5695,6 +5700,56 @@ static int __alloc_contig_migrate_range(unsigned long start, unsigned long end)
> return ret > 0 ? 0 : ret;
> }
>
> +/*
> + * Update zone's cma pages counter used for watermark level calculation.
> + */
> +static inline void __update_cma_watermarks(struct zone *zone, int count)
> +{
> + unsigned long flags;
> + spin_lock_irqsave(&zone->lock, flags);
> + zone->min_cma_pages += count;
> + spin_unlock_irqrestore(&zone->lock, flags);
> + setup_per_zone_wmarks();
> +}
> +
> +/*
> + * Trigger memory pressure bump to reclaim some pages in order to be able to
> + * allocate 'count' pages in single page units. Does similar work as
> + *__alloc_pages_slowpath() function.
> + */
> +static int __reclaim_pages(struct zone *zone, gfp_t gfp_mask, int count)
> +{
> + enum zone_type high_zoneidx = gfp_zone(gfp_mask);
> + struct zonelist *zonelist = node_zonelist(0, gfp_mask);
> + int did_some_progress = 0;
> + int order = 1;
> + unsigned long watermark;
> +
> + /*
> + * Increase level of watermarks to force kswapd do his job
> + * to stabilise at new watermark level.
> + */
> + __update_cma_watermarks(zone, count);
> +
> + /* Obey watermarks as if the page was being allocated */
> + watermark = low_wmark_pages(zone) + count;
> + while (!zone_watermark_ok(zone, 0, watermark, 0, 0)) {
> + wake_all_kswapd(order, zonelist, high_zoneidx, zone_idx(zone));
> +
> + did_some_progress = __perform_reclaim(gfp_mask, order, zonelist,
> + NULL);
> + if (!did_some_progress) {
> + /* Exhausted what can be done so it's blamo time */
> + out_of_memory(zonelist, gfp_mask, order, NULL);
out_of_memory() has got another param in the newest next/master tree,
out_of_memory(zonelist, gfp_mask, order, NULL, false) should be OK.
-barry
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists