[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20170803093701.icju4mxmto3ls3ch@techsingularity.net>
Date: Thu, 3 Aug 2017 10:37:01 +0100
From: Mel Gorman <mgorman@...hsingularity.net>
To: Michal Hocko <mhocko@...nel.org>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
David Rientjes <rientjes@...gle.com>,
Johannes Weiner <hannes@...xchg.org>,
Roman Gushchin <guro@...com>,
Tetsuo Handa <penguin-kernel@...ove.SAKURA.ne.jp>,
linux-mm@...ck.org, LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2 1/2] mm, oom: do not rely on TIF_MEMDIE for memory
reserves access
On Wed, Aug 02, 2017 at 10:29:14AM +0200, Michal Hocko wrote:
> For ages we have been relying on TIF_MEMDIE thread flag to mark OOM
> victims and then, among other things, to give these threads full
> access to memory reserves. There are few shortcomings of this
> implementation, though.
>
I believe the original intent was that allocations from the exit path
would be small and relatively short-lived.
> First of all and the most serious one is that the full access to memory
> reserves is quite dangerous because we leave no safety room for the
> system to operate and potentially do last emergency steps to move on.
>
> Secondly this flag is per task_struct while the OOM killer operates
> on mm_struct granularity so all processes sharing the given mm are
> killed. Giving the full access to all these task_structs could lead to
> a quick memory reserves depletion.
This is a valid concern.
> We have tried to reduce this risk by
> giving TIF_MEMDIE only to the main thread and the currently allocating
> task but that doesn't really solve this problem while it surely opens up
> a room for corner cases - e.g. GFP_NO{FS,IO} requests might loop inside
> the allocator without access to memory reserves because a particular
> thread was not the group leader.
>
> Now that we have the oom reaper and that all oom victims are reapable
> after 1b51e65eab64 ("oom, oom_reaper: allow to reap mm shared by the
> kthreads") we can be more conservative and grant only partial access to
> memory reserves because there are reasonable chances of the parallel
> memory freeing. We still want some access to reserves because we do not
> want other consumers to eat up the victim's freed memory. oom victims
> will still contend with __GFP_HIGH users but those shouldn't be so
> aggressive to starve oom victims completely.
>
> Introduce ALLOC_OOM flag and give all tsk_is_oom_victim tasks access to
> the half of the reserves. This makes the access to reserves independent
> on which task has passed through mark_oom_victim. Also drop any
> usage of TIF_MEMDIE from the page allocator proper and replace it by
> tsk_is_oom_victim as well which will make page_alloc.c completely
> TIF_MEMDIE free finally.
>
> CONFIG_MMU=n doesn't have oom reaper so let's stick to the original
> ALLOC_NO_WATERMARKS approach.
>
> There is a demand to make the oom killer memcg aware which will imply
> many tasks killed at once. This change will allow such a usecase without
> worrying about complete memory reserves depletion.
>
> Changes since v1
> - do not play tricks with nommu and grant access to memory reserves as
> long as TIF_MEMDIE is set
> - break out from allocation properly for oom victims as per Tetsuo
> - distinguish oom victims from other consumers of memory reserves in
> __gfp_pfmemalloc_flags - per Tetsuo
>
> Signed-off-by: Michal Hocko <mhocko@...e.com>
>
> diff --git a/mm/internal.h b/mm/internal.h
> index 24d88f084705..1ebcb1ed01b5 100644
> --- a/mm/internal.h
> +++ b/mm/internal.h
> @@ -480,6 +480,17 @@ unsigned long reclaim_clean_pages_from_list(struct zone *zone,
> /* Mask to get the watermark bits */
> #define ALLOC_WMARK_MASK (ALLOC_NO_WATERMARKS-1)
>
> +/*
> + * Only MMU archs have async oom victim reclaim - aka oom_reaper so we
> + * cannot assume a reduced access to memory reserves is sufficient for
> + * !MMU
> + */
> +#ifdef CONFIG_MMU
> +#define ALLOC_OOM 0x08
> +#else
> +#define ALLOC_OOM ALLOC_NO_WATERMARKS
> +#endif
> +
> #define ALLOC_HARDER 0x10 /* try to alloc harder */
> #define ALLOC_HIGH 0x20 /* __GFP_HIGH set */
> #define ALLOC_CPUSET 0x40 /* check for correct cpuset */
Ok, no collision with the wmark indexes so that should be fine. While I
didn't check, I suspect that !MMU users also have relatively few CPUs to
allow major contention.
> diff --git a/mm/oom_kill.c b/mm/oom_kill.c
> index 9e8b4f030c1c..c9f3569a76c7 100644
> --- a/mm/oom_kill.c
> +++ b/mm/oom_kill.c
> @@ -824,7 +824,8 @@ static void oom_kill_process(struct oom_control *oc, const char *message)
>
> /*
> * If the task is already exiting, don't alarm the sysadmin or kill
> - * its children or threads, just set TIF_MEMDIE so it can die quickly
> + * its children or threads, just give it access to memory reserves
> + * so it can die quickly
> */
> task_lock(p);
> if (task_will_free_mem(p)) {
> @@ -889,9 +890,9 @@ static void oom_kill_process(struct oom_control *oc, const char *message)
> count_memcg_event_mm(mm, OOM_KILL);
>
> /*
> - * We should send SIGKILL before setting TIF_MEMDIE in order to prevent
> - * the OOM victim from depleting the memory reserves from the user
> - * space under its control.
> + * We should send SIGKILL before granting access to memory reserves
> + * in order to prevent the OOM victim from depleting the memory
> + * reserves from the user space under its control.
> */
> do_send_sig_info(SIGKILL, SEND_SIG_FORCED, victim, true);
> mark_oom_victim(victim);
There is the secondary effect that some operations gets aborted entirely
if a SIGKILL is pending but that's neither here nor there.
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 80e4adb4c360..7ae0f6d45614 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -2930,7 +2930,7 @@ bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> {
> long min = mark;
> int o;
> - const bool alloc_harder = (alloc_flags & ALLOC_HARDER);
> + const bool alloc_harder = (alloc_flags & (ALLOC_HARDER|ALLOC_OOM));
>
> /* free_pages may go negative - that's OK */
> free_pages -= (1 << order) - 1;
> @@ -2943,10 +2943,19 @@ bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> * the high-atomic reserves. This will over-estimate the size of the
> * atomic reserve but it avoids a search.
> */
> - if (likely(!alloc_harder))
> + if (likely(!alloc_harder)) {
> free_pages -= z->nr_reserved_highatomic;
> - else
> - min -= min / 4;
> + } else {
> + /*
> + * OOM victims can try even harder than normal ALLOC_HARDER
> + * users
> + */
> + if (alloc_flags & ALLOC_OOM)
> + min -= min / 2;
> + else
> + min -= min / 4;
> + }
> +
>
> #ifdef CONFIG_CMA
> /* If allocation can't use CMA areas don't use free CMA pages */
I see no problem here although the comment could be slightly better. It
suggests at OOM victims can try harder but not why
/*
* OOM victims can try even harder than normal ALLOC_HARDER users on the
* grounds that it's definitely going to be in the exit path shortly and
* free memory. Any allocation it makes during the free path will be
* small and short-lived.
*/
> @@ -3184,7 +3193,7 @@ static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask)
> * of allowed nodes.
> */
> if (!(gfp_mask & __GFP_NOMEMALLOC))
> - if (test_thread_flag(TIF_MEMDIE) ||
> + if (tsk_is_oom_victim(current) ||
> (current->flags & (PF_MEMALLOC | PF_EXITING)))
> filter &= ~SHOW_MEM_FILTER_NODES;
> if (in_interrupt() || !(gfp_mask & __GFP_DIRECT_RECLAIM))
> @@ -3603,21 +3612,46 @@ gfp_to_alloc_flags(gfp_t gfp_mask)
> return alloc_flags;
> }
>
> -bool gfp_pfmemalloc_allowed(gfp_t gfp_mask)
> +static bool oom_reserves_allowed(struct task_struct *tsk)
> {
> - if (unlikely(gfp_mask & __GFP_NOMEMALLOC))
> + if (!tsk_is_oom_victim(tsk))
> + return false;
> +
> + /*
> + * !MMU doesn't have oom reaper so give access to memory reserves
> + * only to the thread with TIF_MEMDIE set
> + */
> + if (!IS_ENABLED(CONFIG_MMU) && !test_thread_flag(TIF_MEMDIE))
> return false;
>
> + return true;
> +}
> +
Ok, there is a chance that a task selected as an OOM kill victim may be
in the middle of a __GFP_NOMEMALLOC allocation but I can't actually see a
problem wiith that. __GFP_NOMEMALLOC users are not going to be in the exit
path (which we care about for an OOM killed task) and the caller should
always be able to handle a failure.
> +/*
> + * Distinguish requests which really need access to full memory
> + * reserves from oom victims which can live with a portion of it
> + */
> +static inline int __gfp_pfmemalloc_flags(gfp_t gfp_mask)
> +{
> + if (unlikely(gfp_mask & __GFP_NOMEMALLOC))
> + return 0;
> if (gfp_mask & __GFP_MEMALLOC)
> - return true;
> + return ALLOC_NO_WATERMARKS;
> if (in_serving_softirq() && (current->flags & PF_MEMALLOC))
> - return true;
> - if (!in_interrupt() &&
> - ((current->flags & PF_MEMALLOC) ||
> - unlikely(test_thread_flag(TIF_MEMDIE))))
> - return true;
> + return ALLOC_NO_WATERMARKS;
> + if (!in_interrupt()) {
> + if (current->flags & PF_MEMALLOC)
> + return ALLOC_NO_WATERMARKS;
> + else if (oom_reserves_allowed(current))
> + return ALLOC_OOM;
> + }
>
> - return false;
> + return 0;
> +}
> +
> +bool gfp_pfmemalloc_allowed(gfp_t gfp_mask)
> +{
> + return __gfp_pfmemalloc_flags(gfp_mask) > 0;
> }
Very subtle sign casing error here. If the flags ever use the high bit,
this wraps and fails. It "shouldn't be possible" but you could just remove
the "> 0" there to be on the safe side or have __gfp_pfmemalloc_flags
return unsigned.
>
> /*
> @@ -3770,6 +3804,7 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> unsigned long alloc_start = jiffies;
> unsigned int stall_timeout = 10 * HZ;
> unsigned int cpuset_mems_cookie;
> + int reserves;
>
This should be explicitly named to indicate it's about flags and not the
number of reserve pages or something else wacky.
> /*
> * In the slowpath, we sanity check order to avoid ever trying to
> @@ -3875,15 +3910,16 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> if (gfp_mask & __GFP_KSWAPD_RECLAIM)
> wake_all_kswapds(order, ac);
>
> - if (gfp_pfmemalloc_allowed(gfp_mask))
> - alloc_flags = ALLOC_NO_WATERMARKS;
> + reserves = __gfp_pfmemalloc_flags(gfp_mask);
> + if (reserves)
> + alloc_flags = reserves;
>
And if it's reserve_flags you can save a branch with
reserve_flags = __gfp_pfmemalloc_flags(gfp_mask);
alloc_pags |= reserve_flags;
It won't make much difference considering how branch-intensive the allocator
is anyway.
> /*
> * Reset the zonelist iterators if memory policies can be ignored.
> * These allocations are high priority and system rather than user
> * orientated.
> */
> - if (!(alloc_flags & ALLOC_CPUSET) || (alloc_flags & ALLOC_NO_WATERMARKS)) {
> + if (!(alloc_flags & ALLOC_CPUSET) || reserves) {
> ac->zonelist = node_zonelist(numa_node_id(), gfp_mask);
> ac->preferred_zoneref = first_zones_zonelist(ac->zonelist,
> ac->high_zoneidx, ac->nodemask);
> @@ -3960,8 +3996,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> goto got_pg;
>
> /* Avoid allocations with no watermarks from looping endlessly */
> - if (test_thread_flag(TIF_MEMDIE) &&
> - (alloc_flags == ALLOC_NO_WATERMARKS ||
> + if (tsk_is_oom_victim(current) &&
> + (alloc_flags == ALLOC_OOM ||
> (gfp_mask & __GFP_NOMEMALLOC)))
> goto nopage;
>
Mostly I only found nit-picks so whether you address them or not
Acked-by: Mel Gorman <mgorman@...hsingularity.net>
--
Mel Gorman
SUSE Labs
Powered by blists - more mailing lists