[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <20190327005948.24263-1-cai@lca.pw>
Date: Tue, 26 Mar 2019 20:59:48 -0400
From: Qian Cai <cai@....pw>
To: akpm@...ux-foundation.org
Cc: catalin.marinas@....com, cl@...ux.com, mhocko@...nel.org,
willy@...radead.org, penberg@...nel.org, rientjes@...gle.com,
iamjoonsoo.kim@....com, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, Qian Cai <cai@....pw>
Subject: [PATCH v4] kmemleak: survive in a low-memory situation
Kmemleak could quickly fail to allocate an object structure and then
disable itself below in a low-memory situation. For example, running a
mmap() workload triggering swapping and OOM. This is especially
problematic for running things like LTP testsuite where one OOM test
case would disable the whole kmemleak and render the rest of test cases
without kmemleak watching for leaking.
Kmemleak allocation could fail even though the tracked memory is
succeeded. Hence, it could still try to start a direct reclaim if it is
not executed in an atomic context (spinlock, irq-handler etc), or a
high-priority allocation in an atomic context as a last-ditch effort.
Since kmemleak is a debug feature, it is unlikely to be used in
production that memory resources is scarce where direct reclaim or
high-priority atomic allocations should not be granted lightly.
Unless there is a brave soul to reimplement the kmemleak to embed it's
metadata into the tracked memory itself in a foreseeable future, this
provides a good balance between enabling kmemleak in a low-memory
situation and not introducing too much hackiness into the existing
code for now. Another approach is to fail back the original allocation
once kmemleak_alloc() failed, but there are too many call sites to
deal with which makes it error-prone.
kmemleak: Cannot allocate a kmemleak_object structure
kmemleak: Kernel memory leak detector disabled
kmemleak: Automatic memory scanning thread ended
RIP: 0010:__alloc_pages_nodemask+0x242a/0x2ab0
Call Trace:
allocate_slab+0x4d9/0x930
new_slab+0x46/0x70
___slab_alloc+0x5d3/0x9c0
__slab_alloc+0x12/0x20
kmem_cache_alloc+0x30a/0x360
create_object+0x96/0x9a0
kmemleak_alloc+0x71/0xa0
kmem_cache_alloc+0x254/0x360
mempool_alloc_slab+0x3f/0x60
mempool_alloc+0x120/0x329
bio_alloc_bioset+0x1a8/0x510
get_swap_bio+0x107/0x470
__swap_writepage+0xab4/0x1650
swap_writepage+0x86/0xe0
Signed-off-by: Qian Cai <cai@....pw>
---
v4: Update the commit log.
Fix a typo in comments per Christ.
Consolidate the allocation.
v3: Update the commit log.
Simplify the code inspired by graph_trace_open() from ftrace.
v2: Remove the needless checking for NULL objects in slab_post_alloc_hook()
per Catalin.
mm/kmemleak.c | 11 ++++++++++-
1 file changed, 10 insertions(+), 1 deletion(-)
diff --git a/mm/kmemleak.c b/mm/kmemleak.c
index a2d894d3de07..7f4545ab1f84 100644
--- a/mm/kmemleak.c
+++ b/mm/kmemleak.c
@@ -580,7 +580,16 @@ static struct kmemleak_object *create_object(unsigned long ptr, size_t size,
struct rb_node **link, *rb_parent;
unsigned long untagged_ptr;
- object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp));
+ /*
+ * The tracked memory was allocated successful, if the kmemleak object
+ * failed to allocate for some reasons, it ends up with the whole
+ * kmemleak disabled, so try it harder.
+ */
+ gfp = (in_atomic() || irqs_disabled()) ?
+ gfp_kmemleak_mask(gfp) | GFP_ATOMIC :
+ gfp_kmemleak_mask(gfp) | __GFP_DIRECT_RECLAIM;
+
+ object = kmem_cache_alloc(object_cache, gfp);
if (!object) {
pr_warn("Cannot allocate a kmemleak_object structure\n");
kmemleak_disable();
--
2.17.2 (Apple Git-113)
Powered by blists - more mailing lists