[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <20220727234241.1423357-1-imran.f.khan@oracle.com>
Date: Thu, 28 Jul 2022 09:42:41 +1000
From: Imran Khan <imran.f.khan@...cle.com>
To: glider@...gle.com, elver@...gle.com, dvyukov@...gle.com,
cl@...ux.com, penberg@...nel.org, rientjes@...gle.com,
iamjoonsoo.kim@....com, akpm@...ux-foundation.org, vbabka@...e.cz,
roman.gushchin@...ux.dev, 42.hyeyoo@...il.com, corbet@....net
Cc: linux-doc@...r.kernel.org, linux-kernel@...r.kernel.org,
kasan-dev@...glegroups.com, linux-mm@...ck.org
Subject: [RFC PATCH] mm/kfence: Introduce kernel parameter for selective usage of kfence.
By default kfence allocation can happen for any slub object, whose size
is up to PAGE_SIZE, as long as that allocation is the first allocation
after expiration of kfence sample interval. But in certain debugging
scenarios we may be interested in debugging corruptions involving
some specific slub objects like dentry or ext4_* etc. In such cases
limiting kfence for allocations involving only specific slub objects
will increase the probablity of catching the issue since kfence pool
will not be consumed by other slub objects.
This patch introduces a kernel parameter slub_kfence that can be used
to specify a comma separated list of slabs for which kfence allocations
will happen. Also introduce a sysfs parameter that can be used to re-enable
kfence for all slabs.
Signed-off-by: Imran Khan <imran.f.khan@...cle.com>
---
I am also working on getting kfence enabled for specific slabs using
/sys/kernel/slab/<slab_name>/kfence interface but in the meanwhile
I am sharing this RFC patch to get some early feedback. Especially
if this feature makes sense or if there is any better/existing way to
achieve similar end results.
.../admin-guide/kernel-parameters.txt | 5 ++
include/linux/kfence.h | 1 +
include/linux/slab.h | 6 ++
mm/kfence/core.c | 86 +++++++++++++++++++
mm/slub.c | 47 ++++++++++
5 files changed, 145 insertions(+)
diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
index 98e5cb91faab..d66f555df7ba 100644
--- a/Documentation/admin-guide/kernel-parameters.txt
+++ b/Documentation/admin-guide/kernel-parameters.txt
@@ -5553,6 +5553,11 @@
last alloc / free. For more information see
Documentation/mm/slub.rst.
+ slub_kfence[=slabs][,slabs]]...] [MM, SLUB]
+ Specifies the slabs for which kfence debug mechanism
+ can be used. For more information about kfence see
+ Documentation/dev-tools/kfence.rst.
+
slub_max_order= [MM, SLUB]
Determines the maximum allowed order for slabs.
A high setting may cause OOMs due to memory
diff --git a/include/linux/kfence.h b/include/linux/kfence.h
index 726857a4b680..140fc4fe87e1 100644
--- a/include/linux/kfence.h
+++ b/include/linux/kfence.h
@@ -125,6 +125,7 @@ static __always_inline void *kfence_alloc(struct kmem_cache *s, size_t size, gfp
#endif
if (likely(atomic_read(&kfence_allocation_gate)))
return NULL;
+
return __kfence_alloc(s, size, flags);
}
diff --git a/include/linux/slab.h b/include/linux/slab.h
index 0fefdf528e0d..b0def74d9fa1 100644
--- a/include/linux/slab.h
+++ b/include/linux/slab.h
@@ -119,6 +119,12 @@
*/
#define SLAB_NO_USER_FLAGS ((slab_flags_t __force)0x10000000U)
+#ifdef CONFIG_KFENCE
+#define SLAB_KFENCE ((slab_flags_t __force)0x20000000U)
+#else
+#define SLAB_KFENCE 0
+#endif
+
/* The following flags affect the page allocator grouping pages by mobility */
/* Objects are reclaimable */
#define SLAB_RECLAIM_ACCOUNT ((slab_flags_t __force)0x00020000U)
diff --git a/mm/kfence/core.c b/mm/kfence/core.c
index c252081b11df..017ea87b495b 100644
--- a/mm/kfence/core.c
+++ b/mm/kfence/core.c
@@ -132,6 +132,8 @@ DEFINE_STATIC_KEY_FALSE(kfence_allocation_key);
/* Gates the allocation, ensuring only one succeeds in a given period. */
atomic_t kfence_allocation_gate = ATOMIC_INIT(1);
+/* Determines if kfence allocation happens only for selected slabs. */
+atomic_t kfence_global_alloc = ATOMIC_INIT(1);
/*
* A Counting Bloom filter of allocation coverage: limits currently covered
* allocations of the same source filling up the pool.
@@ -1003,6 +1005,14 @@ void *__kfence_alloc(struct kmem_cache *s, size_t size, gfp_t flags)
return NULL;
}
+ /*
+ * Skip allocation if kfence has been enable for selected slabs
+ * and this slab is not one of the selected slabs.
+ */
+ if (unlikely(!atomic_read(&kfence_global_alloc)
+ && !(s->flags & SLAB_KFENCE)))
+ return NULL;
+
if (atomic_inc_return(&kfence_allocation_gate) > 1)
return NULL;
#ifdef CONFIG_KFENCE_STATIC_KEYS
@@ -1156,3 +1166,79 @@ bool kfence_handle_page_fault(unsigned long addr, bool is_write, struct pt_regs
return kfence_unprotect(addr); /* Unprotect and let access proceed. */
}
+
+#ifdef CONFIG_SYSFS
+static ssize_t kfence_global_alloc_enabled_show(struct kobject *kobj,
+ struct kobj_attribute *attr, char *buf)
+{
+ return sysfs_emit(buf, "%d\n", atomic_read(&kfence_global_alloc));
+}
+
+static ssize_t kfence_global_alloc_enabled_store(struct kobject *kobj,
+ struct kobj_attribute *attr,
+ const char *buf, size_t count)
+{
+ struct kmem_cache *s;
+ ssize_t ret;
+ int val;
+
+ ret = kstrtoint(buf, 10, &val);
+ if (ret)
+ return ret;
+
+ if (val != 1)
+ return -EINVAL;
+
+ atomic_set(&kfence_global_alloc, val);
+
+ /*
+ * If kfence is re-enabled for all slabs from sysfs, disable
+ * slab specific usage of kfence.
+ */
+ mutex_lock(&slab_mutex);
+ list_for_each_entry(s, &slab_caches, list)
+ if (s->flags & SLAB_KFENCE)
+ s->flags &= ~SLAB_KFENCE;
+ mutex_unlock(&slab_mutex);
+
+ return count;
+}
+
+static struct kobj_attribute kfence_global_alloc_enabled_attr =
+ __ATTR(kfence_global_alloc_enabled,
+ 0644,
+ kfence_global_alloc_enabled_show,
+ kfence_global_alloc_enabled_store);
+
+static struct attribute *kfence_attrs[] = {
+ &kfence_global_alloc_enabled_attr.attr,
+ NULL,
+};
+
+static const struct attribute_group kfence_attr_group = {
+ .attrs = kfence_attrs,
+};
+
+static int __init kfence_init_sysfs(void)
+{
+ int err;
+ struct kobject *kfence_kobj;
+
+ kfence_kobj = kobject_create_and_add("kfence", mm_kobj);
+ if (!kfence_kobj) {
+ pr_err("failed to create kfence_global_alloc_enabled kobject\n");
+ return -ENOMEM;
+ }
+ err = sysfs_create_group(kfence_kobj, &kfence_attr_group);
+ if (err) {
+ pr_err("failed to register numa group\n");
+ goto delete_obj;
+ }
+ return 0;
+
+delete_obj:
+ kobject_put(kfence_kobj);
+ return err;
+}
+subsys_initcall(kfence_init_sysfs);
+#endif /* CONFIG_SYSFS */
diff --git a/mm/slub.c b/mm/slub.c
index 862dbd9af4f5..7ee67ba5097c 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -645,6 +645,7 @@ static slab_flags_t slub_debug;
#endif
static char *slub_debug_string;
+static char *slub_kfence_list;
static int disable_higher_order_debug;
/*
@@ -1589,6 +1590,27 @@ static int __init setup_slub_debug(char *str)
__setup("slub_debug", setup_slub_debug);
+#ifdef CONFIG_KFENCE
+extern atomic_t kfence_global_alloc;
+
+static int __init setup_slub_kfence(char *str)
+{
+ if (*str++ != '=' || !*str)
+ return 1;
+
+ slub_kfence_list = str;
+
+ /*
+ * Disable global kfence usage if specific slabs
+ * were specified in bootargs.
+ */
+ atomic_set(&kfence_global_alloc, 0);
+
+ return 1;
+}
+__setup("slub_kfence", setup_slub_kfence);
+#endif
+
/*
* kmem_cache_flags - apply debugging options to the cache
* @object_size: the size of an object without meta data
@@ -1653,6 +1675,31 @@ slab_flags_t kmem_cache_flags(unsigned int object_size,
}
}
+ /* Check if kfence has been enabled for this slab */
+ iter = slub_kfence_list;
+
+ while (iter && *iter) {
+ char *end, *glob;
+ size_t cmplen;
+
+ end = strchrnul(iter, ',');
+
+ glob = strnchr(iter, end - iter, '*');
+
+ if (glob)
+ cmplen = glob - iter;
+ else
+ cmplen = end - iter;
+
+ if (!strncmp(iter, name, cmplen))
+ flags |= SLAB_KFENCE;
+
+ if (!*end)
+ break;
+
+ iter = end + 1;
+ }
+
return flags | slub_debug_local;
}
#else /* !CONFIG_SLUB_DEBUG */
--
2.30.2
Powered by blists - more mailing lists