lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAJD7tkYss7aTgQ18F9JHndEx6LOZ8NOoH97p=iCDeDBAiBv4yw@mail.gmail.com>
Date:   Sat, 1 Apr 2023 03:25:00 -0700
From:   Yosry Ahmed <yosryahmed@...gle.com>
To:     Vlastimil Babka <vbabka@...e.cz>
Cc:     linux-mm@...ck.org, linux-kernel@...r.kernel.org,
        patches@...ts.linux.dev
Subject: Re: [PATCH] mm: remove all the slab allocators

On Sat, Apr 1, 2023 at 2:47 AM Vlastimil Babka <vbabka@...e.cz> wrote:
>
> As the SLOB removal is on track and the SLAB removal is planned, I have
> realized - why should we stop there and not remove also SLUB? What's a
> slab allocator good for in 2023? The RAM sizes are getting larger and
> the modules cheaper [1]. The object constructor trick was perhaps
> interesting in 1994, but not with contemporary CPUs. So all the slab
> allocator does today is just adding an unnecessary layer of complexity
> over the page allocator.
>
> Thus, with this patch, all three slab allocators are removed, and only a
> layer that passes everything to the page allocator remains in the slab.h
> and mm/slab_common.c files. This will allow users to gradually
> transition away and use the page allocator directly. To summarize the
> advantages:
>
> - Less code to maintain: over 13k lines are removed by this patch, and
>   more could be removed if I wast^Wspent more time on this, and later as
>   users are transitioned from the legacy layer. This no longer needs a
>   separate subsystem so remove it from MAINTAINERS (I hope I can keep the
>   kernel.org account anyway, though).
>
> - Simplified MEMCG_KMEM accounting: while I was lazy and just marked it
>   BROKEN in this patch, it should be trivial to use the page memcg
>   accounting now that we use the page allocator. The per-object
>   accounting went through several iterations in the past and was always
>   complex and added overhead. Page accounting is much simpler by
>   comparison.
>
> - Simplified KASAN and friends: also was lazy in this patch so they
>   can't be enabled but should be easy to fix up and work just on the
>   page level.
>
> - Simpler debugging: just use debug_pagealloc=on, no need to look up the
>   exact syntax of the absurdly complex slub_debug parameter.
>
> - Speed: didn't measure, but for the page allocator we have pcplists, so
>   it should scale just fine. No need for the crazy SLUB's cmpxchg_double()
>   craziness. Maybe that thing could be now removed too? Yeah I can see
>   just two remaining users.
>
> Any downsides? Let's look at memory usage after virtme boot:
>
> Before (with SLUB):
> Slab:              26304 kB
>
> After:
> Slab:             295592 kB

Good catch. Easy 10x win in memory footprint.

Acked-by: Donald Duck <donald@...k.com>

>
> Well, that's not so bad, see [1].
>
> [1] https://www.theregister.com/2023/03/29/dram_prices_crash/
> ---
>  MAINTAINERS              |   15 -
>  include/linux/slab.h     |  211 +-
>  include/linux/slab_def.h |  124 -
>  include/linux/slub_def.h |  198 --
>  init/Kconfig             |    2 +-
>  mm/Kconfig               |  134 +-
>  mm/Makefile              |   10 -
>  mm/slab.c                | 4046 ------------------------
>  mm/slab.h                |  426 ---
>  mm/slab_common.c         |  876 ++---
>  mm/slob.c                |  757 -----
>  mm/slub.c                | 6506 --------------------------------------
>  12 files changed, 228 insertions(+), 13077 deletions(-)
>  delete mode 100644 include/linux/slab_def.h
>  delete mode 100644 include/linux/slub_def.h
>  delete mode 100644 mm/slab.c
>  delete mode 100644 mm/slob.c
>  delete mode 100644 mm/slub.c
>
> diff --git a/MAINTAINERS b/MAINTAINERS
> index 1dc8bd26b6cf..40b05ad03cd0 100644
> --- a/MAINTAINERS
> +++ b/MAINTAINERS
> @@ -19183,21 +19183,6 @@ F:     drivers/irqchip/irq-sl28cpld.c
>  F:     drivers/pwm/pwm-sl28cpld.c
>  F:     drivers/watchdog/sl28cpld_wdt.c
>
> -SLAB ALLOCATOR
> -M:     Christoph Lameter <cl@...ux.com>
> -M:     Pekka Enberg <penberg@...nel.org>
> -M:     David Rientjes <rientjes@...gle.com>
> -M:     Joonsoo Kim <iamjoonsoo.kim@....com>
> -M:     Andrew Morton <akpm@...ux-foundation.org>
> -M:     Vlastimil Babka <vbabka@...e.cz>
> -R:     Roman Gushchin <roman.gushchin@...ux.dev>
> -R:     Hyeonggon Yoo <42.hyeyoo@...il.com>
> -L:     linux-mm@...ck.org
> -S:     Maintained
> -T:     git git://git.kernel.org/pub/scm/linux/kernel/git/vbabka/slab.git
> -F:     include/linux/sl?b*.h
> -F:     mm/sl?b*
> -
>  SLCAN CAN NETWORK DRIVER
>  M:     Dario Binacchi <dario.binacchi@...rulasolutions.com>
>  L:     linux-can@...r.kernel.org
> diff --git a/include/linux/slab.h b/include/linux/slab.h
> index 45af70315a94..61602d54b1d0 100644
> --- a/include/linux/slab.h
> +++ b/include/linux/slab.h
> @@ -140,13 +140,14 @@
>
>  /* The following flags affect the page allocator grouping pages by mobility */
>  /* Objects are reclaimable */
> -#ifndef CONFIG_SLUB_TINY
>  #define SLAB_RECLAIM_ACCOUNT   ((slab_flags_t __force)0x00020000U)
> -#else
> -#define SLAB_RECLAIM_ACCOUNT   ((slab_flags_t __force)0)
> -#endif
>  #define SLAB_TEMPORARY         SLAB_RECLAIM_ACCOUNT    /* Objects are short-lived */
>
> +#define KMALLOC_NOT_NORMAL_BITS                                        \
> +       (__GFP_RECLAIMABLE |                                    \
> +       (IS_ENABLED(CONFIG_ZONE_DMA)   ? __GFP_DMA : 0) |       \
> +       (IS_ENABLED(CONFIG_MEMCG_KMEM) ? __GFP_ACCOUNT : 0))
> +
>  /*
>   * ZERO_SIZE_PTR will be returned for zero sized kmalloc requests.
>   *
> @@ -278,38 +279,11 @@ static inline unsigned int arch_slab_minalign(void)
>   * Kmalloc array related definitions
>   */
>
> -#ifdef CONFIG_SLAB
> -/*
> - * SLAB and SLUB directly allocates requests fitting in to an order-1 page
> - * (PAGE_SIZE*2).  Larger requests are passed to the page allocator.
> - */
> -#define KMALLOC_SHIFT_HIGH     (PAGE_SHIFT + 1)
> -#define KMALLOC_SHIFT_MAX      (MAX_ORDER + PAGE_SHIFT - 1)
> -#ifndef KMALLOC_SHIFT_LOW
> -#define KMALLOC_SHIFT_LOW      5
> -#endif
> -#endif
> -
> -#ifdef CONFIG_SLUB
> -#define KMALLOC_SHIFT_HIGH     (PAGE_SHIFT + 1)
> -#define KMALLOC_SHIFT_MAX      (MAX_ORDER + PAGE_SHIFT - 1)
> -#ifndef KMALLOC_SHIFT_LOW
> -#define KMALLOC_SHIFT_LOW      3
> -#endif
> -#endif
> -
> -#ifdef CONFIG_SLOB
> -/*
> - * SLOB passes all requests larger than one page to the page allocator.
> - * No kmalloc array is necessary since objects of different sizes can
> - * be allocated from the same page.
> - */
>  #define KMALLOC_SHIFT_HIGH     PAGE_SHIFT
>  #define KMALLOC_SHIFT_MAX      (MAX_ORDER + PAGE_SHIFT - 1)
>  #ifndef KMALLOC_SHIFT_LOW
>  #define KMALLOC_SHIFT_LOW      3
>  #endif
> -#endif
>
>  /* Maximum allocatable size */
>  #define KMALLOC_MAX_SIZE       (1UL << KMALLOC_SHIFT_MAX)
> @@ -336,130 +310,6 @@ static inline unsigned int arch_slab_minalign(void)
>  #define SLAB_OBJ_MIN_SIZE      (KMALLOC_MIN_SIZE < 16 ? \
>                                 (KMALLOC_MIN_SIZE) : 16)
>
> -/*
> - * Whenever changing this, take care of that kmalloc_type() and
> - * create_kmalloc_caches() still work as intended.
> - *
> - * KMALLOC_NORMAL can contain only unaccounted objects whereas KMALLOC_CGROUP
> - * is for accounted but unreclaimable and non-dma objects. All the other
> - * kmem caches can have both accounted and unaccounted objects.
> - */
> -enum kmalloc_cache_type {
> -       KMALLOC_NORMAL = 0,
> -#ifndef CONFIG_ZONE_DMA
> -       KMALLOC_DMA = KMALLOC_NORMAL,
> -#endif
> -#ifndef CONFIG_MEMCG_KMEM
> -       KMALLOC_CGROUP = KMALLOC_NORMAL,
> -#endif
> -#ifdef CONFIG_SLUB_TINY
> -       KMALLOC_RECLAIM = KMALLOC_NORMAL,
> -#else
> -       KMALLOC_RECLAIM,
> -#endif
> -#ifdef CONFIG_ZONE_DMA
> -       KMALLOC_DMA,
> -#endif
> -#ifdef CONFIG_MEMCG_KMEM
> -       KMALLOC_CGROUP,
> -#endif
> -       NR_KMALLOC_TYPES
> -};
> -
> -#ifndef CONFIG_SLOB
> -extern struct kmem_cache *
> -kmalloc_caches[NR_KMALLOC_TYPES][KMALLOC_SHIFT_HIGH + 1];
> -
> -/*
> - * Define gfp bits that should not be set for KMALLOC_NORMAL.
> - */
> -#define KMALLOC_NOT_NORMAL_BITS                                        \
> -       (__GFP_RECLAIMABLE |                                    \
> -       (IS_ENABLED(CONFIG_ZONE_DMA)   ? __GFP_DMA : 0) |       \
> -       (IS_ENABLED(CONFIG_MEMCG_KMEM) ? __GFP_ACCOUNT : 0))
> -
> -static __always_inline enum kmalloc_cache_type kmalloc_type(gfp_t flags)
> -{
> -       /*
> -        * The most common case is KMALLOC_NORMAL, so test for it
> -        * with a single branch for all the relevant flags.
> -        */
> -       if (likely((flags & KMALLOC_NOT_NORMAL_BITS) == 0))
> -               return KMALLOC_NORMAL;
> -
> -       /*
> -        * At least one of the flags has to be set. Their priorities in
> -        * decreasing order are:
> -        *  1) __GFP_DMA
> -        *  2) __GFP_RECLAIMABLE
> -        *  3) __GFP_ACCOUNT
> -        */
> -       if (IS_ENABLED(CONFIG_ZONE_DMA) && (flags & __GFP_DMA))
> -               return KMALLOC_DMA;
> -       if (!IS_ENABLED(CONFIG_MEMCG_KMEM) || (flags & __GFP_RECLAIMABLE))
> -               return KMALLOC_RECLAIM;
> -       else
> -               return KMALLOC_CGROUP;
> -}
> -
> -/*
> - * Figure out which kmalloc slab an allocation of a certain size
> - * belongs to.
> - * 0 = zero alloc
> - * 1 =  65 .. 96 bytes
> - * 2 = 129 .. 192 bytes
> - * n = 2^(n-1)+1 .. 2^n
> - *
> - * Note: __kmalloc_index() is compile-time optimized, and not runtime optimized;
> - * typical usage is via kmalloc_index() and therefore evaluated at compile-time.
> - * Callers where !size_is_constant should only be test modules, where runtime
> - * overheads of __kmalloc_index() can be tolerated.  Also see kmalloc_slab().
> - */
> -static __always_inline unsigned int __kmalloc_index(size_t size,
> -                                                   bool size_is_constant)
> -{
> -       if (!size)
> -               return 0;
> -
> -       if (size <= KMALLOC_MIN_SIZE)
> -               return KMALLOC_SHIFT_LOW;
> -
> -       if (KMALLOC_MIN_SIZE <= 32 && size > 64 && size <= 96)
> -               return 1;
> -       if (KMALLOC_MIN_SIZE <= 64 && size > 128 && size <= 192)
> -               return 2;
> -       if (size <=          8) return 3;
> -       if (size <=         16) return 4;
> -       if (size <=         32) return 5;
> -       if (size <=         64) return 6;
> -       if (size <=        128) return 7;
> -       if (size <=        256) return 8;
> -       if (size <=        512) return 9;
> -       if (size <=       1024) return 10;
> -       if (size <=   2 * 1024) return 11;
> -       if (size <=   4 * 1024) return 12;
> -       if (size <=   8 * 1024) return 13;
> -       if (size <=  16 * 1024) return 14;
> -       if (size <=  32 * 1024) return 15;
> -       if (size <=  64 * 1024) return 16;
> -       if (size <= 128 * 1024) return 17;
> -       if (size <= 256 * 1024) return 18;
> -       if (size <= 512 * 1024) return 19;
> -       if (size <= 1024 * 1024) return 20;
> -       if (size <=  2 * 1024 * 1024) return 21;
> -
> -       if (!IS_ENABLED(CONFIG_PROFILE_ALL_BRANCHES) && size_is_constant)
> -               BUILD_BUG_ON_MSG(1, "unexpected size in kmalloc_index()");
> -       else
> -               BUG();
> -
> -       /* Will never be reached. Needed because the compiler may complain */
> -       return -1;
> -}
> -static_assert(PAGE_SHIFT <= 20);
> -#define kmalloc_index(s) __kmalloc_index(s, true)
> -#endif /* !CONFIG_SLOB */
> -
>  void *__kmalloc(size_t size, gfp_t flags) __assume_kmalloc_alignment __alloc_size(1);
>
>  /**
> @@ -567,57 +417,15 @@ void *kmalloc_large_node(size_t size, gfp_t flags, int node) __assume_page_align
>   *     Try really hard to succeed the allocation but fail
>   *     eventually.
>   */
> -#ifndef CONFIG_SLOB
> -static __always_inline __alloc_size(1) void *kmalloc(size_t size, gfp_t flags)
> -{
> -       if (__builtin_constant_p(size) && size) {
> -               unsigned int index;
> -
> -               if (size > KMALLOC_MAX_CACHE_SIZE)
> -                       return kmalloc_large(size, flags);
> -
> -               index = kmalloc_index(size);
> -               return kmalloc_trace(
> -                               kmalloc_caches[kmalloc_type(flags)][index],
> -                               flags, size);
> -       }
> -       return __kmalloc(size, flags);
> -}
> -#else
>  static __always_inline __alloc_size(1) void *kmalloc(size_t size, gfp_t flags)
>  {
> -       if (__builtin_constant_p(size) && size > KMALLOC_MAX_CACHE_SIZE)
> -               return kmalloc_large(size, flags);
> -
> -       return __kmalloc(size, flags);
> +       return kmalloc_large(size, flags);
>  }
> -#endif
>
> -#ifndef CONFIG_SLOB
>  static __always_inline __alloc_size(1) void *kmalloc_node(size_t size, gfp_t flags, int node)
>  {
> -       if (__builtin_constant_p(size) && size) {
> -               unsigned int index;
> -
> -               if (size > KMALLOC_MAX_CACHE_SIZE)
> -                       return kmalloc_large_node(size, flags, node);
> -
> -               index = kmalloc_index(size);
> -               return kmalloc_node_trace(
> -                               kmalloc_caches[kmalloc_type(flags)][index],
> -                               flags, node, size);
> -       }
> -       return __kmalloc_node(size, flags, node);
> +       return kmalloc_large_node(size, flags, node);
>  }
> -#else
> -static __always_inline __alloc_size(1) void *kmalloc_node(size_t size, gfp_t flags, int node)
> -{
> -       if (__builtin_constant_p(size) && size > KMALLOC_MAX_CACHE_SIZE)
> -               return kmalloc_large_node(size, flags, node);
> -
> -       return __kmalloc_node(size, flags, node);
> -}
> -#endif
>
>  /**
>   * kmalloc_array - allocate memory for an array.
> @@ -785,12 +593,7 @@ size_t kmalloc_size_roundup(size_t size);
>
>  void __init kmem_cache_init_late(void);
>
> -#if defined(CONFIG_SMP) && defined(CONFIG_SLAB)
> -int slab_prepare_cpu(unsigned int cpu);
> -int slab_dead_cpu(unsigned int cpu);
> -#else
>  #define slab_prepare_cpu       NULL
>  #define slab_dead_cpu          NULL
> -#endif
>
>  #endif /* _LINUX_SLAB_H */
> diff --git a/include/linux/slab_def.h b/include/linux/slab_def.h
> deleted file mode 100644
> index a61e7d55d0d3..000000000000
> --- a/include/linux/slab_def.h
> +++ /dev/null
> @@ -1,124 +0,0 @@
> -/* SPDX-License-Identifier: GPL-2.0 */
> -#ifndef _LINUX_SLAB_DEF_H
> -#define        _LINUX_SLAB_DEF_H
> -
> -#include <linux/kfence.h>
> -#include <linux/reciprocal_div.h>
> -
> -/*
> - * Definitions unique to the original Linux SLAB allocator.
> - */
> -
> -struct kmem_cache {
> -       struct array_cache __percpu *cpu_cache;
> -
> -/* 1) Cache tunables. Protected by slab_mutex */
> -       unsigned int batchcount;
> -       unsigned int limit;
> -       unsigned int shared;
> -
> -       unsigned int size;
> -       struct reciprocal_value reciprocal_buffer_size;
> -/* 2) touched by every alloc & free from the backend */
> -
> -       slab_flags_t flags;             /* constant flags */
> -       unsigned int num;               /* # of objs per slab */
> -
> -/* 3) cache_grow/shrink */
> -       /* order of pgs per slab (2^n) */
> -       unsigned int gfporder;
> -
> -       /* force GFP flags, e.g. GFP_DMA */
> -       gfp_t allocflags;
> -
> -       size_t colour;                  /* cache colouring range */
> -       unsigned int colour_off;        /* colour offset */
> -       unsigned int freelist_size;
> -
> -       /* constructor func */
> -       void (*ctor)(void *obj);
> -
> -/* 4) cache creation/removal */
> -       const char *name;
> -       struct list_head list;
> -       int refcount;
> -       int object_size;
> -       int align;
> -
> -/* 5) statistics */
> -#ifdef CONFIG_DEBUG_SLAB
> -       unsigned long num_active;
> -       unsigned long num_allocations;
> -       unsigned long high_mark;
> -       unsigned long grown;
> -       unsigned long reaped;
> -       unsigned long errors;
> -       unsigned long max_freeable;
> -       unsigned long node_allocs;
> -       unsigned long node_frees;
> -       unsigned long node_overflow;
> -       atomic_t allochit;
> -       atomic_t allocmiss;
> -       atomic_t freehit;
> -       atomic_t freemiss;
> -
> -       /*
> -        * If debugging is enabled, then the allocator can add additional
> -        * fields and/or padding to every object. 'size' contains the total
> -        * object size including these internal fields, while 'obj_offset'
> -        * and 'object_size' contain the offset to the user object and its
> -        * size.
> -        */
> -       int obj_offset;
> -#endif /* CONFIG_DEBUG_SLAB */
> -
> -#ifdef CONFIG_KASAN_GENERIC
> -       struct kasan_cache kasan_info;
> -#endif
> -
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -       unsigned int *random_seq;
> -#endif
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -       unsigned int useroffset;        /* Usercopy region offset */
> -       unsigned int usersize;          /* Usercopy region size */
> -#endif
> -
> -       struct kmem_cache_node *node[MAX_NUMNODES];
> -};
> -
> -static inline void *nearest_obj(struct kmem_cache *cache, const struct slab *slab,
> -                               void *x)
> -{
> -       void *object = x - (x - slab->s_mem) % cache->size;
> -       void *last_object = slab->s_mem + (cache->num - 1) * cache->size;
> -
> -       if (unlikely(object > last_object))
> -               return last_object;
> -       else
> -               return object;
> -}
> -
> -/*
> - * We want to avoid an expensive divide : (offset / cache->size)
> - *   Using the fact that size is a constant for a particular cache,
> - *   we can replace (offset / cache->size) by
> - *   reciprocal_divide(offset, cache->reciprocal_buffer_size)
> - */
> -static inline unsigned int obj_to_index(const struct kmem_cache *cache,
> -                                       const struct slab *slab, void *obj)
> -{
> -       u32 offset = (obj - slab->s_mem);
> -       return reciprocal_divide(offset, cache->reciprocal_buffer_size);
> -}
> -
> -static inline int objs_per_slab(const struct kmem_cache *cache,
> -                                    const struct slab *slab)
> -{
> -       if (is_kfence_address(slab_address(slab)))
> -               return 1;
> -       return cache->num;
> -}
> -
> -#endif /* _LINUX_SLAB_DEF_H */
> diff --git a/include/linux/slub_def.h b/include/linux/slub_def.h
> deleted file mode 100644
> index f6df03f934e5..000000000000
> --- a/include/linux/slub_def.h
> +++ /dev/null
> @@ -1,198 +0,0 @@
> -/* SPDX-License-Identifier: GPL-2.0 */
> -#ifndef _LINUX_SLUB_DEF_H
> -#define _LINUX_SLUB_DEF_H
> -
> -/*
> - * SLUB : A Slab allocator without object queues.
> - *
> - * (C) 2007 SGI, Christoph Lameter
> - */
> -#include <linux/kfence.h>
> -#include <linux/kobject.h>
> -#include <linux/reciprocal_div.h>
> -#include <linux/local_lock.h>
> -
> -enum stat_item {
> -       ALLOC_FASTPATH,         /* Allocation from cpu slab */
> -       ALLOC_SLOWPATH,         /* Allocation by getting a new cpu slab */
> -       FREE_FASTPATH,          /* Free to cpu slab */
> -       FREE_SLOWPATH,          /* Freeing not to cpu slab */
> -       FREE_FROZEN,            /* Freeing to frozen slab */
> -       FREE_ADD_PARTIAL,       /* Freeing moves slab to partial list */
> -       FREE_REMOVE_PARTIAL,    /* Freeing removes last object */
> -       ALLOC_FROM_PARTIAL,     /* Cpu slab acquired from node partial list */
> -       ALLOC_SLAB,             /* Cpu slab acquired from page allocator */
> -       ALLOC_REFILL,           /* Refill cpu slab from slab freelist */
> -       ALLOC_NODE_MISMATCH,    /* Switching cpu slab */
> -       FREE_SLAB,              /* Slab freed to the page allocator */
> -       CPUSLAB_FLUSH,          /* Abandoning of the cpu slab */
> -       DEACTIVATE_FULL,        /* Cpu slab was full when deactivated */
> -       DEACTIVATE_EMPTY,       /* Cpu slab was empty when deactivated */
> -       DEACTIVATE_TO_HEAD,     /* Cpu slab was moved to the head of partials */
> -       DEACTIVATE_TO_TAIL,     /* Cpu slab was moved to the tail of partials */
> -       DEACTIVATE_REMOTE_FREES,/* Slab contained remotely freed objects */
> -       DEACTIVATE_BYPASS,      /* Implicit deactivation */
> -       ORDER_FALLBACK,         /* Number of times fallback was necessary */
> -       CMPXCHG_DOUBLE_CPU_FAIL,/* Failure of this_cpu_cmpxchg_double */
> -       CMPXCHG_DOUBLE_FAIL,    /* Number of times that cmpxchg double did not match */
> -       CPU_PARTIAL_ALLOC,      /* Used cpu partial on alloc */
> -       CPU_PARTIAL_FREE,       /* Refill cpu partial on free */
> -       CPU_PARTIAL_NODE,       /* Refill cpu partial from node partial */
> -       CPU_PARTIAL_DRAIN,      /* Drain cpu partial to node partial */
> -       NR_SLUB_STAT_ITEMS };
> -
> -#ifndef CONFIG_SLUB_TINY
> -/*
> - * When changing the layout, make sure freelist and tid are still compatible
> - * with this_cpu_cmpxchg_double() alignment requirements.
> - */
> -struct kmem_cache_cpu {
> -       void **freelist;        /* Pointer to next available object */
> -       unsigned long tid;      /* Globally unique transaction id */
> -       struct slab *slab;      /* The slab from which we are allocating */
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       struct slab *partial;   /* Partially allocated frozen slabs */
> -#endif
> -       local_lock_t lock;      /* Protects the fields above */
> -#ifdef CONFIG_SLUB_STATS
> -       unsigned stat[NR_SLUB_STAT_ITEMS];
> -#endif
> -};
> -#endif /* CONFIG_SLUB_TINY */
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -#define slub_percpu_partial(c)         ((c)->partial)
> -
> -#define slub_set_percpu_partial(c, p)          \
> -({                                             \
> -       slub_percpu_partial(c) = (p)->next;     \
> -})
> -
> -#define slub_percpu_partial_read_once(c)     READ_ONCE(slub_percpu_partial(c))
> -#else
> -#define slub_percpu_partial(c)                 NULL
> -
> -#define slub_set_percpu_partial(c, p)
> -
> -#define slub_percpu_partial_read_once(c)       NULL
> -#endif // CONFIG_SLUB_CPU_PARTIAL
> -
> -/*
> - * Word size structure that can be atomically updated or read and that
> - * contains both the order and the number of objects that a slab of the
> - * given order would contain.
> - */
> -struct kmem_cache_order_objects {
> -       unsigned int x;
> -};
> -
> -/*
> - * Slab cache management.
> - */
> -struct kmem_cache {
> -#ifndef CONFIG_SLUB_TINY
> -       struct kmem_cache_cpu __percpu *cpu_slab;
> -#endif
> -       /* Used for retrieving partial slabs, etc. */
> -       slab_flags_t flags;
> -       unsigned long min_partial;
> -       unsigned int size;      /* The size of an object including metadata */
> -       unsigned int object_size;/* The size of an object without metadata */
> -       struct reciprocal_value reciprocal_size;
> -       unsigned int offset;    /* Free pointer offset */
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       /* Number of per cpu partial objects to keep around */
> -       unsigned int cpu_partial;
> -       /* Number of per cpu partial slabs to keep around */
> -       unsigned int cpu_partial_slabs;
> -#endif
> -       struct kmem_cache_order_objects oo;
> -
> -       /* Allocation and freeing of slabs */
> -       struct kmem_cache_order_objects min;
> -       gfp_t allocflags;       /* gfp flags to use on each alloc */
> -       int refcount;           /* Refcount for slab cache destroy */
> -       void (*ctor)(void *);
> -       unsigned int inuse;             /* Offset to metadata */
> -       unsigned int align;             /* Alignment */
> -       unsigned int red_left_pad;      /* Left redzone padding size */
> -       const char *name;       /* Name (only for display!) */
> -       struct list_head list;  /* List of slab caches */
> -#ifdef CONFIG_SYSFS
> -       struct kobject kobj;    /* For sysfs */
> -#endif
> -#ifdef CONFIG_SLAB_FREELIST_HARDENED
> -       unsigned long random;
> -#endif
> -
> -#ifdef CONFIG_NUMA
> -       /*
> -        * Defragmentation by allocating from a remote node.
> -        */
> -       unsigned int remote_node_defrag_ratio;
> -#endif
> -
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -       unsigned int *random_seq;
> -#endif
> -
> -#ifdef CONFIG_KASAN_GENERIC
> -       struct kasan_cache kasan_info;
> -#endif
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -       unsigned int useroffset;        /* Usercopy region offset */
> -       unsigned int usersize;          /* Usercopy region size */
> -#endif
> -
> -       struct kmem_cache_node *node[MAX_NUMNODES];
> -};
> -
> -#if defined(CONFIG_SYSFS) && !defined(CONFIG_SLUB_TINY)
> -#define SLAB_SUPPORTS_SYSFS
> -void sysfs_slab_unlink(struct kmem_cache *);
> -void sysfs_slab_release(struct kmem_cache *);
> -#else
> -static inline void sysfs_slab_unlink(struct kmem_cache *s)
> -{
> -}
> -static inline void sysfs_slab_release(struct kmem_cache *s)
> -{
> -}
> -#endif
> -
> -void *fixup_red_left(struct kmem_cache *s, void *p);
> -
> -static inline void *nearest_obj(struct kmem_cache *cache, const struct slab *slab,
> -                               void *x) {
> -       void *object = x - (x - slab_address(slab)) % cache->size;
> -       void *last_object = slab_address(slab) +
> -               (slab->objects - 1) * cache->size;
> -       void *result = (unlikely(object > last_object)) ? last_object : object;
> -
> -       result = fixup_red_left(cache, result);
> -       return result;
> -}
> -
> -/* Determine object index from a given position */
> -static inline unsigned int __obj_to_index(const struct kmem_cache *cache,
> -                                         void *addr, void *obj)
> -{
> -       return reciprocal_divide(kasan_reset_tag(obj) - addr,
> -                                cache->reciprocal_size);
> -}
> -
> -static inline unsigned int obj_to_index(const struct kmem_cache *cache,
> -                                       const struct slab *slab, void *obj)
> -{
> -       if (is_kfence_address(obj))
> -               return 0;
> -       return __obj_to_index(cache, slab_address(slab), obj);
> -}
> -
> -static inline int objs_per_slab(const struct kmem_cache *cache,
> -                                    const struct slab *slab)
> -{
> -       return slab->objects;
> -}
> -#endif /* _LINUX_SLUB_DEF_H */
> diff --git a/init/Kconfig b/init/Kconfig
> index 1fb5f313d18f..45be2eedf75c 100644
> --- a/init/Kconfig
> +++ b/init/Kconfig
> @@ -973,7 +973,7 @@ config MEMCG
>
>  config MEMCG_KMEM
>         bool
> -       depends on MEMCG && !SLOB
> +       depends on MEMCG && BROKEN
>         default y
>
>  config BLK_CGROUP
> diff --git a/mm/Kconfig b/mm/Kconfig
> index 4751031f3f05..f07e81bca39e 100644
> --- a/mm/Kconfig
> +++ b/mm/Kconfig
> @@ -210,134 +210,9 @@ config ZSMALLOC_CHAIN_SIZE
>
>           For more information, see zsmalloc documentation.
>
> -menu "SLAB allocator options"
> -
> -choice
> -       prompt "Choose SLAB allocator"
> -       default SLUB
> -       help
> -          This option allows to select a slab allocator.
> -
> -config SLAB
> -       bool "SLAB"
> -       depends on !PREEMPT_RT
> -       select HAVE_HARDENED_USERCOPY_ALLOCATOR
> -       help
> -         The regular slab allocator that is established and known to work
> -         well in all environments. It organizes cache hot objects in
> -         per cpu and per node queues.
> -
> -config SLUB
> -       bool "SLUB (Unqueued Allocator)"
> -       select HAVE_HARDENED_USERCOPY_ALLOCATOR
> -       help
> -          SLUB is a slab allocator that minimizes cache line usage
> -          instead of managing queues of cached objects (SLAB approach).
> -          Per cpu caching is realized using slabs of objects instead
> -          of queues of objects. SLUB can use memory efficiently
> -          and has enhanced diagnostics. SLUB is the default choice for
> -          a slab allocator.
> -
> -config SLOB_DEPRECATED
> -       depends on EXPERT
> -       bool "SLOB (Simple Allocator - DEPRECATED)"
> -       depends on !PREEMPT_RT
> -       help
> -          Deprecated and scheduled for removal in a few cycles. SLUB
> -          recommended as replacement. CONFIG_SLUB_TINY can be considered
> -          on systems with 16MB or less RAM.
> -
> -          If you need SLOB to stay, please contact linux-mm@...ck.org and
> -          people listed in the SLAB ALLOCATOR section of MAINTAINERS file,
> -          with your use case.
> -
> -          SLOB replaces the stock allocator with a drastically simpler
> -          allocator. SLOB is generally more space efficient but
> -          does not perform as well on large systems.
> -
> -endchoice
> -
> -config SLOB
> -       bool
> -       default y
> -       depends on SLOB_DEPRECATED
> -
> -config SLUB_TINY
> -       bool "Configure SLUB for minimal memory footprint"
> -       depends on SLUB && EXPERT
> -       select SLAB_MERGE_DEFAULT
> -       help
> -          Configures the SLUB allocator in a way to achieve minimal memory
> -          footprint, sacrificing scalability, debugging and other features.
> -          This is intended only for the smallest system that had used the
> -          SLOB allocator and is not recommended for systems with more than
> -          16MB RAM.
> -
> -          If unsure, say N.
> -
> -config SLAB_MERGE_DEFAULT
> -       bool "Allow slab caches to be merged"
> -       default y
> -       depends on SLAB || SLUB
> -       help
> -         For reduced kernel memory fragmentation, slab caches can be
> -         merged when they share the same size and other characteristics.
> -         This carries a risk of kernel heap overflows being able to
> -         overwrite objects from merged caches (and more easily control
> -         cache layout), which makes such heap attacks easier to exploit
> -         by attackers. By keeping caches unmerged, these kinds of exploits
> -         can usually only damage objects in the same cache. To disable
> -         merging at runtime, "slab_nomerge" can be passed on the kernel
> -         command line.
> -
> -config SLAB_FREELIST_RANDOM
> -       bool "Randomize slab freelist"
> -       depends on SLAB || (SLUB && !SLUB_TINY)
> -       help
> -         Randomizes the freelist order used on creating new pages. This
> -         security feature reduces the predictability of the kernel slab
> -         allocator against heap overflows.
> -
> -config SLAB_FREELIST_HARDENED
> -       bool "Harden slab freelist metadata"
> -       depends on SLAB || (SLUB && !SLUB_TINY)
> -       help
> -         Many kernel heap attacks try to target slab cache metadata and
> -         other infrastructure. This options makes minor performance
> -         sacrifices to harden the kernel slab allocator against common
> -         freelist exploit methods. Some slab implementations have more
> -         sanity-checking than others. This option is most effective with
> -         CONFIG_SLUB.
> -
> -config SLUB_STATS
> -       default n
> -       bool "Enable SLUB performance statistics"
> -       depends on SLUB && SYSFS && !SLUB_TINY
> -       help
> -         SLUB statistics are useful to debug SLUBs allocation behavior in
> -         order find ways to optimize the allocator. This should never be
> -         enabled for production use since keeping statistics slows down
> -         the allocator by a few percentage points. The slabinfo command
> -         supports the determination of the most active slabs to figure
> -         out which slabs are relevant to a particular load.
> -         Try running: slabinfo -DA
> -
> -config SLUB_CPU_PARTIAL
> -       default y
> -       depends on SLUB && SMP && !SLUB_TINY
> -       bool "SLUB per cpu partial cache"
> -       help
> -         Per cpu partial caches accelerate objects allocation and freeing
> -         that is local to a processor at the price of more indeterminism
> -         in the latency of the free. On overflow these caches will be cleared
> -         which requires the taking of locks that may cause latency spikes.
> -         Typically one would choose no for a realtime system.
> -
> -endmenu # SLAB allocator options
> -
>  config SHUFFLE_PAGE_ALLOCATOR
>         bool "Page allocator randomization"
> -       default SLAB_FREELIST_RANDOM && ACPI_NUMA
> +       default ACPI_NUMA
>         help
>           Randomization of the page allocator improves the average
>           utilization of a direct-mapped memory-side-cache. See section
> @@ -345,10 +220,9 @@ config SHUFFLE_PAGE_ALLOCATOR
>           6.2a specification for an example of how a platform advertises
>           the presence of a memory-side-cache. There are also incidental
>           security benefits as it reduces the predictability of page
> -         allocations to compliment SLAB_FREELIST_RANDOM, but the
> -         default granularity of shuffling on the "MAX_ORDER - 1" i.e,
> -         10th order of pages is selected based on cache utilization
> -         benefits on x86.
> +         allocations, but the default granularity of shuffling on the
> +         "MAX_ORDER - 1" i.e, 10th order of pages is selected based on
> +         cache utilization benefits on x86.
>
>           While the randomization improves cache utilization it may
>           negatively impact workloads on platforms without a cache. For
> diff --git a/mm/Makefile b/mm/Makefile
> index 8e105e5b3e29..18b0bb245fc3 100644
> --- a/mm/Makefile
> +++ b/mm/Makefile
> @@ -4,16 +4,12 @@
>  #
>
>  KASAN_SANITIZE_slab_common.o := n
> -KASAN_SANITIZE_slab.o := n
> -KASAN_SANITIZE_slub.o := n
>  KCSAN_SANITIZE_kmemleak.o := n
>
>  # These produce frequent data race reports: most of them are due to races on
>  # the same word but accesses to different bits of that word. Re-enable KCSAN
>  # for these when we have more consensus on what to do about them.
>  KCSAN_SANITIZE_slab_common.o := n
> -KCSAN_SANITIZE_slab.o := n
> -KCSAN_SANITIZE_slub.o := n
>  KCSAN_SANITIZE_page_alloc.o := n
>  # But enable explicit instrumentation for memory barriers.
>  KCSAN_INSTRUMENT_BARRIERS := y
> @@ -22,9 +18,6 @@ KCSAN_INSTRUMENT_BARRIERS := y
>  # flaky coverage that is not a function of syscall inputs. E.g. slab is out of
>  # free pages, or a task is migrated between nodes.
>  KCOV_INSTRUMENT_slab_common.o := n
> -KCOV_INSTRUMENT_slob.o := n
> -KCOV_INSTRUMENT_slab.o := n
> -KCOV_INSTRUMENT_slub.o := n
>  KCOV_INSTRUMENT_page_alloc.o := n
>  KCOV_INSTRUMENT_debug-pagealloc.o := n
>  KCOV_INSTRUMENT_kmemleak.o := n
> @@ -81,12 +74,9 @@ obj-$(CONFIG_HUGETLB_PAGE_OPTIMIZE_VMEMMAP)  += hugetlb_vmemmap.o
>  obj-$(CONFIG_NUMA)     += mempolicy.o
>  obj-$(CONFIG_SPARSEMEM)        += sparse.o
>  obj-$(CONFIG_SPARSEMEM_VMEMMAP) += sparse-vmemmap.o
> -obj-$(CONFIG_SLOB) += slob.o
>  obj-$(CONFIG_MMU_NOTIFIER) += mmu_notifier.o
>  obj-$(CONFIG_KSM) += ksm.o
>  obj-$(CONFIG_PAGE_POISONING) += page_poison.o
> -obj-$(CONFIG_SLAB) += slab.o
> -obj-$(CONFIG_SLUB) += slub.o
>  obj-$(CONFIG_KASAN)    += kasan/
>  obj-$(CONFIG_KFENCE) += kfence/
>  obj-$(CONFIG_KMSAN)    += kmsan/
> diff --git a/mm/slab.c b/mm/slab.c
> deleted file mode 100644
> index edbe722fb906..000000000000
> --- a/mm/slab.c
> +++ /dev/null
> @@ -1,4046 +0,0 @@
> -// SPDX-License-Identifier: GPL-2.0
> -/*
> - * linux/mm/slab.c
> - * Written by Mark Hemment, 1996/97.
> - * (markhe@...td.demon.co.uk)
> - *
> - * kmem_cache_destroy() + some cleanup - 1999 Andrea Arcangeli
> - *
> - * Major cleanup, different bufctl logic, per-cpu arrays
> - *     (c) 2000 Manfred Spraul
> - *
> - * Cleanup, make the head arrays unconditional, preparation for NUMA
> - *     (c) 2002 Manfred Spraul
> - *
> - * An implementation of the Slab Allocator as described in outline in;
> - *     UNIX Internals: The New Frontiers by Uresh Vahalia
> - *     Pub: Prentice Hall      ISBN 0-13-101908-2
> - * or with a little more detail in;
> - *     The Slab Allocator: An Object-Caching Kernel Memory Allocator
> - *     Jeff Bonwick (Sun Microsystems).
> - *     Presented at: USENIX Summer 1994 Technical Conference
> - *
> - * The memory is organized in caches, one cache for each object type.
> - * (e.g. inode_cache, dentry_cache, buffer_head, vm_area_struct)
> - * Each cache consists out of many slabs (they are small (usually one
> - * page long) and always contiguous), and each slab contains multiple
> - * initialized objects.
> - *
> - * This means, that your constructor is used only for newly allocated
> - * slabs and you must pass objects with the same initializations to
> - * kmem_cache_free.
> - *
> - * Each cache can only support one memory type (GFP_DMA, GFP_HIGHMEM,
> - * normal). If you need a special memory type, then must create a new
> - * cache for that memory type.
> - *
> - * In order to reduce fragmentation, the slabs are sorted in 3 groups:
> - *   full slabs with 0 free objects
> - *   partial slabs
> - *   empty slabs with no allocated objects
> - *
> - * If partial slabs exist, then new allocations come from these slabs,
> - * otherwise from empty slabs or new slabs are allocated.
> - *
> - * kmem_cache_destroy() CAN CRASH if you try to allocate from the cache
> - * during kmem_cache_destroy(). The caller must prevent concurrent allocs.
> - *
> - * Each cache has a short per-cpu head array, most allocs
> - * and frees go into that array, and if that array overflows, then 1/2
> - * of the entries in the array are given back into the global cache.
> - * The head array is strictly LIFO and should improve the cache hit rates.
> - * On SMP, it additionally reduces the spinlock operations.
> - *
> - * The c_cpuarray may not be read with enabled local interrupts -
> - * it's changed with a smp_call_function().
> - *
> - * SMP synchronization:
> - *  constructors and destructors are called without any locking.
> - *  Several members in struct kmem_cache and struct slab never change, they
> - *     are accessed without any locking.
> - *  The per-cpu arrays are never accessed from the wrong cpu, no locking,
> - *     and local interrupts are disabled so slab code is preempt-safe.
> - *  The non-constant members are protected with a per-cache irq spinlock.
> - *
> - * Many thanks to Mark Hemment, who wrote another per-cpu slab patch
> - * in 2000 - many ideas in the current implementation are derived from
> - * his patch.
> - *
> - * Further notes from the original documentation:
> - *
> - * 11 April '97.  Started multi-threading - markhe
> - *     The global cache-chain is protected by the mutex 'slab_mutex'.
> - *     The sem is only needed when accessing/extending the cache-chain, which
> - *     can never happen inside an interrupt (kmem_cache_create(),
> - *     kmem_cache_shrink() and kmem_cache_reap()).
> - *
> - *     At present, each engine can be growing a cache.  This should be blocked.
> - *
> - * 15 March 2005. NUMA slab allocator.
> - *     Shai Fultheim <shai@...lex86.org>.
> - *     Shobhit Dayal <shobhit@...softinc.com>
> - *     Alok N Kataria <alokk@...softinc.com>
> - *     Christoph Lameter <christoph@...eter.com>
> - *
> - *     Modified the slab allocator to be node aware on NUMA systems.
> - *     Each node has its own list of partial, free and full slabs.
> - *     All object allocations for a node occur from node specific slab lists.
> - */
> -
> -#include       <linux/slab.h>
> -#include       <linux/mm.h>
> -#include       <linux/poison.h>
> -#include       <linux/swap.h>
> -#include       <linux/cache.h>
> -#include       <linux/interrupt.h>
> -#include       <linux/init.h>
> -#include       <linux/compiler.h>
> -#include       <linux/cpuset.h>
> -#include       <linux/proc_fs.h>
> -#include       <linux/seq_file.h>
> -#include       <linux/notifier.h>
> -#include       <linux/kallsyms.h>
> -#include       <linux/kfence.h>
> -#include       <linux/cpu.h>
> -#include       <linux/sysctl.h>
> -#include       <linux/module.h>
> -#include       <linux/rcupdate.h>
> -#include       <linux/string.h>
> -#include       <linux/uaccess.h>
> -#include       <linux/nodemask.h>
> -#include       <linux/kmemleak.h>
> -#include       <linux/mempolicy.h>
> -#include       <linux/mutex.h>
> -#include       <linux/fault-inject.h>
> -#include       <linux/rtmutex.h>
> -#include       <linux/reciprocal_div.h>
> -#include       <linux/debugobjects.h>
> -#include       <linux/memory.h>
> -#include       <linux/prefetch.h>
> -#include       <linux/sched/task_stack.h>
> -
> -#include       <net/sock.h>
> -
> -#include       <asm/cacheflush.h>
> -#include       <asm/tlbflush.h>
> -#include       <asm/page.h>
> -
> -#include <trace/events/kmem.h>
> -
> -#include       "internal.h"
> -
> -#include       "slab.h"
> -
> -/*
> - * DEBUG       - 1 for kmem_cache_create() to honour; SLAB_RED_ZONE & SLAB_POISON.
> - *               0 for faster, smaller code (especially in the critical paths).
> - *
> - * STATS       - 1 to collect stats for /proc/slabinfo.
> - *               0 for faster, smaller code (especially in the critical paths).
> - *
> - * FORCED_DEBUG        - 1 enables SLAB_RED_ZONE and SLAB_POISON (if possible)
> - */
> -
> -#ifdef CONFIG_DEBUG_SLAB
> -#define        DEBUG           1
> -#define        STATS           1
> -#define        FORCED_DEBUG    1
> -#else
> -#define        DEBUG           0
> -#define        STATS           0
> -#define        FORCED_DEBUG    0
> -#endif
> -
> -/* Shouldn't this be in a header file somewhere? */
> -#define        BYTES_PER_WORD          sizeof(void *)
> -#define        REDZONE_ALIGN           max(BYTES_PER_WORD, __alignof__(unsigned long long))
> -
> -#ifndef ARCH_KMALLOC_FLAGS
> -#define ARCH_KMALLOC_FLAGS SLAB_HWCACHE_ALIGN
> -#endif
> -
> -#define FREELIST_BYTE_INDEX (((PAGE_SIZE >> BITS_PER_BYTE) \
> -                               <= SLAB_OBJ_MIN_SIZE) ? 1 : 0)
> -
> -#if FREELIST_BYTE_INDEX
> -typedef unsigned char freelist_idx_t;
> -#else
> -typedef unsigned short freelist_idx_t;
> -#endif
> -
> -#define SLAB_OBJ_MAX_NUM ((1 << sizeof(freelist_idx_t) * BITS_PER_BYTE) - 1)
> -
> -/*
> - * struct array_cache
> - *
> - * Purpose:
> - * - LIFO ordering, to hand out cache-warm objects from _alloc
> - * - reduce the number of linked list operations
> - * - reduce spinlock operations
> - *
> - * The limit is stored in the per-cpu structure to reduce the data cache
> - * footprint.
> - *
> - */
> -struct array_cache {
> -       unsigned int avail;
> -       unsigned int limit;
> -       unsigned int batchcount;
> -       unsigned int touched;
> -       void *entry[];  /*
> -                        * Must have this definition in here for the proper
> -                        * alignment of array_cache. Also simplifies accessing
> -                        * the entries.
> -                        */
> -};
> -
> -struct alien_cache {
> -       spinlock_t lock;
> -       struct array_cache ac;
> -};
> -
> -/*
> - * Need this for bootstrapping a per node allocator.
> - */
> -#define NUM_INIT_LISTS (2 * MAX_NUMNODES)
> -static struct kmem_cache_node __initdata init_kmem_cache_node[NUM_INIT_LISTS];
> -#define        CACHE_CACHE 0
> -#define        SIZE_NODE (MAX_NUMNODES)
> -
> -static int drain_freelist(struct kmem_cache *cache,
> -                       struct kmem_cache_node *n, int tofree);
> -static void free_block(struct kmem_cache *cachep, void **objpp, int len,
> -                       int node, struct list_head *list);
> -static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list);
> -static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp);
> -static void cache_reap(struct work_struct *unused);
> -
> -static inline void fixup_objfreelist_debug(struct kmem_cache *cachep,
> -                                               void **list);
> -static inline void fixup_slab_list(struct kmem_cache *cachep,
> -                               struct kmem_cache_node *n, struct slab *slab,
> -                               void **list);
> -
> -#define INDEX_NODE kmalloc_index(sizeof(struct kmem_cache_node))
> -
> -static void kmem_cache_node_init(struct kmem_cache_node *parent)
> -{
> -       INIT_LIST_HEAD(&parent->slabs_full);
> -       INIT_LIST_HEAD(&parent->slabs_partial);
> -       INIT_LIST_HEAD(&parent->slabs_free);
> -       parent->total_slabs = 0;
> -       parent->free_slabs = 0;
> -       parent->shared = NULL;
> -       parent->alien = NULL;
> -       parent->colour_next = 0;
> -       raw_spin_lock_init(&parent->list_lock);
> -       parent->free_objects = 0;
> -       parent->free_touched = 0;
> -}
> -
> -#define MAKE_LIST(cachep, listp, slab, nodeid)                         \
> -       do {                                                            \
> -               INIT_LIST_HEAD(listp);                                  \
> -               list_splice(&get_node(cachep, nodeid)->slab, listp);    \
> -       } while (0)
> -
> -#define        MAKE_ALL_LISTS(cachep, ptr, nodeid)                             \
> -       do {                                                            \
> -       MAKE_LIST((cachep), (&(ptr)->slabs_full), slabs_full, nodeid);  \
> -       MAKE_LIST((cachep), (&(ptr)->slabs_partial), slabs_partial, nodeid); \
> -       MAKE_LIST((cachep), (&(ptr)->slabs_free), slabs_free, nodeid);  \
> -       } while (0)
> -
> -#define CFLGS_OBJFREELIST_SLAB ((slab_flags_t __force)0x40000000U)
> -#define CFLGS_OFF_SLAB         ((slab_flags_t __force)0x80000000U)
> -#define        OBJFREELIST_SLAB(x)     ((x)->flags & CFLGS_OBJFREELIST_SLAB)
> -#define        OFF_SLAB(x)     ((x)->flags & CFLGS_OFF_SLAB)
> -
> -#define BATCHREFILL_LIMIT      16
> -/*
> - * Optimization question: fewer reaps means less probability for unnecessary
> - * cpucache drain/refill cycles.
> - *
> - * OTOH the cpuarrays can contain lots of objects,
> - * which could lock up otherwise freeable slabs.
> - */
> -#define REAPTIMEOUT_AC         (2*HZ)
> -#define REAPTIMEOUT_NODE       (4*HZ)
> -
> -#if STATS
> -#define        STATS_INC_ACTIVE(x)     ((x)->num_active++)
> -#define        STATS_DEC_ACTIVE(x)     ((x)->num_active--)
> -#define        STATS_INC_ALLOCED(x)    ((x)->num_allocations++)
> -#define        STATS_INC_GROWN(x)      ((x)->grown++)
> -#define        STATS_ADD_REAPED(x, y)  ((x)->reaped += (y))
> -#define        STATS_SET_HIGH(x)                                               \
> -       do {                                                            \
> -               if ((x)->num_active > (x)->high_mark)                   \
> -                       (x)->high_mark = (x)->num_active;               \
> -       } while (0)
> -#define        STATS_INC_ERR(x)        ((x)->errors++)
> -#define        STATS_INC_NODEALLOCS(x) ((x)->node_allocs++)
> -#define        STATS_INC_NODEFREES(x)  ((x)->node_frees++)
> -#define STATS_INC_ACOVERFLOW(x)   ((x)->node_overflow++)
> -#define        STATS_SET_FREEABLE(x, i)                                        \
> -       do {                                                            \
> -               if ((x)->max_freeable < i)                              \
> -                       (x)->max_freeable = i;                          \
> -       } while (0)
> -#define STATS_INC_ALLOCHIT(x)  atomic_inc(&(x)->allochit)
> -#define STATS_INC_ALLOCMISS(x) atomic_inc(&(x)->allocmiss)
> -#define STATS_INC_FREEHIT(x)   atomic_inc(&(x)->freehit)
> -#define STATS_INC_FREEMISS(x)  atomic_inc(&(x)->freemiss)
> -#else
> -#define        STATS_INC_ACTIVE(x)     do { } while (0)
> -#define        STATS_DEC_ACTIVE(x)     do { } while (0)
> -#define        STATS_INC_ALLOCED(x)    do { } while (0)
> -#define        STATS_INC_GROWN(x)      do { } while (0)
> -#define        STATS_ADD_REAPED(x, y)  do { (void)(y); } while (0)
> -#define        STATS_SET_HIGH(x)       do { } while (0)
> -#define        STATS_INC_ERR(x)        do { } while (0)
> -#define        STATS_INC_NODEALLOCS(x) do { } while (0)
> -#define        STATS_INC_NODEFREES(x)  do { } while (0)
> -#define STATS_INC_ACOVERFLOW(x)   do { } while (0)
> -#define        STATS_SET_FREEABLE(x, i) do { } while (0)
> -#define STATS_INC_ALLOCHIT(x)  do { } while (0)
> -#define STATS_INC_ALLOCMISS(x) do { } while (0)
> -#define STATS_INC_FREEHIT(x)   do { } while (0)
> -#define STATS_INC_FREEMISS(x)  do { } while (0)
> -#endif
> -
> -#if DEBUG
> -
> -/*
> - * memory layout of objects:
> - * 0           : objp
> - * 0 .. cachep->obj_offset - BYTES_PER_WORD - 1: padding. This ensures that
> - *             the end of an object is aligned with the end of the real
> - *             allocation. Catches writes behind the end of the allocation.
> - * cachep->obj_offset - BYTES_PER_WORD .. cachep->obj_offset - 1:
> - *             redzone word.
> - * cachep->obj_offset: The real object.
> - * cachep->size - 2* BYTES_PER_WORD: redzone word [BYTES_PER_WORD long]
> - * cachep->size - 1* BYTES_PER_WORD: last caller address
> - *                                     [BYTES_PER_WORD long]
> - */
> -static int obj_offset(struct kmem_cache *cachep)
> -{
> -       return cachep->obj_offset;
> -}
> -
> -static unsigned long long *dbg_redzone1(struct kmem_cache *cachep, void *objp)
> -{
> -       BUG_ON(!(cachep->flags & SLAB_RED_ZONE));
> -       return (unsigned long long *) (objp + obj_offset(cachep) -
> -                                     sizeof(unsigned long long));
> -}
> -
> -static unsigned long long *dbg_redzone2(struct kmem_cache *cachep, void *objp)
> -{
> -       BUG_ON(!(cachep->flags & SLAB_RED_ZONE));
> -       if (cachep->flags & SLAB_STORE_USER)
> -               return (unsigned long long *)(objp + cachep->size -
> -                                             sizeof(unsigned long long) -
> -                                             REDZONE_ALIGN);
> -       return (unsigned long long *) (objp + cachep->size -
> -                                      sizeof(unsigned long long));
> -}
> -
> -static void **dbg_userword(struct kmem_cache *cachep, void *objp)
> -{
> -       BUG_ON(!(cachep->flags & SLAB_STORE_USER));
> -       return (void **)(objp + cachep->size - BYTES_PER_WORD);
> -}
> -
> -#else
> -
> -#define obj_offset(x)                  0
> -#define dbg_redzone1(cachep, objp)     ({BUG(); (unsigned long long *)NULL;})
> -#define dbg_redzone2(cachep, objp)     ({BUG(); (unsigned long long *)NULL;})
> -#define dbg_userword(cachep, objp)     ({BUG(); (void **)NULL;})
> -
> -#endif
> -
> -/*
> - * Do not go above this order unless 0 objects fit into the slab or
> - * overridden on the command line.
> - */
> -#define        SLAB_MAX_ORDER_HI       1
> -#define        SLAB_MAX_ORDER_LO       0
> -static int slab_max_order = SLAB_MAX_ORDER_LO;
> -static bool slab_max_order_set __initdata;
> -
> -static inline void *index_to_obj(struct kmem_cache *cache,
> -                                const struct slab *slab, unsigned int idx)
> -{
> -       return slab->s_mem + cache->size * idx;
> -}
> -
> -#define BOOT_CPUCACHE_ENTRIES  1
> -/* internal cache of cache description objs */
> -static struct kmem_cache kmem_cache_boot = {
> -       .batchcount = 1,
> -       .limit = BOOT_CPUCACHE_ENTRIES,
> -       .shared = 1,
> -       .size = sizeof(struct kmem_cache),
> -       .name = "kmem_cache",
> -};
> -
> -static DEFINE_PER_CPU(struct delayed_work, slab_reap_work);
> -
> -static inline struct array_cache *cpu_cache_get(struct kmem_cache *cachep)
> -{
> -       return this_cpu_ptr(cachep->cpu_cache);
> -}
> -
> -/*
> - * Calculate the number of objects and left-over bytes for a given buffer size.
> - */
> -static unsigned int cache_estimate(unsigned long gfporder, size_t buffer_size,
> -               slab_flags_t flags, size_t *left_over)
> -{
> -       unsigned int num;
> -       size_t slab_size = PAGE_SIZE << gfporder;
> -
> -       /*
> -        * The slab management structure can be either off the slab or
> -        * on it. For the latter case, the memory allocated for a
> -        * slab is used for:
> -        *
> -        * - @buffer_size bytes for each object
> -        * - One freelist_idx_t for each object
> -        *
> -        * We don't need to consider alignment of freelist because
> -        * freelist will be at the end of slab page. The objects will be
> -        * at the correct alignment.
> -        *
> -        * If the slab management structure is off the slab, then the
> -        * alignment will already be calculated into the size. Because
> -        * the slabs are all pages aligned, the objects will be at the
> -        * correct alignment when allocated.
> -        */
> -       if (flags & (CFLGS_OBJFREELIST_SLAB | CFLGS_OFF_SLAB)) {
> -               num = slab_size / buffer_size;
> -               *left_over = slab_size % buffer_size;
> -       } else {
> -               num = slab_size / (buffer_size + sizeof(freelist_idx_t));
> -               *left_over = slab_size %
> -                       (buffer_size + sizeof(freelist_idx_t));
> -       }
> -
> -       return num;
> -}
> -
> -#if DEBUG
> -#define slab_error(cachep, msg) __slab_error(__func__, cachep, msg)
> -
> -static void __slab_error(const char *function, struct kmem_cache *cachep,
> -                       char *msg)
> -{
> -       pr_err("slab error in %s(): cache `%s': %s\n",
> -              function, cachep->name, msg);
> -       dump_stack();
> -       add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
> -}
> -#endif
> -
> -/*
> - * By default on NUMA we use alien caches to stage the freeing of
> - * objects allocated from other nodes. This causes massive memory
> - * inefficiencies when using fake NUMA setup to split memory into a
> - * large number of small nodes, so it can be disabled on the command
> - * line
> -  */
> -
> -static int use_alien_caches __read_mostly = 1;
> -static int __init noaliencache_setup(char *s)
> -{
> -       use_alien_caches = 0;
> -       return 1;
> -}
> -__setup("noaliencache", noaliencache_setup);
> -
> -static int __init slab_max_order_setup(char *str)
> -{
> -       get_option(&str, &slab_max_order);
> -       slab_max_order = slab_max_order < 0 ? 0 :
> -                               min(slab_max_order, MAX_ORDER - 1);
> -       slab_max_order_set = true;
> -
> -       return 1;
> -}
> -__setup("slab_max_order=", slab_max_order_setup);
> -
> -#ifdef CONFIG_NUMA
> -/*
> - * Special reaping functions for NUMA systems called from cache_reap().
> - * These take care of doing round robin flushing of alien caches (containing
> - * objects freed on different nodes from which they were allocated) and the
> - * flushing of remote pcps by calling drain_node_pages.
> - */
> -static DEFINE_PER_CPU(unsigned long, slab_reap_node);
> -
> -static void init_reap_node(int cpu)
> -{
> -       per_cpu(slab_reap_node, cpu) = next_node_in(cpu_to_mem(cpu),
> -                                                   node_online_map);
> -}
> -
> -static void next_reap_node(void)
> -{
> -       int node = __this_cpu_read(slab_reap_node);
> -
> -       node = next_node_in(node, node_online_map);
> -       __this_cpu_write(slab_reap_node, node);
> -}
> -
> -#else
> -#define init_reap_node(cpu) do { } while (0)
> -#define next_reap_node(void) do { } while (0)
> -#endif
> -
> -/*
> - * Initiate the reap timer running on the target CPU.  We run at around 1 to 2Hz
> - * via the workqueue/eventd.
> - * Add the CPU number into the expiration time to minimize the possibility of
> - * the CPUs getting into lockstep and contending for the global cache chain
> - * lock.
> - */
> -static void start_cpu_timer(int cpu)
> -{
> -       struct delayed_work *reap_work = &per_cpu(slab_reap_work, cpu);
> -
> -       if (reap_work->work.func == NULL) {
> -               init_reap_node(cpu);
> -               INIT_DEFERRABLE_WORK(reap_work, cache_reap);
> -               schedule_delayed_work_on(cpu, reap_work,
> -                                       __round_jiffies_relative(HZ, cpu));
> -       }
> -}
> -
> -static void init_arraycache(struct array_cache *ac, int limit, int batch)
> -{
> -       if (ac) {
> -               ac->avail = 0;
> -               ac->limit = limit;
> -               ac->batchcount = batch;
> -               ac->touched = 0;
> -       }
> -}
> -
> -static struct array_cache *alloc_arraycache(int node, int entries,
> -                                           int batchcount, gfp_t gfp)
> -{
> -       size_t memsize = sizeof(void *) * entries + sizeof(struct array_cache);
> -       struct array_cache *ac = NULL;
> -
> -       ac = kmalloc_node(memsize, gfp, node);
> -       /*
> -        * The array_cache structures contain pointers to free object.
> -        * However, when such objects are allocated or transferred to another
> -        * cache the pointers are not cleared and they could be counted as
> -        * valid references during a kmemleak scan. Therefore, kmemleak must
> -        * not scan such objects.
> -        */
> -       kmemleak_no_scan(ac);
> -       init_arraycache(ac, entries, batchcount);
> -       return ac;
> -}
> -
> -static noinline void cache_free_pfmemalloc(struct kmem_cache *cachep,
> -                                       struct slab *slab, void *objp)
> -{
> -       struct kmem_cache_node *n;
> -       int slab_node;
> -       LIST_HEAD(list);
> -
> -       slab_node = slab_nid(slab);
> -       n = get_node(cachep, slab_node);
> -
> -       raw_spin_lock(&n->list_lock);
> -       free_block(cachep, &objp, 1, slab_node, &list);
> -       raw_spin_unlock(&n->list_lock);
> -
> -       slabs_destroy(cachep, &list);
> -}
> -
> -/*
> - * Transfer objects in one arraycache to another.
> - * Locking must be handled by the caller.
> - *
> - * Return the number of entries transferred.
> - */
> -static int transfer_objects(struct array_cache *to,
> -               struct array_cache *from, unsigned int max)
> -{
> -       /* Figure out how many entries to transfer */
> -       int nr = min3(from->avail, max, to->limit - to->avail);
> -
> -       if (!nr)
> -               return 0;
> -
> -       memcpy(to->entry + to->avail, from->entry + from->avail - nr,
> -                       sizeof(void *) *nr);
> -
> -       from->avail -= nr;
> -       to->avail += nr;
> -       return nr;
> -}
> -
> -/* &alien->lock must be held by alien callers. */
> -static __always_inline void __free_one(struct array_cache *ac, void *objp)
> -{
> -       /* Avoid trivial double-free. */
> -       if (IS_ENABLED(CONFIG_SLAB_FREELIST_HARDENED) &&
> -           WARN_ON_ONCE(ac->avail > 0 && ac->entry[ac->avail - 1] == objp))
> -               return;
> -       ac->entry[ac->avail++] = objp;
> -}
> -
> -#ifndef CONFIG_NUMA
> -
> -#define drain_alien_cache(cachep, alien) do { } while (0)
> -#define reap_alien(cachep, n) do { } while (0)
> -
> -static inline struct alien_cache **alloc_alien_cache(int node,
> -                                               int limit, gfp_t gfp)
> -{
> -       return NULL;
> -}
> -
> -static inline void free_alien_cache(struct alien_cache **ac_ptr)
> -{
> -}
> -
> -static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
> -{
> -       return 0;
> -}
> -
> -static inline gfp_t gfp_exact_node(gfp_t flags)
> -{
> -       return flags & ~__GFP_NOFAIL;
> -}
> -
> -#else  /* CONFIG_NUMA */
> -
> -static struct alien_cache *__alloc_alien_cache(int node, int entries,
> -                                               int batch, gfp_t gfp)
> -{
> -       size_t memsize = sizeof(void *) * entries + sizeof(struct alien_cache);
> -       struct alien_cache *alc = NULL;
> -
> -       alc = kmalloc_node(memsize, gfp, node);
> -       if (alc) {
> -               kmemleak_no_scan(alc);
> -               init_arraycache(&alc->ac, entries, batch);
> -               spin_lock_init(&alc->lock);
> -       }
> -       return alc;
> -}
> -
> -static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
> -{
> -       struct alien_cache **alc_ptr;
> -       int i;
> -
> -       if (limit > 1)
> -               limit = 12;
> -       alc_ptr = kcalloc_node(nr_node_ids, sizeof(void *), gfp, node);
> -       if (!alc_ptr)
> -               return NULL;
> -
> -       for_each_node(i) {
> -               if (i == node || !node_online(i))
> -                       continue;
> -               alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp);
> -               if (!alc_ptr[i]) {
> -                       for (i--; i >= 0; i--)
> -                               kfree(alc_ptr[i]);
> -                       kfree(alc_ptr);
> -                       return NULL;
> -               }
> -       }
> -       return alc_ptr;
> -}
> -
> -static void free_alien_cache(struct alien_cache **alc_ptr)
> -{
> -       int i;
> -
> -       if (!alc_ptr)
> -               return;
> -       for_each_node(i)
> -           kfree(alc_ptr[i]);
> -       kfree(alc_ptr);
> -}
> -
> -static void __drain_alien_cache(struct kmem_cache *cachep,
> -                               struct array_cache *ac, int node,
> -                               struct list_head *list)
> -{
> -       struct kmem_cache_node *n = get_node(cachep, node);
> -
> -       if (ac->avail) {
> -               raw_spin_lock(&n->list_lock);
> -               /*
> -                * Stuff objects into the remote nodes shared array first.
> -                * That way we could avoid the overhead of putting the objects
> -                * into the free lists and getting them back later.
> -                */
> -               if (n->shared)
> -                       transfer_objects(n->shared, ac, ac->limit);
> -
> -               free_block(cachep, ac->entry, ac->avail, node, list);
> -               ac->avail = 0;
> -               raw_spin_unlock(&n->list_lock);
> -       }
> -}
> -
> -/*
> - * Called from cache_reap() to regularly drain alien caches round robin.
> - */
> -static void reap_alien(struct kmem_cache *cachep, struct kmem_cache_node *n)
> -{
> -       int node = __this_cpu_read(slab_reap_node);
> -
> -       if (n->alien) {
> -               struct alien_cache *alc = n->alien[node];
> -               struct array_cache *ac;
> -
> -               if (alc) {
> -                       ac = &alc->ac;
> -                       if (ac->avail && spin_trylock_irq(&alc->lock)) {
> -                               LIST_HEAD(list);
> -
> -                               __drain_alien_cache(cachep, ac, node, &list);
> -                               spin_unlock_irq(&alc->lock);
> -                               slabs_destroy(cachep, &list);
> -                       }
> -               }
> -       }
> -}
> -
> -static void drain_alien_cache(struct kmem_cache *cachep,
> -                               struct alien_cache **alien)
> -{
> -       int i = 0;
> -       struct alien_cache *alc;
> -       struct array_cache *ac;
> -       unsigned long flags;
> -
> -       for_each_online_node(i) {
> -               alc = alien[i];
> -               if (alc) {
> -                       LIST_HEAD(list);
> -
> -                       ac = &alc->ac;
> -                       spin_lock_irqsave(&alc->lock, flags);
> -                       __drain_alien_cache(cachep, ac, i, &list);
> -                       spin_unlock_irqrestore(&alc->lock, flags);
> -                       slabs_destroy(cachep, &list);
> -               }
> -       }
> -}
> -
> -static int __cache_free_alien(struct kmem_cache *cachep, void *objp,
> -                               int node, int slab_node)
> -{
> -       struct kmem_cache_node *n;
> -       struct alien_cache *alien = NULL;
> -       struct array_cache *ac;
> -       LIST_HEAD(list);
> -
> -       n = get_node(cachep, node);
> -       STATS_INC_NODEFREES(cachep);
> -       if (n->alien && n->alien[slab_node]) {
> -               alien = n->alien[slab_node];
> -               ac = &alien->ac;
> -               spin_lock(&alien->lock);
> -               if (unlikely(ac->avail == ac->limit)) {
> -                       STATS_INC_ACOVERFLOW(cachep);
> -                       __drain_alien_cache(cachep, ac, slab_node, &list);
> -               }
> -               __free_one(ac, objp);
> -               spin_unlock(&alien->lock);
> -               slabs_destroy(cachep, &list);
> -       } else {
> -               n = get_node(cachep, slab_node);
> -               raw_spin_lock(&n->list_lock);
> -               free_block(cachep, &objp, 1, slab_node, &list);
> -               raw_spin_unlock(&n->list_lock);
> -               slabs_destroy(cachep, &list);
> -       }
> -       return 1;
> -}
> -
> -static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
> -{
> -       int slab_node = slab_nid(virt_to_slab(objp));
> -       int node = numa_mem_id();
> -       /*
> -        * Make sure we are not freeing an object from another node to the array
> -        * cache on this cpu.
> -        */
> -       if (likely(node == slab_node))
> -               return 0;
> -
> -       return __cache_free_alien(cachep, objp, node, slab_node);
> -}
> -
> -/*
> - * Construct gfp mask to allocate from a specific node but do not reclaim or
> - * warn about failures.
> - */
> -static inline gfp_t gfp_exact_node(gfp_t flags)
> -{
> -       return (flags | __GFP_THISNODE | __GFP_NOWARN) & ~(__GFP_RECLAIM|__GFP_NOFAIL);
> -}
> -#endif
> -
> -static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)
> -{
> -       struct kmem_cache_node *n;
> -
> -       /*
> -        * Set up the kmem_cache_node for cpu before we can
> -        * begin anything. Make sure some other cpu on this
> -        * node has not already allocated this
> -        */
> -       n = get_node(cachep, node);
> -       if (n) {
> -               raw_spin_lock_irq(&n->list_lock);
> -               n->free_limit = (1 + nr_cpus_node(node)) * cachep->batchcount +
> -                               cachep->num;
> -               raw_spin_unlock_irq(&n->list_lock);
> -
> -               return 0;
> -       }
> -
> -       n = kmalloc_node(sizeof(struct kmem_cache_node), gfp, node);
> -       if (!n)
> -               return -ENOMEM;
> -
> -       kmem_cache_node_init(n);
> -       n->next_reap = jiffies + REAPTIMEOUT_NODE +
> -                   ((unsigned long)cachep) % REAPTIMEOUT_NODE;
> -
> -       n->free_limit =
> -               (1 + nr_cpus_node(node)) * cachep->batchcount + cachep->num;
> -
> -       /*
> -        * The kmem_cache_nodes don't come and go as CPUs
> -        * come and go.  slab_mutex provides sufficient
> -        * protection here.
> -        */
> -       cachep->node[node] = n;
> -
> -       return 0;
> -}
> -
> -#if defined(CONFIG_NUMA) || defined(CONFIG_SMP)
> -/*
> - * Allocates and initializes node for a node on each slab cache, used for
> - * either memory or cpu hotplug.  If memory is being hot-added, the kmem_cache_node
> - * will be allocated off-node since memory is not yet online for the new node.
> - * When hotplugging memory or a cpu, existing nodes are not replaced if
> - * already in use.
> - *
> - * Must hold slab_mutex.
> - */
> -static int init_cache_node_node(int node)
> -{
> -       int ret;
> -       struct kmem_cache *cachep;
> -
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               ret = init_cache_node(cachep, node, GFP_KERNEL);
> -               if (ret)
> -                       return ret;
> -       }
> -
> -       return 0;
> -}
> -#endif
> -
> -static int setup_kmem_cache_node(struct kmem_cache *cachep,
> -                               int node, gfp_t gfp, bool force_change)
> -{
> -       int ret = -ENOMEM;
> -       struct kmem_cache_node *n;
> -       struct array_cache *old_shared = NULL;
> -       struct array_cache *new_shared = NULL;
> -       struct alien_cache **new_alien = NULL;
> -       LIST_HEAD(list);
> -
> -       if (use_alien_caches) {
> -               new_alien = alloc_alien_cache(node, cachep->limit, gfp);
> -               if (!new_alien)
> -                       goto fail;
> -       }
> -
> -       if (cachep->shared) {
> -               new_shared = alloc_arraycache(node,
> -                       cachep->shared * cachep->batchcount, 0xbaadf00d, gfp);
> -               if (!new_shared)
> -                       goto fail;
> -       }
> -
> -       ret = init_cache_node(cachep, node, gfp);
> -       if (ret)
> -               goto fail;
> -
> -       n = get_node(cachep, node);
> -       raw_spin_lock_irq(&n->list_lock);
> -       if (n->shared && force_change) {
> -               free_block(cachep, n->shared->entry,
> -                               n->shared->avail, node, &list);
> -               n->shared->avail = 0;
> -       }
> -
> -       if (!n->shared || force_change) {
> -               old_shared = n->shared;
> -               n->shared = new_shared;
> -               new_shared = NULL;
> -       }
> -
> -       if (!n->alien) {
> -               n->alien = new_alien;
> -               new_alien = NULL;
> -       }
> -
> -       raw_spin_unlock_irq(&n->list_lock);
> -       slabs_destroy(cachep, &list);
> -
> -       /*
> -        * To protect lockless access to n->shared during irq disabled context.
> -        * If n->shared isn't NULL in irq disabled context, accessing to it is
> -        * guaranteed to be valid until irq is re-enabled, because it will be
> -        * freed after synchronize_rcu().
> -        */
> -       if (old_shared && force_change)
> -               synchronize_rcu();
> -
> -fail:
> -       kfree(old_shared);
> -       kfree(new_shared);
> -       free_alien_cache(new_alien);
> -
> -       return ret;
> -}
> -
> -#ifdef CONFIG_SMP
> -
> -static void cpuup_canceled(long cpu)
> -{
> -       struct kmem_cache *cachep;
> -       struct kmem_cache_node *n = NULL;
> -       int node = cpu_to_mem(cpu);
> -       const struct cpumask *mask = cpumask_of_node(node);
> -
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               struct array_cache *nc;
> -               struct array_cache *shared;
> -               struct alien_cache **alien;
> -               LIST_HEAD(list);
> -
> -               n = get_node(cachep, node);
> -               if (!n)
> -                       continue;
> -
> -               raw_spin_lock_irq(&n->list_lock);
> -
> -               /* Free limit for this kmem_cache_node */
> -               n->free_limit -= cachep->batchcount;
> -
> -               /* cpu is dead; no one can alloc from it. */
> -               nc = per_cpu_ptr(cachep->cpu_cache, cpu);
> -               free_block(cachep, nc->entry, nc->avail, node, &list);
> -               nc->avail = 0;
> -
> -               if (!cpumask_empty(mask)) {
> -                       raw_spin_unlock_irq(&n->list_lock);
> -                       goto free_slab;
> -               }
> -
> -               shared = n->shared;
> -               if (shared) {
> -                       free_block(cachep, shared->entry,
> -                                  shared->avail, node, &list);
> -                       n->shared = NULL;
> -               }
> -
> -               alien = n->alien;
> -               n->alien = NULL;
> -
> -               raw_spin_unlock_irq(&n->list_lock);
> -
> -               kfree(shared);
> -               if (alien) {
> -                       drain_alien_cache(cachep, alien);
> -                       free_alien_cache(alien);
> -               }
> -
> -free_slab:
> -               slabs_destroy(cachep, &list);
> -       }
> -       /*
> -        * In the previous loop, all the objects were freed to
> -        * the respective cache's slabs,  now we can go ahead and
> -        * shrink each nodelist to its limit.
> -        */
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               n = get_node(cachep, node);
> -               if (!n)
> -                       continue;
> -               drain_freelist(cachep, n, INT_MAX);
> -       }
> -}
> -
> -static int cpuup_prepare(long cpu)
> -{
> -       struct kmem_cache *cachep;
> -       int node = cpu_to_mem(cpu);
> -       int err;
> -
> -       /*
> -        * We need to do this right in the beginning since
> -        * alloc_arraycache's are going to use this list.
> -        * kmalloc_node allows us to add the slab to the right
> -        * kmem_cache_node and not this cpu's kmem_cache_node
> -        */
> -       err = init_cache_node_node(node);
> -       if (err < 0)
> -               goto bad;
> -
> -       /*
> -        * Now we can go ahead with allocating the shared arrays and
> -        * array caches
> -        */
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               err = setup_kmem_cache_node(cachep, node, GFP_KERNEL, false);
> -               if (err)
> -                       goto bad;
> -       }
> -
> -       return 0;
> -bad:
> -       cpuup_canceled(cpu);
> -       return -ENOMEM;
> -}
> -
> -int slab_prepare_cpu(unsigned int cpu)
> -{
> -       int err;
> -
> -       mutex_lock(&slab_mutex);
> -       err = cpuup_prepare(cpu);
> -       mutex_unlock(&slab_mutex);
> -       return err;
> -}
> -
> -/*
> - * This is called for a failed online attempt and for a successful
> - * offline.
> - *
> - * Even if all the cpus of a node are down, we don't free the
> - * kmem_cache_node of any cache. This is to avoid a race between cpu_down, and
> - * a kmalloc allocation from another cpu for memory from the node of
> - * the cpu going down.  The kmem_cache_node structure is usually allocated from
> - * kmem_cache_create() and gets destroyed at kmem_cache_destroy().
> - */
> -int slab_dead_cpu(unsigned int cpu)
> -{
> -       mutex_lock(&slab_mutex);
> -       cpuup_canceled(cpu);
> -       mutex_unlock(&slab_mutex);
> -       return 0;
> -}
> -#endif
> -
> -static int slab_online_cpu(unsigned int cpu)
> -{
> -       start_cpu_timer(cpu);
> -       return 0;
> -}
> -
> -static int slab_offline_cpu(unsigned int cpu)
> -{
> -       /*
> -        * Shutdown cache reaper. Note that the slab_mutex is held so
> -        * that if cache_reap() is invoked it cannot do anything
> -        * expensive but will only modify reap_work and reschedule the
> -        * timer.
> -        */
> -       cancel_delayed_work_sync(&per_cpu(slab_reap_work, cpu));
> -       /* Now the cache_reaper is guaranteed to be not running. */
> -       per_cpu(slab_reap_work, cpu).work.func = NULL;
> -       return 0;
> -}
> -
> -#if defined(CONFIG_NUMA)
> -/*
> - * Drains freelist for a node on each slab cache, used for memory hot-remove.
> - * Returns -EBUSY if all objects cannot be drained so that the node is not
> - * removed.
> - *
> - * Must hold slab_mutex.
> - */
> -static int __meminit drain_cache_node_node(int node)
> -{
> -       struct kmem_cache *cachep;
> -       int ret = 0;
> -
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               struct kmem_cache_node *n;
> -
> -               n = get_node(cachep, node);
> -               if (!n)
> -                       continue;
> -
> -               drain_freelist(cachep, n, INT_MAX);
> -
> -               if (!list_empty(&n->slabs_full) ||
> -                   !list_empty(&n->slabs_partial)) {
> -                       ret = -EBUSY;
> -                       break;
> -               }
> -       }
> -       return ret;
> -}
> -
> -static int __meminit slab_memory_callback(struct notifier_block *self,
> -                                       unsigned long action, void *arg)
> -{
> -       struct memory_notify *mnb = arg;
> -       int ret = 0;
> -       int nid;
> -
> -       nid = mnb->status_change_nid;
> -       if (nid < 0)
> -               goto out;
> -
> -       switch (action) {
> -       case MEM_GOING_ONLINE:
> -               mutex_lock(&slab_mutex);
> -               ret = init_cache_node_node(nid);
> -               mutex_unlock(&slab_mutex);
> -               break;
> -       case MEM_GOING_OFFLINE:
> -               mutex_lock(&slab_mutex);
> -               ret = drain_cache_node_node(nid);
> -               mutex_unlock(&slab_mutex);
> -               break;
> -       case MEM_ONLINE:
> -       case MEM_OFFLINE:
> -       case MEM_CANCEL_ONLINE:
> -       case MEM_CANCEL_OFFLINE:
> -               break;
> -       }
> -out:
> -       return notifier_from_errno(ret);
> -}
> -#endif /* CONFIG_NUMA */
> -
> -/*
> - * swap the static kmem_cache_node with kmalloced memory
> - */
> -static void __init init_list(struct kmem_cache *cachep, struct kmem_cache_node *list,
> -                               int nodeid)
> -{
> -       struct kmem_cache_node *ptr;
> -
> -       ptr = kmalloc_node(sizeof(struct kmem_cache_node), GFP_NOWAIT, nodeid);
> -       BUG_ON(!ptr);
> -
> -       memcpy(ptr, list, sizeof(struct kmem_cache_node));
> -       /*
> -        * Do not assume that spinlocks can be initialized via memcpy:
> -        */
> -       raw_spin_lock_init(&ptr->list_lock);
> -
> -       MAKE_ALL_LISTS(cachep, ptr, nodeid);
> -       cachep->node[nodeid] = ptr;
> -}
> -
> -/*
> - * For setting up all the kmem_cache_node for cache whose buffer_size is same as
> - * size of kmem_cache_node.
> - */
> -static void __init set_up_node(struct kmem_cache *cachep, int index)
> -{
> -       int node;
> -
> -       for_each_online_node(node) {
> -               cachep->node[node] = &init_kmem_cache_node[index + node];
> -               cachep->node[node]->next_reap = jiffies +
> -                   REAPTIMEOUT_NODE +
> -                   ((unsigned long)cachep) % REAPTIMEOUT_NODE;
> -       }
> -}
> -
> -/*
> - * Initialisation.  Called after the page allocator have been initialised and
> - * before smp_init().
> - */
> -void __init kmem_cache_init(void)
> -{
> -       int i;
> -
> -       kmem_cache = &kmem_cache_boot;
> -
> -       if (!IS_ENABLED(CONFIG_NUMA) || num_possible_nodes() == 1)
> -               use_alien_caches = 0;
> -
> -       for (i = 0; i < NUM_INIT_LISTS; i++)
> -               kmem_cache_node_init(&init_kmem_cache_node[i]);
> -
> -       /*
> -        * Fragmentation resistance on low memory - only use bigger
> -        * page orders on machines with more than 32MB of memory if
> -        * not overridden on the command line.
> -        */
> -       if (!slab_max_order_set && totalram_pages() > (32 << 20) >> PAGE_SHIFT)
> -               slab_max_order = SLAB_MAX_ORDER_HI;
> -
> -       /* Bootstrap is tricky, because several objects are allocated
> -        * from caches that do not exist yet:
> -        * 1) initialize the kmem_cache cache: it contains the struct
> -        *    kmem_cache structures of all caches, except kmem_cache itself:
> -        *    kmem_cache is statically allocated.
> -        *    Initially an __init data area is used for the head array and the
> -        *    kmem_cache_node structures, it's replaced with a kmalloc allocated
> -        *    array at the end of the bootstrap.
> -        * 2) Create the first kmalloc cache.
> -        *    The struct kmem_cache for the new cache is allocated normally.
> -        *    An __init data area is used for the head array.
> -        * 3) Create the remaining kmalloc caches, with minimally sized
> -        *    head arrays.
> -        * 4) Replace the __init data head arrays for kmem_cache and the first
> -        *    kmalloc cache with kmalloc allocated arrays.
> -        * 5) Replace the __init data for kmem_cache_node for kmem_cache and
> -        *    the other cache's with kmalloc allocated memory.
> -        * 6) Resize the head arrays of the kmalloc caches to their final sizes.
> -        */
> -
> -       /* 1) create the kmem_cache */
> -
> -       /*
> -        * struct kmem_cache size depends on nr_node_ids & nr_cpu_ids
> -        */
> -       create_boot_cache(kmem_cache, "kmem_cache",
> -               offsetof(struct kmem_cache, node) +
> -                                 nr_node_ids * sizeof(struct kmem_cache_node *),
> -                                 SLAB_HWCACHE_ALIGN, 0, 0);
> -       list_add(&kmem_cache->list, &slab_caches);
> -       slab_state = PARTIAL;
> -
> -       /*
> -        * Initialize the caches that provide memory for the  kmem_cache_node
> -        * structures first.  Without this, further allocations will bug.
> -        */
> -       kmalloc_caches[KMALLOC_NORMAL][INDEX_NODE] = create_kmalloc_cache(
> -                               kmalloc_info[INDEX_NODE].name[KMALLOC_NORMAL],
> -                               kmalloc_info[INDEX_NODE].size,
> -                               ARCH_KMALLOC_FLAGS, 0,
> -                               kmalloc_info[INDEX_NODE].size);
> -       slab_state = PARTIAL_NODE;
> -       setup_kmalloc_cache_index_table();
> -
> -       /* 5) Replace the bootstrap kmem_cache_node */
> -       {
> -               int nid;
> -
> -               for_each_online_node(nid) {
> -                       init_list(kmem_cache, &init_kmem_cache_node[CACHE_CACHE + nid], nid);
> -
> -                       init_list(kmalloc_caches[KMALLOC_NORMAL][INDEX_NODE],
> -                                         &init_kmem_cache_node[SIZE_NODE + nid], nid);
> -               }
> -       }
> -
> -       create_kmalloc_caches(ARCH_KMALLOC_FLAGS);
> -}
> -
> -void __init kmem_cache_init_late(void)
> -{
> -       struct kmem_cache *cachep;
> -
> -       /* 6) resize the head arrays to their final sizes */
> -       mutex_lock(&slab_mutex);
> -       list_for_each_entry(cachep, &slab_caches, list)
> -               if (enable_cpucache(cachep, GFP_NOWAIT))
> -                       BUG();
> -       mutex_unlock(&slab_mutex);
> -
> -       /* Done! */
> -       slab_state = FULL;
> -
> -#ifdef CONFIG_NUMA
> -       /*
> -        * Register a memory hotplug callback that initializes and frees
> -        * node.
> -        */
> -       hotplug_memory_notifier(slab_memory_callback, SLAB_CALLBACK_PRI);
> -#endif
> -
> -       /*
> -        * The reap timers are started later, with a module init call: That part
> -        * of the kernel is not yet operational.
> -        */
> -}
> -
> -static int __init cpucache_init(void)
> -{
> -       int ret;
> -
> -       /*
> -        * Register the timers that return unneeded pages to the page allocator
> -        */
> -       ret = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "SLAB online",
> -                               slab_online_cpu, slab_offline_cpu);
> -       WARN_ON(ret < 0);
> -
> -       return 0;
> -}
> -__initcall(cpucache_init);
> -
> -static noinline void
> -slab_out_of_memory(struct kmem_cache *cachep, gfp_t gfpflags, int nodeid)
> -{
> -#if DEBUG
> -       struct kmem_cache_node *n;
> -       unsigned long flags;
> -       int node;
> -       static DEFINE_RATELIMIT_STATE(slab_oom_rs, DEFAULT_RATELIMIT_INTERVAL,
> -                                     DEFAULT_RATELIMIT_BURST);
> -
> -       if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slab_oom_rs))
> -               return;
> -
> -       pr_warn("SLAB: Unable to allocate memory on node %d, gfp=%#x(%pGg)\n",
> -               nodeid, gfpflags, &gfpflags);
> -       pr_warn("  cache: %s, object size: %d, order: %d\n",
> -               cachep->name, cachep->size, cachep->gfporder);
> -
> -       for_each_kmem_cache_node(cachep, node, n) {
> -               unsigned long total_slabs, free_slabs, free_objs;
> -
> -               raw_spin_lock_irqsave(&n->list_lock, flags);
> -               total_slabs = n->total_slabs;
> -               free_slabs = n->free_slabs;
> -               free_objs = n->free_objects;
> -               raw_spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -               pr_warn("  node %d: slabs: %ld/%ld, objs: %ld/%ld\n",
> -                       node, total_slabs - free_slabs, total_slabs,
> -                       (total_slabs * cachep->num) - free_objs,
> -                       total_slabs * cachep->num);
> -       }
> -#endif
> -}
> -
> -/*
> - * Interface to system's page allocator. No need to hold the
> - * kmem_cache_node ->list_lock.
> - *
> - * If we requested dmaable memory, we will get it. Even if we
> - * did not request dmaable memory, we might get it, but that
> - * would be relatively rare and ignorable.
> - */
> -static struct slab *kmem_getpages(struct kmem_cache *cachep, gfp_t flags,
> -                                                               int nodeid)
> -{
> -       struct folio *folio;
> -       struct slab *slab;
> -
> -       flags |= cachep->allocflags;
> -
> -       folio = (struct folio *) __alloc_pages_node(nodeid, flags, cachep->gfporder);
> -       if (!folio) {
> -               slab_out_of_memory(cachep, flags, nodeid);
> -               return NULL;
> -       }
> -
> -       slab = folio_slab(folio);
> -
> -       account_slab(slab, cachep->gfporder, cachep, flags);
> -       __folio_set_slab(folio);
> -       /* Make the flag visible before any changes to folio->mapping */
> -       smp_wmb();
> -       /* Record if ALLOC_NO_WATERMARKS was set when allocating the slab */
> -       if (sk_memalloc_socks() && folio_is_pfmemalloc(folio))
> -               slab_set_pfmemalloc(slab);
> -
> -       return slab;
> -}
> -
> -/*
> - * Interface to system's page release.
> - */
> -static void kmem_freepages(struct kmem_cache *cachep, struct slab *slab)
> -{
> -       int order = cachep->gfporder;
> -       struct folio *folio = slab_folio(slab);
> -
> -       BUG_ON(!folio_test_slab(folio));
> -       __slab_clear_pfmemalloc(slab);
> -       page_mapcount_reset(&folio->page);
> -       folio->mapping = NULL;
> -       /* Make the mapping reset visible before clearing the flag */
> -       smp_wmb();
> -       __folio_clear_slab(folio);
> -
> -       if (current->reclaim_state)
> -               current->reclaim_state->reclaimed_slab += 1 << order;
> -       unaccount_slab(slab, order, cachep);
> -       __free_pages(&folio->page, order);
> -}
> -
> -static void kmem_rcu_free(struct rcu_head *head)
> -{
> -       struct kmem_cache *cachep;
> -       struct slab *slab;
> -
> -       slab = container_of(head, struct slab, rcu_head);
> -       cachep = slab->slab_cache;
> -
> -       kmem_freepages(cachep, slab);
> -}
> -
> -#if DEBUG
> -static inline bool is_debug_pagealloc_cache(struct kmem_cache *cachep)
> -{
> -       return debug_pagealloc_enabled_static() && OFF_SLAB(cachep) &&
> -                       ((cachep->size % PAGE_SIZE) == 0);
> -}
> -
> -#ifdef CONFIG_DEBUG_PAGEALLOC
> -static void slab_kernel_map(struct kmem_cache *cachep, void *objp, int map)
> -{
> -       if (!is_debug_pagealloc_cache(cachep))
> -               return;
> -
> -       __kernel_map_pages(virt_to_page(objp), cachep->size / PAGE_SIZE, map);
> -}
> -
> -#else
> -static inline void slab_kernel_map(struct kmem_cache *cachep, void *objp,
> -                               int map) {}
> -
> -#endif
> -
> -static void poison_obj(struct kmem_cache *cachep, void *addr, unsigned char val)
> -{
> -       int size = cachep->object_size;
> -       addr = &((char *)addr)[obj_offset(cachep)];
> -
> -       memset(addr, val, size);
> -       *(unsigned char *)(addr + size - 1) = POISON_END;
> -}
> -
> -static void dump_line(char *data, int offset, int limit)
> -{
> -       int i;
> -       unsigned char error = 0;
> -       int bad_count = 0;
> -
> -       pr_err("%03x: ", offset);
> -       for (i = 0; i < limit; i++) {
> -               if (data[offset + i] != POISON_FREE) {
> -                       error = data[offset + i];
> -                       bad_count++;
> -               }
> -       }
> -       print_hex_dump(KERN_CONT, "", 0, 16, 1,
> -                       &data[offset], limit, 1);
> -
> -       if (bad_count == 1) {
> -               error ^= POISON_FREE;
> -               if (!(error & (error - 1))) {
> -                       pr_err("Single bit error detected. Probably bad RAM.\n");
> -#ifdef CONFIG_X86
> -                       pr_err("Run memtest86+ or a similar memory test tool.\n");
> -#else
> -                       pr_err("Run a memory test tool.\n");
> -#endif
> -               }
> -       }
> -}
> -#endif
> -
> -#if DEBUG
> -
> -static void print_objinfo(struct kmem_cache *cachep, void *objp, int lines)
> -{
> -       int i, size;
> -       char *realobj;
> -
> -       if (cachep->flags & SLAB_RED_ZONE) {
> -               pr_err("Redzone: 0x%llx/0x%llx\n",
> -                      *dbg_redzone1(cachep, objp),
> -                      *dbg_redzone2(cachep, objp));
> -       }
> -
> -       if (cachep->flags & SLAB_STORE_USER)
> -               pr_err("Last user: (%pSR)\n", *dbg_userword(cachep, objp));
> -       realobj = (char *)objp + obj_offset(cachep);
> -       size = cachep->object_size;
> -       for (i = 0; i < size && lines; i += 16, lines--) {
> -               int limit;
> -               limit = 16;
> -               if (i + limit > size)
> -                       limit = size - i;
> -               dump_line(realobj, i, limit);
> -       }
> -}
> -
> -static void check_poison_obj(struct kmem_cache *cachep, void *objp)
> -{
> -       char *realobj;
> -       int size, i;
> -       int lines = 0;
> -
> -       if (is_debug_pagealloc_cache(cachep))
> -               return;
> -
> -       realobj = (char *)objp + obj_offset(cachep);
> -       size = cachep->object_size;
> -
> -       for (i = 0; i < size; i++) {
> -               char exp = POISON_FREE;
> -               if (i == size - 1)
> -                       exp = POISON_END;
> -               if (realobj[i] != exp) {
> -                       int limit;
> -                       /* Mismatch ! */
> -                       /* Print header */
> -                       if (lines == 0) {
> -                               pr_err("Slab corruption (%s): %s start=%px, len=%d\n",
> -                                      print_tainted(), cachep->name,
> -                                      realobj, size);
> -                               print_objinfo(cachep, objp, 0);
> -                       }
> -                       /* Hexdump the affected line */
> -                       i = (i / 16) * 16;
> -                       limit = 16;
> -                       if (i + limit > size)
> -                               limit = size - i;
> -                       dump_line(realobj, i, limit);
> -                       i += 16;
> -                       lines++;
> -                       /* Limit to 5 lines */
> -                       if (lines > 5)
> -                               break;
> -               }
> -       }
> -       if (lines != 0) {
> -               /* Print some data about the neighboring objects, if they
> -                * exist:
> -                */
> -               struct slab *slab = virt_to_slab(objp);
> -               unsigned int objnr;
> -
> -               objnr = obj_to_index(cachep, slab, objp);
> -               if (objnr) {
> -                       objp = index_to_obj(cachep, slab, objnr - 1);
> -                       realobj = (char *)objp + obj_offset(cachep);
> -                       pr_err("Prev obj: start=%px, len=%d\n", realobj, size);
> -                       print_objinfo(cachep, objp, 2);
> -               }
> -               if (objnr + 1 < cachep->num) {
> -                       objp = index_to_obj(cachep, slab, objnr + 1);
> -                       realobj = (char *)objp + obj_offset(cachep);
> -                       pr_err("Next obj: start=%px, len=%d\n", realobj, size);
> -                       print_objinfo(cachep, objp, 2);
> -               }
> -       }
> -}
> -#endif
> -
> -#if DEBUG
> -static void slab_destroy_debugcheck(struct kmem_cache *cachep,
> -                                               struct slab *slab)
> -{
> -       int i;
> -
> -       if (OBJFREELIST_SLAB(cachep) && cachep->flags & SLAB_POISON) {
> -               poison_obj(cachep, slab->freelist - obj_offset(cachep),
> -                       POISON_FREE);
> -       }
> -
> -       for (i = 0; i < cachep->num; i++) {
> -               void *objp = index_to_obj(cachep, slab, i);
> -
> -               if (cachep->flags & SLAB_POISON) {
> -                       check_poison_obj(cachep, objp);
> -                       slab_kernel_map(cachep, objp, 1);
> -               }
> -               if (cachep->flags & SLAB_RED_ZONE) {
> -                       if (*dbg_redzone1(cachep, objp) != RED_INACTIVE)
> -                               slab_error(cachep, "start of a freed object was overwritten");
> -                       if (*dbg_redzone2(cachep, objp) != RED_INACTIVE)
> -                               slab_error(cachep, "end of a freed object was overwritten");
> -               }
> -       }
> -}
> -#else
> -static void slab_destroy_debugcheck(struct kmem_cache *cachep,
> -                                               struct slab *slab)
> -{
> -}
> -#endif
> -
> -/**
> - * slab_destroy - destroy and release all objects in a slab
> - * @cachep: cache pointer being destroyed
> - * @slab: slab being destroyed
> - *
> - * Destroy all the objs in a slab, and release the mem back to the system.
> - * Before calling the slab must have been unlinked from the cache. The
> - * kmem_cache_node ->list_lock is not held/needed.
> - */
> -static void slab_destroy(struct kmem_cache *cachep, struct slab *slab)
> -{
> -       void *freelist;
> -
> -       freelist = slab->freelist;
> -       slab_destroy_debugcheck(cachep, slab);
> -       if (unlikely(cachep->flags & SLAB_TYPESAFE_BY_RCU))
> -               call_rcu(&slab->rcu_head, kmem_rcu_free);
> -       else
> -               kmem_freepages(cachep, slab);
> -
> -       /*
> -        * From now on, we don't use freelist
> -        * although actual page can be freed in rcu context
> -        */
> -       if (OFF_SLAB(cachep))
> -               kfree(freelist);
> -}
> -
> -/*
> - * Update the size of the caches before calling slabs_destroy as it may
> - * recursively call kfree.
> - */
> -static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list)
> -{
> -       struct slab *slab, *n;
> -
> -       list_for_each_entry_safe(slab, n, list, slab_list) {
> -               list_del(&slab->slab_list);
> -               slab_destroy(cachep, slab);
> -       }
> -}
> -
> -/**
> - * calculate_slab_order - calculate size (page order) of slabs
> - * @cachep: pointer to the cache that is being created
> - * @size: size of objects to be created in this cache.
> - * @flags: slab allocation flags
> - *
> - * Also calculates the number of objects per slab.
> - *
> - * This could be made much more intelligent.  For now, try to avoid using
> - * high order pages for slabs.  When the gfp() functions are more friendly
> - * towards high-order requests, this should be changed.
> - *
> - * Return: number of left-over bytes in a slab
> - */
> -static size_t calculate_slab_order(struct kmem_cache *cachep,
> -                               size_t size, slab_flags_t flags)
> -{
> -       size_t left_over = 0;
> -       int gfporder;
> -
> -       for (gfporder = 0; gfporder <= KMALLOC_MAX_ORDER; gfporder++) {
> -               unsigned int num;
> -               size_t remainder;
> -
> -               num = cache_estimate(gfporder, size, flags, &remainder);
> -               if (!num)
> -                       continue;
> -
> -               /* Can't handle number of objects more than SLAB_OBJ_MAX_NUM */
> -               if (num > SLAB_OBJ_MAX_NUM)
> -                       break;
> -
> -               if (flags & CFLGS_OFF_SLAB) {
> -                       struct kmem_cache *freelist_cache;
> -                       size_t freelist_size;
> -                       size_t freelist_cache_size;
> -
> -                       freelist_size = num * sizeof(freelist_idx_t);
> -                       if (freelist_size > KMALLOC_MAX_CACHE_SIZE) {
> -                               freelist_cache_size = PAGE_SIZE << get_order(freelist_size);
> -                       } else {
> -                               freelist_cache = kmalloc_slab(freelist_size, 0u);
> -                               if (!freelist_cache)
> -                                       continue;
> -                               freelist_cache_size = freelist_cache->size;
> -
> -                               /*
> -                                * Needed to avoid possible looping condition
> -                                * in cache_grow_begin()
> -                                */
> -                               if (OFF_SLAB(freelist_cache))
> -                                       continue;
> -                       }
> -
> -                       /* check if off slab has enough benefit */
> -                       if (freelist_cache_size > cachep->size / 2)
> -                               continue;
> -               }
> -
> -               /* Found something acceptable - save it away */
> -               cachep->num = num;
> -               cachep->gfporder = gfporder;
> -               left_over = remainder;
> -
> -               /*
> -                * A VFS-reclaimable slab tends to have most allocations
> -                * as GFP_NOFS and we really don't want to have to be allocating
> -                * higher-order pages when we are unable to shrink dcache.
> -                */
> -               if (flags & SLAB_RECLAIM_ACCOUNT)
> -                       break;
> -
> -               /*
> -                * Large number of objects is good, but very large slabs are
> -                * currently bad for the gfp()s.
> -                */
> -               if (gfporder >= slab_max_order)
> -                       break;
> -
> -               /*
> -                * Acceptable internal fragmentation?
> -                */
> -               if (left_over * 8 <= (PAGE_SIZE << gfporder))
> -                       break;
> -       }
> -       return left_over;
> -}
> -
> -static struct array_cache __percpu *alloc_kmem_cache_cpus(
> -               struct kmem_cache *cachep, int entries, int batchcount)
> -{
> -       int cpu;
> -       size_t size;
> -       struct array_cache __percpu *cpu_cache;
> -
> -       size = sizeof(void *) * entries + sizeof(struct array_cache);
> -       cpu_cache = __alloc_percpu(size, sizeof(void *));
> -
> -       if (!cpu_cache)
> -               return NULL;
> -
> -       for_each_possible_cpu(cpu) {
> -               init_arraycache(per_cpu_ptr(cpu_cache, cpu),
> -                               entries, batchcount);
> -       }
> -
> -       return cpu_cache;
> -}
> -
> -static int __ref setup_cpu_cache(struct kmem_cache *cachep, gfp_t gfp)
> -{
> -       if (slab_state >= FULL)
> -               return enable_cpucache(cachep, gfp);
> -
> -       cachep->cpu_cache = alloc_kmem_cache_cpus(cachep, 1, 1);
> -       if (!cachep->cpu_cache)
> -               return 1;
> -
> -       if (slab_state == DOWN) {
> -               /* Creation of first cache (kmem_cache). */
> -               set_up_node(kmem_cache, CACHE_CACHE);
> -       } else if (slab_state == PARTIAL) {
> -               /* For kmem_cache_node */
> -               set_up_node(cachep, SIZE_NODE);
> -       } else {
> -               int node;
> -
> -               for_each_online_node(node) {
> -                       cachep->node[node] = kmalloc_node(
> -                               sizeof(struct kmem_cache_node), gfp, node);
> -                       BUG_ON(!cachep->node[node]);
> -                       kmem_cache_node_init(cachep->node[node]);
> -               }
> -       }
> -
> -       cachep->node[numa_mem_id()]->next_reap =
> -                       jiffies + REAPTIMEOUT_NODE +
> -                       ((unsigned long)cachep) % REAPTIMEOUT_NODE;
> -
> -       cpu_cache_get(cachep)->avail = 0;
> -       cpu_cache_get(cachep)->limit = BOOT_CPUCACHE_ENTRIES;
> -       cpu_cache_get(cachep)->batchcount = 1;
> -       cpu_cache_get(cachep)->touched = 0;
> -       cachep->batchcount = 1;
> -       cachep->limit = BOOT_CPUCACHE_ENTRIES;
> -       return 0;
> -}
> -
> -slab_flags_t kmem_cache_flags(unsigned int object_size,
> -       slab_flags_t flags, const char *name)
> -{
> -       return flags;
> -}
> -
> -struct kmem_cache *
> -__kmem_cache_alias(const char *name, unsigned int size, unsigned int align,
> -                  slab_flags_t flags, void (*ctor)(void *))
> -{
> -       struct kmem_cache *cachep;
> -
> -       cachep = find_mergeable(size, align, flags, name, ctor);
> -       if (cachep) {
> -               cachep->refcount++;
> -
> -               /*
> -                * Adjust the object sizes so that we clear
> -                * the complete object on kzalloc.
> -                */
> -               cachep->object_size = max_t(int, cachep->object_size, size);
> -       }
> -       return cachep;
> -}
> -
> -static bool set_objfreelist_slab_cache(struct kmem_cache *cachep,
> -                       size_t size, slab_flags_t flags)
> -{
> -       size_t left;
> -
> -       cachep->num = 0;
> -
> -       /*
> -        * If slab auto-initialization on free is enabled, store the freelist
> -        * off-slab, so that its contents don't end up in one of the allocated
> -        * objects.
> -        */
> -       if (unlikely(slab_want_init_on_free(cachep)))
> -               return false;
> -
> -       if (cachep->ctor || flags & SLAB_TYPESAFE_BY_RCU)
> -               return false;
> -
> -       left = calculate_slab_order(cachep, size,
> -                       flags | CFLGS_OBJFREELIST_SLAB);
> -       if (!cachep->num)
> -               return false;
> -
> -       if (cachep->num * sizeof(freelist_idx_t) > cachep->object_size)
> -               return false;
> -
> -       cachep->colour = left / cachep->colour_off;
> -
> -       return true;
> -}
> -
> -static bool set_off_slab_cache(struct kmem_cache *cachep,
> -                       size_t size, slab_flags_t flags)
> -{
> -       size_t left;
> -
> -       cachep->num = 0;
> -
> -       /*
> -        * Always use on-slab management when SLAB_NOLEAKTRACE
> -        * to avoid recursive calls into kmemleak.
> -        */
> -       if (flags & SLAB_NOLEAKTRACE)
> -               return false;
> -
> -       /*
> -        * Size is large, assume best to place the slab management obj
> -        * off-slab (should allow better packing of objs).
> -        */
> -       left = calculate_slab_order(cachep, size, flags | CFLGS_OFF_SLAB);
> -       if (!cachep->num)
> -               return false;
> -
> -       /*
> -        * If the slab has been placed off-slab, and we have enough space then
> -        * move it on-slab. This is at the expense of any extra colouring.
> -        */
> -       if (left >= cachep->num * sizeof(freelist_idx_t))
> -               return false;
> -
> -       cachep->colour = left / cachep->colour_off;
> -
> -       return true;
> -}
> -
> -static bool set_on_slab_cache(struct kmem_cache *cachep,
> -                       size_t size, slab_flags_t flags)
> -{
> -       size_t left;
> -
> -       cachep->num = 0;
> -
> -       left = calculate_slab_order(cachep, size, flags);
> -       if (!cachep->num)
> -               return false;
> -
> -       cachep->colour = left / cachep->colour_off;
> -
> -       return true;
> -}
> -
> -/**
> - * __kmem_cache_create - Create a cache.
> - * @cachep: cache management descriptor
> - * @flags: SLAB flags
> - *
> - * Returns a ptr to the cache on success, NULL on failure.
> - * Cannot be called within an int, but can be interrupted.
> - * The @ctor is run when new pages are allocated by the cache.
> - *
> - * The flags are
> - *
> - * %SLAB_POISON - Poison the slab with a known test pattern (a5a5a5a5)
> - * to catch references to uninitialised memory.
> - *
> - * %SLAB_RED_ZONE - Insert `Red' zones around the allocated memory to check
> - * for buffer overruns.
> - *
> - * %SLAB_HWCACHE_ALIGN - Align the objects in this cache to a hardware
> - * cacheline.  This can be beneficial if you're counting cycles as closely
> - * as davem.
> - *
> - * Return: a pointer to the created cache or %NULL in case of error
> - */
> -int __kmem_cache_create(struct kmem_cache *cachep, slab_flags_t flags)
> -{
> -       size_t ralign = BYTES_PER_WORD;
> -       gfp_t gfp;
> -       int err;
> -       unsigned int size = cachep->size;
> -
> -#if DEBUG
> -#if FORCED_DEBUG
> -       /*
> -        * Enable redzoning and last user accounting, except for caches with
> -        * large objects, if the increased size would increase the object size
> -        * above the next power of two: caches with object sizes just above a
> -        * power of two have a significant amount of internal fragmentation.
> -        */
> -       if (size < 4096 || fls(size - 1) == fls(size-1 + REDZONE_ALIGN +
> -                                               2 * sizeof(unsigned long long)))
> -               flags |= SLAB_RED_ZONE | SLAB_STORE_USER;
> -       if (!(flags & SLAB_TYPESAFE_BY_RCU))
> -               flags |= SLAB_POISON;
> -#endif
> -#endif
> -
> -       /*
> -        * Check that size is in terms of words.  This is needed to avoid
> -        * unaligned accesses for some archs when redzoning is used, and makes
> -        * sure any on-slab bufctl's are also correctly aligned.
> -        */
> -       size = ALIGN(size, BYTES_PER_WORD);
> -
> -       if (flags & SLAB_RED_ZONE) {
> -               ralign = REDZONE_ALIGN;
> -               /* If redzoning, ensure that the second redzone is suitably
> -                * aligned, by adjusting the object size accordingly. */
> -               size = ALIGN(size, REDZONE_ALIGN);
> -       }
> -
> -       /* 3) caller mandated alignment */
> -       if (ralign < cachep->align) {
> -               ralign = cachep->align;
> -       }
> -       /* disable debug if necessary */
> -       if (ralign > __alignof__(unsigned long long))
> -               flags &= ~(SLAB_RED_ZONE | SLAB_STORE_USER);
> -       /*
> -        * 4) Store it.
> -        */
> -       cachep->align = ralign;
> -       cachep->colour_off = cache_line_size();
> -       /* Offset must be a multiple of the alignment. */
> -       if (cachep->colour_off < cachep->align)
> -               cachep->colour_off = cachep->align;
> -
> -       if (slab_is_available())
> -               gfp = GFP_KERNEL;
> -       else
> -               gfp = GFP_NOWAIT;
> -
> -#if DEBUG
> -
> -       /*
> -        * Both debugging options require word-alignment which is calculated
> -        * into align above.
> -        */
> -       if (flags & SLAB_RED_ZONE) {
> -               /* add space for red zone words */
> -               cachep->obj_offset += sizeof(unsigned long long);
> -               size += 2 * sizeof(unsigned long long);
> -       }
> -       if (flags & SLAB_STORE_USER) {
> -               /* user store requires one word storage behind the end of
> -                * the real object. But if the second red zone needs to be
> -                * aligned to 64 bits, we must allow that much space.
> -                */
> -               if (flags & SLAB_RED_ZONE)
> -                       size += REDZONE_ALIGN;
> -               else
> -                       size += BYTES_PER_WORD;
> -       }
> -#endif
> -
> -       kasan_cache_create(cachep, &size, &flags);
> -
> -       size = ALIGN(size, cachep->align);
> -       /*
> -        * We should restrict the number of objects in a slab to implement
> -        * byte sized index. Refer comment on SLAB_OBJ_MIN_SIZE definition.
> -        */
> -       if (FREELIST_BYTE_INDEX && size < SLAB_OBJ_MIN_SIZE)
> -               size = ALIGN(SLAB_OBJ_MIN_SIZE, cachep->align);
> -
> -#if DEBUG
> -       /*
> -        * To activate debug pagealloc, off-slab management is necessary
> -        * requirement. In early phase of initialization, small sized slab
> -        * doesn't get initialized so it would not be possible. So, we need
> -        * to check size >= 256. It guarantees that all necessary small
> -        * sized slab is initialized in current slab initialization sequence.
> -        */
> -       if (debug_pagealloc_enabled_static() && (flags & SLAB_POISON) &&
> -               size >= 256 && cachep->object_size > cache_line_size()) {
> -               if (size < PAGE_SIZE || size % PAGE_SIZE == 0) {
> -                       size_t tmp_size = ALIGN(size, PAGE_SIZE);
> -
> -                       if (set_off_slab_cache(cachep, tmp_size, flags)) {
> -                               flags |= CFLGS_OFF_SLAB;
> -                               cachep->obj_offset += tmp_size - size;
> -                               size = tmp_size;
> -                               goto done;
> -                       }
> -               }
> -       }
> -#endif
> -
> -       if (set_objfreelist_slab_cache(cachep, size, flags)) {
> -               flags |= CFLGS_OBJFREELIST_SLAB;
> -               goto done;
> -       }
> -
> -       if (set_off_slab_cache(cachep, size, flags)) {
> -               flags |= CFLGS_OFF_SLAB;
> -               goto done;
> -       }
> -
> -       if (set_on_slab_cache(cachep, size, flags))
> -               goto done;
> -
> -       return -E2BIG;
> -
> -done:
> -       cachep->freelist_size = cachep->num * sizeof(freelist_idx_t);
> -       cachep->flags = flags;
> -       cachep->allocflags = __GFP_COMP;
> -       if (flags & SLAB_CACHE_DMA)
> -               cachep->allocflags |= GFP_DMA;
> -       if (flags & SLAB_CACHE_DMA32)
> -               cachep->allocflags |= GFP_DMA32;
> -       if (flags & SLAB_RECLAIM_ACCOUNT)
> -               cachep->allocflags |= __GFP_RECLAIMABLE;
> -       cachep->size = size;
> -       cachep->reciprocal_buffer_size = reciprocal_value(size);
> -
> -#if DEBUG
> -       /*
> -        * If we're going to use the generic kernel_map_pages()
> -        * poisoning, then it's going to smash the contents of
> -        * the redzone and userword anyhow, so switch them off.
> -        */
> -       if (IS_ENABLED(CONFIG_PAGE_POISONING) &&
> -               (cachep->flags & SLAB_POISON) &&
> -               is_debug_pagealloc_cache(cachep))
> -               cachep->flags &= ~(SLAB_RED_ZONE | SLAB_STORE_USER);
> -#endif
> -
> -       err = setup_cpu_cache(cachep, gfp);
> -       if (err) {
> -               __kmem_cache_release(cachep);
> -               return err;
> -       }
> -
> -       return 0;
> -}
> -
> -#if DEBUG
> -static void check_irq_off(void)
> -{
> -       BUG_ON(!irqs_disabled());
> -}
> -
> -static void check_irq_on(void)
> -{
> -       BUG_ON(irqs_disabled());
> -}
> -
> -static void check_mutex_acquired(void)
> -{
> -       BUG_ON(!mutex_is_locked(&slab_mutex));
> -}
> -
> -static void check_spinlock_acquired(struct kmem_cache *cachep)
> -{
> -#ifdef CONFIG_SMP
> -       check_irq_off();
> -       assert_raw_spin_locked(&get_node(cachep, numa_mem_id())->list_lock);
> -#endif
> -}
> -
> -static void check_spinlock_acquired_node(struct kmem_cache *cachep, int node)
> -{
> -#ifdef CONFIG_SMP
> -       check_irq_off();
> -       assert_raw_spin_locked(&get_node(cachep, node)->list_lock);
> -#endif
> -}
> -
> -#else
> -#define check_irq_off()        do { } while(0)
> -#define check_irq_on() do { } while(0)
> -#define check_mutex_acquired() do { } while(0)
> -#define check_spinlock_acquired(x) do { } while(0)
> -#define check_spinlock_acquired_node(x, y) do { } while(0)
> -#endif
> -
> -static void drain_array_locked(struct kmem_cache *cachep, struct array_cache *ac,
> -                               int node, bool free_all, struct list_head *list)
> -{
> -       int tofree;
> -
> -       if (!ac || !ac->avail)
> -               return;
> -
> -       tofree = free_all ? ac->avail : (ac->limit + 4) / 5;
> -       if (tofree > ac->avail)
> -               tofree = (ac->avail + 1) / 2;
> -
> -       free_block(cachep, ac->entry, tofree, node, list);
> -       ac->avail -= tofree;
> -       memmove(ac->entry, &(ac->entry[tofree]), sizeof(void *) * ac->avail);
> -}
> -
> -static void do_drain(void *arg)
> -{
> -       struct kmem_cache *cachep = arg;
> -       struct array_cache *ac;
> -       int node = numa_mem_id();
> -       struct kmem_cache_node *n;
> -       LIST_HEAD(list);
> -
> -       check_irq_off();
> -       ac = cpu_cache_get(cachep);
> -       n = get_node(cachep, node);
> -       raw_spin_lock(&n->list_lock);
> -       free_block(cachep, ac->entry, ac->avail, node, &list);
> -       raw_spin_unlock(&n->list_lock);
> -       ac->avail = 0;
> -       slabs_destroy(cachep, &list);
> -}
> -
> -static void drain_cpu_caches(struct kmem_cache *cachep)
> -{
> -       struct kmem_cache_node *n;
> -       int node;
> -       LIST_HEAD(list);
> -
> -       on_each_cpu(do_drain, cachep, 1);
> -       check_irq_on();
> -       for_each_kmem_cache_node(cachep, node, n)
> -               if (n->alien)
> -                       drain_alien_cache(cachep, n->alien);
> -
> -       for_each_kmem_cache_node(cachep, node, n) {
> -               raw_spin_lock_irq(&n->list_lock);
> -               drain_array_locked(cachep, n->shared, node, true, &list);
> -               raw_spin_unlock_irq(&n->list_lock);
> -
> -               slabs_destroy(cachep, &list);
> -       }
> -}
> -
> -/*
> - * Remove slabs from the list of free slabs.
> - * Specify the number of slabs to drain in tofree.
> - *
> - * Returns the actual number of slabs released.
> - */
> -static int drain_freelist(struct kmem_cache *cache,
> -                       struct kmem_cache_node *n, int tofree)
> -{
> -       struct list_head *p;
> -       int nr_freed;
> -       struct slab *slab;
> -
> -       nr_freed = 0;
> -       while (nr_freed < tofree && !list_empty(&n->slabs_free)) {
> -
> -               raw_spin_lock_irq(&n->list_lock);
> -               p = n->slabs_free.prev;
> -               if (p == &n->slabs_free) {
> -                       raw_spin_unlock_irq(&n->list_lock);
> -                       goto out;
> -               }
> -
> -               slab = list_entry(p, struct slab, slab_list);
> -               list_del(&slab->slab_list);
> -               n->free_slabs--;
> -               n->total_slabs--;
> -               /*
> -                * Safe to drop the lock. The slab is no longer linked
> -                * to the cache.
> -                */
> -               n->free_objects -= cache->num;
> -               raw_spin_unlock_irq(&n->list_lock);
> -               slab_destroy(cache, slab);
> -               nr_freed++;
> -
> -               cond_resched();
> -       }
> -out:
> -       return nr_freed;
> -}
> -
> -bool __kmem_cache_empty(struct kmem_cache *s)
> -{
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_kmem_cache_node(s, node, n)
> -               if (!list_empty(&n->slabs_full) ||
> -                   !list_empty(&n->slabs_partial))
> -                       return false;
> -       return true;
> -}
> -
> -int __kmem_cache_shrink(struct kmem_cache *cachep)
> -{
> -       int ret = 0;
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       drain_cpu_caches(cachep);
> -
> -       check_irq_on();
> -       for_each_kmem_cache_node(cachep, node, n) {
> -               drain_freelist(cachep, n, INT_MAX);
> -
> -               ret += !list_empty(&n->slabs_full) ||
> -                       !list_empty(&n->slabs_partial);
> -       }
> -       return (ret ? 1 : 0);
> -}
> -
> -int __kmem_cache_shutdown(struct kmem_cache *cachep)
> -{
> -       return __kmem_cache_shrink(cachep);
> -}
> -
> -void __kmem_cache_release(struct kmem_cache *cachep)
> -{
> -       int i;
> -       struct kmem_cache_node *n;
> -
> -       cache_random_seq_destroy(cachep);
> -
> -       free_percpu(cachep->cpu_cache);
> -
> -       /* NUMA: free the node structures */
> -       for_each_kmem_cache_node(cachep, i, n) {
> -               kfree(n->shared);
> -               free_alien_cache(n->alien);
> -               kfree(n);
> -               cachep->node[i] = NULL;
> -       }
> -}
> -
> -/*
> - * Get the memory for a slab management obj.
> - *
> - * For a slab cache when the slab descriptor is off-slab, the
> - * slab descriptor can't come from the same cache which is being created,
> - * Because if it is the case, that means we defer the creation of
> - * the kmalloc_{dma,}_cache of size sizeof(slab descriptor) to this point.
> - * And we eventually call down to __kmem_cache_create(), which
> - * in turn looks up in the kmalloc_{dma,}_caches for the desired-size one.
> - * This is a "chicken-and-egg" problem.
> - *
> - * So the off-slab slab descriptor shall come from the kmalloc_{dma,}_caches,
> - * which are all initialized during kmem_cache_init().
> - */
> -static void *alloc_slabmgmt(struct kmem_cache *cachep,
> -                                  struct slab *slab, int colour_off,
> -                                  gfp_t local_flags, int nodeid)
> -{
> -       void *freelist;
> -       void *addr = slab_address(slab);
> -
> -       slab->s_mem = addr + colour_off;
> -       slab->active = 0;
> -
> -       if (OBJFREELIST_SLAB(cachep))
> -               freelist = NULL;
> -       else if (OFF_SLAB(cachep)) {
> -               /* Slab management obj is off-slab. */
> -               freelist = kmalloc_node(cachep->freelist_size,
> -                                             local_flags, nodeid);
> -       } else {
> -               /* We will use last bytes at the slab for freelist */
> -               freelist = addr + (PAGE_SIZE << cachep->gfporder) -
> -                               cachep->freelist_size;
> -       }
> -
> -       return freelist;
> -}
> -
> -static inline freelist_idx_t get_free_obj(struct slab *slab, unsigned int idx)
> -{
> -       return ((freelist_idx_t *) slab->freelist)[idx];
> -}
> -
> -static inline void set_free_obj(struct slab *slab,
> -                                       unsigned int idx, freelist_idx_t val)
> -{
> -       ((freelist_idx_t *)(slab->freelist))[idx] = val;
> -}
> -
> -static void cache_init_objs_debug(struct kmem_cache *cachep, struct slab *slab)
> -{
> -#if DEBUG
> -       int i;
> -
> -       for (i = 0; i < cachep->num; i++) {
> -               void *objp = index_to_obj(cachep, slab, i);
> -
> -               if (cachep->flags & SLAB_STORE_USER)
> -                       *dbg_userword(cachep, objp) = NULL;
> -
> -               if (cachep->flags & SLAB_RED_ZONE) {
> -                       *dbg_redzone1(cachep, objp) = RED_INACTIVE;
> -                       *dbg_redzone2(cachep, objp) = RED_INACTIVE;
> -               }
> -               /*
> -                * Constructors are not allowed to allocate memory from the same
> -                * cache which they are a constructor for.  Otherwise, deadlock.
> -                * They must also be threaded.
> -                */
> -               if (cachep->ctor && !(cachep->flags & SLAB_POISON)) {
> -                       kasan_unpoison_object_data(cachep,
> -                                                  objp + obj_offset(cachep));
> -                       cachep->ctor(objp + obj_offset(cachep));
> -                       kasan_poison_object_data(
> -                               cachep, objp + obj_offset(cachep));
> -               }
> -
> -               if (cachep->flags & SLAB_RED_ZONE) {
> -                       if (*dbg_redzone2(cachep, objp) != RED_INACTIVE)
> -                               slab_error(cachep, "constructor overwrote the end of an object");
> -                       if (*dbg_redzone1(cachep, objp) != RED_INACTIVE)
> -                               slab_error(cachep, "constructor overwrote the start of an object");
> -               }
> -               /* need to poison the objs? */
> -               if (cachep->flags & SLAB_POISON) {
> -                       poison_obj(cachep, objp, POISON_FREE);
> -                       slab_kernel_map(cachep, objp, 0);
> -               }
> -       }
> -#endif
> -}
> -
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -/* Hold information during a freelist initialization */
> -union freelist_init_state {
> -       struct {
> -               unsigned int pos;
> -               unsigned int *list;
> -               unsigned int count;
> -       };
> -       struct rnd_state rnd_state;
> -};
> -
> -/*
> - * Initialize the state based on the randomization method available.
> - * return true if the pre-computed list is available, false otherwise.
> - */
> -static bool freelist_state_initialize(union freelist_init_state *state,
> -                               struct kmem_cache *cachep,
> -                               unsigned int count)
> -{
> -       bool ret;
> -       unsigned int rand;
> -
> -       /* Use best entropy available to define a random shift */
> -       rand = get_random_u32();
> -
> -       /* Use a random state if the pre-computed list is not available */
> -       if (!cachep->random_seq) {
> -               prandom_seed_state(&state->rnd_state, rand);
> -               ret = false;
> -       } else {
> -               state->list = cachep->random_seq;
> -               state->count = count;
> -               state->pos = rand % count;
> -               ret = true;
> -       }
> -       return ret;
> -}
> -
> -/* Get the next entry on the list and randomize it using a random shift */
> -static freelist_idx_t next_random_slot(union freelist_init_state *state)
> -{
> -       if (state->pos >= state->count)
> -               state->pos = 0;
> -       return state->list[state->pos++];
> -}
> -
> -/* Swap two freelist entries */
> -static void swap_free_obj(struct slab *slab, unsigned int a, unsigned int b)
> -{
> -       swap(((freelist_idx_t *) slab->freelist)[a],
> -               ((freelist_idx_t *) slab->freelist)[b]);
> -}
> -
> -/*
> - * Shuffle the freelist initialization state based on pre-computed lists.
> - * return true if the list was successfully shuffled, false otherwise.
> - */
> -static bool shuffle_freelist(struct kmem_cache *cachep, struct slab *slab)
> -{
> -       unsigned int objfreelist = 0, i, rand, count = cachep->num;
> -       union freelist_init_state state;
> -       bool precomputed;
> -
> -       if (count < 2)
> -               return false;
> -
> -       precomputed = freelist_state_initialize(&state, cachep, count);
> -
> -       /* Take a random entry as the objfreelist */
> -       if (OBJFREELIST_SLAB(cachep)) {
> -               if (!precomputed)
> -                       objfreelist = count - 1;
> -               else
> -                       objfreelist = next_random_slot(&state);
> -               slab->freelist = index_to_obj(cachep, slab, objfreelist) +
> -                                               obj_offset(cachep);
> -               count--;
> -       }
> -
> -       /*
> -        * On early boot, generate the list dynamically.
> -        * Later use a pre-computed list for speed.
> -        */
> -       if (!precomputed) {
> -               for (i = 0; i < count; i++)
> -                       set_free_obj(slab, i, i);
> -
> -               /* Fisher-Yates shuffle */
> -               for (i = count - 1; i > 0; i--) {
> -                       rand = prandom_u32_state(&state.rnd_state);
> -                       rand %= (i + 1);
> -                       swap_free_obj(slab, i, rand);
> -               }
> -       } else {
> -               for (i = 0; i < count; i++)
> -                       set_free_obj(slab, i, next_random_slot(&state));
> -       }
> -
> -       if (OBJFREELIST_SLAB(cachep))
> -               set_free_obj(slab, cachep->num - 1, objfreelist);
> -
> -       return true;
> -}
> -#else
> -static inline bool shuffle_freelist(struct kmem_cache *cachep,
> -                               struct slab *slab)
> -{
> -       return false;
> -}
> -#endif /* CONFIG_SLAB_FREELIST_RANDOM */
> -
> -static void cache_init_objs(struct kmem_cache *cachep,
> -                           struct slab *slab)
> -{
> -       int i;
> -       void *objp;
> -       bool shuffled;
> -
> -       cache_init_objs_debug(cachep, slab);
> -
> -       /* Try to randomize the freelist if enabled */
> -       shuffled = shuffle_freelist(cachep, slab);
> -
> -       if (!shuffled && OBJFREELIST_SLAB(cachep)) {
> -               slab->freelist = index_to_obj(cachep, slab, cachep->num - 1) +
> -                                               obj_offset(cachep);
> -       }
> -
> -       for (i = 0; i < cachep->num; i++) {
> -               objp = index_to_obj(cachep, slab, i);
> -               objp = kasan_init_slab_obj(cachep, objp);
> -
> -               /* constructor could break poison info */
> -               if (DEBUG == 0 && cachep->ctor) {
> -                       kasan_unpoison_object_data(cachep, objp);
> -                       cachep->ctor(objp);
> -                       kasan_poison_object_data(cachep, objp);
> -               }
> -
> -               if (!shuffled)
> -                       set_free_obj(slab, i, i);
> -       }
> -}
> -
> -static void *slab_get_obj(struct kmem_cache *cachep, struct slab *slab)
> -{
> -       void *objp;
> -
> -       objp = index_to_obj(cachep, slab, get_free_obj(slab, slab->active));
> -       slab->active++;
> -
> -       return objp;
> -}
> -
> -static void slab_put_obj(struct kmem_cache *cachep,
> -                       struct slab *slab, void *objp)
> -{
> -       unsigned int objnr = obj_to_index(cachep, slab, objp);
> -#if DEBUG
> -       unsigned int i;
> -
> -       /* Verify double free bug */
> -       for (i = slab->active; i < cachep->num; i++) {
> -               if (get_free_obj(slab, i) == objnr) {
> -                       pr_err("slab: double free detected in cache '%s', objp %px\n",
> -                              cachep->name, objp);
> -                       BUG();
> -               }
> -       }
> -#endif
> -       slab->active--;
> -       if (!slab->freelist)
> -               slab->freelist = objp + obj_offset(cachep);
> -
> -       set_free_obj(slab, slab->active, objnr);
> -}
> -
> -/*
> - * Grow (by 1) the number of slabs within a cache.  This is called by
> - * kmem_cache_alloc() when there are no active objs left in a cache.
> - */
> -static struct slab *cache_grow_begin(struct kmem_cache *cachep,
> -                               gfp_t flags, int nodeid)
> -{
> -       void *freelist;
> -       size_t offset;
> -       gfp_t local_flags;
> -       int slab_node;
> -       struct kmem_cache_node *n;
> -       struct slab *slab;
> -
> -       /*
> -        * Be lazy and only check for valid flags here,  keeping it out of the
> -        * critical path in kmem_cache_alloc().
> -        */
> -       if (unlikely(flags & GFP_SLAB_BUG_MASK))
> -               flags = kmalloc_fix_flags(flags);
> -
> -       WARN_ON_ONCE(cachep->ctor && (flags & __GFP_ZERO));
> -       local_flags = flags & (GFP_CONSTRAINT_MASK|GFP_RECLAIM_MASK);
> -
> -       check_irq_off();
> -       if (gfpflags_allow_blocking(local_flags))
> -               local_irq_enable();
> -
> -       /*
> -        * Get mem for the objs.  Attempt to allocate a physical page from
> -        * 'nodeid'.
> -        */
> -       slab = kmem_getpages(cachep, local_flags, nodeid);
> -       if (!slab)
> -               goto failed;
> -
> -       slab_node = slab_nid(slab);
> -       n = get_node(cachep, slab_node);
> -
> -       /* Get colour for the slab, and cal the next value. */
> -       n->colour_next++;
> -       if (n->colour_next >= cachep->colour)
> -               n->colour_next = 0;
> -
> -       offset = n->colour_next;
> -       if (offset >= cachep->colour)
> -               offset = 0;
> -
> -       offset *= cachep->colour_off;
> -
> -       /*
> -        * Call kasan_poison_slab() before calling alloc_slabmgmt(), so
> -        * page_address() in the latter returns a non-tagged pointer,
> -        * as it should be for slab pages.
> -        */
> -       kasan_poison_slab(slab);
> -
> -       /* Get slab management. */
> -       freelist = alloc_slabmgmt(cachep, slab, offset,
> -                       local_flags & ~GFP_CONSTRAINT_MASK, slab_node);
> -       if (OFF_SLAB(cachep) && !freelist)
> -               goto opps1;
> -
> -       slab->slab_cache = cachep;
> -       slab->freelist = freelist;
> -
> -       cache_init_objs(cachep, slab);
> -
> -       if (gfpflags_allow_blocking(local_flags))
> -               local_irq_disable();
> -
> -       return slab;
> -
> -opps1:
> -       kmem_freepages(cachep, slab);
> -failed:
> -       if (gfpflags_allow_blocking(local_flags))
> -               local_irq_disable();
> -       return NULL;
> -}
> -
> -static void cache_grow_end(struct kmem_cache *cachep, struct slab *slab)
> -{
> -       struct kmem_cache_node *n;
> -       void *list = NULL;
> -
> -       check_irq_off();
> -
> -       if (!slab)
> -               return;
> -
> -       INIT_LIST_HEAD(&slab->slab_list);
> -       n = get_node(cachep, slab_nid(slab));
> -
> -       raw_spin_lock(&n->list_lock);
> -       n->total_slabs++;
> -       if (!slab->active) {
> -               list_add_tail(&slab->slab_list, &n->slabs_free);
> -               n->free_slabs++;
> -       } else
> -               fixup_slab_list(cachep, n, slab, &list);
> -
> -       STATS_INC_GROWN(cachep);
> -       n->free_objects += cachep->num - slab->active;
> -       raw_spin_unlock(&n->list_lock);
> -
> -       fixup_objfreelist_debug(cachep, &list);
> -}
> -
> -#if DEBUG
> -
> -/*
> - * Perform extra freeing checks:
> - * - detect bad pointers.
> - * - POISON/RED_ZONE checking
> - */
> -static void kfree_debugcheck(const void *objp)
> -{
> -       if (!virt_addr_valid(objp)) {
> -               pr_err("kfree_debugcheck: out of range ptr %lxh\n",
> -                      (unsigned long)objp);
> -               BUG();
> -       }
> -}
> -
> -static inline void verify_redzone_free(struct kmem_cache *cache, void *obj)
> -{
> -       unsigned long long redzone1, redzone2;
> -
> -       redzone1 = *dbg_redzone1(cache, obj);
> -       redzone2 = *dbg_redzone2(cache, obj);
> -
> -       /*
> -        * Redzone is ok.
> -        */
> -       if (redzone1 == RED_ACTIVE && redzone2 == RED_ACTIVE)
> -               return;
> -
> -       if (redzone1 == RED_INACTIVE && redzone2 == RED_INACTIVE)
> -               slab_error(cache, "double free detected");
> -       else
> -               slab_error(cache, "memory outside object was overwritten");
> -
> -       pr_err("%px: redzone 1:0x%llx, redzone 2:0x%llx\n",
> -              obj, redzone1, redzone2);
> -}
> -
> -static void *cache_free_debugcheck(struct kmem_cache *cachep, void *objp,
> -                                  unsigned long caller)
> -{
> -       unsigned int objnr;
> -       struct slab *slab;
> -
> -       BUG_ON(virt_to_cache(objp) != cachep);
> -
> -       objp -= obj_offset(cachep);
> -       kfree_debugcheck(objp);
> -       slab = virt_to_slab(objp);
> -
> -       if (cachep->flags & SLAB_RED_ZONE) {
> -               verify_redzone_free(cachep, objp);
> -               *dbg_redzone1(cachep, objp) = RED_INACTIVE;
> -               *dbg_redzone2(cachep, objp) = RED_INACTIVE;
> -       }
> -       if (cachep->flags & SLAB_STORE_USER)
> -               *dbg_userword(cachep, objp) = (void *)caller;
> -
> -       objnr = obj_to_index(cachep, slab, objp);
> -
> -       BUG_ON(objnr >= cachep->num);
> -       BUG_ON(objp != index_to_obj(cachep, slab, objnr));
> -
> -       if (cachep->flags & SLAB_POISON) {
> -               poison_obj(cachep, objp, POISON_FREE);
> -               slab_kernel_map(cachep, objp, 0);
> -       }
> -       return objp;
> -}
> -
> -#else
> -#define kfree_debugcheck(x) do { } while(0)
> -#define cache_free_debugcheck(x, objp, z) (objp)
> -#endif
> -
> -static inline void fixup_objfreelist_debug(struct kmem_cache *cachep,
> -                                               void **list)
> -{
> -#if DEBUG
> -       void *next = *list;
> -       void *objp;
> -
> -       while (next) {
> -               objp = next - obj_offset(cachep);
> -               next = *(void **)next;
> -               poison_obj(cachep, objp, POISON_FREE);
> -       }
> -#endif
> -}
> -
> -static inline void fixup_slab_list(struct kmem_cache *cachep,
> -                               struct kmem_cache_node *n, struct slab *slab,
> -                               void **list)
> -{
> -       /* move slabp to correct slabp list: */
> -       list_del(&slab->slab_list);
> -       if (slab->active == cachep->num) {
> -               list_add(&slab->slab_list, &n->slabs_full);
> -               if (OBJFREELIST_SLAB(cachep)) {
> -#if DEBUG
> -                       /* Poisoning will be done without holding the lock */
> -                       if (cachep->flags & SLAB_POISON) {
> -                               void **objp = slab->freelist;
> -
> -                               *objp = *list;
> -                               *list = objp;
> -                       }
> -#endif
> -                       slab->freelist = NULL;
> -               }
> -       } else
> -               list_add(&slab->slab_list, &n->slabs_partial);
> -}
> -
> -/* Try to find non-pfmemalloc slab if needed */
> -static noinline struct slab *get_valid_first_slab(struct kmem_cache_node *n,
> -                                       struct slab *slab, bool pfmemalloc)
> -{
> -       if (!slab)
> -               return NULL;
> -
> -       if (pfmemalloc)
> -               return slab;
> -
> -       if (!slab_test_pfmemalloc(slab))
> -               return slab;
> -
> -       /* No need to keep pfmemalloc slab if we have enough free objects */
> -       if (n->free_objects > n->free_limit) {
> -               slab_clear_pfmemalloc(slab);
> -               return slab;
> -       }
> -
> -       /* Move pfmemalloc slab to the end of list to speed up next search */
> -       list_del(&slab->slab_list);
> -       if (!slab->active) {
> -               list_add_tail(&slab->slab_list, &n->slabs_free);
> -               n->free_slabs++;
> -       } else
> -               list_add_tail(&slab->slab_list, &n->slabs_partial);
> -
> -       list_for_each_entry(slab, &n->slabs_partial, slab_list) {
> -               if (!slab_test_pfmemalloc(slab))
> -                       return slab;
> -       }
> -
> -       n->free_touched = 1;
> -       list_for_each_entry(slab, &n->slabs_free, slab_list) {
> -               if (!slab_test_pfmemalloc(slab)) {
> -                       n->free_slabs--;
> -                       return slab;
> -               }
> -       }
> -
> -       return NULL;
> -}
> -
> -static struct slab *get_first_slab(struct kmem_cache_node *n, bool pfmemalloc)
> -{
> -       struct slab *slab;
> -
> -       assert_raw_spin_locked(&n->list_lock);
> -       slab = list_first_entry_or_null(&n->slabs_partial, struct slab,
> -                                       slab_list);
> -       if (!slab) {
> -               n->free_touched = 1;
> -               slab = list_first_entry_or_null(&n->slabs_free, struct slab,
> -                                               slab_list);
> -               if (slab)
> -                       n->free_slabs--;
> -       }
> -
> -       if (sk_memalloc_socks())
> -               slab = get_valid_first_slab(n, slab, pfmemalloc);
> -
> -       return slab;
> -}
> -
> -static noinline void *cache_alloc_pfmemalloc(struct kmem_cache *cachep,
> -                               struct kmem_cache_node *n, gfp_t flags)
> -{
> -       struct slab *slab;
> -       void *obj;
> -       void *list = NULL;
> -
> -       if (!gfp_pfmemalloc_allowed(flags))
> -               return NULL;
> -
> -       raw_spin_lock(&n->list_lock);
> -       slab = get_first_slab(n, true);
> -       if (!slab) {
> -               raw_spin_unlock(&n->list_lock);
> -               return NULL;
> -       }
> -
> -       obj = slab_get_obj(cachep, slab);
> -       n->free_objects--;
> -
> -       fixup_slab_list(cachep, n, slab, &list);
> -
> -       raw_spin_unlock(&n->list_lock);
> -       fixup_objfreelist_debug(cachep, &list);
> -
> -       return obj;
> -}
> -
> -/*
> - * Slab list should be fixed up by fixup_slab_list() for existing slab
> - * or cache_grow_end() for new slab
> - */
> -static __always_inline int alloc_block(struct kmem_cache *cachep,
> -               struct array_cache *ac, struct slab *slab, int batchcount)
> -{
> -       /*
> -        * There must be at least one object available for
> -        * allocation.
> -        */
> -       BUG_ON(slab->active >= cachep->num);
> -
> -       while (slab->active < cachep->num && batchcount--) {
> -               STATS_INC_ALLOCED(cachep);
> -               STATS_INC_ACTIVE(cachep);
> -               STATS_SET_HIGH(cachep);
> -
> -               ac->entry[ac->avail++] = slab_get_obj(cachep, slab);
> -       }
> -
> -       return batchcount;
> -}
> -
> -static void *cache_alloc_refill(struct kmem_cache *cachep, gfp_t flags)
> -{
> -       int batchcount;
> -       struct kmem_cache_node *n;
> -       struct array_cache *ac, *shared;
> -       int node;
> -       void *list = NULL;
> -       struct slab *slab;
> -
> -       check_irq_off();
> -       node = numa_mem_id();
> -
> -       ac = cpu_cache_get(cachep);
> -       batchcount = ac->batchcount;
> -       if (!ac->touched && batchcount > BATCHREFILL_LIMIT) {
> -               /*
> -                * If there was little recent activity on this cache, then
> -                * perform only a partial refill.  Otherwise we could generate
> -                * refill bouncing.
> -                */
> -               batchcount = BATCHREFILL_LIMIT;
> -       }
> -       n = get_node(cachep, node);
> -
> -       BUG_ON(ac->avail > 0 || !n);
> -       shared = READ_ONCE(n->shared);
> -       if (!n->free_objects && (!shared || !shared->avail))
> -               goto direct_grow;
> -
> -       raw_spin_lock(&n->list_lock);
> -       shared = READ_ONCE(n->shared);
> -
> -       /* See if we can refill from the shared array */
> -       if (shared && transfer_objects(ac, shared, batchcount)) {
> -               shared->touched = 1;
> -               goto alloc_done;
> -       }
> -
> -       while (batchcount > 0) {
> -               /* Get slab alloc is to come from. */
> -               slab = get_first_slab(n, false);
> -               if (!slab)
> -                       goto must_grow;
> -
> -               check_spinlock_acquired(cachep);
> -
> -               batchcount = alloc_block(cachep, ac, slab, batchcount);
> -               fixup_slab_list(cachep, n, slab, &list);
> -       }
> -
> -must_grow:
> -       n->free_objects -= ac->avail;
> -alloc_done:
> -       raw_spin_unlock(&n->list_lock);
> -       fixup_objfreelist_debug(cachep, &list);
> -
> -direct_grow:
> -       if (unlikely(!ac->avail)) {
> -               /* Check if we can use obj in pfmemalloc slab */
> -               if (sk_memalloc_socks()) {
> -                       void *obj = cache_alloc_pfmemalloc(cachep, n, flags);
> -
> -                       if (obj)
> -                               return obj;
> -               }
> -
> -               slab = cache_grow_begin(cachep, gfp_exact_node(flags), node);
> -
> -               /*
> -                * cache_grow_begin() can reenable interrupts,
> -                * then ac could change.
> -                */
> -               ac = cpu_cache_get(cachep);
> -               if (!ac->avail && slab)
> -                       alloc_block(cachep, ac, slab, batchcount);
> -               cache_grow_end(cachep, slab);
> -
> -               if (!ac->avail)
> -                       return NULL;
> -       }
> -       ac->touched = 1;
> -
> -       return ac->entry[--ac->avail];
> -}
> -
> -#if DEBUG
> -static void *cache_alloc_debugcheck_after(struct kmem_cache *cachep,
> -                               gfp_t flags, void *objp, unsigned long caller)
> -{
> -       WARN_ON_ONCE(cachep->ctor && (flags & __GFP_ZERO));
> -       if (!objp || is_kfence_address(objp))
> -               return objp;
> -       if (cachep->flags & SLAB_POISON) {
> -               check_poison_obj(cachep, objp);
> -               slab_kernel_map(cachep, objp, 1);
> -               poison_obj(cachep, objp, POISON_INUSE);
> -       }
> -       if (cachep->flags & SLAB_STORE_USER)
> -               *dbg_userword(cachep, objp) = (void *)caller;
> -
> -       if (cachep->flags & SLAB_RED_ZONE) {
> -               if (*dbg_redzone1(cachep, objp) != RED_INACTIVE ||
> -                               *dbg_redzone2(cachep, objp) != RED_INACTIVE) {
> -                       slab_error(cachep, "double free, or memory outside object was overwritten");
> -                       pr_err("%px: redzone 1:0x%llx, redzone 2:0x%llx\n",
> -                              objp, *dbg_redzone1(cachep, objp),
> -                              *dbg_redzone2(cachep, objp));
> -               }
> -               *dbg_redzone1(cachep, objp) = RED_ACTIVE;
> -               *dbg_redzone2(cachep, objp) = RED_ACTIVE;
> -       }
> -
> -       objp += obj_offset(cachep);
> -       if (cachep->ctor && cachep->flags & SLAB_POISON)
> -               cachep->ctor(objp);
> -       if ((unsigned long)objp & (arch_slab_minalign() - 1)) {
> -               pr_err("0x%px: not aligned to arch_slab_minalign()=%u\n", objp,
> -                      arch_slab_minalign());
> -       }
> -       return objp;
> -}
> -#else
> -#define cache_alloc_debugcheck_after(a, b, objp, d) (objp)
> -#endif
> -
> -static inline void *____cache_alloc(struct kmem_cache *cachep, gfp_t flags)
> -{
> -       void *objp;
> -       struct array_cache *ac;
> -
> -       check_irq_off();
> -
> -       ac = cpu_cache_get(cachep);
> -       if (likely(ac->avail)) {
> -               ac->touched = 1;
> -               objp = ac->entry[--ac->avail];
> -
> -               STATS_INC_ALLOCHIT(cachep);
> -               goto out;
> -       }
> -
> -       STATS_INC_ALLOCMISS(cachep);
> -       objp = cache_alloc_refill(cachep, flags);
> -       /*
> -        * the 'ac' may be updated by cache_alloc_refill(),
> -        * and kmemleak_erase() requires its correct value.
> -        */
> -       ac = cpu_cache_get(cachep);
> -
> -out:
> -       /*
> -        * To avoid a false negative, if an object that is in one of the
> -        * per-CPU caches is leaked, we need to make sure kmemleak doesn't
> -        * treat the array pointers as a reference to the object.
> -        */
> -       if (objp)
> -               kmemleak_erase(&ac->entry[ac->avail]);
> -       return objp;
> -}
> -
> -#ifdef CONFIG_NUMA
> -static void *____cache_alloc_node(struct kmem_cache *, gfp_t, int);
> -
> -/*
> - * Try allocating on another node if PFA_SPREAD_SLAB is a mempolicy is set.
> - *
> - * If we are in_interrupt, then process context, including cpusets and
> - * mempolicy, may not apply and should not be used for allocation policy.
> - */
> -static void *alternate_node_alloc(struct kmem_cache *cachep, gfp_t flags)
> -{
> -       int nid_alloc, nid_here;
> -
> -       if (in_interrupt() || (flags & __GFP_THISNODE))
> -               return NULL;
> -       nid_alloc = nid_here = numa_mem_id();
> -       if (cpuset_do_slab_mem_spread() && (cachep->flags & SLAB_MEM_SPREAD))
> -               nid_alloc = cpuset_slab_spread_node();
> -       else if (current->mempolicy)
> -               nid_alloc = mempolicy_slab_node();
> -       if (nid_alloc != nid_here)
> -               return ____cache_alloc_node(cachep, flags, nid_alloc);
> -       return NULL;
> -}
> -
> -/*
> - * Fallback function if there was no memory available and no objects on a
> - * certain node and fall back is permitted. First we scan all the
> - * available node for available objects. If that fails then we
> - * perform an allocation without specifying a node. This allows the page
> - * allocator to do its reclaim / fallback magic. We then insert the
> - * slab into the proper nodelist and then allocate from it.
> - */
> -static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
> -{
> -       struct zonelist *zonelist;
> -       struct zoneref *z;
> -       struct zone *zone;
> -       enum zone_type highest_zoneidx = gfp_zone(flags);
> -       void *obj = NULL;
> -       struct slab *slab;
> -       int nid;
> -       unsigned int cpuset_mems_cookie;
> -
> -       if (flags & __GFP_THISNODE)
> -               return NULL;
> -
> -retry_cpuset:
> -       cpuset_mems_cookie = read_mems_allowed_begin();
> -       zonelist = node_zonelist(mempolicy_slab_node(), flags);
> -
> -retry:
> -       /*
> -        * Look through allowed nodes for objects available
> -        * from existing per node queues.
> -        */
> -       for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) {
> -               nid = zone_to_nid(zone);
> -
> -               if (cpuset_zone_allowed(zone, flags) &&
> -                       get_node(cache, nid) &&
> -                       get_node(cache, nid)->free_objects) {
> -                               obj = ____cache_alloc_node(cache,
> -                                       gfp_exact_node(flags), nid);
> -                               if (obj)
> -                                       break;
> -               }
> -       }
> -
> -       if (!obj) {
> -               /*
> -                * This allocation will be performed within the constraints
> -                * of the current cpuset / memory policy requirements.
> -                * We may trigger various forms of reclaim on the allowed
> -                * set and go into memory reserves if necessary.
> -                */
> -               slab = cache_grow_begin(cache, flags, numa_mem_id());
> -               cache_grow_end(cache, slab);
> -               if (slab) {
> -                       nid = slab_nid(slab);
> -                       obj = ____cache_alloc_node(cache,
> -                               gfp_exact_node(flags), nid);
> -
> -                       /*
> -                        * Another processor may allocate the objects in
> -                        * the slab since we are not holding any locks.
> -                        */
> -                       if (!obj)
> -                               goto retry;
> -               }
> -       }
> -
> -       if (unlikely(!obj && read_mems_allowed_retry(cpuset_mems_cookie)))
> -               goto retry_cpuset;
> -       return obj;
> -}
> -
> -/*
> - * An interface to enable slab creation on nodeid
> - */
> -static void *____cache_alloc_node(struct kmem_cache *cachep, gfp_t flags,
> -                               int nodeid)
> -{
> -       struct slab *slab;
> -       struct kmem_cache_node *n;
> -       void *obj = NULL;
> -       void *list = NULL;
> -
> -       VM_BUG_ON(nodeid < 0 || nodeid >= MAX_NUMNODES);
> -       n = get_node(cachep, nodeid);
> -       BUG_ON(!n);
> -
> -       check_irq_off();
> -       raw_spin_lock(&n->list_lock);
> -       slab = get_first_slab(n, false);
> -       if (!slab)
> -               goto must_grow;
> -
> -       check_spinlock_acquired_node(cachep, nodeid);
> -
> -       STATS_INC_NODEALLOCS(cachep);
> -       STATS_INC_ACTIVE(cachep);
> -       STATS_SET_HIGH(cachep);
> -
> -       BUG_ON(slab->active == cachep->num);
> -
> -       obj = slab_get_obj(cachep, slab);
> -       n->free_objects--;
> -
> -       fixup_slab_list(cachep, n, slab, &list);
> -
> -       raw_spin_unlock(&n->list_lock);
> -       fixup_objfreelist_debug(cachep, &list);
> -       return obj;
> -
> -must_grow:
> -       raw_spin_unlock(&n->list_lock);
> -       slab = cache_grow_begin(cachep, gfp_exact_node(flags), nodeid);
> -       if (slab) {
> -               /* This slab isn't counted yet so don't update free_objects */
> -               obj = slab_get_obj(cachep, slab);
> -       }
> -       cache_grow_end(cachep, slab);
> -
> -       return obj ? obj : fallback_alloc(cachep, flags);
> -}
> -
> -static __always_inline void *
> -__do_cache_alloc(struct kmem_cache *cachep, gfp_t flags, int nodeid)
> -{
> -       void *objp = NULL;
> -       int slab_node = numa_mem_id();
> -
> -       if (nodeid == NUMA_NO_NODE) {
> -               if (current->mempolicy || cpuset_do_slab_mem_spread()) {
> -                       objp = alternate_node_alloc(cachep, flags);
> -                       if (objp)
> -                               goto out;
> -               }
> -               /*
> -                * Use the locally cached objects if possible.
> -                * However ____cache_alloc does not allow fallback
> -                * to other nodes. It may fail while we still have
> -                * objects on other nodes available.
> -                */
> -               objp = ____cache_alloc(cachep, flags);
> -               nodeid = slab_node;
> -       } else if (nodeid == slab_node) {
> -               objp = ____cache_alloc(cachep, flags);
> -       } else if (!get_node(cachep, nodeid)) {
> -               /* Node not bootstrapped yet */
> -               objp = fallback_alloc(cachep, flags);
> -               goto out;
> -       }
> -
> -       /*
> -        * We may just have run out of memory on the local node.
> -        * ____cache_alloc_node() knows how to locate memory on other nodes
> -        */
> -       if (!objp)
> -               objp = ____cache_alloc_node(cachep, flags, nodeid);
> -out:
> -       return objp;
> -}
> -#else
> -
> -static __always_inline void *
> -__do_cache_alloc(struct kmem_cache *cachep, gfp_t flags, int nodeid __maybe_unused)
> -{
> -       return ____cache_alloc(cachep, flags);
> -}
> -
> -#endif /* CONFIG_NUMA */
> -
> -static __always_inline void *
> -slab_alloc_node(struct kmem_cache *cachep, struct list_lru *lru, gfp_t flags,
> -               int nodeid, size_t orig_size, unsigned long caller)
> -{
> -       unsigned long save_flags;
> -       void *objp;
> -       struct obj_cgroup *objcg = NULL;
> -       bool init = false;
> -
> -       flags &= gfp_allowed_mask;
> -       cachep = slab_pre_alloc_hook(cachep, lru, &objcg, 1, flags);
> -       if (unlikely(!cachep))
> -               return NULL;
> -
> -       objp = kfence_alloc(cachep, orig_size, flags);
> -       if (unlikely(objp))
> -               goto out;
> -
> -       local_irq_save(save_flags);
> -       objp = __do_cache_alloc(cachep, flags, nodeid);
> -       local_irq_restore(save_flags);
> -       objp = cache_alloc_debugcheck_after(cachep, flags, objp, caller);
> -       prefetchw(objp);
> -       init = slab_want_init_on_alloc(flags, cachep);
> -
> -out:
> -       slab_post_alloc_hook(cachep, objcg, flags, 1, &objp, init,
> -                               cachep->object_size);
> -       return objp;
> -}
> -
> -static __always_inline void *
> -slab_alloc(struct kmem_cache *cachep, struct list_lru *lru, gfp_t flags,
> -          size_t orig_size, unsigned long caller)
> -{
> -       return slab_alloc_node(cachep, lru, flags, NUMA_NO_NODE, orig_size,
> -                              caller);
> -}
> -
> -/*
> - * Caller needs to acquire correct kmem_cache_node's list_lock
> - * @list: List of detached free slabs should be freed by caller
> - */
> -static void free_block(struct kmem_cache *cachep, void **objpp,
> -                       int nr_objects, int node, struct list_head *list)
> -{
> -       int i;
> -       struct kmem_cache_node *n = get_node(cachep, node);
> -       struct slab *slab;
> -
> -       n->free_objects += nr_objects;
> -
> -       for (i = 0; i < nr_objects; i++) {
> -               void *objp;
> -               struct slab *slab;
> -
> -               objp = objpp[i];
> -
> -               slab = virt_to_slab(objp);
> -               list_del(&slab->slab_list);
> -               check_spinlock_acquired_node(cachep, node);
> -               slab_put_obj(cachep, slab, objp);
> -               STATS_DEC_ACTIVE(cachep);
> -
> -               /* fixup slab chains */
> -               if (slab->active == 0) {
> -                       list_add(&slab->slab_list, &n->slabs_free);
> -                       n->free_slabs++;
> -               } else {
> -                       /* Unconditionally move a slab to the end of the
> -                        * partial list on free - maximum time for the
> -                        * other objects to be freed, too.
> -                        */
> -                       list_add_tail(&slab->slab_list, &n->slabs_partial);
> -               }
> -       }
> -
> -       while (n->free_objects > n->free_limit && !list_empty(&n->slabs_free)) {
> -               n->free_objects -= cachep->num;
> -
> -               slab = list_last_entry(&n->slabs_free, struct slab, slab_list);
> -               list_move(&slab->slab_list, list);
> -               n->free_slabs--;
> -               n->total_slabs--;
> -       }
> -}
> -
> -static void cache_flusharray(struct kmem_cache *cachep, struct array_cache *ac)
> -{
> -       int batchcount;
> -       struct kmem_cache_node *n;
> -       int node = numa_mem_id();
> -       LIST_HEAD(list);
> -
> -       batchcount = ac->batchcount;
> -
> -       check_irq_off();
> -       n = get_node(cachep, node);
> -       raw_spin_lock(&n->list_lock);
> -       if (n->shared) {
> -               struct array_cache *shared_array = n->shared;
> -               int max = shared_array->limit - shared_array->avail;
> -               if (max) {
> -                       if (batchcount > max)
> -                               batchcount = max;
> -                       memcpy(&(shared_array->entry[shared_array->avail]),
> -                              ac->entry, sizeof(void *) * batchcount);
> -                       shared_array->avail += batchcount;
> -                       goto free_done;
> -               }
> -       }
> -
> -       free_block(cachep, ac->entry, batchcount, node, &list);
> -free_done:
> -#if STATS
> -       {
> -               int i = 0;
> -               struct slab *slab;
> -
> -               list_for_each_entry(slab, &n->slabs_free, slab_list) {
> -                       BUG_ON(slab->active);
> -
> -                       i++;
> -               }
> -               STATS_SET_FREEABLE(cachep, i);
> -       }
> -#endif
> -       raw_spin_unlock(&n->list_lock);
> -       ac->avail -= batchcount;
> -       memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail);
> -       slabs_destroy(cachep, &list);
> -}
> -
> -/*
> - * Release an obj back to its cache. If the obj has a constructed state, it must
> - * be in this state _before_ it is released.  Called with disabled ints.
> - */
> -static __always_inline void __cache_free(struct kmem_cache *cachep, void *objp,
> -                                        unsigned long caller)
> -{
> -       bool init;
> -
> -       memcg_slab_free_hook(cachep, virt_to_slab(objp), &objp, 1);
> -
> -       if (is_kfence_address(objp)) {
> -               kmemleak_free_recursive(objp, cachep->flags);
> -               __kfence_free(objp);
> -               return;
> -       }
> -
> -       /*
> -        * As memory initialization might be integrated into KASAN,
> -        * kasan_slab_free and initialization memset must be
> -        * kept together to avoid discrepancies in behavior.
> -        */
> -       init = slab_want_init_on_free(cachep);
> -       if (init && !kasan_has_integrated_init())
> -               memset(objp, 0, cachep->object_size);
> -       /* KASAN might put objp into memory quarantine, delaying its reuse. */
> -       if (kasan_slab_free(cachep, objp, init))
> -               return;
> -
> -       /* Use KCSAN to help debug racy use-after-free. */
> -       if (!(cachep->flags & SLAB_TYPESAFE_BY_RCU))
> -               __kcsan_check_access(objp, cachep->object_size,
> -                                    KCSAN_ACCESS_WRITE | KCSAN_ACCESS_ASSERT);
> -
> -       ___cache_free(cachep, objp, caller);
> -}
> -
> -void ___cache_free(struct kmem_cache *cachep, void *objp,
> -               unsigned long caller)
> -{
> -       struct array_cache *ac = cpu_cache_get(cachep);
> -
> -       check_irq_off();
> -       kmemleak_free_recursive(objp, cachep->flags);
> -       objp = cache_free_debugcheck(cachep, objp, caller);
> -
> -       /*
> -        * Skip calling cache_free_alien() when the platform is not numa.
> -        * This will avoid cache misses that happen while accessing slabp (which
> -        * is per page memory  reference) to get nodeid. Instead use a global
> -        * variable to skip the call, which is mostly likely to be present in
> -        * the cache.
> -        */
> -       if (nr_online_nodes > 1 && cache_free_alien(cachep, objp))
> -               return;
> -
> -       if (ac->avail < ac->limit) {
> -               STATS_INC_FREEHIT(cachep);
> -       } else {
> -               STATS_INC_FREEMISS(cachep);
> -               cache_flusharray(cachep, ac);
> -       }
> -
> -       if (sk_memalloc_socks()) {
> -               struct slab *slab = virt_to_slab(objp);
> -
> -               if (unlikely(slab_test_pfmemalloc(slab))) {
> -                       cache_free_pfmemalloc(cachep, slab, objp);
> -                       return;
> -               }
> -       }
> -
> -       __free_one(ac, objp);
> -}
> -
> -static __always_inline
> -void *__kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru,
> -                            gfp_t flags)
> -{
> -       void *ret = slab_alloc(cachep, lru, flags, cachep->object_size, _RET_IP_);
> -
> -       trace_kmem_cache_alloc(_RET_IP_, ret, cachep, flags, NUMA_NO_NODE);
> -
> -       return ret;
> -}
> -
> -void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags)
> -{
> -       return __kmem_cache_alloc_lru(cachep, NULL, flags);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc);
> -
> -void *kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru,
> -                          gfp_t flags)
> -{
> -       return __kmem_cache_alloc_lru(cachep, lru, flags);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_lru);
> -
> -static __always_inline void
> -cache_alloc_debugcheck_after_bulk(struct kmem_cache *s, gfp_t flags,
> -                                 size_t size, void **p, unsigned long caller)
> -{
> -       size_t i;
> -
> -       for (i = 0; i < size; i++)
> -               p[i] = cache_alloc_debugcheck_after(s, flags, p[i], caller);
> -}
> -
> -int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> -                         void **p)
> -{
> -       struct obj_cgroup *objcg = NULL;
> -       unsigned long irqflags;
> -       size_t i;
> -
> -       s = slab_pre_alloc_hook(s, NULL, &objcg, size, flags);
> -       if (!s)
> -               return 0;
> -
> -       local_irq_save(irqflags);
> -       for (i = 0; i < size; i++) {
> -               void *objp = kfence_alloc(s, s->object_size, flags) ?:
> -                            __do_cache_alloc(s, flags, NUMA_NO_NODE);
> -
> -               if (unlikely(!objp))
> -                       goto error;
> -               p[i] = objp;
> -       }
> -       local_irq_restore(irqflags);
> -
> -       cache_alloc_debugcheck_after_bulk(s, flags, size, p, _RET_IP_);
> -
> -       /*
> -        * memcg and kmem_cache debug support and memory initialization.
> -        * Done outside of the IRQ disabled section.
> -        */
> -       slab_post_alloc_hook(s, objcg, flags, size, p,
> -                       slab_want_init_on_alloc(flags, s), s->object_size);
> -       /* FIXME: Trace call missing. Christoph would like a bulk variant */
> -       return size;
> -error:
> -       local_irq_restore(irqflags);
> -       cache_alloc_debugcheck_after_bulk(s, flags, i, p, _RET_IP_);
> -       slab_post_alloc_hook(s, objcg, flags, i, p, false, s->object_size);
> -       kmem_cache_free_bulk(s, i, p);
> -       return 0;
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_bulk);
> -
> -/**
> - * kmem_cache_alloc_node - Allocate an object on the specified node
> - * @cachep: The cache to allocate from.
> - * @flags: See kmalloc().
> - * @nodeid: node number of the target node.
> - *
> - * Identical to kmem_cache_alloc but it will allocate memory on the given
> - * node, which can improve the performance for cpu bound structures.
> - *
> - * Fallback to other node is possible if __GFP_THISNODE is not set.
> - *
> - * Return: pointer to the new object or %NULL in case of error
> - */
> -void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid)
> -{
> -       void *ret = slab_alloc_node(cachep, NULL, flags, nodeid, cachep->object_size, _RET_IP_);
> -
> -       trace_kmem_cache_alloc(_RET_IP_, ret, cachep, flags, nodeid);
> -
> -       return ret;
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_node);
> -
> -void *__kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t flags,
> -                            int nodeid, size_t orig_size,
> -                            unsigned long caller)
> -{
> -       return slab_alloc_node(cachep, NULL, flags, nodeid,
> -                              orig_size, caller);
> -}
> -
> -#ifdef CONFIG_PRINTK
> -void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> -{
> -       struct kmem_cache *cachep;
> -       unsigned int objnr;
> -       void *objp;
> -
> -       kpp->kp_ptr = object;
> -       kpp->kp_slab = slab;
> -       cachep = slab->slab_cache;
> -       kpp->kp_slab_cache = cachep;
> -       objp = object - obj_offset(cachep);
> -       kpp->kp_data_offset = obj_offset(cachep);
> -       slab = virt_to_slab(objp);
> -       objnr = obj_to_index(cachep, slab, objp);
> -       objp = index_to_obj(cachep, slab, objnr);
> -       kpp->kp_objp = objp;
> -       if (DEBUG && cachep->flags & SLAB_STORE_USER)
> -               kpp->kp_ret = *dbg_userword(cachep, objp);
> -}
> -#endif
> -
> -static __always_inline
> -void __do_kmem_cache_free(struct kmem_cache *cachep, void *objp,
> -                         unsigned long caller)
> -{
> -       unsigned long flags;
> -
> -       local_irq_save(flags);
> -       debug_check_no_locks_freed(objp, cachep->object_size);
> -       if (!(cachep->flags & SLAB_DEBUG_OBJECTS))
> -               debug_check_no_obj_freed(objp, cachep->object_size);
> -       __cache_free(cachep, objp, caller);
> -       local_irq_restore(flags);
> -}
> -
> -void __kmem_cache_free(struct kmem_cache *cachep, void *objp,
> -                      unsigned long caller)
> -{
> -       __do_kmem_cache_free(cachep, objp, caller);
> -}
> -
> -/**
> - * kmem_cache_free - Deallocate an object
> - * @cachep: The cache the allocation was from.
> - * @objp: The previously allocated object.
> - *
> - * Free an object which was previously allocated from this
> - * cache.
> - */
> -void kmem_cache_free(struct kmem_cache *cachep, void *objp)
> -{
> -       cachep = cache_from_obj(cachep, objp);
> -       if (!cachep)
> -               return;
> -
> -       trace_kmem_cache_free(_RET_IP_, objp, cachep);
> -       __do_kmem_cache_free(cachep, objp, _RET_IP_);
> -}
> -EXPORT_SYMBOL(kmem_cache_free);
> -
> -void kmem_cache_free_bulk(struct kmem_cache *orig_s, size_t size, void **p)
> -{
> -       unsigned long flags;
> -
> -       local_irq_save(flags);
> -       for (int i = 0; i < size; i++) {
> -               void *objp = p[i];
> -               struct kmem_cache *s;
> -
> -               if (!orig_s) {
> -                       struct folio *folio = virt_to_folio(objp);
> -
> -                       /* called via kfree_bulk */
> -                       if (!folio_test_slab(folio)) {
> -                               local_irq_restore(flags);
> -                               free_large_kmalloc(folio, objp);
> -                               local_irq_save(flags);
> -                               continue;
> -                       }
> -                       s = folio_slab(folio)->slab_cache;
> -               } else {
> -                       s = cache_from_obj(orig_s, objp);
> -               }
> -
> -               if (!s)
> -                       continue;
> -
> -               debug_check_no_locks_freed(objp, s->object_size);
> -               if (!(s->flags & SLAB_DEBUG_OBJECTS))
> -                       debug_check_no_obj_freed(objp, s->object_size);
> -
> -               __cache_free(s, objp, _RET_IP_);
> -       }
> -       local_irq_restore(flags);
> -
> -       /* FIXME: add tracing */
> -}
> -EXPORT_SYMBOL(kmem_cache_free_bulk);
> -
> -/*
> - * This initializes kmem_cache_node or resizes various caches for all nodes.
> - */
> -static int setup_kmem_cache_nodes(struct kmem_cache *cachep, gfp_t gfp)
> -{
> -       int ret;
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_online_node(node) {
> -               ret = setup_kmem_cache_node(cachep, node, gfp, true);
> -               if (ret)
> -                       goto fail;
> -
> -       }
> -
> -       return 0;
> -
> -fail:
> -       if (!cachep->list.next) {
> -               /* Cache is not active yet. Roll back what we did */
> -               node--;
> -               while (node >= 0) {
> -                       n = get_node(cachep, node);
> -                       if (n) {
> -                               kfree(n->shared);
> -                               free_alien_cache(n->alien);
> -                               kfree(n);
> -                               cachep->node[node] = NULL;
> -                       }
> -                       node--;
> -               }
> -       }
> -       return -ENOMEM;
> -}
> -
> -/* Always called with the slab_mutex held */
> -static int do_tune_cpucache(struct kmem_cache *cachep, int limit,
> -                           int batchcount, int shared, gfp_t gfp)
> -{
> -       struct array_cache __percpu *cpu_cache, *prev;
> -       int cpu;
> -
> -       cpu_cache = alloc_kmem_cache_cpus(cachep, limit, batchcount);
> -       if (!cpu_cache)
> -               return -ENOMEM;
> -
> -       prev = cachep->cpu_cache;
> -       cachep->cpu_cache = cpu_cache;
> -       /*
> -        * Without a previous cpu_cache there's no need to synchronize remote
> -        * cpus, so skip the IPIs.
> -        */
> -       if (prev)
> -               kick_all_cpus_sync();
> -
> -       check_irq_on();
> -       cachep->batchcount = batchcount;
> -       cachep->limit = limit;
> -       cachep->shared = shared;
> -
> -       if (!prev)
> -               goto setup_node;
> -
> -       for_each_online_cpu(cpu) {
> -               LIST_HEAD(list);
> -               int node;
> -               struct kmem_cache_node *n;
> -               struct array_cache *ac = per_cpu_ptr(prev, cpu);
> -
> -               node = cpu_to_mem(cpu);
> -               n = get_node(cachep, node);
> -               raw_spin_lock_irq(&n->list_lock);
> -               free_block(cachep, ac->entry, ac->avail, node, &list);
> -               raw_spin_unlock_irq(&n->list_lock);
> -               slabs_destroy(cachep, &list);
> -       }
> -       free_percpu(prev);
> -
> -setup_node:
> -       return setup_kmem_cache_nodes(cachep, gfp);
> -}
> -
> -/* Called with slab_mutex held always */
> -static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp)
> -{
> -       int err;
> -       int limit = 0;
> -       int shared = 0;
> -       int batchcount = 0;
> -
> -       err = cache_random_seq_create(cachep, cachep->num, gfp);
> -       if (err)
> -               goto end;
> -
> -       /*
> -        * The head array serves three purposes:
> -        * - create a LIFO ordering, i.e. return objects that are cache-warm
> -        * - reduce the number of spinlock operations.
> -        * - reduce the number of linked list operations on the slab and
> -        *   bufctl chains: array operations are cheaper.
> -        * The numbers are guessed, we should auto-tune as described by
> -        * Bonwick.
> -        */
> -       if (cachep->size > 131072)
> -               limit = 1;
> -       else if (cachep->size > PAGE_SIZE)
> -               limit = 8;
> -       else if (cachep->size > 1024)
> -               limit = 24;
> -       else if (cachep->size > 256)
> -               limit = 54;
> -       else
> -               limit = 120;
> -
> -       /*
> -        * CPU bound tasks (e.g. network routing) can exhibit cpu bound
> -        * allocation behaviour: Most allocs on one cpu, most free operations
> -        * on another cpu. For these cases, an efficient object passing between
> -        * cpus is necessary. This is provided by a shared array. The array
> -        * replaces Bonwick's magazine layer.
> -        * On uniprocessor, it's functionally equivalent (but less efficient)
> -        * to a larger limit. Thus disabled by default.
> -        */
> -       shared = 0;
> -       if (cachep->size <= PAGE_SIZE && num_possible_cpus() > 1)
> -               shared = 8;
> -
> -#if DEBUG
> -       /*
> -        * With debugging enabled, large batchcount lead to excessively long
> -        * periods with disabled local interrupts. Limit the batchcount
> -        */
> -       if (limit > 32)
> -               limit = 32;
> -#endif
> -       batchcount = (limit + 1) / 2;
> -       err = do_tune_cpucache(cachep, limit, batchcount, shared, gfp);
> -end:
> -       if (err)
> -               pr_err("enable_cpucache failed for %s, error %d\n",
> -                      cachep->name, -err);
> -       return err;
> -}
> -
> -/*
> - * Drain an array if it contains any elements taking the node lock only if
> - * necessary. Note that the node listlock also protects the array_cache
> - * if drain_array() is used on the shared array.
> - */
> -static void drain_array(struct kmem_cache *cachep, struct kmem_cache_node *n,
> -                        struct array_cache *ac, int node)
> -{
> -       LIST_HEAD(list);
> -
> -       /* ac from n->shared can be freed if we don't hold the slab_mutex. */
> -       check_mutex_acquired();
> -
> -       if (!ac || !ac->avail)
> -               return;
> -
> -       if (ac->touched) {
> -               ac->touched = 0;
> -               return;
> -       }
> -
> -       raw_spin_lock_irq(&n->list_lock);
> -       drain_array_locked(cachep, ac, node, false, &list);
> -       raw_spin_unlock_irq(&n->list_lock);
> -
> -       slabs_destroy(cachep, &list);
> -}
> -
> -/**
> - * cache_reap - Reclaim memory from caches.
> - * @w: work descriptor
> - *
> - * Called from workqueue/eventd every few seconds.
> - * Purpose:
> - * - clear the per-cpu caches for this CPU.
> - * - return freeable pages to the main free memory pool.
> - *
> - * If we cannot acquire the cache chain mutex then just give up - we'll try
> - * again on the next iteration.
> - */
> -static void cache_reap(struct work_struct *w)
> -{
> -       struct kmem_cache *searchp;
> -       struct kmem_cache_node *n;
> -       int node = numa_mem_id();
> -       struct delayed_work *work = to_delayed_work(w);
> -
> -       if (!mutex_trylock(&slab_mutex))
> -               /* Give up. Setup the next iteration. */
> -               goto out;
> -
> -       list_for_each_entry(searchp, &slab_caches, list) {
> -               check_irq_on();
> -
> -               /*
> -                * We only take the node lock if absolutely necessary and we
> -                * have established with reasonable certainty that
> -                * we can do some work if the lock was obtained.
> -                */
> -               n = get_node(searchp, node);
> -
> -               reap_alien(searchp, n);
> -
> -               drain_array(searchp, n, cpu_cache_get(searchp), node);
> -
> -               /*
> -                * These are racy checks but it does not matter
> -                * if we skip one check or scan twice.
> -                */
> -               if (time_after(n->next_reap, jiffies))
> -                       goto next;
> -
> -               n->next_reap = jiffies + REAPTIMEOUT_NODE;
> -
> -               drain_array(searchp, n, n->shared, node);
> -
> -               if (n->free_touched)
> -                       n->free_touched = 0;
> -               else {
> -                       int freed;
> -
> -                       freed = drain_freelist(searchp, n, (n->free_limit +
> -                               5 * searchp->num - 1) / (5 * searchp->num));
> -                       STATS_ADD_REAPED(searchp, freed);
> -               }
> -next:
> -               cond_resched();
> -       }
> -       check_irq_on();
> -       mutex_unlock(&slab_mutex);
> -       next_reap_node();
> -out:
> -       /* Set up the next iteration */
> -       schedule_delayed_work_on(smp_processor_id(), work,
> -                               round_jiffies_relative(REAPTIMEOUT_AC));
> -}
> -
> -void get_slabinfo(struct kmem_cache *cachep, struct slabinfo *sinfo)
> -{
> -       unsigned long active_objs, num_objs, active_slabs;
> -       unsigned long total_slabs = 0, free_objs = 0, shared_avail = 0;
> -       unsigned long free_slabs = 0;
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_kmem_cache_node(cachep, node, n) {
> -               check_irq_on();
> -               raw_spin_lock_irq(&n->list_lock);
> -
> -               total_slabs += n->total_slabs;
> -               free_slabs += n->free_slabs;
> -               free_objs += n->free_objects;
> -
> -               if (n->shared)
> -                       shared_avail += n->shared->avail;
> -
> -               raw_spin_unlock_irq(&n->list_lock);
> -       }
> -       num_objs = total_slabs * cachep->num;
> -       active_slabs = total_slabs - free_slabs;
> -       active_objs = num_objs - free_objs;
> -
> -       sinfo->active_objs = active_objs;
> -       sinfo->num_objs = num_objs;
> -       sinfo->active_slabs = active_slabs;
> -       sinfo->num_slabs = total_slabs;
> -       sinfo->shared_avail = shared_avail;
> -       sinfo->limit = cachep->limit;
> -       sinfo->batchcount = cachep->batchcount;
> -       sinfo->shared = cachep->shared;
> -       sinfo->objects_per_slab = cachep->num;
> -       sinfo->cache_order = cachep->gfporder;
> -}
> -
> -void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *cachep)
> -{
> -#if STATS
> -       {                       /* node stats */
> -               unsigned long high = cachep->high_mark;
> -               unsigned long allocs = cachep->num_allocations;
> -               unsigned long grown = cachep->grown;
> -               unsigned long reaped = cachep->reaped;
> -               unsigned long errors = cachep->errors;
> -               unsigned long max_freeable = cachep->max_freeable;
> -               unsigned long node_allocs = cachep->node_allocs;
> -               unsigned long node_frees = cachep->node_frees;
> -               unsigned long overflows = cachep->node_overflow;
> -
> -               seq_printf(m, " : globalstat %7lu %6lu %5lu %4lu %4lu %4lu %4lu %4lu %4lu",
> -                          allocs, high, grown,
> -                          reaped, errors, max_freeable, node_allocs,
> -                          node_frees, overflows);
> -       }
> -       /* cpu stats */
> -       {
> -               unsigned long allochit = atomic_read(&cachep->allochit);
> -               unsigned long allocmiss = atomic_read(&cachep->allocmiss);
> -               unsigned long freehit = atomic_read(&cachep->freehit);
> -               unsigned long freemiss = atomic_read(&cachep->freemiss);
> -
> -               seq_printf(m, " : cpustat %6lu %6lu %6lu %6lu",
> -                          allochit, allocmiss, freehit, freemiss);
> -       }
> -#endif
> -}
> -
> -#define MAX_SLABINFO_WRITE 128
> -/**
> - * slabinfo_write - Tuning for the slab allocator
> - * @file: unused
> - * @buffer: user buffer
> - * @count: data length
> - * @ppos: unused
> - *
> - * Return: %0 on success, negative error code otherwise.
> - */
> -ssize_t slabinfo_write(struct file *file, const char __user *buffer,
> -                      size_t count, loff_t *ppos)
> -{
> -       char kbuf[MAX_SLABINFO_WRITE + 1], *tmp;
> -       int limit, batchcount, shared, res;
> -       struct kmem_cache *cachep;
> -
> -       if (count > MAX_SLABINFO_WRITE)
> -               return -EINVAL;
> -       if (copy_from_user(&kbuf, buffer, count))
> -               return -EFAULT;
> -       kbuf[MAX_SLABINFO_WRITE] = '\0';
> -
> -       tmp = strchr(kbuf, ' ');
> -       if (!tmp)
> -               return -EINVAL;
> -       *tmp = '\0';
> -       tmp++;
> -       if (sscanf(tmp, " %d %d %d", &limit, &batchcount, &shared) != 3)
> -               return -EINVAL;
> -
> -       /* Find the cache in the chain of caches. */
> -       mutex_lock(&slab_mutex);
> -       res = -EINVAL;
> -       list_for_each_entry(cachep, &slab_caches, list) {
> -               if (!strcmp(cachep->name, kbuf)) {
> -                       if (limit < 1 || batchcount < 1 ||
> -                                       batchcount > limit || shared < 0) {
> -                               res = 0;
> -                       } else {
> -                               res = do_tune_cpucache(cachep, limit,
> -                                                      batchcount, shared,
> -                                                      GFP_KERNEL);
> -                       }
> -                       break;
> -               }
> -       }
> -       mutex_unlock(&slab_mutex);
> -       if (res >= 0)
> -               res = count;
> -       return res;
> -}
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -/*
> - * Rejects incorrectly sized objects and objects that are to be copied
> - * to/from userspace but do not fall entirely within the containing slab
> - * cache's usercopy region.
> - *
> - * Returns NULL if check passes, otherwise const char * to name of cache
> - * to indicate an error.
> - */
> -void __check_heap_object(const void *ptr, unsigned long n,
> -                        const struct slab *slab, bool to_user)
> -{
> -       struct kmem_cache *cachep;
> -       unsigned int objnr;
> -       unsigned long offset;
> -
> -       ptr = kasan_reset_tag(ptr);
> -
> -       /* Find and validate object. */
> -       cachep = slab->slab_cache;
> -       objnr = obj_to_index(cachep, slab, (void *)ptr);
> -       BUG_ON(objnr >= cachep->num);
> -
> -       /* Find offset within object. */
> -       if (is_kfence_address(ptr))
> -               offset = ptr - kfence_object_start(ptr);
> -       else
> -               offset = ptr - index_to_obj(cachep, slab, objnr) - obj_offset(cachep);
> -
> -       /* Allow address range falling entirely within usercopy region. */
> -       if (offset >= cachep->useroffset &&
> -           offset - cachep->useroffset <= cachep->usersize &&
> -           n <= cachep->useroffset - offset + cachep->usersize)
> -               return;
> -
> -       usercopy_abort("SLAB object", cachep->name, to_user, offset, n);
> -}
> -#endif /* CONFIG_HARDENED_USERCOPY */
> diff --git a/mm/slab.h b/mm/slab.h
> index 43966aa5fadf..53e97b7640e9 100644
> --- a/mm/slab.h
> +++ b/mm/slab.h
> @@ -9,60 +9,12 @@
>  struct slab {
>         unsigned long __page_flags;
>
> -#if defined(CONFIG_SLAB)
> -
> -       struct kmem_cache *slab_cache;
> -       union {
> -               struct {
> -                       struct list_head slab_list;
> -                       void *freelist; /* array of free object indexes */
> -                       void *s_mem;    /* first object */
> -               };
> -               struct rcu_head rcu_head;
> -       };
> -       unsigned int active;
> -
> -#elif defined(CONFIG_SLUB)
> -
> -       struct kmem_cache *slab_cache;
> -       union {
> -               struct {
> -                       union {
> -                               struct list_head slab_list;
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -                               struct {
> -                                       struct slab *next;
> -                                       int slabs;      /* Nr of slabs left */
> -                               };
> -#endif
> -                       };
> -                       /* Double-word boundary */
> -                       void *freelist;         /* first free object */
> -                       union {
> -                               unsigned long counters;
> -                               struct {
> -                                       unsigned inuse:16;
> -                                       unsigned objects:15;
> -                                       unsigned frozen:1;
> -                               };
> -                       };
> -               };
> -               struct rcu_head rcu_head;
> -       };
> -       unsigned int __unused;
> -
> -#elif defined(CONFIG_SLOB)
> -
>         struct list_head slab_list;
>         void *__unused_1;
>         void *freelist;         /* first free block */
>         long units;
>         unsigned int __unused_2;
>
> -#else
> -#error "Unexpected slab allocator configured"
> -#endif
> -
>         atomic_t __page_refcount;
>  #ifdef CONFIG_MEMCG
>         unsigned long memcg_data;
> @@ -72,20 +24,13 @@ struct slab {
>  #define SLAB_MATCH(pg, sl)                                             \
>         static_assert(offsetof(struct page, pg) == offsetof(struct slab, sl))
>  SLAB_MATCH(flags, __page_flags);
> -#ifndef CONFIG_SLOB
> -SLAB_MATCH(compound_head, slab_cache); /* Ensure bit 0 is clear */
> -#else
>  SLAB_MATCH(compound_head, slab_list);  /* Ensure bit 0 is clear */
> -#endif
>  SLAB_MATCH(_refcount, __page_refcount);
>  #ifdef CONFIG_MEMCG
>  SLAB_MATCH(memcg_data, memcg_data);
>  #endif
>  #undef SLAB_MATCH
>  static_assert(sizeof(struct slab) <= sizeof(struct page));
> -#if defined(CONFIG_HAVE_CMPXCHG_DOUBLE) && defined(CONFIG_SLUB)
> -static_assert(IS_ALIGNED(offsetof(struct slab, freelist), 2*sizeof(void *)));
> -#endif
>
>  /**
>   * folio_slab - Converts from folio to slab.
> @@ -200,7 +145,6 @@ static inline size_t slab_size(const struct slab *slab)
>         return PAGE_SIZE << slab_order(slab);
>  }
>
> -#ifdef CONFIG_SLOB
>  /*
>   * Common fields provided in kmem_cache by all slab allocators
>   * This struct is either used directly by the allocator (SLOB)
> @@ -223,16 +167,6 @@ struct kmem_cache {
>         struct list_head list;  /* List of all slab caches on the system */
>  };
>
> -#endif /* CONFIG_SLOB */
> -
> -#ifdef CONFIG_SLAB
> -#include <linux/slab_def.h>
> -#endif
> -
> -#ifdef CONFIG_SLUB
> -#include <linux/slub_def.h>
> -#endif
> -
>  #include <linux/memcontrol.h>
>  #include <linux/fault-inject.h>
>  #include <linux/kasan.h>
> @@ -268,26 +202,6 @@ extern struct list_head slab_caches;
>  /* The slab cache that manages slab cache information */
>  extern struct kmem_cache *kmem_cache;
>
> -/* A table of kmalloc cache names and sizes */
> -extern const struct kmalloc_info_struct {
> -       const char *name[NR_KMALLOC_TYPES];
> -       unsigned int size;
> -} kmalloc_info[];
> -
> -#ifndef CONFIG_SLOB
> -/* Kmalloc array related functions */
> -void setup_kmalloc_cache_index_table(void);
> -void create_kmalloc_caches(slab_flags_t);
> -
> -/* Find the kmalloc slab corresponding for a certain size */
> -struct kmem_cache *kmalloc_slab(size_t, gfp_t);
> -
> -void *__kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags,
> -                             int node, size_t orig_size,
> -                             unsigned long caller);
> -void __kmem_cache_free(struct kmem_cache *s, void *x, unsigned long caller);
> -#endif
> -
>  gfp_t kmalloc_fix_flags(gfp_t flags);
>
>  /* Functions provided by the slab allocators */
> @@ -303,14 +217,6 @@ extern void create_boot_cache(struct kmem_cache *, const char *name,
>  int slab_unmergeable(struct kmem_cache *s);
>  struct kmem_cache *find_mergeable(unsigned size, unsigned align,
>                 slab_flags_t flags, const char *name, void (*ctor)(void *));
> -#ifndef CONFIG_SLOB
> -struct kmem_cache *
> -__kmem_cache_alias(const char *name, unsigned int size, unsigned int align,
> -                  slab_flags_t flags, void (*ctor)(void *));
> -
> -slab_flags_t kmem_cache_flags(unsigned int object_size,
> -       slab_flags_t flags, const char *name);
> -#else
>  static inline struct kmem_cache *
>  __kmem_cache_alias(const char *name, unsigned int size, unsigned int align,
>                    slab_flags_t flags, void (*ctor)(void *))
> @@ -321,15 +227,10 @@ static inline slab_flags_t kmem_cache_flags(unsigned int object_size,
>  {
>         return flags;
>  }
> -#endif
>
>  static inline bool is_kmalloc_cache(struct kmem_cache *s)
>  {
> -#ifndef CONFIG_SLOB
> -       return (s->flags & SLAB_KMALLOC);
> -#else
>         return false;
> -#endif
>  }
>
>  /* Legal flag mask for kmem_cache_create(), for various configurations */
> @@ -337,26 +238,9 @@ static inline bool is_kmalloc_cache(struct kmem_cache *s)
>                          SLAB_CACHE_DMA32 | SLAB_PANIC | \
>                          SLAB_TYPESAFE_BY_RCU | SLAB_DEBUG_OBJECTS )
>
> -#if defined(CONFIG_DEBUG_SLAB)
> -#define SLAB_DEBUG_FLAGS (SLAB_RED_ZONE | SLAB_POISON | SLAB_STORE_USER)
> -#elif defined(CONFIG_SLUB_DEBUG)
> -#define SLAB_DEBUG_FLAGS (SLAB_RED_ZONE | SLAB_POISON | SLAB_STORE_USER | \
> -                         SLAB_TRACE | SLAB_CONSISTENCY_CHECKS)
> -#else
>  #define SLAB_DEBUG_FLAGS (0)
> -#endif
>
> -#if defined(CONFIG_SLAB)
> -#define SLAB_CACHE_FLAGS (SLAB_MEM_SPREAD | SLAB_NOLEAKTRACE | \
> -                         SLAB_RECLAIM_ACCOUNT | SLAB_TEMPORARY | \
> -                         SLAB_ACCOUNT)
> -#elif defined(CONFIG_SLUB)
> -#define SLAB_CACHE_FLAGS (SLAB_NOLEAKTRACE | SLAB_RECLAIM_ACCOUNT | \
> -                         SLAB_TEMPORARY | SLAB_ACCOUNT | \
> -                         SLAB_NO_USER_FLAGS | SLAB_KMALLOC)
> -#else
>  #define SLAB_CACHE_FLAGS (SLAB_NOLEAKTRACE)
> -#endif
>
>  /* Common flags available with current configuration */
>  #define CACHE_CREATE_MASK (SLAB_CORE_FLAGS | SLAB_DEBUG_FLAGS | SLAB_CACHE_FLAGS)
> @@ -409,19 +293,6 @@ static inline enum node_stat_item cache_vmstat_idx(struct kmem_cache *s)
>                 NR_SLAB_RECLAIMABLE_B : NR_SLAB_UNRECLAIMABLE_B;
>  }
>
> -#ifdef CONFIG_SLUB_DEBUG
> -#ifdef CONFIG_SLUB_DEBUG_ON
> -DECLARE_STATIC_KEY_TRUE(slub_debug_enabled);
> -#else
> -DECLARE_STATIC_KEY_FALSE(slub_debug_enabled);
> -#endif
> -extern void print_tracking(struct kmem_cache *s, void *object);
> -long validate_slab_cache(struct kmem_cache *s);
> -static inline bool __slub_debug_enabled(void)
> -{
> -       return static_branch_unlikely(&slub_debug_enabled);
> -}
> -#else
>  static inline void print_tracking(struct kmem_cache *s, void *object)
>  {
>  }
> @@ -429,7 +300,6 @@ static inline bool __slub_debug_enabled(void)
>  {
>         return false;
>  }
> -#endif
>
>  /*
>   * Returns true if any of the specified slub_debug flags is enabled for the
> @@ -438,160 +308,9 @@ static inline bool __slub_debug_enabled(void)
>   */
>  static inline bool kmem_cache_debug_flags(struct kmem_cache *s, slab_flags_t flags)
>  {
> -       if (IS_ENABLED(CONFIG_SLUB_DEBUG))
> -               VM_WARN_ON_ONCE(!(flags & SLAB_DEBUG_FLAGS));
> -       if (__slub_debug_enabled())
> -               return s->flags & flags;
>         return false;
>  }
>
> -#ifdef CONFIG_MEMCG_KMEM
> -/*
> - * slab_objcgs - get the object cgroups vector associated with a slab
> - * @slab: a pointer to the slab struct
> - *
> - * Returns a pointer to the object cgroups vector associated with the slab,
> - * or NULL if no such vector has been associated yet.
> - */
> -static inline struct obj_cgroup **slab_objcgs(struct slab *slab)
> -{
> -       unsigned long memcg_data = READ_ONCE(slab->memcg_data);
> -
> -       VM_BUG_ON_PAGE(memcg_data && !(memcg_data & MEMCG_DATA_OBJCGS),
> -                                                       slab_page(slab));
> -       VM_BUG_ON_PAGE(memcg_data & MEMCG_DATA_KMEM, slab_page(slab));
> -
> -       return (struct obj_cgroup **)(memcg_data & ~MEMCG_DATA_FLAGS_MASK);
> -}
> -
> -int memcg_alloc_slab_cgroups(struct slab *slab, struct kmem_cache *s,
> -                                gfp_t gfp, bool new_slab);
> -void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
> -                    enum node_stat_item idx, int nr);
> -
> -static inline void memcg_free_slab_cgroups(struct slab *slab)
> -{
> -       kfree(slab_objcgs(slab));
> -       slab->memcg_data = 0;
> -}
> -
> -static inline size_t obj_full_size(struct kmem_cache *s)
> -{
> -       /*
> -        * For each accounted object there is an extra space which is used
> -        * to store obj_cgroup membership. Charge it too.
> -        */
> -       return s->size + sizeof(struct obj_cgroup *);
> -}
> -
> -/*
> - * Returns false if the allocation should fail.
> - */
> -static inline bool memcg_slab_pre_alloc_hook(struct kmem_cache *s,
> -                                            struct list_lru *lru,
> -                                            struct obj_cgroup **objcgp,
> -                                            size_t objects, gfp_t flags)
> -{
> -       struct obj_cgroup *objcg;
> -
> -       if (!memcg_kmem_online())
> -               return true;
> -
> -       if (!(flags & __GFP_ACCOUNT) && !(s->flags & SLAB_ACCOUNT))
> -               return true;
> -
> -       objcg = get_obj_cgroup_from_current();
> -       if (!objcg)
> -               return true;
> -
> -       if (lru) {
> -               int ret;
> -               struct mem_cgroup *memcg;
> -
> -               memcg = get_mem_cgroup_from_objcg(objcg);
> -               ret = memcg_list_lru_alloc(memcg, lru, flags);
> -               css_put(&memcg->css);
> -
> -               if (ret)
> -                       goto out;
> -       }
> -
> -       if (obj_cgroup_charge(objcg, flags, objects * obj_full_size(s)))
> -               goto out;
> -
> -       *objcgp = objcg;
> -       return true;
> -out:
> -       obj_cgroup_put(objcg);
> -       return false;
> -}
> -
> -static inline void memcg_slab_post_alloc_hook(struct kmem_cache *s,
> -                                             struct obj_cgroup *objcg,
> -                                             gfp_t flags, size_t size,
> -                                             void **p)
> -{
> -       struct slab *slab;
> -       unsigned long off;
> -       size_t i;
> -
> -       if (!memcg_kmem_online() || !objcg)
> -               return;
> -
> -       for (i = 0; i < size; i++) {
> -               if (likely(p[i])) {
> -                       slab = virt_to_slab(p[i]);
> -
> -                       if (!slab_objcgs(slab) &&
> -                           memcg_alloc_slab_cgroups(slab, s, flags,
> -                                                        false)) {
> -                               obj_cgroup_uncharge(objcg, obj_full_size(s));
> -                               continue;
> -                       }
> -
> -                       off = obj_to_index(s, slab, p[i]);
> -                       obj_cgroup_get(objcg);
> -                       slab_objcgs(slab)[off] = objcg;
> -                       mod_objcg_state(objcg, slab_pgdat(slab),
> -                                       cache_vmstat_idx(s), obj_full_size(s));
> -               } else {
> -                       obj_cgroup_uncharge(objcg, obj_full_size(s));
> -               }
> -       }
> -       obj_cgroup_put(objcg);
> -}
> -
> -static inline void memcg_slab_free_hook(struct kmem_cache *s, struct slab *slab,
> -                                       void **p, int objects)
> -{
> -       struct obj_cgroup **objcgs;
> -       int i;
> -
> -       if (!memcg_kmem_online())
> -               return;
> -
> -       objcgs = slab_objcgs(slab);
> -       if (!objcgs)
> -               return;
> -
> -       for (i = 0; i < objects; i++) {
> -               struct obj_cgroup *objcg;
> -               unsigned int off;
> -
> -               off = obj_to_index(s, slab, p[i]);
> -               objcg = objcgs[off];
> -               if (!objcg)
> -                       continue;
> -
> -               objcgs[off] = NULL;
> -               obj_cgroup_uncharge(objcg, obj_full_size(s));
> -               mod_objcg_state(objcg, slab_pgdat(slab), cache_vmstat_idx(s),
> -                               -obj_full_size(s));
> -               obj_cgroup_put(objcg);
> -       }
> -}
> -
> -#else /* CONFIG_MEMCG_KMEM */
>  static inline struct obj_cgroup **slab_objcgs(struct slab *slab)
>  {
>         return NULL;
> @@ -632,90 +351,12 @@ static inline void memcg_slab_free_hook(struct kmem_cache *s, struct slab *slab,
>                                         void **p, int objects)
>  {
>  }
> -#endif /* CONFIG_MEMCG_KMEM */
> -
> -#ifndef CONFIG_SLOB
> -static inline struct kmem_cache *virt_to_cache(const void *obj)
> -{
> -       struct slab *slab;
> -
> -       slab = virt_to_slab(obj);
> -       if (WARN_ONCE(!slab, "%s: Object is not a Slab page!\n",
> -                                       __func__))
> -               return NULL;
> -       return slab->slab_cache;
> -}
> -
> -static __always_inline void account_slab(struct slab *slab, int order,
> -                                        struct kmem_cache *s, gfp_t gfp)
> -{
> -       if (memcg_kmem_online() && (s->flags & SLAB_ACCOUNT))
> -               memcg_alloc_slab_cgroups(slab, s, gfp, true);
> -
> -       mod_node_page_state(slab_pgdat(slab), cache_vmstat_idx(s),
> -                           PAGE_SIZE << order);
> -}
> -
> -static __always_inline void unaccount_slab(struct slab *slab, int order,
> -                                          struct kmem_cache *s)
> -{
> -       if (memcg_kmem_online())
> -               memcg_free_slab_cgroups(slab);
> -
> -       mod_node_page_state(slab_pgdat(slab), cache_vmstat_idx(s),
> -                           -(PAGE_SIZE << order));
> -}
> -
> -static inline struct kmem_cache *cache_from_obj(struct kmem_cache *s, void *x)
> -{
> -       struct kmem_cache *cachep;
> -
> -       if (!IS_ENABLED(CONFIG_SLAB_FREELIST_HARDENED) &&
> -           !kmem_cache_debug_flags(s, SLAB_CONSISTENCY_CHECKS))
> -               return s;
> -
> -       cachep = virt_to_cache(x);
> -       if (WARN(cachep && cachep != s,
> -                 "%s: Wrong slab cache. %s but object is from %s\n",
> -                 __func__, s->name, cachep->name))
> -               print_tracking(cachep, x);
> -       return cachep;
> -}
> -
> -void free_large_kmalloc(struct folio *folio, void *object);
> -
> -#endif /* CONFIG_SLOB */
>
>  size_t __ksize(const void *objp);
>
>  static inline size_t slab_ksize(const struct kmem_cache *s)
>  {
> -#ifndef CONFIG_SLUB
>         return s->object_size;
> -
> -#else /* CONFIG_SLUB */
> -# ifdef CONFIG_SLUB_DEBUG
> -       /*
> -        * Debugging requires use of the padding between object
> -        * and whatever may come after it.
> -        */
> -       if (s->flags & (SLAB_RED_ZONE | SLAB_POISON))
> -               return s->object_size;
> -# endif
> -       if (s->flags & SLAB_KASAN)
> -               return s->object_size;
> -       /*
> -        * If we have the need to store the freelist pointer
> -        * back there or track user information then we can
> -        * only use the space before that information.
> -        */
> -       if (s->flags & (SLAB_TYPESAFE_BY_RCU | SLAB_STORE_USER))
> -               return s->inuse;
> -       /*
> -        * Else we can use all the padding etc for the allocation
> -        */
> -       return s->size;
> -#endif
>  }
>
>  static inline struct kmem_cache *slab_pre_alloc_hook(struct kmem_cache *s,
> @@ -777,77 +418,18 @@ static inline void slab_post_alloc_hook(struct kmem_cache *s,
>         memcg_slab_post_alloc_hook(s, objcg, flags, size, p);
>  }
>
> -#ifndef CONFIG_SLOB
> -/*
> - * The slab lists for all objects.
> - */
> -struct kmem_cache_node {
> -#ifdef CONFIG_SLAB
> -       raw_spinlock_t list_lock;
> -       struct list_head slabs_partial; /* partial list first, better asm code */
> -       struct list_head slabs_full;
> -       struct list_head slabs_free;
> -       unsigned long total_slabs;      /* length of all slab lists */
> -       unsigned long free_slabs;       /* length of free slab list only */
> -       unsigned long free_objects;
> -       unsigned int free_limit;
> -       unsigned int colour_next;       /* Per-node cache coloring */
> -       struct array_cache *shared;     /* shared per node */
> -       struct alien_cache **alien;     /* on other nodes */
> -       unsigned long next_reap;        /* updated without locking */
> -       int free_touched;               /* updated without locking */
> -#endif
> -
> -#ifdef CONFIG_SLUB
> -       spinlock_t list_lock;
> -       unsigned long nr_partial;
> -       struct list_head partial;
> -#ifdef CONFIG_SLUB_DEBUG
> -       atomic_long_t nr_slabs;
> -       atomic_long_t total_objects;
> -       struct list_head full;
> -#endif
> -#endif
> -
> -};
> -
> -static inline struct kmem_cache_node *get_node(struct kmem_cache *s, int node)
> -{
> -       return s->node[node];
> -}
> -
> -/*
> - * Iterator over all nodes. The body will be executed for each node that has
> - * a kmem_cache_node structure allocated (which is true for all online nodes)
> - */
> -#define for_each_kmem_cache_node(__s, __node, __n) \
> -       for (__node = 0; __node < nr_node_ids; __node++) \
> -                if ((__n = get_node(__s, __node)))
> -
> -#endif
> -
> -#if defined(CONFIG_SLAB) || defined(CONFIG_SLUB_DEBUG)
> -void dump_unreclaimable_slab(void);
> -#else
>  static inline void dump_unreclaimable_slab(void)
>  {
>  }
> -#endif
>
>  void ___cache_free(struct kmem_cache *cache, void *x, unsigned long addr);
>
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count,
> -                       gfp_t gfp);
> -void cache_random_seq_destroy(struct kmem_cache *cachep);
> -#else
>  static inline int cache_random_seq_create(struct kmem_cache *cachep,
>                                         unsigned int count, gfp_t gfp)
>  {
>         return 0;
>  }
>  static inline void cache_random_seq_destroy(struct kmem_cache *cachep) { }
> -#endif /* CONFIG_SLAB_FREELIST_RANDOM */
>
>  static inline bool slab_want_init_on_alloc(gfp_t flags, struct kmem_cache *c)
>  {
> @@ -871,11 +453,7 @@ static inline bool slab_want_init_on_free(struct kmem_cache *c)
>         return false;
>  }
>
> -#if defined(CONFIG_DEBUG_FS) && defined(CONFIG_SLUB_DEBUG)
> -void debugfs_slab_release(struct kmem_cache *);
> -#else
>  static inline void debugfs_slab_release(struct kmem_cache *s) { }
> -#endif
>
>  #ifdef CONFIG_PRINTK
>  #define KS_ADDRS_COUNT 16
> @@ -903,8 +481,4 @@ void __check_heap_object(const void *ptr, unsigned long n,
>  }
>  #endif
>
> -#ifdef CONFIG_SLUB_DEBUG
> -void skip_orig_size_check(struct kmem_cache *s, const void *object);
> -#endif
> -
>  #endif /* MM_SLAB_H */
> diff --git a/mm/slab_common.c b/mm/slab_common.c
> index bf4e777cfe90..5f6a59e31abd 100644
> --- a/mm/slab_common.c
> +++ b/mm/slab_common.c
> @@ -69,11 +69,6 @@ static int __init setup_slab_merge(char *str)
>         return 1;
>  }
>
> -#ifdef CONFIG_SLUB
> -__setup_param("slub_nomerge", slub_nomerge, setup_slab_nomerge, 0);
> -__setup_param("slub_merge", slub_merge, setup_slab_merge, 0);
> -#endif
> -
>  __setup("slab_nomerge", setup_slab_nomerge);
>  __setup("slab_merge", setup_slab_merge);
>
> @@ -195,15 +190,29 @@ struct kmem_cache *find_mergeable(unsigned int size, unsigned int align,
>                 if (s->size - size >= sizeof(void *))
>                         continue;
>
> -               if (IS_ENABLED(CONFIG_SLAB) && align &&
> -                       (align > s->align || s->align % align))
> -                       continue;
> -
>                 return s;
>         }
>         return NULL;
>  }
>
> +struct slab_rcu {
> +       struct rcu_head head;
> +       int size;
> +};
> +
> +int __kmem_cache_create(struct kmem_cache *c, slab_flags_t flags)
> +{
> +       if (flags & SLAB_TYPESAFE_BY_RCU) {
> +               /* leave room for rcu footer at the end of object */
> +               c->size += sizeof(struct slab_rcu);
> +       }
> +
> +       /* Actual size allocated */
> +       c->size = PAGE_SIZE << get_order(c->size);
> +       c->flags = flags;
> +       return 0;
> +}
> +
>  static struct kmem_cache *create_cache(const char *name,
>                 unsigned int object_size, unsigned int align,
>                 slab_flags_t flags, unsigned int useroffset,
> @@ -285,20 +294,6 @@ kmem_cache_create_usercopy(const char *name,
>         const char *cache_name;
>         int err;
>
> -#ifdef CONFIG_SLUB_DEBUG
> -       /*
> -        * If no slub_debug was enabled globally, the static key is not yet
> -        * enabled by setup_slub_debug(). Enable it if the cache is being
> -        * created with any of the debugging flags passed explicitly.
> -        * It's also possible that this is the first cache created with
> -        * SLAB_STORE_USER and we should init stack_depot for it.
> -        */
> -       if (flags & SLAB_DEBUG_FLAGS)
> -               static_branch_enable(&slub_debug_enabled);
> -       if (flags & SLAB_STORE_USER)
> -               stack_depot_init();
> -#endif
> -
>         mutex_lock(&slab_mutex);
>
>         err = kmem_cache_sanity_check(name, size);
> @@ -552,7 +547,8 @@ static void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *
>  {
>         if (__kfence_obj_info(kpp, object, slab))
>                 return;
> -       __kmem_obj_info(kpp, object, slab);
> +       kpp->kp_ptr = object;
> +       kpp->kp_slab = slab;
>  }
>
>  /**
> @@ -625,462 +621,6 @@ void kmem_dump_obj(void *object)
>  EXPORT_SYMBOL_GPL(kmem_dump_obj);
>  #endif
>
> -#ifndef CONFIG_SLOB
> -/* Create a cache during boot when no slab services are available yet */
> -void __init create_boot_cache(struct kmem_cache *s, const char *name,
> -               unsigned int size, slab_flags_t flags,
> -               unsigned int useroffset, unsigned int usersize)
> -{
> -       int err;
> -       unsigned int align = ARCH_KMALLOC_MINALIGN;
> -
> -       s->name = name;
> -       s->size = s->object_size = size;
> -
> -       /*
> -        * For power of two sizes, guarantee natural alignment for kmalloc
> -        * caches, regardless of SL*B debugging options.
> -        */
> -       if (is_power_of_2(size))
> -               align = max(align, size);
> -       s->align = calculate_alignment(flags, align, size);
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -       s->useroffset = useroffset;
> -       s->usersize = usersize;
> -#endif
> -
> -       err = __kmem_cache_create(s, flags);
> -
> -       if (err)
> -               panic("Creation of kmalloc slab %s size=%u failed. Reason %d\n",
> -                                       name, size, err);
> -
> -       s->refcount = -1;       /* Exempt from merging for now */
> -}
> -
> -struct kmem_cache *__init create_kmalloc_cache(const char *name,
> -               unsigned int size, slab_flags_t flags,
> -               unsigned int useroffset, unsigned int usersize)
> -{
> -       struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
> -
> -       if (!s)
> -               panic("Out of memory when creating slab %s\n", name);
> -
> -       create_boot_cache(s, name, size, flags | SLAB_KMALLOC, useroffset,
> -                                                               usersize);
> -       list_add(&s->list, &slab_caches);
> -       s->refcount = 1;
> -       return s;
> -}
> -
> -struct kmem_cache *
> -kmalloc_caches[NR_KMALLOC_TYPES][KMALLOC_SHIFT_HIGH + 1] __ro_after_init =
> -{ /* initialization for https://bugs.llvm.org/show_bug.cgi?id=42570 */ };
> -EXPORT_SYMBOL(kmalloc_caches);
> -
> -/*
> - * Conversion table for small slabs sizes / 8 to the index in the
> - * kmalloc array. This is necessary for slabs < 192 since we have non power
> - * of two cache sizes there. The size of larger slabs can be determined using
> - * fls.
> - */
> -static u8 size_index[24] __ro_after_init = {
> -       3,      /* 8 */
> -       4,      /* 16 */
> -       5,      /* 24 */
> -       5,      /* 32 */
> -       6,      /* 40 */
> -       6,      /* 48 */
> -       6,      /* 56 */
> -       6,      /* 64 */
> -       1,      /* 72 */
> -       1,      /* 80 */
> -       1,      /* 88 */
> -       1,      /* 96 */
> -       7,      /* 104 */
> -       7,      /* 112 */
> -       7,      /* 120 */
> -       7,      /* 128 */
> -       2,      /* 136 */
> -       2,      /* 144 */
> -       2,      /* 152 */
> -       2,      /* 160 */
> -       2,      /* 168 */
> -       2,      /* 176 */
> -       2,      /* 184 */
> -       2       /* 192 */
> -};
> -
> -static inline unsigned int size_index_elem(unsigned int bytes)
> -{
> -       return (bytes - 1) / 8;
> -}
> -
> -/*
> - * Find the kmem_cache structure that serves a given size of
> - * allocation
> - */
> -struct kmem_cache *kmalloc_slab(size_t size, gfp_t flags)
> -{
> -       unsigned int index;
> -
> -       if (size <= 192) {
> -               if (!size)
> -                       return ZERO_SIZE_PTR;
> -
> -               index = size_index[size_index_elem(size)];
> -       } else {
> -               if (WARN_ON_ONCE(size > KMALLOC_MAX_CACHE_SIZE))
> -                       return NULL;
> -               index = fls(size - 1);
> -       }
> -
> -       return kmalloc_caches[kmalloc_type(flags)][index];
> -}
> -
> -size_t kmalloc_size_roundup(size_t size)
> -{
> -       struct kmem_cache *c;
> -
> -       /* Short-circuit the 0 size case. */
> -       if (unlikely(size == 0))
> -               return 0;
> -       /* Short-circuit saturated "too-large" case. */
> -       if (unlikely(size == SIZE_MAX))
> -               return SIZE_MAX;
> -       /* Above the smaller buckets, size is a multiple of page size. */
> -       if (size > KMALLOC_MAX_CACHE_SIZE)
> -               return PAGE_SIZE << get_order(size);
> -
> -       /* The flags don't matter since size_index is common to all. */
> -       c = kmalloc_slab(size, GFP_KERNEL);
> -       return c ? c->object_size : 0;
> -}
> -EXPORT_SYMBOL(kmalloc_size_roundup);
> -
> -#ifdef CONFIG_ZONE_DMA
> -#define KMALLOC_DMA_NAME(sz)   .name[KMALLOC_DMA] = "dma-kmalloc-" #sz,
> -#else
> -#define KMALLOC_DMA_NAME(sz)
> -#endif
> -
> -#ifdef CONFIG_MEMCG_KMEM
> -#define KMALLOC_CGROUP_NAME(sz)        .name[KMALLOC_CGROUP] = "kmalloc-cg-" #sz,
> -#else
> -#define KMALLOC_CGROUP_NAME(sz)
> -#endif
> -
> -#ifndef CONFIG_SLUB_TINY
> -#define KMALLOC_RCL_NAME(sz)   .name[KMALLOC_RECLAIM] = "kmalloc-rcl-" #sz,
> -#else
> -#define KMALLOC_RCL_NAME(sz)
> -#endif
> -
> -#define INIT_KMALLOC_INFO(__size, __short_size)                        \
> -{                                                              \
> -       .name[KMALLOC_NORMAL]  = "kmalloc-" #__short_size,      \
> -       KMALLOC_RCL_NAME(__short_size)                          \
> -       KMALLOC_CGROUP_NAME(__short_size)                       \
> -       KMALLOC_DMA_NAME(__short_size)                          \
> -       .size = __size,                                         \
> -}
> -
> -/*
> - * kmalloc_info[] is to make slub_debug=,kmalloc-xx option work at boot time.
> - * kmalloc_index() supports up to 2^21=2MB, so the final entry of the table is
> - * kmalloc-2M.
> - */
> -const struct kmalloc_info_struct kmalloc_info[] __initconst = {
> -       INIT_KMALLOC_INFO(0, 0),
> -       INIT_KMALLOC_INFO(96, 96),
> -       INIT_KMALLOC_INFO(192, 192),
> -       INIT_KMALLOC_INFO(8, 8),
> -       INIT_KMALLOC_INFO(16, 16),
> -       INIT_KMALLOC_INFO(32, 32),
> -       INIT_KMALLOC_INFO(64, 64),
> -       INIT_KMALLOC_INFO(128, 128),
> -       INIT_KMALLOC_INFO(256, 256),
> -       INIT_KMALLOC_INFO(512, 512),
> -       INIT_KMALLOC_INFO(1024, 1k),
> -       INIT_KMALLOC_INFO(2048, 2k),
> -       INIT_KMALLOC_INFO(4096, 4k),
> -       INIT_KMALLOC_INFO(8192, 8k),
> -       INIT_KMALLOC_INFO(16384, 16k),
> -       INIT_KMALLOC_INFO(32768, 32k),
> -       INIT_KMALLOC_INFO(65536, 64k),
> -       INIT_KMALLOC_INFO(131072, 128k),
> -       INIT_KMALLOC_INFO(262144, 256k),
> -       INIT_KMALLOC_INFO(524288, 512k),
> -       INIT_KMALLOC_INFO(1048576, 1M),
> -       INIT_KMALLOC_INFO(2097152, 2M)
> -};
> -
> -/*
> - * Patch up the size_index table if we have strange large alignment
> - * requirements for the kmalloc array. This is only the case for
> - * MIPS it seems. The standard arches will not generate any code here.
> - *
> - * Largest permitted alignment is 256 bytes due to the way we
> - * handle the index determination for the smaller caches.
> - *
> - * Make sure that nothing crazy happens if someone starts tinkering
> - * around with ARCH_KMALLOC_MINALIGN
> - */
> -void __init setup_kmalloc_cache_index_table(void)
> -{
> -       unsigned int i;
> -
> -       BUILD_BUG_ON(KMALLOC_MIN_SIZE > 256 ||
> -               !is_power_of_2(KMALLOC_MIN_SIZE));
> -
> -       for (i = 8; i < KMALLOC_MIN_SIZE; i += 8) {
> -               unsigned int elem = size_index_elem(i);
> -
> -               if (elem >= ARRAY_SIZE(size_index))
> -                       break;
> -               size_index[elem] = KMALLOC_SHIFT_LOW;
> -       }
> -
> -       if (KMALLOC_MIN_SIZE >= 64) {
> -               /*
> -                * The 96 byte sized cache is not used if the alignment
> -                * is 64 byte.
> -                */
> -               for (i = 64 + 8; i <= 96; i += 8)
> -                       size_index[size_index_elem(i)] = 7;
> -
> -       }
> -
> -       if (KMALLOC_MIN_SIZE >= 128) {
> -               /*
> -                * The 192 byte sized cache is not used if the alignment
> -                * is 128 byte. Redirect kmalloc to use the 256 byte cache
> -                * instead.
> -                */
> -               for (i = 128 + 8; i <= 192; i += 8)
> -                       size_index[size_index_elem(i)] = 8;
> -       }
> -}
> -
> -static void __init
> -new_kmalloc_cache(int idx, enum kmalloc_cache_type type, slab_flags_t flags)
> -{
> -       if ((KMALLOC_RECLAIM != KMALLOC_NORMAL) && (type == KMALLOC_RECLAIM)) {
> -               flags |= SLAB_RECLAIM_ACCOUNT;
> -       } else if (IS_ENABLED(CONFIG_MEMCG_KMEM) && (type == KMALLOC_CGROUP)) {
> -               if (mem_cgroup_kmem_disabled()) {
> -                       kmalloc_caches[type][idx] = kmalloc_caches[KMALLOC_NORMAL][idx];
> -                       return;
> -               }
> -               flags |= SLAB_ACCOUNT;
> -       } else if (IS_ENABLED(CONFIG_ZONE_DMA) && (type == KMALLOC_DMA)) {
> -               flags |= SLAB_CACHE_DMA;
> -       }
> -
> -       kmalloc_caches[type][idx] = create_kmalloc_cache(
> -                                       kmalloc_info[idx].name[type],
> -                                       kmalloc_info[idx].size, flags, 0,
> -                                       kmalloc_info[idx].size);
> -
> -       /*
> -        * If CONFIG_MEMCG_KMEM is enabled, disable cache merging for
> -        * KMALLOC_NORMAL caches.
> -        */
> -       if (IS_ENABLED(CONFIG_MEMCG_KMEM) && (type == KMALLOC_NORMAL))
> -               kmalloc_caches[type][idx]->refcount = -1;
> -}
> -
> -/*
> - * Create the kmalloc array. Some of the regular kmalloc arrays
> - * may already have been created because they were needed to
> - * enable allocations for slab creation.
> - */
> -void __init create_kmalloc_caches(slab_flags_t flags)
> -{
> -       int i;
> -       enum kmalloc_cache_type type;
> -
> -       /*
> -        * Including KMALLOC_CGROUP if CONFIG_MEMCG_KMEM defined
> -        */
> -       for (type = KMALLOC_NORMAL; type < NR_KMALLOC_TYPES; type++) {
> -               for (i = KMALLOC_SHIFT_LOW; i <= KMALLOC_SHIFT_HIGH; i++) {
> -                       if (!kmalloc_caches[type][i])
> -                               new_kmalloc_cache(i, type, flags);
> -
> -                       /*
> -                        * Caches that are not of the two-to-the-power-of size.
> -                        * These have to be created immediately after the
> -                        * earlier power of two caches
> -                        */
> -                       if (KMALLOC_MIN_SIZE <= 32 && i == 6 &&
> -                                       !kmalloc_caches[type][1])
> -                               new_kmalloc_cache(1, type, flags);
> -                       if (KMALLOC_MIN_SIZE <= 64 && i == 7 &&
> -                                       !kmalloc_caches[type][2])
> -                               new_kmalloc_cache(2, type, flags);
> -               }
> -       }
> -
> -       /* Kmalloc array is now usable */
> -       slab_state = UP;
> -}
> -
> -void free_large_kmalloc(struct folio *folio, void *object)
> -{
> -       unsigned int order = folio_order(folio);
> -
> -       if (WARN_ON_ONCE(order == 0))
> -               pr_warn_once("object pointer: 0x%p\n", object);
> -
> -       kmemleak_free(object);
> -       kasan_kfree_large(object);
> -       kmsan_kfree_large(object);
> -
> -       mod_lruvec_page_state(folio_page(folio, 0), NR_SLAB_UNRECLAIMABLE_B,
> -                             -(PAGE_SIZE << order));
> -       __free_pages(folio_page(folio, 0), order);
> -}
> -
> -static void *__kmalloc_large_node(size_t size, gfp_t flags, int node);
> -static __always_inline
> -void *__do_kmalloc_node(size_t size, gfp_t flags, int node, unsigned long caller)
> -{
> -       struct kmem_cache *s;
> -       void *ret;
> -
> -       if (unlikely(size > KMALLOC_MAX_CACHE_SIZE)) {
> -               ret = __kmalloc_large_node(size, flags, node);
> -               trace_kmalloc(caller, ret, size,
> -                             PAGE_SIZE << get_order(size), flags, node);
> -               return ret;
> -       }
> -
> -       s = kmalloc_slab(size, flags);
> -
> -       if (unlikely(ZERO_OR_NULL_PTR(s)))
> -               return s;
> -
> -       ret = __kmem_cache_alloc_node(s, flags, node, size, caller);
> -       ret = kasan_kmalloc(s, ret, size, flags);
> -       trace_kmalloc(caller, ret, size, s->size, flags, node);
> -       return ret;
> -}
> -
> -void *__kmalloc_node(size_t size, gfp_t flags, int node)
> -{
> -       return __do_kmalloc_node(size, flags, node, _RET_IP_);
> -}
> -EXPORT_SYMBOL(__kmalloc_node);
> -
> -void *__kmalloc(size_t size, gfp_t flags)
> -{
> -       return __do_kmalloc_node(size, flags, NUMA_NO_NODE, _RET_IP_);
> -}
> -EXPORT_SYMBOL(__kmalloc);
> -
> -void *__kmalloc_node_track_caller(size_t size, gfp_t flags,
> -                                 int node, unsigned long caller)
> -{
> -       return __do_kmalloc_node(size, flags, node, caller);
> -}
> -EXPORT_SYMBOL(__kmalloc_node_track_caller);
> -
> -/**
> - * kfree - free previously allocated memory
> - * @object: pointer returned by kmalloc.
> - *
> - * If @object is NULL, no operation is performed.
> - *
> - * Don't free memory not originally allocated by kmalloc()
> - * or you will run into trouble.
> - */
> -void kfree(const void *object)
> -{
> -       struct folio *folio;
> -       struct slab *slab;
> -       struct kmem_cache *s;
> -
> -       trace_kfree(_RET_IP_, object);
> -
> -       if (unlikely(ZERO_OR_NULL_PTR(object)))
> -               return;
> -
> -       folio = virt_to_folio(object);
> -       if (unlikely(!folio_test_slab(folio))) {
> -               free_large_kmalloc(folio, (void *)object);
> -               return;
> -       }
> -
> -       slab = folio_slab(folio);
> -       s = slab->slab_cache;
> -       __kmem_cache_free(s, (void *)object, _RET_IP_);
> -}
> -EXPORT_SYMBOL(kfree);
> -
> -/**
> - * __ksize -- Report full size of underlying allocation
> - * @object: pointer to the object
> - *
> - * This should only be used internally to query the true size of allocations.
> - * It is not meant to be a way to discover the usable size of an allocation
> - * after the fact. Instead, use kmalloc_size_roundup(). Using memory beyond
> - * the originally requested allocation size may trigger KASAN, UBSAN_BOUNDS,
> - * and/or FORTIFY_SOURCE.
> - *
> - * Return: size of the actual memory used by @object in bytes
> - */
> -size_t __ksize(const void *object)
> -{
> -       struct folio *folio;
> -
> -       if (unlikely(object == ZERO_SIZE_PTR))
> -               return 0;
> -
> -       folio = virt_to_folio(object);
> -
> -       if (unlikely(!folio_test_slab(folio))) {
> -               if (WARN_ON(folio_size(folio) <= KMALLOC_MAX_CACHE_SIZE))
> -                       return 0;
> -               if (WARN_ON(object != folio_address(folio)))
> -                       return 0;
> -               return folio_size(folio);
> -       }
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -       skip_orig_size_check(folio_slab(folio)->slab_cache, object);
> -#endif
> -
> -       return slab_ksize(folio_slab(folio)->slab_cache);
> -}
> -
> -void *kmalloc_trace(struct kmem_cache *s, gfp_t gfpflags, size_t size)
> -{
> -       void *ret = __kmem_cache_alloc_node(s, gfpflags, NUMA_NO_NODE,
> -                                           size, _RET_IP_);
> -
> -       trace_kmalloc(_RET_IP_, ret, size, s->size, gfpflags, NUMA_NO_NODE);
> -
> -       ret = kasan_kmalloc(s, ret, size, gfpflags);
> -       return ret;
> -}
> -EXPORT_SYMBOL(kmalloc_trace);
> -
> -void *kmalloc_node_trace(struct kmem_cache *s, gfp_t gfpflags,
> -                        int node, size_t size)
> -{
> -       void *ret = __kmem_cache_alloc_node(s, gfpflags, node, size, _RET_IP_);
> -
> -       trace_kmalloc(_RET_IP_, ret, size, s->size, gfpflags, node);
> -
> -       ret = kasan_kmalloc(s, ret, size, gfpflags);
> -       return ret;
> -}
> -EXPORT_SYMBOL(kmalloc_node_trace);
> -#endif /* !CONFIG_SLOB */
> -
>  gfp_t kmalloc_fix_flags(gfp_t flags)
>  {
>         gfp_t invalid_mask = flags & GFP_SLAB_BUG_MASK;
> @@ -1144,202 +684,44 @@ void *kmalloc_large_node(size_t size, gfp_t flags, int node)
>  }
>  EXPORT_SYMBOL(kmalloc_large_node);
>
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -/* Randomize a generic freelist */
> -static void freelist_randomize(struct rnd_state *state, unsigned int *list,
> -                              unsigned int count)
> -{
> -       unsigned int rand;
> -       unsigned int i;
> -
> -       for (i = 0; i < count; i++)
> -               list[i] = i;
> -
> -       /* Fisher-Yates shuffle */
> -       for (i = count - 1; i > 0; i--) {
> -               rand = prandom_u32_state(state);
> -               rand %= (i + 1);
> -               swap(list[i], list[rand]);
> -       }
> -}
> -
> -/* Create a random sequence per cache */
> -int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count,
> -                                   gfp_t gfp)
> -{
> -       struct rnd_state state;
> -
> -       if (count < 2 || cachep->random_seq)
> -               return 0;
> -
> -       cachep->random_seq = kcalloc(count, sizeof(unsigned int), gfp);
> -       if (!cachep->random_seq)
> -               return -ENOMEM;
> -
> -       /* Get best entropy at this stage of boot */
> -       prandom_seed_state(&state, get_random_long());
> -
> -       freelist_randomize(&state, cachep->random_seq, count);
> -       return 0;
> -}
> -
> -/* Destroy the per-cache random freelist sequence */
> -void cache_random_seq_destroy(struct kmem_cache *cachep)
> -{
> -       kfree(cachep->random_seq);
> -       cachep->random_seq = NULL;
> -}
> -#endif /* CONFIG_SLAB_FREELIST_RANDOM */
> -
> -#if defined(CONFIG_SLAB) || defined(CONFIG_SLUB_DEBUG)
> -#ifdef CONFIG_SLAB
> -#define SLABINFO_RIGHTS (0600)
> -#else
> -#define SLABINFO_RIGHTS (0400)
> -#endif
> -
> -static void print_slabinfo_header(struct seq_file *m)
> -{
> -       /*
> -        * Output format version, so at least we can change it
> -        * without _too_ many complaints.
> -        */
> -#ifdef CONFIG_DEBUG_SLAB
> -       seq_puts(m, "slabinfo - version: 2.1 (statistics)\n");
> -#else
> -       seq_puts(m, "slabinfo - version: 2.1\n");
> -#endif
> -       seq_puts(m, "# name            <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>");
> -       seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>");
> -       seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail>");
> -#ifdef CONFIG_DEBUG_SLAB
> -       seq_puts(m, " : globalstat <listallocs> <maxobjs> <grown> <reaped> <error> <maxfreeable> <nodeallocs> <remotefrees> <alienoverflow>");
> -       seq_puts(m, " : cpustat <allochit> <allocmiss> <freehit> <freemiss>");
> -#endif
> -       seq_putc(m, '\n');
> -}
> -
> -static void *slab_start(struct seq_file *m, loff_t *pos)
> -{
> -       mutex_lock(&slab_mutex);
> -       return seq_list_start(&slab_caches, *pos);
> -}
> -
> -static void *slab_next(struct seq_file *m, void *p, loff_t *pos)
> -{
> -       return seq_list_next(p, &slab_caches, pos);
> -}
> -
> -static void slab_stop(struct seq_file *m, void *p)
> -{
> -       mutex_unlock(&slab_mutex);
> -}
> -
> -static void cache_show(struct kmem_cache *s, struct seq_file *m)
> +static __always_inline void *
> +__do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller)
>  {
> -       struct slabinfo sinfo;
> +       void *ret;
>
> -       memset(&sinfo, 0, sizeof(sinfo));
> -       get_slabinfo(s, &sinfo);
> +       gfp &= gfp_allowed_mask;
>
> -       seq_printf(m, "%-17s %6lu %6lu %6u %4u %4d",
> -                  s->name, sinfo.active_objs, sinfo.num_objs, s->size,
> -                  sinfo.objects_per_slab, (1 << sinfo.cache_order));
> +       might_alloc(gfp);
>
> -       seq_printf(m, " : tunables %4u %4u %4u",
> -                  sinfo.limit, sinfo.batchcount, sinfo.shared);
> -       seq_printf(m, " : slabdata %6lu %6lu %6lu",
> -                  sinfo.active_slabs, sinfo.num_slabs, sinfo.shared_avail);
> -       slabinfo_show_stats(m, s);
> -       seq_putc(m, '\n');
> -}
> +       if (unlikely(!size))
> +               ret = ZERO_SIZE_PTR;
> +       else
> +               ret = __kmalloc_large_node(size, gfp, node);
>
> -static int slab_show(struct seq_file *m, void *p)
> -{
> -       struct kmem_cache *s = list_entry(p, struct kmem_cache, list);
> +       trace_kmalloc(caller, ret, size, PAGE_SIZE << get_order(size), gfp, node);
>
> -       if (p == slab_caches.next)
> -               print_slabinfo_header(m);
> -       cache_show(s, m);
> -       return 0;
> +       kmemleak_alloc(ret, size, 1, gfp);
> +       return ret;
>  }
>
> -void dump_unreclaimable_slab(void)
> +void *__kmalloc(size_t size, gfp_t gfp)
>  {
> -       struct kmem_cache *s;
> -       struct slabinfo sinfo;
> -
> -       /*
> -        * Here acquiring slab_mutex is risky since we don't prefer to get
> -        * sleep in oom path. But, without mutex hold, it may introduce a
> -        * risk of crash.
> -        * Use mutex_trylock to protect the list traverse, dump nothing
> -        * without acquiring the mutex.
> -        */
> -       if (!mutex_trylock(&slab_mutex)) {
> -               pr_warn("excessive unreclaimable slab but cannot dump stats\n");
> -               return;
> -       }
> -
> -       pr_info("Unreclaimable slab info:\n");
> -       pr_info("Name                      Used          Total\n");
> -
> -       list_for_each_entry(s, &slab_caches, list) {
> -               if (s->flags & SLAB_RECLAIM_ACCOUNT)
> -                       continue;
> -
> -               get_slabinfo(s, &sinfo);
> -
> -               if (sinfo.num_objs > 0)
> -                       pr_info("%-17s %10luKB %10luKB\n", s->name,
> -                               (sinfo.active_objs * s->size) / 1024,
> -                               (sinfo.num_objs * s->size) / 1024);
> -       }
> -       mutex_unlock(&slab_mutex);
> +       return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, _RET_IP_);
>  }
> +EXPORT_SYMBOL(__kmalloc);
>
> -/*
> - * slabinfo_op - iterator that generates /proc/slabinfo
> - *
> - * Output layout:
> - * cache-name
> - * num-active-objs
> - * total-objs
> - * object size
> - * num-active-slabs
> - * total-slabs
> - * num-pages-per-slab
> - * + further values on SMP and with statistics enabled
> - */
> -static const struct seq_operations slabinfo_op = {
> -       .start = slab_start,
> -       .next = slab_next,
> -       .stop = slab_stop,
> -       .show = slab_show,
> -};
> -
> -static int slabinfo_open(struct inode *inode, struct file *file)
> +void *__kmalloc_node(size_t size, gfp_t gfp, int node)
>  {
> -       return seq_open(file, &slabinfo_op);
> +       return __do_kmalloc_node(size, gfp, node, _RET_IP_);
>  }
> +EXPORT_SYMBOL(__kmalloc_node);
>
> -static const struct proc_ops slabinfo_proc_ops = {
> -       .proc_flags     = PROC_ENTRY_PERMANENT,
> -       .proc_open      = slabinfo_open,
> -       .proc_read      = seq_read,
> -       .proc_write     = slabinfo_write,
> -       .proc_lseek     = seq_lseek,
> -       .proc_release   = seq_release,
> -};
> -
> -static int __init slab_proc_init(void)
> +void *__kmalloc_node_track_caller(size_t size, gfp_t gfp,
> +                                       int node, unsigned long caller)
>  {
> -       proc_create("slabinfo", SLABINFO_RIGHTS, NULL, &slabinfo_proc_ops);
> -       return 0;
> +       return __do_kmalloc_node(size, gfp, node, caller);
>  }
> -module_init(slab_proc_init);
> -
> -#endif /* CONFIG_SLAB || CONFIG_SLUB_DEBUG */
> +EXPORT_SYMBOL(__kmalloc_node_track_caller);
>
>  static __always_inline __realloc_size(2) void *
>  __do_krealloc(const void *p, size_t new_size, gfp_t flags)
> @@ -1402,6 +784,27 @@ void *krealloc(const void *p, size_t new_size, gfp_t flags)
>  }
>  EXPORT_SYMBOL(krealloc);
>
> +void kfree(const void *block)
> +{
> +       struct folio *sp;
> +       unsigned int order;
> +
> +       trace_kfree(_RET_IP_, block);
> +
> +       if (unlikely(ZERO_OR_NULL_PTR(block)))
> +               return;
> +       kmemleak_free(block);
> +
> +       sp = virt_to_folio(block);
> +       BUG_ON(folio_test_slab(sp));
> +       order = folio_order(sp);
> +
> +       mod_node_page_state(folio_pgdat(sp), NR_SLAB_UNRECLAIMABLE_B,
> +                           -(PAGE_SIZE << order));
> +       __free_pages(folio_page(sp, 0), order);
> +}
> +EXPORT_SYMBOL(kfree);
> +
>  /**
>   * kfree_sensitive - Clear sensitive information in memory before freeing
>   * @p: object to free memory of
> @@ -1427,6 +830,34 @@ void kfree_sensitive(const void *p)
>  }
>  EXPORT_SYMBOL(kfree_sensitive);
>
> +size_t kmalloc_size_roundup(size_t size)
> +{
> +       /* Short-circuit the 0 size case. */
> +       if (unlikely(size == 0))
> +               return 0;
> +       /* Short-circuit saturated "too-large" case. */
> +       if (unlikely(size == SIZE_MAX))
> +               return SIZE_MAX;
> +
> +       return PAGE_SIZE << get_order(size);
> +}
> +
> +EXPORT_SYMBOL(kmalloc_size_roundup);
> +
> +/* can't use ksize for kmem_cache_alloc memory, only kmalloc */
> +size_t __ksize(const void *block)
> +{
> +       struct folio *folio;
> +
> +       BUG_ON(!block);
> +       if (unlikely(block == ZERO_SIZE_PTR))
> +               return 0;
> +
> +       folio = virt_to_folio(block);
> +       BUG_ON(folio_test_slab(folio));
> +       return folio_size(folio);
> +}
> +
>  size_t ksize(const void *objp)
>  {
>         /*
> @@ -1451,6 +882,131 @@ size_t ksize(const void *objp)
>  }
>  EXPORT_SYMBOL(ksize);
>
> +static void *__kmem_cache_alloc_node(struct kmem_cache *c, gfp_t flags, int node)
> +{
> +       void *b;
> +
> +       flags &= gfp_allowed_mask;
> +
> +       might_alloc(flags);
> +
> +       b = __kmalloc_large_node(c->size, flags, node);
> +       trace_kmem_cache_alloc(_RET_IP_, b, c, flags, node);
> +
> +       if (b && c->ctor) {
> +               WARN_ON_ONCE(flags & __GFP_ZERO);
> +               c->ctor(b);
> +       }
> +
> +       kmemleak_alloc_recursive(b, c->size, 1, c->flags, flags);
> +       return b;
> +}
> +
> +void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags)
> +{
> +       return __kmem_cache_alloc_node(cachep, flags, NUMA_NO_NODE);
> +}
> +EXPORT_SYMBOL(kmem_cache_alloc);
> +
> +
> +void *kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru, gfp_t flags)
> +{
> +       return __kmem_cache_alloc_node(cachep, flags, NUMA_NO_NODE);
> +}
> +EXPORT_SYMBOL(kmem_cache_alloc_lru);
> +
> +void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node)
> +{
> +       return __kmem_cache_alloc_node(cachep, gfp, node);
> +}
> +EXPORT_SYMBOL(kmem_cache_alloc_node);
> +
> +static void kmem_rcu_free(struct rcu_head *head)
> +{
> +       struct slab_rcu *slab_rcu = (struct slab_rcu *)head;
> +       void *b = (void *)slab_rcu - (slab_rcu->size - sizeof(struct slab_rcu));
> +
> +       kfree(b);
> +}
> +
> +void kmem_cache_free(struct kmem_cache *c, void *b)
> +{
> +       kmemleak_free_recursive(b, c->flags);
> +       trace_kmem_cache_free(_RET_IP_, b, c);
> +       if (unlikely(c->flags & SLAB_TYPESAFE_BY_RCU)) {
> +               struct slab_rcu *slab_rcu;
> +               slab_rcu = b + (c->size - sizeof(struct slab_rcu));
> +               slab_rcu->size = c->size;
> +               call_rcu(&slab_rcu->head, kmem_rcu_free);
> +       } else {
> +               kfree(b);
> +       }
> +}
> +EXPORT_SYMBOL(kmem_cache_free);
> +
> +void kmem_cache_free_bulk(struct kmem_cache *s, size_t nr, void **p)
> +{
> +       size_t i;
> +
> +       for (i = 0; i < nr; i++) {
> +               if (s)
> +                       kmem_cache_free(s, p[i]);
> +               else
> +                       kfree(p[i]);
> +       }
> +}
> +EXPORT_SYMBOL(kmem_cache_free_bulk);
> +
> +int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t nr,
> +                                                               void **p)
> +{
> +       size_t i;
> +
> +       for (i = 0; i < nr; i++) {
> +               void *x = p[i] = kmem_cache_alloc(s, flags);
> +
> +               if (!x) {
> +                       kmem_cache_free_bulk(s, i, p);
> +                       return 0;
> +               }
> +       }
> +       return i;
> +}
> +EXPORT_SYMBOL(kmem_cache_alloc_bulk);
> +
> +int __kmem_cache_shutdown(struct kmem_cache *c)
> +{
> +       /* No way to check for remaining objects */
> +       return 0;
> +}
> +
> +void __kmem_cache_release(struct kmem_cache *c)
> +{
> +}
> +
> +int __kmem_cache_shrink(struct kmem_cache *d)
> +{
> +       return 0;
> +}
> +
> +static struct kmem_cache kmem_cache_boot = {
> +       .name = "kmem_cache",
> +       .size = sizeof(struct kmem_cache),
> +       .flags = SLAB_PANIC,
> +       .align = ARCH_KMALLOC_MINALIGN,
> +};
> +
> +void __init kmem_cache_init(void)
> +{
> +       kmem_cache = &kmem_cache_boot;
> +       slab_state = UP;
> +}
> +
> +void __init kmem_cache_init_late(void)
> +{
> +       slab_state = FULL;
> +}
> +
>  /* Tracepoints definitions. */
>  EXPORT_TRACEPOINT_SYMBOL(kmalloc);
>  EXPORT_TRACEPOINT_SYMBOL(kmem_cache_alloc);
> diff --git a/mm/slob.c b/mm/slob.c
> deleted file mode 100644
> index fe567fcfa3a3..000000000000
> --- a/mm/slob.c
> +++ /dev/null
> @@ -1,757 +0,0 @@
> -// SPDX-License-Identifier: GPL-2.0
> -/*
> - * SLOB Allocator: Simple List Of Blocks
> - *
> - * Matt Mackall <mpm@...enic.com> 12/30/03
> - *
> - * NUMA support by Paul Mundt, 2007.
> - *
> - * How SLOB works:
> - *
> - * The core of SLOB is a traditional K&R style heap allocator, with
> - * support for returning aligned objects. The granularity of this
> - * allocator is as little as 2 bytes, however typically most architectures
> - * will require 4 bytes on 32-bit and 8 bytes on 64-bit.
> - *
> - * The slob heap is a set of linked list of pages from alloc_pages(),
> - * and within each page, there is a singly-linked list of free blocks
> - * (slob_t). The heap is grown on demand. To reduce fragmentation,
> - * heap pages are segregated into three lists, with objects less than
> - * 256 bytes, objects less than 1024 bytes, and all other objects.
> - *
> - * Allocation from heap involves first searching for a page with
> - * sufficient free blocks (using a next-fit-like approach) followed by
> - * a first-fit scan of the page. Deallocation inserts objects back
> - * into the free list in address order, so this is effectively an
> - * address-ordered first fit.
> - *
> - * Above this is an implementation of kmalloc/kfree. Blocks returned
> - * from kmalloc are prepended with a 4-byte header with the kmalloc size.
> - * If kmalloc is asked for objects of PAGE_SIZE or larger, it calls
> - * alloc_pages() directly, allocating compound pages so the page order
> - * does not have to be separately tracked.
> - * These objects are detected in kfree() because folio_test_slab()
> - * is false for them.
> - *
> - * SLAB is emulated on top of SLOB by simply calling constructors and
> - * destructors for every SLAB allocation. Objects are returned with the
> - * 4-byte alignment unless the SLAB_HWCACHE_ALIGN flag is set, in which
> - * case the low-level allocator will fragment blocks to create the proper
> - * alignment. Again, objects of page-size or greater are allocated by
> - * calling alloc_pages(). As SLAB objects know their size, no separate
> - * size bookkeeping is necessary and there is essentially no allocation
> - * space overhead, and compound pages aren't needed for multi-page
> - * allocations.
> - *
> - * NUMA support in SLOB is fairly simplistic, pushing most of the real
> - * logic down to the page allocator, and simply doing the node accounting
> - * on the upper levels. In the event that a node id is explicitly
> - * provided, __alloc_pages_node() with the specified node id is used
> - * instead. The common case (or when the node id isn't explicitly provided)
> - * will default to the current node, as per numa_node_id().
> - *
> - * Node aware pages are still inserted in to the global freelist, and
> - * these are scanned for by matching against the node id encoded in the
> - * page flags. As a result, block allocations that can be satisfied from
> - * the freelist will only be done so on pages residing on the same node,
> - * in order to prevent random node placement.
> - */
> -
> -#include <linux/kernel.h>
> -#include <linux/slab.h>
> -
> -#include <linux/mm.h>
> -#include <linux/swap.h> /* struct reclaim_state */
> -#include <linux/cache.h>
> -#include <linux/init.h>
> -#include <linux/export.h>
> -#include <linux/rcupdate.h>
> -#include <linux/list.h>
> -#include <linux/kmemleak.h>
> -
> -#include <trace/events/kmem.h>
> -
> -#include <linux/atomic.h>
> -
> -#include "slab.h"
> -/*
> - * slob_block has a field 'units', which indicates size of block if +ve,
> - * or offset of next block if -ve (in SLOB_UNITs).
> - *
> - * Free blocks of size 1 unit simply contain the offset of the next block.
> - * Those with larger size contain their size in the first SLOB_UNIT of
> - * memory, and the offset of the next free block in the second SLOB_UNIT.
> - */
> -#if PAGE_SIZE <= (32767 * 2)
> -typedef s16 slobidx_t;
> -#else
> -typedef s32 slobidx_t;
> -#endif
> -
> -struct slob_block {
> -       slobidx_t units;
> -};
> -typedef struct slob_block slob_t;
> -
> -/*
> - * All partially free slob pages go on these lists.
> - */
> -#define SLOB_BREAK1 256
> -#define SLOB_BREAK2 1024
> -static LIST_HEAD(free_slob_small);
> -static LIST_HEAD(free_slob_medium);
> -static LIST_HEAD(free_slob_large);
> -
> -/*
> - * slob_page_free: true for pages on free_slob_pages list.
> - */
> -static inline int slob_page_free(struct slab *slab)
> -{
> -       return PageSlobFree(slab_page(slab));
> -}
> -
> -static void set_slob_page_free(struct slab *slab, struct list_head *list)
> -{
> -       list_add(&slab->slab_list, list);
> -       __SetPageSlobFree(slab_page(slab));
> -}
> -
> -static inline void clear_slob_page_free(struct slab *slab)
> -{
> -       list_del(&slab->slab_list);
> -       __ClearPageSlobFree(slab_page(slab));
> -}
> -
> -#define SLOB_UNIT sizeof(slob_t)
> -#define SLOB_UNITS(size) DIV_ROUND_UP(size, SLOB_UNIT)
> -
> -/*
> - * struct slob_rcu is inserted at the tail of allocated slob blocks, which
> - * were created with a SLAB_TYPESAFE_BY_RCU slab. slob_rcu is used to free
> - * the block using call_rcu.
> - */
> -struct slob_rcu {
> -       struct rcu_head head;
> -       int size;
> -};
> -
> -/*
> - * slob_lock protects all slob allocator structures.
> - */
> -static DEFINE_SPINLOCK(slob_lock);
> -
> -/*
> - * Encode the given size and next info into a free slob block s.
> - */
> -static void set_slob(slob_t *s, slobidx_t size, slob_t *next)
> -{
> -       slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK);
> -       slobidx_t offset = next - base;
> -
> -       if (size > 1) {
> -               s[0].units = size;
> -               s[1].units = offset;
> -       } else
> -               s[0].units = -offset;
> -}
> -
> -/*
> - * Return the size of a slob block.
> - */
> -static slobidx_t slob_units(slob_t *s)
> -{
> -       if (s->units > 0)
> -               return s->units;
> -       return 1;
> -}
> -
> -/*
> - * Return the next free slob block pointer after this one.
> - */
> -static slob_t *slob_next(slob_t *s)
> -{
> -       slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK);
> -       slobidx_t next;
> -
> -       if (s[0].units < 0)
> -               next = -s[0].units;
> -       else
> -               next = s[1].units;
> -       return base+next;
> -}
> -
> -/*
> - * Returns true if s is the last free block in its page.
> - */
> -static int slob_last(slob_t *s)
> -{
> -       return !((unsigned long)slob_next(s) & ~PAGE_MASK);
> -}
> -
> -static void *slob_new_pages(gfp_t gfp, int order, int node)
> -{
> -       struct page *page;
> -
> -#ifdef CONFIG_NUMA
> -       if (node != NUMA_NO_NODE)
> -               page = __alloc_pages_node(node, gfp, order);
> -       else
> -#endif
> -               page = alloc_pages(gfp, order);
> -
> -       if (!page)
> -               return NULL;
> -
> -       mod_node_page_state(page_pgdat(page), NR_SLAB_UNRECLAIMABLE_B,
> -                           PAGE_SIZE << order);
> -       return page_address(page);
> -}
> -
> -static void slob_free_pages(void *b, int order)
> -{
> -       struct page *sp = virt_to_page(b);
> -
> -       if (current->reclaim_state)
> -               current->reclaim_state->reclaimed_slab += 1 << order;
> -
> -       mod_node_page_state(page_pgdat(sp), NR_SLAB_UNRECLAIMABLE_B,
> -                           -(PAGE_SIZE << order));
> -       __free_pages(sp, order);
> -}
> -
> -/*
> - * slob_page_alloc() - Allocate a slob block within a given slob_page sp.
> - * @sp: Page to look in.
> - * @size: Size of the allocation.
> - * @align: Allocation alignment.
> - * @align_offset: Offset in the allocated block that will be aligned.
> - * @page_removed_from_list: Return parameter.
> - *
> - * Tries to find a chunk of memory at least @size bytes big within @page.
> - *
> - * Return: Pointer to memory if allocated, %NULL otherwise.  If the
> - *         allocation fills up @page then the page is removed from the
> - *         freelist, in this case @page_removed_from_list will be set to
> - *         true (set to false otherwise).
> - */
> -static void *slob_page_alloc(struct slab *sp, size_t size, int align,
> -                             int align_offset, bool *page_removed_from_list)
> -{
> -       slob_t *prev, *cur, *aligned = NULL;
> -       int delta = 0, units = SLOB_UNITS(size);
> -
> -       *page_removed_from_list = false;
> -       for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) {
> -               slobidx_t avail = slob_units(cur);
> -
> -               /*
> -                * 'aligned' will hold the address of the slob block so that the
> -                * address 'aligned'+'align_offset' is aligned according to the
> -                * 'align' parameter. This is for kmalloc() which prepends the
> -                * allocated block with its size, so that the block itself is
> -                * aligned when needed.
> -                */
> -               if (align) {
> -                       aligned = (slob_t *)
> -                               (ALIGN((unsigned long)cur + align_offset, align)
> -                                - align_offset);
> -                       delta = aligned - cur;
> -               }
> -               if (avail >= units + delta) { /* room enough? */
> -                       slob_t *next;
> -
> -                       if (delta) { /* need to fragment head to align? */
> -                               next = slob_next(cur);
> -                               set_slob(aligned, avail - delta, next);
> -                               set_slob(cur, delta, aligned);
> -                               prev = cur;
> -                               cur = aligned;
> -                               avail = slob_units(cur);
> -                       }
> -
> -                       next = slob_next(cur);
> -                       if (avail == units) { /* exact fit? unlink. */
> -                               if (prev)
> -                                       set_slob(prev, slob_units(prev), next);
> -                               else
> -                                       sp->freelist = next;
> -                       } else { /* fragment */
> -                               if (prev)
> -                                       set_slob(prev, slob_units(prev), cur + units);
> -                               else
> -                                       sp->freelist = cur + units;
> -                               set_slob(cur + units, avail - units, next);
> -                       }
> -
> -                       sp->units -= units;
> -                       if (!sp->units) {
> -                               clear_slob_page_free(sp);
> -                               *page_removed_from_list = true;
> -                       }
> -                       return cur;
> -               }
> -               if (slob_last(cur))
> -                       return NULL;
> -       }
> -}
> -
> -/*
> - * slob_alloc: entry point into the slob allocator.
> - */
> -static void *slob_alloc(size_t size, gfp_t gfp, int align, int node,
> -                                                       int align_offset)
> -{
> -       struct folio *folio;
> -       struct slab *sp;
> -       struct list_head *slob_list;
> -       slob_t *b = NULL;
> -       unsigned long flags;
> -       bool _unused;
> -
> -       if (size < SLOB_BREAK1)
> -               slob_list = &free_slob_small;
> -       else if (size < SLOB_BREAK2)
> -               slob_list = &free_slob_medium;
> -       else
> -               slob_list = &free_slob_large;
> -
> -       spin_lock_irqsave(&slob_lock, flags);
> -       /* Iterate through each partially free page, try to find room */
> -       list_for_each_entry(sp, slob_list, slab_list) {
> -               bool page_removed_from_list = false;
> -#ifdef CONFIG_NUMA
> -               /*
> -                * If there's a node specification, search for a partial
> -                * page with a matching node id in the freelist.
> -                */
> -               if (node != NUMA_NO_NODE && slab_nid(sp) != node)
> -                       continue;
> -#endif
> -               /* Enough room on this page? */
> -               if (sp->units < SLOB_UNITS(size))
> -                       continue;
> -
> -               b = slob_page_alloc(sp, size, align, align_offset, &page_removed_from_list);
> -               if (!b)
> -                       continue;
> -
> -               /*
> -                * If slob_page_alloc() removed sp from the list then we
> -                * cannot call list functions on sp.  If so allocation
> -                * did not fragment the page anyway so optimisation is
> -                * unnecessary.
> -                */
> -               if (!page_removed_from_list) {
> -                       /*
> -                        * Improve fragment distribution and reduce our average
> -                        * search time by starting our next search here. (see
> -                        * Knuth vol 1, sec 2.5, pg 449)
> -                        */
> -                       if (!list_is_first(&sp->slab_list, slob_list))
> -                               list_rotate_to_front(&sp->slab_list, slob_list);
> -               }
> -               break;
> -       }
> -       spin_unlock_irqrestore(&slob_lock, flags);
> -
> -       /* Not enough space: must allocate a new page */
> -       if (!b) {
> -               b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node);
> -               if (!b)
> -                       return NULL;
> -               folio = virt_to_folio(b);
> -               __folio_set_slab(folio);
> -               sp = folio_slab(folio);
> -
> -               spin_lock_irqsave(&slob_lock, flags);
> -               sp->units = SLOB_UNITS(PAGE_SIZE);
> -               sp->freelist = b;
> -               INIT_LIST_HEAD(&sp->slab_list);
> -               set_slob(b, SLOB_UNITS(PAGE_SIZE), b + SLOB_UNITS(PAGE_SIZE));
> -               set_slob_page_free(sp, slob_list);
> -               b = slob_page_alloc(sp, size, align, align_offset, &_unused);
> -               BUG_ON(!b);
> -               spin_unlock_irqrestore(&slob_lock, flags);
> -       }
> -       if (unlikely(gfp & __GFP_ZERO))
> -               memset(b, 0, size);
> -       return b;
> -}
> -
> -/*
> - * slob_free: entry point into the slob allocator.
> - */
> -static void slob_free(void *block, int size)
> -{
> -       struct slab *sp;
> -       slob_t *prev, *next, *b = (slob_t *)block;
> -       slobidx_t units;
> -       unsigned long flags;
> -       struct list_head *slob_list;
> -
> -       if (unlikely(ZERO_OR_NULL_PTR(block)))
> -               return;
> -       BUG_ON(!size);
> -
> -       sp = virt_to_slab(block);
> -       units = SLOB_UNITS(size);
> -
> -       spin_lock_irqsave(&slob_lock, flags);
> -
> -       if (sp->units + units == SLOB_UNITS(PAGE_SIZE)) {
> -               /* Go directly to page allocator. Do not pass slob allocator */
> -               if (slob_page_free(sp))
> -                       clear_slob_page_free(sp);
> -               spin_unlock_irqrestore(&slob_lock, flags);
> -               __folio_clear_slab(slab_folio(sp));
> -               slob_free_pages(b, 0);
> -               return;
> -       }
> -
> -       if (!slob_page_free(sp)) {
> -               /* This slob page is about to become partially free. Easy! */
> -               sp->units = units;
> -               sp->freelist = b;
> -               set_slob(b, units,
> -                       (void *)((unsigned long)(b +
> -                                       SLOB_UNITS(PAGE_SIZE)) & PAGE_MASK));
> -               if (size < SLOB_BREAK1)
> -                       slob_list = &free_slob_small;
> -               else if (size < SLOB_BREAK2)
> -                       slob_list = &free_slob_medium;
> -               else
> -                       slob_list = &free_slob_large;
> -               set_slob_page_free(sp, slob_list);
> -               goto out;
> -       }
> -
> -       /*
> -        * Otherwise the page is already partially free, so find reinsertion
> -        * point.
> -        */
> -       sp->units += units;
> -
> -       if (b < (slob_t *)sp->freelist) {
> -               if (b + units == sp->freelist) {
> -                       units += slob_units(sp->freelist);
> -                       sp->freelist = slob_next(sp->freelist);
> -               }
> -               set_slob(b, units, sp->freelist);
> -               sp->freelist = b;
> -       } else {
> -               prev = sp->freelist;
> -               next = slob_next(prev);
> -               while (b > next) {
> -                       prev = next;
> -                       next = slob_next(prev);
> -               }
> -
> -               if (!slob_last(prev) && b + units == next) {
> -                       units += slob_units(next);
> -                       set_slob(b, units, slob_next(next));
> -               } else
> -                       set_slob(b, units, next);
> -
> -               if (prev + slob_units(prev) == b) {
> -                       units = slob_units(b) + slob_units(prev);
> -                       set_slob(prev, units, slob_next(b));
> -               } else
> -                       set_slob(prev, slob_units(prev), b);
> -       }
> -out:
> -       spin_unlock_irqrestore(&slob_lock, flags);
> -}
> -
> -#ifdef CONFIG_PRINTK
> -void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> -{
> -       kpp->kp_ptr = object;
> -       kpp->kp_slab = slab;
> -}
> -#endif
> -
> -/*
> - * End of slob allocator proper. Begin kmem_cache_alloc and kmalloc frontend.
> - */
> -
> -static __always_inline void *
> -__do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller)
> -{
> -       unsigned int *m;
> -       unsigned int minalign;
> -       void *ret;
> -
> -       minalign = max_t(unsigned int, ARCH_KMALLOC_MINALIGN,
> -                        arch_slab_minalign());
> -       gfp &= gfp_allowed_mask;
> -
> -       might_alloc(gfp);
> -
> -       if (size < PAGE_SIZE - minalign) {
> -               int align = minalign;
> -
> -               /*
> -                * For power of two sizes, guarantee natural alignment for
> -                * kmalloc()'d objects.
> -                */
> -               if (is_power_of_2(size))
> -                       align = max_t(unsigned int, minalign, size);
> -
> -               if (!size)
> -                       return ZERO_SIZE_PTR;
> -
> -               m = slob_alloc(size + minalign, gfp, align, node, minalign);
> -
> -               if (!m)
> -                       return NULL;
> -               *m = size;
> -               ret = (void *)m + minalign;
> -
> -               trace_kmalloc(caller, ret, size, size + minalign, gfp, node);
> -       } else {
> -               unsigned int order = get_order(size);
> -
> -               if (likely(order))
> -                       gfp |= __GFP_COMP;
> -               ret = slob_new_pages(gfp, order, node);
> -
> -               trace_kmalloc(caller, ret, size, PAGE_SIZE << order, gfp, node);
> -       }
> -
> -       kmemleak_alloc(ret, size, 1, gfp);
> -       return ret;
> -}
> -
> -void *__kmalloc(size_t size, gfp_t gfp)
> -{
> -       return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, _RET_IP_);
> -}
> -EXPORT_SYMBOL(__kmalloc);
> -
> -void *__kmalloc_node_track_caller(size_t size, gfp_t gfp,
> -                                       int node, unsigned long caller)
> -{
> -       return __do_kmalloc_node(size, gfp, node, caller);
> -}
> -EXPORT_SYMBOL(__kmalloc_node_track_caller);
> -
> -void kfree(const void *block)
> -{
> -       struct folio *sp;
> -
> -       trace_kfree(_RET_IP_, block);
> -
> -       if (unlikely(ZERO_OR_NULL_PTR(block)))
> -               return;
> -       kmemleak_free(block);
> -
> -       sp = virt_to_folio(block);
> -       if (folio_test_slab(sp)) {
> -               unsigned int align = max_t(unsigned int,
> -                                          ARCH_KMALLOC_MINALIGN,
> -                                          arch_slab_minalign());
> -               unsigned int *m = (unsigned int *)(block - align);
> -
> -               slob_free(m, *m + align);
> -       } else {
> -               unsigned int order = folio_order(sp);
> -
> -               mod_node_page_state(folio_pgdat(sp), NR_SLAB_UNRECLAIMABLE_B,
> -                                   -(PAGE_SIZE << order));
> -               __free_pages(folio_page(sp, 0), order);
> -
> -       }
> -}
> -EXPORT_SYMBOL(kfree);
> -
> -size_t kmalloc_size_roundup(size_t size)
> -{
> -       /* Short-circuit the 0 size case. */
> -       if (unlikely(size == 0))
> -               return 0;
> -       /* Short-circuit saturated "too-large" case. */
> -       if (unlikely(size == SIZE_MAX))
> -               return SIZE_MAX;
> -
> -       return ALIGN(size, ARCH_KMALLOC_MINALIGN);
> -}
> -
> -EXPORT_SYMBOL(kmalloc_size_roundup);
> -
> -/* can't use ksize for kmem_cache_alloc memory, only kmalloc */
> -size_t __ksize(const void *block)
> -{
> -       struct folio *folio;
> -       unsigned int align;
> -       unsigned int *m;
> -
> -       BUG_ON(!block);
> -       if (unlikely(block == ZERO_SIZE_PTR))
> -               return 0;
> -
> -       folio = virt_to_folio(block);
> -       if (unlikely(!folio_test_slab(folio)))
> -               return folio_size(folio);
> -
> -       align = max_t(unsigned int, ARCH_KMALLOC_MINALIGN,
> -                     arch_slab_minalign());
> -       m = (unsigned int *)(block - align);
> -       return SLOB_UNITS(*m) * SLOB_UNIT;
> -}
> -
> -int __kmem_cache_create(struct kmem_cache *c, slab_flags_t flags)
> -{
> -       if (flags & SLAB_TYPESAFE_BY_RCU) {
> -               /* leave room for rcu footer at the end of object */
> -               c->size += sizeof(struct slob_rcu);
> -       }
> -
> -       /* Actual size allocated */
> -       c->size = SLOB_UNITS(c->size) * SLOB_UNIT;
> -       c->flags = flags;
> -       return 0;
> -}
> -
> -static void *slob_alloc_node(struct kmem_cache *c, gfp_t flags, int node)
> -{
> -       void *b;
> -
> -       flags &= gfp_allowed_mask;
> -
> -       might_alloc(flags);
> -
> -       if (c->size < PAGE_SIZE) {
> -               b = slob_alloc(c->size, flags, c->align, node, 0);
> -               trace_kmem_cache_alloc(_RET_IP_, b, c, flags, node);
> -       } else {
> -               b = slob_new_pages(flags, get_order(c->size), node);
> -               trace_kmem_cache_alloc(_RET_IP_, b, c, flags, node);
> -       }
> -
> -       if (b && c->ctor) {
> -               WARN_ON_ONCE(flags & __GFP_ZERO);
> -               c->ctor(b);
> -       }
> -
> -       kmemleak_alloc_recursive(b, c->size, 1, c->flags, flags);
> -       return b;
> -}
> -
> -void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags)
> -{
> -       return slob_alloc_node(cachep, flags, NUMA_NO_NODE);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc);
> -
> -
> -void *kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru, gfp_t flags)
> -{
> -       return slob_alloc_node(cachep, flags, NUMA_NO_NODE);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_lru);
> -
> -void *__kmalloc_node(size_t size, gfp_t gfp, int node)
> -{
> -       return __do_kmalloc_node(size, gfp, node, _RET_IP_);
> -}
> -EXPORT_SYMBOL(__kmalloc_node);
> -
> -void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node)
> -{
> -       return slob_alloc_node(cachep, gfp, node);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_node);
> -
> -static void __kmem_cache_free(void *b, int size)
> -{
> -       if (size < PAGE_SIZE)
> -               slob_free(b, size);
> -       else
> -               slob_free_pages(b, get_order(size));
> -}
> -
> -static void kmem_rcu_free(struct rcu_head *head)
> -{
> -       struct slob_rcu *slob_rcu = (struct slob_rcu *)head;
> -       void *b = (void *)slob_rcu - (slob_rcu->size - sizeof(struct slob_rcu));
> -
> -       __kmem_cache_free(b, slob_rcu->size);
> -}
> -
> -void kmem_cache_free(struct kmem_cache *c, void *b)
> -{
> -       kmemleak_free_recursive(b, c->flags);
> -       trace_kmem_cache_free(_RET_IP_, b, c);
> -       if (unlikely(c->flags & SLAB_TYPESAFE_BY_RCU)) {
> -               struct slob_rcu *slob_rcu;
> -               slob_rcu = b + (c->size - sizeof(struct slob_rcu));
> -               slob_rcu->size = c->size;
> -               call_rcu(&slob_rcu->head, kmem_rcu_free);
> -       } else {
> -               __kmem_cache_free(b, c->size);
> -       }
> -}
> -EXPORT_SYMBOL(kmem_cache_free);
> -
> -void kmem_cache_free_bulk(struct kmem_cache *s, size_t nr, void **p)
> -{
> -       size_t i;
> -
> -       for (i = 0; i < nr; i++) {
> -               if (s)
> -                       kmem_cache_free(s, p[i]);
> -               else
> -                       kfree(p[i]);
> -       }
> -}
> -EXPORT_SYMBOL(kmem_cache_free_bulk);
> -
> -int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t nr,
> -                                                               void **p)
> -{
> -       size_t i;
> -
> -       for (i = 0; i < nr; i++) {
> -               void *x = p[i] = kmem_cache_alloc(s, flags);
> -
> -               if (!x) {
> -                       kmem_cache_free_bulk(s, i, p);
> -                       return 0;
> -               }
> -       }
> -       return i;
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_bulk);
> -
> -int __kmem_cache_shutdown(struct kmem_cache *c)
> -{
> -       /* No way to check for remaining objects */
> -       return 0;
> -}
> -
> -void __kmem_cache_release(struct kmem_cache *c)
> -{
> -}
> -
> -int __kmem_cache_shrink(struct kmem_cache *d)
> -{
> -       return 0;
> -}
> -
> -static struct kmem_cache kmem_cache_boot = {
> -       .name = "kmem_cache",
> -       .size = sizeof(struct kmem_cache),
> -       .flags = SLAB_PANIC,
> -       .align = ARCH_KMALLOC_MINALIGN,
> -};
> -
> -void __init kmem_cache_init(void)
> -{
> -       kmem_cache = &kmem_cache_boot;
> -       slab_state = UP;
> -}
> -
> -void __init kmem_cache_init_late(void)
> -{
> -       slab_state = FULL;
> -}
> diff --git a/mm/slub.c b/mm/slub.c
> deleted file mode 100644
> index 39327e98fce3..000000000000
> --- a/mm/slub.c
> +++ /dev/null
> @@ -1,6506 +0,0 @@
> -// SPDX-License-Identifier: GPL-2.0
> -/*
> - * SLUB: A slab allocator that limits cache line use instead of queuing
> - * objects in per cpu and per node lists.
> - *
> - * The allocator synchronizes using per slab locks or atomic operations
> - * and only uses a centralized lock to manage a pool of partial slabs.
> - *
> - * (C) 2007 SGI, Christoph Lameter
> - * (C) 2011 Linux Foundation, Christoph Lameter
> - */
> -
> -#include <linux/mm.h>
> -#include <linux/swap.h> /* struct reclaim_state */
> -#include <linux/module.h>
> -#include <linux/bit_spinlock.h>
> -#include <linux/interrupt.h>
> -#include <linux/swab.h>
> -#include <linux/bitops.h>
> -#include <linux/slab.h>
> -#include "slab.h"
> -#include <linux/proc_fs.h>
> -#include <linux/seq_file.h>
> -#include <linux/kasan.h>
> -#include <linux/kmsan.h>
> -#include <linux/cpu.h>
> -#include <linux/cpuset.h>
> -#include <linux/mempolicy.h>
> -#include <linux/ctype.h>
> -#include <linux/stackdepot.h>
> -#include <linux/debugobjects.h>
> -#include <linux/kallsyms.h>
> -#include <linux/kfence.h>
> -#include <linux/memory.h>
> -#include <linux/math64.h>
> -#include <linux/fault-inject.h>
> -#include <linux/stacktrace.h>
> -#include <linux/prefetch.h>
> -#include <linux/memcontrol.h>
> -#include <linux/random.h>
> -#include <kunit/test.h>
> -#include <kunit/test-bug.h>
> -#include <linux/sort.h>
> -
> -#include <linux/debugfs.h>
> -#include <trace/events/kmem.h>
> -
> -#include "internal.h"
> -
> -/*
> - * Lock order:
> - *   1. slab_mutex (Global Mutex)
> - *   2. node->list_lock (Spinlock)
> - *   3. kmem_cache->cpu_slab->lock (Local lock)
> - *   4. slab_lock(slab) (Only on some arches)
> - *   5. object_map_lock (Only for debugging)
> - *
> - *   slab_mutex
> - *
> - *   The role of the slab_mutex is to protect the list of all the slabs
> - *   and to synchronize major metadata changes to slab cache structures.
> - *   Also synchronizes memory hotplug callbacks.
> - *
> - *   slab_lock
> - *
> - *   The slab_lock is a wrapper around the page lock, thus it is a bit
> - *   spinlock.
> - *
> - *   The slab_lock is only used on arches that do not have the ability
> - *   to do a cmpxchg_double. It only protects:
> - *
> - *     A. slab->freelist       -> List of free objects in a slab
> - *     B. slab->inuse          -> Number of objects in use
> - *     C. slab->objects        -> Number of objects in slab
> - *     D. slab->frozen         -> frozen state
> - *
> - *   Frozen slabs
> - *
> - *   If a slab is frozen then it is exempt from list management. It is not
> - *   on any list except per cpu partial list. The processor that froze the
> - *   slab is the one who can perform list operations on the slab. Other
> - *   processors may put objects onto the freelist but the processor that
> - *   froze the slab is the only one that can retrieve the objects from the
> - *   slab's freelist.
> - *
> - *   list_lock
> - *
> - *   The list_lock protects the partial and full list on each node and
> - *   the partial slab counter. If taken then no new slabs may be added or
> - *   removed from the lists nor make the number of partial slabs be modified.
> - *   (Note that the total number of slabs is an atomic value that may be
> - *   modified without taking the list lock).
> - *
> - *   The list_lock is a centralized lock and thus we avoid taking it as
> - *   much as possible. As long as SLUB does not have to handle partial
> - *   slabs, operations can continue without any centralized lock. F.e.
> - *   allocating a long series of objects that fill up slabs does not require
> - *   the list lock.
> - *
> - *   For debug caches, all allocations are forced to go through a list_lock
> - *   protected region to serialize against concurrent validation.
> - *
> - *   cpu_slab->lock local lock
> - *
> - *   This locks protect slowpath manipulation of all kmem_cache_cpu fields
> - *   except the stat counters. This is a percpu structure manipulated only by
> - *   the local cpu, so the lock protects against being preempted or interrupted
> - *   by an irq. Fast path operations rely on lockless operations instead.
> - *
> - *   On PREEMPT_RT, the local lock neither disables interrupts nor preemption
> - *   which means the lockless fastpath cannot be used as it might interfere with
> - *   an in-progress slow path operations. In this case the local lock is always
> - *   taken but it still utilizes the freelist for the common operations.
> - *
> - *   lockless fastpaths
> - *
> - *   The fast path allocation (slab_alloc_node()) and freeing (do_slab_free())
> - *   are fully lockless when satisfied from the percpu slab (and when
> - *   cmpxchg_double is possible to use, otherwise slab_lock is taken).
> - *   They also don't disable preemption or migration or irqs. They rely on
> - *   the transaction id (tid) field to detect being preempted or moved to
> - *   another cpu.
> - *
> - *   irq, preemption, migration considerations
> - *
> - *   Interrupts are disabled as part of list_lock or local_lock operations, or
> - *   around the slab_lock operation, in order to make the slab allocator safe
> - *   to use in the context of an irq.
> - *
> - *   In addition, preemption (or migration on PREEMPT_RT) is disabled in the
> - *   allocation slowpath, bulk allocation, and put_cpu_partial(), so that the
> - *   local cpu doesn't change in the process and e.g. the kmem_cache_cpu pointer
> - *   doesn't have to be revalidated in each section protected by the local lock.
> - *
> - * SLUB assigns one slab for allocation to each processor.
> - * Allocations only occur from these slabs called cpu slabs.
> - *
> - * Slabs with free elements are kept on a partial list and during regular
> - * operations no list for full slabs is used. If an object in a full slab is
> - * freed then the slab will show up again on the partial lists.
> - * We track full slabs for debugging purposes though because otherwise we
> - * cannot scan all objects.
> - *
> - * Slabs are freed when they become empty. Teardown and setup is
> - * minimal so we rely on the page allocators per cpu caches for
> - * fast frees and allocs.
> - *
> - * slab->frozen                The slab is frozen and exempt from list processing.
> - *                     This means that the slab is dedicated to a purpose
> - *                     such as satisfying allocations for a specific
> - *                     processor. Objects may be freed in the slab while
> - *                     it is frozen but slab_free will then skip the usual
> - *                     list operations. It is up to the processor holding
> - *                     the slab to integrate the slab into the slab lists
> - *                     when the slab is no longer needed.
> - *
> - *                     One use of this flag is to mark slabs that are
> - *                     used for allocations. Then such a slab becomes a cpu
> - *                     slab. The cpu slab may be equipped with an additional
> - *                     freelist that allows lockless access to
> - *                     free objects in addition to the regular freelist
> - *                     that requires the slab lock.
> - *
> - * SLAB_DEBUG_FLAGS    Slab requires special handling due to debug
> - *                     options set. This moves slab handling out of
> - *                     the fast path and disables lockless freelists.
> - */
> -
> -/*
> - * We could simply use migrate_disable()/enable() but as long as it's a
> - * function call even on !PREEMPT_RT, use inline preempt_disable() there.
> - */
> -#ifndef CONFIG_PREEMPT_RT
> -#define slub_get_cpu_ptr(var)          get_cpu_ptr(var)
> -#define slub_put_cpu_ptr(var)          put_cpu_ptr(var)
> -#define USE_LOCKLESS_FAST_PATH()       (true)
> -#else
> -#define slub_get_cpu_ptr(var)          \
> -({                                     \
> -       migrate_disable();              \
> -       this_cpu_ptr(var);              \
> -})
> -#define slub_put_cpu_ptr(var)          \
> -do {                                   \
> -       (void)(var);                    \
> -       migrate_enable();               \
> -} while (0)
> -#define USE_LOCKLESS_FAST_PATH()       (false)
> -#endif
> -
> -#ifndef CONFIG_SLUB_TINY
> -#define __fastpath_inline __always_inline
> -#else
> -#define __fastpath_inline
> -#endif
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -#ifdef CONFIG_SLUB_DEBUG_ON
> -DEFINE_STATIC_KEY_TRUE(slub_debug_enabled);
> -#else
> -DEFINE_STATIC_KEY_FALSE(slub_debug_enabled);
> -#endif
> -#endif         /* CONFIG_SLUB_DEBUG */
> -
> -/* Structure holding parameters for get_partial() call chain */
> -struct partial_context {
> -       struct slab **slab;
> -       gfp_t flags;
> -       unsigned int orig_size;
> -};
> -
> -static inline bool kmem_cache_debug(struct kmem_cache *s)
> -{
> -       return kmem_cache_debug_flags(s, SLAB_DEBUG_FLAGS);
> -}
> -
> -static inline bool slub_debug_orig_size(struct kmem_cache *s)
> -{
> -       return (kmem_cache_debug_flags(s, SLAB_STORE_USER) &&
> -                       (s->flags & SLAB_KMALLOC));
> -}
> -
> -void *fixup_red_left(struct kmem_cache *s, void *p)
> -{
> -       if (kmem_cache_debug_flags(s, SLAB_RED_ZONE))
> -               p += s->red_left_pad;
> -
> -       return p;
> -}
> -
> -static inline bool kmem_cache_has_cpu_partial(struct kmem_cache *s)
> -{
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       return !kmem_cache_debug(s);
> -#else
> -       return false;
> -#endif
> -}
> -
> -/*
> - * Issues still to be resolved:
> - *
> - * - Support PAGE_ALLOC_DEBUG. Should be easy to do.
> - *
> - * - Variable sizing of the per node arrays
> - */
> -
> -/* Enable to log cmpxchg failures */
> -#undef SLUB_DEBUG_CMPXCHG
> -
> -#ifndef CONFIG_SLUB_TINY
> -/*
> - * Minimum number of partial slabs. These will be left on the partial
> - * lists even if they are empty. kmem_cache_shrink may reclaim them.
> - */
> -#define MIN_PARTIAL 5
> -
> -/*
> - * Maximum number of desirable partial slabs.
> - * The existence of more partial slabs makes kmem_cache_shrink
> - * sort the partial list by the number of objects in use.
> - */
> -#define MAX_PARTIAL 10
> -#else
> -#define MIN_PARTIAL 0
> -#define MAX_PARTIAL 0
> -#endif
> -
> -#define DEBUG_DEFAULT_FLAGS (SLAB_CONSISTENCY_CHECKS | SLAB_RED_ZONE | \
> -                               SLAB_POISON | SLAB_STORE_USER)
> -
> -/*
> - * These debug flags cannot use CMPXCHG because there might be consistency
> - * issues when checking or reading debug information
> - */
> -#define SLAB_NO_CMPXCHG (SLAB_CONSISTENCY_CHECKS | SLAB_STORE_USER | \
> -                               SLAB_TRACE)
> -
> -
> -/*
> - * Debugging flags that require metadata to be stored in the slab.  These get
> - * disabled when slub_debug=O is used and a cache's min order increases with
> - * metadata.
> - */
> -#define DEBUG_METADATA_FLAGS (SLAB_RED_ZONE | SLAB_POISON | SLAB_STORE_USER)
> -
> -#define OO_SHIFT       16
> -#define OO_MASK                ((1 << OO_SHIFT) - 1)
> -#define MAX_OBJS_PER_PAGE      32767 /* since slab.objects is u15 */
> -
> -/* Internal SLUB flags */
> -/* Poison object */
> -#define __OBJECT_POISON                ((slab_flags_t __force)0x80000000U)
> -/* Use cmpxchg_double */
> -#define __CMPXCHG_DOUBLE       ((slab_flags_t __force)0x40000000U)
> -
> -/*
> - * Tracking user of a slab.
> - */
> -#define TRACK_ADDRS_COUNT 16
> -struct track {
> -       unsigned long addr;     /* Called from address */
> -#ifdef CONFIG_STACKDEPOT
> -       depot_stack_handle_t handle;
> -#endif
> -       int cpu;                /* Was running on cpu */
> -       int pid;                /* Pid context */
> -       unsigned long when;     /* When did the operation occur */
> -};
> -
> -enum track_item { TRACK_ALLOC, TRACK_FREE };
> -
> -#ifdef SLAB_SUPPORTS_SYSFS
> -static int sysfs_slab_add(struct kmem_cache *);
> -static int sysfs_slab_alias(struct kmem_cache *, const char *);
> -#else
> -static inline int sysfs_slab_add(struct kmem_cache *s) { return 0; }
> -static inline int sysfs_slab_alias(struct kmem_cache *s, const char *p)
> -                                                       { return 0; }
> -#endif
> -
> -#if defined(CONFIG_DEBUG_FS) && defined(CONFIG_SLUB_DEBUG)
> -static void debugfs_slab_add(struct kmem_cache *);
> -#else
> -static inline void debugfs_slab_add(struct kmem_cache *s) { }
> -#endif
> -
> -static inline void stat(const struct kmem_cache *s, enum stat_item si)
> -{
> -#ifdef CONFIG_SLUB_STATS
> -       /*
> -        * The rmw is racy on a preemptible kernel but this is acceptable, so
> -        * avoid this_cpu_add()'s irq-disable overhead.
> -        */
> -       raw_cpu_inc(s->cpu_slab->stat[si]);
> -#endif
> -}
> -
> -/*
> - * Tracks for which NUMA nodes we have kmem_cache_nodes allocated.
> - * Corresponds to node_state[N_NORMAL_MEMORY], but can temporarily
> - * differ during memory hotplug/hotremove operations.
> - * Protected by slab_mutex.
> - */
> -static nodemask_t slab_nodes;
> -
> -#ifndef CONFIG_SLUB_TINY
> -/*
> - * Workqueue used for flush_cpu_slab().
> - */
> -static struct workqueue_struct *flushwq;
> -#endif
> -
> -/********************************************************************
> - *                     Core slab cache functions
> - *******************************************************************/
> -
> -/*
> - * Returns freelist pointer (ptr). With hardening, this is obfuscated
> - * with an XOR of the address where the pointer is held and a per-cache
> - * random number.
> - */
> -static inline void *freelist_ptr(const struct kmem_cache *s, void *ptr,
> -                                unsigned long ptr_addr)
> -{
> -#ifdef CONFIG_SLAB_FREELIST_HARDENED
> -       /*
> -        * When CONFIG_KASAN_SW/HW_TAGS is enabled, ptr_addr might be tagged.
> -        * Normally, this doesn't cause any issues, as both set_freepointer()
> -        * and get_freepointer() are called with a pointer with the same tag.
> -        * However, there are some issues with CONFIG_SLUB_DEBUG code. For
> -        * example, when __free_slub() iterates over objects in a cache, it
> -        * passes untagged pointers to check_object(). check_object() in turns
> -        * calls get_freepointer() with an untagged pointer, which causes the
> -        * freepointer to be restored incorrectly.
> -        */
> -       return (void *)((unsigned long)ptr ^ s->random ^
> -                       swab((unsigned long)kasan_reset_tag((void *)ptr_addr)));
> -#else
> -       return ptr;
> -#endif
> -}
> -
> -/* Returns the freelist pointer recorded at location ptr_addr. */
> -static inline void *freelist_dereference(const struct kmem_cache *s,
> -                                        void *ptr_addr)
> -{
> -       return freelist_ptr(s, (void *)*(unsigned long *)(ptr_addr),
> -                           (unsigned long)ptr_addr);
> -}
> -
> -static inline void *get_freepointer(struct kmem_cache *s, void *object)
> -{
> -       object = kasan_reset_tag(object);
> -       return freelist_dereference(s, object + s->offset);
> -}
> -
> -#ifndef CONFIG_SLUB_TINY
> -static void prefetch_freepointer(const struct kmem_cache *s, void *object)
> -{
> -       prefetchw(object + s->offset);
> -}
> -#endif
> -
> -/*
> - * When running under KMSAN, get_freepointer_safe() may return an uninitialized
> - * pointer value in the case the current thread loses the race for the next
> - * memory chunk in the freelist. In that case this_cpu_cmpxchg_double() in
> - * slab_alloc_node() will fail, so the uninitialized value won't be used, but
> - * KMSAN will still check all arguments of cmpxchg because of imperfect
> - * handling of inline assembly.
> - * To work around this problem, we apply __no_kmsan_checks to ensure that
> - * get_freepointer_safe() returns initialized memory.
> - */
> -__no_kmsan_checks
> -static inline void *get_freepointer_safe(struct kmem_cache *s, void *object)
> -{
> -       unsigned long freepointer_addr;
> -       void *p;
> -
> -       if (!debug_pagealloc_enabled_static())
> -               return get_freepointer(s, object);
> -
> -       object = kasan_reset_tag(object);
> -       freepointer_addr = (unsigned long)object + s->offset;
> -       copy_from_kernel_nofault(&p, (void **)freepointer_addr, sizeof(p));
> -       return freelist_ptr(s, p, freepointer_addr);
> -}
> -
> -static inline void set_freepointer(struct kmem_cache *s, void *object, void *fp)
> -{
> -       unsigned long freeptr_addr = (unsigned long)object + s->offset;
> -
> -#ifdef CONFIG_SLAB_FREELIST_HARDENED
> -       BUG_ON(object == fp); /* naive detection of double free or corruption */
> -#endif
> -
> -       freeptr_addr = (unsigned long)kasan_reset_tag((void *)freeptr_addr);
> -       *(void **)freeptr_addr = freelist_ptr(s, fp, freeptr_addr);
> -}
> -
> -/* Loop over all objects in a slab */
> -#define for_each_object(__p, __s, __addr, __objects) \
> -       for (__p = fixup_red_left(__s, __addr); \
> -               __p < (__addr) + (__objects) * (__s)->size; \
> -               __p += (__s)->size)
> -
> -static inline unsigned int order_objects(unsigned int order, unsigned int size)
> -{
> -       return ((unsigned int)PAGE_SIZE << order) / size;
> -}
> -
> -static inline struct kmem_cache_order_objects oo_make(unsigned int order,
> -               unsigned int size)
> -{
> -       struct kmem_cache_order_objects x = {
> -               (order << OO_SHIFT) + order_objects(order, size)
> -       };
> -
> -       return x;
> -}
> -
> -static inline unsigned int oo_order(struct kmem_cache_order_objects x)
> -{
> -       return x.x >> OO_SHIFT;
> -}
> -
> -static inline unsigned int oo_objects(struct kmem_cache_order_objects x)
> -{
> -       return x.x & OO_MASK;
> -}
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -static void slub_set_cpu_partial(struct kmem_cache *s, unsigned int nr_objects)
> -{
> -       unsigned int nr_slabs;
> -
> -       s->cpu_partial = nr_objects;
> -
> -       /*
> -        * We take the number of objects but actually limit the number of
> -        * slabs on the per cpu partial list, in order to limit excessive
> -        * growth of the list. For simplicity we assume that the slabs will
> -        * be half-full.
> -        */
> -       nr_slabs = DIV_ROUND_UP(nr_objects * 2, oo_objects(s->oo));
> -       s->cpu_partial_slabs = nr_slabs;
> -}
> -#else
> -static inline void
> -slub_set_cpu_partial(struct kmem_cache *s, unsigned int nr_objects)
> -{
> -}
> -#endif /* CONFIG_SLUB_CPU_PARTIAL */
> -
> -/*
> - * Per slab locking using the pagelock
> - */
> -static __always_inline void slab_lock(struct slab *slab)
> -{
> -       struct page *page = slab_page(slab);
> -
> -       VM_BUG_ON_PAGE(PageTail(page), page);
> -       bit_spin_lock(PG_locked, &page->flags);
> -}
> -
> -static __always_inline void slab_unlock(struct slab *slab)
> -{
> -       struct page *page = slab_page(slab);
> -
> -       VM_BUG_ON_PAGE(PageTail(page), page);
> -       __bit_spin_unlock(PG_locked, &page->flags);
> -}
> -
> -/*
> - * Interrupts must be disabled (for the fallback code to work right), typically
> - * by an _irqsave() lock variant. On PREEMPT_RT the preempt_disable(), which is
> - * part of bit_spin_lock(), is sufficient because the policy is not to allow any
> - * allocation/ free operation in hardirq context. Therefore nothing can
> - * interrupt the operation.
> - */
> -static inline bool __cmpxchg_double_slab(struct kmem_cache *s, struct slab *slab,
> -               void *freelist_old, unsigned long counters_old,
> -               void *freelist_new, unsigned long counters_new,
> -               const char *n)
> -{
> -       if (USE_LOCKLESS_FAST_PATH())
> -               lockdep_assert_irqs_disabled();
> -#if defined(CONFIG_HAVE_CMPXCHG_DOUBLE) && \
> -    defined(CONFIG_HAVE_ALIGNED_STRUCT_PAGE)
> -       if (s->flags & __CMPXCHG_DOUBLE) {
> -               if (cmpxchg_double(&slab->freelist, &slab->counters,
> -                                  freelist_old, counters_old,
> -                                  freelist_new, counters_new))
> -                       return true;
> -       } else
> -#endif
> -       {
> -               slab_lock(slab);
> -               if (slab->freelist == freelist_old &&
> -                                       slab->counters == counters_old) {
> -                       slab->freelist = freelist_new;
> -                       slab->counters = counters_new;
> -                       slab_unlock(slab);
> -                       return true;
> -               }
> -               slab_unlock(slab);
> -       }
> -
> -       cpu_relax();
> -       stat(s, CMPXCHG_DOUBLE_FAIL);
> -
> -#ifdef SLUB_DEBUG_CMPXCHG
> -       pr_info("%s %s: cmpxchg double redo ", n, s->name);
> -#endif
> -
> -       return false;
> -}
> -
> -static inline bool cmpxchg_double_slab(struct kmem_cache *s, struct slab *slab,
> -               void *freelist_old, unsigned long counters_old,
> -               void *freelist_new, unsigned long counters_new,
> -               const char *n)
> -{
> -#if defined(CONFIG_HAVE_CMPXCHG_DOUBLE) && \
> -    defined(CONFIG_HAVE_ALIGNED_STRUCT_PAGE)
> -       if (s->flags & __CMPXCHG_DOUBLE) {
> -               if (cmpxchg_double(&slab->freelist, &slab->counters,
> -                                  freelist_old, counters_old,
> -                                  freelist_new, counters_new))
> -                       return true;
> -       } else
> -#endif
> -       {
> -               unsigned long flags;
> -
> -               local_irq_save(flags);
> -               slab_lock(slab);
> -               if (slab->freelist == freelist_old &&
> -                                       slab->counters == counters_old) {
> -                       slab->freelist = freelist_new;
> -                       slab->counters = counters_new;
> -                       slab_unlock(slab);
> -                       local_irq_restore(flags);
> -                       return true;
> -               }
> -               slab_unlock(slab);
> -               local_irq_restore(flags);
> -       }
> -
> -       cpu_relax();
> -       stat(s, CMPXCHG_DOUBLE_FAIL);
> -
> -#ifdef SLUB_DEBUG_CMPXCHG
> -       pr_info("%s %s: cmpxchg double redo ", n, s->name);
> -#endif
> -
> -       return false;
> -}
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -static unsigned long object_map[BITS_TO_LONGS(MAX_OBJS_PER_PAGE)];
> -static DEFINE_SPINLOCK(object_map_lock);
> -
> -static void __fill_map(unsigned long *obj_map, struct kmem_cache *s,
> -                      struct slab *slab)
> -{
> -       void *addr = slab_address(slab);
> -       void *p;
> -
> -       bitmap_zero(obj_map, slab->objects);
> -
> -       for (p = slab->freelist; p; p = get_freepointer(s, p))
> -               set_bit(__obj_to_index(s, addr, p), obj_map);
> -}
> -
> -#if IS_ENABLED(CONFIG_KUNIT)
> -static bool slab_add_kunit_errors(void)
> -{
> -       struct kunit_resource *resource;
> -
> -       if (!kunit_get_current_test())
> -               return false;
> -
> -       resource = kunit_find_named_resource(current->kunit_test, "slab_errors");
> -       if (!resource)
> -               return false;
> -
> -       (*(int *)resource->data)++;
> -       kunit_put_resource(resource);
> -       return true;
> -}
> -#else
> -static inline bool slab_add_kunit_errors(void) { return false; }
> -#endif
> -
> -static inline unsigned int size_from_object(struct kmem_cache *s)
> -{
> -       if (s->flags & SLAB_RED_ZONE)
> -               return s->size - s->red_left_pad;
> -
> -       return s->size;
> -}
> -
> -static inline void *restore_red_left(struct kmem_cache *s, void *p)
> -{
> -       if (s->flags & SLAB_RED_ZONE)
> -               p -= s->red_left_pad;
> -
> -       return p;
> -}
> -
> -/*
> - * Debug settings:
> - */
> -#if defined(CONFIG_SLUB_DEBUG_ON)
> -static slab_flags_t slub_debug = DEBUG_DEFAULT_FLAGS;
> -#else
> -static slab_flags_t slub_debug;
> -#endif
> -
> -static char *slub_debug_string;
> -static int disable_higher_order_debug;
> -
> -/*
> - * slub is about to manipulate internal object metadata.  This memory lies
> - * outside the range of the allocated object, so accessing it would normally
> - * be reported by kasan as a bounds error.  metadata_access_enable() is used
> - * to tell kasan that these accesses are OK.
> - */
> -static inline void metadata_access_enable(void)
> -{
> -       kasan_disable_current();
> -}
> -
> -static inline void metadata_access_disable(void)
> -{
> -       kasan_enable_current();
> -}
> -
> -/*
> - * Object debugging
> - */
> -
> -/* Verify that a pointer has an address that is valid within a slab page */
> -static inline int check_valid_pointer(struct kmem_cache *s,
> -                               struct slab *slab, void *object)
> -{
> -       void *base;
> -
> -       if (!object)
> -               return 1;
> -
> -       base = slab_address(slab);
> -       object = kasan_reset_tag(object);
> -       object = restore_red_left(s, object);
> -       if (object < base || object >= base + slab->objects * s->size ||
> -               (object - base) % s->size) {
> -               return 0;
> -       }
> -
> -       return 1;
> -}
> -
> -static void print_section(char *level, char *text, u8 *addr,
> -                         unsigned int length)
> -{
> -       metadata_access_enable();
> -       print_hex_dump(level, text, DUMP_PREFIX_ADDRESS,
> -                       16, 1, kasan_reset_tag((void *)addr), length, 1);
> -       metadata_access_disable();
> -}
> -
> -/*
> - * See comment in calculate_sizes().
> - */
> -static inline bool freeptr_outside_object(struct kmem_cache *s)
> -{
> -       return s->offset >= s->inuse;
> -}
> -
> -/*
> - * Return offset of the end of info block which is inuse + free pointer if
> - * not overlapping with object.
> - */
> -static inline unsigned int get_info_end(struct kmem_cache *s)
> -{
> -       if (freeptr_outside_object(s))
> -               return s->inuse + sizeof(void *);
> -       else
> -               return s->inuse;
> -}
> -
> -static struct track *get_track(struct kmem_cache *s, void *object,
> -       enum track_item alloc)
> -{
> -       struct track *p;
> -
> -       p = object + get_info_end(s);
> -
> -       return kasan_reset_tag(p + alloc);
> -}
> -
> -#ifdef CONFIG_STACKDEPOT
> -static noinline depot_stack_handle_t set_track_prepare(void)
> -{
> -       depot_stack_handle_t handle;
> -       unsigned long entries[TRACK_ADDRS_COUNT];
> -       unsigned int nr_entries;
> -
> -       nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 3);
> -       handle = stack_depot_save(entries, nr_entries, GFP_NOWAIT);
> -
> -       return handle;
> -}
> -#else
> -static inline depot_stack_handle_t set_track_prepare(void)
> -{
> -       return 0;
> -}
> -#endif
> -
> -static void set_track_update(struct kmem_cache *s, void *object,
> -                            enum track_item alloc, unsigned long addr,
> -                            depot_stack_handle_t handle)
> -{
> -       struct track *p = get_track(s, object, alloc);
> -
> -#ifdef CONFIG_STACKDEPOT
> -       p->handle = handle;
> -#endif
> -       p->addr = addr;
> -       p->cpu = smp_processor_id();
> -       p->pid = current->pid;
> -       p->when = jiffies;
> -}
> -
> -static __always_inline void set_track(struct kmem_cache *s, void *object,
> -                                     enum track_item alloc, unsigned long addr)
> -{
> -       depot_stack_handle_t handle = set_track_prepare();
> -
> -       set_track_update(s, object, alloc, addr, handle);
> -}
> -
> -static void init_tracking(struct kmem_cache *s, void *object)
> -{
> -       struct track *p;
> -
> -       if (!(s->flags & SLAB_STORE_USER))
> -               return;
> -
> -       p = get_track(s, object, TRACK_ALLOC);
> -       memset(p, 0, 2*sizeof(struct track));
> -}
> -
> -static void print_track(const char *s, struct track *t, unsigned long pr_time)
> -{
> -       depot_stack_handle_t handle __maybe_unused;
> -
> -       if (!t->addr)
> -               return;
> -
> -       pr_err("%s in %pS age=%lu cpu=%u pid=%d\n",
> -              s, (void *)t->addr, pr_time - t->when, t->cpu, t->pid);
> -#ifdef CONFIG_STACKDEPOT
> -       handle = READ_ONCE(t->handle);
> -       if (handle)
> -               stack_depot_print(handle);
> -       else
> -               pr_err("object allocation/free stack trace missing\n");
> -#endif
> -}
> -
> -void print_tracking(struct kmem_cache *s, void *object)
> -{
> -       unsigned long pr_time = jiffies;
> -       if (!(s->flags & SLAB_STORE_USER))
> -               return;
> -
> -       print_track("Allocated", get_track(s, object, TRACK_ALLOC), pr_time);
> -       print_track("Freed", get_track(s, object, TRACK_FREE), pr_time);
> -}
> -
> -static void print_slab_info(const struct slab *slab)
> -{
> -       struct folio *folio = (struct folio *)slab_folio(slab);
> -
> -       pr_err("Slab 0x%p objects=%u used=%u fp=0x%p flags=%pGp\n",
> -              slab, slab->objects, slab->inuse, slab->freelist,
> -              folio_flags(folio, 0));
> -}
> -
> -/*
> - * kmalloc caches has fixed sizes (mostly power of 2), and kmalloc() API
> - * family will round up the real request size to these fixed ones, so
> - * there could be an extra area than what is requested. Save the original
> - * request size in the meta data area, for better debug and sanity check.
> - */
> -static inline void set_orig_size(struct kmem_cache *s,
> -                               void *object, unsigned int orig_size)
> -{
> -       void *p = kasan_reset_tag(object);
> -
> -       if (!slub_debug_orig_size(s))
> -               return;
> -
> -#ifdef CONFIG_KASAN_GENERIC
> -       /*
> -        * KASAN could save its free meta data in object's data area at
> -        * offset 0, if the size is larger than 'orig_size', it will
> -        * overlap the data redzone in [orig_size+1, object_size], and
> -        * the check should be skipped.
> -        */
> -       if (kasan_metadata_size(s, true) > orig_size)
> -               orig_size = s->object_size;
> -#endif
> -
> -       p += get_info_end(s);
> -       p += sizeof(struct track) * 2;
> -
> -       *(unsigned int *)p = orig_size;
> -}
> -
> -static inline unsigned int get_orig_size(struct kmem_cache *s, void *object)
> -{
> -       void *p = kasan_reset_tag(object);
> -
> -       if (!slub_debug_orig_size(s))
> -               return s->object_size;
> -
> -       p += get_info_end(s);
> -       p += sizeof(struct track) * 2;
> -
> -       return *(unsigned int *)p;
> -}
> -
> -void skip_orig_size_check(struct kmem_cache *s, const void *object)
> -{
> -       set_orig_size(s, (void *)object, s->object_size);
> -}
> -
> -static void slab_bug(struct kmem_cache *s, char *fmt, ...)
> -{
> -       struct va_format vaf;
> -       va_list args;
> -
> -       va_start(args, fmt);
> -       vaf.fmt = fmt;
> -       vaf.va = &args;
> -       pr_err("=============================================================================\n");
> -       pr_err("BUG %s (%s): %pV\n", s->name, print_tainted(), &vaf);
> -       pr_err("-----------------------------------------------------------------------------\n\n");
> -       va_end(args);
> -}
> -
> -__printf(2, 3)
> -static void slab_fix(struct kmem_cache *s, char *fmt, ...)
> -{
> -       struct va_format vaf;
> -       va_list args;
> -
> -       if (slab_add_kunit_errors())
> -               return;
> -
> -       va_start(args, fmt);
> -       vaf.fmt = fmt;
> -       vaf.va = &args;
> -       pr_err("FIX %s: %pV\n", s->name, &vaf);
> -       va_end(args);
> -}
> -
> -static void print_trailer(struct kmem_cache *s, struct slab *slab, u8 *p)
> -{
> -       unsigned int off;       /* Offset of last byte */
> -       u8 *addr = slab_address(slab);
> -
> -       print_tracking(s, p);
> -
> -       print_slab_info(slab);
> -
> -       pr_err("Object 0x%p @offset=%tu fp=0x%p\n\n",
> -              p, p - addr, get_freepointer(s, p));
> -
> -       if (s->flags & SLAB_RED_ZONE)
> -               print_section(KERN_ERR, "Redzone  ", p - s->red_left_pad,
> -                             s->red_left_pad);
> -       else if (p > addr + 16)
> -               print_section(KERN_ERR, "Bytes b4 ", p - 16, 16);
> -
> -       print_section(KERN_ERR,         "Object   ", p,
> -                     min_t(unsigned int, s->object_size, PAGE_SIZE));
> -       if (s->flags & SLAB_RED_ZONE)
> -               print_section(KERN_ERR, "Redzone  ", p + s->object_size,
> -                       s->inuse - s->object_size);
> -
> -       off = get_info_end(s);
> -
> -       if (s->flags & SLAB_STORE_USER)
> -               off += 2 * sizeof(struct track);
> -
> -       if (slub_debug_orig_size(s))
> -               off += sizeof(unsigned int);
> -
> -       off += kasan_metadata_size(s, false);
> -
> -       if (off != size_from_object(s))
> -               /* Beginning of the filler is the free pointer */
> -               print_section(KERN_ERR, "Padding  ", p + off,
> -                             size_from_object(s) - off);
> -
> -       dump_stack();
> -}
> -
> -static void object_err(struct kmem_cache *s, struct slab *slab,
> -                       u8 *object, char *reason)
> -{
> -       if (slab_add_kunit_errors())
> -               return;
> -
> -       slab_bug(s, "%s", reason);
> -       print_trailer(s, slab, object);
> -       add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
> -}
> -
> -static bool freelist_corrupted(struct kmem_cache *s, struct slab *slab,
> -                              void **freelist, void *nextfree)
> -{
> -       if ((s->flags & SLAB_CONSISTENCY_CHECKS) &&
> -           !check_valid_pointer(s, slab, nextfree) && freelist) {
> -               object_err(s, slab, *freelist, "Freechain corrupt");
> -               *freelist = NULL;
> -               slab_fix(s, "Isolate corrupted freechain");
> -               return true;
> -       }
> -
> -       return false;
> -}
> -
> -static __printf(3, 4) void slab_err(struct kmem_cache *s, struct slab *slab,
> -                       const char *fmt, ...)
> -{
> -       va_list args;
> -       char buf[100];
> -
> -       if (slab_add_kunit_errors())
> -               return;
> -
> -       va_start(args, fmt);
> -       vsnprintf(buf, sizeof(buf), fmt, args);
> -       va_end(args);
> -       slab_bug(s, "%s", buf);
> -       print_slab_info(slab);
> -       dump_stack();
> -       add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
> -}
> -
> -static void init_object(struct kmem_cache *s, void *object, u8 val)
> -{
> -       u8 *p = kasan_reset_tag(object);
> -       unsigned int poison_size = s->object_size;
> -
> -       if (s->flags & SLAB_RED_ZONE) {
> -               memset(p - s->red_left_pad, val, s->red_left_pad);
> -
> -               if (slub_debug_orig_size(s) && val == SLUB_RED_ACTIVE) {
> -                       /*
> -                        * Redzone the extra allocated space by kmalloc than
> -                        * requested, and the poison size will be limited to
> -                        * the original request size accordingly.
> -                        */
> -                       poison_size = get_orig_size(s, object);
> -               }
> -       }
> -
> -       if (s->flags & __OBJECT_POISON) {
> -               memset(p, POISON_FREE, poison_size - 1);
> -               p[poison_size - 1] = POISON_END;
> -       }
> -
> -       if (s->flags & SLAB_RED_ZONE)
> -               memset(p + poison_size, val, s->inuse - poison_size);
> -}
> -
> -static void restore_bytes(struct kmem_cache *s, char *message, u8 data,
> -                                               void *from, void *to)
> -{
> -       slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x", message, from, to - 1, data);
> -       memset(from, data, to - from);
> -}
> -
> -static int check_bytes_and_report(struct kmem_cache *s, struct slab *slab,
> -                       u8 *object, char *what,
> -                       u8 *start, unsigned int value, unsigned int bytes)
> -{
> -       u8 *fault;
> -       u8 *end;
> -       u8 *addr = slab_address(slab);
> -
> -       metadata_access_enable();
> -       fault = memchr_inv(kasan_reset_tag(start), value, bytes);
> -       metadata_access_disable();
> -       if (!fault)
> -               return 1;
> -
> -       end = start + bytes;
> -       while (end > fault && end[-1] == value)
> -               end--;
> -
> -       if (slab_add_kunit_errors())
> -               goto skip_bug_print;
> -
> -       slab_bug(s, "%s overwritten", what);
> -       pr_err("0x%p-0x%p @offset=%tu. First byte 0x%x instead of 0x%x\n",
> -                                       fault, end - 1, fault - addr,
> -                                       fault[0], value);
> -       print_trailer(s, slab, object);
> -       add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
> -
> -skip_bug_print:
> -       restore_bytes(s, what, value, fault, end);
> -       return 0;
> -}
> -
> -/*
> - * Object layout:
> - *
> - * object address
> - *     Bytes of the object to be managed.
> - *     If the freepointer may overlay the object then the free
> - *     pointer is at the middle of the object.
> - *
> - *     Poisoning uses 0x6b (POISON_FREE) and the last byte is
> - *     0xa5 (POISON_END)
> - *
> - * object + s->object_size
> - *     Padding to reach word boundary. This is also used for Redzoning.
> - *     Padding is extended by another word if Redzoning is enabled and
> - *     object_size == inuse.
> - *
> - *     We fill with 0xbb (RED_INACTIVE) for inactive objects and with
> - *     0xcc (RED_ACTIVE) for objects in use.
> - *
> - * object + s->inuse
> - *     Meta data starts here.
> - *
> - *     A. Free pointer (if we cannot overwrite object on free)
> - *     B. Tracking data for SLAB_STORE_USER
> - *     C. Original request size for kmalloc object (SLAB_STORE_USER enabled)
> - *     D. Padding to reach required alignment boundary or at minimum
> - *             one word if debugging is on to be able to detect writes
> - *             before the word boundary.
> - *
> - *     Padding is done using 0x5a (POISON_INUSE)
> - *
> - * object + s->size
> - *     Nothing is used beyond s->size.
> - *
> - * If slabcaches are merged then the object_size and inuse boundaries are mostly
> - * ignored. And therefore no slab options that rely on these boundaries
> - * may be used with merged slabcaches.
> - */
> -
> -static int check_pad_bytes(struct kmem_cache *s, struct slab *slab, u8 *p)
> -{
> -       unsigned long off = get_info_end(s);    /* The end of info */
> -
> -       if (s->flags & SLAB_STORE_USER) {
> -               /* We also have user information there */
> -               off += 2 * sizeof(struct track);
> -
> -               if (s->flags & SLAB_KMALLOC)
> -                       off += sizeof(unsigned int);
> -       }
> -
> -       off += kasan_metadata_size(s, false);
> -
> -       if (size_from_object(s) == off)
> -               return 1;
> -
> -       return check_bytes_and_report(s, slab, p, "Object padding",
> -                       p + off, POISON_INUSE, size_from_object(s) - off);
> -}
> -
> -/* Check the pad bytes at the end of a slab page */
> -static void slab_pad_check(struct kmem_cache *s, struct slab *slab)
> -{
> -       u8 *start;
> -       u8 *fault;
> -       u8 *end;
> -       u8 *pad;
> -       int length;
> -       int remainder;
> -
> -       if (!(s->flags & SLAB_POISON))
> -               return;
> -
> -       start = slab_address(slab);
> -       length = slab_size(slab);
> -       end = start + length;
> -       remainder = length % s->size;
> -       if (!remainder)
> -               return;
> -
> -       pad = end - remainder;
> -       metadata_access_enable();
> -       fault = memchr_inv(kasan_reset_tag(pad), POISON_INUSE, remainder);
> -       metadata_access_disable();
> -       if (!fault)
> -               return;
> -       while (end > fault && end[-1] == POISON_INUSE)
> -               end--;
> -
> -       slab_err(s, slab, "Padding overwritten. 0x%p-0x%p @offset=%tu",
> -                       fault, end - 1, fault - start);
> -       print_section(KERN_ERR, "Padding ", pad, remainder);
> -
> -       restore_bytes(s, "slab padding", POISON_INUSE, fault, end);
> -}
> -
> -static int check_object(struct kmem_cache *s, struct slab *slab,
> -                                       void *object, u8 val)
> -{
> -       u8 *p = object;
> -       u8 *endobject = object + s->object_size;
> -       unsigned int orig_size;
> -
> -       if (s->flags & SLAB_RED_ZONE) {
> -               if (!check_bytes_and_report(s, slab, object, "Left Redzone",
> -                       object - s->red_left_pad, val, s->red_left_pad))
> -                       return 0;
> -
> -               if (!check_bytes_and_report(s, slab, object, "Right Redzone",
> -                       endobject, val, s->inuse - s->object_size))
> -                       return 0;
> -
> -               if (slub_debug_orig_size(s) && val == SLUB_RED_ACTIVE) {
> -                       orig_size = get_orig_size(s, object);
> -
> -                       if (s->object_size > orig_size  &&
> -                               !check_bytes_and_report(s, slab, object,
> -                                       "kmalloc Redzone", p + orig_size,
> -                                       val, s->object_size - orig_size)) {
> -                               return 0;
> -                       }
> -               }
> -       } else {
> -               if ((s->flags & SLAB_POISON) && s->object_size < s->inuse) {
> -                       check_bytes_and_report(s, slab, p, "Alignment padding",
> -                               endobject, POISON_INUSE,
> -                               s->inuse - s->object_size);
> -               }
> -       }
> -
> -       if (s->flags & SLAB_POISON) {
> -               if (val != SLUB_RED_ACTIVE && (s->flags & __OBJECT_POISON) &&
> -                       (!check_bytes_and_report(s, slab, p, "Poison", p,
> -                                       POISON_FREE, s->object_size - 1) ||
> -                        !check_bytes_and_report(s, slab, p, "End Poison",
> -                               p + s->object_size - 1, POISON_END, 1)))
> -                       return 0;
> -               /*
> -                * check_pad_bytes cleans up on its own.
> -                */
> -               check_pad_bytes(s, slab, p);
> -       }
> -
> -       if (!freeptr_outside_object(s) && val == SLUB_RED_ACTIVE)
> -               /*
> -                * Object and freepointer overlap. Cannot check
> -                * freepointer while object is allocated.
> -                */
> -               return 1;
> -
> -       /* Check free pointer validity */
> -       if (!check_valid_pointer(s, slab, get_freepointer(s, p))) {
> -               object_err(s, slab, p, "Freepointer corrupt");
> -               /*
> -                * No choice but to zap it and thus lose the remainder
> -                * of the free objects in this slab. May cause
> -                * another error because the object count is now wrong.
> -                */
> -               set_freepointer(s, p, NULL);
> -               return 0;
> -       }
> -       return 1;
> -}
> -
> -static int check_slab(struct kmem_cache *s, struct slab *slab)
> -{
> -       int maxobj;
> -
> -       if (!folio_test_slab(slab_folio(slab))) {
> -               slab_err(s, slab, "Not a valid slab page");
> -               return 0;
> -       }
> -
> -       maxobj = order_objects(slab_order(slab), s->size);
> -       if (slab->objects > maxobj) {
> -               slab_err(s, slab, "objects %u > max %u",
> -                       slab->objects, maxobj);
> -               return 0;
> -       }
> -       if (slab->inuse > slab->objects) {
> -               slab_err(s, slab, "inuse %u > max %u",
> -                       slab->inuse, slab->objects);
> -               return 0;
> -       }
> -       /* Slab_pad_check fixes things up after itself */
> -       slab_pad_check(s, slab);
> -       return 1;
> -}
> -
> -/*
> - * Determine if a certain object in a slab is on the freelist. Must hold the
> - * slab lock to guarantee that the chains are in a consistent state.
> - */
> -static int on_freelist(struct kmem_cache *s, struct slab *slab, void *search)
> -{
> -       int nr = 0;
> -       void *fp;
> -       void *object = NULL;
> -       int max_objects;
> -
> -       fp = slab->freelist;
> -       while (fp && nr <= slab->objects) {
> -               if (fp == search)
> -                       return 1;
> -               if (!check_valid_pointer(s, slab, fp)) {
> -                       if (object) {
> -                               object_err(s, slab, object,
> -                                       "Freechain corrupt");
> -                               set_freepointer(s, object, NULL);
> -                       } else {
> -                               slab_err(s, slab, "Freepointer corrupt");
> -                               slab->freelist = NULL;
> -                               slab->inuse = slab->objects;
> -                               slab_fix(s, "Freelist cleared");
> -                               return 0;
> -                       }
> -                       break;
> -               }
> -               object = fp;
> -               fp = get_freepointer(s, object);
> -               nr++;
> -       }
> -
> -       max_objects = order_objects(slab_order(slab), s->size);
> -       if (max_objects > MAX_OBJS_PER_PAGE)
> -               max_objects = MAX_OBJS_PER_PAGE;
> -
> -       if (slab->objects != max_objects) {
> -               slab_err(s, slab, "Wrong number of objects. Found %d but should be %d",
> -                        slab->objects, max_objects);
> -               slab->objects = max_objects;
> -               slab_fix(s, "Number of objects adjusted");
> -       }
> -       if (slab->inuse != slab->objects - nr) {
> -               slab_err(s, slab, "Wrong object count. Counter is %d but counted were %d",
> -                        slab->inuse, slab->objects - nr);
> -               slab->inuse = slab->objects - nr;
> -               slab_fix(s, "Object count adjusted");
> -       }
> -       return search == NULL;
> -}
> -
> -static void trace(struct kmem_cache *s, struct slab *slab, void *object,
> -                                                               int alloc)
> -{
> -       if (s->flags & SLAB_TRACE) {
> -               pr_info("TRACE %s %s 0x%p inuse=%d fp=0x%p\n",
> -                       s->name,
> -                       alloc ? "alloc" : "free",
> -                       object, slab->inuse,
> -                       slab->freelist);
> -
> -               if (!alloc)
> -                       print_section(KERN_INFO, "Object ", (void *)object,
> -                                       s->object_size);
> -
> -               dump_stack();
> -       }
> -}
> -
> -/*
> - * Tracking of fully allocated slabs for debugging purposes.
> - */
> -static void add_full(struct kmem_cache *s,
> -       struct kmem_cache_node *n, struct slab *slab)
> -{
> -       if (!(s->flags & SLAB_STORE_USER))
> -               return;
> -
> -       lockdep_assert_held(&n->list_lock);
> -       list_add(&slab->slab_list, &n->full);
> -}
> -
> -static void remove_full(struct kmem_cache *s, struct kmem_cache_node *n, struct slab *slab)
> -{
> -       if (!(s->flags & SLAB_STORE_USER))
> -               return;
> -
> -       lockdep_assert_held(&n->list_lock);
> -       list_del(&slab->slab_list);
> -}
> -
> -/* Tracking of the number of slabs for debugging purposes */
> -static inline unsigned long slabs_node(struct kmem_cache *s, int node)
> -{
> -       struct kmem_cache_node *n = get_node(s, node);
> -
> -       return atomic_long_read(&n->nr_slabs);
> -}
> -
> -static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
> -{
> -       return atomic_long_read(&n->nr_slabs);
> -}
> -
> -static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects)
> -{
> -       struct kmem_cache_node *n = get_node(s, node);
> -
> -       /*
> -        * May be called early in order to allocate a slab for the
> -        * kmem_cache_node structure. Solve the chicken-egg
> -        * dilemma by deferring the increment of the count during
> -        * bootstrap (see early_kmem_cache_node_alloc).
> -        */
> -       if (likely(n)) {
> -               atomic_long_inc(&n->nr_slabs);
> -               atomic_long_add(objects, &n->total_objects);
> -       }
> -}
> -static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects)
> -{
> -       struct kmem_cache_node *n = get_node(s, node);
> -
> -       atomic_long_dec(&n->nr_slabs);
> -       atomic_long_sub(objects, &n->total_objects);
> -}
> -
> -/* Object debug checks for alloc/free paths */
> -static void setup_object_debug(struct kmem_cache *s, void *object)
> -{
> -       if (!kmem_cache_debug_flags(s, SLAB_STORE_USER|SLAB_RED_ZONE|__OBJECT_POISON))
> -               return;
> -
> -       init_object(s, object, SLUB_RED_INACTIVE);
> -       init_tracking(s, object);
> -}
> -
> -static
> -void setup_slab_debug(struct kmem_cache *s, struct slab *slab, void *addr)
> -{
> -       if (!kmem_cache_debug_flags(s, SLAB_POISON))
> -               return;
> -
> -       metadata_access_enable();
> -       memset(kasan_reset_tag(addr), POISON_INUSE, slab_size(slab));
> -       metadata_access_disable();
> -}
> -
> -static inline int alloc_consistency_checks(struct kmem_cache *s,
> -                                       struct slab *slab, void *object)
> -{
> -       if (!check_slab(s, slab))
> -               return 0;
> -
> -       if (!check_valid_pointer(s, slab, object)) {
> -               object_err(s, slab, object, "Freelist Pointer check fails");
> -               return 0;
> -       }
> -
> -       if (!check_object(s, slab, object, SLUB_RED_INACTIVE))
> -               return 0;
> -
> -       return 1;
> -}
> -
> -static noinline bool alloc_debug_processing(struct kmem_cache *s,
> -                       struct slab *slab, void *object, int orig_size)
> -{
> -       if (s->flags & SLAB_CONSISTENCY_CHECKS) {
> -               if (!alloc_consistency_checks(s, slab, object))
> -                       goto bad;
> -       }
> -
> -       /* Success. Perform special debug activities for allocs */
> -       trace(s, slab, object, 1);
> -       set_orig_size(s, object, orig_size);
> -       init_object(s, object, SLUB_RED_ACTIVE);
> -       return true;
> -
> -bad:
> -       if (folio_test_slab(slab_folio(slab))) {
> -               /*
> -                * If this is a slab page then lets do the best we can
> -                * to avoid issues in the future. Marking all objects
> -                * as used avoids touching the remaining objects.
> -                */
> -               slab_fix(s, "Marking all objects used");
> -               slab->inuse = slab->objects;
> -               slab->freelist = NULL;
> -       }
> -       return false;
> -}
> -
> -static inline int free_consistency_checks(struct kmem_cache *s,
> -               struct slab *slab, void *object, unsigned long addr)
> -{
> -       if (!check_valid_pointer(s, slab, object)) {
> -               slab_err(s, slab, "Invalid object pointer 0x%p", object);
> -               return 0;
> -       }
> -
> -       if (on_freelist(s, slab, object)) {
> -               object_err(s, slab, object, "Object already free");
> -               return 0;
> -       }
> -
> -       if (!check_object(s, slab, object, SLUB_RED_ACTIVE))
> -               return 0;
> -
> -       if (unlikely(s != slab->slab_cache)) {
> -               if (!folio_test_slab(slab_folio(slab))) {
> -                       slab_err(s, slab, "Attempt to free object(0x%p) outside of slab",
> -                                object);
> -               } else if (!slab->slab_cache) {
> -                       pr_err("SLUB <none>: no slab for object 0x%p.\n",
> -                              object);
> -                       dump_stack();
> -               } else
> -                       object_err(s, slab, object,
> -                                       "page slab pointer corrupt.");
> -               return 0;
> -       }
> -       return 1;
> -}
> -
> -/*
> - * Parse a block of slub_debug options. Blocks are delimited by ';'
> - *
> - * @str:    start of block
> - * @flags:  returns parsed flags, or DEBUG_DEFAULT_FLAGS if none specified
> - * @slabs:  return start of list of slabs, or NULL when there's no list
> - * @init:   assume this is initial parsing and not per-kmem-create parsing
> - *
> - * returns the start of next block if there's any, or NULL
> - */
> -static char *
> -parse_slub_debug_flags(char *str, slab_flags_t *flags, char **slabs, bool init)
> -{
> -       bool higher_order_disable = false;
> -
> -       /* Skip any completely empty blocks */
> -       while (*str && *str == ';')
> -               str++;
> -
> -       if (*str == ',') {
> -               /*
> -                * No options but restriction on slabs. This means full
> -                * debugging for slabs matching a pattern.
> -                */
> -               *flags = DEBUG_DEFAULT_FLAGS;
> -               goto check_slabs;
> -       }
> -       *flags = 0;
> -
> -       /* Determine which debug features should be switched on */
> -       for (; *str && *str != ',' && *str != ';'; str++) {
> -               switch (tolower(*str)) {
> -               case '-':
> -                       *flags = 0;
> -                       break;
> -               case 'f':
> -                       *flags |= SLAB_CONSISTENCY_CHECKS;
> -                       break;
> -               case 'z':
> -                       *flags |= SLAB_RED_ZONE;
> -                       break;
> -               case 'p':
> -                       *flags |= SLAB_POISON;
> -                       break;
> -               case 'u':
> -                       *flags |= SLAB_STORE_USER;
> -                       break;
> -               case 't':
> -                       *flags |= SLAB_TRACE;
> -                       break;
> -               case 'a':
> -                       *flags |= SLAB_FAILSLAB;
> -                       break;
> -               case 'o':
> -                       /*
> -                        * Avoid enabling debugging on caches if its minimum
> -                        * order would increase as a result.
> -                        */
> -                       higher_order_disable = true;
> -                       break;
> -               default:
> -                       if (init)
> -                               pr_err("slub_debug option '%c' unknown. skipped\n", *str);
> -               }
> -       }
> -check_slabs:
> -       if (*str == ',')
> -               *slabs = ++str;
> -       else
> -               *slabs = NULL;
> -
> -       /* Skip over the slab list */
> -       while (*str && *str != ';')
> -               str++;
> -
> -       /* Skip any completely empty blocks */
> -       while (*str && *str == ';')
> -               str++;
> -
> -       if (init && higher_order_disable)
> -               disable_higher_order_debug = 1;
> -
> -       if (*str)
> -               return str;
> -       else
> -               return NULL;
> -}
> -
> -static int __init setup_slub_debug(char *str)
> -{
> -       slab_flags_t flags;
> -       slab_flags_t global_flags;
> -       char *saved_str;
> -       char *slab_list;
> -       bool global_slub_debug_changed = false;
> -       bool slab_list_specified = false;
> -
> -       global_flags = DEBUG_DEFAULT_FLAGS;
> -       if (*str++ != '=' || !*str)
> -               /*
> -                * No options specified. Switch on full debugging.
> -                */
> -               goto out;
> -
> -       saved_str = str;
> -       while (str) {
> -               str = parse_slub_debug_flags(str, &flags, &slab_list, true);
> -
> -               if (!slab_list) {
> -                       global_flags = flags;
> -                       global_slub_debug_changed = true;
> -               } else {
> -                       slab_list_specified = true;
> -                       if (flags & SLAB_STORE_USER)
> -                               stack_depot_request_early_init();
> -               }
> -       }
> -
> -       /*
> -        * For backwards compatibility, a single list of flags with list of
> -        * slabs means debugging is only changed for those slabs, so the global
> -        * slub_debug should be unchanged (0 or DEBUG_DEFAULT_FLAGS, depending
> -        * on CONFIG_SLUB_DEBUG_ON). We can extended that to multiple lists as
> -        * long as there is no option specifying flags without a slab list.
> -        */
> -       if (slab_list_specified) {
> -               if (!global_slub_debug_changed)
> -                       global_flags = slub_debug;
> -               slub_debug_string = saved_str;
> -       }
> -out:
> -       slub_debug = global_flags;
> -       if (slub_debug & SLAB_STORE_USER)
> -               stack_depot_request_early_init();
> -       if (slub_debug != 0 || slub_debug_string)
> -               static_branch_enable(&slub_debug_enabled);
> -       else
> -               static_branch_disable(&slub_debug_enabled);
> -       if ((static_branch_unlikely(&init_on_alloc) ||
> -            static_branch_unlikely(&init_on_free)) &&
> -           (slub_debug & SLAB_POISON))
> -               pr_info("mem auto-init: SLAB_POISON will take precedence over init_on_alloc/init_on_free\n");
> -       return 1;
> -}
> -
> -__setup("slub_debug", setup_slub_debug);
> -
> -/*
> - * kmem_cache_flags - apply debugging options to the cache
> - * @object_size:       the size of an object without meta data
> - * @flags:             flags to set
> - * @name:              name of the cache
> - *
> - * Debug option(s) are applied to @flags. In addition to the debug
> - * option(s), if a slab name (or multiple) is specified i.e.
> - * slub_debug=<Debug-Options>,<slab name1>,<slab name2> ...
> - * then only the select slabs will receive the debug option(s).
> - */
> -slab_flags_t kmem_cache_flags(unsigned int object_size,
> -       slab_flags_t flags, const char *name)
> -{
> -       char *iter;
> -       size_t len;
> -       char *next_block;
> -       slab_flags_t block_flags;
> -       slab_flags_t slub_debug_local = slub_debug;
> -
> -       if (flags & SLAB_NO_USER_FLAGS)
> -               return flags;
> -
> -       /*
> -        * If the slab cache is for debugging (e.g. kmemleak) then
> -        * don't store user (stack trace) information by default,
> -        * but let the user enable it via the command line below.
> -        */
> -       if (flags & SLAB_NOLEAKTRACE)
> -               slub_debug_local &= ~SLAB_STORE_USER;
> -
> -       len = strlen(name);
> -       next_block = slub_debug_string;
> -       /* Go through all blocks of debug options, see if any matches our slab's name */
> -       while (next_block) {
> -               next_block = parse_slub_debug_flags(next_block, &block_flags, &iter, false);
> -               if (!iter)
> -                       continue;
> -               /* Found a block that has a slab list, search it */
> -               while (*iter) {
> -                       char *end, *glob;
> -                       size_t cmplen;
> -
> -                       end = strchrnul(iter, ',');
> -                       if (next_block && next_block < end)
> -                               end = next_block - 1;
> -
> -                       glob = strnchr(iter, end - iter, '*');
> -                       if (glob)
> -                               cmplen = glob - iter;
> -                       else
> -                               cmplen = max_t(size_t, len, (end - iter));
> -
> -                       if (!strncmp(name, iter, cmplen)) {
> -                               flags |= block_flags;
> -                               return flags;
> -                       }
> -
> -                       if (!*end || *end == ';')
> -                               break;
> -                       iter = end + 1;
> -               }
> -       }
> -
> -       return flags | slub_debug_local;
> -}
> -#else /* !CONFIG_SLUB_DEBUG */
> -static inline void setup_object_debug(struct kmem_cache *s, void *object) {}
> -static inline
> -void setup_slab_debug(struct kmem_cache *s, struct slab *slab, void *addr) {}
> -
> -static inline bool alloc_debug_processing(struct kmem_cache *s,
> -       struct slab *slab, void *object, int orig_size) { return true; }
> -
> -static inline bool free_debug_processing(struct kmem_cache *s,
> -       struct slab *slab, void *head, void *tail, int *bulk_cnt,
> -       unsigned long addr, depot_stack_handle_t handle) { return true; }
> -
> -static inline void slab_pad_check(struct kmem_cache *s, struct slab *slab) {}
> -static inline int check_object(struct kmem_cache *s, struct slab *slab,
> -                       void *object, u8 val) { return 1; }
> -static inline depot_stack_handle_t set_track_prepare(void) { return 0; }
> -static inline void set_track(struct kmem_cache *s, void *object,
> -                            enum track_item alloc, unsigned long addr) {}
> -static inline void add_full(struct kmem_cache *s, struct kmem_cache_node *n,
> -                                       struct slab *slab) {}
> -static inline void remove_full(struct kmem_cache *s, struct kmem_cache_node *n,
> -                                       struct slab *slab) {}
> -slab_flags_t kmem_cache_flags(unsigned int object_size,
> -       slab_flags_t flags, const char *name)
> -{
> -       return flags;
> -}
> -#define slub_debug 0
> -
> -#define disable_higher_order_debug 0
> -
> -static inline unsigned long slabs_node(struct kmem_cache *s, int node)
> -                                                       { return 0; }
> -static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
> -                                                       { return 0; }
> -static inline void inc_slabs_node(struct kmem_cache *s, int node,
> -                                                       int objects) {}
> -static inline void dec_slabs_node(struct kmem_cache *s, int node,
> -                                                       int objects) {}
> -
> -#ifndef CONFIG_SLUB_TINY
> -static bool freelist_corrupted(struct kmem_cache *s, struct slab *slab,
> -                              void **freelist, void *nextfree)
> -{
> -       return false;
> -}
> -#endif
> -#endif /* CONFIG_SLUB_DEBUG */
> -
> -/*
> - * Hooks for other subsystems that check memory allocations. In a typical
> - * production configuration these hooks all should produce no code at all.
> - */
> -static __always_inline bool slab_free_hook(struct kmem_cache *s,
> -                                               void *x, bool init)
> -{
> -       kmemleak_free_recursive(x, s->flags);
> -       kmsan_slab_free(s, x);
> -
> -       debug_check_no_locks_freed(x, s->object_size);
> -
> -       if (!(s->flags & SLAB_DEBUG_OBJECTS))
> -               debug_check_no_obj_freed(x, s->object_size);
> -
> -       /* Use KCSAN to help debug racy use-after-free. */
> -       if (!(s->flags & SLAB_TYPESAFE_BY_RCU))
> -               __kcsan_check_access(x, s->object_size,
> -                                    KCSAN_ACCESS_WRITE | KCSAN_ACCESS_ASSERT);
> -
> -       /*
> -        * As memory initialization might be integrated into KASAN,
> -        * kasan_slab_free and initialization memset's must be
> -        * kept together to avoid discrepancies in behavior.
> -        *
> -        * The initialization memset's clear the object and the metadata,
> -        * but don't touch the SLAB redzone.
> -        */
> -       if (init) {
> -               int rsize;
> -
> -               if (!kasan_has_integrated_init())
> -                       memset(kasan_reset_tag(x), 0, s->object_size);
> -               rsize = (s->flags & SLAB_RED_ZONE) ? s->red_left_pad : 0;
> -               memset((char *)kasan_reset_tag(x) + s->inuse, 0,
> -                      s->size - s->inuse - rsize);
> -       }
> -       /* KASAN might put x into memory quarantine, delaying its reuse. */
> -       return kasan_slab_free(s, x, init);
> -}
> -
> -static inline bool slab_free_freelist_hook(struct kmem_cache *s,
> -                                          void **head, void **tail,
> -                                          int *cnt)
> -{
> -
> -       void *object;
> -       void *next = *head;
> -       void *old_tail = *tail ? *tail : *head;
> -
> -       if (is_kfence_address(next)) {
> -               slab_free_hook(s, next, false);
> -               return true;
> -       }
> -
> -       /* Head and tail of the reconstructed freelist */
> -       *head = NULL;
> -       *tail = NULL;
> -
> -       do {
> -               object = next;
> -               next = get_freepointer(s, object);
> -
> -               /* If object's reuse doesn't have to be delayed */
> -               if (!slab_free_hook(s, object, slab_want_init_on_free(s))) {
> -                       /* Move object to the new freelist */
> -                       set_freepointer(s, object, *head);
> -                       *head = object;
> -                       if (!*tail)
> -                               *tail = object;
> -               } else {
> -                       /*
> -                        * Adjust the reconstructed freelist depth
> -                        * accordingly if object's reuse is delayed.
> -                        */
> -                       --(*cnt);
> -               }
> -       } while (object != old_tail);
> -
> -       if (*head == *tail)
> -               *tail = NULL;
> -
> -       return *head != NULL;
> -}
> -
> -static void *setup_object(struct kmem_cache *s, void *object)
> -{
> -       setup_object_debug(s, object);
> -       object = kasan_init_slab_obj(s, object);
> -       if (unlikely(s->ctor)) {
> -               kasan_unpoison_object_data(s, object);
> -               s->ctor(object);
> -               kasan_poison_object_data(s, object);
> -       }
> -       return object;
> -}
> -
> -/*
> - * Slab allocation and freeing
> - */
> -static inline struct slab *alloc_slab_page(gfp_t flags, int node,
> -               struct kmem_cache_order_objects oo)
> -{
> -       struct folio *folio;
> -       struct slab *slab;
> -       unsigned int order = oo_order(oo);
> -
> -       if (node == NUMA_NO_NODE)
> -               folio = (struct folio *)alloc_pages(flags, order);
> -       else
> -               folio = (struct folio *)__alloc_pages_node(node, flags, order);
> -
> -       if (!folio)
> -               return NULL;
> -
> -       slab = folio_slab(folio);
> -       __folio_set_slab(folio);
> -       /* Make the flag visible before any changes to folio->mapping */
> -       smp_wmb();
> -       if (folio_is_pfmemalloc(folio))
> -               slab_set_pfmemalloc(slab);
> -
> -       return slab;
> -}
> -
> -#ifdef CONFIG_SLAB_FREELIST_RANDOM
> -/* Pre-initialize the random sequence cache */
> -static int init_cache_random_seq(struct kmem_cache *s)
> -{
> -       unsigned int count = oo_objects(s->oo);
> -       int err;
> -
> -       /* Bailout if already initialised */
> -       if (s->random_seq)
> -               return 0;
> -
> -       err = cache_random_seq_create(s, count, GFP_KERNEL);
> -       if (err) {
> -               pr_err("SLUB: Unable to initialize free list for %s\n",
> -                       s->name);
> -               return err;
> -       }
> -
> -       /* Transform to an offset on the set of pages */
> -       if (s->random_seq) {
> -               unsigned int i;
> -
> -               for (i = 0; i < count; i++)
> -                       s->random_seq[i] *= s->size;
> -       }
> -       return 0;
> -}
> -
> -/* Initialize each random sequence freelist per cache */
> -static void __init init_freelist_randomization(void)
> -{
> -       struct kmem_cache *s;
> -
> -       mutex_lock(&slab_mutex);
> -
> -       list_for_each_entry(s, &slab_caches, list)
> -               init_cache_random_seq(s);
> -
> -       mutex_unlock(&slab_mutex);
> -}
> -
> -/* Get the next entry on the pre-computed freelist randomized */
> -static void *next_freelist_entry(struct kmem_cache *s, struct slab *slab,
> -                               unsigned long *pos, void *start,
> -                               unsigned long page_limit,
> -                               unsigned long freelist_count)
> -{
> -       unsigned int idx;
> -
> -       /*
> -        * If the target page allocation failed, the number of objects on the
> -        * page might be smaller than the usual size defined by the cache.
> -        */
> -       do {
> -               idx = s->random_seq[*pos];
> -               *pos += 1;
> -               if (*pos >= freelist_count)
> -                       *pos = 0;
> -       } while (unlikely(idx >= page_limit));
> -
> -       return (char *)start + idx;
> -}
> -
> -/* Shuffle the single linked freelist based on a random pre-computed sequence */
> -static bool shuffle_freelist(struct kmem_cache *s, struct slab *slab)
> -{
> -       void *start;
> -       void *cur;
> -       void *next;
> -       unsigned long idx, pos, page_limit, freelist_count;
> -
> -       if (slab->objects < 2 || !s->random_seq)
> -               return false;
> -
> -       freelist_count = oo_objects(s->oo);
> -       pos = get_random_u32_below(freelist_count);
> -
> -       page_limit = slab->objects * s->size;
> -       start = fixup_red_left(s, slab_address(slab));
> -
> -       /* First entry is used as the base of the freelist */
> -       cur = next_freelist_entry(s, slab, &pos, start, page_limit,
> -                               freelist_count);
> -       cur = setup_object(s, cur);
> -       slab->freelist = cur;
> -
> -       for (idx = 1; idx < slab->objects; idx++) {
> -               next = next_freelist_entry(s, slab, &pos, start, page_limit,
> -                       freelist_count);
> -               next = setup_object(s, next);
> -               set_freepointer(s, cur, next);
> -               cur = next;
> -       }
> -       set_freepointer(s, cur, NULL);
> -
> -       return true;
> -}
> -#else
> -static inline int init_cache_random_seq(struct kmem_cache *s)
> -{
> -       return 0;
> -}
> -static inline void init_freelist_randomization(void) { }
> -static inline bool shuffle_freelist(struct kmem_cache *s, struct slab *slab)
> -{
> -       return false;
> -}
> -#endif /* CONFIG_SLAB_FREELIST_RANDOM */
> -
> -static struct slab *allocate_slab(struct kmem_cache *s, gfp_t flags, int node)
> -{
> -       struct slab *slab;
> -       struct kmem_cache_order_objects oo = s->oo;
> -       gfp_t alloc_gfp;
> -       void *start, *p, *next;
> -       int idx;
> -       bool shuffle;
> -
> -       flags &= gfp_allowed_mask;
> -
> -       flags |= s->allocflags;
> -
> -       /*
> -        * Let the initial higher-order allocation fail under memory pressure
> -        * so we fall-back to the minimum order allocation.
> -        */
> -       alloc_gfp = (flags | __GFP_NOWARN | __GFP_NORETRY) & ~__GFP_NOFAIL;
> -       if ((alloc_gfp & __GFP_DIRECT_RECLAIM) && oo_order(oo) > oo_order(s->min))
> -               alloc_gfp = (alloc_gfp | __GFP_NOMEMALLOC) & ~__GFP_RECLAIM;
> -
> -       slab = alloc_slab_page(alloc_gfp, node, oo);
> -       if (unlikely(!slab)) {
> -               oo = s->min;
> -               alloc_gfp = flags;
> -               /*
> -                * Allocation may have failed due to fragmentation.
> -                * Try a lower order alloc if possible
> -                */
> -               slab = alloc_slab_page(alloc_gfp, node, oo);
> -               if (unlikely(!slab))
> -                       return NULL;
> -               stat(s, ORDER_FALLBACK);
> -       }
> -
> -       slab->objects = oo_objects(oo);
> -       slab->inuse = 0;
> -       slab->frozen = 0;
> -
> -       account_slab(slab, oo_order(oo), s, flags);
> -
> -       slab->slab_cache = s;
> -
> -       kasan_poison_slab(slab);
> -
> -       start = slab_address(slab);
> -
> -       setup_slab_debug(s, slab, start);
> -
> -       shuffle = shuffle_freelist(s, slab);
> -
> -       if (!shuffle) {
> -               start = fixup_red_left(s, start);
> -               start = setup_object(s, start);
> -               slab->freelist = start;
> -               for (idx = 0, p = start; idx < slab->objects - 1; idx++) {
> -                       next = p + s->size;
> -                       next = setup_object(s, next);
> -                       set_freepointer(s, p, next);
> -                       p = next;
> -               }
> -               set_freepointer(s, p, NULL);
> -       }
> -
> -       return slab;
> -}
> -
> -static struct slab *new_slab(struct kmem_cache *s, gfp_t flags, int node)
> -{
> -       if (unlikely(flags & GFP_SLAB_BUG_MASK))
> -               flags = kmalloc_fix_flags(flags);
> -
> -       WARN_ON_ONCE(s->ctor && (flags & __GFP_ZERO));
> -
> -       return allocate_slab(s,
> -               flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node);
> -}
> -
> -static void __free_slab(struct kmem_cache *s, struct slab *slab)
> -{
> -       struct folio *folio = slab_folio(slab);
> -       int order = folio_order(folio);
> -       int pages = 1 << order;
> -
> -       __slab_clear_pfmemalloc(slab);
> -       folio->mapping = NULL;
> -       /* Make the mapping reset visible before clearing the flag */
> -       smp_wmb();
> -       __folio_clear_slab(folio);
> -       if (current->reclaim_state)
> -               current->reclaim_state->reclaimed_slab += pages;
> -       unaccount_slab(slab, order, s);
> -       __free_pages(&folio->page, order);
> -}
> -
> -static void rcu_free_slab(struct rcu_head *h)
> -{
> -       struct slab *slab = container_of(h, struct slab, rcu_head);
> -
> -       __free_slab(slab->slab_cache, slab);
> -}
> -
> -static void free_slab(struct kmem_cache *s, struct slab *slab)
> -{
> -       if (kmem_cache_debug_flags(s, SLAB_CONSISTENCY_CHECKS)) {
> -               void *p;
> -
> -               slab_pad_check(s, slab);
> -               for_each_object(p, s, slab_address(slab), slab->objects)
> -                       check_object(s, slab, p, SLUB_RED_INACTIVE);
> -       }
> -
> -       if (unlikely(s->flags & SLAB_TYPESAFE_BY_RCU))
> -               call_rcu(&slab->rcu_head, rcu_free_slab);
> -       else
> -               __free_slab(s, slab);
> -}
> -
> -static void discard_slab(struct kmem_cache *s, struct slab *slab)
> -{
> -       dec_slabs_node(s, slab_nid(slab), slab->objects);
> -       free_slab(s, slab);
> -}
> -
> -/*
> - * Management of partially allocated slabs.
> - */
> -static inline void
> -__add_partial(struct kmem_cache_node *n, struct slab *slab, int tail)
> -{
> -       n->nr_partial++;
> -       if (tail == DEACTIVATE_TO_TAIL)
> -               list_add_tail(&slab->slab_list, &n->partial);
> -       else
> -               list_add(&slab->slab_list, &n->partial);
> -}
> -
> -static inline void add_partial(struct kmem_cache_node *n,
> -                               struct slab *slab, int tail)
> -{
> -       lockdep_assert_held(&n->list_lock);
> -       __add_partial(n, slab, tail);
> -}
> -
> -static inline void remove_partial(struct kmem_cache_node *n,
> -                                       struct slab *slab)
> -{
> -       lockdep_assert_held(&n->list_lock);
> -       list_del(&slab->slab_list);
> -       n->nr_partial--;
> -}
> -
> -/*
> - * Called only for kmem_cache_debug() caches instead of acquire_slab(), with a
> - * slab from the n->partial list. Remove only a single object from the slab, do
> - * the alloc_debug_processing() checks and leave the slab on the list, or move
> - * it to full list if it was the last free object.
> - */
> -static void *alloc_single_from_partial(struct kmem_cache *s,
> -               struct kmem_cache_node *n, struct slab *slab, int orig_size)
> -{
> -       void *object;
> -
> -       lockdep_assert_held(&n->list_lock);
> -
> -       object = slab->freelist;
> -       slab->freelist = get_freepointer(s, object);
> -       slab->inuse++;
> -
> -       if (!alloc_debug_processing(s, slab, object, orig_size)) {
> -               remove_partial(n, slab);
> -               return NULL;
> -       }
> -
> -       if (slab->inuse == slab->objects) {
> -               remove_partial(n, slab);
> -               add_full(s, n, slab);
> -       }
> -
> -       return object;
> -}
> -
> -/*
> - * Called only for kmem_cache_debug() caches to allocate from a freshly
> - * allocated slab. Allocate a single object instead of whole freelist
> - * and put the slab to the partial (or full) list.
> - */
> -static void *alloc_single_from_new_slab(struct kmem_cache *s,
> -                                       struct slab *slab, int orig_size)
> -{
> -       int nid = slab_nid(slab);
> -       struct kmem_cache_node *n = get_node(s, nid);
> -       unsigned long flags;
> -       void *object;
> -
> -
> -       object = slab->freelist;
> -       slab->freelist = get_freepointer(s, object);
> -       slab->inuse = 1;
> -
> -       if (!alloc_debug_processing(s, slab, object, orig_size))
> -               /*
> -                * It's not really expected that this would fail on a
> -                * freshly allocated slab, but a concurrent memory
> -                * corruption in theory could cause that.
> -                */
> -               return NULL;
> -
> -       spin_lock_irqsave(&n->list_lock, flags);
> -
> -       if (slab->inuse == slab->objects)
> -               add_full(s, n, slab);
> -       else
> -               add_partial(n, slab, DEACTIVATE_TO_HEAD);
> -
> -       inc_slabs_node(s, nid, slab->objects);
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -       return object;
> -}
> -
> -/*
> - * Remove slab from the partial list, freeze it and
> - * return the pointer to the freelist.
> - *
> - * Returns a list of objects or NULL if it fails.
> - */
> -static inline void *acquire_slab(struct kmem_cache *s,
> -               struct kmem_cache_node *n, struct slab *slab,
> -               int mode)
> -{
> -       void *freelist;
> -       unsigned long counters;
> -       struct slab new;
> -
> -       lockdep_assert_held(&n->list_lock);
> -
> -       /*
> -        * Zap the freelist and set the frozen bit.
> -        * The old freelist is the list of objects for the
> -        * per cpu allocation list.
> -        */
> -       freelist = slab->freelist;
> -       counters = slab->counters;
> -       new.counters = counters;
> -       if (mode) {
> -               new.inuse = slab->objects;
> -               new.freelist = NULL;
> -       } else {
> -               new.freelist = freelist;
> -       }
> -
> -       VM_BUG_ON(new.frozen);
> -       new.frozen = 1;
> -
> -       if (!__cmpxchg_double_slab(s, slab,
> -                       freelist, counters,
> -                       new.freelist, new.counters,
> -                       "acquire_slab"))
> -               return NULL;
> -
> -       remove_partial(n, slab);
> -       WARN_ON(!freelist);
> -       return freelist;
> -}
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -static void put_cpu_partial(struct kmem_cache *s, struct slab *slab, int drain);
> -#else
> -static inline void put_cpu_partial(struct kmem_cache *s, struct slab *slab,
> -                                  int drain) { }
> -#endif
> -static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags);
> -
> -/*
> - * Try to allocate a partial slab from a specific node.
> - */
> -static void *get_partial_node(struct kmem_cache *s, struct kmem_cache_node *n,
> -                             struct partial_context *pc)
> -{
> -       struct slab *slab, *slab2;
> -       void *object = NULL;
> -       unsigned long flags;
> -       unsigned int partial_slabs = 0;
> -
> -       /*
> -        * Racy check. If we mistakenly see no partial slabs then we
> -        * just allocate an empty slab. If we mistakenly try to get a
> -        * partial slab and there is none available then get_partial()
> -        * will return NULL.
> -        */
> -       if (!n || !n->nr_partial)
> -               return NULL;
> -
> -       spin_lock_irqsave(&n->list_lock, flags);
> -       list_for_each_entry_safe(slab, slab2, &n->partial, slab_list) {
> -               void *t;
> -
> -               if (!pfmemalloc_match(slab, pc->flags))
> -                       continue;
> -
> -               if (IS_ENABLED(CONFIG_SLUB_TINY) || kmem_cache_debug(s)) {
> -                       object = alloc_single_from_partial(s, n, slab,
> -                                                       pc->orig_size);
> -                       if (object)
> -                               break;
> -                       continue;
> -               }
> -
> -               t = acquire_slab(s, n, slab, object == NULL);
> -               if (!t)
> -                       break;
> -
> -               if (!object) {
> -                       *pc->slab = slab;
> -                       stat(s, ALLOC_FROM_PARTIAL);
> -                       object = t;
> -               } else {
> -                       put_cpu_partial(s, slab, 0);
> -                       stat(s, CPU_PARTIAL_NODE);
> -                       partial_slabs++;
> -               }
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -               if (!kmem_cache_has_cpu_partial(s)
> -                       || partial_slabs > s->cpu_partial_slabs / 2)
> -                       break;
> -#else
> -               break;
> -#endif
> -
> -       }
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -       return object;
> -}
> -
> -/*
> - * Get a slab from somewhere. Search in increasing NUMA distances.
> - */
> -static void *get_any_partial(struct kmem_cache *s, struct partial_context *pc)
> -{
> -#ifdef CONFIG_NUMA
> -       struct zonelist *zonelist;
> -       struct zoneref *z;
> -       struct zone *zone;
> -       enum zone_type highest_zoneidx = gfp_zone(pc->flags);
> -       void *object;
> -       unsigned int cpuset_mems_cookie;
> -
> -       /*
> -        * The defrag ratio allows a configuration of the tradeoffs between
> -        * inter node defragmentation and node local allocations. A lower
> -        * defrag_ratio increases the tendency to do local allocations
> -        * instead of attempting to obtain partial slabs from other nodes.
> -        *
> -        * If the defrag_ratio is set to 0 then kmalloc() always
> -        * returns node local objects. If the ratio is higher then kmalloc()
> -        * may return off node objects because partial slabs are obtained
> -        * from other nodes and filled up.
> -        *
> -        * If /sys/kernel/slab/xx/remote_node_defrag_ratio is set to 100
> -        * (which makes defrag_ratio = 1000) then every (well almost)
> -        * allocation will first attempt to defrag slab caches on other nodes.
> -        * This means scanning over all nodes to look for partial slabs which
> -        * may be expensive if we do it every time we are trying to find a slab
> -        * with available objects.
> -        */
> -       if (!s->remote_node_defrag_ratio ||
> -                       get_cycles() % 1024 > s->remote_node_defrag_ratio)
> -               return NULL;
> -
> -       do {
> -               cpuset_mems_cookie = read_mems_allowed_begin();
> -               zonelist = node_zonelist(mempolicy_slab_node(), pc->flags);
> -               for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) {
> -                       struct kmem_cache_node *n;
> -
> -                       n = get_node(s, zone_to_nid(zone));
> -
> -                       if (n && cpuset_zone_allowed(zone, pc->flags) &&
> -                                       n->nr_partial > s->min_partial) {
> -                               object = get_partial_node(s, n, pc);
> -                               if (object) {
> -                                       /*
> -                                        * Don't check read_mems_allowed_retry()
> -                                        * here - if mems_allowed was updated in
> -                                        * parallel, that was a harmless race
> -                                        * between allocation and the cpuset
> -                                        * update
> -                                        */
> -                                       return object;
> -                               }
> -                       }
> -               }
> -       } while (read_mems_allowed_retry(cpuset_mems_cookie));
> -#endif /* CONFIG_NUMA */
> -       return NULL;
> -}
> -
> -/*
> - * Get a partial slab, lock it and return it.
> - */
> -static void *get_partial(struct kmem_cache *s, int node, struct partial_context *pc)
> -{
> -       void *object;
> -       int searchnode = node;
> -
> -       if (node == NUMA_NO_NODE)
> -               searchnode = numa_mem_id();
> -
> -       object = get_partial_node(s, get_node(s, searchnode), pc);
> -       if (object || node != NUMA_NO_NODE)
> -               return object;
> -
> -       return get_any_partial(s, pc);
> -}
> -
> -#ifndef CONFIG_SLUB_TINY
> -
> -#ifdef CONFIG_PREEMPTION
> -/*
> - * Calculate the next globally unique transaction for disambiguation
> - * during cmpxchg. The transactions start with the cpu number and are then
> - * incremented by CONFIG_NR_CPUS.
> - */
> -#define TID_STEP  roundup_pow_of_two(CONFIG_NR_CPUS)
> -#else
> -/*
> - * No preemption supported therefore also no need to check for
> - * different cpus.
> - */
> -#define TID_STEP 1
> -#endif /* CONFIG_PREEMPTION */
> -
> -static inline unsigned long next_tid(unsigned long tid)
> -{
> -       return tid + TID_STEP;
> -}
> -
> -#ifdef SLUB_DEBUG_CMPXCHG
> -static inline unsigned int tid_to_cpu(unsigned long tid)
> -{
> -       return tid % TID_STEP;
> -}
> -
> -static inline unsigned long tid_to_event(unsigned long tid)
> -{
> -       return tid / TID_STEP;
> -}
> -#endif
> -
> -static inline unsigned int init_tid(int cpu)
> -{
> -       return cpu;
> -}
> -
> -static inline void note_cmpxchg_failure(const char *n,
> -               const struct kmem_cache *s, unsigned long tid)
> -{
> -#ifdef SLUB_DEBUG_CMPXCHG
> -       unsigned long actual_tid = __this_cpu_read(s->cpu_slab->tid);
> -
> -       pr_info("%s %s: cmpxchg redo ", n, s->name);
> -
> -#ifdef CONFIG_PREEMPTION
> -       if (tid_to_cpu(tid) != tid_to_cpu(actual_tid))
> -               pr_warn("due to cpu change %d -> %d\n",
> -                       tid_to_cpu(tid), tid_to_cpu(actual_tid));
> -       else
> -#endif
> -       if (tid_to_event(tid) != tid_to_event(actual_tid))
> -               pr_warn("due to cpu running other code. Event %ld->%ld\n",
> -                       tid_to_event(tid), tid_to_event(actual_tid));
> -       else
> -               pr_warn("for unknown reason: actual=%lx was=%lx target=%lx\n",
> -                       actual_tid, tid, next_tid(tid));
> -#endif
> -       stat(s, CMPXCHG_DOUBLE_CPU_FAIL);
> -}
> -
> -static void init_kmem_cache_cpus(struct kmem_cache *s)
> -{
> -       int cpu;
> -       struct kmem_cache_cpu *c;
> -
> -       for_each_possible_cpu(cpu) {
> -               c = per_cpu_ptr(s->cpu_slab, cpu);
> -               local_lock_init(&c->lock);
> -               c->tid = init_tid(cpu);
> -       }
> -}
> -
> -/*
> - * Finishes removing the cpu slab. Merges cpu's freelist with slab's freelist,
> - * unfreezes the slabs and puts it on the proper list.
> - * Assumes the slab has been already safely taken away from kmem_cache_cpu
> - * by the caller.
> - */
> -static void deactivate_slab(struct kmem_cache *s, struct slab *slab,
> -                           void *freelist)
> -{
> -       enum slab_modes { M_NONE, M_PARTIAL, M_FREE, M_FULL_NOLIST };
> -       struct kmem_cache_node *n = get_node(s, slab_nid(slab));
> -       int free_delta = 0;
> -       enum slab_modes mode = M_NONE;
> -       void *nextfree, *freelist_iter, *freelist_tail;
> -       int tail = DEACTIVATE_TO_HEAD;
> -       unsigned long flags = 0;
> -       struct slab new;
> -       struct slab old;
> -
> -       if (slab->freelist) {
> -               stat(s, DEACTIVATE_REMOTE_FREES);
> -               tail = DEACTIVATE_TO_TAIL;
> -       }
> -
> -       /*
> -        * Stage one: Count the objects on cpu's freelist as free_delta and
> -        * remember the last object in freelist_tail for later splicing.
> -        */
> -       freelist_tail = NULL;
> -       freelist_iter = freelist;
> -       while (freelist_iter) {
> -               nextfree = get_freepointer(s, freelist_iter);
> -
> -               /*
> -                * If 'nextfree' is invalid, it is possible that the object at
> -                * 'freelist_iter' is already corrupted.  So isolate all objects
> -                * starting at 'freelist_iter' by skipping them.
> -                */
> -               if (freelist_corrupted(s, slab, &freelist_iter, nextfree))
> -                       break;
> -
> -               freelist_tail = freelist_iter;
> -               free_delta++;
> -
> -               freelist_iter = nextfree;
> -       }
> -
> -       /*
> -        * Stage two: Unfreeze the slab while splicing the per-cpu
> -        * freelist to the head of slab's freelist.
> -        *
> -        * Ensure that the slab is unfrozen while the list presence
> -        * reflects the actual number of objects during unfreeze.
> -        *
> -        * We first perform cmpxchg holding lock and insert to list
> -        * when it succeed. If there is mismatch then the slab is not
> -        * unfrozen and number of objects in the slab may have changed.
> -        * Then release lock and retry cmpxchg again.
> -        */
> -redo:
> -
> -       old.freelist = READ_ONCE(slab->freelist);
> -       old.counters = READ_ONCE(slab->counters);
> -       VM_BUG_ON(!old.frozen);
> -
> -       /* Determine target state of the slab */
> -       new.counters = old.counters;
> -       if (freelist_tail) {
> -               new.inuse -= free_delta;
> -               set_freepointer(s, freelist_tail, old.freelist);
> -               new.freelist = freelist;
> -       } else
> -               new.freelist = old.freelist;
> -
> -       new.frozen = 0;
> -
> -       if (!new.inuse && n->nr_partial >= s->min_partial) {
> -               mode = M_FREE;
> -       } else if (new.freelist) {
> -               mode = M_PARTIAL;
> -               /*
> -                * Taking the spinlock removes the possibility that
> -                * acquire_slab() will see a slab that is frozen
> -                */
> -               spin_lock_irqsave(&n->list_lock, flags);
> -       } else {
> -               mode = M_FULL_NOLIST;
> -       }
> -
> -
> -       if (!cmpxchg_double_slab(s, slab,
> -                               old.freelist, old.counters,
> -                               new.freelist, new.counters,
> -                               "unfreezing slab")) {
> -               if (mode == M_PARTIAL)
> -                       spin_unlock_irqrestore(&n->list_lock, flags);
> -               goto redo;
> -       }
> -
> -
> -       if (mode == M_PARTIAL) {
> -               add_partial(n, slab, tail);
> -               spin_unlock_irqrestore(&n->list_lock, flags);
> -               stat(s, tail);
> -       } else if (mode == M_FREE) {
> -               stat(s, DEACTIVATE_EMPTY);
> -               discard_slab(s, slab);
> -               stat(s, FREE_SLAB);
> -       } else if (mode == M_FULL_NOLIST) {
> -               stat(s, DEACTIVATE_FULL);
> -       }
> -}
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -static void __unfreeze_partials(struct kmem_cache *s, struct slab *partial_slab)
> -{
> -       struct kmem_cache_node *n = NULL, *n2 = NULL;
> -       struct slab *slab, *slab_to_discard = NULL;
> -       unsigned long flags = 0;
> -
> -       while (partial_slab) {
> -               struct slab new;
> -               struct slab old;
> -
> -               slab = partial_slab;
> -               partial_slab = slab->next;
> -
> -               n2 = get_node(s, slab_nid(slab));
> -               if (n != n2) {
> -                       if (n)
> -                               spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -                       n = n2;
> -                       spin_lock_irqsave(&n->list_lock, flags);
> -               }
> -
> -               do {
> -
> -                       old.freelist = slab->freelist;
> -                       old.counters = slab->counters;
> -                       VM_BUG_ON(!old.frozen);
> -
> -                       new.counters = old.counters;
> -                       new.freelist = old.freelist;
> -
> -                       new.frozen = 0;
> -
> -               } while (!__cmpxchg_double_slab(s, slab,
> -                               old.freelist, old.counters,
> -                               new.freelist, new.counters,
> -                               "unfreezing slab"));
> -
> -               if (unlikely(!new.inuse && n->nr_partial >= s->min_partial)) {
> -                       slab->next = slab_to_discard;
> -                       slab_to_discard = slab;
> -               } else {
> -                       add_partial(n, slab, DEACTIVATE_TO_TAIL);
> -                       stat(s, FREE_ADD_PARTIAL);
> -               }
> -       }
> -
> -       if (n)
> -               spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -       while (slab_to_discard) {
> -               slab = slab_to_discard;
> -               slab_to_discard = slab_to_discard->next;
> -
> -               stat(s, DEACTIVATE_EMPTY);
> -               discard_slab(s, slab);
> -               stat(s, FREE_SLAB);
> -       }
> -}
> -
> -/*
> - * Unfreeze all the cpu partial slabs.
> - */
> -static void unfreeze_partials(struct kmem_cache *s)
> -{
> -       struct slab *partial_slab;
> -       unsigned long flags;
> -
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -       partial_slab = this_cpu_read(s->cpu_slab->partial);
> -       this_cpu_write(s->cpu_slab->partial, NULL);
> -       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -
> -       if (partial_slab)
> -               __unfreeze_partials(s, partial_slab);
> -}
> -
> -static void unfreeze_partials_cpu(struct kmem_cache *s,
> -                                 struct kmem_cache_cpu *c)
> -{
> -       struct slab *partial_slab;
> -
> -       partial_slab = slub_percpu_partial(c);
> -       c->partial = NULL;
> -
> -       if (partial_slab)
> -               __unfreeze_partials(s, partial_slab);
> -}
> -
> -/*
> - * Put a slab that was just frozen (in __slab_free|get_partial_node) into a
> - * partial slab slot if available.
> - *
> - * If we did not find a slot then simply move all the partials to the
> - * per node partial list.
> - */
> -static void put_cpu_partial(struct kmem_cache *s, struct slab *slab, int drain)
> -{
> -       struct slab *oldslab;
> -       struct slab *slab_to_unfreeze = NULL;
> -       unsigned long flags;
> -       int slabs = 0;
> -
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -
> -       oldslab = this_cpu_read(s->cpu_slab->partial);
> -
> -       if (oldslab) {
> -               if (drain && oldslab->slabs >= s->cpu_partial_slabs) {
> -                       /*
> -                        * Partial array is full. Move the existing set to the
> -                        * per node partial list. Postpone the actual unfreezing
> -                        * outside of the critical section.
> -                        */
> -                       slab_to_unfreeze = oldslab;
> -                       oldslab = NULL;
> -               } else {
> -                       slabs = oldslab->slabs;
> -               }
> -       }
> -
> -       slabs++;
> -
> -       slab->slabs = slabs;
> -       slab->next = oldslab;
> -
> -       this_cpu_write(s->cpu_slab->partial, slab);
> -
> -       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -
> -       if (slab_to_unfreeze) {
> -               __unfreeze_partials(s, slab_to_unfreeze);
> -               stat(s, CPU_PARTIAL_DRAIN);
> -       }
> -}
> -
> -#else  /* CONFIG_SLUB_CPU_PARTIAL */
> -
> -static inline void unfreeze_partials(struct kmem_cache *s) { }
> -static inline void unfreeze_partials_cpu(struct kmem_cache *s,
> -                                 struct kmem_cache_cpu *c) { }
> -
> -#endif /* CONFIG_SLUB_CPU_PARTIAL */
> -
> -static inline void flush_slab(struct kmem_cache *s, struct kmem_cache_cpu *c)
> -{
> -       unsigned long flags;
> -       struct slab *slab;
> -       void *freelist;
> -
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -
> -       slab = c->slab;
> -       freelist = c->freelist;
> -
> -       c->slab = NULL;
> -       c->freelist = NULL;
> -       c->tid = next_tid(c->tid);
> -
> -       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -
> -       if (slab) {
> -               deactivate_slab(s, slab, freelist);
> -               stat(s, CPUSLAB_FLUSH);
> -       }
> -}
> -
> -static inline void __flush_cpu_slab(struct kmem_cache *s, int cpu)
> -{
> -       struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab, cpu);
> -       void *freelist = c->freelist;
> -       struct slab *slab = c->slab;
> -
> -       c->slab = NULL;
> -       c->freelist = NULL;
> -       c->tid = next_tid(c->tid);
> -
> -       if (slab) {
> -               deactivate_slab(s, slab, freelist);
> -               stat(s, CPUSLAB_FLUSH);
> -       }
> -
> -       unfreeze_partials_cpu(s, c);
> -}
> -
> -struct slub_flush_work {
> -       struct work_struct work;
> -       struct kmem_cache *s;
> -       bool skip;
> -};
> -
> -/*
> - * Flush cpu slab.
> - *
> - * Called from CPU work handler with migration disabled.
> - */
> -static void flush_cpu_slab(struct work_struct *w)
> -{
> -       struct kmem_cache *s;
> -       struct kmem_cache_cpu *c;
> -       struct slub_flush_work *sfw;
> -
> -       sfw = container_of(w, struct slub_flush_work, work);
> -
> -       s = sfw->s;
> -       c = this_cpu_ptr(s->cpu_slab);
> -
> -       if (c->slab)
> -               flush_slab(s, c);
> -
> -       unfreeze_partials(s);
> -}
> -
> -static bool has_cpu_slab(int cpu, struct kmem_cache *s)
> -{
> -       struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab, cpu);
> -
> -       return c->slab || slub_percpu_partial(c);
> -}
> -
> -static DEFINE_MUTEX(flush_lock);
> -static DEFINE_PER_CPU(struct slub_flush_work, slub_flush);
> -
> -static void flush_all_cpus_locked(struct kmem_cache *s)
> -{
> -       struct slub_flush_work *sfw;
> -       unsigned int cpu;
> -
> -       lockdep_assert_cpus_held();
> -       mutex_lock(&flush_lock);
> -
> -       for_each_online_cpu(cpu) {
> -               sfw = &per_cpu(slub_flush, cpu);
> -               if (!has_cpu_slab(cpu, s)) {
> -                       sfw->skip = true;
> -                       continue;
> -               }
> -               INIT_WORK(&sfw->work, flush_cpu_slab);
> -               sfw->skip = false;
> -               sfw->s = s;
> -               queue_work_on(cpu, flushwq, &sfw->work);
> -       }
> -
> -       for_each_online_cpu(cpu) {
> -               sfw = &per_cpu(slub_flush, cpu);
> -               if (sfw->skip)
> -                       continue;
> -               flush_work(&sfw->work);
> -       }
> -
> -       mutex_unlock(&flush_lock);
> -}
> -
> -static void flush_all(struct kmem_cache *s)
> -{
> -       cpus_read_lock();
> -       flush_all_cpus_locked(s);
> -       cpus_read_unlock();
> -}
> -
> -/*
> - * Use the cpu notifier to insure that the cpu slabs are flushed when
> - * necessary.
> - */
> -static int slub_cpu_dead(unsigned int cpu)
> -{
> -       struct kmem_cache *s;
> -
> -       mutex_lock(&slab_mutex);
> -       list_for_each_entry(s, &slab_caches, list)
> -               __flush_cpu_slab(s, cpu);
> -       mutex_unlock(&slab_mutex);
> -       return 0;
> -}
> -
> -#else /* CONFIG_SLUB_TINY */
> -static inline void flush_all_cpus_locked(struct kmem_cache *s) { }
> -static inline void flush_all(struct kmem_cache *s) { }
> -static inline void __flush_cpu_slab(struct kmem_cache *s, int cpu) { }
> -static inline int slub_cpu_dead(unsigned int cpu) { return 0; }
> -#endif /* CONFIG_SLUB_TINY */
> -
> -/*
> - * Check if the objects in a per cpu structure fit numa
> - * locality expectations.
> - */
> -static inline int node_match(struct slab *slab, int node)
> -{
> -#ifdef CONFIG_NUMA
> -       if (node != NUMA_NO_NODE && slab_nid(slab) != node)
> -               return 0;
> -#endif
> -       return 1;
> -}
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -static int count_free(struct slab *slab)
> -{
> -       return slab->objects - slab->inuse;
> -}
> -
> -static inline unsigned long node_nr_objs(struct kmem_cache_node *n)
> -{
> -       return atomic_long_read(&n->total_objects);
> -}
> -
> -/* Supports checking bulk free of a constructed freelist */
> -static inline bool free_debug_processing(struct kmem_cache *s,
> -       struct slab *slab, void *head, void *tail, int *bulk_cnt,
> -       unsigned long addr, depot_stack_handle_t handle)
> -{
> -       bool checks_ok = false;
> -       void *object = head;
> -       int cnt = 0;
> -
> -       if (s->flags & SLAB_CONSISTENCY_CHECKS) {
> -               if (!check_slab(s, slab))
> -                       goto out;
> -       }
> -
> -       if (slab->inuse < *bulk_cnt) {
> -               slab_err(s, slab, "Slab has %d allocated objects but %d are to be freed\n",
> -                        slab->inuse, *bulk_cnt);
> -               goto out;
> -       }
> -
> -next_object:
> -
> -       if (++cnt > *bulk_cnt)
> -               goto out_cnt;
> -
> -       if (s->flags & SLAB_CONSISTENCY_CHECKS) {
> -               if (!free_consistency_checks(s, slab, object, addr))
> -                       goto out;
> -       }
> -
> -       if (s->flags & SLAB_STORE_USER)
> -               set_track_update(s, object, TRACK_FREE, addr, handle);
> -       trace(s, slab, object, 0);
> -       /* Freepointer not overwritten by init_object(), SLAB_POISON moved it */
> -       init_object(s, object, SLUB_RED_INACTIVE);
> -
> -       /* Reached end of constructed freelist yet? */
> -       if (object != tail) {
> -               object = get_freepointer(s, object);
> -               goto next_object;
> -       }
> -       checks_ok = true;
> -
> -out_cnt:
> -       if (cnt != *bulk_cnt) {
> -               slab_err(s, slab, "Bulk free expected %d objects but found %d\n",
> -                        *bulk_cnt, cnt);
> -               *bulk_cnt = cnt;
> -       }
> -
> -out:
> -
> -       if (!checks_ok)
> -               slab_fix(s, "Object at 0x%p not freed", object);
> -
> -       return checks_ok;
> -}
> -#endif /* CONFIG_SLUB_DEBUG */
> -
> -#if defined(CONFIG_SLUB_DEBUG) || defined(SLAB_SUPPORTS_SYSFS)
> -static unsigned long count_partial(struct kmem_cache_node *n,
> -                                       int (*get_count)(struct slab *))
> -{
> -       unsigned long flags;
> -       unsigned long x = 0;
> -       struct slab *slab;
> -
> -       spin_lock_irqsave(&n->list_lock, flags);
> -       list_for_each_entry(slab, &n->partial, slab_list)
> -               x += get_count(slab);
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -       return x;
> -}
> -#endif /* CONFIG_SLUB_DEBUG || SLAB_SUPPORTS_SYSFS */
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -static noinline void
> -slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid)
> -{
> -       static DEFINE_RATELIMIT_STATE(slub_oom_rs, DEFAULT_RATELIMIT_INTERVAL,
> -                                     DEFAULT_RATELIMIT_BURST);
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slub_oom_rs))
> -               return;
> -
> -       pr_warn("SLUB: Unable to allocate memory on node %d, gfp=%#x(%pGg)\n",
> -               nid, gfpflags, &gfpflags);
> -       pr_warn("  cache: %s, object size: %u, buffer size: %u, default order: %u, min order: %u\n",
> -               s->name, s->object_size, s->size, oo_order(s->oo),
> -               oo_order(s->min));
> -
> -       if (oo_order(s->min) > get_order(s->object_size))
> -               pr_warn("  %s debugging increased min order, use slub_debug=O to disable.\n",
> -                       s->name);
> -
> -       for_each_kmem_cache_node(s, node, n) {
> -               unsigned long nr_slabs;
> -               unsigned long nr_objs;
> -               unsigned long nr_free;
> -
> -               nr_free  = count_partial(n, count_free);
> -               nr_slabs = node_nr_slabs(n);
> -               nr_objs  = node_nr_objs(n);
> -
> -               pr_warn("  node %d: slabs: %ld, objs: %ld, free: %ld\n",
> -                       node, nr_slabs, nr_objs, nr_free);
> -       }
> -}
> -#else /* CONFIG_SLUB_DEBUG */
> -static inline void
> -slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) { }
> -#endif
> -
> -static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags)
> -{
> -       if (unlikely(slab_test_pfmemalloc(slab)))
> -               return gfp_pfmemalloc_allowed(gfpflags);
> -
> -       return true;
> -}
> -
> -#ifndef CONFIG_SLUB_TINY
> -/*
> - * Check the slab->freelist and either transfer the freelist to the
> - * per cpu freelist or deactivate the slab.
> - *
> - * The slab is still frozen if the return value is not NULL.
> - *
> - * If this function returns NULL then the slab has been unfrozen.
> - */
> -static inline void *get_freelist(struct kmem_cache *s, struct slab *slab)
> -{
> -       struct slab new;
> -       unsigned long counters;
> -       void *freelist;
> -
> -       lockdep_assert_held(this_cpu_ptr(&s->cpu_slab->lock));
> -
> -       do {
> -               freelist = slab->freelist;
> -               counters = slab->counters;
> -
> -               new.counters = counters;
> -               VM_BUG_ON(!new.frozen);
> -
> -               new.inuse = slab->objects;
> -               new.frozen = freelist != NULL;
> -
> -       } while (!__cmpxchg_double_slab(s, slab,
> -               freelist, counters,
> -               NULL, new.counters,
> -               "get_freelist"));
> -
> -       return freelist;
> -}
> -
> -/*
> - * Slow path. The lockless freelist is empty or we need to perform
> - * debugging duties.
> - *
> - * Processing is still very fast if new objects have been freed to the
> - * regular freelist. In that case we simply take over the regular freelist
> - * as the lockless freelist and zap the regular freelist.
> - *
> - * If that is not working then we fall back to the partial lists. We take the
> - * first element of the freelist as the object to allocate now and move the
> - * rest of the freelist to the lockless freelist.
> - *
> - * And if we were unable to get a new slab from the partial slab lists then
> - * we need to allocate a new slab. This is the slowest path since it involves
> - * a call to the page allocator and the setup of a new slab.
> - *
> - * Version of __slab_alloc to use when we know that preemption is
> - * already disabled (which is the case for bulk allocation).
> - */
> -static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
> -                         unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size)
> -{
> -       void *freelist;
> -       struct slab *slab;
> -       unsigned long flags;
> -       struct partial_context pc;
> -
> -       stat(s, ALLOC_SLOWPATH);
> -
> -reread_slab:
> -
> -       slab = READ_ONCE(c->slab);
> -       if (!slab) {
> -               /*
> -                * if the node is not online or has no normal memory, just
> -                * ignore the node constraint
> -                */
> -               if (unlikely(node != NUMA_NO_NODE &&
> -                            !node_isset(node, slab_nodes)))
> -                       node = NUMA_NO_NODE;
> -               goto new_slab;
> -       }
> -redo:
> -
> -       if (unlikely(!node_match(slab, node))) {
> -               /*
> -                * same as above but node_match() being false already
> -                * implies node != NUMA_NO_NODE
> -                */
> -               if (!node_isset(node, slab_nodes)) {
> -                       node = NUMA_NO_NODE;
> -               } else {
> -                       stat(s, ALLOC_NODE_MISMATCH);
> -                       goto deactivate_slab;
> -               }
> -       }
> -
> -       /*
> -        * By rights, we should be searching for a slab page that was
> -        * PFMEMALLOC but right now, we are losing the pfmemalloc
> -        * information when the page leaves the per-cpu allocator
> -        */
> -       if (unlikely(!pfmemalloc_match(slab, gfpflags)))
> -               goto deactivate_slab;
> -
> -       /* must check again c->slab in case we got preempted and it changed */
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -       if (unlikely(slab != c->slab)) {
> -               local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -               goto reread_slab;
> -       }
> -       freelist = c->freelist;
> -       if (freelist)
> -               goto load_freelist;
> -
> -       freelist = get_freelist(s, slab);
> -
> -       if (!freelist) {
> -               c->slab = NULL;
> -               c->tid = next_tid(c->tid);
> -               local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -               stat(s, DEACTIVATE_BYPASS);
> -               goto new_slab;
> -       }
> -
> -       stat(s, ALLOC_REFILL);
> -
> -load_freelist:
> -
> -       lockdep_assert_held(this_cpu_ptr(&s->cpu_slab->lock));
> -
> -       /*
> -        * freelist is pointing to the list of objects to be used.
> -        * slab is pointing to the slab from which the objects are obtained.
> -        * That slab must be frozen for per cpu allocations to work.
> -        */
> -       VM_BUG_ON(!c->slab->frozen);
> -       c->freelist = get_freepointer(s, freelist);
> -       c->tid = next_tid(c->tid);
> -       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -       return freelist;
> -
> -deactivate_slab:
> -
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -       if (slab != c->slab) {
> -               local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -               goto reread_slab;
> -       }
> -       freelist = c->freelist;
> -       c->slab = NULL;
> -       c->freelist = NULL;
> -       c->tid = next_tid(c->tid);
> -       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -       deactivate_slab(s, slab, freelist);
> -
> -new_slab:
> -
> -       if (slub_percpu_partial(c)) {
> -               local_lock_irqsave(&s->cpu_slab->lock, flags);
> -               if (unlikely(c->slab)) {
> -                       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -                       goto reread_slab;
> -               }
> -               if (unlikely(!slub_percpu_partial(c))) {
> -                       local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -                       /* we were preempted and partial list got empty */
> -                       goto new_objects;
> -               }
> -
> -               slab = c->slab = slub_percpu_partial(c);
> -               slub_set_percpu_partial(c, slab);
> -               local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -               stat(s, CPU_PARTIAL_ALLOC);
> -               goto redo;
> -       }
> -
> -new_objects:
> -
> -       pc.flags = gfpflags;
> -       pc.slab = &slab;
> -       pc.orig_size = orig_size;
> -       freelist = get_partial(s, node, &pc);
> -       if (freelist)
> -               goto check_new_slab;
> -
> -       slub_put_cpu_ptr(s->cpu_slab);
> -       slab = new_slab(s, gfpflags, node);
> -       c = slub_get_cpu_ptr(s->cpu_slab);
> -
> -       if (unlikely(!slab)) {
> -               slab_out_of_memory(s, gfpflags, node);
> -               return NULL;
> -       }
> -
> -       stat(s, ALLOC_SLAB);
> -
> -       if (kmem_cache_debug(s)) {
> -               freelist = alloc_single_from_new_slab(s, slab, orig_size);
> -
> -               if (unlikely(!freelist))
> -                       goto new_objects;
> -
> -               if (s->flags & SLAB_STORE_USER)
> -                       set_track(s, freelist, TRACK_ALLOC, addr);
> -
> -               return freelist;
> -       }
> -
> -       /*
> -        * No other reference to the slab yet so we can
> -        * muck around with it freely without cmpxchg
> -        */
> -       freelist = slab->freelist;
> -       slab->freelist = NULL;
> -       slab->inuse = slab->objects;
> -       slab->frozen = 1;
> -
> -       inc_slabs_node(s, slab_nid(slab), slab->objects);
> -
> -check_new_slab:
> -
> -       if (kmem_cache_debug(s)) {
> -               /*
> -                * For debug caches here we had to go through
> -                * alloc_single_from_partial() so just store the tracking info
> -                * and return the object
> -                */
> -               if (s->flags & SLAB_STORE_USER)
> -                       set_track(s, freelist, TRACK_ALLOC, addr);
> -
> -               return freelist;
> -       }
> -
> -       if (unlikely(!pfmemalloc_match(slab, gfpflags))) {
> -               /*
> -                * For !pfmemalloc_match() case we don't load freelist so that
> -                * we don't make further mismatched allocations easier.
> -                */
> -               deactivate_slab(s, slab, get_freepointer(s, freelist));
> -               return freelist;
> -       }
> -
> -retry_load_slab:
> -
> -       local_lock_irqsave(&s->cpu_slab->lock, flags);
> -       if (unlikely(c->slab)) {
> -               void *flush_freelist = c->freelist;
> -               struct slab *flush_slab = c->slab;
> -
> -               c->slab = NULL;
> -               c->freelist = NULL;
> -               c->tid = next_tid(c->tid);
> -
> -               local_unlock_irqrestore(&s->cpu_slab->lock, flags);
> -
> -               deactivate_slab(s, flush_slab, flush_freelist);
> -
> -               stat(s, CPUSLAB_FLUSH);
> -
> -               goto retry_load_slab;
> -       }
> -       c->slab = slab;
> -
> -       goto load_freelist;
> -}
> -
> -/*
> - * A wrapper for ___slab_alloc() for contexts where preemption is not yet
> - * disabled. Compensates for possible cpu changes by refetching the per cpu area
> - * pointer.
> - */
> -static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
> -                         unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size)
> -{
> -       void *p;
> -
> -#ifdef CONFIG_PREEMPT_COUNT
> -       /*
> -        * We may have been preempted and rescheduled on a different
> -        * cpu before disabling preemption. Need to reload cpu area
> -        * pointer.
> -        */
> -       c = slub_get_cpu_ptr(s->cpu_slab);
> -#endif
> -
> -       p = ___slab_alloc(s, gfpflags, node, addr, c, orig_size);
> -#ifdef CONFIG_PREEMPT_COUNT
> -       slub_put_cpu_ptr(s->cpu_slab);
> -#endif
> -       return p;
> -}
> -
> -static __always_inline void *__slab_alloc_node(struct kmem_cache *s,
> -               gfp_t gfpflags, int node, unsigned long addr, size_t orig_size)
> -{
> -       struct kmem_cache_cpu *c;
> -       struct slab *slab;
> -       unsigned long tid;
> -       void *object;
> -
> -redo:
> -       /*
> -        * Must read kmem_cache cpu data via this cpu ptr. Preemption is
> -        * enabled. We may switch back and forth between cpus while
> -        * reading from one cpu area. That does not matter as long
> -        * as we end up on the original cpu again when doing the cmpxchg.
> -        *
> -        * We must guarantee that tid and kmem_cache_cpu are retrieved on the
> -        * same cpu. We read first the kmem_cache_cpu pointer and use it to read
> -        * the tid. If we are preempted and switched to another cpu between the
> -        * two reads, it's OK as the two are still associated with the same cpu
> -        * and cmpxchg later will validate the cpu.
> -        */
> -       c = raw_cpu_ptr(s->cpu_slab);
> -       tid = READ_ONCE(c->tid);
> -
> -       /*
> -        * Irqless object alloc/free algorithm used here depends on sequence
> -        * of fetching cpu_slab's data. tid should be fetched before anything
> -        * on c to guarantee that object and slab associated with previous tid
> -        * won't be used with current tid. If we fetch tid first, object and
> -        * slab could be one associated with next tid and our alloc/free
> -        * request will be failed. In this case, we will retry. So, no problem.
> -        */
> -       barrier();
> -
> -       /*
> -        * The transaction ids are globally unique per cpu and per operation on
> -        * a per cpu queue. Thus they can be guarantee that the cmpxchg_double
> -        * occurs on the right processor and that there was no operation on the
> -        * linked list in between.
> -        */
> -
> -       object = c->freelist;
> -       slab = c->slab;
> -
> -       if (!USE_LOCKLESS_FAST_PATH() ||
> -           unlikely(!object || !slab || !node_match(slab, node))) {
> -               object = __slab_alloc(s, gfpflags, node, addr, c, orig_size);
> -       } else {
> -               void *next_object = get_freepointer_safe(s, object);
> -
> -               /*
> -                * The cmpxchg will only match if there was no additional
> -                * operation and if we are on the right processor.
> -                *
> -                * The cmpxchg does the following atomically (without lock
> -                * semantics!)
> -                * 1. Relocate first pointer to the current per cpu area.
> -                * 2. Verify that tid and freelist have not been changed
> -                * 3. If they were not changed replace tid and freelist
> -                *
> -                * Since this is without lock semantics the protection is only
> -                * against code executing on this cpu *not* from access by
> -                * other cpus.
> -                */
> -               if (unlikely(!this_cpu_cmpxchg_double(
> -                               s->cpu_slab->freelist, s->cpu_slab->tid,
> -                               object, tid,
> -                               next_object, next_tid(tid)))) {
> -
> -                       note_cmpxchg_failure("slab_alloc", s, tid);
> -                       goto redo;
> -               }
> -               prefetch_freepointer(s, next_object);
> -               stat(s, ALLOC_FASTPATH);
> -       }
> -
> -       return object;
> -}
> -#else /* CONFIG_SLUB_TINY */
> -static void *__slab_alloc_node(struct kmem_cache *s,
> -               gfp_t gfpflags, int node, unsigned long addr, size_t orig_size)
> -{
> -       struct partial_context pc;
> -       struct slab *slab;
> -       void *object;
> -
> -       pc.flags = gfpflags;
> -       pc.slab = &slab;
> -       pc.orig_size = orig_size;
> -       object = get_partial(s, node, &pc);
> -
> -       if (object)
> -               return object;
> -
> -       slab = new_slab(s, gfpflags, node);
> -       if (unlikely(!slab)) {
> -               slab_out_of_memory(s, gfpflags, node);
> -               return NULL;
> -       }
> -
> -       object = alloc_single_from_new_slab(s, slab, orig_size);
> -
> -       return object;
> -}
> -#endif /* CONFIG_SLUB_TINY */
> -
> -/*
> - * If the object has been wiped upon free, make sure it's fully initialized by
> - * zeroing out freelist pointer.
> - */
> -static __always_inline void maybe_wipe_obj_freeptr(struct kmem_cache *s,
> -                                                  void *obj)
> -{
> -       if (unlikely(slab_want_init_on_free(s)) && obj)
> -               memset((void *)((char *)kasan_reset_tag(obj) + s->offset),
> -                       0, sizeof(void *));
> -}
> -
> -/*
> - * Inlined fastpath so that allocation functions (kmalloc, kmem_cache_alloc)
> - * have the fastpath folded into their functions. So no function call
> - * overhead for requests that can be satisfied on the fastpath.
> - *
> - * The fastpath works by first checking if the lockless freelist can be used.
> - * If not then __slab_alloc is called for slow processing.
> - *
> - * Otherwise we can simply pick the next object from the lockless free list.
> - */
> -static __fastpath_inline void *slab_alloc_node(struct kmem_cache *s, struct list_lru *lru,
> -               gfp_t gfpflags, int node, unsigned long addr, size_t orig_size)
> -{
> -       void *object;
> -       struct obj_cgroup *objcg = NULL;
> -       bool init = false;
> -
> -       s = slab_pre_alloc_hook(s, lru, &objcg, 1, gfpflags);
> -       if (!s)
> -               return NULL;
> -
> -       object = kfence_alloc(s, orig_size, gfpflags);
> -       if (unlikely(object))
> -               goto out;
> -
> -       object = __slab_alloc_node(s, gfpflags, node, addr, orig_size);
> -
> -       maybe_wipe_obj_freeptr(s, object);
> -       init = slab_want_init_on_alloc(gfpflags, s);
> -
> -out:
> -       /*
> -        * When init equals 'true', like for kzalloc() family, only
> -        * @orig_size bytes might be zeroed instead of s->object_size
> -        */
> -       slab_post_alloc_hook(s, objcg, gfpflags, 1, &object, init, orig_size);
> -
> -       return object;
> -}
> -
> -static __fastpath_inline void *slab_alloc(struct kmem_cache *s, struct list_lru *lru,
> -               gfp_t gfpflags, unsigned long addr, size_t orig_size)
> -{
> -       return slab_alloc_node(s, lru, gfpflags, NUMA_NO_NODE, addr, orig_size);
> -}
> -
> -static __fastpath_inline
> -void *__kmem_cache_alloc_lru(struct kmem_cache *s, struct list_lru *lru,
> -                            gfp_t gfpflags)
> -{
> -       void *ret = slab_alloc(s, lru, gfpflags, _RET_IP_, s->object_size);
> -
> -       trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, NUMA_NO_NODE);
> -
> -       return ret;
> -}
> -
> -void *kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags)
> -{
> -       return __kmem_cache_alloc_lru(s, NULL, gfpflags);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc);
> -
> -void *kmem_cache_alloc_lru(struct kmem_cache *s, struct list_lru *lru,
> -                          gfp_t gfpflags)
> -{
> -       return __kmem_cache_alloc_lru(s, lru, gfpflags);
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_lru);
> -
> -void *__kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags,
> -                             int node, size_t orig_size,
> -                             unsigned long caller)
> -{
> -       return slab_alloc_node(s, NULL, gfpflags, node,
> -                              caller, orig_size);
> -}
> -
> -void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node)
> -{
> -       void *ret = slab_alloc_node(s, NULL, gfpflags, node, _RET_IP_, s->object_size);
> -
> -       trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, node);
> -
> -       return ret;
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_node);
> -
> -static noinline void free_to_partial_list(
> -       struct kmem_cache *s, struct slab *slab,
> -       void *head, void *tail, int bulk_cnt,
> -       unsigned long addr)
> -{
> -       struct kmem_cache_node *n = get_node(s, slab_nid(slab));
> -       struct slab *slab_free = NULL;
> -       int cnt = bulk_cnt;
> -       unsigned long flags;
> -       depot_stack_handle_t handle = 0;
> -
> -       if (s->flags & SLAB_STORE_USER)
> -               handle = set_track_prepare();
> -
> -       spin_lock_irqsave(&n->list_lock, flags);
> -
> -       if (free_debug_processing(s, slab, head, tail, &cnt, addr, handle)) {
> -               void *prior = slab->freelist;
> -
> -               /* Perform the actual freeing while we still hold the locks */
> -               slab->inuse -= cnt;
> -               set_freepointer(s, tail, prior);
> -               slab->freelist = head;
> -
> -               /*
> -                * If the slab is empty, and node's partial list is full,
> -                * it should be discarded anyway no matter it's on full or
> -                * partial list.
> -                */
> -               if (slab->inuse == 0 && n->nr_partial >= s->min_partial)
> -                       slab_free = slab;
> -
> -               if (!prior) {
> -                       /* was on full list */
> -                       remove_full(s, n, slab);
> -                       if (!slab_free) {
> -                               add_partial(n, slab, DEACTIVATE_TO_TAIL);
> -                               stat(s, FREE_ADD_PARTIAL);
> -                       }
> -               } else if (slab_free) {
> -                       remove_partial(n, slab);
> -                       stat(s, FREE_REMOVE_PARTIAL);
> -               }
> -       }
> -
> -       if (slab_free) {
> -               /*
> -                * Update the counters while still holding n->list_lock to
> -                * prevent spurious validation warnings
> -                */
> -               dec_slabs_node(s, slab_nid(slab_free), slab_free->objects);
> -       }
> -
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -       if (slab_free) {
> -               stat(s, FREE_SLAB);
> -               free_slab(s, slab_free);
> -       }
> -}
> -
> -/*
> - * Slow path handling. This may still be called frequently since objects
> - * have a longer lifetime than the cpu slabs in most processing loads.
> - *
> - * So we still attempt to reduce cache line usage. Just take the slab
> - * lock and free the item. If there is no additional partial slab
> - * handling required then we can return immediately.
> - */
> -static void __slab_free(struct kmem_cache *s, struct slab *slab,
> -                       void *head, void *tail, int cnt,
> -                       unsigned long addr)
> -
> -{
> -       void *prior;
> -       int was_frozen;
> -       struct slab new;
> -       unsigned long counters;
> -       struct kmem_cache_node *n = NULL;
> -       unsigned long flags;
> -
> -       stat(s, FREE_SLOWPATH);
> -
> -       if (kfence_free(head))
> -               return;
> -
> -       if (IS_ENABLED(CONFIG_SLUB_TINY) || kmem_cache_debug(s)) {
> -               free_to_partial_list(s, slab, head, tail, cnt, addr);
> -               return;
> -       }
> -
> -       do {
> -               if (unlikely(n)) {
> -                       spin_unlock_irqrestore(&n->list_lock, flags);
> -                       n = NULL;
> -               }
> -               prior = slab->freelist;
> -               counters = slab->counters;
> -               set_freepointer(s, tail, prior);
> -               new.counters = counters;
> -               was_frozen = new.frozen;
> -               new.inuse -= cnt;
> -               if ((!new.inuse || !prior) && !was_frozen) {
> -
> -                       if (kmem_cache_has_cpu_partial(s) && !prior) {
> -
> -                               /*
> -                                * Slab was on no list before and will be
> -                                * partially empty
> -                                * We can defer the list move and instead
> -                                * freeze it.
> -                                */
> -                               new.frozen = 1;
> -
> -                       } else { /* Needs to be taken off a list */
> -
> -                               n = get_node(s, slab_nid(slab));
> -                               /*
> -                                * Speculatively acquire the list_lock.
> -                                * If the cmpxchg does not succeed then we may
> -                                * drop the list_lock without any processing.
> -                                *
> -                                * Otherwise the list_lock will synchronize with
> -                                * other processors updating the list of slabs.
> -                                */
> -                               spin_lock_irqsave(&n->list_lock, flags);
> -
> -                       }
> -               }
> -
> -       } while (!cmpxchg_double_slab(s, slab,
> -               prior, counters,
> -               head, new.counters,
> -               "__slab_free"));
> -
> -       if (likely(!n)) {
> -
> -               if (likely(was_frozen)) {
> -                       /*
> -                        * The list lock was not taken therefore no list
> -                        * activity can be necessary.
> -                        */
> -                       stat(s, FREE_FROZEN);
> -               } else if (new.frozen) {
> -                       /*
> -                        * If we just froze the slab then put it onto the
> -                        * per cpu partial list.
> -                        */
> -                       put_cpu_partial(s, slab, 1);
> -                       stat(s, CPU_PARTIAL_FREE);
> -               }
> -
> -               return;
> -       }
> -
> -       if (unlikely(!new.inuse && n->nr_partial >= s->min_partial))
> -               goto slab_empty;
> -
> -       /*
> -        * Objects left in the slab. If it was not on the partial list before
> -        * then add it.
> -        */
> -       if (!kmem_cache_has_cpu_partial(s) && unlikely(!prior)) {
> -               remove_full(s, n, slab);
> -               add_partial(n, slab, DEACTIVATE_TO_TAIL);
> -               stat(s, FREE_ADD_PARTIAL);
> -       }
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -       return;
> -
> -slab_empty:
> -       if (prior) {
> -               /*
> -                * Slab on the partial list.
> -                */
> -               remove_partial(n, slab);
> -               stat(s, FREE_REMOVE_PARTIAL);
> -       } else {
> -               /* Slab must be on the full list */
> -               remove_full(s, n, slab);
> -       }
> -
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -       stat(s, FREE_SLAB);
> -       discard_slab(s, slab);
> -}
> -
> -#ifndef CONFIG_SLUB_TINY
> -/*
> - * Fastpath with forced inlining to produce a kfree and kmem_cache_free that
> - * can perform fastpath freeing without additional function calls.
> - *
> - * The fastpath is only possible if we are freeing to the current cpu slab
> - * of this processor. This typically the case if we have just allocated
> - * the item before.
> - *
> - * If fastpath is not possible then fall back to __slab_free where we deal
> - * with all sorts of special processing.
> - *
> - * Bulk free of a freelist with several objects (all pointing to the
> - * same slab) possible by specifying head and tail ptr, plus objects
> - * count (cnt). Bulk free indicated by tail pointer being set.
> - */
> -static __always_inline void do_slab_free(struct kmem_cache *s,
> -                               struct slab *slab, void *head, void *tail,
> -                               int cnt, unsigned long addr)
> -{
> -       void *tail_obj = tail ? : head;
> -       struct kmem_cache_cpu *c;
> -       unsigned long tid;
> -       void **freelist;
> -
> -redo:
> -       /*
> -        * Determine the currently cpus per cpu slab.
> -        * The cpu may change afterward. However that does not matter since
> -        * data is retrieved via this pointer. If we are on the same cpu
> -        * during the cmpxchg then the free will succeed.
> -        */
> -       c = raw_cpu_ptr(s->cpu_slab);
> -       tid = READ_ONCE(c->tid);
> -
> -       /* Same with comment on barrier() in slab_alloc_node() */
> -       barrier();
> -
> -       if (unlikely(slab != c->slab)) {
> -               __slab_free(s, slab, head, tail_obj, cnt, addr);
> -               return;
> -       }
> -
> -       if (USE_LOCKLESS_FAST_PATH()) {
> -               freelist = READ_ONCE(c->freelist);
> -
> -               set_freepointer(s, tail_obj, freelist);
> -
> -               if (unlikely(!this_cpu_cmpxchg_double(
> -                               s->cpu_slab->freelist, s->cpu_slab->tid,
> -                               freelist, tid,
> -                               head, next_tid(tid)))) {
> -
> -                       note_cmpxchg_failure("slab_free", s, tid);
> -                       goto redo;
> -               }
> -       } else {
> -               /* Update the free list under the local lock */
> -               local_lock(&s->cpu_slab->lock);
> -               c = this_cpu_ptr(s->cpu_slab);
> -               if (unlikely(slab != c->slab)) {
> -                       local_unlock(&s->cpu_slab->lock);
> -                       goto redo;
> -               }
> -               tid = c->tid;
> -               freelist = c->freelist;
> -
> -               set_freepointer(s, tail_obj, freelist);
> -               c->freelist = head;
> -               c->tid = next_tid(tid);
> -
> -               local_unlock(&s->cpu_slab->lock);
> -       }
> -       stat(s, FREE_FASTPATH);
> -}
> -#else /* CONFIG_SLUB_TINY */
> -static void do_slab_free(struct kmem_cache *s,
> -                               struct slab *slab, void *head, void *tail,
> -                               int cnt, unsigned long addr)
> -{
> -       void *tail_obj = tail ? : head;
> -
> -       __slab_free(s, slab, head, tail_obj, cnt, addr);
> -}
> -#endif /* CONFIG_SLUB_TINY */
> -
> -static __fastpath_inline void slab_free(struct kmem_cache *s, struct slab *slab,
> -                                     void *head, void *tail, void **p, int cnt,
> -                                     unsigned long addr)
> -{
> -       memcg_slab_free_hook(s, slab, p, cnt);
> -       /*
> -        * With KASAN enabled slab_free_freelist_hook modifies the freelist
> -        * to remove objects, whose reuse must be delayed.
> -        */
> -       if (slab_free_freelist_hook(s, &head, &tail, &cnt))
> -               do_slab_free(s, slab, head, tail, cnt, addr);
> -}
> -
> -#ifdef CONFIG_KASAN_GENERIC
> -void ___cache_free(struct kmem_cache *cache, void *x, unsigned long addr)
> -{
> -       do_slab_free(cache, virt_to_slab(x), x, NULL, 1, addr);
> -}
> -#endif
> -
> -void __kmem_cache_free(struct kmem_cache *s, void *x, unsigned long caller)
> -{
> -       slab_free(s, virt_to_slab(x), x, NULL, &x, 1, caller);
> -}
> -
> -void kmem_cache_free(struct kmem_cache *s, void *x)
> -{
> -       s = cache_from_obj(s, x);
> -       if (!s)
> -               return;
> -       trace_kmem_cache_free(_RET_IP_, x, s);
> -       slab_free(s, virt_to_slab(x), x, NULL, &x, 1, _RET_IP_);
> -}
> -EXPORT_SYMBOL(kmem_cache_free);
> -
> -struct detached_freelist {
> -       struct slab *slab;
> -       void *tail;
> -       void *freelist;
> -       int cnt;
> -       struct kmem_cache *s;
> -};
> -
> -/*
> - * This function progressively scans the array with free objects (with
> - * a limited look ahead) and extract objects belonging to the same
> - * slab.  It builds a detached freelist directly within the given
> - * slab/objects.  This can happen without any need for
> - * synchronization, because the objects are owned by running process.
> - * The freelist is build up as a single linked list in the objects.
> - * The idea is, that this detached freelist can then be bulk
> - * transferred to the real freelist(s), but only requiring a single
> - * synchronization primitive.  Look ahead in the array is limited due
> - * to performance reasons.
> - */
> -static inline
> -int build_detached_freelist(struct kmem_cache *s, size_t size,
> -                           void **p, struct detached_freelist *df)
> -{
> -       int lookahead = 3;
> -       void *object;
> -       struct folio *folio;
> -       size_t same;
> -
> -       object = p[--size];
> -       folio = virt_to_folio(object);
> -       if (!s) {
> -               /* Handle kalloc'ed objects */
> -               if (unlikely(!folio_test_slab(folio))) {
> -                       free_large_kmalloc(folio, object);
> -                       df->slab = NULL;
> -                       return size;
> -               }
> -               /* Derive kmem_cache from object */
> -               df->slab = folio_slab(folio);
> -               df->s = df->slab->slab_cache;
> -       } else {
> -               df->slab = folio_slab(folio);
> -               df->s = cache_from_obj(s, object); /* Support for memcg */
> -       }
> -
> -       /* Start new detached freelist */
> -       df->tail = object;
> -       df->freelist = object;
> -       df->cnt = 1;
> -
> -       if (is_kfence_address(object))
> -               return size;
> -
> -       set_freepointer(df->s, object, NULL);
> -
> -       same = size;
> -       while (size) {
> -               object = p[--size];
> -               /* df->slab is always set at this point */
> -               if (df->slab == virt_to_slab(object)) {
> -                       /* Opportunity build freelist */
> -                       set_freepointer(df->s, object, df->freelist);
> -                       df->freelist = object;
> -                       df->cnt++;
> -                       same--;
> -                       if (size != same)
> -                               swap(p[size], p[same]);
> -                       continue;
> -               }
> -
> -               /* Limit look ahead search */
> -               if (!--lookahead)
> -                       break;
> -       }
> -
> -       return same;
> -}
> -
> -/* Note that interrupts must be enabled when calling this function. */
> -void kmem_cache_free_bulk(struct kmem_cache *s, size_t size, void **p)
> -{
> -       if (!size)
> -               return;
> -
> -       do {
> -               struct detached_freelist df;
> -
> -               size = build_detached_freelist(s, size, p, &df);
> -               if (!df.slab)
> -                       continue;
> -
> -               slab_free(df.s, df.slab, df.freelist, df.tail, &p[size], df.cnt,
> -                         _RET_IP_);
> -       } while (likely(size));
> -}
> -EXPORT_SYMBOL(kmem_cache_free_bulk);
> -
> -#ifndef CONFIG_SLUB_TINY
> -static inline int __kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags,
> -                       size_t size, void **p, struct obj_cgroup *objcg)
> -{
> -       struct kmem_cache_cpu *c;
> -       unsigned long irqflags;
> -       int i;
> -
> -       /*
> -        * Drain objects in the per cpu slab, while disabling local
> -        * IRQs, which protects against PREEMPT and interrupts
> -        * handlers invoking normal fastpath.
> -        */
> -       c = slub_get_cpu_ptr(s->cpu_slab);
> -       local_lock_irqsave(&s->cpu_slab->lock, irqflags);
> -
> -       for (i = 0; i < size; i++) {
> -               void *object = kfence_alloc(s, s->object_size, flags);
> -
> -               if (unlikely(object)) {
> -                       p[i] = object;
> -                       continue;
> -               }
> -
> -               object = c->freelist;
> -               if (unlikely(!object)) {
> -                       /*
> -                        * We may have removed an object from c->freelist using
> -                        * the fastpath in the previous iteration; in that case,
> -                        * c->tid has not been bumped yet.
> -                        * Since ___slab_alloc() may reenable interrupts while
> -                        * allocating memory, we should bump c->tid now.
> -                        */
> -                       c->tid = next_tid(c->tid);
> -
> -                       local_unlock_irqrestore(&s->cpu_slab->lock, irqflags);
> -
> -                       /*
> -                        * Invoking slow path likely have side-effect
> -                        * of re-populating per CPU c->freelist
> -                        */
> -                       p[i] = ___slab_alloc(s, flags, NUMA_NO_NODE,
> -                                           _RET_IP_, c, s->object_size);
> -                       if (unlikely(!p[i]))
> -                               goto error;
> -
> -                       c = this_cpu_ptr(s->cpu_slab);
> -                       maybe_wipe_obj_freeptr(s, p[i]);
> -
> -                       local_lock_irqsave(&s->cpu_slab->lock, irqflags);
> -
> -                       continue; /* goto for-loop */
> -               }
> -               c->freelist = get_freepointer(s, object);
> -               p[i] = object;
> -               maybe_wipe_obj_freeptr(s, p[i]);
> -       }
> -       c->tid = next_tid(c->tid);
> -       local_unlock_irqrestore(&s->cpu_slab->lock, irqflags);
> -       slub_put_cpu_ptr(s->cpu_slab);
> -
> -       return i;
> -
> -error:
> -       slub_put_cpu_ptr(s->cpu_slab);
> -       slab_post_alloc_hook(s, objcg, flags, i, p, false, s->object_size);
> -       kmem_cache_free_bulk(s, i, p);
> -       return 0;
> -
> -}
> -#else /* CONFIG_SLUB_TINY */
> -static int __kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags,
> -                       size_t size, void **p, struct obj_cgroup *objcg)
> -{
> -       int i;
> -
> -       for (i = 0; i < size; i++) {
> -               void *object = kfence_alloc(s, s->object_size, flags);
> -
> -               if (unlikely(object)) {
> -                       p[i] = object;
> -                       continue;
> -               }
> -
> -               p[i] = __slab_alloc_node(s, flags, NUMA_NO_NODE,
> -                                        _RET_IP_, s->object_size);
> -               if (unlikely(!p[i]))
> -                       goto error;
> -
> -               maybe_wipe_obj_freeptr(s, p[i]);
> -       }
> -
> -       return i;
> -
> -error:
> -       slab_post_alloc_hook(s, objcg, flags, i, p, false, s->object_size);
> -       kmem_cache_free_bulk(s, i, p);
> -       return 0;
> -}
> -#endif /* CONFIG_SLUB_TINY */
> -
> -/* Note that interrupts must be enabled when calling this function. */
> -int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> -                         void **p)
> -{
> -       int i;
> -       struct obj_cgroup *objcg = NULL;
> -
> -       if (!size)
> -               return 0;
> -
> -       /* memcg and kmem_cache debug support */
> -       s = slab_pre_alloc_hook(s, NULL, &objcg, size, flags);
> -       if (unlikely(!s))
> -               return 0;
> -
> -       i = __kmem_cache_alloc_bulk(s, flags, size, p, objcg);
> -
> -       /*
> -        * memcg and kmem_cache debug support and memory initialization.
> -        * Done outside of the IRQ disabled fastpath loop.
> -        */
> -       if (i != 0)
> -               slab_post_alloc_hook(s, objcg, flags, size, p,
> -                       slab_want_init_on_alloc(flags, s), s->object_size);
> -       return i;
> -}
> -EXPORT_SYMBOL(kmem_cache_alloc_bulk);
> -
> -
> -/*
> - * Object placement in a slab is made very easy because we always start at
> - * offset 0. If we tune the size of the object to the alignment then we can
> - * get the required alignment by putting one properly sized object after
> - * another.
> - *
> - * Notice that the allocation order determines the sizes of the per cpu
> - * caches. Each processor has always one slab available for allocations.
> - * Increasing the allocation order reduces the number of times that slabs
> - * must be moved on and off the partial lists and is therefore a factor in
> - * locking overhead.
> - */
> -
> -/*
> - * Minimum / Maximum order of slab pages. This influences locking overhead
> - * and slab fragmentation. A higher order reduces the number of partial slabs
> - * and increases the number of allocations possible without having to
> - * take the list_lock.
> - */
> -static unsigned int slub_min_order;
> -static unsigned int slub_max_order =
> -       IS_ENABLED(CONFIG_SLUB_TINY) ? 1 : PAGE_ALLOC_COSTLY_ORDER;
> -static unsigned int slub_min_objects;
> -
> -/*
> - * Calculate the order of allocation given an slab object size.
> - *
> - * The order of allocation has significant impact on performance and other
> - * system components. Generally order 0 allocations should be preferred since
> - * order 0 does not cause fragmentation in the page allocator. Larger objects
> - * be problematic to put into order 0 slabs because there may be too much
> - * unused space left. We go to a higher order if more than 1/16th of the slab
> - * would be wasted.
> - *
> - * In order to reach satisfactory performance we must ensure that a minimum
> - * number of objects is in one slab. Otherwise we may generate too much
> - * activity on the partial lists which requires taking the list_lock. This is
> - * less a concern for large slabs though which are rarely used.
> - *
> - * slub_max_order specifies the order where we begin to stop considering the
> - * number of objects in a slab as critical. If we reach slub_max_order then
> - * we try to keep the page order as low as possible. So we accept more waste
> - * of space in favor of a small page order.
> - *
> - * Higher order allocations also allow the placement of more objects in a
> - * slab and thereby reduce object handling overhead. If the user has
> - * requested a higher minimum order then we start with that one instead of
> - * the smallest order which will fit the object.
> - */
> -static inline unsigned int calc_slab_order(unsigned int size,
> -               unsigned int min_objects, unsigned int max_order,
> -               unsigned int fract_leftover)
> -{
> -       unsigned int min_order = slub_min_order;
> -       unsigned int order;
> -
> -       if (order_objects(min_order, size) > MAX_OBJS_PER_PAGE)
> -               return get_order(size * MAX_OBJS_PER_PAGE) - 1;
> -
> -       for (order = max(min_order, (unsigned int)get_order(min_objects * size));
> -                       order <= max_order; order++) {
> -
> -               unsigned int slab_size = (unsigned int)PAGE_SIZE << order;
> -               unsigned int rem;
> -
> -               rem = slab_size % size;
> -
> -               if (rem <= slab_size / fract_leftover)
> -                       break;
> -       }
> -
> -       return order;
> -}
> -
> -static inline int calculate_order(unsigned int size)
> -{
> -       unsigned int order;
> -       unsigned int min_objects;
> -       unsigned int max_objects;
> -       unsigned int nr_cpus;
> -
> -       /*
> -        * Attempt to find best configuration for a slab. This
> -        * works by first attempting to generate a layout with
> -        * the best configuration and backing off gradually.
> -        *
> -        * First we increase the acceptable waste in a slab. Then
> -        * we reduce the minimum objects required in a slab.
> -        */
> -       min_objects = slub_min_objects;
> -       if (!min_objects) {
> -               /*
> -                * Some architectures will only update present cpus when
> -                * onlining them, so don't trust the number if it's just 1. But
> -                * we also don't want to use nr_cpu_ids always, as on some other
> -                * architectures, there can be many possible cpus, but never
> -                * onlined. Here we compromise between trying to avoid too high
> -                * order on systems that appear larger than they are, and too
> -                * low order on systems that appear smaller than they are.
> -                */
> -               nr_cpus = num_present_cpus();
> -               if (nr_cpus <= 1)
> -                       nr_cpus = nr_cpu_ids;
> -               min_objects = 4 * (fls(nr_cpus) + 1);
> -       }
> -       max_objects = order_objects(slub_max_order, size);
> -       min_objects = min(min_objects, max_objects);
> -
> -       while (min_objects > 1) {
> -               unsigned int fraction;
> -
> -               fraction = 16;
> -               while (fraction >= 4) {
> -                       order = calc_slab_order(size, min_objects,
> -                                       slub_max_order, fraction);
> -                       if (order <= slub_max_order)
> -                               return order;
> -                       fraction /= 2;
> -               }
> -               min_objects--;
> -       }
> -
> -       /*
> -        * We were unable to place multiple objects in a slab. Now
> -        * lets see if we can place a single object there.
> -        */
> -       order = calc_slab_order(size, 1, slub_max_order, 1);
> -       if (order <= slub_max_order)
> -               return order;
> -
> -       /*
> -        * Doh this slab cannot be placed using slub_max_order.
> -        */
> -       order = calc_slab_order(size, 1, MAX_ORDER, 1);
> -       if (order < MAX_ORDER)
> -               return order;
> -       return -ENOSYS;
> -}
> -
> -static void
> -init_kmem_cache_node(struct kmem_cache_node *n)
> -{
> -       n->nr_partial = 0;
> -       spin_lock_init(&n->list_lock);
> -       INIT_LIST_HEAD(&n->partial);
> -#ifdef CONFIG_SLUB_DEBUG
> -       atomic_long_set(&n->nr_slabs, 0);
> -       atomic_long_set(&n->total_objects, 0);
> -       INIT_LIST_HEAD(&n->full);
> -#endif
> -}
> -
> -#ifndef CONFIG_SLUB_TINY
> -static inline int alloc_kmem_cache_cpus(struct kmem_cache *s)
> -{
> -       BUILD_BUG_ON(PERCPU_DYNAMIC_EARLY_SIZE <
> -                       NR_KMALLOC_TYPES * KMALLOC_SHIFT_HIGH *
> -                       sizeof(struct kmem_cache_cpu));
> -
> -       /*
> -        * Must align to double word boundary for the double cmpxchg
> -        * instructions to work; see __pcpu_double_call_return_bool().
> -        */
> -       s->cpu_slab = __alloc_percpu(sizeof(struct kmem_cache_cpu),
> -                                    2 * sizeof(void *));
> -
> -       if (!s->cpu_slab)
> -               return 0;
> -
> -       init_kmem_cache_cpus(s);
> -
> -       return 1;
> -}
> -#else
> -static inline int alloc_kmem_cache_cpus(struct kmem_cache *s)
> -{
> -       return 1;
> -}
> -#endif /* CONFIG_SLUB_TINY */
> -
> -static struct kmem_cache *kmem_cache_node;
> -
> -/*
> - * No kmalloc_node yet so do it by hand. We know that this is the first
> - * slab on the node for this slabcache. There are no concurrent accesses
> - * possible.
> - *
> - * Note that this function only works on the kmem_cache_node
> - * when allocating for the kmem_cache_node. This is used for bootstrapping
> - * memory on a fresh node that has no slab structures yet.
> - */
> -static void early_kmem_cache_node_alloc(int node)
> -{
> -       struct slab *slab;
> -       struct kmem_cache_node *n;
> -
> -       BUG_ON(kmem_cache_node->size < sizeof(struct kmem_cache_node));
> -
> -       slab = new_slab(kmem_cache_node, GFP_NOWAIT, node);
> -
> -       BUG_ON(!slab);
> -       inc_slabs_node(kmem_cache_node, slab_nid(slab), slab->objects);
> -       if (slab_nid(slab) != node) {
> -               pr_err("SLUB: Unable to allocate memory from node %d\n", node);
> -               pr_err("SLUB: Allocating a useless per node structure in order to be able to continue\n");
> -       }
> -
> -       n = slab->freelist;
> -       BUG_ON(!n);
> -#ifdef CONFIG_SLUB_DEBUG
> -       init_object(kmem_cache_node, n, SLUB_RED_ACTIVE);
> -       init_tracking(kmem_cache_node, n);
> -#endif
> -       n = kasan_slab_alloc(kmem_cache_node, n, GFP_KERNEL, false);
> -       slab->freelist = get_freepointer(kmem_cache_node, n);
> -       slab->inuse = 1;
> -       kmem_cache_node->node[node] = n;
> -       init_kmem_cache_node(n);
> -       inc_slabs_node(kmem_cache_node, node, slab->objects);
> -
> -       /*
> -        * No locks need to be taken here as it has just been
> -        * initialized and there is no concurrent access.
> -        */
> -       __add_partial(n, slab, DEACTIVATE_TO_HEAD);
> -}
> -
> -static void free_kmem_cache_nodes(struct kmem_cache *s)
> -{
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_kmem_cache_node(s, node, n) {
> -               s->node[node] = NULL;
> -               kmem_cache_free(kmem_cache_node, n);
> -       }
> -}
> -
> -void __kmem_cache_release(struct kmem_cache *s)
> -{
> -       cache_random_seq_destroy(s);
> -#ifndef CONFIG_SLUB_TINY
> -       free_percpu(s->cpu_slab);
> -#endif
> -       free_kmem_cache_nodes(s);
> -}
> -
> -static int init_kmem_cache_nodes(struct kmem_cache *s)
> -{
> -       int node;
> -
> -       for_each_node_mask(node, slab_nodes) {
> -               struct kmem_cache_node *n;
> -
> -               if (slab_state == DOWN) {
> -                       early_kmem_cache_node_alloc(node);
> -                       continue;
> -               }
> -               n = kmem_cache_alloc_node(kmem_cache_node,
> -                                               GFP_KERNEL, node);
> -
> -               if (!n) {
> -                       free_kmem_cache_nodes(s);
> -                       return 0;
> -               }
> -
> -               init_kmem_cache_node(n);
> -               s->node[node] = n;
> -       }
> -       return 1;
> -}
> -
> -static void set_cpu_partial(struct kmem_cache *s)
> -{
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       unsigned int nr_objects;
> -
> -       /*
> -        * cpu_partial determined the maximum number of objects kept in the
> -        * per cpu partial lists of a processor.
> -        *
> -        * Per cpu partial lists mainly contain slabs that just have one
> -        * object freed. If they are used for allocation then they can be
> -        * filled up again with minimal effort. The slab will never hit the
> -        * per node partial lists and therefore no locking will be required.
> -        *
> -        * For backwards compatibility reasons, this is determined as number
> -        * of objects, even though we now limit maximum number of pages, see
> -        * slub_set_cpu_partial()
> -        */
> -       if (!kmem_cache_has_cpu_partial(s))
> -               nr_objects = 0;
> -       else if (s->size >= PAGE_SIZE)
> -               nr_objects = 6;
> -       else if (s->size >= 1024)
> -               nr_objects = 24;
> -       else if (s->size >= 256)
> -               nr_objects = 52;
> -       else
> -               nr_objects = 120;
> -
> -       slub_set_cpu_partial(s, nr_objects);
> -#endif
> -}
> -
> -/*
> - * calculate_sizes() determines the order and the distribution of data within
> - * a slab object.
> - */
> -static int calculate_sizes(struct kmem_cache *s)
> -{
> -       slab_flags_t flags = s->flags;
> -       unsigned int size = s->object_size;
> -       unsigned int order;
> -
> -       /*
> -        * Round up object size to the next word boundary. We can only
> -        * place the free pointer at word boundaries and this determines
> -        * the possible location of the free pointer.
> -        */
> -       size = ALIGN(size, sizeof(void *));
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -       /*
> -        * Determine if we can poison the object itself. If the user of
> -        * the slab may touch the object after free or before allocation
> -        * then we should never poison the object itself.
> -        */
> -       if ((flags & SLAB_POISON) && !(flags & SLAB_TYPESAFE_BY_RCU) &&
> -                       !s->ctor)
> -               s->flags |= __OBJECT_POISON;
> -       else
> -               s->flags &= ~__OBJECT_POISON;
> -
> -
> -       /*
> -        * If we are Redzoning then check if there is some space between the
> -        * end of the object and the free pointer. If not then add an
> -        * additional word to have some bytes to store Redzone information.
> -        */
> -       if ((flags & SLAB_RED_ZONE) && size == s->object_size)
> -               size += sizeof(void *);
> -#endif
> -
> -       /*
> -        * With that we have determined the number of bytes in actual use
> -        * by the object and redzoning.
> -        */
> -       s->inuse = size;
> -
> -       if (slub_debug_orig_size(s) ||
> -           (flags & (SLAB_TYPESAFE_BY_RCU | SLAB_POISON)) ||
> -           ((flags & SLAB_RED_ZONE) && s->object_size < sizeof(void *)) ||
> -           s->ctor) {
> -               /*
> -                * Relocate free pointer after the object if it is not
> -                * permitted to overwrite the first word of the object on
> -                * kmem_cache_free.
> -                *
> -                * This is the case if we do RCU, have a constructor or
> -                * destructor, are poisoning the objects, or are
> -                * redzoning an object smaller than sizeof(void *).
> -                *
> -                * The assumption that s->offset >= s->inuse means free
> -                * pointer is outside of the object is used in the
> -                * freeptr_outside_object() function. If that is no
> -                * longer true, the function needs to be modified.
> -                */
> -               s->offset = size;
> -               size += sizeof(void *);
> -       } else {
> -               /*
> -                * Store freelist pointer near middle of object to keep
> -                * it away from the edges of the object to avoid small
> -                * sized over/underflows from neighboring allocations.
> -                */
> -               s->offset = ALIGN_DOWN(s->object_size / 2, sizeof(void *));
> -       }
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -       if (flags & SLAB_STORE_USER) {
> -               /*
> -                * Need to store information about allocs and frees after
> -                * the object.
> -                */
> -               size += 2 * sizeof(struct track);
> -
> -               /* Save the original kmalloc request size */
> -               if (flags & SLAB_KMALLOC)
> -                       size += sizeof(unsigned int);
> -       }
> -#endif
> -
> -       kasan_cache_create(s, &size, &s->flags);
> -#ifdef CONFIG_SLUB_DEBUG
> -       if (flags & SLAB_RED_ZONE) {
> -               /*
> -                * Add some empty padding so that we can catch
> -                * overwrites from earlier objects rather than let
> -                * tracking information or the free pointer be
> -                * corrupted if a user writes before the start
> -                * of the object.
> -                */
> -               size += sizeof(void *);
> -
> -               s->red_left_pad = sizeof(void *);
> -               s->red_left_pad = ALIGN(s->red_left_pad, s->align);
> -               size += s->red_left_pad;
> -       }
> -#endif
> -
> -       /*
> -        * SLUB stores one object immediately after another beginning from
> -        * offset 0. In order to align the objects we have to simply size
> -        * each object to conform to the alignment.
> -        */
> -       size = ALIGN(size, s->align);
> -       s->size = size;
> -       s->reciprocal_size = reciprocal_value(size);
> -       order = calculate_order(size);
> -
> -       if ((int)order < 0)
> -               return 0;
> -
> -       s->allocflags = 0;
> -       if (order)
> -               s->allocflags |= __GFP_COMP;
> -
> -       if (s->flags & SLAB_CACHE_DMA)
> -               s->allocflags |= GFP_DMA;
> -
> -       if (s->flags & SLAB_CACHE_DMA32)
> -               s->allocflags |= GFP_DMA32;
> -
> -       if (s->flags & SLAB_RECLAIM_ACCOUNT)
> -               s->allocflags |= __GFP_RECLAIMABLE;
> -
> -       /*
> -        * Determine the number of objects per slab
> -        */
> -       s->oo = oo_make(order, size);
> -       s->min = oo_make(get_order(size), size);
> -
> -       return !!oo_objects(s->oo);
> -}
> -
> -static int kmem_cache_open(struct kmem_cache *s, slab_flags_t flags)
> -{
> -       s->flags = kmem_cache_flags(s->size, flags, s->name);
> -#ifdef CONFIG_SLAB_FREELIST_HARDENED
> -       s->random = get_random_long();
> -#endif
> -
> -       if (!calculate_sizes(s))
> -               goto error;
> -       if (disable_higher_order_debug) {
> -               /*
> -                * Disable debugging flags that store metadata if the min slab
> -                * order increased.
> -                */
> -               if (get_order(s->size) > get_order(s->object_size)) {
> -                       s->flags &= ~DEBUG_METADATA_FLAGS;
> -                       s->offset = 0;
> -                       if (!calculate_sizes(s))
> -                               goto error;
> -               }
> -       }
> -
> -#if defined(CONFIG_HAVE_CMPXCHG_DOUBLE) && \
> -    defined(CONFIG_HAVE_ALIGNED_STRUCT_PAGE)
> -       if (system_has_cmpxchg_double() && (s->flags & SLAB_NO_CMPXCHG) == 0)
> -               /* Enable fast mode */
> -               s->flags |= __CMPXCHG_DOUBLE;
> -#endif
> -
> -       /*
> -        * The larger the object size is, the more slabs we want on the partial
> -        * list to avoid pounding the page allocator excessively.
> -        */
> -       s->min_partial = min_t(unsigned long, MAX_PARTIAL, ilog2(s->size) / 2);
> -       s->min_partial = max_t(unsigned long, MIN_PARTIAL, s->min_partial);
> -
> -       set_cpu_partial(s);
> -
> -#ifdef CONFIG_NUMA
> -       s->remote_node_defrag_ratio = 1000;
> -#endif
> -
> -       /* Initialize the pre-computed randomized freelist if slab is up */
> -       if (slab_state >= UP) {
> -               if (init_cache_random_seq(s))
> -                       goto error;
> -       }
> -
> -       if (!init_kmem_cache_nodes(s))
> -               goto error;
> -
> -       if (alloc_kmem_cache_cpus(s))
> -               return 0;
> -
> -error:
> -       __kmem_cache_release(s);
> -       return -EINVAL;
> -}
> -
> -static void list_slab_objects(struct kmem_cache *s, struct slab *slab,
> -                             const char *text)
> -{
> -#ifdef CONFIG_SLUB_DEBUG
> -       void *addr = slab_address(slab);
> -       void *p;
> -
> -       slab_err(s, slab, text, s->name);
> -
> -       spin_lock(&object_map_lock);
> -       __fill_map(object_map, s, slab);
> -
> -       for_each_object(p, s, addr, slab->objects) {
> -
> -               if (!test_bit(__obj_to_index(s, addr, p), object_map)) {
> -                       pr_err("Object 0x%p @offset=%tu\n", p, p - addr);
> -                       print_tracking(s, p);
> -               }
> -       }
> -       spin_unlock(&object_map_lock);
> -#endif
> -}
> -
> -/*
> - * Attempt to free all partial slabs on a node.
> - * This is called from __kmem_cache_shutdown(). We must take list_lock
> - * because sysfs file might still access partial list after the shutdowning.
> - */
> -static void free_partial(struct kmem_cache *s, struct kmem_cache_node *n)
> -{
> -       LIST_HEAD(discard);
> -       struct slab *slab, *h;
> -
> -       BUG_ON(irqs_disabled());
> -       spin_lock_irq(&n->list_lock);
> -       list_for_each_entry_safe(slab, h, &n->partial, slab_list) {
> -               if (!slab->inuse) {
> -                       remove_partial(n, slab);
> -                       list_add(&slab->slab_list, &discard);
> -               } else {
> -                       list_slab_objects(s, slab,
> -                         "Objects remaining in %s on __kmem_cache_shutdown()");
> -               }
> -       }
> -       spin_unlock_irq(&n->list_lock);
> -
> -       list_for_each_entry_safe(slab, h, &discard, slab_list)
> -               discard_slab(s, slab);
> -}
> -
> -bool __kmem_cache_empty(struct kmem_cache *s)
> -{
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_kmem_cache_node(s, node, n)
> -               if (n->nr_partial || slabs_node(s, node))
> -                       return false;
> -       return true;
> -}
> -
> -/*
> - * Release all resources used by a slab cache.
> - */
> -int __kmem_cache_shutdown(struct kmem_cache *s)
> -{
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       flush_all_cpus_locked(s);
> -       /* Attempt to free all objects */
> -       for_each_kmem_cache_node(s, node, n) {
> -               free_partial(s, n);
> -               if (n->nr_partial || slabs_node(s, node))
> -                       return 1;
> -       }
> -       return 0;
> -}
> -
> -#ifdef CONFIG_PRINTK
> -void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> -{
> -       void *base;
> -       int __maybe_unused i;
> -       unsigned int objnr;
> -       void *objp;
> -       void *objp0;
> -       struct kmem_cache *s = slab->slab_cache;
> -       struct track __maybe_unused *trackp;
> -
> -       kpp->kp_ptr = object;
> -       kpp->kp_slab = slab;
> -       kpp->kp_slab_cache = s;
> -       base = slab_address(slab);
> -       objp0 = kasan_reset_tag(object);
> -#ifdef CONFIG_SLUB_DEBUG
> -       objp = restore_red_left(s, objp0);
> -#else
> -       objp = objp0;
> -#endif
> -       objnr = obj_to_index(s, slab, objp);
> -       kpp->kp_data_offset = (unsigned long)((char *)objp0 - (char *)objp);
> -       objp = base + s->size * objnr;
> -       kpp->kp_objp = objp;
> -       if (WARN_ON_ONCE(objp < base || objp >= base + slab->objects * s->size
> -                        || (objp - base) % s->size) ||
> -           !(s->flags & SLAB_STORE_USER))
> -               return;
> -#ifdef CONFIG_SLUB_DEBUG
> -       objp = fixup_red_left(s, objp);
> -       trackp = get_track(s, objp, TRACK_ALLOC);
> -       kpp->kp_ret = (void *)trackp->addr;
> -#ifdef CONFIG_STACKDEPOT
> -       {
> -               depot_stack_handle_t handle;
> -               unsigned long *entries;
> -               unsigned int nr_entries;
> -
> -               handle = READ_ONCE(trackp->handle);
> -               if (handle) {
> -                       nr_entries = stack_depot_fetch(handle, &entries);
> -                       for (i = 0; i < KS_ADDRS_COUNT && i < nr_entries; i++)
> -                               kpp->kp_stack[i] = (void *)entries[i];
> -               }
> -
> -               trackp = get_track(s, objp, TRACK_FREE);
> -               handle = READ_ONCE(trackp->handle);
> -               if (handle) {
> -                       nr_entries = stack_depot_fetch(handle, &entries);
> -                       for (i = 0; i < KS_ADDRS_COUNT && i < nr_entries; i++)
> -                               kpp->kp_free_stack[i] = (void *)entries[i];
> -               }
> -       }
> -#endif
> -#endif
> -}
> -#endif
> -
> -/********************************************************************
> - *             Kmalloc subsystem
> - *******************************************************************/
> -
> -static int __init setup_slub_min_order(char *str)
> -{
> -       get_option(&str, (int *)&slub_min_order);
> -
> -       return 1;
> -}
> -
> -__setup("slub_min_order=", setup_slub_min_order);
> -
> -static int __init setup_slub_max_order(char *str)
> -{
> -       get_option(&str, (int *)&slub_max_order);
> -       slub_max_order = min(slub_max_order, (unsigned int)MAX_ORDER - 1);
> -
> -       return 1;
> -}
> -
> -__setup("slub_max_order=", setup_slub_max_order);
> -
> -static int __init setup_slub_min_objects(char *str)
> -{
> -       get_option(&str, (int *)&slub_min_objects);
> -
> -       return 1;
> -}
> -
> -__setup("slub_min_objects=", setup_slub_min_objects);
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -/*
> - * Rejects incorrectly sized objects and objects that are to be copied
> - * to/from userspace but do not fall entirely within the containing slab
> - * cache's usercopy region.
> - *
> - * Returns NULL if check passes, otherwise const char * to name of cache
> - * to indicate an error.
> - */
> -void __check_heap_object(const void *ptr, unsigned long n,
> -                        const struct slab *slab, bool to_user)
> -{
> -       struct kmem_cache *s;
> -       unsigned int offset;
> -       bool is_kfence = is_kfence_address(ptr);
> -
> -       ptr = kasan_reset_tag(ptr);
> -
> -       /* Find object and usable object size. */
> -       s = slab->slab_cache;
> -
> -       /* Reject impossible pointers. */
> -       if (ptr < slab_address(slab))
> -               usercopy_abort("SLUB object not in SLUB page?!", NULL,
> -                              to_user, 0, n);
> -
> -       /* Find offset within object. */
> -       if (is_kfence)
> -               offset = ptr - kfence_object_start(ptr);
> -       else
> -               offset = (ptr - slab_address(slab)) % s->size;
> -
> -       /* Adjust for redzone and reject if within the redzone. */
> -       if (!is_kfence && kmem_cache_debug_flags(s, SLAB_RED_ZONE)) {
> -               if (offset < s->red_left_pad)
> -                       usercopy_abort("SLUB object in left red zone",
> -                                      s->name, to_user, offset, n);
> -               offset -= s->red_left_pad;
> -       }
> -
> -       /* Allow address range falling entirely within usercopy region. */
> -       if (offset >= s->useroffset &&
> -           offset - s->useroffset <= s->usersize &&
> -           n <= s->useroffset - offset + s->usersize)
> -               return;
> -
> -       usercopy_abort("SLUB object", s->name, to_user, offset, n);
> -}
> -#endif /* CONFIG_HARDENED_USERCOPY */
> -
> -#define SHRINK_PROMOTE_MAX 32
> -
> -/*
> - * kmem_cache_shrink discards empty slabs and promotes the slabs filled
> - * up most to the head of the partial lists. New allocations will then
> - * fill those up and thus they can be removed from the partial lists.
> - *
> - * The slabs with the least items are placed last. This results in them
> - * being allocated from last increasing the chance that the last objects
> - * are freed in them.
> - */
> -static int __kmem_cache_do_shrink(struct kmem_cache *s)
> -{
> -       int node;
> -       int i;
> -       struct kmem_cache_node *n;
> -       struct slab *slab;
> -       struct slab *t;
> -       struct list_head discard;
> -       struct list_head promote[SHRINK_PROMOTE_MAX];
> -       unsigned long flags;
> -       int ret = 0;
> -
> -       for_each_kmem_cache_node(s, node, n) {
> -               INIT_LIST_HEAD(&discard);
> -               for (i = 0; i < SHRINK_PROMOTE_MAX; i++)
> -                       INIT_LIST_HEAD(promote + i);
> -
> -               spin_lock_irqsave(&n->list_lock, flags);
> -
> -               /*
> -                * Build lists of slabs to discard or promote.
> -                *
> -                * Note that concurrent frees may occur while we hold the
> -                * list_lock. slab->inuse here is the upper limit.
> -                */
> -               list_for_each_entry_safe(slab, t, &n->partial, slab_list) {
> -                       int free = slab->objects - slab->inuse;
> -
> -                       /* Do not reread slab->inuse */
> -                       barrier();
> -
> -                       /* We do not keep full slabs on the list */
> -                       BUG_ON(free <= 0);
> -
> -                       if (free == slab->objects) {
> -                               list_move(&slab->slab_list, &discard);
> -                               n->nr_partial--;
> -                               dec_slabs_node(s, node, slab->objects);
> -                       } else if (free <= SHRINK_PROMOTE_MAX)
> -                               list_move(&slab->slab_list, promote + free - 1);
> -               }
> -
> -               /*
> -                * Promote the slabs filled up most to the head of the
> -                * partial list.
> -                */
> -               for (i = SHRINK_PROMOTE_MAX - 1; i >= 0; i--)
> -                       list_splice(promote + i, &n->partial);
> -
> -               spin_unlock_irqrestore(&n->list_lock, flags);
> -
> -               /* Release empty slabs */
> -               list_for_each_entry_safe(slab, t, &discard, slab_list)
> -                       free_slab(s, slab);
> -
> -               if (slabs_node(s, node))
> -                       ret = 1;
> -       }
> -
> -       return ret;
> -}
> -
> -int __kmem_cache_shrink(struct kmem_cache *s)
> -{
> -       flush_all(s);
> -       return __kmem_cache_do_shrink(s);
> -}
> -
> -static int slab_mem_going_offline_callback(void *arg)
> -{
> -       struct kmem_cache *s;
> -
> -       mutex_lock(&slab_mutex);
> -       list_for_each_entry(s, &slab_caches, list) {
> -               flush_all_cpus_locked(s);
> -               __kmem_cache_do_shrink(s);
> -       }
> -       mutex_unlock(&slab_mutex);
> -
> -       return 0;
> -}
> -
> -static void slab_mem_offline_callback(void *arg)
> -{
> -       struct memory_notify *marg = arg;
> -       int offline_node;
> -
> -       offline_node = marg->status_change_nid_normal;
> -
> -       /*
> -        * If the node still has available memory. we need kmem_cache_node
> -        * for it yet.
> -        */
> -       if (offline_node < 0)
> -               return;
> -
> -       mutex_lock(&slab_mutex);
> -       node_clear(offline_node, slab_nodes);
> -       /*
> -        * We no longer free kmem_cache_node structures here, as it would be
> -        * racy with all get_node() users, and infeasible to protect them with
> -        * slab_mutex.
> -        */
> -       mutex_unlock(&slab_mutex);
> -}
> -
> -static int slab_mem_going_online_callback(void *arg)
> -{
> -       struct kmem_cache_node *n;
> -       struct kmem_cache *s;
> -       struct memory_notify *marg = arg;
> -       int nid = marg->status_change_nid_normal;
> -       int ret = 0;
> -
> -       /*
> -        * If the node's memory is already available, then kmem_cache_node is
> -        * already created. Nothing to do.
> -        */
> -       if (nid < 0)
> -               return 0;
> -
> -       /*
> -        * We are bringing a node online. No memory is available yet. We must
> -        * allocate a kmem_cache_node structure in order to bring the node
> -        * online.
> -        */
> -       mutex_lock(&slab_mutex);
> -       list_for_each_entry(s, &slab_caches, list) {
> -               /*
> -                * The structure may already exist if the node was previously
> -                * onlined and offlined.
> -                */
> -               if (get_node(s, nid))
> -                       continue;
> -               /*
> -                * XXX: kmem_cache_alloc_node will fallback to other nodes
> -                *      since memory is not yet available from the node that
> -                *      is brought up.
> -                */
> -               n = kmem_cache_alloc(kmem_cache_node, GFP_KERNEL);
> -               if (!n) {
> -                       ret = -ENOMEM;
> -                       goto out;
> -               }
> -               init_kmem_cache_node(n);
> -               s->node[nid] = n;
> -       }
> -       /*
> -        * Any cache created after this point will also have kmem_cache_node
> -        * initialized for the new node.
> -        */
> -       node_set(nid, slab_nodes);
> -out:
> -       mutex_unlock(&slab_mutex);
> -       return ret;
> -}
> -
> -static int slab_memory_callback(struct notifier_block *self,
> -                               unsigned long action, void *arg)
> -{
> -       int ret = 0;
> -
> -       switch (action) {
> -       case MEM_GOING_ONLINE:
> -               ret = slab_mem_going_online_callback(arg);
> -               break;
> -       case MEM_GOING_OFFLINE:
> -               ret = slab_mem_going_offline_callback(arg);
> -               break;
> -       case MEM_OFFLINE:
> -       case MEM_CANCEL_ONLINE:
> -               slab_mem_offline_callback(arg);
> -               break;
> -       case MEM_ONLINE:
> -       case MEM_CANCEL_OFFLINE:
> -               break;
> -       }
> -       if (ret)
> -               ret = notifier_from_errno(ret);
> -       else
> -               ret = NOTIFY_OK;
> -       return ret;
> -}
> -
> -/********************************************************************
> - *                     Basic setup of slabs
> - *******************************************************************/
> -
> -/*
> - * Used for early kmem_cache structures that were allocated using
> - * the page allocator. Allocate them properly then fix up the pointers
> - * that may be pointing to the wrong kmem_cache structure.
> - */
> -
> -static struct kmem_cache * __init bootstrap(struct kmem_cache *static_cache)
> -{
> -       int node;
> -       struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
> -       struct kmem_cache_node *n;
> -
> -       memcpy(s, static_cache, kmem_cache->object_size);
> -
> -       /*
> -        * This runs very early, and only the boot processor is supposed to be
> -        * up.  Even if it weren't true, IRQs are not up so we couldn't fire
> -        * IPIs around.
> -        */
> -       __flush_cpu_slab(s, smp_processor_id());
> -       for_each_kmem_cache_node(s, node, n) {
> -               struct slab *p;
> -
> -               list_for_each_entry(p, &n->partial, slab_list)
> -                       p->slab_cache = s;
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -               list_for_each_entry(p, &n->full, slab_list)
> -                       p->slab_cache = s;
> -#endif
> -       }
> -       list_add(&s->list, &slab_caches);
> -       return s;
> -}
> -
> -void __init kmem_cache_init(void)
> -{
> -       static __initdata struct kmem_cache boot_kmem_cache,
> -               boot_kmem_cache_node;
> -       int node;
> -
> -       if (debug_guardpage_minorder())
> -               slub_max_order = 0;
> -
> -       /* Print slub debugging pointers without hashing */
> -       if (__slub_debug_enabled())
> -               no_hash_pointers_enable(NULL);
> -
> -       kmem_cache_node = &boot_kmem_cache_node;
> -       kmem_cache = &boot_kmem_cache;
> -
> -       /*
> -        * Initialize the nodemask for which we will allocate per node
> -        * structures. Here we don't need taking slab_mutex yet.
> -        */
> -       for_each_node_state(node, N_NORMAL_MEMORY)
> -               node_set(node, slab_nodes);
> -
> -       create_boot_cache(kmem_cache_node, "kmem_cache_node",
> -               sizeof(struct kmem_cache_node), SLAB_HWCACHE_ALIGN, 0, 0);
> -
> -       hotplug_memory_notifier(slab_memory_callback, SLAB_CALLBACK_PRI);
> -
> -       /* Able to allocate the per node structures */
> -       slab_state = PARTIAL;
> -
> -       create_boot_cache(kmem_cache, "kmem_cache",
> -                       offsetof(struct kmem_cache, node) +
> -                               nr_node_ids * sizeof(struct kmem_cache_node *),
> -                      SLAB_HWCACHE_ALIGN, 0, 0);
> -
> -       kmem_cache = bootstrap(&boot_kmem_cache);
> -       kmem_cache_node = bootstrap(&boot_kmem_cache_node);
> -
> -       /* Now we can use the kmem_cache to allocate kmalloc slabs */
> -       setup_kmalloc_cache_index_table();
> -       create_kmalloc_caches(0);
> -
> -       /* Setup random freelists for each cache */
> -       init_freelist_randomization();
> -
> -       cpuhp_setup_state_nocalls(CPUHP_SLUB_DEAD, "slub:dead", NULL,
> -                                 slub_cpu_dead);
> -
> -       pr_info("SLUB: HWalign=%d, Order=%u-%u, MinObjects=%u, CPUs=%u, Nodes=%u\n",
> -               cache_line_size(),
> -               slub_min_order, slub_max_order, slub_min_objects,
> -               nr_cpu_ids, nr_node_ids);
> -}
> -
> -void __init kmem_cache_init_late(void)
> -{
> -#ifndef CONFIG_SLUB_TINY
> -       flushwq = alloc_workqueue("slub_flushwq", WQ_MEM_RECLAIM, 0);
> -       WARN_ON(!flushwq);
> -#endif
> -}
> -
> -struct kmem_cache *
> -__kmem_cache_alias(const char *name, unsigned int size, unsigned int align,
> -                  slab_flags_t flags, void (*ctor)(void *))
> -{
> -       struct kmem_cache *s;
> -
> -       s = find_mergeable(size, align, flags, name, ctor);
> -       if (s) {
> -               if (sysfs_slab_alias(s, name))
> -                       return NULL;
> -
> -               s->refcount++;
> -
> -               /*
> -                * Adjust the object sizes so that we clear
> -                * the complete object on kzalloc.
> -                */
> -               s->object_size = max(s->object_size, size);
> -               s->inuse = max(s->inuse, ALIGN(size, sizeof(void *)));
> -       }
> -
> -       return s;
> -}
> -
> -int __kmem_cache_create(struct kmem_cache *s, slab_flags_t flags)
> -{
> -       int err;
> -
> -       err = kmem_cache_open(s, flags);
> -       if (err)
> -               return err;
> -
> -       /* Mutex is not taken during early boot */
> -       if (slab_state <= UP)
> -               return 0;
> -
> -       err = sysfs_slab_add(s);
> -       if (err) {
> -               __kmem_cache_release(s);
> -               return err;
> -       }
> -
> -       if (s->flags & SLAB_STORE_USER)
> -               debugfs_slab_add(s);
> -
> -       return 0;
> -}
> -
> -#ifdef SLAB_SUPPORTS_SYSFS
> -static int count_inuse(struct slab *slab)
> -{
> -       return slab->inuse;
> -}
> -
> -static int count_total(struct slab *slab)
> -{
> -       return slab->objects;
> -}
> -#endif
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -static void validate_slab(struct kmem_cache *s, struct slab *slab,
> -                         unsigned long *obj_map)
> -{
> -       void *p;
> -       void *addr = slab_address(slab);
> -
> -       if (!check_slab(s, slab) || !on_freelist(s, slab, NULL))
> -               return;
> -
> -       /* Now we know that a valid freelist exists */
> -       __fill_map(obj_map, s, slab);
> -       for_each_object(p, s, addr, slab->objects) {
> -               u8 val = test_bit(__obj_to_index(s, addr, p), obj_map) ?
> -                        SLUB_RED_INACTIVE : SLUB_RED_ACTIVE;
> -
> -               if (!check_object(s, slab, p, val))
> -                       break;
> -       }
> -}
> -
> -static int validate_slab_node(struct kmem_cache *s,
> -               struct kmem_cache_node *n, unsigned long *obj_map)
> -{
> -       unsigned long count = 0;
> -       struct slab *slab;
> -       unsigned long flags;
> -
> -       spin_lock_irqsave(&n->list_lock, flags);
> -
> -       list_for_each_entry(slab, &n->partial, slab_list) {
> -               validate_slab(s, slab, obj_map);
> -               count++;
> -       }
> -       if (count != n->nr_partial) {
> -               pr_err("SLUB %s: %ld partial slabs counted but counter=%ld\n",
> -                      s->name, count, n->nr_partial);
> -               slab_add_kunit_errors();
> -       }
> -
> -       if (!(s->flags & SLAB_STORE_USER))
> -               goto out;
> -
> -       list_for_each_entry(slab, &n->full, slab_list) {
> -               validate_slab(s, slab, obj_map);
> -               count++;
> -       }
> -       if (count != atomic_long_read(&n->nr_slabs)) {
> -               pr_err("SLUB: %s %ld slabs counted but counter=%ld\n",
> -                      s->name, count, atomic_long_read(&n->nr_slabs));
> -               slab_add_kunit_errors();
> -       }
> -
> -out:
> -       spin_unlock_irqrestore(&n->list_lock, flags);
> -       return count;
> -}
> -
> -long validate_slab_cache(struct kmem_cache *s)
> -{
> -       int node;
> -       unsigned long count = 0;
> -       struct kmem_cache_node *n;
> -       unsigned long *obj_map;
> -
> -       obj_map = bitmap_alloc(oo_objects(s->oo), GFP_KERNEL);
> -       if (!obj_map)
> -               return -ENOMEM;
> -
> -       flush_all(s);
> -       for_each_kmem_cache_node(s, node, n)
> -               count += validate_slab_node(s, n, obj_map);
> -
> -       bitmap_free(obj_map);
> -
> -       return count;
> -}
> -EXPORT_SYMBOL(validate_slab_cache);
> -
> -#ifdef CONFIG_DEBUG_FS
> -/*
> - * Generate lists of code addresses where slabcache objects are allocated
> - * and freed.
> - */
> -
> -struct location {
> -       depot_stack_handle_t handle;
> -       unsigned long count;
> -       unsigned long addr;
> -       unsigned long waste;
> -       long long sum_time;
> -       long min_time;
> -       long max_time;
> -       long min_pid;
> -       long max_pid;
> -       DECLARE_BITMAP(cpus, NR_CPUS);
> -       nodemask_t nodes;
> -};
> -
> -struct loc_track {
> -       unsigned long max;
> -       unsigned long count;
> -       struct location *loc;
> -       loff_t idx;
> -};
> -
> -static struct dentry *slab_debugfs_root;
> -
> -static void free_loc_track(struct loc_track *t)
> -{
> -       if (t->max)
> -               free_pages((unsigned long)t->loc,
> -                       get_order(sizeof(struct location) * t->max));
> -}
> -
> -static int alloc_loc_track(struct loc_track *t, unsigned long max, gfp_t flags)
> -{
> -       struct location *l;
> -       int order;
> -
> -       order = get_order(sizeof(struct location) * max);
> -
> -       l = (void *)__get_free_pages(flags, order);
> -       if (!l)
> -               return 0;
> -
> -       if (t->count) {
> -               memcpy(l, t->loc, sizeof(struct location) * t->count);
> -               free_loc_track(t);
> -       }
> -       t->max = max;
> -       t->loc = l;
> -       return 1;
> -}
> -
> -static int add_location(struct loc_track *t, struct kmem_cache *s,
> -                               const struct track *track,
> -                               unsigned int orig_size)
> -{
> -       long start, end, pos;
> -       struct location *l;
> -       unsigned long caddr, chandle, cwaste;
> -       unsigned long age = jiffies - track->when;
> -       depot_stack_handle_t handle = 0;
> -       unsigned int waste = s->object_size - orig_size;
> -
> -#ifdef CONFIG_STACKDEPOT
> -       handle = READ_ONCE(track->handle);
> -#endif
> -       start = -1;
> -       end = t->count;
> -
> -       for ( ; ; ) {
> -               pos = start + (end - start + 1) / 2;
> -
> -               /*
> -                * There is nothing at "end". If we end up there
> -                * we need to add something to before end.
> -                */
> -               if (pos == end)
> -                       break;
> -
> -               l = &t->loc[pos];
> -               caddr = l->addr;
> -               chandle = l->handle;
> -               cwaste = l->waste;
> -               if ((track->addr == caddr) && (handle == chandle) &&
> -                       (waste == cwaste)) {
> -
> -                       l->count++;
> -                       if (track->when) {
> -                               l->sum_time += age;
> -                               if (age < l->min_time)
> -                                       l->min_time = age;
> -                               if (age > l->max_time)
> -                                       l->max_time = age;
> -
> -                               if (track->pid < l->min_pid)
> -                                       l->min_pid = track->pid;
> -                               if (track->pid > l->max_pid)
> -                                       l->max_pid = track->pid;
> -
> -                               cpumask_set_cpu(track->cpu,
> -                                               to_cpumask(l->cpus));
> -                       }
> -                       node_set(page_to_nid(virt_to_page(track)), l->nodes);
> -                       return 1;
> -               }
> -
> -               if (track->addr < caddr)
> -                       end = pos;
> -               else if (track->addr == caddr && handle < chandle)
> -                       end = pos;
> -               else if (track->addr == caddr && handle == chandle &&
> -                               waste < cwaste)
> -                       end = pos;
> -               else
> -                       start = pos;
> -       }
> -
> -       /*
> -        * Not found. Insert new tracking element.
> -        */
> -       if (t->count >= t->max && !alloc_loc_track(t, 2 * t->max, GFP_ATOMIC))
> -               return 0;
> -
> -       l = t->loc + pos;
> -       if (pos < t->count)
> -               memmove(l + 1, l,
> -                       (t->count - pos) * sizeof(struct location));
> -       t->count++;
> -       l->count = 1;
> -       l->addr = track->addr;
> -       l->sum_time = age;
> -       l->min_time = age;
> -       l->max_time = age;
> -       l->min_pid = track->pid;
> -       l->max_pid = track->pid;
> -       l->handle = handle;
> -       l->waste = waste;
> -       cpumask_clear(to_cpumask(l->cpus));
> -       cpumask_set_cpu(track->cpu, to_cpumask(l->cpus));
> -       nodes_clear(l->nodes);
> -       node_set(page_to_nid(virt_to_page(track)), l->nodes);
> -       return 1;
> -}
> -
> -static void process_slab(struct loc_track *t, struct kmem_cache *s,
> -               struct slab *slab, enum track_item alloc,
> -               unsigned long *obj_map)
> -{
> -       void *addr = slab_address(slab);
> -       bool is_alloc = (alloc == TRACK_ALLOC);
> -       void *p;
> -
> -       __fill_map(obj_map, s, slab);
> -
> -       for_each_object(p, s, addr, slab->objects)
> -               if (!test_bit(__obj_to_index(s, addr, p), obj_map))
> -                       add_location(t, s, get_track(s, p, alloc),
> -                                    is_alloc ? get_orig_size(s, p) :
> -                                               s->object_size);
> -}
> -#endif  /* CONFIG_DEBUG_FS   */
> -#endif /* CONFIG_SLUB_DEBUG */
> -
> -#ifdef SLAB_SUPPORTS_SYSFS
> -enum slab_stat_type {
> -       SL_ALL,                 /* All slabs */
> -       SL_PARTIAL,             /* Only partially allocated slabs */
> -       SL_CPU,                 /* Only slabs used for cpu caches */
> -       SL_OBJECTS,             /* Determine allocated objects not slabs */
> -       SL_TOTAL                /* Determine object capacity not slabs */
> -};
> -
> -#define SO_ALL         (1 << SL_ALL)
> -#define SO_PARTIAL     (1 << SL_PARTIAL)
> -#define SO_CPU         (1 << SL_CPU)
> -#define SO_OBJECTS     (1 << SL_OBJECTS)
> -#define SO_TOTAL       (1 << SL_TOTAL)
> -
> -static ssize_t show_slab_objects(struct kmem_cache *s,
> -                                char *buf, unsigned long flags)
> -{
> -       unsigned long total = 0;
> -       int node;
> -       int x;
> -       unsigned long *nodes;
> -       int len = 0;
> -
> -       nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL);
> -       if (!nodes)
> -               return -ENOMEM;
> -
> -       if (flags & SO_CPU) {
> -               int cpu;
> -
> -               for_each_possible_cpu(cpu) {
> -                       struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,
> -                                                              cpu);
> -                       int node;
> -                       struct slab *slab;
> -
> -                       slab = READ_ONCE(c->slab);
> -                       if (!slab)
> -                               continue;
> -
> -                       node = slab_nid(slab);
> -                       if (flags & SO_TOTAL)
> -                               x = slab->objects;
> -                       else if (flags & SO_OBJECTS)
> -                               x = slab->inuse;
> -                       else
> -                               x = 1;
> -
> -                       total += x;
> -                       nodes[node] += x;
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -                       slab = slub_percpu_partial_read_once(c);
> -                       if (slab) {
> -                               node = slab_nid(slab);
> -                               if (flags & SO_TOTAL)
> -                                       WARN_ON_ONCE(1);
> -                               else if (flags & SO_OBJECTS)
> -                                       WARN_ON_ONCE(1);
> -                               else
> -                                       x = slab->slabs;
> -                               total += x;
> -                               nodes[node] += x;
> -                       }
> -#endif
> -               }
> -       }
> -
> -       /*
> -        * It is impossible to take "mem_hotplug_lock" here with "kernfs_mutex"
> -        * already held which will conflict with an existing lock order:
> -        *
> -        * mem_hotplug_lock->slab_mutex->kernfs_mutex
> -        *
> -        * We don't really need mem_hotplug_lock (to hold off
> -        * slab_mem_going_offline_callback) here because slab's memory hot
> -        * unplug code doesn't destroy the kmem_cache->node[] data.
> -        */
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -       if (flags & SO_ALL) {
> -               struct kmem_cache_node *n;
> -
> -               for_each_kmem_cache_node(s, node, n) {
> -
> -                       if (flags & SO_TOTAL)
> -                               x = atomic_long_read(&n->total_objects);
> -                       else if (flags & SO_OBJECTS)
> -                               x = atomic_long_read(&n->total_objects) -
> -                                       count_partial(n, count_free);
> -                       else
> -                               x = atomic_long_read(&n->nr_slabs);
> -                       total += x;
> -                       nodes[node] += x;
> -               }
> -
> -       } else
> -#endif
> -       if (flags & SO_PARTIAL) {
> -               struct kmem_cache_node *n;
> -
> -               for_each_kmem_cache_node(s, node, n) {
> -                       if (flags & SO_TOTAL)
> -                               x = count_partial(n, count_total);
> -                       else if (flags & SO_OBJECTS)
> -                               x = count_partial(n, count_inuse);
> -                       else
> -                               x = n->nr_partial;
> -                       total += x;
> -                       nodes[node] += x;
> -               }
> -       }
> -
> -       len += sysfs_emit_at(buf, len, "%lu", total);
> -#ifdef CONFIG_NUMA
> -       for (node = 0; node < nr_node_ids; node++) {
> -               if (nodes[node])
> -                       len += sysfs_emit_at(buf, len, " N%d=%lu",
> -                                            node, nodes[node]);
> -       }
> -#endif
> -       len += sysfs_emit_at(buf, len, "\n");
> -       kfree(nodes);
> -
> -       return len;
> -}
> -
> -#define to_slab_attr(n) container_of(n, struct slab_attribute, attr)
> -#define to_slab(n) container_of(n, struct kmem_cache, kobj)
> -
> -struct slab_attribute {
> -       struct attribute attr;
> -       ssize_t (*show)(struct kmem_cache *s, char *buf);
> -       ssize_t (*store)(struct kmem_cache *s, const char *x, size_t count);
> -};
> -
> -#define SLAB_ATTR_RO(_name) \
> -       static struct slab_attribute _name##_attr = __ATTR_RO_MODE(_name, 0400)
> -
> -#define SLAB_ATTR(_name) \
> -       static struct slab_attribute _name##_attr = __ATTR_RW_MODE(_name, 0600)
> -
> -static ssize_t slab_size_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", s->size);
> -}
> -SLAB_ATTR_RO(slab_size);
> -
> -static ssize_t align_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", s->align);
> -}
> -SLAB_ATTR_RO(align);
> -
> -static ssize_t object_size_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", s->object_size);
> -}
> -SLAB_ATTR_RO(object_size);
> -
> -static ssize_t objs_per_slab_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", oo_objects(s->oo));
> -}
> -SLAB_ATTR_RO(objs_per_slab);
> -
> -static ssize_t order_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", oo_order(s->oo));
> -}
> -SLAB_ATTR_RO(order);
> -
> -static ssize_t min_partial_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%lu\n", s->min_partial);
> -}
> -
> -static ssize_t min_partial_store(struct kmem_cache *s, const char *buf,
> -                                size_t length)
> -{
> -       unsigned long min;
> -       int err;
> -
> -       err = kstrtoul(buf, 10, &min);
> -       if (err)
> -               return err;
> -
> -       s->min_partial = min;
> -       return length;
> -}
> -SLAB_ATTR(min_partial);
> -
> -static ssize_t cpu_partial_show(struct kmem_cache *s, char *buf)
> -{
> -       unsigned int nr_partial = 0;
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       nr_partial = s->cpu_partial;
> -#endif
> -
> -       return sysfs_emit(buf, "%u\n", nr_partial);
> -}
> -
> -static ssize_t cpu_partial_store(struct kmem_cache *s, const char *buf,
> -                                size_t length)
> -{
> -       unsigned int objects;
> -       int err;
> -
> -       err = kstrtouint(buf, 10, &objects);
> -       if (err)
> -               return err;
> -       if (objects && !kmem_cache_has_cpu_partial(s))
> -               return -EINVAL;
> -
> -       slub_set_cpu_partial(s, objects);
> -       flush_all(s);
> -       return length;
> -}
> -SLAB_ATTR(cpu_partial);
> -
> -static ssize_t ctor_show(struct kmem_cache *s, char *buf)
> -{
> -       if (!s->ctor)
> -               return 0;
> -       return sysfs_emit(buf, "%pS\n", s->ctor);
> -}
> -SLAB_ATTR_RO(ctor);
> -
> -static ssize_t aliases_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", s->refcount < 0 ? 0 : s->refcount - 1);
> -}
> -SLAB_ATTR_RO(aliases);
> -
> -static ssize_t partial_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_PARTIAL);
> -}
> -SLAB_ATTR_RO(partial);
> -
> -static ssize_t cpu_slabs_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_CPU);
> -}
> -SLAB_ATTR_RO(cpu_slabs);
> -
> -static ssize_t objects_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_ALL|SO_OBJECTS);
> -}
> -SLAB_ATTR_RO(objects);
> -
> -static ssize_t objects_partial_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_PARTIAL|SO_OBJECTS);
> -}
> -SLAB_ATTR_RO(objects_partial);
> -
> -static ssize_t slabs_cpu_partial_show(struct kmem_cache *s, char *buf)
> -{
> -       int objects = 0;
> -       int slabs = 0;
> -       int cpu __maybe_unused;
> -       int len = 0;
> -
> -#ifdef CONFIG_SLUB_CPU_PARTIAL
> -       for_each_online_cpu(cpu) {
> -               struct slab *slab;
> -
> -               slab = slub_percpu_partial(per_cpu_ptr(s->cpu_slab, cpu));
> -
> -               if (slab)
> -                       slabs += slab->slabs;
> -       }
> -#endif
> -
> -       /* Approximate half-full slabs, see slub_set_cpu_partial() */
> -       objects = (slabs * oo_objects(s->oo)) / 2;
> -       len += sysfs_emit_at(buf, len, "%d(%d)", objects, slabs);
> -
> -#if defined(CONFIG_SLUB_CPU_PARTIAL) && defined(CONFIG_SMP)
> -       for_each_online_cpu(cpu) {
> -               struct slab *slab;
> -
> -               slab = slub_percpu_partial(per_cpu_ptr(s->cpu_slab, cpu));
> -               if (slab) {
> -                       slabs = READ_ONCE(slab->slabs);
> -                       objects = (slabs * oo_objects(s->oo)) / 2;
> -                       len += sysfs_emit_at(buf, len, " C%d=%d(%d)",
> -                                            cpu, objects, slabs);
> -               }
> -       }
> -#endif
> -       len += sysfs_emit_at(buf, len, "\n");
> -
> -       return len;
> -}
> -SLAB_ATTR_RO(slabs_cpu_partial);
> -
> -static ssize_t reclaim_account_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_RECLAIM_ACCOUNT));
> -}
> -SLAB_ATTR_RO(reclaim_account);
> -
> -static ssize_t hwcache_align_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_HWCACHE_ALIGN));
> -}
> -SLAB_ATTR_RO(hwcache_align);
> -
> -#ifdef CONFIG_ZONE_DMA
> -static ssize_t cache_dma_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_CACHE_DMA));
> -}
> -SLAB_ATTR_RO(cache_dma);
> -#endif
> -
> -#ifdef CONFIG_HARDENED_USERCOPY
> -static ssize_t usersize_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", s->usersize);
> -}
> -SLAB_ATTR_RO(usersize);
> -#endif
> -
> -static ssize_t destroy_by_rcu_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_TYPESAFE_BY_RCU));
> -}
> -SLAB_ATTR_RO(destroy_by_rcu);
> -
> -#ifdef CONFIG_SLUB_DEBUG
> -static ssize_t slabs_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_ALL);
> -}
> -SLAB_ATTR_RO(slabs);
> -
> -static ssize_t total_objects_show(struct kmem_cache *s, char *buf)
> -{
> -       return show_slab_objects(s, buf, SO_ALL|SO_TOTAL);
> -}
> -SLAB_ATTR_RO(total_objects);
> -
> -static ssize_t sanity_checks_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_CONSISTENCY_CHECKS));
> -}
> -SLAB_ATTR_RO(sanity_checks);
> -
> -static ssize_t trace_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_TRACE));
> -}
> -SLAB_ATTR_RO(trace);
> -
> -static ssize_t red_zone_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_RED_ZONE));
> -}
> -
> -SLAB_ATTR_RO(red_zone);
> -
> -static ssize_t poison_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_POISON));
> -}
> -
> -SLAB_ATTR_RO(poison);
> -
> -static ssize_t store_user_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_STORE_USER));
> -}
> -
> -SLAB_ATTR_RO(store_user);
> -
> -static ssize_t validate_show(struct kmem_cache *s, char *buf)
> -{
> -       return 0;
> -}
> -
> -static ssize_t validate_store(struct kmem_cache *s,
> -                       const char *buf, size_t length)
> -{
> -       int ret = -EINVAL;
> -
> -       if (buf[0] == '1' && kmem_cache_debug(s)) {
> -               ret = validate_slab_cache(s);
> -               if (ret >= 0)
> -                       ret = length;
> -       }
> -       return ret;
> -}
> -SLAB_ATTR(validate);
> -
> -#endif /* CONFIG_SLUB_DEBUG */
> -
> -#ifdef CONFIG_FAILSLAB
> -static ssize_t failslab_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_FAILSLAB));
> -}
> -
> -static ssize_t failslab_store(struct kmem_cache *s, const char *buf,
> -                               size_t length)
> -{
> -       if (s->refcount > 1)
> -               return -EINVAL;
> -
> -       if (buf[0] == '1')
> -               WRITE_ONCE(s->flags, s->flags | SLAB_FAILSLAB);
> -       else
> -               WRITE_ONCE(s->flags, s->flags & ~SLAB_FAILSLAB);
> -
> -       return length;
> -}
> -SLAB_ATTR(failslab);
> -#endif
> -
> -static ssize_t shrink_show(struct kmem_cache *s, char *buf)
> -{
> -       return 0;
> -}
> -
> -static ssize_t shrink_store(struct kmem_cache *s,
> -                       const char *buf, size_t length)
> -{
> -       if (buf[0] == '1')
> -               kmem_cache_shrink(s);
> -       else
> -               return -EINVAL;
> -       return length;
> -}
> -SLAB_ATTR(shrink);
> -
> -#ifdef CONFIG_NUMA
> -static ssize_t remote_node_defrag_ratio_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%u\n", s->remote_node_defrag_ratio / 10);
> -}
> -
> -static ssize_t remote_node_defrag_ratio_store(struct kmem_cache *s,
> -                               const char *buf, size_t length)
> -{
> -       unsigned int ratio;
> -       int err;
> -
> -       err = kstrtouint(buf, 10, &ratio);
> -       if (err)
> -               return err;
> -       if (ratio > 100)
> -               return -ERANGE;
> -
> -       s->remote_node_defrag_ratio = ratio * 10;
> -
> -       return length;
> -}
> -SLAB_ATTR(remote_node_defrag_ratio);
> -#endif
> -
> -#ifdef CONFIG_SLUB_STATS
> -static int show_stat(struct kmem_cache *s, char *buf, enum stat_item si)
> -{
> -       unsigned long sum  = 0;
> -       int cpu;
> -       int len = 0;
> -       int *data = kmalloc_array(nr_cpu_ids, sizeof(int), GFP_KERNEL);
> -
> -       if (!data)
> -               return -ENOMEM;
> -
> -       for_each_online_cpu(cpu) {
> -               unsigned x = per_cpu_ptr(s->cpu_slab, cpu)->stat[si];
> -
> -               data[cpu] = x;
> -               sum += x;
> -       }
> -
> -       len += sysfs_emit_at(buf, len, "%lu", sum);
> -
> -#ifdef CONFIG_SMP
> -       for_each_online_cpu(cpu) {
> -               if (data[cpu])
> -                       len += sysfs_emit_at(buf, len, " C%d=%u",
> -                                            cpu, data[cpu]);
> -       }
> -#endif
> -       kfree(data);
> -       len += sysfs_emit_at(buf, len, "\n");
> -
> -       return len;
> -}
> -
> -static void clear_stat(struct kmem_cache *s, enum stat_item si)
> -{
> -       int cpu;
> -
> -       for_each_online_cpu(cpu)
> -               per_cpu_ptr(s->cpu_slab, cpu)->stat[si] = 0;
> -}
> -
> -#define STAT_ATTR(si, text)                                    \
> -static ssize_t text##_show(struct kmem_cache *s, char *buf)    \
> -{                                                              \
> -       return show_stat(s, buf, si);                           \
> -}                                                              \
> -static ssize_t text##_store(struct kmem_cache *s,              \
> -                               const char *buf, size_t length) \
> -{                                                              \
> -       if (buf[0] != '0')                                      \
> -               return -EINVAL;                                 \
> -       clear_stat(s, si);                                      \
> -       return length;                                          \
> -}                                                              \
> -SLAB_ATTR(text);                                               \
> -
> -STAT_ATTR(ALLOC_FASTPATH, alloc_fastpath);
> -STAT_ATTR(ALLOC_SLOWPATH, alloc_slowpath);
> -STAT_ATTR(FREE_FASTPATH, free_fastpath);
> -STAT_ATTR(FREE_SLOWPATH, free_slowpath);
> -STAT_ATTR(FREE_FROZEN, free_frozen);
> -STAT_ATTR(FREE_ADD_PARTIAL, free_add_partial);
> -STAT_ATTR(FREE_REMOVE_PARTIAL, free_remove_partial);
> -STAT_ATTR(ALLOC_FROM_PARTIAL, alloc_from_partial);
> -STAT_ATTR(ALLOC_SLAB, alloc_slab);
> -STAT_ATTR(ALLOC_REFILL, alloc_refill);
> -STAT_ATTR(ALLOC_NODE_MISMATCH, alloc_node_mismatch);
> -STAT_ATTR(FREE_SLAB, free_slab);
> -STAT_ATTR(CPUSLAB_FLUSH, cpuslab_flush);
> -STAT_ATTR(DEACTIVATE_FULL, deactivate_full);
> -STAT_ATTR(DEACTIVATE_EMPTY, deactivate_empty);
> -STAT_ATTR(DEACTIVATE_TO_HEAD, deactivate_to_head);
> -STAT_ATTR(DEACTIVATE_TO_TAIL, deactivate_to_tail);
> -STAT_ATTR(DEACTIVATE_REMOTE_FREES, deactivate_remote_frees);
> -STAT_ATTR(DEACTIVATE_BYPASS, deactivate_bypass);
> -STAT_ATTR(ORDER_FALLBACK, order_fallback);
> -STAT_ATTR(CMPXCHG_DOUBLE_CPU_FAIL, cmpxchg_double_cpu_fail);
> -STAT_ATTR(CMPXCHG_DOUBLE_FAIL, cmpxchg_double_fail);
> -STAT_ATTR(CPU_PARTIAL_ALLOC, cpu_partial_alloc);
> -STAT_ATTR(CPU_PARTIAL_FREE, cpu_partial_free);
> -STAT_ATTR(CPU_PARTIAL_NODE, cpu_partial_node);
> -STAT_ATTR(CPU_PARTIAL_DRAIN, cpu_partial_drain);
> -#endif /* CONFIG_SLUB_STATS */
> -
> -#ifdef CONFIG_KFENCE
> -static ssize_t skip_kfence_show(struct kmem_cache *s, char *buf)
> -{
> -       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_SKIP_KFENCE));
> -}
> -
> -static ssize_t skip_kfence_store(struct kmem_cache *s,
> -                       const char *buf, size_t length)
> -{
> -       int ret = length;
> -
> -       if (buf[0] == '0')
> -               s->flags &= ~SLAB_SKIP_KFENCE;
> -       else if (buf[0] == '1')
> -               s->flags |= SLAB_SKIP_KFENCE;
> -       else
> -               ret = -EINVAL;
> -
> -       return ret;
> -}
> -SLAB_ATTR(skip_kfence);
> -#endif
> -
> -static struct attribute *slab_attrs[] = {
> -       &slab_size_attr.attr,
> -       &object_size_attr.attr,
> -       &objs_per_slab_attr.attr,
> -       &order_attr.attr,
> -       &min_partial_attr.attr,
> -       &cpu_partial_attr.attr,
> -       &objects_attr.attr,
> -       &objects_partial_attr.attr,
> -       &partial_attr.attr,
> -       &cpu_slabs_attr.attr,
> -       &ctor_attr.attr,
> -       &aliases_attr.attr,
> -       &align_attr.attr,
> -       &hwcache_align_attr.attr,
> -       &reclaim_account_attr.attr,
> -       &destroy_by_rcu_attr.attr,
> -       &shrink_attr.attr,
> -       &slabs_cpu_partial_attr.attr,
> -#ifdef CONFIG_SLUB_DEBUG
> -       &total_objects_attr.attr,
> -       &slabs_attr.attr,
> -       &sanity_checks_attr.attr,
> -       &trace_attr.attr,
> -       &red_zone_attr.attr,
> -       &poison_attr.attr,
> -       &store_user_attr.attr,
> -       &validate_attr.attr,
> -#endif
> -#ifdef CONFIG_ZONE_DMA
> -       &cache_dma_attr.attr,
> -#endif
> -#ifdef CONFIG_NUMA
> -       &remote_node_defrag_ratio_attr.attr,
> -#endif
> -#ifdef CONFIG_SLUB_STATS
> -       &alloc_fastpath_attr.attr,
> -       &alloc_slowpath_attr.attr,
> -       &free_fastpath_attr.attr,
> -       &free_slowpath_attr.attr,
> -       &free_frozen_attr.attr,
> -       &free_add_partial_attr.attr,
> -       &free_remove_partial_attr.attr,
> -       &alloc_from_partial_attr.attr,
> -       &alloc_slab_attr.attr,
> -       &alloc_refill_attr.attr,
> -       &alloc_node_mismatch_attr.attr,
> -       &free_slab_attr.attr,
> -       &cpuslab_flush_attr.attr,
> -       &deactivate_full_attr.attr,
> -       &deactivate_empty_attr.attr,
> -       &deactivate_to_head_attr.attr,
> -       &deactivate_to_tail_attr.attr,
> -       &deactivate_remote_frees_attr.attr,
> -       &deactivate_bypass_attr.attr,
> -       &order_fallback_attr.attr,
> -       &cmpxchg_double_fail_attr.attr,
> -       &cmpxchg_double_cpu_fail_attr.attr,
> -       &cpu_partial_alloc_attr.attr,
> -       &cpu_partial_free_attr.attr,
> -       &cpu_partial_node_attr.attr,
> -       &cpu_partial_drain_attr.attr,
> -#endif
> -#ifdef CONFIG_FAILSLAB
> -       &failslab_attr.attr,
> -#endif
> -#ifdef CONFIG_HARDENED_USERCOPY
> -       &usersize_attr.attr,
> -#endif
> -#ifdef CONFIG_KFENCE
> -       &skip_kfence_attr.attr,
> -#endif
> -
> -       NULL
> -};
> -
> -static const struct attribute_group slab_attr_group = {
> -       .attrs = slab_attrs,
> -};
> -
> -static ssize_t slab_attr_show(struct kobject *kobj,
> -                               struct attribute *attr,
> -                               char *buf)
> -{
> -       struct slab_attribute *attribute;
> -       struct kmem_cache *s;
> -
> -       attribute = to_slab_attr(attr);
> -       s = to_slab(kobj);
> -
> -       if (!attribute->show)
> -               return -EIO;
> -
> -       return attribute->show(s, buf);
> -}
> -
> -static ssize_t slab_attr_store(struct kobject *kobj,
> -                               struct attribute *attr,
> -                               const char *buf, size_t len)
> -{
> -       struct slab_attribute *attribute;
> -       struct kmem_cache *s;
> -
> -       attribute = to_slab_attr(attr);
> -       s = to_slab(kobj);
> -
> -       if (!attribute->store)
> -               return -EIO;
> -
> -       return attribute->store(s, buf, len);
> -}
> -
> -static void kmem_cache_release(struct kobject *k)
> -{
> -       slab_kmem_cache_release(to_slab(k));
> -}
> -
> -static const struct sysfs_ops slab_sysfs_ops = {
> -       .show = slab_attr_show,
> -       .store = slab_attr_store,
> -};
> -
> -static struct kobj_type slab_ktype = {
> -       .sysfs_ops = &slab_sysfs_ops,
> -       .release = kmem_cache_release,
> -};
> -
> -static struct kset *slab_kset;
> -
> -static inline struct kset *cache_kset(struct kmem_cache *s)
> -{
> -       return slab_kset;
> -}
> -
> -#define ID_STR_LENGTH 32
> -
> -/* Create a unique string id for a slab cache:
> - *
> - * Format      :[flags-]size
> - */
> -static char *create_unique_id(struct kmem_cache *s)
> -{
> -       char *name = kmalloc(ID_STR_LENGTH, GFP_KERNEL);
> -       char *p = name;
> -
> -       if (!name)
> -               return ERR_PTR(-ENOMEM);
> -
> -       *p++ = ':';
> -       /*
> -        * First flags affecting slabcache operations. We will only
> -        * get here for aliasable slabs so we do not need to support
> -        * too many flags. The flags here must cover all flags that
> -        * are matched during merging to guarantee that the id is
> -        * unique.
> -        */
> -       if (s->flags & SLAB_CACHE_DMA)
> -               *p++ = 'd';
> -       if (s->flags & SLAB_CACHE_DMA32)
> -               *p++ = 'D';
> -       if (s->flags & SLAB_RECLAIM_ACCOUNT)
> -               *p++ = 'a';
> -       if (s->flags & SLAB_CONSISTENCY_CHECKS)
> -               *p++ = 'F';
> -       if (s->flags & SLAB_ACCOUNT)
> -               *p++ = 'A';
> -       if (p != name + 1)
> -               *p++ = '-';
> -       p += snprintf(p, ID_STR_LENGTH - (p - name), "%07u", s->size);
> -
> -       if (WARN_ON(p > name + ID_STR_LENGTH - 1)) {
> -               kfree(name);
> -               return ERR_PTR(-EINVAL);
> -       }
> -       kmsan_unpoison_memory(name, p - name);
> -       return name;
> -}
> -
> -static int sysfs_slab_add(struct kmem_cache *s)
> -{
> -       int err;
> -       const char *name;
> -       struct kset *kset = cache_kset(s);
> -       int unmergeable = slab_unmergeable(s);
> -
> -       if (!unmergeable && disable_higher_order_debug &&
> -                       (slub_debug & DEBUG_METADATA_FLAGS))
> -               unmergeable = 1;
> -
> -       if (unmergeable) {
> -               /*
> -                * Slabcache can never be merged so we can use the name proper.
> -                * This is typically the case for debug situations. In that
> -                * case we can catch duplicate names easily.
> -                */
> -               sysfs_remove_link(&slab_kset->kobj, s->name);
> -               name = s->name;
> -       } else {
> -               /*
> -                * Create a unique name for the slab as a target
> -                * for the symlinks.
> -                */
> -               name = create_unique_id(s);
> -               if (IS_ERR(name))
> -                       return PTR_ERR(name);
> -       }
> -
> -       s->kobj.kset = kset;
> -       err = kobject_init_and_add(&s->kobj, &slab_ktype, NULL, "%s", name);
> -       if (err)
> -               goto out;
> -
> -       err = sysfs_create_group(&s->kobj, &slab_attr_group);
> -       if (err)
> -               goto out_del_kobj;
> -
> -       if (!unmergeable) {
> -               /* Setup first alias */
> -               sysfs_slab_alias(s, s->name);
> -       }
> -out:
> -       if (!unmergeable)
> -               kfree(name);
> -       return err;
> -out_del_kobj:
> -       kobject_del(&s->kobj);
> -       goto out;
> -}
> -
> -void sysfs_slab_unlink(struct kmem_cache *s)
> -{
> -       if (slab_state >= FULL)
> -               kobject_del(&s->kobj);
> -}
> -
> -void sysfs_slab_release(struct kmem_cache *s)
> -{
> -       if (slab_state >= FULL)
> -               kobject_put(&s->kobj);
> -}
> -
> -/*
> - * Need to buffer aliases during bootup until sysfs becomes
> - * available lest we lose that information.
> - */
> -struct saved_alias {
> -       struct kmem_cache *s;
> -       const char *name;
> -       struct saved_alias *next;
> -};
> -
> -static struct saved_alias *alias_list;
> -
> -static int sysfs_slab_alias(struct kmem_cache *s, const char *name)
> -{
> -       struct saved_alias *al;
> -
> -       if (slab_state == FULL) {
> -               /*
> -                * If we have a leftover link then remove it.
> -                */
> -               sysfs_remove_link(&slab_kset->kobj, name);
> -               return sysfs_create_link(&slab_kset->kobj, &s->kobj, name);
> -       }
> -
> -       al = kmalloc(sizeof(struct saved_alias), GFP_KERNEL);
> -       if (!al)
> -               return -ENOMEM;
> -
> -       al->s = s;
> -       al->name = name;
> -       al->next = alias_list;
> -       alias_list = al;
> -       kmsan_unpoison_memory(al, sizeof(*al));
> -       return 0;
> -}
> -
> -static int __init slab_sysfs_init(void)
> -{
> -       struct kmem_cache *s;
> -       int err;
> -
> -       mutex_lock(&slab_mutex);
> -
> -       slab_kset = kset_create_and_add("slab", NULL, kernel_kobj);
> -       if (!slab_kset) {
> -               mutex_unlock(&slab_mutex);
> -               pr_err("Cannot register slab subsystem.\n");
> -               return -ENOSYS;
> -       }
> -
> -       slab_state = FULL;
> -
> -       list_for_each_entry(s, &slab_caches, list) {
> -               err = sysfs_slab_add(s);
> -               if (err)
> -                       pr_err("SLUB: Unable to add boot slab %s to sysfs\n",
> -                              s->name);
> -       }
> -
> -       while (alias_list) {
> -               struct saved_alias *al = alias_list;
> -
> -               alias_list = alias_list->next;
> -               err = sysfs_slab_alias(al->s, al->name);
> -               if (err)
> -                       pr_err("SLUB: Unable to add boot slab alias %s to sysfs\n",
> -                              al->name);
> -               kfree(al);
> -       }
> -
> -       mutex_unlock(&slab_mutex);
> -       return 0;
> -}
> -late_initcall(slab_sysfs_init);
> -#endif /* SLAB_SUPPORTS_SYSFS */
> -
> -#if defined(CONFIG_SLUB_DEBUG) && defined(CONFIG_DEBUG_FS)
> -static int slab_debugfs_show(struct seq_file *seq, void *v)
> -{
> -       struct loc_track *t = seq->private;
> -       struct location *l;
> -       unsigned long idx;
> -
> -       idx = (unsigned long) t->idx;
> -       if (idx < t->count) {
> -               l = &t->loc[idx];
> -
> -               seq_printf(seq, "%7ld ", l->count);
> -
> -               if (l->addr)
> -                       seq_printf(seq, "%pS", (void *)l->addr);
> -               else
> -                       seq_puts(seq, "<not-available>");
> -
> -               if (l->waste)
> -                       seq_printf(seq, " waste=%lu/%lu",
> -                               l->count * l->waste, l->waste);
> -
> -               if (l->sum_time != l->min_time) {
> -                       seq_printf(seq, " age=%ld/%llu/%ld",
> -                               l->min_time, div_u64(l->sum_time, l->count),
> -                               l->max_time);
> -               } else
> -                       seq_printf(seq, " age=%ld", l->min_time);
> -
> -               if (l->min_pid != l->max_pid)
> -                       seq_printf(seq, " pid=%ld-%ld", l->min_pid, l->max_pid);
> -               else
> -                       seq_printf(seq, " pid=%ld",
> -                               l->min_pid);
> -
> -               if (num_online_cpus() > 1 && !cpumask_empty(to_cpumask(l->cpus)))
> -                       seq_printf(seq, " cpus=%*pbl",
> -                                cpumask_pr_args(to_cpumask(l->cpus)));
> -
> -               if (nr_online_nodes > 1 && !nodes_empty(l->nodes))
> -                       seq_printf(seq, " nodes=%*pbl",
> -                                nodemask_pr_args(&l->nodes));
> -
> -#ifdef CONFIG_STACKDEPOT
> -               {
> -                       depot_stack_handle_t handle;
> -                       unsigned long *entries;
> -                       unsigned int nr_entries, j;
> -
> -                       handle = READ_ONCE(l->handle);
> -                       if (handle) {
> -                               nr_entries = stack_depot_fetch(handle, &entries);
> -                               seq_puts(seq, "\n");
> -                               for (j = 0; j < nr_entries; j++)
> -                                       seq_printf(seq, "        %pS\n", (void *)entries[j]);
> -                       }
> -               }
> -#endif
> -               seq_puts(seq, "\n");
> -       }
> -
> -       if (!idx && !t->count)
> -               seq_puts(seq, "No data\n");
> -
> -       return 0;
> -}
> -
> -static void slab_debugfs_stop(struct seq_file *seq, void *v)
> -{
> -}
> -
> -static void *slab_debugfs_next(struct seq_file *seq, void *v, loff_t *ppos)
> -{
> -       struct loc_track *t = seq->private;
> -
> -       t->idx = ++(*ppos);
> -       if (*ppos <= t->count)
> -               return ppos;
> -
> -       return NULL;
> -}
> -
> -static int cmp_loc_by_count(const void *a, const void *b, const void *data)
> -{
> -       struct location *loc1 = (struct location *)a;
> -       struct location *loc2 = (struct location *)b;
> -
> -       if (loc1->count > loc2->count)
> -               return -1;
> -       else
> -               return 1;
> -}
> -
> -static void *slab_debugfs_start(struct seq_file *seq, loff_t *ppos)
> -{
> -       struct loc_track *t = seq->private;
> -
> -       t->idx = *ppos;
> -       return ppos;
> -}
> -
> -static const struct seq_operations slab_debugfs_sops = {
> -       .start  = slab_debugfs_start,
> -       .next   = slab_debugfs_next,
> -       .stop   = slab_debugfs_stop,
> -       .show   = slab_debugfs_show,
> -};
> -
> -static int slab_debug_trace_open(struct inode *inode, struct file *filep)
> -{
> -
> -       struct kmem_cache_node *n;
> -       enum track_item alloc;
> -       int node;
> -       struct loc_track *t = __seq_open_private(filep, &slab_debugfs_sops,
> -                                               sizeof(struct loc_track));
> -       struct kmem_cache *s = file_inode(filep)->i_private;
> -       unsigned long *obj_map;
> -
> -       if (!t)
> -               return -ENOMEM;
> -
> -       obj_map = bitmap_alloc(oo_objects(s->oo), GFP_KERNEL);
> -       if (!obj_map) {
> -               seq_release_private(inode, filep);
> -               return -ENOMEM;
> -       }
> -
> -       if (strcmp(filep->f_path.dentry->d_name.name, "alloc_traces") == 0)
> -               alloc = TRACK_ALLOC;
> -       else
> -               alloc = TRACK_FREE;
> -
> -       if (!alloc_loc_track(t, PAGE_SIZE / sizeof(struct location), GFP_KERNEL)) {
> -               bitmap_free(obj_map);
> -               seq_release_private(inode, filep);
> -               return -ENOMEM;
> -       }
> -
> -       for_each_kmem_cache_node(s, node, n) {
> -               unsigned long flags;
> -               struct slab *slab;
> -
> -               if (!atomic_long_read(&n->nr_slabs))
> -                       continue;
> -
> -               spin_lock_irqsave(&n->list_lock, flags);
> -               list_for_each_entry(slab, &n->partial, slab_list)
> -                       process_slab(t, s, slab, alloc, obj_map);
> -               list_for_each_entry(slab, &n->full, slab_list)
> -                       process_slab(t, s, slab, alloc, obj_map);
> -               spin_unlock_irqrestore(&n->list_lock, flags);
> -       }
> -
> -       /* Sort locations by count */
> -       sort_r(t->loc, t->count, sizeof(struct location),
> -               cmp_loc_by_count, NULL, NULL);
> -
> -       bitmap_free(obj_map);
> -       return 0;
> -}
> -
> -static int slab_debug_trace_release(struct inode *inode, struct file *file)
> -{
> -       struct seq_file *seq = file->private_data;
> -       struct loc_track *t = seq->private;
> -
> -       free_loc_track(t);
> -       return seq_release_private(inode, file);
> -}
> -
> -static const struct file_operations slab_debugfs_fops = {
> -       .open    = slab_debug_trace_open,
> -       .read    = seq_read,
> -       .llseek  = seq_lseek,
> -       .release = slab_debug_trace_release,
> -};
> -
> -static void debugfs_slab_add(struct kmem_cache *s)
> -{
> -       struct dentry *slab_cache_dir;
> -
> -       if (unlikely(!slab_debugfs_root))
> -               return;
> -
> -       slab_cache_dir = debugfs_create_dir(s->name, slab_debugfs_root);
> -
> -       debugfs_create_file("alloc_traces", 0400,
> -               slab_cache_dir, s, &slab_debugfs_fops);
> -
> -       debugfs_create_file("free_traces", 0400,
> -               slab_cache_dir, s, &slab_debugfs_fops);
> -}
> -
> -void debugfs_slab_release(struct kmem_cache *s)
> -{
> -       debugfs_lookup_and_remove(s->name, slab_debugfs_root);
> -}
> -
> -static int __init slab_debugfs_init(void)
> -{
> -       struct kmem_cache *s;
> -
> -       slab_debugfs_root = debugfs_create_dir("slab", NULL);
> -
> -       list_for_each_entry(s, &slab_caches, list)
> -               if (s->flags & SLAB_STORE_USER)
> -                       debugfs_slab_add(s);
> -
> -       return 0;
> -
> -}
> -__initcall(slab_debugfs_init);
> -#endif
> -/*
> - * The /proc/slabinfo ABI
> - */
> -#ifdef CONFIG_SLUB_DEBUG
> -void get_slabinfo(struct kmem_cache *s, struct slabinfo *sinfo)
> -{
> -       unsigned long nr_slabs = 0;
> -       unsigned long nr_objs = 0;
> -       unsigned long nr_free = 0;
> -       int node;
> -       struct kmem_cache_node *n;
> -
> -       for_each_kmem_cache_node(s, node, n) {
> -               nr_slabs += node_nr_slabs(n);
> -               nr_objs += node_nr_objs(n);
> -               nr_free += count_partial(n, count_free);
> -       }
> -
> -       sinfo->active_objs = nr_objs - nr_free;
> -       sinfo->num_objs = nr_objs;
> -       sinfo->active_slabs = nr_slabs;
> -       sinfo->num_slabs = nr_slabs;
> -       sinfo->objects_per_slab = oo_objects(s->oo);
> -       sinfo->cache_order = oo_order(s->oo);
> -}
> -
> -void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *s)
> -{
> -}
> -
> -ssize_t slabinfo_write(struct file *file, const char __user *buffer,
> -                      size_t count, loff_t *ppos)
> -{
> -       return -EIO;
> -}
> -#endif /* CONFIG_SLUB_DEBUG */
> --
> 2.40.0
>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ