lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <d4606a20-b92b-495f-4b76-a4633cf805c7@bytedance.com>
Date:   Wed, 16 Nov 2022 18:42:43 +0800
From:   wuqiang <wuqiang.matt@...edance.com>
To:     "Masami Hiramatsu (Google)" <mhiramat@...nel.org>
Cc:     davem@...emloft.net, anil.s.keshavamurthy@...el.com,
        naveen.n.rao@...ux.ibm.com, rostedt@...dmis.org,
        peterz@...radead.org, akpm@...ux-foundation.org,
        sander@...nheule.net, ebiggers@...gle.com,
        dan.j.williams@...el.com, jpoimboe@...nel.org,
        linux-kernel@...r.kernel.org, lkp@...el.com, mattwu@....com
Subject: Re: [PATCH v6 1/4] lib: objpool added: ring-array based lockless MPMC
 queue

On 2022/11/14 23:54, Masami Hiramatsu (Google) wrote:
> On Tue,  8 Nov 2022 15:14:40 +0800
> wuqiang <wuqiang.matt@...edance.com> wrote:
> 
>> The object pool is a scalable implementaion of high performance queue
>> for objects allocation and reclamation, such as kretprobe instances.
>>
>> With leveraging per-cpu ring-array to mitigate the hot spots of memory
>> contention, it could deliver near-linear scalability for high parallel
>> scenarios. The ring-array is compactly managed in a single cache-line
>> to benefit from warmed L1 cache for most cases (<= 4 objects per-core).
>> The body of pre-allocated objects is stored in continuous cache-lines
>> just after the ring-array.
>>
>> The object pool is interrupt safe. Both allocation and reclamation
>> (object pop and push operations) can be preemptible or interruptable.
>>
>> It's best suited for following cases:
>> 1) Memory allocation or reclamation are prohibited or too expensive
>> 2) Consumers are of different priorities, such as irqs and threads
>>
>> Limitations:
>> 1) Maximum objects (capacity) is determined during pool initializing
>> 2) The memory of objects won't be freed until the poll is finalized
>> 3) Object allocation (pop) may fail after trying all cpu slots
>> 4) Object reclamation (push) won't fail but may take long time to
>>     finish for imbalanced scenarios. You can try larger max_entries
>>     to mitigate, or ( >= CPUS * nr_objs) to avoid
>>
>> Signed-off-by: wuqiang <wuqiang.matt@...edance.com>
>> ---
>>   include/linux/objpool.h | 153 +++++++++++++
>>   lib/Makefile            |   2 +-
>>   lib/objpool.c           | 487 ++++++++++++++++++++++++++++++++++++++++
>>   3 files changed, 641 insertions(+), 1 deletion(-)
>>   create mode 100644 include/linux/objpool.h
>>   create mode 100644 lib/objpool.c
>>
>> diff --git a/include/linux/objpool.h b/include/linux/objpool.h
>> new file mode 100644
>> index 000000000000..7899b054b50c
>> --- /dev/null
>> +++ b/include/linux/objpool.h
>> @@ -0,0 +1,153 @@
>> +/* SPDX-License-Identifier: GPL-2.0 */
>> +
>> +#ifndef _LINUX_OBJPOOL_H
>> +#define _LINUX_OBJPOOL_H
>> +
>> +#include <linux/types.h>
>> +
>> +/*
>> + * objpool: ring-array based lockless MPMC queue
>> + *
>> + * Copyright: wuqiang.matt@...edance.com
>> + *
>> + * The object pool is a scalable implementaion of high performance queue
>> + * for objects allocation and reclamation, such as kretprobe instances.
>> + *
>> + * With leveraging per-cpu ring-array to mitigate the hot spots of memory
>> + * contention, it could deliver near-linear scalability for high parallel
>> + * scenarios. The ring-array is compactly managed in a single cache-line
>> + * to benefit from warmed L1 cache for most cases (<= 4 objects per-core).
>> + * The body of pre-allocated objects is stored in continuous cache-lines
>> + * just after the ring-array.
>> + *
>> + * The object pool is interrupt safe. Both allocation and reclamation
>> + * (object pop and push operations) can be preemptible or interruptable.
>> + *
>> + * It's best suited for following cases:
>> + * 1) Memory allocation or reclamation are prohibited or too expensive
>> + * 2) Consumers are of different priorities, such as irqs and threads
>> + *
>> + * Limitations:
>> + * 1) Maximum objects (capacity) is determined during pool initializing
>> + * 2) The memory of objects won't be freed until the poll is finalized
>> + * 3) Object allocation (pop) may fail after trying all cpu slots
>> + * 4) Object reclamation (push) won't fail but may take long time to
>> + *    finish for imbalanced scenarios. You can try larger max_entries
>> + *    to mitigate, or ( >= CPUS * nr_objs) to avoid
>> + */
>> +
>> +/*
>> + * objpool_slot: per-cpu ring array
>> + *
>> + * Represents a cpu-local array-based ring buffer, its size is specialized
>> + * during initialization of object pool.
>> + *
>> + * The objpool_slot is allocated from local memory for NUMA system, and to
>> + * be kept compact in a single cacheline. ages[] is stored just after the
>> + * body of objpool_slot, and then entries[]. The Array of ages[] describes
>> + * revision of each item, solely used to avoid ABA. And array of entries[]
>> + * contains the pointers of objects.
>> + *
>> + * The default size of objpool_slot is a single cache-line, aka. 64 bytes.
>> + *
>> + * 64bit:
>> + *        4      8      12     16        32                 64
>> + * | head | tail | size | mask | ages[4] | ents[4]: (8 * 4) | objects
>> + *
>> + * 32bit:
>> + *        4      8      12     16        32        48       64
>> + * | head | tail | size | mask | ages[4] | ents[4] | unused | objects
>> + *
>> + */
>> +
>> +struct objpool_slot {
>> +	uint32_t                head;	/* head of ring array */
>> +	uint32_t                tail;	/* tail of ring array */
>> +	uint32_t                size;	/* array size, pow of 2 */
>> +	uint32_t                mask;	/* size - 1 */
>> +} __attribute__((packed));
>> +
>> +/* caller-specified object initial callback to setup each object, only called once */
>> +typedef int (*objpool_init_obj_cb)(void *context, void *obj);
> 
> It seems a bit confused that this "initialize object" callback
> don't have the @obj as the first argument.

Sure, will update in next version.

>> +
>> +/* caller-specified cleanup callback for private objects/pool/context */
>> +typedef int (*objpool_release_cb)(void *context, void *ptr, uint32_t flags);
> 
> Do you have any use-case for this release callback?
> If not, until actual use-case comes up, I recommend you to defer
> implementing it.

No actual use-case for now, since both kretprobe and rethook use internal
objects. It's mainly for user-managed objects and asynchronous finilization.

I'll reconsider your advice. Thanks.

>> +
>> +/* called for object releasing: ptr points to an object */
>> +#define OBJPOOL_FLAG_NODE        (0x00000001)
>> +/* for user pool and context releasing, ptr could be NULL */
>> +#define OBJPOOL_FLAG_POOL        (0x00001000)
>> +/* the object or pool to be released is user-managed */
>> +#define OBJPOOL_FLAG_USER        (0x00008000)
> 
> Ditto.
> 
>> +
>> +/*
>> + * objpool_head: object pooling metadata
>> + */
>> +
>> +struct objpool_head {
>> +	unsigned int            obj_size;	/* object & element size */
>> +	unsigned int            nr_objs;	/* total objs (to be pre-allocated) */
>> +	unsigned int            nr_cpus;	/* num of possible cpus */
>> +	unsigned int            capacity;	/* max objects per cpuslot */
>> +	unsigned long           flags;		/* flags for objpool management */
>> +	gfp_t                   gfp;		/* gfp flags for kmalloc & vmalloc */
>> +	unsigned int            pool_size;	/* user pool size in byes */
>> +	void                   *pool;		/* user managed memory pool */
>> +	struct objpool_slot   **cpu_slots;	/* array of percpu slots */
>> +	unsigned int           *slot_sizes;	/* size in bytes of slots */
>> +	objpool_release_cb      release;	/* resource cleanup callback */
>> +	void                   *context;	/* caller-provided context */
>> +};
>> +
>> +#define OBJPOOL_FROM_VMALLOC	(0x800000000)	/* objpool allocated from vmalloc area */
>> +#define OBJPOOL_HAVE_OBJECTS	(0x400000000)	/* objects allocated along with objpool */
> 
> This also doesn't need at this moment. Please start from simple
> design for review.
> 
>> +
>> +/* initialize object pool and pre-allocate objects */
>> +int objpool_init(struct objpool_head *head, unsigned int nr_objs,
>> +		 unsigned int max_objs, unsigned int object_size,
>> +		 gfp_t gfp, void *context, objpool_init_obj_cb objinit,
>> +		 objpool_release_cb release);
>> +
>> +/* add objects in batch from user provided pool */
>> +int objpool_populate(struct objpool_head *head, void *pool,
>> +		     unsigned int size, unsigned int object_size,
>> +		     void *context, objpool_init_obj_cb objinit);
>> +
>> +/* add pre-allocated object (managed by user) to objpool */
>> +int objpool_add(void *obj, struct objpool_head *head);
>> +
>> +/* allocate an object from objects pool */
>> +void *objpool_pop(struct objpool_head *head);
>> +
>> +/* reclaim an object to objects pool */
>> +int objpool_push(void *node, struct objpool_head *head);
>> +
>> +/* cleanup the whole object pool (objects including) */
>> +void objpool_fini(struct objpool_head *head);
>> +
>> +/* whether the object is pre-allocated with percpu slots */
>> +static inline int objpool_is_inslot(void *obj, struct objpool_head *head)
>> +{
>> +	void *slot;
>> +	int i;
>> +
>> +	if (!obj || !(head->flags & OBJPOOL_HAVE_OBJECTS))
>> +		return 0;
>> +
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		slot = head->cpu_slots[i];
>> +		if (obj >= slot && obj < slot + head->slot_sizes[i])
>> +			return 1;
>> +	}
>> +
>> +	return 0;
>> +}
> 
> Ditto.
> 
> It is too complicated to mix the internal allocated objects
> and external ones. This will expose the implementation of the
> objpool (users must understand they have to free the object
> only outside of slot)

Mixing is NOT recommended and normally it should be one of internal / user
oool / external. But as a general lib of objpool, mixing is supported just
because we don't know what use-cases real users would face.

> You can add it afterwards if it is really needed :)

Sure, I'll rethink of it. For kretprobe and rethook, a performance-oriented
version (with internal objects) should be simplely enough and pretty compact.

>> +
>> +/* whether the object is from user pool (batched adding) */
>> +static inline int objpool_is_inpool(void *obj, struct objpool_head *head)
>> +{
>> +	return (obj && head->pool && obj >= head->pool &&
>> +		obj < head->pool + head->pool_size);
>> +}
>> +
>> +#endif /* _LINUX_OBJPOOL_H */
>> diff --git a/lib/Makefile b/lib/Makefile
>> index 161d6a724ff7..e938703a321f 100644
>> --- a/lib/Makefile
>> +++ b/lib/Makefile
>> @@ -34,7 +34,7 @@ lib-y := ctype.o string.o vsprintf.o cmdline.o \
>>   	 is_single_threaded.o plist.o decompress.o kobject_uevent.o \
>>   	 earlycpio.o seq_buf.o siphash.o dec_and_lock.o \
>>   	 nmi_backtrace.o win_minmax.o memcat_p.o \
>> -	 buildid.o
>> +	 buildid.o objpool.o
>>   
>>   lib-$(CONFIG_PRINTK) += dump_stack.o
>>   lib-$(CONFIG_SMP) += cpumask.o
>> diff --git a/lib/objpool.c b/lib/objpool.c
>> new file mode 100644
>> index 000000000000..ecffa0795f3d
>> --- /dev/null
>> +++ b/lib/objpool.c
>> @@ -0,0 +1,487 @@
>> +// SPDX-License-Identifier: GPL-2.0
>> +
>> +#include <linux/objpool.h>
>> +#include <linux/slab.h>
>> +#include <linux/vmalloc.h>
>> +#include <linux/atomic.h>
>> +#include <linux/prefetch.h>
>> +
>> +/*
>> + * objpool: ring-array based lockless MPMC/FIFO queues
>> + *
>> + * Copyright: wuqiang.matt@...edance.com
>> + */
>> +
>> +/* compute the suitable num of objects to be managed by slot */
>> +static inline unsigned int __objpool_num_of_objs(unsigned int size)
>> +{
>> +	return rounddown_pow_of_two((size - sizeof(struct objpool_slot)) /
>> +			(sizeof(uint32_t) + sizeof(void *)));
>> +}
>> +
>> +#define SLOT_AGES(s) ((uint32_t *)((char *)(s) + sizeof(struct objpool_slot)))
>> +#define SLOT_ENTS(s) ((void **)((char *)(s) + sizeof(struct objpool_slot) + \
>> +			sizeof(uint32_t) * (s)->size))
>> +#define SLOT_OBJS(s) ((void *)((char *)(s) + sizeof(struct objpool_slot) + \
>> +			(sizeof(uint32_t) + sizeof(void *)) * (s)->size))
>> +
>> +/* allocate and initialize percpu slots */
>> +static inline int
>> +__objpool_init_percpu_slots(struct objpool_head *head, unsigned int nobjs,
>> +			void *context, objpool_init_obj_cb objinit)
>> +{
>> +	unsigned int i, j, n, size, objsz, nents = head->capacity;
>> +
>> +	/* aligned object size by sizeof(void *) */
>> +	objsz = ALIGN(head->obj_size, sizeof(void *));
>> +	/* shall we allocate objects along with objpool_slot */
>> +	if (objsz)
>> +		head->flags |= OBJPOOL_HAVE_OBJECTS;
>> +
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		struct objpool_slot *os;
>> +
>> +		/* compute how many objects to be managed by this slot */
>> +		n = nobjs / head->nr_cpus;
>> +		if (i < (nobjs % head->nr_cpus))
>> +			n++;
>> +		size = sizeof(struct objpool_slot) + sizeof(void *) * nents +
>> +		       sizeof(uint32_t) * nents + objsz * n;
>> +
>> +		/* decide memory area for cpu-slot allocation */
>> +		if (!i && !(head->gfp & GFP_ATOMIC) && size > PAGE_SIZE / 2)
>> +			head->flags |= OBJPOOL_FROM_VMALLOC;
>> +
>> +		/* allocate percpu slot & objects from local memory */
>> +		if (head->flags & OBJPOOL_FROM_VMALLOC)
>> +			os = __vmalloc_node(size, sizeof(void *), head->gfp,
>> +				cpu_to_node(i), __builtin_return_address(0));
>> +		else
>> +			os = kmalloc_node(size, head->gfp, cpu_to_node(i));
>> +		if (!os)
>> +			return -ENOMEM;
>> +
>> +		/* initialize percpu slot for the i-th cpu */
>> +		memset(os, 0, size);
>> +		os->size = head->capacity;
>> +		os->mask = os->size - 1;
>> +		head->cpu_slots[i] = os;
>> +		head->slot_sizes[i] = size;
>> +
>> +		/*
>> +		 * start from 2nd round to avoid conflict of 1st item.
>> +		 * we assume that the head item is ready for retrieval
>> +		 * iff head is equal to ages[head & mask]. but ages is
>> +		 * initialized as 0, so in view of the caller of pop(),
>> +		 * the 1st item (0th) is always ready, but fact could
>> +		 * be: push() is stalled before the final update, thus
>> +		 * the item being inserted will be lost forever.
>> +		 */
>> +		os->head = os->tail = head->capacity;
>> +
>> +		if (!objsz)
>> +			continue;
>> +
>> +		for (j = 0; j < n; j++) {
>> +			uint32_t *ages = SLOT_AGES(os);
>> +			void **ents = SLOT_ENTS(os);
>> +			void *obj = SLOT_OBJS(os) + j * objsz;
>> +			uint32_t ie = os->tail & os->mask;
>> +
>> +			/* perform object initialization */
>> +			if (objinit) {
>> +				int rc = objinit(context, obj);
>> +				if (rc)
>> +					return rc;
>> +			}
>> +
>> +			/* add obj into the ring array */
>> +			ents[ie] = obj;
>> +			ages[ie] = os->tail;
>> +			os->tail++;
>> +			head->nr_objs++;
>> +		}
>> +	}
>> +
>> +	return 0;
>> +}
>> +
>> +/* cleanup all percpu slots of the object pool */
>> +static inline void __objpool_fini_percpu_slots(struct objpool_head *head)
>> +{
>> +	unsigned int i;
>> +
>> +	if (!head->cpu_slots)
>> +		return;
>> +
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		if (!head->cpu_slots[i])
>> +			continue;
>> +		if (head->flags & OBJPOOL_FROM_VMALLOC)
>> +			vfree(head->cpu_slots[i]);
>> +		else
>> +			kfree(head->cpu_slots[i]);
>> +	}
>> +	kfree(head->cpu_slots);
>> +	head->cpu_slots = NULL;
>> +	head->slot_sizes = NULL;
>> +}
>> +
>> +/**
>> + * objpool_init: initialize object pool and pre-allocate objects
>> + *
>> + * args:
>> + * @head:    the object pool to be initialized, declared by caller
>> + * @nr_objs: total objects to be pre-allocated by this object pool
>> + * @max_objs: max entries (object pool capacity), use nr_objs if 0
>> + * @object_size: size of an object, no objects pre-allocated if 0
>> + * @gfp:     flags for memory allocation (via kmalloc or vmalloc)
>> + * @context: user context for object initialization callback
>> + * @objinit: object initialization callback for extra setting-up
>> + * @release: cleanup callback for private objects/pool/context
>> + *
>> + * return:
>> + *         0 for success, otherwise error code
>> + *
>> + * All pre-allocated objects are to be zeroed. Caller could do extra
>> + * initialization in objinit callback. The objinit callback will be
>> + * called once and only once after the slot allocation. Then objpool
>> + * won't touch any content of the objects since then. It's caller's
>> + * duty to perform reinitialization after object allocation (pop) or
>> + * clearance before object reclamation (push) if required.
>> + */
>> +int objpool_init(struct objpool_head *head, unsigned int nr_objs,
>> +		unsigned int max_objs, unsigned int object_size,
>> +		gfp_t gfp, void *context, objpool_init_obj_cb objinit,
>> +		objpool_release_cb release)
>> +{
>> +	unsigned int nents, ncpus = num_possible_cpus();
>> +	int rc;
>> +
>> +	/* calculate percpu slot size (rounded to pow of 2) */
>> +	if (max_objs < nr_objs)
> 
> This should be an error case.
> 
> 	if (!max_objs)
> 
>> +		max_objs = nr_objs;
> 
> 	else if (max_objs < nr_objs)
> 		return -EINVAL;

Got it.

> But to simplify that, I think it should use only nr_objs.
> I mean, if we can pass the @objinit, there seems no reason to
> have both nr_objs and max_objs.

I kept both them just to give user the flexibility to best meet the
requirements. For cases that objects are in imbalanced distribution
among CPUs, max_objs should be bigger enough (nr_cpus * nr_objs) to
deliver good performance, with a cost of a little more memory.

>> +	nents = max_objs / ncpus;
>> +	if (nents < __objpool_num_of_objs(L1_CACHE_BYTES))
>> +		nents = __objpool_num_of_objs(L1_CACHE_BYTES);
>> +	nents = roundup_pow_of_two(nents);
>> +	while (nents * ncpus < nr_objs)
>> +		nents = nents << 1;
>> +
>> +	memset(head, 0, sizeof(struct objpool_head));
>> +	head->nr_cpus = ncpus;
>> +	head->obj_size = object_size;
>> +	head->capacity = nents;
>> +	head->gfp = gfp & ~__GFP_ZERO;
>> +	head->context = context;
>> +	head->release = release;
>> +
>> +	/* allocate array for percpu slots */
>> +	head->cpu_slots = kzalloc(head->nr_cpus * sizeof(void *) +
>> +			       head->nr_cpus * sizeof(uint32_t), head->gfp);
>> +	if (!head->cpu_slots)
>> +		return -ENOMEM;
>> +	head->slot_sizes = (uint32_t *)&head->cpu_slots[head->nr_cpus];
>> +
>> +	/* initialize per-cpu slots */
>> +	rc = __objpool_init_percpu_slots(head, nr_objs, context, objinit);
>> +	if (rc)
>> +		__objpool_fini_percpu_slots(head);
>> +
>> +	return rc;
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_init);
>> +
>> +/* adding object to slot tail, the given slot must NOT be full */
>> +static inline int __objpool_add_slot(void *obj, struct objpool_slot *os)
>> +{
>> +	uint32_t *ages = SLOT_AGES(os);
>> +	void **ents = SLOT_ENTS(os);
>> +	uint32_t tail = atomic_inc_return((atomic_t *)&os->tail) - 1;
>> +
>> +	WRITE_ONCE(ents[tail & os->mask], obj);
>> +
>> +	/* order matters: obj must be updated before tail updating */
>> +	smp_store_release(&ages[tail & os->mask], tail);
>> +	return 0;
>> +}
>> +
>> +/* adding object to slot, abort if the slot was already full */
>> +static inline int __objpool_try_add_slot(void *obj, struct objpool_slot *os)
>> +{
>> +	uint32_t *ages = SLOT_AGES(os);
>> +	void **ents = SLOT_ENTS(os);
>> +	uint32_t head, tail;
>> +
>> +	do {
>> +		/* perform memory loading for both head and tail */
>> +		head = READ_ONCE(os->head);
>> +		tail = READ_ONCE(os->tail);
>> +		/* just abort if slot is full */
>> +		if (tail >= head + os->size)
>> +			return -ENOENT;
>> +		/* try to extend tail by 1 using CAS to avoid races */
>> +		if (try_cmpxchg_acquire(&os->tail, &tail, tail + 1))
>> +			break;
>> +	} while (1);
>> +
>> +	/* the tail-th of slot is reserved for the given obj */
>> +	WRITE_ONCE(ents[tail & os->mask], obj);
>> +	/* update epoch id to make this object available for pop() */
>> +	smp_store_release(&ages[tail & os->mask], tail);
>> +	return 0;
>> +}
>> +
>> +/**
>> + * objpool_populate: add objects from user provided pool in batch
>> + *
>> + * args:
>> + * @head:  object pool
>> + * @pool: user buffer for pre-allocated objects
>> + * @size: size of user buffer
>> + * @object_size: size of object & element
>> + * @context: user context for objinit callback
>> + * @objinit: object initialization callback
>> + *
>> + * return: 0 or error code
>> + */
>> +int objpool_populate(struct objpool_head *head, void *pool,
>> +		unsigned int size, unsigned int object_size,
>> +		void *context, objpool_init_obj_cb objinit)
>> +{
>> +	unsigned int n = head->nr_objs, used = 0, i;
>> +
>> +	if (head->pool || !pool || size < object_size)
>> +		return -EINVAL;
>> +	if (head->obj_size && head->obj_size != object_size)
>> +		return -EINVAL;
>> +	if (head->context && context && head->context != context)
>> +		return -EINVAL;
>> +	if (head->nr_objs >= head->nr_cpus * head->capacity)
>> +		return -ENOENT;
>> +
>> +	WARN_ON_ONCE(((unsigned long)pool) & (sizeof(void *) - 1));
>> +	WARN_ON_ONCE(((uint32_t)object_size) & (sizeof(void *) - 1));
>> +
>> +	/* align object size by sizeof(void *) */
>> +	head->obj_size = object_size;
>> +	object_size = ALIGN(object_size, sizeof(void *));
>> +	if (object_size == 0)
>> +		return -EINVAL;
>> +
>> +	while (used + object_size <= size) {
>> +		void *obj = pool + used;
>> +
>> +		/* perform object initialization */
>> +		if (objinit) {
>> +			int rc = objinit(context, obj);
>> +			if (rc)
>> +				return rc;
>> +		}
>> +
>> +		/* insert obj to its corresponding objpool slot */
>> +		i = (n + used * head->nr_cpus/size) % head->nr_cpus;
>> +		if (!__objpool_try_add_slot(obj, head->cpu_slots[i]))
>> +			head->nr_objs++;
>> +
>> +		used += object_size;
>> +	}
>> +
>> +	if (!used)
>> +		return -ENOENT;
>> +
>> +	head->context = context;
>> +	head->pool = pool;
>> +	head->pool_size = size;
>> +
>> +	return 0;
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_populate);
>> +
>> +/**
>> + * objpool_add: add pre-allocated object to objpool during pool
>> + * initialization
>> + *
>> + * args:
>> + * @obj:  object pointer to be added to objpool
>> + * @head: object pool to be inserted into
>> + *
>> + * return:
>> + *     0 or error code
>> + *
>> + * objpool_add_node doesn't handle race conditions, can only be
>> + * called during objpool initialization
>> + */
>> +int objpool_add(void *obj, struct objpool_head *head)
>> +{
>> +	unsigned int i, cpu;
>> +
>> +	if (!obj)
>> +		return -EINVAL;
>> +	if (head->nr_objs >= head->nr_cpus * head->capacity)
>> +		return -ENOENT;
>> +
>> +	cpu = head->nr_objs % head->nr_cpus;
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		if (!__objpool_try_add_slot(obj, head->cpu_slots[cpu])) {
>> +			head->nr_objs++;
>> +			return 0;
>> +		}
>> +
>> +		if (++cpu >= head->nr_cpus)
>> +			cpu = 0;
>> +	}
>> +
>> +	return -ENOENT;
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_add);
>> +
>> +/**
>> + * objpool_push: reclaim the object and return back to objects pool
>> + *
>> + * args:
>> + * @obj:  object pointer to be pushed to object pool
>> + * @head: object pool
>> + *
>> + * return:
>> + *     0 or error code: it fails only when objects pool are full
>> + *
>> + * objpool_push is non-blockable, and can be nested
>> + */
>> +int objpool_push(void *obj, struct objpool_head *head)
>> +{
>> +	unsigned int cpu = raw_smp_processor_id() % head->nr_cpus;
>> +
>> +	do {
>> +		if (head->nr_objs > head->capacity) {
>> +			if (!__objpool_try_add_slot(obj, head->cpu_slots[cpu]))
>> +				return 0;
>> +		} else {
>> +			if (!__objpool_add_slot(obj, head->cpu_slots[cpu]))
>> +				return 0;
>> +		}
>> +		if (++cpu >= head->nr_cpus)
>> +			cpu = 0;
>> +	} while (1);
>> +
>> +	return -ENOENT;
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_push);
>> +
>> +/* try to retrieve object from slot */
>> +static inline void *__objpool_try_get_slot(struct objpool_slot *os)
>> +{
>> +	uint32_t *ages = SLOT_AGES(os);
>> +	void **ents = SLOT_ENTS(os);
>> +	/* do memory load of head to local head */
>> +	uint32_t head = smp_load_acquire(&os->head);
>> +
>> +	/* loop if slot isn't empty */
>> +	while (head != READ_ONCE(os->tail)) {
>> +		uint32_t id = head & os->mask, prev = head;
>> +
>> +		/* do prefetching of object ents */
>> +		prefetch(&ents[id]);
>> +
>> +		/*
>> +		 * check whether this item was ready for retrieval ? There's
>> +		 * possibility * in theory * we might retrieve wrong object,
>> +		 * in case ages[id] overflows when current task is sleeping,
>> +		 * but it will take very very long to overflow an uint32_t
>> +		 */
>> +		if (smp_load_acquire(&ages[id]) == head) {
>> +			/* node must have been udpated by push() */
>> +			void *node = READ_ONCE(ents[id]);
>> +			/* commit and move forward head of the slot */
>> +			if (try_cmpxchg_release(&os->head, &head, head + 1))
>> +				return node;
>> +		}
>> +
>> +		/* re-load head from memory continue trying */
>> +		head = READ_ONCE(os->head);
>> +		/*
>> +		 * head stays unchanged, so it's very likely current pop()
>> +		 * just preempted/interrupted an ongoing push() operation
>> +		 */
>> +		if (head == prev)
>> +			break;
>> +	}
>> +
>> +	return NULL;
>> +}
>> +
>> +/**
>> + * objpool_pop: allocate an object from objects pool
>> + *
>> + * args:
>> + * @oh:  object pool
>> + *
>> + * return:
>> + *   object: NULL if failed (object pool is empty)
>> + *
>> + * objpool_pop can be nested, so can be used in any context.
>> + */
>> +void *objpool_pop(struct objpool_head *head)
>> +{
>> +	unsigned int i, cpu;
>> +	void *obj = NULL;
>> +
>> +	cpu = raw_smp_processor_id() % head->nr_cpus;
> 
> (Not sure, do we really need this?)

V6 (this version) needs it, since it's using num_possible_cpus() to manage
the array of cpu_slots.

Last week I finished an improved version, which takes account of holes in
cpu_possible_mask. The testings are going well so far.

> Thank you,

Your advices would be greatly appreciated. Thank you for your time.

> 
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		struct objpool_slot *slot = head->cpu_slots[cpu];
>> +		obj = __objpool_try_get_slot(slot);
>> +		if (obj)
>> +			break;
>> +		if (++cpu >= head->nr_cpus)
>> +			cpu = 0;
>> +	}
>> +
>> +	return obj;
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_pop);
>> +
>> +/**
>> + * objpool_fini: cleanup the whole object pool (releasing all objects)
>> + *
>> + * args:
>> + * @head: object pool to be released
>> + *
>> + */
>> +void objpool_fini(struct objpool_head *head)
>> +{
>> +	uint32_t i, flags;
>> +
>> +	if (!head->cpu_slots)
>> +		return;
>> +
>> +	if (!head->release) {
>> +		__objpool_fini_percpu_slots(head);
>> +		return;
>> +	}
>> +
>> +	/* cleanup all objects remained in objpool */
>> +	for (i = 0; i < head->nr_cpus; i++) {
>> +		void *obj;
>> +		do {
>> +			flags = OBJPOOL_FLAG_NODE;
>> +			obj = __objpool_try_get_slot(head->cpu_slots[i]);
>> +			if (!obj)
>> +				break;
>> +			if (!objpool_is_inpool(obj, head) &&
>> +			    !objpool_is_inslot(obj, head)) {
>> +				flags |= OBJPOOL_FLAG_USER;
>> +			}
>> +			head->release(head->context, obj, flags);
>> +		} while (obj);
>> +	}
>> +
>> +	/* release percpu slots */
>> +	__objpool_fini_percpu_slots(head);
>> +
>> +	/* cleanup user private pool and related context */
>> +	flags = OBJPOOL_FLAG_POOL;
>> +	if (head->pool)
>> +		flags |= OBJPOOL_FLAG_USER;
>> +	head->release(head->context, head->pool, flags);
>> +}
>> +EXPORT_SYMBOL_GPL(objpool_fini);
>> -- 
>> 2.34.1
>>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ