| // SPDX-License-Identifier: GPL-2.0 |
| |
| #include <linux/objpool.h> |
| #include <linux/slab.h> |
| #include <linux/vmalloc.h> |
| #include <linux/atomic.h> |
| #include <linux/irqflags.h> |
| #include <linux/cpumask.h> |
| #include <linux/log2.h> |
| |
| /* |
| * objpool: ring-array based lockless MPMC/FIFO queues |
| * |
| * Copyright: wuqiang.matt@bytedance.com,mhiramat@kernel.org |
| */ |
| |
| /* initialize percpu objpool_slot */ |
| static int |
| objpool_init_percpu_slot(struct objpool_head *pool, |
| struct objpool_slot *slot, |
| int nodes, void *context, |
| objpool_init_obj_cb objinit) |
| { |
| void *obj = (void *)&slot->entries[pool->capacity]; |
| int i; |
| |
| /* initialize elements of percpu objpool_slot */ |
| slot->mask = pool->capacity - 1; |
| |
| for (i = 0; i < nodes; i++) { |
| if (objinit) { |
| int rc = objinit(obj, context); |
| if (rc) |
| return rc; |
| } |
| slot->entries[slot->tail & slot->mask] = obj; |
| obj = obj + pool->obj_size; |
| slot->tail++; |
| slot->last = slot->tail; |
| pool->nr_objs++; |
| } |
| |
| return 0; |
| } |
| |
| /* allocate and initialize percpu slots */ |
| static int |
| objpool_init_percpu_slots(struct objpool_head *pool, int nr_objs, |
| void *context, objpool_init_obj_cb objinit) |
| { |
| int i, cpu_count = 0; |
| |
| for (i = 0; i < pool->nr_cpus; i++) { |
| |
| struct objpool_slot *slot; |
| int nodes, size, rc; |
| |
| /* skip the cpu node which could never be present */ |
| if (!cpu_possible(i)) |
| continue; |
| |
| /* compute how many objects to be allocated with this slot */ |
| nodes = nr_objs / num_possible_cpus(); |
| if (cpu_count < (nr_objs % num_possible_cpus())) |
| nodes++; |
| cpu_count++; |
| |
| size = struct_size(slot, entries, pool->capacity) + |
| pool->obj_size * nodes; |
| |
| /* |
| * here we allocate percpu-slot & objs together in a single |
| * allocation to make it more compact, taking advantage of |
| * warm caches and TLB hits. in default vmalloc is used to |
| * reduce the pressure of kernel slab system. as we know, |
| * mimimal size of vmalloc is one page since vmalloc would |
| * always align the requested size to page size |
| */ |
| if (pool->gfp & GFP_ATOMIC) |
| slot = kmalloc_node(size, pool->gfp, cpu_to_node(i)); |
| else |
| slot = __vmalloc_node(size, sizeof(void *), pool->gfp, |
| cpu_to_node(i), __builtin_return_address(0)); |
| if (!slot) |
| return -ENOMEM; |
| memset(slot, 0, size); |
| pool->cpu_slots[i] = slot; |
| |
| /* initialize the objpool_slot of cpu node i */ |
| rc = objpool_init_percpu_slot(pool, slot, nodes, context, objinit); |
| if (rc) |
| return rc; |
| } |
| |
| return 0; |
| } |
| |
| /* cleanup all percpu slots of the object pool */ |
| static void objpool_fini_percpu_slots(struct objpool_head *pool) |
| { |
| int i; |
| |
| if (!pool->cpu_slots) |
| return; |
| |
| for (i = 0; i < pool->nr_cpus; i++) |
| kvfree(pool->cpu_slots[i]); |
| kfree(pool->cpu_slots); |
| } |
| |
| /* initialize object pool and pre-allocate objects */ |
| int objpool_init(struct objpool_head *pool, int nr_objs, int object_size, |
| gfp_t gfp, void *context, objpool_init_obj_cb objinit, |
| objpool_fini_cb release) |
| { |
| int rc, capacity, slot_size; |
| |
| /* check input parameters */ |
| if (nr_objs <= 0 || nr_objs > OBJPOOL_NR_OBJECT_MAX || |
| object_size <= 0 || object_size > OBJPOOL_OBJECT_SIZE_MAX) |
| return -EINVAL; |
| |
| /* align up to unsigned long size */ |
| object_size = ALIGN(object_size, sizeof(long)); |
| |
| /* calculate capacity of percpu objpool_slot */ |
| capacity = roundup_pow_of_two(nr_objs); |
| if (!capacity) |
| return -EINVAL; |
| |
| /* initialize objpool pool */ |
| memset(pool, 0, sizeof(struct objpool_head)); |
| pool->nr_cpus = nr_cpu_ids; |
| pool->obj_size = object_size; |
| pool->capacity = capacity; |
| pool->gfp = gfp & ~__GFP_ZERO; |
| pool->context = context; |
| pool->release = release; |
| slot_size = pool->nr_cpus * sizeof(struct objpool_slot); |
| pool->cpu_slots = kzalloc(slot_size, pool->gfp); |
| if (!pool->cpu_slots) |
| return -ENOMEM; |
| |
| /* initialize per-cpu slots */ |
| rc = objpool_init_percpu_slots(pool, nr_objs, context, objinit); |
| if (rc) |
| objpool_fini_percpu_slots(pool); |
| else |
| refcount_set(&pool->ref, pool->nr_objs + 1); |
| |
| return rc; |
| } |
| EXPORT_SYMBOL_GPL(objpool_init); |
| |
| /* adding object to slot, abort if the slot was already full */ |
| static inline int |
| objpool_try_add_slot(void *obj, struct objpool_head *pool, int cpu) |
| { |
| struct objpool_slot *slot = pool->cpu_slots[cpu]; |
| uint32_t head, tail; |
| |
| /* loading tail and head as a local snapshot, tail first */ |
| tail = READ_ONCE(slot->tail); |
| |
| do { |
| head = READ_ONCE(slot->head); |
| /* fault caught: something must be wrong */ |
| WARN_ON_ONCE(tail - head > pool->nr_objs); |
| } while (!try_cmpxchg_acquire(&slot->tail, &tail, tail + 1)); |
| |
| /* now the tail position is reserved for the given obj */ |
| WRITE_ONCE(slot->entries[tail & slot->mask], obj); |
| /* update sequence to make this obj available for pop() */ |
| smp_store_release(&slot->last, tail + 1); |
| |
| return 0; |
| } |
| |
| /* reclaim an object to object pool */ |
| int objpool_push(void *obj, struct objpool_head *pool) |
| { |
| unsigned long flags; |
| int rc; |
| |
| /* disable local irq to avoid preemption & interruption */ |
| raw_local_irq_save(flags); |
| rc = objpool_try_add_slot(obj, pool, raw_smp_processor_id()); |
| raw_local_irq_restore(flags); |
| |
| return rc; |
| } |
| EXPORT_SYMBOL_GPL(objpool_push); |
| |
| /* try to retrieve object from slot */ |
| static inline void *objpool_try_get_slot(struct objpool_head *pool, int cpu) |
| { |
| struct objpool_slot *slot = pool->cpu_slots[cpu]; |
| /* load head snapshot, other cpus may change it */ |
| uint32_t head = smp_load_acquire(&slot->head); |
| |
| while (head != READ_ONCE(slot->last)) { |
| void *obj; |
| |
| /* obj must be retrieved before moving forward head */ |
| obj = READ_ONCE(slot->entries[head & slot->mask]); |
| |
| /* move head forward to mark it's consumption */ |
| if (try_cmpxchg_release(&slot->head, &head, head + 1)) |
| return obj; |
| } |
| |
| return NULL; |
| } |
| |
| /* allocate an object from object pool */ |
| void *objpool_pop(struct objpool_head *pool) |
| { |
| void *obj = NULL; |
| unsigned long flags; |
| int i, cpu; |
| |
| /* disable local irq to avoid preemption & interruption */ |
| raw_local_irq_save(flags); |
| |
| cpu = raw_smp_processor_id(); |
| for (i = 0; i < num_possible_cpus(); i++) { |
| obj = objpool_try_get_slot(pool, cpu); |
| if (obj) |
| break; |
| cpu = cpumask_next_wrap(cpu, cpu_possible_mask, -1, 1); |
| } |
| raw_local_irq_restore(flags); |
| |
| return obj; |
| } |
| EXPORT_SYMBOL_GPL(objpool_pop); |
| |
| /* release whole objpool forcely */ |
| void objpool_free(struct objpool_head *pool) |
| { |
| if (!pool->cpu_slots) |
| return; |
| |
| /* release percpu slots */ |
| objpool_fini_percpu_slots(pool); |
| |
| /* call user's cleanup callback if provided */ |
| if (pool->release) |
| pool->release(pool, pool->context); |
| } |
| EXPORT_SYMBOL_GPL(objpool_free); |
| |
| /* drop the allocated object, rather reclaim it to objpool */ |
| int objpool_drop(void *obj, struct objpool_head *pool) |
| { |
| if (!obj || !pool) |
| return -EINVAL; |
| |
| if (refcount_dec_and_test(&pool->ref)) { |
| objpool_free(pool); |
| return 0; |
| } |
| |
| return -EAGAIN; |
| } |
| EXPORT_SYMBOL_GPL(objpool_drop); |
| |
| /* drop unused objects and defref objpool for releasing */ |
| void objpool_fini(struct objpool_head *pool) |
| { |
| int count = 1; /* extra ref for objpool itself */ |
| |
| /* drop all remained objects from objpool */ |
| while (objpool_pop(pool)) |
| count++; |
| |
| if (refcount_sub_and_test(count, &pool->ref)) |
| objpool_free(pool); |
| } |
| EXPORT_SYMBOL_GPL(objpool_fini); |