// SPDX-License-Identifier: GPL-2.0 #include #include #include #include #include #include #include /* * objpool: ring-array based lockless MPMC/FIFO queues * * Copyright: wuqiang.matt@bytedance.com,mhiramat@kernel.org */ /* initialize percpu objpool_slot */ static int objpool_init_percpu_slot(struct objpool_head *pool, struct objpool_slot *slot, int nodes, void *context, objpool_init_obj_cb objinit) { void *obj = (void *)&slot->entries[pool->capacity]; int i; /* initialize elements of percpu objpool_slot */ slot->mask = pool->capacity - 1; for (i = 0; i < nodes; i++) { if (objinit) { int rc = objinit(obj, context); if (rc) return rc; } slot->entries[slot->tail & slot->mask] = obj; obj = obj + pool->obj_size; slot->tail++; slot->last = slot->tail; pool->nr_objs++; } return 0; } /* allocate and initialize percpu slots */ static int objpool_init_percpu_slots(struct objpool_head *pool, int nr_objs, void *context, objpool_init_obj_cb objinit) { int i, cpu_count = 0; for (i = 0; i < pool->nr_cpus; i++) { struct objpool_slot *slot; int nodes, size, rc; /* skip the cpu node which could never be present */ if (!cpu_possible(i)) continue; /* compute how many objects to be allocated with this slot */ nodes = nr_objs / num_possible_cpus(); if (cpu_count < (nr_objs % num_possible_cpus())) nodes++; cpu_count++; size = struct_size(slot, entries, pool->capacity) + pool->obj_size * nodes; /* * here we allocate percpu-slot & objs together in a single * allocation to make it more compact, taking advantage of * warm caches and TLB hits. in default vmalloc is used to * reduce the pressure of kernel slab system. as we know, * mimimal size of vmalloc is one page since vmalloc would * always align the requested size to page size */ if (pool->gfp & GFP_ATOMIC) slot = kmalloc_node(size, pool->gfp, cpu_to_node(i)); else slot = __vmalloc_node(size, sizeof(void *), pool->gfp, cpu_to_node(i), __builtin_return_address(0)); if (!slot) return -ENOMEM; memset(slot, 0, size); pool->cpu_slots[i] = slot; /* initialize the objpool_slot of cpu node i */ rc = objpool_init_percpu_slot(pool, slot, nodes, context, objinit); if (rc) return rc; } return 0; } /* cleanup all percpu slots of the object pool */ static void objpool_fini_percpu_slots(struct objpool_head *pool) { int i; if (!pool->cpu_slots) return; for (i = 0; i < pool->nr_cpus; i++) kvfree(pool->cpu_slots[i]); kfree(pool->cpu_slots); } /* initialize object pool and pre-allocate objects */ int objpool_init(struct objpool_head *pool, int nr_objs, int object_size, gfp_t gfp, void *context, objpool_init_obj_cb objinit, objpool_fini_cb release) { int rc, capacity, slot_size; /* check input parameters */ if (nr_objs <= 0 || nr_objs > OBJPOOL_NR_OBJECT_MAX || object_size <= 0 || object_size > OBJPOOL_OBJECT_SIZE_MAX) return -EINVAL; /* align up to unsigned long size */ object_size = ALIGN(object_size, sizeof(long)); /* calculate capacity of percpu objpool_slot */ capacity = roundup_pow_of_two(nr_objs); if (!capacity) return -EINVAL; /* initialize objpool pool */ memset(pool, 0, sizeof(struct objpool_head)); pool->nr_cpus = nr_cpu_ids; pool->obj_size = object_size; pool->capacity = capacity; pool->gfp = gfp & ~__GFP_ZERO; pool->context = context; pool->release = release; slot_size = pool->nr_cpus * sizeof(struct objpool_slot); pool->cpu_slots = kzalloc(slot_size, pool->gfp); if (!pool->cpu_slots) return -ENOMEM; /* initialize per-cpu slots */ rc = objpool_init_percpu_slots(pool, nr_objs, context, objinit); if (rc) objpool_fini_percpu_slots(pool); else refcount_set(&pool->ref, pool->nr_objs + 1); return rc; } EXPORT_SYMBOL_GPL(objpool_init); /* adding object to slot, abort if the slot was already full */ static inline int objpool_try_add_slot(void *obj, struct objpool_head *pool, int cpu) { struct objpool_slot *slot = pool->cpu_slots[cpu]; uint32_t head, tail; /* loading tail and head as a local snapshot, tail first */ tail = READ_ONCE(slot->tail); do { head = READ_ONCE(slot->head); /* fault caught: something must be wrong */ WARN_ON_ONCE(tail - head > pool->nr_objs); } while (!try_cmpxchg_acquire(&slot->tail, &tail, tail + 1)); /* now the tail position is reserved for the given obj */ WRITE_ONCE(slot->entries[tail & slot->mask], obj); /* update sequence to make this obj available for pop() */ smp_store_release(&slot->last, tail + 1); return 0; } /* reclaim an object to object pool */ int objpool_push(void *obj, struct objpool_head *pool) { unsigned long flags; int rc; /* disable local irq to avoid preemption & interruption */ raw_local_irq_save(flags); rc = objpool_try_add_slot(obj, pool, raw_smp_processor_id()); raw_local_irq_restore(flags); return rc; } EXPORT_SYMBOL_GPL(objpool_push); /* try to retrieve object from slot */ static inline void *objpool_try_get_slot(struct objpool_head *pool, int cpu) { struct objpool_slot *slot = pool->cpu_slots[cpu]; /* load head snapshot, other cpus may change it */ uint32_t head = smp_load_acquire(&slot->head); while (head != READ_ONCE(slot->last)) { void *obj; /* * data visibility of 'last' and 'head' could be out of * order since memory updating of 'last' and 'head' are * performed in push() and pop() independently * * before any retrieving attempts, pop() must guarantee * 'last' is behind 'head', that is to say, there must * be available objects in slot, which could be ensured * by condition 'last != head && last - head <= nr_objs' * that is equivalent to 'last - head - 1 < nr_objs' as * 'last' and 'head' are both unsigned int32 */ if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) { head = READ_ONCE(slot->head); continue; } /* obj must be retrieved before moving forward head */ obj = READ_ONCE(slot->entries[head & slot->mask]); /* move head forward to mark it's consumption */ if (try_cmpxchg_release(&slot->head, &head, head + 1)) return obj; } return NULL; } /* allocate an object from object pool */ void *objpool_pop(struct objpool_head *pool) { void *obj = NULL; unsigned long flags; int i, cpu; /* disable local irq to avoid preemption & interruption */ raw_local_irq_save(flags); cpu = raw_smp_processor_id(); for (i = 0; i < num_possible_cpus(); i++) { obj = objpool_try_get_slot(pool, cpu); if (obj) break; cpu = cpumask_next_wrap(cpu, cpu_possible_mask, -1, 1); } raw_local_irq_restore(flags); return obj; } EXPORT_SYMBOL_GPL(objpool_pop); /* release whole objpool forcely */ void objpool_free(struct objpool_head *pool) { if (!pool->cpu_slots) return; /* release percpu slots */ objpool_fini_percpu_slots(pool); /* call user's cleanup callback if provided */ if (pool->release) pool->release(pool, pool->context); } EXPORT_SYMBOL_GPL(objpool_free); /* drop the allocated object, rather reclaim it to objpool */ int objpool_drop(void *obj, struct objpool_head *pool) { if (!obj || !pool) return -EINVAL; if (refcount_dec_and_test(&pool->ref)) { objpool_free(pool); return 0; } return -EAGAIN; } EXPORT_SYMBOL_GPL(objpool_drop); /* drop unused objects and defref objpool for releasing */ void objpool_fini(struct objpool_head *pool) { int count = 1; /* extra ref for objpool itself */ /* drop all remained objects from objpool */ while (objpool_pop(pool)) count++; if (refcount_sub_and_test(count, &pool->ref)) objpool_free(pool); } EXPORT_SYMBOL_GPL(objpool_fini);