| #ifndef __PERCPU_IDA_H__ |
| #define __PERCPU_IDA_H__ |
| |
| #include <linux/types.h> |
| #include <linux/bitops.h> |
| #include <linux/init.h> |
| #include <linux/sched.h> |
| #include <linux/spinlock_types.h> |
| #include <linux/wait.h> |
| #include <linux/cpumask.h> |
| |
| struct percpu_ida_cpu; |
| |
| struct percpu_ida { |
| /* |
| * number of tags available to be allocated, as passed to |
| * percpu_ida_init() |
| */ |
| unsigned nr_tags; |
| unsigned percpu_max_size; |
| unsigned percpu_batch_size; |
| |
| struct percpu_ida_cpu __percpu *tag_cpu; |
| |
| /* |
| * Bitmap of cpus that (may) have tags on their percpu freelists: |
| * steal_tags() uses this to decide when to steal tags, and which cpus |
| * to try stealing from. |
| * |
| * It's ok for a freelist to be empty when its bit is set - steal_tags() |
| * will just keep looking - but the bitmap _must_ be set whenever a |
| * percpu freelist does have tags. |
| */ |
| cpumask_t cpus_have_tags; |
| |
| struct { |
| spinlock_t lock; |
| /* |
| * When we go to steal tags from another cpu (see steal_tags()), |
| * we want to pick a cpu at random. Cycling through them every |
| * time we steal is a bit easier and more or less equivalent: |
| */ |
| unsigned cpu_last_stolen; |
| |
| /* For sleeping on allocation failure */ |
| wait_queue_head_t wait; |
| |
| /* |
| * Global freelist - it's a stack where nr_free points to the |
| * top |
| */ |
| unsigned nr_free; |
| unsigned *freelist; |
| } ____cacheline_aligned_in_smp; |
| }; |
| |
| /* |
| * Number of tags we move between the percpu freelist and the global freelist at |
| * a time |
| */ |
| #define IDA_DEFAULT_PCPU_BATCH_MOVE 32U |
| /* Max size of percpu freelist, */ |
| #define IDA_DEFAULT_PCPU_SIZE ((IDA_DEFAULT_PCPU_BATCH_MOVE * 3) / 2) |
| |
| int percpu_ida_alloc(struct percpu_ida *pool, int state); |
| void percpu_ida_free(struct percpu_ida *pool, unsigned tag); |
| |
| void percpu_ida_destroy(struct percpu_ida *pool); |
| int __percpu_ida_init(struct percpu_ida *pool, unsigned long nr_tags, |
| unsigned long max_size, unsigned long batch_size); |
| static inline int percpu_ida_init(struct percpu_ida *pool, unsigned long nr_tags) |
| { |
| return __percpu_ida_init(pool, nr_tags, IDA_DEFAULT_PCPU_SIZE, |
| IDA_DEFAULT_PCPU_BATCH_MOVE); |
| } |
| |
| typedef int (*percpu_ida_cb)(unsigned, void *); |
| int percpu_ida_for_each_free(struct percpu_ida *pool, percpu_ida_cb fn, |
| void *data); |
| |
| unsigned percpu_ida_free_tags(struct percpu_ida *pool, int cpu); |
| #endif /* __PERCPU_IDA_H__ */ |