| #ifndef __LINUX_PERCPU_H |
| #define __LINUX_PERCPU_H |
| #include <linux/spinlock.h> /* For preempt_disable() */ |
| #include <linux/slab.h> /* For kmalloc() */ |
| #include <linux/smp.h> |
| #include <linux/string.h> /* For memset() */ |
| #include <asm/percpu.h> |
| |
| /* Enough to cover all DEFINE_PER_CPUs in kernel, including modules. */ |
| #ifndef PERCPU_ENOUGH_ROOM |
| #define PERCPU_ENOUGH_ROOM 32768 |
| #endif |
| |
| /* Must be an lvalue. */ |
| #define get_cpu_var(var) (*({ preempt_disable(); &__get_cpu_var(var); })) |
| #define put_cpu_var(var) preempt_enable() |
| |
| #ifdef CONFIG_SMP |
| |
| struct percpu_data { |
| void *ptrs[NR_CPUS]; |
| void *blkp; |
| }; |
| |
| /* |
| * Use this to get to a cpu's version of the per-cpu object allocated using |
| * alloc_percpu. Non-atomic access to the current CPU's version should |
| * probably be combined with get_cpu()/put_cpu(). |
| */ |
| #define per_cpu_ptr(ptr, cpu) \ |
| ({ \ |
| struct percpu_data *__p = (struct percpu_data *)~(unsigned long)(ptr); \ |
| (__typeof__(ptr))__p->ptrs[(cpu)]; \ |
| }) |
| |
| extern void *__alloc_percpu(size_t size, size_t align); |
| extern void free_percpu(const void *); |
| |
| #else /* CONFIG_SMP */ |
| |
| #define per_cpu_ptr(ptr, cpu) (ptr) |
| |
| static inline void *__alloc_percpu(size_t size, size_t align) |
| { |
| void *ret = kmalloc(size, GFP_KERNEL); |
| if (ret) |
| memset(ret, 0, size); |
| return ret; |
| } |
| static inline void free_percpu(const void *ptr) |
| { |
| kfree(ptr); |
| } |
| |
| #endif /* CONFIG_SMP */ |
| |
| /* Simple wrapper for the common case: zeros memory. */ |
| #define alloc_percpu(type) \ |
| ((type *)(__alloc_percpu(sizeof(type), __alignof__(type)))) |
| |
| #endif /* __LINUX_PERCPU_H */ |