| /* |
| * Basic general purpose allocator for managing special purpose memory |
| * not managed by the regular kmalloc/kfree interface. |
| * Uses for this includes on-device special memory, uncached memory |
| * etc. |
| * |
| * This code is based on the buddy allocator found in the sym53c8xx_2 |
| * driver Copyright (C) 1999-2001 Gerard Roudier <groudier@free.fr>, |
| * and adapted for general purpose use. |
| * |
| * Copyright 2005 (C) Jes Sorensen <jes@trained-monkey.org> |
| * |
| * This source code is licensed under the GNU General Public License, |
| * Version 2. See the file COPYING for more details. |
| */ |
| |
| #include <linux/module.h> |
| #include <linux/stddef.h> |
| #include <linux/kernel.h> |
| #include <linux/string.h> |
| #include <linux/slab.h> |
| #include <linux/init.h> |
| #include <linux/mm.h> |
| #include <linux/spinlock.h> |
| #include <linux/genalloc.h> |
| |
| #include <asm/page.h> |
| |
| |
| struct gen_pool *gen_pool_create(int nr_chunks, int max_chunk_shift, |
| unsigned long (*fp)(struct gen_pool *), |
| unsigned long data) |
| { |
| struct gen_pool *poolp; |
| unsigned long tmp; |
| int i; |
| |
| /* |
| * This is really an arbitrary limit, +10 is enough for |
| * IA64_GRANULE_SHIFT, aka 16MB. If anyone needs a large limit |
| * this can be increased without problems. |
| */ |
| if ((max_chunk_shift > (PAGE_SHIFT + 10)) || |
| ((max_chunk_shift < ALLOC_MIN_SHIFT) && max_chunk_shift)) |
| return NULL; |
| |
| if (!max_chunk_shift) |
| max_chunk_shift = PAGE_SHIFT; |
| |
| poolp = kmalloc(sizeof(struct gen_pool), GFP_KERNEL); |
| if (!poolp) |
| return NULL; |
| memset(poolp, 0, sizeof(struct gen_pool)); |
| poolp->h = kmalloc(sizeof(struct gen_pool_link) * |
| (max_chunk_shift - ALLOC_MIN_SHIFT + 1), |
| GFP_KERNEL); |
| if (!poolp->h) { |
| printk(KERN_WARNING "gen_pool_alloc() failed to allocate\n"); |
| kfree(poolp); |
| return NULL; |
| } |
| memset(poolp->h, 0, sizeof(struct gen_pool_link) * |
| (max_chunk_shift - ALLOC_MIN_SHIFT + 1)); |
| |
| spin_lock_init(&poolp->lock); |
| poolp->get_new_chunk = fp; |
| poolp->max_chunk_shift = max_chunk_shift; |
| poolp->private = data; |
| |
| for (i = 0; i < nr_chunks; i++) { |
| tmp = poolp->get_new_chunk(poolp); |
| printk(KERN_INFO "allocated %lx\n", tmp); |
| if (!tmp) |
| break; |
| gen_pool_free(poolp, tmp, (1 << poolp->max_chunk_shift)); |
| } |
| |
| return poolp; |
| } |
| EXPORT_SYMBOL(gen_pool_create); |
| |
| |
| /* |
| * Simple power of two buddy-like generic allocator. |
| * Provides naturally aligned memory chunks. |
| */ |
| unsigned long gen_pool_alloc(struct gen_pool *poolp, int size) |
| { |
| int j, i, s, max_chunk_size; |
| unsigned long a, flags; |
| struct gen_pool_link *h = poolp->h; |
| |
| max_chunk_size = 1 << poolp->max_chunk_shift; |
| |
| if (size > max_chunk_size) |
| return 0; |
| |
| i = 0; |
| |
| size = max(size, 1 << ALLOC_MIN_SHIFT); |
| s = roundup_pow_of_two(size); |
| |
| j = i; |
| |
| spin_lock_irqsave(&poolp->lock, flags); |
| while (!h[j].next) { |
| if (s == max_chunk_size) { |
| struct gen_pool_link *ptr; |
| spin_unlock_irqrestore(&poolp->lock, flags); |
| ptr = (struct gen_pool_link *)poolp->get_new_chunk(poolp); |
| spin_lock_irqsave(&poolp->lock, flags); |
| h[j].next = ptr; |
| if (h[j].next) |
| h[j].next->next = NULL; |
| break; |
| } |
| j++; |
| s <<= 1; |
| } |
| a = (unsigned long) h[j].next; |
| if (a) { |
| h[j].next = h[j].next->next; |
| /* |
| * This should be split into a seperate function doing |
| * the chunk split in order to support custom |
| * handling memory not physically accessible by host |
| */ |
| while (j > i) { |
| j -= 1; |
| s >>= 1; |
| h[j].next = (struct gen_pool_link *) (a + s); |
| h[j].next->next = NULL; |
| } |
| } |
| spin_unlock_irqrestore(&poolp->lock, flags); |
| return a; |
| } |
| EXPORT_SYMBOL(gen_pool_alloc); |
| |
| |
| /* |
| * Counter-part of the generic allocator. |
| */ |
| void gen_pool_free(struct gen_pool *poolp, unsigned long ptr, int size) |
| { |
| struct gen_pool_link *q; |
| struct gen_pool_link *h = poolp->h; |
| unsigned long a, b, flags; |
| int i, s, max_chunk_size; |
| |
| max_chunk_size = 1 << poolp->max_chunk_shift; |
| |
| if (size > max_chunk_size) |
| return; |
| |
| i = 0; |
| |
| size = max(size, 1 << ALLOC_MIN_SHIFT); |
| s = roundup_pow_of_two(size); |
| |
| a = ptr; |
| |
| spin_lock_irqsave(&poolp->lock, flags); |
| while (1) { |
| if (s == max_chunk_size) { |
| ((struct gen_pool_link *)a)->next = h[i].next; |
| h[i].next = (struct gen_pool_link *)a; |
| break; |
| } |
| b = a ^ s; |
| q = &h[i]; |
| |
| while (q->next && q->next != (struct gen_pool_link *)b) |
| q = q->next; |
| |
| if (!q->next) { |
| ((struct gen_pool_link *)a)->next = h[i].next; |
| h[i].next = (struct gen_pool_link *)a; |
| break; |
| } |
| q->next = q->next->next; |
| a = a & b; |
| s <<= 1; |
| i++; |
| } |
| spin_unlock_irqrestore(&poolp->lock, flags); |
| } |
| EXPORT_SYMBOL(gen_pool_free); |