Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Quicklist support. |
| 3 | * |
| 4 | * Quicklists are light weight lists of pages that have a defined state |
| 5 | * on alloc and free. Pages must be in the quicklist specific defined state |
| 6 | * (zero by default) when the page is freed. It seems that the initial idea |
| 7 | * for such lists first came from Dave Miller and then various other people |
| 8 | * improved on it. |
| 9 | * |
| 10 | * Copyright (C) 2007 SGI, |
| 11 | * Christoph Lameter <clameter@sgi.com> |
| 12 | * Generalized, added support for multiple lists and |
| 13 | * constructors / destructors. |
| 14 | */ |
| 15 | #include <linux/kernel.h> |
| 16 | |
| 17 | #include <linux/mm.h> |
| 18 | #include <linux/mmzone.h> |
| 19 | #include <linux/module.h> |
| 20 | #include <linux/quicklist.h> |
| 21 | |
| 22 | DEFINE_PER_CPU(struct quicklist, quicklist)[CONFIG_NR_QUICK]; |
| 23 | |
| 24 | #define FRACTION_OF_NODE_MEM 16 |
| 25 | |
| 26 | static unsigned long max_pages(unsigned long min_pages) |
| 27 | { |
| 28 | unsigned long node_free_pages, max; |
Christoph Lameter | 96990a4 | 2008-01-14 00:55:14 -0800 | [diff] [blame] | 29 | struct zone *zones = NODE_DATA(numa_node_id())->node_zones; |
Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 30 | |
Christoph Lameter | 96990a4 | 2008-01-14 00:55:14 -0800 | [diff] [blame] | 31 | node_free_pages = |
| 32 | #ifdef CONFIG_ZONE_DMA |
| 33 | zone_page_state(&zones[ZONE_DMA], NR_FREE_PAGES) + |
| 34 | #endif |
| 35 | #ifdef CONFIG_ZONE_DMA32 |
| 36 | zone_page_state(&zones[ZONE_DMA32], NR_FREE_PAGES) + |
| 37 | #endif |
| 38 | zone_page_state(&zones[ZONE_NORMAL], NR_FREE_PAGES); |
| 39 | |
Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 40 | max = node_free_pages / FRACTION_OF_NODE_MEM; |
| 41 | return max(max, min_pages); |
| 42 | } |
| 43 | |
| 44 | static long min_pages_to_free(struct quicklist *q, |
| 45 | unsigned long min_pages, long max_free) |
| 46 | { |
| 47 | long pages_to_free; |
| 48 | |
| 49 | pages_to_free = q->nr_pages - max_pages(min_pages); |
| 50 | |
| 51 | return min(pages_to_free, max_free); |
| 52 | } |
| 53 | |
| 54 | /* |
| 55 | * Trim down the number of pages in the quicklist |
| 56 | */ |
| 57 | void quicklist_trim(int nr, void (*dtor)(void *), |
| 58 | unsigned long min_pages, unsigned long max_free) |
| 59 | { |
| 60 | long pages_to_free; |
| 61 | struct quicklist *q; |
| 62 | |
| 63 | q = &get_cpu_var(quicklist)[nr]; |
| 64 | if (q->nr_pages > min_pages) { |
| 65 | pages_to_free = min_pages_to_free(q, min_pages, max_free); |
| 66 | |
| 67 | while (pages_to_free > 0) { |
| 68 | /* |
| 69 | * We pass a gfp_t of 0 to quicklist_alloc here |
| 70 | * because we will never call into the page allocator. |
| 71 | */ |
| 72 | void *p = quicklist_alloc(nr, 0, NULL); |
| 73 | |
| 74 | if (dtor) |
| 75 | dtor(p); |
| 76 | free_page((unsigned long)p); |
| 77 | pages_to_free--; |
| 78 | } |
| 79 | } |
| 80 | put_cpu_var(quicklist); |
| 81 | } |
| 82 | |
| 83 | unsigned long quicklist_total_size(void) |
| 84 | { |
| 85 | unsigned long count = 0; |
| 86 | int cpu; |
| 87 | struct quicklist *ql, *q; |
| 88 | |
| 89 | for_each_online_cpu(cpu) { |
| 90 | ql = per_cpu(quicklist, cpu); |
| 91 | for (q = ql; q < ql + CONFIG_NR_QUICK; q++) |
| 92 | count += q->nr_pages; |
| 93 | } |
| 94 | return count; |
| 95 | } |
| 96 | |