| /* |
| * Quicklist support. |
| * |
| * Quicklists are light weight lists of pages that have a defined state |
| * on alloc and free. Pages must be in the quicklist specific defined state |
| * (zero by default) when the page is freed. It seems that the initial idea |
| * for such lists first came from Dave Miller and then various other people |
| * improved on it. |
| * |
| * Copyright (C) 2007 SGI, |
| * Christoph Lameter <clameter@sgi.com> |
| * Generalized, added support for multiple lists and |
| * constructors / destructors. |
| */ |
| #include <linux/kernel.h> |
| |
| #include <linux/gfp.h> |
| #include <linux/mm.h> |
| #include <linux/mmzone.h> |
| #include <linux/quicklist.h> |
| |
| DEFINE_PER_CPU(struct quicklist [CONFIG_NR_QUICK], quicklist); |
| |
| #define FRACTION_OF_NODE_MEM 16 |
| |
| static unsigned long max_pages(unsigned long min_pages) |
| { |
| unsigned long node_free_pages, max; |
| int node = numa_node_id(); |
| struct zone *zones = NODE_DATA(node)->node_zones; |
| int num_cpus_on_node; |
| |
| node_free_pages = |
| #ifdef CONFIG_ZONE_DMA |
| zone_page_state(&zones[ZONE_DMA], NR_FREE_PAGES) + |
| #endif |
| #ifdef CONFIG_ZONE_DMA32 |
| zone_page_state(&zones[ZONE_DMA32], NR_FREE_PAGES) + |
| #endif |
| zone_page_state(&zones[ZONE_NORMAL], NR_FREE_PAGES); |
| |
| max = node_free_pages / FRACTION_OF_NODE_MEM; |
| |
| num_cpus_on_node = cpumask_weight(cpumask_of_node(node)); |
| max /= num_cpus_on_node; |
| |
| return max(max, min_pages); |
| } |
| |
| static long min_pages_to_free(struct quicklist *q, |
| unsigned long min_pages, long max_free) |
| { |
| long pages_to_free; |
| |
| pages_to_free = q->nr_pages - max_pages(min_pages); |
| |
| return min(pages_to_free, max_free); |
| } |
| |
| /* |
| * Trim down the number of pages in the quicklist |
| */ |
| void quicklist_trim(int nr, void (*dtor)(void *), |
| unsigned long min_pages, unsigned long max_free) |
| { |
| long pages_to_free; |
| struct quicklist *q; |
| |
| q = &get_cpu_var(quicklist)[nr]; |
| if (q->nr_pages > min_pages) { |
| pages_to_free = min_pages_to_free(q, min_pages, max_free); |
| |
| while (pages_to_free > 0) { |
| /* |
| * We pass a gfp_t of 0 to quicklist_alloc here |
| * because we will never call into the page allocator. |
| */ |
| void *p = quicklist_alloc(nr, 0, NULL); |
| |
| if (dtor) |
| dtor(p); |
| free_page((unsigned long)p); |
| pages_to_free--; |
| } |
| } |
| put_cpu_var(quicklist); |
| } |
| |
| unsigned long quicklist_total_size(void) |
| { |
| unsigned long count = 0; |
| int cpu; |
| struct quicklist *ql, *q; |
| |
| for_each_online_cpu(cpu) { |
| ql = per_cpu(quicklist, cpu); |
| for (q = ql; q < ql + CONFIG_NR_QUICK; q++) |
| count += q->nr_pages; |
| } |
| return count; |
| } |
| |