| /* |
| * Discontiguous memory support, Kanoj Sarcar, SGI, Nov 1999 |
| */ |
| #ifndef _LINUX_BOOTMEM_H |
| #define _LINUX_BOOTMEM_H |
| |
| #include <linux/mmzone.h> |
| #include <linux/mm_types.h> |
| #include <asm/dma.h> |
| |
| /* |
| * simple boot-time physical memory area allocator. |
| */ |
| |
| extern unsigned long max_low_pfn; |
| extern unsigned long min_low_pfn; |
| |
| /* |
| * highest page |
| */ |
| extern unsigned long max_pfn; |
| |
| #ifndef CONFIG_NO_BOOTMEM |
| /* |
| * node_bootmem_map is a map pointer - the bits represent all physical |
| * memory pages (including holes) on the node. |
| */ |
| typedef struct bootmem_data { |
| unsigned long node_min_pfn; |
| unsigned long node_low_pfn; |
| void *node_bootmem_map; |
| unsigned long last_end_off; |
| unsigned long hint_idx; |
| struct list_head list; |
| } bootmem_data_t; |
| |
| extern bootmem_data_t bootmem_node_data[]; |
| #endif |
| |
| extern unsigned long bootmem_bootmap_pages(unsigned long); |
| |
| extern unsigned long init_bootmem_node(pg_data_t *pgdat, |
| unsigned long freepfn, |
| unsigned long startpfn, |
| unsigned long endpfn); |
| extern unsigned long init_bootmem(unsigned long addr, unsigned long memend); |
| |
| extern unsigned long free_all_bootmem(void); |
| extern void reset_all_zones_managed_pages(void); |
| |
| extern void free_bootmem_node(pg_data_t *pgdat, |
| unsigned long addr, |
| unsigned long size); |
| extern void free_bootmem(unsigned long physaddr, unsigned long size); |
| extern void free_bootmem_late(unsigned long physaddr, unsigned long size); |
| |
| /* |
| * Flags for reserve_bootmem (also if CONFIG_HAVE_ARCH_BOOTMEM_NODE, |
| * the architecture-specific code should honor this). |
| * |
| * If flags is 0, then the return value is always 0 (success). If |
| * flags contains BOOTMEM_EXCLUSIVE, then -EBUSY is returned if the |
| * memory already was reserved. |
| */ |
| #define BOOTMEM_DEFAULT 0 |
| #define BOOTMEM_EXCLUSIVE (1<<0) |
| |
| extern int reserve_bootmem(unsigned long addr, |
| unsigned long size, |
| int flags); |
| extern int reserve_bootmem_node(pg_data_t *pgdat, |
| unsigned long physaddr, |
| unsigned long size, |
| int flags); |
| |
| extern void *__alloc_bootmem(unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| extern void *__alloc_bootmem_nopanic(unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| extern void *__alloc_bootmem_node(pg_data_t *pgdat, |
| unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| void *__alloc_bootmem_node_high(pg_data_t *pgdat, |
| unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| extern void *__alloc_bootmem_node_nopanic(pg_data_t *pgdat, |
| unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| void *___alloc_bootmem_node_nopanic(pg_data_t *pgdat, |
| unsigned long size, |
| unsigned long align, |
| unsigned long goal, |
| unsigned long limit); |
| extern void *__alloc_bootmem_low(unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| void *__alloc_bootmem_low_nopanic(unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| extern void *__alloc_bootmem_low_node(pg_data_t *pgdat, |
| unsigned long size, |
| unsigned long align, |
| unsigned long goal); |
| |
| #ifdef CONFIG_NO_BOOTMEM |
| /* We are using top down, so it is safe to use 0 here */ |
| #define BOOTMEM_LOW_LIMIT 0 |
| #else |
| #define BOOTMEM_LOW_LIMIT __pa(MAX_DMA_ADDRESS) |
| #endif |
| |
| #define alloc_bootmem(x) \ |
| __alloc_bootmem(x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_align(x, align) \ |
| __alloc_bootmem(x, align, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_nopanic(x) \ |
| __alloc_bootmem_nopanic(x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_pages(x) \ |
| __alloc_bootmem(x, PAGE_SIZE, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_pages_nopanic(x) \ |
| __alloc_bootmem_nopanic(x, PAGE_SIZE, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_node(pgdat, x) \ |
| __alloc_bootmem_node(pgdat, x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_node_nopanic(pgdat, x) \ |
| __alloc_bootmem_node_nopanic(pgdat, x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_pages_node(pgdat, x) \ |
| __alloc_bootmem_node(pgdat, x, PAGE_SIZE, BOOTMEM_LOW_LIMIT) |
| #define alloc_bootmem_pages_node_nopanic(pgdat, x) \ |
| __alloc_bootmem_node_nopanic(pgdat, x, PAGE_SIZE, BOOTMEM_LOW_LIMIT) |
| |
| #define alloc_bootmem_low(x) \ |
| __alloc_bootmem_low(x, SMP_CACHE_BYTES, 0) |
| #define alloc_bootmem_low_pages_nopanic(x) \ |
| __alloc_bootmem_low_nopanic(x, PAGE_SIZE, 0) |
| #define alloc_bootmem_low_pages(x) \ |
| __alloc_bootmem_low(x, PAGE_SIZE, 0) |
| #define alloc_bootmem_low_pages_node(pgdat, x) \ |
| __alloc_bootmem_low_node(pgdat, x, PAGE_SIZE, 0) |
| |
| |
| #if defined(CONFIG_HAVE_MEMBLOCK) && defined(CONFIG_NO_BOOTMEM) |
| |
| /* FIXME: use MEMBLOCK_ALLOC_* variants here */ |
| #define BOOTMEM_ALLOC_ACCESSIBLE 0 |
| #define BOOTMEM_ALLOC_ANYWHERE (~(phys_addr_t)0) |
| |
| /* FIXME: Move to memblock.h at a point where we remove nobootmem.c */ |
| void *memblock_virt_alloc_try_nid_nopanic(phys_addr_t size, |
| phys_addr_t align, phys_addr_t min_addr, |
| phys_addr_t max_addr, int nid); |
| void *memblock_virt_alloc_try_nid(phys_addr_t size, phys_addr_t align, |
| phys_addr_t min_addr, phys_addr_t max_addr, int nid); |
| void __memblock_free_early(phys_addr_t base, phys_addr_t size); |
| void __memblock_free_late(phys_addr_t base, phys_addr_t size); |
| |
| static inline void * __init memblock_virt_alloc( |
| phys_addr_t size, phys_addr_t align) |
| { |
| return memblock_virt_alloc_try_nid(size, align, BOOTMEM_LOW_LIMIT, |
| BOOTMEM_ALLOC_ACCESSIBLE, |
| NUMA_NO_NODE); |
| } |
| |
| static inline void * __init memblock_virt_alloc_nopanic( |
| phys_addr_t size, phys_addr_t align) |
| { |
| return memblock_virt_alloc_try_nid_nopanic(size, align, |
| BOOTMEM_LOW_LIMIT, |
| BOOTMEM_ALLOC_ACCESSIBLE, |
| NUMA_NO_NODE); |
| } |
| |
| #ifndef ARCH_LOW_ADDRESS_LIMIT |
| #define ARCH_LOW_ADDRESS_LIMIT 0xffffffffUL |
| #endif |
| |
| static inline void * __init memblock_virt_alloc_low( |
| phys_addr_t size, phys_addr_t align) |
| { |
| return memblock_virt_alloc_try_nid(size, align, |
| BOOTMEM_LOW_LIMIT, |
| ARCH_LOW_ADDRESS_LIMIT, |
| NUMA_NO_NODE); |
| } |
| static inline void * __init memblock_virt_alloc_low_nopanic( |
| phys_addr_t size, phys_addr_t align) |
| { |
| return memblock_virt_alloc_try_nid_nopanic(size, align, |
| BOOTMEM_LOW_LIMIT, |
| ARCH_LOW_ADDRESS_LIMIT, |
| NUMA_NO_NODE); |
| } |
| |
| static inline void * __init memblock_virt_alloc_from_nopanic( |
| phys_addr_t size, phys_addr_t align, phys_addr_t min_addr) |
| { |
| return memblock_virt_alloc_try_nid_nopanic(size, align, min_addr, |
| BOOTMEM_ALLOC_ACCESSIBLE, |
| NUMA_NO_NODE); |
| } |
| |
| static inline void * __init memblock_virt_alloc_node( |
| phys_addr_t size, int nid) |
| { |
| return memblock_virt_alloc_try_nid(size, 0, BOOTMEM_LOW_LIMIT, |
| BOOTMEM_ALLOC_ACCESSIBLE, nid); |
| } |
| |
| static inline void * __init memblock_virt_alloc_node_nopanic( |
| phys_addr_t size, int nid) |
| { |
| return memblock_virt_alloc_try_nid_nopanic(size, 0, BOOTMEM_LOW_LIMIT, |
| BOOTMEM_ALLOC_ACCESSIBLE, |
| nid); |
| } |
| |
| static inline void __init memblock_free_early( |
| phys_addr_t base, phys_addr_t size) |
| { |
| __memblock_free_early(base, size); |
| } |
| |
| static inline void __init memblock_free_early_nid( |
| phys_addr_t base, phys_addr_t size, int nid) |
| { |
| __memblock_free_early(base, size); |
| } |
| |
| static inline void __init memblock_free_late( |
| phys_addr_t base, phys_addr_t size) |
| { |
| __memblock_free_late(base, size); |
| } |
| |
| #else |
| |
| #define BOOTMEM_ALLOC_ACCESSIBLE 0 |
| |
| |
| /* Fall back to all the existing bootmem APIs */ |
| static inline void * __init memblock_virt_alloc( |
| phys_addr_t size, phys_addr_t align) |
| { |
| if (!align) |
| align = SMP_CACHE_BYTES; |
| return __alloc_bootmem(size, align, BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_nopanic( |
| phys_addr_t size, phys_addr_t align) |
| { |
| if (!align) |
| align = SMP_CACHE_BYTES; |
| return __alloc_bootmem_nopanic(size, align, BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_low( |
| phys_addr_t size, phys_addr_t align) |
| { |
| if (!align) |
| align = SMP_CACHE_BYTES; |
| return __alloc_bootmem_low(size, align, BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_low_nopanic( |
| phys_addr_t size, phys_addr_t align) |
| { |
| if (!align) |
| align = SMP_CACHE_BYTES; |
| return __alloc_bootmem_low_nopanic(size, align, BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_from_nopanic( |
| phys_addr_t size, phys_addr_t align, phys_addr_t min_addr) |
| { |
| return __alloc_bootmem_nopanic(size, align, min_addr); |
| } |
| |
| static inline void * __init memblock_virt_alloc_node( |
| phys_addr_t size, int nid) |
| { |
| return __alloc_bootmem_node(NODE_DATA(nid), size, SMP_CACHE_BYTES, |
| BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_node_nopanic( |
| phys_addr_t size, int nid) |
| { |
| return __alloc_bootmem_node_nopanic(NODE_DATA(nid), size, |
| SMP_CACHE_BYTES, |
| BOOTMEM_LOW_LIMIT); |
| } |
| |
| static inline void * __init memblock_virt_alloc_try_nid(phys_addr_t size, |
| phys_addr_t align, phys_addr_t min_addr, phys_addr_t max_addr, int nid) |
| { |
| return __alloc_bootmem_node_high(NODE_DATA(nid), size, align, |
| min_addr); |
| } |
| |
| static inline void * __init memblock_virt_alloc_try_nid_nopanic( |
| phys_addr_t size, phys_addr_t align, |
| phys_addr_t min_addr, phys_addr_t max_addr, int nid) |
| { |
| return ___alloc_bootmem_node_nopanic(NODE_DATA(nid), size, align, |
| min_addr, max_addr); |
| } |
| |
| static inline void __init memblock_free_early( |
| phys_addr_t base, phys_addr_t size) |
| { |
| free_bootmem(base, size); |
| } |
| |
| static inline void __init memblock_free_early_nid( |
| phys_addr_t base, phys_addr_t size, int nid) |
| { |
| free_bootmem_node(NODE_DATA(nid), base, size); |
| } |
| |
| static inline void __init memblock_free_late( |
| phys_addr_t base, phys_addr_t size) |
| { |
| free_bootmem_late(base, size); |
| } |
| #endif /* defined(CONFIG_HAVE_MEMBLOCK) && defined(CONFIG_NO_BOOTMEM) */ |
| |
| #ifdef CONFIG_HAVE_ARCH_ALLOC_REMAP |
| extern void *alloc_remap(int nid, unsigned long size); |
| #else |
| static inline void *alloc_remap(int nid, unsigned long size) |
| { |
| return NULL; |
| } |
| #endif /* CONFIG_HAVE_ARCH_ALLOC_REMAP */ |
| |
| extern void *alloc_large_system_hash(const char *tablename, |
| unsigned long bucketsize, |
| unsigned long numentries, |
| int scale, |
| int flags, |
| unsigned int *_hash_shift, |
| unsigned int *_hash_mask, |
| unsigned long low_limit, |
| unsigned long high_limit); |
| |
| #define HASH_EARLY 0x00000001 /* Allocating during early boot? */ |
| #define HASH_SMALL 0x00000002 /* sub-page allocation allowed, min |
| * shift passed via *_hash_shift */ |
| |
| /* Only NUMA needs hash distribution. 64bit NUMA architectures have |
| * sufficient vmalloc space. |
| */ |
| #if defined(CONFIG_NUMA) && defined(CONFIG_64BIT) |
| #define HASHDIST_DEFAULT 1 |
| #else |
| #define HASHDIST_DEFAULT 0 |
| #endif |
| extern int hashdist; /* Distribute hashes across NUMA nodes? */ |
| |
| |
| #endif /* _LINUX_BOOTMEM_H */ |