blob: 5148150cc80b02d9ff440de4827ab7a173067446 [file] [log] [blame]
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -08001#ifndef __ASM_MEMORY_MODEL_H
2#define __ASM_MEMORY_MODEL_H
3
Chen Gang8f235d12016-01-14 15:18:33 -08004#include <linux/pfn.h>
5
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -08006#ifndef __ASSEMBLY__
7
8#if defined(CONFIG_FLATMEM)
9
10#ifndef ARCH_PFN_OFFSET
11#define ARCH_PFN_OFFSET (0UL)
12#endif
13
14#elif defined(CONFIG_DISCONTIGMEM)
15
16#ifndef arch_pfn_to_nid
17#define arch_pfn_to_nid(pfn) pfn_to_nid(pfn)
18#endif
19
20#ifndef arch_local_page_offset
21#define arch_local_page_offset(pfn, nid) \
22 ((pfn) - NODE_DATA(nid)->node_start_pfn)
23#endif
24
25#endif /* CONFIG_DISCONTIGMEM */
26
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080027/*
28 * supports 3 memory models.
29 */
30#if defined(CONFIG_FLATMEM)
31
Andy Whitcroft67de6482006-06-23 02:03:12 -070032#define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET))
33#define __page_to_pfn(page) ((unsigned long)((page) - mem_map) + \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080034 ARCH_PFN_OFFSET)
35#elif defined(CONFIG_DISCONTIGMEM)
36
Andy Whitcroft67de6482006-06-23 02:03:12 -070037#define __pfn_to_page(pfn) \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080038({ unsigned long __pfn = (pfn); \
Rafael J. Wysockic5d71242008-11-08 13:53:33 +010039 unsigned long __nid = arch_pfn_to_nid(__pfn); \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080040 NODE_DATA(__nid)->node_mem_map + arch_local_page_offset(__pfn, __nid);\
41})
42
Andy Whitcroft67de6482006-06-23 02:03:12 -070043#define __page_to_pfn(pg) \
Ian Campbellaa462ab2011-08-17 17:40:33 +010044({ const struct page *__pg = (pg); \
KAMEZAWA Hiroyukia0140c12006-03-27 01:15:55 -080045 struct pglist_data *__pgdat = NODE_DATA(page_to_nid(__pg)); \
46 (unsigned long)(__pg - __pgdat->node_mem_map) + \
47 __pgdat->node_start_pfn; \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080048})
49
Christoph Lameter8f6aac42007-10-16 01:24:13 -070050#elif defined(CONFIG_SPARSEMEM_VMEMMAP)
51
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020052/* memmap is virtually contiguous. */
Christoph Lameter8f6aac42007-10-16 01:24:13 -070053#define __pfn_to_page(pfn) (vmemmap + (pfn))
Martin Schwidefsky32272a22008-12-25 13:38:59 +010054#define __page_to_pfn(page) (unsigned long)((page) - vmemmap)
Christoph Lameter8f6aac42007-10-16 01:24:13 -070055
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080056#elif defined(CONFIG_SPARSEMEM)
57/*
Zhang Yanfei1a491232013-10-03 19:38:14 +080058 * Note: section's mem_map is encoded to reflect its start_pfn.
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080059 * section[i].section_mem_map == mem_map's address - start_pfn;
60 */
Andy Whitcroft67de6482006-06-23 02:03:12 -070061#define __page_to_pfn(pg) \
Ian Campbellaa462ab2011-08-17 17:40:33 +010062({ const struct page *__pg = (pg); \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080063 int __sec = page_to_section(__pg); \
Randy Dunlapf05b6282007-02-10 01:42:59 -080064 (unsigned long)(__pg - __section_mem_map_addr(__nr_to_section(__sec))); \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080065})
66
Andy Whitcroft67de6482006-06-23 02:03:12 -070067#define __pfn_to_page(pfn) \
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080068({ unsigned long __pfn = (pfn); \
69 struct mem_section *__sec = __pfn_to_section(__pfn); \
70 __section_mem_map_addr(__sec) + __pfn; \
71})
72#endif /* CONFIG_FLATMEM/DISCONTIGMEM/SPARSEMEM */
Andy Whitcroft67de6482006-06-23 02:03:12 -070073
Christoph Hellwig012dcef2015-08-07 17:41:01 -040074/*
75 * Convert a physical address to a Page Frame Number and back
76 */
Chen Gang8f235d12016-01-14 15:18:33 -080077#define __phys_to_pfn(paddr) PHYS_PFN(paddr)
Tyler Bakerae4f97692015-09-19 03:58:10 -040078#define __pfn_to_phys(pfn) PFN_PHYS(pfn)
Christoph Hellwig012dcef2015-08-07 17:41:01 -040079
Andy Whitcroft67de6482006-06-23 02:03:12 -070080#define page_to_pfn __page_to_pfn
81#define pfn_to_page __pfn_to_page
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080082
83#endif /* __ASSEMBLY__ */
KAMEZAWA Hiroyukia117e662006-03-27 01:15:25 -080084
85#endif