Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 1 | #ifndef _SPARC64_PAGE_H |
| 2 | #define _SPARC64_PAGE_H |
| 3 | |
| 4 | #include <linux/const.h> |
| 5 | |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 6 | #define PAGE_SHIFT 13 |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 7 | |
| 8 | #define PAGE_SIZE (_AC(1,UL) << PAGE_SHIFT) |
| 9 | #define PAGE_MASK (~(PAGE_SIZE-1)) |
| 10 | |
| 11 | /* Flushing for D-cache alias handling is only needed if |
| 12 | * the page size is smaller than 16K. |
| 13 | */ |
| 14 | #if PAGE_SHIFT < 14 |
| 15 | #define DCACHE_ALIASING_POSSIBLE |
| 16 | #endif |
| 17 | |
David S. Miller | 37b3a8f | 2013-09-25 13:48:49 -0700 | [diff] [blame] | 18 | #define HPAGE_SHIFT 23 |
| 19 | #define REAL_HPAGE_SHIFT 22 |
Nitin Gupta | c7d9f77 | 2017-02-01 16:16:36 -0800 | [diff] [blame] | 20 | #define HPAGE_256MB_SHIFT 28 |
Nitin Gupta | dcd1912 | 2017-02-06 12:33:26 -0800 | [diff] [blame] | 21 | #define HPAGE_64K_SHIFT 16 |
David S. Miller | 37b3a8f | 2013-09-25 13:48:49 -0700 | [diff] [blame] | 22 | #define REAL_HPAGE_SIZE (_AC(1,UL) << REAL_HPAGE_SHIFT) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 23 | |
David Miller | 9e695d2 | 2012-10-08 16:34:29 -0700 | [diff] [blame] | 24 | #if defined(CONFIG_HUGETLB_PAGE) || defined(CONFIG_TRANSPARENT_HUGEPAGE) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 25 | #define HPAGE_SIZE (_AC(1,UL) << HPAGE_SHIFT) |
| 26 | #define HPAGE_MASK (~(HPAGE_SIZE - 1UL)) |
| 27 | #define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT) |
| 28 | #define HAVE_ARCH_HUGETLB_UNMAPPED_AREA |
Mike Kravetz | 1e953d8 | 2016-08-31 13:48:19 -0700 | [diff] [blame] | 29 | #define REAL_HPAGE_PER_HPAGE (_AC(1,UL) << (HPAGE_SHIFT - REAL_HPAGE_SHIFT)) |
Nitin Gupta | dcd1912 | 2017-02-06 12:33:26 -0800 | [diff] [blame] | 30 | #define HUGE_MAX_HSTATE 3 |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 31 | #endif |
| 32 | |
| 33 | #ifndef __ASSEMBLY__ |
| 34 | |
David Miller | 9e695d2 | 2012-10-08 16:34:29 -0700 | [diff] [blame] | 35 | #if defined(CONFIG_HUGETLB_PAGE) || defined(CONFIG_TRANSPARENT_HUGEPAGE) |
David S. Miller | 0fbebed | 2013-02-19 22:34:10 -0800 | [diff] [blame] | 36 | struct pt_regs; |
Sam Ravnborg | f05a686 | 2014-05-16 23:25:50 +0200 | [diff] [blame] | 37 | void hugetlb_setup(struct pt_regs *regs); |
David Miller | 9e695d2 | 2012-10-08 16:34:29 -0700 | [diff] [blame] | 38 | #endif |
| 39 | |
David S. Miller | b0f1e79 | 2008-09-11 23:36:32 -0700 | [diff] [blame] | 40 | #define WANT_PAGE_VIRTUAL |
| 41 | |
Sam Ravnborg | f05a686 | 2014-05-16 23:25:50 +0200 | [diff] [blame] | 42 | void _clear_page(void *page); |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 43 | #define clear_page(X) _clear_page((void *)(X)) |
| 44 | struct page; |
Sam Ravnborg | f05a686 | 2014-05-16 23:25:50 +0200 | [diff] [blame] | 45 | void clear_user_page(void *addr, unsigned long vaddr, struct page *page); |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 46 | #define copy_page(X,Y) memcpy((void *)(X), (void *)(Y), PAGE_SIZE) |
Sam Ravnborg | f05a686 | 2014-05-16 23:25:50 +0200 | [diff] [blame] | 47 | void copy_user_page(void *to, void *from, unsigned long vaddr, struct page *topage); |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 48 | |
| 49 | /* Unlike sparc32, sparc64's parameter passing API is more |
| 50 | * sane in that structures which as small enough are passed |
| 51 | * in registers instead of on the stack. Thus, setting |
| 52 | * STRICT_MM_TYPECHECKS does not generate worse code so |
| 53 | * let's enable it to get the type checking. |
| 54 | */ |
| 55 | |
| 56 | #define STRICT_MM_TYPECHECKS |
| 57 | |
| 58 | #ifdef STRICT_MM_TYPECHECKS |
| 59 | /* These are used to make use of C type-checking.. */ |
| 60 | typedef struct { unsigned long pte; } pte_t; |
| 61 | typedef struct { unsigned long iopte; } iopte_t; |
David S. Miller | 2b77933 | 2013-09-25 14:33:16 -0700 | [diff] [blame] | 62 | typedef struct { unsigned long pmd; } pmd_t; |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 63 | typedef struct { unsigned long pud; } pud_t; |
David S. Miller | 2b77933 | 2013-09-25 14:33:16 -0700 | [diff] [blame] | 64 | typedef struct { unsigned long pgd; } pgd_t; |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 65 | typedef struct { unsigned long pgprot; } pgprot_t; |
| 66 | |
| 67 | #define pte_val(x) ((x).pte) |
| 68 | #define iopte_val(x) ((x).iopte) |
| 69 | #define pmd_val(x) ((x).pmd) |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 70 | #define pud_val(x) ((x).pud) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 71 | #define pgd_val(x) ((x).pgd) |
| 72 | #define pgprot_val(x) ((x).pgprot) |
| 73 | |
| 74 | #define __pte(x) ((pte_t) { (x) } ) |
| 75 | #define __iopte(x) ((iopte_t) { (x) } ) |
| 76 | #define __pmd(x) ((pmd_t) { (x) } ) |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 77 | #define __pud(x) ((pud_t) { (x) } ) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 78 | #define __pgd(x) ((pgd_t) { (x) } ) |
| 79 | #define __pgprot(x) ((pgprot_t) { (x) } ) |
| 80 | |
| 81 | #else |
| 82 | /* .. while these make it easier on the compiler */ |
| 83 | typedef unsigned long pte_t; |
| 84 | typedef unsigned long iopte_t; |
David S. Miller | 2b77933 | 2013-09-25 14:33:16 -0700 | [diff] [blame] | 85 | typedef unsigned long pmd_t; |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 86 | typedef unsigned long pud_t; |
David S. Miller | 2b77933 | 2013-09-25 14:33:16 -0700 | [diff] [blame] | 87 | typedef unsigned long pgd_t; |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 88 | typedef unsigned long pgprot_t; |
| 89 | |
| 90 | #define pte_val(x) (x) |
| 91 | #define iopte_val(x) (x) |
| 92 | #define pmd_val(x) (x) |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 93 | #define pud_val(x) (x) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 94 | #define pgd_val(x) (x) |
| 95 | #define pgprot_val(x) (x) |
| 96 | |
| 97 | #define __pte(x) (x) |
| 98 | #define __iopte(x) (x) |
| 99 | #define __pmd(x) (x) |
David S. Miller | ac55c76 | 2014-09-26 21:19:46 -0700 | [diff] [blame] | 100 | #define __pud(x) (x) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 101 | #define __pgd(x) (x) |
| 102 | #define __pgprot(x) (x) |
| 103 | |
| 104 | #endif /* (STRICT_MM_TYPECHECKS) */ |
| 105 | |
David Miller | c460bec | 2012-10-08 16:34:22 -0700 | [diff] [blame] | 106 | typedef pte_t *pgtable_t; |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 107 | |
David S. Miller | 4397bed | 2014-09-26 21:58:33 -0700 | [diff] [blame] | 108 | extern unsigned long sparc64_va_hole_top; |
| 109 | extern unsigned long sparc64_va_hole_bottom; |
David S. Miller | c920745 | 2013-09-18 11:58:32 -0700 | [diff] [blame] | 110 | |
David S. Miller | 2b77933 | 2013-09-25 14:33:16 -0700 | [diff] [blame] | 111 | /* The next two defines specify the actual exclusion region we |
| 112 | * enforce, wherein we use a 4GB red zone on each side of the VA hole. |
| 113 | */ |
David S. Miller | 4397bed | 2014-09-26 21:58:33 -0700 | [diff] [blame] | 114 | #define VA_EXCLUDE_START (sparc64_va_hole_bottom - (1UL << 32UL)) |
| 115 | #define VA_EXCLUDE_END (sparc64_va_hole_top + (1UL << 32UL)) |
David S. Miller | c920745 | 2013-09-18 11:58:32 -0700 | [diff] [blame] | 116 | |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 117 | #define TASK_UNMAPPED_BASE (test_thread_flag(TIF_32BIT) ? \ |
David S. Miller | c920745 | 2013-09-18 11:58:32 -0700 | [diff] [blame] | 118 | _AC(0x0000000070000000,UL) : \ |
| 119 | VA_EXCLUDE_END) |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 120 | |
| 121 | #include <asm-generic/memory_model.h> |
| 122 | |
David S. Miller | b2d4383 | 2013-09-20 21:50:41 -0700 | [diff] [blame] | 123 | extern unsigned long PAGE_OFFSET; |
| 124 | |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 125 | #endif /* !(__ASSEMBLY__) */ |
| 126 | |
David S. Miller | 7c0fa0f | 2014-09-24 21:49:29 -0700 | [diff] [blame] | 127 | /* The maximum number of physical memory address bits we support. The |
| 128 | * largest value we can support is whatever "KPGD_SHIFT + KPTE_BITS" |
| 129 | * evaluates to. |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 130 | */ |
David S. Miller | 7c0fa0f | 2014-09-24 21:49:29 -0700 | [diff] [blame] | 131 | #define MAX_PHYS_ADDRESS_BITS 53 |
David S. Miller | bb7b435 | 2013-09-18 15:39:06 -0700 | [diff] [blame] | 132 | |
David S. Miller | bb7b435 | 2013-09-18 15:39:06 -0700 | [diff] [blame] | 133 | #define ILOG2_4MB 22 |
| 134 | #define ILOG2_256MB 28 |
| 135 | |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 136 | #ifndef __ASSEMBLY__ |
| 137 | |
| 138 | #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) |
| 139 | #define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET)) |
| 140 | |
| 141 | #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) |
| 142 | |
| 143 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT) |
| 144 | |
| 145 | #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) |
| 146 | |
| 147 | #define virt_to_phys __pa |
| 148 | #define phys_to_virt __va |
| 149 | |
| 150 | #endif /* !(__ASSEMBLY__) */ |
| 151 | |
| 152 | #define VM_DATA_DEFAULT_FLAGS (VM_READ | VM_WRITE | VM_EXEC | \ |
| 153 | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) |
| 154 | |
Arnd Bergmann | 5b17e1c | 2009-05-13 22:56:30 +0000 | [diff] [blame] | 155 | #include <asm-generic/getorder.h> |
Sam Ravnborg | f5e706a | 2008-07-17 21:55:51 -0700 | [diff] [blame] | 156 | |
| 157 | #endif /* _SPARC64_PAGE_H */ |