blob: 0727a55d87d99a17c5e7fd0d8b6d8e2fe3b403a8 [file] [log] [blame]
Gerald Schaefer53492b12008-04-30 13:38:46 +02001/*
2 * IBM System z Huge TLB Page Support for Kernel.
3 *
Heiko Carstensa53c8fa2012-07-20 11:15:04 +02004 * Copyright IBM Corp. 2007
Gerald Schaefer53492b12008-04-30 13:38:46 +02005 * Author(s): Gerald Schaefer <gerald.schaefer@de.ibm.com>
6 */
7
8#include <linux/mm.h>
9#include <linux/hugetlb.h>
10
Martin Schwidefskye5098612013-07-23 20:57:57 +020011static inline pmd_t __pte_to_pmd(pte_t pte)
12{
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020013 int none, young, prot;
Martin Schwidefskye5098612013-07-23 20:57:57 +020014 pmd_t pmd;
15
16 /*
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020017 * Convert encoding pte bits pmd bits
18 * .IR...wrdytp ..R...I...y.
19 * empty .10...000000 -> ..0...1...0.
20 * prot-none, clean, old .11...000001 -> ..0...1...1.
21 * prot-none, clean, young .11...000101 -> ..1...1...1.
22 * prot-none, dirty, old .10...001001 -> ..0...1...1.
23 * prot-none, dirty, young .10...001101 -> ..1...1...1.
24 * read-only, clean, old .11...010001 -> ..1...1...0.
25 * read-only, clean, young .01...010101 -> ..1...0...1.
26 * read-only, dirty, old .11...011001 -> ..1...1...0.
27 * read-only, dirty, young .01...011101 -> ..1...0...1.
28 * read-write, clean, old .11...110001 -> ..0...1...0.
29 * read-write, clean, young .01...110101 -> ..0...0...1.
30 * read-write, dirty, old .10...111001 -> ..0...1...0.
31 * read-write, dirty, young .00...111101 -> ..0...0...1.
Martin Schwidefskye5098612013-07-23 20:57:57 +020032 * Huge ptes are dirty by definition, a clean pte is made dirty
33 * by the conversion.
34 */
35 if (pte_present(pte)) {
36 pmd_val(pmd) = pte_val(pte) & PAGE_MASK;
37 if (pte_val(pte) & _PAGE_INVALID)
38 pmd_val(pmd) |= _SEGMENT_ENTRY_INVALID;
39 none = (pte_val(pte) & _PAGE_PRESENT) &&
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020040 !(pte_val(pte) & _PAGE_READ) &&
41 !(pte_val(pte) & _PAGE_WRITE);
42 prot = (pte_val(pte) & _PAGE_PROTECT) &&
43 !(pte_val(pte) & _PAGE_WRITE);
44 young = pte_val(pte) & _PAGE_YOUNG;
45 if (none || young)
46 pmd_val(pmd) |= _SEGMENT_ENTRY_YOUNG;
47 if (prot || (none && young))
Martin Schwidefskye5098612013-07-23 20:57:57 +020048 pmd_val(pmd) |= _SEGMENT_ENTRY_PROTECT;
49 } else
50 pmd_val(pmd) = _SEGMENT_ENTRY_INVALID;
51 return pmd;
52}
53
54static inline pte_t __pmd_to_pte(pmd_t pmd)
55{
56 pte_t pte;
57
58 /*
59 * Convert encoding pmd bits pte bits
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020060 * ..R...I...y. .IR...wrdytp
61 * empty ..0...1...0. -> .10...000000
62 * prot-none, old ..0...1...1. -> .10...001001
63 * prot-none, young ..1...1...1. -> .10...001101
64 * read-only, old ..1...1...0. -> .11...011001
65 * read-only, young ..1...0...1. -> .01...011101
66 * read-write, old ..0...1...0. -> .10...111001
67 * read-write, young ..0...0...1. -> .00...111101
Martin Schwidefskye5098612013-07-23 20:57:57 +020068 * Huge ptes are dirty by definition
69 */
70 if (pmd_present(pmd)) {
71 pte_val(pte) = _PAGE_PRESENT | _PAGE_LARGE | _PAGE_DIRTY |
72 (pmd_val(pmd) & PAGE_MASK);
73 if (pmd_val(pmd) & _SEGMENT_ENTRY_INVALID)
74 pte_val(pte) |= _PAGE_INVALID;
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020075 if (pmd_prot_none(pmd)) {
76 if (pmd_val(pmd) & _SEGMENT_ENTRY_PROTECT)
77 pte_val(pte) |= _PAGE_YOUNG;
78 } else {
79 pte_val(pte) |= _PAGE_READ;
Martin Schwidefskye5098612013-07-23 20:57:57 +020080 if (pmd_val(pmd) & _SEGMENT_ENTRY_PROTECT)
81 pte_val(pte) |= _PAGE_PROTECT;
82 else
83 pte_val(pte) |= _PAGE_WRITE;
Martin Schwidefsky0944fe32013-07-23 22:11:42 +020084 if (pmd_val(pmd) & _SEGMENT_ENTRY_YOUNG)
85 pte_val(pte) |= _PAGE_YOUNG;
Martin Schwidefskye5098612013-07-23 20:57:57 +020086 }
87 } else
88 pte_val(pte) = _PAGE_INVALID;
89 return pte;
90}
Gerald Schaefer53492b12008-04-30 13:38:46 +020091
92void set_huge_pte_at(struct mm_struct *mm, unsigned long addr,
Martin Schwidefskye5098612013-07-23 20:57:57 +020093 pte_t *ptep, pte_t pte)
Gerald Schaefer53492b12008-04-30 13:38:46 +020094{
Martin Schwidefskye5098612013-07-23 20:57:57 +020095 pmd_t pmd;
Gerald Schaefer53492b12008-04-30 13:38:46 +020096
Martin Schwidefskye5098612013-07-23 20:57:57 +020097 pmd = __pte_to_pmd(pte);
Gerald Schaefer53492b12008-04-30 13:38:46 +020098 if (!MACHINE_HAS_HPAGE) {
Martin Schwidefskye5098612013-07-23 20:57:57 +020099 pmd_val(pmd) &= ~_SEGMENT_ENTRY_ORIGIN;
100 pmd_val(pmd) |= pte_page(pte)[1].index;
101 } else
102 pmd_val(pmd) |= _SEGMENT_ENTRY_LARGE | _SEGMENT_ENTRY_CO;
103 *(pmd_t *) ptep = pmd;
104}
Gerald Schaefer53492b12008-04-30 13:38:46 +0200105
Martin Schwidefskye5098612013-07-23 20:57:57 +0200106pte_t huge_ptep_get(pte_t *ptep)
107{
108 unsigned long origin;
109 pmd_t pmd;
110
111 pmd = *(pmd_t *) ptep;
112 if (!MACHINE_HAS_HPAGE && pmd_present(pmd)) {
113 origin = pmd_val(pmd) & _SEGMENT_ENTRY_ORIGIN;
114 pmd_val(pmd) &= ~_SEGMENT_ENTRY_ORIGIN;
115 pmd_val(pmd) |= *(unsigned long *) origin;
116 }
117 return __pmd_to_pte(pmd);
118}
119
120pte_t huge_ptep_get_and_clear(struct mm_struct *mm,
121 unsigned long addr, pte_t *ptep)
122{
123 pmd_t *pmdp = (pmd_t *) ptep;
124 pte_t pte = huge_ptep_get(ptep);
125
Martin Schwidefsky1b948d62014-04-03 13:55:01 +0200126 pmdp_flush_direct(mm, addr, pmdp);
Martin Schwidefskye5098612013-07-23 20:57:57 +0200127 pmd_val(*pmdp) = _SEGMENT_ENTRY_EMPTY;
128 return pte;
Gerald Schaefer53492b12008-04-30 13:38:46 +0200129}
130
131int arch_prepare_hugepage(struct page *page)
132{
133 unsigned long addr = page_to_phys(page);
134 pte_t pte;
135 pte_t *ptep;
136 int i;
137
138 if (MACHINE_HAS_HPAGE)
139 return 0;
140
Martin Schwidefskye5992f22011-07-24 10:48:20 +0200141 ptep = (pte_t *) pte_alloc_one(&init_mm, addr);
Gerald Schaefer53492b12008-04-30 13:38:46 +0200142 if (!ptep)
143 return -ENOMEM;
144
Gerald Schaefer106c9922013-04-29 15:07:23 -0700145 pte_val(pte) = addr;
Gerald Schaefer53492b12008-04-30 13:38:46 +0200146 for (i = 0; i < PTRS_PER_PTE; i++) {
147 set_pte_at(&init_mm, addr + i * PAGE_SIZE, ptep + i, pte);
148 pte_val(pte) += PAGE_SIZE;
149 }
150 page[1].index = (unsigned long) ptep;
151 return 0;
152}
153
154void arch_release_hugepage(struct page *page)
155{
156 pte_t *ptep;
157
158 if (MACHINE_HAS_HPAGE)
159 return;
160
161 ptep = (pte_t *) page[1].index;
162 if (!ptep)
163 return;
Martin Schwidefskye5098612013-07-23 20:57:57 +0200164 clear_table((unsigned long *) ptep, _PAGE_INVALID,
Gerald Schaefera6864252012-05-09 16:27:37 +0200165 PTRS_PER_PTE * sizeof(pte_t));
Martin Schwidefsky80217142010-10-25 16:10:11 +0200166 page_table_free(&init_mm, (unsigned long *) ptep);
Gerald Schaefer53492b12008-04-30 13:38:46 +0200167 page[1].index = 0;
168}
169
Andi Kleena5516432008-07-23 21:27:41 -0700170pte_t *huge_pte_alloc(struct mm_struct *mm,
171 unsigned long addr, unsigned long sz)
Gerald Schaefer53492b12008-04-30 13:38:46 +0200172{
173 pgd_t *pgdp;
174 pud_t *pudp;
175 pmd_t *pmdp = NULL;
176
177 pgdp = pgd_offset(mm, addr);
178 pudp = pud_alloc(mm, pgdp, addr);
179 if (pudp)
180 pmdp = pmd_alloc(mm, pudp, addr);
181 return (pte_t *) pmdp;
182}
183
184pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr)
185{
186 pgd_t *pgdp;
187 pud_t *pudp;
188 pmd_t *pmdp = NULL;
189
190 pgdp = pgd_offset(mm, addr);
191 if (pgd_present(*pgdp)) {
192 pudp = pud_offset(pgdp, addr);
193 if (pud_present(*pudp))
194 pmdp = pmd_offset(pudp, addr);
195 }
196 return (pte_t *) pmdp;
197}
198
199int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep)
200{
201 return 0;
202}
203
204struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address,
205 int write)
206{
207 return ERR_PTR(-EINVAL);
208}
209
210int pmd_huge(pmd_t pmd)
211{
212 if (!MACHINE_HAS_HPAGE)
213 return 0;
214
215 return !!(pmd_val(pmd) & _SEGMENT_ENTRY_LARGE);
216}
217
Andi Kleenceb86872008-07-23 21:27:50 -0700218int pud_huge(pud_t pud)
219{
220 return 0;
221}
222
Naoya Horiguchi83467ef2013-09-11 14:22:11 -0700223int pmd_huge_support(void)
224{
225 return 1;
226}
227
Gerald Schaefer53492b12008-04-30 13:38:46 +0200228struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address,
229 pmd_t *pmdp, int write)
230{
231 struct page *page;
232
233 if (!MACHINE_HAS_HPAGE)
234 return NULL;
235
236 page = pmd_page(*pmdp);
237 if (page)
238 page += ((address & ~HPAGE_MASK) >> PAGE_SHIFT);
239 return page;
240}