Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 1 | /* |
| 2 | * This code is used on x86_64 to create page table identity mappings on |
| 3 | * demand by building up a new set of page tables (or appending to the |
| 4 | * existing ones), and then switching over to them when ready. |
Kees Cook | 11fdf97 | 2016-05-25 15:45:31 -0700 | [diff] [blame^] | 5 | * |
| 6 | * Copyright (C) 2015-2016 Yinghai Lu |
| 7 | * Copyright (C) 2016 Kees Cook |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 8 | */ |
| 9 | |
| 10 | /* |
| 11 | * Since we're dealing with identity mappings, physical and virtual |
| 12 | * addresses are the same, so override these defines which are ultimately |
| 13 | * used by the headers in misc.h. |
| 14 | */ |
| 15 | #define __pa(x) ((unsigned long)(x)) |
| 16 | #define __va(x) ((void *)((unsigned long)(x))) |
| 17 | |
| 18 | #include "misc.h" |
| 19 | |
| 20 | /* These actually do the work of building the kernel identity maps. */ |
| 21 | #include <asm/init.h> |
| 22 | #include <asm/pgtable.h> |
| 23 | #include "../../mm/ident_map.c" |
| 24 | |
| 25 | /* Used by pgtable.h asm code to force instruction serialization. */ |
| 26 | unsigned long __force_order; |
| 27 | |
| 28 | /* Used to track our page table allocation area. */ |
| 29 | struct alloc_pgt_data { |
| 30 | unsigned char *pgt_buf; |
| 31 | unsigned long pgt_buf_size; |
| 32 | unsigned long pgt_buf_offset; |
| 33 | }; |
| 34 | |
| 35 | /* |
| 36 | * Allocates space for a page table entry, using struct alloc_pgt_data |
| 37 | * above. Besides the local callers, this is used as the allocation |
| 38 | * callback in mapping_info below. |
| 39 | */ |
| 40 | static void *alloc_pgt_page(void *context) |
| 41 | { |
| 42 | struct alloc_pgt_data *pages = (struct alloc_pgt_data *)context; |
| 43 | unsigned char *entry; |
| 44 | |
| 45 | /* Validate there is space available for a new page. */ |
| 46 | if (pages->pgt_buf_offset >= pages->pgt_buf_size) { |
| 47 | debug_putstr("out of pgt_buf in " __FILE__ "!?\n"); |
| 48 | debug_putaddr(pages->pgt_buf_offset); |
| 49 | debug_putaddr(pages->pgt_buf_size); |
| 50 | return NULL; |
| 51 | } |
| 52 | |
| 53 | entry = pages->pgt_buf + pages->pgt_buf_offset; |
| 54 | pages->pgt_buf_offset += PAGE_SIZE; |
| 55 | |
| 56 | return entry; |
| 57 | } |
| 58 | |
| 59 | /* Used to track our allocated page tables. */ |
| 60 | static struct alloc_pgt_data pgt_data; |
| 61 | |
| 62 | /* The top level page table entry pointer. */ |
| 63 | static unsigned long level4p; |
| 64 | |
Kees Cook | 11fdf97 | 2016-05-25 15:45:31 -0700 | [diff] [blame^] | 65 | /* |
| 66 | * Mapping information structure passed to kernel_ident_mapping_init(). |
| 67 | * Due to relocation, pointers must be assigned at run time not build time. |
| 68 | */ |
| 69 | static struct x86_mapping_info mapping_info = { |
| 70 | .pmd_flag = __PAGE_KERNEL_LARGE_EXEC, |
| 71 | }; |
| 72 | |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 73 | /* Locates and clears a region for a new top level page table. */ |
Kees Cook | 11fdf97 | 2016-05-25 15:45:31 -0700 | [diff] [blame^] | 74 | void initialize_identity_maps(void) |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 75 | { |
Kees Cook | 11fdf97 | 2016-05-25 15:45:31 -0700 | [diff] [blame^] | 76 | /* Init mapping_info with run-time function/buffer pointers. */ |
| 77 | mapping_info.alloc_pgt_page = alloc_pgt_page; |
| 78 | mapping_info.context = &pgt_data; |
| 79 | |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 80 | /* |
| 81 | * It should be impossible for this not to already be true, |
| 82 | * but since calling this a second time would rewind the other |
| 83 | * counters, let's just make sure this is reset too. |
| 84 | */ |
| 85 | pgt_data.pgt_buf_offset = 0; |
| 86 | |
| 87 | /* |
| 88 | * If we came here via startup_32(), cr3 will be _pgtable already |
| 89 | * and we must append to the existing area instead of entirely |
| 90 | * overwriting it. |
| 91 | */ |
| 92 | level4p = read_cr3(); |
| 93 | if (level4p == (unsigned long)_pgtable) { |
| 94 | debug_putstr("booted via startup_32()\n"); |
| 95 | pgt_data.pgt_buf = _pgtable + BOOT_INIT_PGT_SIZE; |
| 96 | pgt_data.pgt_buf_size = BOOT_PGT_SIZE - BOOT_INIT_PGT_SIZE; |
| 97 | memset(pgt_data.pgt_buf, 0, pgt_data.pgt_buf_size); |
| 98 | } else { |
| 99 | debug_putstr("booted via startup_64()\n"); |
| 100 | pgt_data.pgt_buf = _pgtable; |
| 101 | pgt_data.pgt_buf_size = BOOT_PGT_SIZE; |
| 102 | memset(pgt_data.pgt_buf, 0, pgt_data.pgt_buf_size); |
| 103 | level4p = (unsigned long)alloc_pgt_page(&pgt_data); |
| 104 | } |
| 105 | } |
| 106 | |
| 107 | /* |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 108 | * Adds the specified range to what will become the new identity mappings. |
| 109 | * Once all ranges have been added, the new mapping is activated by calling |
| 110 | * finalize_identity_maps() below. |
| 111 | */ |
| 112 | void add_identity_map(unsigned long start, unsigned long size) |
| 113 | { |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 114 | unsigned long end = start + size; |
| 115 | |
Kees Cook | 3a94707 | 2016-05-06 15:01:35 -0700 | [diff] [blame] | 116 | /* Align boundary to 2M. */ |
| 117 | start = round_down(start, PMD_SIZE); |
| 118 | end = round_up(end, PMD_SIZE); |
| 119 | if (start >= end) |
| 120 | return; |
| 121 | |
| 122 | /* Build the mapping. */ |
| 123 | kernel_ident_mapping_init(&mapping_info, (pgd_t *)level4p, |
| 124 | start, end); |
| 125 | } |
| 126 | |
| 127 | /* |
| 128 | * This switches the page tables to the new level4 that has been built |
| 129 | * via calls to add_identity_map() above. If booted via startup_32(), |
| 130 | * this is effectively a no-op. |
| 131 | */ |
| 132 | void finalize_identity_maps(void) |
| 133 | { |
| 134 | write_cr3(level4p); |
| 135 | } |