Vineet Gupta | cc562d2 | 2013-01-18 15:12:19 +0530 | [diff] [blame^] | 1 | /* |
| 2 | * TLB Exception Handling for ARC |
| 3 | * |
| 4 | * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com) |
| 5 | * |
| 6 | * This program is free software; you can redistribute it and/or modify |
| 7 | * it under the terms of the GNU General Public License version 2 as |
| 8 | * published by the Free Software Foundation. |
| 9 | * |
| 10 | * Vineetg: April 2011 : |
| 11 | * -MMU v1: moved out legacy code into a seperate file |
| 12 | * -MMU v3: PD{0,1} bits layout changed: They don't overlap anymore, |
| 13 | * helps avoid a shift when preparing PD0 from PTE |
| 14 | * |
| 15 | * Vineetg: July 2009 |
| 16 | * -For MMU V2, we need not do heuristics at the time of commiting a D-TLB |
| 17 | * entry, so that it doesn't knock out it's I-TLB entry |
| 18 | * -Some more fine tuning: |
| 19 | * bmsk instead of add, asl.cc instead of branch, delay slot utilise etc |
| 20 | * |
| 21 | * Vineetg: July 2009 |
| 22 | * -Practically rewrote the I/D TLB Miss handlers |
| 23 | * Now 40 and 135 instructions a peice as compared to 131 and 449 resp. |
| 24 | * Hence Leaner by 1.5 K |
| 25 | * Used Conditional arithmetic to replace excessive branching |
| 26 | * Also used short instructions wherever possible |
| 27 | * |
| 28 | * Vineetg: Aug 13th 2008 |
| 29 | * -Passing ECR (Exception Cause REG) to do_page_fault( ) for printing |
| 30 | * more information in case of a Fatality |
| 31 | * |
| 32 | * Vineetg: March 25th Bug #92690 |
| 33 | * -Added Debug Code to check if sw-ASID == hw-ASID |
| 34 | |
| 35 | * Rahul Trivedi, Amit Bhor: Codito Technologies 2004 |
| 36 | */ |
| 37 | |
| 38 | .cpu A7 |
| 39 | |
| 40 | #include <linux/linkage.h> |
| 41 | #include <asm/entry.h> |
| 42 | #include <asm/tlb.h> |
| 43 | #include <asm/pgtable.h> |
| 44 | #include <asm/arcregs.h> |
| 45 | #include <asm/cache.h> |
| 46 | #include <asm/processor.h> |
| 47 | #if (CONFIG_ARC_MMU_VER == 1) |
| 48 | #include <asm/tlb-mmu1.h> |
| 49 | #endif |
| 50 | |
| 51 | ;-------------------------------------------------------------------------- |
| 52 | ; scratch memory to save the registers (r0-r3) used to code TLB refill Handler |
| 53 | ; For details refer to comments before TLBMISS_FREEUP_REGS below |
| 54 | ;-------------------------------------------------------------------------- |
| 55 | |
| 56 | .section .data |
| 57 | .global ex_saved_reg1 |
| 58 | .align 1 << L1_CACHE_SHIFT ; IMP: Must be Cache Line aligned |
| 59 | .type ex_saved_reg1, @object |
| 60 | .size ex_saved_reg1, 16 |
| 61 | ex_saved_reg1: |
| 62 | .zero 16 |
| 63 | |
| 64 | ;============================================================================ |
| 65 | ; Troubleshooting Stuff |
| 66 | ;============================================================================ |
| 67 | |
| 68 | ; Linux keeps ASID (Address Space ID) in task->active_mm->context.asid |
| 69 | ; When Creating TLB Entries, instead of doing 3 dependent loads from memory, |
| 70 | ; we use the MMU PID Reg to get current ASID. |
| 71 | ; In bizzare scenrios SW and HW ASID can get out-of-sync which is trouble. |
| 72 | ; So we try to detect this in TLB Mis shandler |
| 73 | |
| 74 | |
| 75 | .macro DBG_ASID_MISMATCH |
| 76 | |
| 77 | #ifdef CONFIG_ARC_DBG_TLB_PARANOIA |
| 78 | |
| 79 | ; make sure h/w ASID is same as s/w ASID |
| 80 | |
| 81 | GET_CURR_TASK_ON_CPU r3 |
| 82 | ld r0, [r3, TASK_ACT_MM] |
| 83 | ld r0, [r0, MM_CTXT+MM_CTXT_ASID] |
| 84 | |
| 85 | lr r1, [ARC_REG_PID] |
| 86 | and r1, r1, 0xFF |
| 87 | breq r1, r0, 5f |
| 88 | |
| 89 | ; Error if H/w and S/w ASID don't match, but NOT if in kernel mode |
| 90 | lr r0, [erstatus] |
| 91 | bbit0 r0, STATUS_U_BIT, 5f |
| 92 | |
| 93 | ; We sure are in troubled waters, Flag the error, but to do so |
| 94 | ; need to switch to kernel mode stack to call error routine |
| 95 | GET_TSK_STACK_BASE r3, sp |
| 96 | |
| 97 | ; Call printk to shoutout aloud |
| 98 | mov r0, 1 |
| 99 | j print_asid_mismatch |
| 100 | |
| 101 | 5: ; ASIDs match so proceed normally |
| 102 | nop |
| 103 | |
| 104 | #endif |
| 105 | |
| 106 | .endm |
| 107 | |
| 108 | ;============================================================================ |
| 109 | ;TLB Miss handling Code |
| 110 | ;============================================================================ |
| 111 | |
| 112 | ;----------------------------------------------------------------------------- |
| 113 | ; This macro does the page-table lookup for the faulting address. |
| 114 | ; OUT: r0 = PTE faulted on, r1 = ptr to PTE, r2 = Faulting V-address |
| 115 | .macro LOAD_FAULT_PTE |
| 116 | |
| 117 | lr r2, [efa] |
| 118 | |
| 119 | lr r1, [ARC_REG_SCRATCH_DATA0] ; current pgd |
| 120 | |
| 121 | lsr r0, r2, PGDIR_SHIFT ; Bits for indexing into PGD |
| 122 | ld.as r1, [r1, r0] ; PGD entry corresp to faulting addr |
| 123 | and.f r1, r1, PAGE_MASK ; Ignoring protection and other flags |
| 124 | ; contains Ptr to Page Table |
| 125 | bz.d do_slow_path_pf ; if no Page Table, do page fault |
| 126 | |
| 127 | ; Get the PTE entry: The idea is |
| 128 | ; (1) x = addr >> PAGE_SHIFT -> masks page-off bits from @fault-addr |
| 129 | ; (2) y = x & (PTRS_PER_PTE - 1) -> to get index |
| 130 | ; (3) z = pgtbl[y] |
| 131 | ; To avoid the multiply by in end, we do the -2, <<2 below |
| 132 | |
| 133 | lsr r0, r2, (PAGE_SHIFT - 2) |
| 134 | and r0, r0, ( (PTRS_PER_PTE - 1) << 2) |
| 135 | ld.aw r0, [r1, r0] ; get PTE and PTE ptr for fault addr |
| 136 | |
| 137 | .endm |
| 138 | |
| 139 | ;----------------------------------------------------------------- |
| 140 | ; Convert Linux PTE entry into TLB entry |
| 141 | ; A one-word PTE entry is programmed as two-word TLB Entry [PD0:PD1] in mmu |
| 142 | ; IN: r0 = PTE, r1 = ptr to PTE |
| 143 | |
| 144 | .macro CONV_PTE_TO_TLB |
| 145 | and r3, r0, PTE_BITS_IN_PD1 ; Extract permission flags+PFN from PTE |
| 146 | sr r3, [ARC_REG_TLBPD1] ; these go in PD1 |
| 147 | |
| 148 | and r2, r0, PTE_BITS_IN_PD0 ; Extract other PTE flags: (V)alid, (G)lb |
| 149 | #if (CONFIG_ARC_MMU_VER <= 2) /* Neednot be done with v3 onwards */ |
| 150 | lsr r2, r2 ; shift PTE flags to match layout in PD0 |
| 151 | #endif |
| 152 | |
| 153 | lr r3,[ARC_REG_TLBPD0] ; MMU prepares PD0 with vaddr and asid |
| 154 | |
| 155 | or r3, r3, r2 ; S | vaddr | {sasid|asid} |
| 156 | sr r3,[ARC_REG_TLBPD0] ; rewrite PD0 |
| 157 | .endm |
| 158 | |
| 159 | ;----------------------------------------------------------------- |
| 160 | ; Commit the TLB entry into MMU |
| 161 | |
| 162 | .macro COMMIT_ENTRY_TO_MMU |
| 163 | |
| 164 | /* Get free TLB slot: Set = computed from vaddr, way = random */ |
| 165 | sr TLBGetIndex, [ARC_REG_TLBCOMMAND] |
| 166 | |
| 167 | /* Commit the Write */ |
| 168 | #if (CONFIG_ARC_MMU_VER >= 2) /* introduced in v2 */ |
| 169 | sr TLBWriteNI, [ARC_REG_TLBCOMMAND] |
| 170 | #else |
| 171 | sr TLBWrite, [ARC_REG_TLBCOMMAND] |
| 172 | #endif |
| 173 | .endm |
| 174 | |
| 175 | ;----------------------------------------------------------------- |
| 176 | ; ARC700 Exception Handling doesn't auto-switch stack and it only provides |
| 177 | ; ONE scratch AUX reg "ARC_REG_SCRATCH_DATA0" |
| 178 | ; |
| 179 | ; For Non-SMP, the scratch AUX reg is repurposed to cache task PGD, so a |
| 180 | ; "global" is used to free-up FIRST core reg to be able to code the rest of |
| 181 | ; exception prologue (IRQ auto-disabled on Exceptions, so it's IRQ-safe). |
| 182 | ; Since the Fast Path TLB Miss handler is coded with 4 regs, the remaining 3 |
| 183 | ; need to be saved as well by extending the "global" to be 4 words. Hence |
| 184 | ; ".size ex_saved_reg1, 16" |
| 185 | ; [All of this dance is to avoid stack switching for each TLB Miss, since we |
| 186 | ; only need to save only a handful of regs, as opposed to complete reg file] |
| 187 | |
| 188 | ; As simple as that.... |
| 189 | |
| 190 | .macro TLBMISS_FREEUP_REGS |
| 191 | st r0, [@ex_saved_reg1] |
| 192 | mov_s r0, @ex_saved_reg1 |
| 193 | st_s r1, [r0, 4] |
| 194 | st_s r2, [r0, 8] |
| 195 | st_s r3, [r0, 12] |
| 196 | |
| 197 | ; VERIFY if the ASID in MMU-PID Reg is same as |
| 198 | ; one in Linux data structures |
| 199 | |
| 200 | DBG_ASID_MISMATCH |
| 201 | .endm |
| 202 | |
| 203 | ;----------------------------------------------------------------- |
| 204 | .macro TLBMISS_RESTORE_REGS |
| 205 | mov_s r0, @ex_saved_reg1 |
| 206 | ld_s r3, [r0,12] |
| 207 | ld_s r2, [r0, 8] |
| 208 | ld_s r1, [r0, 4] |
| 209 | ld_s r0, [r0] |
| 210 | .endm |
| 211 | |
| 212 | .section .text, "ax",@progbits ;Fast Path Code, candidate for ICCM |
| 213 | |
| 214 | ;----------------------------------------------------------------------------- |
| 215 | ; I-TLB Miss Exception Handler |
| 216 | ;----------------------------------------------------------------------------- |
| 217 | |
| 218 | ARC_ENTRY EV_TLBMissI |
| 219 | |
| 220 | TLBMISS_FREEUP_REGS |
| 221 | |
| 222 | ;---------------------------------------------------------------- |
| 223 | ; Get the PTE corresponding to V-addr accessed |
| 224 | LOAD_FAULT_PTE |
| 225 | |
| 226 | ;---------------------------------------------------------------- |
| 227 | ; VERIFY_PTE: Check if PTE permissions approp for executing code |
| 228 | cmp_s r2, VMALLOC_START |
| 229 | mov.lo r2, (_PAGE_PRESENT | _PAGE_READ | _PAGE_EXECUTE) |
| 230 | mov.hs r2, (_PAGE_PRESENT | _PAGE_K_READ | _PAGE_K_EXECUTE) |
| 231 | |
| 232 | and r3, r0, r2 ; Mask out NON Flag bits from PTE |
| 233 | xor.f r3, r3, r2 ; check ( ( pte & flags_test ) == flags_test ) |
| 234 | bnz do_slow_path_pf |
| 235 | |
| 236 | ; Let Linux VM know that the page was accessed |
| 237 | or r0, r0, (_PAGE_PRESENT | _PAGE_ACCESSED) ; set Accessed Bit |
| 238 | st_s r0, [r1] ; Write back PTE |
| 239 | |
| 240 | CONV_PTE_TO_TLB |
| 241 | COMMIT_ENTRY_TO_MMU |
| 242 | TLBMISS_RESTORE_REGS |
| 243 | rtie |
| 244 | |
| 245 | ARC_EXIT EV_TLBMissI |
| 246 | |
| 247 | ;----------------------------------------------------------------------------- |
| 248 | ; D-TLB Miss Exception Handler |
| 249 | ;----------------------------------------------------------------------------- |
| 250 | |
| 251 | ARC_ENTRY EV_TLBMissD |
| 252 | |
| 253 | TLBMISS_FREEUP_REGS |
| 254 | |
| 255 | ;---------------------------------------------------------------- |
| 256 | ; Get the PTE corresponding to V-addr accessed |
| 257 | ; If PTE exists, it will setup, r0 = PTE, r1 = Ptr to PTE |
| 258 | LOAD_FAULT_PTE |
| 259 | |
| 260 | ;---------------------------------------------------------------- |
| 261 | ; VERIFY_PTE: Chk if PTE permissions approp for data access (R/W/R+W) |
| 262 | |
| 263 | mov_s r2, 0 |
| 264 | lr r3, [ecr] |
| 265 | btst_s r3, ECR_C_BIT_DTLB_LD_MISS ; Read Access |
| 266 | or.nz r2, r2, _PAGE_READ ; chk for Read flag in PTE |
| 267 | btst_s r3, ECR_C_BIT_DTLB_ST_MISS ; Write Access |
| 268 | or.nz r2, r2, _PAGE_WRITE ; chk for Write flag in PTE |
| 269 | ; Above laddering takes care of XCHG access |
| 270 | ; which is both Read and Write |
| 271 | |
| 272 | ; If kernel mode access, ; make _PAGE_xx flags as _PAGE_K_xx |
| 273 | ; For copy_(to|from)_user, despite exception taken in kernel mode, |
| 274 | ; this code is not hit, because EFA would still be the user mode |
| 275 | ; address (EFA < 0x6000_0000). |
| 276 | ; This code is for legit kernel mode faults, vmalloc specifically |
| 277 | ; (EFA: 0x7000_0000 to 0x7FFF_FFFF) |
| 278 | |
| 279 | lr r3, [efa] |
| 280 | cmp r3, VMALLOC_START - 1 ; If kernel mode access |
| 281 | asl.hi r2, r2, 3 ; make _PAGE_xx flags as _PAGE_K_xx |
| 282 | or r2, r2, _PAGE_PRESENT ; Common flag for K/U mode |
| 283 | |
| 284 | ; By now, r2 setup with all the Flags we need to check in PTE |
| 285 | and r3, r0, r2 ; Mask out NON Flag bits from PTE |
| 286 | brne.d r3, r2, do_slow_path_pf ; is ((pte & flags_test) == flags_test) |
| 287 | |
| 288 | ;---------------------------------------------------------------- |
| 289 | ; UPDATE_PTE: Let Linux VM know that page was accessed/dirty |
| 290 | lr r3, [ecr] |
| 291 | or r0, r0, (_PAGE_PRESENT | _PAGE_ACCESSED) ; Accessed bit always |
| 292 | btst_s r3, ECR_C_BIT_DTLB_ST_MISS ; See if it was a Write Access ? |
| 293 | or.nz r0, r0, _PAGE_MODIFIED ; if Write, set Dirty bit as well |
| 294 | st_s r0, [r1] ; Write back PTE |
| 295 | |
| 296 | CONV_PTE_TO_TLB |
| 297 | |
| 298 | #if (CONFIG_ARC_MMU_VER == 1) |
| 299 | ; MMU with 2 way set assoc J-TLB, needs some help in pathetic case of |
| 300 | ; memcpy where 3 parties contend for 2 ways, ensuing a livelock. |
| 301 | ; But only for old MMU or one with Metal Fix |
| 302 | TLB_WRITE_HEURISTICS |
| 303 | #endif |
| 304 | |
| 305 | COMMIT_ENTRY_TO_MMU |
| 306 | TLBMISS_RESTORE_REGS |
| 307 | rtie |
| 308 | |
| 309 | ;-------- Common routine to call Linux Page Fault Handler ----------- |
| 310 | do_slow_path_pf: |
| 311 | |
| 312 | ; Restore the 4-scratch regs saved by fast path miss handler |
| 313 | TLBMISS_RESTORE_REGS |
| 314 | |
| 315 | ; Slow path TLB Miss handled as a regular ARC Exception |
| 316 | ; (stack switching / save the complete reg-file). |
| 317 | ; That requires freeing up r9 |
| 318 | EXCPN_PROLOG_FREEUP_REG r9 |
| 319 | |
| 320 | lr r9, [erstatus] |
| 321 | |
| 322 | SWITCH_TO_KERNEL_STK |
| 323 | SAVE_ALL_SYS |
| 324 | |
| 325 | ; ------- setup args for Linux Page fault Hanlder --------- |
| 326 | mov_s r0, sp |
| 327 | lr r2, [efa] |
| 328 | lr r3, [ecr] |
| 329 | |
| 330 | ; Both st and ex imply WRITE access of some sort, hence do_page_fault( ) |
| 331 | ; invoked with write=1 for DTLB-st/ex Miss and write=0 for ITLB miss or |
| 332 | ; DTLB-ld Miss |
| 333 | ; DTLB Miss Cause code is ld = 0x01 , st = 0x02, ex = 0x03 |
| 334 | ; Following code uses that fact that st/ex have one bit in common |
| 335 | |
| 336 | btst_s r3, ECR_C_BIT_DTLB_ST_MISS |
| 337 | mov.z r1, 0 |
| 338 | mov.nz r1, 1 |
| 339 | |
| 340 | ; We don't want exceptions to be disabled while the fault is handled. |
| 341 | ; Now that we have saved the context we return from exception hence |
| 342 | ; exceptions get re-enable |
| 343 | |
| 344 | FAKE_RET_FROM_EXCPN r9 |
| 345 | |
| 346 | bl do_page_fault |
| 347 | b ret_from_exception |
| 348 | |
| 349 | ARC_EXIT EV_TLBMissD |
| 350 | |
| 351 | ARC_ENTRY EV_TLBMissB ; Bogus entry to measure sz of DTLBMiss hdlr |