| /* |
| * arch/arm/kernel/crunch-bits.S |
| * Cirrus MaverickCrunch context switching and handling |
| * |
| * Copyright (C) 2006 Lennert Buytenhek <buytenh@wantstofly.org> |
| * |
| * Shamelessly stolen from the iWMMXt code by Nicolas Pitre, which is |
| * Copyright (c) 2003-2004, MontaVista Software, Inc. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 as |
| * published by the Free Software Foundation. |
| */ |
| |
| #include <linux/linkage.h> |
| #include <asm/ptrace.h> |
| #include <asm/thread_info.h> |
| #include <asm/asm-offsets.h> |
| #include <mach/ep93xx-regs.h> |
| |
| /* |
| * We can't use hex constants here due to a bug in gas. |
| */ |
| #define CRUNCH_MVDX0 0 |
| #define CRUNCH_MVDX1 8 |
| #define CRUNCH_MVDX2 16 |
| #define CRUNCH_MVDX3 24 |
| #define CRUNCH_MVDX4 32 |
| #define CRUNCH_MVDX5 40 |
| #define CRUNCH_MVDX6 48 |
| #define CRUNCH_MVDX7 56 |
| #define CRUNCH_MVDX8 64 |
| #define CRUNCH_MVDX9 72 |
| #define CRUNCH_MVDX10 80 |
| #define CRUNCH_MVDX11 88 |
| #define CRUNCH_MVDX12 96 |
| #define CRUNCH_MVDX13 104 |
| #define CRUNCH_MVDX14 112 |
| #define CRUNCH_MVDX15 120 |
| #define CRUNCH_MVAX0L 128 |
| #define CRUNCH_MVAX0M 132 |
| #define CRUNCH_MVAX0H 136 |
| #define CRUNCH_MVAX1L 140 |
| #define CRUNCH_MVAX1M 144 |
| #define CRUNCH_MVAX1H 148 |
| #define CRUNCH_MVAX2L 152 |
| #define CRUNCH_MVAX2M 156 |
| #define CRUNCH_MVAX2H 160 |
| #define CRUNCH_MVAX3L 164 |
| #define CRUNCH_MVAX3M 168 |
| #define CRUNCH_MVAX3H 172 |
| #define CRUNCH_DSPSC 176 |
| |
| #define CRUNCH_SIZE 184 |
| |
| .text |
| |
| /* |
| * Lazy switching of crunch coprocessor context |
| * |
| * r10 = struct thread_info pointer |
| * r9 = ret_from_exception |
| * lr = undefined instr exit |
| * |
| * called from prefetch exception handler with interrupts disabled |
| */ |
| ENTRY(crunch_task_enable) |
| ldr r8, =(EP93XX_APB_VIRT_BASE + 0x00130000) @ syscon addr |
| |
| ldr r1, [r8, #0x80] |
| tst r1, #0x00800000 @ access to crunch enabled? |
| movne pc, lr @ if so no business here |
| mov r3, #0xaa @ unlock syscon swlock |
| str r3, [r8, #0xc0] |
| orr r1, r1, #0x00800000 @ enable access to crunch |
| str r1, [r8, #0x80] |
| |
| ldr r3, =crunch_owner |
| add r0, r10, #TI_CRUNCH_STATE @ get task crunch save area |
| ldr r2, [sp, #60] @ current task pc value |
| ldr r1, [r3] @ get current crunch owner |
| str r0, [r3] @ this task now owns crunch |
| sub r2, r2, #4 @ adjust pc back |
| str r2, [sp, #60] |
| |
| ldr r2, [r8, #0x80] |
| mov r2, r2 @ flush out enable (@@@) |
| |
| teq r1, #0 @ test for last ownership |
| mov lr, r9 @ normal exit from exception |
| beq crunch_load @ no owner, skip save |
| |
| crunch_save: |
| cfstr64 mvdx0, [r1, #CRUNCH_MVDX0] @ save 64b registers |
| cfstr64 mvdx1, [r1, #CRUNCH_MVDX1] |
| cfstr64 mvdx2, [r1, #CRUNCH_MVDX2] |
| cfstr64 mvdx3, [r1, #CRUNCH_MVDX3] |
| cfstr64 mvdx4, [r1, #CRUNCH_MVDX4] |
| cfstr64 mvdx5, [r1, #CRUNCH_MVDX5] |
| cfstr64 mvdx6, [r1, #CRUNCH_MVDX6] |
| cfstr64 mvdx7, [r1, #CRUNCH_MVDX7] |
| cfstr64 mvdx8, [r1, #CRUNCH_MVDX8] |
| cfstr64 mvdx9, [r1, #CRUNCH_MVDX9] |
| cfstr64 mvdx10, [r1, #CRUNCH_MVDX10] |
| cfstr64 mvdx11, [r1, #CRUNCH_MVDX11] |
| cfstr64 mvdx12, [r1, #CRUNCH_MVDX12] |
| cfstr64 mvdx13, [r1, #CRUNCH_MVDX13] |
| cfstr64 mvdx14, [r1, #CRUNCH_MVDX14] |
| cfstr64 mvdx15, [r1, #CRUNCH_MVDX15] |
| |
| #ifdef __ARMEB__ |
| #error fix me for ARMEB |
| #endif |
| |
| cfmv32al mvfx0, mvax0 @ save 72b accumulators |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX0L] |
| cfmv32am mvfx0, mvax0 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX0M] |
| cfmv32ah mvfx0, mvax0 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX0H] |
| cfmv32al mvfx0, mvax1 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX1L] |
| cfmv32am mvfx0, mvax1 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX1M] |
| cfmv32ah mvfx0, mvax1 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX1H] |
| cfmv32al mvfx0, mvax2 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX2L] |
| cfmv32am mvfx0, mvax2 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX2M] |
| cfmv32ah mvfx0, mvax2 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX2H] |
| cfmv32al mvfx0, mvax3 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX3L] |
| cfmv32am mvfx0, mvax3 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX3M] |
| cfmv32ah mvfx0, mvax3 |
| cfstr32 mvfx0, [r1, #CRUNCH_MVAX3H] |
| |
| cfmv32sc mvdx0, dspsc @ save status word |
| cfstr64 mvdx0, [r1, #CRUNCH_DSPSC] |
| |
| teq r0, #0 @ anything to load? |
| cfldr64eq mvdx0, [r1, #CRUNCH_MVDX0] @ mvdx0 was clobbered |
| moveq pc, lr |
| |
| crunch_load: |
| cfldr64 mvdx0, [r0, #CRUNCH_DSPSC] @ load status word |
| cfmvsc32 dspsc, mvdx0 |
| |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX0L] @ load 72b accumulators |
| cfmval32 mvax0, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX0M] |
| cfmvam32 mvax0, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX0H] |
| cfmvah32 mvax0, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX1L] |
| cfmval32 mvax1, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX1M] |
| cfmvam32 mvax1, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX1H] |
| cfmvah32 mvax1, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX2L] |
| cfmval32 mvax2, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX2M] |
| cfmvam32 mvax2, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX2H] |
| cfmvah32 mvax2, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX3L] |
| cfmval32 mvax3, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX3M] |
| cfmvam32 mvax3, mvfx0 |
| cfldr32 mvfx0, [r0, #CRUNCH_MVAX3H] |
| cfmvah32 mvax3, mvfx0 |
| |
| cfldr64 mvdx0, [r0, #CRUNCH_MVDX0] @ load 64b registers |
| cfldr64 mvdx1, [r0, #CRUNCH_MVDX1] |
| cfldr64 mvdx2, [r0, #CRUNCH_MVDX2] |
| cfldr64 mvdx3, [r0, #CRUNCH_MVDX3] |
| cfldr64 mvdx4, [r0, #CRUNCH_MVDX4] |
| cfldr64 mvdx5, [r0, #CRUNCH_MVDX5] |
| cfldr64 mvdx6, [r0, #CRUNCH_MVDX6] |
| cfldr64 mvdx7, [r0, #CRUNCH_MVDX7] |
| cfldr64 mvdx8, [r0, #CRUNCH_MVDX8] |
| cfldr64 mvdx9, [r0, #CRUNCH_MVDX9] |
| cfldr64 mvdx10, [r0, #CRUNCH_MVDX10] |
| cfldr64 mvdx11, [r0, #CRUNCH_MVDX11] |
| cfldr64 mvdx12, [r0, #CRUNCH_MVDX12] |
| cfldr64 mvdx13, [r0, #CRUNCH_MVDX13] |
| cfldr64 mvdx14, [r0, #CRUNCH_MVDX14] |
| cfldr64 mvdx15, [r0, #CRUNCH_MVDX15] |
| |
| mov pc, lr |
| |
| /* |
| * Back up crunch regs to save area and disable access to them |
| * (mainly for gdb or sleep mode usage) |
| * |
| * r0 = struct thread_info pointer of target task or NULL for any |
| */ |
| ENTRY(crunch_task_disable) |
| stmfd sp!, {r4, r5, lr} |
| |
| mrs ip, cpsr |
| orr r2, ip, #PSR_I_BIT @ disable interrupts |
| msr cpsr_c, r2 |
| |
| ldr r4, =(EP93XX_APB_VIRT_BASE + 0x00130000) @ syscon addr |
| |
| ldr r3, =crunch_owner |
| add r2, r0, #TI_CRUNCH_STATE @ get task crunch save area |
| ldr r1, [r3] @ get current crunch owner |
| teq r1, #0 @ any current owner? |
| beq 1f @ no: quit |
| teq r0, #0 @ any owner? |
| teqne r1, r2 @ or specified one? |
| bne 1f @ no: quit |
| |
| ldr r5, [r4, #0x80] @ enable access to crunch |
| mov r2, #0xaa |
| str r2, [r4, #0xc0] |
| orr r5, r5, #0x00800000 |
| str r5, [r4, #0x80] |
| |
| mov r0, #0 @ nothing to load |
| str r0, [r3] @ no more current owner |
| ldr r2, [r4, #0x80] @ flush out enable (@@@) |
| mov r2, r2 |
| bl crunch_save |
| |
| mov r2, #0xaa @ disable access to crunch |
| str r2, [r4, #0xc0] |
| bic r5, r5, #0x00800000 |
| str r5, [r4, #0x80] |
| ldr r5, [r4, #0x80] @ flush out enable (@@@) |
| mov r5, r5 |
| |
| 1: msr cpsr_c, ip @ restore interrupt mode |
| ldmfd sp!, {r4, r5, pc} |
| |
| /* |
| * Copy crunch state to given memory address |
| * |
| * r0 = struct thread_info pointer of target task |
| * r1 = memory address where to store crunch state |
| * |
| * this is called mainly in the creation of signal stack frames |
| */ |
| ENTRY(crunch_task_copy) |
| mrs ip, cpsr |
| orr r2, ip, #PSR_I_BIT @ disable interrupts |
| msr cpsr_c, r2 |
| |
| ldr r3, =crunch_owner |
| add r2, r0, #TI_CRUNCH_STATE @ get task crunch save area |
| ldr r3, [r3] @ get current crunch owner |
| teq r2, r3 @ does this task own it... |
| beq 1f |
| |
| @ current crunch values are in the task save area |
| msr cpsr_c, ip @ restore interrupt mode |
| mov r0, r1 |
| mov r1, r2 |
| mov r2, #CRUNCH_SIZE |
| b memcpy |
| |
| 1: @ this task owns crunch regs -- grab a copy from there |
| mov r0, #0 @ nothing to load |
| mov r3, lr @ preserve return address |
| bl crunch_save |
| msr cpsr_c, ip @ restore interrupt mode |
| mov pc, r3 |
| |
| /* |
| * Restore crunch state from given memory address |
| * |
| * r0 = struct thread_info pointer of target task |
| * r1 = memory address where to get crunch state from |
| * |
| * this is used to restore crunch state when unwinding a signal stack frame |
| */ |
| ENTRY(crunch_task_restore) |
| mrs ip, cpsr |
| orr r2, ip, #PSR_I_BIT @ disable interrupts |
| msr cpsr_c, r2 |
| |
| ldr r3, =crunch_owner |
| add r2, r0, #TI_CRUNCH_STATE @ get task crunch save area |
| ldr r3, [r3] @ get current crunch owner |
| teq r2, r3 @ does this task own it... |
| beq 1f |
| |
| @ this task doesn't own crunch regs -- use its save area |
| msr cpsr_c, ip @ restore interrupt mode |
| mov r0, r2 |
| mov r2, #CRUNCH_SIZE |
| b memcpy |
| |
| 1: @ this task owns crunch regs -- load them directly |
| mov r0, r1 |
| mov r1, #0 @ nothing to save |
| mov r3, lr @ preserve return address |
| bl crunch_load |
| msr cpsr_c, ip @ restore interrupt mode |
| mov pc, r3 |