| /* |
| * Copyright (C) 2013 Imagination Technologies |
| * Author: Paul Burton <paul.burton@imgtec.com> |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License as published by the |
| * Free Software Foundation; either version 2 of the License, or (at your |
| * option) any later version. |
| */ |
| #ifndef _ASM_MSA_H |
| #define _ASM_MSA_H |
| |
| #include <asm/mipsregs.h> |
| |
| #ifndef __ASSEMBLY__ |
| |
| #include <asm/inst.h> |
| |
| extern void _save_msa(struct task_struct *); |
| extern void _restore_msa(struct task_struct *); |
| extern void _init_msa_upper(void); |
| |
| extern void read_msa_wr_b(unsigned idx, union fpureg *to); |
| extern void read_msa_wr_h(unsigned idx, union fpureg *to); |
| extern void read_msa_wr_w(unsigned idx, union fpureg *to); |
| extern void read_msa_wr_d(unsigned idx, union fpureg *to); |
| |
| /** |
| * read_msa_wr() - Read a single MSA vector register |
| * @idx: The index of the vector register to read |
| * @to: The FPU register union to store the registers value in |
| * @fmt: The format of the data in the vector register |
| * |
| * Read the value of MSA vector register idx into the FPU register |
| * union to, using the format fmt. |
| */ |
| static inline void read_msa_wr(unsigned idx, union fpureg *to, |
| enum msa_2b_fmt fmt) |
| { |
| switch (fmt) { |
| case msa_fmt_b: |
| read_msa_wr_b(idx, to); |
| break; |
| |
| case msa_fmt_h: |
| read_msa_wr_h(idx, to); |
| break; |
| |
| case msa_fmt_w: |
| read_msa_wr_w(idx, to); |
| break; |
| |
| case msa_fmt_d: |
| read_msa_wr_d(idx, to); |
| break; |
| |
| default: |
| BUG(); |
| } |
| } |
| |
| extern void write_msa_wr_b(unsigned idx, union fpureg *from); |
| extern void write_msa_wr_h(unsigned idx, union fpureg *from); |
| extern void write_msa_wr_w(unsigned idx, union fpureg *from); |
| extern void write_msa_wr_d(unsigned idx, union fpureg *from); |
| |
| /** |
| * write_msa_wr() - Write a single MSA vector register |
| * @idx: The index of the vector register to write |
| * @from: The FPU register union to take the registers value from |
| * @fmt: The format of the data in the vector register |
| * |
| * Write the value from the FPU register union from into MSA vector |
| * register idx, using the format fmt. |
| */ |
| static inline void write_msa_wr(unsigned idx, union fpureg *from, |
| enum msa_2b_fmt fmt) |
| { |
| switch (fmt) { |
| case msa_fmt_b: |
| write_msa_wr_b(idx, from); |
| break; |
| |
| case msa_fmt_h: |
| write_msa_wr_h(idx, from); |
| break; |
| |
| case msa_fmt_w: |
| write_msa_wr_w(idx, from); |
| break; |
| |
| case msa_fmt_d: |
| write_msa_wr_d(idx, from); |
| break; |
| |
| default: |
| BUG(); |
| } |
| } |
| |
| static inline void enable_msa(void) |
| { |
| if (cpu_has_msa) { |
| set_c0_config5(MIPS_CONF5_MSAEN); |
| enable_fpu_hazard(); |
| } |
| } |
| |
| static inline void disable_msa(void) |
| { |
| if (cpu_has_msa) { |
| clear_c0_config5(MIPS_CONF5_MSAEN); |
| disable_fpu_hazard(); |
| } |
| } |
| |
| static inline int is_msa_enabled(void) |
| { |
| if (!cpu_has_msa) |
| return 0; |
| |
| return read_c0_config5() & MIPS_CONF5_MSAEN; |
| } |
| |
| static inline int thread_msa_context_live(void) |
| { |
| /* |
| * Check cpu_has_msa only if it's a constant. This will allow the |
| * compiler to optimise out code for CPUs without MSA without adding |
| * an extra redundant check for CPUs with MSA. |
| */ |
| if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa) |
| return 0; |
| |
| return test_thread_flag(TIF_MSA_CTX_LIVE); |
| } |
| |
| static inline void save_msa(struct task_struct *t) |
| { |
| if (cpu_has_msa) |
| _save_msa(t); |
| } |
| |
| static inline void restore_msa(struct task_struct *t) |
| { |
| if (cpu_has_msa) |
| _restore_msa(t); |
| } |
| |
| #ifdef TOOLCHAIN_SUPPORTS_MSA |
| |
| #define __BUILD_MSA_CTL_REG(name, cs) \ |
| static inline unsigned int read_msa_##name(void) \ |
| { \ |
| unsigned int reg; \ |
| __asm__ __volatile__( \ |
| " .set push\n" \ |
| " .set msa\n" \ |
| " cfcmsa %0, $" #cs "\n" \ |
| " .set pop\n" \ |
| : "=r"(reg)); \ |
| return reg; \ |
| } \ |
| \ |
| static inline void write_msa_##name(unsigned int val) \ |
| { \ |
| __asm__ __volatile__( \ |
| " .set push\n" \ |
| " .set msa\n" \ |
| " ctcmsa $" #cs ", %0\n" \ |
| " .set pop\n" \ |
| : : "r"(val)); \ |
| } |
| |
| #else /* !TOOLCHAIN_SUPPORTS_MSA */ |
| |
| /* |
| * Define functions using .word for the c[ft]cmsa instructions in order to |
| * allow compilation with toolchains that do not support MSA. Once all |
| * toolchains in use support MSA these can be removed. |
| */ |
| #ifdef CONFIG_CPU_MICROMIPS |
| #define CFC_MSA_INSN 0x587e0056 |
| #define CTC_MSA_INSN 0x583e0816 |
| #else |
| #define CFC_MSA_INSN 0x787e0059 |
| #define CTC_MSA_INSN 0x783e0819 |
| #endif |
| |
| #define __BUILD_MSA_CTL_REG(name, cs) \ |
| static inline unsigned int read_msa_##name(void) \ |
| { \ |
| unsigned int reg; \ |
| __asm__ __volatile__( \ |
| " .set push\n" \ |
| " .set noat\n" \ |
| " .insn\n" \ |
| " .word %1 | (" #cs " << 11)\n" \ |
| " move %0, $1\n" \ |
| " .set pop\n" \ |
| : "=r"(reg) : "i"(CFC_MSA_INSN)); \ |
| return reg; \ |
| } \ |
| \ |
| static inline void write_msa_##name(unsigned int val) \ |
| { \ |
| __asm__ __volatile__( \ |
| " .set push\n" \ |
| " .set noat\n" \ |
| " move $1, %0\n" \ |
| " .insn\n" \ |
| " .word %1 | (" #cs " << 6)\n" \ |
| " .set pop\n" \ |
| : : "r"(val), "i"(CTC_MSA_INSN)); \ |
| } |
| |
| #endif /* !TOOLCHAIN_SUPPORTS_MSA */ |
| |
| __BUILD_MSA_CTL_REG(ir, 0) |
| __BUILD_MSA_CTL_REG(csr, 1) |
| __BUILD_MSA_CTL_REG(access, 2) |
| __BUILD_MSA_CTL_REG(save, 3) |
| __BUILD_MSA_CTL_REG(modify, 4) |
| __BUILD_MSA_CTL_REG(request, 5) |
| __BUILD_MSA_CTL_REG(map, 6) |
| __BUILD_MSA_CTL_REG(unmap, 7) |
| |
| #endif /* !__ASSEMBLY__ */ |
| |
| #define MSA_IR 0 |
| #define MSA_CSR 1 |
| #define MSA_ACCESS 2 |
| #define MSA_SAVE 3 |
| #define MSA_MODIFY 4 |
| #define MSA_REQUEST 5 |
| #define MSA_MAP 6 |
| #define MSA_UNMAP 7 |
| |
| /* MSA Implementation Register (MSAIR) */ |
| #define MSA_IR_REVB 0 |
| #define MSA_IR_REVF (_ULCAST_(0xff) << MSA_IR_REVB) |
| #define MSA_IR_PROCB 8 |
| #define MSA_IR_PROCF (_ULCAST_(0xff) << MSA_IR_PROCB) |
| #define MSA_IR_WRPB 16 |
| #define MSA_IR_WRPF (_ULCAST_(0x1) << MSA_IR_WRPB) |
| |
| /* MSA Control & Status Register (MSACSR) */ |
| #define MSA_CSR_RMB 0 |
| #define MSA_CSR_RMF (_ULCAST_(0x3) << MSA_CSR_RMB) |
| #define MSA_CSR_RM_NEAREST 0 |
| #define MSA_CSR_RM_TO_ZERO 1 |
| #define MSA_CSR_RM_TO_POS 2 |
| #define MSA_CSR_RM_TO_NEG 3 |
| #define MSA_CSR_FLAGSB 2 |
| #define MSA_CSR_FLAGSF (_ULCAST_(0x1f) << MSA_CSR_FLAGSB) |
| #define MSA_CSR_FLAGS_IB 2 |
| #define MSA_CSR_FLAGS_IF (_ULCAST_(0x1) << MSA_CSR_FLAGS_IB) |
| #define MSA_CSR_FLAGS_UB 3 |
| #define MSA_CSR_FLAGS_UF (_ULCAST_(0x1) << MSA_CSR_FLAGS_UB) |
| #define MSA_CSR_FLAGS_OB 4 |
| #define MSA_CSR_FLAGS_OF (_ULCAST_(0x1) << MSA_CSR_FLAGS_OB) |
| #define MSA_CSR_FLAGS_ZB 5 |
| #define MSA_CSR_FLAGS_ZF (_ULCAST_(0x1) << MSA_CSR_FLAGS_ZB) |
| #define MSA_CSR_FLAGS_VB 6 |
| #define MSA_CSR_FLAGS_VF (_ULCAST_(0x1) << MSA_CSR_FLAGS_VB) |
| #define MSA_CSR_ENABLESB 7 |
| #define MSA_CSR_ENABLESF (_ULCAST_(0x1f) << MSA_CSR_ENABLESB) |
| #define MSA_CSR_ENABLES_IB 7 |
| #define MSA_CSR_ENABLES_IF (_ULCAST_(0x1) << MSA_CSR_ENABLES_IB) |
| #define MSA_CSR_ENABLES_UB 8 |
| #define MSA_CSR_ENABLES_UF (_ULCAST_(0x1) << MSA_CSR_ENABLES_UB) |
| #define MSA_CSR_ENABLES_OB 9 |
| #define MSA_CSR_ENABLES_OF (_ULCAST_(0x1) << MSA_CSR_ENABLES_OB) |
| #define MSA_CSR_ENABLES_ZB 10 |
| #define MSA_CSR_ENABLES_ZF (_ULCAST_(0x1) << MSA_CSR_ENABLES_ZB) |
| #define MSA_CSR_ENABLES_VB 11 |
| #define MSA_CSR_ENABLES_VF (_ULCAST_(0x1) << MSA_CSR_ENABLES_VB) |
| #define MSA_CSR_CAUSEB 12 |
| #define MSA_CSR_CAUSEF (_ULCAST_(0x3f) << MSA_CSR_CAUSEB) |
| #define MSA_CSR_CAUSE_IB 12 |
| #define MSA_CSR_CAUSE_IF (_ULCAST_(0x1) << MSA_CSR_CAUSE_IB) |
| #define MSA_CSR_CAUSE_UB 13 |
| #define MSA_CSR_CAUSE_UF (_ULCAST_(0x1) << MSA_CSR_CAUSE_UB) |
| #define MSA_CSR_CAUSE_OB 14 |
| #define MSA_CSR_CAUSE_OF (_ULCAST_(0x1) << MSA_CSR_CAUSE_OB) |
| #define MSA_CSR_CAUSE_ZB 15 |
| #define MSA_CSR_CAUSE_ZF (_ULCAST_(0x1) << MSA_CSR_CAUSE_ZB) |
| #define MSA_CSR_CAUSE_VB 16 |
| #define MSA_CSR_CAUSE_VF (_ULCAST_(0x1) << MSA_CSR_CAUSE_VB) |
| #define MSA_CSR_CAUSE_EB 17 |
| #define MSA_CSR_CAUSE_EF (_ULCAST_(0x1) << MSA_CSR_CAUSE_EB) |
| #define MSA_CSR_NXB 18 |
| #define MSA_CSR_NXF (_ULCAST_(0x1) << MSA_CSR_NXB) |
| #define MSA_CSR_FSB 24 |
| #define MSA_CSR_FSF (_ULCAST_(0x1) << MSA_CSR_FSB) |
| |
| #endif /* _ASM_MSA_H */ |