| |
| /*---------------------------------------------------------------*/ |
| /*--- begin host_ppc_isel.c ---*/ |
| /*---------------------------------------------------------------*/ |
| |
| /* |
| This file is part of Valgrind, a dynamic binary instrumentation |
| framework. |
| |
| Copyright (C) 2004-2012 OpenWorks LLP |
| info@open-works.net |
| |
| This program is free software; you can redistribute it and/or |
| modify it under the terms of the GNU General Public License as |
| published by the Free Software Foundation; either version 2 of the |
| License, or (at your option) any later version. |
| |
| This program is distributed in the hope that it will be useful, but |
| WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| General Public License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with this program; if not, write to the Free Software |
| Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA |
| 02110-1301, USA. |
| |
| The GNU General Public License is contained in the file COPYING. |
| |
| Neither the names of the U.S. Department of Energy nor the |
| University of California nor the names of its contributors may be |
| used to endorse or promote products derived from this software |
| without prior written permission. |
| */ |
| |
| #include "libvex_basictypes.h" |
| #include "libvex_ir.h" |
| #include "libvex.h" |
| |
| #include "ir_match.h" |
| #include "main_util.h" |
| #include "main_globals.h" |
| #include "host_generic_regs.h" |
| #include "host_generic_simd64.h" |
| #include "host_ppc_defs.h" |
| |
| /* GPR register class for ppc32/64 */ |
| #define HRcGPR(__mode64) (__mode64 ? HRcInt64 : HRcInt32) |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- Register Usage Conventions ---*/ |
| /*---------------------------------------------------------*/ |
| /* |
| Integer Regs |
| ------------ |
| GPR0 Reserved |
| GPR1 Stack Pointer |
| GPR2 not used - TOC pointer |
| GPR3:10 Allocateable |
| GPR11 if mode64: not used - calls by ptr / env ptr for some langs |
| GPR12 if mode64: not used - exceptions / global linkage code |
| GPR13 not used - Thread-specific pointer |
| GPR14:28 Allocateable |
| GPR29 Unused by us (reserved for the dispatcher) |
| GPR30 AltiVec temp spill register |
| GPR31 GuestStatePointer |
| |
| Of Allocateable regs: |
| if (mode64) |
| GPR3:10 Caller-saved regs |
| else |
| GPR3:12 Caller-saved regs |
| GPR14:29 Callee-saved regs |
| |
| GPR3 [Return | Parameter] - carrying reg |
| GPR4:10 Parameter-carrying regs |
| |
| |
| Floating Point Regs |
| ------------------- |
| FPR0:31 Allocateable |
| |
| FPR0 Caller-saved - scratch reg |
| if (mode64) |
| FPR1:13 Caller-saved - param & return regs |
| else |
| FPR1:8 Caller-saved - param & return regs |
| FPR9:13 Caller-saved regs |
| FPR14:31 Callee-saved regs |
| |
| |
| Vector Regs (on processors with the VMX feature) |
| ----------- |
| VR0-VR1 Volatile scratch registers |
| VR2-VR13 Volatile vector parameters registers |
| VR14-VR19 Volatile scratch registers |
| VR20-VR31 Non-volatile registers |
| VRSAVE Non-volatile 32-bit register |
| */ |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- PPC FP Status & Control Register Conventions ---*/ |
| /*---------------------------------------------------------*/ |
| /* |
| Vex-generated code expects to run with the FPU set as follows: all |
| exceptions masked. The rounding mode is set appropriately before |
| each floating point insn emitted (or left unchanged if known to be |
| correct already). There are a few fp insns (fmr,fneg,fabs,fnabs), |
| which are unaffected by the rm and so the rounding mode is not set |
| prior to them. |
| |
| At least on MPC7447A (Mac Mini), frsqrte is also not affected by |
| rounding mode. At some point the ppc docs get sufficiently vague |
| that the only way to find out is to write test programs. |
| */ |
| /* Notes on the FP instruction set, 6 Feb 06. |
| |
| What exns -> CR1 ? Sets FPRF ? Observes RM ? |
| ------------------------------------------------------------- |
| |
| fmr[.] if . n n |
| fneg[.] if . n n |
| fabs[.] if . n n |
| fnabs[.] if . n n |
| |
| fadd[.] if . y y |
| fadds[.] if . y y |
| fcfid[.] (Si64->dbl) if . y y |
| fcfidU[.] (Ui64->dbl) if . y y |
| fcfids[.] (Si64->sngl) if . Y Y |
| fcfidus[.] (Ui64->sngl) if . Y Y |
| fcmpo (cmp, result n n n |
| fcmpu to crfD) n n n |
| fctid[.] (dbl->i64) if . ->undef y |
| fctidz[.] (dbl->i64) if . ->undef rounds-to-zero |
| fctiw[.] (dbl->i32) if . ->undef y |
| fctiwz[.] (dbl->i32) if . ->undef rounds-to-zero |
| fdiv[.] if . y y |
| fdivs[.] if . y y |
| fmadd[.] if . y y |
| fmadds[.] if . y y |
| fmsub[.] if . y y |
| fmsubs[.] if . y y |
| fmul[.] if . y y |
| fmuls[.] if . y y |
| |
| (note: for fnm*, rounding happens before final negation) |
| fnmadd[.] if . y y |
| fnmadds[.] if . y y |
| fnmsub[.] if . y y |
| fnmsubs[.] if . y y |
| |
| fre[.] if . y y |
| fres[.] if . y y |
| |
| frsqrte[.] if . y apparently not |
| |
| fsqrt[.] if . y y |
| fsqrts[.] if . y y |
| fsub[.] if . y y |
| fsubs[.] if . y y |
| |
| |
| fpscr: bits 30-31 (ibm) is RM |
| 24-29 (ibm) are exnmasks/non-IEEE bit, all zero |
| 15-19 (ibm) is FPRF: class, <, =, >, UNord |
| |
| ppc fe(guest) makes fpscr read as all zeros except RM (and maybe FPRF |
| in future) |
| |
| mcrfs - move fpscr field to CR field |
| mtfsfi[.] - 4 bit imm moved to fpscr field |
| mtfsf[.] - move frS[low 1/2] to fpscr but using 8-bit field mask |
| mtfsb1[.] - set given fpscr bit |
| mtfsb0[.] - clear given fpscr bit |
| mffs[.] - move all fpscr to frD[low 1/2] |
| |
| For [.] presumably cr1 is set with exn summary bits, as per |
| main FP insns |
| |
| A single precision store truncates/denormalises the in-register value, |
| but does not round it. This is so that flds followed by fsts is |
| always the identity. |
| */ |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- misc helpers ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* These are duplicated in guest-ppc/toIR.c */ |
| static IRExpr* unop ( IROp op, IRExpr* a ) |
| { |
| return IRExpr_Unop(op, a); |
| } |
| |
| static IRExpr* mkU32 ( UInt i ) |
| { |
| return IRExpr_Const(IRConst_U32(i)); |
| } |
| |
| static IRExpr* bind ( Int binder ) |
| { |
| return IRExpr_Binder(binder); |
| } |
| |
| static Bool isZeroU8 ( IRExpr* e ) |
| { |
| return e->tag == Iex_Const |
| && e->Iex.Const.con->tag == Ico_U8 |
| && e->Iex.Const.con->Ico.U8 == 0; |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISelEnv ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* This carries around: |
| |
| - A mapping from IRTemp to IRType, giving the type of any IRTemp we |
| might encounter. This is computed before insn selection starts, |
| and does not change. |
| |
| - A mapping from IRTemp to HReg. This tells the insn selector |
| which virtual register(s) are associated with each IRTemp |
| temporary. This is computed before insn selection starts, and |
| does not change. We expect this mapping to map precisely the |
| same set of IRTemps as the type mapping does. |
| |
| - vregmapLo holds the primary register for the IRTemp. |
| - vregmapMedLo holds the secondary register for the IRTemp, |
| if any is needed. That's only for Ity_I64 temps |
| in 32 bit mode or Ity_I128 temps in 64-bit mode. |
| - vregmapMedHi is only for dealing with Ity_I128 temps in |
| 32 bit mode. It holds bits 95:64 (Intel numbering) |
| of the IRTemp. |
| - vregmapHi is also only for dealing with Ity_I128 temps |
| in 32 bit mode. It holds the most significant bits |
| (127:96 in Intel numbering) of the IRTemp. |
| |
| - The code array, that is, the insns selected so far. |
| |
| - A counter, for generating new virtual registers. |
| |
| - The host subarchitecture we are selecting insns for. |
| This is set at the start and does not change. |
| |
| - A Bool to tell us if the host is 32 or 64bit. |
| This is set at the start and does not change. |
| |
| - An IRExpr*, which may be NULL, holding the IR expression (an |
| IRRoundingMode-encoded value) to which the FPU's rounding mode |
| was most recently set. Setting to NULL is always safe. Used to |
| avoid redundant settings of the FPU's rounding mode, as |
| described in set_FPU_rounding_mode below. |
| |
| - A VexMiscInfo*, needed for knowing how to generate |
| function calls for this target. |
| |
| - The maximum guest address of any guest insn in this block. |
| Actually, the address of the highest-addressed byte from any |
| insn in this block. Is set at the start and does not change. |
| This is used for detecting jumps which are definitely |
| forward-edges from this block, and therefore can be made |
| (chained) to the fast entry point of the destination, thereby |
| avoiding the destination's event check. |
| */ |
| |
| typedef |
| struct { |
| /* Constant -- are set at the start and do not change. */ |
| IRTypeEnv* type_env; |
| // 64-bit mode 32-bit mode |
| HReg* vregmapLo; // Low 64-bits [63:0] Low 32-bits [31:0] |
| HReg* vregmapMedLo; // high 64-bits[127:64] Next 32-bits [63:32] |
| HReg* vregmapMedHi; // unused Next 32-bits [95:64] |
| HReg* vregmapHi; // unused highest 32-bits [127:96] |
| Int n_vregmap; |
| |
| /* 27 Jan 06: Not currently used, but should be */ |
| UInt hwcaps; |
| |
| Bool mode64; |
| |
| VexAbiInfo* vbi; |
| |
| Bool chainingAllowed; |
| Addr64 max_ga; |
| |
| /* These are modified as we go along. */ |
| HInstrArray* code; |
| Int vreg_ctr; |
| |
| IRExpr* previous_rm; |
| } |
| ISelEnv; |
| |
| |
| static HReg lookupIRTemp ( ISelEnv* env, IRTemp tmp ) |
| { |
| vassert(tmp >= 0); |
| vassert(tmp < env->n_vregmap); |
| return env->vregmapLo[tmp]; |
| } |
| |
| static void lookupIRTempPair ( HReg* vrHI, HReg* vrLO, |
| ISelEnv* env, IRTemp tmp ) |
| { |
| vassert(tmp >= 0); |
| vassert(tmp < env->n_vregmap); |
| vassert(! hregIsInvalid(env->vregmapMedLo[tmp])); |
| *vrLO = env->vregmapLo[tmp]; |
| *vrHI = env->vregmapMedLo[tmp]; |
| } |
| |
| /* Only for used in 32-bit mode */ |
| static void lookupIRTempQuad ( HReg* vrHi, HReg* vrMedHi, HReg* vrMedLo, |
| HReg* vrLo, ISelEnv* env, IRTemp tmp ) |
| { |
| vassert(!env->mode64); |
| vassert(tmp >= 0); |
| vassert(tmp < env->n_vregmap); |
| vassert(! hregIsInvalid(env->vregmapMedLo[tmp])); |
| *vrHi = env->vregmapHi[tmp]; |
| *vrMedHi = env->vregmapMedHi[tmp]; |
| *vrMedLo = env->vregmapMedLo[tmp]; |
| *vrLo = env->vregmapLo[tmp]; |
| } |
| |
| static void addInstr ( ISelEnv* env, PPCInstr* instr ) |
| { |
| addHInstr(env->code, instr); |
| if (vex_traceflags & VEX_TRACE_VCODE) { |
| ppPPCInstr(instr, env->mode64); |
| vex_printf("\n"); |
| } |
| } |
| |
| static HReg newVRegI ( ISelEnv* env ) |
| { |
| HReg reg = mkHReg(env->vreg_ctr, HRcGPR(env->mode64), |
| True/*virtual reg*/); |
| env->vreg_ctr++; |
| return reg; |
| } |
| |
| static HReg newVRegF ( ISelEnv* env ) |
| { |
| HReg reg = mkHReg(env->vreg_ctr, HRcFlt64, True/*virtual reg*/); |
| env->vreg_ctr++; |
| return reg; |
| } |
| |
| static HReg newVRegV ( ISelEnv* env ) |
| { |
| HReg reg = mkHReg(env->vreg_ctr, HRcVec128, True/*virtual reg*/); |
| env->vreg_ctr++; |
| return reg; |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Forward declarations ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* These are organised as iselXXX and iselXXX_wrk pairs. The |
| iselXXX_wrk do the real work, but are not to be called directly. |
| For each XXX, iselXXX calls its iselXXX_wrk counterpart, then |
| checks that all returned registers are virtual. You should not |
| call the _wrk version directly. |
| |
| 'Word' refers to the size of the native machine word, that is, |
| 32-bit int in 32-bit mode and 64-bit int in 64-bit mode. '2Word' |
| therefore refers to a double-width (64/128-bit) quantity in two |
| integer registers. |
| */ |
| /* 32-bit mode: compute an I8/I16/I32 into a GPR. |
| 64-bit mode: compute an I8/I16/I32/I64 into a GPR. */ |
| static HReg iselWordExpr_R_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselWordExpr_R ( ISelEnv* env, IRExpr* e ); |
| |
| /* 32-bit mode: Compute an I8/I16/I32 into a RH |
| (reg-or-halfword-immediate). |
| 64-bit mode: Compute an I8/I16/I32/I64 into a RH |
| (reg-or-halfword-immediate). |
| It's important to specify whether the immediate is to be regarded |
| as signed or not. If yes, this will never return -32768 as an |
| immediate; this guaranteed that all signed immediates that are |
| return can have their sign inverted if need be. |
| */ |
| static PPCRH* iselWordExpr_RH_wrk ( ISelEnv* env, |
| Bool syned, IRExpr* e ); |
| static PPCRH* iselWordExpr_RH ( ISelEnv* env, |
| Bool syned, IRExpr* e ); |
| |
| /* 32-bit mode: compute an I32 into a RI (reg or 32-bit immediate). |
| 64-bit mode: compute an I64 into a RI (reg or 64-bit immediate). */ |
| static PPCRI* iselWordExpr_RI_wrk ( ISelEnv* env, IRExpr* e ); |
| static PPCRI* iselWordExpr_RI ( ISelEnv* env, IRExpr* e ); |
| |
| /* In 32 bit mode ONLY, compute an I8 into a |
| reg-or-5-bit-unsigned-immediate, the latter being an immediate in |
| the range 1 .. 31 inclusive. Used for doing shift amounts. */ |
| static PPCRH* iselWordExpr_RH5u_wrk ( ISelEnv* env, IRExpr* e ); |
| static PPCRH* iselWordExpr_RH5u ( ISelEnv* env, IRExpr* e ); |
| |
| /* In 64-bit mode ONLY, compute an I8 into a |
| reg-or-6-bit-unsigned-immediate, the latter being an immediate in |
| the range 1 .. 63 inclusive. Used for doing shift amounts. */ |
| static PPCRH* iselWordExpr_RH6u_wrk ( ISelEnv* env, IRExpr* e ); |
| static PPCRH* iselWordExpr_RH6u ( ISelEnv* env, IRExpr* e ); |
| |
| /* 32-bit mode: compute an I32 into an AMode. |
| 64-bit mode: compute an I64 into an AMode. |
| |
| Requires to know (xferTy) the type of data to be loaded/stored |
| using this amode. That is so that, for 64-bit code generation, any |
| PPCAMode_IR returned will have an index (immediate offset) field |
| that is guaranteed to be 4-aligned, if there is any chance that the |
| amode is to be used in ld/ldu/lda/std/stdu. |
| |
| Since there are no such restrictions on 32-bit insns, xferTy is |
| ignored for 32-bit code generation. */ |
| static PPCAMode* iselWordExpr_AMode_wrk ( ISelEnv* env, IRExpr* e, IRType xferTy ); |
| static PPCAMode* iselWordExpr_AMode ( ISelEnv* env, IRExpr* e, IRType xferTy ); |
| |
| static void iselInt128Expr_to_32x4_wrk ( HReg* rHi, HReg* rMedHi, |
| HReg* rMedLo, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| static void iselInt128Expr_to_32x4 ( HReg* rHi, HReg* rMedHi, |
| HReg* rMedLo, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| |
| |
| /* 32-bit mode ONLY: compute an I64 into a GPR pair. */ |
| static void iselInt64Expr_wrk ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| static void iselInt64Expr ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| |
| /* 64-bit mode ONLY: compute an I128 into a GPR64 pair. */ |
| static void iselInt128Expr_wrk ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| static void iselInt128Expr ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ); |
| |
| static PPCCondCode iselCondCode_wrk ( ISelEnv* env, IRExpr* e ); |
| static PPCCondCode iselCondCode ( ISelEnv* env, IRExpr* e ); |
| |
| static HReg iselDblExpr_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselDblExpr ( ISelEnv* env, IRExpr* e ); |
| |
| static HReg iselFltExpr_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselFltExpr ( ISelEnv* env, IRExpr* e ); |
| |
| static HReg iselVecExpr_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselVecExpr ( ISelEnv* env, IRExpr* e ); |
| |
| /* 64-bit mode ONLY. */ |
| static HReg iselDfp32Expr_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselDfp32Expr ( ISelEnv* env, IRExpr* e ); |
| static HReg iselDfp64Expr_wrk ( ISelEnv* env, IRExpr* e ); |
| static HReg iselDfp64Expr ( ISelEnv* env, IRExpr* e ); |
| |
| /* 64-bit mode ONLY: compute an D128 into a GPR64 pair. */ |
| static void iselDfp128Expr_wrk ( HReg* rHi, HReg* rLo, ISelEnv* env, |
| IRExpr* e ); |
| static void iselDfp128Expr ( HReg* rHi, HReg* rLo, ISelEnv* env, |
| IRExpr* e ); |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Misc helpers ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Make an int reg-reg move. */ |
| |
| static PPCInstr* mk_iMOVds_RR ( HReg r_dst, HReg r_src ) |
| { |
| vassert(hregClass(r_dst) == hregClass(r_src)); |
| vassert(hregClass(r_src) == HRcInt32 || |
| hregClass(r_src) == HRcInt64); |
| return PPCInstr_Alu(Palu_OR, r_dst, r_src, PPCRH_Reg(r_src)); |
| } |
| |
| /* Advance/retreat %r1 by n. */ |
| |
| static void add_to_sp ( ISelEnv* env, UInt n ) |
| { |
| HReg sp = StackFramePtr(env->mode64); |
| vassert(n <= 1024 && (n%16) == 0); |
| addInstr(env, PPCInstr_Alu( Palu_ADD, sp, sp, |
| PPCRH_Imm(True,toUShort(n)) )); |
| } |
| |
| static void sub_from_sp ( ISelEnv* env, UInt n ) |
| { |
| HReg sp = StackFramePtr(env->mode64); |
| vassert(n <= 1024 && (n%16) == 0); |
| addInstr(env, PPCInstr_Alu( Palu_SUB, sp, sp, |
| PPCRH_Imm(True,toUShort(n)) )); |
| } |
| |
| /* |
| returns a quadword aligned address on the stack |
| - copies SP, adds 16bytes, aligns to quadword. |
| use sub_from_sp(32) before calling this, |
| as expects to have 32 bytes to play with. |
| */ |
| static HReg get_sp_aligned16 ( ISelEnv* env ) |
| { |
| HReg r = newVRegI(env); |
| HReg align16 = newVRegI(env); |
| addInstr(env, mk_iMOVds_RR(r, StackFramePtr(env->mode64))); |
| // add 16 |
| addInstr(env, PPCInstr_Alu( Palu_ADD, r, r, |
| PPCRH_Imm(True,toUShort(16)) )); |
| // mask to quadword |
| addInstr(env, |
| PPCInstr_LI(align16, 0xFFFFFFFFFFFFFFF0ULL, env->mode64)); |
| addInstr(env, PPCInstr_Alu(Palu_AND, r,r, PPCRH_Reg(align16))); |
| return r; |
| } |
| |
| |
| |
| /* Load 2*I32 regs to fp reg */ |
| static HReg mk_LoadRR32toFPR ( ISelEnv* env, |
| HReg r_srcHi, HReg r_srcLo ) |
| { |
| HReg fr_dst = newVRegF(env); |
| PPCAMode *am_addr0, *am_addr1; |
| |
| vassert(!env->mode64); |
| vassert(hregClass(r_srcHi) == HRcInt32); |
| vassert(hregClass(r_srcLo) == HRcInt32); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr0 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| am_addr1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| // store hi,lo as Ity_I32's |
| addInstr(env, PPCInstr_Store( 4, am_addr0, r_srcHi, env->mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_addr1, r_srcLo, env->mode64 )); |
| |
| // load as float |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_dst, am_addr0)); |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return fr_dst; |
| } |
| |
| /* Load I64 reg to fp reg */ |
| static HReg mk_LoadR64toFPR ( ISelEnv* env, HReg r_src ) |
| { |
| HReg fr_dst = newVRegF(env); |
| PPCAMode *am_addr0; |
| |
| vassert(env->mode64); |
| vassert(hregClass(r_src) == HRcInt64); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr0 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| // store as Ity_I64 |
| addInstr(env, PPCInstr_Store( 8, am_addr0, r_src, env->mode64 )); |
| |
| // load as float |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_dst, am_addr0)); |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return fr_dst; |
| } |
| |
| |
| /* Given an amode, return one which references 4 bytes further |
| along. */ |
| |
| static PPCAMode* advance4 ( ISelEnv* env, PPCAMode* am ) |
| { |
| PPCAMode* am4 = dopyPPCAMode( am ); |
| if (am4->tag == Pam_IR |
| && am4->Pam.IR.index + 4 <= 32767) { |
| am4->Pam.IR.index += 4; |
| } else { |
| vpanic("advance4(ppc,host)"); |
| } |
| return am4; |
| } |
| |
| |
| /* Given a guest-state array descriptor, an index expression and a |
| bias, generate a PPCAMode pointing at the relevant piece of |
| guest state. */ |
| static |
| PPCAMode* genGuestArrayOffset ( ISelEnv* env, IRRegArray* descr, |
| IRExpr* off, Int bias ) |
| { |
| HReg rtmp, roff; |
| Int elemSz = sizeofIRType(descr->elemTy); |
| Int nElems = descr->nElems; |
| Int shift = 0; |
| |
| /* Throw out any cases we don't need. In theory there might be a |
| day where we need to handle others, but not today. */ |
| |
| if (nElems != 16 && nElems != 32) |
| vpanic("genGuestArrayOffset(ppc host)(1)"); |
| |
| switch (elemSz) { |
| case 4: shift = 2; break; |
| case 8: shift = 3; break; |
| default: vpanic("genGuestArrayOffset(ppc host)(2)"); |
| } |
| |
| if (bias < -100 || bias > 100) /* somewhat arbitrarily */ |
| vpanic("genGuestArrayOffset(ppc host)(3)"); |
| if (descr->base < 0 || descr->base > 5000) /* somewhat arbitrarily */ |
| vpanic("genGuestArrayOffset(ppc host)(4)"); |
| |
| /* Compute off into a reg, %off. Then return: |
| |
| addi %tmp, %off, bias (if bias != 0) |
| andi %tmp, nElems-1 |
| sldi %tmp, shift |
| addi %tmp, %tmp, base |
| ... Baseblockptr + %tmp ... |
| */ |
| roff = iselWordExpr_R(env, off); |
| rtmp = newVRegI(env); |
| addInstr(env, PPCInstr_Alu( |
| Palu_ADD, |
| rtmp, roff, |
| PPCRH_Imm(True/*signed*/, toUShort(bias)))); |
| addInstr(env, PPCInstr_Alu( |
| Palu_AND, |
| rtmp, rtmp, |
| PPCRH_Imm(False/*unsigned*/, toUShort(nElems-1)))); |
| addInstr(env, PPCInstr_Shft( |
| Pshft_SHL, |
| env->mode64 ? False : True/*F:64-bit, T:32-bit shift*/, |
| rtmp, rtmp, |
| PPCRH_Imm(False/*unsigned*/, toUShort(shift)))); |
| addInstr(env, PPCInstr_Alu( |
| Palu_ADD, |
| rtmp, rtmp, |
| PPCRH_Imm(True/*signed*/, toUShort(descr->base)))); |
| return |
| PPCAMode_RR( GuestStatePtr(env->mode64), rtmp ); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Function call helpers ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Used only in doHelperCall. See big comment in doHelperCall re |
| handling of register-parameter args. This function figures out |
| whether evaluation of an expression might require use of a fixed |
| register. If in doubt return True (safe but suboptimal). |
| */ |
| static |
| Bool mightRequireFixedRegs ( IRExpr* e ) |
| { |
| switch (e->tag) { |
| case Iex_RdTmp: case Iex_Const: case Iex_Get: |
| return False; |
| default: |
| return True; |
| } |
| } |
| |
| |
| /* Do a complete function call. |guard| is a Ity_Bit expression |
| indicating whether or not the call happens. If guard==NULL, the |
| call is unconditional. |retloc| is set to indicate where the |
| return value is after the call. The caller (of this fn) must |
| generate code to add |stackAdjustAfterCall| to the stack pointer |
| after the call is done. */ |
| |
| static |
| void doHelperCall ( /*OUT*/UInt* stackAdjustAfterCall, |
| /*OUT*/RetLoc* retloc, |
| ISelEnv* env, |
| IRExpr* guard, |
| IRCallee* cee, IRType retTy, IRExpr** args ) |
| { |
| PPCCondCode cc; |
| HReg argregs[PPC_N_REGPARMS]; |
| HReg tmpregs[PPC_N_REGPARMS]; |
| Bool go_fast; |
| Int n_args, i, argreg; |
| UInt argiregs; |
| Bool mode64 = env->mode64; |
| |
| /* Set default returns. We'll update them later if needed. */ |
| *stackAdjustAfterCall = 0; |
| *retloc = mk_RetLoc_INVALID(); |
| |
| /* These are used for cross-checking that IR-level constraints on |
| the use of IRExpr_VECRET() and IRExpr_BBPTR() are observed. */ |
| UInt nVECRETs = 0; |
| UInt nBBPTRs = 0; |
| |
| /* Do we need to force use of an odd-even reg pair for 64-bit args? |
| JRS 31-07-2013: is this still relevant, now that we are not |
| generating code for 32-bit AIX ? */ |
| Bool regalign_int64s |
| = (!mode64) && env->vbi->host_ppc32_regalign_int64_args; |
| |
| /* Marshal args for a call and do the call. |
| |
| This function only deals with a tiny set of possibilities, which |
| cover all helpers in practice. The restrictions are that only |
| arguments in registers are supported, hence only PPC_N_REGPARMS x |
| (mode32:32 | mode64:64) integer bits in total can be passed. |
| In fact the only supported arg type is (mode32:I32 | mode64:I64). |
| |
| The return type can be I{64,32,16,8} or V{128,256}. In the |
| latter two cases, it is expected that |args| will contain the |
| special node IRExpr_VECRET(), in which case this routine |
| generates code to allocate space on the stack for the vector |
| return value. Since we are not passing any scalars on the |
| stack, it is enough to preallocate the return space before |
| marshalling any arguments, in this case. |
| |
| |args| may also contain IRExpr_BBPTR(), in which case the value |
| in the guest state pointer register is passed as the |
| corresponding argument. |
| |
| Generating code which is both efficient and correct when |
| parameters are to be passed in registers is difficult, for the |
| reasons elaborated in detail in comments attached to |
| doHelperCall() in priv/host-x86/isel.c. Here, we use a variant |
| of the method described in those comments. |
| |
| The problem is split into two cases: the fast scheme and the |
| slow scheme. In the fast scheme, arguments are computed |
| directly into the target (real) registers. This is only safe |
| when we can be sure that computation of each argument will not |
| trash any real registers set by computation of any other |
| argument. |
| |
| In the slow scheme, all args are first computed into vregs, and |
| once they are all done, they are moved to the relevant real |
| regs. This always gives correct code, but it also gives a bunch |
| of vreg-to-rreg moves which are usually redundant but are hard |
| for the register allocator to get rid of. |
| |
| To decide which scheme to use, all argument expressions are |
| first examined. If they are all so simple that it is clear they |
| will be evaluated without use of any fixed registers, use the |
| fast scheme, else use the slow scheme. Note also that only |
| unconditional calls may use the fast scheme, since having to |
| compute a condition expression could itself trash real |
| registers. |
| |
| Note this requires being able to examine an expression and |
| determine whether or not evaluation of it might use a fixed |
| register. That requires knowledge of how the rest of this insn |
| selector works. Currently just the following 3 are regarded as |
| safe -- hopefully they cover the majority of arguments in |
| practice: IRExpr_Tmp IRExpr_Const IRExpr_Get. |
| */ |
| |
| /* Note that the cee->regparms field is meaningless on PPC32/64 host |
| (since there is only one calling convention) and so we always |
| ignore it. */ |
| |
| n_args = 0; |
| for (i = 0; args[i]; i++) |
| n_args++; |
| |
| if (n_args > PPC_N_REGPARMS) { |
| vpanic("doHelperCall(PPC): cannot currently handle > 8 args"); |
| // PPC_N_REGPARMS |
| } |
| |
| /* This is kind of stupid .. the arrays are sized as PPC_N_REGPARMS |
| but we then assume that that value is 8. */ |
| vassert(PPC_N_REGPARMS == 8); |
| |
| argregs[0] = hregPPC_GPR3(mode64); |
| argregs[1] = hregPPC_GPR4(mode64); |
| argregs[2] = hregPPC_GPR5(mode64); |
| argregs[3] = hregPPC_GPR6(mode64); |
| argregs[4] = hregPPC_GPR7(mode64); |
| argregs[5] = hregPPC_GPR8(mode64); |
| argregs[6] = hregPPC_GPR9(mode64); |
| argregs[7] = hregPPC_GPR10(mode64); |
| argiregs = 0; |
| |
| tmpregs[0] = tmpregs[1] = tmpregs[2] = |
| tmpregs[3] = tmpregs[4] = tmpregs[5] = |
| tmpregs[6] = tmpregs[7] = INVALID_HREG; |
| |
| /* First decide which scheme (slow or fast) is to be used. First |
| assume the fast scheme, and select slow if any contraindications |
| (wow) appear. */ |
| |
| go_fast = True; |
| |
| /* We'll need space on the stack for the return value. Avoid |
| possible complications with nested calls by using the slow |
| scheme. */ |
| if (retTy == Ity_V128 || retTy == Ity_V256) |
| go_fast = False; |
| |
| if (go_fast && guard) { |
| if (guard->tag == Iex_Const |
| && guard->Iex.Const.con->tag == Ico_U1 |
| && guard->Iex.Const.con->Ico.U1 == True) { |
| /* unconditional */ |
| } else { |
| /* Not manifestly unconditional -- be conservative. */ |
| go_fast = False; |
| } |
| } |
| |
| if (go_fast) { |
| for (i = 0; i < n_args; i++) { |
| IRExpr* arg = args[i]; |
| if (UNLIKELY(arg->tag == Iex_BBPTR)) { |
| /* that's OK */ |
| } |
| else if (UNLIKELY(arg->tag == Iex_VECRET)) { |
| /* This implies ill-formed IR, since if the IR was |
| well-formed, the return-type test above would have |
| filtered it out. */ |
| vpanic("doHelperCall(PPC): invalid IR"); |
| } |
| else if (mightRequireFixedRegs(arg)) { |
| go_fast = False; |
| break; |
| } |
| } |
| } |
| |
| /* At this point the scheme to use has been established. Generate |
| code to get the arg values into the argument rregs. */ |
| |
| if (go_fast) { |
| |
| /* FAST SCHEME */ |
| argreg = 0; |
| |
| for (i = 0; i < n_args; i++) { |
| IRExpr* arg = args[i]; |
| vassert(argreg < PPC_N_REGPARMS); |
| |
| if (arg->tag == Iex_BBPTR) { |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg], |
| GuestStatePtr(mode64) )); |
| argreg++; |
| } else { |
| vassert(arg->tag != Iex_VECRET); |
| IRType ty = typeOfIRExpr(env->type_env, arg); |
| vassert(ty == Ity_I32 || ty == Ity_I64); |
| if (!mode64) { |
| if (ty == Ity_I32) { |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, |
| mk_iMOVds_RR( argregs[argreg], |
| iselWordExpr_R(env, arg) )); |
| } else { // Ity_I64 in 32-bit mode |
| HReg rHi, rLo; |
| if (regalign_int64s && (argreg%2) == 1) |
| // ppc32 ELF abi spec for passing LONG_LONG |
| argreg++; // XXX: odd argreg => even rN |
| vassert(argreg < PPC_N_REGPARMS-1); |
| iselInt64Expr(&rHi,&rLo, env, arg); |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg++], rHi )); |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg], rLo)); |
| } |
| } else { // mode64 |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg], |
| iselWordExpr_R(env, arg) )); |
| } |
| argreg++; |
| } /* if (arg == IRExprP__BBPR) */ |
| } |
| |
| /* Fast scheme only applies for unconditional calls. Hence: */ |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| |
| } else { |
| |
| /* SLOW SCHEME; move via temporaries */ |
| argreg = 0; |
| |
| /* If we have a vector return type, allocate a place for it on |
| the stack and record its address. Rather than figure out the |
| complexities of PPC{32,64} ELF ABI stack frame layout, simply |
| drop the SP by 1024 and allocate the return point in the |
| middle. I think this should comfortably clear any ABI |
| mandated register save areas. Note that it doesn't maintain |
| the backchain as it should, since we're not doing st{d,w}u to |
| adjust the SP, but .. that doesn't seem to be a big deal. |
| Since we're not expecting to have to unwind out of here. */ |
| HReg r_vecRetAddr = INVALID_HREG; |
| if (retTy == Ity_V128) { |
| r_vecRetAddr = newVRegI(env); |
| sub_from_sp(env, 512); |
| addInstr(env, mk_iMOVds_RR( r_vecRetAddr, StackFramePtr(mode64) )); |
| sub_from_sp(env, 512); |
| } |
| else if (retTy == Ity_V256) { |
| vassert(0); //ATC |
| r_vecRetAddr = newVRegI(env); |
| sub_from_sp(env, 512); |
| addInstr(env, mk_iMOVds_RR( r_vecRetAddr, StackFramePtr(mode64) )); |
| sub_from_sp(env, 512); |
| } |
| |
| vassert(n_args >= 0 && n_args <= 8); |
| for (i = 0; i < n_args; i++) { |
| IRExpr* arg = args[i]; |
| vassert(argreg < PPC_N_REGPARMS); |
| if (UNLIKELY(arg->tag == Iex_BBPTR)) { |
| tmpregs[argreg] = newVRegI(env); |
| addInstr(env, mk_iMOVds_RR( tmpregs[argreg], |
| GuestStatePtr(mode64) )); |
| nBBPTRs++; |
| } |
| else if (UNLIKELY(arg->tag == Iex_VECRET)) { |
| /* We stashed the address of the return slot earlier, so just |
| retrieve it now. */ |
| vassert(!hregIsInvalid(r_vecRetAddr)); |
| tmpregs[i] = r_vecRetAddr; |
| nVECRETs++; |
| } |
| else { |
| IRType ty = typeOfIRExpr(env->type_env, arg); |
| vassert(ty == Ity_I32 || ty == Ity_I64); |
| if (!mode64) { |
| if (ty == Ity_I32) { |
| tmpregs[argreg] = iselWordExpr_R(env, arg); |
| } else { // Ity_I64 in 32-bit mode |
| HReg rHi, rLo; |
| if (regalign_int64s && (argreg%2) == 1) |
| // ppc32 ELF abi spec for passing LONG_LONG |
| argreg++; // XXX: odd argreg => even rN |
| vassert(argreg < PPC_N_REGPARMS-1); |
| iselInt64Expr(&rHi,&rLo, env, arg); |
| tmpregs[argreg++] = rHi; |
| tmpregs[argreg] = rLo; |
| } |
| } else { // mode64 |
| tmpregs[argreg] = iselWordExpr_R(env, arg); |
| } |
| } |
| argreg++; |
| } |
| |
| /* Now we can compute the condition. We can't do it earlier |
| because the argument computations could trash the condition |
| codes. Be a bit clever to handle the common case where the |
| guard is 1:Bit. */ |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| if (guard) { |
| if (guard->tag == Iex_Const |
| && guard->Iex.Const.con->tag == Ico_U1 |
| && guard->Iex.Const.con->Ico.U1 == True) { |
| /* unconditional -- do nothing */ |
| } else { |
| cc = iselCondCode( env, guard ); |
| } |
| } |
| |
| /* Move the args to their final destinations. */ |
| for (i = 0; i < argreg; i++) { |
| if (hregIsInvalid(tmpregs[i])) // Skip invalid regs |
| continue; |
| /* None of these insns, including any spill code that might |
| be generated, may alter the condition codes. */ |
| argiregs |= (1 << (i+3)); |
| addInstr( env, mk_iMOVds_RR( argregs[i], tmpregs[i] ) ); |
| } |
| |
| } |
| |
| /* Do final checks, set the return values, and generate the call |
| instruction proper. */ |
| if (retTy == Ity_V128 || retTy == Ity_V256) { |
| vassert(nVECRETs == 1); |
| } else { |
| vassert(nVECRETs == 0); |
| } |
| |
| vassert(nBBPTRs == 0 || nBBPTRs == 1); |
| |
| vassert(*stackAdjustAfterCall == 0); |
| vassert(is_RetLoc_INVALID(*retloc)); |
| switch (retTy) { |
| case Ity_INVALID: |
| /* Function doesn't return a value. */ |
| *retloc = mk_RetLoc_simple(RLPri_None); |
| break; |
| case Ity_I64: |
| *retloc = mk_RetLoc_simple(mode64 ? RLPri_Int : RLPri_2Int); |
| break; |
| case Ity_I32: case Ity_I16: case Ity_I8: |
| *retloc = mk_RetLoc_simple(RLPri_Int); |
| break; |
| case Ity_V128: |
| /* Result is 512 bytes up the stack, and after it has been |
| retrieved, adjust SP upwards by 1024. */ |
| *retloc = mk_RetLoc_spRel(RLPri_V128SpRel, 512); |
| *stackAdjustAfterCall = 1024; |
| break; |
| case Ity_V256: |
| vassert(0); // ATC |
| /* Ditto */ |
| *retloc = mk_RetLoc_spRel(RLPri_V256SpRel, 512); |
| *stackAdjustAfterCall = 1024; |
| break; |
| default: |
| /* IR can denote other possible return types, but we don't |
| handle those here. */ |
| vassert(0); |
| } |
| |
| /* Finally, generate the call itself. This needs the *retloc value |
| set in the switch above, which is why it's at the end. */ |
| |
| ULong target = mode64 ? Ptr_to_ULong(cee->addr) |
| : toUInt(Ptr_to_ULong(cee->addr)); |
| addInstr(env, PPCInstr_Call( cc, (Addr64)target, argiregs, *retloc )); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: FP rounding mode helpers ---*/ |
| /*---------------------------------------------------------*/ |
| |
| ///* Set FPU's rounding mode to the default */ |
| //static |
| //void set_FPU_rounding_default ( ISelEnv* env ) |
| //{ |
| // HReg fr_src = newVRegF(env); |
| // HReg r_src = newVRegI(env); |
| // |
| // /* Default rounding mode = 0x0 |
| // Only supporting the rounding-mode bits - the rest of FPSCR is 0x0 |
| // - so we can set the whole register at once (faster) |
| // note: upper 32 bits ignored by FpLdFPSCR |
| // */ |
| // addInstr(env, PPCInstr_LI(r_src, 0x0, env->mode64)); |
| // if (env->mode64) { |
| // fr_src = mk_LoadR64toFPR( env, r_src ); // 1*I64 -> F64 |
| // } else { |
| // fr_src = mk_LoadRR32toFPR( env, r_src, r_src ); // 2*I32 -> F64 |
| // } |
| // addInstr(env, PPCInstr_FpLdFPSCR( fr_src )); |
| //} |
| |
| /* Convert IR rounding mode to PPC encoding */ |
| static HReg roundModeIRtoPPC ( ISelEnv* env, HReg r_rmIR ) |
| { |
| /* |
| rounding mode | PPC | IR |
| ----------------------------------------------- |
| to nearest, ties to even | 000 | 000 |
| to zero | 001 | 011 |
| to +infinity | 010 | 010 |
| to -infinity | 011 | 001 |
| +++++ Below are the extended rounding modes for decimal floating point +++++ |
| to nearest, ties away from 0 | 100 | 100 |
| to nearest, ties toward 0 | 101 | 111 |
| to away from 0 | 110 | 110 |
| to prepare for shorter precision | 111 | 101 |
| */ |
| HReg r_rmPPC = newVRegI(env); |
| HReg r_tmp1 = newVRegI(env); |
| HReg r_tmp2 = newVRegI(env); |
| |
| vassert(hregClass(r_rmIR) == HRcGPR(env->mode64)); |
| |
| // r_rmPPC = XOR(r_rmIR, r_rmIR << 1) & 3 |
| // |
| // slwi tmp1, r_rmIR, 1 |
| // xor tmp1, r_rmIR, tmp1 |
| // andi r_rmPPC, tmp1, 3 |
| |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| r_tmp1, r_rmIR, PPCRH_Imm(False,1))); |
| |
| addInstr( env, PPCInstr_Alu( Palu_AND, |
| r_tmp2, r_tmp1, PPCRH_Imm( False, 3 ) ) ); |
| |
| addInstr( env, PPCInstr_Alu( Palu_XOR, |
| r_rmPPC, r_rmIR, PPCRH_Reg( r_tmp2 ) ) ); |
| |
| return r_rmPPC; |
| } |
| |
| |
| /* Set the FPU's rounding mode: 'mode' is an I32-typed expression |
| denoting a value in the range 0 .. 7, indicating a round mode |
| encoded as per type IRRoundingMode. Set the PPC FPSCR to have the |
| same rounding. When the dfp_rm arg is True, set the decimal |
| floating point rounding mode bits (29:31); otherwise, set the |
| binary floating point rounding mode bits (62:63). |
| |
| For speed & simplicity, we're setting the *entire* FPSCR here. |
| |
| Setting the rounding mode is expensive. So this function tries to |
| avoid repeatedly setting the rounding mode to the same thing by |
| first comparing 'mode' to the 'mode' tree supplied in the previous |
| call to this function, if any. (The previous value is stored in |
| env->previous_rm.) If 'mode' is a single IR temporary 't' and |
| env->previous_rm is also just 't', then the setting is skipped. |
| |
| This is safe because of the SSA property of IR: an IR temporary can |
| only be defined once and so will have the same value regardless of |
| where it appears in the block. Cool stuff, SSA. |
| |
| A safety condition: all attempts to set the RM must be aware of |
| this mechanism - by being routed through the functions here. |
| |
| Of course this only helps if blocks where the RM is set more than |
| once and it is set to the same value each time, *and* that value is |
| held in the same IR temporary each time. In order to assure the |
| latter as much as possible, the IR optimiser takes care to do CSE |
| on any block with any sign of floating point activity. |
| */ |
| static |
| void _set_FPU_rounding_mode ( ISelEnv* env, IRExpr* mode, Bool dfp_rm ) |
| { |
| HReg fr_src = newVRegF(env); |
| HReg r_src; |
| |
| vassert(typeOfIRExpr(env->type_env,mode) == Ity_I32); |
| |
| /* Do we need to do anything? */ |
| if (env->previous_rm |
| && env->previous_rm->tag == Iex_RdTmp |
| && mode->tag == Iex_RdTmp |
| && env->previous_rm->Iex.RdTmp.tmp == mode->Iex.RdTmp.tmp) { |
| /* no - setting it to what it was before. */ |
| vassert(typeOfIRExpr(env->type_env, env->previous_rm) == Ity_I32); |
| return; |
| } |
| |
| /* No luck - we better set it, and remember what we set it to. */ |
| env->previous_rm = mode; |
| |
| /* Only supporting the rounding-mode bits - the rest of FPSCR is |
| 0x0 - so we can set the whole register at once (faster). */ |
| |
| // Resolve rounding mode and convert to PPC representation |
| r_src = roundModeIRtoPPC( env, iselWordExpr_R(env, mode) ); |
| |
| // gpr -> fpr |
| if (env->mode64) { |
| if (dfp_rm) { |
| HReg r_tmp1 = newVRegI( env ); |
| addInstr( env, |
| PPCInstr_Shft( Pshft_SHL, False/*64bit shift*/, |
| r_tmp1, r_src, PPCRH_Imm( False, 32 ) ) ); |
| fr_src = mk_LoadR64toFPR( env, r_tmp1 ); |
| } else { |
| fr_src = mk_LoadR64toFPR( env, r_src ); // 1*I64 -> F64 |
| } |
| } else { |
| if (dfp_rm) { |
| HReg r_zero = newVRegI( env ); |
| addInstr( env, PPCInstr_LI( r_zero, 0, env->mode64 ) ); |
| fr_src = mk_LoadRR32toFPR( env, r_src, r_zero ); |
| } else { |
| fr_src = mk_LoadRR32toFPR( env, r_src, r_src ); // 2*I32 -> F64 |
| } |
| } |
| |
| // Move to FPSCR |
| addInstr(env, PPCInstr_FpLdFPSCR( fr_src, dfp_rm )); |
| } |
| |
| static void set_FPU_rounding_mode ( ISelEnv* env, IRExpr* mode ) |
| { |
| _set_FPU_rounding_mode(env, mode, False); |
| } |
| |
| static void set_FPU_DFP_rounding_mode ( ISelEnv* env, IRExpr* mode ) |
| { |
| _set_FPU_rounding_mode(env, mode, True); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: vector helpers ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Generate all-zeroes into a new vector register. |
| */ |
| static HReg generate_zeroes_V128 ( ISelEnv* env ) |
| { |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(Pav_XOR, dst, dst, dst)); |
| return dst; |
| } |
| |
| /* Generate all-ones into a new vector register. |
| */ |
| static HReg generate_ones_V128 ( ISelEnv* env ) |
| { |
| HReg dst = newVRegV(env); |
| PPCVI5s * src = PPCVI5s_Imm(-1); |
| addInstr(env, PPCInstr_AvSplat(8, dst, src)); |
| return dst; |
| } |
| |
| |
| /* |
| Generates code for AvSplat |
| - takes in IRExpr* of type 8|16|32 |
| returns vector reg of duplicated lanes of input |
| - uses AvSplat(imm) for imms up to simm6. |
| otherwise must use store reg & load vector |
| */ |
| static HReg mk_AvDuplicateRI( ISelEnv* env, IRExpr* e ) |
| { |
| HReg r_src; |
| HReg dst = newVRegV(env); |
| PPCRI* ri = iselWordExpr_RI(env, e); |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| UInt sz = (ty == Ity_I8) ? 8 : (ty == Ity_I16) ? 16 : 32; |
| vassert(ty == Ity_I8 || ty == Ity_I16 || ty == Ity_I32); |
| |
| /* special case: immediate */ |
| if (ri->tag == Pri_Imm) { |
| Int simm32 = (Int)ri->Pri.Imm; |
| |
| /* figure out if it's do-able with imm splats. */ |
| if (simm32 >= -32 && simm32 <= 31) { |
| Char simm6 = (Char)simm32; |
| if (simm6 > 15) { /* 16:31 inclusive */ |
| HReg v1 = newVRegV(env); |
| HReg v2 = newVRegV(env); |
| addInstr(env, PPCInstr_AvSplat(sz, v1, PPCVI5s_Imm(-16))); |
| addInstr(env, PPCInstr_AvSplat(sz, v2, PPCVI5s_Imm(simm6-16))); |
| addInstr(env, |
| (sz== 8) ? PPCInstr_AvBin8x16(Pav_SUBU, dst, v2, v1) : |
| (sz==16) ? PPCInstr_AvBin16x8(Pav_SUBU, dst, v2, v1) |
| : PPCInstr_AvBin32x4(Pav_SUBU, dst, v2, v1) ); |
| return dst; |
| } |
| if (simm6 < -16) { /* -32:-17 inclusive */ |
| HReg v1 = newVRegV(env); |
| HReg v2 = newVRegV(env); |
| addInstr(env, PPCInstr_AvSplat(sz, v1, PPCVI5s_Imm(-16))); |
| addInstr(env, PPCInstr_AvSplat(sz, v2, PPCVI5s_Imm(simm6+16))); |
| addInstr(env, |
| (sz== 8) ? PPCInstr_AvBin8x16(Pav_ADDU, dst, v2, v1) : |
| (sz==16) ? PPCInstr_AvBin16x8(Pav_ADDU, dst, v2, v1) |
| : PPCInstr_AvBin32x4(Pav_ADDU, dst, v2, v1) ); |
| return dst; |
| } |
| /* simplest form: -16:15 inclusive */ |
| addInstr(env, PPCInstr_AvSplat(sz, dst, PPCVI5s_Imm(simm6))); |
| return dst; |
| } |
| |
| /* no luck; use the Slow way. */ |
| r_src = newVRegI(env); |
| addInstr(env, PPCInstr_LI(r_src, (Long)simm32, env->mode64)); |
| } |
| else { |
| r_src = ri->Pri.Reg; |
| } |
| |
| /* default case: store r_src in lowest lane of 16-aligned mem, |
| load vector, splat lowest lane to dst */ |
| { |
| /* CAB: Maybe faster to store r_src multiple times (sz dependent), |
| and simply load the vector? */ |
| HReg r_aligned16; |
| HReg v_src = newVRegV(env); |
| PPCAMode *am_off12; |
| |
| sub_from_sp( env, 32 ); // Move SP down |
| /* Get a 16-aligned address within our stack space */ |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off12 = PPCAMode_IR( 12, r_aligned16 ); |
| |
| /* Store r_src in low word of 16-aligned mem */ |
| addInstr(env, PPCInstr_Store( 4, am_off12, r_src, env->mode64 )); |
| |
| /* Load src to vector[low lane] */ |
| addInstr(env, PPCInstr_AvLdSt( True/*ld*/, 4, v_src, am_off12 ) ); |
| add_to_sp( env, 32 ); // Reset SP |
| |
| /* Finally, splat v_src[low_lane] to dst */ |
| addInstr(env, PPCInstr_AvSplat(sz, dst, PPCVI5s_Reg(v_src))); |
| return dst; |
| } |
| } |
| |
| |
| /* for each lane of vSrc: lane == nan ? laneX = all 1's : all 0's */ |
| static HReg isNan ( ISelEnv* env, HReg vSrc ) |
| { |
| HReg zeros, msk_exp, msk_mnt, expt, mnts, vIsNan; |
| |
| vassert(hregClass(vSrc) == HRcVec128); |
| |
| zeros = mk_AvDuplicateRI(env, mkU32(0)); |
| msk_exp = mk_AvDuplicateRI(env, mkU32(0x7F800000)); |
| msk_mnt = mk_AvDuplicateRI(env, mkU32(0x7FFFFF)); |
| expt = newVRegV(env); |
| mnts = newVRegV(env); |
| vIsNan = newVRegV(env); |
| |
| /* 32bit float => sign(1) | exponent(8) | mantissa(23) |
| nan => exponent all ones, mantissa > 0 */ |
| |
| addInstr(env, PPCInstr_AvBinary(Pav_AND, expt, vSrc, msk_exp)); |
| addInstr(env, PPCInstr_AvBin32x4(Pav_CMPEQU, expt, expt, msk_exp)); |
| addInstr(env, PPCInstr_AvBinary(Pav_AND, mnts, vSrc, msk_mnt)); |
| addInstr(env, PPCInstr_AvBin32x4(Pav_CMPGTU, mnts, mnts, zeros)); |
| addInstr(env, PPCInstr_AvBinary(Pav_AND, vIsNan, expt, mnts)); |
| return vIsNan; |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Integer expressions (64/32/16/8 bit) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Select insns for an integer-typed expression, and add them to the |
| code list. Return a reg holding the result. This reg will be a |
| virtual register. THE RETURNED REG MUST NOT BE MODIFIED. If you |
| want to modify it, ask for a new vreg, copy it in there, and modify |
| the copy. The register allocator will do its best to map both |
| vregs to the same real register, so the copies will often disappear |
| later in the game. |
| |
| This should handle expressions of 64, 32, 16 and 8-bit type. |
| All results are returned in a (mode64 ? 64bit : 32bit) register. |
| For 16- and 8-bit expressions, the upper (32/48/56 : 16/24) bits |
| are arbitrary, so you should mask or sign extend partial values |
| if necessary. |
| */ |
| |
| static HReg iselWordExpr_R ( ISelEnv* env, IRExpr* e ) |
| { |
| HReg r = iselWordExpr_R_wrk(env, e); |
| /* sanity checks ... */ |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| |
| vassert(hregClass(r) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(r)); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static HReg iselWordExpr_R_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| Bool mode64 = env->mode64; |
| MatchInfo mi; |
| DECLARE_PATTERN(p_32to1_then_1Uto8); |
| |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && mode64)); |
| |
| switch (e->tag) { |
| |
| /* --------- TEMP --------- */ |
| case Iex_RdTmp: |
| return lookupIRTemp(env, e->Iex.RdTmp.tmp); |
| |
| /* --------- LOAD --------- */ |
| case Iex_Load: { |
| HReg r_dst; |
| PPCAMode* am_addr; |
| if (e->Iex.Load.end != Iend_BE) |
| goto irreducible; |
| r_dst = newVRegI(env); |
| am_addr = iselWordExpr_AMode( env, e->Iex.Load.addr, ty/*of xfer*/ ); |
| addInstr(env, PPCInstr_Load( toUChar(sizeofIRType(ty)), |
| r_dst, am_addr, mode64 )); |
| return r_dst; |
| /*NOTREACHED*/ |
| } |
| |
| /* --------- BINARY OP --------- */ |
| case Iex_Binop: { |
| PPCAluOp aluOp; |
| PPCShftOp shftOp; |
| |
| /* Is it an addition or logical style op? */ |
| switch (e->Iex.Binop.op) { |
| case Iop_Add8: case Iop_Add16: case Iop_Add32: case Iop_Add64: |
| aluOp = Palu_ADD; break; |
| case Iop_Sub8: case Iop_Sub16: case Iop_Sub32: case Iop_Sub64: |
| aluOp = Palu_SUB; break; |
| case Iop_And8: case Iop_And16: case Iop_And32: case Iop_And64: |
| aluOp = Palu_AND; break; |
| case Iop_Or8: case Iop_Or16: case Iop_Or32: case Iop_Or64: |
| aluOp = Palu_OR; break; |
| case Iop_Xor8: case Iop_Xor16: case Iop_Xor32: case Iop_Xor64: |
| aluOp = Palu_XOR; break; |
| default: |
| aluOp = Palu_INVALID; break; |
| } |
| /* For commutative ops we assume any literal |
| values are on the second operand. */ |
| if (aluOp != Palu_INVALID) { |
| HReg r_dst = newVRegI(env); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* ri_srcR = NULL; |
| /* get right arg into an RH, in the appropriate way */ |
| switch (aluOp) { |
| case Palu_ADD: case Palu_SUB: |
| ri_srcR = iselWordExpr_RH(env, True/*signed*/, |
| e->Iex.Binop.arg2); |
| break; |
| case Palu_AND: case Palu_OR: case Palu_XOR: |
| ri_srcR = iselWordExpr_RH(env, False/*signed*/, |
| e->Iex.Binop.arg2); |
| break; |
| default: |
| vpanic("iselWordExpr_R_wrk-aluOp-arg2"); |
| } |
| addInstr(env, PPCInstr_Alu(aluOp, r_dst, r_srcL, ri_srcR)); |
| return r_dst; |
| } |
| |
| /* a shift? */ |
| switch (e->Iex.Binop.op) { |
| case Iop_Shl8: case Iop_Shl16: case Iop_Shl32: case Iop_Shl64: |
| shftOp = Pshft_SHL; break; |
| case Iop_Shr8: case Iop_Shr16: case Iop_Shr32: case Iop_Shr64: |
| shftOp = Pshft_SHR; break; |
| case Iop_Sar8: case Iop_Sar16: case Iop_Sar32: case Iop_Sar64: |
| shftOp = Pshft_SAR; break; |
| default: |
| shftOp = Pshft_INVALID; break; |
| } |
| /* we assume any literal values are on the second operand. */ |
| if (shftOp != Pshft_INVALID) { |
| HReg r_dst = newVRegI(env); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* ri_srcR = NULL; |
| /* get right arg into an RH, in the appropriate way */ |
| switch (shftOp) { |
| case Pshft_SHL: case Pshft_SHR: case Pshft_SAR: |
| if (!mode64) |
| ri_srcR = iselWordExpr_RH5u(env, e->Iex.Binop.arg2); |
| else |
| ri_srcR = iselWordExpr_RH6u(env, e->Iex.Binop.arg2); |
| break; |
| default: |
| vpanic("iselIntExpr_R_wrk-shftOp-arg2"); |
| } |
| /* widen the left arg if needed */ |
| if (shftOp == Pshft_SHR || shftOp == Pshft_SAR) { |
| if (ty == Ity_I8 || ty == Ity_I16) { |
| PPCRH* amt = PPCRH_Imm(False, |
| toUShort(ty == Ity_I8 ? 24 : 16)); |
| HReg tmp = newVRegI(env); |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, |
| True/*32bit shift*/, |
| tmp, r_srcL, amt)); |
| addInstr(env, PPCInstr_Shft(shftOp, |
| True/*32bit shift*/, |
| tmp, tmp, amt)); |
| r_srcL = tmp; |
| vassert(0); /* AWAITING TEST CASE */ |
| } |
| } |
| /* Only 64 expressions need 64bit shifts, |
| 32bit shifts are fine for all others */ |
| if (ty == Ity_I64) { |
| vassert(mode64); |
| addInstr(env, PPCInstr_Shft(shftOp, False/*64bit shift*/, |
| r_dst, r_srcL, ri_srcR)); |
| } else { |
| addInstr(env, PPCInstr_Shft(shftOp, True/*32bit shift*/, |
| r_dst, r_srcL, ri_srcR)); |
| } |
| return r_dst; |
| } |
| |
| /* How about a div? */ |
| if (e->Iex.Binop.op == Iop_DivS32 || |
| e->Iex.Binop.op == Iop_DivU32 || |
| e->Iex.Binop.op == Iop_DivS32E || |
| e->Iex.Binop.op == Iop_DivU32E) { |
| Bool syned = toBool((e->Iex.Binop.op == Iop_DivS32) || (e->Iex.Binop.op == Iop_DivS32E)); |
| HReg r_dst = newVRegI(env); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| addInstr( env, |
| PPCInstr_Div( ( ( e->Iex.Binop.op == Iop_DivU32E ) |
| || ( e->Iex.Binop.op == Iop_DivS32E ) ) ? True |
| : False, |
| syned, |
| True/*32bit div*/, |
| r_dst, |
| r_srcL, |
| r_srcR ) ); |
| return r_dst; |
| } |
| if (e->Iex.Binop.op == Iop_DivS64 || |
| e->Iex.Binop.op == Iop_DivU64 || e->Iex.Binop.op == Iop_DivS64E |
| || e->Iex.Binop.op == Iop_DivU64E ) { |
| Bool syned = toBool((e->Iex.Binop.op == Iop_DivS64) ||(e->Iex.Binop.op == Iop_DivS64E)); |
| HReg r_dst = newVRegI(env); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| vassert(mode64); |
| addInstr( env, |
| PPCInstr_Div( ( ( e->Iex.Binop.op == Iop_DivS64E ) |
| || ( e->Iex.Binop.op |
| == Iop_DivU64E ) ) ? True |
| : False, |
| syned, |
| False/*64bit div*/, |
| r_dst, |
| r_srcL, |
| r_srcR ) ); |
| return r_dst; |
| } |
| |
| /* No? Anyone for a mul? */ |
| if (e->Iex.Binop.op == Iop_Mul32 |
| || e->Iex.Binop.op == Iop_Mul64) { |
| Bool syned = False; |
| Bool sz32 = (e->Iex.Binop.op != Iop_Mul64); |
| HReg r_dst = newVRegI(env); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_MulL(syned, False/*lo32*/, sz32, |
| r_dst, r_srcL, r_srcR)); |
| return r_dst; |
| } |
| |
| /* 32 x 32 -> 64 multiply */ |
| if (mode64 |
| && (e->Iex.Binop.op == Iop_MullU32 |
| || e->Iex.Binop.op == Iop_MullS32)) { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg r_dst = newVRegI(env); |
| Bool syned = toBool(e->Iex.Binop.op == Iop_MullS32); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_MulL(False/*signedness irrelevant*/, |
| False/*lo32*/, True/*32bit mul*/, |
| tLo, r_srcL, r_srcR)); |
| addInstr(env, PPCInstr_MulL(syned, |
| True/*hi32*/, True/*32bit mul*/, |
| tHi, r_srcL, r_srcR)); |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, False/*64bit shift*/, |
| r_dst, tHi, PPCRH_Imm(False,32))); |
| addInstr(env, PPCInstr_Alu(Palu_OR, |
| r_dst, r_dst, PPCRH_Reg(tLo))); |
| return r_dst; |
| } |
| |
| /* El-mutanto 3-way compare? */ |
| if (e->Iex.Binop.op == Iop_CmpORD32S |
| || e->Iex.Binop.op == Iop_CmpORD32U) { |
| Bool syned = toBool(e->Iex.Binop.op == Iop_CmpORD32S); |
| HReg dst = newVRegI(env); |
| HReg srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* srcR = iselWordExpr_RH(env, syned, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Cmp(syned, True/*32bit cmp*/, |
| 7/*cr*/, srcL, srcR)); |
| addInstr(env, PPCInstr_MfCR(dst)); |
| addInstr(env, PPCInstr_Alu(Palu_AND, dst, dst, |
| PPCRH_Imm(False,7<<1))); |
| return dst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_CmpORD64S |
| || e->Iex.Binop.op == Iop_CmpORD64U) { |
| Bool syned = toBool(e->Iex.Binop.op == Iop_CmpORD64S); |
| HReg dst = newVRegI(env); |
| HReg srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* srcR = iselWordExpr_RH(env, syned, e->Iex.Binop.arg2); |
| vassert(mode64); |
| addInstr(env, PPCInstr_Cmp(syned, False/*64bit cmp*/, |
| 7/*cr*/, srcL, srcR)); |
| addInstr(env, PPCInstr_MfCR(dst)); |
| addInstr(env, PPCInstr_Alu(Palu_AND, dst, dst, |
| PPCRH_Imm(False,7<<1))); |
| return dst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_Max32U) { |
| HReg r1 = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r2 = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| HReg rdst = newVRegI(env); |
| PPCCondCode cc = mk_PPCCondCode( Pct_TRUE, Pcf_7LT ); |
| addInstr(env, mk_iMOVds_RR(rdst, r1)); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, rdst, PPCRH_Reg(r2))); |
| addInstr(env, PPCInstr_CMov(cc, rdst, PPCRI_Reg(r2))); |
| return rdst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_32HLto64) { |
| HReg r_Hi = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_Lo = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| HReg r_Tmp = newVRegI(env); |
| HReg r_dst = newVRegI(env); |
| HReg msk = newVRegI(env); |
| vassert(mode64); |
| /* r_dst = OR( r_Hi<<32, r_Lo ) */ |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, False/*64bit shift*/, |
| r_dst, r_Hi, PPCRH_Imm(False,32))); |
| addInstr(env, PPCInstr_LI(msk, 0xFFFFFFFF, mode64)); |
| addInstr(env, PPCInstr_Alu( Palu_AND, r_Tmp, r_Lo, |
| PPCRH_Reg(msk) )); |
| addInstr(env, PPCInstr_Alu( Palu_OR, r_dst, r_dst, |
| PPCRH_Reg(r_Tmp) )); |
| return r_dst; |
| } |
| |
| if ((e->Iex.Binop.op == Iop_CmpF64) || |
| (e->Iex.Binop.op == Iop_CmpD64) || |
| (e->Iex.Binop.op == Iop_CmpD128)) { |
| HReg fr_srcL; |
| HReg fr_srcL_lo; |
| HReg fr_srcR; |
| HReg fr_srcR_lo; |
| |
| HReg r_ccPPC = newVRegI(env); |
| HReg r_ccIR = newVRegI(env); |
| HReg r_ccIR_b0 = newVRegI(env); |
| HReg r_ccIR_b2 = newVRegI(env); |
| HReg r_ccIR_b6 = newVRegI(env); |
| |
| if (e->Iex.Binop.op == Iop_CmpF64) { |
| fr_srcL = iselDblExpr(env, e->Iex.Binop.arg1); |
| fr_srcR = iselDblExpr(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_FpCmp(r_ccPPC, fr_srcL, fr_srcR)); |
| |
| } else if (e->Iex.Binop.op == Iop_CmpD64) { |
| fr_srcL = iselDfp64Expr(env, e->Iex.Binop.arg1); |
| fr_srcR = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Dfp64Cmp(r_ccPPC, fr_srcL, fr_srcR)); |
| |
| } else { // e->Iex.Binop.op == Iop_CmpD128 |
| iselDfp128Expr(&fr_srcL, &fr_srcL_lo, env, e->Iex.Binop.arg1); |
| iselDfp128Expr(&fr_srcR, &fr_srcR_lo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Dfp128Cmp(r_ccPPC, fr_srcL, fr_srcL_lo, |
| fr_srcR, fr_srcR_lo)); |
| } |
| |
| /* Map compare result from PPC to IR, |
| conforming to CmpF64 definition. */ |
| /* |
| FP cmp result | PPC | IR |
| -------------------------- |
| UN | 0x1 | 0x45 |
| EQ | 0x2 | 0x40 |
| GT | 0x4 | 0x00 |
| LT | 0x8 | 0x01 |
| */ |
| |
| // r_ccIR_b0 = r_ccPPC[0] | r_ccPPC[3] |
| addInstr(env, PPCInstr_Shft(Pshft_SHR, True/*32bit shift*/, |
| r_ccIR_b0, r_ccPPC, |
| PPCRH_Imm(False,0x3))); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_ccIR_b0, |
| r_ccPPC, PPCRH_Reg(r_ccIR_b0))); |
| addInstr(env, PPCInstr_Alu(Palu_AND, r_ccIR_b0, |
| r_ccIR_b0, PPCRH_Imm(False,0x1))); |
| |
| // r_ccIR_b2 = r_ccPPC[0] |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| r_ccIR_b2, r_ccPPC, |
| PPCRH_Imm(False,0x2))); |
| addInstr(env, PPCInstr_Alu(Palu_AND, r_ccIR_b2, |
| r_ccIR_b2, PPCRH_Imm(False,0x4))); |
| |
| // r_ccIR_b6 = r_ccPPC[0] | r_ccPPC[1] |
| addInstr(env, PPCInstr_Shft(Pshft_SHR, True/*32bit shift*/, |
| r_ccIR_b6, r_ccPPC, |
| PPCRH_Imm(False,0x1))); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_ccIR_b6, |
| r_ccPPC, PPCRH_Reg(r_ccIR_b6))); |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| r_ccIR_b6, r_ccIR_b6, |
| PPCRH_Imm(False,0x6))); |
| addInstr(env, PPCInstr_Alu(Palu_AND, r_ccIR_b6, |
| r_ccIR_b6, PPCRH_Imm(False,0x40))); |
| |
| // r_ccIR = r_ccIR_b0 | r_ccIR_b2 | r_ccIR_b6 |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_ccIR, |
| r_ccIR_b0, PPCRH_Reg(r_ccIR_b2))); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_ccIR, |
| r_ccIR, PPCRH_Reg(r_ccIR_b6))); |
| return r_ccIR; |
| } |
| |
| if ( e->Iex.Binop.op == Iop_F64toI32S || |
| e->Iex.Binop.op == Iop_F64toI32U ) { |
| /* This works in both mode64 and mode32. */ |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| HReg fsrc = iselDblExpr(env, e->Iex.Binop.arg2); |
| HReg ftmp = newVRegF(env); |
| HReg idst = newVRegI(env); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpCftI(False/*F->I*/, True/*int32*/, |
| e->Iex.Binop.op == Iop_F64toI32S ? True/*syned*/ |
| : False, |
| True/*flt64*/, |
| ftmp, fsrc)); |
| addInstr(env, PPCInstr_FpSTFIW(r1, ftmp)); |
| addInstr(env, PPCInstr_Load(4, idst, zero_r1, mode64)); |
| |
| /* in 64-bit mode we need to sign-widen idst. */ |
| if (mode64) |
| addInstr(env, PPCInstr_Unary(Pun_EXTSW, idst, idst)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return idst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_F64toI64S || e->Iex.Binop.op == Iop_F64toI64U ) { |
| if (mode64) { |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| HReg fsrc = iselDblExpr(env, e->Iex.Binop.arg2); |
| HReg idst = newVRegI(env); |
| HReg ftmp = newVRegF(env); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpCftI(False/*F->I*/, False/*int64*/, |
| ( e->Iex.Binop.op == Iop_F64toI64S ) ? True |
| : False, |
| True, ftmp, fsrc)); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, ftmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(8, idst, zero_r1, True/*mode64*/)); |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return idst; |
| } |
| } |
| |
| if (e->Iex.Binop.op == Iop_D64toI64S ) { |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| HReg idst = newVRegI(env); |
| HReg ftmp = newVRegF(env); |
| |
| /* Set host rounding mode */ |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DCTFIX, ftmp, fr_src)); |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, ftmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(8, idst, zero_r1, mode64)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return idst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_D128toI64S ) { |
| PPCFpOp fpop = Pfp_DCTFIXQ; |
| HReg r_srcHi = newVRegF(env); |
| HReg r_srcLo = newVRegF(env); |
| HReg idst = newVRegI(env); |
| HReg ftmp = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_DfpD128toD64(fpop, ftmp, r_srcHi, r_srcLo)); |
| |
| // put the D64 result into an integer register |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, ftmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(8, idst, zero_r1, True/*mode64*/)); |
| add_to_sp( env, 16 ); |
| return idst; |
| } |
| break; |
| } |
| |
| /* --------- UNARY OP --------- */ |
| case Iex_Unop: { |
| IROp op_unop = e->Iex.Unop.op; |
| |
| /* 1Uto8(32to1(expr32)) */ |
| DEFINE_PATTERN(p_32to1_then_1Uto8, |
| unop(Iop_1Uto8,unop(Iop_32to1,bind(0)))); |
| if (matchIRExpr(&mi,p_32to1_then_1Uto8,e)) { |
| IRExpr* expr32 = mi.bindee[0]; |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, expr32); |
| addInstr(env, PPCInstr_Alu(Palu_AND, r_dst, |
| r_src, PPCRH_Imm(False,1))); |
| return r_dst; |
| } |
| |
| /* 16Uto32(LDbe:I16(expr32)) */ |
| { |
| DECLARE_PATTERN(p_LDbe16_then_16Uto32); |
| DEFINE_PATTERN(p_LDbe16_then_16Uto32, |
| unop(Iop_16Uto32, |
| IRExpr_Load(Iend_BE,Ity_I16,bind(0))) ); |
| if (matchIRExpr(&mi,p_LDbe16_then_16Uto32,e)) { |
| HReg r_dst = newVRegI(env); |
| PPCAMode* amode |
| = iselWordExpr_AMode( env, mi.bindee[0], Ity_I16/*xfer*/ ); |
| addInstr(env, PPCInstr_Load(2,r_dst,amode, mode64)); |
| return r_dst; |
| } |
| } |
| |
| switch (op_unop) { |
| case Iop_8Uto16: |
| case Iop_8Uto32: |
| case Iop_8Uto64: |
| case Iop_16Uto32: |
| case Iop_16Uto64: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| UShort mask = toUShort(op_unop==Iop_16Uto64 ? 0xFFFF : |
| op_unop==Iop_16Uto32 ? 0xFFFF : 0xFF); |
| addInstr(env, PPCInstr_Alu(Palu_AND,r_dst,r_src, |
| PPCRH_Imm(False,mask))); |
| return r_dst; |
| } |
| case Iop_32Uto64: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| vassert(mode64); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHL, False/*64bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,32))); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHR, False/*64bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,32))); |
| return r_dst; |
| } |
| case Iop_8Sto16: |
| case Iop_8Sto32: |
| case Iop_16Sto32: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| UShort amt = toUShort(op_unop==Iop_16Sto32 ? 16 : 24); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,amt))); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,amt))); |
| return r_dst; |
| } |
| case Iop_8Sto64: |
| case Iop_16Sto64: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| UShort amt = toUShort(op_unop==Iop_8Sto64 ? 56 : 48); |
| vassert(mode64); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHL, False/*64bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,amt))); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SAR, False/*64bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,amt))); |
| return r_dst; |
| } |
| case Iop_32Sto64: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| vassert(mode64); |
| /* According to the IBM docs, in 64 bit mode, srawi r,r,0 |
| sign extends the lower 32 bits into the upper 32 bits. */ |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,0))); |
| return r_dst; |
| } |
| case Iop_Not8: |
| case Iop_Not16: |
| case Iop_Not32: |
| case Iop_Not64: { |
| if (op_unop == Iop_Not64) vassert(mode64); |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Unary(Pun_NOT,r_dst,r_src)); |
| return r_dst; |
| } |
| case Iop_64HIto32: { |
| if (!mode64) { |
| HReg rHi, rLo; |
| iselInt64Expr(&rHi,&rLo, env, e->Iex.Unop.arg); |
| return rHi; /* and abandon rLo .. poor wee thing :-) */ |
| } else { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHR, False/*64bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,32))); |
| return r_dst; |
| } |
| } |
| case Iop_64to32: { |
| if (!mode64) { |
| HReg rHi, rLo; |
| iselInt64Expr(&rHi,&rLo, env, e->Iex.Unop.arg); |
| return rLo; /* similar stupid comment to the above ... */ |
| } else { |
| /* This is a no-op. */ |
| return iselWordExpr_R(env, e->Iex.Unop.arg); |
| } |
| } |
| case Iop_64to16: { |
| if (mode64) { /* This is a no-op. */ |
| return iselWordExpr_R(env, e->Iex.Unop.arg); |
| } |
| break; /* evidently not used in 32-bit mode */ |
| } |
| case Iop_16HIto8: |
| case Iop_32HIto16: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| UShort shift = toUShort(op_unop == Iop_16HIto8 ? 8 : 16); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHR, True/*32bit shift*/, |
| r_dst, r_src, PPCRH_Imm(False,shift))); |
| return r_dst; |
| } |
| case Iop_128HIto64: |
| if (mode64) { |
| HReg rHi, rLo; |
| iselInt128Expr(&rHi,&rLo, env, e->Iex.Unop.arg); |
| return rHi; /* and abandon rLo .. poor wee thing :-) */ |
| } |
| break; |
| case Iop_128to64: |
| if (mode64) { |
| HReg rHi, rLo; |
| iselInt128Expr(&rHi,&rLo, env, e->Iex.Unop.arg); |
| return rLo; /* similar stupid comment to the above ... */ |
| } |
| break; |
| case Iop_1Uto64: |
| case Iop_1Uto32: |
| case Iop_1Uto8: |
| if ((op_unop != Iop_1Uto64) || mode64) { |
| HReg r_dst = newVRegI(env); |
| PPCCondCode cond = iselCondCode(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Set(cond,r_dst)); |
| return r_dst; |
| } |
| break; |
| case Iop_1Sto8: |
| case Iop_1Sto16: |
| case Iop_1Sto32: { |
| /* could do better than this, but for now ... */ |
| HReg r_dst = newVRegI(env); |
| PPCCondCode cond = iselCondCode(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Set(cond,r_dst)); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,31))); |
| addInstr(env, |
| PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,31))); |
| return r_dst; |
| } |
| case Iop_1Sto64: |
| if (mode64) { |
| /* could do better than this, but for now ... */ |
| HReg r_dst = newVRegI(env); |
| PPCCondCode cond = iselCondCode(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Set(cond,r_dst)); |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, False/*64bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,63))); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, False/*64bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False,63))); |
| return r_dst; |
| } |
| break; |
| case Iop_Clz32: |
| case Iop_Clz64: { |
| HReg r_src, r_dst; |
| PPCUnaryOp op_clz = (op_unop == Iop_Clz32) ? Pun_CLZ32 : |
| Pun_CLZ64; |
| if (op_unop == Iop_Clz64 && !mode64) |
| goto irreducible; |
| /* Count leading zeroes. */ |
| r_dst = newVRegI(env); |
| r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Unary(op_clz,r_dst,r_src)); |
| return r_dst; |
| } |
| |
| case Iop_Left8: |
| case Iop_Left16: |
| case Iop_Left32: |
| case Iop_Left64: { |
| HReg r_src, r_dst; |
| if (op_unop == Iop_Left64 && !mode64) |
| goto irreducible; |
| r_dst = newVRegI(env); |
| r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Unary(Pun_NEG,r_dst,r_src)); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_dst, r_dst, PPCRH_Reg(r_src))); |
| return r_dst; |
| } |
| |
| case Iop_CmpwNEZ32: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Unary(Pun_NEG,r_dst,r_src)); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_dst, r_dst, PPCRH_Reg(r_src))); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False, 31))); |
| return r_dst; |
| } |
| |
| case Iop_CmpwNEZ64: { |
| HReg r_dst = newVRegI(env); |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| if (!mode64) goto irreducible; |
| addInstr(env, PPCInstr_Unary(Pun_NEG,r_dst,r_src)); |
| addInstr(env, PPCInstr_Alu(Palu_OR, r_dst, r_dst, PPCRH_Reg(r_src))); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, False/*64bit shift*/, |
| r_dst, r_dst, PPCRH_Imm(False, 63))); |
| return r_dst; |
| } |
| |
| case Iop_V128to32: { |
| HReg r_aligned16; |
| HReg dst = newVRegI(env); |
| HReg vec = iselVecExpr(env, e->Iex.Unop.arg); |
| PPCAMode *am_off0, *am_off12; |
| sub_from_sp( env, 32 ); // Move SP down 32 bytes |
| |
| // get a quadword aligned address within our stack space |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_off12 = PPCAMode_IR( 12,r_aligned16 ); |
| |
| // store vec, load low word to dst |
| addInstr(env, |
| PPCInstr_AvLdSt( False/*store*/, 16, vec, am_off0 )); |
| addInstr(env, |
| PPCInstr_Load( 4, dst, am_off12, mode64 )); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| return dst; |
| } |
| |
| case Iop_V128to64: |
| case Iop_V128HIto64: |
| if (mode64) { |
| HReg r_aligned16; |
| HReg dst = newVRegI(env); |
| HReg vec = iselVecExpr(env, e->Iex.Unop.arg); |
| PPCAMode *am_off0, *am_off8; |
| sub_from_sp( env, 32 ); // Move SP down 32 bytes |
| |
| // get a quadword aligned address within our stack space |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_off8 = PPCAMode_IR( 8 ,r_aligned16 ); |
| |
| // store vec, load low word (+8) or high (+0) to dst |
| addInstr(env, |
| PPCInstr_AvLdSt( False/*store*/, 16, vec, am_off0 )); |
| addInstr(env, |
| PPCInstr_Load( |
| 8, dst, |
| op_unop == Iop_V128HIto64 ? am_off0 : am_off8, |
| mode64 )); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| return dst; |
| } |
| break; |
| case Iop_16to8: |
| case Iop_32to8: |
| case Iop_32to16: |
| case Iop_64to8: |
| /* These are no-ops. */ |
| return iselWordExpr_R(env, e->Iex.Unop.arg); |
| |
| /* ReinterpF64asI64(e) */ |
| /* Given an IEEE754 double, produce an I64 with the same bit |
| pattern. */ |
| case Iop_ReinterpF64asI64: |
| if (mode64) { |
| PPCAMode *am_addr; |
| HReg fr_src = iselDblExpr(env, e->Iex.Unop.arg); |
| HReg r_dst = newVRegI(env); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr = PPCAMode_IR( 0, StackFramePtr(mode64) ); |
| |
| // store as F64 |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr )); |
| // load as Ity_I64 |
| addInstr(env, PPCInstr_Load( 8, r_dst, am_addr, mode64 )); |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return r_dst; |
| } |
| break; |
| |
| /* ReinterpF32asI32(e) */ |
| /* Given an IEEE754 float, produce an I32 with the same bit |
| pattern. */ |
| case Iop_ReinterpF32asI32: { |
| /* I believe this generates correct code for both 32- and |
| 64-bit hosts. */ |
| PPCAMode *am_addr; |
| HReg fr_src = iselFltExpr(env, e->Iex.Unop.arg); |
| HReg r_dst = newVRegI(env); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr = PPCAMode_IR( 0, StackFramePtr(mode64) ); |
| |
| // store as F32 |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 4, |
| fr_src, am_addr )); |
| // load as Ity_I32 |
| addInstr(env, PPCInstr_Load( 4, r_dst, am_addr, mode64 )); |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return r_dst; |
| } |
| break; |
| |
| case Iop_ReinterpD64asI64: |
| if (mode64) { |
| PPCAMode *am_addr; |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Unop.arg); |
| HReg r_dst = newVRegI(env); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr = PPCAMode_IR( 0, StackFramePtr(mode64) ); |
| |
| // store as D64 |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr )); |
| // load as Ity_I64 |
| addInstr(env, PPCInstr_Load( 8, r_dst, am_addr, mode64 )); |
| add_to_sp( env, 16 ); // Reset SP |
| return r_dst; |
| } |
| break; |
| |
| case Iop_BCDtoDPB: { |
| /* the following is only valid in 64 bit mode */ |
| if (!mode64) break; |
| |
| PPCCondCode cc; |
| UInt argiregs; |
| HReg argregs[1]; |
| HReg r_dst = newVRegI(env); |
| Int argreg; |
| HWord* fdescr; |
| |
| argiregs = 0; |
| argreg = 0; |
| argregs[0] = hregPPC_GPR3(mode64); |
| |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg++], |
| iselWordExpr_R(env, e->Iex.Unop.arg) ) ); |
| |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| |
| fdescr = (HWord*)h_calc_BCDtoDPB; |
| addInstr(env, PPCInstr_Call( cc, (Addr64)(fdescr[0]), |
| argiregs, mk_RetLoc_simple(RLPri_Int)) ); |
| |
| addInstr(env, mk_iMOVds_RR(r_dst, argregs[0])); |
| return r_dst; |
| } |
| |
| case Iop_DPBtoBCD: { |
| /* the following is only valid in 64 bit mode */ |
| if (!mode64) break; |
| |
| PPCCondCode cc; |
| UInt argiregs; |
| HReg argregs[1]; |
| HReg r_dst = newVRegI(env); |
| Int argreg; |
| HWord* fdescr; |
| |
| argiregs = 0; |
| argreg = 0; |
| argregs[0] = hregPPC_GPR3(mode64); |
| |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg++], |
| iselWordExpr_R(env, e->Iex.Unop.arg) ) ); |
| |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| |
| fdescr = (HWord*)h_calc_DPBtoBCD; |
| addInstr(env, PPCInstr_Call( cc, (Addr64)(fdescr[0]), |
| argiregs, mk_RetLoc_simple(RLPri_Int) ) ); |
| |
| addInstr(env, mk_iMOVds_RR(r_dst, argregs[0])); |
| return r_dst; |
| } |
| |
| default: |
| break; |
| } |
| |
| switch (e->Iex.Unop.op) { |
| case Iop_ExtractExpD64: { |
| |
| HReg fr_dst = newVRegI(env); |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Unop.arg); |
| HReg tmp = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DXEX, tmp, fr_src)); |
| |
| // put the D64 result into a integer register |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, tmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(8, fr_dst, zero_r1, env->mode64)); |
| add_to_sp( env, 16 ); |
| return fr_dst; |
| } |
| case Iop_ExtractExpD128: { |
| HReg fr_dst = newVRegI(env); |
| HReg r_srcHi; |
| HReg r_srcLo; |
| HReg tmp = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_ExtractExpD128(Pfp_DXEXQ, tmp, |
| r_srcHi, r_srcLo)); |
| |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, tmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(8, fr_dst, zero_r1, env->mode64)); |
| add_to_sp( env, 16 ); |
| return fr_dst; |
| } |
| default: |
| break; |
| } |
| |
| break; |
| } |
| |
| /* --------- GET --------- */ |
| case Iex_Get: { |
| if (ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && mode64)) { |
| HReg r_dst = newVRegI(env); |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(mode64) ); |
| addInstr(env, PPCInstr_Load( toUChar(sizeofIRType(ty)), |
| r_dst, am_addr, mode64 )); |
| return r_dst; |
| } |
| break; |
| } |
| |
| case Iex_GetI: { |
| PPCAMode* src_am |
| = genGuestArrayOffset( env, e->Iex.GetI.descr, |
| e->Iex.GetI.ix, e->Iex.GetI.bias ); |
| HReg r_dst = newVRegI(env); |
| if (mode64 && ty == Ity_I64) { |
| addInstr(env, PPCInstr_Load( toUChar(8), |
| r_dst, src_am, mode64 )); |
| return r_dst; |
| } |
| if ((!mode64) && ty == Ity_I32) { |
| addInstr(env, PPCInstr_Load( toUChar(4), |
| r_dst, src_am, mode64 )); |
| return r_dst; |
| } |
| break; |
| } |
| |
| /* --------- CCALL --------- */ |
| case Iex_CCall: { |
| HReg r_dst = newVRegI(env); |
| vassert(ty == Ity_I32); |
| |
| /* be very restrictive for now. Only 32/64-bit ints allowed for |
| args, and 32 bits for return type. */ |
| if (e->Iex.CCall.retty != Ity_I32) |
| goto irreducible; |
| |
| /* Marshal args, do the call, clear stack. */ |
| UInt addToSp = 0; |
| RetLoc rloc = mk_RetLoc_INVALID(); |
| doHelperCall( &addToSp, &rloc, env, NULL/*guard*/, |
| e->Iex.CCall.cee, e->Iex.CCall.retty, e->Iex.CCall.args ); |
| vassert(is_sane_RetLoc(rloc)); |
| vassert(rloc.pri == RLPri_Int); |
| vassert(addToSp == 0); |
| |
| /* GPR3 now holds the destination address from Pin_Goto */ |
| addInstr(env, mk_iMOVds_RR(r_dst, hregPPC_GPR3(mode64))); |
| return r_dst; |
| } |
| |
| /* --------- LITERAL --------- */ |
| /* 32/16/8-bit literals */ |
| case Iex_Const: { |
| Long l; |
| HReg r_dst = newVRegI(env); |
| IRConst* con = e->Iex.Const.con; |
| switch (con->tag) { |
| case Ico_U64: if (!mode64) goto irreducible; |
| l = (Long) con->Ico.U64; break; |
| case Ico_U32: l = (Long)(Int) con->Ico.U32; break; |
| case Ico_U16: l = (Long)(Int)(Short)con->Ico.U16; break; |
| case Ico_U8: l = (Long)(Int)(Char )con->Ico.U8; break; |
| default: vpanic("iselIntExpr_R.const(ppc)"); |
| } |
| addInstr(env, PPCInstr_LI(r_dst, (ULong)l, mode64)); |
| return r_dst; |
| } |
| |
| /* --------- MULTIPLEX --------- */ |
| case Iex_ITE: { // VFD |
| if ((ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && mode64)) && |
| typeOfIRExpr(env->type_env,e->Iex.ITE.cond) == Ity_I1) { |
| PPCRI* r1 = iselWordExpr_RI(env, e->Iex.ITE.iftrue); |
| HReg r0 = iselWordExpr_R(env, e->Iex.ITE.iffalse); |
| HReg r_dst = newVRegI(env); |
| addInstr(env, mk_iMOVds_RR(r_dst,r0)); |
| PPCCondCode cc = iselCondCode(env, e->Iex.ITE.cond); |
| addInstr(env, PPCInstr_CMov(cc, r_dst, r1)); |
| return r_dst; |
| } |
| break; |
| } |
| |
| default: |
| break; |
| } /* switch (e->tag) */ |
| |
| |
| /* We get here if no pattern matched. */ |
| irreducible: |
| ppIRExpr(e); |
| vpanic("iselIntExpr_R(ppc): cannot reduce tree"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Integer expression auxiliaries ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* --------------------- AMODEs --------------------- */ |
| |
| /* Return an AMode which computes the value of the specified |
| expression, possibly also adding insns to the code list as a |
| result. The expression may only be a word-size one. |
| */ |
| |
| static Bool uInt_fits_in_16_bits ( UInt u ) |
| { |
| /* Is u the same as the sign-extend of its lower 16 bits? */ |
| Int i = u & 0xFFFF; |
| i <<= 16; |
| i >>= 16; |
| return toBool(u == (UInt)i); |
| } |
| |
| static Bool uLong_fits_in_16_bits ( ULong u ) |
| { |
| /* Is u the same as the sign-extend of its lower 16 bits? */ |
| Long i = u & 0xFFFFULL; |
| i <<= 48; |
| i >>= 48; |
| return toBool(u == (ULong)i); |
| } |
| |
| static Bool uLong_is_4_aligned ( ULong u ) |
| { |
| return toBool((u & 3ULL) == 0); |
| } |
| |
| static Bool sane_AMode ( ISelEnv* env, PPCAMode* am ) |
| { |
| Bool mode64 = env->mode64; |
| switch (am->tag) { |
| case Pam_IR: |
| /* Using uInt_fits_in_16_bits in 64-bit mode seems a bit bogus, |
| somehow, but I think it's OK. */ |
| return toBool( hregClass(am->Pam.IR.base) == HRcGPR(mode64) && |
| hregIsVirtual(am->Pam.IR.base) && |
| uInt_fits_in_16_bits(am->Pam.IR.index) ); |
| case Pam_RR: |
| return toBool( hregClass(am->Pam.RR.base) == HRcGPR(mode64) && |
| hregIsVirtual(am->Pam.RR.base) && |
| hregClass(am->Pam.RR.index) == HRcGPR(mode64) && |
| hregIsVirtual(am->Pam.RR.index) ); |
| default: |
| vpanic("sane_AMode: unknown ppc amode tag"); |
| } |
| } |
| |
| static |
| PPCAMode* iselWordExpr_AMode ( ISelEnv* env, IRExpr* e, IRType xferTy ) |
| { |
| PPCAMode* am = iselWordExpr_AMode_wrk(env, e, xferTy); |
| vassert(sane_AMode(env, am)); |
| return am; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCAMode* iselWordExpr_AMode_wrk ( ISelEnv* env, IRExpr* e, IRType xferTy ) |
| { |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| |
| if (env->mode64) { |
| |
| /* If the data load/store type is I32 or I64, this amode might |
| be destined for use in ld/ldu/lwa/st/stu. In which case |
| insist that if it comes out as an _IR, the immediate must |
| have its bottom two bits be zero. This does assume that for |
| any other type (I8/I16/I128/F32/F64/V128) the amode will not |
| be parked in any such instruction. But that seems a |
| reasonable assumption. */ |
| Bool aligned4imm = toBool(xferTy == Ity_I32 || xferTy == Ity_I64); |
| |
| vassert(ty == Ity_I64); |
| |
| /* Add64(expr,i), where i == sign-extend of (i & 0xFFFF) */ |
| if (e->tag == Iex_Binop |
| && e->Iex.Binop.op == Iop_Add64 |
| && e->Iex.Binop.arg2->tag == Iex_Const |
| && e->Iex.Binop.arg2->Iex.Const.con->tag == Ico_U64 |
| && (aligned4imm ? uLong_is_4_aligned(e->Iex.Binop.arg2 |
| ->Iex.Const.con->Ico.U64) |
| : True) |
| && uLong_fits_in_16_bits(e->Iex.Binop.arg2 |
| ->Iex.Const.con->Ico.U64)) { |
| return PPCAMode_IR( (Int)e->Iex.Binop.arg2->Iex.Const.con->Ico.U64, |
| iselWordExpr_R(env, e->Iex.Binop.arg1) ); |
| } |
| |
| /* Add64(expr,expr) */ |
| if (e->tag == Iex_Binop |
| && e->Iex.Binop.op == Iop_Add64) { |
| HReg r_base = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_idx = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| return PPCAMode_RR( r_idx, r_base ); |
| } |
| |
| } else { |
| |
| vassert(ty == Ity_I32); |
| |
| /* Add32(expr,i), where i == sign-extend of (i & 0xFFFF) */ |
| if (e->tag == Iex_Binop |
| && e->Iex.Binop.op == Iop_Add32 |
| && e->Iex.Binop.arg2->tag == Iex_Const |
| && e->Iex.Binop.arg2->Iex.Const.con->tag == Ico_U32 |
| && uInt_fits_in_16_bits(e->Iex.Binop.arg2 |
| ->Iex.Const.con->Ico.U32)) { |
| return PPCAMode_IR( (Int)e->Iex.Binop.arg2->Iex.Const.con->Ico.U32, |
| iselWordExpr_R(env, e->Iex.Binop.arg1) ); |
| } |
| |
| /* Add32(expr,expr) */ |
| if (e->tag == Iex_Binop |
| && e->Iex.Binop.op == Iop_Add32) { |
| HReg r_base = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_idx = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| return PPCAMode_RR( r_idx, r_base ); |
| } |
| |
| } |
| |
| /* Doesn't match anything in particular. Generate it into |
| a register and use that. */ |
| return PPCAMode_IR( 0, iselWordExpr_R(env,e) ); |
| } |
| |
| |
| /* --------------------- RH --------------------- */ |
| |
| /* Compute an I8/I16/I32 (and I64, in 64-bit mode) into a RH |
| (reg-or-halfword-immediate). It's important to specify whether the |
| immediate is to be regarded as signed or not. If yes, this will |
| never return -32768 as an immediate; this guaranteed that all |
| signed immediates that are return can have their sign inverted if |
| need be. */ |
| |
| static PPCRH* iselWordExpr_RH ( ISelEnv* env, Bool syned, IRExpr* e ) |
| { |
| PPCRH* ri = iselWordExpr_RH_wrk(env, syned, e); |
| /* sanity checks ... */ |
| switch (ri->tag) { |
| case Prh_Imm: |
| vassert(ri->Prh.Imm.syned == syned); |
| if (syned) |
| vassert(ri->Prh.Imm.imm16 != 0x8000); |
| return ri; |
| case Prh_Reg: |
| vassert(hregClass(ri->Prh.Reg.reg) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(ri->Prh.Reg.reg)); |
| return ri; |
| default: |
| vpanic("iselIntExpr_RH: unknown ppc RH tag"); |
| } |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCRH* iselWordExpr_RH_wrk ( ISelEnv* env, Bool syned, IRExpr* e ) |
| { |
| ULong u; |
| Long l; |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && env->mode64)); |
| |
| /* special case: immediate */ |
| if (e->tag == Iex_Const) { |
| IRConst* con = e->Iex.Const.con; |
| /* What value are we aiming to generate? */ |
| switch (con->tag) { |
| /* Note: Not sign-extending - we carry 'syned' around */ |
| case Ico_U64: vassert(env->mode64); |
| u = con->Ico.U64; break; |
| case Ico_U32: u = 0xFFFFFFFF & con->Ico.U32; break; |
| case Ico_U16: u = 0x0000FFFF & con->Ico.U16; break; |
| case Ico_U8: u = 0x000000FF & con->Ico.U8; break; |
| default: vpanic("iselIntExpr_RH.Iex_Const(ppch)"); |
| } |
| l = (Long)u; |
| /* Now figure out if it's representable. */ |
| if (!syned && u <= 65535) { |
| return PPCRH_Imm(False/*unsigned*/, toUShort(u & 0xFFFF)); |
| } |
| if (syned && l >= -32767 && l <= 32767) { |
| return PPCRH_Imm(True/*signed*/, toUShort(u & 0xFFFF)); |
| } |
| /* no luck; use the Slow Way. */ |
| } |
| |
| /* default case: calculate into a register and return that */ |
| return PPCRH_Reg( iselWordExpr_R ( env, e ) ); |
| } |
| |
| |
| /* --------------------- RIs --------------------- */ |
| |
| /* Calculate an expression into an PPCRI operand. As with |
| iselIntExpr_R, the expression can have type 32, 16 or 8 bits, or, |
| in 64-bit mode, 64 bits. */ |
| |
| static PPCRI* iselWordExpr_RI ( ISelEnv* env, IRExpr* e ) |
| { |
| PPCRI* ri = iselWordExpr_RI_wrk(env, e); |
| /* sanity checks ... */ |
| switch (ri->tag) { |
| case Pri_Imm: |
| return ri; |
| case Pri_Reg: |
| vassert(hregClass(ri->Pri.Reg) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(ri->Pri.Reg)); |
| return ri; |
| default: |
| vpanic("iselIntExpr_RI: unknown ppc RI tag"); |
| } |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCRI* iselWordExpr_RI_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| Long l; |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && env->mode64)); |
| |
| /* special case: immediate */ |
| if (e->tag == Iex_Const) { |
| IRConst* con = e->Iex.Const.con; |
| switch (con->tag) { |
| case Ico_U64: vassert(env->mode64); |
| l = (Long) con->Ico.U64; break; |
| case Ico_U32: l = (Long)(Int) con->Ico.U32; break; |
| case Ico_U16: l = (Long)(Int)(Short)con->Ico.U16; break; |
| case Ico_U8: l = (Long)(Int)(Char )con->Ico.U8; break; |
| default: vpanic("iselIntExpr_RI.Iex_Const(ppch)"); |
| } |
| return PPCRI_Imm((ULong)l); |
| } |
| |
| /* default case: calculate into a register and return that */ |
| return PPCRI_Reg( iselWordExpr_R ( env, e ) ); |
| } |
| |
| |
| /* --------------------- RH5u --------------------- */ |
| |
| /* Compute an I8 into a reg-or-5-bit-unsigned-immediate, the latter |
| being an immediate in the range 1 .. 31 inclusive. Used for doing |
| shift amounts. Only used in 32-bit mode. */ |
| |
| static PPCRH* iselWordExpr_RH5u ( ISelEnv* env, IRExpr* e ) |
| { |
| PPCRH* ri; |
| vassert(!env->mode64); |
| ri = iselWordExpr_RH5u_wrk(env, e); |
| /* sanity checks ... */ |
| switch (ri->tag) { |
| case Prh_Imm: |
| vassert(ri->Prh.Imm.imm16 >= 1 && ri->Prh.Imm.imm16 <= 31); |
| vassert(!ri->Prh.Imm.syned); |
| return ri; |
| case Prh_Reg: |
| vassert(hregClass(ri->Prh.Reg.reg) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(ri->Prh.Reg.reg)); |
| return ri; |
| default: |
| vpanic("iselIntExpr_RH5u: unknown ppc RI tag"); |
| } |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCRH* iselWordExpr_RH5u_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_I8); |
| |
| /* special case: immediate */ |
| if (e->tag == Iex_Const |
| && e->Iex.Const.con->tag == Ico_U8 |
| && e->Iex.Const.con->Ico.U8 >= 1 |
| && e->Iex.Const.con->Ico.U8 <= 31) { |
| return PPCRH_Imm(False/*unsigned*/, e->Iex.Const.con->Ico.U8); |
| } |
| |
| /* default case: calculate into a register and return that */ |
| return PPCRH_Reg( iselWordExpr_R ( env, e ) ); |
| } |
| |
| |
| /* --------------------- RH6u --------------------- */ |
| |
| /* Compute an I8 into a reg-or-6-bit-unsigned-immediate, the latter |
| being an immediate in the range 1 .. 63 inclusive. Used for doing |
| shift amounts. Only used in 64-bit mode. */ |
| |
| static PPCRH* iselWordExpr_RH6u ( ISelEnv* env, IRExpr* e ) |
| { |
| PPCRH* ri; |
| vassert(env->mode64); |
| ri = iselWordExpr_RH6u_wrk(env, e); |
| /* sanity checks ... */ |
| switch (ri->tag) { |
| case Prh_Imm: |
| vassert(ri->Prh.Imm.imm16 >= 1 && ri->Prh.Imm.imm16 <= 63); |
| vassert(!ri->Prh.Imm.syned); |
| return ri; |
| case Prh_Reg: |
| vassert(hregClass(ri->Prh.Reg.reg) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(ri->Prh.Reg.reg)); |
| return ri; |
| default: |
| vpanic("iselIntExpr_RH6u: unknown ppc64 RI tag"); |
| } |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCRH* iselWordExpr_RH6u_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_I8); |
| |
| /* special case: immediate */ |
| if (e->tag == Iex_Const |
| && e->Iex.Const.con->tag == Ico_U8 |
| && e->Iex.Const.con->Ico.U8 >= 1 |
| && e->Iex.Const.con->Ico.U8 <= 63) { |
| return PPCRH_Imm(False/*unsigned*/, e->Iex.Const.con->Ico.U8); |
| } |
| |
| /* default case: calculate into a register and return that */ |
| return PPCRH_Reg( iselWordExpr_R ( env, e ) ); |
| } |
| |
| |
| /* --------------------- CONDCODE --------------------- */ |
| |
| /* Generate code to evaluated a bit-typed expression, returning the |
| condition code which would correspond when the expression would |
| notionally have returned 1. */ |
| |
| static PPCCondCode iselCondCode ( ISelEnv* env, IRExpr* e ) |
| { |
| /* Uh, there's nothing we can sanity check here, unfortunately. */ |
| return iselCondCode_wrk(env,e); |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static PPCCondCode iselCondCode_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| vassert(e); |
| vassert(typeOfIRExpr(env->type_env,e) == Ity_I1); |
| |
| /* Constant 1:Bit */ |
| if (e->tag == Iex_Const && e->Iex.Const.con->Ico.U1 == True) { |
| // Make a compare that will always be true: |
| HReg r_zero = newVRegI(env); |
| addInstr(env, PPCInstr_LI(r_zero, 0, env->mode64)); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, r_zero, PPCRH_Reg(r_zero))); |
| return mk_PPCCondCode( Pct_TRUE, Pcf_7EQ ); |
| } |
| |
| /* Not1(...) */ |
| if (e->tag == Iex_Unop && e->Iex.Unop.op == Iop_Not1) { |
| /* Generate code for the arg, and negate the test condition */ |
| PPCCondCode cond = iselCondCode(env, e->Iex.Unop.arg); |
| cond.test = invertCondTest(cond.test); |
| return cond; |
| } |
| |
| /* --- patterns rooted at: 32to1 or 64to1 --- */ |
| |
| /* 32to1, 64to1 */ |
| if (e->tag == Iex_Unop && |
| (e->Iex.Unop.op == Iop_32to1 || e->Iex.Unop.op == Iop_64to1)) { |
| HReg src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| HReg tmp = newVRegI(env); |
| /* could do better, probably -- andi. */ |
| addInstr(env, PPCInstr_Alu(Palu_AND, tmp, |
| src, PPCRH_Imm(False,1))); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, tmp, PPCRH_Imm(False,1))); |
| return mk_PPCCondCode( Pct_TRUE, Pcf_7EQ ); |
| } |
| |
| /* --- patterns rooted at: CmpNEZ8 --- */ |
| |
| /* CmpNEZ8(x) */ |
| /* Note this cloned as CmpNE8(x,0) below. */ |
| /* could do better -- andi. */ |
| if (e->tag == Iex_Unop |
| && e->Iex.Unop.op == Iop_CmpNEZ8) { |
| HReg arg = iselWordExpr_R(env, e->Iex.Unop.arg); |
| HReg tmp = newVRegI(env); |
| addInstr(env, PPCInstr_Alu(Palu_AND, tmp, arg, |
| PPCRH_Imm(False,0xFF))); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, tmp, PPCRH_Imm(False,0))); |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| } |
| |
| /* --- patterns rooted at: CmpNEZ32 --- */ |
| |
| /* CmpNEZ32(x) */ |
| if (e->tag == Iex_Unop |
| && e->Iex.Unop.op == Iop_CmpNEZ32) { |
| HReg r1 = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, r1, PPCRH_Imm(False,0))); |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| } |
| |
| /* --- patterns rooted at: Cmp*32* --- */ |
| |
| /* Cmp*32*(x,y) */ |
| if (e->tag == Iex_Binop |
| && (e->Iex.Binop.op == Iop_CmpEQ32 |
| || e->Iex.Binop.op == Iop_CmpNE32 |
| || e->Iex.Binop.op == Iop_CmpLT32S |
| || e->Iex.Binop.op == Iop_CmpLT32U |
| || e->Iex.Binop.op == Iop_CmpLE32S |
| || e->Iex.Binop.op == Iop_CmpLE32U)) { |
| Bool syned = (e->Iex.Binop.op == Iop_CmpLT32S || |
| e->Iex.Binop.op == Iop_CmpLE32S); |
| HReg r1 = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* ri2 = iselWordExpr_RH(env, syned, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Cmp(syned, True/*32bit cmp*/, |
| 7/*cr*/, r1, ri2)); |
| |
| switch (e->Iex.Binop.op) { |
| case Iop_CmpEQ32: return mk_PPCCondCode( Pct_TRUE, Pcf_7EQ ); |
| case Iop_CmpNE32: return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| case Iop_CmpLT32U: case Iop_CmpLT32S: |
| return mk_PPCCondCode( Pct_TRUE, Pcf_7LT ); |
| case Iop_CmpLE32U: case Iop_CmpLE32S: |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7GT ); |
| default: vpanic("iselCondCode(ppc): CmpXX32"); |
| } |
| } |
| |
| /* --- patterns rooted at: CmpNEZ64 --- */ |
| |
| /* CmpNEZ64 */ |
| if (e->tag == Iex_Unop |
| && e->Iex.Unop.op == Iop_CmpNEZ64) { |
| if (!env->mode64) { |
| HReg hi, lo; |
| HReg tmp = newVRegI(env); |
| iselInt64Expr( &hi, &lo, env, e->Iex.Unop.arg ); |
| addInstr(env, PPCInstr_Alu(Palu_OR, tmp, lo, PPCRH_Reg(hi))); |
| addInstr(env, PPCInstr_Cmp(False/*sign*/, True/*32bit cmp*/, |
| 7/*cr*/, tmp,PPCRH_Imm(False,0))); |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| } else { // mode64 |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Cmp(False/*sign*/, False/*64bit cmp*/, |
| 7/*cr*/, r_src,PPCRH_Imm(False,0))); |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| } |
| } |
| |
| /* --- patterns rooted at: Cmp*64* --- */ |
| |
| /* Cmp*64*(x,y) */ |
| if (e->tag == Iex_Binop |
| && (e->Iex.Binop.op == Iop_CmpEQ64 |
| || e->Iex.Binop.op == Iop_CmpNE64 |
| || e->Iex.Binop.op == Iop_CmpLT64S |
| || e->Iex.Binop.op == Iop_CmpLT64U |
| || e->Iex.Binop.op == Iop_CmpLE64S |
| || e->Iex.Binop.op == Iop_CmpLE64U)) { |
| Bool syned = (e->Iex.Binop.op == Iop_CmpLT64S || |
| e->Iex.Binop.op == Iop_CmpLE64S); |
| HReg r1 = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| PPCRH* ri2 = iselWordExpr_RH(env, syned, e->Iex.Binop.arg2); |
| vassert(env->mode64); |
| addInstr(env, PPCInstr_Cmp(syned, False/*64bit cmp*/, |
| 7/*cr*/, r1, ri2)); |
| |
| switch (e->Iex.Binop.op) { |
| case Iop_CmpEQ64: return mk_PPCCondCode( Pct_TRUE, Pcf_7EQ ); |
| case Iop_CmpNE64: return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| case Iop_CmpLT64U: return mk_PPCCondCode( Pct_TRUE, Pcf_7LT ); |
| case Iop_CmpLE64U: return mk_PPCCondCode( Pct_FALSE, Pcf_7GT ); |
| default: vpanic("iselCondCode(ppc): CmpXX64"); |
| } |
| } |
| |
| /* --- patterns rooted at: CmpNE8 --- */ |
| |
| /* CmpNE8(x,0) */ |
| /* Note this is a direct copy of CmpNEZ8 above. */ |
| /* could do better -- andi. */ |
| if (e->tag == Iex_Binop |
| && e->Iex.Binop.op == Iop_CmpNE8 |
| && isZeroU8(e->Iex.Binop.arg2)) { |
| HReg arg = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg tmp = newVRegI(env); |
| addInstr(env, PPCInstr_Alu(Palu_AND, tmp, arg, |
| PPCRH_Imm(False,0xFF))); |
| addInstr(env, PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, tmp, PPCRH_Imm(False,0))); |
| return mk_PPCCondCode( Pct_FALSE, Pcf_7EQ ); |
| } |
| |
| /* var */ |
| if (e->tag == Iex_RdTmp) { |
| HReg r_src = lookupIRTemp(env, e->Iex.RdTmp.tmp); |
| HReg src_masked = newVRegI(env); |
| addInstr(env, |
| PPCInstr_Alu(Palu_AND, src_masked, |
| r_src, PPCRH_Imm(False,1))); |
| addInstr(env, |
| PPCInstr_Cmp(False/*unsigned*/, True/*32bit cmp*/, |
| 7/*cr*/, src_masked, PPCRH_Imm(False,1))); |
| return mk_PPCCondCode( Pct_TRUE, Pcf_7EQ ); |
| } |
| |
| vex_printf("iselCondCode(ppc): No such tag(%u)\n", e->tag); |
| ppIRExpr(e); |
| vpanic("iselCondCode(ppc)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Integer expressions (128 bit) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* 64-bit mode ONLY: compute a 128-bit value into a register pair, |
| which is returned as the first two parameters. As with |
| iselWordExpr_R, these may be either real or virtual regs; in any |
| case they must not be changed by subsequent code emitted by the |
| caller. */ |
| |
| static void iselInt128Expr ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ) |
| { |
| vassert(env->mode64); |
| iselInt128Expr_wrk(rHi, rLo, env, e); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(*rHi) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(*rHi)); |
| vassert(hregClass(*rLo) == HRcGPR(env->mode64)); |
| vassert(hregIsVirtual(*rLo)); |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static void iselInt128Expr_wrk ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ) |
| { |
| vassert(e); |
| vassert(typeOfIRExpr(env->type_env,e) == Ity_I128); |
| |
| /* read 128-bit IRTemp */ |
| if (e->tag == Iex_RdTmp) { |
| lookupIRTempPair( rHi, rLo, env, e->Iex.RdTmp.tmp); |
| return; |
| } |
| |
| /* --------- BINARY ops --------- */ |
| if (e->tag == Iex_Binop) { |
| switch (e->Iex.Binop.op) { |
| /* 64 x 64 -> 128 multiply */ |
| case Iop_MullU64: |
| case Iop_MullS64: { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| Bool syned = toBool(e->Iex.Binop.op == Iop_MullS64); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_MulL(False/*signedness irrelevant*/, |
| False/*lo64*/, False/*64bit mul*/, |
| tLo, r_srcL, r_srcR)); |
| addInstr(env, PPCInstr_MulL(syned, |
| True/*hi64*/, False/*64bit mul*/, |
| tHi, r_srcL, r_srcR)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* 64HLto128(e1,e2) */ |
| case Iop_64HLto128: |
| *rHi = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| *rLo = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| return; |
| default: |
| break; |
| } |
| } /* if (e->tag == Iex_Binop) */ |
| |
| |
| /* --------- UNARY ops --------- */ |
| if (e->tag == Iex_Unop) { |
| switch (e->Iex.Unop.op) { |
| default: |
| break; |
| } |
| } /* if (e->tag == Iex_Unop) */ |
| |
| vex_printf("iselInt128Expr(ppc64): No such tag(%u)\n", e->tag); |
| ppIRExpr(e); |
| vpanic("iselInt128Expr(ppc64)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Integer expressions (64 bit) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* 32-bit mode ONLY: compute a 128-bit value into a register quad */ |
| static void iselInt128Expr_to_32x4 ( HReg* rHi, HReg* rMedHi, HReg* rMedLo, |
| HReg* rLo, ISelEnv* env, IRExpr* e ) |
| { |
| vassert(!env->mode64); |
| iselInt128Expr_to_32x4_wrk(rHi, rMedHi, rMedLo, rLo, env, e); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(*rHi) == HRcInt32); |
| vassert(hregIsVirtual(*rHi)); |
| vassert(hregClass(*rMedHi) == HRcInt32); |
| vassert(hregIsVirtual(*rMedHi)); |
| vassert(hregClass(*rMedLo) == HRcInt32); |
| vassert(hregIsVirtual(*rMedLo)); |
| vassert(hregClass(*rLo) == HRcInt32); |
| vassert(hregIsVirtual(*rLo)); |
| } |
| |
| static void iselInt128Expr_to_32x4_wrk ( HReg* rHi, HReg* rMedHi, |
| HReg* rMedLo, HReg* rLo, |
| ISelEnv* env, IRExpr* e ) |
| { |
| vassert(e); |
| vassert(typeOfIRExpr(env->type_env,e) == Ity_I128); |
| |
| /* read 128-bit IRTemp */ |
| if (e->tag == Iex_RdTmp) { |
| lookupIRTempQuad( rHi, rMedHi, rMedLo, rLo, env, e->Iex.RdTmp.tmp); |
| return; |
| } |
| |
| if (e->tag == Iex_Binop) { |
| |
| IROp op_binop = e->Iex.Binop.op; |
| switch (op_binop) { |
| case Iop_64HLto128: |
| iselInt64Expr(rHi, rMedHi, env, e->Iex.Binop.arg1); |
| iselInt64Expr(rMedLo, rLo, env, e->Iex.Binop.arg2); |
| return; |
| default: |
| vex_printf("iselInt128Expr_to_32x4_wrk: Binop case 0x%x not found\n", |
| op_binop); |
| break; |
| } |
| } |
| |
| vex_printf("iselInt128Expr_to_32x4_wrk: e->tag 0x%x not found\n", e->tag); |
| return; |
| } |
| |
| /* 32-bit mode ONLY: compute a 64-bit value into a register pair, |
| which is returned as the first two parameters. As with |
| iselIntExpr_R, these may be either real or virtual regs; in any |
| case they must not be changed by subsequent code emitted by the |
| caller. */ |
| |
| static void iselInt64Expr ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ) |
| { |
| vassert(!env->mode64); |
| iselInt64Expr_wrk(rHi, rLo, env, e); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(*rHi) == HRcInt32); |
| vassert(hregIsVirtual(*rHi)); |
| vassert(hregClass(*rLo) == HRcInt32); |
| vassert(hregIsVirtual(*rLo)); |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY ! */ |
| static void iselInt64Expr_wrk ( HReg* rHi, HReg* rLo, |
| ISelEnv* env, IRExpr* e ) |
| { |
| vassert(e); |
| vassert(typeOfIRExpr(env->type_env,e) == Ity_I64); |
| |
| /* 64-bit load */ |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg r_addr = iselWordExpr_R(env, e->Iex.Load.addr); |
| vassert(!env->mode64); |
| addInstr(env, PPCInstr_Load( 4/*byte-load*/, |
| tHi, PPCAMode_IR( 0, r_addr ), |
| False/*32-bit insn please*/) ); |
| addInstr(env, PPCInstr_Load( 4/*byte-load*/, |
| tLo, PPCAMode_IR( 4, r_addr ), |
| False/*32-bit insn please*/) ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* 64-bit literal */ |
| if (e->tag == Iex_Const) { |
| ULong w64 = e->Iex.Const.con->Ico.U64; |
| UInt wHi = ((UInt)(w64 >> 32)) & 0xFFFFFFFF; |
| UInt wLo = ((UInt)w64) & 0xFFFFFFFF; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| vassert(e->Iex.Const.con->tag == Ico_U64); |
| addInstr(env, PPCInstr_LI(tHi, (Long)(Int)wHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_LI(tLo, (Long)(Int)wLo, False/*mode32*/)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* read 64-bit IRTemp */ |
| if (e->tag == Iex_RdTmp) { |
| lookupIRTempPair( rHi, rLo, env, e->Iex.RdTmp.tmp); |
| return; |
| } |
| |
| /* 64-bit GET */ |
| if (e->tag == Iex_Get) { |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(False/*mode32*/) ); |
| PPCAMode* am_addr4 = advance4(env, am_addr); |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| addInstr(env, PPCInstr_Load( 4, tHi, am_addr, False/*mode32*/ )); |
| addInstr(env, PPCInstr_Load( 4, tLo, am_addr4, False/*mode32*/ )); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* 64-bit ITE */ |
| if (e->tag == Iex_ITE) { // VFD |
| HReg e0Lo, e0Hi, eXLo, eXHi; |
| iselInt64Expr(&eXHi, &eXLo, env, e->Iex.ITE.iftrue); |
| iselInt64Expr(&e0Hi, &e0Lo, env, e->Iex.ITE.iffalse); |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| addInstr(env, mk_iMOVds_RR(tHi,e0Hi)); |
| addInstr(env, mk_iMOVds_RR(tLo,e0Lo)); |
| PPCCondCode cc = iselCondCode(env, e->Iex.ITE.cond); |
| addInstr(env, PPCInstr_CMov(cc,tHi,PPCRI_Reg(eXHi))); |
| addInstr(env, PPCInstr_CMov(cc,tLo,PPCRI_Reg(eXLo))); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* --------- BINARY ops --------- */ |
| if (e->tag == Iex_Binop) { |
| IROp op_binop = e->Iex.Binop.op; |
| switch (op_binop) { |
| /* 32 x 32 -> 64 multiply */ |
| case Iop_MullU32: |
| case Iop_MullS32: { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| Bool syned = toBool(op_binop == Iop_MullS32); |
| HReg r_srcL = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg r_srcR = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_MulL(False/*signedness irrelevant*/, |
| False/*lo32*/, True/*32bit mul*/, |
| tLo, r_srcL, r_srcR)); |
| addInstr(env, PPCInstr_MulL(syned, |
| True/*hi32*/, True/*32bit mul*/, |
| tHi, r_srcL, r_srcR)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* Or64/And64/Xor64 */ |
| case Iop_Or64: |
| case Iop_And64: |
| case Iop_Xor64: { |
| HReg xLo, xHi, yLo, yHi; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| PPCAluOp op = (op_binop == Iop_Or64) ? Palu_OR : |
| (op_binop == Iop_And64) ? Palu_AND : Palu_XOR; |
| iselInt64Expr(&xHi, &xLo, env, e->Iex.Binop.arg1); |
| iselInt64Expr(&yHi, &yLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Alu(op, tHi, xHi, PPCRH_Reg(yHi))); |
| addInstr(env, PPCInstr_Alu(op, tLo, xLo, PPCRH_Reg(yLo))); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* Add64 */ |
| case Iop_Add64: { |
| HReg xLo, xHi, yLo, yHi; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| iselInt64Expr(&xHi, &xLo, env, e->Iex.Binop.arg1); |
| iselInt64Expr(&yHi, &yLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_AddSubC( True/*add*/, True /*set carry*/, |
| tLo, xLo, yLo)); |
| addInstr(env, PPCInstr_AddSubC( True/*add*/, False/*read carry*/, |
| tHi, xHi, yHi)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* 32HLto64(e1,e2) */ |
| case Iop_32HLto64: |
| *rHi = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| *rLo = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| return; |
| |
| /* F64toI64[S|U] */ |
| case Iop_F64toI64S: case Iop_F64toI64U: { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, r1 ); |
| HReg fsrc = iselDblExpr(env, e->Iex.Binop.arg2); |
| HReg ftmp = newVRegF(env); |
| |
| vassert(!env->mode64); |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpCftI(False/*F->I*/, False/*int64*/, |
| (op_binop == Iop_F64toI64S) ? True : False, |
| True, ftmp, fsrc)); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, ftmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(4, tHi, zero_r1, False/*mode32*/)); |
| addInstr(env, PPCInstr_Load(4, tLo, four_r1, False/*mode32*/)); |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| case Iop_D64toI64S: { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, r1 ); |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| HReg tmp = newVRegF(env); |
| |
| vassert(!env->mode64); |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DCTFIX, tmp, fr_src)); |
| |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, tmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(4, tHi, zero_r1, False/*mode32*/)); |
| addInstr(env, PPCInstr_Load(4, tLo, four_r1, False/*mode32*/)); |
| add_to_sp( env, 16 ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| case Iop_D128toI64S: { |
| PPCFpOp fpop = Pfp_DCTFIXQ; |
| HReg r_srcHi = newVRegF(env); |
| HReg r_srcLo = newVRegF(env); |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg ftmp = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_DfpD128toD64(fpop, ftmp, r_srcHi, r_srcLo)); |
| |
| // put the D64 result into an integer register pair |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, ftmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(4, tHi, zero_r1, False/*mode32*/)); |
| addInstr(env, PPCInstr_Load(4, tLo, four_r1, False/*mode32*/)); |
| add_to_sp( env, 16 ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| default: |
| break; |
| } |
| } /* if (e->tag == Iex_Binop) */ |
| |
| |
| /* --------- UNARY ops --------- */ |
| if (e->tag == Iex_Unop) { |
| switch (e->Iex.Unop.op) { |
| |
| /* CmpwNEZ64(e) */ |
| case Iop_CmpwNEZ64: { |
| HReg argHi, argLo; |
| HReg tmp1 = newVRegI(env); |
| HReg tmp2 = newVRegI(env); |
| iselInt64Expr(&argHi, &argLo, env, e->Iex.Unop.arg); |
| /* tmp1 = argHi | argLo */ |
| addInstr(env, PPCInstr_Alu(Palu_OR, tmp1, argHi, PPCRH_Reg(argLo))); |
| /* tmp2 = (tmp1 | -tmp1) >>s 31 */ |
| addInstr(env, PPCInstr_Unary(Pun_NEG,tmp2,tmp1)); |
| addInstr(env, PPCInstr_Alu(Palu_OR, tmp2, tmp2, PPCRH_Reg(tmp1))); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| tmp2, tmp2, PPCRH_Imm(False, 31))); |
| *rHi = tmp2; |
| *rLo = tmp2; /* yes, really tmp2 */ |
| return; |
| } |
| |
| /* Left64 */ |
| case Iop_Left64: { |
| HReg argHi, argLo; |
| HReg zero32 = newVRegI(env); |
| HReg resHi = newVRegI(env); |
| HReg resLo = newVRegI(env); |
| iselInt64Expr(&argHi, &argLo, env, e->Iex.Unop.arg); |
| vassert(env->mode64 == False); |
| addInstr(env, PPCInstr_LI(zero32, 0, env->mode64)); |
| /* resHi:resLo = - argHi:argLo */ |
| addInstr(env, PPCInstr_AddSubC( False/*sub*/, True/*set carry*/, |
| resLo, zero32, argLo )); |
| addInstr(env, PPCInstr_AddSubC( False/*sub*/, False/*read carry*/, |
| resHi, zero32, argHi )); |
| /* resHi:resLo |= srcHi:srcLo */ |
| addInstr(env, PPCInstr_Alu(Palu_OR, resLo, resLo, PPCRH_Reg(argLo))); |
| addInstr(env, PPCInstr_Alu(Palu_OR, resHi, resHi, PPCRH_Reg(argHi))); |
| *rHi = resHi; |
| *rLo = resLo; |
| return; |
| } |
| |
| /* 32Sto64(e) */ |
| case Iop_32Sto64: { |
| HReg tHi = newVRegI(env); |
| HReg src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| tHi, src, PPCRH_Imm(False,31))); |
| *rHi = tHi; |
| *rLo = src; |
| return; |
| } |
| case Iop_ExtractExpD64: { |
| HReg tmp = newVRegF(env); |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Unop.arg); |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DXEX, tmp, fr_src)); |
| |
| // put the D64 result into a integer register pair |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, tmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(4, tHi, zero_r1, False/*mode32*/)); |
| addInstr(env, PPCInstr_Load(4, tLo, four_r1, False/*mode32*/)); |
| add_to_sp( env, 16 ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| case Iop_ExtractExpD128: { |
| HReg r_srcHi; |
| HReg r_srcLo; |
| HReg tmp = newVRegF(env); |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_ExtractExpD128(Pfp_DXEXQ, tmp, |
| r_srcHi, r_srcLo)); |
| |
| // put the D64 result into a integer register pair |
| sub_from_sp( env, 16 ); |
| addInstr(env, PPCInstr_FpLdSt(False/*store*/, 8, tmp, zero_r1)); |
| addInstr(env, PPCInstr_Load(4, tHi, zero_r1, False/*mode32*/)); |
| addInstr(env, PPCInstr_Load(4, tLo, four_r1, False/*mode32*/)); |
| add_to_sp( env, 16 ); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* 32Uto64(e) */ |
| case Iop_32Uto64: { |
| HReg tHi = newVRegI(env); |
| HReg tLo = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_LI(tHi, 0, False/*mode32*/)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| case Iop_128to64: { |
| /* Narrow, return the low 64-bit half as a 32-bit |
| * register pair */ |
| HReg r_Hi = INVALID_HREG; |
| HReg r_MedHi = INVALID_HREG; |
| HReg r_MedLo = INVALID_HREG; |
| HReg r_Lo = INVALID_HREG; |
| |
| iselInt128Expr_to_32x4(&r_Hi, &r_MedHi, &r_MedLo, &r_Lo, |
| env, e->Iex.Unop.arg); |
| *rHi = r_MedLo; |
| *rLo = r_Lo; |
| return; |
| } |
| |
| case Iop_128HIto64: { |
| /* Narrow, return the high 64-bit half as a 32-bit |
| * register pair */ |
| HReg r_Hi = INVALID_HREG; |
| HReg r_MedHi = INVALID_HREG; |
| HReg r_MedLo = INVALID_HREG; |
| HReg r_Lo = INVALID_HREG; |
| |
| iselInt128Expr_to_32x4(&r_Hi, &r_MedHi, &r_MedLo, &r_Lo, |
| env, e->Iex.Unop.arg); |
| *rHi = r_Hi; |
| *rLo = r_MedHi; |
| return; |
| } |
| |
| /* V128{HI}to64 */ |
| case Iop_V128HIto64: |
| case Iop_V128to64: { |
| HReg r_aligned16; |
| Int off = e->Iex.Unop.op==Iop_V128HIto64 ? 0 : 8; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg vec = iselVecExpr(env, e->Iex.Unop.arg); |
| PPCAMode *am_off0, *am_offLO, *am_offHI; |
| sub_from_sp( env, 32 ); // Move SP down 32 bytes |
| |
| // get a quadword aligned address within our stack space |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_offHI = PPCAMode_IR( off, r_aligned16 ); |
| am_offLO = PPCAMode_IR( off+4, r_aligned16 ); |
| |
| // store as Vec128 |
| addInstr(env, |
| PPCInstr_AvLdSt( False/*store*/, 16, vec, am_off0 )); |
| |
| // load hi,lo words (of hi/lo half of vec) as Ity_I32's |
| addInstr(env, |
| PPCInstr_Load( 4, tHi, am_offHI, False/*mode32*/ )); |
| addInstr(env, |
| PPCInstr_Load( 4, tLo, am_offLO, False/*mode32*/ )); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| /* could do better than this, but for now ... */ |
| case Iop_1Sto64: { |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| PPCCondCode cond = iselCondCode(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Set(cond,tLo)); |
| addInstr(env, PPCInstr_Shft(Pshft_SHL, True/*32bit shift*/, |
| tLo, tLo, PPCRH_Imm(False,31))); |
| addInstr(env, PPCInstr_Shft(Pshft_SAR, True/*32bit shift*/, |
| tLo, tLo, PPCRH_Imm(False,31))); |
| addInstr(env, mk_iMOVds_RR(tHi, tLo)); |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| case Iop_Not64: { |
| HReg xLo, xHi; |
| HReg tmpLo = newVRegI(env); |
| HReg tmpHi = newVRegI(env); |
| iselInt64Expr(&xHi, &xLo, env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Unary(Pun_NOT,tmpLo,xLo)); |
| addInstr(env, PPCInstr_Unary(Pun_NOT,tmpHi,xHi)); |
| *rHi = tmpHi; |
| *rLo = tmpLo; |
| return; |
| } |
| |
| /* ReinterpF64asI64(e) */ |
| /* Given an IEEE754 double, produce an I64 with the same bit |
| pattern. */ |
| case Iop_ReinterpF64asI64: { |
| PPCAMode *am_addr0, *am_addr1; |
| HReg fr_src = iselDblExpr(env, e->Iex.Unop.arg); |
| HReg r_dstLo = newVRegI(env); |
| HReg r_dstHi = newVRegI(env); |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr0 = PPCAMode_IR( 0, StackFramePtr(False/*mode32*/) ); |
| am_addr1 = PPCAMode_IR( 4, StackFramePtr(False/*mode32*/) ); |
| |
| // store as F64 |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr0 )); |
| |
| // load hi,lo as Ity_I32's |
| addInstr(env, PPCInstr_Load( 4, r_dstHi, |
| am_addr0, False/*mode32*/ )); |
| addInstr(env, PPCInstr_Load( 4, r_dstLo, |
| am_addr1, False/*mode32*/ )); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return; |
| } |
| |
| case Iop_ReinterpD64asI64: { |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Unop.arg); |
| PPCAMode *am_addr0, *am_addr1; |
| HReg r_dstLo = newVRegI(env); |
| HReg r_dstHi = newVRegI(env); |
| |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr0 = PPCAMode_IR( 0, StackFramePtr(False/*mode32*/) ); |
| am_addr1 = PPCAMode_IR( 4, StackFramePtr(False/*mode32*/) ); |
| |
| // store as D64 |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr0 )); |
| |
| // load hi,lo as Ity_I32's |
| addInstr(env, PPCInstr_Load( 4, r_dstHi, |
| am_addr0, False/*mode32*/ )); |
| addInstr(env, PPCInstr_Load( 4, r_dstLo, |
| am_addr1, False/*mode32*/ )); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| |
| add_to_sp( env, 16 ); // Reset SP |
| |
| return; |
| } |
| |
| case Iop_BCDtoDPB: { |
| PPCCondCode cc; |
| UInt argiregs; |
| HReg argregs[2]; |
| Int argreg; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg tmpHi; |
| HReg tmpLo; |
| ULong target; |
| Bool mode64 = env->mode64; |
| |
| argregs[0] = hregPPC_GPR3(mode64); |
| argregs[1] = hregPPC_GPR4(mode64); |
| |
| argiregs = 0; |
| argreg = 0; |
| |
| iselInt64Expr( &tmpHi, &tmpLo, env, e->Iex.Unop.arg ); |
| |
| argiregs |= ( 1 << (argreg+3 ) ); |
| addInstr( env, mk_iMOVds_RR( argregs[argreg++], tmpHi ) ); |
| |
| argiregs |= ( 1 << (argreg+3 ) ); |
| addInstr( env, mk_iMOVds_RR( argregs[argreg], tmpLo ) ); |
| |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| target = toUInt( Ptr_to_ULong(h_calc_BCDtoDPB ) ); |
| |
| addInstr( env, PPCInstr_Call( cc, (Addr64)target, |
| argiregs, |
| mk_RetLoc_simple(RLPri_2Int) ) ); |
| addInstr( env, mk_iMOVds_RR( tHi, argregs[argreg-1] ) ); |
| addInstr( env, mk_iMOVds_RR( tLo, argregs[argreg] ) ); |
| |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| case Iop_DPBtoBCD: { |
| PPCCondCode cc; |
| UInt argiregs; |
| HReg argregs[2]; |
| Int argreg; |
| HReg tLo = newVRegI(env); |
| HReg tHi = newVRegI(env); |
| HReg tmpHi; |
| HReg tmpLo; |
| ULong target; |
| Bool mode64 = env->mode64; |
| |
| argregs[0] = hregPPC_GPR3(mode64); |
| argregs[1] = hregPPC_GPR4(mode64); |
| |
| argiregs = 0; |
| argreg = 0; |
| |
| iselInt64Expr(&tmpHi, &tmpLo, env, e->Iex.Unop.arg); |
| |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg++], tmpHi )); |
| |
| argiregs |= (1 << (argreg+3)); |
| addInstr(env, mk_iMOVds_RR( argregs[argreg], tmpLo)); |
| |
| cc = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| |
| target = toUInt( Ptr_to_ULong( h_calc_DPBtoBCD ) ); |
| |
| addInstr(env, PPCInstr_Call( cc, (Addr64)target, argiregs, |
| mk_RetLoc_simple(RLPri_2Int) ) ); |
| addInstr(env, mk_iMOVds_RR(tHi, argregs[argreg-1])); |
| addInstr(env, mk_iMOVds_RR(tLo, argregs[argreg])); |
| |
| *rHi = tHi; |
| *rLo = tLo; |
| return; |
| } |
| |
| default: |
| break; |
| } |
| } /* if (e->tag == Iex_Unop) */ |
| |
| vex_printf("iselInt64Expr(ppc): No such tag(%u)\n", e->tag); |
| ppIRExpr(e); |
| vpanic("iselInt64Expr(ppc)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Floating point expressions (32 bit) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Nothing interesting here; really just wrappers for |
| 64-bit stuff. */ |
| |
| static HReg iselFltExpr ( ISelEnv* env, IRExpr* e ) |
| { |
| HReg r = iselFltExpr_wrk( env, e ); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(r) == HRcFlt64); /* yes, really Flt64 */ |
| vassert(hregIsVirtual(r)); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static HReg iselFltExpr_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| Bool mode64 = env->mode64; |
| |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(ty == Ity_F32); |
| |
| if (e->tag == Iex_RdTmp) { |
| return lookupIRTemp(env, e->Iex.RdTmp.tmp); |
| } |
| |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| PPCAMode* am_addr; |
| HReg r_dst = newVRegF(env); |
| vassert(e->Iex.Load.ty == Ity_F32); |
| am_addr = iselWordExpr_AMode(env, e->Iex.Load.addr, Ity_F32/*xfer*/); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 4, r_dst, am_addr)); |
| return r_dst; |
| } |
| |
| if (e->tag == Iex_Get) { |
| HReg r_dst = newVRegF(env); |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(env->mode64) ); |
| addInstr(env, PPCInstr_FpLdSt( True/*load*/, 4, r_dst, am_addr )); |
| return r_dst; |
| } |
| |
| if (e->tag == Iex_Unop && e->Iex.Unop.op == Iop_TruncF64asF32) { |
| /* This is quite subtle. The only way to do the relevant |
| truncation is to do a single-precision store and then a |
| double precision load to get it back into a register. The |
| problem is, if the data is then written to memory a second |
| time, as in |
| |
| STbe(...) = TruncF64asF32(...) |
| |
| then will the second truncation further alter the value? The |
| answer is no: flds (as generated here) followed by fsts |
| (generated for the STbe) is the identity function on 32-bit |
| floats, so we are safe. |
| |
| Another upshot of this is that if iselStmt can see the |
| entirety of |
| |
| STbe(...) = TruncF64asF32(arg) |
| |
| then it can short circuit having to deal with TruncF64asF32 |
| individually; instead just compute arg into a 64-bit FP |
| register and do 'fsts' (since that itself does the |
| truncation). |
| |
| We generate pretty poor code here (should be ok both for |
| 32-bit and 64-bit mode); but it is expected that for the most |
| part the latter optimisation will apply and hence this code |
| will not often be used. |
| */ |
| HReg fsrc = iselDblExpr(env, e->Iex.Unop.arg); |
| HReg fdst = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| sub_from_sp( env, 16 ); |
| // store as F32, hence truncating |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 4, |
| fsrc, zero_r1 )); |
| // and reload. Good huh?! (sigh) |
| addInstr(env, PPCInstr_FpLdSt( True/*load*/, 4, |
| fdst, zero_r1 )); |
| add_to_sp( env, 16 ); |
| return fdst; |
| } |
| |
| if (e->tag == Iex_Binop && e->Iex.Binop.op == Iop_I64UtoF32) { |
| if (mode64) { |
| HReg fdst = newVRegF(env); |
| HReg isrc = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| |
| addInstr(env, PPCInstr_Store(8, zero_r1, isrc, True/*mode64*/)); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fdst, zero_r1)); |
| addInstr(env, PPCInstr_FpCftI(True/*I->F*/, False/*int64*/, |
| False, False, |
| fdst, fdst)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return fdst; |
| } else { |
| /* 32-bit mode */ |
| HReg fdst = newVRegF(env); |
| HReg isrcHi, isrcLo; |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, r1 ); |
| |
| iselInt64Expr(&isrcHi, &isrcLo, env, e->Iex.Binop.arg2); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| |
| addInstr(env, PPCInstr_Store(4, zero_r1, isrcHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, isrcLo, False/*mode32*/)); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fdst, zero_r1)); |
| addInstr(env, PPCInstr_FpCftI(True/*I->F*/, False/*int64*/, |
| False, False, |
| fdst, fdst)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return fdst; |
| } |
| |
| } |
| |
| vex_printf("iselFltExpr(ppc): No such tag(%u)\n", e->tag); |
| ppIRExpr(e); |
| vpanic("iselFltExpr_wrk(ppc)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Floating point expressions (64 bit) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Compute a 64-bit floating point value into a register, the identity |
| of which is returned. As with iselIntExpr_R, the reg may be either |
| real or virtual; in any case it must not be changed by subsequent |
| code emitted by the caller. */ |
| |
| /* IEEE 754 formats. From http://www.freesoft.org/CIE/RFC/1832/32.htm: |
| |
| Type S (1 bit) E (11 bits) F (52 bits) |
| ---- --------- ----------- ----------- |
| signalling NaN u 2047 (max) .0uuuuu---u |
| (with at least |
| one 1 bit) |
| quiet NaN u 2047 (max) .1uuuuu---u |
| |
| negative infinity 1 2047 (max) .000000---0 |
| |
| positive infinity 0 2047 (max) .000000---0 |
| |
| negative zero 1 0 .000000---0 |
| |
| positive zero 0 0 .000000---0 |
| */ |
| |
| static HReg iselDblExpr ( ISelEnv* env, IRExpr* e ) |
| { |
| HReg r = iselDblExpr_wrk( env, e ); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(r) == HRcFlt64); |
| vassert(hregIsVirtual(r)); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static HReg iselDblExpr_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| Bool mode64 = env->mode64; |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(e); |
| vassert(ty == Ity_F64); |
| |
| if (e->tag == Iex_RdTmp) { |
| return lookupIRTemp(env, e->Iex.RdTmp.tmp); |
| } |
| |
| /* --------- LITERAL --------- */ |
| if (e->tag == Iex_Const) { |
| union { UInt u32x2[2]; ULong u64; Double f64; } u; |
| vassert(sizeof(u) == 8); |
| vassert(sizeof(u.u64) == 8); |
| vassert(sizeof(u.f64) == 8); |
| vassert(sizeof(u.u32x2) == 8); |
| |
| if (e->Iex.Const.con->tag == Ico_F64) { |
| u.f64 = e->Iex.Const.con->Ico.F64; |
| } |
| else if (e->Iex.Const.con->tag == Ico_F64i) { |
| u.u64 = e->Iex.Const.con->Ico.F64i; |
| } |
| else |
| vpanic("iselDblExpr(ppc): const"); |
| |
| if (!mode64) { |
| HReg r_srcHi = newVRegI(env); |
| HReg r_srcLo = newVRegI(env); |
| addInstr(env, PPCInstr_LI(r_srcHi, u.u32x2[0], mode64)); |
| addInstr(env, PPCInstr_LI(r_srcLo, u.u32x2[1], mode64)); |
| return mk_LoadRR32toFPR( env, r_srcHi, r_srcLo ); |
| } else { // mode64 |
| HReg r_src = newVRegI(env); |
| addInstr(env, PPCInstr_LI(r_src, u.u64, mode64)); |
| return mk_LoadR64toFPR( env, r_src ); // 1*I64 -> F64 |
| } |
| } |
| |
| /* --------- LOAD --------- */ |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| HReg r_dst = newVRegF(env); |
| PPCAMode* am_addr; |
| vassert(e->Iex.Load.ty == Ity_F64); |
| am_addr = iselWordExpr_AMode(env, e->Iex.Load.addr, Ity_F64/*xfer*/); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_dst, am_addr)); |
| return r_dst; |
| } |
| |
| /* --------- GET --------- */ |
| if (e->tag == Iex_Get) { |
| HReg r_dst = newVRegF(env); |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(mode64) ); |
| addInstr(env, PPCInstr_FpLdSt( True/*load*/, 8, r_dst, am_addr )); |
| return r_dst; |
| } |
| |
| /* --------- OPS --------- */ |
| if (e->tag == Iex_Qop) { |
| PPCFpOp fpop = Pfp_INVALID; |
| switch (e->Iex.Qop.details->op) { |
| case Iop_MAddF64: fpop = Pfp_MADDD; break; |
| case Iop_MAddF64r32: fpop = Pfp_MADDS; break; |
| case Iop_MSubF64: fpop = Pfp_MSUBD; break; |
| case Iop_MSubF64r32: fpop = Pfp_MSUBS; break; |
| default: break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg r_dst = newVRegF(env); |
| HReg r_srcML = iselDblExpr(env, e->Iex.Qop.details->arg2); |
| HReg r_srcMR = iselDblExpr(env, e->Iex.Qop.details->arg3); |
| HReg r_srcAcc = iselDblExpr(env, e->Iex.Qop.details->arg4); |
| set_FPU_rounding_mode( env, e->Iex.Qop.details->arg1 ); |
| addInstr(env, PPCInstr_FpMulAcc(fpop, r_dst, |
| r_srcML, r_srcMR, r_srcAcc)); |
| return r_dst; |
| } |
| } |
| |
| if (e->tag == Iex_Triop) { |
| IRTriop *triop = e->Iex.Triop.details; |
| PPCFpOp fpop = Pfp_INVALID; |
| switch (triop->op) { |
| case Iop_AddF64: fpop = Pfp_ADDD; break; |
| case Iop_SubF64: fpop = Pfp_SUBD; break; |
| case Iop_MulF64: fpop = Pfp_MULD; break; |
| case Iop_DivF64: fpop = Pfp_DIVD; break; |
| case Iop_AddF64r32: fpop = Pfp_ADDS; break; |
| case Iop_SubF64r32: fpop = Pfp_SUBS; break; |
| case Iop_MulF64r32: fpop = Pfp_MULS; break; |
| case Iop_DivF64r32: fpop = Pfp_DIVS; break; |
| default: break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg r_dst = newVRegF(env); |
| HReg r_srcL = iselDblExpr(env, triop->arg2); |
| HReg r_srcR = iselDblExpr(env, triop->arg3); |
| set_FPU_rounding_mode( env, triop->arg1 ); |
| addInstr(env, PPCInstr_FpBinary(fpop, r_dst, r_srcL, r_srcR)); |
| return r_dst; |
| } |
| } |
| |
| if (e->tag == Iex_Binop) { |
| PPCFpOp fpop = Pfp_INVALID; |
| switch (e->Iex.Binop.op) { |
| case Iop_SqrtF64: fpop = Pfp_SQRT; break; |
| default: break; |
| } |
| if (fpop == Pfp_SQRT) { |
| HReg fr_dst = newVRegF(env); |
| HReg fr_src = iselDblExpr(env, e->Iex.Binop.arg2); |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_FpUnary(fpop, fr_dst, fr_src)); |
| return fr_dst; |
| } |
| } |
| |
| if (e->tag == Iex_Binop) { |
| |
| if (e->Iex.Binop.op == Iop_RoundF64toF32) { |
| HReg r_dst = newVRegF(env); |
| HReg r_src = iselDblExpr(env, e->Iex.Binop.arg2); |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_FpRSP(r_dst, r_src)); |
| //set_FPU_rounding_default( env ); |
| return r_dst; |
| } |
| |
| if (e->Iex.Binop.op == Iop_I64StoF64 || e->Iex.Binop.op == Iop_I64UtoF64) { |
| if (mode64) { |
| HReg fdst = newVRegF(env); |
| HReg isrc = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| |
| addInstr(env, PPCInstr_Store(8, zero_r1, isrc, True/*mode64*/)); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fdst, zero_r1)); |
| addInstr(env, PPCInstr_FpCftI(True/*I->F*/, False/*int64*/, |
| e->Iex.Binop.op == Iop_I64StoF64, |
| True/*fdst is 64 bit*/, |
| fdst, fdst)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return fdst; |
| } else { |
| /* 32-bit mode */ |
| HReg fdst = newVRegF(env); |
| HReg isrcHi, isrcLo; |
| HReg r1 = StackFramePtr(env->mode64); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, r1 ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, r1 ); |
| |
| iselInt64Expr(&isrcHi, &isrcLo, env, e->Iex.Binop.arg2); |
| |
| /* Set host rounding mode */ |
| set_FPU_rounding_mode( env, e->Iex.Binop.arg1 ); |
| |
| sub_from_sp( env, 16 ); |
| |
| addInstr(env, PPCInstr_Store(4, zero_r1, isrcHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, isrcLo, False/*mode32*/)); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fdst, zero_r1)); |
| addInstr(env, PPCInstr_FpCftI(True/*I->F*/, False/*int64*/, |
| e->Iex.Binop.op == Iop_I64StoF64, |
| True/*fdst is 64 bit*/, |
| fdst, fdst)); |
| |
| add_to_sp( env, 16 ); |
| |
| ///* Restore default FPU rounding. */ |
| //set_FPU_rounding_default( env ); |
| return fdst; |
| } |
| } |
| |
| } |
| |
| if (e->tag == Iex_Unop) { |
| PPCFpOp fpop = Pfp_INVALID; |
| switch (e->Iex.Unop.op) { |
| case Iop_NegF64: fpop = Pfp_NEG; break; |
| case Iop_AbsF64: fpop = Pfp_ABS; break; |
| case Iop_Est5FRSqrt: fpop = Pfp_RSQRTE; break; |
| case Iop_RoundF64toF64_NegINF: fpop = Pfp_FRIM; break; |
| case Iop_RoundF64toF64_PosINF: fpop = Pfp_FRIP; break; |
| case Iop_RoundF64toF64_NEAREST: fpop = Pfp_FRIN; break; |
| case Iop_RoundF64toF64_ZERO: fpop = Pfp_FRIZ; break; |
| default: break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg fr_dst = newVRegF(env); |
| HReg fr_src = iselDblExpr(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_FpUnary(fpop, fr_dst, fr_src)); |
| return fr_dst; |
| } |
| } |
| |
| if (e->tag == Iex_Unop) { |
| switch (e->Iex.Unop.op) { |
| case Iop_ReinterpI64asF64: { |
| /* Given an I64, produce an IEEE754 double with the same |
| bit pattern. */ |
| if (!mode64) { |
| HReg r_srcHi, r_srcLo; |
| iselInt64Expr( &r_srcHi, &r_srcLo, env, e->Iex.Unop.arg); |
| return mk_LoadRR32toFPR( env, r_srcHi, r_srcLo ); |
| } else { |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| return mk_LoadR64toFPR( env, r_src ); |
| } |
| } |
| |
| case Iop_F32toF64: { |
| if (e->Iex.Unop.arg->tag == Iex_Unop && |
| e->Iex.Unop.arg->Iex.Unop.op == Iop_ReinterpI32asF32 ) { |
| e = e->Iex.Unop.arg; |
| |
| HReg src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| HReg fr_dst = newVRegF(env); |
| PPCAMode *am_addr; |
| |
| sub_from_sp( env, 16 ); // Move SP down 16 bytes |
| am_addr = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| // store src as Ity_I32's |
| addInstr(env, PPCInstr_Store( 4, am_addr, src, env->mode64 )); |
| |
| // load single precision float, but the end results loads into a |
| // 64-bit FP register -- i.e., F64. |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 4, fr_dst, am_addr)); |
| |
| add_to_sp( env, 16 ); // Reset SP |
| return fr_dst; |
| } |
| |
| |
| /* this is a no-op */ |
| HReg res = iselFltExpr(env, e->Iex.Unop.arg); |
| return res; |
| } |
| default: |
| break; |
| } |
| } |
| |
| /* --------- MULTIPLEX --------- */ |
| if (e->tag == Iex_ITE) { // VFD |
| if (ty == Ity_F64 |
| && typeOfIRExpr(env->type_env,e->Iex.ITE.cond) == Ity_I1) { |
| HReg fr1 = iselDblExpr(env, e->Iex.ITE.iftrue); |
| HReg fr0 = iselDblExpr(env, e->Iex.ITE.iffalse); |
| HReg fr_dst = newVRegF(env); |
| addInstr(env, PPCInstr_FpUnary( Pfp_MOV, fr_dst, fr0 )); |
| PPCCondCode cc = iselCondCode(env, e->Iex.ITE.cond); |
| addInstr(env, PPCInstr_FpCMov( cc, fr_dst, fr1 )); |
| return fr_dst; |
| } |
| } |
| |
| vex_printf("iselDblExpr(ppc): No such tag(%u)\n", e->tag); |
| ppIRExpr(e); |
| vpanic("iselDblExpr_wrk(ppc)"); |
| } |
| |
| static HReg iselDfp32Expr(ISelEnv* env, IRExpr* e) |
| { |
| HReg r = iselDfp32Expr_wrk( env, e ); |
| vassert(hregClass(r) == HRcFlt64); |
| vassert( hregIsVirtual(r) ); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static HReg iselDfp32Expr_wrk(ISelEnv* env, IRExpr* e) |
| { |
| Bool mode64 = env->mode64; |
| IRType ty = typeOfIRExpr( env->type_env, e ); |
| |
| vassert( e ); |
| vassert( ty == Ity_D32 ); |
| |
| /* --------- GET --------- */ |
| if (e->tag == Iex_Get) { |
| HReg r_dst = newVRegF( env ); |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(mode64) ); |
| addInstr( env, PPCInstr_FpLdSt( True/*load*/, 8, r_dst, am_addr ) ); |
| return r_dst; |
| } |
| |
| /* --------- LOAD --------- */ |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| PPCAMode* am_addr; |
| HReg r_dst = newVRegF(env); |
| vassert(e->Iex.Load.ty == Ity_D32); |
| am_addr = iselWordExpr_AMode(env, e->Iex.Load.addr, Ity_D32/*xfer*/); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 4, r_dst, am_addr)); |
| return r_dst; |
| } |
| |
| /* --------- OPS --------- */ |
| if (e->tag == Iex_Binop) { |
| if (e->Iex.Binop.op == Iop_D64toD32) { |
| HReg fr_dst = newVRegF(env); |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DRSP, fr_dst, fr_src)); |
| return fr_dst; |
| } |
| } |
| |
| ppIRExpr( e ); |
| vpanic( "iselDfp32Expr_wrk(ppc)" ); |
| } |
| |
| static HReg iselDfp64Expr(ISelEnv* env, IRExpr* e) |
| { |
| HReg r = iselDfp64Expr_wrk( env, e ); |
| vassert(hregClass(r) == HRcFlt64); |
| vassert( hregIsVirtual(r) ); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static HReg iselDfp64Expr_wrk(ISelEnv* env, IRExpr* e) |
| { |
| Bool mode64 = env->mode64; |
| IRType ty = typeOfIRExpr( env->type_env, e ); |
| HReg r_dstHi, r_dstLo; |
| |
| vassert( e ); |
| vassert( ty == Ity_D64 ); |
| |
| if (e->tag == Iex_RdTmp) { |
| return lookupIRTemp( env, e->Iex.RdTmp.tmp ); |
| } |
| |
| /* --------- GET --------- */ |
| if (e->tag == Iex_Get) { |
| HReg r_dst = newVRegF( env ); |
| PPCAMode* am_addr = PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(mode64) ); |
| addInstr( env, PPCInstr_FpLdSt( True/*load*/, 8, r_dst, am_addr ) ); |
| return r_dst; |
| } |
| |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| PPCAMode* am_addr; |
| HReg r_dst = newVRegF(env); |
| vassert(e->Iex.Load.ty == Ity_D64); |
| am_addr = iselWordExpr_AMode(env, e->Iex.Load.addr, Ity_D64/*xfer*/); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_dst, am_addr)); |
| return r_dst; |
| } |
| |
| /* --------- OPS --------- */ |
| if (e->tag == Iex_Qop) { |
| HReg r_dst = newVRegF( env ); |
| return r_dst; |
| } |
| |
| if (e->tag == Iex_Unop) { |
| HReg fr_dst = newVRegF(env); |
| switch (e->Iex.Unop.op) { |
| case Iop_ReinterpI64asD64: { |
| /* Given an I64, produce an IEEE754 DFP with the same |
| bit pattern. */ |
| if (!mode64) { |
| HReg r_srcHi, r_srcLo; |
| iselInt64Expr( &r_srcHi, &r_srcLo, env, e->Iex.Unop.arg); |
| return mk_LoadRR32toFPR( env, r_srcHi, r_srcLo ); |
| } else { |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| return mk_LoadR64toFPR( env, r_src ); |
| } |
| } |
| case Iop_D32toD64: { |
| HReg fr_src = iselDfp32Expr(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_DCTDP, fr_dst, fr_src)); |
| return fr_dst; |
| } |
| case Iop_D128HItoD64: |
| iselDfp128Expr( &r_dstHi, &r_dstLo, env, e->Iex.Unop.arg ); |
| return r_dstHi; |
| case Iop_D128LOtoD64: |
| iselDfp128Expr( &r_dstHi, &r_dstLo, env, e->Iex.Unop.arg ); |
| return r_dstLo; |
| case Iop_InsertExpD64: { |
| HReg fr_srcL = iselDblExpr(env, e->Iex.Binop.arg1); |
| HReg fr_srcR = iselDblExpr(env, e->Iex.Binop.arg2); |
| |
| addInstr(env, PPCInstr_Dfp64Binary(Pfp_DIEX, fr_dst, fr_srcL, |
| fr_srcR)); |
| return fr_dst; |
| } |
| default: |
| vex_printf( "ERROR: iselDfp64Expr_wrk, UNKNOWN unop case %d\n", |
| e->Iex.Unop.op ); |
| } |
| } |
| |
| if (e->tag == Iex_Binop) { |
| PPCFpOp fpop = Pfp_INVALID; |
| HReg fr_dst = newVRegF(env); |
| |
| switch (e->Iex.Binop.op) { |
| case Iop_D128toD64: fpop = Pfp_DRDPQ; break; |
| case Iop_D64toD32: fpop = Pfp_DRSP; break; |
| case Iop_I64StoD64: fpop = Pfp_DCFFIX; break; |
| case Iop_RoundD64toInt: fpop = Pfp_DRINTN; break; |
| default: break; |
| } |
| if (fpop == Pfp_DRDPQ) { |
| HReg r_srcHi = newVRegF(env); |
| HReg r_srcLo = newVRegF(env); |
| |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_DfpD128toD64(fpop, fr_dst, r_srcHi, r_srcLo)); |
| return fr_dst; |
| |
| } else if (fpop == Pfp_DRINTN) { |
| HReg fr_src = newVRegF(env); |
| PPCRI* r_rmc = iselWordExpr_RI(env, e->Iex.Binop.arg1); |
| |
| /* NOTE, this IOP takes a DFP value and rounds to the |
| * neares floating point integer value, i.e. fractional part |
| * is zero. The result is a decimal floating point number. |
| * the INT in the name is a bit misleading. |
| */ |
| fr_src = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_DfpRound(fr_dst, fr_src, r_rmc)); |
| return fr_dst; |
| |
| } else if (fpop == Pfp_DRSP) { |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| addInstr(env, PPCInstr_Dfp64Unary(fpop, fr_dst, fr_src)); |
| return fr_dst; |
| |
| } else if (fpop == Pfp_DCFFIX) { |
| HReg fr_src = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| sub_from_sp( env, 16 ); |
| |
| // put the I64 value into a floating point register |
| if (mode64) { |
| HReg tmp = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| |
| addInstr(env, PPCInstr_Store(8, zero_r1, tmp, True/*mode64*/)); |
| } else { |
| HReg tmpHi, tmpLo; |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| iselInt64Expr(&tmpHi, &tmpLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Store(4, zero_r1, tmpHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, tmpLo, False/*mode32*/)); |
| } |
| |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_src, zero_r1)); |
| addInstr(env, PPCInstr_Dfp64Unary(fpop, fr_dst, fr_src)); |
| add_to_sp( env, 16 ); |
| return fr_dst; |
| } |
| |
| switch (e->Iex.Binop.op) { |
| /* shift instructions D64, I32 -> D64 */ |
| case Iop_ShlD64: fpop = Pfp_DSCLI; break; |
| case Iop_ShrD64: fpop = Pfp_DSCRI; break; |
| default: break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg fr_src = iselDfp64Expr(env, e->Iex.Binop.arg1); |
| PPCRI* shift = iselWordExpr_RI(env, e->Iex.Binop.arg2); |
| |
| /* shift value must be an immediate value */ |
| vassert(shift->tag == Pri_Imm); |
| |
| addInstr(env, PPCInstr_DfpShift(fpop, fr_dst, fr_src, shift)); |
| return fr_dst; |
| } |
| |
| switch (e->Iex.Binop.op) { |
| case Iop_InsertExpD64: |
| fpop = Pfp_DIEX; |
| break; |
| default: break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg fr_srcL = newVRegF(env); |
| HReg fr_srcR = iselDfp64Expr(env, e->Iex.Binop.arg2); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| sub_from_sp( env, 16 ); |
| |
| if (env->mode64) { |
| // put the I64 value into a floating point reg |
| HReg tmp = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| |
| addInstr(env, PPCInstr_Store(8, zero_r1, tmp, True/*mode64*/)); |
| } else { |
| // put the I64 register pair into a floating point reg |
| HReg tmpHi; |
| HReg tmpLo; |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| iselInt64Expr(&tmpHi, &tmpLo, env, e->Iex.Binop.arg1); |
| addInstr(env, PPCInstr_Store(4, zero_r1, tmpHi, False/*!mode64*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, tmpLo, False/*!mode64*/)); |
| } |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_srcL, zero_r1)); |
| addInstr(env, PPCInstr_Dfp64Binary(fpop, fr_dst, fr_srcL, |
| fr_srcR)); |
| add_to_sp( env, 16 ); |
| return fr_dst; |
| } |
| } |
| |
| if (e->tag == Iex_Triop) { |
| IRTriop *triop = e->Iex.Triop.details; |
| PPCFpOp fpop = Pfp_INVALID; |
| |
| switch (triop->op) { |
| case Iop_AddD64: |
| fpop = Pfp_DFPADD; |
| break; |
| case Iop_SubD64: |
| fpop = Pfp_DFPSUB; |
| break; |
| case Iop_MulD64: |
| fpop = Pfp_DFPMUL; |
| break; |
| case Iop_DivD64: |
| fpop = Pfp_DFPDIV; |
| break; |
| default: |
| break; |
| } |
| if (fpop != Pfp_INVALID) { |
| HReg r_dst = newVRegF( env ); |
| HReg r_srcL = iselDfp64Expr( env, triop->arg2 ); |
| HReg r_srcR = iselDfp64Expr( env, triop->arg3 ); |
| |
| set_FPU_DFP_rounding_mode( env, triop->arg1 ); |
| addInstr( env, PPCInstr_Dfp64Binary( fpop, r_dst, r_srcL, r_srcR ) ); |
| return r_dst; |
| } |
| |
| switch (triop->op) { |
| case Iop_QuantizeD64: fpop = Pfp_DQUA; break; |
| case Iop_SignificanceRoundD64: fpop = Pfp_RRDTR; break; |
| default: break; |
| } |
| if (fpop == Pfp_DQUA) { |
| HReg r_dst = newVRegF(env); |
| HReg r_srcL = iselDfp64Expr(env, triop->arg2); |
| HReg r_srcR = iselDfp64Expr(env, triop->arg3); |
| PPCRI* rmc = iselWordExpr_RI(env, triop->arg1); |
| addInstr(env, PPCInstr_DfpQuantize(fpop, r_dst, r_srcL, r_srcR, |
| rmc)); |
| return r_dst; |
| |
| } else if (fpop == Pfp_RRDTR) { |
| HReg r_dst = newVRegF(env); |
| HReg r_srcL = newVRegF(env); |
| HReg r_srcR = iselDfp64Expr(env, triop->arg3); |
| PPCRI* rmc = iselWordExpr_RI(env, triop->arg1); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| HReg i8_val = iselWordExpr_R(env, triop->arg2); |
| |
| /* Move I8 to float register to issue instruction */ |
| sub_from_sp( env, 16 ); |
| if (mode64) |
| addInstr(env, PPCInstr_Store(8, zero_r1, i8_val, True/*mode64*/)); |
| else |
| addInstr(env, PPCInstr_Store(4, zero_r1, i8_val, False/*mode32*/)); |
| |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_srcL, zero_r1)); |
| add_to_sp( env, 16 ); |
| |
| // will set TE and RMC when issuing instruction |
| addInstr(env, PPCInstr_DfpQuantize(fpop, r_dst, r_srcL, r_srcR, rmc)); |
| return r_dst; |
| } |
| } |
| |
| ppIRExpr( e ); |
| vpanic( "iselDfp64Expr_wrk(ppc)" ); |
| } |
| |
| static void iselDfp128Expr(HReg* rHi, HReg* rLo, ISelEnv* env, IRExpr* e) |
| { |
| iselDfp128Expr_wrk( rHi, rLo, env, e ); |
| vassert( hregIsVirtual(*rHi) ); |
| vassert( hregIsVirtual(*rLo) ); |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static void iselDfp128Expr_wrk(HReg* rHi, HReg *rLo, ISelEnv* env, IRExpr* e) |
| { |
| vassert( e ); |
| vassert( typeOfIRExpr(env->type_env,e) == Ity_D128 ); |
| |
| /* read 128-bit IRTemp */ |
| if (e->tag == Iex_RdTmp) { |
| lookupIRTempPair( rHi, rLo, env, e->Iex.RdTmp.tmp ); |
| return; |
| } |
| |
| if (e->tag == Iex_Unop) { |
| HReg r_dstHi = newVRegF(env); |
| HReg r_dstLo = newVRegF(env); |
| |
| if (e->Iex.Unop.op == Iop_I64StoD128) { |
| HReg fr_src = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| |
| // put the I64 value into a floating point reg |
| if (env->mode64) { |
| HReg tmp = iselWordExpr_R(env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Store(8, zero_r1, tmp, True/*mode64*/)); |
| } else { |
| HReg tmpHi, tmpLo; |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| iselInt64Expr(&tmpHi, &tmpLo, env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Store(4, zero_r1, tmpHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, tmpLo, False/*mode32*/)); |
| } |
| |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_src, zero_r1)); |
| addInstr(env, PPCInstr_DfpI64StoD128(Pfp_DCFFIXQ, r_dstHi, r_dstLo, |
| fr_src)); |
| } |
| |
| if (e->Iex.Unop.op == Iop_D64toD128) { |
| HReg r_src = iselDfp64Expr(env, e->Iex.Unop.arg); |
| |
| /* Source is 64bit, result is 128 bit. High 64bit source arg, |
| * is ignored by the instruction. Set high arg to r_src just |
| * to meet the vassert tests. |
| */ |
| addInstr(env, PPCInstr_Dfp128Unary(Pfp_DCTQPQ, r_dstHi, r_dstLo, |
| r_src, r_src)); |
| } |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| } |
| |
| /* --------- OPS --------- */ |
| if (e->tag == Iex_Binop) { |
| HReg r_srcHi; |
| HReg r_srcLo; |
| |
| switch (e->Iex.Binop.op) { |
| case Iop_D64HLtoD128: |
| r_srcHi = iselDfp64Expr( env, e->Iex.Binop.arg1 ); |
| r_srcLo = iselDfp64Expr( env, e->Iex.Binop.arg2 ); |
| *rHi = r_srcHi; |
| *rLo = r_srcLo; |
| return; |
| break; |
| case Iop_D128toD64: { |
| PPCFpOp fpop = Pfp_DRDPQ; |
| HReg fr_dst = newVRegF(env); |
| |
| set_FPU_DFP_rounding_mode( env, e->Iex.Binop.arg1 ); |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_DfpD128toD64(fpop, fr_dst, r_srcHi, r_srcLo)); |
| |
| /* Need to meet the interface spec but the result is |
| * just 64-bits so send the result back in both halfs. |
| */ |
| *rHi = fr_dst; |
| *rLo = fr_dst; |
| return; |
| } |
| case Iop_ShlD128: |
| case Iop_ShrD128: { |
| HReg fr_dst_hi = newVRegF(env); |
| HReg fr_dst_lo = newVRegF(env); |
| PPCRI* shift = iselWordExpr_RI(env, e->Iex.Binop.arg2); |
| PPCFpOp fpop = Pfp_DSCLIQ; /* fix later if necessary */ |
| |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg1); |
| |
| if (e->Iex.Binop.op == Iop_ShrD128) |
| fpop = Pfp_DSCRIQ; |
| |
| addInstr(env, PPCInstr_DfpShift128(fpop, fr_dst_hi, fr_dst_lo, |
| r_srcHi, r_srcLo, shift)); |
| |
| *rHi = fr_dst_hi; |
| *rLo = fr_dst_lo; |
| return; |
| } |
| case Iop_RoundD128toInt: { |
| HReg r_dstHi = newVRegF(env); |
| HReg r_dstLo = newVRegF(env); |
| PPCRI* r_rmc = iselWordExpr_RI(env, e->Iex.Binop.arg1); |
| |
| // will set R and RMC when issuing instruction |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| |
| addInstr(env, PPCInstr_DfpRound128(r_dstHi, r_dstLo, |
| r_srcHi, r_srcLo, r_rmc)); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| } |
| case Iop_InsertExpD128: { |
| HReg r_dstHi = newVRegF(env); |
| HReg r_dstLo = newVRegF(env); |
| HReg r_srcL = newVRegF(env); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| r_srcHi = newVRegF(env); |
| r_srcLo = newVRegF(env); |
| |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, e->Iex.Binop.arg2); |
| |
| /* Move I64 to float register to issue instruction */ |
| if (env->mode64) { |
| HReg tmp = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| addInstr(env, PPCInstr_Store(8, zero_r1, tmp, True/*mode64*/)); |
| } else { |
| HReg tmpHi, tmpLo; |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| |
| iselInt64Expr(&tmpHi, &tmpLo, env, e->Iex.Unop.arg); |
| addInstr(env, PPCInstr_Store(4, zero_r1, tmpHi, False/*mode32*/)); |
| addInstr(env, PPCInstr_Store(4, four_r1, tmpLo, False/*mode32*/)); |
| } |
| |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_srcL, zero_r1)); |
| addInstr(env, PPCInstr_InsertExpD128(Pfp_DIEXQ, |
| r_dstHi, r_dstLo, |
| r_srcL, r_srcHi, r_srcLo)); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| } |
| default: |
| vex_printf( "ERROR: iselDfp128Expr_wrk, UNKNOWN binop case %d\n", |
| e->Iex.Binop.op ); |
| break; |
| } |
| } |
| |
| if (e->tag == Iex_Triop) { |
| IRTriop *triop = e->Iex.Triop.details; |
| PPCFpOp fpop = Pfp_INVALID; |
| HReg r_dstHi = newVRegF(env); |
| HReg r_dstLo = newVRegF(env); |
| |
| switch (triop->op) { |
| case Iop_AddD128: |
| fpop = Pfp_DFPADDQ; |
| break; |
| case Iop_SubD128: |
| fpop = Pfp_DFPSUBQ; |
| break; |
| case Iop_MulD128: |
| fpop = Pfp_DFPMULQ; |
| break; |
| case Iop_DivD128: |
| fpop = Pfp_DFPDIVQ; |
| break; |
| default: |
| break; |
| } |
| |
| if (fpop != Pfp_INVALID) { |
| HReg r_srcRHi = newVRegV( env ); |
| HReg r_srcRLo = newVRegV( env ); |
| |
| /* dst will be used to pass in the left operand and get the result. */ |
| iselDfp128Expr( &r_dstHi, &r_dstLo, env, triop->arg2 ); |
| iselDfp128Expr( &r_srcRHi, &r_srcRLo, env, triop->arg3 ); |
| set_FPU_DFP_rounding_mode( env, triop->arg1 ); |
| addInstr( env, |
| PPCInstr_Dfp128Binary( fpop, r_dstHi, r_dstLo, |
| r_srcRHi, r_srcRLo ) ); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| } |
| switch (triop->op) { |
| case Iop_QuantizeD128: fpop = Pfp_DQUAQ; break; |
| case Iop_SignificanceRoundD128: fpop = Pfp_DRRNDQ; break; |
| default: break; |
| } |
| if (fpop == Pfp_DQUAQ) { |
| HReg r_srcHi = newVRegF(env); |
| HReg r_srcLo = newVRegF(env); |
| PPCRI* rmc = iselWordExpr_RI(env, triop->arg1); |
| |
| /* dst will be used to pass in the left operand and get the result */ |
| iselDfp128Expr(&r_dstHi, &r_dstLo, env, triop->arg2); |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, triop->arg3); |
| |
| // will set RMC when issuing instruction |
| addInstr(env, PPCInstr_DfpQuantize128(fpop, r_dstHi, r_dstLo, |
| r_srcHi, r_srcLo, rmc)); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| |
| } else if (fpop == Pfp_DRRNDQ) { |
| HReg r_srcHi = newVRegF(env); |
| HReg r_srcLo = newVRegF(env); |
| PPCRI* rmc = iselWordExpr_RI(env, triop->arg1); |
| PPCAMode* zero_r1 = PPCAMode_IR( 0, StackFramePtr(env->mode64) ); |
| PPCAMode* four_r1 = PPCAMode_IR( 4, StackFramePtr(env->mode64) ); |
| HReg i8_val = iselWordExpr_R(env, triop->arg2); |
| HReg r_zero = newVRegI( env ); |
| |
| iselDfp128Expr(&r_srcHi, &r_srcLo, env, triop->arg3); |
| |
| /* dst will be used to pass in the left operand and get the result */ |
| /* Move I8 to float register to issue instruction. Note, the |
| * instruction only looks at the bottom 6 bits so we really don't |
| * have to clear the upper bits since the iselWordExpr_R sets the |
| * bottom 8-bits. |
| */ |
| sub_from_sp( env, 16 ); |
| |
| if (env->mode64) |
| addInstr(env, PPCInstr_Store(4, four_r1, i8_val, True/*mode64*/)); |
| else |
| addInstr(env, PPCInstr_Store(4, four_r1, i8_val, False/*mode32*/)); |
| |
| /* Have to write to the upper bits to ensure they have been |
| * initialized. The instruction ignores all but the lower 6-bits. |
| */ |
| addInstr( env, PPCInstr_LI( r_zero, 0, env->mode64 ) ); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_dstHi, zero_r1)); |
| addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, r_dstLo, zero_r1)); |
| |
| add_to_sp( env, 16 ); |
| |
| // will set RMC when issuing instruction |
| addInstr(env, PPCInstr_DfpQuantize128(fpop, r_dstHi, r_dstLo, |
| r_srcHi, r_srcLo, rmc)); |
| *rHi = r_dstHi; |
| *rLo = r_dstLo; |
| return; |
| } |
| } |
| |
| ppIRExpr( e ); |
| vpanic( "iselDfp128Expr(ppc64)" ); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: SIMD (Vector) expressions, 128 bit. ---*/ |
| /*---------------------------------------------------------*/ |
| |
| static HReg iselVecExpr ( ISelEnv* env, IRExpr* e ) |
| { |
| HReg r = iselVecExpr_wrk( env, e ); |
| # if 0 |
| vex_printf("\n"); ppIRExpr(e); vex_printf("\n"); |
| # endif |
| vassert(hregClass(r) == HRcVec128); |
| vassert(hregIsVirtual(r)); |
| return r; |
| } |
| |
| /* DO NOT CALL THIS DIRECTLY */ |
| static HReg iselVecExpr_wrk ( ISelEnv* env, IRExpr* e ) |
| { |
| Bool mode64 = env->mode64; |
| PPCAvOp op = Pav_INVALID; |
| PPCAvFpOp fpop = Pavfp_INVALID; |
| IRType ty = typeOfIRExpr(env->type_env,e); |
| vassert(e); |
| vassert(ty == Ity_V128); |
| |
| if (e->tag == Iex_RdTmp) { |
| return lookupIRTemp(env, e->Iex.RdTmp.tmp); |
| } |
| |
| if (e->tag == Iex_Get) { |
| /* Guest state vectors are 16byte aligned, |
| so don't need to worry here */ |
| HReg dst = newVRegV(env); |
| addInstr(env, |
| PPCInstr_AvLdSt( True/*load*/, 16, dst, |
| PPCAMode_IR( e->Iex.Get.offset, |
| GuestStatePtr(mode64) ))); |
| return dst; |
| } |
| |
| if (e->tag == Iex_Load && e->Iex.Load.end == Iend_BE) { |
| PPCAMode* am_addr; |
| HReg v_dst = newVRegV(env); |
| vassert(e->Iex.Load.ty == Ity_V128); |
| am_addr = iselWordExpr_AMode(env, e->Iex.Load.addr, Ity_V128/*xfer*/); |
| addInstr(env, PPCInstr_AvLdSt( True/*load*/, 16, v_dst, am_addr)); |
| return v_dst; |
| } |
| |
| if (e->tag == Iex_Unop) { |
| switch (e->Iex.Unop.op) { |
| |
| case Iop_NotV128: { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, arg)); |
| return dst; |
| } |
| |
| case Iop_CmpNEZ8x16: { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg zero = newVRegV(env); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(Pav_XOR, zero, zero, zero)); |
| addInstr(env, PPCInstr_AvBin8x16(Pav_CMPEQU, dst, arg, zero)); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, dst)); |
| return dst; |
| } |
| |
| case Iop_CmpNEZ16x8: { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg zero = newVRegV(env); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(Pav_XOR, zero, zero, zero)); |
| addInstr(env, PPCInstr_AvBin16x8(Pav_CMPEQU, dst, arg, zero)); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, dst)); |
| return dst; |
| } |
| |
| case Iop_CmpNEZ32x4: { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg zero = newVRegV(env); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(Pav_XOR, zero, zero, zero)); |
| addInstr(env, PPCInstr_AvBin32x4(Pav_CMPEQU, dst, arg, zero)); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, dst)); |
| return dst; |
| } |
| |
| case Iop_CmpNEZ64x2: { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg zero = newVRegV(env); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(Pav_XOR, zero, zero, zero)); |
| addInstr(env, PPCInstr_AvBin64x2(Pav_CMPEQU, dst, arg, zero)); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, dst)); |
| return dst; |
| } |
| |
| case Iop_Recip32Fx4: fpop = Pavfp_RCPF; goto do_32Fx4_unary; |
| case Iop_RSqrt32Fx4: fpop = Pavfp_RSQRTF; goto do_32Fx4_unary; |
| case Iop_I32UtoFx4: fpop = Pavfp_CVTU2F; goto do_32Fx4_unary; |
| case Iop_I32StoFx4: fpop = Pavfp_CVTS2F; goto do_32Fx4_unary; |
| case Iop_QFtoI32Ux4_RZ: fpop = Pavfp_QCVTF2U; goto do_32Fx4_unary; |
| case Iop_QFtoI32Sx4_RZ: fpop = Pavfp_QCVTF2S; goto do_32Fx4_unary; |
| case Iop_RoundF32x4_RM: fpop = Pavfp_ROUNDM; goto do_32Fx4_unary; |
| case Iop_RoundF32x4_RP: fpop = Pavfp_ROUNDP; goto do_32Fx4_unary; |
| case Iop_RoundF32x4_RN: fpop = Pavfp_ROUNDN; goto do_32Fx4_unary; |
| case Iop_RoundF32x4_RZ: fpop = Pavfp_ROUNDZ; goto do_32Fx4_unary; |
| do_32Fx4_unary: |
| { |
| HReg arg = iselVecExpr(env, e->Iex.Unop.arg); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvUn32Fx4(fpop, dst, arg)); |
| return dst; |
| } |
| |
| case Iop_32UtoV128: { |
| HReg r_aligned16, r_zeros; |
| HReg r_src = iselWordExpr_R(env, e->Iex.Unop.arg); |
| HReg dst = newVRegV(env); |
| PPCAMode *am_off0, *am_off4, *am_off8, *am_off12; |
| sub_from_sp( env, 32 ); // Move SP down |
| |
| /* Get a quadword aligned address within our stack space */ |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_off4 = PPCAMode_IR( 4, r_aligned16 ); |
| am_off8 = PPCAMode_IR( 8, r_aligned16 ); |
| am_off12 = PPCAMode_IR( 12, r_aligned16 ); |
| |
| /* Store zeros */ |
| r_zeros = newVRegI(env); |
| addInstr(env, PPCInstr_LI(r_zeros, 0x0, mode64)); |
| addInstr(env, PPCInstr_Store( 4, am_off0, r_zeros, mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_off4, r_zeros, mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_off8, r_zeros, mode64 )); |
| |
| /* Store r_src in low word of quadword-aligned mem */ |
| addInstr(env, PPCInstr_Store( 4, am_off12, r_src, mode64 )); |
| |
| /* Load word into low word of quadword vector reg */ |
| addInstr(env, PPCInstr_AvLdSt( True/*ld*/, 4, dst, am_off12 )); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| return dst; |
| } |
| |
| case Iop_Dup8x16: |
| case Iop_Dup16x8: |
| case Iop_Dup32x4: |
| return mk_AvDuplicateRI(env, e->Iex.Unop.arg); |
| |
| default: |
| break; |
| } /* switch (e->Iex.Unop.op) */ |
| } /* if (e->tag == Iex_Unop) */ |
| |
| if (e->tag == Iex_Binop) { |
| switch (e->Iex.Binop.op) { |
| |
| case Iop_64HLtoV128: { |
| if (!mode64) { |
| HReg r3, r2, r1, r0, r_aligned16; |
| PPCAMode *am_off0, *am_off4, *am_off8, *am_off12; |
| HReg dst = newVRegV(env); |
| /* do this via the stack (easy, convenient, etc) */ |
| sub_from_sp( env, 32 ); // Move SP down |
| |
| // get a quadword aligned address within our stack space |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_off4 = PPCAMode_IR( 4, r_aligned16 ); |
| am_off8 = PPCAMode_IR( 8, r_aligned16 ); |
| am_off12 = PPCAMode_IR( 12, r_aligned16 ); |
| |
| /* Do the less significant 64 bits */ |
| iselInt64Expr(&r1, &r0, env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_Store( 4, am_off12, r0, mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_off8, r1, mode64 )); |
| /* Do the more significant 64 bits */ |
| iselInt64Expr(&r3, &r2, env, e->Iex.Binop.arg1); |
| addInstr(env, PPCInstr_Store( 4, am_off4, r2, mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_off0, r3, mode64 )); |
| |
| /* Fetch result back from stack. */ |
| addInstr(env, PPCInstr_AvLdSt(True/*ld*/, 16, dst, am_off0)); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| return dst; |
| } else { |
| HReg rHi = iselWordExpr_R(env, e->Iex.Binop.arg1); |
| HReg rLo = iselWordExpr_R(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| HReg r_aligned16; |
| PPCAMode *am_off0, *am_off8; |
| /* do this via the stack (easy, convenient, etc) */ |
| sub_from_sp( env, 32 ); // Move SP down |
| |
| // get a quadword aligned address within our stack space |
| r_aligned16 = get_sp_aligned16( env ); |
| am_off0 = PPCAMode_IR( 0, r_aligned16 ); |
| am_off8 = PPCAMode_IR( 8, r_aligned16 ); |
| |
| /* Store 2*I64 to stack */ |
| addInstr(env, PPCInstr_Store( 8, am_off0, rHi, mode64 )); |
| addInstr(env, PPCInstr_Store( 8, am_off8, rLo, mode64 )); |
| |
| /* Fetch result back from stack. */ |
| addInstr(env, PPCInstr_AvLdSt(True/*ld*/, 16, dst, am_off0)); |
| |
| add_to_sp( env, 32 ); // Reset SP |
| return dst; |
| } |
| } |
| |
| case Iop_Add32Fx4: fpop = Pavfp_ADDF; goto do_32Fx4; |
| case Iop_Sub32Fx4: fpop = Pavfp_SUBF; goto do_32Fx4; |
| case Iop_Max32Fx4: fpop = Pavfp_MAXF; goto do_32Fx4; |
| case Iop_Min32Fx4: fpop = Pavfp_MINF; goto do_32Fx4; |
| case Iop_Mul32Fx4: fpop = Pavfp_MULF; goto do_32Fx4; |
| case Iop_CmpEQ32Fx4: fpop = Pavfp_CMPEQF; goto do_32Fx4; |
| case Iop_CmpGT32Fx4: fpop = Pavfp_CMPGTF; goto do_32Fx4; |
| case Iop_CmpGE32Fx4: fpop = Pavfp_CMPGEF; goto do_32Fx4; |
| do_32Fx4: |
| { |
| HReg argL = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg argR = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBin32Fx4(fpop, dst, argL, argR)); |
| return dst; |
| } |
| |
| case Iop_CmpLE32Fx4: { |
| HReg argL = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg argR = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| |
| /* stay consistent with native ppc compares: |
| if a left/right lane holds a nan, return zeros for that lane |
| so: le == NOT(gt OR isNan) |
| */ |
| HReg isNanLR = newVRegV(env); |
| HReg isNanL = isNan(env, argL); |
| HReg isNanR = isNan(env, argR); |
| addInstr(env, PPCInstr_AvBinary(Pav_OR, isNanLR, |
| isNanL, isNanR)); |
| |
| addInstr(env, PPCInstr_AvBin32Fx4(Pavfp_CMPGTF, dst, |
| argL, argR)); |
| addInstr(env, PPCInstr_AvBinary(Pav_OR, dst, dst, isNanLR)); |
| addInstr(env, PPCInstr_AvUnary(Pav_NOT, dst, dst)); |
| return dst; |
| } |
| |
| case Iop_AndV128: op = Pav_AND; goto do_AvBin; |
| case Iop_OrV128: op = Pav_OR; goto do_AvBin; |
| case Iop_XorV128: op = Pav_XOR; goto do_AvBin; |
| do_AvBin: { |
| HReg arg1 = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg arg2 = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBinary(op, dst, arg1, arg2)); |
| return dst; |
| } |
| |
| case Iop_Shl8x16: op = Pav_SHL; goto do_AvBin8x16; |
| case Iop_Shr8x16: op = Pav_SHR; goto do_AvBin8x16; |
| case Iop_Sar8x16: op = Pav_SAR; goto do_AvBin8x16; |
| case Iop_Rol8x16: op = Pav_ROTL; goto do_AvBin8x16; |
| case Iop_InterleaveHI8x16: op = Pav_MRGHI; goto do_AvBin8x16; |
| case Iop_InterleaveLO8x16: op = Pav_MRGLO; goto do_AvBin8x16; |
| case Iop_Add8x16: op = Pav_ADDU; goto do_AvBin8x16; |
| case Iop_QAdd8Ux16: op = Pav_QADDU; goto do_AvBin8x16; |
| case Iop_QAdd8Sx16: op = Pav_QADDS; goto do_AvBin8x16; |
| case Iop_Sub8x16: op = Pav_SUBU; goto do_AvBin8x16; |
| case Iop_QSub8Ux16: op = Pav_QSUBU; goto do_AvBin8x16; |
| case Iop_QSub8Sx16: op = Pav_QSUBS; goto do_AvBin8x16; |
| case Iop_Avg8Ux16: op = Pav_AVGU; goto do_AvBin8x16; |
| case Iop_Avg8Sx16: op = Pav_AVGS; goto do_AvBin8x16; |
| case Iop_Max8Ux16: op = Pav_MAXU; goto do_AvBin8x16; |
| case Iop_Max8Sx16: op = Pav_MAXS; goto do_AvBin8x16; |
| case Iop_Min8Ux16: op = Pav_MINU; goto do_AvBin8x16; |
| case Iop_Min8Sx16: op = Pav_MINS; goto do_AvBin8x16; |
| case Iop_MullEven8Ux16: op = Pav_OMULU; goto do_AvBin8x16; |
| case Iop_MullEven8Sx16: op = Pav_OMULS; goto do_AvBin8x16; |
| case Iop_CmpEQ8x16: op = Pav_CMPEQU; goto do_AvBin8x16; |
| case Iop_CmpGT8Ux16: op = Pav_CMPGTU; goto do_AvBin8x16; |
| case Iop_CmpGT8Sx16: op = Pav_CMPGTS; goto do_AvBin8x16; |
| do_AvBin8x16: { |
| HReg arg1 = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg arg2 = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBin8x16(op, dst, arg1, arg2)); |
| return dst; |
| } |
| |
| case Iop_Shl16x8: op = Pav_SHL; goto do_AvBin16x8; |
| case Iop_Shr16x8: op = Pav_SHR; goto do_AvBin16x8; |
| case Iop_Sar16x8: op = Pav_SAR; goto do_AvBin16x8; |
| case Iop_Rol16x8: op = Pav_ROTL; goto do_AvBin16x8; |
| case Iop_NarrowBin16to8x16: op = Pav_PACKUU; goto do_AvBin16x8; |
| case Iop_QNarrowBin16Uto8Ux16: op = Pav_QPACKUU; goto do_AvBin16x8; |
| case Iop_QNarrowBin16Sto8Sx16: op = Pav_QPACKSS; goto do_AvBin16x8; |
| case Iop_InterleaveHI16x8: op = Pav_MRGHI; goto do_AvBin16x8; |
| case Iop_InterleaveLO16x8: op = Pav_MRGLO; goto do_AvBin16x8; |
| case Iop_Add16x8: op = Pav_ADDU; goto do_AvBin16x8; |
| case Iop_QAdd16Ux8: op = Pav_QADDU; goto do_AvBin16x8; |
| case Iop_QAdd16Sx8: op = Pav_QADDS; goto do_AvBin16x8; |
| case Iop_Sub16x8: op = Pav_SUBU; goto do_AvBin16x8; |
| case Iop_QSub16Ux8: op = Pav_QSUBU; goto do_AvBin16x8; |
| case Iop_QSub16Sx8: op = Pav_QSUBS; goto do_AvBin16x8; |
| case Iop_Avg16Ux8: op = Pav_AVGU; goto do_AvBin16x8; |
| case Iop_Avg16Sx8: op = Pav_AVGS; goto do_AvBin16x8; |
| case Iop_Max16Ux8: op = Pav_MAXU; goto do_AvBin16x8; |
| case Iop_Max16Sx8: op = Pav_MAXS; goto do_AvBin16x8; |
| case Iop_Min16Ux8: op = Pav_MINU; goto do_AvBin16x8; |
| case Iop_Min16Sx8: op = Pav_MINS; goto do_AvBin16x8; |
| case Iop_MullEven16Ux8: op = Pav_OMULU; goto do_AvBin16x8; |
| case Iop_MullEven16Sx8: op = Pav_OMULS; goto do_AvBin16x8; |
| case Iop_CmpEQ16x8: op = Pav_CMPEQU; goto do_AvBin16x8; |
| case Iop_CmpGT16Ux8: op = Pav_CMPGTU; goto do_AvBin16x8; |
| case Iop_CmpGT16Sx8: op = Pav_CMPGTS; goto do_AvBin16x8; |
| do_AvBin16x8: { |
| HReg arg1 = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg arg2 = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBin16x8(op, dst, arg1, arg2)); |
| return dst; |
| } |
| |
| case Iop_Shl32x4: op = Pav_SHL; goto do_AvBin32x4; |
| case Iop_Shr32x4: op = Pav_SHR; goto do_AvBin32x4; |
| case Iop_Sar32x4: op = Pav_SAR; goto do_AvBin32x4; |
| case Iop_Rol32x4: op = Pav_ROTL; goto do_AvBin32x4; |
| case Iop_NarrowBin32to16x8: op = Pav_PACKUU; goto do_AvBin32x4; |
| case Iop_QNarrowBin32Uto16Ux8: op = Pav_QPACKUU; goto do_AvBin32x4; |
| case Iop_QNarrowBin32Sto16Sx8: op = Pav_QPACKSS; goto do_AvBin32x4; |
| case Iop_InterleaveHI32x4: op = Pav_MRGHI; goto do_AvBin32x4; |
| case Iop_InterleaveLO32x4: op = Pav_MRGLO; goto do_AvBin32x4; |
| case Iop_Add32x4: op = Pav_ADDU; goto do_AvBin32x4; |
| case Iop_QAdd32Ux4: op = Pav_QADDU; goto do_AvBin32x4; |
| case Iop_QAdd32Sx4: op = Pav_QADDS; goto do_AvBin32x4; |
| case Iop_Sub32x4: op = Pav_SUBU; goto do_AvBin32x4; |
| case Iop_QSub32Ux4: op = Pav_QSUBU; goto do_AvBin32x4; |
| case Iop_QSub32Sx4: op = Pav_QSUBS; goto do_AvBin32x4; |
| case Iop_Avg32Ux4: op = Pav_AVGU; goto do_AvBin32x4; |
| case Iop_Avg32Sx4: op = Pav_AVGS; goto do_AvBin32x4; |
| case Iop_Max32Ux4: op = Pav_MAXU; goto do_AvBin32x4; |
| case Iop_Max32Sx4: op = Pav_MAXS; goto do_AvBin32x4; |
| case Iop_Min32Ux4: op = Pav_MINU; goto do_AvBin32x4; |
| case Iop_Min32Sx4: op = Pav_MINS; goto do_AvBin32x4; |
| case Iop_CmpEQ32x4: op = Pav_CMPEQU; goto do_AvBin32x4; |
| case Iop_CmpGT32Ux4: op = Pav_CMPGTU; goto do_AvBin32x4; |
| case Iop_CmpGT32Sx4: op = Pav_CMPGTS; goto do_AvBin32x4; |
| do_AvBin32x4: { |
| HReg arg1 = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg arg2 = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBin32x4(op, dst, arg1, arg2)); |
| return dst; |
| } |
| |
| case Iop_NarrowBin64to32x4: op = Pav_PACKUU; goto do_AvBin64x2; |
| case Iop_Add64x2: op = Pav_ADDU; goto do_AvBin64x2; |
| do_AvBin64x2: { |
| HReg arg1 = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg arg2 = iselVecExpr(env, e->Iex.Binop.arg2); |
| HReg dst = newVRegV(env); |
| addInstr(env, PPCInstr_AvBin64x2(op, dst, arg1, arg2)); |
| return dst; |
| } |
| |
| case Iop_ShlN8x16: op = Pav_SHL; goto do_AvShift8x16; |
| case Iop_SarN8x16: op = Pav_SAR; goto do_AvShift8x16; |
| do_AvShift8x16: { |
| HReg r_src = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg dst = newVRegV(env); |
| HReg v_shft = mk_AvDuplicateRI(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_AvBin8x16(op, dst, r_src, v_shft)); |
| return dst; |
| } |
| |
| case Iop_ShlN16x8: op = Pav_SHL; goto do_AvShift16x8; |
| case Iop_ShrN16x8: op = Pav_SHR; goto do_AvShift16x8; |
| case Iop_SarN16x8: op = Pav_SAR; goto do_AvShift16x8; |
| do_AvShift16x8: { |
| HReg r_src = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg dst = newVRegV(env); |
| HReg v_shft = mk_AvDuplicateRI(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_AvBin16x8(op, dst, r_src, v_shft)); |
| return dst; |
| } |
| |
| case Iop_ShlN32x4: op = Pav_SHL; goto do_AvShift32x4; |
| case Iop_ShrN32x4: op = Pav_SHR; goto do_AvShift32x4; |
| case Iop_SarN32x4: op = Pav_SAR; goto do_AvShift32x4; |
| do_AvShift32x4: { |
| HReg r_src = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg dst = newVRegV(env); |
| HReg v_shft = mk_AvDuplicateRI(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_AvBin32x4(op, dst, r_src, v_shft)); |
| return dst; |
| } |
| |
| case Iop_ShrV128: op = Pav_SHR; goto do_AvShiftV128; |
| case Iop_ShlV128: op = Pav_SHL; goto do_AvShiftV128; |
| do_AvShiftV128: { |
| HReg dst = newVRegV(env); |
| HReg r_src = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg v_shft = mk_AvDuplicateRI(env, e->Iex.Binop.arg2); |
| /* Note: shift value gets masked by 127 */ |
| addInstr(env, PPCInstr_AvBinary(op, dst, r_src, v_shft)); |
| return dst; |
| } |
| |
| case Iop_Perm8x16: { |
| HReg dst = newVRegV(env); |
| HReg v_src = iselVecExpr(env, e->Iex.Binop.arg1); |
| HReg v_ctl = iselVecExpr(env, e->Iex.Binop.arg2); |
| addInstr(env, PPCInstr_AvPerm(dst, v_src, v_src, v_ctl)); |
| return dst; |
| } |
| |
| default: |
| break; |
| } /* switch (e->Iex.Binop.op) */ |
| } /* if (e->tag == Iex_Binop) */ |
| |
| if (e->tag == Iex_Const ) { |
| vassert(e->Iex.Const.con->tag == Ico_V128); |
| if (e->Iex.Const.con->Ico.V128 == 0x0000) { |
| return generate_zeroes_V128(env); |
| } |
| else if (e->Iex.Const.con->Ico.V128 == 0xffff) { |
| return generate_ones_V128(env); |
| } |
| } |
| |
| vex_printf("iselVecExpr(ppc) (subarch = %s): can't reduce\n", |
| LibVEX_ppVexHwCaps(mode64 ? VexArchPPC64 : VexArchPPC32, |
| env->hwcaps)); |
| ppIRExpr(e); |
| vpanic("iselVecExpr_wrk(ppc)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Statements ---*/ |
| /*---------------------------------------------------------*/ |
| |
| static void iselStmt ( ISelEnv* env, IRStmt* stmt ) |
| { |
| Bool mode64 = env->mode64; |
| if (vex_traceflags & VEX_TRACE_VCODE) { |
| vex_printf("\n -- "); |
| ppIRStmt(stmt); |
| vex_printf("\n"); |
| } |
| |
| switch (stmt->tag) { |
| |
| /* --------- STORE --------- */ |
| case Ist_Store: { |
| IRType tya = typeOfIRExpr(env->type_env, stmt->Ist.Store.addr); |
| IRType tyd = typeOfIRExpr(env->type_env, stmt->Ist.Store.data); |
| IREndness end = stmt->Ist.Store.end; |
| |
| if (end != Iend_BE) |
| goto stmt_fail; |
| if (!mode64 && (tya != Ity_I32)) |
| goto stmt_fail; |
| if (mode64 && (tya != Ity_I64)) |
| goto stmt_fail; |
| |
| if (tyd == Ity_I8 || tyd == Ity_I16 || tyd == Ity_I32 || |
| (mode64 && (tyd == Ity_I64))) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg r_src = iselWordExpr_R(env, stmt->Ist.Store.data); |
| addInstr(env, PPCInstr_Store( toUChar(sizeofIRType(tyd)), |
| am_addr, r_src, mode64 )); |
| return; |
| } |
| if (tyd == Ity_F64) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg fr_src = iselDblExpr(env, stmt->Ist.Store.data); |
| addInstr(env, |
| PPCInstr_FpLdSt(False/*store*/, 8, fr_src, am_addr)); |
| return; |
| } |
| if (tyd == Ity_F32) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg fr_src = iselFltExpr(env, stmt->Ist.Store.data); |
| addInstr(env, |
| PPCInstr_FpLdSt(False/*store*/, 4, fr_src, am_addr)); |
| return; |
| } |
| if (tyd == Ity_D64) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg fr_src = iselDfp64Expr(env, stmt->Ist.Store.data); |
| addInstr(env, |
| PPCInstr_FpLdSt(False/*store*/, 8, fr_src, am_addr)); |
| return; |
| } |
| if (tyd == Ity_D32) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg fr_src = iselDfp32Expr(env, stmt->Ist.Store.data); |
| addInstr(env, |
| PPCInstr_FpLdSt(False/*store*/, 4, fr_src, am_addr)); |
| return; |
| } |
| if (tyd == Ity_V128) { |
| PPCAMode* am_addr |
| = iselWordExpr_AMode(env, stmt->Ist.Store.addr, tyd/*of xfer*/); |
| HReg v_src = iselVecExpr(env, stmt->Ist.Store.data); |
| addInstr(env, |
| PPCInstr_AvLdSt(False/*store*/, 16, v_src, am_addr)); |
| return; |
| } |
| if (tyd == Ity_I64 && !mode64) { |
| /* Just calculate the address in the register. Life is too |
| short to arse around trying and possibly failing to adjust |
| the offset in a 'reg+offset' style amode. */ |
| HReg rHi32, rLo32; |
| HReg r_addr = iselWordExpr_R(env, stmt->Ist.Store.addr); |
| iselInt64Expr( &rHi32, &rLo32, env, stmt->Ist.Store.data ); |
| addInstr(env, PPCInstr_Store( 4/*byte-store*/, |
| PPCAMode_IR( 0, r_addr ), |
| rHi32, |
| False/*32-bit insn please*/) ); |
| addInstr(env, PPCInstr_Store( 4/*byte-store*/, |
| PPCAMode_IR( 4, r_addr ), |
| rLo32, |
| False/*32-bit insn please*/) ); |
| return; |
| } |
| break; |
| } |
| |
| /* --------- PUT --------- */ |
| case Ist_Put: { |
| IRType ty = typeOfIRExpr(env->type_env, stmt->Ist.Put.data); |
| if (ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && mode64)) { |
| HReg r_src = iselWordExpr_R(env, stmt->Ist.Put.data); |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| addInstr(env, PPCInstr_Store( toUChar(sizeofIRType(ty)), |
| am_addr, r_src, mode64 )); |
| return; |
| } |
| if (!mode64 && ty == Ity_I64) { |
| HReg rHi, rLo; |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| PPCAMode* am_addr4 = advance4(env, am_addr); |
| iselInt64Expr(&rHi,&rLo, env, stmt->Ist.Put.data); |
| addInstr(env, PPCInstr_Store( 4, am_addr, rHi, mode64 )); |
| addInstr(env, PPCInstr_Store( 4, am_addr4, rLo, mode64 )); |
| return; |
| } |
| if (ty == Ity_V128) { |
| /* Guest state vectors are 16byte aligned, |
| so don't need to worry here */ |
| HReg v_src = iselVecExpr(env, stmt->Ist.Put.data); |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| addInstr(env, |
| PPCInstr_AvLdSt(False/*store*/, 16, v_src, am_addr)); |
| return; |
| } |
| if (ty == Ity_F64) { |
| HReg fr_src = iselDblExpr(env, stmt->Ist.Put.data); |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| addInstr(env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr )); |
| return; |
| } |
| if (ty == Ity_D32) { |
| /* The 32-bit value is stored in a 64-bit register */ |
| HReg fr_src = iselDfp32Expr( env, stmt->Ist.Put.data ); |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| addInstr( env, PPCInstr_FpLdSt( False/*store*/, 8, |
| fr_src, am_addr ) ); |
| return; |
| } |
| if (ty == Ity_D64) { |
| HReg fr_src = iselDfp64Expr( env, stmt->Ist.Put.data ); |
| PPCAMode* am_addr = PPCAMode_IR( stmt->Ist.Put.offset, |
| GuestStatePtr(mode64) ); |
| addInstr( env, PPCInstr_FpLdSt( False/*store*/, 8, fr_src, am_addr ) ); |
| return; |
| } |
| break; |
| } |
| |
| /* --------- Indexed PUT --------- */ |
| case Ist_PutI: { |
| IRPutI *puti = stmt->Ist.PutI.details; |
| |
| PPCAMode* dst_am |
| = genGuestArrayOffset( |
| env, puti->descr, |
| puti->ix, puti->bias ); |
| IRType ty = typeOfIRExpr(env->type_env, puti->data); |
| if (mode64 && ty == Ity_I64) { |
| HReg r_src = iselWordExpr_R(env, puti->data); |
| addInstr(env, PPCInstr_Store( toUChar(8), |
| dst_am, r_src, mode64 )); |
| return; |
| } |
| if ((!mode64) && ty == Ity_I32) { |
| HReg r_src = iselWordExpr_R(env, puti->data); |
| addInstr(env, PPCInstr_Store( toUChar(4), |
| dst_am, r_src, mode64 )); |
| return; |
| } |
| break; |
| } |
| |
| /* --------- TMP --------- */ |
| case Ist_WrTmp: { |
| IRTemp tmp = stmt->Ist.WrTmp.tmp; |
| IRType ty = typeOfIRTemp(env->type_env, tmp); |
| if (ty == Ity_I8 || ty == Ity_I16 || |
| ty == Ity_I32 || ((ty == Ity_I64) && mode64)) { |
| HReg r_dst = lookupIRTemp(env, tmp); |
| HReg r_src = iselWordExpr_R(env, stmt->Ist.WrTmp.data); |
| addInstr(env, mk_iMOVds_RR( r_dst, r_src )); |
| return; |
| } |
| if (!mode64 && ty == Ity_I64) { |
| HReg r_srcHi, r_srcLo, r_dstHi, r_dstLo; |
| |
| iselInt64Expr(&r_srcHi,&r_srcLo, env, stmt->Ist.WrTmp.data); |
| lookupIRTempPair( &r_dstHi, &r_dstLo, env, tmp); |
| addInstr(env, mk_iMOVds_RR(r_dstHi, r_srcHi) ); |
| addInstr(env, mk_iMOVds_RR(r_dstLo, r_srcLo) ); |
| return; |
| } |
| if (mode64 && ty == Ity_I128) { |
| HReg r_srcHi, r_srcLo, r_dstHi, r_dstLo; |
| iselInt128Expr(&r_srcHi,&r_srcLo, env, stmt->Ist.WrTmp.data); |
| lookupIRTempPair( &r_dstHi, &r_dstLo, env, tmp); |
| addInstr(env, mk_iMOVds_RR(r_dstHi, r_srcHi) ); |
| addInstr(env, mk_iMOVds_RR(r_dstLo, r_srcLo) ); |
| return; |
| } |
| if (!mode64 && ty == Ity_I128) { |
| HReg r_srcHi, r_srcMedHi, r_srcMedLo, r_srcLo; |
| HReg r_dstHi, r_dstMedHi, r_dstMedLo, r_dstLo; |
| |
| iselInt128Expr_to_32x4(&r_srcHi, &r_srcMedHi, |
| &r_srcMedLo, &r_srcLo, |
| env, stmt->Ist.WrTmp.data); |
| |
| lookupIRTempQuad( &r_dstHi, &r_dstMedHi, &r_dstMedLo, |
| &r_dstLo, env, tmp); |
| |
| addInstr(env, mk_iMOVds_RR(r_dstHi, r_srcHi) ); |
| addInstr(env, mk_iMOVds_RR(r_dstMedHi, r_srcMedHi) ); |
| addInstr(env, mk_iMOVds_RR(r_dstMedLo, r_srcMedLo) ); |
| addInstr(env, mk_iMOVds_RR(r_dstLo, r_srcLo) ); |
| return; |
| } |
| if (ty == Ity_I1) { |
| PPCCondCode cond = iselCondCode(env, stmt->Ist.WrTmp.data); |
| HReg r_dst = lookupIRTemp(env, tmp); |
| addInstr(env, PPCInstr_Set(cond, r_dst)); |
| return; |
| } |
| if (ty == Ity_F64) { |
| HReg fr_dst = lookupIRTemp(env, tmp); |
| HReg fr_src = iselDblExpr(env, stmt->Ist.WrTmp.data); |
| addInstr(env, PPCInstr_FpUnary(Pfp_MOV, fr_dst, fr_src)); |
| return; |
| } |
| if (ty == Ity_F32) { |
| HReg fr_dst = lookupIRTemp(env, tmp); |
| HReg fr_src = iselFltExpr(env, stmt->Ist.WrTmp.data); |
| addInstr(env, PPCInstr_FpUnary(Pfp_MOV, fr_dst, fr_src)); |
| return; |
| } |
| if (ty == Ity_D32) { |
| HReg fr_dst = lookupIRTemp(env, tmp); |
| HReg fr_src = iselDfp32Expr(env, stmt->Ist.WrTmp.data); |
| addInstr(env, PPCInstr_Dfp64Unary(Pfp_MOV, fr_dst, fr_src)); |
| return; |
| } |
| if (ty == Ity_V128) { |
| HReg v_dst = lookupIRTemp(env, tmp); |
| HReg v_src = iselVecExpr(env, stmt->Ist.WrTmp.data); |
| addInstr(env, PPCInstr_AvUnary(Pav_MOV, v_dst, v_src)); |
| return; |
| } |
| if (ty == Ity_D64) { |
| HReg fr_dst = lookupIRTemp( env, tmp ); |
| HReg fr_src = iselDfp64Expr( env, stmt->Ist.WrTmp.data ); |
| addInstr( env, PPCInstr_Dfp64Unary( Pfp_MOV, fr_dst, fr_src ) ); |
| return; |
| } |
| if (ty == Ity_D128) { |
| HReg fr_srcHi, fr_srcLo, fr_dstHi, fr_dstLo; |
| // lookupDfp128IRTempPair( &fr_dstHi, &fr_dstLo, env, tmp ); |
| lookupIRTempPair( &fr_dstHi, &fr_dstLo, env, tmp ); |
| iselDfp128Expr( &fr_srcHi, &fr_srcLo, env, stmt->Ist.WrTmp.data ); |
| addInstr( env, PPCInstr_Dfp64Unary( Pfp_MOV, fr_dstHi, fr_srcHi ) ); |
| addInstr( env, PPCInstr_Dfp64Unary( Pfp_MOV, fr_dstLo, fr_srcLo ) ); |
| return; |
| } |
| break; |
| } |
| |
| /* --------- Load Linked or Store Conditional --------- */ |
| case Ist_LLSC: { |
| IRTemp res = stmt->Ist.LLSC.result; |
| IRType tyRes = typeOfIRTemp(env->type_env, res); |
| IRType tyAddr = typeOfIRExpr(env->type_env, stmt->Ist.LLSC.addr); |
| |
| if (stmt->Ist.LLSC.end != Iend_BE) |
| goto stmt_fail; |
| if (!mode64 && (tyAddr != Ity_I32)) |
| goto stmt_fail; |
| if (mode64 && (tyAddr != Ity_I64)) |
| goto stmt_fail; |
| |
| if (stmt->Ist.LLSC.storedata == NULL) { |
| /* LL */ |
| HReg r_addr = iselWordExpr_R( env, stmt->Ist.LLSC.addr ); |
| HReg r_dst = lookupIRTemp(env, res); |
| if (tyRes == Ity_I32) { |
| addInstr(env, PPCInstr_LoadL( 4, r_dst, r_addr, mode64 )); |
| return; |
| } |
| if (tyRes == Ity_I64 && mode64) { |
| addInstr(env, PPCInstr_LoadL( 8, r_dst, r_addr, mode64 )); |
| return; |
| } |
| /* fallthru */; |
| } else { |
| /* SC */ |
| HReg r_res = lookupIRTemp(env, res); /* :: Ity_I1 */ |
| HReg r_a = iselWordExpr_R(env, stmt->Ist.LLSC.addr); |
| HReg r_src = iselWordExpr_R(env, stmt->Ist.LLSC.storedata); |
| HReg r_tmp = newVRegI(env); |
| IRType tyData = typeOfIRExpr(env->type_env, |
| stmt->Ist.LLSC.storedata); |
| vassert(tyRes == Ity_I1); |
| if (tyData == Ity_I32 || (tyData == Ity_I64 && mode64)) { |
| addInstr(env, PPCInstr_StoreC( tyData==Ity_I32 ? 4 : 8, |
| r_a, r_src, mode64 )); |
| addInstr(env, PPCInstr_MfCR( r_tmp )); |
| addInstr(env, PPCInstr_Shft( |
| Pshft_SHR, |
| env->mode64 ? False : True |
| /*F:64-bit, T:32-bit shift*/, |
| r_tmp, r_tmp, |
| PPCRH_Imm(False/*unsigned*/, 29))); |
| /* Probably unnecessary, since the IR dest type is Ity_I1, |
| and so we are entitled to leave whatever junk we like |
| drifting round in the upper 31 or 63 bits of r_res. |
| However, for the sake of conservativeness .. */ |
| addInstr(env, PPCInstr_Alu( |
| Palu_AND, |
| r_res, r_tmp, |
| PPCRH_Imm(False/*signed*/, 1))); |
| return; |
| } |
| /* fallthru */ |
| } |
| goto stmt_fail; |
| /*NOTREACHED*/ |
| } |
| |
| /* --------- Call to DIRTY helper --------- */ |
| case Ist_Dirty: { |
| IRDirty* d = stmt->Ist.Dirty.details; |
| |
| /* Figure out the return type, if any. */ |
| IRType retty = Ity_INVALID; |
| if (d->tmp != IRTemp_INVALID) |
| retty = typeOfIRTemp(env->type_env, d->tmp); |
| |
| /* Throw out any return types we don't know about. */ |
| Bool retty_ok = False; |
| if (mode64) { |
| switch (retty) { |
| case Ity_INVALID: /* function doesn't return anything */ |
| case Ity_V128: |
| case Ity_I64: case Ity_I32: case Ity_I16: case Ity_I8: |
| retty_ok = True; break; |
| default: |
| break; |
| } |
| } else { |
| switch (retty) { |
| case Ity_INVALID: /* function doesn't return anything */ |
| case Ity_V128: |
| case Ity_I64: case Ity_I32: case Ity_I16: case Ity_I8: |
| retty_ok = True; break; |
| default: |
| break; |
| } |
| } |
| if (!retty_ok) |
| break; /* will go to stmt_fail: */ |
| |
| /* Marshal args, do the call, clear stack, set the return value |
| to 0x555..555 if this is a conditional call that returns a |
| value and the call is skipped. */ |
| UInt addToSp = 0; |
| RetLoc rloc = mk_RetLoc_INVALID(); |
| doHelperCall( &addToSp, &rloc, env, d->guard, d->cee, retty, d->args ); |
| vassert(is_sane_RetLoc(rloc)); |
| |
| /* Now figure out what to do with the returned value, if any. */ |
| switch (retty) { |
| case Ity_INVALID: { |
| /* No return value. Nothing to do. */ |
| vassert(d->tmp == IRTemp_INVALID); |
| vassert(rloc.pri == RLPri_None); |
| vassert(addToSp == 0); |
| return; |
| } |
| case Ity_I32: case Ity_I16: case Ity_I8: { |
| /* The returned value is in %r3. Park it in the register |
| associated with tmp. */ |
| HReg r_dst = lookupIRTemp(env, d->tmp); |
| addInstr(env, mk_iMOVds_RR(r_dst, hregPPC_GPR3(mode64))); |
| vassert(rloc.pri == RLPri_Int); |
| vassert(addToSp == 0); |
| return; |
| } |
| case Ity_I64: |
| if (mode64) { |
| /* The returned value is in %r3. Park it in the register |
| associated with tmp. */ |
| HReg r_dst = lookupIRTemp(env, d->tmp); |
| addInstr(env, mk_iMOVds_RR(r_dst, hregPPC_GPR3(mode64))); |
| vassert(rloc.pri == RLPri_Int); |
| vassert(addToSp == 0); |
| } else { |
| /* The returned value is in %r3:%r4. Park it in the |
| register-pair associated with tmp. */ |
| HReg r_dstHi = INVALID_HREG; |
| HReg r_dstLo = INVALID_HREG; |
| lookupIRTempPair( &r_dstHi, &r_dstLo, env, d->tmp); |
| addInstr(env, mk_iMOVds_RR(r_dstHi, hregPPC_GPR3(mode64))); |
| addInstr(env, mk_iMOVds_RR(r_dstLo, hregPPC_GPR4(mode64))); |
| vassert(rloc.pri == RLPri_2Int); |
| vassert(addToSp == 0); |
| } |
| return; |
| case Ity_V128: { |
| /* The returned value is on the stack, and *retloc tells |
| us where. Fish it off the stack and then move the |
| stack pointer upwards to clear it, as directed by |
| doHelperCall. */ |
| vassert(rloc.pri == RLPri_V128SpRel); |
| vassert(addToSp >= 16); |
| HReg dst = lookupIRTemp(env, d->tmp); |
| PPCAMode* am = PPCAMode_IR(rloc.spOff, StackFramePtr(mode64)); |
| addInstr(env, PPCInstr_AvLdSt( True/*load*/, 16, dst, am )); |
| add_to_sp(env, addToSp); |
| return; |
| } |
| default: |
| /*NOTREACHED*/ |
| vassert(0); |
| } |
| } |
| |
| /* --------- MEM FENCE --------- */ |
| case Ist_MBE: |
| switch (stmt->Ist.MBE.event) { |
| case Imbe_Fence: |
| addInstr(env, PPCInstr_MFence()); |
| return; |
| default: |
| break; |
| } |
| break; |
| |
| /* --------- INSTR MARK --------- */ |
| /* Doesn't generate any executable code ... */ |
| case Ist_IMark: |
| return; |
| |
| /* --------- ABI HINT --------- */ |
| /* These have no meaning (denotation in the IR) and so we ignore |
| them ... if any actually made it this far. */ |
| case Ist_AbiHint: |
| return; |
| |
| /* --------- NO-OP --------- */ |
| /* Fairly self-explanatory, wouldn't you say? */ |
| case Ist_NoOp: |
| return; |
| |
| /* --------- EXIT --------- */ |
| case Ist_Exit: { |
| IRConst* dst = stmt->Ist.Exit.dst; |
| if (!mode64 && dst->tag != Ico_U32) |
| vpanic("iselStmt(ppc): Ist_Exit: dst is not a 32-bit value"); |
| if (mode64 && dst->tag != Ico_U64) |
| vpanic("iselStmt(ppc64): Ist_Exit: dst is not a 64-bit value"); |
| |
| PPCCondCode cc = iselCondCode(env, stmt->Ist.Exit.guard); |
| PPCAMode* amCIA = PPCAMode_IR(stmt->Ist.Exit.offsIP, |
| hregPPC_GPR31(mode64)); |
| |
| /* Case: boring transfer to known address */ |
| if (stmt->Ist.Exit.jk == Ijk_Boring |
| || stmt->Ist.Exit.jk == Ijk_Call |
| /* || stmt->Ist.Exit.jk == Ijk_Ret */) { |
| if (env->chainingAllowed) { |
| /* .. almost always true .. */ |
| /* Skip the event check at the dst if this is a forwards |
| edge. */ |
| Bool toFastEP |
| = mode64 |
| ? (((Addr64)stmt->Ist.Exit.dst->Ico.U64) > (Addr64)env->max_ga) |
| : (((Addr32)stmt->Ist.Exit.dst->Ico.U32) > (Addr32)env->max_ga); |
| if (0) vex_printf("%s", toFastEP ? "Y" : ","); |
| addInstr(env, PPCInstr_XDirect( |
| mode64 ? (Addr64)stmt->Ist.Exit.dst->Ico.U64 |
| : (Addr64)stmt->Ist.Exit.dst->Ico.U32, |
| amCIA, cc, toFastEP)); |
| } else { |
| /* .. very occasionally .. */ |
| /* We can't use chaining, so ask for an assisted transfer, |
| as that's the only alternative that is allowable. */ |
| HReg r = iselWordExpr_R(env, IRExpr_Const(stmt->Ist.Exit.dst)); |
| addInstr(env, PPCInstr_XAssisted(r, amCIA, cc, Ijk_Boring)); |
| } |
| return; |
| } |
| |
| /* Case: assisted transfer to arbitrary address */ |
| switch (stmt->Ist.Exit.jk) { |
| /* Keep this list in sync with that in iselNext below */ |
| case Ijk_ClientReq: |
| case Ijk_EmFail: |
| case Ijk_EmWarn: |
| case Ijk_NoDecode: |
| case Ijk_NoRedir: |
| case Ijk_SigBUS: |
| case Ijk_SigTRAP: |
| case Ijk_Sys_syscall: |
| case Ijk_TInval: |
| { |
| HReg r = iselWordExpr_R(env, IRExpr_Const(stmt->Ist.Exit.dst)); |
| addInstr(env, PPCInstr_XAssisted(r, amCIA, cc, |
| stmt->Ist.Exit.jk)); |
| return; |
| } |
| default: |
| break; |
| } |
| |
| /* Do we ever expect to see any other kind? */ |
| goto stmt_fail; |
| } |
| |
| default: break; |
| } |
| stmt_fail: |
| ppIRStmt(stmt); |
| vpanic("iselStmt(ppc)"); |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- ISEL: Basic block terminators (Nexts) ---*/ |
| /*---------------------------------------------------------*/ |
| |
| static void iselNext ( ISelEnv* env, |
| IRExpr* next, IRJumpKind jk, Int offsIP ) |
| { |
| if (vex_traceflags & VEX_TRACE_VCODE) { |
| vex_printf( "\n-- PUT(%d) = ", offsIP); |
| ppIRExpr( next ); |
| vex_printf( "; exit-"); |
| ppIRJumpKind(jk); |
| vex_printf( "\n"); |
| } |
| |
| PPCCondCode always = mk_PPCCondCode( Pct_ALWAYS, Pcf_NONE ); |
| |
| /* Case: boring transfer to known address */ |
| if (next->tag == Iex_Const) { |
| IRConst* cdst = next->Iex.Const.con; |
| vassert(cdst->tag == (env->mode64 ? Ico_U64 :Ico_U32)); |
| if (jk == Ijk_Boring || jk == Ijk_Call) { |
| /* Boring transfer to known address */ |
| PPCAMode* amCIA = PPCAMode_IR(offsIP, hregPPC_GPR31(env->mode64)); |
| if (env->chainingAllowed) { |
| /* .. almost always true .. */ |
| /* Skip the event check at the dst if this is a forwards |
| edge. */ |
| Bool toFastEP |
| = env->mode64 |
| ? (((Addr64)cdst->Ico.U64) > (Addr64)env->max_ga) |
| : (((Addr32)cdst->Ico.U32) > (Addr32)env->max_ga); |
| if (0) vex_printf("%s", toFastEP ? "X" : "."); |
| addInstr(env, PPCInstr_XDirect( |
| env->mode64 ? (Addr64)cdst->Ico.U64 |
| : (Addr64)cdst->Ico.U32, |
| amCIA, always, toFastEP)); |
| } else { |
| /* .. very occasionally .. */ |
| /* We can't use chaining, so ask for an assisted transfer, |
| as that's the only alternative that is allowable. */ |
| HReg r = iselWordExpr_R(env, next); |
| addInstr(env, PPCInstr_XAssisted(r, amCIA, always, |
| Ijk_Boring)); |
| } |
| return; |
| } |
| } |
| |
| /* Case: call/return (==boring) transfer to any address */ |
| switch (jk) { |
| case Ijk_Boring: case Ijk_Ret: case Ijk_Call: { |
| HReg r = iselWordExpr_R(env, next); |
| PPCAMode* amCIA = PPCAMode_IR(offsIP, hregPPC_GPR31(env->mode64)); |
| if (env->chainingAllowed) { |
| addInstr(env, PPCInstr_XIndir(r, amCIA, always)); |
| } else { |
| addInstr(env, PPCInstr_XAssisted(r, amCIA, always, |
| Ijk_Boring)); |
| } |
| return; |
| } |
| default: |
| break; |
| } |
| |
| /* Case: assisted transfer to arbitrary address */ |
| switch (jk) { |
| /* Keep this list in sync with that for Ist_Exit above */ |
| case Ijk_ClientReq: |
| case Ijk_EmFail: |
| case Ijk_EmWarn: |
| case Ijk_NoDecode: |
| case Ijk_NoRedir: |
| case Ijk_SigBUS: |
| case Ijk_SigTRAP: |
| case Ijk_Sys_syscall: |
| case Ijk_TInval: |
| { |
| HReg r = iselWordExpr_R(env, next); |
| PPCAMode* amCIA = PPCAMode_IR(offsIP, hregPPC_GPR31(env->mode64)); |
| addInstr(env, PPCInstr_XAssisted(r, amCIA, always, jk)); |
| return; |
| } |
| default: |
| break; |
| } |
| |
| vex_printf( "\n-- PUT(%d) = ", offsIP); |
| ppIRExpr( next ); |
| vex_printf( "; exit-"); |
| ppIRJumpKind(jk); |
| vex_printf( "\n"); |
| vassert(0); // are we expecting any other kind? |
| } |
| |
| |
| /*---------------------------------------------------------*/ |
| /*--- Insn selector top-level ---*/ |
| /*---------------------------------------------------------*/ |
| |
| /* Translate an entire SB to ppc code. */ |
| HInstrArray* iselSB_PPC ( IRSB* bb, |
| VexArch arch_host, |
| VexArchInfo* archinfo_host, |
| VexAbiInfo* vbi, |
| Int offs_Host_EvC_Counter, |
| Int offs_Host_EvC_FailAddr, |
| Bool chainingAllowed, |
| Bool addProfInc, |
| Addr64 max_ga ) |
| { |
| Int i, j; |
| HReg hregLo, hregMedLo, hregMedHi, hregHi; |
| ISelEnv* env; |
| UInt hwcaps_host = archinfo_host->hwcaps; |
| Bool mode64 = False; |
| UInt mask32, mask64; |
| PPCAMode *amCounter, *amFailAddr; |
| |
| |
| vassert(arch_host == VexArchPPC32 || arch_host == VexArchPPC64); |
| mode64 = arch_host == VexArchPPC64; |
| if (!mode64) vassert(max_ga <= 0xFFFFFFFFULL); |
| |
| /* do some sanity checks */ |
| mask32 = VEX_HWCAPS_PPC32_F | VEX_HWCAPS_PPC32_V |
| | VEX_HWCAPS_PPC32_FX | VEX_HWCAPS_PPC32_GX | VEX_HWCAPS_PPC32_VX |
| | VEX_HWCAPS_PPC32_DFP | VEX_HWCAPS_PPC32_ISA2_07; |
| |
| |
| mask64 = VEX_HWCAPS_PPC64_V | VEX_HWCAPS_PPC64_FX |
| | VEX_HWCAPS_PPC64_GX | VEX_HWCAPS_PPC64_VX | VEX_HWCAPS_PPC64_DFP |
| | VEX_HWCAPS_PPC64_ISA2_07; |
| |
| if (mode64) { |
| vassert((hwcaps_host & mask32) == 0); |
| } else { |
| vassert((hwcaps_host & mask64) == 0); |
| } |
| |
| /* Make up an initial environment to use. */ |
| env = LibVEX_Alloc(sizeof(ISelEnv)); |
| env->vreg_ctr = 0; |
| |
| /* Are we being ppc32 or ppc64? */ |
| env->mode64 = mode64; |
| |
| /* Set up output code array. */ |
| env->code = newHInstrArray(); |
| |
| /* Copy BB's type env. */ |
| env->type_env = bb->tyenv; |
| |
| /* Make up an IRTemp -> virtual HReg mapping. This doesn't |
| * change as we go along. |
| * |
| * vregmap2 and vregmap3 are only used in 32 bit mode |
| * for supporting I128 in 32-bit mode |
| */ |
| env->n_vregmap = bb->tyenv->types_used; |
| env->vregmapLo = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); |
| env->vregmapMedLo = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); |
| if (mode64) { |
| env->vregmapMedHi = NULL; |
| env->vregmapHi = NULL; |
| } else { |
| env->vregmapMedHi = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); |
| env->vregmapHi = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); |
| } |
| |
| /* and finally ... */ |
| env->chainingAllowed = chainingAllowed; |
| env->max_ga = max_ga; |
| env->hwcaps = hwcaps_host; |
| env->previous_rm = NULL; |
| env->vbi = vbi; |
| |
| /* For each IR temporary, allocate a suitably-kinded virtual |
| register. */ |
| j = 0; |
| for (i = 0; i < env->n_vregmap; i++) { |
| hregLo = hregMedLo = hregMedHi = hregHi = INVALID_HREG; |
| switch (bb->tyenv->types[i]) { |
| case Ity_I1: |
| case Ity_I8: |
| case Ity_I16: |
| case Ity_I32: |
| if (mode64) { hregLo = mkHReg(j++, HRcInt64, True); break; |
| } else { hregLo = mkHReg(j++, HRcInt32, True); break; |
| } |
| case Ity_I64: |
| if (mode64) { hregLo = mkHReg(j++, HRcInt64, True); break; |
| } else { hregLo = mkHReg(j++, HRcInt32, True); |
| hregMedLo = mkHReg(j++, HRcInt32, True); break; |
| } |
| case Ity_I128: |
| if (mode64) { hregLo = mkHReg(j++, HRcInt64, True); |
| hregMedLo = mkHReg(j++, HRcInt64, True); break; |
| } else { hregLo = mkHReg(j++, HRcInt32, True); |
| hregMedLo = mkHReg(j++, HRcInt32, True); |
| hregMedHi = mkHReg(j++, HRcInt32, True); |
| hregHi = mkHReg(j++, HRcInt32, True); break; |
| } |
| case Ity_F32: |
| case Ity_F64: hregLo = mkHReg(j++, HRcFlt64, True); break; |
| case Ity_V128: hregLo = mkHReg(j++, HRcVec128, True); break; |
| case Ity_D32: |
| case Ity_D64: hregLo = mkHReg(j++, HRcFlt64, True); break; |
| case Ity_D128: hregLo = mkHReg(j++, HRcFlt64, True); |
| hregMedLo = mkHReg(j++, HRcFlt64, True); break; |
| default: |
| ppIRType(bb->tyenv->types[i]); |
| vpanic("iselBB(ppc): IRTemp type"); |
| } |
| env->vregmapLo[i] = hregLo; |
| env->vregmapMedLo[i] = hregMedLo; |
| if (!mode64) { |
| env->vregmapMedHi[i] = hregMedHi; |
| env->vregmapHi[i] = hregHi; |
| } |
| } |
| env->vreg_ctr = j; |
| |
| /* The very first instruction must be an event check. */ |
| amCounter = PPCAMode_IR(offs_Host_EvC_Counter, hregPPC_GPR31(mode64)); |
| amFailAddr = PPCAMode_IR(offs_Host_EvC_FailAddr, hregPPC_GPR31(mode64)); |
| addInstr(env, PPCInstr_EvCheck(amCounter, amFailAddr)); |
| |
| /* Possibly a block counter increment (for profiling). At this |
| point we don't know the address of the counter, so just pretend |
| it is zero. It will have to be patched later, but before this |
| translation is used, by a call to LibVEX_patchProfCtr. */ |
| if (addProfInc) { |
| addInstr(env, PPCInstr_ProfInc()); |
| } |
| |
| /* Ok, finally we can iterate over the statements. */ |
| for (i = 0; i < bb->stmts_used; i++) |
| iselStmt(env, bb->stmts[i]); |
| |
| iselNext(env, bb->next, bb->jumpkind, bb->offsIP); |
| |
| /* record the number of vregs we used. */ |
| env->code->n_vregs = env->vreg_ctr; |
| return env->code; |
| } |
| |
| |
| /*---------------------------------------------------------------*/ |
| /*--- end host_ppc_isel.c ---*/ |
| /*---------------------------------------------------------------*/ |