|  | //===- SIInstrInfo.h - SI Instruction Info Interface ------------*- C++ -*-===// | 
|  | // | 
|  | //                     The LLVM Compiler Infrastructure | 
|  | // | 
|  | // This file is distributed under the University of Illinois Open Source | 
|  | // License. See LICENSE.TXT for details. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  | // | 
|  | /// \file | 
|  | /// Interface definition for SIInstrInfo. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #ifndef LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H | 
|  | #define LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H | 
|  |  | 
|  | #include "AMDGPUInstrInfo.h" | 
|  | #include "SIDefines.h" | 
|  | #include "SIRegisterInfo.h" | 
|  | #include "Utils/AMDGPUBaseInfo.h" | 
|  | #include "llvm/ADT/ArrayRef.h" | 
|  | #include "llvm/ADT/SetVector.h" | 
|  | #include "llvm/CodeGen/MachineBasicBlock.h" | 
|  | #include "llvm/CodeGen/MachineFunction.h" | 
|  | #include "llvm/CodeGen/MachineInstr.h" | 
|  | #include "llvm/CodeGen/MachineInstrBuilder.h" | 
|  | #include "llvm/CodeGen/MachineOperand.h" | 
|  | #include "llvm/MC/MCInstrDesc.h" | 
|  | #include "llvm/Support/Compiler.h" | 
|  | #include <cassert> | 
|  | #include <cstdint> | 
|  |  | 
|  | #define GET_INSTRINFO_HEADER | 
|  | #include "AMDGPUGenInstrInfo.inc" | 
|  |  | 
|  | namespace llvm { | 
|  |  | 
|  | class APInt; | 
|  | class MachineRegisterInfo; | 
|  | class RegScavenger; | 
|  | class SISubtarget; | 
|  | class TargetRegisterClass; | 
|  |  | 
|  | class SIInstrInfo final : public AMDGPUGenInstrInfo { | 
|  | private: | 
|  | const SIRegisterInfo RI; | 
|  | const SISubtarget &ST; | 
|  |  | 
|  | // The inverse predicate should have the negative value. | 
|  | enum BranchPredicate { | 
|  | INVALID_BR = 0, | 
|  | SCC_TRUE = 1, | 
|  | SCC_FALSE = -1, | 
|  | VCCNZ = 2, | 
|  | VCCZ = -2, | 
|  | EXECNZ = -3, | 
|  | EXECZ = 3 | 
|  | }; | 
|  |  | 
|  | using SetVectorType = SmallSetVector<MachineInstr *, 32>; | 
|  |  | 
|  | static unsigned getBranchOpcode(BranchPredicate Cond); | 
|  | static BranchPredicate getBranchPredicate(unsigned Opcode); | 
|  |  | 
|  | public: | 
|  | unsigned buildExtractSubReg(MachineBasicBlock::iterator MI, | 
|  | MachineRegisterInfo &MRI, | 
|  | MachineOperand &SuperReg, | 
|  | const TargetRegisterClass *SuperRC, | 
|  | unsigned SubIdx, | 
|  | const TargetRegisterClass *SubRC) const; | 
|  | MachineOperand buildExtractSubRegOrImm(MachineBasicBlock::iterator MI, | 
|  | MachineRegisterInfo &MRI, | 
|  | MachineOperand &SuperReg, | 
|  | const TargetRegisterClass *SuperRC, | 
|  | unsigned SubIdx, | 
|  | const TargetRegisterClass *SubRC) const; | 
|  | private: | 
|  | void swapOperands(MachineInstr &Inst) const; | 
|  |  | 
|  | bool moveScalarAddSub(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  |  | 
|  | void lowerScalarAbs(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  |  | 
|  | void lowerScalarXnor(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  |  | 
|  | void splitScalar64BitUnaryOp(SetVectorType &Worklist, | 
|  | MachineInstr &Inst, unsigned Opcode) const; | 
|  |  | 
|  | void splitScalar64BitAddSub(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  |  | 
|  | void splitScalar64BitBinaryOp(SetVectorType &Worklist, | 
|  | MachineInstr &Inst, unsigned Opcode) const; | 
|  |  | 
|  | void splitScalar64BitBCNT(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  | void splitScalar64BitBFE(SetVectorType &Worklist, | 
|  | MachineInstr &Inst) const; | 
|  | void movePackToVALU(SetVectorType &Worklist, | 
|  | MachineRegisterInfo &MRI, | 
|  | MachineInstr &Inst) const; | 
|  |  | 
|  | void addUsersToMoveToVALUWorklist(unsigned Reg, MachineRegisterInfo &MRI, | 
|  | SetVectorType &Worklist) const; | 
|  |  | 
|  | void | 
|  | addSCCDefUsersToVALUWorklist(MachineInstr &SCCDefInst, | 
|  | SetVectorType &Worklist) const; | 
|  |  | 
|  | const TargetRegisterClass * | 
|  | getDestEquivalentVGPRClass(const MachineInstr &Inst) const; | 
|  |  | 
|  | bool checkInstOffsetsDoNotOverlap(MachineInstr &MIa, MachineInstr &MIb) const; | 
|  |  | 
|  | unsigned findUsedSGPR(const MachineInstr &MI, int OpIndices[3]) const; | 
|  |  | 
|  | protected: | 
|  | bool swapSourceModifiers(MachineInstr &MI, | 
|  | MachineOperand &Src0, unsigned Src0OpName, | 
|  | MachineOperand &Src1, unsigned Src1OpName) const; | 
|  |  | 
|  | MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI, | 
|  | unsigned OpIdx0, | 
|  | unsigned OpIdx1) const override; | 
|  |  | 
|  | public: | 
|  | enum TargetOperandFlags { | 
|  | MO_MASK = 0x7, | 
|  |  | 
|  | MO_NONE = 0, | 
|  | // MO_GOTPCREL -> symbol@GOTPCREL -> R_AMDGPU_GOTPCREL. | 
|  | MO_GOTPCREL = 1, | 
|  | // MO_GOTPCREL32_LO -> symbol@gotpcrel32@lo -> R_AMDGPU_GOTPCREL32_LO. | 
|  | MO_GOTPCREL32 = 2, | 
|  | MO_GOTPCREL32_LO = 2, | 
|  | // MO_GOTPCREL32_HI -> symbol@gotpcrel32@hi -> R_AMDGPU_GOTPCREL32_HI. | 
|  | MO_GOTPCREL32_HI = 3, | 
|  | // MO_REL32_LO -> symbol@rel32@lo -> R_AMDGPU_REL32_LO. | 
|  | MO_REL32 = 4, | 
|  | MO_REL32_LO = 4, | 
|  | // MO_REL32_HI -> symbol@rel32@hi -> R_AMDGPU_REL32_HI. | 
|  | MO_REL32_HI = 5 | 
|  | }; | 
|  |  | 
|  | explicit SIInstrInfo(const SISubtarget &ST); | 
|  |  | 
|  | const SIRegisterInfo &getRegisterInfo() const { | 
|  | return RI; | 
|  | } | 
|  |  | 
|  | bool isReallyTriviallyReMaterializable(const MachineInstr &MI, | 
|  | AliasAnalysis *AA) const override; | 
|  |  | 
|  | bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, | 
|  | int64_t &Offset1, | 
|  | int64_t &Offset2) const override; | 
|  |  | 
|  | bool getMemOpBaseRegImmOfs(MachineInstr &LdSt, unsigned &BaseReg, | 
|  | int64_t &Offset, | 
|  | const TargetRegisterInfo *TRI) const final; | 
|  |  | 
|  | bool shouldClusterMemOps(MachineInstr &FirstLdSt, unsigned BaseReg1, | 
|  | MachineInstr &SecondLdSt, unsigned BaseReg2, | 
|  | unsigned NumLoads) const override; | 
|  |  | 
|  | bool shouldScheduleLoadsNear(SDNode *Load0, SDNode *Load1, int64_t Offset0, | 
|  | int64_t Offset1, unsigned NumLoads) const override; | 
|  |  | 
|  | void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, | 
|  | const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, | 
|  | bool KillSrc) const override; | 
|  |  | 
|  | unsigned calculateLDSSpillAddress(MachineBasicBlock &MBB, MachineInstr &MI, | 
|  | RegScavenger *RS, unsigned TmpReg, | 
|  | unsigned Offset, unsigned Size) const; | 
|  |  | 
|  | void materializeImmediate(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MI, | 
|  | const DebugLoc &DL, | 
|  | unsigned DestReg, | 
|  | int64_t Value) const; | 
|  |  | 
|  | const TargetRegisterClass *getPreferredSelectRegClass( | 
|  | unsigned Size) const; | 
|  |  | 
|  | unsigned insertNE(MachineBasicBlock *MBB, | 
|  | MachineBasicBlock::iterator I, const DebugLoc &DL, | 
|  | unsigned SrcReg, int Value) const; | 
|  |  | 
|  | unsigned insertEQ(MachineBasicBlock *MBB, | 
|  | MachineBasicBlock::iterator I, const DebugLoc &DL, | 
|  | unsigned SrcReg, int Value)  const; | 
|  |  | 
|  | void storeRegToStackSlot(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MI, unsigned SrcReg, | 
|  | bool isKill, int FrameIndex, | 
|  | const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const override; | 
|  |  | 
|  | void loadRegFromStackSlot(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MI, unsigned DestReg, | 
|  | int FrameIndex, const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const override; | 
|  |  | 
|  | bool expandPostRAPseudo(MachineInstr &MI) const override; | 
|  |  | 
|  | // Returns an opcode that can be used to move a value to a \p DstRC | 
|  | // register.  If there is no hardware instruction that can store to \p | 
|  | // DstRC, then AMDGPU::COPY is returned. | 
|  | unsigned getMovOpcode(const TargetRegisterClass *DstRC) const; | 
|  |  | 
|  | LLVM_READONLY | 
|  | int commuteOpcode(unsigned Opc) const; | 
|  |  | 
|  | LLVM_READONLY | 
|  | inline int commuteOpcode(const MachineInstr &MI) const { | 
|  | return commuteOpcode(MI.getOpcode()); | 
|  | } | 
|  |  | 
|  | bool findCommutedOpIndices(MachineInstr &MI, unsigned &SrcOpIdx1, | 
|  | unsigned &SrcOpIdx2) const override; | 
|  |  | 
|  | bool isBranchOffsetInRange(unsigned BranchOpc, | 
|  | int64_t BrOffset) const override; | 
|  |  | 
|  | MachineBasicBlock *getBranchDestBlock(const MachineInstr &MI) const override; | 
|  |  | 
|  | unsigned insertIndirectBranch(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock &NewDestBB, | 
|  | const DebugLoc &DL, | 
|  | int64_t BrOffset, | 
|  | RegScavenger *RS = nullptr) const override; | 
|  |  | 
|  | bool analyzeBranchImpl(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator I, | 
|  | MachineBasicBlock *&TBB, | 
|  | MachineBasicBlock *&FBB, | 
|  | SmallVectorImpl<MachineOperand> &Cond, | 
|  | bool AllowModify) const; | 
|  |  | 
|  | bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, | 
|  | MachineBasicBlock *&FBB, | 
|  | SmallVectorImpl<MachineOperand> &Cond, | 
|  | bool AllowModify = false) const override; | 
|  |  | 
|  | unsigned removeBranch(MachineBasicBlock &MBB, | 
|  | int *BytesRemoved = nullptr) const override; | 
|  |  | 
|  | unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, | 
|  | MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond, | 
|  | const DebugLoc &DL, | 
|  | int *BytesAdded = nullptr) const override; | 
|  |  | 
|  | bool reverseBranchCondition( | 
|  | SmallVectorImpl<MachineOperand> &Cond) const override; | 
|  |  | 
|  | bool canInsertSelect(const MachineBasicBlock &MBB, | 
|  | ArrayRef<MachineOperand> Cond, | 
|  | unsigned TrueReg, unsigned FalseReg, | 
|  | int &CondCycles, | 
|  | int &TrueCycles, int &FalseCycles) const override; | 
|  |  | 
|  | void insertSelect(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator I, const DebugLoc &DL, | 
|  | unsigned DstReg, ArrayRef<MachineOperand> Cond, | 
|  | unsigned TrueReg, unsigned FalseReg) const override; | 
|  |  | 
|  | void insertVectorSelect(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator I, const DebugLoc &DL, | 
|  | unsigned DstReg, ArrayRef<MachineOperand> Cond, | 
|  | unsigned TrueReg, unsigned FalseReg) const; | 
|  |  | 
|  | unsigned getAddressSpaceForPseudoSourceKind( | 
|  | PseudoSourceValue::PSVKind Kind) const override; | 
|  |  | 
|  | bool | 
|  | areMemAccessesTriviallyDisjoint(MachineInstr &MIa, MachineInstr &MIb, | 
|  | AliasAnalysis *AA = nullptr) const override; | 
|  |  | 
|  | bool isFoldableCopy(const MachineInstr &MI) const; | 
|  |  | 
|  | bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, unsigned Reg, | 
|  | MachineRegisterInfo *MRI) const final; | 
|  |  | 
|  | unsigned getMachineCSELookAheadLimit() const override { return 500; } | 
|  |  | 
|  | MachineInstr *convertToThreeAddress(MachineFunction::iterator &MBB, | 
|  | MachineInstr &MI, | 
|  | LiveVariables *LV) const override; | 
|  |  | 
|  | bool isSchedulingBoundary(const MachineInstr &MI, | 
|  | const MachineBasicBlock *MBB, | 
|  | const MachineFunction &MF) const override; | 
|  |  | 
|  | static bool isSALU(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SALU; | 
|  | } | 
|  |  | 
|  | bool isSALU(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SALU; | 
|  | } | 
|  |  | 
|  | static bool isVALU(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VALU; | 
|  | } | 
|  |  | 
|  | bool isVALU(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VALU; | 
|  | } | 
|  |  | 
|  | static bool isVMEM(const MachineInstr &MI) { | 
|  | return isMUBUF(MI) || isMTBUF(MI) || isMIMG(MI); | 
|  | } | 
|  |  | 
|  | bool isVMEM(uint16_t Opcode) const { | 
|  | return isMUBUF(Opcode) || isMTBUF(Opcode) || isMIMG(Opcode); | 
|  | } | 
|  |  | 
|  | static bool isSOP1(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOP1; | 
|  | } | 
|  |  | 
|  | bool isSOP1(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOP1; | 
|  | } | 
|  |  | 
|  | static bool isSOP2(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOP2; | 
|  | } | 
|  |  | 
|  | bool isSOP2(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOP2; | 
|  | } | 
|  |  | 
|  | static bool isSOPC(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOPC; | 
|  | } | 
|  |  | 
|  | bool isSOPC(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOPC; | 
|  | } | 
|  |  | 
|  | static bool isSOPK(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOPK; | 
|  | } | 
|  |  | 
|  | bool isSOPK(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOPK; | 
|  | } | 
|  |  | 
|  | static bool isSOPP(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOPP; | 
|  | } | 
|  |  | 
|  | bool isSOPP(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOPP; | 
|  | } | 
|  |  | 
|  | static bool isVOP1(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VOP1; | 
|  | } | 
|  |  | 
|  | bool isVOP1(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VOP1; | 
|  | } | 
|  |  | 
|  | static bool isVOP2(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VOP2; | 
|  | } | 
|  |  | 
|  | bool isVOP2(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VOP2; | 
|  | } | 
|  |  | 
|  | static bool isVOP3(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VOP3; | 
|  | } | 
|  |  | 
|  | bool isVOP3(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VOP3; | 
|  | } | 
|  |  | 
|  | static bool isSDWA(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SDWA; | 
|  | } | 
|  |  | 
|  | bool isSDWA(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SDWA; | 
|  | } | 
|  |  | 
|  | static bool isVOPC(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VOPC; | 
|  | } | 
|  |  | 
|  | bool isVOPC(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VOPC; | 
|  | } | 
|  |  | 
|  | static bool isMUBUF(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::MUBUF; | 
|  | } | 
|  |  | 
|  | bool isMUBUF(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::MUBUF; | 
|  | } | 
|  |  | 
|  | static bool isMTBUF(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::MTBUF; | 
|  | } | 
|  |  | 
|  | bool isMTBUF(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::MTBUF; | 
|  | } | 
|  |  | 
|  | static bool isSMRD(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SMRD; | 
|  | } | 
|  |  | 
|  | bool isSMRD(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SMRD; | 
|  | } | 
|  |  | 
|  | bool isBufferSMRD(const MachineInstr &MI) const; | 
|  |  | 
|  | static bool isDS(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::DS; | 
|  | } | 
|  |  | 
|  | bool isDS(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::DS; | 
|  | } | 
|  |  | 
|  | static bool isMIMG(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::MIMG; | 
|  | } | 
|  |  | 
|  | bool isMIMG(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::MIMG; | 
|  | } | 
|  |  | 
|  | static bool isGather4(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::Gather4; | 
|  | } | 
|  |  | 
|  | bool isGather4(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::Gather4; | 
|  | } | 
|  |  | 
|  | static bool isFLAT(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::FLAT; | 
|  | } | 
|  |  | 
|  | // Is a FLAT encoded instruction which accesses a specific segment, | 
|  | // i.e. global_* or scratch_*. | 
|  | static bool isSegmentSpecificFLAT(const MachineInstr &MI) { | 
|  | auto Flags = MI.getDesc().TSFlags; | 
|  | return (Flags & SIInstrFlags::FLAT) && !(Flags & SIInstrFlags::LGKM_CNT); | 
|  | } | 
|  |  | 
|  | // Any FLAT encoded instruction, including global_* and scratch_*. | 
|  | bool isFLAT(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::FLAT; | 
|  | } | 
|  |  | 
|  | static bool isEXP(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::EXP; | 
|  | } | 
|  |  | 
|  | bool isEXP(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::EXP; | 
|  | } | 
|  |  | 
|  | static bool isWQM(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::WQM; | 
|  | } | 
|  |  | 
|  | bool isWQM(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::WQM; | 
|  | } | 
|  |  | 
|  | static bool isDisableWQM(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::DisableWQM; | 
|  | } | 
|  |  | 
|  | bool isDisableWQM(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::DisableWQM; | 
|  | } | 
|  |  | 
|  | static bool isVGPRSpill(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VGPRSpill; | 
|  | } | 
|  |  | 
|  | bool isVGPRSpill(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VGPRSpill; | 
|  | } | 
|  |  | 
|  | static bool isSGPRSpill(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SGPRSpill; | 
|  | } | 
|  |  | 
|  | bool isSGPRSpill(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SGPRSpill; | 
|  | } | 
|  |  | 
|  | static bool isDPP(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::DPP; | 
|  | } | 
|  |  | 
|  | bool isDPP(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::DPP; | 
|  | } | 
|  |  | 
|  | static bool isVOP3P(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VOP3P; | 
|  | } | 
|  |  | 
|  | bool isVOP3P(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VOP3P; | 
|  | } | 
|  |  | 
|  | static bool isVINTRP(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VINTRP; | 
|  | } | 
|  |  | 
|  | bool isVINTRP(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::VINTRP; | 
|  | } | 
|  |  | 
|  | static bool isScalarUnit(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & (SIInstrFlags::SALU | SIInstrFlags::SMRD); | 
|  | } | 
|  |  | 
|  | static bool usesVM_CNT(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::VM_CNT; | 
|  | } | 
|  |  | 
|  | static bool usesLGKM_CNT(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::LGKM_CNT; | 
|  | } | 
|  |  | 
|  | static bool sopkIsZext(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SOPK_ZEXT; | 
|  | } | 
|  |  | 
|  | bool sopkIsZext(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SOPK_ZEXT; | 
|  | } | 
|  |  | 
|  | /// \returns true if this is an s_store_dword* instruction. This is more | 
|  | /// specific than than isSMEM && mayStore. | 
|  | static bool isScalarStore(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::SCALAR_STORE; | 
|  | } | 
|  |  | 
|  | bool isScalarStore(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::SCALAR_STORE; | 
|  | } | 
|  |  | 
|  | static bool isFixedSize(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::FIXED_SIZE; | 
|  | } | 
|  |  | 
|  | bool isFixedSize(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::FIXED_SIZE; | 
|  | } | 
|  |  | 
|  | static bool hasFPClamp(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::FPClamp; | 
|  | } | 
|  |  | 
|  | bool hasFPClamp(uint16_t Opcode) const { | 
|  | return get(Opcode).TSFlags & SIInstrFlags::FPClamp; | 
|  | } | 
|  |  | 
|  | static bool hasIntClamp(const MachineInstr &MI) { | 
|  | return MI.getDesc().TSFlags & SIInstrFlags::IntClamp; | 
|  | } | 
|  |  | 
|  | uint64_t getClampMask(const MachineInstr &MI) const { | 
|  | const uint64_t ClampFlags = SIInstrFlags::FPClamp | | 
|  | SIInstrFlags::IntClamp | | 
|  | SIInstrFlags::ClampLo | | 
|  | SIInstrFlags::ClampHi; | 
|  | return MI.getDesc().TSFlags & ClampFlags; | 
|  | } | 
|  |  | 
|  | bool isVGPRCopy(const MachineInstr &MI) const { | 
|  | assert(MI.isCopy()); | 
|  | unsigned Dest = MI.getOperand(0).getReg(); | 
|  | const MachineFunction &MF = *MI.getParent()->getParent(); | 
|  | const MachineRegisterInfo &MRI = MF.getRegInfo(); | 
|  | return !RI.isSGPRReg(MRI, Dest); | 
|  | } | 
|  |  | 
|  | bool isInlineConstant(const APInt &Imm) const; | 
|  |  | 
|  | bool isInlineConstant(const MachineOperand &MO, uint8_t OperandType) const; | 
|  |  | 
|  | bool isInlineConstant(const MachineOperand &MO, | 
|  | const MCOperandInfo &OpInfo) const { | 
|  | return isInlineConstant(MO, OpInfo.OperandType); | 
|  | } | 
|  |  | 
|  | /// \p returns true if \p UseMO is substituted with \p DefMO in \p MI it would | 
|  | /// be an inline immediate. | 
|  | bool isInlineConstant(const MachineInstr &MI, | 
|  | const MachineOperand &UseMO, | 
|  | const MachineOperand &DefMO) const { | 
|  | assert(UseMO.getParent() == &MI); | 
|  | int OpIdx = MI.getOperandNo(&UseMO); | 
|  | if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands) { | 
|  | return false; | 
|  | } | 
|  |  | 
|  | return isInlineConstant(DefMO, MI.getDesc().OpInfo[OpIdx]); | 
|  | } | 
|  |  | 
|  | /// \p returns true if the operand \p OpIdx in \p MI is a valid inline | 
|  | /// immediate. | 
|  | bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx) const { | 
|  | const MachineOperand &MO = MI.getOperand(OpIdx); | 
|  | return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType); | 
|  | } | 
|  |  | 
|  | bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx, | 
|  | const MachineOperand &MO) const { | 
|  | if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands) | 
|  | return false; | 
|  |  | 
|  | if (MI.isCopy()) { | 
|  | unsigned Size = getOpSize(MI, OpIdx); | 
|  | assert(Size == 8 || Size == 4); | 
|  |  | 
|  | uint8_t OpType = (Size == 8) ? | 
|  | AMDGPU::OPERAND_REG_IMM_INT64 : AMDGPU::OPERAND_REG_IMM_INT32; | 
|  | return isInlineConstant(MO, OpType); | 
|  | } | 
|  |  | 
|  | return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType); | 
|  | } | 
|  |  | 
|  | bool isInlineConstant(const MachineOperand &MO) const { | 
|  | const MachineInstr *Parent = MO.getParent(); | 
|  | return isInlineConstant(*Parent, Parent->getOperandNo(&MO)); | 
|  | } | 
|  |  | 
|  | bool isLiteralConstant(const MachineOperand &MO, | 
|  | const MCOperandInfo &OpInfo) const { | 
|  | return MO.isImm() && !isInlineConstant(MO, OpInfo.OperandType); | 
|  | } | 
|  |  | 
|  | bool isLiteralConstant(const MachineInstr &MI, int OpIdx) const { | 
|  | const MachineOperand &MO = MI.getOperand(OpIdx); | 
|  | return MO.isImm() && !isInlineConstant(MI, OpIdx); | 
|  | } | 
|  |  | 
|  | // Returns true if this operand could potentially require a 32-bit literal | 
|  | // operand, but not necessarily. A FrameIndex for example could resolve to an | 
|  | // inline immediate value that will not require an additional 4-bytes; this | 
|  | // assumes that it will. | 
|  | bool isLiteralConstantLike(const MachineOperand &MO, | 
|  | const MCOperandInfo &OpInfo) const; | 
|  |  | 
|  | bool isImmOperandLegal(const MachineInstr &MI, unsigned OpNo, | 
|  | const MachineOperand &MO) const; | 
|  |  | 
|  | /// Return true if this 64-bit VALU instruction has a 32-bit encoding. | 
|  | /// This function will return false if you pass it a 32-bit instruction. | 
|  | bool hasVALU32BitEncoding(unsigned Opcode) const; | 
|  |  | 
|  | /// Returns true if this operand uses the constant bus. | 
|  | bool usesConstantBus(const MachineRegisterInfo &MRI, | 
|  | const MachineOperand &MO, | 
|  | const MCOperandInfo &OpInfo) const; | 
|  |  | 
|  | /// Return true if this instruction has any modifiers. | 
|  | ///  e.g. src[012]_mod, omod, clamp. | 
|  | bool hasModifiers(unsigned Opcode) const; | 
|  |  | 
|  | bool hasModifiersSet(const MachineInstr &MI, | 
|  | unsigned OpName) const; | 
|  | bool hasAnyModifiersSet(const MachineInstr &MI) const; | 
|  |  | 
|  | bool verifyInstruction(const MachineInstr &MI, | 
|  | StringRef &ErrInfo) const override; | 
|  |  | 
|  | unsigned getVALUOp(const MachineInstr &MI) const; | 
|  |  | 
|  | /// Return the correct register class for \p OpNo.  For target-specific | 
|  | /// instructions, this will return the register class that has been defined | 
|  | /// in tablegen.  For generic instructions, like REG_SEQUENCE it will return | 
|  | /// the register class of its machine operand. | 
|  | /// to infer the correct register class base on the other operands. | 
|  | const TargetRegisterClass *getOpRegClass(const MachineInstr &MI, | 
|  | unsigned OpNo) const; | 
|  |  | 
|  | /// Return the size in bytes of the operand OpNo on the given | 
|  | // instruction opcode. | 
|  | unsigned getOpSize(uint16_t Opcode, unsigned OpNo) const { | 
|  | const MCOperandInfo &OpInfo = get(Opcode).OpInfo[OpNo]; | 
|  |  | 
|  | if (OpInfo.RegClass == -1) { | 
|  | // If this is an immediate operand, this must be a 32-bit literal. | 
|  | assert(OpInfo.OperandType == MCOI::OPERAND_IMMEDIATE); | 
|  | return 4; | 
|  | } | 
|  |  | 
|  | return RI.getRegSizeInBits(*RI.getRegClass(OpInfo.RegClass)) / 8; | 
|  | } | 
|  |  | 
|  | /// This form should usually be preferred since it handles operands | 
|  | /// with unknown register classes. | 
|  | unsigned getOpSize(const MachineInstr &MI, unsigned OpNo) const { | 
|  | return RI.getRegSizeInBits(*getOpRegClass(MI, OpNo)) / 8; | 
|  | } | 
|  |  | 
|  | /// \returns true if it is legal for the operand at index \p OpNo | 
|  | /// to read a VGPR. | 
|  | bool canReadVGPR(const MachineInstr &MI, unsigned OpNo) const; | 
|  |  | 
|  | /// Legalize the \p OpIndex operand of this instruction by inserting | 
|  | /// a MOV.  For example: | 
|  | /// ADD_I32_e32 VGPR0, 15 | 
|  | /// to | 
|  | /// MOV VGPR1, 15 | 
|  | /// ADD_I32_e32 VGPR0, VGPR1 | 
|  | /// | 
|  | /// If the operand being legalized is a register, then a COPY will be used | 
|  | /// instead of MOV. | 
|  | void legalizeOpWithMove(MachineInstr &MI, unsigned OpIdx) const; | 
|  |  | 
|  | /// Check if \p MO is a legal operand if it was the \p OpIdx Operand | 
|  | /// for \p MI. | 
|  | bool isOperandLegal(const MachineInstr &MI, unsigned OpIdx, | 
|  | const MachineOperand *MO = nullptr) const; | 
|  |  | 
|  | /// Check if \p MO would be a valid operand for the given operand | 
|  | /// definition \p OpInfo. Note this does not attempt to validate constant bus | 
|  | /// restrictions (e.g. literal constant usage). | 
|  | bool isLegalVSrcOperand(const MachineRegisterInfo &MRI, | 
|  | const MCOperandInfo &OpInfo, | 
|  | const MachineOperand &MO) const; | 
|  |  | 
|  | /// Check if \p MO (a register operand) is a legal register for the | 
|  | /// given operand description. | 
|  | bool isLegalRegOperand(const MachineRegisterInfo &MRI, | 
|  | const MCOperandInfo &OpInfo, | 
|  | const MachineOperand &MO) const; | 
|  |  | 
|  | /// Legalize operands in \p MI by either commuting it or inserting a | 
|  | /// copy of src1. | 
|  | void legalizeOperandsVOP2(MachineRegisterInfo &MRI, MachineInstr &MI) const; | 
|  |  | 
|  | /// Fix operands in \p MI to satisfy constant bus requirements. | 
|  | void legalizeOperandsVOP3(MachineRegisterInfo &MRI, MachineInstr &MI) const; | 
|  |  | 
|  | /// Copy a value from a VGPR (\p SrcReg) to SGPR.  This function can only | 
|  | /// be used when it is know that the value in SrcReg is same across all | 
|  | /// threads in the wave. | 
|  | /// \returns The SGPR register that \p SrcReg was copied to. | 
|  | unsigned readlaneVGPRToSGPR(unsigned SrcReg, MachineInstr &UseMI, | 
|  | MachineRegisterInfo &MRI) const; | 
|  |  | 
|  | void legalizeOperandsSMRD(MachineRegisterInfo &MRI, MachineInstr &MI) const; | 
|  |  | 
|  | void legalizeGenericOperand(MachineBasicBlock &InsertMBB, | 
|  | MachineBasicBlock::iterator I, | 
|  | const TargetRegisterClass *DstRC, | 
|  | MachineOperand &Op, MachineRegisterInfo &MRI, | 
|  | const DebugLoc &DL) const; | 
|  |  | 
|  | /// Legalize all operands in this instruction.  This function may | 
|  | /// create new instruction and insert them before \p MI. | 
|  | void legalizeOperands(MachineInstr &MI) const; | 
|  |  | 
|  | /// Replace this instruction's opcode with the equivalent VALU | 
|  | /// opcode.  This function will also move the users of \p MI to the | 
|  | /// VALU if necessary. | 
|  | void moveToVALU(MachineInstr &MI) const; | 
|  |  | 
|  | void insertWaitStates(MachineBasicBlock &MBB,MachineBasicBlock::iterator MI, | 
|  | int Count) const; | 
|  |  | 
|  | void insertNoop(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MI) const override; | 
|  |  | 
|  | void insertReturn(MachineBasicBlock &MBB) const; | 
|  | /// Return the number of wait states that result from executing this | 
|  | /// instruction. | 
|  | unsigned getNumWaitStates(const MachineInstr &MI) const; | 
|  |  | 
|  | /// Returns the operand named \p Op.  If \p MI does not have an | 
|  | /// operand named \c Op, this function returns nullptr. | 
|  | LLVM_READONLY | 
|  | MachineOperand *getNamedOperand(MachineInstr &MI, unsigned OperandName) const; | 
|  |  | 
|  | LLVM_READONLY | 
|  | const MachineOperand *getNamedOperand(const MachineInstr &MI, | 
|  | unsigned OpName) const { | 
|  | return getNamedOperand(const_cast<MachineInstr &>(MI), OpName); | 
|  | } | 
|  |  | 
|  | /// Get required immediate operand | 
|  | int64_t getNamedImmOperand(const MachineInstr &MI, unsigned OpName) const { | 
|  | int Idx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), OpName); | 
|  | return MI.getOperand(Idx).getImm(); | 
|  | } | 
|  |  | 
|  | uint64_t getDefaultRsrcDataFormat() const; | 
|  | uint64_t getScratchRsrcWords23() const; | 
|  |  | 
|  | bool isLowLatencyInstruction(const MachineInstr &MI) const; | 
|  | bool isHighLatencyInstruction(const MachineInstr &MI) const; | 
|  |  | 
|  | /// Return the descriptor of the target-specific machine instruction | 
|  | /// that corresponds to the specified pseudo or native opcode. | 
|  | const MCInstrDesc &getMCOpcodeFromPseudo(unsigned Opcode) const { | 
|  | return get(pseudoToMCOpcode(Opcode)); | 
|  | } | 
|  |  | 
|  | unsigned isStackAccess(const MachineInstr &MI, int &FrameIndex) const; | 
|  | unsigned isSGPRStackAccess(const MachineInstr &MI, int &FrameIndex) const; | 
|  |  | 
|  | unsigned isLoadFromStackSlot(const MachineInstr &MI, | 
|  | int &FrameIndex) const override; | 
|  | unsigned isStoreToStackSlot(const MachineInstr &MI, | 
|  | int &FrameIndex) const override; | 
|  |  | 
|  | unsigned getInstBundleSize(const MachineInstr &MI) const; | 
|  | unsigned getInstSizeInBytes(const MachineInstr &MI) const override; | 
|  |  | 
|  | bool mayAccessFlatAddressSpace(const MachineInstr &MI) const; | 
|  |  | 
|  | bool isNonUniformBranchInstr(MachineInstr &Instr) const; | 
|  |  | 
|  | void convertNonUniformIfRegion(MachineBasicBlock *IfEntry, | 
|  | MachineBasicBlock *IfEnd) const; | 
|  |  | 
|  | void convertNonUniformLoopRegion(MachineBasicBlock *LoopEntry, | 
|  | MachineBasicBlock *LoopEnd) const; | 
|  |  | 
|  | std::pair<unsigned, unsigned> | 
|  | decomposeMachineOperandsTargetFlags(unsigned TF) const override; | 
|  |  | 
|  | ArrayRef<std::pair<int, const char *>> | 
|  | getSerializableTargetIndices() const override; | 
|  |  | 
|  | ArrayRef<std::pair<unsigned, const char *>> | 
|  | getSerializableDirectMachineOperandTargetFlags() const override; | 
|  |  | 
|  | ScheduleHazardRecognizer * | 
|  | CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II, | 
|  | const ScheduleDAG *DAG) const override; | 
|  |  | 
|  | ScheduleHazardRecognizer * | 
|  | CreateTargetPostRAHazardRecognizer(const MachineFunction &MF) const override; | 
|  |  | 
|  | bool isBasicBlockPrologue(const MachineInstr &MI) const override; | 
|  |  | 
|  | /// Return a partially built integer add instruction without carry. | 
|  | /// Caller must add source operands. | 
|  | /// For pre-GFX9 it will generate unused carry destination operand. | 
|  | /// TODO: After GFX9 it should return a no-carry operation. | 
|  | MachineInstrBuilder getAddNoCarry(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator I, | 
|  | const DebugLoc &DL, | 
|  | unsigned DestReg) const; | 
|  |  | 
|  | static bool isKillTerminator(unsigned Opcode); | 
|  | const MCInstrDesc &getKillTerminatorFromPseudo(unsigned Opcode) const; | 
|  |  | 
|  | static bool isLegalMUBUFImmOffset(unsigned Imm) { | 
|  | return isUInt<12>(Imm); | 
|  | } | 
|  |  | 
|  | /// \brief Return a target-specific opcode if Opcode is a pseudo instruction. | 
|  | /// Return -1 if the target-specific opcode for the pseudo instruction does | 
|  | /// not exist. If Opcode is not a pseudo instruction, this is identity. | 
|  | int pseudoToMCOpcode(int Opcode) const; | 
|  |  | 
|  | }; | 
|  |  | 
|  | namespace AMDGPU { | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getVOPe64(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getVOPe32(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getSDWAOp(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getBasicFromSDWAOp(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getCommuteRev(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getCommuteOrig(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getAddr64Inst(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getMUBUFNoLdsInst(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getAtomicRetOp(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getAtomicNoRetOp(uint16_t Opcode); | 
|  |  | 
|  | LLVM_READONLY | 
|  | int getSOPKOp(uint16_t Opcode); | 
|  |  | 
|  | const uint64_t RSRC_DATA_FORMAT = 0xf00000000000LL; | 
|  | const uint64_t RSRC_ELEMENT_SIZE_SHIFT = (32 + 19); | 
|  | const uint64_t RSRC_INDEX_STRIDE_SHIFT = (32 + 21); | 
|  | const uint64_t RSRC_TID_ENABLE = UINT64_C(1) << (32 + 23); | 
|  |  | 
|  | // For MachineOperands. | 
|  | enum TargetFlags { | 
|  | TF_LONG_BRANCH_FORWARD = 1 << 0, | 
|  | TF_LONG_BRANCH_BACKWARD = 1 << 1 | 
|  | }; | 
|  |  | 
|  | } // end namespace AMDGPU | 
|  |  | 
|  | namespace SI { | 
|  | namespace KernelInputOffsets { | 
|  |  | 
|  | /// Offsets in bytes from the start of the input buffer | 
|  | enum Offsets { | 
|  | NGROUPS_X = 0, | 
|  | NGROUPS_Y = 4, | 
|  | NGROUPS_Z = 8, | 
|  | GLOBAL_SIZE_X = 12, | 
|  | GLOBAL_SIZE_Y = 16, | 
|  | GLOBAL_SIZE_Z = 20, | 
|  | LOCAL_SIZE_X = 24, | 
|  | LOCAL_SIZE_Y = 28, | 
|  | LOCAL_SIZE_Z = 32 | 
|  | }; | 
|  |  | 
|  | } // end namespace KernelInputOffsets | 
|  | } // end namespace SI | 
|  |  | 
|  | } // end namespace llvm | 
|  |  | 
|  | #endif // LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H |