|  | //===- AArch64MacroFusion.cpp - AArch64 Macro Fusion ----------------------===// | 
|  | // | 
|  | //                     The LLVM Compiler Infrastructure | 
|  | // | 
|  | // This file is distributed under the University of Illinois Open Source | 
|  | // License. See LICENSE.TXT for details. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  | // | 
|  | /// \file This file contains the AArch64 implementation of the DAG scheduling | 
|  | ///  mutation to pair instructions back to back. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #include "AArch64Subtarget.h" | 
|  | #include "llvm/CodeGen/MacroFusion.h" | 
|  | #include "llvm/CodeGen/TargetInstrInfo.h" | 
|  |  | 
|  | using namespace llvm; | 
|  |  | 
|  | namespace { | 
|  |  | 
|  | // Fuse CMN, CMP, TST followed by Bcc. | 
|  | static bool isArithmeticBccPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | if (SecondMI.getOpcode() == AArch64::Bcc) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | if (!FirstMI) | 
|  | return true; | 
|  |  | 
|  | switch (FirstMI->getOpcode()) { | 
|  | case AArch64::ADDSWri: | 
|  | case AArch64::ADDSWrr: | 
|  | case AArch64::ADDSXri: | 
|  | case AArch64::ADDSXrr: | 
|  | case AArch64::ANDSWri: | 
|  | case AArch64::ANDSWrr: | 
|  | case AArch64::ANDSXri: | 
|  | case AArch64::ANDSXrr: | 
|  | case AArch64::SUBSWri: | 
|  | case AArch64::SUBSWrr: | 
|  | case AArch64::SUBSXri: | 
|  | case AArch64::SUBSXrr: | 
|  | case AArch64::BICSWrr: | 
|  | case AArch64::BICSXrr: | 
|  | return true; | 
|  | case AArch64::ADDSWrs: | 
|  | case AArch64::ADDSXrs: | 
|  | case AArch64::ANDSWrs: | 
|  | case AArch64::ANDSXrs: | 
|  | case AArch64::SUBSWrs: | 
|  | case AArch64::SUBSXrs: | 
|  | case AArch64::BICSWrs: | 
|  | case AArch64::BICSXrs: | 
|  | // Shift value can be 0 making these behave like the "rr" variant... | 
|  | return (!AArch64InstrInfo::hasShiftedReg(*FirstMI)); | 
|  | } | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Fuse ALU operations followed by CBZ/CBNZ. | 
|  | static bool isArithmeticCbzPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | unsigned SecondOpcode = SecondMI.getOpcode(); | 
|  |  | 
|  | if (SecondOpcode == AArch64::CBNZW || SecondOpcode == AArch64::CBNZX || | 
|  | SecondOpcode == AArch64::CBZW  || SecondOpcode == AArch64::CBZX) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | if (!FirstMI) | 
|  | return true; | 
|  |  | 
|  | switch (FirstMI->getOpcode()) { | 
|  | case AArch64::ADDWri: | 
|  | case AArch64::ADDWrr: | 
|  | case AArch64::ADDXri: | 
|  | case AArch64::ADDXrr: | 
|  | case AArch64::ANDWri: | 
|  | case AArch64::ANDWrr: | 
|  | case AArch64::ANDXri: | 
|  | case AArch64::ANDXrr: | 
|  | case AArch64::EORWri: | 
|  | case AArch64::EORWrr: | 
|  | case AArch64::EORXri: | 
|  | case AArch64::EORXrr: | 
|  | case AArch64::ORRWri: | 
|  | case AArch64::ORRWrr: | 
|  | case AArch64::ORRXri: | 
|  | case AArch64::ORRXrr: | 
|  | case AArch64::SUBWri: | 
|  | case AArch64::SUBWrr: | 
|  | case AArch64::SUBXri: | 
|  | case AArch64::SUBXrr: | 
|  | return true; | 
|  | case AArch64::ADDWrs: | 
|  | case AArch64::ADDXrs: | 
|  | case AArch64::ANDWrs: | 
|  | case AArch64::ANDXrs: | 
|  | case AArch64::SUBWrs: | 
|  | case AArch64::SUBXrs: | 
|  | case AArch64::BICWrs: | 
|  | case AArch64::BICXrs: | 
|  | // Shift value can be 0 making these behave like the "rr" variant... | 
|  | return (!AArch64InstrInfo::hasShiftedReg(*FirstMI)); | 
|  | } | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Fuse AES crypto encoding or decoding. | 
|  | static bool isAESPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | unsigned FirstOpcode = | 
|  | FirstMI ? FirstMI->getOpcode() | 
|  | : static_cast<unsigned>(AArch64::INSTRUCTION_LIST_END); | 
|  | unsigned SecondOpcode = SecondMI.getOpcode(); | 
|  |  | 
|  | // AES encode. | 
|  | if ((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | FirstOpcode == AArch64::AESErr) && | 
|  | (SecondOpcode == AArch64::AESMCrr || | 
|  | SecondOpcode == AArch64::AESMCrrTied)) | 
|  | return true; | 
|  | // AES decode. | 
|  | else if ((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | FirstOpcode == AArch64::AESDrr) && | 
|  | (SecondOpcode == AArch64::AESIMCrr || | 
|  | SecondOpcode == AArch64::AESIMCrrTied)) | 
|  | return true; | 
|  |  | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Fuse literal generation. | 
|  | static bool isLiteralsPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | unsigned FirstOpcode = | 
|  | FirstMI ? FirstMI->getOpcode() | 
|  | : static_cast<unsigned>(AArch64::INSTRUCTION_LIST_END); | 
|  | unsigned SecondOpcode = SecondMI.getOpcode(); | 
|  |  | 
|  | // PC relative address. | 
|  | if ((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | FirstOpcode == AArch64::ADRP) && | 
|  | SecondOpcode == AArch64::ADDXri) | 
|  | return true; | 
|  | // 32 bit immediate. | 
|  | else if ((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | FirstOpcode == AArch64::MOVZWi) && | 
|  | (SecondOpcode == AArch64::MOVKWi && | 
|  | SecondMI.getOperand(3).getImm() == 16)) | 
|  | return true; | 
|  | // Lower half of 64 bit immediate. | 
|  | else if((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | FirstOpcode == AArch64::MOVZXi) && | 
|  | (SecondOpcode == AArch64::MOVKXi && | 
|  | SecondMI.getOperand(3).getImm() == 16)) | 
|  | return true; | 
|  | // Upper half of 64 bit immediate. | 
|  | else if ((FirstOpcode == AArch64::INSTRUCTION_LIST_END || | 
|  | (FirstOpcode == AArch64::MOVKXi && | 
|  | FirstMI->getOperand(3).getImm() == 32)) && | 
|  | (SecondOpcode == AArch64::MOVKXi && | 
|  | SecondMI.getOperand(3).getImm() == 48)) | 
|  | return true; | 
|  |  | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Fuse address generation and loads or stores. | 
|  | static bool isAddressLdStPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | unsigned SecondOpcode = SecondMI.getOpcode(); | 
|  |  | 
|  | switch (SecondOpcode) { | 
|  | case AArch64::STRBBui: | 
|  | case AArch64::STRBui: | 
|  | case AArch64::STRDui: | 
|  | case AArch64::STRHHui: | 
|  | case AArch64::STRHui: | 
|  | case AArch64::STRQui: | 
|  | case AArch64::STRSui: | 
|  | case AArch64::STRWui: | 
|  | case AArch64::STRXui: | 
|  | case AArch64::LDRBBui: | 
|  | case AArch64::LDRBui: | 
|  | case AArch64::LDRDui: | 
|  | case AArch64::LDRHHui: | 
|  | case AArch64::LDRHui: | 
|  | case AArch64::LDRQui: | 
|  | case AArch64::LDRSui: | 
|  | case AArch64::LDRWui: | 
|  | case AArch64::LDRXui: | 
|  | case AArch64::LDRSBWui: | 
|  | case AArch64::LDRSBXui: | 
|  | case AArch64::LDRSHWui: | 
|  | case AArch64::LDRSHXui: | 
|  | case AArch64::LDRSWui: | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | if (!FirstMI) | 
|  | return true; | 
|  |  | 
|  | switch (FirstMI->getOpcode()) { | 
|  | case AArch64::ADR: | 
|  | return (SecondMI.getOperand(2).getImm() == 0); | 
|  | case AArch64::ADRP: | 
|  | return true; | 
|  | } | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Fuse compare and conditional select. | 
|  | static bool isCCSelectPair(const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | unsigned SecondOpcode = SecondMI.getOpcode(); | 
|  |  | 
|  | // 32 bits | 
|  | if (SecondOpcode == AArch64::CSELWr) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | if (!FirstMI) | 
|  | return true; | 
|  |  | 
|  | if (FirstMI->definesRegister(AArch64::WZR)) | 
|  | switch (FirstMI->getOpcode()) { | 
|  | case AArch64::SUBSWrs: | 
|  | return (!AArch64InstrInfo::hasShiftedReg(*FirstMI)); | 
|  | case AArch64::SUBSWrx: | 
|  | return (!AArch64InstrInfo::hasExtendedReg(*FirstMI)); | 
|  | case AArch64::SUBSWrr: | 
|  | case AArch64::SUBSWri: | 
|  | return true; | 
|  | } | 
|  | } | 
|  | // 64 bits | 
|  | else if (SecondOpcode == AArch64::CSELXr) { | 
|  | // Assume the 1st instr to be a wildcard if it is unspecified. | 
|  | if (!FirstMI) | 
|  | return true; | 
|  |  | 
|  | if (FirstMI->definesRegister(AArch64::XZR)) | 
|  | switch (FirstMI->getOpcode()) { | 
|  | case AArch64::SUBSXrs: | 
|  | return (!AArch64InstrInfo::hasShiftedReg(*FirstMI)); | 
|  | case AArch64::SUBSXrx: | 
|  | case AArch64::SUBSXrx64: | 
|  | return (!AArch64InstrInfo::hasExtendedReg(*FirstMI)); | 
|  | case AArch64::SUBSXrr: | 
|  | case AArch64::SUBSXri: | 
|  | return true; | 
|  | } | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | /// Check if the instr pair, FirstMI and SecondMI, should be fused | 
|  | /// together. Given SecondMI, when FirstMI is unspecified, then check if | 
|  | /// SecondMI may be part of a fused pair at all. | 
|  | static bool shouldScheduleAdjacent(const TargetInstrInfo &TII, | 
|  | const TargetSubtargetInfo &TSI, | 
|  | const MachineInstr *FirstMI, | 
|  | const MachineInstr &SecondMI) { | 
|  | const AArch64Subtarget &ST = static_cast<const AArch64Subtarget&>(TSI); | 
|  |  | 
|  | if (ST.hasArithmeticBccFusion() && isArithmeticBccPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  | if (ST.hasArithmeticCbzFusion() && isArithmeticCbzPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  | if (ST.hasFuseAES() && isAESPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  | if (ST.hasFuseLiterals() && isLiteralsPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  | if (ST.hasFuseAddress() && isAddressLdStPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  | if (ST.hasFuseCCSelect() && isCCSelectPair(FirstMI, SecondMI)) | 
|  | return true; | 
|  |  | 
|  | return false; | 
|  | } | 
|  |  | 
|  | } // end namespace | 
|  |  | 
|  |  | 
|  | namespace llvm { | 
|  |  | 
|  | std::unique_ptr<ScheduleDAGMutation> createAArch64MacroFusionDAGMutation () { | 
|  | return createMacroFusionDAGMutation(shouldScheduleAdjacent); | 
|  | } | 
|  |  | 
|  | } // end namespace llvm |