Fix a long-standing wart in the code generator: two-address instruction lowering
actually *removes* one of the operands, instead of just assigning both operands
the same register. This make reasoning about instructions unnecessarily complex,
because you need to know if you are before or after register allocation to match
up operand #'s with the target description file.
Changing this also gets rid of a bunch of hacky code in various places.
This patch also includes changes to fold loads into cmp/test instructions in
the X86 backend, along with a significant simplification to the X86 spill
folding code.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30108 91177308-0d34-0410-b5e6-96231b3b80d8
diff --git a/lib/CodeGen/LiveIntervalAnalysis.cpp b/lib/CodeGen/LiveIntervalAnalysis.cpp
index 5bc03bc..237dcd3 100644
--- a/lib/CodeGen/LiveIntervalAnalysis.cpp
+++ b/lib/CodeGen/LiveIntervalAnalysis.cpp
@@ -262,23 +262,11 @@
MachineInstr *MI = getInstructionFromIndex(index);
- // NewRegLiveIn - This instruction might have multiple uses of the spilled
- // register. In this case, for the first use, keep track of the new vreg
- // that we reload it into. If we see a second use, reuse this vreg
- // instead of creating live ranges for two reloads.
- unsigned NewRegLiveIn = 0;
-
- for_operand:
+ RestartInstruction:
for (unsigned i = 0; i != MI->getNumOperands(); ++i) {
MachineOperand& mop = MI->getOperand(i);
if (mop.isRegister() && mop.getReg() == li.reg) {
- if (NewRegLiveIn && mop.isUse()) {
- // We already emitted a reload of this value, reuse it for
- // subsequent operands.
- MI->getOperand(i).setReg(NewRegLiveIn);
- DEBUG(std::cerr << "\t\t\t\treused reload into reg" << NewRegLiveIn
- << " for operand #" << i << '\n');
- } else if (MachineInstr* fmi = mri_->foldMemoryOperand(MI, i, slot)) {
+ if (MachineInstr *fmi = mri_->foldMemoryOperand(MI, i, slot)) {
// Attempt to fold the memory reference into the instruction. If we
// can do this, we don't need to insert spill code.
if (lv_)
@@ -292,47 +280,63 @@
++numFolded;
// Folding the load/store can completely change the instruction in
// unpredictable ways, rescan it from the beginning.
- goto for_operand;
+ goto RestartInstruction;
} else {
- // This is tricky. We need to add information in the interval about
- // the spill code so we have to use our extra load/store slots.
+ // Create a new virtual register for the spill interval.
+ unsigned NewVReg = mf_->getSSARegMap()->createVirtualRegister(rc);
+
+ // Scan all of the operands of this instruction rewriting operands
+ // to use NewVReg instead of li.reg as appropriate. We do this for
+ // two reasons:
//
- // If we have a use we are going to have a load so we start the
- // interval from the load slot onwards. Otherwise we start from the
- // def slot.
- unsigned start = (mop.isUse() ?
- getLoadIndex(index) :
- getDefIndex(index));
- // If we have a def we are going to have a store right after it so
- // we end the interval after the use of the next
- // instruction. Otherwise we end after the use of this instruction.
- unsigned end = 1 + (mop.isDef() ?
- getStoreIndex(index) :
- getUseIndex(index));
+ // 1. If the instr reads the same spilled vreg multiple times, we
+ // want to reuse the NewVReg.
+ // 2. If the instr is a two-addr instruction, we are required to
+ // keep the src/dst regs pinned.
+ //
+ // Keep track of whether we replace a use and/or def so that we can
+ // create the spill interval with the appropriate range.
+ mop.setReg(NewVReg);
+
+ bool HasUse = mop.isUse();
+ bool HasDef = mop.isDef();
+ for (unsigned j = i+1, e = MI->getNumOperands(); j != e; ++j) {
+ if (MI->getOperand(j).isReg() &&
+ MI->getOperand(j).getReg() == li.reg) {
+ MI->getOperand(j).setReg(NewVReg);
+ HasUse |= MI->getOperand(j).isUse();
+ HasDef |= MI->getOperand(j).isDef();
+ }
+ }
// create a new register for this spill
- NewRegLiveIn = mf_->getSSARegMap()->createVirtualRegister(rc);
- MI->getOperand(i).setReg(NewRegLiveIn);
vrm.grow();
- vrm.assignVirt2StackSlot(NewRegLiveIn, slot);
- LiveInterval& nI = getOrCreateInterval(NewRegLiveIn);
+ vrm.assignVirt2StackSlot(NewVReg, slot);
+ LiveInterval &nI = getOrCreateInterval(NewVReg);
assert(nI.empty());
// the spill weight is now infinity as it
// cannot be spilled again
nI.weight = float(HUGE_VAL);
- LiveRange LR(start, end, nI.getNextValue(~0U, 0));
- DEBUG(std::cerr << " +" << LR);
- nI.addRange(LR);
+
+ if (HasUse) {
+ LiveRange LR(getLoadIndex(index), getUseIndex(index),
+ nI.getNextValue(~0U, 0));
+ DEBUG(std::cerr << " +" << LR);
+ nI.addRange(LR);
+ }
+ if (HasDef) {
+ LiveRange LR(getDefIndex(index), getStoreIndex(index),
+ nI.getNextValue(~0U, 0));
+ DEBUG(std::cerr << " +" << LR);
+ nI.addRange(LR);
+ }
+
added.push_back(&nI);
// update live variables if it is available
if (lv_)
- lv_->addVirtualRegisterKilled(NewRegLiveIn, MI);
-
- // If this is a live in, reuse it for subsequent live-ins. If it's
- // a def, we can't do this.
- if (!mop.isUse()) NewRegLiveIn = 0;
+ lv_->addVirtualRegisterKilled(NewVReg, MI);
DEBUG(std::cerr << "\t\t\t\tadded new interval: ";
nI.print(std::cerr, mri_); std::cerr << '\n');
@@ -445,7 +449,9 @@
// operand, and is a def-and-use.
if (mi->getOperand(0).isRegister() &&
mi->getOperand(0).getReg() == interval.reg &&
- mi->getOperand(0).isDef() && mi->getOperand(0).isUse()) {
+ mi->getNumOperands() > 1 && mi->getOperand(1).isRegister() &&
+ mi->getOperand(1).getReg() == interval.reg &&
+ mi->getOperand(0).isDef() && mi->getOperand(1).isUse()) {
// If this is a two-address definition, then we have already processed
// the live range. The only problem is that we didn't realize there
// are actually two values in the live interval. Because of this we
diff --git a/lib/CodeGen/RegAllocSimple.cpp b/lib/CodeGen/RegAllocSimple.cpp
index b28e21a..ad09f82 100644
--- a/lib/CodeGen/RegAllocSimple.cpp
+++ b/lib/CodeGen/RegAllocSimple.cpp
@@ -203,17 +203,13 @@
physReg = getFreeReg(virtualReg);
} else {
// must be same register number as the first operand
- // This maps a = b + c into b += c, and saves b into a's spot
+ // This maps a = b + c into b = b + c, and saves b into a's spot.
assert(MI->getOperand(1).isRegister() &&
MI->getOperand(1).getReg() &&
MI->getOperand(1).isUse() &&
"Two address instruction invalid!");
physReg = MI->getOperand(1).getReg();
- spillVirtReg(MBB, next(MI), virtualReg, physReg);
- MI->getOperand(1).setDef();
- MI->RemoveOperand(0);
- break; // This is the last operand to process
}
spillVirtReg(MBB, next(MI), virtualReg, physReg);
} else {
diff --git a/lib/CodeGen/TwoAddressInstructionPass.cpp b/lib/CodeGen/TwoAddressInstructionPass.cpp
index f1e41d2..7db9958 100644
--- a/lib/CodeGen/TwoAddressInstructionPass.cpp
+++ b/lib/CodeGen/TwoAddressInstructionPass.cpp
@@ -206,9 +206,8 @@
}
}
- assert(mi->getOperand(0).isDef());
- mi->getOperand(0).setUse();
- mi->RemoveOperand(1);
+ assert(mi->getOperand(0).isDef() && mi->getOperand(1).isUse());
+ mi->getOperand(1).setReg(mi->getOperand(0).getReg());
MadeChange = true;
DEBUG(std::cerr << "\t\trewrite to:\t"; mi->print(std::cerr, &TM));
diff --git a/lib/CodeGen/VirtRegMap.cpp b/lib/CodeGen/VirtRegMap.cpp
index ce9a050..ea1794c 100644
--- a/lib/CodeGen/VirtRegMap.cpp
+++ b/lib/CodeGen/VirtRegMap.cpp
@@ -57,6 +57,12 @@
// VirtRegMap implementation
//===----------------------------------------------------------------------===//
+VirtRegMap::VirtRegMap(MachineFunction &mf)
+ : TII(*mf.getTarget().getInstrInfo()), MF(mf),
+ Virt2PhysMap(NO_PHYS_REG), Virt2StackSlotMap(NO_STACK_SLOT) {
+ grow();
+}
+
void VirtRegMap::grow() {
Virt2PhysMap.grow(MF.getSSARegMap()->getLastVirtReg());
Virt2StackSlotMap.grow(MF.getSSARegMap()->getLastVirtReg());
@@ -92,11 +98,13 @@
}
ModRef MRInfo;
- if (!OldMI->getOperand(OpNo).isDef()) {
- assert(OldMI->getOperand(OpNo).isUse() && "Operand is not use or def?");
- MRInfo = isRef;
+ if (OpNo < 2 && TII.isTwoAddrInstr(OldMI->getOpcode())) {
+ // Folded a two-address operand.
+ MRInfo = isModRef;
+ } else if (OldMI->getOperand(OpNo).isDef()) {
+ MRInfo = isMod;
} else {
- MRInfo = OldMI->getOperand(OpNo).isUse() ? isModRef : isMod;
+ MRInfo = isRef;
}
// add new memory reference
@@ -492,11 +500,6 @@
// that we can choose to reuse the physregs instead of emitting reloads.
AvailableSpills Spills(MRI, TII);
- // DefAndUseVReg - When we see a def&use operand that is spilled, keep track
- // of it. ".first" is the machine operand index (should always be 0 for now),
- // and ".second" is the virtual register that is spilled.
- std::vector<std::pair<unsigned, unsigned> > DefAndUseVReg;
-
// MaybeDeadStores - When we need to write a value back into a stack slot,
// keep track of the inserted store. If the stack slot value is never read
// (because the value was used from some available register, for example), and
@@ -516,8 +519,6 @@
/// reuse.
ReuseInfo ReusedOperands(MI);
- DefAndUseVReg.clear();
-
// Process all of the spilled uses and all non spilled reg references.
for (unsigned i = 0, e = MI.getNumOperands(); i != e; ++i) {
MachineOperand &MO = MI.getOperand(i);
@@ -547,24 +548,27 @@
if (!MO.isUse())
continue; // Handle defs in the loop below (handle use&def here though)
- // If this is both a def and a use, we need to emit a store to the
- // stack slot after the instruction. Keep track of D&U operands
- // because we are about to change it to a physreg here.
- if (MO.isDef()) {
- // Remember that this was a def-and-use operand, and that the
- // stack slot is live after this instruction executes.
- DefAndUseVReg.push_back(std::make_pair(i, VirtReg));
- }
-
int StackSlot = VRM.getStackSlot(VirtReg);
unsigned PhysReg;
// Check to see if this stack slot is available.
if ((PhysReg = Spills.getSpillSlotPhysReg(StackSlot))) {
- // Don't reuse it for a def&use operand if we aren't allowed to change
- // the physreg!
- if (!MO.isDef() || Spills.canClobberPhysReg(StackSlot)) {
+ // This spilled operand might be part of a two-address operand. If this
+ // is the case, then changing it will necessarily require changing the
+ // def part of the instruction as well. However, in some cases, we
+ // aren't allowed to modify the reused register. If none of these cases
+ // apply, reuse it.
+ bool CanReuse = true;
+ if (i == 1 && MI.getOperand(0).isReg() &&
+ MI.getOperand(0).getReg() == VirtReg &&
+ TII->isTwoAddrInstr(MI.getOpcode())) {
+ // Okay, we have a two address operand. We can reuse this physreg as
+ // long as we are allowed to clobber the value.
+ CanReuse = Spills.canClobberPhysReg(StackSlot);
+ }
+
+ if (CanReuse) {
// If this stack slot value is already available, reuse it!
DEBUG(std::cerr << "Reusing SS#" << StackSlot << " from physreg "
<< MRI->getName(PhysReg) << " for vreg"
@@ -777,47 +781,32 @@
unsigned VirtReg = MO.getReg();
if (!MRegisterInfo::isVirtualRegister(VirtReg)) {
- // Check to see if this is a def-and-use vreg operand that we do need
- // to insert a store for.
- bool OpTakenCareOf = false;
- if (MO.isUse() && !DefAndUseVReg.empty()) {
- for (unsigned dau = 0, e = DefAndUseVReg.size(); dau != e; ++dau)
- if (DefAndUseVReg[dau].first == i) {
- VirtReg = DefAndUseVReg[dau].second;
- OpTakenCareOf = true;
- break;
- }
+ // Check to see if this is a noop copy. If so, eliminate the
+ // instruction before considering the dest reg to be changed.
+ unsigned Src, Dst;
+ if (TII->isMoveInstr(MI, Src, Dst) && Src == Dst) {
+ ++NumDCE;
+ DEBUG(std::cerr << "Removing now-noop copy: " << MI);
+ MBB.erase(&MI);
+ VRM.RemoveFromFoldedVirtMap(&MI);
+ goto ProcessNextInst;
}
-
- if (!OpTakenCareOf) {
- // Check to see if this is a noop copy. If so, eliminate the
- // instruction before considering the dest reg to be changed.
- unsigned Src, Dst;
- if (TII->isMoveInstr(MI, Src, Dst) && Src == Dst) {
- ++NumDCE;
- DEBUG(std::cerr << "Removing now-noop copy: " << MI);
- MBB.erase(&MI);
- VRM.RemoveFromFoldedVirtMap(&MI);
- goto ProcessNextInst;
- }
- Spills.ClobberPhysReg(VirtReg);
- continue;
- }
+ Spills.ClobberPhysReg(VirtReg);
+ continue;
}
// The only vregs left are stack slot definitions.
int StackSlot = VRM.getStackSlot(VirtReg);
const TargetRegisterClass *RC =
MBB.getParent()->getSSARegMap()->getRegClass(VirtReg);
- unsigned PhysReg;
- // If this is a def&use operand, and we used a different physreg for
- // it than the one assigned, make sure to execute the store from the
- // correct physical register.
- if (MO.getReg() == VirtReg)
- PhysReg = VRM.getPhys(VirtReg);
+ // If this def is part of a two-address operand, make sure to execute
+ // the store from the correct physical register.
+ unsigned PhysReg;
+ if (i == 0 && TII->isTwoAddrInstr(MI.getOpcode()))
+ PhysReg = MI.getOperand(1).getReg();
else
- PhysReg = MO.getReg();
+ PhysReg = VRM.getPhys(VirtReg);
PhysRegsUsed[PhysReg] = true;
MRI->storeRegToStackSlot(MBB, next(MII), PhysReg, StackSlot, RC);
diff --git a/lib/CodeGen/VirtRegMap.h b/lib/CodeGen/VirtRegMap.h
index 83d5aad..426d1cf 100644
--- a/lib/CodeGen/VirtRegMap.h
+++ b/lib/CodeGen/VirtRegMap.h
@@ -23,6 +23,7 @@
namespace llvm {
class MachineInstr;
+ class TargetInstrInfo;
class VirtRegMap {
public:
@@ -31,6 +32,8 @@
std::pair<unsigned, ModRef> > MI2VirtMapTy;
private:
+ const TargetInstrInfo &TII;
+
MachineFunction &MF;
/// Virt2PhysMap - This is a virtual to physical register
/// mapping. Each virtual register is required to have an entry in
@@ -58,10 +61,7 @@
};
public:
- VirtRegMap(MachineFunction &mf)
- : MF(mf), Virt2PhysMap(NO_PHYS_REG), Virt2StackSlotMap(NO_STACK_SLOT) {
- grow();
- }
+ VirtRegMap(MachineFunction &mf);
void grow();
diff --git a/lib/Target/X86/X86InstrInfo.cpp b/lib/Target/X86/X86InstrInfo.cpp
index 47388ea..9d8eab8 100644
--- a/lib/Target/X86/X86InstrInfo.cpp
+++ b/lib/Target/X86/X86InstrInfo.cpp
@@ -22,7 +22,7 @@
X86InstrInfo::X86InstrInfo(X86TargetMachine &tm)
: TargetInstrInfo(X86Insts, sizeof(X86Insts)/sizeof(X86Insts[0])),
- TM(tm) {
+ TM(tm), RI(*this) {
}
diff --git a/lib/Target/X86/X86RegisterInfo.cpp b/lib/Target/X86/X86RegisterInfo.cpp
index 90ca91d..89cc8f2 100644
--- a/lib/Target/X86/X86RegisterInfo.cpp
+++ b/lib/Target/X86/X86RegisterInfo.cpp
@@ -46,8 +46,8 @@
cl::Hidden);
}
-X86RegisterInfo::X86RegisterInfo()
- : X86GenRegisterInfo(X86::ADJCALLSTACKDOWN, X86::ADJCALLSTACKUP) {}
+X86RegisterInfo::X86RegisterInfo(const TargetInstrInfo &tii)
+ : X86GenRegisterInfo(X86::ADJCALLSTACKDOWN, X86::ADJCALLSTACKUP), TII(tii) {}
void X86RegisterInfo::storeRegToStackSlot(MachineBasicBlock &MBB,
MachineBasicBlock::iterator MI,
@@ -139,39 +139,46 @@
BuildMI(MBB, MI, Opc, 1, DestReg).addReg(SrcReg);
}
-
-static MachineInstr *MakeMInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- return addFrameReference(BuildMI(Opcode, 4), FrameIndex);
+static MachineInstr *FuseTwoAddrInst(unsigned Opcode, unsigned FrameIndex,
+ MachineInstr *MI) {
+ unsigned NumOps = MI->getNumOperands()-2;
+ // Create the base instruction with the memory operand as the first part.
+ MachineInstrBuilder MIB = addFrameReference(BuildMI(Opcode, 4+NumOps),
+ FrameIndex);
+
+ // Loop over the rest of the ri operands, converting them over.
+ for (unsigned i = 0; i != NumOps; ++i) {
+ if (MI->getOperand(i+2).isReg())
+ MIB = MIB.addReg(MI->getOperand(i+2).getReg());
+ else {
+ assert(MI->getOperand(i+2).isImm() && "Unknown operand type!");
+ MIB = MIB.addImm(MI->getOperand(i+2).getImm());
+ }
+ }
+ return MIB;
}
-static MachineInstr *MakeMRInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- return addFrameReference(BuildMI(Opcode, 5), FrameIndex)
- .addReg(MI->getOperand(1).getReg());
-}
-
-static MachineInstr *MakeMRIInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- return addFrameReference(BuildMI(Opcode, 6), FrameIndex)
- .addReg(MI->getOperand(1).getReg())
- .addImm(MI->getOperand(2).getImmedValue());
-}
-
-static MachineInstr *MakeMIInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- if (MI->getOperand(1).isImmediate())
- return addFrameReference(BuildMI(Opcode, 5), FrameIndex)
- .addImm(MI->getOperand(1).getImmedValue());
- else if (MI->getOperand(1).isGlobalAddress())
- return addFrameReference(BuildMI(Opcode, 5), FrameIndex)
- .addGlobalAddress(MI->getOperand(1).getGlobal(),
- MI->getOperand(1).getOffset());
- else if (MI->getOperand(1).isJumpTableIndex())
- return addFrameReference(BuildMI(Opcode, 5), FrameIndex)
- .addJumpTableIndex(MI->getOperand(1).getJumpTableIndex());
- assert(0 && "Unknown operand for MakeMI!");
- return 0;
+static MachineInstr *FuseInst(unsigned Opcode, unsigned OpNo,
+ unsigned FrameIndex, MachineInstr *MI) {
+ MachineInstrBuilder MIB = BuildMI(Opcode, MI->getNumOperands()+3);
+
+ for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) {
+ MachineOperand &MO = MI->getOperand(i);
+ if (i == OpNo) {
+ assert(MO.isReg() && "Expected to fold into reg operand!");
+ MIB = addFrameReference(MIB, FrameIndex);
+ } else if (MO.isReg())
+ MIB = MIB.addReg(MO.getReg(), MO.getUseType());
+ else if (MO.isImm())
+ MIB = MIB.addImm(MO.getImm());
+ else if (MO.isGlobalAddress())
+ MIB = MIB.addGlobalAddress(MO.getGlobal(), MO.getOffset());
+ else if (MO.isJumpTableIndex())
+ MIB = MIB.addJumpTableIndex(MO.getJumpTableIndex());
+ else
+ assert(0 && "Unknown operand for FuseInst!");
+ }
+ return MIB;
}
static MachineInstr *MakeM0Inst(unsigned Opcode, unsigned FrameIndex,
@@ -179,20 +186,6 @@
return addFrameReference(BuildMI(Opcode, 5), FrameIndex).addImm(0);
}
-static MachineInstr *MakeRMInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- const MachineOperand& op = MI->getOperand(0);
- return addFrameReference(BuildMI(Opcode, 5, op.getReg(), op.getUseType()),
- FrameIndex);
-}
-
-static MachineInstr *MakeRMIInst(unsigned Opcode, unsigned FrameIndex,
- MachineInstr *MI) {
- const MachineOperand& op = MI->getOperand(0);
- return addFrameReference(BuildMI(Opcode, 6, op.getReg(), op.getUseType()),
- FrameIndex).addImm(MI->getOperand(2).getImmedValue());
-}
-
//===----------------------------------------------------------------------===//
// Efficient Lookup Table Support
@@ -204,8 +197,6 @@
struct TableEntry {
unsigned from; // Original opcode.
unsigned to; // New opcode.
- unsigned make; // Form of make required to produce the
- // new instruction.
// less operators used by STL search.
bool operator<(const TableEntry &TE) const { return from < TE.from; }
@@ -257,451 +248,451 @@
#endif
-MachineInstr* X86RegisterInfo::foldMemoryOperand(MachineInstr* MI,
+MachineInstr* X86RegisterInfo::foldMemoryOperand(MachineInstr *MI,
unsigned i,
int FrameIndex) const {
// Check switch flag
if (NoFusing) return NULL;
- // Selection of instruction makes
- enum {
- makeM0Inst,
- makeMIInst,
- makeMInst,
- makeMRIInst,
- makeMRInst,
- makeRMIInst,
- makeRMInst
- };
-
// Table (and size) to search
const TableEntry *OpcodeTablePtr = NULL;
unsigned OpcodeTableSize = 0;
+ bool isTwoAddrFold = false;
- if (i == 0) { // If operand 0
+ // Folding a memory location into the two-address part of a two-address
+ // instruction is different than folding it other places. It requires
+ // replacing the *two* registers with the memory location.
+ if (MI->getNumOperands() >= 2 && MI->getOperand(0).isReg() &&
+ MI->getOperand(1).isReg() && i < 2 &&
+ MI->getOperand(0).getReg() == MI->getOperand(1).getReg() &&
+ TII.isTwoAddrInstr(MI->getOpcode())) {
static const TableEntry OpcodeTable[] = {
- { X86::ADC32ri, X86::ADC32mi, makeMIInst },
- { X86::ADC32ri8, X86::ADC32mi8, makeMIInst },
- { X86::ADC32rr, X86::ADC32mr, makeMRInst },
- { X86::ADD16ri, X86::ADD16mi, makeMIInst },
- { X86::ADD16ri8, X86::ADD16mi8, makeMIInst },
- { X86::ADD16rr, X86::ADD16mr, makeMRInst },
- { X86::ADD32ri, X86::ADD32mi, makeMIInst },
- { X86::ADD32ri8, X86::ADD32mi8, makeMIInst },
- { X86::ADD32rr, X86::ADD32mr, makeMRInst },
- { X86::ADD8ri, X86::ADD8mi, makeMIInst },
- { X86::ADD8rr, X86::ADD8mr, makeMRInst },
- { X86::AND16ri, X86::AND16mi, makeMIInst },
- { X86::AND16ri8, X86::AND16mi8, makeMIInst },
- { X86::AND16rr, X86::AND16mr, makeMRInst },
- { X86::AND32ri, X86::AND32mi, makeMIInst },
- { X86::AND32ri8, X86::AND32mi8, makeMIInst },
- { X86::AND32rr, X86::AND32mr, makeMRInst },
- { X86::AND8ri, X86::AND8mi, makeMIInst },
- { X86::AND8rr, X86::AND8mr, makeMRInst },
- { X86::DEC16r, X86::DEC16m, makeMInst },
- { X86::DEC32r, X86::DEC32m, makeMInst },
- { X86::DEC8r, X86::DEC8m, makeMInst },
- { X86::DIV16r, X86::DIV16m, makeMInst },
- { X86::DIV32r, X86::DIV32m, makeMInst },
- { X86::DIV8r, X86::DIV8m, makeMInst },
- { X86::FsMOVAPDrr, X86::MOVSDmr, makeMRInst },
- { X86::FsMOVAPSrr, X86::MOVSSmr, makeMRInst },
- { X86::IDIV16r, X86::IDIV16m, makeMInst },
- { X86::IDIV32r, X86::IDIV32m, makeMInst },
- { X86::IDIV8r, X86::IDIV8m, makeMInst },
- { X86::IMUL16r, X86::IMUL16m, makeMInst },
- { X86::IMUL32r, X86::IMUL32m, makeMInst },
- { X86::IMUL8r, X86::IMUL8m, makeMInst },
- { X86::INC16r, X86::INC16m, makeMInst },
- { X86::INC32r, X86::INC32m, makeMInst },
- { X86::INC8r, X86::INC8m, makeMInst },
- { X86::MOV16r0, X86::MOV16mi, makeM0Inst },
- { X86::MOV16ri, X86::MOV16mi, makeMIInst },
- { X86::MOV16rr, X86::MOV16mr, makeMRInst },
- { X86::MOV32r0, X86::MOV32mi, makeM0Inst },
- { X86::MOV32ri, X86::MOV32mi, makeMIInst },
- { X86::MOV32rr, X86::MOV32mr, makeMRInst },
- { X86::MOV8r0, X86::MOV8mi, makeM0Inst },
- { X86::MOV8ri, X86::MOV8mi, makeMIInst },
- { X86::MOV8rr, X86::MOV8mr, makeMRInst },
- { X86::MOVAPDrr, X86::MOVAPDmr, makeMRInst },
- { X86::MOVAPSrr, X86::MOVAPSmr, makeMRInst },
- { X86::MOVPDI2DIrr, X86::MOVPDI2DImr, makeMRInst },
- { X86::MOVPS2SSrr, X86::MOVPS2SSmr, makeMRInst },
- { X86::MOVSDrr, X86::MOVSDmr, makeMRInst },
- { X86::MOVSSrr, X86::MOVSSmr, makeMRInst },
- { X86::MOVUPDrr, X86::MOVUPDmr, makeMRInst },
- { X86::MOVUPSrr, X86::MOVUPSmr, makeMRInst },
- { X86::MUL16r, X86::MUL16m, makeMInst },
- { X86::MUL32r, X86::MUL32m, makeMInst },
- { X86::MUL8r, X86::MUL8m, makeMInst },
- { X86::NEG16r, X86::NEG16m, makeMInst },
- { X86::NEG32r, X86::NEG32m, makeMInst },
- { X86::NEG8r, X86::NEG8m, makeMInst },
- { X86::NOT16r, X86::NOT16m, makeMInst },
- { X86::NOT32r, X86::NOT32m, makeMInst },
- { X86::NOT8r, X86::NOT8m, makeMInst },
- { X86::OR16ri, X86::OR16mi, makeMIInst },
- { X86::OR16ri8, X86::OR16mi8, makeMIInst },
- { X86::OR16rr, X86::OR16mr, makeMRInst },
- { X86::OR32ri, X86::OR32mi, makeMIInst },
- { X86::OR32ri8, X86::OR32mi8, makeMIInst },
- { X86::OR32rr, X86::OR32mr, makeMRInst },
- { X86::OR8ri, X86::OR8mi, makeMIInst },
- { X86::OR8rr, X86::OR8mr, makeMRInst },
- { X86::ROL16r1, X86::ROL16m1, makeMInst },
- { X86::ROL16rCL, X86::ROL16mCL, makeMInst },
- { X86::ROL16ri, X86::ROL16mi, makeMIInst },
- { X86::ROL32r1, X86::ROL32m1, makeMInst },
- { X86::ROL32rCL, X86::ROL32mCL, makeMInst },
- { X86::ROL32ri, X86::ROL32mi, makeMIInst },
- { X86::ROL8r1, X86::ROL8m1, makeMInst },
- { X86::ROL8rCL, X86::ROL8mCL, makeMInst },
- { X86::ROL8ri, X86::ROL8mi, makeMIInst },
- { X86::ROR16r1, X86::ROR16m1, makeMInst },
- { X86::ROR16rCL, X86::ROR16mCL, makeMInst },
- { X86::ROR16ri, X86::ROR16mi, makeMIInst },
- { X86::ROR32r1, X86::ROR32m1, makeMInst },
- { X86::ROR32rCL, X86::ROR32mCL, makeMInst },
- { X86::ROR32ri, X86::ROR32mi, makeMIInst },
- { X86::ROR8r1, X86::ROR8m1, makeMInst },
- { X86::ROR8rCL, X86::ROR8mCL, makeMInst },
- { X86::ROR8ri, X86::ROR8mi, makeMIInst },
- { X86::SAR16r1, X86::SAR16m1, makeMInst },
- { X86::SAR16rCL, X86::SAR16mCL, makeMInst },
- { X86::SAR16ri, X86::SAR16mi, makeMIInst },
- { X86::SAR32r1, X86::SAR32m1, makeMInst },
- { X86::SAR32rCL, X86::SAR32mCL, makeMInst },
- { X86::SAR32ri, X86::SAR32mi, makeMIInst },
- { X86::SAR8r1, X86::SAR8m1, makeMInst },
- { X86::SAR8rCL, X86::SAR8mCL, makeMInst },
- { X86::SAR8ri, X86::SAR8mi, makeMIInst },
- { X86::SBB32ri, X86::SBB32mi, makeMIInst },
- { X86::SBB32ri8, X86::SBB32mi8, makeMIInst },
- { X86::SBB32rr, X86::SBB32mr, makeMRInst },
- { X86::SETAEr, X86::SETAEm, makeMInst },
- { X86::SETAr, X86::SETAm, makeMInst },
- { X86::SETBEr, X86::SETBEm, makeMInst },
- { X86::SETBr, X86::SETBm, makeMInst },
- { X86::SETEr, X86::SETEm, makeMInst },
- { X86::SETGEr, X86::SETGEm, makeMInst },
- { X86::SETGr, X86::SETGm, makeMInst },
- { X86::SETLEr, X86::SETLEm, makeMInst },
- { X86::SETLr, X86::SETLm, makeMInst },
- { X86::SETNEr, X86::SETNEm, makeMInst },
- { X86::SETNPr, X86::SETNPm, makeMInst },
- { X86::SETNSr, X86::SETNSm, makeMInst },
- { X86::SETPr, X86::SETPm, makeMInst },
- { X86::SETSr, X86::SETSm, makeMInst },
- { X86::SHL16r1, X86::SHL16m1, makeMInst },
- { X86::SHL16rCL, X86::SHL16mCL, makeMInst },
- { X86::SHL16ri, X86::SHL16mi, makeMIInst },
- { X86::SHL32r1, X86::SHL32m1, makeMInst },
- { X86::SHL32rCL, X86::SHL32mCL, makeMInst },
- { X86::SHL32ri, X86::SHL32mi, makeMIInst },
- { X86::SHL8r1, X86::SHL8m1, makeMInst },
- { X86::SHL8rCL, X86::SHL8mCL, makeMInst },
- { X86::SHL8ri, X86::SHL8mi, makeMIInst },
- { X86::SHLD16rrCL, X86::SHLD16mrCL, makeMRInst },
- { X86::SHLD16rri8, X86::SHLD16mri8, makeMRIInst },
- { X86::SHLD32rrCL, X86::SHLD32mrCL, makeMRInst },
- { X86::SHLD32rri8, X86::SHLD32mri8, makeMRIInst },
- { X86::SHR16r1, X86::SHR16m1, makeMInst },
- { X86::SHR16rCL, X86::SHR16mCL, makeMInst },
- { X86::SHR16ri, X86::SHR16mi, makeMIInst },
- { X86::SHR32r1, X86::SHR32m1, makeMInst },
- { X86::SHR32rCL, X86::SHR32mCL, makeMInst },
- { X86::SHR32ri, X86::SHR32mi, makeMIInst },
- { X86::SHR8r1, X86::SHR8m1, makeMInst },
- { X86::SHR8rCL, X86::SHR8mCL, makeMInst },
- { X86::SHR8ri, X86::SHR8mi, makeMIInst },
- { X86::SHRD16rrCL, X86::SHRD16mrCL, makeMRInst },
- { X86::SHRD16rri8, X86::SHRD16mri8, makeMRIInst },
- { X86::SHRD32rrCL, X86::SHRD32mrCL, makeMRInst },
- { X86::SHRD32rri8, X86::SHRD32mri8, makeMRIInst },
- { X86::SUB16ri, X86::SUB16mi, makeMIInst },
- { X86::SUB16ri8, X86::SUB16mi8, makeMIInst },
- { X86::SUB16rr, X86::SUB16mr, makeMRInst },
- { X86::SUB32ri, X86::SUB32mi, makeMIInst },
- { X86::SUB32ri8, X86::SUB32mi8, makeMIInst },
- { X86::SUB32rr, X86::SUB32mr, makeMRInst },
- { X86::SUB8ri, X86::SUB8mi, makeMIInst },
- { X86::SUB8rr, X86::SUB8mr, makeMRInst },
- { X86::XCHG16rr, X86::XCHG16mr, makeMRInst },
- { X86::XCHG32rr, X86::XCHG32mr, makeMRInst },
- { X86::XCHG8rr, X86::XCHG8mr, makeMRInst },
- { X86::XOR16ri, X86::XOR16mi, makeMIInst },
- { X86::XOR16ri8, X86::XOR16mi8, makeMIInst },
- { X86::XOR16rr, X86::XOR16mr, makeMRInst },
- { X86::XOR32ri, X86::XOR32mi, makeMIInst },
- { X86::XOR32ri8, X86::XOR32mi8, makeMIInst },
- { X86::XOR32rr, X86::XOR32mr, makeMRInst },
- { X86::XOR8ri, X86::XOR8mi, makeMIInst },
- { X86::XOR8rr, X86::XOR8mr, makeMRInst }
+ { X86::ADC32ri, X86::ADC32mi },
+ { X86::ADC32ri8, X86::ADC32mi8 },
+ { X86::ADC32rr, X86::ADC32mr },
+ { X86::ADD16ri, X86::ADD16mi },
+ { X86::ADD16ri8, X86::ADD16mi8 },
+ { X86::ADD16rr, X86::ADD16mr },
+ { X86::ADD32ri, X86::ADD32mi },
+ { X86::ADD32ri8, X86::ADD32mi8 },
+ { X86::ADD32rr, X86::ADD32mr },
+ { X86::ADD8ri, X86::ADD8mi },
+ { X86::ADD8rr, X86::ADD8mr },
+ { X86::AND16ri, X86::AND16mi },
+ { X86::AND16ri8, X86::AND16mi8 },
+ { X86::AND16rr, X86::AND16mr },
+ { X86::AND32ri, X86::AND32mi },
+ { X86::AND32ri8, X86::AND32mi8 },
+ { X86::AND32rr, X86::AND32mr },
+ { X86::AND8ri, X86::AND8mi },
+ { X86::AND8rr, X86::AND8mr },
+ { X86::DEC16r, X86::DEC16m },
+ { X86::DEC32r, X86::DEC32m },
+ { X86::DEC8r, X86::DEC8m },
+ { X86::INC16r, X86::INC16m },
+ { X86::INC32r, X86::INC32m },
+ { X86::INC8r, X86::INC8m },
+ { X86::NEG16r, X86::NEG16m },
+ { X86::NEG32r, X86::NEG32m },
+ { X86::NEG8r, X86::NEG8m },
+ { X86::NOT16r, X86::NOT16m },
+ { X86::NOT32r, X86::NOT32m },
+ { X86::NOT8r, X86::NOT8m },
+ { X86::OR16ri, X86::OR16mi },
+ { X86::OR16ri8, X86::OR16mi8 },
+ { X86::OR16rr, X86::OR16mr },
+ { X86::OR32ri, X86::OR32mi },
+ { X86::OR32ri8, X86::OR32mi8 },
+ { X86::OR32rr, X86::OR32mr },
+ { X86::OR8ri, X86::OR8mi },
+ { X86::OR8rr, X86::OR8mr },
+ { X86::ROL16r1, X86::ROL16m1 },
+ { X86::ROL16rCL, X86::ROL16mCL },
+ { X86::ROL16ri, X86::ROL16mi },
+ { X86::ROL32r1, X86::ROL32m1 },
+ { X86::ROL32rCL, X86::ROL32mCL },
+ { X86::ROL32ri, X86::ROL32mi },
+ { X86::ROL8r1, X86::ROL8m1 },
+ { X86::ROL8rCL, X86::ROL8mCL },
+ { X86::ROL8ri, X86::ROL8mi },
+ { X86::ROR16r1, X86::ROR16m1 },
+ { X86::ROR16rCL, X86::ROR16mCL },
+ { X86::ROR16ri, X86::ROR16mi },
+ { X86::ROR32r1, X86::ROR32m1 },
+ { X86::ROR32rCL, X86::ROR32mCL },
+ { X86::ROR32ri, X86::ROR32mi },
+ { X86::ROR8r1, X86::ROR8m1 },
+ { X86::ROR8rCL, X86::ROR8mCL },
+ { X86::ROR8ri, X86::ROR8mi },
+ { X86::SAR16r1, X86::SAR16m1 },
+ { X86::SAR16rCL, X86::SAR16mCL },
+ { X86::SAR16ri, X86::SAR16mi },
+ { X86::SAR32r1, X86::SAR32m1 },
+ { X86::SAR32rCL, X86::SAR32mCL },
+ { X86::SAR32ri, X86::SAR32mi },
+ { X86::SAR8r1, X86::SAR8m1 },
+ { X86::SAR8rCL, X86::SAR8mCL },
+ { X86::SAR8ri, X86::SAR8mi },
+ { X86::SBB32ri, X86::SBB32mi },
+ { X86::SBB32ri8, X86::SBB32mi8 },
+ { X86::SBB32rr, X86::SBB32mr },
+ { X86::SHL16r1, X86::SHL16m1 },
+ { X86::SHL16rCL, X86::SHL16mCL },
+ { X86::SHL16ri, X86::SHL16mi },
+ { X86::SHL32r1, X86::SHL32m1 },
+ { X86::SHL32rCL, X86::SHL32mCL },
+ { X86::SHL32ri, X86::SHL32mi },
+ { X86::SHL8r1, X86::SHL8m1 },
+ { X86::SHL8rCL, X86::SHL8mCL },
+ { X86::SHL8ri, X86::SHL8mi },
+ { X86::SHLD16rrCL, X86::SHLD16mrCL },
+ { X86::SHLD16rri8, X86::SHLD16mri8 },
+ { X86::SHLD32rrCL, X86::SHLD32mrCL },
+ { X86::SHLD32rri8, X86::SHLD32mri8 },
+ { X86::SHR16r1, X86::SHR16m1 },
+ { X86::SHR16rCL, X86::SHR16mCL },
+ { X86::SHR16ri, X86::SHR16mi },
+ { X86::SHR32r1, X86::SHR32m1 },
+ { X86::SHR32rCL, X86::SHR32mCL },
+ { X86::SHR32ri, X86::SHR32mi },
+ { X86::SHR8r1, X86::SHR8m1 },
+ { X86::SHR8rCL, X86::SHR8mCL },
+ { X86::SHR8ri, X86::SHR8mi },
+ { X86::SHRD16rrCL, X86::SHRD16mrCL },
+ { X86::SHRD16rri8, X86::SHRD16mri8 },
+ { X86::SHRD32rrCL, X86::SHRD32mrCL },
+ { X86::SHRD32rri8, X86::SHRD32mri8 },
+ { X86::SUB16ri, X86::SUB16mi },
+ { X86::SUB16ri8, X86::SUB16mi8 },
+ { X86::SUB16rr, X86::SUB16mr },
+ { X86::SUB32ri, X86::SUB32mi },
+ { X86::SUB32ri8, X86::SUB32mi8 },
+ { X86::SUB32rr, X86::SUB32mr },
+ { X86::SUB8ri, X86::SUB8mi },
+ { X86::SUB8rr, X86::SUB8mr },
+ { X86::XOR16ri, X86::XOR16mi },
+ { X86::XOR16ri8, X86::XOR16mi8 },
+ { X86::XOR16rr, X86::XOR16mr },
+ { X86::XOR32ri, X86::XOR32mi },
+ { X86::XOR32ri8, X86::XOR32mi8 },
+ { X86::XOR32rr, X86::XOR32mr },
+ { X86::XOR8ri, X86::XOR8mi },
+ { X86::XOR8rr, X86::XOR8mr }
+ };
+ ASSERT_SORTED(OpcodeTable);
+ OpcodeTablePtr = OpcodeTable;
+ OpcodeTableSize = ARRAY_SIZE(OpcodeTable);
+ isTwoAddrFold = true;
+ } else if (i == 0) { // If operand 0
+ if (MI->getOpcode() == X86::MOV16r0)
+ return MakeM0Inst(X86::MOV16mi, FrameIndex, MI);
+ else if (MI->getOpcode() == X86::MOV32r0)
+ return MakeM0Inst(X86::MOV32mi, FrameIndex, MI);
+ else if (MI->getOpcode() == X86::MOV8r0)
+ return MakeM0Inst(X86::MOV8mi, FrameIndex, MI);
+
+ static const TableEntry OpcodeTable[] = {
+ { X86::CMP16ri, X86::CMP16mi },
+ { X86::CMP16ri8, X86::CMP16mi8 },
+ { X86::CMP32ri, X86::CMP32mi },
+ { X86::CMP32ri8, X86::CMP32mi8 },
+ { X86::CMP8ri, X86::CMP8mi },
+ { X86::DIV16r, X86::DIV16m },
+ { X86::DIV32r, X86::DIV32m },
+ { X86::DIV8r, X86::DIV8m },
+ { X86::FsMOVAPDrr, X86::MOVSDmr },
+ { X86::FsMOVAPSrr, X86::MOVSSmr },
+ { X86::IDIV16r, X86::IDIV16m },
+ { X86::IDIV32r, X86::IDIV32m },
+ { X86::IDIV8r, X86::IDIV8m },
+ { X86::IMUL16r, X86::IMUL16m },
+ { X86::IMUL32r, X86::IMUL32m },
+ { X86::IMUL8r, X86::IMUL8m },
+ { X86::MOV16ri, X86::MOV16mi },
+ { X86::MOV16rr, X86::MOV16mr },
+ { X86::MOV32ri, X86::MOV32mi },
+ { X86::MOV32rr, X86::MOV32mr },
+ { X86::MOV8ri, X86::MOV8mi },
+ { X86::MOV8rr, X86::MOV8mr },
+ { X86::MOVAPDrr, X86::MOVAPDmr },
+ { X86::MOVAPSrr, X86::MOVAPSmr },
+ { X86::MOVPDI2DIrr, X86::MOVPDI2DImr },
+ { X86::MOVPS2SSrr, X86::MOVPS2SSmr },
+ { X86::MOVSDrr, X86::MOVSDmr },
+ { X86::MOVSSrr, X86::MOVSSmr },
+ { X86::MOVUPDrr, X86::MOVUPDmr },
+ { X86::MOVUPSrr, X86::MOVUPSmr },
+ { X86::MUL16r, X86::MUL16m },
+ { X86::MUL32r, X86::MUL32m },
+ { X86::MUL8r, X86::MUL8m },
+ { X86::SETAEr, X86::SETAEm },
+ { X86::SETAr, X86::SETAm },
+ { X86::SETBEr, X86::SETBEm },
+ { X86::SETBr, X86::SETBm },
+ { X86::SETEr, X86::SETEm },
+ { X86::SETGEr, X86::SETGEm },
+ { X86::SETGr, X86::SETGm },
+ { X86::SETLEr, X86::SETLEm },
+ { X86::SETLr, X86::SETLm },
+ { X86::SETNEr, X86::SETNEm },
+ { X86::SETNPr, X86::SETNPm },
+ { X86::SETNSr, X86::SETNSm },
+ { X86::SETPr, X86::SETPm },
+ { X86::SETSr, X86::SETSm },
+ { X86::TEST16ri, X86::TEST16mi },
+ { X86::TEST32ri, X86::TEST32mi },
+ { X86::TEST8ri, X86::TEST8mi },
+ { X86::XCHG16rr, X86::XCHG16mr },
+ { X86::XCHG32rr, X86::XCHG32mr },
+ { X86::XCHG8rr, X86::XCHG8mr }
};
ASSERT_SORTED(OpcodeTable);
OpcodeTablePtr = OpcodeTable;
OpcodeTableSize = ARRAY_SIZE(OpcodeTable);
} else if (i == 1) {
static const TableEntry OpcodeTable[] = {
- { X86::ADC32rr, X86::ADC32rm, makeRMInst },
- { X86::ADD16rr, X86::ADD16rm, makeRMInst },
- { X86::ADD32rr, X86::ADD32rm, makeRMInst },
- { X86::ADD8rr, X86::ADD8rm, makeRMInst },
- { X86::ADDPDrr, X86::ADDPDrm, makeRMInst },
- { X86::ADDPSrr, X86::ADDPSrm, makeRMInst },
- { X86::ADDSDrr, X86::ADDSDrm, makeRMInst },
- { X86::ADDSSrr, X86::ADDSSrm, makeRMInst },
- { X86::ADDSUBPDrr, X86::ADDSUBPDrm, makeRMInst },
- { X86::ADDSUBPSrr, X86::ADDSUBPSrm, makeRMInst },
- { X86::AND16rr, X86::AND16rm, makeRMInst },
- { X86::AND32rr, X86::AND32rm, makeRMInst },
- { X86::AND8rr, X86::AND8rm, makeRMInst },
- { X86::ANDNPDrr, X86::ANDNPDrm, makeRMInst },
- { X86::ANDNPSrr, X86::ANDNPSrm, makeRMInst },
- { X86::ANDPDrr, X86::ANDPDrm, makeRMInst },
- { X86::ANDPSrr, X86::ANDPSrm, makeRMInst },
- { X86::CMOVA16rr, X86::CMOVA16rm, makeRMInst },
- { X86::CMOVA32rr, X86::CMOVA32rm, makeRMInst },
- { X86::CMOVAE16rr, X86::CMOVAE16rm, makeRMInst },
- { X86::CMOVAE32rr, X86::CMOVAE32rm, makeRMInst },
- { X86::CMOVB16rr, X86::CMOVB16rm, makeRMInst },
- { X86::CMOVB32rr, X86::CMOVB32rm, makeRMInst },
- { X86::CMOVBE16rr, X86::CMOVBE16rm, makeRMInst },
- { X86::CMOVBE32rr, X86::CMOVBE32rm, makeRMInst },
- { X86::CMOVE16rr, X86::CMOVE16rm, makeRMInst },
- { X86::CMOVE32rr, X86::CMOVE32rm, makeRMInst },
- { X86::CMOVG16rr, X86::CMOVG16rm, makeRMInst },
- { X86::CMOVG32rr, X86::CMOVG32rm, makeRMInst },
- { X86::CMOVGE16rr, X86::CMOVGE16rm, makeRMInst },
- { X86::CMOVGE32rr, X86::CMOVGE32rm, makeRMInst },
- { X86::CMOVL16rr, X86::CMOVL16rm, makeRMInst },
- { X86::CMOVL32rr, X86::CMOVL32rm, makeRMInst },
- { X86::CMOVLE16rr, X86::CMOVLE16rm, makeRMInst },
- { X86::CMOVLE32rr, X86::CMOVLE32rm, makeRMInst },
- { X86::CMOVNE16rr, X86::CMOVNE16rm, makeRMInst },
- { X86::CMOVNE32rr, X86::CMOVNE32rm, makeRMInst },
- { X86::CMOVNP16rr, X86::CMOVNP16rm, makeRMInst },
- { X86::CMOVNP32rr, X86::CMOVNP32rm, makeRMInst },
- { X86::CMOVNS16rr, X86::CMOVNS16rm, makeRMInst },
- { X86::CMOVNS32rr, X86::CMOVNS32rm, makeRMInst },
- { X86::CMOVP16rr, X86::CMOVP16rm, makeRMInst },
- { X86::CMOVP32rr, X86::CMOVP32rm, makeRMInst },
- { X86::CMOVS16rr, X86::CMOVS16rm, makeRMInst },
- { X86::CMOVS32rr, X86::CMOVS32rm, makeRMInst },
- { X86::CMP16ri, X86::CMP16mi, makeMIInst },
- { X86::CMP16ri8, X86::CMP16mi8, makeMIInst },
- { X86::CMP16rr, X86::CMP16rm, makeRMInst },
- { X86::CMP32ri, X86::CMP32mi, makeMIInst },
- { X86::CMP32ri8, X86::CMP32mi8, makeRMInst },
- { X86::CMP32rr, X86::CMP32rm, makeRMInst },
- { X86::CMP8ri, X86::CMP8mi, makeRMInst },
- { X86::CMP8rr, X86::CMP8rm, makeRMInst },
- { X86::CMPPDrri, X86::CMPPDrmi, makeRMIInst },
- { X86::CMPPSrri, X86::CMPPSrmi, makeRMIInst },
- { X86::CMPSDrr, X86::CMPSDrm, makeRMInst },
- { X86::CMPSSrr, X86::CMPSSrm, makeRMInst },
- { X86::CVTSD2SSrr, X86::CVTSD2SSrm, makeRMInst },
- { X86::CVTSI2SDrr, X86::CVTSI2SDrm, makeRMInst },
- { X86::CVTSI2SSrr, X86::CVTSI2SSrm, makeRMInst },
- { X86::CVTSS2SDrr, X86::CVTSS2SDrm, makeRMInst },
- { X86::CVTTSD2SIrr, X86::CVTTSD2SIrm, makeRMInst },
- { X86::CVTTSS2SIrr, X86::CVTTSS2SIrm, makeRMInst },
- { X86::DIVPDrr, X86::DIVPDrm, makeRMInst },
- { X86::DIVPSrr, X86::DIVPSrm, makeRMInst },
- { X86::DIVSDrr, X86::DIVSDrm, makeRMInst },
- { X86::DIVSSrr, X86::DIVSSrm, makeRMInst },
- { X86::FsMOVAPDrr, X86::MOVSDrm, makeRMInst },
- { X86::FsMOVAPSrr, X86::MOVSSrm, makeRMInst },
- { X86::HADDPDrr, X86::HADDPDrm, makeRMInst },
- { X86::HADDPSrr, X86::HADDPSrm, makeRMInst },
- { X86::HSUBPDrr, X86::HSUBPDrm, makeRMInst },
- { X86::HSUBPSrr, X86::HSUBPSrm, makeRMInst },
- { X86::IMUL16rr, X86::IMUL16rm, makeRMInst },
- { X86::IMUL16rri, X86::IMUL16rmi, makeRMIInst },
- { X86::IMUL16rri8, X86::IMUL16rmi8, makeRMIInst },
- { X86::IMUL32rr, X86::IMUL32rm, makeRMInst },
- { X86::IMUL32rri, X86::IMUL32rmi, makeRMIInst },
- { X86::IMUL32rri8, X86::IMUL32rmi8, makeRMIInst },
- { X86::Int_CMPSDrr, X86::Int_CMPSDrm, makeRMInst },
- { X86::Int_CMPSSrr, X86::Int_CMPSSrm, makeRMInst },
- { X86::Int_COMISDrr, X86::Int_COMISDrm, makeRMInst },
- { X86::Int_COMISSrr, X86::Int_COMISSrm, makeRMInst },
- { X86::Int_CVTDQ2PDrr, X86::Int_CVTDQ2PDrm, makeRMInst },
- { X86::Int_CVTDQ2PSrr, X86::Int_CVTDQ2PSrm, makeRMInst },
- { X86::Int_CVTPD2DQrr, X86::Int_CVTPD2DQrm, makeRMInst },
- { X86::Int_CVTPD2PSrr, X86::Int_CVTPD2PSrm, makeRMInst },
- { X86::Int_CVTPS2DQrr, X86::Int_CVTPS2DQrm, makeRMInst },
- { X86::Int_CVTPS2PDrr, X86::Int_CVTPS2PDrm, makeRMInst },
- { X86::Int_CVTSD2SIrr, X86::Int_CVTSD2SIrm, makeRMInst },
- { X86::Int_CVTSD2SSrr, X86::Int_CVTSD2SSrm, makeRMInst },
- { X86::Int_CVTSI2SDrr, X86::Int_CVTSI2SDrm, makeRMInst },
- { X86::Int_CVTSI2SSrr, X86::Int_CVTSI2SSrm, makeRMInst },
- { X86::Int_CVTSS2SDrr, X86::Int_CVTSS2SDrm, makeRMInst },
- { X86::Int_CVTSS2SIrr, X86::Int_CVTSS2SIrm, makeRMInst },
- { X86::Int_CVTTPD2DQrr, X86::Int_CVTTPD2DQrm, makeRMInst },
- { X86::Int_CVTTPS2DQrr, X86::Int_CVTTPS2DQrm, makeRMInst },
- { X86::Int_CVTTSD2SIrr, X86::Int_CVTTSD2SIrm, makeRMInst },
- { X86::Int_CVTTSS2SIrr, X86::Int_CVTTSS2SIrm, makeRMInst },
- { X86::Int_UCOMISDrr, X86::Int_UCOMISDrm, makeRMInst },
- { X86::Int_UCOMISSrr, X86::Int_UCOMISSrm, makeRMInst },
- { X86::MAXPDrr, X86::MAXPDrm, makeRMInst },
- { X86::MAXPSrr, X86::MAXPSrm, makeRMInst },
- { X86::MINPDrr, X86::MINPDrm, makeRMInst },
- { X86::MINPSrr, X86::MINPSrm, makeRMInst },
- { X86::MOV16rr, X86::MOV16rm, makeRMInst },
- { X86::MOV32rr, X86::MOV32rm, makeRMInst },
- { X86::MOV8rr, X86::MOV8rm, makeRMInst },
- { X86::MOVAPDrr, X86::MOVAPDrm, makeRMInst },
- { X86::MOVAPSrr, X86::MOVAPSrm, makeRMInst },
- { X86::MOVDDUPrr, X86::MOVDDUPrm, makeRMInst },
- { X86::MOVDI2PDIrr, X86::MOVDI2PDIrm, makeRMInst },
- { X86::MOVQI2PQIrr, X86::MOVQI2PQIrm, makeRMInst },
- { X86::MOVSD2PDrr, X86::MOVSD2PDrm, makeRMInst },
- { X86::MOVSDrr, X86::MOVSDrm, makeRMInst },
- { X86::MOVSHDUPrr, X86::MOVSHDUPrm, makeRMInst },
- { X86::MOVSLDUPrr, X86::MOVSLDUPrm, makeRMInst },
- { X86::MOVSS2PSrr, X86::MOVSS2PSrm, makeRMInst },
- { X86::MOVSSrr, X86::MOVSSrm, makeRMInst },
- { X86::MOVSX16rr8, X86::MOVSX16rm8, makeRMInst },
- { X86::MOVSX32rr16, X86::MOVSX32rm16, makeRMInst },
- { X86::MOVSX32rr8, X86::MOVSX32rm8, makeRMInst },
- { X86::MOVUPDrr, X86::MOVUPDrm, makeRMInst },
- { X86::MOVUPSrr, X86::MOVUPSrm, makeRMInst },
- { X86::MOVZX16rr8, X86::MOVZX16rm8, makeRMInst },
- { X86::MOVZX32rr16, X86::MOVZX32rm16, makeRMInst },
- { X86::MOVZX32rr8, X86::MOVZX32rm8, makeRMInst },
- { X86::MULPDrr, X86::MULPDrm, makeRMInst },
- { X86::MULPSrr, X86::MULPSrm, makeRMInst },
- { X86::MULSDrr, X86::MULSDrm, makeRMInst },
- { X86::MULSSrr, X86::MULSSrm, makeRMInst },
- { X86::OR16rr, X86::OR16rm, makeRMInst },
- { X86::OR32rr, X86::OR32rm, makeRMInst },
- { X86::OR8rr, X86::OR8rm, makeRMInst },
- { X86::ORPDrr, X86::ORPDrm, makeRMInst },
- { X86::ORPSrr, X86::ORPSrm, makeRMInst },
- { X86::PACKSSDWrr, X86::PACKSSDWrm, makeRMInst },
- { X86::PACKSSWBrr, X86::PACKSSWBrm, makeRMInst },
- { X86::PACKUSWBrr, X86::PACKUSWBrm, makeRMInst },
- { X86::PADDBrr, X86::PADDBrm, makeRMInst },
- { X86::PADDDrr, X86::PADDDrm, makeRMInst },
- { X86::PADDSBrr, X86::PADDSBrm, makeRMInst },
- { X86::PADDSWrr, X86::PADDSWrm, makeRMInst },
- { X86::PADDWrr, X86::PADDWrm, makeRMInst },
- { X86::PANDNrr, X86::PANDNrm, makeRMInst },
- { X86::PANDrr, X86::PANDrm, makeRMInst },
- { X86::PAVGBrr, X86::PAVGBrm, makeRMInst },
- { X86::PAVGWrr, X86::PAVGWrm, makeRMInst },
- { X86::PCMPEQBrr, X86::PCMPEQBrm, makeRMInst },
- { X86::PCMPEQDrr, X86::PCMPEQDrm, makeRMInst },
- { X86::PCMPEQWrr, X86::PCMPEQWrm, makeRMInst },
- { X86::PCMPGTBrr, X86::PCMPGTBrm, makeRMInst },
- { X86::PCMPGTDrr, X86::PCMPGTDrm, makeRMInst },
- { X86::PCMPGTWrr, X86::PCMPGTWrm, makeRMInst },
- { X86::PINSRWrri, X86::PINSRWrmi, makeRMIInst },
- { X86::PMADDWDrr, X86::PMADDWDrm, makeRMInst },
- { X86::PMAXSWrr, X86::PMAXSWrm, makeRMInst },
- { X86::PMAXUBrr, X86::PMAXUBrm, makeRMInst },
- { X86::PMINSWrr, X86::PMINSWrm, makeRMInst },
- { X86::PMINUBrr, X86::PMINUBrm, makeRMInst },
- { X86::PMULHUWrr, X86::PMULHUWrm, makeRMInst },
- { X86::PMULHWrr, X86::PMULHWrm, makeRMInst },
- { X86::PMULLWrr, X86::PMULLWrm, makeRMInst },
- { X86::PMULUDQrr, X86::PMULUDQrm, makeRMInst },
- { X86::PORrr, X86::PORrm, makeRMInst },
- { X86::PSADBWrr, X86::PSADBWrm, makeRMInst },
- { X86::PSHUFDri, X86::PSHUFDmi, makeRMIInst },
- { X86::PSHUFHWri, X86::PSHUFHWmi, makeRMIInst },
- { X86::PSHUFLWri, X86::PSHUFLWmi, makeRMIInst },
- { X86::PSLLDrr, X86::PSLLDrm, makeRMInst },
- { X86::PSLLQrr, X86::PSLLQrm, makeRMInst },
- { X86::PSLLWrr, X86::PSLLWrm, makeRMInst },
- { X86::PSRADrr, X86::PSRADrm, makeRMInst },
- { X86::PSRAWrr, X86::PSRAWrm, makeRMInst },
- { X86::PSRLDrr, X86::PSRLDrm, makeRMInst },
- { X86::PSRLQrr, X86::PSRLQrm, makeRMInst },
- { X86::PSRLWrr, X86::PSRLWrm, makeRMInst },
- { X86::PSUBBrr, X86::PSUBBrm, makeRMInst },
- { X86::PSUBDrr, X86::PSUBDrm, makeRMInst },
- { X86::PSUBSBrr, X86::PSUBSBrm, makeRMInst },
- { X86::PSUBSWrr, X86::PSUBSWrm, makeRMInst },
- { X86::PSUBWrr, X86::PSUBWrm, makeRMInst },
- { X86::PUNPCKHBWrr, X86::PUNPCKHBWrm, makeRMInst },
- { X86::PUNPCKHDQrr, X86::PUNPCKHDQrm, makeRMInst },
- { X86::PUNPCKHQDQrr, X86::PUNPCKHQDQrm, makeRMInst },
- { X86::PUNPCKHWDrr, X86::PUNPCKHWDrm, makeRMInst },
- { X86::PUNPCKLBWrr, X86::PUNPCKLBWrm, makeRMInst },
- { X86::PUNPCKLDQrr, X86::PUNPCKLDQrm, makeRMInst },
- { X86::PUNPCKLQDQrr, X86::PUNPCKLQDQrm, makeRMInst },
- { X86::PUNPCKLWDrr, X86::PUNPCKLWDrm, makeRMInst },
- { X86::PXORrr, X86::PXORrm, makeRMInst },
- { X86::RCPPSr, X86::RCPPSm, makeRMInst },
- { X86::RSQRTPSr, X86::RSQRTPSm, makeRMInst },
- { X86::SBB32rr, X86::SBB32rm, makeRMInst },
- { X86::SHUFPDrri, X86::SHUFPDrmi, makeRMIInst },
- { X86::SHUFPSrri, X86::SHUFPSrmi, makeRMIInst },
- { X86::SQRTPDr, X86::SQRTPDm, makeRMInst },
- { X86::SQRTPSr, X86::SQRTPSm, makeRMInst },
- { X86::SQRTSDr, X86::SQRTSDm, makeRMInst },
- { X86::SQRTSSr, X86::SQRTSSm, makeRMInst },
- { X86::SUB16rr, X86::SUB16rm, makeRMInst },
- { X86::SUB32rr, X86::SUB32rm, makeRMInst },
- { X86::SUB8rr, X86::SUB8rm, makeRMInst },
- { X86::SUBPDrr, X86::SUBPDrm, makeRMInst },
- { X86::SUBPSrr, X86::SUBPSrm, makeRMInst },
- { X86::SUBSDrr, X86::SUBSDrm, makeRMInst },
- { X86::SUBSSrr, X86::SUBSSrm, makeRMInst },
- { X86::TEST16ri, X86::TEST16mi, makeMIInst },
- { X86::TEST16rr, X86::TEST16rm, makeRMInst },
- { X86::TEST32ri, X86::TEST32mi, makeMIInst },
- { X86::TEST32rr, X86::TEST32rm, makeRMInst },
- { X86::TEST8ri, X86::TEST8mi, makeMIInst },
- { X86::TEST8rr, X86::TEST8rm, makeRMInst },
- { X86::UCOMISDrr, X86::UCOMISDrm, makeRMInst },
- { X86::UCOMISSrr, X86::UCOMISSrm, makeRMInst },
- { X86::UNPCKHPDrr, X86::UNPCKHPDrm, makeRMInst },
- { X86::UNPCKHPSrr, X86::UNPCKHPSrm, makeRMInst },
- { X86::UNPCKLPDrr, X86::UNPCKLPDrm, makeRMInst },
- { X86::UNPCKLPSrr, X86::UNPCKLPSrm, makeRMInst },
- { X86::XCHG16rr, X86::XCHG16rm, makeRMInst },
- { X86::XCHG32rr, X86::XCHG32rm, makeRMInst },
- { X86::XCHG8rr, X86::XCHG8rm, makeRMInst },
- { X86::XOR16rr, X86::XOR16rm, makeRMInst },
- { X86::XOR32rr, X86::XOR32rm, makeRMInst },
- { X86::XOR8rr, X86::XOR8rm, makeRMInst },
- { X86::XORPDrr, X86::XORPDrm, makeRMInst },
- { X86::XORPSrr, X86::XORPSrm, makeRMInst }
+ { X86::CMP16rr, X86::CMP16rm },
+ { X86::CMP32rr, X86::CMP32rm },
+ { X86::CMP8rr, X86::CMP8rm },
+ { X86::CMPPDrri, X86::CMPPDrmi },
+ { X86::CMPPSrri, X86::CMPPSrmi },
+ { X86::CMPSDrr, X86::CMPSDrm },
+ { X86::CMPSSrr, X86::CMPSSrm },
+ { X86::CVTSD2SSrr, X86::CVTSD2SSrm },
+ { X86::CVTSI2SDrr, X86::CVTSI2SDrm },
+ { X86::CVTSI2SSrr, X86::CVTSI2SSrm },
+ { X86::CVTSS2SDrr, X86::CVTSS2SDrm },
+ { X86::CVTTSD2SIrr, X86::CVTTSD2SIrm },
+ { X86::CVTTSS2SIrr, X86::CVTTSS2SIrm },
+ { X86::FsMOVAPDrr, X86::MOVSDrm },
+ { X86::FsMOVAPSrr, X86::MOVSSrm },
+ { X86::IMUL16rri, X86::IMUL16rmi },
+ { X86::IMUL16rri8, X86::IMUL16rmi8 },
+ { X86::IMUL32rri, X86::IMUL32rmi },
+ { X86::IMUL32rri8, X86::IMUL32rmi8 },
+ { X86::Int_CMPSDrr, X86::Int_CMPSDrm },
+ { X86::Int_CMPSSrr, X86::Int_CMPSSrm },
+ { X86::Int_COMISDrr, X86::Int_COMISDrm },
+ { X86::Int_COMISSrr, X86::Int_COMISSrm },
+ { X86::Int_CVTDQ2PDrr, X86::Int_CVTDQ2PDrm },
+ { X86::Int_CVTDQ2PSrr, X86::Int_CVTDQ2PSrm },
+ { X86::Int_CVTPD2DQrr, X86::Int_CVTPD2DQrm },
+ { X86::Int_CVTPD2PSrr, X86::Int_CVTPD2PSrm },
+ { X86::Int_CVTPS2DQrr, X86::Int_CVTPS2DQrm },
+ { X86::Int_CVTPS2PDrr, X86::Int_CVTPS2PDrm },
+ { X86::Int_CVTSD2SIrr, X86::Int_CVTSD2SIrm },
+ { X86::Int_CVTSD2SSrr, X86::Int_CVTSD2SSrm },
+ { X86::Int_CVTSI2SDrr, X86::Int_CVTSI2SDrm },
+ { X86::Int_CVTSI2SSrr, X86::Int_CVTSI2SSrm },
+ { X86::Int_CVTSS2SDrr, X86::Int_CVTSS2SDrm },
+ { X86::Int_CVTSS2SIrr, X86::Int_CVTSS2SIrm },
+ { X86::Int_CVTTPD2DQrr, X86::Int_CVTTPD2DQrm },
+ { X86::Int_CVTTPS2DQrr, X86::Int_CVTTPS2DQrm },
+ { X86::Int_CVTTSD2SIrr, X86::Int_CVTTSD2SIrm },
+ { X86::Int_CVTTSS2SIrr, X86::Int_CVTTSS2SIrm },
+ { X86::Int_UCOMISDrr, X86::Int_UCOMISDrm },
+ { X86::Int_UCOMISSrr, X86::Int_UCOMISSrm },
+ { X86::MOV16rr, X86::MOV16rm },
+ { X86::MOV32rr, X86::MOV32rm },
+ { X86::MOV8rr, X86::MOV8rm },
+ { X86::MOVAPDrr, X86::MOVAPDrm },
+ { X86::MOVAPSrr, X86::MOVAPSrm },
+ { X86::MOVDDUPrr, X86::MOVDDUPrm },
+ { X86::MOVDI2PDIrr, X86::MOVDI2PDIrm },
+ { X86::MOVQI2PQIrr, X86::MOVQI2PQIrm },
+ { X86::MOVSD2PDrr, X86::MOVSD2PDrm },
+ { X86::MOVSDrr, X86::MOVSDrm },
+ { X86::MOVSHDUPrr, X86::MOVSHDUPrm },
+ { X86::MOVSLDUPrr, X86::MOVSLDUPrm },
+ { X86::MOVSS2PSrr, X86::MOVSS2PSrm },
+ { X86::MOVSSrr, X86::MOVSSrm },
+ { X86::MOVSX16rr8, X86::MOVSX16rm8 },
+ { X86::MOVSX32rr16, X86::MOVSX32rm16 },
+ { X86::MOVSX32rr8, X86::MOVSX32rm8 },
+ { X86::MOVUPDrr, X86::MOVUPDrm },
+ { X86::MOVUPSrr, X86::MOVUPSrm },
+ { X86::MOVZX16rr8, X86::MOVZX16rm8 },
+ { X86::MOVZX32rr16, X86::MOVZX32rm16 },
+ { X86::MOVZX32rr8, X86::MOVZX32rm8 },
+ { X86::PSHUFDri, X86::PSHUFDmi },
+ { X86::PSHUFHWri, X86::PSHUFHWmi },
+ { X86::PSHUFLWri, X86::PSHUFLWmi },
+ { X86::TEST16rr, X86::TEST16rm },
+ { X86::TEST32rr, X86::TEST32rm },
+ { X86::TEST8rr, X86::TEST8rm },
+ { X86::UCOMISDrr, X86::UCOMISDrm },
+ { X86::UCOMISSrr, X86::UCOMISSrm },
+ { X86::XCHG16rr, X86::XCHG16rm },
+ { X86::XCHG32rr, X86::XCHG32rm },
+ { X86::XCHG8rr, X86::XCHG8rm }
+ };
+ ASSERT_SORTED(OpcodeTable);
+ OpcodeTablePtr = OpcodeTable;
+ OpcodeTableSize = ARRAY_SIZE(OpcodeTable);
+ } else if (i == 2) {
+ static const TableEntry OpcodeTable[] = {
+ { X86::ADC32rr, X86::ADC32rm },
+ { X86::ADD16rr, X86::ADD16rm },
+ { X86::ADD32rr, X86::ADD32rm },
+ { X86::ADD8rr, X86::ADD8rm },
+ { X86::ADDPDrr, X86::ADDPDrm },
+ { X86::ADDPSrr, X86::ADDPSrm },
+ { X86::ADDSDrr, X86::ADDSDrm },
+ { X86::ADDSSrr, X86::ADDSSrm },
+ { X86::ADDSUBPDrr, X86::ADDSUBPDrm },
+ { X86::ADDSUBPSrr, X86::ADDSUBPSrm },
+ { X86::AND16rr, X86::AND16rm },
+ { X86::AND32rr, X86::AND32rm },
+ { X86::AND8rr, X86::AND8rm },
+ { X86::ANDNPDrr, X86::ANDNPDrm },
+ { X86::ANDNPSrr, X86::ANDNPSrm },
+ { X86::ANDPDrr, X86::ANDPDrm },
+ { X86::ANDPSrr, X86::ANDPSrm },
+ { X86::CMOVA16rr, X86::CMOVA16rm },
+ { X86::CMOVA32rr, X86::CMOVA32rm },
+ { X86::CMOVAE16rr, X86::CMOVAE16rm },
+ { X86::CMOVAE32rr, X86::CMOVAE32rm },
+ { X86::CMOVB16rr, X86::CMOVB16rm },
+ { X86::CMOVB32rr, X86::CMOVB32rm },
+ { X86::CMOVBE16rr, X86::CMOVBE16rm },
+ { X86::CMOVBE32rr, X86::CMOVBE32rm },
+ { X86::CMOVE16rr, X86::CMOVE16rm },
+ { X86::CMOVE32rr, X86::CMOVE32rm },
+ { X86::CMOVG16rr, X86::CMOVG16rm },
+ { X86::CMOVG32rr, X86::CMOVG32rm },
+ { X86::CMOVGE16rr, X86::CMOVGE16rm },
+ { X86::CMOVGE32rr, X86::CMOVGE32rm },
+ { X86::CMOVL16rr, X86::CMOVL16rm },
+ { X86::CMOVL32rr, X86::CMOVL32rm },
+ { X86::CMOVLE16rr, X86::CMOVLE16rm },
+ { X86::CMOVLE32rr, X86::CMOVLE32rm },
+ { X86::CMOVNE16rr, X86::CMOVNE16rm },
+ { X86::CMOVNE32rr, X86::CMOVNE32rm },
+ { X86::CMOVNP16rr, X86::CMOVNP16rm },
+ { X86::CMOVNP32rr, X86::CMOVNP32rm },
+ { X86::CMOVNS16rr, X86::CMOVNS16rm },
+ { X86::CMOVNS32rr, X86::CMOVNS32rm },
+ { X86::CMOVP16rr, X86::CMOVP16rm },
+ { X86::CMOVP32rr, X86::CMOVP32rm },
+ { X86::CMOVS16rr, X86::CMOVS16rm },
+ { X86::CMOVS32rr, X86::CMOVS32rm },
+ { X86::DIVPDrr, X86::DIVPDrm },
+ { X86::DIVPSrr, X86::DIVPSrm },
+ { X86::DIVSDrr, X86::DIVSDrm },
+ { X86::DIVSSrr, X86::DIVSSrm },
+ { X86::HADDPDrr, X86::HADDPDrm },
+ { X86::HADDPSrr, X86::HADDPSrm },
+ { X86::HSUBPDrr, X86::HSUBPDrm },
+ { X86::HSUBPSrr, X86::HSUBPSrm },
+ { X86::IMUL16rr, X86::IMUL16rm },
+ { X86::IMUL32rr, X86::IMUL32rm },
+ { X86::MAXPDrr, X86::MAXPDrm },
+ { X86::MAXPSrr, X86::MAXPSrm },
+ { X86::MINPDrr, X86::MINPDrm },
+ { X86::MINPSrr, X86::MINPSrm },
+ { X86::MULPDrr, X86::MULPDrm },
+ { X86::MULPSrr, X86::MULPSrm },
+ { X86::MULSDrr, X86::MULSDrm },
+ { X86::MULSSrr, X86::MULSSrm },
+ { X86::OR16rr, X86::OR16rm },
+ { X86::OR32rr, X86::OR32rm },
+ { X86::OR8rr, X86::OR8rm },
+ { X86::ORPDrr, X86::ORPDrm },
+ { X86::ORPSrr, X86::ORPSrm },
+ { X86::PACKSSDWrr, X86::PACKSSDWrm },
+ { X86::PACKSSWBrr, X86::PACKSSWBrm },
+ { X86::PACKUSWBrr, X86::PACKUSWBrm },
+ { X86::PADDBrr, X86::PADDBrm },
+ { X86::PADDDrr, X86::PADDDrm },
+ { X86::PADDSBrr, X86::PADDSBrm },
+ { X86::PADDSWrr, X86::PADDSWrm },
+ { X86::PADDWrr, X86::PADDWrm },
+ { X86::PANDNrr, X86::PANDNrm },
+ { X86::PANDrr, X86::PANDrm },
+ { X86::PAVGBrr, X86::PAVGBrm },
+ { X86::PAVGWrr, X86::PAVGWrm },
+ { X86::PCMPEQBrr, X86::PCMPEQBrm },
+ { X86::PCMPEQDrr, X86::PCMPEQDrm },
+ { X86::PCMPEQWrr, X86::PCMPEQWrm },
+ { X86::PCMPGTBrr, X86::PCMPGTBrm },
+ { X86::PCMPGTDrr, X86::PCMPGTDrm },
+ { X86::PCMPGTWrr, X86::PCMPGTWrm },
+ { X86::PINSRWrri, X86::PINSRWrmi },
+ { X86::PMADDWDrr, X86::PMADDWDrm },
+ { X86::PMAXSWrr, X86::PMAXSWrm },
+ { X86::PMAXUBrr, X86::PMAXUBrm },
+ { X86::PMINSWrr, X86::PMINSWrm },
+ { X86::PMINUBrr, X86::PMINUBrm },
+ { X86::PMULHUWrr, X86::PMULHUWrm },
+ { X86::PMULHWrr, X86::PMULHWrm },
+ { X86::PMULLWrr, X86::PMULLWrm },
+ { X86::PMULUDQrr, X86::PMULUDQrm },
+ { X86::PORrr, X86::PORrm },
+ { X86::PSADBWrr, X86::PSADBWrm },
+ { X86::PSLLDrr, X86::PSLLDrm },
+ { X86::PSLLQrr, X86::PSLLQrm },
+ { X86::PSLLWrr, X86::PSLLWrm },
+ { X86::PSRADrr, X86::PSRADrm },
+ { X86::PSRAWrr, X86::PSRAWrm },
+ { X86::PSRLDrr, X86::PSRLDrm },
+ { X86::PSRLQrr, X86::PSRLQrm },
+ { X86::PSRLWrr, X86::PSRLWrm },
+ { X86::PSUBBrr, X86::PSUBBrm },
+ { X86::PSUBDrr, X86::PSUBDrm },
+ { X86::PSUBSBrr, X86::PSUBSBrm },
+ { X86::PSUBSWrr, X86::PSUBSWrm },
+ { X86::PSUBWrr, X86::PSUBWrm },
+ { X86::PUNPCKHBWrr, X86::PUNPCKHBWrm },
+ { X86::PUNPCKHDQrr, X86::PUNPCKHDQrm },
+ { X86::PUNPCKHQDQrr, X86::PUNPCKHQDQrm },
+ { X86::PUNPCKHWDrr, X86::PUNPCKHWDrm },
+ { X86::PUNPCKLBWrr, X86::PUNPCKLBWrm },
+ { X86::PUNPCKLDQrr, X86::PUNPCKLDQrm },
+ { X86::PUNPCKLQDQrr, X86::PUNPCKLQDQrm },
+ { X86::PUNPCKLWDrr, X86::PUNPCKLWDrm },
+ { X86::PXORrr, X86::PXORrm },
+ { X86::RCPPSr, X86::RCPPSm },
+ { X86::RSQRTPSr, X86::RSQRTPSm },
+ { X86::SBB32rr, X86::SBB32rm },
+ { X86::SHUFPDrri, X86::SHUFPDrmi },
+ { X86::SHUFPSrri, X86::SHUFPSrmi },
+ { X86::SQRTPDr, X86::SQRTPDm },
+ { X86::SQRTPSr, X86::SQRTPSm },
+ { X86::SQRTSDr, X86::SQRTSDm },
+ { X86::SQRTSSr, X86::SQRTSSm },
+ { X86::SUB16rr, X86::SUB16rm },
+ { X86::SUB32rr, X86::SUB32rm },
+ { X86::SUB8rr, X86::SUB8rm },
+ { X86::SUBPDrr, X86::SUBPDrm },
+ { X86::SUBPSrr, X86::SUBPSrm },
+ { X86::SUBSDrr, X86::SUBSDrm },
+ { X86::SUBSSrr, X86::SUBSSrm },
+ { X86::UNPCKHPDrr, X86::UNPCKHPDrm },
+ { X86::UNPCKHPSrr, X86::UNPCKHPSrm },
+ { X86::UNPCKLPDrr, X86::UNPCKLPDrm },
+ { X86::UNPCKLPSrr, X86::UNPCKLPSrm },
+ { X86::XOR16rr, X86::XOR16rm },
+ { X86::XOR32rr, X86::XOR32rm },
+ { X86::XOR8rr, X86::XOR8rm },
+ { X86::XORPDrr, X86::XORPDrm },
+ { X86::XORPSrr, X86::XORPSrm }
};
ASSERT_SORTED(OpcodeTable);
OpcodeTablePtr = OpcodeTable;
OpcodeTableSize = ARRAY_SIZE(OpcodeTable);
}
- // If table selected
+ // If table selected...
if (OpcodeTablePtr) {
- // Opcode to fuse
+ // Find the Opcode to fuse
unsigned fromOpcode = MI->getOpcode();
// Lookup fromOpcode in table
- const TableEntry *entry = TableLookup(OpcodeTablePtr, OpcodeTableSize,
- fromOpcode);
-
- // If opcode found in table
- if (entry) {
- // Fused opcode
- unsigned toOpcode = entry->to;
+ if (const TableEntry *Entry = TableLookup(OpcodeTablePtr, OpcodeTableSize,
+ fromOpcode)) {
+ if (isTwoAddrFold)
+ return FuseTwoAddrInst(Entry->to, FrameIndex, MI);
- // Make new instruction
- switch (entry->make) {
- case makeM0Inst: return MakeM0Inst(toOpcode, FrameIndex, MI);
- case makeMIInst: return MakeMIInst(toOpcode, FrameIndex, MI);
- case makeMInst: return MakeMInst(toOpcode, FrameIndex, MI);
- case makeMRIInst: return MakeMRIInst(toOpcode, FrameIndex, MI);
- case makeMRInst: return MakeMRInst(toOpcode, FrameIndex, MI);
- case makeRMIInst: return MakeRMIInst(toOpcode, FrameIndex, MI);
- case makeRMInst: return MakeRMInst(toOpcode, FrameIndex, MI);
- default: assert(0 && "Unknown instruction make");
- }
+ return FuseInst(Entry->to, i, FrameIndex, MI);
}
}
@@ -761,8 +752,7 @@
MachineInstr *New = 0;
if (Old->getOpcode() == X86::ADJCALLSTACKDOWN) {
- New=BuildMI(X86::SUB32ri, 1, X86::ESP, MachineOperand::UseAndDef)
- .addImm(Amount);
+ New=BuildMI(X86::SUB32ri, 2, X86::ESP).addReg(X86::ESP).addImm(Amount);
} else {
assert(Old->getOpcode() == X86::ADJCALLSTACKUP);
// factor out the amount the callee already popped.
@@ -770,8 +760,7 @@
Amount -= CalleeAmt;
if (Amount) {
unsigned Opc = Amount < 128 ? X86::ADD32ri8 : X86::ADD32ri;
- New = BuildMI(Opc, 1, X86::ESP,
- MachineOperand::UseAndDef).addImm(Amount);
+ New = BuildMI(Opc, 2, X86::ESP).addReg(X86::ESP).addImm(Amount);
}
}
@@ -785,8 +774,7 @@
if (unsigned CalleeAmt = I->getOperand(1).getImmedValue()) {
unsigned Opc = CalleeAmt < 128 ? X86::SUB32ri8 : X86::SUB32ri;
MachineInstr *New =
- BuildMI(Opc, 1, X86::ESP,
- MachineOperand::UseAndDef).addImm(CalleeAmt);
+ BuildMI(Opc, 1, X86::ESP).addReg(X86::ESP).addImm(CalleeAmt);
MBB.insert(I, New);
}
}
@@ -870,7 +858,7 @@
MBB.insert(MBBI, MI);
} else {
unsigned Opc = NumBytes < 128 ? X86::SUB32ri8 : X86::SUB32ri;
- MI = BuildMI(Opc, 1, X86::ESP,MachineOperand::UseAndDef).addImm(NumBytes);
+ MI = BuildMI(Opc, 2, X86::ESP).addReg(X86::ESP).addImm(NumBytes);
MBB.insert(MBBI, MI);
}
}
@@ -897,7 +885,7 @@
// If it's main() on Cygwin\Mingw32 we should align stack as well
if (Fn->hasExternalLinkage() && Fn->getName() == "main" &&
Subtarget->TargetType == X86Subtarget::isCygwin) {
- MI = BuildMI(X86::AND32ri, 2, X86::ESP).addImm(-Align);
+ MI = BuildMI(X86::AND32ri, 2, X86::ESP).addReg(X86::ESP).addImm(-Align);
MBB.insert(MBBI, MI);
// Probe the stack
@@ -929,7 +917,7 @@
int EBPOffset = MFI->getObjectOffset(MFI->getObjectIndexEnd()-1)+4;
// mov ESP, EBP
- BuildMI(MBB, MBBI, X86::MOV32rr, 1,X86::ESP).addReg(X86::EBP);
+ BuildMI(MBB, MBBI, X86::MOV32rr, 1, X86::ESP).addReg(X86::EBP);
// pop EBP
BuildMI(MBB, MBBI, X86::POP32r, 0, X86::EBP);
@@ -960,12 +948,10 @@
if (NumBytes > 0) {
unsigned Opc = NumBytes < 128 ? X86::ADD32ri8 : X86::ADD32ri;
- BuildMI(MBB, MBBI, Opc, 2)
- .addReg(X86::ESP, MachineOperand::UseAndDef).addImm(NumBytes);
+ BuildMI(MBB, MBBI, Opc, 2, X86::ESP).addReg(X86::ESP).addImm(NumBytes);
} else if ((int)NumBytes < 0) {
unsigned Opc = -NumBytes < 128 ? X86::SUB32ri8 : X86::SUB32ri;
- BuildMI(MBB, MBBI, Opc, 2)
- .addReg(X86::ESP, MachineOperand::UseAndDef).addImm(-NumBytes);
+ BuildMI(MBB, MBBI, Opc, 2, X86::ESP).addReg(X86::ESP).addImm(-NumBytes);
}
}
}
diff --git a/lib/Target/X86/X86RegisterInfo.h b/lib/Target/X86/X86RegisterInfo.h
index 8650d06..e86cc28 100644
--- a/lib/Target/X86/X86RegisterInfo.h
+++ b/lib/Target/X86/X86RegisterInfo.h
@@ -15,15 +15,15 @@
#define X86REGISTERINFO_H
#include "llvm/Target/MRegisterInfo.h"
-
-class llvm::Type;
-
#include "X86GenRegisterInfo.h.inc"
namespace llvm {
+ class Type;
+ class TargetInstrInfo;
struct X86RegisterInfo : public X86GenRegisterInfo {
- X86RegisterInfo();
+ const TargetInstrInfo &TII;
+ X86RegisterInfo(const TargetInstrInfo &tii);
/// Code Generation virtual methods...
void storeRegToStackSlot(MachineBasicBlock &MBB,
diff --git a/utils/TableGen/AsmWriterEmitter.cpp b/utils/TableGen/AsmWriterEmitter.cpp
index 02773d2..2ea7449 100644
--- a/utils/TableGen/AsmWriterEmitter.cpp
+++ b/utils/TableGen/AsmWriterEmitter.cpp
@@ -211,15 +211,12 @@
unsigned OpNo = CGI.getOperandNamed(VarName);
CodeGenInstruction::OperandInfo OpInfo = CGI.OperandList[OpNo];
- // If this is a two-address instruction and we are not accessing the
- // 0th operand, remove an operand.
+ // If this is a two-address instruction, verify the second operand isn't
+ // used.
unsigned MIOp = OpInfo.MIOperandNo;
- if (CGI.isTwoAddress && MIOp != 0) {
- if (MIOp == 1)
- throw "Should refer to operand #0 instead of #1 for two-address"
- " instruction '" + CGI.TheDef->getName() + "'!";
- --MIOp;
- }
+ if (CGI.isTwoAddress && MIOp == 1)
+ throw "Should refer to operand #0 instead of #1 for two-address"
+ " instruction '" + CGI.TheDef->getName() + "'!";
if (CurVariant == Variant || CurVariant == ~0U)
Operands.push_back(AsmWriterOperand(OpInfo.PrinterMethodName, MIOp,