blob: 028039d0b0ccbbf384c189e4c17d366f978f7cac [file] [log] [blame]
Igor Bregerf7359d82017-02-22 12:25:09 +00001//===- X86InstructionSelector.cpp ----------------------------*- C++ -*-==//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9/// \file
10/// This file implements the targeting of the InstructionSelector class for
11/// X86.
12/// \todo This should be generated by TableGen.
13//===----------------------------------------------------------------------===//
14
Igor Bregera8ba5722017-03-23 15:25:57 +000015#include "X86InstrBuilder.h"
Igor Bregerf7359d82017-02-22 12:25:09 +000016#include "X86InstrInfo.h"
17#include "X86RegisterBankInfo.h"
18#include "X86RegisterInfo.h"
19#include "X86Subtarget.h"
20#include "X86TargetMachine.h"
Igor Breger3b97ea32017-04-12 12:54:54 +000021#include "llvm/CodeGen/GlobalISel/InstructionSelector.h"
Igor Breger28f290f2017-05-17 12:48:08 +000022#include "llvm/CodeGen/GlobalISel/Utils.h"
Igor Bregerf7359d82017-02-22 12:25:09 +000023#include "llvm/CodeGen/MachineBasicBlock.h"
24#include "llvm/CodeGen/MachineFunction.h"
25#include "llvm/CodeGen/MachineInstr.h"
26#include "llvm/CodeGen/MachineInstrBuilder.h"
Daniel Sanders0b5293f2017-04-06 09:49:34 +000027#include "llvm/CodeGen/MachineOperand.h"
Igor Bregerf7359d82017-02-22 12:25:09 +000028#include "llvm/CodeGen/MachineRegisterInfo.h"
29#include "llvm/IR/Type.h"
30#include "llvm/Support/Debug.h"
31#include "llvm/Support/raw_ostream.h"
32
33#define DEBUG_TYPE "X86-isel"
34
Daniel Sanders6ab0daa2017-07-04 14:35:06 +000035#include "llvm/CodeGen/GlobalISel/InstructionSelectorImpl.h"
36
Igor Bregerf7359d82017-02-22 12:25:09 +000037using namespace llvm;
38
Daniel Sanders0b5293f2017-04-06 09:49:34 +000039namespace {
40
Daniel Sanderse7b0d662017-04-21 15:59:56 +000041#define GET_GLOBALISEL_PREDICATE_BITSET
42#include "X86GenGlobalISel.inc"
43#undef GET_GLOBALISEL_PREDICATE_BITSET
44
Daniel Sanders0b5293f2017-04-06 09:49:34 +000045class X86InstructionSelector : public InstructionSelector {
46public:
Daniel Sanderse7b0d662017-04-21 15:59:56 +000047 X86InstructionSelector(const X86TargetMachine &TM, const X86Subtarget &STI,
Daniel Sanders0b5293f2017-04-06 09:49:34 +000048 const X86RegisterBankInfo &RBI);
49
50 bool select(MachineInstr &I) const override;
51
52private:
53 /// tblgen-erated 'select' implementation, used as the initial selector for
54 /// the patterns that don't require complex C++.
55 bool selectImpl(MachineInstr &I) const;
56
Hiroshi Inouebb703e82017-07-02 03:24:54 +000057 // TODO: remove after supported by Tablegen-erated instruction selection.
Daniel Sanders0b5293f2017-04-06 09:49:34 +000058 unsigned getLoadStoreOp(LLT &Ty, const RegisterBank &RB, unsigned Opc,
59 uint64_t Alignment) const;
60
Daniel Sanders0b5293f2017-04-06 09:49:34 +000061 bool selectLoadStoreOp(MachineInstr &I, MachineRegisterInfo &MRI,
62 MachineFunction &MF) const;
Igor Breger810c6252017-05-08 09:40:43 +000063 bool selectFrameIndexOrGep(MachineInstr &I, MachineRegisterInfo &MRI,
64 MachineFunction &MF) const;
Igor Breger717bd362017-07-02 08:58:29 +000065 bool selectGlobalValue(MachineInstr &I, MachineRegisterInfo &MRI,
66 MachineFunction &MF) const;
Igor Breger3b97ea32017-04-12 12:54:54 +000067 bool selectConstant(MachineInstr &I, MachineRegisterInfo &MRI,
68 MachineFunction &MF) const;
Igor Breger4fdf1e42017-04-19 11:34:59 +000069 bool selectTrunc(MachineInstr &I, MachineRegisterInfo &MRI,
70 MachineFunction &MF) const;
Igor Bregerfda31e62017-05-10 06:52:58 +000071 bool selectZext(MachineInstr &I, MachineRegisterInfo &MRI,
72 MachineFunction &MF) const;
Igor Bregerc7b59772017-05-11 07:17:40 +000073 bool selectCmp(MachineInstr &I, MachineRegisterInfo &MRI,
74 MachineFunction &MF) const;
Igor Breger28f290f2017-05-17 12:48:08 +000075 bool selectUadde(MachineInstr &I, MachineRegisterInfo &MRI,
76 MachineFunction &MF) const;
Igor Breger1dcd5e82017-06-20 09:15:10 +000077 bool selectCopy(MachineInstr &I, MachineRegisterInfo &MRI) const;
Igor Bregerb186a692017-07-02 08:15:49 +000078 bool selectUnmergeValues(MachineInstr &I, MachineRegisterInfo &MRI,
79 MachineFunction &MF) const;
Igor Breger0cddd342017-06-29 12:08:28 +000080 bool selectMergeValues(MachineInstr &I, MachineRegisterInfo &MRI,
81 MachineFunction &MF) const;
Igor Breger1c29be72017-06-22 09:43:35 +000082 bool selectInsert(MachineInstr &I, MachineRegisterInfo &MRI,
83 MachineFunction &MF) const;
Igor Bregerf5035d62017-06-25 11:42:17 +000084 bool selectExtract(MachineInstr &I, MachineRegisterInfo &MRI,
85 MachineFunction &MF) const;
Igor Breger1c29be72017-06-22 09:43:35 +000086
87 // emit insert subreg instruction and insert it before MachineInstr &I
88 bool emitInsertSubreg(unsigned DstReg, unsigned SrcReg, MachineInstr &I,
89 MachineRegisterInfo &MRI, MachineFunction &MF) const;
Igor Bregerf5035d62017-06-25 11:42:17 +000090 // emit extract subreg instruction and insert it before MachineInstr &I
91 bool emitExtractSubreg(unsigned DstReg, unsigned SrcReg, MachineInstr &I,
92 MachineRegisterInfo &MRI, MachineFunction &MF) const;
Igor Breger1dcd5e82017-06-20 09:15:10 +000093
94 const TargetRegisterClass *getRegClass(LLT Ty, const RegisterBank &RB) const;
95 const TargetRegisterClass *getRegClass(LLT Ty, unsigned Reg,
96 MachineRegisterInfo &MRI) const;
Igor Breger28f290f2017-05-17 12:48:08 +000097
Daniel Sanderse7b0d662017-04-21 15:59:56 +000098 const X86TargetMachine &TM;
Daniel Sanders0b5293f2017-04-06 09:49:34 +000099 const X86Subtarget &STI;
100 const X86InstrInfo &TII;
101 const X86RegisterInfo &TRI;
102 const X86RegisterBankInfo &RBI;
Daniel Sanderse7b0d662017-04-21 15:59:56 +0000103
Daniel Sanderse9fdba32017-04-29 17:30:09 +0000104#define GET_GLOBALISEL_PREDICATES_DECL
105#include "X86GenGlobalISel.inc"
106#undef GET_GLOBALISEL_PREDICATES_DECL
Daniel Sanders0b5293f2017-04-06 09:49:34 +0000107
108#define GET_GLOBALISEL_TEMPORARIES_DECL
109#include "X86GenGlobalISel.inc"
110#undef GET_GLOBALISEL_TEMPORARIES_DECL
111};
112
113} // end anonymous namespace
114
Daniel Sanders8a4bae92017-03-14 21:32:08 +0000115#define GET_GLOBALISEL_IMPL
Igor Bregerf7359d82017-02-22 12:25:09 +0000116#include "X86GenGlobalISel.inc"
Daniel Sanders8a4bae92017-03-14 21:32:08 +0000117#undef GET_GLOBALISEL_IMPL
Igor Bregerf7359d82017-02-22 12:25:09 +0000118
Daniel Sanderse7b0d662017-04-21 15:59:56 +0000119X86InstructionSelector::X86InstructionSelector(const X86TargetMachine &TM,
120 const X86Subtarget &STI,
Igor Bregerf7359d82017-02-22 12:25:09 +0000121 const X86RegisterBankInfo &RBI)
Daniel Sanderse7b0d662017-04-21 15:59:56 +0000122 : InstructionSelector(), TM(TM), STI(STI), TII(*STI.getInstrInfo()),
Daniel Sanderse9fdba32017-04-29 17:30:09 +0000123 TRI(*STI.getRegisterInfo()), RBI(RBI),
124#define GET_GLOBALISEL_PREDICATES_INIT
125#include "X86GenGlobalISel.inc"
126#undef GET_GLOBALISEL_PREDICATES_INIT
Daniel Sanders8a4bae92017-03-14 21:32:08 +0000127#define GET_GLOBALISEL_TEMPORARIES_INIT
128#include "X86GenGlobalISel.inc"
129#undef GET_GLOBALISEL_TEMPORARIES_INIT
130{
131}
Igor Bregerf7359d82017-02-22 12:25:09 +0000132
133// FIXME: This should be target-independent, inferred from the types declared
134// for each class in the bank.
Igor Breger1dcd5e82017-06-20 09:15:10 +0000135const TargetRegisterClass *
136X86InstructionSelector::getRegClass(LLT Ty, const RegisterBank &RB) const {
Igor Bregerf7359d82017-02-22 12:25:09 +0000137 if (RB.getID() == X86::GPRRegBankID) {
Igor Breger4fdf1e42017-04-19 11:34:59 +0000138 if (Ty.getSizeInBits() <= 8)
139 return &X86::GR8RegClass;
140 if (Ty.getSizeInBits() == 16)
141 return &X86::GR16RegClass;
Igor Breger321cf3c2017-03-03 08:06:46 +0000142 if (Ty.getSizeInBits() == 32)
Igor Bregerf7359d82017-02-22 12:25:09 +0000143 return &X86::GR32RegClass;
144 if (Ty.getSizeInBits() == 64)
145 return &X86::GR64RegClass;
146 }
Igor Breger321cf3c2017-03-03 08:06:46 +0000147 if (RB.getID() == X86::VECRRegBankID) {
148 if (Ty.getSizeInBits() == 32)
Igor Breger1dcd5e82017-06-20 09:15:10 +0000149 return STI.hasAVX512() ? &X86::FR32XRegClass : &X86::FR32RegClass;
Igor Breger321cf3c2017-03-03 08:06:46 +0000150 if (Ty.getSizeInBits() == 64)
Igor Breger1dcd5e82017-06-20 09:15:10 +0000151 return STI.hasAVX512() ? &X86::FR64XRegClass : &X86::FR64RegClass;
Igor Breger321cf3c2017-03-03 08:06:46 +0000152 if (Ty.getSizeInBits() == 128)
Igor Breger1dcd5e82017-06-20 09:15:10 +0000153 return STI.hasAVX512() ? &X86::VR128XRegClass : &X86::VR128RegClass;
Igor Breger321cf3c2017-03-03 08:06:46 +0000154 if (Ty.getSizeInBits() == 256)
Igor Breger1dcd5e82017-06-20 09:15:10 +0000155 return STI.hasAVX512() ? &X86::VR256XRegClass : &X86::VR256RegClass;
Igor Breger321cf3c2017-03-03 08:06:46 +0000156 if (Ty.getSizeInBits() == 512)
157 return &X86::VR512RegClass;
158 }
Igor Bregerf7359d82017-02-22 12:25:09 +0000159
160 llvm_unreachable("Unknown RegBank!");
161}
162
Igor Breger1dcd5e82017-06-20 09:15:10 +0000163const TargetRegisterClass *
164X86InstructionSelector::getRegClass(LLT Ty, unsigned Reg,
165 MachineRegisterInfo &MRI) const {
166 const RegisterBank &RegBank = *RBI.getRegBank(Reg, MRI, TRI);
167 return getRegClass(Ty, RegBank);
168}
169
Igor Bregerb3a860a2017-08-20 07:14:40 +0000170unsigned getSubRegIndex(const TargetRegisterClass *RC) {
171 unsigned SubIdx = X86::NoSubRegister;
172 if (RC == &X86::GR32RegClass) {
173 SubIdx = X86::sub_32bit;
174 } else if (RC == &X86::GR16RegClass) {
175 SubIdx = X86::sub_16bit;
176 } else if (RC == &X86::GR8RegClass) {
177 SubIdx = X86::sub_8bit;
178 }
179
180 return SubIdx;
181}
182
183const TargetRegisterClass *getRegClassFromGRPhysReg(unsigned Reg) {
184 assert(TargetRegisterInfo::isPhysicalRegister(Reg));
185 if (X86::GR64RegClass.contains(Reg))
186 return &X86::GR64RegClass;
187 if (X86::GR32RegClass.contains(Reg))
188 return &X86::GR32RegClass;
189 if (X86::GR16RegClass.contains(Reg))
190 return &X86::GR16RegClass;
191 if (X86::GR8RegClass.contains(Reg))
192 return &X86::GR8RegClass;
193
194 llvm_unreachable("Unknown RegClass for PhysReg!");
195}
196
Igor Bregerf7359d82017-02-22 12:25:09 +0000197// Set X86 Opcode and constrain DestReg.
Igor Breger1dcd5e82017-06-20 09:15:10 +0000198bool X86InstructionSelector::selectCopy(MachineInstr &I,
199 MachineRegisterInfo &MRI) const {
Igor Bregerf7359d82017-02-22 12:25:09 +0000200
201 unsigned DstReg = I.getOperand(0).getReg();
Igor Bregerb3a860a2017-08-20 07:14:40 +0000202 const unsigned DstSize = RBI.getSizeInBits(DstReg, MRI, TRI);
203 const RegisterBank &DstRegBank = *RBI.getRegBank(DstReg, MRI, TRI);
Igor Bregerf7359d82017-02-22 12:25:09 +0000204
Igor Bregerf7359d82017-02-22 12:25:09 +0000205 unsigned SrcReg = I.getOperand(1).getReg();
206 const unsigned SrcSize = RBI.getSizeInBits(SrcReg, MRI, TRI);
Igor Bregerb3a860a2017-08-20 07:14:40 +0000207 const RegisterBank &SrcRegBank = *RBI.getRegBank(SrcReg, MRI, TRI);
208
209 if (TargetRegisterInfo::isPhysicalRegister(DstReg)) {
210 assert(I.isCopy() && "Generic operators do not allow physical registers");
211
212 if (DstSize > SrcSize && SrcRegBank.getID() == X86::GPRRegBankID &&
213 DstRegBank.getID() == X86::GPRRegBankID) {
214
215 const TargetRegisterClass *SrcRC =
216 getRegClass(MRI.getType(SrcReg), SrcRegBank);
217 const TargetRegisterClass *DstRC = getRegClassFromGRPhysReg(DstReg);
218
219 if (SrcRC != DstRC) {
220 // This case can be generated by ABI lowering, performe anyext
221 unsigned ExtSrc = MRI.createVirtualRegister(DstRC);
222 BuildMI(*I.getParent(), I, I.getDebugLoc(),
223 TII.get(TargetOpcode::SUBREG_TO_REG))
224 .addDef(ExtSrc)
225 .addImm(0)
226 .addReg(SrcReg)
227 .addImm(getSubRegIndex(SrcRC));
228
229 I.getOperand(1).setReg(ExtSrc);
230 }
231 }
232
233 return true;
234 }
Igor Breger360d0f22017-04-27 08:02:03 +0000235
Igor Bregerf7359d82017-02-22 12:25:09 +0000236 assert((!TargetRegisterInfo::isPhysicalRegister(SrcReg) || I.isCopy()) &&
237 "No phys reg on generic operators");
238 assert((DstSize == SrcSize ||
239 // Copies are a mean to setup initial types, the number of
240 // bits may not exactly match.
241 (TargetRegisterInfo::isPhysicalRegister(SrcReg) &&
242 DstSize <= RBI.getSizeInBits(SrcReg, MRI, TRI))) &&
243 "Copy with different width?!");
244
Igor Bregerb3a860a2017-08-20 07:14:40 +0000245 const TargetRegisterClass *DstRC =
246 getRegClass(MRI.getType(DstReg), DstRegBank);
Igor Bregerf7359d82017-02-22 12:25:09 +0000247
Igor Bregerb3a860a2017-08-20 07:14:40 +0000248 if (SrcRegBank.getID() == X86::GPRRegBankID &&
249 DstRegBank.getID() == X86::GPRRegBankID && SrcSize > DstSize &&
250 TargetRegisterInfo::isPhysicalRegister(SrcReg)) {
251 // Change the physical register to performe truncate.
Igor Breger360d0f22017-04-27 08:02:03 +0000252
Igor Bregerb3a860a2017-08-20 07:14:40 +0000253 const TargetRegisterClass *SrcRC = getRegClassFromGRPhysReg(SrcReg);
Igor Breger360d0f22017-04-27 08:02:03 +0000254
Igor Bregerb3a860a2017-08-20 07:14:40 +0000255 if (DstRC != SrcRC) {
256 I.getOperand(1).setSubReg(getSubRegIndex(DstRC));
Igor Breger360d0f22017-04-27 08:02:03 +0000257 I.getOperand(1).substPhysReg(SrcReg, TRI);
258 }
Igor Bregerf7359d82017-02-22 12:25:09 +0000259 }
260
261 // No need to constrain SrcReg. It will get constrained when
262 // we hit another of its use or its defs.
263 // Copies do not have constraints.
Igor Breger8a924be2017-03-23 12:13:29 +0000264 const TargetRegisterClass *OldRC = MRI.getRegClassOrNull(DstReg);
Igor Bregerb3a860a2017-08-20 07:14:40 +0000265 if (!OldRC || !DstRC->hasSubClassEq(OldRC)) {
266 if (!RBI.constrainGenericRegister(DstReg, *DstRC, MRI)) {
Igor Breger8a924be2017-03-23 12:13:29 +0000267 DEBUG(dbgs() << "Failed to constrain " << TII.getName(I.getOpcode())
268 << " operand\n");
269 return false;
270 }
Igor Bregerf7359d82017-02-22 12:25:09 +0000271 }
272 I.setDesc(TII.get(X86::COPY));
273 return true;
274}
275
276bool X86InstructionSelector::select(MachineInstr &I) const {
277 assert(I.getParent() && "Instruction should be in a basic block!");
278 assert(I.getParent()->getParent() && "Instruction should be in a function!");
279
280 MachineBasicBlock &MBB = *I.getParent();
281 MachineFunction &MF = *MBB.getParent();
282 MachineRegisterInfo &MRI = MF.getRegInfo();
283
284 unsigned Opcode = I.getOpcode();
285 if (!isPreISelGenericOpcode(Opcode)) {
286 // Certain non-generic instructions also need some special handling.
287
288 if (I.isCopy())
Igor Breger1dcd5e82017-06-20 09:15:10 +0000289 return selectCopy(I, MRI);
Igor Bregerf7359d82017-02-22 12:25:09 +0000290
291 // TODO: handle more cases - LOAD_STACK_GUARD, PHI
292 return true;
293 }
294
Benjamin Kramer5a7e0f82017-02-22 12:59:47 +0000295 assert(I.getNumOperands() == I.getNumExplicitOperands() &&
296 "Generic instruction has unexpected implicit operands\n");
Igor Bregerf7359d82017-02-22 12:25:09 +0000297
Igor Breger2452ef02017-05-01 07:06:08 +0000298 if (selectImpl(I))
Igor Bregerfda31e62017-05-10 06:52:58 +0000299 return true;
Igor Breger2452ef02017-05-01 07:06:08 +0000300
301 DEBUG(dbgs() << " C++ instruction selection: "; I.print(dbgs()));
302
303 // TODO: This should be implemented by tblgen.
Igor Bregera8ba5722017-03-23 15:25:57 +0000304 if (selectLoadStoreOp(I, MRI, MF))
Igor Breger321cf3c2017-03-03 08:06:46 +0000305 return true;
Igor Breger810c6252017-05-08 09:40:43 +0000306 if (selectFrameIndexOrGep(I, MRI, MF))
Igor Breger531a2032017-03-26 08:11:12 +0000307 return true;
Igor Breger717bd362017-07-02 08:58:29 +0000308 if (selectGlobalValue(I, MRI, MF))
309 return true;
Igor Breger3b97ea32017-04-12 12:54:54 +0000310 if (selectConstant(I, MRI, MF))
311 return true;
Igor Breger4fdf1e42017-04-19 11:34:59 +0000312 if (selectTrunc(I, MRI, MF))
313 return true;
Igor Bregerfda31e62017-05-10 06:52:58 +0000314 if (selectZext(I, MRI, MF))
315 return true;
Igor Bregerc7b59772017-05-11 07:17:40 +0000316 if (selectCmp(I, MRI, MF))
317 return true;
Igor Breger28f290f2017-05-17 12:48:08 +0000318 if (selectUadde(I, MRI, MF))
319 return true;
Igor Bregerb186a692017-07-02 08:15:49 +0000320 if (selectUnmergeValues(I, MRI, MF))
321 return true;
Igor Breger0cddd342017-06-29 12:08:28 +0000322 if (selectMergeValues(I, MRI, MF))
323 return true;
Igor Bregerf5035d62017-06-25 11:42:17 +0000324 if (selectExtract(I, MRI, MF))
325 return true;
Igor Breger1c29be72017-06-22 09:43:35 +0000326 if (selectInsert(I, MRI, MF))
327 return true;
Igor Breger321cf3c2017-03-03 08:06:46 +0000328
Igor Breger2452ef02017-05-01 07:06:08 +0000329 return false;
Igor Bregerf7359d82017-02-22 12:25:09 +0000330}
Igor Breger321cf3c2017-03-03 08:06:46 +0000331
Igor Bregera8ba5722017-03-23 15:25:57 +0000332unsigned X86InstructionSelector::getLoadStoreOp(LLT &Ty, const RegisterBank &RB,
333 unsigned Opc,
334 uint64_t Alignment) const {
335 bool Isload = (Opc == TargetOpcode::G_LOAD);
336 bool HasAVX = STI.hasAVX();
337 bool HasAVX512 = STI.hasAVX512();
338 bool HasVLX = STI.hasVLX();
339
340 if (Ty == LLT::scalar(8)) {
341 if (X86::GPRRegBankID == RB.getID())
342 return Isload ? X86::MOV8rm : X86::MOV8mr;
343 } else if (Ty == LLT::scalar(16)) {
344 if (X86::GPRRegBankID == RB.getID())
345 return Isload ? X86::MOV16rm : X86::MOV16mr;
Igor Bregera9edb882017-05-01 06:08:32 +0000346 } else if (Ty == LLT::scalar(32) || Ty == LLT::pointer(0, 32)) {
Igor Bregera8ba5722017-03-23 15:25:57 +0000347 if (X86::GPRRegBankID == RB.getID())
348 return Isload ? X86::MOV32rm : X86::MOV32mr;
349 if (X86::VECRRegBankID == RB.getID())
350 return Isload ? (HasAVX512 ? X86::VMOVSSZrm
351 : HasAVX ? X86::VMOVSSrm : X86::MOVSSrm)
352 : (HasAVX512 ? X86::VMOVSSZmr
353 : HasAVX ? X86::VMOVSSmr : X86::MOVSSmr);
Igor Bregera9edb882017-05-01 06:08:32 +0000354 } else if (Ty == LLT::scalar(64) || Ty == LLT::pointer(0, 64)) {
Igor Bregera8ba5722017-03-23 15:25:57 +0000355 if (X86::GPRRegBankID == RB.getID())
356 return Isload ? X86::MOV64rm : X86::MOV64mr;
357 if (X86::VECRRegBankID == RB.getID())
358 return Isload ? (HasAVX512 ? X86::VMOVSDZrm
359 : HasAVX ? X86::VMOVSDrm : X86::MOVSDrm)
360 : (HasAVX512 ? X86::VMOVSDZmr
361 : HasAVX ? X86::VMOVSDmr : X86::MOVSDmr);
362 } else if (Ty.isVector() && Ty.getSizeInBits() == 128) {
363 if (Alignment >= 16)
364 return Isload ? (HasVLX ? X86::VMOVAPSZ128rm
365 : HasAVX512
366 ? X86::VMOVAPSZ128rm_NOVLX
367 : HasAVX ? X86::VMOVAPSrm : X86::MOVAPSrm)
368 : (HasVLX ? X86::VMOVAPSZ128mr
369 : HasAVX512
370 ? X86::VMOVAPSZ128mr_NOVLX
371 : HasAVX ? X86::VMOVAPSmr : X86::MOVAPSmr);
372 else
373 return Isload ? (HasVLX ? X86::VMOVUPSZ128rm
374 : HasAVX512
375 ? X86::VMOVUPSZ128rm_NOVLX
376 : HasAVX ? X86::VMOVUPSrm : X86::MOVUPSrm)
377 : (HasVLX ? X86::VMOVUPSZ128mr
378 : HasAVX512
379 ? X86::VMOVUPSZ128mr_NOVLX
380 : HasAVX ? X86::VMOVUPSmr : X86::MOVUPSmr);
Igor Breger617be6e2017-05-23 08:23:51 +0000381 } else if (Ty.isVector() && Ty.getSizeInBits() == 256) {
382 if (Alignment >= 32)
383 return Isload ? (HasVLX ? X86::VMOVAPSZ256rm
384 : HasAVX512 ? X86::VMOVAPSZ256rm_NOVLX
385 : X86::VMOVAPSYrm)
386 : (HasVLX ? X86::VMOVAPSZ256mr
387 : HasAVX512 ? X86::VMOVAPSZ256mr_NOVLX
388 : X86::VMOVAPSYmr);
389 else
390 return Isload ? (HasVLX ? X86::VMOVUPSZ256rm
391 : HasAVX512 ? X86::VMOVUPSZ256rm_NOVLX
392 : X86::VMOVUPSYrm)
393 : (HasVLX ? X86::VMOVUPSZ256mr
394 : HasAVX512 ? X86::VMOVUPSZ256mr_NOVLX
395 : X86::VMOVUPSYmr);
396 } else if (Ty.isVector() && Ty.getSizeInBits() == 512) {
397 if (Alignment >= 64)
398 return Isload ? X86::VMOVAPSZrm : X86::VMOVAPSZmr;
399 else
400 return Isload ? X86::VMOVUPSZrm : X86::VMOVUPSZmr;
Igor Bregera8ba5722017-03-23 15:25:57 +0000401 }
402 return Opc;
403}
404
Igor Bregerbd2deda2017-06-19 13:12:57 +0000405// Fill in an address from the given instruction.
406void X86SelectAddress(const MachineInstr &I, const MachineRegisterInfo &MRI,
407 X86AddressMode &AM) {
408
409 assert(I.getOperand(0).isReg() && "unsupported opperand.");
410 assert(MRI.getType(I.getOperand(0).getReg()).isPointer() &&
411 "unsupported type.");
412
413 if (I.getOpcode() == TargetOpcode::G_GEP) {
414 if (auto COff = getConstantVRegVal(I.getOperand(2).getReg(), MRI)) {
415 int64_t Imm = *COff;
416 if (isInt<32>(Imm)) { // Check for displacement overflow.
417 AM.Disp = static_cast<int32_t>(Imm);
418 AM.Base.Reg = I.getOperand(1).getReg();
419 return;
420 }
421 }
422 } else if (I.getOpcode() == TargetOpcode::G_FRAME_INDEX) {
423 AM.Base.FrameIndex = I.getOperand(1).getIndex();
424 AM.BaseType = X86AddressMode::FrameIndexBase;
425 return;
426 }
427
428 // Default behavior.
429 AM.Base.Reg = I.getOperand(0).getReg();
430 return;
431}
432
Igor Bregera8ba5722017-03-23 15:25:57 +0000433bool X86InstructionSelector::selectLoadStoreOp(MachineInstr &I,
434 MachineRegisterInfo &MRI,
435 MachineFunction &MF) const {
436
437 unsigned Opc = I.getOpcode();
438
439 if (Opc != TargetOpcode::G_STORE && Opc != TargetOpcode::G_LOAD)
440 return false;
441
442 const unsigned DefReg = I.getOperand(0).getReg();
443 LLT Ty = MRI.getType(DefReg);
444 const RegisterBank &RB = *RBI.getRegBank(DefReg, MRI, TRI);
445
446 auto &MemOp = **I.memoperands_begin();
Igor Bregerbd2deda2017-06-19 13:12:57 +0000447 if (MemOp.getOrdering() != AtomicOrdering::NotAtomic) {
448 DEBUG(dbgs() << "Atomic load/store not supported yet\n");
449 return false;
450 }
451
Igor Bregera8ba5722017-03-23 15:25:57 +0000452 unsigned NewOpc = getLoadStoreOp(Ty, RB, Opc, MemOp.getAlignment());
453 if (NewOpc == Opc)
454 return false;
455
Igor Bregerbd2deda2017-06-19 13:12:57 +0000456 X86AddressMode AM;
457 X86SelectAddress(*MRI.getVRegDef(I.getOperand(1).getReg()), MRI, AM);
458
Igor Bregera8ba5722017-03-23 15:25:57 +0000459 I.setDesc(TII.get(NewOpc));
460 MachineInstrBuilder MIB(MF, I);
Igor Bregerbd2deda2017-06-19 13:12:57 +0000461 if (Opc == TargetOpcode::G_LOAD) {
462 I.RemoveOperand(1);
463 addFullAddress(MIB, AM);
464 } else {
Igor Bregera8ba5722017-03-23 15:25:57 +0000465 // G_STORE (VAL, Addr), X86Store instruction (Addr, VAL)
Igor Bregerbd2deda2017-06-19 13:12:57 +0000466 I.RemoveOperand(1);
Igor Bregera8ba5722017-03-23 15:25:57 +0000467 I.RemoveOperand(0);
Igor Bregerbd2deda2017-06-19 13:12:57 +0000468 addFullAddress(MIB, AM).addUse(DefReg);
Igor Bregera8ba5722017-03-23 15:25:57 +0000469 }
470 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
471}
472
Igor Breger717bd362017-07-02 08:58:29 +0000473static unsigned getLeaOP(LLT Ty, const X86Subtarget &STI) {
474 if (Ty == LLT::pointer(0, 64))
475 return X86::LEA64r;
476 else if (Ty == LLT::pointer(0, 32))
477 return STI.isTarget64BitILP32() ? X86::LEA64_32r : X86::LEA32r;
478 else
479 llvm_unreachable("Can't get LEA opcode. Unsupported type.");
480}
481
Igor Breger810c6252017-05-08 09:40:43 +0000482bool X86InstructionSelector::selectFrameIndexOrGep(MachineInstr &I,
483 MachineRegisterInfo &MRI,
484 MachineFunction &MF) const {
485 unsigned Opc = I.getOpcode();
486
487 if (Opc != TargetOpcode::G_FRAME_INDEX && Opc != TargetOpcode::G_GEP)
Igor Breger531a2032017-03-26 08:11:12 +0000488 return false;
489
490 const unsigned DefReg = I.getOperand(0).getReg();
491 LLT Ty = MRI.getType(DefReg);
492
Igor Breger810c6252017-05-08 09:40:43 +0000493 // Use LEA to calculate frame index and GEP
Igor Breger717bd362017-07-02 08:58:29 +0000494 unsigned NewOpc = getLeaOP(Ty, STI);
Igor Breger531a2032017-03-26 08:11:12 +0000495 I.setDesc(TII.get(NewOpc));
496 MachineInstrBuilder MIB(MF, I);
Igor Breger810c6252017-05-08 09:40:43 +0000497
498 if (Opc == TargetOpcode::G_FRAME_INDEX) {
499 addOffset(MIB, 0);
500 } else {
501 MachineOperand &InxOp = I.getOperand(2);
502 I.addOperand(InxOp); // set IndexReg
503 InxOp.ChangeToImmediate(1); // set Scale
504 MIB.addImm(0).addReg(0);
505 }
Igor Breger531a2032017-03-26 08:11:12 +0000506
507 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
508}
Daniel Sanders0b5293f2017-04-06 09:49:34 +0000509
Igor Breger717bd362017-07-02 08:58:29 +0000510bool X86InstructionSelector::selectGlobalValue(MachineInstr &I,
511 MachineRegisterInfo &MRI,
512 MachineFunction &MF) const {
513 unsigned Opc = I.getOpcode();
514
515 if (Opc != TargetOpcode::G_GLOBAL_VALUE)
516 return false;
517
518 auto GV = I.getOperand(1).getGlobal();
519 if (GV->isThreadLocal()) {
520 return false; // TODO: we don't support TLS yet.
521 }
522
523 // Can't handle alternate code models yet.
524 if (TM.getCodeModel() != CodeModel::Small)
525 return 0;
526
527 X86AddressMode AM;
528 AM.GV = GV;
529 AM.GVOpFlags = STI.classifyGlobalReference(GV);
530
531 // TODO: The ABI requires an extra load. not supported yet.
532 if (isGlobalStubReference(AM.GVOpFlags))
533 return false;
534
535 // TODO: This reference is relative to the pic base. not supported yet.
536 if (isGlobalRelativeToPICBase(AM.GVOpFlags))
537 return false;
538
539 if (STI.isPICStyleRIPRel()) {
540 // Use rip-relative addressing.
541 assert(AM.Base.Reg == 0 && AM.IndexReg == 0);
542 AM.Base.Reg = X86::RIP;
543 }
544
545 const unsigned DefReg = I.getOperand(0).getReg();
546 LLT Ty = MRI.getType(DefReg);
547 unsigned NewOpc = getLeaOP(Ty, STI);
548
549 I.setDesc(TII.get(NewOpc));
550 MachineInstrBuilder MIB(MF, I);
551
552 I.RemoveOperand(1);
553 addFullAddress(MIB, AM);
554
555 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
556}
557
Igor Breger3b97ea32017-04-12 12:54:54 +0000558bool X86InstructionSelector::selectConstant(MachineInstr &I,
559 MachineRegisterInfo &MRI,
560 MachineFunction &MF) const {
561 if (I.getOpcode() != TargetOpcode::G_CONSTANT)
562 return false;
563
564 const unsigned DefReg = I.getOperand(0).getReg();
565 LLT Ty = MRI.getType(DefReg);
566
Igor Breger5c787ab2017-07-03 11:06:54 +0000567 if (RBI.getRegBank(DefReg, MRI, TRI)->getID() != X86::GPRRegBankID)
568 return false;
Igor Breger3b97ea32017-04-12 12:54:54 +0000569
570 uint64_t Val = 0;
571 if (I.getOperand(1).isCImm()) {
572 Val = I.getOperand(1).getCImm()->getZExtValue();
573 I.getOperand(1).ChangeToImmediate(Val);
574 } else if (I.getOperand(1).isImm()) {
575 Val = I.getOperand(1).getImm();
576 } else
577 llvm_unreachable("Unsupported operand type.");
578
579 unsigned NewOpc;
580 switch (Ty.getSizeInBits()) {
581 case 8:
582 NewOpc = X86::MOV8ri;
583 break;
584 case 16:
585 NewOpc = X86::MOV16ri;
586 break;
587 case 32:
588 NewOpc = X86::MOV32ri;
589 break;
590 case 64: {
591 // TODO: in case isUInt<32>(Val), X86::MOV32ri can be used
592 if (isInt<32>(Val))
593 NewOpc = X86::MOV64ri32;
594 else
595 NewOpc = X86::MOV64ri;
596 break;
597 }
598 default:
599 llvm_unreachable("Can't select G_CONSTANT, unsupported type.");
600 }
601
602 I.setDesc(TII.get(NewOpc));
603 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
604}
605
Igor Breger4fdf1e42017-04-19 11:34:59 +0000606bool X86InstructionSelector::selectTrunc(MachineInstr &I,
607 MachineRegisterInfo &MRI,
608 MachineFunction &MF) const {
609 if (I.getOpcode() != TargetOpcode::G_TRUNC)
610 return false;
611
612 const unsigned DstReg = I.getOperand(0).getReg();
613 const unsigned SrcReg = I.getOperand(1).getReg();
614
615 const LLT DstTy = MRI.getType(DstReg);
616 const LLT SrcTy = MRI.getType(SrcReg);
617
618 const RegisterBank &DstRB = *RBI.getRegBank(DstReg, MRI, TRI);
619 const RegisterBank &SrcRB = *RBI.getRegBank(SrcReg, MRI, TRI);
620
621 if (DstRB.getID() != SrcRB.getID()) {
622 DEBUG(dbgs() << "G_TRUNC input/output on different banks\n");
623 return false;
624 }
625
626 if (DstRB.getID() != X86::GPRRegBankID)
627 return false;
628
Igor Breger1dcd5e82017-06-20 09:15:10 +0000629 const TargetRegisterClass *DstRC = getRegClass(DstTy, DstRB);
Igor Breger4fdf1e42017-04-19 11:34:59 +0000630 if (!DstRC)
631 return false;
632
Igor Breger1dcd5e82017-06-20 09:15:10 +0000633 const TargetRegisterClass *SrcRC = getRegClass(SrcTy, SrcRB);
Igor Breger4fdf1e42017-04-19 11:34:59 +0000634 if (!SrcRC)
635 return false;
636
Igor Breger014fc562017-05-21 11:13:56 +0000637 unsigned SubIdx;
638 if (DstRC == SrcRC) {
639 // Nothing to be done
640 SubIdx = X86::NoSubRegister;
641 } else if (DstRC == &X86::GR32RegClass) {
642 SubIdx = X86::sub_32bit;
643 } else if (DstRC == &X86::GR16RegClass) {
644 SubIdx = X86::sub_16bit;
645 } else if (DstRC == &X86::GR8RegClass) {
646 SubIdx = X86::sub_8bit;
647 } else {
648 return false;
649 }
650
651 SrcRC = TRI.getSubClassWithSubReg(SrcRC, SubIdx);
652
Igor Breger4fdf1e42017-04-19 11:34:59 +0000653 if (!RBI.constrainGenericRegister(SrcReg, *SrcRC, MRI) ||
654 !RBI.constrainGenericRegister(DstReg, *DstRC, MRI)) {
655 DEBUG(dbgs() << "Failed to constrain G_TRUNC\n");
656 return false;
657 }
658
Igor Breger014fc562017-05-21 11:13:56 +0000659 I.getOperand(1).setSubReg(SubIdx);
Igor Breger4fdf1e42017-04-19 11:34:59 +0000660
661 I.setDesc(TII.get(X86::COPY));
662 return true;
663}
664
Igor Bregerfda31e62017-05-10 06:52:58 +0000665bool X86InstructionSelector::selectZext(MachineInstr &I,
666 MachineRegisterInfo &MRI,
667 MachineFunction &MF) const {
668 if (I.getOpcode() != TargetOpcode::G_ZEXT)
669 return false;
670
671 const unsigned DstReg = I.getOperand(0).getReg();
672 const unsigned SrcReg = I.getOperand(1).getReg();
673
674 const LLT DstTy = MRI.getType(DstReg);
675 const LLT SrcTy = MRI.getType(SrcReg);
676
Igor Bregerd48c5e42017-07-10 09:07:34 +0000677 if (SrcTy != LLT::scalar(1))
678 return false;
Igor Bregerfda31e62017-05-10 06:52:58 +0000679
Igor Bregerd48c5e42017-07-10 09:07:34 +0000680 unsigned AndOpc;
681 if (DstTy == LLT::scalar(8))
Igor Breger324d3792017-07-11 08:04:51 +0000682 AndOpc = X86::AND8ri;
Igor Bregerd48c5e42017-07-10 09:07:34 +0000683 else if (DstTy == LLT::scalar(16))
684 AndOpc = X86::AND16ri8;
685 else if (DstTy == LLT::scalar(32))
686 AndOpc = X86::AND32ri8;
687 else if (DstTy == LLT::scalar(64))
688 AndOpc = X86::AND64ri8;
689 else
690 return false;
Igor Bregerfda31e62017-05-10 06:52:58 +0000691
Igor Bregerd48c5e42017-07-10 09:07:34 +0000692 unsigned DefReg = SrcReg;
693 if (DstTy != LLT::scalar(8)) {
694 DefReg = MRI.createVirtualRegister(getRegClass(DstTy, DstReg, MRI));
Igor Bregerfda31e62017-05-10 06:52:58 +0000695 BuildMI(*I.getParent(), I, I.getDebugLoc(),
696 TII.get(TargetOpcode::SUBREG_TO_REG), DefReg)
697 .addImm(0)
698 .addReg(SrcReg)
699 .addImm(X86::sub_8bit);
Igor Bregerfda31e62017-05-10 06:52:58 +0000700 }
701
Igor Bregerd48c5e42017-07-10 09:07:34 +0000702 MachineInstr &AndInst =
703 *BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(AndOpc), DstReg)
704 .addReg(DefReg)
705 .addImm(1);
706
707 constrainSelectedInstRegOperands(AndInst, TII, TRI, RBI);
708
709 I.eraseFromParent();
710 return true;
Igor Bregerfda31e62017-05-10 06:52:58 +0000711}
712
Igor Bregerc7b59772017-05-11 07:17:40 +0000713bool X86InstructionSelector::selectCmp(MachineInstr &I,
714 MachineRegisterInfo &MRI,
715 MachineFunction &MF) const {
716 if (I.getOpcode() != TargetOpcode::G_ICMP)
717 return false;
718
719 X86::CondCode CC;
720 bool SwapArgs;
721 std::tie(CC, SwapArgs) = X86::getX86ConditionCode(
722 (CmpInst::Predicate)I.getOperand(1).getPredicate());
723 unsigned OpSet = X86::getSETFromCond(CC);
724
725 unsigned LHS = I.getOperand(2).getReg();
726 unsigned RHS = I.getOperand(3).getReg();
727
728 if (SwapArgs)
729 std::swap(LHS, RHS);
730
731 unsigned OpCmp;
732 LLT Ty = MRI.getType(LHS);
733
734 switch (Ty.getSizeInBits()) {
735 default:
736 return false;
737 case 8:
738 OpCmp = X86::CMP8rr;
739 break;
740 case 16:
741 OpCmp = X86::CMP16rr;
742 break;
743 case 32:
744 OpCmp = X86::CMP32rr;
745 break;
746 case 64:
747 OpCmp = X86::CMP64rr;
748 break;
749 }
750
751 MachineInstr &CmpInst =
752 *BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(OpCmp))
753 .addReg(LHS)
754 .addReg(RHS);
755
756 MachineInstr &SetInst = *BuildMI(*I.getParent(), I, I.getDebugLoc(),
757 TII.get(OpSet), I.getOperand(0).getReg());
758
759 constrainSelectedInstRegOperands(CmpInst, TII, TRI, RBI);
760 constrainSelectedInstRegOperands(SetInst, TII, TRI, RBI);
761
762 I.eraseFromParent();
763 return true;
764}
765
Igor Breger28f290f2017-05-17 12:48:08 +0000766bool X86InstructionSelector::selectUadde(MachineInstr &I,
767 MachineRegisterInfo &MRI,
768 MachineFunction &MF) const {
769 if (I.getOpcode() != TargetOpcode::G_UADDE)
770 return false;
771
772 const unsigned DstReg = I.getOperand(0).getReg();
773 const unsigned CarryOutReg = I.getOperand(1).getReg();
774 const unsigned Op0Reg = I.getOperand(2).getReg();
775 const unsigned Op1Reg = I.getOperand(3).getReg();
776 unsigned CarryInReg = I.getOperand(4).getReg();
777
778 const LLT DstTy = MRI.getType(DstReg);
779
780 if (DstTy != LLT::scalar(32))
781 return false;
782
783 // find CarryIn def instruction.
784 MachineInstr *Def = MRI.getVRegDef(CarryInReg);
785 while (Def->getOpcode() == TargetOpcode::G_TRUNC) {
786 CarryInReg = Def->getOperand(1).getReg();
787 Def = MRI.getVRegDef(CarryInReg);
788 }
789
790 unsigned Opcode;
791 if (Def->getOpcode() == TargetOpcode::G_UADDE) {
792 // carry set by prev ADD.
793
794 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(X86::COPY), X86::EFLAGS)
795 .addReg(CarryInReg);
796
797 if (!RBI.constrainGenericRegister(CarryInReg, X86::GR32RegClass, MRI))
798 return false;
799
800 Opcode = X86::ADC32rr;
801 } else if (auto val = getConstantVRegVal(CarryInReg, MRI)) {
802 // carry is constant, support only 0.
803 if (*val != 0)
804 return false;
805
806 Opcode = X86::ADD32rr;
807 } else
808 return false;
809
810 MachineInstr &AddInst =
811 *BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(Opcode), DstReg)
812 .addReg(Op0Reg)
813 .addReg(Op1Reg);
814
815 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(X86::COPY), CarryOutReg)
816 .addReg(X86::EFLAGS);
817
818 if (!constrainSelectedInstRegOperands(AddInst, TII, TRI, RBI) ||
819 !RBI.constrainGenericRegister(CarryOutReg, X86::GR32RegClass, MRI))
820 return false;
821
822 I.eraseFromParent();
823 return true;
824}
825
Igor Bregerf5035d62017-06-25 11:42:17 +0000826bool X86InstructionSelector::selectExtract(MachineInstr &I,
827 MachineRegisterInfo &MRI,
828 MachineFunction &MF) const {
829
830 if (I.getOpcode() != TargetOpcode::G_EXTRACT)
831 return false;
832
833 const unsigned DstReg = I.getOperand(0).getReg();
834 const unsigned SrcReg = I.getOperand(1).getReg();
835 int64_t Index = I.getOperand(2).getImm();
836
837 const LLT DstTy = MRI.getType(DstReg);
838 const LLT SrcTy = MRI.getType(SrcReg);
839
840 // Meanwile handle vector type only.
841 if (!DstTy.isVector())
842 return false;
843
844 if (Index % DstTy.getSizeInBits() != 0)
845 return false; // Not extract subvector.
846
847 if (Index == 0) {
848 // Replace by extract subreg copy.
849 if (!emitExtractSubreg(DstReg, SrcReg, I, MRI, MF))
850 return false;
851
852 I.eraseFromParent();
853 return true;
854 }
855
856 bool HasAVX = STI.hasAVX();
857 bool HasAVX512 = STI.hasAVX512();
858 bool HasVLX = STI.hasVLX();
859
860 if (SrcTy.getSizeInBits() == 256 && DstTy.getSizeInBits() == 128) {
861 if (HasVLX)
862 I.setDesc(TII.get(X86::VEXTRACTF32x4Z256rr));
863 else if (HasAVX)
864 I.setDesc(TII.get(X86::VEXTRACTF128rr));
865 else
866 return false;
867 } else if (SrcTy.getSizeInBits() == 512 && HasAVX512) {
868 if (DstTy.getSizeInBits() == 128)
869 I.setDesc(TII.get(X86::VEXTRACTF32x4Zrr));
870 else if (DstTy.getSizeInBits() == 256)
871 I.setDesc(TII.get(X86::VEXTRACTF64x4Zrr));
872 else
873 return false;
874 } else
875 return false;
876
877 // Convert to X86 VEXTRACT immediate.
878 Index = Index / DstTy.getSizeInBits();
879 I.getOperand(2).setImm(Index);
880
881 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
882}
883
884bool X86InstructionSelector::emitExtractSubreg(unsigned DstReg, unsigned SrcReg,
885 MachineInstr &I,
886 MachineRegisterInfo &MRI,
887 MachineFunction &MF) const {
888
889 const LLT DstTy = MRI.getType(DstReg);
890 const LLT SrcTy = MRI.getType(SrcReg);
891 unsigned SubIdx = X86::NoSubRegister;
892
893 if (!DstTy.isVector() || !SrcTy.isVector())
894 return false;
895
896 assert(SrcTy.getSizeInBits() > DstTy.getSizeInBits() &&
897 "Incorrect Src/Dst register size");
898
899 if (DstTy.getSizeInBits() == 128)
900 SubIdx = X86::sub_xmm;
901 else if (DstTy.getSizeInBits() == 256)
902 SubIdx = X86::sub_ymm;
903 else
904 return false;
905
906 const TargetRegisterClass *DstRC = getRegClass(DstTy, DstReg, MRI);
907 const TargetRegisterClass *SrcRC = getRegClass(SrcTy, SrcReg, MRI);
908
909 SrcRC = TRI.getSubClassWithSubReg(SrcRC, SubIdx);
910
911 if (!RBI.constrainGenericRegister(SrcReg, *SrcRC, MRI) ||
912 !RBI.constrainGenericRegister(DstReg, *DstRC, MRI)) {
913 DEBUG(dbgs() << "Failed to constrain G_TRUNC\n");
914 return false;
915 }
916
917 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(X86::COPY), DstReg)
918 .addReg(SrcReg, 0, SubIdx);
919
920 return true;
921}
922
Igor Breger1c29be72017-06-22 09:43:35 +0000923bool X86InstructionSelector::emitInsertSubreg(unsigned DstReg, unsigned SrcReg,
924 MachineInstr &I,
925 MachineRegisterInfo &MRI,
926 MachineFunction &MF) const {
927
928 const LLT DstTy = MRI.getType(DstReg);
929 const LLT SrcTy = MRI.getType(SrcReg);
930 unsigned SubIdx = X86::NoSubRegister;
931
932 // TODO: support scalar types
933 if (!DstTy.isVector() || !SrcTy.isVector())
934 return false;
935
936 assert(SrcTy.getSizeInBits() < DstTy.getSizeInBits() &&
937 "Incorrect Src/Dst register size");
938
939 if (SrcTy.getSizeInBits() == 128)
940 SubIdx = X86::sub_xmm;
941 else if (SrcTy.getSizeInBits() == 256)
942 SubIdx = X86::sub_ymm;
943 else
944 return false;
945
946 const TargetRegisterClass *SrcRC = getRegClass(SrcTy, SrcReg, MRI);
947 const TargetRegisterClass *DstRC = getRegClass(DstTy, DstReg, MRI);
948
949 if (!RBI.constrainGenericRegister(SrcReg, *SrcRC, MRI) ||
950 !RBI.constrainGenericRegister(DstReg, *DstRC, MRI)) {
951 DEBUG(dbgs() << "Failed to constrain INSERT_SUBREG\n");
952 return false;
953 }
954
955 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(X86::COPY))
956 .addReg(DstReg, RegState::DefineNoRead, SubIdx)
957 .addReg(SrcReg);
958
959 return true;
960}
961
962bool X86InstructionSelector::selectInsert(MachineInstr &I,
963 MachineRegisterInfo &MRI,
964 MachineFunction &MF) const {
965
966 if (I.getOpcode() != TargetOpcode::G_INSERT)
967 return false;
968
969 const unsigned DstReg = I.getOperand(0).getReg();
970 const unsigned SrcReg = I.getOperand(1).getReg();
971 const unsigned InsertReg = I.getOperand(2).getReg();
972 int64_t Index = I.getOperand(3).getImm();
973
974 const LLT DstTy = MRI.getType(DstReg);
975 const LLT InsertRegTy = MRI.getType(InsertReg);
976
977 // Meanwile handle vector type only.
978 if (!DstTy.isVector())
979 return false;
980
981 if (Index % InsertRegTy.getSizeInBits() != 0)
982 return false; // Not insert subvector.
983
984 if (Index == 0 && MRI.getVRegDef(SrcReg)->isImplicitDef()) {
985 // Replace by subreg copy.
986 if (!emitInsertSubreg(DstReg, InsertReg, I, MRI, MF))
987 return false;
988
989 I.eraseFromParent();
990 return true;
991 }
992
993 bool HasAVX = STI.hasAVX();
994 bool HasAVX512 = STI.hasAVX512();
995 bool HasVLX = STI.hasVLX();
996
997 if (DstTy.getSizeInBits() == 256 && InsertRegTy.getSizeInBits() == 128) {
998 if (HasVLX)
999 I.setDesc(TII.get(X86::VINSERTF32x4Z256rr));
1000 else if (HasAVX)
1001 I.setDesc(TII.get(X86::VINSERTF128rr));
1002 else
1003 return false;
1004 } else if (DstTy.getSizeInBits() == 512 && HasAVX512) {
1005 if (InsertRegTy.getSizeInBits() == 128)
1006 I.setDesc(TII.get(X86::VINSERTF32x4Zrr));
1007 else if (InsertRegTy.getSizeInBits() == 256)
1008 I.setDesc(TII.get(X86::VINSERTF64x4Zrr));
1009 else
1010 return false;
1011 } else
1012 return false;
1013
1014 // Convert to X86 VINSERT immediate.
1015 Index = Index / InsertRegTy.getSizeInBits();
1016
1017 I.getOperand(3).setImm(Index);
1018
1019 return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
1020}
1021
Igor Bregerb186a692017-07-02 08:15:49 +00001022bool X86InstructionSelector::selectUnmergeValues(MachineInstr &I,
1023 MachineRegisterInfo &MRI,
1024 MachineFunction &MF) const {
1025 if (I.getOpcode() != TargetOpcode::G_UNMERGE_VALUES)
1026 return false;
1027
1028 // Split to extracts.
1029 unsigned NumDefs = I.getNumOperands() - 1;
1030 unsigned SrcReg = I.getOperand(NumDefs).getReg();
1031 unsigned DefSize = MRI.getType(I.getOperand(0).getReg()).getSizeInBits();
1032
1033 for (unsigned Idx = 0; Idx < NumDefs; ++Idx) {
1034
1035 MachineInstr &ExtrInst =
1036 *BuildMI(*I.getParent(), I, I.getDebugLoc(),
1037 TII.get(TargetOpcode::G_EXTRACT), I.getOperand(Idx).getReg())
1038 .addReg(SrcReg)
1039 .addImm(Idx * DefSize);
1040
1041 if (!select(ExtrInst))
1042 return false;
1043 }
1044
1045 I.eraseFromParent();
1046 return true;
1047}
1048
Igor Breger0cddd342017-06-29 12:08:28 +00001049bool X86InstructionSelector::selectMergeValues(MachineInstr &I,
1050 MachineRegisterInfo &MRI,
1051 MachineFunction &MF) const {
1052 if (I.getOpcode() != TargetOpcode::G_MERGE_VALUES)
1053 return false;
1054
1055 // Split to inserts.
1056 unsigned DstReg = I.getOperand(0).getReg();
1057 unsigned SrcReg0 = I.getOperand(1).getReg();
1058
1059 const LLT DstTy = MRI.getType(DstReg);
1060 const LLT SrcTy = MRI.getType(SrcReg0);
1061 unsigned SrcSize = SrcTy.getSizeInBits();
1062
1063 const RegisterBank &RegBank = *RBI.getRegBank(DstReg, MRI, TRI);
1064
1065 // For the first src use insertSubReg.
1066 unsigned DefReg = MRI.createGenericVirtualRegister(DstTy);
1067 MRI.setRegBank(DefReg, RegBank);
1068 if (!emitInsertSubreg(DefReg, I.getOperand(1).getReg(), I, MRI, MF))
1069 return false;
1070
1071 for (unsigned Idx = 2; Idx < I.getNumOperands(); ++Idx) {
1072
1073 unsigned Tmp = MRI.createGenericVirtualRegister(DstTy);
1074 MRI.setRegBank(Tmp, RegBank);
1075
1076 MachineInstr &InsertInst = *BuildMI(*I.getParent(), I, I.getDebugLoc(),
1077 TII.get(TargetOpcode::G_INSERT), Tmp)
1078 .addReg(DefReg)
1079 .addReg(I.getOperand(Idx).getReg())
1080 .addImm((Idx - 1) * SrcSize);
1081
1082 DefReg = Tmp;
1083
1084 if (!select(InsertInst))
1085 return false;
1086 }
1087
1088 MachineInstr &CopyInst = *BuildMI(*I.getParent(), I, I.getDebugLoc(),
1089 TII.get(TargetOpcode::COPY), DstReg)
1090 .addReg(DefReg);
1091
1092 if (!select(CopyInst))
1093 return false;
1094
1095 I.eraseFromParent();
1096 return true;
1097}
Daniel Sanders0b5293f2017-04-06 09:49:34 +00001098InstructionSelector *
Daniel Sanderse7b0d662017-04-21 15:59:56 +00001099llvm::createX86InstructionSelector(const X86TargetMachine &TM,
1100 X86Subtarget &Subtarget,
Daniel Sanders0b5293f2017-04-06 09:49:34 +00001101 X86RegisterBankInfo &RBI) {
Daniel Sanderse7b0d662017-04-21 15:59:56 +00001102 return new X86InstructionSelector(TM, Subtarget, RBI);
Daniel Sanders0b5293f2017-04-06 09:49:34 +00001103}