blob: 2d705c9d18af3ee9c20fc5c650e54bc23ea3cbbd [file] [log] [blame]
Clement Courbet44b4c542018-06-19 11:28:59 +00001//===-- Target.cpp ----------------------------------------------*- C++ -*-===//
2//
Chandler Carruth2946cd72019-01-19 08:50:56 +00003// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
Clement Courbet44b4c542018-06-19 11:28:59 +00006//
7//===----------------------------------------------------------------------===//
8#include "../Target.h"
9
Clement Courbet4860b982018-06-26 08:49:30 +000010#include "../Latency.h"
11#include "../Uops.h"
Clement Courbet717c9762018-06-28 07:41:16 +000012#include "MCTargetDesc/X86BaseInfo.h"
Clement Courbeta51efc22018-06-25 13:12:02 +000013#include "MCTargetDesc/X86MCTargetDesc.h"
Clement Courbet6fd00e32018-06-20 11:54:35 +000014#include "X86.h"
Clement Courbeta51efc22018-06-25 13:12:02 +000015#include "X86RegisterInfo.h"
Clement Courbete7851692018-07-03 06:17:05 +000016#include "X86Subtarget.h"
Clement Courbeta51efc22018-06-25 13:12:02 +000017#include "llvm/MC/MCInstBuilder.h"
Clement Courbet6fd00e32018-06-20 11:54:35 +000018
Fangrui Song32401af2018-10-22 17:10:47 +000019namespace llvm {
Clement Courbet44b4c542018-06-19 11:28:59 +000020namespace exegesis {
21
Clement Courbetc5448382018-11-07 16:14:55 +000022// Returns an error if we cannot handle the memory references in this
23// instruction.
Jinsong Ji56c74cf2018-11-20 14:41:59 +000024static Error isInvalidMemoryInstr(const Instruction &Instr) {
Clement Courbetc5448382018-11-07 16:14:55 +000025 switch (Instr.Description->TSFlags & X86II::FormMask) {
26 default:
27 llvm_unreachable("Unknown FormMask value");
28 // These have no memory access.
29 case X86II::Pseudo:
30 case X86II::RawFrm:
31 case X86II::MRMDestReg:
32 case X86II::MRMSrcReg:
33 case X86II::MRMSrcReg4VOp3:
34 case X86II::MRMSrcRegOp4:
35 case X86II::MRMXr:
36 case X86II::MRM0r:
37 case X86II::MRM1r:
38 case X86II::MRM2r:
39 case X86II::MRM3r:
40 case X86II::MRM4r:
41 case X86II::MRM5r:
42 case X86II::MRM6r:
43 case X86II::MRM7r:
44 case X86II::MRM_C0:
45 case X86II::MRM_C1:
46 case X86II::MRM_C2:
47 case X86II::MRM_C3:
48 case X86II::MRM_C4:
49 case X86II::MRM_C5:
50 case X86II::MRM_C6:
51 case X86II::MRM_C7:
52 case X86II::MRM_C8:
53 case X86II::MRM_C9:
54 case X86II::MRM_CA:
55 case X86II::MRM_CB:
56 case X86II::MRM_CC:
57 case X86II::MRM_CD:
58 case X86II::MRM_CE:
59 case X86II::MRM_CF:
60 case X86II::MRM_D0:
61 case X86II::MRM_D1:
62 case X86II::MRM_D2:
63 case X86II::MRM_D3:
64 case X86II::MRM_D4:
65 case X86II::MRM_D5:
66 case X86II::MRM_D6:
67 case X86II::MRM_D7:
68 case X86II::MRM_D8:
69 case X86II::MRM_D9:
70 case X86II::MRM_DA:
71 case X86II::MRM_DB:
72 case X86II::MRM_DC:
73 case X86II::MRM_DD:
74 case X86II::MRM_DE:
75 case X86II::MRM_DF:
76 case X86II::MRM_E0:
77 case X86II::MRM_E1:
78 case X86II::MRM_E2:
79 case X86II::MRM_E3:
80 case X86II::MRM_E4:
81 case X86II::MRM_E5:
82 case X86II::MRM_E6:
83 case X86II::MRM_E7:
84 case X86II::MRM_E8:
85 case X86II::MRM_E9:
86 case X86II::MRM_EA:
87 case X86II::MRM_EB:
88 case X86II::MRM_EC:
89 case X86II::MRM_ED:
90 case X86II::MRM_EE:
91 case X86II::MRM_EF:
92 case X86II::MRM_F0:
93 case X86II::MRM_F1:
94 case X86II::MRM_F2:
95 case X86II::MRM_F3:
96 case X86II::MRM_F4:
97 case X86II::MRM_F5:
98 case X86II::MRM_F6:
99 case X86II::MRM_F7:
100 case X86II::MRM_F8:
101 case X86II::MRM_F9:
102 case X86II::MRM_FA:
103 case X86II::MRM_FB:
104 case X86II::MRM_FC:
105 case X86II::MRM_FD:
106 case X86II::MRM_FE:
107 case X86II::MRM_FF:
108 case X86II::RawFrmImm8:
109 return Error::success();
110 case X86II::AddRegFrm:
111 return (Instr.Description->Opcode == X86::POP16r || Instr.Description->Opcode == X86::POP32r ||
112 Instr.Description->Opcode == X86::PUSH16r || Instr.Description->Opcode == X86::PUSH32r)
113 ? make_error<BenchmarkFailure>(
114 "unsupported opcode: unsupported memory access")
115 : Error::success();
116 // These access memory and are handled.
117 case X86II::MRMDestMem:
118 case X86II::MRMSrcMem:
119 case X86II::MRMSrcMem4VOp3:
120 case X86II::MRMSrcMemOp4:
121 case X86II::MRMXm:
122 case X86II::MRM0m:
123 case X86II::MRM1m:
124 case X86II::MRM2m:
125 case X86II::MRM3m:
126 case X86II::MRM4m:
127 case X86II::MRM5m:
128 case X86II::MRM6m:
129 case X86II::MRM7m:
130 return Error::success();
131 // These access memory and are not handled yet.
132 case X86II::RawFrmImm16:
133 case X86II::RawFrmMemOffs:
134 case X86II::RawFrmSrc:
135 case X86II::RawFrmDst:
136 case X86II::RawFrmDstSrc:
137 return make_error<BenchmarkFailure>(
138 "unsupported opcode: non uniform memory access");
Guillaume Chatelet3c639f32018-10-22 14:46:08 +0000139 }
Guillaume Chatelet3c639f32018-10-22 14:46:08 +0000140}
141
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000142static llvm::Error IsInvalidOpcode(const Instruction &Instr) {
143 const auto OpcodeName = Instr.Name;
Clement Courbet003e08f2018-11-06 14:11:58 +0000144 if ((Instr.Description->TSFlags & X86II::FormMask) == X86II::Pseudo)
145 return llvm::make_error<BenchmarkFailure>(
146 "unsupported opcode: pseudo instruction");
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000147 if (OpcodeName.startswith("POPF") || OpcodeName.startswith("PUSHF") ||
148 OpcodeName.startswith("ADJCALLSTACK"))
149 return llvm::make_error<BenchmarkFailure>(
Clement Courbet8d0dd0b2018-10-19 12:24:49 +0000150 "unsupported opcode: Push/Pop/AdjCallStack");
Clement Courbetc5448382018-11-07 16:14:55 +0000151 if (llvm::Error Error = isInvalidMemoryInstr(Instr))
Clement Courbet5b0d7832018-11-07 16:52:50 +0000152 return Error;
Guillaume Chatelet3c639f32018-10-22 14:46:08 +0000153 // We do not handle instructions with OPERAND_PCREL.
154 for (const Operand &Op : Instr.Operands)
155 if (Op.isExplicit() &&
156 Op.getExplicitOperandInfo().OperandType == llvm::MCOI::OPERAND_PCREL)
157 return llvm::make_error<BenchmarkFailure>(
158 "unsupported opcode: PC relative operand");
Clement Courbet8d0dd0b2018-10-19 12:24:49 +0000159 // We do not handle second-form X87 instructions. We only handle first-form
160 // ones (_Fp), see comment in X86InstrFPStack.td.
161 for (const Operand &Op : Instr.Operands)
162 if (Op.isReg() && Op.isExplicit() &&
163 Op.getExplicitOperandInfo().RegClass == llvm::X86::RSTRegClassID)
164 return llvm::make_error<BenchmarkFailure>(
165 "unsupported second-form X87 instruction");
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000166 return llvm::Error::success();
167}
168
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000169static unsigned getX86FPFlags(const Instruction &Instr) {
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000170 return Instr.Description->TSFlags & llvm::X86II::FPTypeMask;
171}
172
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000173namespace {
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000174class X86LatencySnippetGenerator : public LatencySnippetGenerator {
175public:
176 using LatencySnippetGenerator::LatencySnippetGenerator;
Clement Courbet4860b982018-06-26 08:49:30 +0000177
Guillaume Chatelet296a8622018-10-15 09:09:19 +0000178 llvm::Expected<std::vector<CodeTemplate>>
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000179 generateCodeTemplates(const Instruction &Instr) const override;
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000180};
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000181} // namespace
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000182
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000183llvm::Expected<std::vector<CodeTemplate>>
184X86LatencySnippetGenerator::generateCodeTemplates(
185 const Instruction &Instr) const {
186 if (auto E = IsInvalidOpcode(Instr))
187 return std::move(E);
188
189 switch (getX86FPFlags(Instr)) {
190 case llvm::X86II::NotFP:
191 return LatencySnippetGenerator::generateCodeTemplates(Instr);
192 case llvm::X86II::ZeroArgFP:
193 case llvm::X86II::OneArgFP:
194 case llvm::X86II::SpecialFP:
195 case llvm::X86II::CompareFP:
196 case llvm::X86II::CondMovFP:
197 return llvm::make_error<BenchmarkFailure>("Unsupported x87 Instruction");
198 case llvm::X86II::OneArgFPRW:
199 case llvm::X86II::TwoArgFP:
200 // These are instructions like
201 // - `ST(0) = fsqrt(ST(0))` (OneArgFPRW)
202 // - `ST(0) = ST(0) + ST(i)` (TwoArgFP)
203 // They are intrinsically serial and do not modify the state of the stack.
204 return generateSelfAliasingCodeTemplates(Instr);
205 default:
206 llvm_unreachable("Unknown FP Type!");
207 }
208}
209
210namespace {
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000211class X86UopsSnippetGenerator : public UopsSnippetGenerator {
212public:
213 using UopsSnippetGenerator::UopsSnippetGenerator;
214
Guillaume Chatelet296a8622018-10-15 09:09:19 +0000215 llvm::Expected<std::vector<CodeTemplate>>
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000216 generateCodeTemplates(const Instruction &Instr) const override;
Clement Courbet4860b982018-06-26 08:49:30 +0000217};
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000218} // namespace
Clement Courbet4860b982018-06-26 08:49:30 +0000219
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000220llvm::Expected<std::vector<CodeTemplate>>
221X86UopsSnippetGenerator::generateCodeTemplates(
222 const Instruction &Instr) const {
223 if (auto E = IsInvalidOpcode(Instr))
224 return std::move(E);
225
226 switch (getX86FPFlags(Instr)) {
227 case llvm::X86II::NotFP:
228 return UopsSnippetGenerator::generateCodeTemplates(Instr);
229 case llvm::X86II::ZeroArgFP:
230 case llvm::X86II::OneArgFP:
231 case llvm::X86II::SpecialFP:
232 return llvm::make_error<BenchmarkFailure>("Unsupported x87 Instruction");
233 case llvm::X86II::OneArgFPRW:
234 case llvm::X86II::TwoArgFP:
235 // These are instructions like
236 // - `ST(0) = fsqrt(ST(0))` (OneArgFPRW)
237 // - `ST(0) = ST(0) + ST(i)` (TwoArgFP)
238 // They are intrinsically serial and do not modify the state of the stack.
239 // We generate the same code for latency and uops.
240 return generateSelfAliasingCodeTemplates(Instr);
241 case llvm::X86II::CompareFP:
242 case llvm::X86II::CondMovFP:
243 // We can compute uops for any FP instruction that does not grow or shrink
244 // the stack (either do not touch the stack or push as much as they pop).
245 return generateUnconstrainedCodeTemplates(
246 Instr, "instruction does not grow/shrink the FP stack");
247 default:
248 llvm_unreachable("Unknown FP Type!");
249 }
250}
251
252static unsigned getLoadImmediateOpcode(unsigned RegBitWidth) {
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000253 switch (RegBitWidth) {
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000254 case 8:
255 return llvm::X86::MOV8ri;
256 case 16:
257 return llvm::X86::MOV16ri;
258 case 32:
259 return llvm::X86::MOV32ri;
260 case 64:
261 return llvm::X86::MOV64ri;
262 }
263 llvm_unreachable("Invalid Value Width");
264}
265
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000266// Generates instruction to load an immediate value into a register.
267static llvm::MCInst loadImmediate(unsigned Reg, unsigned RegBitWidth,
268 const llvm::APInt &Value) {
269 if (Value.getBitWidth() > RegBitWidth)
270 llvm_unreachable("Value must fit in the Register");
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000271 return llvm::MCInstBuilder(getLoadImmediateOpcode(RegBitWidth))
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000272 .addReg(Reg)
273 .addImm(Value.getZExtValue());
274}
275
276// Allocates scratch memory on the stack.
277static llvm::MCInst allocateStackSpace(unsigned Bytes) {
278 return llvm::MCInstBuilder(llvm::X86::SUB64ri8)
279 .addReg(llvm::X86::RSP)
280 .addReg(llvm::X86::RSP)
281 .addImm(Bytes);
282}
283
284// Fills scratch memory at offset `OffsetBytes` with value `Imm`.
285static llvm::MCInst fillStackSpace(unsigned MovOpcode, unsigned OffsetBytes,
286 uint64_t Imm) {
287 return llvm::MCInstBuilder(MovOpcode)
288 // Address = ESP
289 .addReg(llvm::X86::RSP) // BaseReg
290 .addImm(1) // ScaleAmt
291 .addReg(0) // IndexReg
292 .addImm(OffsetBytes) // Disp
293 .addReg(0) // Segment
294 // Immediate.
295 .addImm(Imm);
296}
297
298// Loads scratch memory into register `Reg` using opcode `RMOpcode`.
299static llvm::MCInst loadToReg(unsigned Reg, unsigned RMOpcode) {
300 return llvm::MCInstBuilder(RMOpcode)
301 .addReg(Reg)
302 // Address = ESP
303 .addReg(llvm::X86::RSP) // BaseReg
304 .addImm(1) // ScaleAmt
305 .addReg(0) // IndexReg
306 .addImm(0) // Disp
307 .addReg(0); // Segment
308}
309
310// Releases scratch memory.
311static llvm::MCInst releaseStackSpace(unsigned Bytes) {
312 return llvm::MCInstBuilder(llvm::X86::ADD64ri8)
313 .addReg(llvm::X86::RSP)
314 .addReg(llvm::X86::RSP)
315 .addImm(Bytes);
316}
317
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000318// Reserves some space on the stack, fills it with the content of the provided
319// constant and provide methods to load the stack value into a register.
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000320namespace {
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000321struct ConstantInliner {
Clement Courbet78b2e732018-09-25 07:31:44 +0000322 explicit ConstantInliner(const llvm::APInt &Constant) : Constant_(Constant) {}
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000323
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000324 std::vector<llvm::MCInst> loadAndFinalize(unsigned Reg, unsigned RegBitWidth,
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000325 unsigned Opcode);
Guillaume Chatelet8721ad92018-09-18 11:26:35 +0000326
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000327 std::vector<llvm::MCInst> loadX87STAndFinalize(unsigned Reg);
Clement Courbetc51f4522018-10-19 09:56:54 +0000328
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000329 std::vector<llvm::MCInst> loadX87FPAndFinalize(unsigned Reg);
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000330
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000331 std::vector<llvm::MCInst> popFlagAndFinalize();
Simon Pilgrimf652ef32018-09-18 15:38:16 +0000332
333private:
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000334 ConstantInliner &add(const llvm::MCInst &Inst) {
335 Instructions.push_back(Inst);
336 return *this;
337 }
338
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000339 void initStack(unsigned Bytes);
340
341 static constexpr const unsigned kF80Bytes = 10; // 80 bits.
Clement Courbet78b2e732018-09-25 07:31:44 +0000342
343 llvm::APInt Constant_;
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000344 std::vector<llvm::MCInst> Instructions;
345};
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000346} // namespace
347
348std::vector<llvm::MCInst> ConstantInliner::loadAndFinalize(unsigned Reg,
349 unsigned RegBitWidth,
350 unsigned Opcode) {
351 assert((RegBitWidth & 7) == 0 && "RegBitWidth must be a multiple of 8 bits");
352 initStack(RegBitWidth / 8);
353 add(loadToReg(Reg, Opcode));
354 add(releaseStackSpace(RegBitWidth / 8));
355 return std::move(Instructions);
356}
357
358std::vector<llvm::MCInst> ConstantInliner::loadX87STAndFinalize(unsigned Reg) {
359 initStack(kF80Bytes);
360 add(llvm::MCInstBuilder(llvm::X86::LD_F80m)
361 // Address = ESP
362 .addReg(llvm::X86::RSP) // BaseReg
363 .addImm(1) // ScaleAmt
364 .addReg(0) // IndexReg
365 .addImm(0) // Disp
366 .addReg(0)); // Segment
367 if (Reg != llvm::X86::ST0)
368 add(llvm::MCInstBuilder(llvm::X86::ST_Frr).addReg(Reg));
369 add(releaseStackSpace(kF80Bytes));
370 return std::move(Instructions);
371}
372
373std::vector<llvm::MCInst> ConstantInliner::loadX87FPAndFinalize(unsigned Reg) {
374 initStack(kF80Bytes);
375 add(llvm::MCInstBuilder(llvm::X86::LD_Fp80m)
376 .addReg(Reg)
377 // Address = ESP
378 .addReg(llvm::X86::RSP) // BaseReg
379 .addImm(1) // ScaleAmt
380 .addReg(0) // IndexReg
381 .addImm(0) // Disp
382 .addReg(0)); // Segment
383 add(releaseStackSpace(kF80Bytes));
384 return std::move(Instructions);
385}
386
387std::vector<llvm::MCInst> ConstantInliner::popFlagAndFinalize() {
388 initStack(8);
389 add(llvm::MCInstBuilder(llvm::X86::POPF64));
390 return std::move(Instructions);
391}
392
393void ConstantInliner::initStack(unsigned Bytes) {
394 assert(Constant_.getBitWidth() <= Bytes * 8 &&
395 "Value does not have the correct size");
396 const llvm::APInt WideConstant = Constant_.getBitWidth() < Bytes * 8
397 ? Constant_.sext(Bytes * 8)
398 : Constant_;
399 add(allocateStackSpace(Bytes));
400 size_t ByteOffset = 0;
401 for (; Bytes - ByteOffset >= 4; ByteOffset += 4)
402 add(fillStackSpace(
403 llvm::X86::MOV32mi, ByteOffset,
404 WideConstant.extractBits(32, ByteOffset * 8).getZExtValue()));
405 if (Bytes - ByteOffset >= 2) {
406 add(fillStackSpace(
407 llvm::X86::MOV16mi, ByteOffset,
408 WideConstant.extractBits(16, ByteOffset * 8).getZExtValue()));
409 ByteOffset += 2;
410 }
411 if (Bytes - ByteOffset >= 1)
412 add(fillStackSpace(
413 llvm::X86::MOV8mi, ByteOffset,
414 WideConstant.extractBits(8, ByteOffset * 8).getZExtValue()));
415}
Guillaume Chatelet5ad29092018-09-18 11:26:27 +0000416
Clement Courbet41c8af32018-10-25 07:44:01 +0000417#include "X86GenExegesis.inc"
418
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000419namespace {
Clement Courbet44b4c542018-06-19 11:28:59 +0000420class ExegesisX86Target : public ExegesisTarget {
Clement Courbet41c8af32018-10-25 07:44:01 +0000421public:
422 ExegesisX86Target() : ExegesisTarget(X86CpuPfmCounters) {}
423
424private:
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000425 void addTargetSpecificPasses(llvm::PassManagerBase &PM) const override;
Clement Courbet6fd00e32018-06-20 11:54:35 +0000426
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000427 unsigned getScratchMemoryRegister(const llvm::Triple &TT) const override;
Guillaume Chateletfb943542018-08-01 14:41:45 +0000428
429 unsigned getMaxMemoryAccessSize() const override { return 64; }
430
Guillaume Chatelet70ac0192018-09-27 09:23:04 +0000431 void fillMemoryOperands(InstructionTemplate &IT, unsigned Reg,
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000432 unsigned Offset) const override;
Guillaume Chateletfb943542018-08-01 14:41:45 +0000433
Guillaume Chateletc96a97b2018-09-20 12:22:18 +0000434 std::vector<llvm::MCInst> setRegTo(const llvm::MCSubtargetInfo &STI,
435 unsigned Reg,
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000436 const llvm::APInt &Value) const override;
Clement Courbeta51efc22018-06-25 13:12:02 +0000437
Clement Courbetd939f6d2018-09-13 07:40:53 +0000438 std::unique_ptr<SnippetGenerator>
439 createLatencySnippetGenerator(const LLVMState &State) const override {
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000440 return llvm::make_unique<X86LatencySnippetGenerator>(State);
Clement Courbet4860b982018-06-26 08:49:30 +0000441 }
442
Clement Courbetd939f6d2018-09-13 07:40:53 +0000443 std::unique_ptr<SnippetGenerator>
444 createUopsSnippetGenerator(const LLVMState &State) const override {
Guillaume Chatelet946fb052018-10-12 15:12:22 +0000445 return llvm::make_unique<X86UopsSnippetGenerator>(State);
Clement Courbet4860b982018-06-26 08:49:30 +0000446 }
447
Clement Courbet44b4c542018-06-19 11:28:59 +0000448 bool matchesArch(llvm::Triple::ArchType Arch) const override {
449 return Arch == llvm::Triple::x86_64 || Arch == llvm::Triple::x86;
450 }
451};
Clement Courbet44b4c542018-06-19 11:28:59 +0000452} // namespace
453
Jinsong Ji56c74cf2018-11-20 14:41:59 +0000454void ExegesisX86Target::addTargetSpecificPasses(
455 llvm::PassManagerBase &PM) const {
456 // Lowers FP pseudo-instructions, e.g. ABS_Fp32 -> ABS_F.
457 PM.add(llvm::createX86FloatingPointStackifierPass());
458}
459
460unsigned
461ExegesisX86Target::getScratchMemoryRegister(const llvm::Triple &TT) const {
462 if (!TT.isArch64Bit()) {
463 // FIXME: This would require popping from the stack, so we would have to
464 // add some additional setup code.
465 return 0;
466 }
467 return TT.isOSWindows() ? llvm::X86::RCX : llvm::X86::RDI;
468}
469
470void ExegesisX86Target::fillMemoryOperands(InstructionTemplate &IT,
471 unsigned Reg,
472 unsigned Offset) const {
473 assert(!isInvalidMemoryInstr(IT.Instr) &&
474 "fillMemoryOperands requires a valid memory instruction");
475 int MemOpIdx = X86II::getMemoryOperandNo(IT.Instr.Description->TSFlags);
476 assert(MemOpIdx >= 0 && "invalid memory operand index");
477 // getMemoryOperandNo() ignores tied operands, so we have to add them back.
478 for (unsigned I = 0; I <= static_cast<unsigned>(MemOpIdx); ++I) {
479 const auto &Op = IT.Instr.Operands[I];
480 if (Op.isTied() && Op.getTiedToIndex() < I) {
481 ++MemOpIdx;
482 }
483 }
484 // Now fill in the memory operands.
485 const auto SetOp = [&IT](int OpIdx, const MCOperand &OpVal) {
486 const auto Op = IT.Instr.Operands[OpIdx];
487 assert(Op.isMemory() && Op.isExplicit() && "invalid memory pattern");
488 IT.getValueFor(Op) = OpVal;
489 };
490 SetOp(MemOpIdx + 0, MCOperand::createReg(Reg)); // BaseReg
491 SetOp(MemOpIdx + 1, MCOperand::createImm(1)); // ScaleAmt
492 SetOp(MemOpIdx + 2, MCOperand::createReg(0)); // IndexReg
493 SetOp(MemOpIdx + 3, MCOperand::createImm(Offset)); // Disp
494 SetOp(MemOpIdx + 4, MCOperand::createReg(0)); // Segment
495}
496
497std::vector<llvm::MCInst>
498ExegesisX86Target::setRegTo(const llvm::MCSubtargetInfo &STI, unsigned Reg,
499 const llvm::APInt &Value) const {
500 if (llvm::X86::GR8RegClass.contains(Reg))
501 return {loadImmediate(Reg, 8, Value)};
502 if (llvm::X86::GR16RegClass.contains(Reg))
503 return {loadImmediate(Reg, 16, Value)};
504 if (llvm::X86::GR32RegClass.contains(Reg))
505 return {loadImmediate(Reg, 32, Value)};
506 if (llvm::X86::GR64RegClass.contains(Reg))
507 return {loadImmediate(Reg, 64, Value)};
508 ConstantInliner CI(Value);
509 if (llvm::X86::VR64RegClass.contains(Reg))
510 return CI.loadAndFinalize(Reg, 64, llvm::X86::MMX_MOVQ64rm);
511 if (llvm::X86::VR128XRegClass.contains(Reg)) {
512 if (STI.getFeatureBits()[llvm::X86::FeatureAVX512])
513 return CI.loadAndFinalize(Reg, 128, llvm::X86::VMOVDQU32Z128rm);
514 if (STI.getFeatureBits()[llvm::X86::FeatureAVX])
515 return CI.loadAndFinalize(Reg, 128, llvm::X86::VMOVDQUrm);
516 return CI.loadAndFinalize(Reg, 128, llvm::X86::MOVDQUrm);
517 }
518 if (llvm::X86::VR256XRegClass.contains(Reg)) {
519 if (STI.getFeatureBits()[llvm::X86::FeatureAVX512])
520 return CI.loadAndFinalize(Reg, 256, llvm::X86::VMOVDQU32Z256rm);
521 if (STI.getFeatureBits()[llvm::X86::FeatureAVX])
522 return CI.loadAndFinalize(Reg, 256, llvm::X86::VMOVDQUYrm);
523 }
524 if (llvm::X86::VR512RegClass.contains(Reg))
525 if (STI.getFeatureBits()[llvm::X86::FeatureAVX512])
526 return CI.loadAndFinalize(Reg, 512, llvm::X86::VMOVDQU32Zrm);
527 if (llvm::X86::RSTRegClass.contains(Reg)) {
528 return CI.loadX87STAndFinalize(Reg);
529 }
530 if (llvm::X86::RFP32RegClass.contains(Reg) ||
531 llvm::X86::RFP64RegClass.contains(Reg) ||
532 llvm::X86::RFP80RegClass.contains(Reg)) {
533 return CI.loadX87FPAndFinalize(Reg);
534 }
535 if (Reg == llvm::X86::EFLAGS)
536 return CI.popFlagAndFinalize();
537 return {}; // Not yet implemented.
538}
539
Clement Courbetcff2caa2018-06-25 11:22:23 +0000540static ExegesisTarget *getTheExegesisX86Target() {
Clement Courbet44b4c542018-06-19 11:28:59 +0000541 static ExegesisX86Target Target;
542 return &Target;
543}
544
545void InitializeX86ExegesisTarget() {
546 ExegesisTarget::registerTarget(getTheExegesisX86Target());
547}
548
Clement Courbetcff2caa2018-06-25 11:22:23 +0000549} // namespace exegesis
Fangrui Song32401af2018-10-22 17:10:47 +0000550} // namespace llvm