blob: 966c6fec20c63fa71fc972cc917788f59f11571b [file] [log] [blame]
Tom Stellarde1818af2016-02-18 03:42:32 +00001//===-- AMDGPUDisassembler.cpp - Disassembler for AMDGPU ISA --------------===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10//===----------------------------------------------------------------------===//
11//
12/// \file
13///
14/// This file contains definition for AMDGPU ISA disassembler
15//
16//===----------------------------------------------------------------------===//
17
18// ToDo: What to do with instruction suffixes (v_mov_b32 vs v_mov_b32_e32)?
19
20#include "AMDGPUDisassembler.h"
21#include "AMDGPU.h"
22#include "AMDGPURegisterInfo.h"
Chandler Carruth6bda14b2017-06-06 11:49:48 +000023#include "MCTargetDesc/AMDGPUMCTargetDesc.h"
Artem Tamazov212a2512016-05-24 12:05:16 +000024#include "SIDefines.h"
Tom Stellarde1818af2016-02-18 03:42:32 +000025#include "Utils/AMDGPUBaseInfo.h"
26
Zachary Turner264b5d92017-06-07 03:48:56 +000027#include "llvm/BinaryFormat/ELF.h"
Nikolay Haustovac106ad2016-03-01 13:57:29 +000028#include "llvm/MC/MCContext.h"
Tom Stellarde1818af2016-02-18 03:42:32 +000029#include "llvm/MC/MCFixedLenDisassembler.h"
30#include "llvm/MC/MCInst.h"
31#include "llvm/MC/MCInstrDesc.h"
32#include "llvm/MC/MCSubtargetInfo.h"
Chandler Carruth6bda14b2017-06-06 11:49:48 +000033#include "llvm/Support/Debug.h"
Nikolay Haustovac106ad2016-03-01 13:57:29 +000034#include "llvm/Support/Endian.h"
Tom Stellarde1818af2016-02-18 03:42:32 +000035#include "llvm/Support/TargetRegistry.h"
36
Tom Stellarde1818af2016-02-18 03:42:32 +000037using namespace llvm;
38
39#define DEBUG_TYPE "amdgpu-disassembler"
40
41typedef llvm::MCDisassembler::DecodeStatus DecodeStatus;
42
43
Nikolay Haustovac106ad2016-03-01 13:57:29 +000044inline static MCDisassembler::DecodeStatus
45addOperand(MCInst &Inst, const MCOperand& Opnd) {
46 Inst.addOperand(Opnd);
47 return Opnd.isValid() ?
48 MCDisassembler::Success :
49 MCDisassembler::SoftFail;
Tom Stellarde1818af2016-02-18 03:42:32 +000050}
51
Sam Kolton549c89d2017-06-21 08:53:38 +000052static int insertNamedMCOperand(MCInst &MI, const MCOperand &Op,
53 uint16_t NameIdx) {
54 int OpIdx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), NameIdx);
55 if (OpIdx != -1) {
56 auto I = MI.begin();
57 std::advance(I, OpIdx);
58 MI.insert(I, Op);
59 }
60 return OpIdx;
61}
62
Sam Kolton3381d7a2016-10-06 13:46:08 +000063static DecodeStatus decodeSoppBrTarget(MCInst &Inst, unsigned Imm,
64 uint64_t Addr, const void *Decoder) {
65 auto DAsm = static_cast<const AMDGPUDisassembler*>(Decoder);
66
67 APInt SignedOffset(18, Imm * 4, true);
68 int64_t Offset = (SignedOffset.sext(64) + 4 + Addr).getSExtValue();
69
70 if (DAsm->tryAddingSymbolicOperand(Inst, Offset, Addr, true, 2, 2))
71 return MCDisassembler::Success;
Matt Arsenaultf3dd8632016-11-01 00:55:14 +000072 return addOperand(Inst, MCOperand::createImm(Imm));
Sam Kolton3381d7a2016-10-06 13:46:08 +000073}
74
Sam Kolton363f47a2017-05-26 15:52:00 +000075#define DECODE_OPERAND(StaticDecoderName, DecoderName) \
76static DecodeStatus StaticDecoderName(MCInst &Inst, \
77 unsigned Imm, \
78 uint64_t /*Addr*/, \
79 const void *Decoder) { \
Nikolay Haustovac106ad2016-03-01 13:57:29 +000080 auto DAsm = static_cast<const AMDGPUDisassembler*>(Decoder); \
Sam Kolton363f47a2017-05-26 15:52:00 +000081 return addOperand(Inst, DAsm->DecoderName(Imm)); \
Tom Stellarde1818af2016-02-18 03:42:32 +000082}
83
Sam Kolton363f47a2017-05-26 15:52:00 +000084#define DECODE_OPERAND_REG(RegClass) \
85DECODE_OPERAND(Decode##RegClass##RegisterClass, decodeOperand_##RegClass)
Tom Stellarde1818af2016-02-18 03:42:32 +000086
Sam Kolton363f47a2017-05-26 15:52:00 +000087DECODE_OPERAND_REG(VGPR_32)
88DECODE_OPERAND_REG(VS_32)
89DECODE_OPERAND_REG(VS_64)
Dmitry Preobrazhensky30fc5232017-07-18 13:12:48 +000090DECODE_OPERAND_REG(VS_128)
Nikolay Haustov161a1582016-02-25 16:09:14 +000091
Sam Kolton363f47a2017-05-26 15:52:00 +000092DECODE_OPERAND_REG(VReg_64)
93DECODE_OPERAND_REG(VReg_96)
94DECODE_OPERAND_REG(VReg_128)
Tom Stellarde1818af2016-02-18 03:42:32 +000095
Sam Kolton363f47a2017-05-26 15:52:00 +000096DECODE_OPERAND_REG(SReg_32)
97DECODE_OPERAND_REG(SReg_32_XM0_XEXEC)
98DECODE_OPERAND_REG(SReg_64)
99DECODE_OPERAND_REG(SReg_64_XEXEC)
100DECODE_OPERAND_REG(SReg_128)
101DECODE_OPERAND_REG(SReg_256)
102DECODE_OPERAND_REG(SReg_512)
Tom Stellarde1818af2016-02-18 03:42:32 +0000103
Matt Arsenault4bd72362016-12-10 00:39:12 +0000104
105static DecodeStatus decodeOperand_VSrc16(MCInst &Inst,
106 unsigned Imm,
107 uint64_t Addr,
108 const void *Decoder) {
109 auto DAsm = static_cast<const AMDGPUDisassembler*>(Decoder);
110 return addOperand(Inst, DAsm->decodeOperand_VSrc16(Imm));
111}
112
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000113static DecodeStatus decodeOperand_VSrcV216(MCInst &Inst,
114 unsigned Imm,
115 uint64_t Addr,
116 const void *Decoder) {
117 auto DAsm = static_cast<const AMDGPUDisassembler*>(Decoder);
118 return addOperand(Inst, DAsm->decodeOperand_VSrcV216(Imm));
119}
120
Sam Kolton549c89d2017-06-21 08:53:38 +0000121#define DECODE_SDWA(DecName) \
122DECODE_OPERAND(decodeSDWA##DecName, decodeSDWA##DecName)
Sam Kolton363f47a2017-05-26 15:52:00 +0000123
Sam Kolton549c89d2017-06-21 08:53:38 +0000124DECODE_SDWA(Src32)
125DECODE_SDWA(Src16)
126DECODE_SDWA(VopcDst)
Sam Kolton363f47a2017-05-26 15:52:00 +0000127
Tom Stellarde1818af2016-02-18 03:42:32 +0000128#include "AMDGPUGenDisassemblerTables.inc"
129
130//===----------------------------------------------------------------------===//
131//
132//===----------------------------------------------------------------------===//
133
Sam Kolton1048fb12016-03-31 14:15:04 +0000134template <typename T> static inline T eatBytes(ArrayRef<uint8_t>& Bytes) {
135 assert(Bytes.size() >= sizeof(T));
136 const auto Res = support::endian::read<T, support::endianness::little>(Bytes.data());
137 Bytes = Bytes.slice(sizeof(T));
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000138 return Res;
139}
140
141DecodeStatus AMDGPUDisassembler::tryDecodeInst(const uint8_t* Table,
142 MCInst &MI,
143 uint64_t Inst,
144 uint64_t Address) const {
145 assert(MI.getOpcode() == 0);
146 assert(MI.getNumOperands() == 0);
147 MCInst TmpInst;
Dmitry Preobrazhenskyce941c92017-05-19 14:27:52 +0000148 HasLiteral = false;
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000149 const auto SavedBytes = Bytes;
150 if (decodeInstruction(Table, TmpInst, Inst, Address, this, STI)) {
151 MI = TmpInst;
152 return MCDisassembler::Success;
153 }
154 Bytes = SavedBytes;
155 return MCDisassembler::Fail;
156}
157
Tom Stellarde1818af2016-02-18 03:42:32 +0000158DecodeStatus AMDGPUDisassembler::getInstruction(MCInst &MI, uint64_t &Size,
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000159 ArrayRef<uint8_t> Bytes_,
Nikolay Haustov161a1582016-02-25 16:09:14 +0000160 uint64_t Address,
Tom Stellarde1818af2016-02-18 03:42:32 +0000161 raw_ostream &WS,
162 raw_ostream &CS) const {
163 CommentStream = &CS;
Sam Kolton549c89d2017-06-21 08:53:38 +0000164 bool IsSDWA = false;
Tom Stellarde1818af2016-02-18 03:42:32 +0000165
166 // ToDo: AMDGPUDisassembler supports only VI ISA.
Matt Arsenaultd122abe2017-02-15 21:50:34 +0000167 if (!STI.getFeatureBits()[AMDGPU::FeatureGCN3Encoding])
168 report_fatal_error("Disassembly not yet supported for subtarget");
Tom Stellarde1818af2016-02-18 03:42:32 +0000169
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000170 const unsigned MaxInstBytesNum = (std::min)((size_t)8, Bytes_.size());
171 Bytes = Bytes_.slice(0, MaxInstBytesNum);
Nikolay Haustov161a1582016-02-25 16:09:14 +0000172
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000173 DecodeStatus Res = MCDisassembler::Fail;
174 do {
Valery Pykhtin824e8042016-03-04 10:59:50 +0000175 // ToDo: better to switch encoding length using some bit predicate
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000176 // but it is unknown yet, so try all we can
Matt Arsenault37fefd62016-06-10 02:18:02 +0000177
Sam Koltonc9bdcb72016-06-09 11:04:45 +0000178 // Try to decode DPP and SDWA first to solve conflict with VOP1 and VOP2
179 // encodings
Sam Kolton1048fb12016-03-31 14:15:04 +0000180 if (Bytes.size() >= 8) {
181 const uint64_t QW = eatBytes<uint64_t>(Bytes);
182 Res = tryDecodeInst(DecoderTableDPP64, MI, QW, Address);
183 if (Res) break;
Sam Koltonc9bdcb72016-06-09 11:04:45 +0000184
185 Res = tryDecodeInst(DecoderTableSDWA64, MI, QW, Address);
Sam Kolton549c89d2017-06-21 08:53:38 +0000186 if (Res) { IsSDWA = true; break; }
Sam Kolton363f47a2017-05-26 15:52:00 +0000187
188 Res = tryDecodeInst(DecoderTableSDWA964, MI, QW, Address);
Sam Kolton549c89d2017-06-21 08:53:38 +0000189 if (Res) { IsSDWA = true; break; }
Sam Kolton1048fb12016-03-31 14:15:04 +0000190 }
191
192 // Reinitialize Bytes as DPP64 could have eaten too much
193 Bytes = Bytes_.slice(0, MaxInstBytesNum);
194
195 // Try decode 32-bit instruction
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000196 if (Bytes.size() < 4) break;
Sam Kolton1048fb12016-03-31 14:15:04 +0000197 const uint32_t DW = eatBytes<uint32_t>(Bytes);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000198 Res = tryDecodeInst(DecoderTableVI32, MI, DW, Address);
199 if (Res) break;
Tom Stellarde1818af2016-02-18 03:42:32 +0000200
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000201 Res = tryDecodeInst(DecoderTableAMDGPU32, MI, DW, Address);
202 if (Res) break;
Tom Stellarde1818af2016-02-18 03:42:32 +0000203
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000204 if (Bytes.size() < 4) break;
Sam Kolton1048fb12016-03-31 14:15:04 +0000205 const uint64_t QW = ((uint64_t)eatBytes<uint32_t>(Bytes) << 32) | DW;
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000206 Res = tryDecodeInst(DecoderTableVI64, MI, QW, Address);
207 if (Res) break;
208
209 Res = tryDecodeInst(DecoderTableAMDGPU64, MI, QW, Address);
210 } while (false);
211
Matt Arsenault678e1112017-04-10 17:58:06 +0000212 if (Res && (MI.getOpcode() == AMDGPU::V_MAC_F32_e64_vi ||
213 MI.getOpcode() == AMDGPU::V_MAC_F32_e64_si ||
214 MI.getOpcode() == AMDGPU::V_MAC_F16_e64_vi)) {
215 // Insert dummy unused src2_modifiers.
Sam Kolton549c89d2017-06-21 08:53:38 +0000216 insertNamedMCOperand(MI, MCOperand::createImm(0),
217 AMDGPU::OpName::src2_modifiers);
Matt Arsenault678e1112017-04-10 17:58:06 +0000218 }
219
Sam Kolton549c89d2017-06-21 08:53:38 +0000220 if (Res && IsSDWA)
221 Res = convertSDWAInst(MI);
222
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000223 Size = Res ? (MaxInstBytesNum - Bytes.size()) : 0;
224 return Res;
Tom Stellarde1818af2016-02-18 03:42:32 +0000225}
226
Sam Kolton549c89d2017-06-21 08:53:38 +0000227DecodeStatus AMDGPUDisassembler::convertSDWAInst(MCInst &MI) const {
228 if (STI.getFeatureBits()[AMDGPU::FeatureGFX9]) {
229 if (AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::sdst) != -1)
230 // VOPC - insert clamp
231 insertNamedMCOperand(MI, MCOperand::createImm(0), AMDGPU::OpName::clamp);
232 } else if (STI.getFeatureBits()[AMDGPU::FeatureVolcanicIslands]) {
233 int SDst = AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::sdst);
234 if (SDst != -1) {
235 // VOPC - insert VCC register as sdst
236 insertNamedMCOperand(MI, MCOperand::createReg(AMDGPU::VCC),
237 AMDGPU::OpName::sdst);
238 } else {
239 // VOP1/2 - insert omod if present in instruction
240 insertNamedMCOperand(MI, MCOperand::createImm(0), AMDGPU::OpName::omod);
241 }
242 }
243 return MCDisassembler::Success;
244}
245
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000246const char* AMDGPUDisassembler::getRegClassName(unsigned RegClassID) const {
247 return getContext().getRegisterInfo()->
248 getRegClassName(&AMDGPUMCRegisterClasses[RegClassID]);
Tom Stellarde1818af2016-02-18 03:42:32 +0000249}
250
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000251inline
252MCOperand AMDGPUDisassembler::errOperand(unsigned V,
253 const Twine& ErrMsg) const {
254 *CommentStream << "Error: " + ErrMsg;
255
256 // ToDo: add support for error operands to MCInst.h
257 // return MCOperand::createError(V);
258 return MCOperand();
Nikolay Haustov161a1582016-02-25 16:09:14 +0000259}
260
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000261inline
262MCOperand AMDGPUDisassembler::createRegOperand(unsigned int RegId) const {
263 return MCOperand::createReg(RegId);
Tom Stellarde1818af2016-02-18 03:42:32 +0000264}
265
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000266inline
267MCOperand AMDGPUDisassembler::createRegOperand(unsigned RegClassID,
268 unsigned Val) const {
269 const auto& RegCl = AMDGPUMCRegisterClasses[RegClassID];
270 if (Val >= RegCl.getNumRegs())
271 return errOperand(Val, Twine(getRegClassName(RegClassID)) +
272 ": unknown register " + Twine(Val));
273 return createRegOperand(RegCl.getRegister(Val));
Tom Stellarde1818af2016-02-18 03:42:32 +0000274}
275
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000276inline
277MCOperand AMDGPUDisassembler::createSRegOperand(unsigned SRegClassID,
278 unsigned Val) const {
Tom Stellarde1818af2016-02-18 03:42:32 +0000279 // ToDo: SI/CI have 104 SGPRs, VI - 102
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000280 // Valery: here we accepting as much as we can, let assembler sort it out
281 int shift = 0;
282 switch (SRegClassID) {
283 case AMDGPU::SGPR_32RegClassID:
Artem Tamazov212a2512016-05-24 12:05:16 +0000284 case AMDGPU::TTMP_32RegClassID:
285 break;
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000286 case AMDGPU::SGPR_64RegClassID:
Artem Tamazov212a2512016-05-24 12:05:16 +0000287 case AMDGPU::TTMP_64RegClassID:
288 shift = 1;
289 break;
290 case AMDGPU::SGPR_128RegClassID:
291 case AMDGPU::TTMP_128RegClassID:
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000292 // ToDo: unclear if s[100:104] is available on VI. Can we use VCC as SGPR in
293 // this bundle?
294 case AMDGPU::SReg_256RegClassID:
295 // ToDo: unclear if s[96:104] is available on VI. Can we use VCC as SGPR in
296 // this bundle?
Artem Tamazov212a2512016-05-24 12:05:16 +0000297 case AMDGPU::SReg_512RegClassID:
298 shift = 2;
299 break;
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000300 // ToDo: unclear if s[88:104] is available on VI. Can we use VCC as SGPR in
301 // this bundle?
Artem Tamazov212a2512016-05-24 12:05:16 +0000302 default:
Matt Arsenault92b355b2016-11-15 19:34:37 +0000303 llvm_unreachable("unhandled register class");
Tom Stellarde1818af2016-02-18 03:42:32 +0000304 }
Matt Arsenault92b355b2016-11-15 19:34:37 +0000305
306 if (Val % (1 << shift)) {
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000307 *CommentStream << "Warning: " << getRegClassName(SRegClassID)
308 << ": scalar reg isn't aligned " << Val;
Matt Arsenault92b355b2016-11-15 19:34:37 +0000309 }
310
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000311 return createRegOperand(SRegClassID, Val >> shift);
Tom Stellarde1818af2016-02-18 03:42:32 +0000312}
313
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000314MCOperand AMDGPUDisassembler::decodeOperand_VS_32(unsigned Val) const {
Artem Tamazov212a2512016-05-24 12:05:16 +0000315 return decodeSrcOp(OPW32, Val);
Tom Stellarde1818af2016-02-18 03:42:32 +0000316}
317
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000318MCOperand AMDGPUDisassembler::decodeOperand_VS_64(unsigned Val) const {
Artem Tamazov212a2512016-05-24 12:05:16 +0000319 return decodeSrcOp(OPW64, Val);
Nikolay Haustov161a1582016-02-25 16:09:14 +0000320}
321
Dmitry Preobrazhensky30fc5232017-07-18 13:12:48 +0000322MCOperand AMDGPUDisassembler::decodeOperand_VS_128(unsigned Val) const {
323 return decodeSrcOp(OPW128, Val);
324}
325
Matt Arsenault4bd72362016-12-10 00:39:12 +0000326MCOperand AMDGPUDisassembler::decodeOperand_VSrc16(unsigned Val) const {
327 return decodeSrcOp(OPW16, Val);
328}
329
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000330MCOperand AMDGPUDisassembler::decodeOperand_VSrcV216(unsigned Val) const {
331 return decodeSrcOp(OPWV216, Val);
332}
333
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000334MCOperand AMDGPUDisassembler::decodeOperand_VGPR_32(unsigned Val) const {
Matt Arsenaultcb540bc2016-07-19 00:35:03 +0000335 // Some instructions have operand restrictions beyond what the encoding
336 // allows. Some ordinarily VSrc_32 operands are VGPR_32, so clear the extra
337 // high bit.
338 Val &= 255;
339
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000340 return createRegOperand(AMDGPU::VGPR_32RegClassID, Val);
341}
342
343MCOperand AMDGPUDisassembler::decodeOperand_VReg_64(unsigned Val) const {
344 return createRegOperand(AMDGPU::VReg_64RegClassID, Val);
345}
346
347MCOperand AMDGPUDisassembler::decodeOperand_VReg_96(unsigned Val) const {
348 return createRegOperand(AMDGPU::VReg_96RegClassID, Val);
349}
350
351MCOperand AMDGPUDisassembler::decodeOperand_VReg_128(unsigned Val) const {
352 return createRegOperand(AMDGPU::VReg_128RegClassID, Val);
353}
354
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000355MCOperand AMDGPUDisassembler::decodeOperand_SReg_32(unsigned Val) const {
356 // table-gen generated disassembler doesn't care about operand types
357 // leaving only registry class so SSrc_32 operand turns into SReg_32
358 // and therefore we accept immediates and literals here as well
Artem Tamazov212a2512016-05-24 12:05:16 +0000359 return decodeSrcOp(OPW32, Val);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000360}
361
Matt Arsenault640c44b2016-11-29 19:39:53 +0000362MCOperand AMDGPUDisassembler::decodeOperand_SReg_32_XM0_XEXEC(
363 unsigned Val) const {
364 // SReg_32_XM0 is SReg_32 without M0 or EXEC_LO/EXEC_HI
Artem Tamazov38e496b2016-04-29 17:04:50 +0000365 return decodeOperand_SReg_32(Val);
366}
367
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000368MCOperand AMDGPUDisassembler::decodeOperand_SReg_64(unsigned Val) const {
Matt Arsenault640c44b2016-11-29 19:39:53 +0000369 return decodeSrcOp(OPW64, Val);
370}
371
372MCOperand AMDGPUDisassembler::decodeOperand_SReg_64_XEXEC(unsigned Val) const {
Artem Tamazov212a2512016-05-24 12:05:16 +0000373 return decodeSrcOp(OPW64, Val);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000374}
375
376MCOperand AMDGPUDisassembler::decodeOperand_SReg_128(unsigned Val) const {
Artem Tamazov212a2512016-05-24 12:05:16 +0000377 return decodeSrcOp(OPW128, Val);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000378}
379
380MCOperand AMDGPUDisassembler::decodeOperand_SReg_256(unsigned Val) const {
381 return createSRegOperand(AMDGPU::SReg_256RegClassID, Val);
382}
383
384MCOperand AMDGPUDisassembler::decodeOperand_SReg_512(unsigned Val) const {
385 return createSRegOperand(AMDGPU::SReg_512RegClassID, Val);
386}
387
388
389MCOperand AMDGPUDisassembler::decodeLiteralConstant() const {
Nikolay Haustov161a1582016-02-25 16:09:14 +0000390 // For now all literal constants are supposed to be unsigned integer
391 // ToDo: deal with signed/unsigned 64-bit integer constants
392 // ToDo: deal with float/double constants
Dmitry Preobrazhenskyce941c92017-05-19 14:27:52 +0000393 if (!HasLiteral) {
394 if (Bytes.size() < 4) {
395 return errOperand(0, "cannot read literal, inst bytes left " +
396 Twine(Bytes.size()));
397 }
398 HasLiteral = true;
399 Literal = eatBytes<uint32_t>(Bytes);
400 }
401 return MCOperand::createImm(Literal);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000402}
403
404MCOperand AMDGPUDisassembler::decodeIntImmed(unsigned Imm) {
Artem Tamazov212a2512016-05-24 12:05:16 +0000405 using namespace AMDGPU::EncValues;
406 assert(Imm >= INLINE_INTEGER_C_MIN && Imm <= INLINE_INTEGER_C_MAX);
407 return MCOperand::createImm((Imm <= INLINE_INTEGER_C_POSITIVE_MAX) ?
408 (static_cast<int64_t>(Imm) - INLINE_INTEGER_C_MIN) :
409 (INLINE_INTEGER_C_POSITIVE_MAX - static_cast<int64_t>(Imm)));
410 // Cast prevents negative overflow.
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000411}
412
Matt Arsenault4bd72362016-12-10 00:39:12 +0000413static int64_t getInlineImmVal32(unsigned Imm) {
414 switch (Imm) {
415 case 240:
416 return FloatToBits(0.5f);
417 case 241:
418 return FloatToBits(-0.5f);
419 case 242:
420 return FloatToBits(1.0f);
421 case 243:
422 return FloatToBits(-1.0f);
423 case 244:
424 return FloatToBits(2.0f);
425 case 245:
426 return FloatToBits(-2.0f);
427 case 246:
428 return FloatToBits(4.0f);
429 case 247:
430 return FloatToBits(-4.0f);
431 case 248: // 1 / (2 * PI)
432 return 0x3e22f983;
433 default:
434 llvm_unreachable("invalid fp inline imm");
435 }
436}
437
438static int64_t getInlineImmVal64(unsigned Imm) {
439 switch (Imm) {
440 case 240:
441 return DoubleToBits(0.5);
442 case 241:
443 return DoubleToBits(-0.5);
444 case 242:
445 return DoubleToBits(1.0);
446 case 243:
447 return DoubleToBits(-1.0);
448 case 244:
449 return DoubleToBits(2.0);
450 case 245:
451 return DoubleToBits(-2.0);
452 case 246:
453 return DoubleToBits(4.0);
454 case 247:
455 return DoubleToBits(-4.0);
456 case 248: // 1 / (2 * PI)
457 return 0x3fc45f306dc9c882;
458 default:
459 llvm_unreachable("invalid fp inline imm");
460 }
461}
462
463static int64_t getInlineImmVal16(unsigned Imm) {
464 switch (Imm) {
465 case 240:
466 return 0x3800;
467 case 241:
468 return 0xB800;
469 case 242:
470 return 0x3C00;
471 case 243:
472 return 0xBC00;
473 case 244:
474 return 0x4000;
475 case 245:
476 return 0xC000;
477 case 246:
478 return 0x4400;
479 case 247:
480 return 0xC400;
481 case 248: // 1 / (2 * PI)
482 return 0x3118;
483 default:
484 llvm_unreachable("invalid fp inline imm");
485 }
486}
487
488MCOperand AMDGPUDisassembler::decodeFPImmed(OpWidthTy Width, unsigned Imm) {
Artem Tamazov212a2512016-05-24 12:05:16 +0000489 assert(Imm >= AMDGPU::EncValues::INLINE_FLOATING_C_MIN
490 && Imm <= AMDGPU::EncValues::INLINE_FLOATING_C_MAX);
Matt Arsenault4bd72362016-12-10 00:39:12 +0000491
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000492 // ToDo: case 248: 1/(2*PI) - is allowed only on VI
Matt Arsenault4bd72362016-12-10 00:39:12 +0000493 switch (Width) {
494 case OPW32:
495 return MCOperand::createImm(getInlineImmVal32(Imm));
496 case OPW64:
497 return MCOperand::createImm(getInlineImmVal64(Imm));
498 case OPW16:
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000499 case OPWV216:
Matt Arsenault4bd72362016-12-10 00:39:12 +0000500 return MCOperand::createImm(getInlineImmVal16(Imm));
501 default:
502 llvm_unreachable("implement me");
Nikolay Haustov161a1582016-02-25 16:09:14 +0000503 }
Nikolay Haustov161a1582016-02-25 16:09:14 +0000504}
505
Artem Tamazov212a2512016-05-24 12:05:16 +0000506unsigned AMDGPUDisassembler::getVgprClassId(const OpWidthTy Width) const {
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000507 using namespace AMDGPU;
Artem Tamazov212a2512016-05-24 12:05:16 +0000508 assert(OPW_FIRST_ <= Width && Width < OPW_LAST_);
509 switch (Width) {
510 default: // fall
Matt Arsenault4bd72362016-12-10 00:39:12 +0000511 case OPW32:
512 case OPW16:
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000513 case OPWV216:
Matt Arsenault4bd72362016-12-10 00:39:12 +0000514 return VGPR_32RegClassID;
Artem Tamazov212a2512016-05-24 12:05:16 +0000515 case OPW64: return VReg_64RegClassID;
516 case OPW128: return VReg_128RegClassID;
517 }
518}
519
520unsigned AMDGPUDisassembler::getSgprClassId(const OpWidthTy Width) const {
521 using namespace AMDGPU;
522 assert(OPW_FIRST_ <= Width && Width < OPW_LAST_);
523 switch (Width) {
524 default: // fall
Matt Arsenault4bd72362016-12-10 00:39:12 +0000525 case OPW32:
526 case OPW16:
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000527 case OPWV216:
Matt Arsenault4bd72362016-12-10 00:39:12 +0000528 return SGPR_32RegClassID;
Artem Tamazov212a2512016-05-24 12:05:16 +0000529 case OPW64: return SGPR_64RegClassID;
530 case OPW128: return SGPR_128RegClassID;
531 }
532}
533
534unsigned AMDGPUDisassembler::getTtmpClassId(const OpWidthTy Width) const {
535 using namespace AMDGPU;
536 assert(OPW_FIRST_ <= Width && Width < OPW_LAST_);
537 switch (Width) {
538 default: // fall
Matt Arsenault4bd72362016-12-10 00:39:12 +0000539 case OPW32:
540 case OPW16:
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000541 case OPWV216:
Matt Arsenault4bd72362016-12-10 00:39:12 +0000542 return TTMP_32RegClassID;
Artem Tamazov212a2512016-05-24 12:05:16 +0000543 case OPW64: return TTMP_64RegClassID;
544 case OPW128: return TTMP_128RegClassID;
545 }
546}
547
548MCOperand AMDGPUDisassembler::decodeSrcOp(const OpWidthTy Width, unsigned Val) const {
549 using namespace AMDGPU::EncValues;
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000550 assert(Val < 512); // enum9
551
Artem Tamazov212a2512016-05-24 12:05:16 +0000552 if (VGPR_MIN <= Val && Val <= VGPR_MAX) {
553 return createRegOperand(getVgprClassId(Width), Val - VGPR_MIN);
554 }
Artem Tamazovb49c3362016-05-26 15:52:16 +0000555 if (Val <= SGPR_MAX) {
556 assert(SGPR_MIN == 0); // "SGPR_MIN <= Val" is always true and causes compilation warning.
Artem Tamazov212a2512016-05-24 12:05:16 +0000557 return createSRegOperand(getSgprClassId(Width), Val - SGPR_MIN);
558 }
559 if (TTMP_MIN <= Val && Val <= TTMP_MAX) {
560 return createSRegOperand(getTtmpClassId(Width), Val - TTMP_MIN);
561 }
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000562
Artem Tamazov212a2512016-05-24 12:05:16 +0000563 if (INLINE_INTEGER_C_MIN <= Val && Val <= INLINE_INTEGER_C_MAX)
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000564 return decodeIntImmed(Val);
565
Artem Tamazov212a2512016-05-24 12:05:16 +0000566 if (INLINE_FLOATING_C_MIN <= Val && Val <= INLINE_FLOATING_C_MAX)
Matt Arsenault4bd72362016-12-10 00:39:12 +0000567 return decodeFPImmed(Width, Val);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000568
Artem Tamazov212a2512016-05-24 12:05:16 +0000569 if (Val == LITERAL_CONST)
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000570 return decodeLiteralConstant();
571
Matt Arsenault4bd72362016-12-10 00:39:12 +0000572 switch (Width) {
573 case OPW32:
574 case OPW16:
Matt Arsenault9be7b0d2017-02-27 18:49:11 +0000575 case OPWV216:
Matt Arsenault4bd72362016-12-10 00:39:12 +0000576 return decodeSpecialReg32(Val);
577 case OPW64:
578 return decodeSpecialReg64(Val);
579 default:
580 llvm_unreachable("unexpected immediate type");
581 }
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000582}
583
584MCOperand AMDGPUDisassembler::decodeSpecialReg32(unsigned Val) const {
585 using namespace AMDGPU;
586 switch (Val) {
587 case 102: return createRegOperand(getMCReg(FLAT_SCR_LO, STI));
588 case 103: return createRegOperand(getMCReg(FLAT_SCR_HI, STI));
589 // ToDo: no support for xnack_mask_lo/_hi register
590 case 104:
591 case 105: break;
592 case 106: return createRegOperand(VCC_LO);
593 case 107: return createRegOperand(VCC_HI);
Artem Tamazov212a2512016-05-24 12:05:16 +0000594 case 108: return createRegOperand(TBA_LO);
595 case 109: return createRegOperand(TBA_HI);
596 case 110: return createRegOperand(TMA_LO);
597 case 111: return createRegOperand(TMA_HI);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000598 case 124: return createRegOperand(M0);
599 case 126: return createRegOperand(EXEC_LO);
600 case 127: return createRegOperand(EXEC_HI);
Matt Arsenaulta3b3b482017-02-18 18:41:41 +0000601 case 235: return createRegOperand(SRC_SHARED_BASE);
602 case 236: return createRegOperand(SRC_SHARED_LIMIT);
603 case 237: return createRegOperand(SRC_PRIVATE_BASE);
604 case 238: return createRegOperand(SRC_PRIVATE_LIMIT);
605 // TODO: SRC_POPS_EXITING_WAVE_ID
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000606 // ToDo: no support for vccz register
607 case 251: break;
608 // ToDo: no support for execz register
609 case 252: break;
610 case 253: return createRegOperand(SCC);
611 default: break;
Tom Stellarde1818af2016-02-18 03:42:32 +0000612 }
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000613 return errOperand(Val, "unknown operand encoding " + Twine(Val));
Tom Stellarde1818af2016-02-18 03:42:32 +0000614}
615
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000616MCOperand AMDGPUDisassembler::decodeSpecialReg64(unsigned Val) const {
617 using namespace AMDGPU;
618 switch (Val) {
619 case 102: return createRegOperand(getMCReg(FLAT_SCR, STI));
620 case 106: return createRegOperand(VCC);
Artem Tamazov212a2512016-05-24 12:05:16 +0000621 case 108: return createRegOperand(TBA);
622 case 110: return createRegOperand(TMA);
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000623 case 126: return createRegOperand(EXEC);
624 default: break;
Tom Stellarde1818af2016-02-18 03:42:32 +0000625 }
Nikolay Haustovac106ad2016-03-01 13:57:29 +0000626 return errOperand(Val, "unknown operand encoding " + Twine(Val));
Tom Stellarde1818af2016-02-18 03:42:32 +0000627}
628
Sam Kolton549c89d2017-06-21 08:53:38 +0000629MCOperand AMDGPUDisassembler::decodeSDWASrc(const OpWidthTy Width,
630 unsigned Val) const {
Sam Kolton363f47a2017-05-26 15:52:00 +0000631 using namespace AMDGPU::SDWA;
632
Sam Kolton549c89d2017-06-21 08:53:38 +0000633 if (STI.getFeatureBits()[AMDGPU::FeatureGFX9]) {
Sam Koltona179d252017-06-27 15:02:23 +0000634 // XXX: static_cast<int> is needed to avoid stupid warning:
635 // compare with unsigned is always true
636 if (SDWA9EncValues::SRC_VGPR_MIN <= static_cast<int>(Val) &&
Sam Kolton549c89d2017-06-21 08:53:38 +0000637 Val <= SDWA9EncValues::SRC_VGPR_MAX) {
638 return createRegOperand(getVgprClassId(Width),
639 Val - SDWA9EncValues::SRC_VGPR_MIN);
640 }
641 if (SDWA9EncValues::SRC_SGPR_MIN <= Val &&
642 Val <= SDWA9EncValues::SRC_SGPR_MAX) {
643 return createSRegOperand(getSgprClassId(Width),
644 Val - SDWA9EncValues::SRC_SGPR_MIN);
645 }
646
647 return decodeSpecialReg32(Val - SDWA9EncValues::SRC_SGPR_MIN);
648 } else if (STI.getFeatureBits()[AMDGPU::FeatureVolcanicIslands]) {
649 return createRegOperand(getVgprClassId(Width), Val);
Sam Kolton363f47a2017-05-26 15:52:00 +0000650 }
Sam Kolton549c89d2017-06-21 08:53:38 +0000651 llvm_unreachable("unsupported target");
Sam Kolton363f47a2017-05-26 15:52:00 +0000652}
653
Sam Kolton549c89d2017-06-21 08:53:38 +0000654MCOperand AMDGPUDisassembler::decodeSDWASrc16(unsigned Val) const {
655 return decodeSDWASrc(OPW16, Val);
Sam Kolton363f47a2017-05-26 15:52:00 +0000656}
657
Sam Kolton549c89d2017-06-21 08:53:38 +0000658MCOperand AMDGPUDisassembler::decodeSDWASrc32(unsigned Val) const {
659 return decodeSDWASrc(OPW32, Val);
Sam Kolton363f47a2017-05-26 15:52:00 +0000660}
661
662
Sam Kolton549c89d2017-06-21 08:53:38 +0000663MCOperand AMDGPUDisassembler::decodeSDWAVopcDst(unsigned Val) const {
Sam Kolton363f47a2017-05-26 15:52:00 +0000664 using namespace AMDGPU::SDWA;
665
Sam Kolton549c89d2017-06-21 08:53:38 +0000666 assert(STI.getFeatureBits()[AMDGPU::FeatureGFX9] &&
667 "SDWAVopcDst should be present only on GFX9");
Sam Kolton363f47a2017-05-26 15:52:00 +0000668 if (Val & SDWA9EncValues::VOPC_DST_VCC_MASK) {
669 Val &= SDWA9EncValues::VOPC_DST_SGPR_MASK;
670 if (Val > AMDGPU::EncValues::SGPR_MAX) {
671 return decodeSpecialReg64(Val);
672 } else {
673 return createSRegOperand(getSgprClassId(OPW64), Val);
674 }
675 } else {
676 return createRegOperand(AMDGPU::VCC);
677 }
678}
679
Sam Kolton3381d7a2016-10-06 13:46:08 +0000680//===----------------------------------------------------------------------===//
681// AMDGPUSymbolizer
682//===----------------------------------------------------------------------===//
Matt Arsenaultf3dd8632016-11-01 00:55:14 +0000683
Sam Kolton3381d7a2016-10-06 13:46:08 +0000684// Try to find symbol name for specified label
685bool AMDGPUSymbolizer::tryAddingSymbolicOperand(MCInst &Inst,
686 raw_ostream &/*cStream*/, int64_t Value,
687 uint64_t /*Address*/, bool IsBranch,
688 uint64_t /*Offset*/, uint64_t /*InstSize*/) {
689 typedef std::tuple<uint64_t, StringRef, uint8_t> SymbolInfoTy;
690 typedef std::vector<SymbolInfoTy> SectionSymbolsTy;
691
692 if (!IsBranch) {
693 return false;
694 }
695
696 auto *Symbols = static_cast<SectionSymbolsTy *>(DisInfo);
697 auto Result = std::find_if(Symbols->begin(), Symbols->end(),
698 [Value](const SymbolInfoTy& Val) {
699 return std::get<0>(Val) == static_cast<uint64_t>(Value)
700 && std::get<2>(Val) == ELF::STT_NOTYPE;
701 });
702 if (Result != Symbols->end()) {
703 auto *Sym = Ctx.getOrCreateSymbol(std::get<1>(*Result));
704 const auto *Add = MCSymbolRefExpr::create(Sym, Ctx);
705 Inst.addOperand(MCOperand::createExpr(Add));
706 return true;
707 }
708 return false;
709}
710
Matt Arsenault92b355b2016-11-15 19:34:37 +0000711void AMDGPUSymbolizer::tryAddingPcLoadReferenceComment(raw_ostream &cStream,
712 int64_t Value,
713 uint64_t Address) {
714 llvm_unreachable("unimplemented");
715}
716
Sam Kolton3381d7a2016-10-06 13:46:08 +0000717//===----------------------------------------------------------------------===//
718// Initialization
719//===----------------------------------------------------------------------===//
720
721static MCSymbolizer *createAMDGPUSymbolizer(const Triple &/*TT*/,
722 LLVMOpInfoCallback /*GetOpInfo*/,
723 LLVMSymbolLookupCallback /*SymbolLookUp*/,
Matt Arsenaultf3dd8632016-11-01 00:55:14 +0000724 void *DisInfo,
Sam Kolton3381d7a2016-10-06 13:46:08 +0000725 MCContext *Ctx,
726 std::unique_ptr<MCRelocationInfo> &&RelInfo) {
727 return new AMDGPUSymbolizer(*Ctx, std::move(RelInfo), DisInfo);
728}
729
Tom Stellarde1818af2016-02-18 03:42:32 +0000730static MCDisassembler *createAMDGPUDisassembler(const Target &T,
731 const MCSubtargetInfo &STI,
732 MCContext &Ctx) {
733 return new AMDGPUDisassembler(STI, Ctx);
734}
735
736extern "C" void LLVMInitializeAMDGPUDisassembler() {
Mehdi Aminif42454b2016-10-09 23:00:34 +0000737 TargetRegistry::RegisterMCDisassembler(getTheGCNTarget(),
738 createAMDGPUDisassembler);
739 TargetRegistry::RegisterMCSymbolizer(getTheGCNTarget(),
740 createAMDGPUSymbolizer);
Tom Stellarde1818af2016-02-18 03:42:32 +0000741}