blob: 2719ea2490ebf3abb8888f33b48f94fab9d03dbd [file] [log] [blame]
Tom Stellard75aadc22012-12-11 21:25:42 +00001//===-- SIInstrInfo.cpp - SI Instruction Information ---------------------===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10/// \file
11/// \brief SI Implementation of TargetInstrInfo.
12//
13//===----------------------------------------------------------------------===//
14
15
16#include "SIInstrInfo.h"
17#include "AMDGPUTargetMachine.h"
Tom Stellard16a9a202013-08-14 23:24:17 +000018#include "SIDefines.h"
Tom Stellard75aadc22012-12-11 21:25:42 +000019#include "llvm/CodeGen/MachineInstrBuilder.h"
20#include "llvm/CodeGen/MachineRegisterInfo.h"
21#include "llvm/MC/MCInstrDesc.h"
Tom Stellard75aadc22012-12-11 21:25:42 +000022#include <stdio.h>
23
24using namespace llvm;
25
26SIInstrInfo::SIInstrInfo(AMDGPUTargetMachine &tm)
27 : AMDGPUInstrInfo(tm),
Bill Wendling37e9adb2013-06-07 20:28:55 +000028 RI(tm)
Tom Stellard75aadc22012-12-11 21:25:42 +000029 { }
30
31const SIRegisterInfo &SIInstrInfo::getRegisterInfo() const {
32 return RI;
33}
34
35void
36SIInstrInfo::copyPhysReg(MachineBasicBlock &MBB,
Christian Konigd0e3da12013-03-01 09:46:27 +000037 MachineBasicBlock::iterator MI, DebugLoc DL,
38 unsigned DestReg, unsigned SrcReg,
39 bool KillSrc) const {
40
Tom Stellard75aadc22012-12-11 21:25:42 +000041 // If we are trying to copy to or from SCC, there is a bug somewhere else in
42 // the backend. While it may be theoretically possible to do this, it should
43 // never be necessary.
44 assert(DestReg != AMDGPU::SCC && SrcReg != AMDGPU::SCC);
45
Craig Topper0afd0ab2013-07-15 06:39:13 +000046 static const int16_t Sub0_15[] = {
Christian Konigd0e3da12013-03-01 09:46:27 +000047 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3,
48 AMDGPU::sub4, AMDGPU::sub5, AMDGPU::sub6, AMDGPU::sub7,
49 AMDGPU::sub8, AMDGPU::sub9, AMDGPU::sub10, AMDGPU::sub11,
50 AMDGPU::sub12, AMDGPU::sub13, AMDGPU::sub14, AMDGPU::sub15, 0
51 };
52
Craig Topper0afd0ab2013-07-15 06:39:13 +000053 static const int16_t Sub0_7[] = {
Christian Konigd0e3da12013-03-01 09:46:27 +000054 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3,
55 AMDGPU::sub4, AMDGPU::sub5, AMDGPU::sub6, AMDGPU::sub7, 0
56 };
57
Craig Topper0afd0ab2013-07-15 06:39:13 +000058 static const int16_t Sub0_3[] = {
Christian Konigd0e3da12013-03-01 09:46:27 +000059 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3, 0
60 };
61
Craig Topper0afd0ab2013-07-15 06:39:13 +000062 static const int16_t Sub0_2[] = {
Christian Konig8b1ed282013-04-10 08:39:16 +000063 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, 0
64 };
65
Craig Topper0afd0ab2013-07-15 06:39:13 +000066 static const int16_t Sub0_1[] = {
Christian Konigd0e3da12013-03-01 09:46:27 +000067 AMDGPU::sub0, AMDGPU::sub1, 0
68 };
69
70 unsigned Opcode;
71 const int16_t *SubIndices;
72
Christian Konig082c6612013-03-26 14:04:12 +000073 if (AMDGPU::M0 == DestReg) {
74 // Check if M0 isn't already set to this value
75 for (MachineBasicBlock::reverse_iterator E = MBB.rend(),
76 I = MachineBasicBlock::reverse_iterator(MI); I != E; ++I) {
77
78 if (!I->definesRegister(AMDGPU::M0))
79 continue;
80
81 unsigned Opc = I->getOpcode();
82 if (Opc != TargetOpcode::COPY && Opc != AMDGPU::S_MOV_B32)
83 break;
84
85 if (!I->readsRegister(SrcReg))
86 break;
87
88 // The copy isn't necessary
89 return;
90 }
91 }
92
Christian Konigd0e3da12013-03-01 09:46:27 +000093 if (AMDGPU::SReg_32RegClass.contains(DestReg)) {
94 assert(AMDGPU::SReg_32RegClass.contains(SrcReg));
95 BuildMI(MBB, MI, DL, get(AMDGPU::S_MOV_B32), DestReg)
96 .addReg(SrcReg, getKillRegState(KillSrc));
97 return;
98
Tom Stellardaac18892013-02-07 19:39:43 +000099 } else if (AMDGPU::SReg_64RegClass.contains(DestReg)) {
Tom Stellard75aadc22012-12-11 21:25:42 +0000100 assert(AMDGPU::SReg_64RegClass.contains(SrcReg));
101 BuildMI(MBB, MI, DL, get(AMDGPU::S_MOV_B64), DestReg)
102 .addReg(SrcReg, getKillRegState(KillSrc));
Christian Konigd0e3da12013-03-01 09:46:27 +0000103 return;
104
105 } else if (AMDGPU::SReg_128RegClass.contains(DestReg)) {
106 assert(AMDGPU::SReg_128RegClass.contains(SrcReg));
107 Opcode = AMDGPU::S_MOV_B32;
108 SubIndices = Sub0_3;
109
110 } else if (AMDGPU::SReg_256RegClass.contains(DestReg)) {
111 assert(AMDGPU::SReg_256RegClass.contains(SrcReg));
112 Opcode = AMDGPU::S_MOV_B32;
113 SubIndices = Sub0_7;
114
115 } else if (AMDGPU::SReg_512RegClass.contains(DestReg)) {
116 assert(AMDGPU::SReg_512RegClass.contains(SrcReg));
117 Opcode = AMDGPU::S_MOV_B32;
118 SubIndices = Sub0_15;
119
Tom Stellard75aadc22012-12-11 21:25:42 +0000120 } else if (AMDGPU::VReg_32RegClass.contains(DestReg)) {
121 assert(AMDGPU::VReg_32RegClass.contains(SrcReg) ||
Christian Konigd0e3da12013-03-01 09:46:27 +0000122 AMDGPU::SReg_32RegClass.contains(SrcReg));
Tom Stellard75aadc22012-12-11 21:25:42 +0000123 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DestReg)
124 .addReg(SrcReg, getKillRegState(KillSrc));
Christian Konigd0e3da12013-03-01 09:46:27 +0000125 return;
126
127 } else if (AMDGPU::VReg_64RegClass.contains(DestReg)) {
128 assert(AMDGPU::VReg_64RegClass.contains(SrcReg) ||
129 AMDGPU::SReg_64RegClass.contains(SrcReg));
130 Opcode = AMDGPU::V_MOV_B32_e32;
131 SubIndices = Sub0_1;
132
Christian Konig8b1ed282013-04-10 08:39:16 +0000133 } else if (AMDGPU::VReg_96RegClass.contains(DestReg)) {
134 assert(AMDGPU::VReg_96RegClass.contains(SrcReg));
135 Opcode = AMDGPU::V_MOV_B32_e32;
136 SubIndices = Sub0_2;
137
Christian Konigd0e3da12013-03-01 09:46:27 +0000138 } else if (AMDGPU::VReg_128RegClass.contains(DestReg)) {
139 assert(AMDGPU::VReg_128RegClass.contains(SrcReg) ||
140 AMDGPU::SReg_128RegClass.contains(SrcReg));
141 Opcode = AMDGPU::V_MOV_B32_e32;
142 SubIndices = Sub0_3;
143
144 } else if (AMDGPU::VReg_256RegClass.contains(DestReg)) {
145 assert(AMDGPU::VReg_256RegClass.contains(SrcReg) ||
146 AMDGPU::SReg_256RegClass.contains(SrcReg));
147 Opcode = AMDGPU::V_MOV_B32_e32;
148 SubIndices = Sub0_7;
149
150 } else if (AMDGPU::VReg_512RegClass.contains(DestReg)) {
151 assert(AMDGPU::VReg_512RegClass.contains(SrcReg) ||
152 AMDGPU::SReg_512RegClass.contains(SrcReg));
153 Opcode = AMDGPU::V_MOV_B32_e32;
154 SubIndices = Sub0_15;
155
Tom Stellard75aadc22012-12-11 21:25:42 +0000156 } else {
Christian Konigd0e3da12013-03-01 09:46:27 +0000157 llvm_unreachable("Can't copy register!");
158 }
159
160 while (unsigned SubIdx = *SubIndices++) {
161 MachineInstrBuilder Builder = BuildMI(MBB, MI, DL,
162 get(Opcode), RI.getSubReg(DestReg, SubIdx));
163
164 Builder.addReg(RI.getSubReg(SrcReg, SubIdx), getKillRegState(KillSrc));
165
166 if (*SubIndices)
167 Builder.addReg(DestReg, RegState::Define | RegState::Implicit);
Tom Stellard75aadc22012-12-11 21:25:42 +0000168 }
169}
170
Christian Konig3c145802013-03-27 09:12:59 +0000171unsigned SIInstrInfo::commuteOpcode(unsigned Opcode) const {
172
173 int NewOpc;
174
175 // Try to map original to commuted opcode
176 if ((NewOpc = AMDGPU::getCommuteRev(Opcode)) != -1)
177 return NewOpc;
178
179 // Try to map commuted to original opcode
180 if ((NewOpc = AMDGPU::getCommuteOrig(Opcode)) != -1)
181 return NewOpc;
182
183 return Opcode;
184}
185
Christian Konig76edd4f2013-02-26 17:52:29 +0000186MachineInstr *SIInstrInfo::commuteInstruction(MachineInstr *MI,
187 bool NewMI) const {
188
189 if (MI->getNumOperands() < 3 || !MI->getOperand(1).isReg() ||
190 !MI->getOperand(2).isReg())
191 return 0;
192
Christian Konig3c145802013-03-27 09:12:59 +0000193 MI = TargetInstrInfo::commuteInstruction(MI, NewMI);
194
195 if (MI)
196 MI->setDesc(get(commuteOpcode(MI->getOpcode())));
197
198 return MI;
Christian Konig76edd4f2013-02-26 17:52:29 +0000199}
200
Tom Stellard75aadc22012-12-11 21:25:42 +0000201MachineInstr * SIInstrInfo::getMovImmInstr(MachineFunction *MF, unsigned DstReg,
202 int64_t Imm) const {
Christian Konigc756cb992013-02-16 11:28:22 +0000203 MachineInstr * MI = MF->CreateMachineInstr(get(AMDGPU::V_MOV_B32_e32), DebugLoc());
NAKAMURA Takumi2a0b40f2012-12-20 00:22:11 +0000204 MachineInstrBuilder MIB(*MF, MI);
205 MIB.addReg(DstReg, RegState::Define);
206 MIB.addImm(Imm);
Tom Stellard75aadc22012-12-11 21:25:42 +0000207
208 return MI;
209
210}
211
212bool SIInstrInfo::isMov(unsigned Opcode) const {
213 switch(Opcode) {
214 default: return false;
215 case AMDGPU::S_MOV_B32:
216 case AMDGPU::S_MOV_B64:
217 case AMDGPU::V_MOV_B32_e32:
218 case AMDGPU::V_MOV_B32_e64:
Tom Stellard75aadc22012-12-11 21:25:42 +0000219 return true;
220 }
221}
222
223bool
224SIInstrInfo::isSafeToMoveRegClassDefs(const TargetRegisterClass *RC) const {
225 return RC != &AMDGPU::EXECRegRegClass;
226}
Tom Stellardf3b2a1e2013-02-06 17:32:29 +0000227
Tom Stellard16a9a202013-08-14 23:24:17 +0000228int SIInstrInfo::isMIMG(uint16_t Opcode) const {
229 return get(Opcode).TSFlags & SIInstrFlags::MIMG;
230}
231
Michel Danzer20680b12013-08-16 16:19:24 +0000232int SIInstrInfo::isSMRD(uint16_t Opcode) const {
233 return get(Opcode).TSFlags & SIInstrFlags::SMRD;
234}
235
Tom Stellardf3b2a1e2013-02-06 17:32:29 +0000236//===----------------------------------------------------------------------===//
237// Indirect addressing callbacks
238//===----------------------------------------------------------------------===//
239
240unsigned SIInstrInfo::calculateIndirectAddress(unsigned RegIndex,
241 unsigned Channel) const {
242 assert(Channel == 0);
243 return RegIndex;
244}
245
246
247int SIInstrInfo::getIndirectIndexBegin(const MachineFunction &MF) const {
248 llvm_unreachable("Unimplemented");
249}
250
251int SIInstrInfo::getIndirectIndexEnd(const MachineFunction &MF) const {
252 llvm_unreachable("Unimplemented");
253}
254
255const TargetRegisterClass *SIInstrInfo::getIndirectAddrStoreRegClass(
256 unsigned SourceReg) const {
257 llvm_unreachable("Unimplemented");
258}
259
260const TargetRegisterClass *SIInstrInfo::getIndirectAddrLoadRegClass() const {
261 llvm_unreachable("Unimplemented");
262}
263
264MachineInstrBuilder SIInstrInfo::buildIndirectWrite(
265 MachineBasicBlock *MBB,
266 MachineBasicBlock::iterator I,
267 unsigned ValueReg,
268 unsigned Address, unsigned OffsetReg) const {
269 llvm_unreachable("Unimplemented");
270}
271
272MachineInstrBuilder SIInstrInfo::buildIndirectRead(
273 MachineBasicBlock *MBB,
274 MachineBasicBlock::iterator I,
275 unsigned ValueReg,
276 unsigned Address, unsigned OffsetReg) const {
277 llvm_unreachable("Unimplemented");
278}
279
280const TargetRegisterClass *SIInstrInfo::getSuperIndirectRegClass() const {
281 llvm_unreachable("Unimplemented");
282}