blob: 8ca85b54e01c5f355b7b822df9c3cc40073af2f0 [file] [log] [blame]
Matt Arsenault0c90e952015-11-06 18:17:45 +00001//===----------------------- SIFrameLowering.cpp --------------------------===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//==-----------------------------------------------------------------------===//
9
10#include "SIFrameLowering.h"
Chandler Carruth6bda14b2017-06-06 11:49:48 +000011#include "AMDGPUSubtarget.h"
Matt Arsenault0e3d3892015-11-30 21:15:53 +000012#include "SIInstrInfo.h"
13#include "SIMachineFunctionInfo.h"
Matt Arsenault0c90e952015-11-06 18:17:45 +000014#include "SIRegisterInfo.h"
Matt Arsenault43e92fe2016-06-24 06:30:11 +000015
Matt Arsenault03ae3992018-03-29 21:30:06 +000016#include "llvm/CodeGen/LivePhysRegs.h"
Matt Arsenault0c90e952015-11-06 18:17:45 +000017#include "llvm/CodeGen/MachineFrameInfo.h"
18#include "llvm/CodeGen/MachineFunction.h"
Matt Arsenault0e3d3892015-11-30 21:15:53 +000019#include "llvm/CodeGen/MachineInstrBuilder.h"
Matt Arsenault0c90e952015-11-06 18:17:45 +000020#include "llvm/CodeGen/RegisterScavenging.h"
21
22using namespace llvm;
23
Matt Arsenault0e3d3892015-11-30 21:15:53 +000024
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +000025static ArrayRef<MCPhysReg> getAllSGPR128(const SISubtarget &ST,
26 const MachineFunction &MF) {
Matt Arsenaultab3429c2016-05-18 15:19:50 +000027 return makeArrayRef(AMDGPU::SGPR_128RegClass.begin(),
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +000028 ST.getMaxNumSGPRs(MF) / 4);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +000029}
30
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +000031static ArrayRef<MCPhysReg> getAllSGPRs(const SISubtarget &ST,
32 const MachineFunction &MF) {
Matt Arsenault26f8f3d2015-11-30 21:16:03 +000033 return makeArrayRef(AMDGPU::SGPR_32RegClass.begin(),
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +000034 ST.getMaxNumSGPRs(MF));
Matt Arsenault26f8f3d2015-11-30 21:16:03 +000035}
36
Matt Arsenaulte823d922017-02-18 18:29:53 +000037void SIFrameLowering::emitFlatScratchInit(const SISubtarget &ST,
Matt Arsenault57bc4322016-08-31 21:52:21 +000038 MachineFunction &MF,
39 MachineBasicBlock &MBB) const {
Matt Arsenaulte823d922017-02-18 18:29:53 +000040 const SIInstrInfo *TII = ST.getInstrInfo();
41 const SIRegisterInfo* TRI = &TII->getRegisterInfo();
Matt Arsenault8623e8d2017-08-03 23:00:29 +000042 const SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
Matt Arsenaulte823d922017-02-18 18:29:53 +000043
Matt Arsenault57bc4322016-08-31 21:52:21 +000044 // We don't need this if we only have spills since there is no user facing
45 // scratch.
46
47 // TODO: If we know we don't have flat instructions earlier, we can omit
48 // this from the input registers.
49 //
50 // TODO: We only need to know if we access scratch space through a flat
51 // pointer. Because we only detect if flat instructions are used at all,
52 // this will be used more often than necessary on VI.
53
54 // Debug location must be unknown since the first debug location is used to
55 // determine the end of the prologue.
56 DebugLoc DL;
57 MachineBasicBlock::iterator I = MBB.begin();
58
59 unsigned FlatScratchInitReg
Matt Arsenault8623e8d2017-08-03 23:00:29 +000060 = MFI->getPreloadedReg(AMDGPUFunctionArgInfo::FLAT_SCRATCH_INIT);
Matt Arsenault57bc4322016-08-31 21:52:21 +000061
62 MachineRegisterInfo &MRI = MF.getRegInfo();
63 MRI.addLiveIn(FlatScratchInitReg);
64 MBB.addLiveIn(FlatScratchInitReg);
65
Matt Arsenault57bc4322016-08-31 21:52:21 +000066 unsigned FlatScrInitLo = TRI->getSubReg(FlatScratchInitReg, AMDGPU::sub0);
Matt Arsenaulte823d922017-02-18 18:29:53 +000067 unsigned FlatScrInitHi = TRI->getSubReg(FlatScratchInitReg, AMDGPU::sub1);
Matt Arsenault57bc4322016-08-31 21:52:21 +000068
Matt Arsenault57bc4322016-08-31 21:52:21 +000069 unsigned ScratchWaveOffsetReg = MFI->getScratchWaveOffsetReg();
70
Matt Arsenaulte823d922017-02-18 18:29:53 +000071 // Do a 64-bit pointer add.
72 if (ST.flatScratchIsPointer()) {
73 BuildMI(MBB, I, DL, TII->get(AMDGPU::S_ADD_U32), AMDGPU::FLAT_SCR_LO)
74 .addReg(FlatScrInitLo)
75 .addReg(ScratchWaveOffsetReg);
76 BuildMI(MBB, I, DL, TII->get(AMDGPU::S_ADDC_U32), AMDGPU::FLAT_SCR_HI)
77 .addReg(FlatScrInitHi)
78 .addImm(0);
79
80 return;
81 }
82
83 // Copy the size in bytes.
84 BuildMI(MBB, I, DL, TII->get(AMDGPU::COPY), AMDGPU::FLAT_SCR_LO)
85 .addReg(FlatScrInitHi, RegState::Kill);
86
Matt Arsenault57bc4322016-08-31 21:52:21 +000087 // Add wave offset in bytes to private base offset.
88 // See comment in AMDKernelCodeT.h for enable_sgpr_flat_scratch_init.
89 BuildMI(MBB, I, DL, TII->get(AMDGPU::S_ADD_U32), FlatScrInitLo)
90 .addReg(FlatScrInitLo)
91 .addReg(ScratchWaveOffsetReg);
92
93 // Convert offset to 256-byte units.
94 BuildMI(MBB, I, DL, TII->get(AMDGPU::S_LSHR_B32), AMDGPU::FLAT_SCR_HI)
95 .addReg(FlatScrInitLo, RegState::Kill)
96 .addImm(8);
97}
98
99unsigned SIFrameLowering::getReservedPrivateSegmentBufferReg(
100 const SISubtarget &ST,
101 const SIInstrInfo *TII,
102 const SIRegisterInfo *TRI,
103 SIMachineFunctionInfo *MFI,
104 MachineFunction &MF) const {
Matt Arsenaulte2218492017-04-24 21:08:32 +0000105 MachineRegisterInfo &MRI = MF.getRegInfo();
Matt Arsenault57bc4322016-08-31 21:52:21 +0000106
107 // We need to insert initialization of the scratch resource descriptor.
108 unsigned ScratchRsrcReg = MFI->getScratchRSrcReg();
Matt Arsenaulte2218492017-04-24 21:08:32 +0000109 if (ScratchRsrcReg == AMDGPU::NoRegister ||
110 !MRI.isPhysRegUsed(ScratchRsrcReg))
Matt Arsenault08906a32016-10-28 19:43:31 +0000111 return AMDGPU::NoRegister;
Matt Arsenault57bc4322016-08-31 21:52:21 +0000112
113 if (ST.hasSGPRInitBug() ||
114 ScratchRsrcReg != TRI->reservedPrivateSegmentBufferReg(MF))
115 return ScratchRsrcReg;
116
117 // We reserved the last registers for this. Shift it down to the end of those
118 // which were actually used.
119 //
120 // FIXME: It might be safer to use a pseudoregister before replacement.
121
122 // FIXME: We should be able to eliminate unused input registers. We only
123 // cannot do this for the resources required for scratch access. For now we
124 // skip over user SGPRs and may leave unused holes.
125
126 // We find the resource first because it has an alignment requirement.
127
Matt Arsenault08906a32016-10-28 19:43:31 +0000128 unsigned NumPreloaded = (MFI->getNumPreloadedSGPRs() + 3) / 4;
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +0000129 ArrayRef<MCPhysReg> AllSGPR128s = getAllSGPR128(ST, MF);
Matt Arsenault08906a32016-10-28 19:43:31 +0000130 AllSGPR128s = AllSGPR128s.slice(std::min(static_cast<unsigned>(AllSGPR128s.size()), NumPreloaded));
131
Matt Arsenaulte0bf7d02017-02-21 19:12:08 +0000132 // Skip the last N reserved elements because they should have already been
133 // reserved for VCC etc.
Matt Arsenault08906a32016-10-28 19:43:31 +0000134 for (MCPhysReg Reg : AllSGPR128s) {
Matt Arsenault57bc4322016-08-31 21:52:21 +0000135 // Pick the first unallocated one. Make sure we don't clobber the other
136 // reserved input we needed.
Matt Arsenault08906a32016-10-28 19:43:31 +0000137 if (!MRI.isPhysRegUsed(Reg) && MRI.isAllocatable(Reg)) {
Matt Arsenault57bc4322016-08-31 21:52:21 +0000138 MRI.replaceRegWith(ScratchRsrcReg, Reg);
139 MFI->setScratchRSrcReg(Reg);
140 return Reg;
141 }
142 }
143
144 return ScratchRsrcReg;
145}
146
Matt Arsenault36c31222017-04-25 23:40:57 +0000147// Shift down registers reserved for the scratch wave offset and stack pointer
148// SGPRs.
149std::pair<unsigned, unsigned>
150SIFrameLowering::getReservedPrivateSegmentWaveByteOffsetReg(
Matt Arsenault57bc4322016-08-31 21:52:21 +0000151 const SISubtarget &ST,
152 const SIInstrInfo *TII,
153 const SIRegisterInfo *TRI,
154 SIMachineFunctionInfo *MFI,
155 MachineFunction &MF) const {
Matt Arsenaulte2218492017-04-24 21:08:32 +0000156 MachineRegisterInfo &MRI = MF.getRegInfo();
Matt Arsenault57bc4322016-08-31 21:52:21 +0000157 unsigned ScratchWaveOffsetReg = MFI->getScratchWaveOffsetReg();
Matt Arsenaulte2218492017-04-24 21:08:32 +0000158
159 // No replacement necessary.
160 if (ScratchWaveOffsetReg == AMDGPU::NoRegister ||
Matt Arsenault36c31222017-04-25 23:40:57 +0000161 !MRI.isPhysRegUsed(ScratchWaveOffsetReg)) {
Matt Arsenault1cc47f82017-07-18 16:44:56 +0000162 assert(MFI->getStackPtrOffsetReg() == AMDGPU::SP_REG);
Matt Arsenault36c31222017-04-25 23:40:57 +0000163 return std::make_pair(AMDGPU::NoRegister, AMDGPU::NoRegister);
164 }
Matt Arsenaulte2218492017-04-24 21:08:32 +0000165
Matt Arsenault36c31222017-04-25 23:40:57 +0000166 unsigned SPReg = MFI->getStackPtrOffsetReg();
167 if (ST.hasSGPRInitBug())
168 return std::make_pair(ScratchWaveOffsetReg, SPReg);
Matt Arsenault57bc4322016-08-31 21:52:21 +0000169
Matt Arsenault57bc4322016-08-31 21:52:21 +0000170 unsigned NumPreloaded = MFI->getNumPreloadedSGPRs();
171
Konstantin Zhuravlyove03b1d72017-02-08 13:02:33 +0000172 ArrayRef<MCPhysReg> AllSGPRs = getAllSGPRs(ST, MF);
Matt Arsenault08906a32016-10-28 19:43:31 +0000173 if (NumPreloaded > AllSGPRs.size())
Matt Arsenault36c31222017-04-25 23:40:57 +0000174 return std::make_pair(ScratchWaveOffsetReg, SPReg);
Matt Arsenault08906a32016-10-28 19:43:31 +0000175
176 AllSGPRs = AllSGPRs.slice(NumPreloaded);
177
Matt Arsenault57bc4322016-08-31 21:52:21 +0000178 // We need to drop register from the end of the list that we cannot use
179 // for the scratch wave offset.
180 // + 2 s102 and s103 do not exist on VI.
181 // + 2 for vcc
182 // + 2 for xnack_mask
183 // + 2 for flat_scratch
184 // + 4 for registers reserved for scratch resource register
185 // + 1 for register reserved for scratch wave offset. (By exluding this
186 // register from the list to consider, it means that when this
187 // register is being used for the scratch wave offset and there
188 // are no other free SGPRs, then the value will stay in this register.
Matt Arsenault36c31222017-04-25 23:40:57 +0000189 // + 1 if stack pointer is used.
Matt Arsenault57bc4322016-08-31 21:52:21 +0000190 // ----
Matt Arsenault36c31222017-04-25 23:40:57 +0000191 // 13 (+1)
192 unsigned ReservedRegCount = 13;
Matt Arsenault08906a32016-10-28 19:43:31 +0000193
Matt Arsenault36c31222017-04-25 23:40:57 +0000194 if (AllSGPRs.size() < ReservedRegCount)
195 return std::make_pair(ScratchWaveOffsetReg, SPReg);
196
197 bool HandledScratchWaveOffsetReg =
198 ScratchWaveOffsetReg != TRI->reservedPrivateSegmentWaveByteOffsetReg(MF);
199
200 for (MCPhysReg Reg : AllSGPRs.drop_back(ReservedRegCount)) {
Matt Arsenault57bc4322016-08-31 21:52:21 +0000201 // Pick the first unallocated SGPR. Be careful not to pick an alias of the
202 // scratch descriptor, since we haven’t added its uses yet.
Matt Arsenaulte2218492017-04-24 21:08:32 +0000203 if (!MRI.isPhysRegUsed(Reg) && MRI.isAllocatable(Reg)) {
Matt Arsenault36c31222017-04-25 23:40:57 +0000204 if (!HandledScratchWaveOffsetReg) {
205 HandledScratchWaveOffsetReg = true;
206
207 MRI.replaceRegWith(ScratchWaveOffsetReg, Reg);
208 MFI->setScratchWaveOffsetReg(Reg);
209 ScratchWaveOffsetReg = Reg;
Matt Arsenault36c31222017-04-25 23:40:57 +0000210 break;
211 }
Matt Arsenault57bc4322016-08-31 21:52:21 +0000212 }
213 }
214
Matt Arsenault36c31222017-04-25 23:40:57 +0000215 return std::make_pair(ScratchWaveOffsetReg, SPReg);
Matt Arsenault57bc4322016-08-31 21:52:21 +0000216}
217
Matt Arsenault2b1f9aa2017-05-17 21:56:25 +0000218void SIFrameLowering::emitEntryFunctionPrologue(MachineFunction &MF,
219 MachineBasicBlock &MBB) const {
Konstantin Zhuravlyovf2f3d142016-06-25 03:11:28 +0000220 // Emit debugger prologue if "amdgpu-debugger-emit-prologue" attribute was
221 // specified.
222 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
223 if (ST.debuggerEmitPrologue())
224 emitDebuggerPrologue(MF, MBB);
225
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000226 assert(&MF.front() == &MBB && "Shrink-wrapping not yet supported");
227
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000228 SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000229
230 // If we only have SGPR spills, we won't actually be using scratch memory
231 // since these spill to VGPRs.
232 //
233 // FIXME: We should be cleaning up these unused SGPR spill frame indices
234 // somewhere.
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000235
Matt Arsenault43e92fe2016-06-24 06:30:11 +0000236 const SIInstrInfo *TII = ST.getInstrInfo();
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000237 const SIRegisterInfo *TRI = &TII->getRegisterInfo();
Matt Arsenault296b8492016-02-12 06:31:30 +0000238 MachineRegisterInfo &MRI = MF.getRegInfo();
Matt Arsenault57bc4322016-08-31 21:52:21 +0000239
Matt Arsenault08906a32016-10-28 19:43:31 +0000240 // We need to do the replacement of the private segment buffer and wave offset
241 // register even if there are no stack objects. There could be stores to undef
242 // or a constant without an associated object.
243
244 // FIXME: We still have implicit uses on SGPR spill instructions in case they
245 // need to spill to vector memory. It's likely that will not happen, but at
246 // this point it appears we need the setup. This part of the prolog should be
247 // emitted after frame indices are eliminated.
248
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000249 if (MFI->hasFlatScratchInit())
Matt Arsenaulte823d922017-02-18 18:29:53 +0000250 emitFlatScratchInit(ST, MF, MBB);
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000251
Matt Arsenault36c31222017-04-25 23:40:57 +0000252 unsigned SPReg = MFI->getStackPtrOffsetReg();
Matt Arsenault1cc47f82017-07-18 16:44:56 +0000253 if (SPReg != AMDGPU::SP_REG) {
254 assert(MRI.isReserved(SPReg) && "SPReg used but not reserved");
255
Matt Arsenault36c31222017-04-25 23:40:57 +0000256 DebugLoc DL;
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000257 const MachineFrameInfo &FrameInfo = MF.getFrameInfo();
258 int64_t StackSize = FrameInfo.getStackSize();
Matt Arsenault36c31222017-04-25 23:40:57 +0000259
260 if (StackSize == 0) {
261 BuildMI(MBB, MBB.begin(), DL, TII->get(AMDGPU::COPY), SPReg)
262 .addReg(MFI->getScratchWaveOffsetReg());
263 } else {
264 BuildMI(MBB, MBB.begin(), DL, TII->get(AMDGPU::S_ADD_U32), SPReg)
265 .addReg(MFI->getScratchWaveOffsetReg())
266 .addImm(StackSize * ST.getWavefrontSize());
267 }
268 }
269
Matt Arsenaulte2218492017-04-24 21:08:32 +0000270 unsigned ScratchRsrcReg
271 = getReservedPrivateSegmentBufferReg(ST, TII, TRI, MFI, MF);
Matt Arsenault36c31222017-04-25 23:40:57 +0000272
273 unsigned ScratchWaveOffsetReg;
274 std::tie(ScratchWaveOffsetReg, SPReg)
Matt Arsenaulte2218492017-04-24 21:08:32 +0000275 = getReservedPrivateSegmentWaveByteOffsetReg(ST, TII, TRI, MFI, MF);
276
277 // It's possible to have uses of only ScratchWaveOffsetReg without
278 // ScratchRsrcReg if it's only used for the initialization of flat_scratch,
279 // but the inverse is not true.
280 if (ScratchWaveOffsetReg == AMDGPU::NoRegister) {
281 assert(ScratchRsrcReg == AMDGPU::NoRegister);
282 return;
283 }
284
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000285 // We need to insert initialization of the scratch resource descriptor.
Matt Arsenault8623e8d2017-08-03 23:00:29 +0000286 unsigned PreloadedScratchWaveOffsetReg = MFI->getPreloadedReg(
287 AMDGPUFunctionArgInfo::PRIVATE_SEGMENT_WAVE_BYTE_OFFSET);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000288
289 unsigned PreloadedPrivateBufferReg = AMDGPU::NoRegister;
Matt Arsenault10fc0622017-06-26 03:01:31 +0000290 if (ST.isAmdCodeObjectV2(MF)) {
Matt Arsenault8623e8d2017-08-03 23:00:29 +0000291 PreloadedPrivateBufferReg = MFI->getPreloadedReg(
292 AMDGPUFunctionArgInfo::PRIVATE_SEGMENT_BUFFER);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000293 }
294
Matt Arsenaulte2218492017-04-24 21:08:32 +0000295 bool OffsetRegUsed = MRI.isPhysRegUsed(ScratchWaveOffsetReg);
296 bool ResourceRegUsed = ScratchRsrcReg != AMDGPU::NoRegister &&
297 MRI.isPhysRegUsed(ScratchRsrcReg);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000298
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000299 // We added live-ins during argument lowering, but since they were not used
300 // they were deleted. We're adding the uses now, so add them back.
Matt Arsenault08906a32016-10-28 19:43:31 +0000301 if (OffsetRegUsed) {
302 assert(PreloadedScratchWaveOffsetReg != AMDGPU::NoRegister &&
303 "scratch wave offset input is required");
304 MRI.addLiveIn(PreloadedScratchWaveOffsetReg);
305 MBB.addLiveIn(PreloadedScratchWaveOffsetReg);
306 }
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000307
Matt Arsenault08906a32016-10-28 19:43:31 +0000308 if (ResourceRegUsed && PreloadedPrivateBufferReg != AMDGPU::NoRegister) {
Tom Stellard2f3f9852017-01-25 01:25:13 +0000309 assert(ST.isAmdCodeObjectV2(MF) || ST.isMesaGfxShader(MF));
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000310 MRI.addLiveIn(PreloadedPrivateBufferReg);
311 MBB.addLiveIn(PreloadedPrivateBufferReg);
312 }
313
Matt Arsenault57bc4322016-08-31 21:52:21 +0000314 // Make the register selected live throughout the function.
315 for (MachineBasicBlock &OtherBB : MF) {
316 if (&OtherBB == &MBB)
317 continue;
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000318
Matt Arsenault08906a32016-10-28 19:43:31 +0000319 if (OffsetRegUsed)
320 OtherBB.addLiveIn(ScratchWaveOffsetReg);
321
322 if (ResourceRegUsed)
323 OtherBB.addLiveIn(ScratchRsrcReg);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000324 }
325
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000326 DebugLoc DL;
Matt Arsenault57bc4322016-08-31 21:52:21 +0000327 MachineBasicBlock::iterator I = MBB.begin();
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000328
Matt Arsenault08906a32016-10-28 19:43:31 +0000329 // If we reserved the original input registers, we don't need to copy to the
330 // reserved registers.
331
332 bool CopyBuffer = ResourceRegUsed &&
333 PreloadedPrivateBufferReg != AMDGPU::NoRegister &&
Tom Stellard2f3f9852017-01-25 01:25:13 +0000334 ST.isAmdCodeObjectV2(MF) &&
Matt Arsenault08906a32016-10-28 19:43:31 +0000335 ScratchRsrcReg != PreloadedPrivateBufferReg;
336
337 // This needs to be careful of the copying order to avoid overwriting one of
338 // the input registers before it's been copied to it's final
339 // destination. Usually the offset should be copied first.
340 bool CopyBufferFirst = TRI->isSubRegisterEq(PreloadedPrivateBufferReg,
341 ScratchWaveOffsetReg);
342 if (CopyBuffer && CopyBufferFirst) {
343 BuildMI(MBB, I, DL, TII->get(AMDGPU::COPY), ScratchRsrcReg)
344 .addReg(PreloadedPrivateBufferReg, RegState::Kill);
345 }
346
347 if (OffsetRegUsed &&
348 PreloadedScratchWaveOffsetReg != ScratchWaveOffsetReg) {
Matt Arsenault1d215172016-08-31 21:52:25 +0000349 BuildMI(MBB, I, DL, TII->get(AMDGPU::COPY), ScratchWaveOffsetReg)
Marek Olsak584d2c02017-05-04 22:25:20 +0000350 .addReg(PreloadedScratchWaveOffsetReg,
351 MRI.isPhysRegUsed(ScratchWaveOffsetReg) ? 0 : RegState::Kill);
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000352 }
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000353
Matt Arsenault08906a32016-10-28 19:43:31 +0000354 if (CopyBuffer && !CopyBufferFirst) {
Matt Arsenault1d215172016-08-31 21:52:25 +0000355 BuildMI(MBB, I, DL, TII->get(AMDGPU::COPY), ScratchRsrcReg)
356 .addReg(PreloadedPrivateBufferReg, RegState::Kill);
Matt Arsenault08906a32016-10-28 19:43:31 +0000357 }
358
Tim Renouf13229152017-09-29 09:49:35 +0000359 if (ResourceRegUsed)
360 emitEntryFunctionScratchSetup(ST, MF, MBB, MFI, I,
361 PreloadedPrivateBufferReg, ScratchRsrcReg);
362}
363
364// Emit scratch setup code for AMDPAL or Mesa, assuming ResourceRegUsed is set.
365void SIFrameLowering::emitEntryFunctionScratchSetup(const SISubtarget &ST,
366 MachineFunction &MF, MachineBasicBlock &MBB, SIMachineFunctionInfo *MFI,
367 MachineBasicBlock::iterator I, unsigned PreloadedPrivateBufferReg,
368 unsigned ScratchRsrcReg) const {
369
370 const SIInstrInfo *TII = ST.getInstrInfo();
371 const SIRegisterInfo *TRI = &TII->getRegisterInfo();
372 DebugLoc DL;
Tim Renouf13229152017-09-29 09:49:35 +0000373
374 if (ST.isAmdPalOS()) {
375 // The pointer to the GIT is formed from the offset passed in and either
376 // the amdgpu-git-ptr-high function attribute or the top part of the PC
377 unsigned RsrcLo = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub0);
378 unsigned RsrcHi = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub1);
379 unsigned Rsrc01 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub0_sub1);
380
381 const MCInstrDesc &SMovB32 = TII->get(AMDGPU::S_MOV_B32);
382
383 if (MFI->getGITPtrHigh() != 0xffffffff) {
384 BuildMI(MBB, I, DL, SMovB32, RsrcHi)
385 .addImm(MFI->getGITPtrHigh())
386 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
387 } else {
388 const MCInstrDesc &GetPC64 = TII->get(AMDGPU::S_GETPC_B64);
389 BuildMI(MBB, I, DL, GetPC64, Rsrc01);
390 }
Tim Renouf832f90f2018-02-26 14:46:43 +0000391 auto GitPtrLo = AMDGPU::SGPR0; // Low GIT address passed in
392 if (ST.hasMergedShaders()) {
393 switch (MF.getFunction().getCallingConv()) {
394 case CallingConv::AMDGPU_HS:
395 case CallingConv::AMDGPU_GS:
396 // Low GIT address is passed in s8 rather than s0 for an LS+HS or
397 // ES+GS merged shader on gfx9+.
398 GitPtrLo = AMDGPU::SGPR8;
399 break;
400 default:
401 break;
402 }
403 }
Tim Renouf7190a462018-04-10 11:25:15 +0000404 MF.getRegInfo().addLiveIn(GitPtrLo);
405 MF.front().addLiveIn(GitPtrLo);
Tim Renouf13229152017-09-29 09:49:35 +0000406 BuildMI(MBB, I, DL, SMovB32, RsrcLo)
Tim Renouf832f90f2018-02-26 14:46:43 +0000407 .addReg(GitPtrLo)
Tim Renouf13229152017-09-29 09:49:35 +0000408 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
409
410 // We now have the GIT ptr - now get the scratch descriptor from the entry
Tim Renouf7190a462018-04-10 11:25:15 +0000411 // at offset 0 (or offset 16 for a compute shader).
Tim Renouf13229152017-09-29 09:49:35 +0000412 PointerType *PtrTy =
Matthias Braunf1caa282017-12-15 22:22:58 +0000413 PointerType::get(Type::getInt64Ty(MF.getFunction().getContext()),
Tim Renouf13229152017-09-29 09:49:35 +0000414 AMDGPUAS::CONSTANT_ADDRESS);
415 MachinePointerInfo PtrInfo(UndefValue::get(PtrTy));
416 const MCInstrDesc &LoadDwordX4 = TII->get(AMDGPU::S_LOAD_DWORDX4_IMM);
417 auto MMO = MF.getMachineMemOperand(PtrInfo,
418 MachineMemOperand::MOLoad |
419 MachineMemOperand::MOInvariant |
420 MachineMemOperand::MODereferenceable,
421 0, 0);
Tim Renouf7190a462018-04-10 11:25:15 +0000422 unsigned Offset
423 = MF.getFunction().getCallingConv() == CallingConv::AMDGPU_CS ? 16 : 0;
Tim Renouf13229152017-09-29 09:49:35 +0000424 BuildMI(MBB, I, DL, LoadDwordX4, ScratchRsrcReg)
425 .addReg(Rsrc01)
Tim Renouf7190a462018-04-10 11:25:15 +0000426 .addImm(Offset) // offset
Tim Renouf13229152017-09-29 09:49:35 +0000427 .addImm(0) // glc
428 .addReg(ScratchRsrcReg, RegState::ImplicitDefine)
429 .addMemOperand(MMO);
430 return;
431 }
432 if (ST.isMesaGfxShader(MF)
433 || (PreloadedPrivateBufferReg == AMDGPU::NoRegister)) {
Tom Stellard2f3f9852017-01-25 01:25:13 +0000434 assert(!ST.isAmdCodeObjectV2(MF));
Matt Arsenault1d215172016-08-31 21:52:25 +0000435 const MCInstrDesc &SMovB32 = TII->get(AMDGPU::S_MOV_B32);
436
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000437 unsigned Rsrc2 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub2);
438 unsigned Rsrc3 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub3);
439
440 // Use relocations to get the pointer, and setup the other bits manually.
441 uint64_t Rsrc23 = TII->getScratchRsrcWords23();
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000442
Matt Arsenault10fc0622017-06-26 03:01:31 +0000443 if (MFI->hasImplicitBufferPtr()) {
Tom Stellard2f3f9852017-01-25 01:25:13 +0000444 unsigned Rsrc01 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub0_sub1);
445
Matthias Braunf1caa282017-12-15 22:22:58 +0000446 if (AMDGPU::isCompute(MF.getFunction().getCallingConv())) {
Tom Stellard2f3f9852017-01-25 01:25:13 +0000447 const MCInstrDesc &Mov64 = TII->get(AMDGPU::S_MOV_B64);
448
449 BuildMI(MBB, I, DL, Mov64, Rsrc01)
Matt Arsenault10fc0622017-06-26 03:01:31 +0000450 .addReg(MFI->getImplicitBufferPtrUserSGPR())
Tom Stellard2f3f9852017-01-25 01:25:13 +0000451 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
452 } else {
453 const MCInstrDesc &LoadDwordX2 = TII->get(AMDGPU::S_LOAD_DWORDX2_IMM);
454
455 PointerType *PtrTy =
Matthias Braunf1caa282017-12-15 22:22:58 +0000456 PointerType::get(Type::getInt64Ty(MF.getFunction().getContext()),
Konstantin Zhuravlyov435151a2017-11-01 19:12:38 +0000457 AMDGPUAS::CONSTANT_ADDRESS);
Tom Stellard2f3f9852017-01-25 01:25:13 +0000458 MachinePointerInfo PtrInfo(UndefValue::get(PtrTy));
459 auto MMO = MF.getMachineMemOperand(PtrInfo,
460 MachineMemOperand::MOLoad |
461 MachineMemOperand::MOInvariant |
462 MachineMemOperand::MODereferenceable,
463 0, 0);
464 BuildMI(MBB, I, DL, LoadDwordX2, Rsrc01)
Matt Arsenault10fc0622017-06-26 03:01:31 +0000465 .addReg(MFI->getImplicitBufferPtrUserSGPR())
Tom Stellard2f3f9852017-01-25 01:25:13 +0000466 .addImm(0) // offset
467 .addImm(0) // glc
468 .addMemOperand(MMO)
469 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
470 }
471 } else {
472 unsigned Rsrc0 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub0);
473 unsigned Rsrc1 = TRI->getSubReg(ScratchRsrcReg, AMDGPU::sub1);
474
475 BuildMI(MBB, I, DL, SMovB32, Rsrc0)
476 .addExternalSymbol("SCRATCH_RSRC_DWORD0")
477 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
478
479 BuildMI(MBB, I, DL, SMovB32, Rsrc1)
480 .addExternalSymbol("SCRATCH_RSRC_DWORD1")
481 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
482
483 }
Matt Arsenault26f8f3d2015-11-30 21:16:03 +0000484
485 BuildMI(MBB, I, DL, SMovB32, Rsrc2)
486 .addImm(Rsrc23 & 0xffffffff)
487 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
488
489 BuildMI(MBB, I, DL, SMovB32, Rsrc3)
490 .addImm(Rsrc23 >> 32)
491 .addReg(ScratchRsrcReg, RegState::ImplicitDefine);
492 }
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000493}
494
Matt Arsenault03ae3992018-03-29 21:30:06 +0000495// Find a scratch register that we can use at the start of the prologue to
496// re-align the stack pointer. We avoid using callee-save registers since they
497// may appear to be free when this is called from canUseAsPrologue (during
498// shrink wrapping), but then no longer be free when this is called from
499// emitPrologue.
500//
501// FIXME: This is a bit conservative, since in the above case we could use one
502// of the callee-save registers as a scratch temp to re-align the stack pointer,
503// but we would then have to make sure that we were in fact saving at least one
504// callee-save register in the prologue, which is additional complexity that
505// doesn't seem worth the benefit.
506static unsigned findScratchNonCalleeSaveRegister(MachineBasicBlock &MBB) {
507 MachineFunction *MF = MBB.getParent();
508
509 const SISubtarget &Subtarget = MF->getSubtarget<SISubtarget>();
510 const SIRegisterInfo &TRI = *Subtarget.getRegisterInfo();
511 LivePhysRegs LiveRegs(TRI);
512 LiveRegs.addLiveIns(MBB);
513
514 // Mark callee saved registers as used so we will not choose them.
515 const MCPhysReg *CSRegs = TRI.getCalleeSavedRegs(MF);
516 for (unsigned i = 0; CSRegs[i]; ++i)
517 LiveRegs.addReg(CSRegs[i]);
518
519 MachineRegisterInfo &MRI = MF->getRegInfo();
520
521 for (unsigned Reg : AMDGPU::SReg_32_XM0RegClass) {
522 if (LiveRegs.available(MRI, Reg))
523 return Reg;
524 }
525
526 return AMDGPU::NoRegister;
527}
528
Matt Arsenault2b1f9aa2017-05-17 21:56:25 +0000529void SIFrameLowering::emitPrologue(MachineFunction &MF,
530 MachineBasicBlock &MBB) const {
Matt Arsenault03ae3992018-03-29 21:30:06 +0000531 SIMachineFunctionInfo *FuncInfo = MF.getInfo<SIMachineFunctionInfo>();
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000532 if (FuncInfo->isEntryFunction()) {
Matt Arsenault2b1f9aa2017-05-17 21:56:25 +0000533 emitEntryFunctionPrologue(MF, MBB);
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000534 return;
535 }
536
537 const MachineFrameInfo &MFI = MF.getFrameInfo();
538 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
539 const SIInstrInfo *TII = ST.getInstrInfo();
Matt Arsenault03ae3992018-03-29 21:30:06 +0000540 const SIRegisterInfo &TRI = TII->getRegisterInfo();
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000541
542 unsigned StackPtrReg = FuncInfo->getStackPtrOffsetReg();
543 unsigned FramePtrReg = FuncInfo->getFrameOffsetReg();
544
545 MachineBasicBlock::iterator MBBI = MBB.begin();
546 DebugLoc DL;
547
Matt Arsenault03ae3992018-03-29 21:30:06 +0000548 // XXX - Is this the right predicate?
549
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000550 bool NeedFP = hasFP(MF);
Matt Arsenault03ae3992018-03-29 21:30:06 +0000551 uint32_t NumBytes = MFI.getStackSize();
552 uint32_t RoundedSize = NumBytes;
553 const bool NeedsRealignment = TRI.needsStackRealignment(MF);
554
555 if (NeedsRealignment) {
556 assert(NeedFP);
557 const unsigned Alignment = MFI.getMaxAlignment();
Matt Arsenault03ae3992018-03-29 21:30:06 +0000558
559 RoundedSize += Alignment;
560
561 unsigned ScratchSPReg = findScratchNonCalleeSaveRegister(MBB);
562 assert(ScratchSPReg != AMDGPU::NoRegister);
563
564 // s_add_u32 tmp_reg, s32, NumBytes
565 // s_and_b32 s32, tmp_reg, 0b111...0000
566 BuildMI(MBB, MBBI, DL, TII->get(AMDGPU::S_ADD_U32), ScratchSPReg)
567 .addReg(StackPtrReg)
568 .addImm((Alignment - 1) * ST.getWavefrontSize())
569 .setMIFlag(MachineInstr::FrameSetup);
570 BuildMI(MBB, MBBI, DL, TII->get(AMDGPU::S_AND_B32), FramePtrReg)
571 .addReg(ScratchSPReg, RegState::Kill)
572 .addImm(-Alignment * ST.getWavefrontSize())
573 .setMIFlag(MachineInstr::FrameSetup);
574 FuncInfo->setIsStackRealigned(true);
575 } else if (NeedFP) {
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000576 // If we need a base pointer, set it up here. It's whatever the value of
577 // the stack pointer is at this point. Any variable size objects will be
578 // allocated after this, so we can still use the base pointer to reference
579 // locals.
580 BuildMI(MBB, MBBI, DL, TII->get(AMDGPU::COPY), FramePtrReg)
581 .addReg(StackPtrReg)
582 .setMIFlag(MachineInstr::FrameSetup);
583 }
584
Matt Arsenault03ae3992018-03-29 21:30:06 +0000585 if (RoundedSize != 0 && hasSP(MF)) {
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000586 BuildMI(MBB, MBBI, DL, TII->get(AMDGPU::S_ADD_U32), StackPtrReg)
587 .addReg(StackPtrReg)
Matt Arsenault03ae3992018-03-29 21:30:06 +0000588 .addImm(RoundedSize * ST.getWavefrontSize())
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000589 .setMIFlag(MachineInstr::FrameSetup);
590 }
Matt Arsenault8e8f8f42017-08-02 01:52:45 +0000591
592 for (const SIMachineFunctionInfo::SGPRSpillVGPRCSR &Reg
593 : FuncInfo->getSGPRSpillVGPRs()) {
594 if (!Reg.FI.hasValue())
595 continue;
596 TII->storeRegToStackSlot(MBB, MBBI, Reg.VGPR, true,
597 Reg.FI.getValue(), &AMDGPU::VGPR_32RegClass,
598 &TII->getRegisterInfo());
599 }
Matt Arsenault2b1f9aa2017-05-17 21:56:25 +0000600}
601
Matt Arsenault43e92fe2016-06-24 06:30:11 +0000602void SIFrameLowering::emitEpilogue(MachineFunction &MF,
603 MachineBasicBlock &MBB) const {
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000604 const SIMachineFunctionInfo *FuncInfo = MF.getInfo<SIMachineFunctionInfo>();
605 if (FuncInfo->isEntryFunction())
606 return;
Matt Arsenault43e92fe2016-06-24 06:30:11 +0000607
Matt Arsenault8e8f8f42017-08-02 01:52:45 +0000608 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
609 const SIInstrInfo *TII = ST.getInstrInfo();
610 MachineBasicBlock::iterator MBBI = MBB.getFirstTerminator();
611
612 for (const SIMachineFunctionInfo::SGPRSpillVGPRCSR &Reg
613 : FuncInfo->getSGPRSpillVGPRs()) {
614 if (!Reg.FI.hasValue())
615 continue;
616 TII->loadRegFromStackSlot(MBB, MBBI, Reg.VGPR,
617 Reg.FI.getValue(), &AMDGPU::VGPR_32RegClass,
618 &TII->getRegisterInfo());
619 }
620
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000621 unsigned StackPtrReg = FuncInfo->getStackPtrOffsetReg();
622 if (StackPtrReg == AMDGPU::NoRegister)
623 return;
624
625 const MachineFrameInfo &MFI = MF.getFrameInfo();
626 uint32_t NumBytes = MFI.getStackSize();
627
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000628 DebugLoc DL;
629
630 // FIXME: Clarify distinction between no set SP and SP. For callee functions,
631 // it's really whether we need SP to be accurate or not.
632
633 if (NumBytes != 0 && hasSP(MF)) {
Matt Arsenault03ae3992018-03-29 21:30:06 +0000634 uint32_t RoundedSize = FuncInfo->isStackRealigned() ?
635 NumBytes + MFI.getMaxAlignment() : NumBytes;
636
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000637 BuildMI(MBB, MBBI, DL, TII->get(AMDGPU::S_SUB_U32), StackPtrReg)
638 .addReg(StackPtrReg)
Matt Arsenault03ae3992018-03-29 21:30:06 +0000639 .addImm(RoundedSize * ST.getWavefrontSize());
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000640 }
Matt Arsenault43e92fe2016-06-24 06:30:11 +0000641}
642
Matt Arsenault7b6c5d22017-02-22 22:23:32 +0000643static bool allStackObjectsAreDead(const MachineFrameInfo &MFI) {
644 for (int I = MFI.getObjectIndexBegin(), E = MFI.getObjectIndexEnd();
645 I != E; ++I) {
646 if (!MFI.isDeadObjectIndex(I))
647 return false;
648 }
649
650 return true;
651}
652
Konstantin Zhuravlyovffdb00e2017-03-10 19:39:07 +0000653int SIFrameLowering::getFrameIndexReference(const MachineFunction &MF, int FI,
654 unsigned &FrameReg) const {
655 const SIRegisterInfo *RI = MF.getSubtarget<SISubtarget>().getRegisterInfo();
656
657 FrameReg = RI->getFrameRegister(MF);
658 return MF.getFrameInfo().getObjectOffset(FI);
659}
660
Matt Arsenault0c90e952015-11-06 18:17:45 +0000661void SIFrameLowering::processFunctionBeforeFrameFinalized(
662 MachineFunction &MF,
663 RegScavenger *RS) const {
Matthias Braun941a7052016-07-28 18:40:00 +0000664 MachineFrameInfo &MFI = MF.getFrameInfo();
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000665
Matthias Braun941a7052016-07-28 18:40:00 +0000666 if (!MFI.hasStackObjects())
Matt Arsenault0e3d3892015-11-30 21:15:53 +0000667 return;
668
Matt Arsenault7b6c5d22017-02-22 22:23:32 +0000669 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
670 const SIInstrInfo *TII = ST.getInstrInfo();
671 const SIRegisterInfo &TRI = TII->getRegisterInfo();
672 SIMachineFunctionInfo *FuncInfo = MF.getInfo<SIMachineFunctionInfo>();
673 bool AllSGPRSpilledToVGPRs = false;
674
675 if (TRI.spillSGPRToVGPR() && FuncInfo->hasSpilledSGPRs()) {
676 AllSGPRSpilledToVGPRs = true;
677
678 // Process all SGPR spills before frame offsets are finalized. Ideally SGPRs
679 // are spilled to VGPRs, in which case we can eliminate the stack usage.
680 //
681 // XXX - This operates under the assumption that only other SGPR spills are
682 // users of the frame index. I'm not 100% sure this is correct. The
683 // StackColoring pass has a comment saying a future improvement would be to
684 // merging of allocas with spill slots, but for now according to
685 // MachineFrameInfo isSpillSlot can't alias any other object.
686 for (MachineBasicBlock &MBB : MF) {
687 MachineBasicBlock::iterator Next;
688 for (auto I = MBB.begin(), E = MBB.end(); I != E; I = Next) {
689 MachineInstr &MI = *I;
690 Next = std::next(I);
691
692 if (TII->isSGPRSpill(MI)) {
693 int FI = TII->getNamedOperand(MI, AMDGPU::OpName::addr)->getIndex();
694 if (FuncInfo->allocateSGPRSpillToVGPR(MF, FI)) {
695 bool Spilled = TRI.eliminateSGPRToVGPRSpillFrameIndex(MI, FI, RS);
696 (void)Spilled;
697 assert(Spilled && "failed to spill SGPR to VGPR when allocated");
698 } else
699 AllSGPRSpilledToVGPRs = false;
700 }
701 }
702 }
703
704 FuncInfo->removeSGPRToVGPRFrameIndices(MFI);
705 }
706
707 // FIXME: The other checks should be redundant with allStackObjectsAreDead,
708 // but currently hasNonSpillStackObjects is set only from source
709 // allocas. Stack temps produced from legalization are not counted currently.
710 if (FuncInfo->hasNonSpillStackObjects() || FuncInfo->hasSpilledVGPRs() ||
711 !AllSGPRSpilledToVGPRs || !allStackObjectsAreDead(MFI)) {
712 assert(RS && "RegScavenger required if spilling");
713
Matt Arsenault707780b2017-02-22 21:05:25 +0000714 // We force this to be at offset 0 so no user object ever has 0 as an
715 // address, so we may use 0 as an invalid pointer value. This is because
716 // LLVM assumes 0 is an invalid pointer in address space 0. Because alloca
717 // is required to be address space 0, we are forced to accept this for
718 // now. Ideally we could have the stack in another address space with 0 as a
719 // valid pointer, and -1 as the null value.
720 //
721 // This will also waste additional space when user stack objects require > 4
722 // byte alignment.
723 //
724 // The main cost here is losing the offset for addressing modes. However
725 // this also ensures we shouldn't need a register for the offset when
726 // emergency scavenging.
727 int ScavengeFI = MFI.CreateFixedObject(
Krzysztof Parzyszek44e25f32017-04-24 18:55:33 +0000728 TRI.getSpillSize(AMDGPU::SGPR_32RegClass), 0, false);
Matt Arsenault707780b2017-02-22 21:05:25 +0000729 RS->addScavengingFrameIndex(ScavengeFI);
730 }
Matt Arsenault0c90e952015-11-06 18:17:45 +0000731}
Konstantin Zhuravlyovf2f3d142016-06-25 03:11:28 +0000732
Matt Arsenaultecb43ef2017-09-13 23:47:01 +0000733void SIFrameLowering::determineCalleeSaves(MachineFunction &MF, BitVector &SavedRegs,
734 RegScavenger *RS) const {
735 TargetFrameLowering::determineCalleeSaves(MF, SavedRegs, RS);
736 const SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
737
738 // The SP is specifically managed and we don't want extra spills of it.
739 SavedRegs.reset(MFI->getStackPtrOffsetReg());
740}
741
Matt Arsenaultb62a4eb2017-08-01 19:54:18 +0000742MachineBasicBlock::iterator SIFrameLowering::eliminateCallFramePseudoInstr(
743 MachineFunction &MF,
744 MachineBasicBlock &MBB,
745 MachineBasicBlock::iterator I) const {
746 int64_t Amount = I->getOperand(0).getImm();
747 if (Amount == 0)
748 return MBB.erase(I);
749
750 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
751 const SIInstrInfo *TII = ST.getInstrInfo();
752 const DebugLoc &DL = I->getDebugLoc();
753 unsigned Opc = I->getOpcode();
754 bool IsDestroy = Opc == TII->getCallFrameDestroyOpcode();
755 uint64_t CalleePopAmount = IsDestroy ? I->getOperand(1).getImm() : 0;
756
757 const TargetFrameLowering *TFI = MF.getSubtarget().getFrameLowering();
758 if (!TFI->hasReservedCallFrame(MF)) {
759 unsigned Align = getStackAlignment();
760
761 Amount = alignTo(Amount, Align);
762 assert(isUInt<32>(Amount) && "exceeded stack address space size");
763 const SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
764 unsigned SPReg = MFI->getStackPtrOffsetReg();
765
766 unsigned Op = IsDestroy ? AMDGPU::S_SUB_U32 : AMDGPU::S_ADD_U32;
767 BuildMI(MBB, I, DL, TII->get(Op), SPReg)
768 .addReg(SPReg)
769 .addImm(Amount * ST.getWavefrontSize());
770 } else if (CalleePopAmount != 0) {
771 llvm_unreachable("is this used?");
772 }
773
774 return MBB.erase(I);
775}
776
Konstantin Zhuravlyovf2f3d142016-06-25 03:11:28 +0000777void SIFrameLowering::emitDebuggerPrologue(MachineFunction &MF,
778 MachineBasicBlock &MBB) const {
779 const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
780 const SIInstrInfo *TII = ST.getInstrInfo();
781 const SIRegisterInfo *TRI = &TII->getRegisterInfo();
782 const SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
783
784 MachineBasicBlock::iterator I = MBB.begin();
785 DebugLoc DL;
786
787 // For each dimension:
788 for (unsigned i = 0; i < 3; ++i) {
789 // Get work group ID SGPR, and make it live-in again.
790 unsigned WorkGroupIDSGPR = MFI->getWorkGroupIDSGPR(i);
791 MF.getRegInfo().addLiveIn(WorkGroupIDSGPR);
792 MBB.addLiveIn(WorkGroupIDSGPR);
793
794 // Since SGPRs are spilled into VGPRs, copy work group ID SGPR to VGPR in
795 // order to spill it to scratch.
796 unsigned WorkGroupIDVGPR =
797 MF.getRegInfo().createVirtualRegister(&AMDGPU::VGPR_32RegClass);
798 BuildMI(MBB, I, DL, TII->get(AMDGPU::V_MOV_B32_e32), WorkGroupIDVGPR)
799 .addReg(WorkGroupIDSGPR);
800
801 // Spill work group ID.
802 int WorkGroupIDObjectIdx = MFI->getDebuggerWorkGroupIDStackObjectIndex(i);
803 TII->storeRegToStackSlot(MBB, I, WorkGroupIDVGPR, false,
804 WorkGroupIDObjectIdx, &AMDGPU::VGPR_32RegClass, TRI);
805
806 // Get work item ID VGPR, and make it live-in again.
807 unsigned WorkItemIDVGPR = MFI->getWorkItemIDVGPR(i);
808 MF.getRegInfo().addLiveIn(WorkItemIDVGPR);
809 MBB.addLiveIn(WorkItemIDVGPR);
810
811 // Spill work item ID.
812 int WorkItemIDObjectIdx = MFI->getDebuggerWorkItemIDStackObjectIndex(i);
813 TII->storeRegToStackSlot(MBB, I, WorkItemIDVGPR, false,
814 WorkItemIDObjectIdx, &AMDGPU::VGPR_32RegClass, TRI);
815 }
816}
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000817
818bool SIFrameLowering::hasFP(const MachineFunction &MF) const {
819 // All stack operations are relative to the frame offset SGPR.
820 // TODO: Still want to eliminate sometimes.
821 const MachineFrameInfo &MFI = MF.getFrameInfo();
822
823 // XXX - Is this only called after frame is finalized? Should be able to check
824 // frame size.
825 return MFI.hasStackObjects() && !allStackObjectsAreDead(MFI);
826}
827
828bool SIFrameLowering::hasSP(const MachineFunction &MF) const {
Matt Arsenault03ae3992018-03-29 21:30:06 +0000829 const SIRegisterInfo *TRI = MF.getSubtarget<SISubtarget>().getRegisterInfo();
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000830 // All stack operations are relative to the frame offset SGPR.
831 const MachineFrameInfo &MFI = MF.getFrameInfo();
Matt Arsenault03ae3992018-03-29 21:30:06 +0000832 return MFI.hasCalls() || MFI.hasVarSizedObjects() || TRI->needsStackRealignment(MF);
Matt Arsenaultf28683c2017-06-26 17:53:59 +0000833}