blob: 528a773ef2179ac0dc6936885ec9aaa0315457dd [file] [log] [blame]
Tom Stellardd8ea85a2016-12-21 19:06:24 +00001//===-- llvm/lib/Target/AMDGPU/AMDGPUCallLowering.cpp - Call lowering -----===//
Tom Stellard000c5af2016-04-14 19:09:28 +00002//
Chandler Carruth2946cd72019-01-19 08:50:56 +00003// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
Tom Stellard000c5af2016-04-14 19:09:28 +00006//
7//===----------------------------------------------------------------------===//
8///
9/// \file
10/// This file implements the lowering of LLVM calls to machine code calls for
11/// GlobalISel.
12///
13//===----------------------------------------------------------------------===//
14
15#include "AMDGPUCallLowering.h"
Tom Stellardca166212017-01-30 21:56:46 +000016#include "AMDGPU.h"
Tom Stellard000c5af2016-04-14 19:09:28 +000017#include "AMDGPUISelLowering.h"
Tom Stellardca166212017-01-30 21:56:46 +000018#include "AMDGPUSubtarget.h"
19#include "SIISelLowering.h"
Tom Stellardca166212017-01-30 21:56:46 +000020#include "SIMachineFunctionInfo.h"
Chandler Carruth6bda14b2017-06-06 11:49:48 +000021#include "SIRegisterInfo.h"
Tom Stellard44b30b42018-05-22 02:03:23 +000022#include "MCTargetDesc/AMDGPUMCTargetDesc.h"
Tom Stellard206b9922019-04-09 02:26:03 +000023#include "llvm/CodeGen/Analysis.h"
Tom Stellardca166212017-01-30 21:56:46 +000024#include "llvm/CodeGen/CallingConvLower.h"
Tom Stellard000c5af2016-04-14 19:09:28 +000025#include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h"
26#include "llvm/CodeGen/MachineInstrBuilder.h"
Tom Stellard206b9922019-04-09 02:26:03 +000027#include "llvm/Support/LowLevelTypeImpl.h"
Tom Stellard000c5af2016-04-14 19:09:28 +000028
29using namespace llvm;
30
Tom Stellard206b9922019-04-09 02:26:03 +000031namespace {
32
33struct OutgoingArgHandler : public CallLowering::ValueHandler {
34 OutgoingArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI,
35 MachineInstrBuilder MIB, CCAssignFn *AssignFn)
36 : ValueHandler(MIRBuilder, MRI, AssignFn), MIB(MIB) {}
37
38 MachineInstrBuilder MIB;
39
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +000040 Register getStackAddress(uint64_t Size, int64_t Offset,
Tom Stellard206b9922019-04-09 02:26:03 +000041 MachinePointerInfo &MPO) override {
42 llvm_unreachable("not implemented");
43 }
44
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +000045 void assignValueToAddress(Register ValVReg, Register Addr, uint64_t Size,
Tom Stellard206b9922019-04-09 02:26:03 +000046 MachinePointerInfo &MPO, CCValAssign &VA) override {
47 llvm_unreachable("not implemented");
48 }
49
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +000050 void assignValueToReg(Register ValVReg, Register PhysReg,
Tom Stellard206b9922019-04-09 02:26:03 +000051 CCValAssign &VA) override {
52 MIB.addUse(PhysReg);
53 MIRBuilder.buildCopy(PhysReg, ValVReg);
54 }
55
56 bool assignArg(unsigned ValNo, MVT ValVT, MVT LocVT,
57 CCValAssign::LocInfo LocInfo,
58 const CallLowering::ArgInfo &Info,
59 CCState &State) override {
60 return AssignFn(ValNo, ValVT, LocVT, LocInfo, Info.Flags, State);
61 }
62};
63
64}
65
Tom Stellard000c5af2016-04-14 19:09:28 +000066AMDGPUCallLowering::AMDGPUCallLowering(const AMDGPUTargetLowering &TLI)
Matt Arsenault0da63502018-08-31 05:49:54 +000067 : CallLowering(&TLI) {
Tom Stellard000c5af2016-04-14 19:09:28 +000068}
69
70bool AMDGPUCallLowering::lowerReturn(MachineIRBuilder &MIRBuilder,
Alexander Ivchenko49168f62018-08-02 08:33:31 +000071 const Value *Val,
Matt Arsenaulte3a676e2019-06-24 15:50:29 +000072 ArrayRef<Register> VRegs) const {
Tom Stellard206b9922019-04-09 02:26:03 +000073
74 MachineFunction &MF = MIRBuilder.getMF();
75 MachineRegisterInfo &MRI = MF.getRegInfo();
76 SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
77 MFI->setIfReturnsVoid(!Val);
78
79 if (!Val) {
80 MIRBuilder.buildInstr(AMDGPU::S_ENDPGM).addImm(0);
81 return true;
82 }
83
Matt Arsenaulte3a676e2019-06-24 15:50:29 +000084 Register VReg = VRegs[0];
Tom Stellard206b9922019-04-09 02:26:03 +000085
86 const Function &F = MF.getFunction();
87 auto &DL = F.getParent()->getDataLayout();
88 if (!AMDGPU::isShader(F.getCallingConv()))
Tom Stellard257882f2018-04-24 21:29:36 +000089 return false;
90
Tom Stellard206b9922019-04-09 02:26:03 +000091
92 const AMDGPUTargetLowering &TLI = *getTLI<AMDGPUTargetLowering>();
93 SmallVector<EVT, 4> SplitVTs;
94 SmallVector<uint64_t, 4> Offsets;
95 ArgInfo OrigArg{VReg, Val->getType()};
96 setArgFlags(OrigArg, AttributeList::ReturnIndex, DL, F);
97 ComputeValueVTs(TLI, DL, OrigArg.Ty, SplitVTs, &Offsets, 0);
98
99 SmallVector<ArgInfo, 8> SplitArgs;
100 CCAssignFn *AssignFn = CCAssignFnForReturn(F.getCallingConv(), false);
101 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) {
102 Type *SplitTy = SplitVTs[i].getTypeForEVT(F.getContext());
103 SplitArgs.push_back({VRegs[i], SplitTy, OrigArg.Flags, OrigArg.IsFixed});
104 }
105 auto RetInstr = MIRBuilder.buildInstrNoInsert(AMDGPU::SI_RETURN_TO_EPILOG);
106 OutgoingArgHandler Handler(MIRBuilder, MRI, RetInstr, AssignFn);
107 if (!handleAssignments(MIRBuilder, SplitArgs, Handler))
108 return false;
109 MIRBuilder.insertInstr(RetInstr);
110
Tom Stellard000c5af2016-04-14 19:09:28 +0000111 return true;
112}
113
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000114Register AMDGPUCallLowering::lowerParameterPtr(MachineIRBuilder &MIRBuilder,
Tom Stellardca166212017-01-30 21:56:46 +0000115 Type *ParamTy,
Matt Arsenault29f30372018-07-05 17:01:20 +0000116 uint64_t Offset) const {
Tom Stellardca166212017-01-30 21:56:46 +0000117
118 MachineFunction &MF = MIRBuilder.getMF();
Matt Arsenault8623e8d2017-08-03 23:00:29 +0000119 const SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
Tom Stellardca166212017-01-30 21:56:46 +0000120 MachineRegisterInfo &MRI = MF.getRegInfo();
Matthias Braunf1caa282017-12-15 22:22:58 +0000121 const Function &F = MF.getFunction();
Tom Stellardca166212017-01-30 21:56:46 +0000122 const DataLayout &DL = F.getParent()->getDataLayout();
Matt Arsenault0da63502018-08-31 05:49:54 +0000123 PointerType *PtrTy = PointerType::get(ParamTy, AMDGPUAS::CONSTANT_ADDRESS);
Daniel Sanders52b4ce72017-03-07 23:20:35 +0000124 LLT PtrType = getLLTForType(*PtrTy, DL);
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000125 Register DstReg = MRI.createGenericVirtualRegister(PtrType);
126 Register KernArgSegmentPtr =
Matt Arsenault8623e8d2017-08-03 23:00:29 +0000127 MFI->getPreloadedReg(AMDGPUFunctionArgInfo::KERNARG_SEGMENT_PTR);
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000128 Register KernArgSegmentVReg = MRI.getLiveInVirtReg(KernArgSegmentPtr);
Tom Stellardca166212017-01-30 21:56:46 +0000129
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000130 Register OffsetReg = MRI.createGenericVirtualRegister(LLT::scalar(64));
Tom Stellardca166212017-01-30 21:56:46 +0000131 MIRBuilder.buildConstant(OffsetReg, Offset);
132
133 MIRBuilder.buildGEP(DstReg, KernArgSegmentVReg, OffsetReg);
134
135 return DstReg;
136}
137
138void AMDGPUCallLowering::lowerParameter(MachineIRBuilder &MIRBuilder,
Matt Arsenault29f30372018-07-05 17:01:20 +0000139 Type *ParamTy, uint64_t Offset,
140 unsigned Align,
Matt Arsenaulte3a676e2019-06-24 15:50:29 +0000141 Register DstReg) const {
Tom Stellardca166212017-01-30 21:56:46 +0000142 MachineFunction &MF = MIRBuilder.getMF();
Matthias Braunf1caa282017-12-15 22:22:58 +0000143 const Function &F = MF.getFunction();
Tom Stellardca166212017-01-30 21:56:46 +0000144 const DataLayout &DL = F.getParent()->getDataLayout();
Matt Arsenault0da63502018-08-31 05:49:54 +0000145 PointerType *PtrTy = PointerType::get(ParamTy, AMDGPUAS::CONSTANT_ADDRESS);
Tom Stellardca166212017-01-30 21:56:46 +0000146 MachinePointerInfo PtrInfo(UndefValue::get(PtrTy));
147 unsigned TypeSize = DL.getTypeStoreSize(ParamTy);
Matt Arsenaulte3a676e2019-06-24 15:50:29 +0000148 Register PtrReg = lowerParameterPtr(MIRBuilder, ParamTy, Offset);
Tom Stellardca166212017-01-30 21:56:46 +0000149
150 MachineMemOperand *MMO =
151 MF.getMachineMemOperand(PtrInfo, MachineMemOperand::MOLoad |
152 MachineMemOperand::MONonTemporal |
153 MachineMemOperand::MOInvariant,
154 TypeSize, Align);
155
156 MIRBuilder.buildLoad(DstReg, PtrReg, *MMO);
157}
158
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000159static Register findFirstFreeSGPR(CCState &CCInfo) {
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000160 unsigned NumSGPRs = AMDGPU::SGPR_32RegClass.getNumRegs();
161 for (unsigned Reg = 0; Reg < NumSGPRs; ++Reg) {
162 if (!CCInfo.isAllocated(AMDGPU::SGPR0 + Reg)) {
163 return AMDGPU::SGPR0 + Reg;
164 }
165 }
166 llvm_unreachable("Cannot allocate sgpr");
167}
168
Matt Arsenaulte2c86cc2019-07-01 18:45:36 +0000169static void allocateSpecialEntryInputVGPRs(CCState &CCInfo,
170 MachineFunction &MF,
171 const SIRegisterInfo &TRI,
172 SIMachineFunctionInfo &Info) {
173 const LLT S32 = LLT::scalar(32);
174 MachineRegisterInfo &MRI = MF.getRegInfo();
175
176 if (Info.hasWorkItemIDX()) {
177 Register Reg = AMDGPU::VGPR0;
178 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::VGPR_32RegClass), S32);
179
180 CCInfo.AllocateReg(Reg);
181 Info.setWorkItemIDX(ArgDescriptor::createRegister(Reg));
182 }
183
184 if (Info.hasWorkItemIDY()) {
185 Register Reg = AMDGPU::VGPR1;
186 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::VGPR_32RegClass), S32);
187
188 CCInfo.AllocateReg(Reg);
189 Info.setWorkItemIDY(ArgDescriptor::createRegister(Reg));
190 }
191
192 if (Info.hasWorkItemIDZ()) {
193 Register Reg = AMDGPU::VGPR2;
194 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::VGPR_32RegClass), S32);
195
196 CCInfo.AllocateReg(Reg);
197 Info.setWorkItemIDZ(ArgDescriptor::createRegister(Reg));
198 }
199}
200
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000201static void allocateSystemSGPRs(CCState &CCInfo,
202 MachineFunction &MF,
203 SIMachineFunctionInfo &Info,
204 CallingConv::ID CallConv,
205 bool IsShader) {
Matt Arsenault756d8192019-07-01 18:47:22 +0000206 const LLT S32 = LLT::scalar(32);
207 MachineRegisterInfo &MRI = MF.getRegInfo();
208
209 if (Info.hasWorkGroupIDX()) {
210 Register Reg = Info.addWorkGroupIDX();
211 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::SReg_32_XM0RegClass), S32);
212 CCInfo.AllocateReg(Reg);
213 }
214
215 if (Info.hasWorkGroupIDY()) {
216 Register Reg = Info.addWorkGroupIDY();
217 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::SReg_32_XM0RegClass), S32);
218 CCInfo.AllocateReg(Reg);
219 }
220
221 if (Info.hasWorkGroupIDZ()) {
222 unsigned Reg = Info.addWorkGroupIDZ();
223 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::SReg_32_XM0RegClass), S32);
224 CCInfo.AllocateReg(Reg);
225 }
226
227 if (Info.hasWorkGroupInfo()) {
228 unsigned Reg = Info.addWorkGroupInfo();
229 MRI.setType(MF.addLiveIn(Reg, &AMDGPU::SReg_32_XM0RegClass), S32);
230 CCInfo.AllocateReg(Reg);
231 }
232
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000233 if (Info.hasPrivateSegmentWaveByteOffset()) {
234 // Scratch wave offset passed in system SGPR.
235 unsigned PrivateSegmentWaveByteOffsetReg;
236
237 if (IsShader) {
238 PrivateSegmentWaveByteOffsetReg =
239 Info.getPrivateSegmentWaveByteOffsetSystemSGPR();
240
241 // This is true if the scratch wave byte offset doesn't have a fixed
242 // location.
243 if (PrivateSegmentWaveByteOffsetReg == AMDGPU::NoRegister) {
244 PrivateSegmentWaveByteOffsetReg = findFirstFreeSGPR(CCInfo);
245 Info.setPrivateSegmentWaveByteOffset(PrivateSegmentWaveByteOffsetReg);
246 }
247 } else
248 PrivateSegmentWaveByteOffsetReg = Info.addPrivateSegmentWaveByteOffset();
249
250 MF.addLiveIn(PrivateSegmentWaveByteOffsetReg, &AMDGPU::SGPR_32RegClass);
251 CCInfo.AllocateReg(PrivateSegmentWaveByteOffsetReg);
252 }
253}
254
Diana Picusc3dbe232019-06-27 08:54:17 +0000255bool AMDGPUCallLowering::lowerFormalArguments(
256 MachineIRBuilder &MIRBuilder, const Function &F,
257 ArrayRef<ArrayRef<Register>> VRegs) const {
Tom Stellard37444282018-05-07 22:17:54 +0000258 // AMDGPU_GS and AMDGP_HS are not supported yet.
259 if (F.getCallingConv() == CallingConv::AMDGPU_GS ||
260 F.getCallingConv() == CallingConv::AMDGPU_HS)
Tom Stellard6c814182018-04-30 15:15:23 +0000261 return false;
Tom Stellardca166212017-01-30 21:56:46 +0000262
263 MachineFunction &MF = MIRBuilder.getMF();
Tom Stellard5bfbae52018-07-11 20:59:01 +0000264 const GCNSubtarget *Subtarget = &MF.getSubtarget<GCNSubtarget>();
Tom Stellardca166212017-01-30 21:56:46 +0000265 MachineRegisterInfo &MRI = MF.getRegInfo();
266 SIMachineFunctionInfo *Info = MF.getInfo<SIMachineFunctionInfo>();
Tom Stellard5bfbae52018-07-11 20:59:01 +0000267 const SIRegisterInfo *TRI = MF.getSubtarget<GCNSubtarget>().getRegisterInfo();
Tom Stellardca166212017-01-30 21:56:46 +0000268 const DataLayout &DL = F.getParent()->getDataLayout();
269
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000270 bool IsShader = AMDGPU::isShader(F.getCallingConv());
271
Tom Stellardca166212017-01-30 21:56:46 +0000272 SmallVector<CCValAssign, 16> ArgLocs;
273 CCState CCInfo(F.getCallingConv(), F.isVarArg(), MF, ArgLocs, F.getContext());
274
275 // FIXME: How should these inputs interact with inreg / custom SGPR inputs?
276 if (Info->hasPrivateSegmentBuffer()) {
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000277 Register PrivateSegmentBufferReg = Info->addPrivateSegmentBuffer(*TRI);
Tom Stellardca166212017-01-30 21:56:46 +0000278 MF.addLiveIn(PrivateSegmentBufferReg, &AMDGPU::SReg_128RegClass);
279 CCInfo.AllocateReg(PrivateSegmentBufferReg);
280 }
281
282 if (Info->hasDispatchPtr()) {
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000283 Register DispatchPtrReg = Info->addDispatchPtr(*TRI);
Tom Stellardca166212017-01-30 21:56:46 +0000284 // FIXME: Need to add reg as live-in
285 CCInfo.AllocateReg(DispatchPtrReg);
286 }
287
288 if (Info->hasQueuePtr()) {
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000289 Register QueuePtrReg = Info->addQueuePtr(*TRI);
Tom Stellardca166212017-01-30 21:56:46 +0000290 // FIXME: Need to add reg as live-in
291 CCInfo.AllocateReg(QueuePtrReg);
292 }
293
294 if (Info->hasKernargSegmentPtr()) {
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000295 Register InputPtrReg = Info->addKernargSegmentPtr(*TRI);
Yaxun Liu0124b542018-02-13 18:00:25 +0000296 const LLT P2 = LLT::pointer(AMDGPUAS::CONSTANT_ADDRESS, 64);
Matt Arsenaultfaeaedf2019-06-24 16:16:12 +0000297 Register VReg = MRI.createGenericVirtualRegister(P2);
Tom Stellardca166212017-01-30 21:56:46 +0000298 MRI.addLiveIn(InputPtrReg, VReg);
299 MIRBuilder.getMBB().addLiveIn(InputPtrReg);
300 MIRBuilder.buildCopy(VReg, InputPtrReg);
301 CCInfo.AllocateReg(InputPtrReg);
302 }
303
304 if (Info->hasDispatchID()) {
305 unsigned DispatchIDReg = Info->addDispatchID(*TRI);
306 // FIXME: Need to add reg as live-in
307 CCInfo.AllocateReg(DispatchIDReg);
308 }
309
310 if (Info->hasFlatScratchInit()) {
311 unsigned FlatScratchInitReg = Info->addFlatScratchInit(*TRI);
312 // FIXME: Need to add reg as live-in
313 CCInfo.AllocateReg(FlatScratchInitReg);
314 }
315
Matt Arsenault29f30372018-07-05 17:01:20 +0000316 // The infrastructure for normal calling convention lowering is essentially
317 // useless for kernels. We want to avoid any kind of legalization or argument
318 // splitting.
319 if (F.getCallingConv() == CallingConv::AMDGPU_KERNEL) {
320 unsigned i = 0;
321 const unsigned KernArgBaseAlign = 16;
322 const unsigned BaseOffset = Subtarget->getExplicitKernelArgOffset(F);
323 uint64_t ExplicitArgOffset = 0;
324
325 // TODO: Align down to dword alignment and extract bits for extending loads.
326 for (auto &Arg : F.args()) {
327 Type *ArgTy = Arg.getType();
328 unsigned AllocSize = DL.getTypeAllocSize(ArgTy);
329 if (AllocSize == 0)
330 continue;
331
332 unsigned ABIAlign = DL.getABITypeAlignment(ArgTy);
333
334 uint64_t ArgOffset = alignTo(ExplicitArgOffset, ABIAlign) + BaseOffset;
335 ExplicitArgOffset = alignTo(ExplicitArgOffset, ABIAlign) + AllocSize;
336
Diana Picusc3dbe232019-06-27 08:54:17 +0000337 ArrayRef<Register> OrigArgRegs = VRegs[i];
338 Register ArgReg =
339 OrigArgRegs.size() == 1
340 ? OrigArgRegs[0]
341 : MRI.createGenericVirtualRegister(getLLTForType(*ArgTy, DL));
Matt Arsenault29f30372018-07-05 17:01:20 +0000342 unsigned Align = MinAlign(KernArgBaseAlign, ArgOffset);
343 ArgOffset = alignTo(ArgOffset, DL.getABITypeAlignment(ArgTy));
Diana Picusc3dbe232019-06-27 08:54:17 +0000344 lowerParameter(MIRBuilder, ArgTy, ArgOffset, Align, ArgReg);
345 if (OrigArgRegs.size() > 1)
346 unpackRegs(OrigArgRegs, ArgReg, ArgTy, MIRBuilder);
Matt Arsenault29f30372018-07-05 17:01:20 +0000347 ++i;
348 }
349
Matt Arsenaulte2c86cc2019-07-01 18:45:36 +0000350 allocateSpecialEntryInputVGPRs(CCInfo, MF, *TRI, *Info);
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000351 allocateSystemSGPRs(CCInfo, MF, *Info, F.getCallingConv(), IsShader);
Matt Arsenault29f30372018-07-05 17:01:20 +0000352 return true;
353 }
354
Tom Stellardca166212017-01-30 21:56:46 +0000355 unsigned NumArgs = F.arg_size();
356 Function::const_arg_iterator CurOrigArg = F.arg_begin();
357 const AMDGPUTargetLowering &TLI = *getTLI<AMDGPUTargetLowering>();
Tom Stellardc7709e12018-04-24 20:51:28 +0000358 unsigned PSInputNum = 0;
359 BitVector Skipped(NumArgs);
Tom Stellardca166212017-01-30 21:56:46 +0000360 for (unsigned i = 0; i != NumArgs; ++i, ++CurOrigArg) {
Tom Stellard9d8337d2017-08-01 12:38:33 +0000361 EVT ValEVT = TLI.getValueType(DL, CurOrigArg->getType());
362
363 // We can only hanlde simple value types at the moment.
Tom Stellardca166212017-01-30 21:56:46 +0000364 ISD::ArgFlagsTy Flags;
Diana Picusc3dbe232019-06-27 08:54:17 +0000365 assert(VRegs[i].size() == 1 && "Can't lower into more than one register");
366 ArgInfo OrigArg{VRegs[i][0], CurOrigArg->getType()};
Tom Stellard9d8337d2017-08-01 12:38:33 +0000367 setArgFlags(OrigArg, i + 1, DL, F);
Tom Stellardca166212017-01-30 21:56:46 +0000368 Flags.setOrigAlign(DL.getABITypeAlignment(CurOrigArg->getType()));
Tom Stellardc7709e12018-04-24 20:51:28 +0000369
370 if (F.getCallingConv() == CallingConv::AMDGPU_PS &&
371 !OrigArg.Flags.isInReg() && !OrigArg.Flags.isByVal() &&
372 PSInputNum <= 15) {
373 if (CurOrigArg->use_empty() && !Info->isPSInputAllocated(PSInputNum)) {
374 Skipped.set(i);
375 ++PSInputNum;
376 continue;
377 }
378
379 Info->markPSInputAllocated(PSInputNum);
380 if (!CurOrigArg->use_empty())
381 Info->markPSInputEnabled(PSInputNum);
382
383 ++PSInputNum;
384 }
385
Tom Stellardca166212017-01-30 21:56:46 +0000386 CCAssignFn *AssignFn = CCAssignFnForCall(F.getCallingConv(),
387 /*IsVarArg=*/false);
Tom Stellard9d8337d2017-08-01 12:38:33 +0000388
Tom Stellardc7709e12018-04-24 20:51:28 +0000389 if (ValEVT.isVector()) {
390 EVT ElemVT = ValEVT.getVectorElementType();
391 if (!ValEVT.isSimple())
392 return false;
393 MVT ValVT = ElemVT.getSimpleVT();
394 bool Res = AssignFn(i, ValVT, ValVT, CCValAssign::Full,
395 OrigArg.Flags, CCInfo);
396 if (!Res)
397 return false;
398 } else {
399 MVT ValVT = ValEVT.getSimpleVT();
400 if (!ValEVT.isSimple())
401 return false;
402 bool Res =
403 AssignFn(i, ValVT, ValVT, CCValAssign::Full, OrigArg.Flags, CCInfo);
404
405 // Fail if we don't know how to handle this type.
406 if (Res)
407 return false;
408 }
Tom Stellardca166212017-01-30 21:56:46 +0000409 }
410
411 Function::const_arg_iterator Arg = F.arg_begin();
Tom Stellard9d8337d2017-08-01 12:38:33 +0000412
Tom Stellardc7709e12018-04-24 20:51:28 +0000413 if (F.getCallingConv() == CallingConv::AMDGPU_VS ||
414 F.getCallingConv() == CallingConv::AMDGPU_PS) {
415 for (unsigned i = 0, OrigArgIdx = 0;
416 OrigArgIdx != NumArgs && i != ArgLocs.size(); ++Arg, ++OrigArgIdx) {
417 if (Skipped.test(OrigArgIdx))
418 continue;
Diana Picusc3dbe232019-06-27 08:54:17 +0000419 assert(VRegs[OrigArgIdx].size() == 1 &&
420 "Can't lower into more than 1 reg");
421 CCValAssign &VA = ArgLocs[i++];
422 MRI.addLiveIn(VA.getLocReg(), VRegs[OrigArgIdx][0]);
423 MIRBuilder.getMBB().addLiveIn(VA.getLocReg());
424 MIRBuilder.buildCopy(VRegs[OrigArgIdx][0], VA.getLocReg());
Tom Stellard9d8337d2017-08-01 12:38:33 +0000425 }
Matt Arsenaulte0a4da82019-05-30 19:33:18 +0000426
427 allocateSystemSGPRs(CCInfo, MF, *Info, F.getCallingConv(), IsShader);
Tom Stellard9d8337d2017-08-01 12:38:33 +0000428 return true;
429 }
430
Matt Arsenault29f30372018-07-05 17:01:20 +0000431 return false;
Tom Stellard000c5af2016-04-14 19:09:28 +0000432}