blob: ce17202f3414ab226a92e1f362518521dc40a89c [file] [log] [blame]
Eugene Zelenkofa6434b2017-08-31 21:56:16 +00001//===- AMDGPUAnnotateKernelFeaturesPass.cpp -------------------------------===//
Matt Arsenault39319482015-11-06 18:01:57 +00002//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10/// \file This pass adds target attributes to functions which use intrinsics
11/// which will impact calling convention lowering.
12//
13//===----------------------------------------------------------------------===//
14
15#include "AMDGPU.h"
Matt Arsenaulte823d922017-02-18 18:29:53 +000016#include "AMDGPUSubtarget.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000017#include "Utils/AMDGPUBaseInfo.h"
18#include "llvm/ADT/SmallPtrSet.h"
19#include "llvm/ADT/SmallVector.h"
20#include "llvm/ADT/StringRef.h"
Matt Arsenault2ffe8fd2016-08-11 19:18:50 +000021#include "llvm/ADT/Triple.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000022#include "llvm/Analysis/CallGraph.h"
Matt Arsenault6b930462017-07-13 21:43:42 +000023#include "llvm/Analysis/CallGraphSCCPass.h"
Francis Visoiu Mistrih8b617642017-05-18 17:21:13 +000024#include "llvm/CodeGen/TargetPassConfig.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000025#include "llvm/IR/CallSite.h"
26#include "llvm/IR/Constant.h"
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000027#include "llvm/IR/Constants.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000028#include "llvm/IR/Function.h"
29#include "llvm/IR/Instruction.h"
Matt Arsenault39319482015-11-06 18:01:57 +000030#include "llvm/IR/Instructions.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000031#include "llvm/IR/Intrinsics.h"
Matt Arsenault39319482015-11-06 18:01:57 +000032#include "llvm/IR/Module.h"
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000033#include "llvm/IR/Type.h"
34#include "llvm/IR/Use.h"
35#include "llvm/Pass.h"
36#include "llvm/Support/Casting.h"
37#include "llvm/Support/ErrorHandling.h"
38#include "llvm/Target/TargetMachine.h"
Matt Arsenault39319482015-11-06 18:01:57 +000039
40#define DEBUG_TYPE "amdgpu-annotate-kernel-features"
41
42using namespace llvm;
43
44namespace {
45
Matt Arsenault6b930462017-07-13 21:43:42 +000046class AMDGPUAnnotateKernelFeatures : public CallGraphSCCPass {
Matt Arsenault39319482015-11-06 18:01:57 +000047private:
Matt Arsenault6b930462017-07-13 21:43:42 +000048 const TargetMachine *TM = nullptr;
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000049 AMDGPUAS AS;
Matt Arsenault99c14522016-04-25 19:27:24 +000050
Matt Arsenault6b930462017-07-13 21:43:42 +000051 bool addFeatureAttributes(Function &F);
52
Matt Arsenault39319482015-11-06 18:01:57 +000053public:
54 static char ID;
55
Matt Arsenault6b930462017-07-13 21:43:42 +000056 AMDGPUAnnotateKernelFeatures() : CallGraphSCCPass(ID) {}
57
58 bool doInitialization(CallGraph &CG) override;
59 bool runOnSCC(CallGraphSCC &SCC) override;
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000060
Mehdi Amini117296c2016-10-01 02:56:57 +000061 StringRef getPassName() const override {
Matt Arsenault39319482015-11-06 18:01:57 +000062 return "AMDGPU Annotate Kernel Features";
63 }
64
65 void getAnalysisUsage(AnalysisUsage &AU) const override {
66 AU.setPreservesAll();
Matt Arsenault6b930462017-07-13 21:43:42 +000067 CallGraphSCCPass::getAnalysisUsage(AU);
Matt Arsenault39319482015-11-06 18:01:57 +000068 }
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000069
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000070 static bool visitConstantExpr(const ConstantExpr *CE, AMDGPUAS AS);
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000071 static bool visitConstantExprsRecursively(
72 const Constant *EntryC,
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000073 SmallPtrSet<const Constant *, 8> &ConstantExprVisited,
74 AMDGPUAS AS);
Matt Arsenault39319482015-11-06 18:01:57 +000075};
76
Eugene Zelenkofa6434b2017-08-31 21:56:16 +000077} // end anonymous namespace
Matt Arsenault39319482015-11-06 18:01:57 +000078
79char AMDGPUAnnotateKernelFeatures::ID = 0;
80
81char &llvm::AMDGPUAnnotateKernelFeaturesID = AMDGPUAnnotateKernelFeatures::ID;
82
Matt Arsenault99c14522016-04-25 19:27:24 +000083INITIALIZE_PASS(AMDGPUAnnotateKernelFeatures, DEBUG_TYPE,
84 "Add AMDGPU function attributes", false, false)
Matt Arsenault39319482015-11-06 18:01:57 +000085
Matt Arsenault39319482015-11-06 18:01:57 +000086
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000087// The queue ptr is only needed when casting to flat, not from it.
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000088static bool castRequiresQueuePtr(unsigned SrcAS, const AMDGPUAS &AS) {
89 return SrcAS == AS.LOCAL_ADDRESS || SrcAS == AS.PRIVATE_ADDRESS;
Matt Arsenault99c14522016-04-25 19:27:24 +000090}
91
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000092static bool castRequiresQueuePtr(const AddrSpaceCastInst *ASC,
93 const AMDGPUAS &AS) {
94 return castRequiresQueuePtr(ASC->getSrcAddressSpace(), AS);
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000095}
96
Yaxun Liu1a14bfa2017-03-27 14:04:01 +000097bool AMDGPUAnnotateKernelFeatures::visitConstantExpr(const ConstantExpr *CE,
98 AMDGPUAS AS) {
Matt Arsenault3b2e2a52016-06-06 20:03:31 +000099 if (CE->getOpcode() == Instruction::AddrSpaceCast) {
100 unsigned SrcAS = CE->getOperand(0)->getType()->getPointerAddressSpace();
Yaxun Liu1a14bfa2017-03-27 14:04:01 +0000101 return castRequiresQueuePtr(SrcAS, AS);
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000102 }
103
104 return false;
105}
106
107bool AMDGPUAnnotateKernelFeatures::visitConstantExprsRecursively(
108 const Constant *EntryC,
Yaxun Liu1a14bfa2017-03-27 14:04:01 +0000109 SmallPtrSet<const Constant *, 8> &ConstantExprVisited,
110 AMDGPUAS AS) {
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000111
112 if (!ConstantExprVisited.insert(EntryC).second)
113 return false;
114
115 SmallVector<const Constant *, 16> Stack;
116 Stack.push_back(EntryC);
117
118 while (!Stack.empty()) {
119 const Constant *C = Stack.pop_back_val();
120
121 // Check this constant expression.
122 if (const auto *CE = dyn_cast<ConstantExpr>(C)) {
Yaxun Liu1a14bfa2017-03-27 14:04:01 +0000123 if (visitConstantExpr(CE, AS))
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000124 return true;
125 }
126
127 // Visit all sub-expressions.
128 for (const Use &U : C->operands()) {
129 const auto *OpC = dyn_cast<Constant>(U);
130 if (!OpC)
131 continue;
132
133 if (!ConstantExprVisited.insert(OpC).second)
134 continue;
135
136 Stack.push_back(OpC);
137 }
138 }
139
140 return false;
141}
142
Matt Arsenault6b930462017-07-13 21:43:42 +0000143// We do not need to note the x workitem or workgroup id because they are always
144// initialized.
145//
146// TODO: We should not add the attributes if the known compile time workgroup
147// size is 1 for y/z.
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000148static StringRef intrinsicToAttrName(Intrinsic::ID ID,
149 bool &NonKernelOnly,
150 bool &IsQueuePtr) {
Matt Arsenault6b930462017-07-13 21:43:42 +0000151 switch (ID) {
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000152 case Intrinsic::amdgcn_workitem_id_x:
153 NonKernelOnly = true;
154 return "amdgpu-work-item-id-x";
155 case Intrinsic::amdgcn_workgroup_id_x:
156 NonKernelOnly = true;
157 return "amdgpu-work-group-id-x";
Matt Arsenault6b930462017-07-13 21:43:42 +0000158 case Intrinsic::amdgcn_workitem_id_y:
159 case Intrinsic::r600_read_tidig_y:
160 return "amdgpu-work-item-id-y";
161 case Intrinsic::amdgcn_workitem_id_z:
162 case Intrinsic::r600_read_tidig_z:
163 return "amdgpu-work-item-id-z";
164 case Intrinsic::amdgcn_workgroup_id_y:
165 case Intrinsic::r600_read_tgid_y:
166 return "amdgpu-work-group-id-y";
167 case Intrinsic::amdgcn_workgroup_id_z:
168 case Intrinsic::r600_read_tgid_z:
169 return "amdgpu-work-group-id-z";
170 case Intrinsic::amdgcn_dispatch_ptr:
171 return "amdgpu-dispatch-ptr";
172 case Intrinsic::amdgcn_dispatch_id:
173 return "amdgpu-dispatch-id";
Matt Arsenault23e4df62017-07-14 00:11:13 +0000174 case Intrinsic::amdgcn_kernarg_segment_ptr:
Matt Arsenault23e4df62017-07-14 00:11:13 +0000175 return "amdgpu-kernarg-segment-ptr";
Matt Arsenault9166ce82017-07-28 15:52:08 +0000176 case Intrinsic::amdgcn_implicitarg_ptr:
177 return "amdgpu-implicitarg-ptr";
Matt Arsenault6b930462017-07-13 21:43:42 +0000178 case Intrinsic::amdgcn_queue_ptr:
179 case Intrinsic::trap:
180 case Intrinsic::debugtrap:
181 IsQueuePtr = true;
182 return "amdgpu-queue-ptr";
183 default:
184 return "";
185 }
186}
187
188static bool handleAttr(Function &Parent, const Function &Callee,
189 StringRef Name) {
190 if (Callee.hasFnAttribute(Name)) {
191 Parent.addFnAttr(Name);
192 return true;
193 }
194
195 return false;
196}
197
198static void copyFeaturesToFunction(Function &Parent, const Function &Callee,
199 bool &NeedQueuePtr) {
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000200 // X ids unnecessarily propagated to kernels.
Matt Arsenault6b930462017-07-13 21:43:42 +0000201 static const StringRef AttrNames[] = {
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000202 { "amdgpu-work-item-id-x" },
Matt Arsenault6b930462017-07-13 21:43:42 +0000203 { "amdgpu-work-item-id-y" },
204 { "amdgpu-work-item-id-z" },
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000205 { "amdgpu-work-group-id-x" },
Matt Arsenault6b930462017-07-13 21:43:42 +0000206 { "amdgpu-work-group-id-y" },
207 { "amdgpu-work-group-id-z" },
208 { "amdgpu-dispatch-ptr" },
Matt Arsenault23e4df62017-07-14 00:11:13 +0000209 { "amdgpu-dispatch-id" },
Matt Arsenault9166ce82017-07-28 15:52:08 +0000210 { "amdgpu-kernarg-segment-ptr" },
211 { "amdgpu-implicitarg-ptr" }
Matt Arsenault6b930462017-07-13 21:43:42 +0000212 };
213
214 if (handleAttr(Parent, Callee, "amdgpu-queue-ptr"))
215 NeedQueuePtr = true;
216
217 for (StringRef AttrName : AttrNames)
218 handleAttr(Parent, Callee, AttrName);
219}
220
221bool AMDGPUAnnotateKernelFeatures::addFeatureAttributes(Function &F) {
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000222 const AMDGPUSubtarget &ST = TM->getSubtarget<AMDGPUSubtarget>(F);
223 bool HasFlat = ST.hasFlatAddressSpace();
224 bool HasApertureRegs = ST.hasApertureRegs();
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000225 SmallPtrSet<const Constant *, 8> ConstantExprVisited;
226
Matt Arsenault6b930462017-07-13 21:43:42 +0000227 bool Changed = false;
228 bool NeedQueuePtr = false;
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000229 bool HaveCall = false;
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000230 bool IsFunc = !AMDGPU::isEntryFunctionCC(F.getCallingConv());
Matt Arsenault6b930462017-07-13 21:43:42 +0000231
232 for (BasicBlock &BB : F) {
233 for (Instruction &I : BB) {
234 CallSite CS(&I);
235 if (CS) {
236 Function *Callee = CS.getCalledFunction();
237
238 // TODO: Do something with indirect calls.
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000239 if (!Callee) {
240 if (!CS.isInlineAsm())
241 HaveCall = true;
Matt Arsenault6b930462017-07-13 21:43:42 +0000242 continue;
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000243 }
Matt Arsenault6b930462017-07-13 21:43:42 +0000244
245 Intrinsic::ID IID = Callee->getIntrinsicID();
246 if (IID == Intrinsic::not_intrinsic) {
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000247 HaveCall = true;
Matt Arsenault6b930462017-07-13 21:43:42 +0000248 copyFeaturesToFunction(F, *Callee, NeedQueuePtr);
249 Changed = true;
250 } else {
Matt Arsenaulte15855d2017-07-17 22:35:50 +0000251 bool NonKernelOnly = false;
252 StringRef AttrName = intrinsicToAttrName(IID,
253 NonKernelOnly, NeedQueuePtr);
254 if (!AttrName.empty() && (IsFunc || !NonKernelOnly)) {
Matt Arsenault6b930462017-07-13 21:43:42 +0000255 F.addFnAttr(AttrName);
256 Changed = true;
257 }
258 }
259 }
260
261 if (NeedQueuePtr || HasApertureRegs)
262 continue;
263
Matt Arsenault99c14522016-04-25 19:27:24 +0000264 if (const AddrSpaceCastInst *ASC = dyn_cast<AddrSpaceCastInst>(&I)) {
Matt Arsenault6b930462017-07-13 21:43:42 +0000265 if (castRequiresQueuePtr(ASC, AS)) {
266 NeedQueuePtr = true;
267 continue;
268 }
Matt Arsenault99c14522016-04-25 19:27:24 +0000269 }
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000270
271 for (const Use &U : I.operands()) {
272 const auto *OpC = dyn_cast<Constant>(U);
273 if (!OpC)
274 continue;
275
Matt Arsenault6b930462017-07-13 21:43:42 +0000276 if (visitConstantExprsRecursively(OpC, ConstantExprVisited, AS)) {
277 NeedQueuePtr = true;
278 break;
279 }
Matt Arsenault3b2e2a52016-06-06 20:03:31 +0000280 }
Matt Arsenault99c14522016-04-25 19:27:24 +0000281 }
282 }
283
Matt Arsenault6b930462017-07-13 21:43:42 +0000284 if (NeedQueuePtr) {
285 F.addFnAttr("amdgpu-queue-ptr");
286 Changed = true;
287 }
288
Matt Arsenault254ad3d2017-07-18 16:44:58 +0000289 // TODO: We could refine this to captured pointers that could possibly be
290 // accessed by flat instructions. For now this is mostly a poor way of
291 // estimating whether there are calls before argument lowering.
292 if (HasFlat && !IsFunc && HaveCall) {
293 F.addFnAttr("amdgpu-flat-scratch");
294 Changed = true;
295 }
296
Matt Arsenault6b930462017-07-13 21:43:42 +0000297 return Changed;
Matt Arsenault99c14522016-04-25 19:27:24 +0000298}
Matt Arsenault39319482015-11-06 18:01:57 +0000299
Matt Arsenault6b930462017-07-13 21:43:42 +0000300bool AMDGPUAnnotateKernelFeatures::runOnSCC(CallGraphSCC &SCC) {
301 Module &M = SCC.getCallGraph().getModule();
Matt Arsenault39319482015-11-06 18:01:57 +0000302 Triple TT(M.getTargetTriple());
303
Matt Arsenault6b930462017-07-13 21:43:42 +0000304 bool Changed = false;
305 for (CallGraphNode *I : SCC) {
306 Function *F = I->getFunction();
307 if (!F || F->isDeclaration())
308 continue;
Matt Arsenault43976df2016-01-30 04:25:19 +0000309
Matt Arsenault6b930462017-07-13 21:43:42 +0000310 Changed |= addFeatureAttributes(*F);
Matt Arsenault99c14522016-04-25 19:27:24 +0000311 }
312
Matt Arsenault39319482015-11-06 18:01:57 +0000313 return Changed;
314}
315
Matt Arsenault6b930462017-07-13 21:43:42 +0000316bool AMDGPUAnnotateKernelFeatures::doInitialization(CallGraph &CG) {
317 auto *TPC = getAnalysisIfAvailable<TargetPassConfig>();
318 if (!TPC)
319 report_fatal_error("TargetMachine is required");
320
321 AS = AMDGPU::getAMDGPUAS(CG.getModule());
322 TM = &TPC->getTM<TargetMachine>();
323 return false;
324}
325
326Pass *llvm::createAMDGPUAnnotateKernelFeaturesPass() {
Francis Visoiu Mistrih8b617642017-05-18 17:21:13 +0000327 return new AMDGPUAnnotateKernelFeatures();
Matt Arsenault39319482015-11-06 18:01:57 +0000328}