blob: b6c3997dae3fa04c1f07951dfca850006aa26b48 [file] [log] [blame]
Dan Gohman4552e3c2009-10-13 18:30:07 +00001//===- InlineCost.cpp - Cost analysis for inliner -------------------------===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10// This file implements inline cost analysis.
11//
12//===----------------------------------------------------------------------===//
13
14#include "llvm/Analysis/InlineCost.h"
Chandler Carruthed0881b2012-12-03 16:50:05 +000015#include "llvm/ADT/STLExtras.h"
16#include "llvm/ADT/SetVector.h"
17#include "llvm/ADT/SmallPtrSet.h"
18#include "llvm/ADT/SmallVector.h"
19#include "llvm/ADT/Statistic.h"
Daniel Jasperaec2fa32016-12-19 08:22:17 +000020#include "llvm/Analysis/AssumptionCache.h"
Hal Finkel57f03dd2014-09-07 13:49:57 +000021#include "llvm/Analysis/CodeMetrics.h"
Chandler Carruthd9903882015-01-14 11:23:27 +000022#include "llvm/Analysis/ConstantFolding.h"
Chandler Carruth0539c072012-03-31 12:42:41 +000023#include "llvm/Analysis/InstructionSimplify.h"
Easwaran Raman71069cf2016-06-09 22:23:21 +000024#include "llvm/Analysis/ProfileSummaryInfo.h"
Chandler Carruth42f3dce2013-01-21 11:55:09 +000025#include "llvm/Analysis/TargetTransformInfo.h"
Chandler Carruth219b89b2014-03-04 11:01:28 +000026#include "llvm/IR/CallSite.h"
Chandler Carruth9fb823b2013-01-02 11:36:10 +000027#include "llvm/IR/CallingConv.h"
28#include "llvm/IR/DataLayout.h"
Chandler Carruth03eb0de2014-03-04 10:40:04 +000029#include "llvm/IR/GetElementPtrTypeIterator.h"
Chandler Carruth9fb823b2013-01-02 11:36:10 +000030#include "llvm/IR/GlobalAlias.h"
Chandler Carruth7da14f12014-03-06 03:23:41 +000031#include "llvm/IR/InstVisitor.h"
Chandler Carruth9fb823b2013-01-02 11:36:10 +000032#include "llvm/IR/IntrinsicInst.h"
33#include "llvm/IR/Operator.h"
Chandler Carruth0539c072012-03-31 12:42:41 +000034#include "llvm/Support/Debug.h"
Chandler Carruth0539c072012-03-31 12:42:41 +000035#include "llvm/Support/raw_ostream.h"
Eric Christopher2dfbd7e2011-02-05 00:49:15 +000036
Dan Gohman4552e3c2009-10-13 18:30:07 +000037using namespace llvm;
38
Chandler Carruthf1221bd2014-04-22 02:48:03 +000039#define DEBUG_TYPE "inline-cost"
40
Chandler Carruth7ae90d42012-04-11 10:15:10 +000041STATISTIC(NumCallsAnalyzed, "Number of call sites analyzed");
42
Easwaran Raman1c57cc22016-08-10 00:48:04 +000043static cl::opt<int> InlineThreshold(
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +000044 "inline-threshold", cl::Hidden, cl::init(225), cl::ZeroOrMore,
45 cl::desc("Control the amount of inlining to perform (default = 225)"));
46
47static cl::opt<int> HintThreshold(
48 "inlinehint-threshold", cl::Hidden, cl::init(325),
49 cl::desc("Threshold for inlining functions with inline hint"));
50
51// We introduce this threshold to help performance of instrumentation based
52// PGO before we actually hook up inliner with analysis passes such as BPI and
53// BFI.
54static cl::opt<int> ColdThreshold(
55 "inlinecold-threshold", cl::Hidden, cl::init(225),
56 cl::desc("Threshold for inlining functions with cold attribute"));
57
Dehao Chende39cb92016-08-05 20:28:41 +000058static cl::opt<int>
59 HotCallSiteThreshold("hot-callsite-threshold", cl::Hidden, cl::init(3000),
60 cl::ZeroOrMore,
61 cl::desc("Threshold for hot callsites "));
62
Chandler Carruth0539c072012-03-31 12:42:41 +000063namespace {
Chandler Carrutha3089552012-03-14 07:32:53 +000064
Chandler Carruth0539c072012-03-31 12:42:41 +000065class CallAnalyzer : public InstVisitor<CallAnalyzer, bool> {
66 typedef InstVisitor<CallAnalyzer, bool> Base;
67 friend class InstVisitor<CallAnalyzer, bool>;
Owen Andersona08318a2010-09-09 16:56:42 +000068
Chandler Carruth42f3dce2013-01-21 11:55:09 +000069 /// The TargetTransformInfo available for this compilation.
70 const TargetTransformInfo &TTI;
71
Daniel Jasperaec2fa32016-12-19 08:22:17 +000072 /// Getter for the cache of @llvm.assume intrinsics.
73 std::function<AssumptionCache &(Function &)> &GetAssumptionCache;
74
Easwaran Raman71069cf2016-06-09 22:23:21 +000075 /// Profile summary information.
76 ProfileSummaryInfo *PSI;
77
Piotr Padlewskif3d122c2016-09-30 21:05:49 +000078 /// The called function.
Chandler Carruth0539c072012-03-31 12:42:41 +000079 Function &F;
Owen Andersona08318a2010-09-09 16:56:42 +000080
Piotr Padlewskif3d122c2016-09-30 21:05:49 +000081 /// The candidate callsite being analyzed. Please do not use this to do
82 /// analysis in the caller function; we want the inline cost query to be
83 /// easily cacheable. Instead, use the cover function paramHasAttr.
Philip Reames9b5c9582015-06-26 20:51:17 +000084 CallSite CandidateCS;
85
Piotr Padlewskif3d122c2016-09-30 21:05:49 +000086 /// Tunable parameters that control the analysis.
Easwaran Raman1c57cc22016-08-10 00:48:04 +000087 const InlineParams &Params;
88
Chandler Carruth0539c072012-03-31 12:42:41 +000089 int Threshold;
90 int Cost;
Owen Andersona08318a2010-09-09 16:56:42 +000091
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +000092 bool IsCallerRecursive;
93 bool IsRecursiveCall;
Chandler Carruth0539c072012-03-31 12:42:41 +000094 bool ExposesReturnsTwice;
95 bool HasDynamicAlloca;
James Molloy4f6fb952012-12-20 16:04:27 +000096 bool ContainsNoDuplicateCall;
Chandler Carruth0814d2a2013-12-13 07:59:56 +000097 bool HasReturn;
98 bool HasIndirectBr;
Reid Kleckner223de262015-04-14 20:38:14 +000099 bool HasFrameEscape;
James Molloy4f6fb952012-12-20 16:04:27 +0000100
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +0000101 /// Number of bytes allocated statically by the callee.
102 uint64_t AllocatedSize;
Chandler Carruth0539c072012-03-31 12:42:41 +0000103 unsigned NumInstructions, NumVectorInstructions;
104 int FiftyPercentVectorBonus, TenPercentVectorBonus;
105 int VectorBonus;
106
Piotr Padlewskif3d122c2016-09-30 21:05:49 +0000107 /// While we walk the potentially-inlined instructions, we build up and
108 /// maintain a mapping of simplified values specific to this callsite. The
109 /// idea is to propagate any special information we have about arguments to
110 /// this call through the inlinable section of the function, and account for
111 /// likely simplifications post-inlining. The most important aspect we track
112 /// is CFG altering simplifications -- when we prove a basic block dead, that
113 /// can cause dramatic shifts in the cost of inlining a function.
Chandler Carruth0539c072012-03-31 12:42:41 +0000114 DenseMap<Value *, Constant *> SimplifiedValues;
115
Piotr Padlewskif3d122c2016-09-30 21:05:49 +0000116 /// Keep track of the values which map back (through function arguments) to
117 /// allocas on the caller stack which could be simplified through SROA.
Chandler Carruth0539c072012-03-31 12:42:41 +0000118 DenseMap<Value *, Value *> SROAArgValues;
119
Piotr Padlewskif3d122c2016-09-30 21:05:49 +0000120 /// The mapping of caller Alloca values to their accumulated cost savings. If
121 /// we have to disable SROA for one of the allocas, this tells us how much
122 /// cost must be added.
Chandler Carruth0539c072012-03-31 12:42:41 +0000123 DenseMap<Value *, int> SROAArgCosts;
124
Piotr Padlewskif3d122c2016-09-30 21:05:49 +0000125 /// Keep track of values which map to a pointer base and constant offset.
Chad Rosier567556a2016-04-28 14:47:23 +0000126 DenseMap<Value *, std::pair<Value *, APInt>> ConstantOffsetPtrs;
Chandler Carruth0539c072012-03-31 12:42:41 +0000127
128 // Custom simplification helper routines.
129 bool isAllocaDerivedArg(Value *V);
130 bool lookupSROAArgAndCost(Value *V, Value *&Arg,
131 DenseMap<Value *, int>::iterator &CostIt);
132 void disableSROA(DenseMap<Value *, int>::iterator CostIt);
133 void disableSROA(Value *V);
134 void accumulateSROACost(DenseMap<Value *, int>::iterator CostIt,
135 int InstructionCost);
Chandler Carruth0539c072012-03-31 12:42:41 +0000136 bool isGEPOffsetConstant(GetElementPtrInst &GEP);
Haicheng Wu201b1912017-01-20 18:51:22 +0000137 bool isGEPFree(GetElementPtrInst &GEP);
Chandler Carruth0539c072012-03-31 12:42:41 +0000138 bool accumulateGEPOffset(GEPOperator &GEP, APInt &Offset);
Chandler Carruth753e21d2012-12-28 14:23:32 +0000139 bool simplifyCallSite(Function *F, CallSite CS);
Chandler Carruth0539c072012-03-31 12:42:41 +0000140 ConstantInt *stripAndComputeInBoundsConstantOffsets(Value *&V);
141
Philip Reames9b5c9582015-06-26 20:51:17 +0000142 /// Return true if the given argument to the function being considered for
143 /// inlining has the given attribute set either at the call site or the
144 /// function declaration. Primarily used to inspect call site specific
145 /// attributes since these can be more precise than the ones on the callee
Easwaran Raman3676da42015-12-03 19:03:20 +0000146 /// itself.
Philip Reames9b5c9582015-06-26 20:51:17 +0000147 bool paramHasAttr(Argument *A, Attribute::AttrKind Attr);
Chad Rosier567556a2016-04-28 14:47:23 +0000148
Philip Reames9b5c9582015-06-26 20:51:17 +0000149 /// Return true if the given value is known non null within the callee if
Easwaran Raman3676da42015-12-03 19:03:20 +0000150 /// inlined through this particular callsite.
Philip Reames9b5c9582015-06-26 20:51:17 +0000151 bool isKnownNonNullInCallee(Value *V);
152
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +0000153 /// Update Threshold based on callsite properties such as callee
154 /// attributes and callee hotness for PGO builds. The Callee is explicitly
155 /// passed to support analyzing indirect calls whose target is inferred by
156 /// analysis.
157 void updateThreshold(CallSite CS, Function &Callee);
158
Easwaran Raman9a3fc172016-04-08 21:28:02 +0000159 /// Return true if size growth is allowed when inlining the callee at CS.
160 bool allowSizeGrowth(CallSite CS);
161
Chandler Carruth0539c072012-03-31 12:42:41 +0000162 // Custom analysis routines.
Hal Finkel57f03dd2014-09-07 13:49:57 +0000163 bool analyzeBlock(BasicBlock *BB, SmallPtrSetImpl<const Value *> &EphValues);
Chandler Carruth0539c072012-03-31 12:42:41 +0000164
165 // Disable several entry points to the visitor so we don't accidentally use
166 // them by declaring but not defining them here.
Chad Rosier567556a2016-04-28 14:47:23 +0000167 void visit(Module *);
168 void visit(Module &);
169 void visit(Function *);
170 void visit(Function &);
171 void visit(BasicBlock *);
172 void visit(BasicBlock &);
Chandler Carruth0539c072012-03-31 12:42:41 +0000173
174 // Provide base case for our instruction visit.
175 bool visitInstruction(Instruction &I);
176
177 // Our visit overrides.
178 bool visitAlloca(AllocaInst &I);
179 bool visitPHI(PHINode &I);
180 bool visitGetElementPtr(GetElementPtrInst &I);
181 bool visitBitCast(BitCastInst &I);
182 bool visitPtrToInt(PtrToIntInst &I);
183 bool visitIntToPtr(IntToPtrInst &I);
184 bool visitCastInst(CastInst &I);
185 bool visitUnaryInstruction(UnaryInstruction &I);
Matt Arsenault727aa342013-07-20 04:09:00 +0000186 bool visitCmpInst(CmpInst &I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000187 bool visitSub(BinaryOperator &I);
188 bool visitBinaryOperator(BinaryOperator &I);
189 bool visitLoad(LoadInst &I);
190 bool visitStore(StoreInst &I);
Chandler Carruth753e21d2012-12-28 14:23:32 +0000191 bool visitExtractValue(ExtractValueInst &I);
192 bool visitInsertValue(InsertValueInst &I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000193 bool visitCallSite(CallSite CS);
Chandler Carruth0814d2a2013-12-13 07:59:56 +0000194 bool visitReturnInst(ReturnInst &RI);
195 bool visitBranchInst(BranchInst &BI);
196 bool visitSwitchInst(SwitchInst &SI);
197 bool visitIndirectBrInst(IndirectBrInst &IBI);
198 bool visitResumeInst(ResumeInst &RI);
David Majnemer654e1302015-07-31 17:58:14 +0000199 bool visitCleanupReturnInst(CleanupReturnInst &RI);
200 bool visitCatchReturnInst(CatchReturnInst &RI);
Chandler Carruth0814d2a2013-12-13 07:59:56 +0000201 bool visitUnreachableInst(UnreachableInst &I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000202
203public:
Sean Silvaab6a6832016-07-23 04:22:50 +0000204 CallAnalyzer(const TargetTransformInfo &TTI,
Daniel Jasperaec2fa32016-12-19 08:22:17 +0000205 std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
Easwaran Raman1c57cc22016-08-10 00:48:04 +0000206 ProfileSummaryInfo *PSI, Function &Callee, CallSite CSArg,
207 const InlineParams &Params)
Daniel Jasperaec2fa32016-12-19 08:22:17 +0000208 : TTI(TTI), GetAssumptionCache(GetAssumptionCache), PSI(PSI), F(Callee),
209 CandidateCS(CSArg), Params(Params), Threshold(Params.DefaultThreshold),
210 Cost(0), IsCallerRecursive(false), IsRecursiveCall(false),
211 ExposesReturnsTwice(false), HasDynamicAlloca(false),
212 ContainsNoDuplicateCall(false), HasReturn(false), HasIndirectBr(false),
213 HasFrameEscape(false), AllocatedSize(0), NumInstructions(0),
214 NumVectorInstructions(0), FiftyPercentVectorBonus(0),
215 TenPercentVectorBonus(0), VectorBonus(0), NumConstantArgs(0),
216 NumConstantOffsetPtrArgs(0), NumAllocaArgs(0), NumConstantPtrCmps(0),
217 NumConstantPtrDiffs(0), NumInstructionsSimplified(0),
218 SROACostSavings(0), SROACostSavingsLost(0) {}
Chandler Carruth0539c072012-03-31 12:42:41 +0000219
220 bool analyzeCall(CallSite CS);
221
222 int getThreshold() { return Threshold; }
223 int getCost() { return Cost; }
224
225 // Keep a bunch of stats about the cost savings found so we can print them
226 // out when debugging.
227 unsigned NumConstantArgs;
228 unsigned NumConstantOffsetPtrArgs;
229 unsigned NumAllocaArgs;
230 unsigned NumConstantPtrCmps;
231 unsigned NumConstantPtrDiffs;
232 unsigned NumInstructionsSimplified;
233 unsigned SROACostSavings;
234 unsigned SROACostSavingsLost;
235
236 void dump();
237};
238
239} // namespace
240
241/// \brief Test whether the given value is an Alloca-derived function argument.
242bool CallAnalyzer::isAllocaDerivedArg(Value *V) {
243 return SROAArgValues.count(V);
Owen Andersona08318a2010-09-09 16:56:42 +0000244}
245
Chandler Carruth0539c072012-03-31 12:42:41 +0000246/// \brief Lookup the SROA-candidate argument and cost iterator which V maps to.
247/// Returns false if V does not map to a SROA-candidate.
248bool CallAnalyzer::lookupSROAArgAndCost(
249 Value *V, Value *&Arg, DenseMap<Value *, int>::iterator &CostIt) {
250 if (SROAArgValues.empty() || SROAArgCosts.empty())
251 return false;
Chandler Carruth783b7192012-03-09 02:49:36 +0000252
Chandler Carruth0539c072012-03-31 12:42:41 +0000253 DenseMap<Value *, Value *>::iterator ArgIt = SROAArgValues.find(V);
254 if (ArgIt == SROAArgValues.end())
255 return false;
Chandler Carruth783b7192012-03-09 02:49:36 +0000256
Chandler Carruth0539c072012-03-31 12:42:41 +0000257 Arg = ArgIt->second;
258 CostIt = SROAArgCosts.find(Arg);
259 return CostIt != SROAArgCosts.end();
Chandler Carruth783b7192012-03-09 02:49:36 +0000260}
261
Chandler Carruth0539c072012-03-31 12:42:41 +0000262/// \brief Disable SROA for the candidate marked by this cost iterator.
Chandler Carruth783b7192012-03-09 02:49:36 +0000263///
Benjamin Kramerbde91762012-06-02 10:20:22 +0000264/// This marks the candidate as no longer viable for SROA, and adds the cost
Chandler Carruth0539c072012-03-31 12:42:41 +0000265/// savings associated with it back into the inline cost measurement.
266void CallAnalyzer::disableSROA(DenseMap<Value *, int>::iterator CostIt) {
267 // If we're no longer able to perform SROA we need to undo its cost savings
268 // and prevent subsequent analysis.
269 Cost += CostIt->second;
270 SROACostSavings -= CostIt->second;
271 SROACostSavingsLost += CostIt->second;
272 SROAArgCosts.erase(CostIt);
273}
274
275/// \brief If 'V' maps to a SROA candidate, disable SROA for it.
276void CallAnalyzer::disableSROA(Value *V) {
277 Value *SROAArg;
278 DenseMap<Value *, int>::iterator CostIt;
279 if (lookupSROAArgAndCost(V, SROAArg, CostIt))
280 disableSROA(CostIt);
281}
282
283/// \brief Accumulate the given cost for a particular SROA candidate.
284void CallAnalyzer::accumulateSROACost(DenseMap<Value *, int>::iterator CostIt,
285 int InstructionCost) {
286 CostIt->second += InstructionCost;
287 SROACostSavings += InstructionCost;
288}
289
Chandler Carruth0539c072012-03-31 12:42:41 +0000290/// \brief Check whether a GEP's indices are all constant.
291///
292/// Respects any simplified values known during the analysis of this callsite.
293bool CallAnalyzer::isGEPOffsetConstant(GetElementPtrInst &GEP) {
294 for (User::op_iterator I = GEP.idx_begin(), E = GEP.idx_end(); I != E; ++I)
295 if (!isa<Constant>(*I) && !SimplifiedValues.lookup(*I))
Chandler Carruth783b7192012-03-09 02:49:36 +0000296 return false;
Chandler Carruth783b7192012-03-09 02:49:36 +0000297
Chandler Carruth0539c072012-03-31 12:42:41 +0000298 return true;
299}
300
301/// \brief Accumulate a constant GEP offset into an APInt if possible.
302///
303/// Returns false if unable to compute the offset for any reason. Respects any
304/// simplified values known during the analysis of this callsite.
305bool CallAnalyzer::accumulateGEPOffset(GEPOperator &GEP, APInt &Offset) {
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000306 const DataLayout &DL = F.getParent()->getDataLayout();
307 unsigned IntPtrWidth = DL.getPointerSizeInBits();
Chandler Carruth0539c072012-03-31 12:42:41 +0000308 assert(IntPtrWidth == Offset.getBitWidth());
309
310 for (gep_type_iterator GTI = gep_type_begin(GEP), GTE = gep_type_end(GEP);
311 GTI != GTE; ++GTI) {
312 ConstantInt *OpC = dyn_cast<ConstantInt>(GTI.getOperand());
313 if (!OpC)
314 if (Constant *SimpleOp = SimplifiedValues.lookup(GTI.getOperand()))
315 OpC = dyn_cast<ConstantInt>(SimpleOp);
316 if (!OpC)
Chandler Carruth783b7192012-03-09 02:49:36 +0000317 return false;
Chad Rosier567556a2016-04-28 14:47:23 +0000318 if (OpC->isZero())
319 continue;
Chandler Carruth783b7192012-03-09 02:49:36 +0000320
Chandler Carruth0539c072012-03-31 12:42:41 +0000321 // Handle a struct index, which adds its field offset to the pointer.
Peter Collingbourneab85225b2016-12-02 02:24:42 +0000322 if (StructType *STy = GTI.getStructTypeOrNull()) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000323 unsigned ElementIdx = OpC->getZExtValue();
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000324 const StructLayout *SL = DL.getStructLayout(STy);
Chandler Carruth0539c072012-03-31 12:42:41 +0000325 Offset += APInt(IntPtrWidth, SL->getElementOffset(ElementIdx));
326 continue;
Chandler Carruth783b7192012-03-09 02:49:36 +0000327 }
Chandler Carruth783b7192012-03-09 02:49:36 +0000328
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000329 APInt TypeSize(IntPtrWidth, DL.getTypeAllocSize(GTI.getIndexedType()));
Chandler Carruth0539c072012-03-31 12:42:41 +0000330 Offset += OpC->getValue().sextOrTrunc(IntPtrWidth) * TypeSize;
331 }
332 return true;
333}
334
Haicheng Wu201b1912017-01-20 18:51:22 +0000335/// \brief Use TTI to check whether a GEP is free.
336///
337/// Respects any simplified values known during the analysis of this callsite.
338bool CallAnalyzer::isGEPFree(GetElementPtrInst &GEP) {
339 SmallVector<Value *, 4> Indices;
340 for (User::op_iterator I = GEP.idx_begin(), E = GEP.idx_end(); I != E; ++I)
341 if (Constant *SimpleOp = SimplifiedValues.lookup(*I))
342 Indices.push_back(SimpleOp);
343 else
344 Indices.push_back(*I);
345 return TargetTransformInfo::TCC_Free ==
346 TTI.getGEPCost(GEP.getSourceElementType(), GEP.getPointerOperand(),
347 Indices);
348}
349
Chandler Carruth0539c072012-03-31 12:42:41 +0000350bool CallAnalyzer::visitAlloca(AllocaInst &I) {
Eric Christopherbeb2cd62014-04-07 13:36:21 +0000351 // Check whether inlining will turn a dynamic alloca into a static
Sanjay Patel0f153422016-05-09 21:51:53 +0000352 // alloca and handle that case.
Eric Christopherbeb2cd62014-04-07 13:36:21 +0000353 if (I.isArrayAllocation()) {
Sanjay Patel0f153422016-05-09 21:51:53 +0000354 Constant *Size = SimplifiedValues.lookup(I.getArraySize());
355 if (auto *AllocSize = dyn_cast_or_null<ConstantInt>(Size)) {
Easwaran Raman22eb80a2016-06-27 22:31:53 +0000356 const DataLayout &DL = F.getParent()->getDataLayout();
Eric Christopherbeb2cd62014-04-07 13:36:21 +0000357 Type *Ty = I.getAllocatedType();
Easwaran Raman22eb80a2016-06-27 22:31:53 +0000358 AllocatedSize = SaturatingMultiplyAdd(
359 AllocSize->getLimitedValue(), DL.getTypeAllocSize(Ty), AllocatedSize);
Eric Christopherbeb2cd62014-04-07 13:36:21 +0000360 return Base::visitAlloca(I);
361 }
362 }
Chandler Carruth0539c072012-03-31 12:42:41 +0000363
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +0000364 // Accumulate the allocated size.
365 if (I.isStaticAlloca()) {
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000366 const DataLayout &DL = F.getParent()->getDataLayout();
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +0000367 Type *Ty = I.getAllocatedType();
Easwaran Raman22eb80a2016-06-27 22:31:53 +0000368 AllocatedSize = SaturatingAdd(DL.getTypeAllocSize(Ty), AllocatedSize);
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +0000369 }
370
Bob Wilsona5b0dc82012-11-19 07:04:35 +0000371 // We will happily inline static alloca instructions.
372 if (I.isStaticAlloca())
Chandler Carruth0539c072012-03-31 12:42:41 +0000373 return Base::visitAlloca(I);
374
375 // FIXME: This is overly conservative. Dynamic allocas are inefficient for
376 // a variety of reasons, and so we would like to not inline them into
377 // functions which don't currently have a dynamic alloca. This simply
378 // disables inlining altogether in the presence of a dynamic alloca.
379 HasDynamicAlloca = true;
380 return false;
381}
382
383bool CallAnalyzer::visitPHI(PHINode &I) {
384 // FIXME: We should potentially be tracking values through phi nodes,
385 // especially when they collapse to a single value due to deleted CFG edges
386 // during inlining.
387
388 // FIXME: We need to propagate SROA *disabling* through phi nodes, even
389 // though we don't want to propagate it's bonuses. The idea is to disable
390 // SROA if it *might* be used in an inappropriate manner.
391
392 // Phi nodes are always zero-cost.
393 return true;
394}
395
396bool CallAnalyzer::visitGetElementPtr(GetElementPtrInst &I) {
397 Value *SROAArg;
398 DenseMap<Value *, int>::iterator CostIt;
Chad Rosier567556a2016-04-28 14:47:23 +0000399 bool SROACandidate =
400 lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt);
Chandler Carruth0539c072012-03-31 12:42:41 +0000401
402 // Try to fold GEPs of constant-offset call site argument pointers. This
403 // requires target data and inbounds GEPs.
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000404 if (I.isInBounds()) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000405 // Check if we have a base + offset for the pointer.
406 Value *Ptr = I.getPointerOperand();
407 std::pair<Value *, APInt> BaseAndOffset = ConstantOffsetPtrs.lookup(Ptr);
408 if (BaseAndOffset.first) {
409 // Check if the offset of this GEP is constant, and if so accumulate it
410 // into Offset.
411 if (!accumulateGEPOffset(cast<GEPOperator>(I), BaseAndOffset.second)) {
412 // Non-constant GEPs aren't folded, and disable SROA.
413 if (SROACandidate)
414 disableSROA(CostIt);
Haicheng Wu201b1912017-01-20 18:51:22 +0000415 return isGEPFree(I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000416 }
417
418 // Add the result as a new mapping to Base + Offset.
419 ConstantOffsetPtrs[&I] = BaseAndOffset;
420
421 // Also handle SROA candidates here, we already know that the GEP is
422 // all-constant indexed.
423 if (SROACandidate)
424 SROAArgValues[&I] = SROAArg;
425
Chandler Carruth783b7192012-03-09 02:49:36 +0000426 return true;
427 }
428 }
429
Chandler Carruth0539c072012-03-31 12:42:41 +0000430 if (isGEPOffsetConstant(I)) {
431 if (SROACandidate)
432 SROAArgValues[&I] = SROAArg;
433
434 // Constant GEPs are modeled as free.
435 return true;
436 }
437
438 // Variable GEPs will require math and will disable SROA.
439 if (SROACandidate)
440 disableSROA(CostIt);
Haicheng Wu201b1912017-01-20 18:51:22 +0000441 return isGEPFree(I);
Chandler Carruth783b7192012-03-09 02:49:36 +0000442}
443
Chandler Carruth0539c072012-03-31 12:42:41 +0000444bool CallAnalyzer::visitBitCast(BitCastInst &I) {
445 // Propagate constants through bitcasts.
Chandler Carruth86ed5302012-12-28 14:43:42 +0000446 Constant *COp = dyn_cast<Constant>(I.getOperand(0));
447 if (!COp)
448 COp = SimplifiedValues.lookup(I.getOperand(0));
449 if (COp)
Chandler Carruth0539c072012-03-31 12:42:41 +0000450 if (Constant *C = ConstantExpr::getBitCast(COp, I.getType())) {
451 SimplifiedValues[&I] = C;
452 return true;
Owen Andersona08318a2010-09-09 16:56:42 +0000453 }
Owen Andersona08318a2010-09-09 16:56:42 +0000454
Chandler Carruth0539c072012-03-31 12:42:41 +0000455 // Track base/offsets through casts
Chad Rosier567556a2016-04-28 14:47:23 +0000456 std::pair<Value *, APInt> BaseAndOffset =
457 ConstantOffsetPtrs.lookup(I.getOperand(0));
Chandler Carruth0539c072012-03-31 12:42:41 +0000458 // Casts don't change the offset, just wrap it up.
459 if (BaseAndOffset.first)
460 ConstantOffsetPtrs[&I] = BaseAndOffset;
461
462 // Also look for SROA candidates here.
463 Value *SROAArg;
464 DenseMap<Value *, int>::iterator CostIt;
465 if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt))
466 SROAArgValues[&I] = SROAArg;
467
468 // Bitcasts are always zero cost.
469 return true;
Owen Andersona08318a2010-09-09 16:56:42 +0000470}
471
Chandler Carruth0539c072012-03-31 12:42:41 +0000472bool CallAnalyzer::visitPtrToInt(PtrToIntInst &I) {
473 // Propagate constants through ptrtoint.
Chandler Carruth86ed5302012-12-28 14:43:42 +0000474 Constant *COp = dyn_cast<Constant>(I.getOperand(0));
475 if (!COp)
476 COp = SimplifiedValues.lookup(I.getOperand(0));
477 if (COp)
Chandler Carruth0539c072012-03-31 12:42:41 +0000478 if (Constant *C = ConstantExpr::getPtrToInt(COp, I.getType())) {
479 SimplifiedValues[&I] = C;
480 return true;
Chandler Carruth4d1d34f2012-03-14 23:19:53 +0000481 }
Chandler Carruth0539c072012-03-31 12:42:41 +0000482
483 // Track base/offset pairs when converted to a plain integer provided the
484 // integer is large enough to represent the pointer.
485 unsigned IntegerSize = I.getType()->getScalarSizeInBits();
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000486 const DataLayout &DL = F.getParent()->getDataLayout();
Mehdi Amini46a43552015-03-04 18:43:29 +0000487 if (IntegerSize >= DL.getPointerSizeInBits()) {
Chad Rosier567556a2016-04-28 14:47:23 +0000488 std::pair<Value *, APInt> BaseAndOffset =
489 ConstantOffsetPtrs.lookup(I.getOperand(0));
Chandler Carruth0539c072012-03-31 12:42:41 +0000490 if (BaseAndOffset.first)
491 ConstantOffsetPtrs[&I] = BaseAndOffset;
492 }
493
494 // This is really weird. Technically, ptrtoint will disable SROA. However,
495 // unless that ptrtoint is *used* somewhere in the live basic blocks after
496 // inlining, it will be nuked, and SROA should proceed. All of the uses which
497 // would block SROA would also block SROA if applied directly to a pointer,
498 // and so we can just add the integer in here. The only places where SROA is
499 // preserved either cannot fire on an integer, or won't in-and-of themselves
500 // disable SROA (ext) w/o some later use that we would see and disable.
501 Value *SROAArg;
502 DenseMap<Value *, int>::iterator CostIt;
503 if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt))
504 SROAArgValues[&I] = SROAArg;
505
Chandler Carruthb8cf5102013-01-21 12:05:16 +0000506 return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
Chandler Carruth4d1d34f2012-03-14 23:19:53 +0000507}
508
Chandler Carruth0539c072012-03-31 12:42:41 +0000509bool CallAnalyzer::visitIntToPtr(IntToPtrInst &I) {
510 // Propagate constants through ptrtoint.
Chandler Carruth86ed5302012-12-28 14:43:42 +0000511 Constant *COp = dyn_cast<Constant>(I.getOperand(0));
512 if (!COp)
513 COp = SimplifiedValues.lookup(I.getOperand(0));
514 if (COp)
Chandler Carruth0539c072012-03-31 12:42:41 +0000515 if (Constant *C = ConstantExpr::getIntToPtr(COp, I.getType())) {
516 SimplifiedValues[&I] = C;
517 return true;
518 }
Dan Gohman4552e3c2009-10-13 18:30:07 +0000519
Chandler Carruth0539c072012-03-31 12:42:41 +0000520 // Track base/offset pairs when round-tripped through a pointer without
521 // modifications provided the integer is not too large.
522 Value *Op = I.getOperand(0);
523 unsigned IntegerSize = Op->getType()->getScalarSizeInBits();
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000524 const DataLayout &DL = F.getParent()->getDataLayout();
Mehdi Amini46a43552015-03-04 18:43:29 +0000525 if (IntegerSize <= DL.getPointerSizeInBits()) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000526 std::pair<Value *, APInt> BaseAndOffset = ConstantOffsetPtrs.lookup(Op);
527 if (BaseAndOffset.first)
528 ConstantOffsetPtrs[&I] = BaseAndOffset;
529 }
Dan Gohman4552e3c2009-10-13 18:30:07 +0000530
Chandler Carruth0539c072012-03-31 12:42:41 +0000531 // "Propagate" SROA here in the same manner as we do for ptrtoint above.
532 Value *SROAArg;
533 DenseMap<Value *, int>::iterator CostIt;
534 if (lookupSROAArgAndCost(Op, SROAArg, CostIt))
535 SROAArgValues[&I] = SROAArg;
Chandler Carruth4d1d34f2012-03-14 23:19:53 +0000536
Chandler Carruthb8cf5102013-01-21 12:05:16 +0000537 return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000538}
539
540bool CallAnalyzer::visitCastInst(CastInst &I) {
541 // Propagate constants through ptrtoint.
Chandler Carruth86ed5302012-12-28 14:43:42 +0000542 Constant *COp = dyn_cast<Constant>(I.getOperand(0));
543 if (!COp)
544 COp = SimplifiedValues.lookup(I.getOperand(0));
545 if (COp)
Chandler Carruth0539c072012-03-31 12:42:41 +0000546 if (Constant *C = ConstantExpr::getCast(I.getOpcode(), COp, I.getType())) {
547 SimplifiedValues[&I] = C;
548 return true;
549 }
550
551 // Disable SROA in the face of arbitrary casts we don't whitelist elsewhere.
552 disableSROA(I.getOperand(0));
553
Chandler Carruthb8cf5102013-01-21 12:05:16 +0000554 return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
Chandler Carruth0539c072012-03-31 12:42:41 +0000555}
556
557bool CallAnalyzer::visitUnaryInstruction(UnaryInstruction &I) {
558 Value *Operand = I.getOperand(0);
Jakub Staszak7b9e0b92013-03-07 20:01:19 +0000559 Constant *COp = dyn_cast<Constant>(Operand);
560 if (!COp)
561 COp = SimplifiedValues.lookup(Operand);
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000562 if (COp) {
563 const DataLayout &DL = F.getParent()->getDataLayout();
Manuel Jacobe9024592016-01-21 06:33:22 +0000564 if (Constant *C = ConstantFoldInstOperands(&I, COp, DL)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000565 SimplifiedValues[&I] = C;
566 return true;
567 }
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000568 }
Chandler Carruth0539c072012-03-31 12:42:41 +0000569
570 // Disable any SROA on the argument to arbitrary unary operators.
571 disableSROA(Operand);
572
573 return false;
574}
575
Philip Reames9b5c9582015-06-26 20:51:17 +0000576bool CallAnalyzer::paramHasAttr(Argument *A, Attribute::AttrKind Attr) {
577 unsigned ArgNo = A->getArgNo();
Chad Rosier567556a2016-04-28 14:47:23 +0000578 return CandidateCS.paramHasAttr(ArgNo + 1, Attr);
Philip Reames9b5c9582015-06-26 20:51:17 +0000579}
580
581bool CallAnalyzer::isKnownNonNullInCallee(Value *V) {
582 // Does the *call site* have the NonNull attribute set on an argument? We
583 // use the attribute on the call site to memoize any analysis done in the
584 // caller. This will also trip if the callee function has a non-null
585 // parameter attribute, but that's a less interesting case because hopefully
586 // the callee would already have been simplified based on that.
587 if (Argument *A = dyn_cast<Argument>(V))
588 if (paramHasAttr(A, Attribute::NonNull))
589 return true;
Chad Rosier567556a2016-04-28 14:47:23 +0000590
Philip Reames9b5c9582015-06-26 20:51:17 +0000591 // Is this an alloca in the caller? This is distinct from the attribute case
592 // above because attributes aren't updated within the inliner itself and we
593 // always want to catch the alloca derived case.
594 if (isAllocaDerivedArg(V))
595 // We can actually predict the result of comparisons between an
596 // alloca-derived value and null. Note that this fires regardless of
597 // SROA firing.
598 return true;
Chad Rosier567556a2016-04-28 14:47:23 +0000599
Philip Reames9b5c9582015-06-26 20:51:17 +0000600 return false;
601}
602
Easwaran Raman9a3fc172016-04-08 21:28:02 +0000603bool CallAnalyzer::allowSizeGrowth(CallSite CS) {
604 // If the normal destination of the invoke or the parent block of the call
605 // site is unreachable-terminated, there is little point in inlining this
606 // unless there is literally zero cost.
607 // FIXME: Note that it is possible that an unreachable-terminated block has a
608 // hot entry. For example, in below scenario inlining hot_call_X() may be
609 // beneficial :
610 // main() {
611 // hot_call_1();
612 // ...
613 // hot_call_N()
614 // exit(0);
615 // }
616 // For now, we are not handling this corner case here as it is rare in real
617 // code. In future, we should elaborate this based on BPI and BFI in more
618 // general threshold adjusting heuristics in updateThreshold().
619 Instruction *Instr = CS.getInstruction();
620 if (InvokeInst *II = dyn_cast<InvokeInst>(Instr)) {
621 if (isa<UnreachableInst>(II->getNormalDest()->getTerminator()))
622 return false;
623 } else if (isa<UnreachableInst>(Instr->getParent()->getTerminator()))
624 return false;
625
626 return true;
627}
628
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +0000629void CallAnalyzer::updateThreshold(CallSite CS, Function &Callee) {
Easwaran Raman9a3fc172016-04-08 21:28:02 +0000630 // If no size growth is allowed for this inlining, set Threshold to 0.
631 if (!allowSizeGrowth(CS)) {
632 Threshold = 0;
633 return;
634 }
635
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +0000636 Function *Caller = CS.getCaller();
Easwaran Raman1c57cc22016-08-10 00:48:04 +0000637
638 // return min(A, B) if B is valid.
639 auto MinIfValid = [](int A, Optional<int> B) {
640 return B ? std::min(A, B.getValue()) : A;
641 };
642
Easwaran Raman0d58fca2016-08-11 03:58:05 +0000643 // return max(A, B) if B is valid.
644 auto MaxIfValid = [](int A, Optional<int> B) {
645 return B ? std::max(A, B.getValue()) : A;
646 };
647
Easwaran Raman1c57cc22016-08-10 00:48:04 +0000648 // Use the OptMinSizeThreshold or OptSizeThreshold knob if they are available
649 // and reduce the threshold if the caller has the necessary attribute.
650 if (Caller->optForMinSize())
651 Threshold = MinIfValid(Threshold, Params.OptMinSizeThreshold);
652 else if (Caller->optForSize())
653 Threshold = MinIfValid(Threshold, Params.OptSizeThreshold);
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +0000654
Easwaran Ramane08b1392017-01-09 21:56:26 +0000655 // Adjust the threshold based on inlinehint attribute and profile based
656 // hotness information if the caller does not have MinSize attribute.
657 if (!Caller->optForMinSize()) {
658 if (Callee.hasFnAttribute(Attribute::InlineHint))
659 Threshold = MaxIfValid(Threshold, Params.HintThreshold);
660 if (PSI) {
661 uint64_t TotalWeight;
662 if (CS.getInstruction()->extractProfTotalWeight(TotalWeight) &&
663 PSI->isHotCount(TotalWeight)) {
664 Threshold = MaxIfValid(Threshold, Params.HotCallSiteThreshold);
665 } else if (PSI->isFunctionEntryHot(&Callee)) {
666 // If callsite hotness can not be determined, we may still know
667 // that the callee is hot and treat it as a weaker hint for threshold
668 // increase.
669 Threshold = MaxIfValid(Threshold, Params.HintThreshold);
670 } else if (PSI->isFunctionEntryCold(&Callee)) {
671 Threshold = MinIfValid(Threshold, Params.ColdThreshold);
672 }
673 }
Dehao Chene1c7c572016-08-05 20:49:04 +0000674 }
Dehao Chen9232f982016-07-11 16:48:54 +0000675
Justin Lebar8650a4d2016-04-15 01:38:48 +0000676 // Finally, take the target-specific inlining threshold multiplier into
677 // account.
678 Threshold *= TTI.getInliningThresholdMultiplier();
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +0000679}
680
Matt Arsenault727aa342013-07-20 04:09:00 +0000681bool CallAnalyzer::visitCmpInst(CmpInst &I) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000682 Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
683 // First try to handle simplified comparisons.
684 if (!isa<Constant>(LHS))
685 if (Constant *SimpleLHS = SimplifiedValues.lookup(LHS))
686 LHS = SimpleLHS;
687 if (!isa<Constant>(RHS))
688 if (Constant *SimpleRHS = SimplifiedValues.lookup(RHS))
689 RHS = SimpleRHS;
Matt Arsenault727aa342013-07-20 04:09:00 +0000690 if (Constant *CLHS = dyn_cast<Constant>(LHS)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000691 if (Constant *CRHS = dyn_cast<Constant>(RHS))
Chad Rosier567556a2016-04-28 14:47:23 +0000692 if (Constant *C =
693 ConstantExpr::getCompare(I.getPredicate(), CLHS, CRHS)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000694 SimplifiedValues[&I] = C;
695 return true;
696 }
Matt Arsenault727aa342013-07-20 04:09:00 +0000697 }
698
699 if (I.getOpcode() == Instruction::FCmp)
700 return false;
Chandler Carruth0539c072012-03-31 12:42:41 +0000701
702 // Otherwise look for a comparison between constant offset pointers with
703 // a common base.
704 Value *LHSBase, *RHSBase;
705 APInt LHSOffset, RHSOffset;
Benjamin Kramerd6f1f842014-03-02 13:30:33 +0000706 std::tie(LHSBase, LHSOffset) = ConstantOffsetPtrs.lookup(LHS);
Chandler Carruth0539c072012-03-31 12:42:41 +0000707 if (LHSBase) {
Benjamin Kramerd6f1f842014-03-02 13:30:33 +0000708 std::tie(RHSBase, RHSOffset) = ConstantOffsetPtrs.lookup(RHS);
Chandler Carruth0539c072012-03-31 12:42:41 +0000709 if (RHSBase && LHSBase == RHSBase) {
710 // We have common bases, fold the icmp to a constant based on the
711 // offsets.
712 Constant *CLHS = ConstantInt::get(LHS->getContext(), LHSOffset);
713 Constant *CRHS = ConstantInt::get(RHS->getContext(), RHSOffset);
714 if (Constant *C = ConstantExpr::getICmp(I.getPredicate(), CLHS, CRHS)) {
715 SimplifiedValues[&I] = C;
716 ++NumConstantPtrCmps;
717 return true;
718 }
719 }
720 }
721
722 // If the comparison is an equality comparison with null, we can simplify it
Philip Reames9b5c9582015-06-26 20:51:17 +0000723 // if we know the value (argument) can't be null
724 if (I.isEquality() && isa<ConstantPointerNull>(I.getOperand(1)) &&
725 isKnownNonNullInCallee(I.getOperand(0))) {
726 bool IsNotEqual = I.getPredicate() == CmpInst::ICMP_NE;
727 SimplifiedValues[&I] = IsNotEqual ? ConstantInt::getTrue(I.getType())
728 : ConstantInt::getFalse(I.getType());
729 return true;
730 }
Chandler Carruth0539c072012-03-31 12:42:41 +0000731 // Finally check for SROA candidates in comparisons.
732 Value *SROAArg;
733 DenseMap<Value *, int>::iterator CostIt;
734 if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt)) {
735 if (isa<ConstantPointerNull>(I.getOperand(1))) {
736 accumulateSROACost(CostIt, InlineConstants::InstrCost);
737 return true;
738 }
739
740 disableSROA(CostIt);
741 }
742
743 return false;
744}
745
746bool CallAnalyzer::visitSub(BinaryOperator &I) {
747 // Try to handle a special case: we can fold computing the difference of two
748 // constant-related pointers.
749 Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
750 Value *LHSBase, *RHSBase;
751 APInt LHSOffset, RHSOffset;
Benjamin Kramerd6f1f842014-03-02 13:30:33 +0000752 std::tie(LHSBase, LHSOffset) = ConstantOffsetPtrs.lookup(LHS);
Chandler Carruth0539c072012-03-31 12:42:41 +0000753 if (LHSBase) {
Benjamin Kramerd6f1f842014-03-02 13:30:33 +0000754 std::tie(RHSBase, RHSOffset) = ConstantOffsetPtrs.lookup(RHS);
Chandler Carruth0539c072012-03-31 12:42:41 +0000755 if (RHSBase && LHSBase == RHSBase) {
756 // We have common bases, fold the subtract to a constant based on the
757 // offsets.
758 Constant *CLHS = ConstantInt::get(LHS->getContext(), LHSOffset);
759 Constant *CRHS = ConstantInt::get(RHS->getContext(), RHSOffset);
760 if (Constant *C = ConstantExpr::getSub(CLHS, CRHS)) {
761 SimplifiedValues[&I] = C;
762 ++NumConstantPtrDiffs;
763 return true;
764 }
765 }
766 }
767
768 // Otherwise, fall back to the generic logic for simplifying and handling
769 // instructions.
770 return Base::visitSub(I);
771}
772
773bool CallAnalyzer::visitBinaryOperator(BinaryOperator &I) {
774 Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
Mehdi Aminia28d91d2015-03-10 02:37:25 +0000775 const DataLayout &DL = F.getParent()->getDataLayout();
Chandler Carruth0539c072012-03-31 12:42:41 +0000776 if (!isa<Constant>(LHS))
777 if (Constant *SimpleLHS = SimplifiedValues.lookup(LHS))
778 LHS = SimpleLHS;
779 if (!isa<Constant>(RHS))
780 if (Constant *SimpleRHS = SimplifiedValues.lookup(RHS))
781 RHS = SimpleRHS;
Michael Zolotukhin4e8598e2015-02-06 20:02:51 +0000782 Value *SimpleV = nullptr;
783 if (auto FI = dyn_cast<FPMathOperator>(&I))
784 SimpleV =
785 SimplifyFPBinOp(I.getOpcode(), LHS, RHS, FI->getFastMathFlags(), DL);
786 else
787 SimpleV = SimplifyBinOp(I.getOpcode(), LHS, RHS, DL);
788
Chandler Carruth0539c072012-03-31 12:42:41 +0000789 if (Constant *C = dyn_cast_or_null<Constant>(SimpleV)) {
790 SimplifiedValues[&I] = C;
791 return true;
792 }
793
794 // Disable any SROA on arguments to arbitrary, unsimplified binary operators.
795 disableSROA(LHS);
796 disableSROA(RHS);
797
798 return false;
799}
800
801bool CallAnalyzer::visitLoad(LoadInst &I) {
802 Value *SROAArg;
803 DenseMap<Value *, int>::iterator CostIt;
Wei Mi6c428d62015-03-20 18:33:12 +0000804 if (lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000805 if (I.isSimple()) {
806 accumulateSROACost(CostIt, InlineConstants::InstrCost);
807 return true;
808 }
809
810 disableSROA(CostIt);
811 }
812
813 return false;
814}
815
816bool CallAnalyzer::visitStore(StoreInst &I) {
817 Value *SROAArg;
818 DenseMap<Value *, int>::iterator CostIt;
Wei Mi6c428d62015-03-20 18:33:12 +0000819 if (lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000820 if (I.isSimple()) {
821 accumulateSROACost(CostIt, InlineConstants::InstrCost);
822 return true;
823 }
824
825 disableSROA(CostIt);
826 }
827
828 return false;
829}
830
Chandler Carruth753e21d2012-12-28 14:23:32 +0000831bool CallAnalyzer::visitExtractValue(ExtractValueInst &I) {
832 // Constant folding for extract value is trivial.
833 Constant *C = dyn_cast<Constant>(I.getAggregateOperand());
834 if (!C)
835 C = SimplifiedValues.lookup(I.getAggregateOperand());
836 if (C) {
837 SimplifiedValues[&I] = ConstantExpr::getExtractValue(C, I.getIndices());
838 return true;
839 }
840
841 // SROA can look through these but give them a cost.
842 return false;
843}
844
845bool CallAnalyzer::visitInsertValue(InsertValueInst &I) {
846 // Constant folding for insert value is trivial.
847 Constant *AggC = dyn_cast<Constant>(I.getAggregateOperand());
848 if (!AggC)
849 AggC = SimplifiedValues.lookup(I.getAggregateOperand());
850 Constant *InsertedC = dyn_cast<Constant>(I.getInsertedValueOperand());
851 if (!InsertedC)
852 InsertedC = SimplifiedValues.lookup(I.getInsertedValueOperand());
853 if (AggC && InsertedC) {
Chad Rosier567556a2016-04-28 14:47:23 +0000854 SimplifiedValues[&I] =
855 ConstantExpr::getInsertValue(AggC, InsertedC, I.getIndices());
Chandler Carruth753e21d2012-12-28 14:23:32 +0000856 return true;
857 }
858
859 // SROA can look through these but give them a cost.
860 return false;
861}
862
863/// \brief Try to simplify a call site.
864///
865/// Takes a concrete function and callsite and tries to actually simplify it by
866/// analyzing the arguments and call itself with instsimplify. Returns true if
867/// it has simplified the callsite to some other entity (a constant), making it
868/// free.
869bool CallAnalyzer::simplifyCallSite(Function *F, CallSite CS) {
870 // FIXME: Using the instsimplify logic directly for this is inefficient
871 // because we have to continually rebuild the argument list even when no
872 // simplifications can be performed. Until that is fixed with remapping
873 // inside of instsimplify, directly constant fold calls here.
874 if (!canConstantFoldCallTo(F))
875 return false;
876
877 // Try to re-map the arguments to constants.
878 SmallVector<Constant *, 4> ConstantArgs;
879 ConstantArgs.reserve(CS.arg_size());
Chad Rosier567556a2016-04-28 14:47:23 +0000880 for (CallSite::arg_iterator I = CS.arg_begin(), E = CS.arg_end(); I != E;
881 ++I) {
Chandler Carruth753e21d2012-12-28 14:23:32 +0000882 Constant *C = dyn_cast<Constant>(*I);
883 if (!C)
884 C = dyn_cast_or_null<Constant>(SimplifiedValues.lookup(*I));
885 if (!C)
886 return false; // This argument doesn't map to a constant.
887
888 ConstantArgs.push_back(C);
889 }
890 if (Constant *C = ConstantFoldCall(F, ConstantArgs)) {
891 SimplifiedValues[CS.getInstruction()] = C;
892 return true;
893 }
894
895 return false;
896}
897
Chandler Carruth0539c072012-03-31 12:42:41 +0000898bool CallAnalyzer::visitCallSite(CallSite CS) {
Chandler Carruth37d25de2013-12-13 08:00:01 +0000899 if (CS.hasFnAttr(Attribute::ReturnsTwice) &&
Duncan P. N. Exon Smithb3fc83c2015-02-14 00:12:15 +0000900 !F.hasFnAttribute(Attribute::ReturnsTwice)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000901 // This aborts the entire analysis.
902 ExposesReturnsTwice = true;
903 return false;
904 }
Chad Rosier567556a2016-04-28 14:47:23 +0000905 if (CS.isCall() && cast<CallInst>(CS.getInstruction())->cannotDuplicate())
James Molloy4f6fb952012-12-20 16:04:27 +0000906 ContainsNoDuplicateCall = true;
Chandler Carruth0539c072012-03-31 12:42:41 +0000907
Chandler Carruth0539c072012-03-31 12:42:41 +0000908 if (Function *F = CS.getCalledFunction()) {
Chandler Carruth753e21d2012-12-28 14:23:32 +0000909 // When we have a concrete function, first try to simplify it directly.
910 if (simplifyCallSite(F, CS))
911 return true;
912
913 // Next check if it is an intrinsic we know about.
914 // FIXME: Lift this into part of the InstVisitor.
915 if (IntrinsicInst *II = dyn_cast<IntrinsicInst>(CS.getInstruction())) {
916 switch (II->getIntrinsicID()) {
917 default:
918 return Base::visitCallSite(CS);
919
Peter Collingbourne7dd8dbf2016-04-22 21:18:02 +0000920 case Intrinsic::load_relative:
921 // This is normally lowered to 4 LLVM instructions.
922 Cost += 3 * InlineConstants::InstrCost;
923 return false;
924
Chandler Carruth753e21d2012-12-28 14:23:32 +0000925 case Intrinsic::memset:
926 case Intrinsic::memcpy:
927 case Intrinsic::memmove:
928 // SROA can usually chew through these intrinsics, but they aren't free.
929 return false;
Reid Kleckner60381792015-07-07 22:25:32 +0000930 case Intrinsic::localescape:
Reid Kleckner223de262015-04-14 20:38:14 +0000931 HasFrameEscape = true;
932 return false;
Chandler Carruth753e21d2012-12-28 14:23:32 +0000933 }
934 }
935
Chandler Carruth0539c072012-03-31 12:42:41 +0000936 if (F == CS.getInstruction()->getParent()->getParent()) {
937 // This flag will fully abort the analysis, so don't bother with anything
938 // else.
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +0000939 IsRecursiveCall = true;
Chandler Carruth0539c072012-03-31 12:42:41 +0000940 return false;
941 }
942
Chandler Carruth0ba8db42013-01-22 11:26:02 +0000943 if (TTI.isLoweredToCall(F)) {
Chandler Carruth0539c072012-03-31 12:42:41 +0000944 // We account for the average 1 instruction per call argument setup
945 // here.
946 Cost += CS.arg_size() * InlineConstants::InstrCost;
947
948 // Everything other than inline ASM will also have a significant cost
949 // merely from making the call.
950 if (!isa<InlineAsm>(CS.getCalledValue()))
951 Cost += InlineConstants::CallPenalty;
952 }
953
954 return Base::visitCallSite(CS);
955 }
956
957 // Otherwise we're in a very special case -- an indirect function call. See
958 // if we can be particularly clever about this.
959 Value *Callee = CS.getCalledValue();
960
961 // First, pay the price of the argument setup. We account for the average
962 // 1 instruction per call argument setup here.
963 Cost += CS.arg_size() * InlineConstants::InstrCost;
964
965 // Next, check if this happens to be an indirect function call to a known
966 // function in this inline context. If not, we've done all we can.
967 Function *F = dyn_cast_or_null<Function>(SimplifiedValues.lookup(Callee));
968 if (!F)
969 return Base::visitCallSite(CS);
970
971 // If we have a constant that we are calling as a function, we can peer
972 // through it and see the function target. This happens not infrequently
973 // during devirtualization and so we want to give it a hefty bonus for
974 // inlining, but cap that bonus in the event that inlining wouldn't pan
975 // out. Pretend to inline the function, with a custom threshold.
Easwaran Raman1c57cc22016-08-10 00:48:04 +0000976 auto IndirectCallParams = Params;
977 IndirectCallParams.DefaultThreshold = InlineConstants::IndirectCallThreshold;
Daniel Jasperaec2fa32016-12-19 08:22:17 +0000978 CallAnalyzer CA(TTI, GetAssumptionCache, PSI, *F, CS, IndirectCallParams);
Chandler Carruth0539c072012-03-31 12:42:41 +0000979 if (CA.analyzeCall(CS)) {
980 // We were able to inline the indirect call! Subtract the cost from the
Easwaran Raman6d90d9f2015-12-07 21:21:20 +0000981 // threshold to get the bonus we want to apply, but don't go below zero.
982 Cost -= std::max(0, CA.getThreshold() - CA.getCost());
Chandler Carruth0539c072012-03-31 12:42:41 +0000983 }
984
985 return Base::visitCallSite(CS);
986}
987
Chandler Carruth0814d2a2013-12-13 07:59:56 +0000988bool CallAnalyzer::visitReturnInst(ReturnInst &RI) {
989 // At least one return instruction will be free after inlining.
990 bool Free = !HasReturn;
991 HasReturn = true;
992 return Free;
993}
994
995bool CallAnalyzer::visitBranchInst(BranchInst &BI) {
996 // We model unconditional branches as essentially free -- they really
997 // shouldn't exist at all, but handling them makes the behavior of the
998 // inliner more regular and predictable. Interestingly, conditional branches
999 // which will fold away are also free.
1000 return BI.isUnconditional() || isa<ConstantInt>(BI.getCondition()) ||
1001 dyn_cast_or_null<ConstantInt>(
1002 SimplifiedValues.lookup(BI.getCondition()));
1003}
1004
1005bool CallAnalyzer::visitSwitchInst(SwitchInst &SI) {
1006 // We model unconditional switches as free, see the comments on handling
1007 // branches.
Chandler Carruthe01fd5f2014-04-28 08:52:44 +00001008 if (isa<ConstantInt>(SI.getCondition()))
1009 return true;
1010 if (Value *V = SimplifiedValues.lookup(SI.getCondition()))
1011 if (isa<ConstantInt>(V))
1012 return true;
1013
1014 // Otherwise, we need to accumulate a cost proportional to the number of
1015 // distinct successor blocks. This fan-out in the CFG cannot be represented
1016 // for free even if we can represent the core switch as a jumptable that
1017 // takes a single instruction.
1018 //
1019 // NB: We convert large switches which are just used to initialize large phi
1020 // nodes to lookup tables instead in simplify-cfg, so this shouldn't prevent
1021 // inlining those. It will prevent inlining in cases where the optimization
1022 // does not (yet) fire.
1023 SmallPtrSet<BasicBlock *, 8> SuccessorBlocks;
1024 SuccessorBlocks.insert(SI.getDefaultDest());
1025 for (auto I = SI.case_begin(), E = SI.case_end(); I != E; ++I)
1026 SuccessorBlocks.insert(I.getCaseSuccessor());
1027 // Add cost corresponding to the number of distinct destinations. The first
1028 // we model as free because of fallthrough.
1029 Cost += (SuccessorBlocks.size() - 1) * InlineConstants::InstrCost;
1030 return false;
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001031}
1032
1033bool CallAnalyzer::visitIndirectBrInst(IndirectBrInst &IBI) {
1034 // We never want to inline functions that contain an indirectbr. This is
1035 // incorrect because all the blockaddress's (in static global initializers
1036 // for example) would be referring to the original function, and this
1037 // indirect jump would jump from the inlined copy of the function into the
1038 // original function which is extremely undefined behavior.
1039 // FIXME: This logic isn't really right; we can safely inline functions with
1040 // indirectbr's as long as no other function or global references the
Gerolf Hoflehner734f4c82014-07-01 00:19:34 +00001041 // blockaddress of a block within the current function.
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001042 HasIndirectBr = true;
1043 return false;
1044}
1045
1046bool CallAnalyzer::visitResumeInst(ResumeInst &RI) {
1047 // FIXME: It's not clear that a single instruction is an accurate model for
1048 // the inline cost of a resume instruction.
1049 return false;
1050}
1051
David Majnemer654e1302015-07-31 17:58:14 +00001052bool CallAnalyzer::visitCleanupReturnInst(CleanupReturnInst &CRI) {
1053 // FIXME: It's not clear that a single instruction is an accurate model for
1054 // the inline cost of a cleanupret instruction.
1055 return false;
1056}
1057
1058bool CallAnalyzer::visitCatchReturnInst(CatchReturnInst &CRI) {
1059 // FIXME: It's not clear that a single instruction is an accurate model for
Joseph Tremoulet8220bcc2015-08-23 00:26:33 +00001060 // the inline cost of a catchret instruction.
David Majnemer654e1302015-07-31 17:58:14 +00001061 return false;
1062}
1063
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001064bool CallAnalyzer::visitUnreachableInst(UnreachableInst &I) {
1065 // FIXME: It might be reasonably to discount the cost of instructions leading
1066 // to unreachable as they have the lowest possible impact on both runtime and
1067 // code size.
1068 return true; // No actual code is needed for unreachable.
1069}
1070
Chandler Carruth0539c072012-03-31 12:42:41 +00001071bool CallAnalyzer::visitInstruction(Instruction &I) {
Chandler Carruthda7513a2012-05-04 00:58:03 +00001072 // Some instructions are free. All of the free intrinsics can also be
1073 // handled by SROA, etc.
Chandler Carruthb8cf5102013-01-21 12:05:16 +00001074 if (TargetTransformInfo::TCC_Free == TTI.getUserCost(&I))
Chandler Carruthda7513a2012-05-04 00:58:03 +00001075 return true;
1076
Chandler Carruth0539c072012-03-31 12:42:41 +00001077 // We found something we don't understand or can't handle. Mark any SROA-able
1078 // values in the operand list as no longer viable.
1079 for (User::op_iterator OI = I.op_begin(), OE = I.op_end(); OI != OE; ++OI)
1080 disableSROA(*OI);
1081
1082 return false;
1083}
1084
Chandler Carruth0539c072012-03-31 12:42:41 +00001085/// \brief Analyze a basic block for its contribution to the inline cost.
1086///
1087/// This method walks the analyzer over every instruction in the given basic
1088/// block and accounts for their cost during inlining at this callsite. It
1089/// aborts early if the threshold has been exceeded or an impossible to inline
1090/// construct has been detected. It returns false if inlining is no longer
1091/// viable, and true if inlining remains viable.
Hal Finkel57f03dd2014-09-07 13:49:57 +00001092bool CallAnalyzer::analyzeBlock(BasicBlock *BB,
1093 SmallPtrSetImpl<const Value *> &EphValues) {
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001094 for (BasicBlock::iterator I = BB->begin(), E = BB->end(); I != E; ++I) {
Chandler Carruth6b4cc8b2014-02-01 10:38:17 +00001095 // FIXME: Currently, the number of instructions in a function regardless of
1096 // our ability to simplify them during inline to constants or dead code,
1097 // are actually used by the vector bonus heuristic. As long as that's true,
1098 // we have to special case debug intrinsics here to prevent differences in
1099 // inlining due to debug symbols. Eventually, the number of unsimplified
1100 // instructions shouldn't factor into the cost computation, but until then,
1101 // hack around it here.
1102 if (isa<DbgInfoIntrinsic>(I))
1103 continue;
1104
Hal Finkel57f03dd2014-09-07 13:49:57 +00001105 // Skip ephemeral values.
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001106 if (EphValues.count(&*I))
Hal Finkel57f03dd2014-09-07 13:49:57 +00001107 continue;
1108
Chandler Carruth0539c072012-03-31 12:42:41 +00001109 ++NumInstructions;
1110 if (isa<ExtractElementInst>(I) || I->getType()->isVectorTy())
1111 ++NumVectorInstructions;
1112
Sanjay Patele9434e82015-09-15 15:26:25 +00001113 // If the instruction is floating point, and the target says this operation
1114 // is expensive or the function has the "use-soft-float" attribute, this may
1115 // eventually become a library call. Treat the cost as such.
Cameron Esfahani17177d12015-02-05 02:09:33 +00001116 if (I->getType()->isFloatingPointTy()) {
1117 bool hasSoftFloatAttr = false;
1118
Sanjay Patele9434e82015-09-15 15:26:25 +00001119 // If the function has the "use-soft-float" attribute, mark it as
1120 // expensive.
Cameron Esfahani17177d12015-02-05 02:09:33 +00001121 if (F.hasFnAttribute("use-soft-float")) {
1122 Attribute Attr = F.getFnAttribute("use-soft-float");
1123 StringRef Val = Attr.getValueAsString();
1124 if (Val == "true")
1125 hasSoftFloatAttr = true;
1126 }
1127
1128 if (TTI.getFPOpCost(I->getType()) == TargetTransformInfo::TCC_Expensive ||
1129 hasSoftFloatAttr)
1130 Cost += InlineConstants::CallPenalty;
1131 }
1132
Chandler Carruth0539c072012-03-31 12:42:41 +00001133 // If the instruction simplified to a constant, there is no cost to this
1134 // instruction. Visit the instructions using our InstVisitor to account for
1135 // all of the per-instruction logic. The visit tree returns true if we
1136 // consumed the instruction in any way, and false if the instruction's base
1137 // cost should count against inlining.
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001138 if (Base::visit(&*I))
Chandler Carruth0539c072012-03-31 12:42:41 +00001139 ++NumInstructionsSimplified;
1140 else
1141 Cost += InlineConstants::InstrCost;
1142
1143 // If the visit this instruction detected an uninlinable pattern, abort.
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001144 if (IsRecursiveCall || ExposesReturnsTwice || HasDynamicAlloca ||
Reid Kleckner223de262015-04-14 20:38:14 +00001145 HasIndirectBr || HasFrameEscape)
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +00001146 return false;
1147
1148 // If the caller is a recursive function then we don't want to inline
1149 // functions which allocate a lot of stack space because it would increase
1150 // the caller stack usage dramatically.
1151 if (IsCallerRecursive &&
1152 AllocatedSize > InlineConstants::TotalAllocaSizeRecursiveCaller)
Chandler Carruth0539c072012-03-31 12:42:41 +00001153 return false;
1154
Chandler Carrutha004f222015-05-27 02:49:05 +00001155 // Check if we've past the maximum possible threshold so we don't spin in
1156 // huge basic blocks that will never inline.
1157 if (Cost > Threshold)
Chandler Carruth0539c072012-03-31 12:42:41 +00001158 return false;
1159 }
1160
1161 return true;
1162}
1163
1164/// \brief Compute the base pointer and cumulative constant offsets for V.
1165///
1166/// This strips all constant offsets off of V, leaving it the base pointer, and
1167/// accumulates the total constant offset applied in the returned constant. It
1168/// returns 0 if V is not a pointer, and returns the constant '0' if there are
1169/// no constant offsets applied.
1170ConstantInt *CallAnalyzer::stripAndComputeInBoundsConstantOffsets(Value *&V) {
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001171 if (!V->getType()->isPointerTy())
Craig Topper353eda42014-04-24 06:44:33 +00001172 return nullptr;
Chandler Carruth0539c072012-03-31 12:42:41 +00001173
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001174 const DataLayout &DL = F.getParent()->getDataLayout();
1175 unsigned IntPtrWidth = DL.getPointerSizeInBits();
Chandler Carruth0539c072012-03-31 12:42:41 +00001176 APInt Offset = APInt::getNullValue(IntPtrWidth);
1177
1178 // Even though we don't look through PHI nodes, we could be called on an
1179 // instruction in an unreachable block, which may be on a cycle.
1180 SmallPtrSet<Value *, 4> Visited;
1181 Visited.insert(V);
1182 do {
1183 if (GEPOperator *GEP = dyn_cast<GEPOperator>(V)) {
1184 if (!GEP->isInBounds() || !accumulateGEPOffset(*GEP, Offset))
Craig Topper353eda42014-04-24 06:44:33 +00001185 return nullptr;
Chandler Carruth0539c072012-03-31 12:42:41 +00001186 V = GEP->getPointerOperand();
1187 } else if (Operator::getOpcode(V) == Instruction::BitCast) {
1188 V = cast<Operator>(V)->getOperand(0);
1189 } else if (GlobalAlias *GA = dyn_cast<GlobalAlias>(V)) {
Sanjoy Das5ce32722016-04-08 00:48:30 +00001190 if (GA->isInterposable())
Chandler Carruth0539c072012-03-31 12:42:41 +00001191 break;
1192 V = GA->getAliasee();
1193 } else {
1194 break;
1195 }
1196 assert(V->getType()->isPointerTy() && "Unexpected operand type!");
David Blaikie70573dc2014-11-19 07:49:26 +00001197 } while (Visited.insert(V).second);
Chandler Carruth0539c072012-03-31 12:42:41 +00001198
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001199 Type *IntPtrTy = DL.getIntPtrType(V->getContext());
Chandler Carruth0539c072012-03-31 12:42:41 +00001200 return cast<ConstantInt>(ConstantInt::get(IntPtrTy, Offset));
1201}
1202
1203/// \brief Analyze a call site for potential inlining.
1204///
1205/// Returns true if inlining this call is viable, and false if it is not
1206/// viable. It computes the cost and adjusts the threshold based on numerous
1207/// factors and heuristics. If this method returns false but the computed cost
1208/// is below the computed threshold, then inlining was forcibly disabled by
Bob Wilson266802d2012-11-19 07:04:30 +00001209/// some artifact of the routine.
Chandler Carruth0539c072012-03-31 12:42:41 +00001210bool CallAnalyzer::analyzeCall(CallSite CS) {
Chandler Carruth7ae90d42012-04-11 10:15:10 +00001211 ++NumCallsAnalyzed;
1212
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001213 // Perform some tweaks to the cost and threshold based on the direct
1214 // callsite information.
Chandler Carruth0539c072012-03-31 12:42:41 +00001215
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001216 // We want to more aggressively inline vector-dense kernels, so up the
1217 // threshold, and we'll lower it if the % of vector instructions gets too
Chandler Carrutha004f222015-05-27 02:49:05 +00001218 // low. Note that these bonuses are some what arbitrary and evolved over time
1219 // by accident as much as because they are principled bonuses.
1220 //
1221 // FIXME: It would be nice to remove all such bonuses. At least it would be
1222 // nice to base the bonus values on something more scientific.
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001223 assert(NumInstructions == 0);
1224 assert(NumVectorInstructions == 0);
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +00001225
1226 // Update the threshold based on callsite properties
1227 updateThreshold(CS, F);
1228
Chandler Carrutha004f222015-05-27 02:49:05 +00001229 FiftyPercentVectorBonus = 3 * Threshold / 2;
1230 TenPercentVectorBonus = 3 * Threshold / 4;
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001231 const DataLayout &DL = F.getParent()->getDataLayout();
Benjamin Kramerc99d0e92012-08-07 11:13:19 +00001232
Chandler Carrutha004f222015-05-27 02:49:05 +00001233 // Track whether the post-inlining function would have more than one basic
1234 // block. A single basic block is often intended for inlining. Balloon the
1235 // threshold by 50% until we pass the single-BB phase.
1236 bool SingleBB = true;
1237 int SingleBBBonus = Threshold / 2;
1238
1239 // Speculatively apply all possible bonuses to Threshold. If cost exceeds
1240 // this Threshold any time, and cost cannot decrease, we can stop processing
1241 // the rest of the function body.
1242 Threshold += (SingleBBBonus + FiftyPercentVectorBonus);
1243
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001244 // Give out bonuses per argument, as the instructions setting them up will
1245 // be gone after inlining.
1246 for (unsigned I = 0, E = CS.arg_size(); I != E; ++I) {
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001247 if (CS.isByValArgument(I)) {
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001248 // We approximate the number of loads and stores needed by dividing the
1249 // size of the byval type by the target's pointer size.
1250 PointerType *PTy = cast<PointerType>(CS.getArgument(I)->getType());
Mehdi Aminia28d91d2015-03-10 02:37:25 +00001251 unsigned TypeSize = DL.getTypeSizeInBits(PTy->getElementType());
1252 unsigned PointerSize = DL.getPointerSizeInBits();
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001253 // Ceiling division.
1254 unsigned NumStores = (TypeSize + PointerSize - 1) / PointerSize;
Benjamin Kramerc99d0e92012-08-07 11:13:19 +00001255
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001256 // If it generates more than 8 stores it is likely to be expanded as an
1257 // inline memcpy so we take that as an upper bound. Otherwise we assume
1258 // one load and one store per word copied.
1259 // FIXME: The maxStoresPerMemcpy setting from the target should be used
1260 // here instead of a magic number of 8, but it's not available via
1261 // DataLayout.
1262 NumStores = std::min(NumStores, 8U);
1263
1264 Cost -= 2 * NumStores * InlineConstants::InstrCost;
1265 } else {
1266 // For non-byval arguments subtract off one instruction per call
1267 // argument.
1268 Cost -= InlineConstants::InstrCost;
Benjamin Kramerc99d0e92012-08-07 11:13:19 +00001269 }
Chandler Carruth0539c072012-03-31 12:42:41 +00001270 }
James Molloy6df8f272016-11-14 11:14:41 +00001271 // The call instruction also disappears after inlining.
1272 Cost -= InlineConstants::InstrCost + InlineConstants::CallPenalty;
1273
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001274 // If there is only one call of the function, and it has internal linkage,
1275 // the cost of inlining it drops dramatically.
Chad Rosier567556a2016-04-28 14:47:23 +00001276 bool OnlyOneCallAndLocalLinkage =
1277 F.hasLocalLinkage() && F.hasOneUse() && &F == CS.getCalledFunction();
James Molloy4f6fb952012-12-20 16:04:27 +00001278 if (OnlyOneCallAndLocalLinkage)
Piotr Padlewskid89875c2016-08-10 21:15:22 +00001279 Cost -= InlineConstants::LastCallToStaticBonus;
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001280
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001281 // If this function uses the coldcc calling convention, prefer not to inline
1282 // it.
1283 if (F.getCallingConv() == CallingConv::Cold)
1284 Cost += InlineConstants::ColdccPenalty;
1285
1286 // Check if we're done. This can happen due to bonuses and penalties.
1287 if (Cost > Threshold)
1288 return false;
1289
Chandler Carruth0539c072012-03-31 12:42:41 +00001290 if (F.empty())
1291 return true;
1292
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +00001293 Function *Caller = CS.getInstruction()->getParent()->getParent();
1294 // Check if the caller function is recursive itself.
Chandler Carruthcdf47882014-03-09 03:16:01 +00001295 for (User *U : Caller->users()) {
1296 CallSite Site(U);
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +00001297 if (!Site)
1298 continue;
1299 Instruction *I = Site.getInstruction();
1300 if (I->getParent()->getParent() == Caller) {
1301 IsCallerRecursive = true;
1302 break;
1303 }
1304 }
1305
Chandler Carruth0539c072012-03-31 12:42:41 +00001306 // Populate our simplified values by mapping from function arguments to call
1307 // arguments with known important simplifications.
1308 CallSite::arg_iterator CAI = CS.arg_begin();
1309 for (Function::arg_iterator FAI = F.arg_begin(), FAE = F.arg_end();
1310 FAI != FAE; ++FAI, ++CAI) {
1311 assert(CAI != CS.arg_end());
1312 if (Constant *C = dyn_cast<Constant>(CAI))
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001313 SimplifiedValues[&*FAI] = C;
Chandler Carruth0539c072012-03-31 12:42:41 +00001314
1315 Value *PtrArg = *CAI;
1316 if (ConstantInt *C = stripAndComputeInBoundsConstantOffsets(PtrArg)) {
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001317 ConstantOffsetPtrs[&*FAI] = std::make_pair(PtrArg, C->getValue());
Chandler Carruth0539c072012-03-31 12:42:41 +00001318
1319 // We can SROA any pointer arguments derived from alloca instructions.
1320 if (isa<AllocaInst>(PtrArg)) {
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001321 SROAArgValues[&*FAI] = PtrArg;
Chandler Carruth0539c072012-03-31 12:42:41 +00001322 SROAArgCosts[PtrArg] = 0;
1323 }
1324 }
1325 }
1326 NumConstantArgs = SimplifiedValues.size();
1327 NumConstantOffsetPtrArgs = ConstantOffsetPtrs.size();
1328 NumAllocaArgs = SROAArgValues.size();
1329
Hal Finkel57f03dd2014-09-07 13:49:57 +00001330 // FIXME: If a caller has multiple calls to a callee, we end up recomputing
1331 // the ephemeral values multiple times (and they're completely determined by
1332 // the callee, so this is purely duplicate work).
1333 SmallPtrSet<const Value *, 32> EphValues;
Daniel Jasperaec2fa32016-12-19 08:22:17 +00001334 CodeMetrics::collectEphemeralValues(&F, &GetAssumptionCache(F), EphValues);
Hal Finkel57f03dd2014-09-07 13:49:57 +00001335
Chandler Carruth0539c072012-03-31 12:42:41 +00001336 // The worklist of live basic blocks in the callee *after* inlining. We avoid
1337 // adding basic blocks of the callee which can be proven to be dead for this
1338 // particular call site in order to get more accurate cost estimates. This
1339 // requires a somewhat heavyweight iteration pattern: we need to walk the
1340 // basic blocks in a breadth-first order as we insert live successors. To
1341 // accomplish this, prioritizing for small iterations because we exit after
1342 // crossing our threshold, we use a small-size optimized SetVector.
1343 typedef SetVector<BasicBlock *, SmallVector<BasicBlock *, 16>,
Chad Rosier567556a2016-04-28 14:47:23 +00001344 SmallPtrSet<BasicBlock *, 16>>
1345 BBSetVector;
Chandler Carruth0539c072012-03-31 12:42:41 +00001346 BBSetVector BBWorklist;
1347 BBWorklist.insert(&F.getEntryBlock());
1348 // Note that we *must not* cache the size, this loop grows the worklist.
1349 for (unsigned Idx = 0; Idx != BBWorklist.size(); ++Idx) {
1350 // Bail out the moment we cross the threshold. This means we'll under-count
1351 // the cost, but only when undercounting doesn't matter.
Chandler Carrutha004f222015-05-27 02:49:05 +00001352 if (Cost > Threshold)
Chandler Carruth0539c072012-03-31 12:42:41 +00001353 break;
1354
1355 BasicBlock *BB = BBWorklist[Idx];
1356 if (BB->empty())
Chandler Carruth4d1d34f2012-03-14 23:19:53 +00001357 continue;
Dan Gohman4552e3c2009-10-13 18:30:07 +00001358
Gerolf Hoflehner734f4c82014-07-01 00:19:34 +00001359 // Disallow inlining a blockaddress. A blockaddress only has defined
1360 // behavior for an indirect branch in the same function, and we do not
1361 // currently support inlining indirect branches. But, the inliner may not
1362 // see an indirect branch that ends up being dead code at a particular call
1363 // site. If the blockaddress escapes the function, e.g., via a global
1364 // variable, inlining may lead to an invalid cross-function reference.
1365 if (BB->hasAddressTaken())
1366 return false;
1367
Chandler Carruth0539c072012-03-31 12:42:41 +00001368 // Analyze the cost of this block. If we blow through the threshold, this
1369 // returns false, and we can bail on out.
Easwaran Ramand295b002016-04-13 21:20:22 +00001370 if (!analyzeBlock(BB, EphValues))
1371 return false;
Eric Christopher46308e62011-02-01 01:16:32 +00001372
Chandler Carruth0814d2a2013-12-13 07:59:56 +00001373 TerminatorInst *TI = BB->getTerminator();
1374
Chandler Carruth0539c072012-03-31 12:42:41 +00001375 // Add in the live successors by first checking whether we have terminator
1376 // that may be simplified based on the values simplified by this call.
1377 if (BranchInst *BI = dyn_cast<BranchInst>(TI)) {
1378 if (BI->isConditional()) {
1379 Value *Cond = BI->getCondition();
Chad Rosier567556a2016-04-28 14:47:23 +00001380 if (ConstantInt *SimpleCond =
1381 dyn_cast_or_null<ConstantInt>(SimplifiedValues.lookup(Cond))) {
Chandler Carruth0539c072012-03-31 12:42:41 +00001382 BBWorklist.insert(BI->getSuccessor(SimpleCond->isZero() ? 1 : 0));
1383 continue;
Eric Christopher46308e62011-02-01 01:16:32 +00001384 }
Chandler Carruth0539c072012-03-31 12:42:41 +00001385 }
1386 } else if (SwitchInst *SI = dyn_cast<SwitchInst>(TI)) {
1387 Value *Cond = SI->getCondition();
Chad Rosier567556a2016-04-28 14:47:23 +00001388 if (ConstantInt *SimpleCond =
1389 dyn_cast_or_null<ConstantInt>(SimplifiedValues.lookup(Cond))) {
Chandler Carruth0539c072012-03-31 12:42:41 +00001390 BBWorklist.insert(SI->findCaseValue(SimpleCond).getCaseSuccessor());
1391 continue;
1392 }
1393 }
Eric Christopher46308e62011-02-01 01:16:32 +00001394
Chandler Carruth0539c072012-03-31 12:42:41 +00001395 // If we're unable to select a particular successor, just count all of
1396 // them.
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +00001397 for (unsigned TIdx = 0, TSize = TI->getNumSuccessors(); TIdx != TSize;
1398 ++TIdx)
Chandler Carruth0539c072012-03-31 12:42:41 +00001399 BBWorklist.insert(TI->getSuccessor(TIdx));
1400
1401 // If we had any successors at this point, than post-inlining is likely to
1402 // have them as well. Note that we assume any basic blocks which existed
1403 // due to branches or switches which folded above will also fold after
1404 // inlining.
1405 if (SingleBB && TI->getNumSuccessors() > 1) {
1406 // Take off the bonus we applied to the threshold.
1407 Threshold -= SingleBBBonus;
1408 SingleBB = false;
Eric Christopher46308e62011-02-01 01:16:32 +00001409 }
1410 }
Andrew Trickcaa500b2011-10-01 01:27:56 +00001411
Chandler Carruthcb5beb32013-12-12 11:59:26 +00001412 // If this is a noduplicate call, we can still inline as long as
James Molloy4f6fb952012-12-20 16:04:27 +00001413 // inlining this would cause the removal of the caller (so the instruction
1414 // is not actually duplicated, just moved).
1415 if (!OnlyOneCallAndLocalLinkage && ContainsNoDuplicateCall)
1416 return false;
1417
Chandler Carrutha004f222015-05-27 02:49:05 +00001418 // We applied the maximum possible vector bonus at the beginning. Now,
1419 // subtract the excess bonus, if any, from the Threshold before
1420 // comparing against Cost.
1421 if (NumVectorInstructions <= NumInstructions / 10)
1422 Threshold -= FiftyPercentVectorBonus;
1423 else if (NumVectorInstructions <= NumInstructions / 2)
1424 Threshold -= (FiftyPercentVectorBonus - TenPercentVectorBonus);
Chandler Carruth0539c072012-03-31 12:42:41 +00001425
Hans Wennborg00ab73d2016-02-05 20:32:42 +00001426 return Cost < std::max(1, Threshold);
Eric Christopher2dfbd7e2011-02-05 00:49:15 +00001427}
1428
Manman Ren49d684e2012-09-12 05:06:18 +00001429#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
Chandler Carruth0539c072012-03-31 12:42:41 +00001430/// \brief Dump stats about this call's analysis.
Yaron Kereneb2a2542016-01-29 20:50:44 +00001431LLVM_DUMP_METHOD void CallAnalyzer::dump() {
Eric Christophera13839f2014-02-26 23:27:16 +00001432#define DEBUG_PRINT_STAT(x) dbgs() << " " #x ": " << x << "\n"
Chandler Carruth0539c072012-03-31 12:42:41 +00001433 DEBUG_PRINT_STAT(NumConstantArgs);
1434 DEBUG_PRINT_STAT(NumConstantOffsetPtrArgs);
1435 DEBUG_PRINT_STAT(NumAllocaArgs);
1436 DEBUG_PRINT_STAT(NumConstantPtrCmps);
1437 DEBUG_PRINT_STAT(NumConstantPtrDiffs);
1438 DEBUG_PRINT_STAT(NumInstructionsSimplified);
Chandler Carrutha004f222015-05-27 02:49:05 +00001439 DEBUG_PRINT_STAT(NumInstructions);
Chandler Carruth0539c072012-03-31 12:42:41 +00001440 DEBUG_PRINT_STAT(SROACostSavings);
1441 DEBUG_PRINT_STAT(SROACostSavingsLost);
James Molloy4f6fb952012-12-20 16:04:27 +00001442 DEBUG_PRINT_STAT(ContainsNoDuplicateCall);
Chandler Carruth394e34f2014-01-31 22:32:32 +00001443 DEBUG_PRINT_STAT(Cost);
1444 DEBUG_PRINT_STAT(Threshold);
Chandler Carruth0539c072012-03-31 12:42:41 +00001445#undef DEBUG_PRINT_STAT
Eric Christopher2dfbd7e2011-02-05 00:49:15 +00001446}
Manman Renc3366cc2012-09-06 19:55:56 +00001447#endif
Eric Christopher2dfbd7e2011-02-05 00:49:15 +00001448
Akira Hatanaka5af7ace2015-11-13 01:44:32 +00001449/// \brief Test that two functions either have or have not the given attribute
1450/// at the same time.
Chad Rosier567556a2016-04-28 14:47:23 +00001451template <typename AttrKind>
Akira Hatanaka5af7ace2015-11-13 01:44:32 +00001452static bool attributeMatches(Function *F1, Function *F2, AttrKind Attr) {
1453 return F1->getFnAttribute(Attr) == F2->getFnAttribute(Attr);
1454}
1455
Evgeniy Stepanov2ad36982013-08-08 08:22:39 +00001456/// \brief Test that there are no attribute conflicts between Caller and Callee
1457/// that prevent inlining.
1458static bool functionsHaveCompatibleAttributes(Function *Caller,
Eric Christopher4371b132015-07-02 01:11:47 +00001459 Function *Callee,
1460 TargetTransformInfo &TTI) {
Eric Christopherd566fb12015-07-29 22:09:48 +00001461 return TTI.areInlineCompatible(Caller, Callee) &&
Akira Hatanaka1cb242e2015-12-22 23:57:37 +00001462 AttributeFuncs::areInlineCompatible(*Caller, *Callee);
Evgeniy Stepanov2ad36982013-08-08 08:22:39 +00001463}
1464
Sean Silvaab6a6832016-07-23 04:22:50 +00001465InlineCost llvm::getInlineCost(
Easwaran Raman1c57cc22016-08-10 00:48:04 +00001466 CallSite CS, const InlineParams &Params, TargetTransformInfo &CalleeTTI,
Daniel Jasperaec2fa32016-12-19 08:22:17 +00001467 std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
Sean Silvaab6a6832016-07-23 04:22:50 +00001468 ProfileSummaryInfo *PSI) {
Daniel Jasperaec2fa32016-12-19 08:22:17 +00001469 return getInlineCost(CS, CS.getCalledFunction(), Params, CalleeTTI,
1470 GetAssumptionCache, PSI);
Easwaran Ramanb9f71202015-12-28 20:28:19 +00001471}
1472
Sean Silvaab6a6832016-07-23 04:22:50 +00001473InlineCost llvm::getInlineCost(
Easwaran Raman1c57cc22016-08-10 00:48:04 +00001474 CallSite CS, Function *Callee, const InlineParams &Params,
Daniel Jasperaec2fa32016-12-19 08:22:17 +00001475 TargetTransformInfo &CalleeTTI,
1476 std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
1477 ProfileSummaryInfo *PSI) {
Easwaran Ramanf4bb2f02016-01-14 23:16:29 +00001478
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001479 // Cannot inline indirect calls.
1480 if (!Callee)
1481 return llvm::InlineCost::getNever();
1482
1483 // Calls to functions with always-inline attributes should be inlined
1484 // whenever possible.
Peter Collingbourne68a88972014-05-19 18:25:54 +00001485 if (CS.hasFnAttr(Attribute::AlwaysInline)) {
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001486 if (isInlineViable(*Callee))
1487 return llvm::InlineCost::getAlways();
1488 return llvm::InlineCost::getNever();
1489 }
1490
Evgeniy Stepanov2ad36982013-08-08 08:22:39 +00001491 // Never inline functions with conflicting attributes (unless callee has
1492 // always-inline attribute).
Easwaran Ramanb9f71202015-12-28 20:28:19 +00001493 if (!functionsHaveCompatibleAttributes(CS.getCaller(), Callee, CalleeTTI))
Evgeniy Stepanov2ad36982013-08-08 08:22:39 +00001494 return llvm::InlineCost::getNever();
1495
Paul Robinsondcbe35b2013-11-18 21:44:03 +00001496 // Don't inline this call if the caller has the optnone attribute.
1497 if (CS.getCaller()->hasFnAttribute(Attribute::OptimizeNone))
1498 return llvm::InlineCost::getNever();
1499
Sanjoy Das5ce32722016-04-08 00:48:30 +00001500 // Don't inline functions which can be interposed at link-time. Don't inline
1501 // functions marked noinline or call sites marked noinline.
Craig Topper107b1872016-12-09 02:18:04 +00001502 // Note: inlining non-exact non-interposable functions is fine, since we know
Sanjoy Das5ce32722016-04-08 00:48:30 +00001503 // we have *a* correct implementation of the source level function.
Chad Rosier567556a2016-04-28 14:47:23 +00001504 if (Callee->isInterposable() || Callee->hasFnAttribute(Attribute::NoInline) ||
1505 CS.isNoInline())
Dan Gohman4552e3c2009-10-13 18:30:07 +00001506 return llvm::InlineCost::getNever();
1507
Nadav Rotem4eb3d4b2012-09-19 08:08:04 +00001508 DEBUG(llvm::dbgs() << " Analyzing call of " << Callee->getName()
Chad Rosier567556a2016-04-28 14:47:23 +00001509 << "...\n");
Andrew Trickcaa500b2011-10-01 01:27:56 +00001510
Daniel Jasperaec2fa32016-12-19 08:22:17 +00001511 CallAnalyzer CA(CalleeTTI, GetAssumptionCache, PSI, *Callee, CS, Params);
Chandler Carruth0539c072012-03-31 12:42:41 +00001512 bool ShouldInline = CA.analyzeCall(CS);
Dan Gohman4552e3c2009-10-13 18:30:07 +00001513
Chandler Carruth0539c072012-03-31 12:42:41 +00001514 DEBUG(CA.dump());
1515
1516 // Check if there was a reason to force inlining or no inlining.
1517 if (!ShouldInline && CA.getCost() < CA.getThreshold())
Dan Gohman4552e3c2009-10-13 18:30:07 +00001518 return InlineCost::getNever();
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001519 if (ShouldInline && CA.getCost() >= CA.getThreshold())
Dan Gohman4552e3c2009-10-13 18:30:07 +00001520 return InlineCost::getAlways();
Andrew Trickcaa500b2011-10-01 01:27:56 +00001521
Chandler Carruth0539c072012-03-31 12:42:41 +00001522 return llvm::InlineCost::get(CA.getCost(), CA.getThreshold());
Dan Gohman4552e3c2009-10-13 18:30:07 +00001523}
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001524
Easwaran Ramanb9f71202015-12-28 20:28:19 +00001525bool llvm::isInlineViable(Function &F) {
Duncan P. N. Exon Smithb3fc83c2015-02-14 00:12:15 +00001526 bool ReturnsTwice = F.hasFnAttribute(Attribute::ReturnsTwice);
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001527 for (Function::iterator BI = F.begin(), BE = F.end(); BI != BE; ++BI) {
Gerolf Hoflehner734f4c82014-07-01 00:19:34 +00001528 // Disallow inlining of functions which contain indirect branches or
1529 // blockaddresses.
1530 if (isa<IndirectBrInst>(BI->getTerminator()) || BI->hasAddressTaken())
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001531 return false;
1532
Duncan P. N. Exon Smith5a82c912015-10-10 00:53:03 +00001533 for (auto &II : *BI) {
1534 CallSite CS(&II);
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001535 if (!CS)
1536 continue;
1537
1538 // Disallow recursive calls.
1539 if (&F == CS.getCalledFunction())
1540 return false;
1541
1542 // Disallow calls which expose returns-twice to a function not previously
1543 // attributed as such.
1544 if (!ReturnsTwice && CS.isCall() &&
1545 cast<CallInst>(CS.getInstruction())->canReturnTwice())
1546 return false;
Reid Kleckner223de262015-04-14 20:38:14 +00001547
Reid Kleckner60381792015-07-07 22:25:32 +00001548 // Disallow inlining functions that call @llvm.localescape. Doing this
Reid Kleckner223de262015-04-14 20:38:14 +00001549 // correctly would require major changes to the inliner.
1550 if (CS.getCalledFunction() &&
1551 CS.getCalledFunction()->getIntrinsicID() ==
Reid Kleckner60381792015-07-07 22:25:32 +00001552 llvm::Intrinsic::localescape)
Reid Kleckner223de262015-04-14 20:38:14 +00001553 return false;
Bob Wilsona5b0dc82012-11-19 07:04:35 +00001554 }
1555 }
1556
1557 return true;
1558}
Easwaran Raman1c57cc22016-08-10 00:48:04 +00001559
1560// APIs to create InlineParams based on command line flags and/or other
1561// parameters.
1562
1563InlineParams llvm::getInlineParams(int Threshold) {
1564 InlineParams Params;
1565
1566 // This field is the threshold to use for a callee by default. This is
1567 // derived from one or more of:
1568 // * optimization or size-optimization levels,
1569 // * a value passed to createFunctionInliningPass function, or
1570 // * the -inline-threshold flag.
1571 // If the -inline-threshold flag is explicitly specified, that is used
1572 // irrespective of anything else.
1573 if (InlineThreshold.getNumOccurrences() > 0)
1574 Params.DefaultThreshold = InlineThreshold;
1575 else
1576 Params.DefaultThreshold = Threshold;
1577
1578 // Set the HintThreshold knob from the -inlinehint-threshold.
1579 Params.HintThreshold = HintThreshold;
1580
1581 // Set the HotCallSiteThreshold knob from the -hot-callsite-threshold.
1582 Params.HotCallSiteThreshold = HotCallSiteThreshold;
1583
1584 // Set the OptMinSizeThreshold and OptSizeThreshold params only if the
1585 // Set the OptMinSizeThreshold and OptSizeThreshold params only if the
1586 // -inlinehint-threshold commandline option is not explicitly given. If that
1587 // option is present, then its value applies even for callees with size and
1588 // minsize attributes.
1589 // If the -inline-threshold is not specified, set the ColdThreshold from the
1590 // -inlinecold-threshold even if it is not explicitly passed. If
1591 // -inline-threshold is specified, then -inlinecold-threshold needs to be
1592 // explicitly specified to set the ColdThreshold knob
1593 if (InlineThreshold.getNumOccurrences() == 0) {
1594 Params.OptMinSizeThreshold = InlineConstants::OptMinSizeThreshold;
1595 Params.OptSizeThreshold = InlineConstants::OptSizeThreshold;
1596 Params.ColdThreshold = ColdThreshold;
1597 } else if (ColdThreshold.getNumOccurrences() > 0) {
1598 Params.ColdThreshold = ColdThreshold;
1599 }
1600 return Params;
1601}
1602
1603InlineParams llvm::getInlineParams() {
1604 return getInlineParams(InlineThreshold);
1605}
1606
1607// Compute the default threshold for inlining based on the opt level and the
1608// size opt level.
1609static int computeThresholdFromOptLevels(unsigned OptLevel,
1610 unsigned SizeOptLevel) {
1611 if (OptLevel > 2)
1612 return InlineConstants::OptAggressiveThreshold;
1613 if (SizeOptLevel == 1) // -Os
1614 return InlineConstants::OptSizeThreshold;
1615 if (SizeOptLevel == 2) // -Oz
1616 return InlineConstants::OptMinSizeThreshold;
1617 return InlineThreshold;
1618}
1619
1620InlineParams llvm::getInlineParams(unsigned OptLevel, unsigned SizeOptLevel) {
1621 return getInlineParams(computeThresholdFromOptLevels(OptLevel, SizeOptLevel));
1622}