1 //===- InlineCost.cpp - Cost analysis for inliner -------------------------===//
2 //
3 //                     The LLVM Compiler Infrastructure
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 //===----------------------------------------------------------------------===//
9 //
10 // This file implements inline cost analysis.
11 //
12 //===----------------------------------------------------------------------===//
13 
14 #include "llvm/Analysis/InlineCost.h"
15 #include "llvm/ADT/STLExtras.h"
16 #include "llvm/ADT/SetVector.h"
17 #include "llvm/ADT/SmallPtrSet.h"
18 #include "llvm/ADT/SmallVector.h"
19 #include "llvm/ADT/Statistic.h"
20 #include "llvm/Analysis/AssumptionCache.h"
21 #include "llvm/Analysis/BlockFrequencyInfo.h"
22 #include "llvm/Analysis/CodeMetrics.h"
23 #include "llvm/Analysis/ConstantFolding.h"
24 #include "llvm/Analysis/InstructionSimplify.h"
25 #include "llvm/Analysis/ProfileSummaryInfo.h"
26 #include "llvm/Analysis/TargetTransformInfo.h"
27 #include "llvm/IR/CallSite.h"
28 #include "llvm/IR/CallingConv.h"
29 #include "llvm/IR/DataLayout.h"
30 #include "llvm/IR/GetElementPtrTypeIterator.h"
31 #include "llvm/IR/GlobalAlias.h"
32 #include "llvm/IR/InstVisitor.h"
33 #include "llvm/IR/IntrinsicInst.h"
34 #include "llvm/IR/Operator.h"
35 #include "llvm/Support/Debug.h"
36 #include "llvm/Support/raw_ostream.h"
37 
38 using namespace llvm;
39 
40 #define DEBUG_TYPE "inline-cost"
41 
42 STATISTIC(NumCallsAnalyzed, "Number of call sites analyzed");
43 
44 static cl::opt<int> InlineThreshold(
45     "inline-threshold", cl::Hidden, cl::init(225), cl::ZeroOrMore,
46     cl::desc("Control the amount of inlining to perform (default = 225)"));
47 
48 static cl::opt<int> HintThreshold(
49     "inlinehint-threshold", cl::Hidden, cl::init(325),
50     cl::desc("Threshold for inlining functions with inline hint"));
51 
52 static cl::opt<int>
53     ColdCallSiteThreshold("inline-cold-callsite-threshold", cl::Hidden,
54                           cl::init(45),
55                           cl::desc("Threshold for inlining cold callsites"));
56 
57 // We introduce this threshold to help performance of instrumentation based
58 // PGO before we actually hook up inliner with analysis passes such as BPI and
59 // BFI.
60 static cl::opt<int> ColdThreshold(
61     "inlinecold-threshold", cl::Hidden, cl::init(45),
62     cl::desc("Threshold for inlining functions with cold attribute"));
63 
64 static cl::opt<int>
65     HotCallSiteThreshold("hot-callsite-threshold", cl::Hidden, cl::init(3000),
66                          cl::ZeroOrMore,
67                          cl::desc("Threshold for hot callsites "));
68 
69 namespace {
70 
71 class CallAnalyzer : public InstVisitor<CallAnalyzer, bool> {
72   typedef InstVisitor<CallAnalyzer, bool> Base;
73   friend class InstVisitor<CallAnalyzer, bool>;
74 
75   /// The TargetTransformInfo available for this compilation.
76   const TargetTransformInfo &TTI;
77 
78   /// Getter for the cache of @llvm.assume intrinsics.
79   std::function<AssumptionCache &(Function &)> &GetAssumptionCache;
80 
81   /// Getter for BlockFrequencyInfo
82   Optional<function_ref<BlockFrequencyInfo &(Function &)>> &GetBFI;
83 
84   /// Profile summary information.
85   ProfileSummaryInfo *PSI;
86 
87   /// The called function.
88   Function &F;
89 
90   // Cache the DataLayout since we use it a lot.
91   const DataLayout &DL;
92 
93   /// The candidate callsite being analyzed. Please do not use this to do
94   /// analysis in the caller function; we want the inline cost query to be
95   /// easily cacheable. Instead, use the cover function paramHasAttr.
96   CallSite CandidateCS;
97 
98   /// Tunable parameters that control the analysis.
99   const InlineParams &Params;
100 
101   int Threshold;
102   int Cost;
103 
104   bool IsCallerRecursive;
105   bool IsRecursiveCall;
106   bool ExposesReturnsTwice;
107   bool HasDynamicAlloca;
108   bool ContainsNoDuplicateCall;
109   bool HasReturn;
110   bool HasIndirectBr;
111   bool HasFrameEscape;
112 
113   /// Number of bytes allocated statically by the callee.
114   uint64_t AllocatedSize;
115   unsigned NumInstructions, NumVectorInstructions;
116   int FiftyPercentVectorBonus, TenPercentVectorBonus;
117   int VectorBonus;
118 
119   /// While we walk the potentially-inlined instructions, we build up and
120   /// maintain a mapping of simplified values specific to this callsite. The
121   /// idea is to propagate any special information we have about arguments to
122   /// this call through the inlinable section of the function, and account for
123   /// likely simplifications post-inlining. The most important aspect we track
124   /// is CFG altering simplifications -- when we prove a basic block dead, that
125   /// can cause dramatic shifts in the cost of inlining a function.
126   DenseMap<Value *, Constant *> SimplifiedValues;
127 
128   /// Keep track of the values which map back (through function arguments) to
129   /// allocas on the caller stack which could be simplified through SROA.
130   DenseMap<Value *, Value *> SROAArgValues;
131 
132   /// The mapping of caller Alloca values to their accumulated cost savings. If
133   /// we have to disable SROA for one of the allocas, this tells us how much
134   /// cost must be added.
135   DenseMap<Value *, int> SROAArgCosts;
136 
137   /// Keep track of values which map to a pointer base and constant offset.
138   DenseMap<Value *, std::pair<Value *, APInt>> ConstantOffsetPtrs;
139 
140   // Custom simplification helper routines.
141   bool isAllocaDerivedArg(Value *V);
142   bool lookupSROAArgAndCost(Value *V, Value *&Arg,
143                             DenseMap<Value *, int>::iterator &CostIt);
144   void disableSROA(DenseMap<Value *, int>::iterator CostIt);
145   void disableSROA(Value *V);
146   void accumulateSROACost(DenseMap<Value *, int>::iterator CostIt,
147                           int InstructionCost);
148   bool isGEPFree(GetElementPtrInst &GEP);
149   bool accumulateGEPOffset(GEPOperator &GEP, APInt &Offset);
150   bool simplifyCallSite(Function *F, CallSite CS);
151   template <typename Callable>
152   bool simplifyInstruction(Instruction &I, Callable Evaluate);
153   ConstantInt *stripAndComputeInBoundsConstantOffsets(Value *&V);
154 
155   /// Return true if the given argument to the function being considered for
156   /// inlining has the given attribute set either at the call site or the
157   /// function declaration.  Primarily used to inspect call site specific
158   /// attributes since these can be more precise than the ones on the callee
159   /// itself.
160   bool paramHasAttr(Argument *A, Attribute::AttrKind Attr);
161 
162   /// Return true if the given value is known non null within the callee if
163   /// inlined through this particular callsite.
164   bool isKnownNonNullInCallee(Value *V);
165 
166   /// Update Threshold based on callsite properties such as callee
167   /// attributes and callee hotness for PGO builds. The Callee is explicitly
168   /// passed to support analyzing indirect calls whose target is inferred by
169   /// analysis.
170   void updateThreshold(CallSite CS, Function &Callee);
171 
172   /// Return true if size growth is allowed when inlining the callee at CS.
173   bool allowSizeGrowth(CallSite CS);
174 
175   // Custom analysis routines.
176   bool analyzeBlock(BasicBlock *BB, SmallPtrSetImpl<const Value *> &EphValues);
177 
178   // Disable several entry points to the visitor so we don't accidentally use
179   // them by declaring but not defining them here.
180   void visit(Module *);
181   void visit(Module &);
182   void visit(Function *);
183   void visit(Function &);
184   void visit(BasicBlock *);
185   void visit(BasicBlock &);
186 
187   // Provide base case for our instruction visit.
188   bool visitInstruction(Instruction &I);
189 
190   // Our visit overrides.
191   bool visitAlloca(AllocaInst &I);
192   bool visitPHI(PHINode &I);
193   bool visitGetElementPtr(GetElementPtrInst &I);
194   bool visitBitCast(BitCastInst &I);
195   bool visitPtrToInt(PtrToIntInst &I);
196   bool visitIntToPtr(IntToPtrInst &I);
197   bool visitCastInst(CastInst &I);
198   bool visitUnaryInstruction(UnaryInstruction &I);
199   bool visitCmpInst(CmpInst &I);
200   bool visitSub(BinaryOperator &I);
201   bool visitBinaryOperator(BinaryOperator &I);
202   bool visitLoad(LoadInst &I);
203   bool visitStore(StoreInst &I);
204   bool visitExtractValue(ExtractValueInst &I);
205   bool visitInsertValue(InsertValueInst &I);
206   bool visitCallSite(CallSite CS);
207   bool visitReturnInst(ReturnInst &RI);
208   bool visitBranchInst(BranchInst &BI);
209   bool visitSwitchInst(SwitchInst &SI);
210   bool visitIndirectBrInst(IndirectBrInst &IBI);
211   bool visitResumeInst(ResumeInst &RI);
212   bool visitCleanupReturnInst(CleanupReturnInst &RI);
213   bool visitCatchReturnInst(CatchReturnInst &RI);
214   bool visitUnreachableInst(UnreachableInst &I);
215 
216 public:
217   CallAnalyzer(const TargetTransformInfo &TTI,
218                std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
219                Optional<function_ref<BlockFrequencyInfo &(Function &)>> &GetBFI,
220                ProfileSummaryInfo *PSI, Function &Callee, CallSite CSArg,
221                const InlineParams &Params)
222       : TTI(TTI), GetAssumptionCache(GetAssumptionCache), GetBFI(GetBFI),
223         PSI(PSI), F(Callee), DL(F.getParent()->getDataLayout()),
224         CandidateCS(CSArg), Params(Params), Threshold(Params.DefaultThreshold),
225         Cost(0), IsCallerRecursive(false), IsRecursiveCall(false),
226         ExposesReturnsTwice(false), HasDynamicAlloca(false),
227         ContainsNoDuplicateCall(false), HasReturn(false), HasIndirectBr(false),
228         HasFrameEscape(false), AllocatedSize(0), NumInstructions(0),
229         NumVectorInstructions(0), FiftyPercentVectorBonus(0),
230         TenPercentVectorBonus(0), VectorBonus(0), NumConstantArgs(0),
231         NumConstantOffsetPtrArgs(0), NumAllocaArgs(0), NumConstantPtrCmps(0),
232         NumConstantPtrDiffs(0), NumInstructionsSimplified(0),
233         SROACostSavings(0), SROACostSavingsLost(0) {}
234 
235   bool analyzeCall(CallSite CS);
236 
237   int getThreshold() { return Threshold; }
238   int getCost() { return Cost; }
239 
240   // Keep a bunch of stats about the cost savings found so we can print them
241   // out when debugging.
242   unsigned NumConstantArgs;
243   unsigned NumConstantOffsetPtrArgs;
244   unsigned NumAllocaArgs;
245   unsigned NumConstantPtrCmps;
246   unsigned NumConstantPtrDiffs;
247   unsigned NumInstructionsSimplified;
248   unsigned SROACostSavings;
249   unsigned SROACostSavingsLost;
250 
251   void dump();
252 };
253 
254 } // namespace
255 
256 /// \brief Test whether the given value is an Alloca-derived function argument.
257 bool CallAnalyzer::isAllocaDerivedArg(Value *V) {
258   return SROAArgValues.count(V);
259 }
260 
261 /// \brief Lookup the SROA-candidate argument and cost iterator which V maps to.
262 /// Returns false if V does not map to a SROA-candidate.
263 bool CallAnalyzer::lookupSROAArgAndCost(
264     Value *V, Value *&Arg, DenseMap<Value *, int>::iterator &CostIt) {
265   if (SROAArgValues.empty() || SROAArgCosts.empty())
266     return false;
267 
268   DenseMap<Value *, Value *>::iterator ArgIt = SROAArgValues.find(V);
269   if (ArgIt == SROAArgValues.end())
270     return false;
271 
272   Arg = ArgIt->second;
273   CostIt = SROAArgCosts.find(Arg);
274   return CostIt != SROAArgCosts.end();
275 }
276 
277 /// \brief Disable SROA for the candidate marked by this cost iterator.
278 ///
279 /// This marks the candidate as no longer viable for SROA, and adds the cost
280 /// savings associated with it back into the inline cost measurement.
281 void CallAnalyzer::disableSROA(DenseMap<Value *, int>::iterator CostIt) {
282   // If we're no longer able to perform SROA we need to undo its cost savings
283   // and prevent subsequent analysis.
284   Cost += CostIt->second;
285   SROACostSavings -= CostIt->second;
286   SROACostSavingsLost += CostIt->second;
287   SROAArgCosts.erase(CostIt);
288 }
289 
290 /// \brief If 'V' maps to a SROA candidate, disable SROA for it.
291 void CallAnalyzer::disableSROA(Value *V) {
292   Value *SROAArg;
293   DenseMap<Value *, int>::iterator CostIt;
294   if (lookupSROAArgAndCost(V, SROAArg, CostIt))
295     disableSROA(CostIt);
296 }
297 
298 /// \brief Accumulate the given cost for a particular SROA candidate.
299 void CallAnalyzer::accumulateSROACost(DenseMap<Value *, int>::iterator CostIt,
300                                       int InstructionCost) {
301   CostIt->second += InstructionCost;
302   SROACostSavings += InstructionCost;
303 }
304 
305 /// \brief Accumulate a constant GEP offset into an APInt if possible.
306 ///
307 /// Returns false if unable to compute the offset for any reason. Respects any
308 /// simplified values known during the analysis of this callsite.
309 bool CallAnalyzer::accumulateGEPOffset(GEPOperator &GEP, APInt &Offset) {
310   unsigned IntPtrWidth = DL.getPointerSizeInBits();
311   assert(IntPtrWidth == Offset.getBitWidth());
312 
313   for (gep_type_iterator GTI = gep_type_begin(GEP), GTE = gep_type_end(GEP);
314        GTI != GTE; ++GTI) {
315     ConstantInt *OpC = dyn_cast<ConstantInt>(GTI.getOperand());
316     if (!OpC)
317       if (Constant *SimpleOp = SimplifiedValues.lookup(GTI.getOperand()))
318         OpC = dyn_cast<ConstantInt>(SimpleOp);
319     if (!OpC)
320       return false;
321     if (OpC->isZero())
322       continue;
323 
324     // Handle a struct index, which adds its field offset to the pointer.
325     if (StructType *STy = GTI.getStructTypeOrNull()) {
326       unsigned ElementIdx = OpC->getZExtValue();
327       const StructLayout *SL = DL.getStructLayout(STy);
328       Offset += APInt(IntPtrWidth, SL->getElementOffset(ElementIdx));
329       continue;
330     }
331 
332     APInt TypeSize(IntPtrWidth, DL.getTypeAllocSize(GTI.getIndexedType()));
333     Offset += OpC->getValue().sextOrTrunc(IntPtrWidth) * TypeSize;
334   }
335   return true;
336 }
337 
338 /// \brief Use TTI to check whether a GEP is free.
339 ///
340 /// Respects any simplified values known during the analysis of this callsite.
341 bool CallAnalyzer::isGEPFree(GetElementPtrInst &GEP) {
342   SmallVector<Value *, 4> Indices;
343   for (User::op_iterator I = GEP.idx_begin(), E = GEP.idx_end(); I != E; ++I)
344     if (Constant *SimpleOp = SimplifiedValues.lookup(*I))
345        Indices.push_back(SimpleOp);
346      else
347        Indices.push_back(*I);
348   return TargetTransformInfo::TCC_Free ==
349          TTI.getGEPCost(GEP.getSourceElementType(), GEP.getPointerOperand(),
350                         Indices);
351 }
352 
353 bool CallAnalyzer::visitAlloca(AllocaInst &I) {
354   // Check whether inlining will turn a dynamic alloca into a static
355   // alloca and handle that case.
356   if (I.isArrayAllocation()) {
357     Constant *Size = SimplifiedValues.lookup(I.getArraySize());
358     if (auto *AllocSize = dyn_cast_or_null<ConstantInt>(Size)) {
359       Type *Ty = I.getAllocatedType();
360       AllocatedSize = SaturatingMultiplyAdd(
361           AllocSize->getLimitedValue(), DL.getTypeAllocSize(Ty), AllocatedSize);
362       return Base::visitAlloca(I);
363     }
364   }
365 
366   // Accumulate the allocated size.
367   if (I.isStaticAlloca()) {
368     Type *Ty = I.getAllocatedType();
369     AllocatedSize = SaturatingAdd(DL.getTypeAllocSize(Ty), AllocatedSize);
370   }
371 
372   // We will happily inline static alloca instructions.
373   if (I.isStaticAlloca())
374     return Base::visitAlloca(I);
375 
376   // FIXME: This is overly conservative. Dynamic allocas are inefficient for
377   // a variety of reasons, and so we would like to not inline them into
378   // functions which don't currently have a dynamic alloca. This simply
379   // disables inlining altogether in the presence of a dynamic alloca.
380   HasDynamicAlloca = true;
381   return false;
382 }
383 
384 bool CallAnalyzer::visitPHI(PHINode &I) {
385   // FIXME: We should potentially be tracking values through phi nodes,
386   // especially when they collapse to a single value due to deleted CFG edges
387   // during inlining.
388 
389   // FIXME: We need to propagate SROA *disabling* through phi nodes, even
390   // though we don't want to propagate it's bonuses. The idea is to disable
391   // SROA if it *might* be used in an inappropriate manner.
392 
393   // Phi nodes are always zero-cost.
394   return true;
395 }
396 
397 bool CallAnalyzer::visitGetElementPtr(GetElementPtrInst &I) {
398   Value *SROAArg;
399   DenseMap<Value *, int>::iterator CostIt;
400   bool SROACandidate =
401       lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt);
402 
403   // Try to fold GEPs of constant-offset call site argument pointers. This
404   // requires target data and inbounds GEPs.
405   if (I.isInBounds()) {
406     // Check if we have a base + offset for the pointer.
407     Value *Ptr = I.getPointerOperand();
408     std::pair<Value *, APInt> BaseAndOffset = ConstantOffsetPtrs.lookup(Ptr);
409     if (BaseAndOffset.first) {
410       // Check if the offset of this GEP is constant, and if so accumulate it
411       // into Offset.
412       if (!accumulateGEPOffset(cast<GEPOperator>(I), BaseAndOffset.second)) {
413         // Non-constant GEPs aren't folded, and disable SROA.
414         if (SROACandidate)
415           disableSROA(CostIt);
416         return isGEPFree(I);
417       }
418 
419       // Add the result as a new mapping to Base + Offset.
420       ConstantOffsetPtrs[&I] = BaseAndOffset;
421 
422       // Also handle SROA candidates here, we already know that the GEP is
423       // all-constant indexed.
424       if (SROACandidate)
425         SROAArgValues[&I] = SROAArg;
426 
427       return true;
428     }
429   }
430 
431   // Lambda to check whether a GEP's indices are all constant.
432   auto IsGEPOffsetConstant = [&](GetElementPtrInst &GEP) {
433     for (User::op_iterator I = GEP.idx_begin(), E = GEP.idx_end(); I != E; ++I)
434       if (!isa<Constant>(*I) && !SimplifiedValues.lookup(*I))
435         return false;
436     return true;
437   };
438 
439   if (IsGEPOffsetConstant(I)) {
440     if (SROACandidate)
441       SROAArgValues[&I] = SROAArg;
442 
443     // Constant GEPs are modeled as free.
444     return true;
445   }
446 
447   // Variable GEPs will require math and will disable SROA.
448   if (SROACandidate)
449     disableSROA(CostIt);
450   return isGEPFree(I);
451 }
452 
453 /// Simplify \p I if its operands are constants and update SimplifiedValues.
454 /// \p Evaluate is a callable specific to instruction type that evaluates the
455 /// instruction when all the operands are constants.
456 template <typename Callable>
457 bool CallAnalyzer::simplifyInstruction(Instruction &I, Callable Evaluate) {
458   SmallVector<Constant *, 2> COps;
459   for (Value *Op : I.operands()) {
460     Constant *COp = dyn_cast<Constant>(Op);
461     if (!COp)
462       COp = SimplifiedValues.lookup(Op);
463     if (!COp)
464       return false;
465     COps.push_back(COp);
466   }
467   auto *C = Evaluate(COps);
468   if (!C)
469     return false;
470   SimplifiedValues[&I] = C;
471   return true;
472 }
473 
474 bool CallAnalyzer::visitBitCast(BitCastInst &I) {
475   // Propagate constants through bitcasts.
476   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
477         return ConstantExpr::getBitCast(COps[0], I.getType());
478       }))
479     return true;
480 
481   // Track base/offsets through casts
482   std::pair<Value *, APInt> BaseAndOffset =
483       ConstantOffsetPtrs.lookup(I.getOperand(0));
484   // Casts don't change the offset, just wrap it up.
485   if (BaseAndOffset.first)
486     ConstantOffsetPtrs[&I] = BaseAndOffset;
487 
488   // Also look for SROA candidates here.
489   Value *SROAArg;
490   DenseMap<Value *, int>::iterator CostIt;
491   if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt))
492     SROAArgValues[&I] = SROAArg;
493 
494   // Bitcasts are always zero cost.
495   return true;
496 }
497 
498 bool CallAnalyzer::visitPtrToInt(PtrToIntInst &I) {
499   // Propagate constants through ptrtoint.
500   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
501         return ConstantExpr::getPtrToInt(COps[0], I.getType());
502       }))
503     return true;
504 
505   // Track base/offset pairs when converted to a plain integer provided the
506   // integer is large enough to represent the pointer.
507   unsigned IntegerSize = I.getType()->getScalarSizeInBits();
508   if (IntegerSize >= DL.getPointerSizeInBits()) {
509     std::pair<Value *, APInt> BaseAndOffset =
510         ConstantOffsetPtrs.lookup(I.getOperand(0));
511     if (BaseAndOffset.first)
512       ConstantOffsetPtrs[&I] = BaseAndOffset;
513   }
514 
515   // This is really weird. Technically, ptrtoint will disable SROA. However,
516   // unless that ptrtoint is *used* somewhere in the live basic blocks after
517   // inlining, it will be nuked, and SROA should proceed. All of the uses which
518   // would block SROA would also block SROA if applied directly to a pointer,
519   // and so we can just add the integer in here. The only places where SROA is
520   // preserved either cannot fire on an integer, or won't in-and-of themselves
521   // disable SROA (ext) w/o some later use that we would see and disable.
522   Value *SROAArg;
523   DenseMap<Value *, int>::iterator CostIt;
524   if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt))
525     SROAArgValues[&I] = SROAArg;
526 
527   return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
528 }
529 
530 bool CallAnalyzer::visitIntToPtr(IntToPtrInst &I) {
531   // Propagate constants through ptrtoint.
532   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
533         return ConstantExpr::getIntToPtr(COps[0], I.getType());
534       }))
535     return true;
536 
537   // Track base/offset pairs when round-tripped through a pointer without
538   // modifications provided the integer is not too large.
539   Value *Op = I.getOperand(0);
540   unsigned IntegerSize = Op->getType()->getScalarSizeInBits();
541   if (IntegerSize <= DL.getPointerSizeInBits()) {
542     std::pair<Value *, APInt> BaseAndOffset = ConstantOffsetPtrs.lookup(Op);
543     if (BaseAndOffset.first)
544       ConstantOffsetPtrs[&I] = BaseAndOffset;
545   }
546 
547   // "Propagate" SROA here in the same manner as we do for ptrtoint above.
548   Value *SROAArg;
549   DenseMap<Value *, int>::iterator CostIt;
550   if (lookupSROAArgAndCost(Op, SROAArg, CostIt))
551     SROAArgValues[&I] = SROAArg;
552 
553   return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
554 }
555 
556 bool CallAnalyzer::visitCastInst(CastInst &I) {
557   // Propagate constants through ptrtoint.
558   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
559         return ConstantExpr::getCast(I.getOpcode(), COps[0], I.getType());
560       }))
561     return true;
562 
563   // Disable SROA in the face of arbitrary casts we don't whitelist elsewhere.
564   disableSROA(I.getOperand(0));
565 
566   return TargetTransformInfo::TCC_Free == TTI.getUserCost(&I);
567 }
568 
569 bool CallAnalyzer::visitUnaryInstruction(UnaryInstruction &I) {
570   Value *Operand = I.getOperand(0);
571   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
572         return ConstantFoldInstOperands(&I, COps[0], DL);
573       }))
574     return true;
575 
576   // Disable any SROA on the argument to arbitrary unary operators.
577   disableSROA(Operand);
578 
579   return false;
580 }
581 
582 bool CallAnalyzer::paramHasAttr(Argument *A, Attribute::AttrKind Attr) {
583   return CandidateCS.paramHasAttr(A->getArgNo(), Attr);
584 }
585 
586 bool CallAnalyzer::isKnownNonNullInCallee(Value *V) {
587   // Does the *call site* have the NonNull attribute set on an argument?  We
588   // use the attribute on the call site to memoize any analysis done in the
589   // caller. This will also trip if the callee function has a non-null
590   // parameter attribute, but that's a less interesting case because hopefully
591   // the callee would already have been simplified based on that.
592   if (Argument *A = dyn_cast<Argument>(V))
593     if (paramHasAttr(A, Attribute::NonNull))
594       return true;
595 
596   // Is this an alloca in the caller?  This is distinct from the attribute case
597   // above because attributes aren't updated within the inliner itself and we
598   // always want to catch the alloca derived case.
599   if (isAllocaDerivedArg(V))
600     // We can actually predict the result of comparisons between an
601     // alloca-derived value and null. Note that this fires regardless of
602     // SROA firing.
603     return true;
604 
605   return false;
606 }
607 
608 bool CallAnalyzer::allowSizeGrowth(CallSite CS) {
609   // If the normal destination of the invoke or the parent block of the call
610   // site is unreachable-terminated, there is little point in inlining this
611   // unless there is literally zero cost.
612   // FIXME: Note that it is possible that an unreachable-terminated block has a
613   // hot entry. For example, in below scenario inlining hot_call_X() may be
614   // beneficial :
615   // main() {
616   //   hot_call_1();
617   //   ...
618   //   hot_call_N()
619   //   exit(0);
620   // }
621   // For now, we are not handling this corner case here as it is rare in real
622   // code. In future, we should elaborate this based on BPI and BFI in more
623   // general threshold adjusting heuristics in updateThreshold().
624   Instruction *Instr = CS.getInstruction();
625   if (InvokeInst *II = dyn_cast<InvokeInst>(Instr)) {
626     if (isa<UnreachableInst>(II->getNormalDest()->getTerminator()))
627       return false;
628   } else if (isa<UnreachableInst>(Instr->getParent()->getTerminator()))
629     return false;
630 
631   return true;
632 }
633 
634 void CallAnalyzer::updateThreshold(CallSite CS, Function &Callee) {
635   // If no size growth is allowed for this inlining, set Threshold to 0.
636   if (!allowSizeGrowth(CS)) {
637     Threshold = 0;
638     return;
639   }
640 
641   Function *Caller = CS.getCaller();
642 
643   // return min(A, B) if B is valid.
644   auto MinIfValid = [](int A, Optional<int> B) {
645     return B ? std::min(A, B.getValue()) : A;
646   };
647 
648   // return max(A, B) if B is valid.
649   auto MaxIfValid = [](int A, Optional<int> B) {
650     return B ? std::max(A, B.getValue()) : A;
651   };
652 
653   // Use the OptMinSizeThreshold or OptSizeThreshold knob if they are available
654   // and reduce the threshold if the caller has the necessary attribute.
655   if (Caller->optForMinSize())
656     Threshold = MinIfValid(Threshold, Params.OptMinSizeThreshold);
657   else if (Caller->optForSize())
658     Threshold = MinIfValid(Threshold, Params.OptSizeThreshold);
659 
660   // Adjust the threshold based on inlinehint attribute and profile based
661   // hotness information if the caller does not have MinSize attribute.
662   if (!Caller->optForMinSize()) {
663     if (Callee.hasFnAttribute(Attribute::InlineHint))
664       Threshold = MaxIfValid(Threshold, Params.HintThreshold);
665     if (PSI) {
666       BlockFrequencyInfo *CallerBFI = GetBFI ? &((*GetBFI)(*Caller)) : nullptr;
667       // FIXME: After switching to the new passmanager, simplify the logic below
668       // by checking only the callsite hotness/coldness. The check for CallerBFI
669       // exists only because we do not have BFI available with the old PM.
670       //
671       // Use callee's hotness information only if we have no way of determining
672       // callsite's hotness information. Callsite hotness can be determined if
673       // sample profile is used (which adds hotness metadata to calls) or if
674       // caller's BlockFrequencyInfo is available.
675       if (CallerBFI || PSI->hasSampleProfile()) {
676         if (PSI->isHotCallSite(CS, CallerBFI)) {
677           DEBUG(dbgs() << "Hot callsite.\n");
678           Threshold = Params.HotCallSiteThreshold.getValue();
679         } else if (PSI->isColdCallSite(CS, CallerBFI)) {
680           DEBUG(dbgs() << "Cold callsite.\n");
681           Threshold = MinIfValid(Threshold, Params.ColdCallSiteThreshold);
682         }
683       } else {
684         if (PSI->isFunctionEntryHot(&Callee)) {
685           DEBUG(dbgs() << "Hot callee.\n");
686           // If callsite hotness can not be determined, we may still know
687           // that the callee is hot and treat it as a weaker hint for threshold
688           // increase.
689           Threshold = MaxIfValid(Threshold, Params.HintThreshold);
690         } else if (PSI->isFunctionEntryCold(&Callee)) {
691           DEBUG(dbgs() << "Cold callee.\n");
692           Threshold = MinIfValid(Threshold, Params.ColdThreshold);
693         }
694       }
695     }
696   }
697 
698   // Finally, take the target-specific inlining threshold multiplier into
699   // account.
700   Threshold *= TTI.getInliningThresholdMultiplier();
701 }
702 
703 bool CallAnalyzer::visitCmpInst(CmpInst &I) {
704   Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
705   // First try to handle simplified comparisons.
706   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
707         return ConstantExpr::getCompare(I.getPredicate(), COps[0], COps[1]);
708       }))
709     return true;
710 
711   if (I.getOpcode() == Instruction::FCmp)
712     return false;
713 
714   // Otherwise look for a comparison between constant offset pointers with
715   // a common base.
716   Value *LHSBase, *RHSBase;
717   APInt LHSOffset, RHSOffset;
718   std::tie(LHSBase, LHSOffset) = ConstantOffsetPtrs.lookup(LHS);
719   if (LHSBase) {
720     std::tie(RHSBase, RHSOffset) = ConstantOffsetPtrs.lookup(RHS);
721     if (RHSBase && LHSBase == RHSBase) {
722       // We have common bases, fold the icmp to a constant based on the
723       // offsets.
724       Constant *CLHS = ConstantInt::get(LHS->getContext(), LHSOffset);
725       Constant *CRHS = ConstantInt::get(RHS->getContext(), RHSOffset);
726       if (Constant *C = ConstantExpr::getICmp(I.getPredicate(), CLHS, CRHS)) {
727         SimplifiedValues[&I] = C;
728         ++NumConstantPtrCmps;
729         return true;
730       }
731     }
732   }
733 
734   // If the comparison is an equality comparison with null, we can simplify it
735   // if we know the value (argument) can't be null
736   if (I.isEquality() && isa<ConstantPointerNull>(I.getOperand(1)) &&
737       isKnownNonNullInCallee(I.getOperand(0))) {
738     bool IsNotEqual = I.getPredicate() == CmpInst::ICMP_NE;
739     SimplifiedValues[&I] = IsNotEqual ? ConstantInt::getTrue(I.getType())
740                                       : ConstantInt::getFalse(I.getType());
741     return true;
742   }
743   // Finally check for SROA candidates in comparisons.
744   Value *SROAArg;
745   DenseMap<Value *, int>::iterator CostIt;
746   if (lookupSROAArgAndCost(I.getOperand(0), SROAArg, CostIt)) {
747     if (isa<ConstantPointerNull>(I.getOperand(1))) {
748       accumulateSROACost(CostIt, InlineConstants::InstrCost);
749       return true;
750     }
751 
752     disableSROA(CostIt);
753   }
754 
755   return false;
756 }
757 
758 bool CallAnalyzer::visitSub(BinaryOperator &I) {
759   // Try to handle a special case: we can fold computing the difference of two
760   // constant-related pointers.
761   Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
762   Value *LHSBase, *RHSBase;
763   APInt LHSOffset, RHSOffset;
764   std::tie(LHSBase, LHSOffset) = ConstantOffsetPtrs.lookup(LHS);
765   if (LHSBase) {
766     std::tie(RHSBase, RHSOffset) = ConstantOffsetPtrs.lookup(RHS);
767     if (RHSBase && LHSBase == RHSBase) {
768       // We have common bases, fold the subtract to a constant based on the
769       // offsets.
770       Constant *CLHS = ConstantInt::get(LHS->getContext(), LHSOffset);
771       Constant *CRHS = ConstantInt::get(RHS->getContext(), RHSOffset);
772       if (Constant *C = ConstantExpr::getSub(CLHS, CRHS)) {
773         SimplifiedValues[&I] = C;
774         ++NumConstantPtrDiffs;
775         return true;
776       }
777     }
778   }
779 
780   // Otherwise, fall back to the generic logic for simplifying and handling
781   // instructions.
782   return Base::visitSub(I);
783 }
784 
785 bool CallAnalyzer::visitBinaryOperator(BinaryOperator &I) {
786   Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
787   auto Evaluate = [&](SmallVectorImpl<Constant *> &COps) {
788     Value *SimpleV = nullptr;
789     if (auto FI = dyn_cast<FPMathOperator>(&I))
790       SimpleV = SimplifyFPBinOp(I.getOpcode(), COps[0], COps[1],
791                                 FI->getFastMathFlags(), DL);
792     else
793       SimpleV = SimplifyBinOp(I.getOpcode(), COps[0], COps[1], DL);
794     return dyn_cast_or_null<Constant>(SimpleV);
795   };
796 
797   if (simplifyInstruction(I, Evaluate))
798     return true;
799 
800   // Disable any SROA on arguments to arbitrary, unsimplified binary operators.
801   disableSROA(LHS);
802   disableSROA(RHS);
803 
804   return false;
805 }
806 
807 bool CallAnalyzer::visitLoad(LoadInst &I) {
808   Value *SROAArg;
809   DenseMap<Value *, int>::iterator CostIt;
810   if (lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt)) {
811     if (I.isSimple()) {
812       accumulateSROACost(CostIt, InlineConstants::InstrCost);
813       return true;
814     }
815 
816     disableSROA(CostIt);
817   }
818 
819   return false;
820 }
821 
822 bool CallAnalyzer::visitStore(StoreInst &I) {
823   Value *SROAArg;
824   DenseMap<Value *, int>::iterator CostIt;
825   if (lookupSROAArgAndCost(I.getPointerOperand(), SROAArg, CostIt)) {
826     if (I.isSimple()) {
827       accumulateSROACost(CostIt, InlineConstants::InstrCost);
828       return true;
829     }
830 
831     disableSROA(CostIt);
832   }
833 
834   return false;
835 }
836 
837 bool CallAnalyzer::visitExtractValue(ExtractValueInst &I) {
838   // Constant folding for extract value is trivial.
839   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
840         return ConstantExpr::getExtractValue(COps[0], I.getIndices());
841       }))
842     return true;
843 
844   // SROA can look through these but give them a cost.
845   return false;
846 }
847 
848 bool CallAnalyzer::visitInsertValue(InsertValueInst &I) {
849   // Constant folding for insert value is trivial.
850   if (simplifyInstruction(I, [&](SmallVectorImpl<Constant *> &COps) {
851         return ConstantExpr::getInsertValue(/*AggregateOperand*/ COps[0],
852                                             /*InsertedValueOperand*/ COps[1],
853                                             I.getIndices());
854       }))
855     return true;
856 
857   // SROA can look through these but give them a cost.
858   return false;
859 }
860 
861 /// \brief Try to simplify a call site.
862 ///
863 /// Takes a concrete function and callsite and tries to actually simplify it by
864 /// analyzing the arguments and call itself with instsimplify. Returns true if
865 /// it has simplified the callsite to some other entity (a constant), making it
866 /// free.
867 bool CallAnalyzer::simplifyCallSite(Function *F, CallSite CS) {
868   // FIXME: Using the instsimplify logic directly for this is inefficient
869   // because we have to continually rebuild the argument list even when no
870   // simplifications can be performed. Until that is fixed with remapping
871   // inside of instsimplify, directly constant fold calls here.
872   if (!canConstantFoldCallTo(CS, F))
873     return false;
874 
875   // Try to re-map the arguments to constants.
876   SmallVector<Constant *, 4> ConstantArgs;
877   ConstantArgs.reserve(CS.arg_size());
878   for (CallSite::arg_iterator I = CS.arg_begin(), E = CS.arg_end(); I != E;
879        ++I) {
880     Constant *C = dyn_cast<Constant>(*I);
881     if (!C)
882       C = dyn_cast_or_null<Constant>(SimplifiedValues.lookup(*I));
883     if (!C)
884       return false; // This argument doesn't map to a constant.
885 
886     ConstantArgs.push_back(C);
887   }
888   if (Constant *C = ConstantFoldCall(CS, F, ConstantArgs)) {
889     SimplifiedValues[CS.getInstruction()] = C;
890     return true;
891   }
892 
893   return false;
894 }
895 
896 bool CallAnalyzer::visitCallSite(CallSite CS) {
897   if (CS.hasFnAttr(Attribute::ReturnsTwice) &&
898       !F.hasFnAttribute(Attribute::ReturnsTwice)) {
899     // This aborts the entire analysis.
900     ExposesReturnsTwice = true;
901     return false;
902   }
903   if (CS.isCall() && cast<CallInst>(CS.getInstruction())->cannotDuplicate())
904     ContainsNoDuplicateCall = true;
905 
906   if (Function *F = CS.getCalledFunction()) {
907     // When we have a concrete function, first try to simplify it directly.
908     if (simplifyCallSite(F, CS))
909       return true;
910 
911     // Next check if it is an intrinsic we know about.
912     // FIXME: Lift this into part of the InstVisitor.
913     if (IntrinsicInst *II = dyn_cast<IntrinsicInst>(CS.getInstruction())) {
914       switch (II->getIntrinsicID()) {
915       default:
916         return Base::visitCallSite(CS);
917 
918       case Intrinsic::load_relative:
919         // This is normally lowered to 4 LLVM instructions.
920         Cost += 3 * InlineConstants::InstrCost;
921         return false;
922 
923       case Intrinsic::memset:
924       case Intrinsic::memcpy:
925       case Intrinsic::memmove:
926         // SROA can usually chew through these intrinsics, but they aren't free.
927         return false;
928       case Intrinsic::localescape:
929         HasFrameEscape = true;
930         return false;
931       }
932     }
933 
934     if (F == CS.getInstruction()->getParent()->getParent()) {
935       // This flag will fully abort the analysis, so don't bother with anything
936       // else.
937       IsRecursiveCall = true;
938       return false;
939     }
940 
941     if (TTI.isLoweredToCall(F)) {
942       // We account for the average 1 instruction per call argument setup
943       // here.
944       Cost += CS.arg_size() * InlineConstants::InstrCost;
945 
946       // Everything other than inline ASM will also have a significant cost
947       // merely from making the call.
948       if (!isa<InlineAsm>(CS.getCalledValue()))
949         Cost += InlineConstants::CallPenalty;
950     }
951 
952     return Base::visitCallSite(CS);
953   }
954 
955   // Otherwise we're in a very special case -- an indirect function call. See
956   // if we can be particularly clever about this.
957   Value *Callee = CS.getCalledValue();
958 
959   // First, pay the price of the argument setup. We account for the average
960   // 1 instruction per call argument setup here.
961   Cost += CS.arg_size() * InlineConstants::InstrCost;
962 
963   // Next, check if this happens to be an indirect function call to a known
964   // function in this inline context. If not, we've done all we can.
965   Function *F = dyn_cast_or_null<Function>(SimplifiedValues.lookup(Callee));
966   if (!F)
967     return Base::visitCallSite(CS);
968 
969   // If we have a constant that we are calling as a function, we can peer
970   // through it and see the function target. This happens not infrequently
971   // during devirtualization and so we want to give it a hefty bonus for
972   // inlining, but cap that bonus in the event that inlining wouldn't pan
973   // out. Pretend to inline the function, with a custom threshold.
974   auto IndirectCallParams = Params;
975   IndirectCallParams.DefaultThreshold = InlineConstants::IndirectCallThreshold;
976   CallAnalyzer CA(TTI, GetAssumptionCache, GetBFI, PSI, *F, CS,
977                   IndirectCallParams);
978   if (CA.analyzeCall(CS)) {
979     // We were able to inline the indirect call! Subtract the cost from the
980     // threshold to get the bonus we want to apply, but don't go below zero.
981     Cost -= std::max(0, CA.getThreshold() - CA.getCost());
982   }
983 
984   return Base::visitCallSite(CS);
985 }
986 
987 bool CallAnalyzer::visitReturnInst(ReturnInst &RI) {
988   // At least one return instruction will be free after inlining.
989   bool Free = !HasReturn;
990   HasReturn = true;
991   return Free;
992 }
993 
994 bool CallAnalyzer::visitBranchInst(BranchInst &BI) {
995   // We model unconditional branches as essentially free -- they really
996   // shouldn't exist at all, but handling them makes the behavior of the
997   // inliner more regular and predictable. Interestingly, conditional branches
998   // which will fold away are also free.
999   return BI.isUnconditional() || isa<ConstantInt>(BI.getCondition()) ||
1000          dyn_cast_or_null<ConstantInt>(
1001              SimplifiedValues.lookup(BI.getCondition()));
1002 }
1003 
1004 bool CallAnalyzer::visitSwitchInst(SwitchInst &SI) {
1005   // We model unconditional switches as free, see the comments on handling
1006   // branches.
1007   if (isa<ConstantInt>(SI.getCondition()))
1008     return true;
1009   if (Value *V = SimplifiedValues.lookup(SI.getCondition()))
1010     if (isa<ConstantInt>(V))
1011       return true;
1012 
1013   // Assume the most general case where the swith is lowered into
1014   // either a jump table, bit test, or a balanced binary tree consisting of
1015   // case clusters without merging adjacent clusters with the same
1016   // destination. We do not consider the switches that are lowered with a mix
1017   // of jump table/bit test/binary search tree. The cost of the switch is
1018   // proportional to the size of the tree or the size of jump table range.
1019   //
1020   // NB: We convert large switches which are just used to initialize large phi
1021   // nodes to lookup tables instead in simplify-cfg, so this shouldn't prevent
1022   // inlining those. It will prevent inlining in cases where the optimization
1023   // does not (yet) fire.
1024 
1025   // Maximum valid cost increased in this function.
1026   int CostUpperBound = INT_MAX - InlineConstants::InstrCost - 1;
1027 
1028   // Exit early for a large switch, assuming one case needs at least one
1029   // instruction.
1030   // FIXME: This is not true for a bit test, but ignore such case for now to
1031   // save compile-time.
1032   int64_t CostLowerBound =
1033       std::min((int64_t)CostUpperBound,
1034                (int64_t)SI.getNumCases() * InlineConstants::InstrCost + Cost);
1035 
1036   if (CostLowerBound > Threshold) {
1037     Cost = CostLowerBound;
1038     return false;
1039   }
1040 
1041   unsigned JumpTableSize = 0;
1042   unsigned NumCaseCluster =
1043       TTI.getEstimatedNumberOfCaseClusters(SI, JumpTableSize);
1044 
1045   // If suitable for a jump table, consider the cost for the table size and
1046   // branch to destination.
1047   if (JumpTableSize) {
1048     int64_t JTCost = (int64_t)JumpTableSize * InlineConstants::InstrCost +
1049                      4 * InlineConstants::InstrCost;
1050 
1051     Cost = std::min((int64_t)CostUpperBound, JTCost + Cost);
1052     return false;
1053   }
1054 
1055   // Considering forming a binary search, we should find the number of nodes
1056   // which is same as the number of comparisons when lowered. For a given
1057   // number of clusters, n, we can define a recursive function, f(n), to find
1058   // the number of nodes in the tree. The recursion is :
1059   // f(n) = 1 + f(n/2) + f (n - n/2), when n > 3,
1060   // and f(n) = n, when n <= 3.
1061   // This will lead a binary tree where the leaf should be either f(2) or f(3)
1062   // when n > 3.  So, the number of comparisons from leaves should be n, while
1063   // the number of non-leaf should be :
1064   //   2^(log2(n) - 1) - 1
1065   //   = 2^log2(n) * 2^-1 - 1
1066   //   = n / 2 - 1.
1067   // Considering comparisons from leaf and non-leaf nodes, we can estimate the
1068   // number of comparisons in a simple closed form :
1069   //   n + n / 2 - 1 = n * 3 / 2 - 1
1070   if (NumCaseCluster <= 3) {
1071     // Suppose a comparison includes one compare and one conditional branch.
1072     Cost += NumCaseCluster * 2 * InlineConstants::InstrCost;
1073     return false;
1074   }
1075 
1076   int64_t ExpectedNumberOfCompare = 3 * (int64_t)NumCaseCluster / 2 - 1;
1077   int64_t SwitchCost =
1078       ExpectedNumberOfCompare * 2 * InlineConstants::InstrCost;
1079 
1080   Cost = std::min((int64_t)CostUpperBound, SwitchCost + Cost);
1081   return false;
1082 }
1083 
1084 bool CallAnalyzer::visitIndirectBrInst(IndirectBrInst &IBI) {
1085   // We never want to inline functions that contain an indirectbr.  This is
1086   // incorrect because all the blockaddress's (in static global initializers
1087   // for example) would be referring to the original function, and this
1088   // indirect jump would jump from the inlined copy of the function into the
1089   // original function which is extremely undefined behavior.
1090   // FIXME: This logic isn't really right; we can safely inline functions with
1091   // indirectbr's as long as no other function or global references the
1092   // blockaddress of a block within the current function.
1093   HasIndirectBr = true;
1094   return false;
1095 }
1096 
1097 bool CallAnalyzer::visitResumeInst(ResumeInst &RI) {
1098   // FIXME: It's not clear that a single instruction is an accurate model for
1099   // the inline cost of a resume instruction.
1100   return false;
1101 }
1102 
1103 bool CallAnalyzer::visitCleanupReturnInst(CleanupReturnInst &CRI) {
1104   // FIXME: It's not clear that a single instruction is an accurate model for
1105   // the inline cost of a cleanupret instruction.
1106   return false;
1107 }
1108 
1109 bool CallAnalyzer::visitCatchReturnInst(CatchReturnInst &CRI) {
1110   // FIXME: It's not clear that a single instruction is an accurate model for
1111   // the inline cost of a catchret instruction.
1112   return false;
1113 }
1114 
1115 bool CallAnalyzer::visitUnreachableInst(UnreachableInst &I) {
1116   // FIXME: It might be reasonably to discount the cost of instructions leading
1117   // to unreachable as they have the lowest possible impact on both runtime and
1118   // code size.
1119   return true; // No actual code is needed for unreachable.
1120 }
1121 
1122 bool CallAnalyzer::visitInstruction(Instruction &I) {
1123   // Some instructions are free. All of the free intrinsics can also be
1124   // handled by SROA, etc.
1125   if (TargetTransformInfo::TCC_Free == TTI.getUserCost(&I))
1126     return true;
1127 
1128   // We found something we don't understand or can't handle. Mark any SROA-able
1129   // values in the operand list as no longer viable.
1130   for (User::op_iterator OI = I.op_begin(), OE = I.op_end(); OI != OE; ++OI)
1131     disableSROA(*OI);
1132 
1133   return false;
1134 }
1135 
1136 /// \brief Analyze a basic block for its contribution to the inline cost.
1137 ///
1138 /// This method walks the analyzer over every instruction in the given basic
1139 /// block and accounts for their cost during inlining at this callsite. It
1140 /// aborts early if the threshold has been exceeded or an impossible to inline
1141 /// construct has been detected. It returns false if inlining is no longer
1142 /// viable, and true if inlining remains viable.
1143 bool CallAnalyzer::analyzeBlock(BasicBlock *BB,
1144                                 SmallPtrSetImpl<const Value *> &EphValues) {
1145   for (BasicBlock::iterator I = BB->begin(), E = BB->end(); I != E; ++I) {
1146     // FIXME: Currently, the number of instructions in a function regardless of
1147     // our ability to simplify them during inline to constants or dead code,
1148     // are actually used by the vector bonus heuristic. As long as that's true,
1149     // we have to special case debug intrinsics here to prevent differences in
1150     // inlining due to debug symbols. Eventually, the number of unsimplified
1151     // instructions shouldn't factor into the cost computation, but until then,
1152     // hack around it here.
1153     if (isa<DbgInfoIntrinsic>(I))
1154       continue;
1155 
1156     // Skip ephemeral values.
1157     if (EphValues.count(&*I))
1158       continue;
1159 
1160     ++NumInstructions;
1161     if (isa<ExtractElementInst>(I) || I->getType()->isVectorTy())
1162       ++NumVectorInstructions;
1163 
1164     // If the instruction is floating point, and the target says this operation
1165     // is expensive or the function has the "use-soft-float" attribute, this may
1166     // eventually become a library call. Treat the cost as such.
1167     if (I->getType()->isFloatingPointTy()) {
1168       // If the function has the "use-soft-float" attribute, mark it as
1169       // expensive.
1170       if (TTI.getFPOpCost(I->getType()) == TargetTransformInfo::TCC_Expensive ||
1171           (F.getFnAttribute("use-soft-float").getValueAsString() == "true"))
1172         Cost += InlineConstants::CallPenalty;
1173     }
1174 
1175     // If the instruction simplified to a constant, there is no cost to this
1176     // instruction. Visit the instructions using our InstVisitor to account for
1177     // all of the per-instruction logic. The visit tree returns true if we
1178     // consumed the instruction in any way, and false if the instruction's base
1179     // cost should count against inlining.
1180     if (Base::visit(&*I))
1181       ++NumInstructionsSimplified;
1182     else
1183       Cost += InlineConstants::InstrCost;
1184 
1185     // If the visit this instruction detected an uninlinable pattern, abort.
1186     if (IsRecursiveCall || ExposesReturnsTwice || HasDynamicAlloca ||
1187         HasIndirectBr || HasFrameEscape)
1188       return false;
1189 
1190     // If the caller is a recursive function then we don't want to inline
1191     // functions which allocate a lot of stack space because it would increase
1192     // the caller stack usage dramatically.
1193     if (IsCallerRecursive &&
1194         AllocatedSize > InlineConstants::TotalAllocaSizeRecursiveCaller)
1195       return false;
1196 
1197     // Check if we've past the maximum possible threshold so we don't spin in
1198     // huge basic blocks that will never inline.
1199     if (Cost > Threshold)
1200       return false;
1201   }
1202 
1203   return true;
1204 }
1205 
1206 /// \brief Compute the base pointer and cumulative constant offsets for V.
1207 ///
1208 /// This strips all constant offsets off of V, leaving it the base pointer, and
1209 /// accumulates the total constant offset applied in the returned constant. It
1210 /// returns 0 if V is not a pointer, and returns the constant '0' if there are
1211 /// no constant offsets applied.
1212 ConstantInt *CallAnalyzer::stripAndComputeInBoundsConstantOffsets(Value *&V) {
1213   if (!V->getType()->isPointerTy())
1214     return nullptr;
1215 
1216   unsigned IntPtrWidth = DL.getPointerSizeInBits();
1217   APInt Offset = APInt::getNullValue(IntPtrWidth);
1218 
1219   // Even though we don't look through PHI nodes, we could be called on an
1220   // instruction in an unreachable block, which may be on a cycle.
1221   SmallPtrSet<Value *, 4> Visited;
1222   Visited.insert(V);
1223   do {
1224     if (GEPOperator *GEP = dyn_cast<GEPOperator>(V)) {
1225       if (!GEP->isInBounds() || !accumulateGEPOffset(*GEP, Offset))
1226         return nullptr;
1227       V = GEP->getPointerOperand();
1228     } else if (Operator::getOpcode(V) == Instruction::BitCast) {
1229       V = cast<Operator>(V)->getOperand(0);
1230     } else if (GlobalAlias *GA = dyn_cast<GlobalAlias>(V)) {
1231       if (GA->isInterposable())
1232         break;
1233       V = GA->getAliasee();
1234     } else {
1235       break;
1236     }
1237     assert(V->getType()->isPointerTy() && "Unexpected operand type!");
1238   } while (Visited.insert(V).second);
1239 
1240   Type *IntPtrTy = DL.getIntPtrType(V->getContext());
1241   return cast<ConstantInt>(ConstantInt::get(IntPtrTy, Offset));
1242 }
1243 
1244 /// \brief Analyze a call site for potential inlining.
1245 ///
1246 /// Returns true if inlining this call is viable, and false if it is not
1247 /// viable. It computes the cost and adjusts the threshold based on numerous
1248 /// factors and heuristics. If this method returns false but the computed cost
1249 /// is below the computed threshold, then inlining was forcibly disabled by
1250 /// some artifact of the routine.
1251 bool CallAnalyzer::analyzeCall(CallSite CS) {
1252   ++NumCallsAnalyzed;
1253 
1254   // Perform some tweaks to the cost and threshold based on the direct
1255   // callsite information.
1256 
1257   // We want to more aggressively inline vector-dense kernels, so up the
1258   // threshold, and we'll lower it if the % of vector instructions gets too
1259   // low. Note that these bonuses are some what arbitrary and evolved over time
1260   // by accident as much as because they are principled bonuses.
1261   //
1262   // FIXME: It would be nice to remove all such bonuses. At least it would be
1263   // nice to base the bonus values on something more scientific.
1264   assert(NumInstructions == 0);
1265   assert(NumVectorInstructions == 0);
1266 
1267   // Update the threshold based on callsite properties
1268   updateThreshold(CS, F);
1269 
1270   FiftyPercentVectorBonus = 3 * Threshold / 2;
1271   TenPercentVectorBonus = 3 * Threshold / 4;
1272 
1273   // Track whether the post-inlining function would have more than one basic
1274   // block. A single basic block is often intended for inlining. Balloon the
1275   // threshold by 50% until we pass the single-BB phase.
1276   bool SingleBB = true;
1277   int SingleBBBonus = Threshold / 2;
1278 
1279   // Speculatively apply all possible bonuses to Threshold. If cost exceeds
1280   // this Threshold any time, and cost cannot decrease, we can stop processing
1281   // the rest of the function body.
1282   Threshold += (SingleBBBonus + FiftyPercentVectorBonus);
1283 
1284   // Give out bonuses for the callsite, as the instructions setting them up
1285   // will be gone after inlining.
1286   Cost -= getCallsiteCost(CS, DL);
1287 
1288   // If there is only one call of the function, and it has internal linkage,
1289   // the cost of inlining it drops dramatically.
1290   bool OnlyOneCallAndLocalLinkage =
1291       F.hasLocalLinkage() && F.hasOneUse() && &F == CS.getCalledFunction();
1292   if (OnlyOneCallAndLocalLinkage)
1293     Cost -= InlineConstants::LastCallToStaticBonus;
1294 
1295   // If this function uses the coldcc calling convention, prefer not to inline
1296   // it.
1297   if (F.getCallingConv() == CallingConv::Cold)
1298     Cost += InlineConstants::ColdccPenalty;
1299 
1300   // Check if we're done. This can happen due to bonuses and penalties.
1301   if (Cost > Threshold)
1302     return false;
1303 
1304   if (F.empty())
1305     return true;
1306 
1307   Function *Caller = CS.getInstruction()->getParent()->getParent();
1308   // Check if the caller function is recursive itself.
1309   for (User *U : Caller->users()) {
1310     CallSite Site(U);
1311     if (!Site)
1312       continue;
1313     Instruction *I = Site.getInstruction();
1314     if (I->getParent()->getParent() == Caller) {
1315       IsCallerRecursive = true;
1316       break;
1317     }
1318   }
1319 
1320   // Populate our simplified values by mapping from function arguments to call
1321   // arguments with known important simplifications.
1322   CallSite::arg_iterator CAI = CS.arg_begin();
1323   for (Function::arg_iterator FAI = F.arg_begin(), FAE = F.arg_end();
1324        FAI != FAE; ++FAI, ++CAI) {
1325     assert(CAI != CS.arg_end());
1326     if (Constant *C = dyn_cast<Constant>(CAI))
1327       SimplifiedValues[&*FAI] = C;
1328 
1329     Value *PtrArg = *CAI;
1330     if (ConstantInt *C = stripAndComputeInBoundsConstantOffsets(PtrArg)) {
1331       ConstantOffsetPtrs[&*FAI] = std::make_pair(PtrArg, C->getValue());
1332 
1333       // We can SROA any pointer arguments derived from alloca instructions.
1334       if (isa<AllocaInst>(PtrArg)) {
1335         SROAArgValues[&*FAI] = PtrArg;
1336         SROAArgCosts[PtrArg] = 0;
1337       }
1338     }
1339   }
1340   NumConstantArgs = SimplifiedValues.size();
1341   NumConstantOffsetPtrArgs = ConstantOffsetPtrs.size();
1342   NumAllocaArgs = SROAArgValues.size();
1343 
1344   // FIXME: If a caller has multiple calls to a callee, we end up recomputing
1345   // the ephemeral values multiple times (and they're completely determined by
1346   // the callee, so this is purely duplicate work).
1347   SmallPtrSet<const Value *, 32> EphValues;
1348   CodeMetrics::collectEphemeralValues(&F, &GetAssumptionCache(F), EphValues);
1349 
1350   // The worklist of live basic blocks in the callee *after* inlining. We avoid
1351   // adding basic blocks of the callee which can be proven to be dead for this
1352   // particular call site in order to get more accurate cost estimates. This
1353   // requires a somewhat heavyweight iteration pattern: we need to walk the
1354   // basic blocks in a breadth-first order as we insert live successors. To
1355   // accomplish this, prioritizing for small iterations because we exit after
1356   // crossing our threshold, we use a small-size optimized SetVector.
1357   typedef SetVector<BasicBlock *, SmallVector<BasicBlock *, 16>,
1358                     SmallPtrSet<BasicBlock *, 16>>
1359       BBSetVector;
1360   BBSetVector BBWorklist;
1361   BBWorklist.insert(&F.getEntryBlock());
1362   // Note that we *must not* cache the size, this loop grows the worklist.
1363   for (unsigned Idx = 0; Idx != BBWorklist.size(); ++Idx) {
1364     // Bail out the moment we cross the threshold. This means we'll under-count
1365     // the cost, but only when undercounting doesn't matter.
1366     if (Cost > Threshold)
1367       break;
1368 
1369     BasicBlock *BB = BBWorklist[Idx];
1370     if (BB->empty())
1371       continue;
1372 
1373     // Disallow inlining a blockaddress. A blockaddress only has defined
1374     // behavior for an indirect branch in the same function, and we do not
1375     // currently support inlining indirect branches. But, the inliner may not
1376     // see an indirect branch that ends up being dead code at a particular call
1377     // site. If the blockaddress escapes the function, e.g., via a global
1378     // variable, inlining may lead to an invalid cross-function reference.
1379     if (BB->hasAddressTaken())
1380       return false;
1381 
1382     // Analyze the cost of this block. If we blow through the threshold, this
1383     // returns false, and we can bail on out.
1384     if (!analyzeBlock(BB, EphValues))
1385       return false;
1386 
1387     TerminatorInst *TI = BB->getTerminator();
1388 
1389     // Add in the live successors by first checking whether we have terminator
1390     // that may be simplified based on the values simplified by this call.
1391     if (BranchInst *BI = dyn_cast<BranchInst>(TI)) {
1392       if (BI->isConditional()) {
1393         Value *Cond = BI->getCondition();
1394         if (ConstantInt *SimpleCond =
1395                 dyn_cast_or_null<ConstantInt>(SimplifiedValues.lookup(Cond))) {
1396           BBWorklist.insert(BI->getSuccessor(SimpleCond->isZero() ? 1 : 0));
1397           continue;
1398         }
1399       }
1400     } else if (SwitchInst *SI = dyn_cast<SwitchInst>(TI)) {
1401       Value *Cond = SI->getCondition();
1402       if (ConstantInt *SimpleCond =
1403               dyn_cast_or_null<ConstantInt>(SimplifiedValues.lookup(Cond))) {
1404         BBWorklist.insert(SI->findCaseValue(SimpleCond)->getCaseSuccessor());
1405         continue;
1406       }
1407     }
1408 
1409     // If we're unable to select a particular successor, just count all of
1410     // them.
1411     for (unsigned TIdx = 0, TSize = TI->getNumSuccessors(); TIdx != TSize;
1412          ++TIdx)
1413       BBWorklist.insert(TI->getSuccessor(TIdx));
1414 
1415     // If we had any successors at this point, than post-inlining is likely to
1416     // have them as well. Note that we assume any basic blocks which existed
1417     // due to branches or switches which folded above will also fold after
1418     // inlining.
1419     if (SingleBB && TI->getNumSuccessors() > 1) {
1420       // Take off the bonus we applied to the threshold.
1421       Threshold -= SingleBBBonus;
1422       SingleBB = false;
1423     }
1424   }
1425 
1426   // If this is a noduplicate call, we can still inline as long as
1427   // inlining this would cause the removal of the caller (so the instruction
1428   // is not actually duplicated, just moved).
1429   if (!OnlyOneCallAndLocalLinkage && ContainsNoDuplicateCall)
1430     return false;
1431 
1432   // We applied the maximum possible vector bonus at the beginning. Now,
1433   // subtract the excess bonus, if any, from the Threshold before
1434   // comparing against Cost.
1435   if (NumVectorInstructions <= NumInstructions / 10)
1436     Threshold -= FiftyPercentVectorBonus;
1437   else if (NumVectorInstructions <= NumInstructions / 2)
1438     Threshold -= (FiftyPercentVectorBonus - TenPercentVectorBonus);
1439 
1440   return Cost < std::max(1, Threshold);
1441 }
1442 
1443 #if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1444 /// \brief Dump stats about this call's analysis.
1445 LLVM_DUMP_METHOD void CallAnalyzer::dump() {
1446 #define DEBUG_PRINT_STAT(x) dbgs() << "      " #x ": " << x << "\n"
1447   DEBUG_PRINT_STAT(NumConstantArgs);
1448   DEBUG_PRINT_STAT(NumConstantOffsetPtrArgs);
1449   DEBUG_PRINT_STAT(NumAllocaArgs);
1450   DEBUG_PRINT_STAT(NumConstantPtrCmps);
1451   DEBUG_PRINT_STAT(NumConstantPtrDiffs);
1452   DEBUG_PRINT_STAT(NumInstructionsSimplified);
1453   DEBUG_PRINT_STAT(NumInstructions);
1454   DEBUG_PRINT_STAT(SROACostSavings);
1455   DEBUG_PRINT_STAT(SROACostSavingsLost);
1456   DEBUG_PRINT_STAT(ContainsNoDuplicateCall);
1457   DEBUG_PRINT_STAT(Cost);
1458   DEBUG_PRINT_STAT(Threshold);
1459 #undef DEBUG_PRINT_STAT
1460 }
1461 #endif
1462 
1463 /// \brief Test that there are no attribute conflicts between Caller and Callee
1464 ///        that prevent inlining.
1465 static bool functionsHaveCompatibleAttributes(Function *Caller,
1466                                               Function *Callee,
1467                                               TargetTransformInfo &TTI) {
1468   return TTI.areInlineCompatible(Caller, Callee) &&
1469          AttributeFuncs::areInlineCompatible(*Caller, *Callee);
1470 }
1471 
1472 int llvm::getCallsiteCost(CallSite CS, const DataLayout &DL) {
1473   int Cost = 0;
1474   for (unsigned I = 0, E = CS.arg_size(); I != E; ++I) {
1475     if (CS.isByValArgument(I)) {
1476       // We approximate the number of loads and stores needed by dividing the
1477       // size of the byval type by the target's pointer size.
1478       PointerType *PTy = cast<PointerType>(CS.getArgument(I)->getType());
1479       unsigned TypeSize = DL.getTypeSizeInBits(PTy->getElementType());
1480       unsigned PointerSize = DL.getPointerSizeInBits();
1481       // Ceiling division.
1482       unsigned NumStores = (TypeSize + PointerSize - 1) / PointerSize;
1483 
1484       // If it generates more than 8 stores it is likely to be expanded as an
1485       // inline memcpy so we take that as an upper bound. Otherwise we assume
1486       // one load and one store per word copied.
1487       // FIXME: The maxStoresPerMemcpy setting from the target should be used
1488       // here instead of a magic number of 8, but it's not available via
1489       // DataLayout.
1490       NumStores = std::min(NumStores, 8U);
1491 
1492       Cost += 2 * NumStores * InlineConstants::InstrCost;
1493     } else {
1494       // For non-byval arguments subtract off one instruction per call
1495       // argument.
1496       Cost += InlineConstants::InstrCost;
1497     }
1498   }
1499   // The call instruction also disappears after inlining.
1500   Cost += InlineConstants::InstrCost + InlineConstants::CallPenalty;
1501   return Cost;
1502 }
1503 
1504 InlineCost llvm::getInlineCost(
1505     CallSite CS, const InlineParams &Params, TargetTransformInfo &CalleeTTI,
1506     std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
1507     Optional<function_ref<BlockFrequencyInfo &(Function &)>> GetBFI,
1508     ProfileSummaryInfo *PSI) {
1509   return getInlineCost(CS, CS.getCalledFunction(), Params, CalleeTTI,
1510                        GetAssumptionCache, GetBFI, PSI);
1511 }
1512 
1513 InlineCost llvm::getInlineCost(
1514     CallSite CS, Function *Callee, const InlineParams &Params,
1515     TargetTransformInfo &CalleeTTI,
1516     std::function<AssumptionCache &(Function &)> &GetAssumptionCache,
1517     Optional<function_ref<BlockFrequencyInfo &(Function &)>> GetBFI,
1518     ProfileSummaryInfo *PSI) {
1519 
1520   // Cannot inline indirect calls.
1521   if (!Callee)
1522     return llvm::InlineCost::getNever();
1523 
1524   // Calls to functions with always-inline attributes should be inlined
1525   // whenever possible.
1526   if (CS.hasFnAttr(Attribute::AlwaysInline)) {
1527     if (isInlineViable(*Callee))
1528       return llvm::InlineCost::getAlways();
1529     return llvm::InlineCost::getNever();
1530   }
1531 
1532   // Never inline functions with conflicting attributes (unless callee has
1533   // always-inline attribute).
1534   if (!functionsHaveCompatibleAttributes(CS.getCaller(), Callee, CalleeTTI))
1535     return llvm::InlineCost::getNever();
1536 
1537   // Don't inline this call if the caller has the optnone attribute.
1538   if (CS.getCaller()->hasFnAttribute(Attribute::OptimizeNone))
1539     return llvm::InlineCost::getNever();
1540 
1541   // Don't inline functions which can be interposed at link-time.  Don't inline
1542   // functions marked noinline or call sites marked noinline.
1543   // Note: inlining non-exact non-interposable functions is fine, since we know
1544   // we have *a* correct implementation of the source level function.
1545   if (Callee->isInterposable() || Callee->hasFnAttribute(Attribute::NoInline) ||
1546       CS.isNoInline())
1547     return llvm::InlineCost::getNever();
1548 
1549   DEBUG(llvm::dbgs() << "      Analyzing call of " << Callee->getName()
1550                      << "...\n");
1551 
1552   CallAnalyzer CA(CalleeTTI, GetAssumptionCache, GetBFI, PSI, *Callee, CS,
1553                   Params);
1554   bool ShouldInline = CA.analyzeCall(CS);
1555 
1556   DEBUG(CA.dump());
1557 
1558   // Check if there was a reason to force inlining or no inlining.
1559   if (!ShouldInline && CA.getCost() < CA.getThreshold())
1560     return InlineCost::getNever();
1561   if (ShouldInline && CA.getCost() >= CA.getThreshold())
1562     return InlineCost::getAlways();
1563 
1564   return llvm::InlineCost::get(CA.getCost(), CA.getThreshold());
1565 }
1566 
1567 bool llvm::isInlineViable(Function &F) {
1568   bool ReturnsTwice = F.hasFnAttribute(Attribute::ReturnsTwice);
1569   for (Function::iterator BI = F.begin(), BE = F.end(); BI != BE; ++BI) {
1570     // Disallow inlining of functions which contain indirect branches or
1571     // blockaddresses.
1572     if (isa<IndirectBrInst>(BI->getTerminator()) || BI->hasAddressTaken())
1573       return false;
1574 
1575     for (auto &II : *BI) {
1576       CallSite CS(&II);
1577       if (!CS)
1578         continue;
1579 
1580       // Disallow recursive calls.
1581       if (&F == CS.getCalledFunction())
1582         return false;
1583 
1584       // Disallow calls which expose returns-twice to a function not previously
1585       // attributed as such.
1586       if (!ReturnsTwice && CS.isCall() &&
1587           cast<CallInst>(CS.getInstruction())->canReturnTwice())
1588         return false;
1589 
1590       // Disallow inlining functions that call @llvm.localescape. Doing this
1591       // correctly would require major changes to the inliner.
1592       if (CS.getCalledFunction() &&
1593           CS.getCalledFunction()->getIntrinsicID() ==
1594               llvm::Intrinsic::localescape)
1595         return false;
1596     }
1597   }
1598 
1599   return true;
1600 }
1601 
1602 // APIs to create InlineParams based on command line flags and/or other
1603 // parameters.
1604 
1605 InlineParams llvm::getInlineParams(int Threshold) {
1606   InlineParams Params;
1607 
1608   // This field is the threshold to use for a callee by default. This is
1609   // derived from one or more of:
1610   //  * optimization or size-optimization levels,
1611   //  * a value passed to createFunctionInliningPass function, or
1612   //  * the -inline-threshold flag.
1613   //  If the -inline-threshold flag is explicitly specified, that is used
1614   //  irrespective of anything else.
1615   if (InlineThreshold.getNumOccurrences() > 0)
1616     Params.DefaultThreshold = InlineThreshold;
1617   else
1618     Params.DefaultThreshold = Threshold;
1619 
1620   // Set the HintThreshold knob from the -inlinehint-threshold.
1621   Params.HintThreshold = HintThreshold;
1622 
1623   // Set the HotCallSiteThreshold knob from the -hot-callsite-threshold.
1624   Params.HotCallSiteThreshold = HotCallSiteThreshold;
1625 
1626   // Set the ColdCallSiteThreshold knob from the -inline-cold-callsite-threshold.
1627   Params.ColdCallSiteThreshold = ColdCallSiteThreshold;
1628 
1629   // Set the OptMinSizeThreshold and OptSizeThreshold params only if the
1630   // -inlinehint-threshold commandline option is not explicitly given. If that
1631   // option is present, then its value applies even for callees with size and
1632   // minsize attributes.
1633   // If the -inline-threshold is not specified, set the ColdThreshold from the
1634   // -inlinecold-threshold even if it is not explicitly passed. If
1635   // -inline-threshold is specified, then -inlinecold-threshold needs to be
1636   // explicitly specified to set the ColdThreshold knob
1637   if (InlineThreshold.getNumOccurrences() == 0) {
1638     Params.OptMinSizeThreshold = InlineConstants::OptMinSizeThreshold;
1639     Params.OptSizeThreshold = InlineConstants::OptSizeThreshold;
1640     Params.ColdThreshold = ColdThreshold;
1641   } else if (ColdThreshold.getNumOccurrences() > 0) {
1642     Params.ColdThreshold = ColdThreshold;
1643   }
1644   return Params;
1645 }
1646 
1647 InlineParams llvm::getInlineParams() {
1648   return getInlineParams(InlineThreshold);
1649 }
1650 
1651 // Compute the default threshold for inlining based on the opt level and the
1652 // size opt level.
1653 static int computeThresholdFromOptLevels(unsigned OptLevel,
1654                                          unsigned SizeOptLevel) {
1655   if (OptLevel > 2)
1656     return InlineConstants::OptAggressiveThreshold;
1657   if (SizeOptLevel == 1) // -Os
1658     return InlineConstants::OptSizeThreshold;
1659   if (SizeOptLevel == 2) // -Oz
1660     return InlineConstants::OptMinSizeThreshold;
1661   return InlineThreshold;
1662 }
1663 
1664 InlineParams llvm::getInlineParams(unsigned OptLevel, unsigned SizeOptLevel) {
1665   return getInlineParams(computeThresholdFromOptLevels(OptLevel, SizeOptLevel));
1666 }
1667