209 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			209 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C++
		
	
	
	
| //===- AMDGPUInline.cpp - Code to perform simple function inlining --------===//
 | |
| //
 | |
| //                     The LLVM Compiler Infrastructure
 | |
| //
 | |
| // This file is distributed under the University of Illinois Open Source
 | |
| // License. See LICENSE.TXT for details.
 | |
| //
 | |
| //===----------------------------------------------------------------------===//
 | |
| //
 | |
| /// \file
 | |
| /// This is AMDGPU specific replacement of the standard inliner.
 | |
| /// The main purpose is to account for the fact that calls not only expensive
 | |
| /// on the AMDGPU, but much more expensive if a private memory pointer is
 | |
| /// passed to a function as an argument. In this situation, we are unable to
 | |
| /// eliminate private memory in the caller unless inlined and end up with slow
 | |
| /// and expensive scratch access. Thus, we boost the inline threshold for such
 | |
| /// functions here.
 | |
| ///
 | |
| //===----------------------------------------------------------------------===//
 | |
| 
 | |
| 
 | |
| #include "AMDGPU.h"
 | |
| #include "llvm/Transforms/IPO.h"
 | |
| #include "llvm/Analysis/AssumptionCache.h"
 | |
| #include "llvm/Analysis/CallGraph.h"
 | |
| #include "llvm/Analysis/InlineCost.h"
 | |
| #include "llvm/Analysis/ValueTracking.h"
 | |
| #include "llvm/Analysis/TargetTransformInfo.h"
 | |
| #include "llvm/IR/CallSite.h"
 | |
| #include "llvm/IR/DataLayout.h"
 | |
| #include "llvm/IR/Instructions.h"
 | |
| #include "llvm/IR/Module.h"
 | |
| #include "llvm/IR/Type.h"
 | |
| #include "llvm/Support/CommandLine.h"
 | |
| #include "llvm/Support/Debug.h"
 | |
| #include "llvm/Transforms/IPO/Inliner.h"
 | |
| 
 | |
| using namespace llvm;
 | |
| 
 | |
| #define DEBUG_TYPE "inline"
 | |
| 
 | |
| static cl::opt<int>
 | |
| ArgAllocaCost("amdgpu-inline-arg-alloca-cost", cl::Hidden, cl::init(2200),
 | |
|               cl::desc("Cost of alloca argument"));
 | |
| 
 | |
| // If the amount of scratch memory to eliminate exceeds our ability to allocate
 | |
| // it into registers we gain nothing by agressively inlining functions for that
 | |
| // heuristic.
 | |
| static cl::opt<unsigned>
 | |
| ArgAllocaCutoff("amdgpu-inline-arg-alloca-cutoff", cl::Hidden, cl::init(256),
 | |
|                 cl::desc("Maximum alloca size to use for inline cost"));
 | |
| 
 | |
| namespace {
 | |
| 
 | |
| class AMDGPUInliner : public LegacyInlinerBase {
 | |
| 
 | |
| public:
 | |
|   AMDGPUInliner() : LegacyInlinerBase(ID) {
 | |
|     initializeAMDGPUInlinerPass(*PassRegistry::getPassRegistry());
 | |
|     Params = getInlineParams();
 | |
|   }
 | |
| 
 | |
|   static char ID; // Pass identification, replacement for typeid
 | |
| 
 | |
|   unsigned getInlineThreshold(CallSite CS) const;
 | |
| 
 | |
|   InlineCost getInlineCost(CallSite CS) override;
 | |
| 
 | |
|   bool runOnSCC(CallGraphSCC &SCC) override;
 | |
| 
 | |
|   void getAnalysisUsage(AnalysisUsage &AU) const override;
 | |
| 
 | |
| private:
 | |
|   TargetTransformInfoWrapperPass *TTIWP;
 | |
| 
 | |
|   InlineParams Params;
 | |
| };
 | |
| 
 | |
| } // end anonymous namespace
 | |
| 
 | |
| char AMDGPUInliner::ID = 0;
 | |
| INITIALIZE_PASS_BEGIN(AMDGPUInliner, "amdgpu-inline",
 | |
|                 "AMDGPU Function Integration/Inlining", false, false)
 | |
| INITIALIZE_PASS_DEPENDENCY(AssumptionCacheTracker)
 | |
| INITIALIZE_PASS_DEPENDENCY(CallGraphWrapperPass)
 | |
| INITIALIZE_PASS_DEPENDENCY(ProfileSummaryInfoWrapperPass)
 | |
| INITIALIZE_PASS_DEPENDENCY(TargetTransformInfoWrapperPass)
 | |
| INITIALIZE_PASS_DEPENDENCY(TargetLibraryInfoWrapperPass)
 | |
| INITIALIZE_PASS_END(AMDGPUInliner, "amdgpu-inline",
 | |
|                 "AMDGPU Function Integration/Inlining", false, false)
 | |
| 
 | |
| Pass *llvm::createAMDGPUFunctionInliningPass() { return new AMDGPUInliner(); }
 | |
| 
 | |
| bool AMDGPUInliner::runOnSCC(CallGraphSCC &SCC) {
 | |
|   TTIWP = &getAnalysis<TargetTransformInfoWrapperPass>();
 | |
|   return LegacyInlinerBase::runOnSCC(SCC);
 | |
| }
 | |
| 
 | |
| void AMDGPUInliner::getAnalysisUsage(AnalysisUsage &AU) const {
 | |
|   AU.addRequired<TargetTransformInfoWrapperPass>();
 | |
|   LegacyInlinerBase::getAnalysisUsage(AU);
 | |
| }
 | |
| 
 | |
| unsigned AMDGPUInliner::getInlineThreshold(CallSite CS) const {
 | |
|   int Thres = Params.DefaultThreshold;
 | |
| 
 | |
|   Function *Caller = CS.getCaller();
 | |
|   // Listen to the inlinehint attribute when it would increase the threshold
 | |
|   // and the caller does not need to minimize its size.
 | |
|   Function *Callee = CS.getCalledFunction();
 | |
|   bool InlineHint = Callee && !Callee->isDeclaration() &&
 | |
|     Callee->hasFnAttribute(Attribute::InlineHint);
 | |
|   if (InlineHint && Params.HintThreshold && Params.HintThreshold > Thres
 | |
|       && !Caller->hasFnAttribute(Attribute::MinSize))
 | |
|     Thres = Params.HintThreshold.getValue();
 | |
| 
 | |
|   const DataLayout &DL = Caller->getParent()->getDataLayout();
 | |
|   if (!Callee)
 | |
|     return (unsigned)Thres;
 | |
| 
 | |
|   const AMDGPUAS AS = AMDGPU::getAMDGPUAS(*Caller->getParent());
 | |
| 
 | |
|   // If we have a pointer to private array passed into a function
 | |
|   // it will not be optimized out, leaving scratch usage.
 | |
|   // Increase the inline threshold to allow inliniting in this case.
 | |
|   uint64_t AllocaSize = 0;
 | |
|   SmallPtrSet<const AllocaInst *, 8> AIVisited;
 | |
|   for (Value *PtrArg : CS.args()) {
 | |
|     Type *Ty = PtrArg->getType();
 | |
|     if (!Ty->isPointerTy() ||
 | |
|         Ty->getPointerAddressSpace() != AS.PRIVATE_ADDRESS)
 | |
|       continue;
 | |
|     PtrArg = GetUnderlyingObject(PtrArg, DL);
 | |
|     if (const AllocaInst *AI = dyn_cast<AllocaInst>(PtrArg)) {
 | |
|       if (!AI->isStaticAlloca() || !AIVisited.insert(AI).second)
 | |
|         continue;
 | |
|       AllocaSize += DL.getTypeAllocSize(AI->getAllocatedType());
 | |
|       // If the amount of stack memory is excessive we will not be able
 | |
|       // to get rid of the scratch anyway, bail out.
 | |
|       if (AllocaSize > ArgAllocaCutoff) {
 | |
|         AllocaSize = 0;
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
|   }
 | |
|   if (AllocaSize)
 | |
|     Thres += ArgAllocaCost;
 | |
| 
 | |
|   return (unsigned)Thres;
 | |
| }
 | |
| 
 | |
| // Check if call is just a wrapper around another call.
 | |
| // In this case we only have call and ret instructions.
 | |
| static bool isWrapperOnlyCall(CallSite CS) {
 | |
|   Function *Callee = CS.getCalledFunction();
 | |
|   if (!Callee || Callee->size() != 1)
 | |
|     return false;
 | |
|   const BasicBlock &BB = Callee->getEntryBlock();
 | |
|   if (const Instruction *I = BB.getFirstNonPHI()) {
 | |
|     if (!isa<CallInst>(I)) {
 | |
|       return false;
 | |
|     }
 | |
|     if (isa<ReturnInst>(*std::next(I->getIterator()))) {
 | |
|       LLVM_DEBUG(dbgs() << "    Wrapper only call detected: "
 | |
|                         << Callee->getName() << '\n');
 | |
|       return true;
 | |
|     }
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| InlineCost AMDGPUInliner::getInlineCost(CallSite CS) {
 | |
|   Function *Callee = CS.getCalledFunction();
 | |
|   Function *Caller = CS.getCaller();
 | |
|   TargetTransformInfo &TTI = TTIWP->getTTI(*Callee);
 | |
| 
 | |
|   if (!Callee || Callee->isDeclaration() || CS.isNoInline() ||
 | |
|       !TTI.areInlineCompatible(Caller, Callee))
 | |
|     return llvm::InlineCost::getNever();
 | |
| 
 | |
|   if (CS.hasFnAttr(Attribute::AlwaysInline)) {
 | |
|     if (isInlineViable(*Callee))
 | |
|       return llvm::InlineCost::getAlways();
 | |
|     return llvm::InlineCost::getNever();
 | |
|   }
 | |
| 
 | |
|   if (isWrapperOnlyCall(CS))
 | |
|     return llvm::InlineCost::getAlways();
 | |
| 
 | |
|   InlineParams LocalParams = Params;
 | |
|   LocalParams.DefaultThreshold = (int)getInlineThreshold(CS);
 | |
|   bool RemarksEnabled = false;
 | |
|   const auto &BBs = Caller->getBasicBlockList();
 | |
|   if (!BBs.empty()) {
 | |
|     auto DI = OptimizationRemark(DEBUG_TYPE, "", DebugLoc(), &BBs.front());
 | |
|     if (DI.isEnabled())
 | |
|       RemarksEnabled = true;
 | |
|   }
 | |
| 
 | |
|   OptimizationRemarkEmitter ORE(Caller);
 | |
|   std::function<AssumptionCache &(Function &)> GetAssumptionCache =
 | |
|       [this](Function &F) -> AssumptionCache & {
 | |
|     return ACT->getAssumptionCache(F);
 | |
|   };
 | |
| 
 | |
|   return llvm::getInlineCost(CS, Callee, LocalParams, TTI, GetAssumptionCache,
 | |
|                              None, PSI, RemarksEnabled ? &ORE : nullptr);
 | |
| }
 |