Start refactoring the inline cost estimation code so that it can be used

for purposes other than inlining.

llvm-svn: 83997
This commit is contained in:
Dan Gohman 2009-10-13 18:24:11 +00:00
parent 19788ca686
commit 5b3e05bcaa
2 changed files with 87 additions and 79 deletions

View File

@ -23,6 +23,7 @@ namespace llvm {
class Value; class Value;
class Function; class Function;
class BasicBlock;
class CallSite; class CallSite;
template<class PtrType, unsigned SmallSize> template<class PtrType, unsigned SmallSize>
class SmallPtrSet; class SmallPtrSet;
@ -96,9 +97,9 @@ namespace llvm {
: ConstantWeight(CWeight), AllocaWeight(AWeight) {} : ConstantWeight(CWeight), AllocaWeight(AWeight) {}
}; };
// FunctionInfo - For each function, calculate the size of it in blocks and // RegionInfo - Calculate size and a few related metrics for a set of
// instructions. // basic blocks.
struct FunctionInfo { struct RegionInfo {
/// NeverInline - True if this callee should never be inlined into a /// NeverInline - True if this callee should never be inlined into a
/// caller. /// caller.
bool NeverInline; bool NeverInline;
@ -115,17 +116,24 @@ namespace llvm {
/// kernels. /// kernels.
unsigned NumVectorInsts; unsigned NumVectorInsts;
/// NumRets - Keep track of how many Ret instructions the block contains.
unsigned NumRets;
/// ArgumentWeights - Each formal argument of the function is inspected to /// ArgumentWeights - Each formal argument of the function is inspected to
/// see if it is used in any contexts where making it a constant or alloca /// see if it is used in any contexts where making it a constant or alloca
/// would reduce the code size. If so, we add some value to the argument /// would reduce the code size. If so, we add some value to the argument
/// entry here. /// entry here.
std::vector<ArgInfo> ArgumentWeights; std::vector<ArgInfo> ArgumentWeights;
FunctionInfo() : NeverInline(false), usesDynamicAlloca(false), NumInsts(0), RegionInfo() : NeverInline(false), usesDynamicAlloca(false), NumInsts(0),
NumBlocks(0), NumVectorInsts(0) {} NumBlocks(0), NumVectorInsts(0), NumRets(0) {}
/// analyzeFunction - Fill in the current structure with information /// analyzeBasicBlock - Add information about the specified basic block
/// gleaned from the specified function. /// to the current structure.
void analyzeBasicBlock(const BasicBlock *BB);
/// analyzeFunction - Add information about the specified function
/// to the current structure.
void analyzeFunction(Function *F); void analyzeFunction(Function *F);
/// CountCodeReductionForConstant - Figure out an approximation for how /// CountCodeReductionForConstant - Figure out an approximation for how
@ -140,7 +148,7 @@ namespace llvm {
unsigned CountCodeReductionForAlloca(Value *V); unsigned CountCodeReductionForAlloca(Value *V);
}; };
std::map<const Function *, FunctionInfo> CachedFunctionInfo; std::map<const Function *, RegionInfo> CachedFunctionInfo;
public: public:

View File

@ -21,7 +21,7 @@ using namespace llvm;
// CountCodeReductionForConstant - Figure out an approximation for how many // CountCodeReductionForConstant - Figure out an approximation for how many
// instructions will be constant folded if the specified value is constant. // instructions will be constant folded if the specified value is constant.
// //
unsigned InlineCostAnalyzer::FunctionInfo:: unsigned InlineCostAnalyzer::RegionInfo::
CountCodeReductionForConstant(Value *V) { CountCodeReductionForConstant(Value *V) {
unsigned Reduction = 0; unsigned Reduction = 0;
for (Value::use_iterator UI = V->use_begin(), E = V->use_end(); UI != E; ++UI) for (Value::use_iterator UI = V->use_begin(), E = V->use_end(); UI != E; ++UI)
@ -77,7 +77,7 @@ unsigned InlineCostAnalyzer::FunctionInfo::
// the function will be if it is inlined into a context where an argument // the function will be if it is inlined into a context where an argument
// becomes an alloca. // becomes an alloca.
// //
unsigned InlineCostAnalyzer::FunctionInfo:: unsigned InlineCostAnalyzer::RegionInfo::
CountCodeReductionForAlloca(Value *V) { CountCodeReductionForAlloca(Value *V) {
if (!isa<PointerType>(V->getType())) return 0; // Not a pointer if (!isa<PointerType>(V->getType())) return 0; // Not a pointer
unsigned Reduction = 0; unsigned Reduction = 0;
@ -99,85 +99,85 @@ unsigned InlineCostAnalyzer::FunctionInfo::
return Reduction; return Reduction;
} }
/// analyzeFunction - Fill in the current structure with information gleaned /// analyzeBasicBlock - Fill in the current structure with information gleaned
/// from the specified function. /// from the specified block.
void InlineCostAnalyzer::FunctionInfo::analyzeFunction(Function *F) { void InlineCostAnalyzer::RegionInfo::analyzeBasicBlock(const BasicBlock *BB) {
unsigned NumInsts = 0, NumBlocks = 0, NumVectorInsts = 0, NumRets = 0; ++NumBlocks;
// Look at the size of the callee. Each basic block counts as 20 units, and for (BasicBlock::const_iterator II = BB->begin(), E = BB->end();
// each instruction counts as 5. II != E; ++II) {
for (Function::const_iterator BB = F->begin(), E = F->end(); BB != E; ++BB) { if (isa<PHINode>(II)) continue; // PHI nodes don't count.
for (BasicBlock::const_iterator II = BB->begin(), E = BB->end();
II != E; ++II) {
if (isa<PHINode>(II)) continue; // PHI nodes don't count.
// Special handling for calls. // Special handling for calls.
if (isa<CallInst>(II) || isa<InvokeInst>(II)) { if (isa<CallInst>(II) || isa<InvokeInst>(II)) {
if (isa<DbgInfoIntrinsic>(II)) if (isa<DbgInfoIntrinsic>(II))
continue; // Debug intrinsics don't count as size. continue; // Debug intrinsics don't count as size.
CallSite CS = CallSite::get(const_cast<Instruction*>(&*II));
// If this function contains a call to setjmp or _setjmp, never inline
// it. This is a hack because we depend on the user marking their local
// variables as volatile if they are live across a setjmp call, and they
// probably won't do this in callers.
if (Function *F = CS.getCalledFunction())
if (F->isDeclaration() &&
(F->getName() == "setjmp" || F->getName() == "_setjmp")) {
NeverInline = true;
return;
}
// Calls often compile into many machine instructions. Bump up their
// cost to reflect this.
if (!isa<IntrinsicInst>(II))
NumInsts += InlineConstants::CallPenalty;
}
// These, too, are calls. CallSite CS = CallSite::get(const_cast<Instruction*>(&*II));
if (isa<MallocInst>(II) || isa<FreeInst>(II))
NumInsts += InlineConstants::CallPenalty;
if (const AllocaInst *AI = dyn_cast<AllocaInst>(II)) {
if (!AI->isStaticAlloca())
this->usesDynamicAlloca = true;
}
if (isa<ExtractElementInst>(II) || isa<VectorType>(II->getType()))
++NumVectorInsts;
// Noop casts, including ptr <-> int, don't count. // If this function contains a call to setjmp or _setjmp, never inline
if (const CastInst *CI = dyn_cast<CastInst>(II)) { // it. This is a hack because we depend on the user marking their local
if (CI->isLosslessCast() || isa<IntToPtrInst>(CI) || // variables as volatile if they are live across a setjmp call, and they
isa<PtrToIntInst>(CI)) // probably won't do this in callers.
continue; if (Function *F = CS.getCalledFunction())
} else if (const GetElementPtrInst *GEPI = if (F->isDeclaration() &&
dyn_cast<GetElementPtrInst>(II)) { (F->getName() == "setjmp" || F->getName() == "_setjmp")) {
// If a GEP has all constant indices, it will probably be folded with NeverInline = true;
// a load/store. return;
if (GEPI->hasAllConstantIndices()) }
continue;
}
if (isa<ReturnInst>(II)) // Calls often compile into many machine instructions. Bump up their
++NumRets; // cost to reflect this.
if (!isa<IntrinsicInst>(II))
++NumInsts; NumInsts += InlineConstants::CallPenalty;
}
// These, too, are calls.
if (isa<MallocInst>(II) || isa<FreeInst>(II))
NumInsts += InlineConstants::CallPenalty;
if (const AllocaInst *AI = dyn_cast<AllocaInst>(II)) {
if (!AI->isStaticAlloca())
this->usesDynamicAlloca = true;
} }
++NumBlocks; if (isa<ExtractElementInst>(II) || isa<VectorType>(II->getType()))
++NumVectorInsts;
// Noop casts, including ptr <-> int, don't count.
if (const CastInst *CI = dyn_cast<CastInst>(II)) {
if (CI->isLosslessCast() || isa<IntToPtrInst>(CI) ||
isa<PtrToIntInst>(CI))
continue;
} else if (const GetElementPtrInst *GEPI =
dyn_cast<GetElementPtrInst>(II)) {
// If a GEP has all constant indices, it will probably be folded with
// a load/store.
if (GEPI->hasAllConstantIndices())
continue;
}
if (isa<ReturnInst>(II))
++NumRets;
++NumInsts;
} }
}
/// analyzeFunction - Fill in the current structure with information gleaned
/// from the specified function.
void InlineCostAnalyzer::RegionInfo::analyzeFunction(Function *F) {
// Look at the size of the callee. Each basic block counts as 20 units, and
// each instruction counts as 5.
for (Function::const_iterator BB = F->begin(), E = F->end(); BB != E; ++BB)
analyzeBasicBlock(&*BB);
// A function with exactly one return has it removed during the inlining // A function with exactly one return has it removed during the inlining
// process (see InlineFunction), so don't count it. // process (see InlineFunction), so don't count it.
// FIXME: This knowledge should really be encoded outside of RegionInfo.
if (NumRets==1) if (NumRets==1)
--NumInsts; --NumInsts;
this->NumBlocks = NumBlocks;
this->NumInsts = NumInsts;
this->NumVectorInsts = NumVectorInsts;
// Check out all of the arguments to the function, figuring out how much // Check out all of the arguments to the function, figuring out how much
// code can be eliminated if one of the arguments is a constant. // code can be eliminated if one of the arguments is a constant.
for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end(); I != E; ++I) for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end(); I != E; ++I)
@ -229,7 +229,7 @@ InlineCost InlineCostAnalyzer::getInlineCost(CallSite CS,
InlineCost += InlineConstants::NoreturnPenalty; InlineCost += InlineConstants::NoreturnPenalty;
// Get information about the callee... // Get information about the callee...
FunctionInfo &CalleeFI = CachedFunctionInfo[Callee]; RegionInfo &CalleeFI = CachedFunctionInfo[Callee];
// If we haven't calculated this information yet, do so now. // If we haven't calculated this information yet, do so now.
if (CalleeFI.NumBlocks == 0) if (CalleeFI.NumBlocks == 0)
@ -240,7 +240,7 @@ InlineCost InlineCostAnalyzer::getInlineCost(CallSite CS,
return InlineCost::getNever(); return InlineCost::getNever();
// FIXME: It would be nice to kill off CalleeFI.NeverInline. Then we // FIXME: It would be nice to kill off CalleeFI.NeverInline. Then we
// could move this up and avoid computing the FunctionInfo for // could move this up and avoid computing the RegionInfo for
// things we are going to just return always inline for. This // things we are going to just return always inline for. This
// requires handling setjmp somewhere else, however. // requires handling setjmp somewhere else, however.
if (!Callee->isDeclaration() && Callee->hasFnAttr(Attribute::AlwaysInline)) if (!Callee->isDeclaration() && Callee->hasFnAttr(Attribute::AlwaysInline))
@ -248,7 +248,7 @@ InlineCost InlineCostAnalyzer::getInlineCost(CallSite CS,
if (CalleeFI.usesDynamicAlloca) { if (CalleeFI.usesDynamicAlloca) {
// Get infomation about the caller... // Get infomation about the caller...
FunctionInfo &CallerFI = CachedFunctionInfo[Caller]; RegionInfo &CallerFI = CachedFunctionInfo[Caller];
// If we haven't calculated this information yet, do so now. // If we haven't calculated this information yet, do so now.
if (CallerFI.NumBlocks == 0) if (CallerFI.NumBlocks == 0)
@ -316,7 +316,7 @@ float InlineCostAnalyzer::getInlineFudgeFactor(CallSite CS) {
Function *Callee = CS.getCalledFunction(); Function *Callee = CS.getCalledFunction();
// Get information about the callee... // Get information about the callee...
FunctionInfo &CalleeFI = CachedFunctionInfo[Callee]; RegionInfo &CalleeFI = CachedFunctionInfo[Callee];
// If we haven't calculated this information yet, do so now. // If we haven't calculated this information yet, do so now.
if (CalleeFI.NumBlocks == 0) if (CalleeFI.NumBlocks == 0)