2004-02-28 11:26:20 +08:00
|
|
|
//===- CodeExtractor.cpp - Pull code region into a new function -----------===//
|
2005-04-22 07:48:37 +08:00
|
|
|
//
|
2019-01-19 16:50:56 +08:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
2005-04-22 07:48:37 +08:00
|
|
|
//
|
2004-02-28 11:26:20 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file implements the interface to tear out a code region, such as an
|
|
|
|
// individual loop or a parallel section, into a new function, replacing it with
|
|
|
|
// a call to the new function.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2012-05-04 18:18:49 +08:00
|
|
|
#include "llvm/Transforms/Utils/CodeExtractor.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/ADT/ArrayRef.h"
|
|
|
|
#include "llvm/ADT/DenseMap.h"
|
|
|
|
#include "llvm/ADT/Optional.h"
|
2013-02-09 09:04:28 +08:00
|
|
|
#include "llvm/ADT/STLExtras.h"
|
2014-01-07 19:48:04 +08:00
|
|
|
#include "llvm/ADT/SetVector.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/ADT/SmallPtrSet.h"
|
|
|
|
#include "llvm/ADT/SmallVector.h"
|
2019-02-08 14:55:18 +08:00
|
|
|
#include "llvm/Analysis/AssumptionCache.h"
|
2016-08-02 10:15:45 +08:00
|
|
|
#include "llvm/Analysis/BlockFrequencyInfo.h"
|
|
|
|
#include "llvm/Analysis/BlockFrequencyInfoImpl.h"
|
|
|
|
#include "llvm/Analysis/BranchProbabilityInfo.h"
|
2012-12-04 00:50:05 +08:00
|
|
|
#include "llvm/Analysis/LoopInfo.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/IR/Argument.h"
|
|
|
|
#include "llvm/IR/Attributes.h"
|
|
|
|
#include "llvm/IR/BasicBlock.h"
|
|
|
|
#include "llvm/IR/CFG.h"
|
|
|
|
#include "llvm/IR/Constant.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/Constants.h"
|
2020-01-16 03:26:34 +08:00
|
|
|
#include "llvm/IR/DIBuilder.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/IR/DataLayout.h"
|
2020-01-16 03:26:34 +08:00
|
|
|
#include "llvm/IR/DebugInfoMetadata.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/DerivedTypes.h"
|
2014-01-13 17:26:24 +08:00
|
|
|
#include "llvm/IR/Dominators.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/IR/Function.h"
|
|
|
|
#include "llvm/IR/GlobalValue.h"
|
2020-01-16 03:26:34 +08:00
|
|
|
#include "llvm/IR/InstIterator.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/IR/InstrTypes.h"
|
|
|
|
#include "llvm/IR/Instruction.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/Instructions.h"
|
2017-05-31 05:22:18 +08:00
|
|
|
#include "llvm/IR/IntrinsicInst.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/Intrinsics.h"
|
|
|
|
#include "llvm/IR/LLVMContext.h"
|
2016-08-02 10:15:45 +08:00
|
|
|
#include "llvm/IR/MDBuilder.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/Module.h"
|
2019-02-08 14:55:18 +08:00
|
|
|
#include "llvm/IR/PatternMatch.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/IR/Type.h"
|
|
|
|
#include "llvm/IR/User.h"
|
|
|
|
#include "llvm/IR/Value.h"
|
2014-01-13 17:26:24 +08:00
|
|
|
#include "llvm/IR/Verifier.h"
|
2004-02-28 11:26:20 +08:00
|
|
|
#include "llvm/Pass.h"
|
2016-08-02 10:15:45 +08:00
|
|
|
#include "llvm/Support/BlockFrequency.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include "llvm/Support/BranchProbability.h"
|
|
|
|
#include "llvm/Support/Casting.h"
|
2004-09-02 06:55:40 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
|
|
|
#include "llvm/Support/Debug.h"
|
2009-07-11 21:10:19 +08:00
|
|
|
#include "llvm/Support/ErrorHandling.h"
|
2009-08-23 12:37:46 +08:00
|
|
|
#include "llvm/Support/raw_ostream.h"
|
2012-12-04 00:50:05 +08:00
|
|
|
#include "llvm/Transforms/Utils/BasicBlockUtils.h"
|
2018-11-07 03:05:53 +08:00
|
|
|
#include "llvm/Transforms/Utils/Local.h"
|
2017-10-12 05:41:43 +08:00
|
|
|
#include <cassert>
|
|
|
|
#include <cstdint>
|
|
|
|
#include <iterator>
|
|
|
|
#include <map>
|
2004-03-15 06:34:55 +08:00
|
|
|
#include <set>
|
2017-10-12 05:41:43 +08:00
|
|
|
#include <utility>
|
|
|
|
#include <vector>
|
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
using namespace llvm;
|
2019-02-08 14:55:18 +08:00
|
|
|
using namespace llvm::PatternMatch;
|
2018-01-18 06:24:23 +08:00
|
|
|
using ProfileCount = Function::ProfileCount;
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2014-04-22 06:55:11 +08:00
|
|
|
#define DEBUG_TYPE "code-extractor"
|
|
|
|
|
2004-04-24 07:54:17 +08:00
|
|
|
// Provide a command-line option to aggregate function arguments into a struct
|
2008-12-13 13:21:37 +08:00
|
|
|
// for functions produced by the code extractor. This is useful when converting
|
2004-04-24 07:54:17 +08:00
|
|
|
// extracted functions to pthread-based code, as only one argument (void*) can
|
|
|
|
// be passed in to pthread_create().
|
|
|
|
static cl::opt<bool>
|
|
|
|
AggregateArgsOpt("aggregate-extracted-args", cl::Hidden,
|
|
|
|
cl::desc("Aggregate arguments to code-extracted functions"));
|
|
|
|
|
2018-05-01 23:54:18 +08:00
|
|
|
/// Test whether a block is valid for extraction.
|
2018-05-12 06:49:49 +08:00
|
|
|
static bool isBlockValidForExtraction(const BasicBlock &BB,
|
|
|
|
const SetVector<BasicBlock *> &Result,
|
|
|
|
bool AllowVarArgs, bool AllowAlloca) {
|
2017-06-28 02:57:53 +08:00
|
|
|
// taking the address of a basic block moved to another function is illegal
|
|
|
|
if (BB.hasAddressTaken())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// don't hoist code that uses another basicblock address, as it's likely to
|
|
|
|
// lead to unexpected behavior, like cross-function jumps
|
|
|
|
SmallPtrSet<User const *, 16> Visited;
|
|
|
|
SmallVector<User const *, 16> ToVisit;
|
|
|
|
|
|
|
|
for (Instruction const &Inst : BB)
|
|
|
|
ToVisit.push_back(&Inst);
|
|
|
|
|
|
|
|
while (!ToVisit.empty()) {
|
|
|
|
User const *Curr = ToVisit.pop_back_val();
|
|
|
|
if (!Visited.insert(Curr).second)
|
|
|
|
continue;
|
|
|
|
if (isa<BlockAddress const>(Curr))
|
|
|
|
return false; // even a reference to self is likely to be not compatible
|
|
|
|
|
|
|
|
if (isa<Instruction>(Curr) && cast<Instruction>(Curr)->getParent() != &BB)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
for (auto const &U : Curr->operands()) {
|
|
|
|
if (auto *UU = dyn_cast<User>(U))
|
|
|
|
ToVisit.push_back(UU);
|
|
|
|
}
|
|
|
|
}
|
2005-04-22 07:48:37 +08:00
|
|
|
|
2018-05-12 06:49:49 +08:00
|
|
|
// If explicitly requested, allow vastart and alloca. For invoke instructions
|
|
|
|
// verify that extraction is valid.
|
2012-05-04 18:18:49 +08:00
|
|
|
for (BasicBlock::const_iterator I = BB.begin(), E = BB.end(); I != E; ++I) {
|
2018-05-12 06:49:49 +08:00
|
|
|
if (isa<AllocaInst>(I)) {
|
|
|
|
if (!AllowAlloca)
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (const auto *II = dyn_cast<InvokeInst>(I)) {
|
|
|
|
// Unwind destination (either a landingpad, catchswitch, or cleanuppad)
|
|
|
|
// must be a part of the subgraph which is being extracted.
|
|
|
|
if (auto *UBB = II->getUnwindDest())
|
|
|
|
if (!Result.count(UBB))
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// All catch handlers of a catchswitch instruction as well as the unwind
|
|
|
|
// destination must be in the subgraph.
|
|
|
|
if (const auto *CSI = dyn_cast<CatchSwitchInst>(I)) {
|
|
|
|
if (auto *UBB = CSI->getUnwindDest())
|
|
|
|
if (!Result.count(UBB))
|
|
|
|
return false;
|
|
|
|
for (auto *HBB : CSI->handlers())
|
|
|
|
if (!Result.count(const_cast<BasicBlock*>(HBB)))
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure that entire catch handler is within subgraph. It is sufficient
|
|
|
|
// to check that catch return's block is in the list.
|
|
|
|
if (const auto *CPI = dyn_cast<CatchPadInst>(I)) {
|
|
|
|
for (const auto *U : CPI->users())
|
|
|
|
if (const auto *CRI = dyn_cast<CatchReturnInst>(U))
|
|
|
|
if (!Result.count(const_cast<BasicBlock*>(CRI->getParent())))
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// And do similar checks for cleanup handler - the entire handler must be
|
|
|
|
// in subgraph which is going to be extracted. For cleanup return should
|
|
|
|
// additionally check that the unwind destination is also in the subgraph.
|
|
|
|
if (const auto *CPI = dyn_cast<CleanupPadInst>(I)) {
|
|
|
|
for (const auto *U : CPI->users())
|
|
|
|
if (const auto *CRI = dyn_cast<CleanupReturnInst>(U))
|
|
|
|
if (!Result.count(const_cast<BasicBlock*>(CRI->getParent())))
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (const auto *CRI = dyn_cast<CleanupReturnInst>(I)) {
|
|
|
|
if (auto *UBB = CRI->getUnwindDest())
|
|
|
|
if (!Result.count(UBB))
|
|
|
|
return false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2018-11-07 03:06:08 +08:00
|
|
|
if (const CallInst *CI = dyn_cast<CallInst>(I)) {
|
|
|
|
if (const Function *F = CI->getCalledFunction()) {
|
|
|
|
auto IID = F->getIntrinsicID();
|
|
|
|
if (IID == Intrinsic::vastart) {
|
2017-11-13 18:35:52 +08:00
|
|
|
if (AllowVarArgs)
|
|
|
|
continue;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
}
|
2018-11-07 03:06:08 +08:00
|
|
|
|
|
|
|
// Currently, we miscompile outlined copies of eh_typid_for. There are
|
|
|
|
// proposals for fixing this in llvm.org/PR39545.
|
|
|
|
if (IID == Intrinsic::eh_typeid_for)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
2012-05-04 18:18:49 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-05-01 23:54:18 +08:00
|
|
|
/// Build a set of blocks to extract if the input blocks are viable.
|
2017-04-21 08:21:09 +08:00
|
|
|
static SetVector<BasicBlock *>
|
2017-11-13 18:35:52 +08:00
|
|
|
buildExtractionBlockSet(ArrayRef<BasicBlock *> BBs, DominatorTree *DT,
|
2018-05-12 06:49:49 +08:00
|
|
|
bool AllowVarArgs, bool AllowAlloca) {
|
2017-04-21 12:25:00 +08:00
|
|
|
assert(!BBs.empty() && "The set of blocks to extract must be non-empty");
|
2017-04-21 08:21:09 +08:00
|
|
|
SetVector<BasicBlock *> Result;
|
2012-05-04 18:26:45 +08:00
|
|
|
|
2012-05-04 18:18:49 +08:00
|
|
|
// Loop over the blocks, adding them to our set-vector, and aborting with an
|
|
|
|
// empty set if we encounter invalid blocks.
|
2017-04-21 12:25:00 +08:00
|
|
|
for (BasicBlock *BB : BBs) {
|
|
|
|
// If this block is dead, don't process it.
|
|
|
|
if (DT && !DT->isReachableFromEntry(BB))
|
|
|
|
continue;
|
2012-05-04 18:18:49 +08:00
|
|
|
|
2017-04-21 12:25:00 +08:00
|
|
|
if (!Result.insert(BB))
|
|
|
|
llvm_unreachable("Repeated basic blocks in extraction input");
|
|
|
|
}
|
2004-05-12 14:01:40 +08:00
|
|
|
|
2019-04-16 10:12:05 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "Region front block: " << Result.front()->getName()
|
|
|
|
<< '\n');
|
|
|
|
|
2018-05-12 06:49:49 +08:00
|
|
|
for (auto *BB : Result) {
|
|
|
|
if (!isBlockValidForExtraction(*BB, Result, AllowVarArgs, AllowAlloca))
|
|
|
|
return {};
|
|
|
|
|
|
|
|
// Make sure that the first block is not a landing pad.
|
|
|
|
if (BB == Result.front()) {
|
|
|
|
if (BB->isEHPad()) {
|
2018-05-14 20:53:11 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "The first block cannot be an unwind block\n");
|
2018-05-12 06:49:49 +08:00
|
|
|
return {};
|
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// All blocks other than the first must not have predecessors outside of
|
|
|
|
// the subgraph which is being extracted.
|
|
|
|
for (auto *PBB : predecessors(BB))
|
|
|
|
if (!Result.count(PBB)) {
|
2019-04-16 10:12:05 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "No blocks in this region may have entries from "
|
|
|
|
"outside the region except for the first block!\n"
|
|
|
|
<< "Problematic source BB: " << BB->getName() << "\n"
|
|
|
|
<< "Problematic destination BB: " << PBB->getName()
|
|
|
|
<< "\n");
|
2018-05-12 06:49:49 +08:00
|
|
|
return {};
|
|
|
|
}
|
|
|
|
}
|
2012-05-04 18:26:45 +08:00
|
|
|
|
2012-05-04 18:18:49 +08:00
|
|
|
return Result;
|
|
|
|
}
|
|
|
|
|
|
|
|
CodeExtractor::CodeExtractor(ArrayRef<BasicBlock *> BBs, DominatorTree *DT,
|
2016-08-02 10:15:45 +08:00
|
|
|
bool AggregateArgs, BlockFrequencyInfo *BFI,
|
2019-02-08 14:55:18 +08:00
|
|
|
BranchProbabilityInfo *BPI, AssumptionCache *AC,
|
|
|
|
bool AllowVarArgs, bool AllowAlloca,
|
|
|
|
std::string Suffix)
|
2016-08-02 10:15:45 +08:00
|
|
|
: DT(DT), AggregateArgs(AggregateArgs || AggregateArgsOpt), BFI(BFI),
|
2019-02-08 14:55:18 +08:00
|
|
|
BPI(BPI), AC(AC), AllowVarArgs(AllowVarArgs),
|
2018-10-25 02:53:47 +08:00
|
|
|
Blocks(buildExtractionBlockSet(BBs, DT, AllowVarArgs, AllowAlloca)),
|
|
|
|
Suffix(Suffix) {}
|
2016-08-02 10:15:45 +08:00
|
|
|
|
|
|
|
CodeExtractor::CodeExtractor(DominatorTree &DT, Loop &L, bool AggregateArgs,
|
|
|
|
BlockFrequencyInfo *BFI,
|
2019-02-08 14:55:18 +08:00
|
|
|
BranchProbabilityInfo *BPI, AssumptionCache *AC,
|
|
|
|
std::string Suffix)
|
2016-08-02 10:15:45 +08:00
|
|
|
: DT(&DT), AggregateArgs(AggregateArgs || AggregateArgsOpt), BFI(BFI),
|
2019-02-08 14:55:18 +08:00
|
|
|
BPI(BPI), AC(AC), AllowVarArgs(false),
|
2017-11-13 19:08:47 +08:00
|
|
|
Blocks(buildExtractionBlockSet(L.getBlocks(), &DT,
|
2018-05-12 06:49:49 +08:00
|
|
|
/* AllowVarArgs */ false,
|
2018-10-25 02:53:47 +08:00
|
|
|
/* AllowAlloca */ false)),
|
|
|
|
Suffix(Suffix) {}
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2012-05-04 18:18:49 +08:00
|
|
|
/// definedInRegion - Return true if the specified value is defined in the
|
|
|
|
/// extracted region.
|
|
|
|
static bool definedInRegion(const SetVector<BasicBlock *> &Blocks, Value *V) {
|
|
|
|
if (Instruction *I = dyn_cast<Instruction>(V))
|
|
|
|
if (Blocks.count(I->getParent()))
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// definedInCaller - Return true if the specified value is defined in the
|
|
|
|
/// function being code extracted, but not in the region being extracted.
|
|
|
|
/// These values must be passed in as live-ins to the function.
|
|
|
|
static bool definedInCaller(const SetVector<BasicBlock *> &Blocks, Value *V) {
|
|
|
|
if (isa<Argument>(V)) return true;
|
|
|
|
if (Instruction *I = dyn_cast<Instruction>(V))
|
|
|
|
if (!Blocks.count(I->getParent()))
|
|
|
|
return true;
|
|
|
|
return false;
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
|
|
|
|
2017-06-12 04:46:05 +08:00
|
|
|
static BasicBlock *getCommonExitBlock(const SetVector<BasicBlock *> &Blocks) {
|
|
|
|
BasicBlock *CommonExitBlock = nullptr;
|
|
|
|
auto hasNonCommonExitSucc = [&](BasicBlock *Block) {
|
|
|
|
for (auto *Succ : successors(Block)) {
|
|
|
|
// Internal edges, ok.
|
|
|
|
if (Blocks.count(Succ))
|
|
|
|
continue;
|
|
|
|
if (!CommonExitBlock) {
|
|
|
|
CommonExitBlock = Succ;
|
|
|
|
continue;
|
|
|
|
}
|
2019-10-02 20:15:17 +08:00
|
|
|
if (CommonExitBlock != Succ)
|
|
|
|
return true;
|
2017-06-12 04:46:05 +08:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
};
|
|
|
|
|
|
|
|
if (any_of(Blocks, hasNonCommonExitSucc))
|
|
|
|
return nullptr;
|
|
|
|
|
|
|
|
return CommonExitBlock;
|
|
|
|
}
|
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
CodeExtractorAnalysisCache::CodeExtractorAnalysisCache(Function &F) {
|
|
|
|
for (BasicBlock &BB : F) {
|
|
|
|
for (Instruction &II : BB.instructionsWithoutDebug())
|
|
|
|
if (auto *AI = dyn_cast<AllocaInst>(&II))
|
|
|
|
Allocas.push_back(AI);
|
2017-06-12 04:46:05 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
findSideEffectInfoForBlock(BB);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void CodeExtractorAnalysisCache::findSideEffectInfoForBlock(BasicBlock &BB) {
|
|
|
|
for (Instruction &II : BB.instructionsWithoutDebug()) {
|
|
|
|
unsigned Opcode = II.getOpcode();
|
|
|
|
Value *MemAddr = nullptr;
|
|
|
|
switch (Opcode) {
|
|
|
|
case Instruction::Store:
|
|
|
|
case Instruction::Load: {
|
|
|
|
if (Opcode == Instruction::Store) {
|
|
|
|
StoreInst *SI = cast<StoreInst>(&II);
|
|
|
|
MemAddr = SI->getPointerOperand();
|
|
|
|
} else {
|
|
|
|
LoadInst *LI = cast<LoadInst>(&II);
|
|
|
|
MemAddr = LI->getPointerOperand();
|
|
|
|
}
|
|
|
|
// Global variable can not be aliased with locals.
|
|
|
|
if (dyn_cast<Constant>(MemAddr))
|
2017-06-12 04:46:05 +08:00
|
|
|
break;
|
2019-10-09 01:17:51 +08:00
|
|
|
Value *Base = MemAddr->stripInBoundsConstantOffsets();
|
|
|
|
if (!isa<AllocaInst>(Base)) {
|
|
|
|
SideEffectingBlocks.insert(&BB);
|
|
|
|
return;
|
2017-06-12 04:46:05 +08:00
|
|
|
}
|
2019-10-09 01:17:51 +08:00
|
|
|
BaseMemAddrs[&BB].insert(Base);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default: {
|
|
|
|
IntrinsicInst *IntrInst = dyn_cast<IntrinsicInst>(&II);
|
|
|
|
if (IntrInst) {
|
|
|
|
if (IntrInst->isLifetimeStartOrEnd())
|
|
|
|
break;
|
|
|
|
SideEffectingBlocks.insert(&BB);
|
|
|
|
return;
|
2017-06-12 04:46:05 +08:00
|
|
|
}
|
2019-10-09 01:17:51 +08:00
|
|
|
// Treat all the other cases conservatively if it has side effects.
|
|
|
|
if (II.mayHaveSideEffects()) {
|
|
|
|
SideEffectingBlocks.insert(&BB);
|
|
|
|
return;
|
2017-06-12 04:46:05 +08:00
|
|
|
}
|
|
|
|
}
|
2019-10-09 01:17:51 +08:00
|
|
|
}
|
2017-06-12 04:46:05 +08:00
|
|
|
}
|
2019-10-09 01:17:51 +08:00
|
|
|
}
|
2017-06-12 04:46:05 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
bool CodeExtractorAnalysisCache::doesBlockContainClobberOfAddr(
|
|
|
|
BasicBlock &BB, AllocaInst *Addr) const {
|
|
|
|
if (SideEffectingBlocks.count(&BB))
|
|
|
|
return true;
|
|
|
|
auto It = BaseMemAddrs.find(&BB);
|
|
|
|
if (It != BaseMemAddrs.end())
|
|
|
|
return It->second.count(Addr);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool CodeExtractor::isLegalToShrinkwrapLifetimeMarkers(
|
|
|
|
const CodeExtractorAnalysisCache &CEAC, Instruction *Addr) const {
|
|
|
|
AllocaInst *AI = cast<AllocaInst>(Addr->stripInBoundsConstantOffsets());
|
|
|
|
Function *Func = (*Blocks.begin())->getParent();
|
|
|
|
for (BasicBlock &BB : *Func) {
|
|
|
|
if (Blocks.count(&BB))
|
|
|
|
continue;
|
|
|
|
if (CEAC.doesBlockContainClobberOfAddr(BB, AI))
|
|
|
|
return false;
|
|
|
|
}
|
2017-06-12 04:46:05 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
BasicBlock *
|
|
|
|
CodeExtractor::findOrCreateBlockForHoisting(BasicBlock *CommonExitBlock) {
|
|
|
|
BasicBlock *SinglePredFromOutlineRegion = nullptr;
|
|
|
|
assert(!Blocks.count(CommonExitBlock) &&
|
|
|
|
"Expect a block outside the region!");
|
|
|
|
for (auto *Pred : predecessors(CommonExitBlock)) {
|
|
|
|
if (!Blocks.count(Pred))
|
|
|
|
continue;
|
|
|
|
if (!SinglePredFromOutlineRegion) {
|
|
|
|
SinglePredFromOutlineRegion = Pred;
|
|
|
|
} else if (SinglePredFromOutlineRegion != Pred) {
|
|
|
|
SinglePredFromOutlineRegion = nullptr;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (SinglePredFromOutlineRegion)
|
|
|
|
return SinglePredFromOutlineRegion;
|
|
|
|
|
|
|
|
#ifndef NDEBUG
|
|
|
|
auto getFirstPHI = [](BasicBlock *BB) {
|
|
|
|
BasicBlock::iterator I = BB->begin();
|
|
|
|
PHINode *FirstPhi = nullptr;
|
|
|
|
while (I != BB->end()) {
|
|
|
|
PHINode *Phi = dyn_cast<PHINode>(I);
|
|
|
|
if (!Phi)
|
|
|
|
break;
|
|
|
|
if (!FirstPhi) {
|
|
|
|
FirstPhi = Phi;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return FirstPhi;
|
|
|
|
};
|
|
|
|
// If there are any phi nodes, the single pred either exists or has already
|
|
|
|
// be created before code extraction.
|
|
|
|
assert(!getFirstPHI(CommonExitBlock) && "Phi not expected");
|
|
|
|
#endif
|
|
|
|
|
|
|
|
BasicBlock *NewExitBlock = CommonExitBlock->splitBasicBlock(
|
|
|
|
CommonExitBlock->getFirstNonPHI()->getIterator());
|
|
|
|
|
2017-11-01 17:48:12 +08:00
|
|
|
for (auto PI = pred_begin(CommonExitBlock), PE = pred_end(CommonExitBlock);
|
|
|
|
PI != PE;) {
|
|
|
|
BasicBlock *Pred = *PI++;
|
2017-06-12 04:46:05 +08:00
|
|
|
if (Blocks.count(Pred))
|
|
|
|
continue;
|
|
|
|
Pred->getTerminator()->replaceUsesOfWith(CommonExitBlock, NewExitBlock);
|
|
|
|
}
|
|
|
|
// Now add the old exit block to the outline region.
|
|
|
|
Blocks.insert(CommonExitBlock);
|
|
|
|
return CommonExitBlock;
|
|
|
|
}
|
|
|
|
|
2019-07-11 00:32:16 +08:00
|
|
|
// Find the pair of life time markers for address 'Addr' that are either
|
|
|
|
// defined inside the outline region or can legally be shrinkwrapped into the
|
|
|
|
// outline region. If there are not other untracked uses of the address, return
|
|
|
|
// the pair of markers if found; otherwise return a pair of nullptr.
|
|
|
|
CodeExtractor::LifetimeMarkerInfo
|
2019-10-09 01:17:51 +08:00
|
|
|
CodeExtractor::getLifetimeMarkers(const CodeExtractorAnalysisCache &CEAC,
|
|
|
|
Instruction *Addr,
|
2019-07-11 00:32:16 +08:00
|
|
|
BasicBlock *ExitBlock) const {
|
|
|
|
LifetimeMarkerInfo Info;
|
|
|
|
|
|
|
|
for (User *U : Addr->users()) {
|
|
|
|
IntrinsicInst *IntrInst = dyn_cast<IntrinsicInst>(U);
|
|
|
|
if (IntrInst) {
|
|
|
|
if (IntrInst->getIntrinsicID() == Intrinsic::lifetime_start) {
|
|
|
|
// Do not handle the case where Addr has multiple start markers.
|
|
|
|
if (Info.LifeStart)
|
|
|
|
return {};
|
|
|
|
Info.LifeStart = IntrInst;
|
|
|
|
}
|
|
|
|
if (IntrInst->getIntrinsicID() == Intrinsic::lifetime_end) {
|
|
|
|
if (Info.LifeEnd)
|
|
|
|
return {};
|
|
|
|
Info.LifeEnd = IntrInst;
|
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
// Find untracked uses of the address, bail.
|
|
|
|
if (!definedInRegion(Blocks, U))
|
|
|
|
return {};
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Info.LifeStart || !Info.LifeEnd)
|
|
|
|
return {};
|
|
|
|
|
|
|
|
Info.SinkLifeStart = !definedInRegion(Blocks, Info.LifeStart);
|
|
|
|
Info.HoistLifeEnd = !definedInRegion(Blocks, Info.LifeEnd);
|
|
|
|
// Do legality check.
|
|
|
|
if ((Info.SinkLifeStart || Info.HoistLifeEnd) &&
|
2019-10-09 01:17:51 +08:00
|
|
|
!isLegalToShrinkwrapLifetimeMarkers(CEAC, Addr))
|
2019-07-11 00:32:16 +08:00
|
|
|
return {};
|
|
|
|
|
|
|
|
// Check to see if we have a place to do hoisting, if not, bail.
|
|
|
|
if (Info.HoistLifeEnd && !ExitBlock)
|
|
|
|
return {};
|
|
|
|
|
|
|
|
return Info;
|
|
|
|
}
|
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
void CodeExtractor::findAllocas(const CodeExtractorAnalysisCache &CEAC,
|
|
|
|
ValueSet &SinkCands, ValueSet &HoistCands,
|
2017-06-12 04:46:05 +08:00
|
|
|
BasicBlock *&ExitBlock) const {
|
2017-05-31 05:22:18 +08:00
|
|
|
Function *Func = (*Blocks.begin())->getParent();
|
2017-06-12 04:46:05 +08:00
|
|
|
ExitBlock = getCommonExitBlock(Blocks);
|
|
|
|
|
2019-07-11 00:32:16 +08:00
|
|
|
auto moveOrIgnoreLifetimeMarkers =
|
|
|
|
[&](const LifetimeMarkerInfo &LMI) -> bool {
|
|
|
|
if (!LMI.LifeStart)
|
|
|
|
return false;
|
|
|
|
if (LMI.SinkLifeStart) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Sinking lifetime.start: " << *LMI.LifeStart
|
|
|
|
<< "\n");
|
|
|
|
SinkCands.insert(LMI.LifeStart);
|
|
|
|
}
|
|
|
|
if (LMI.HoistLifeEnd) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Hoisting lifetime.end: " << *LMI.LifeEnd << "\n");
|
|
|
|
HoistCands.insert(LMI.LifeEnd);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
};
|
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
// Look up allocas in the original function in CodeExtractorAnalysisCache, as
|
|
|
|
// this is much faster than walking all the instructions.
|
|
|
|
for (AllocaInst *AI : CEAC.getAllocas()) {
|
|
|
|
BasicBlock *BB = AI->getParent();
|
|
|
|
if (Blocks.count(BB))
|
2017-05-31 05:22:18 +08:00
|
|
|
continue;
|
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
// As a prior call to extractCodeRegion() may have shrinkwrapped the alloca,
|
|
|
|
// check whether it is actually still in the original function.
|
|
|
|
Function *AIFunc = BB->getParent();
|
|
|
|
if (AIFunc != Func)
|
|
|
|
continue;
|
2017-05-31 05:22:18 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
LifetimeMarkerInfo MarkerInfo = getLifetimeMarkers(CEAC, AI, ExitBlock);
|
|
|
|
bool Moved = moveOrIgnoreLifetimeMarkers(MarkerInfo);
|
|
|
|
if (Moved) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Sinking alloca: " << *AI << "\n");
|
|
|
|
SinkCands.insert(AI);
|
|
|
|
continue;
|
|
|
|
}
|
2017-06-12 04:46:05 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
// Follow any bitcasts.
|
|
|
|
SmallVector<Instruction *, 2> Bitcasts;
|
|
|
|
SmallVector<LifetimeMarkerInfo, 2> BitcastLifetimeInfo;
|
|
|
|
for (User *U : AI->users()) {
|
|
|
|
if (U->stripInBoundsConstantOffsets() == AI) {
|
|
|
|
Instruction *Bitcast = cast<Instruction>(U);
|
|
|
|
LifetimeMarkerInfo LMI = getLifetimeMarkers(CEAC, Bitcast, ExitBlock);
|
|
|
|
if (LMI.LifeStart) {
|
|
|
|
Bitcasts.push_back(Bitcast);
|
|
|
|
BitcastLifetimeInfo.push_back(LMI);
|
|
|
|
continue;
|
2017-05-31 05:22:18 +08:00
|
|
|
}
|
|
|
|
}
|
2017-06-12 04:46:05 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
// Found unknown use of AI.
|
|
|
|
if (!definedInRegion(Blocks, U)) {
|
|
|
|
Bitcasts.clear();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2019-07-11 00:32:16 +08:00
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
// Either no bitcasts reference the alloca or there are unknown uses.
|
|
|
|
if (Bitcasts.empty())
|
|
|
|
continue;
|
|
|
|
|
|
|
|
LLVM_DEBUG(dbgs() << "Sinking alloca (via bitcast): " << *AI << "\n");
|
|
|
|
SinkCands.insert(AI);
|
|
|
|
for (unsigned I = 0, E = Bitcasts.size(); I != E; ++I) {
|
|
|
|
Instruction *BitcastAddr = Bitcasts[I];
|
|
|
|
const LifetimeMarkerInfo &LMI = BitcastLifetimeInfo[I];
|
|
|
|
assert(LMI.LifeStart &&
|
|
|
|
"Unsafe to sink bitcast without lifetime markers");
|
|
|
|
moveOrIgnoreLifetimeMarkers(LMI);
|
|
|
|
if (!definedInRegion(Blocks, BitcastAddr)) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Sinking bitcast-of-alloca: " << *BitcastAddr
|
|
|
|
<< "\n");
|
|
|
|
SinkCands.insert(BitcastAddr);
|
2019-07-11 00:32:20 +08:00
|
|
|
}
|
2017-05-31 05:22:18 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-10-02 23:36:39 +08:00
|
|
|
bool CodeExtractor::isEligible() const {
|
|
|
|
if (Blocks.empty())
|
|
|
|
return false;
|
|
|
|
BasicBlock *Header = *Blocks.begin();
|
|
|
|
Function *F = Header->getParent();
|
|
|
|
|
|
|
|
// For functions with varargs, check that varargs handling is only done in the
|
|
|
|
// outlined function, i.e vastart and vaend are only used in outlined blocks.
|
|
|
|
if (AllowVarArgs && F->getFunctionType()->isVarArg()) {
|
|
|
|
auto containsVarArgIntrinsic = [](const Instruction &I) {
|
|
|
|
if (const CallInst *CI = dyn_cast<CallInst>(&I))
|
|
|
|
if (const Function *Callee = CI->getCalledFunction())
|
|
|
|
return Callee->getIntrinsicID() == Intrinsic::vastart ||
|
|
|
|
Callee->getIntrinsicID() == Intrinsic::vaend;
|
|
|
|
return false;
|
|
|
|
};
|
|
|
|
|
|
|
|
for (auto &BB : *F) {
|
|
|
|
if (Blocks.count(&BB))
|
|
|
|
continue;
|
|
|
|
if (llvm::any_of(BB, containsVarArgIntrinsic))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2017-05-31 05:22:18 +08:00
|
|
|
void CodeExtractor::findInputsOutputs(ValueSet &Inputs, ValueSet &Outputs,
|
|
|
|
const ValueSet &SinkCands) const {
|
2016-06-26 20:28:59 +08:00
|
|
|
for (BasicBlock *BB : Blocks) {
|
2012-05-04 19:20:27 +08:00
|
|
|
// If a used value is defined outside the region, it's an input. If an
|
|
|
|
// instruction is used outside the region, it's an output.
|
2016-06-26 20:28:59 +08:00
|
|
|
for (Instruction &II : *BB) {
|
2019-10-02 23:36:39 +08:00
|
|
|
for (auto &OI : II.operands()) {
|
|
|
|
Value *V = OI;
|
2017-05-31 05:22:18 +08:00
|
|
|
if (!SinkCands.count(V) && definedInCaller(Blocks, V))
|
|
|
|
Inputs.insert(V);
|
|
|
|
}
|
2012-05-04 19:20:27 +08:00
|
|
|
|
2016-06-26 20:28:59 +08:00
|
|
|
for (User *U : II.users())
|
2014-03-09 11:16:01 +08:00
|
|
|
if (!definedInRegion(Blocks, U)) {
|
2016-06-26 20:28:59 +08:00
|
|
|
Outputs.insert(&II);
|
2012-05-04 19:20:27 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-12-04 06:40:21 +08:00
|
|
|
/// severSplitPHINodesOfEntry - If a PHI node has multiple inputs from outside
|
|
|
|
/// of the region, we need to split the entry block of the region so that the
|
|
|
|
/// PHI node is easier to deal with.
|
|
|
|
void CodeExtractor::severSplitPHINodesOfEntry(BasicBlock *&Header) {
|
2011-03-30 19:19:20 +08:00
|
|
|
unsigned NumPredsFromRegion = 0;
|
2004-05-12 23:29:13 +08:00
|
|
|
unsigned NumPredsOutsideRegion = 0;
|
|
|
|
|
2007-03-23 00:38:57 +08:00
|
|
|
if (Header != &Header->getParent()->getEntryBlock()) {
|
2004-05-12 23:29:13 +08:00
|
|
|
PHINode *PN = dyn_cast<PHINode>(Header->begin());
|
|
|
|
if (!PN) return; // No PHI nodes.
|
|
|
|
|
|
|
|
// If the header node contains any PHI nodes, check to see if there is more
|
|
|
|
// than one entry from outside the region. If so, we need to sever the
|
|
|
|
// header block into two.
|
|
|
|
for (unsigned i = 0, e = PN->getNumIncomingValues(); i != e; ++i)
|
2012-05-04 18:18:49 +08:00
|
|
|
if (Blocks.count(PN->getIncomingBlock(i)))
|
2011-03-30 19:19:20 +08:00
|
|
|
++NumPredsFromRegion;
|
2004-05-12 23:29:13 +08:00
|
|
|
else
|
|
|
|
++NumPredsOutsideRegion;
|
|
|
|
|
|
|
|
// If there is one (or fewer) predecessor from outside the region, we don't
|
|
|
|
// need to do anything special.
|
|
|
|
if (NumPredsOutsideRegion <= 1) return;
|
|
|
|
}
|
2004-05-12 14:01:40 +08:00
|
|
|
|
2004-05-12 23:29:13 +08:00
|
|
|
// Otherwise, we need to split the header block into two pieces: one
|
|
|
|
// containing PHI nodes merging values from outside of the region, and a
|
|
|
|
// second that contains all of the code for the block and merges back any
|
|
|
|
// incoming values from inside of the region.
|
2017-10-12 05:41:43 +08:00
|
|
|
BasicBlock *NewBB = SplitBlock(Header, Header->getFirstNonPHI(), DT);
|
2004-05-12 23:29:13 +08:00
|
|
|
|
|
|
|
// We only want to code extract the second block now, and it becomes the new
|
|
|
|
// header of the region.
|
|
|
|
BasicBlock *OldPred = Header;
|
2012-05-04 18:18:49 +08:00
|
|
|
Blocks.remove(OldPred);
|
|
|
|
Blocks.insert(NewBB);
|
2004-05-12 23:29:13 +08:00
|
|
|
Header = NewBB;
|
|
|
|
|
|
|
|
// Okay, now we need to adjust the PHI nodes and any branches from within the
|
|
|
|
// region to go to the new header block instead of the old header block.
|
2011-03-30 19:19:20 +08:00
|
|
|
if (NumPredsFromRegion) {
|
2004-05-12 23:29:13 +08:00
|
|
|
PHINode *PN = cast<PHINode>(OldPred->begin());
|
|
|
|
// Loop over all of the predecessors of OldPred that are in the region,
|
|
|
|
// changing them to branch to NewBB instead.
|
|
|
|
for (unsigned i = 0, e = PN->getNumIncomingValues(); i != e; ++i)
|
2012-05-04 18:18:49 +08:00
|
|
|
if (Blocks.count(PN->getIncomingBlock(i))) {
|
2018-10-15 18:04:59 +08:00
|
|
|
Instruction *TI = PN->getIncomingBlock(i)->getTerminator();
|
2004-05-12 23:29:13 +08:00
|
|
|
TI->replaceUsesOfWith(OldPred, NewBB);
|
|
|
|
}
|
|
|
|
|
2011-04-15 13:18:47 +08:00
|
|
|
// Okay, everything within the region is now branching to the right block, we
|
2004-05-12 23:29:13 +08:00
|
|
|
// just have to update the PHI nodes now, inserting PHI nodes into NewBB.
|
2017-04-21 05:40:22 +08:00
|
|
|
BasicBlock::iterator AfterPHIs;
|
2004-09-16 01:06:42 +08:00
|
|
|
for (AfterPHIs = OldPred->begin(); isa<PHINode>(AfterPHIs); ++AfterPHIs) {
|
|
|
|
PHINode *PN = cast<PHINode>(AfterPHIs);
|
2004-05-12 23:29:13 +08:00
|
|
|
// Create a new PHI node in the new region, which has an incoming value
|
|
|
|
// from OldPred of PN.
|
2011-03-30 19:28:46 +08:00
|
|
|
PHINode *NewPN = PHINode::Create(PN->getType(), 1 + NumPredsFromRegion,
|
2015-10-13 10:39:05 +08:00
|
|
|
PN->getName() + ".ce", &NewBB->front());
|
2017-04-25 12:51:19 +08:00
|
|
|
PN->replaceAllUsesWith(NewPN);
|
2004-05-12 23:29:13 +08:00
|
|
|
NewPN->addIncoming(PN, OldPred);
|
|
|
|
|
|
|
|
// Loop over all of the incoming value in PN, moving them to NewPN if they
|
|
|
|
// are from the extracted region.
|
|
|
|
for (unsigned i = 0; i != PN->getNumIncomingValues(); ++i) {
|
2012-05-04 18:18:49 +08:00
|
|
|
if (Blocks.count(PN->getIncomingBlock(i))) {
|
2004-05-12 23:29:13 +08:00
|
|
|
NewPN->addIncoming(PN->getIncomingValue(i), PN->getIncomingBlock(i));
|
|
|
|
PN->removeIncomingValue(i);
|
|
|
|
--i;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2004-05-13 00:07:41 +08:00
|
|
|
}
|
2004-05-12 23:29:13 +08:00
|
|
|
|
2018-12-04 06:40:21 +08:00
|
|
|
/// severSplitPHINodesOfExits - if PHI nodes in exit blocks have inputs from
|
|
|
|
/// outlined region, we split these PHIs on two: one with inputs from region
|
|
|
|
/// and other with remaining incoming blocks; then first PHIs are placed in
|
|
|
|
/// outlined region.
|
|
|
|
void CodeExtractor::severSplitPHINodesOfExits(
|
|
|
|
const SmallPtrSetImpl<BasicBlock *> &Exits) {
|
|
|
|
for (BasicBlock *ExitBB : Exits) {
|
|
|
|
BasicBlock *NewBB = nullptr;
|
|
|
|
|
|
|
|
for (PHINode &PN : ExitBB->phis()) {
|
|
|
|
// Find all incoming values from the outlining region.
|
|
|
|
SmallVector<unsigned, 2> IncomingVals;
|
|
|
|
for (unsigned i = 0; i < PN.getNumIncomingValues(); ++i)
|
|
|
|
if (Blocks.count(PN.getIncomingBlock(i)))
|
|
|
|
IncomingVals.push_back(i);
|
|
|
|
|
|
|
|
// Do not process PHI if there is one (or fewer) predecessor from region.
|
|
|
|
// If PHI has exactly one predecessor from region, only this one incoming
|
|
|
|
// will be replaced on codeRepl block, so it should be safe to skip PHI.
|
|
|
|
if (IncomingVals.size() <= 1)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Create block for new PHIs and add it to the list of outlined if it
|
|
|
|
// wasn't done before.
|
|
|
|
if (!NewBB) {
|
|
|
|
NewBB = BasicBlock::Create(ExitBB->getContext(),
|
|
|
|
ExitBB->getName() + ".split",
|
|
|
|
ExitBB->getParent(), ExitBB);
|
|
|
|
SmallVector<BasicBlock *, 4> Preds(pred_begin(ExitBB),
|
|
|
|
pred_end(ExitBB));
|
|
|
|
for (BasicBlock *PredBB : Preds)
|
|
|
|
if (Blocks.count(PredBB))
|
|
|
|
PredBB->getTerminator()->replaceUsesOfWith(ExitBB, NewBB);
|
|
|
|
BranchInst::Create(ExitBB, NewBB);
|
|
|
|
Blocks.insert(NewBB);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Split this PHI.
|
|
|
|
PHINode *NewPN =
|
|
|
|
PHINode::Create(PN.getType(), IncomingVals.size(),
|
|
|
|
PN.getName() + ".ce", NewBB->getFirstNonPHI());
|
|
|
|
for (unsigned i : IncomingVals)
|
|
|
|
NewPN->addIncoming(PN.getIncomingValue(i), PN.getIncomingBlock(i));
|
|
|
|
for (unsigned i : reverse(IncomingVals))
|
|
|
|
PN.removeIncomingValue(i, false);
|
|
|
|
PN.addIncoming(NewPN, NewBB);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-05-13 00:07:41 +08:00
|
|
|
void CodeExtractor::splitReturnBlocks() {
|
2016-06-26 20:28:59 +08:00
|
|
|
for (BasicBlock *Block : Blocks)
|
|
|
|
if (ReturnInst *RI = dyn_cast<ReturnInst>(Block->getTerminator())) {
|
2015-10-13 10:39:05 +08:00
|
|
|
BasicBlock *New =
|
2016-06-26 20:28:59 +08:00
|
|
|
Block->splitBasicBlock(RI->getIterator(), Block->getName() + ".ret");
|
2009-08-25 07:32:14 +08:00
|
|
|
if (DT) {
|
2010-09-11 06:25:58 +08:00
|
|
|
// Old dominates New. New node dominates all other nodes dominated
|
|
|
|
// by Old.
|
2016-06-26 20:28:59 +08:00
|
|
|
DomTreeNode *OldNode = DT->getNode(Block);
|
|
|
|
SmallVector<DomTreeNode *, 8> Children(OldNode->begin(),
|
|
|
|
OldNode->end());
|
2009-08-25 07:32:14 +08:00
|
|
|
|
2016-06-26 20:28:59 +08:00
|
|
|
DomTreeNode *NewNode = DT->addNewBlock(New, Block);
|
2009-08-25 07:32:14 +08:00
|
|
|
|
2016-06-26 20:28:59 +08:00
|
|
|
for (DomTreeNode *I : Children)
|
|
|
|
DT->changeImmediateDominator(I, NewNode);
|
2009-08-25 07:32:14 +08:00
|
|
|
}
|
|
|
|
}
|
2004-05-12 14:01:40 +08:00
|
|
|
}
|
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
/// constructFunction - make a function based on inputs and outputs, as follows:
|
|
|
|
/// f(in0, ..., inN, out0, ..., outN)
|
2012-05-04 18:18:49 +08:00
|
|
|
Function *CodeExtractor::constructFunction(const ValueSet &inputs,
|
|
|
|
const ValueSet &outputs,
|
2004-03-18 13:28:49 +08:00
|
|
|
BasicBlock *header,
|
2004-02-28 11:26:20 +08:00
|
|
|
BasicBlock *newRootNode,
|
|
|
|
BasicBlock *newHeader,
|
2004-03-18 13:28:49 +08:00
|
|
|
Function *oldFunction,
|
|
|
|
Module *M) {
|
2018-05-14 20:53:11 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "inputs: " << inputs.size() << "\n");
|
|
|
|
LLVM_DEBUG(dbgs() << "outputs: " << outputs.size() << "\n");
|
2004-02-28 11:26:20 +08:00
|
|
|
|
|
|
|
// This function returns unsigned, outputs will go back by reference.
|
2004-05-12 12:14:24 +08:00
|
|
|
switch (NumExitBlocks) {
|
|
|
|
case 0:
|
2009-08-14 05:58:54 +08:00
|
|
|
case 1: RetTy = Type::getVoidTy(header->getContext()); break;
|
|
|
|
case 2: RetTy = Type::getInt1Ty(header->getContext()); break;
|
|
|
|
default: RetTy = Type::getInt16Ty(header->getContext()); break;
|
2004-05-12 12:14:24 +08:00
|
|
|
}
|
|
|
|
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<Type *> paramTy;
|
2004-02-28 11:26:20 +08:00
|
|
|
|
|
|
|
// Add the types of the input values to the function's argument list
|
2016-06-26 20:28:59 +08:00
|
|
|
for (Value *value : inputs) {
|
2018-05-14 20:53:11 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "value used in func: " << *value << "\n");
|
2004-02-28 11:26:20 +08:00
|
|
|
paramTy.push_back(value->getType());
|
|
|
|
}
|
|
|
|
|
2004-03-18 11:49:40 +08:00
|
|
|
// Add the types of the output values to the function's argument list.
|
2016-06-26 20:28:59 +08:00
|
|
|
for (Value *output : outputs) {
|
2018-05-14 20:53:11 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "instr used in func: " << *output << "\n");
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs)
|
2016-06-26 20:28:59 +08:00
|
|
|
paramTy.push_back(output->getType());
|
2004-04-24 07:54:17 +08:00
|
|
|
else
|
2016-06-26 20:28:59 +08:00
|
|
|
paramTy.push_back(PointerType::getUnqual(output->getType()));
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
|
|
|
|
2018-05-14 20:53:11 +08:00
|
|
|
LLVM_DEBUG({
|
2016-06-26 21:39:33 +08:00
|
|
|
dbgs() << "Function type: " << *RetTy << " f(";
|
|
|
|
for (Type *i : paramTy)
|
|
|
|
dbgs() << *i << ", ";
|
|
|
|
dbgs() << ")\n";
|
|
|
|
});
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2020-01-11 22:44:25 +08:00
|
|
|
StructType *StructTy = nullptr;
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs && (inputs.size() + outputs.size() > 0)) {
|
2015-03-14 09:53:18 +08:00
|
|
|
StructTy = StructType::get(M->getContext(), paramTy);
|
2004-04-24 07:54:17 +08:00
|
|
|
paramTy.clear();
|
2015-03-14 09:53:18 +08:00
|
|
|
paramTy.push_back(PointerType::getUnqual(StructTy));
|
2004-04-24 07:54:17 +08:00
|
|
|
}
|
2011-07-18 12:54:35 +08:00
|
|
|
FunctionType *funcType =
|
2017-11-13 18:35:52 +08:00
|
|
|
FunctionType::get(RetTy, paramTy,
|
|
|
|
AllowVarArgs && oldFunction->isVarArg());
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2018-10-25 02:53:47 +08:00
|
|
|
std::string SuffixToUse =
|
|
|
|
Suffix.empty()
|
|
|
|
? (header->getName().empty() ? "extracted" : header->getName().str())
|
|
|
|
: Suffix;
|
2004-02-28 11:26:20 +08:00
|
|
|
// Create the new function
|
2018-08-23 17:25:17 +08:00
|
|
|
Function *newFunction = Function::Create(
|
|
|
|
funcType, GlobalValue::InternalLinkage, oldFunction->getAddressSpace(),
|
2018-10-25 02:53:47 +08:00
|
|
|
oldFunction->getName() + "." + SuffixToUse, M);
|
2008-12-18 13:52:56 +08:00
|
|
|
// If the old function is no-throw, so is the new one.
|
|
|
|
if (oldFunction->doesNotThrow())
|
2012-10-10 11:12:49 +08:00
|
|
|
newFunction->setDoesNotThrow();
|
2016-08-01 11:15:32 +08:00
|
|
|
|
|
|
|
// Inherit the uwtable attribute if we need to.
|
|
|
|
if (oldFunction->hasUWTable())
|
|
|
|
newFunction->setHasUWTable();
|
|
|
|
|
2018-01-07 19:22:25 +08:00
|
|
|
// Inherit all of the target dependent attributes and white-listed
|
|
|
|
// target independent attributes.
|
2016-08-01 11:15:32 +08:00
|
|
|
// (e.g. If the extracted region contains a call to an x86.sse
|
|
|
|
// instruction we need to make sure that the extracted region has the
|
|
|
|
// "target-features" attribute allowing it to be lowered.
|
|
|
|
// FIXME: This should be changed to check to see if a specific
|
|
|
|
// attribute can not be inherited.
|
2018-01-07 19:22:25 +08:00
|
|
|
for (const auto &Attr : oldFunction->getAttributes().getFnAttributes()) {
|
|
|
|
if (Attr.isStringAttribute()) {
|
|
|
|
if (Attr.getKindAsString() == "thunk")
|
|
|
|
continue;
|
|
|
|
} else
|
|
|
|
switch (Attr.getKindAsEnum()) {
|
|
|
|
// Those attributes cannot be propagated safely. Explicitly list them
|
|
|
|
// here so we get a warning if new attributes are added. This list also
|
|
|
|
// includes non-function attributes.
|
|
|
|
case Attribute::Alignment:
|
|
|
|
case Attribute::AllocSize:
|
|
|
|
case Attribute::ArgMemOnly:
|
|
|
|
case Attribute::Builtin:
|
|
|
|
case Attribute::ByVal:
|
|
|
|
case Attribute::Convergent:
|
|
|
|
case Attribute::Dereferenceable:
|
|
|
|
case Attribute::DereferenceableOrNull:
|
|
|
|
case Attribute::InAlloca:
|
|
|
|
case Attribute::InReg:
|
|
|
|
case Attribute::InaccessibleMemOnly:
|
|
|
|
case Attribute::InaccessibleMemOrArgMemOnly:
|
|
|
|
case Attribute::JumpTable:
|
|
|
|
case Attribute::Naked:
|
|
|
|
case Attribute::Nest:
|
|
|
|
case Attribute::NoAlias:
|
|
|
|
case Attribute::NoBuiltin:
|
|
|
|
case Attribute::NoCapture:
|
|
|
|
case Attribute::NoReturn:
|
[Attributor] Deduce "nosync" function attribute.
Introduce and deduce "nosync" function attribute to indicate that a function
does not synchronize with another thread in a way that other thread might free memory.
Reviewers: jdoerfert, jfb, nhaehnle, arsenm
Subscribers: wdng, hfinkel, nhaenhle, mehdi_amini, steven_wu,
dexonsmith, arsenm, uenoku, hiraditya, jfb, llvm-commits
Differential Revision: https://reviews.llvm.org/D62766
llvm-svn: 365830
2019-07-12 05:37:40 +08:00
|
|
|
case Attribute::NoSync:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::None:
|
|
|
|
case Attribute::NonNull:
|
|
|
|
case Attribute::ReadNone:
|
|
|
|
case Attribute::ReadOnly:
|
|
|
|
case Attribute::Returned:
|
|
|
|
case Attribute::ReturnsTwice:
|
|
|
|
case Attribute::SExt:
|
|
|
|
case Attribute::Speculatable:
|
|
|
|
case Attribute::StackAlignment:
|
|
|
|
case Attribute::StructRet:
|
|
|
|
case Attribute::SwiftError:
|
|
|
|
case Attribute::SwiftSelf:
|
2019-06-27 23:51:40 +08:00
|
|
|
case Attribute::WillReturn:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::WriteOnly:
|
|
|
|
case Attribute::ZExt:
|
2019-03-13 05:02:54 +08:00
|
|
|
case Attribute::ImmArg:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::EndAttrKinds:
|
2020-03-04 00:42:57 +08:00
|
|
|
case Attribute::EmptyKey:
|
|
|
|
case Attribute::TombstoneKey:
|
2018-01-07 19:22:25 +08:00
|
|
|
continue;
|
|
|
|
// Those attributes should be safe to propagate to the extracted function.
|
|
|
|
case Attribute::AlwaysInline:
|
|
|
|
case Attribute::Cold:
|
|
|
|
case Attribute::NoRecurse:
|
|
|
|
case Attribute::InlineHint:
|
|
|
|
case Attribute::MinSize:
|
|
|
|
case Attribute::NoDuplicate:
|
2019-07-08 23:57:56 +08:00
|
|
|
case Attribute::NoFree:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::NoImplicitFloat:
|
|
|
|
case Attribute::NoInline:
|
|
|
|
case Attribute::NonLazyBind:
|
|
|
|
case Attribute::NoRedZone:
|
|
|
|
case Attribute::NoUnwind:
|
[SimplifyCFG] Create attribute for fuzzing-specific optimizations.
Summary:
When building with libFuzzer, converting control flow to selects or
obscuring the original operands of CMPs reduces the effectiveness of
libFuzzer's heuristics.
This patch provides an attribute to disable or modify certain optimizations
for optimal fuzzing signal.
Provides a less aggressive alternative to https://reviews.llvm.org/D44057.
Reviewers: vitalybuka, davide, arsenm, hfinkel
Reviewed By: vitalybuka
Subscribers: junbuml, mehdi_amini, wdng, javed.absar, hiraditya, llvm-commits, kcc
Differential Revision: https://reviews.llvm.org/D44232
llvm-svn: 328214
2018-03-23 01:07:51 +08:00
|
|
|
case Attribute::OptForFuzzing:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::OptimizeNone:
|
|
|
|
case Attribute::OptimizeForSize:
|
|
|
|
case Attribute::SafeStack:
|
2018-04-04 04:10:40 +08:00
|
|
|
case Attribute::ShadowCallStack:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::SanitizeAddress:
|
|
|
|
case Attribute::SanitizeMemory:
|
|
|
|
case Attribute::SanitizeThread:
|
|
|
|
case Attribute::SanitizeHWAddress:
|
ARM MTE stack sanitizer.
Add "memtag" sanitizer that detects and mitigates stack memory issues
using armv8.5 Memory Tagging Extension.
It is similar in principle to HWASan, which is a software implementation
of the same idea, but there are enough differencies to warrant a new
sanitizer type IMHO. It is also expected to have very different
performance properties.
The new sanitizer does not have a runtime library (it may grow one
later, along with a "debugging" mode). Similar to SafeStack and
StackProtector, the instrumentation pass (in a follow up change) will be
inserted in all cases, but will only affect functions marked with the
new sanitize_memtag attribute.
Reviewers: pcc, hctim, vitalybuka, ostannard
Subscribers: srhines, mehdi_amini, javed.absar, kristof.beyls, hiraditya, cryptoad, steven_wu, dexonsmith, cfe-commits, llvm-commits
Tags: #clang, #llvm
Differential Revision: https://reviews.llvm.org/D64169
llvm-svn: 366123
2019-07-16 04:02:23 +08:00
|
|
|
case Attribute::SanitizeMemTag:
|
2018-09-04 20:38:00 +08:00
|
|
|
case Attribute::SpeculativeLoadHardening:
|
2018-01-07 19:22:25 +08:00
|
|
|
case Attribute::StackProtect:
|
|
|
|
case Attribute::StackProtectReq:
|
|
|
|
case Attribute::StackProtectStrong:
|
|
|
|
case Attribute::StrictFP:
|
|
|
|
case Attribute::UWTable:
|
2018-03-17 21:29:46 +08:00
|
|
|
case Attribute::NoCfCheck:
|
2018-01-07 19:22:25 +08:00
|
|
|
break;
|
|
|
|
}
|
2016-08-01 11:15:32 +08:00
|
|
|
|
2018-01-07 19:22:25 +08:00
|
|
|
newFunction->addFnAttr(Attr);
|
|
|
|
}
|
2004-02-28 11:26:20 +08:00
|
|
|
newFunction->getBasicBlockList().push_back(newRootNode);
|
|
|
|
|
2004-03-18 11:49:40 +08:00
|
|
|
// Create an iterator to name all of the arguments we inserted.
|
2005-03-15 12:54:21 +08:00
|
|
|
Function::arg_iterator AI = newFunction->arg_begin();
|
2004-03-18 11:49:40 +08:00
|
|
|
|
|
|
|
// Rewrite all users of the inputs in the extracted region to use the
|
2004-04-24 07:54:17 +08:00
|
|
|
// arguments (or appropriate addressing into struct) instead.
|
|
|
|
for (unsigned i = 0, e = inputs.size(); i != e; ++i) {
|
|
|
|
Value *RewriteVal;
|
|
|
|
if (AggregateArgs) {
|
2007-09-04 23:46:09 +08:00
|
|
|
Value *Idx[2];
|
2009-08-14 05:58:54 +08:00
|
|
|
Idx[0] = Constant::getNullValue(Type::getInt32Ty(header->getContext()));
|
|
|
|
Idx[1] = ConstantInt::get(Type::getInt32Ty(header->getContext()), i);
|
2018-10-15 18:04:59 +08:00
|
|
|
Instruction *TI = newFunction->begin()->getTerminator();
|
2015-03-14 09:53:18 +08:00
|
|
|
GetElementPtrInst *GEP = GetElementPtrInst::Create(
|
2015-10-13 10:39:05 +08:00
|
|
|
StructTy, &*AI, Idx, "gep_" + inputs[i]->getName(), TI);
|
2019-02-02 04:44:24 +08:00
|
|
|
RewriteVal = new LoadInst(StructTy->getElementType(i), GEP,
|
|
|
|
"loadgep_" + inputs[i]->getName(), TI);
|
2004-04-24 07:54:17 +08:00
|
|
|
} else
|
2015-10-13 10:39:05 +08:00
|
|
|
RewriteVal = &*AI++;
|
2004-04-24 07:54:17 +08:00
|
|
|
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<User *> Users(inputs[i]->user_begin(), inputs[i]->user_end());
|
2016-06-26 20:28:59 +08:00
|
|
|
for (User *use : Users)
|
|
|
|
if (Instruction *inst = dyn_cast<Instruction>(use))
|
2012-05-04 18:18:49 +08:00
|
|
|
if (Blocks.count(inst->getParent()))
|
2004-04-24 07:54:17 +08:00
|
|
|
inst->replaceUsesOfWith(inputs[i], RewriteVal);
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
|
|
|
|
2004-04-24 07:54:17 +08:00
|
|
|
// Set names for input and output arguments.
|
|
|
|
if (!AggregateArgs) {
|
2005-03-15 12:54:21 +08:00
|
|
|
AI = newFunction->arg_begin();
|
2004-04-24 07:54:17 +08:00
|
|
|
for (unsigned i = 0, e = inputs.size(); i != e; ++i, ++AI)
|
2008-04-15 01:38:21 +08:00
|
|
|
AI->setName(inputs[i]->getName());
|
2004-04-24 07:54:17 +08:00
|
|
|
for (unsigned i = 0, e = outputs.size(); i != e; ++i, ++AI)
|
2005-04-22 07:48:37 +08:00
|
|
|
AI->setName(outputs[i]->getName()+".out");
|
2004-04-24 07:54:17 +08:00
|
|
|
}
|
2004-03-18 11:49:40 +08:00
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
// Rewrite branches to basic blocks outside of the loop to new dummy blocks
|
|
|
|
// within the new function. This must be done before we lose track of which
|
|
|
|
// blocks were originally in the code region.
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<User *> Users(header->user_begin(), header->user_end());
|
2019-10-16 09:50:21 +08:00
|
|
|
for (auto &U : Users)
|
2004-03-18 13:28:49 +08:00
|
|
|
// The BasicBlock which contains the branch is not in the region
|
|
|
|
// modify the branch target to a new block
|
2019-10-16 09:50:21 +08:00
|
|
|
if (Instruction *I = dyn_cast<Instruction>(U))
|
|
|
|
if (I->isTerminator() && I->getFunction() == oldFunction &&
|
|
|
|
!Blocks.count(I->getParent()))
|
2018-10-18 08:38:54 +08:00
|
|
|
I->replaceUsesOfWith(header, newHeader);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
|
|
|
return newFunction;
|
|
|
|
}
|
|
|
|
|
2019-02-14 03:53:38 +08:00
|
|
|
/// Erase lifetime.start markers which reference inputs to the extraction
|
2019-02-16 02:46:58 +08:00
|
|
|
/// region, and insert the referenced memory into \p LifetimesStart.
|
2019-01-19 10:37:59 +08:00
|
|
|
///
|
|
|
|
/// The extraction region is defined by a set of blocks (\p Blocks), and a set
|
|
|
|
/// of allocas which will be moved from the caller function into the extracted
|
|
|
|
/// function (\p SunkAllocas).
|
2019-02-14 03:53:38 +08:00
|
|
|
static void eraseLifetimeMarkersOnInputs(const SetVector<BasicBlock *> &Blocks,
|
|
|
|
const SetVector<Value *> &SunkAllocas,
|
2019-02-16 02:46:58 +08:00
|
|
|
SetVector<Value *> &LifetimesStart) {
|
2019-01-19 10:37:59 +08:00
|
|
|
for (BasicBlock *BB : Blocks) {
|
|
|
|
for (auto It = BB->begin(), End = BB->end(); It != End;) {
|
|
|
|
auto *II = dyn_cast<IntrinsicInst>(&*It);
|
|
|
|
++It;
|
|
|
|
if (!II || !II->isLifetimeStartOrEnd())
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Get the memory operand of the lifetime marker. If the underlying
|
|
|
|
// object is a sunk alloca, or is otherwise defined in the extraction
|
|
|
|
// region, the lifetime marker must not be erased.
|
|
|
|
Value *Mem = II->getOperand(1)->stripInBoundsOffsets();
|
|
|
|
if (SunkAllocas.count(Mem) || definedInRegion(Blocks, Mem))
|
|
|
|
continue;
|
|
|
|
|
2019-02-14 03:53:38 +08:00
|
|
|
if (II->getIntrinsicID() == Intrinsic::lifetime_start)
|
|
|
|
LifetimesStart.insert(Mem);
|
2019-01-19 10:37:59 +08:00
|
|
|
II->eraseFromParent();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Insert lifetime start/end markers surrounding the call to the new function
|
|
|
|
/// for objects defined in the caller.
|
2019-02-14 03:53:38 +08:00
|
|
|
static void insertLifetimeMarkersSurroundingCall(
|
|
|
|
Module *M, ArrayRef<Value *> LifetimesStart, ArrayRef<Value *> LifetimesEnd,
|
|
|
|
CallInst *TheCall) {
|
2019-01-19 10:37:59 +08:00
|
|
|
LLVMContext &Ctx = M->getContext();
|
|
|
|
auto Int8PtrTy = Type::getInt8PtrTy(Ctx);
|
|
|
|
auto NegativeOne = ConstantInt::getSigned(Type::getInt64Ty(Ctx), -1);
|
|
|
|
Instruction *Term = TheCall->getParent()->getTerminator();
|
|
|
|
|
2019-02-14 03:53:38 +08:00
|
|
|
// The memory argument to a lifetime marker must be a i8*. Cache any bitcasts
|
|
|
|
// needed to satisfy this requirement so they may be reused.
|
|
|
|
DenseMap<Value *, Value *> Bitcasts;
|
|
|
|
|
|
|
|
// Emit lifetime markers for the pointers given in \p Objects. Insert the
|
|
|
|
// markers before the call if \p InsertBefore, and after the call otherwise.
|
|
|
|
auto insertMarkers = [&](Function *MarkerFunc, ArrayRef<Value *> Objects,
|
|
|
|
bool InsertBefore) {
|
|
|
|
for (Value *Mem : Objects) {
|
|
|
|
assert((!isa<Instruction>(Mem) || cast<Instruction>(Mem)->getFunction() ==
|
|
|
|
TheCall->getFunction()) &&
|
|
|
|
"Input memory not defined in original function");
|
|
|
|
Value *&MemAsI8Ptr = Bitcasts[Mem];
|
|
|
|
if (!MemAsI8Ptr) {
|
|
|
|
if (Mem->getType() == Int8PtrTy)
|
|
|
|
MemAsI8Ptr = Mem;
|
|
|
|
else
|
|
|
|
MemAsI8Ptr =
|
|
|
|
CastInst::CreatePointerCast(Mem, Int8PtrTy, "lt.cast", TheCall);
|
|
|
|
}
|
|
|
|
|
|
|
|
auto Marker = CallInst::Create(MarkerFunc, {NegativeOne, MemAsI8Ptr});
|
|
|
|
if (InsertBefore)
|
|
|
|
Marker->insertBefore(TheCall);
|
|
|
|
else
|
|
|
|
Marker->insertBefore(Term);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
if (!LifetimesStart.empty()) {
|
|
|
|
auto StartFn = llvm::Intrinsic::getDeclaration(
|
|
|
|
M, llvm::Intrinsic::lifetime_start, Int8PtrTy);
|
|
|
|
insertMarkers(StartFn, LifetimesStart, /*InsertBefore=*/true);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!LifetimesEnd.empty()) {
|
|
|
|
auto EndFn = llvm::Intrinsic::getDeclaration(
|
|
|
|
M, llvm::Intrinsic::lifetime_end, Int8PtrTy);
|
|
|
|
insertMarkers(EndFn, LifetimesEnd, /*InsertBefore=*/false);
|
2019-01-19 10:37:59 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-05-12 14:01:40 +08:00
|
|
|
/// emitCallAndSwitchStatement - This method sets up the caller side by adding
|
|
|
|
/// the call instruction, splitting any PHI nodes in the header block as
|
|
|
|
/// necessary.
|
2019-01-05 01:43:22 +08:00
|
|
|
CallInst *CodeExtractor::emitCallAndSwitchStatement(Function *newFunction,
|
|
|
|
BasicBlock *codeReplacer,
|
|
|
|
ValueSet &inputs,
|
|
|
|
ValueSet &outputs) {
|
2004-05-12 14:01:40 +08:00
|
|
|
// Emit a call to the new function, passing in: *pointer to struct (if
|
|
|
|
// aggregating parameters), or plan inputs and allocated memory for outputs
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<Value *> params, StructValues, ReloadOutputs, Reloads;
|
2017-04-11 06:27:50 +08:00
|
|
|
|
|
|
|
Module *M = newFunction->getParent();
|
|
|
|
LLVMContext &Context = M->getContext();
|
|
|
|
const DataLayout &DL = M->getDataLayout();
|
2019-01-05 01:43:22 +08:00
|
|
|
CallInst *call = nullptr;
|
2004-04-24 07:54:17 +08:00
|
|
|
|
|
|
|
// Add inputs as params, or to be filled into the struct
|
2019-01-29 03:13:37 +08:00
|
|
|
unsigned ArgNo = 0;
|
|
|
|
SmallVector<unsigned, 1> SwiftErrorArgs;
|
|
|
|
for (Value *input : inputs) {
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs)
|
2016-06-26 20:28:59 +08:00
|
|
|
StructValues.push_back(input);
|
2019-01-29 03:13:37 +08:00
|
|
|
else {
|
2016-06-26 20:28:59 +08:00
|
|
|
params.push_back(input);
|
2019-01-29 03:13:37 +08:00
|
|
|
if (input->isSwiftError())
|
|
|
|
SwiftErrorArgs.push_back(ArgNo);
|
|
|
|
}
|
|
|
|
++ArgNo;
|
|
|
|
}
|
2004-04-24 07:54:17 +08:00
|
|
|
|
|
|
|
// Create allocas for the outputs
|
2016-06-26 20:28:59 +08:00
|
|
|
for (Value *output : outputs) {
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs) {
|
2016-06-26 20:28:59 +08:00
|
|
|
StructValues.push_back(output);
|
2004-04-24 07:54:17 +08:00
|
|
|
} else {
|
|
|
|
AllocaInst *alloca =
|
2017-04-11 06:27:50 +08:00
|
|
|
new AllocaInst(output->getType(), DL.getAllocaAddrSpace(),
|
|
|
|
nullptr, output->getName() + ".loc",
|
|
|
|
&codeReplacer->getParent()->front().front());
|
2004-04-24 07:54:17 +08:00
|
|
|
ReloadOutputs.push_back(alloca);
|
|
|
|
params.push_back(alloca);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-03-14 09:53:18 +08:00
|
|
|
StructType *StructArgTy = nullptr;
|
2014-04-25 13:29:35 +08:00
|
|
|
AllocaInst *Struct = nullptr;
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs && (inputs.size() + outputs.size() > 0)) {
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<Type *> ArgTypes;
|
2012-05-04 18:18:49 +08:00
|
|
|
for (ValueSet::iterator v = StructValues.begin(),
|
2004-04-24 07:54:17 +08:00
|
|
|
ve = StructValues.end(); v != ve; ++v)
|
|
|
|
ArgTypes.push_back((*v)->getType());
|
|
|
|
|
|
|
|
// Allocate a struct at the beginning of this function
|
2015-03-14 09:53:18 +08:00
|
|
|
StructArgTy = StructType::get(newFunction->getContext(), ArgTypes);
|
2017-04-11 06:27:50 +08:00
|
|
|
Struct = new AllocaInst(StructArgTy, DL.getAllocaAddrSpace(), nullptr,
|
|
|
|
"structArg",
|
2015-10-13 10:39:05 +08:00
|
|
|
&codeReplacer->getParent()->front().front());
|
2004-04-24 07:54:17 +08:00
|
|
|
params.push_back(Struct);
|
|
|
|
|
|
|
|
for (unsigned i = 0, e = inputs.size(); i != e; ++i) {
|
2007-09-04 23:46:09 +08:00
|
|
|
Value *Idx[2];
|
2009-08-14 05:58:54 +08:00
|
|
|
Idx[0] = Constant::getNullValue(Type::getInt32Ty(Context));
|
|
|
|
Idx[1] = ConstantInt::get(Type::getInt32Ty(Context), i);
|
2015-03-14 09:53:18 +08:00
|
|
|
GetElementPtrInst *GEP = GetElementPtrInst::Create(
|
|
|
|
StructArgTy, Struct, Idx, "gep_" + StructValues[i]->getName());
|
2004-04-24 07:54:17 +08:00
|
|
|
codeReplacer->getInstList().push_back(GEP);
|
|
|
|
StoreInst *SI = new StoreInst(StructValues[i], GEP);
|
|
|
|
codeReplacer->getInstList().push_back(SI);
|
|
|
|
}
|
2005-04-22 07:48:37 +08:00
|
|
|
}
|
2004-04-24 07:54:17 +08:00
|
|
|
|
|
|
|
// Emit the call to the function
|
2019-01-05 01:43:22 +08:00
|
|
|
call = CallInst::Create(newFunction, params,
|
|
|
|
NumExitBlocks > 1 ? "targetBlock" : "");
|
[CodeExtractor] Add debug locations for new call and branch instrs.
Summary:
If a partially inlined function has debug info, we have to add debug
locations to the call instruction calling the outlined function.
We use the debug location of the first instruction in the outlined
function, as the introduced call transfers control to this statement and
there is no other equivalent line in the source code.
We also use the same debug location for the branch instruction added
to jump from artificial entry block for the outlined function, which just
jumps to the first actual basic block of the outlined function.
Reviewers: davide, aprantl, rriddle, dblaikie, danielcdh, wmi
Reviewed By: aprantl, rriddle, danielcdh
Subscribers: eraman, JDevlieghere, llvm-commits
Differential Revision: https://reviews.llvm.org/D40413
llvm-svn: 320199
2017-12-09 05:49:03 +08:00
|
|
|
// Add debug location to the new call, if the original function has debug
|
|
|
|
// info. In that case, the terminator of the entry block of the extracted
|
|
|
|
// function contains the first debug location of the extracted function,
|
|
|
|
// set in extractCodeRegion.
|
|
|
|
if (codeReplacer->getParent()->getSubprogram()) {
|
|
|
|
if (auto DL = newFunction->getEntryBlock().getTerminator()->getDebugLoc())
|
|
|
|
call->setDebugLoc(DL);
|
|
|
|
}
|
2004-04-24 07:54:17 +08:00
|
|
|
codeReplacer->getInstList().push_back(call);
|
|
|
|
|
2019-01-29 03:13:37 +08:00
|
|
|
// Set swifterror parameter attributes.
|
|
|
|
for (unsigned SwiftErrArgNo : SwiftErrorArgs) {
|
|
|
|
call->addParamAttr(SwiftErrArgNo, Attribute::SwiftError);
|
|
|
|
newFunction->addParamAttr(SwiftErrArgNo, Attribute::SwiftError);
|
|
|
|
}
|
|
|
|
|
2005-03-15 12:54:21 +08:00
|
|
|
Function::arg_iterator OutputArgBegin = newFunction->arg_begin();
|
2004-04-24 07:54:17 +08:00
|
|
|
unsigned FirstOut = inputs.size();
|
|
|
|
if (!AggregateArgs)
|
|
|
|
std::advance(OutputArgBegin, inputs.size());
|
2004-03-18 12:12:05 +08:00
|
|
|
|
[CodeExtractor] Fix multiple bugs under certain shape of extracted region
Summary:
If the extracted region has multiple exported data flows toward the same BB which is not included in the region, correct resotre instructions and PHI nodes won't be generated inside the exitStub. The solution is simply put the restore instructions right after the definition of output values instead of putting in exitStub.
Unittest for this bug is included.
Author: myhsu
Reviewers: chandlerc, davide, lattner, silvas, davidxl, wmi, kuhar
Subscribers: dberlin, kuhar, mgorny, llvm-commits
Differential Revision: https://reviews.llvm.org/D37902
llvm-svn: 315041
2017-10-06 11:37:06 +08:00
|
|
|
// Reload the outputs passed in by reference.
|
2004-03-18 11:49:40 +08:00
|
|
|
for (unsigned i = 0, e = outputs.size(); i != e; ++i) {
|
2014-04-25 13:29:35 +08:00
|
|
|
Value *Output = nullptr;
|
2004-04-24 07:54:17 +08:00
|
|
|
if (AggregateArgs) {
|
2007-09-04 23:46:09 +08:00
|
|
|
Value *Idx[2];
|
2009-08-14 05:58:54 +08:00
|
|
|
Idx[0] = Constant::getNullValue(Type::getInt32Ty(Context));
|
|
|
|
Idx[1] = ConstantInt::get(Type::getInt32Ty(Context), FirstOut + i);
|
2015-03-14 09:53:18 +08:00
|
|
|
GetElementPtrInst *GEP = GetElementPtrInst::Create(
|
|
|
|
StructArgTy, Struct, Idx, "gep_reload_" + outputs[i]->getName());
|
2004-04-24 07:54:17 +08:00
|
|
|
codeReplacer->getInstList().push_back(GEP);
|
|
|
|
Output = GEP;
|
|
|
|
} else {
|
|
|
|
Output = ReloadOutputs[i];
|
|
|
|
}
|
2019-02-02 04:44:24 +08:00
|
|
|
LoadInst *load = new LoadInst(outputs[i]->getType(), Output,
|
|
|
|
outputs[i]->getName() + ".reload");
|
2009-08-25 08:54:39 +08:00
|
|
|
Reloads.push_back(load);
|
2004-03-18 11:49:40 +08:00
|
|
|
codeReplacer->getInstList().push_back(load);
|
2017-10-12 05:41:43 +08:00
|
|
|
std::vector<User *> Users(outputs[i]->user_begin(), outputs[i]->user_end());
|
2004-03-18 11:49:40 +08:00
|
|
|
for (unsigned u = 0, e = Users.size(); u != e; ++u) {
|
|
|
|
Instruction *inst = cast<Instruction>(Users[u]);
|
2012-05-04 18:18:49 +08:00
|
|
|
if (!Blocks.count(inst->getParent()))
|
2004-03-18 11:49:40 +08:00
|
|
|
inst->replaceUsesOfWith(outputs[i], load);
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
|
|
|
}
|
2004-03-15 07:05:49 +08:00
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
// Now we can emit a switch statement using the call as a value.
|
2004-05-12 12:14:24 +08:00
|
|
|
SwitchInst *TheSwitch =
|
2009-08-14 05:58:54 +08:00
|
|
|
SwitchInst::Create(Constant::getNullValue(Type::getInt16Ty(Context)),
|
2008-04-07 04:25:17 +08:00
|
|
|
codeReplacer, 0, codeReplacer);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
|
|
|
// Since there may be multiple exits from the original region, make the new
|
2004-03-15 07:05:49 +08:00
|
|
|
// function return an unsigned, switch on that number. This loop iterates
|
|
|
|
// over all of the blocks in the extracted region, updating any terminator
|
|
|
|
// instructions in the to-be-extracted region that branch to blocks that are
|
|
|
|
// not in the region to be extracted.
|
2017-10-12 05:41:43 +08:00
|
|
|
std::map<BasicBlock *, BasicBlock *> ExitBlockMap;
|
2004-03-15 07:05:49 +08:00
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
unsigned switchVal = 0;
|
2016-06-26 20:28:59 +08:00
|
|
|
for (BasicBlock *Block : Blocks) {
|
2018-10-15 18:04:59 +08:00
|
|
|
Instruction *TI = Block->getTerminator();
|
2004-03-15 07:05:49 +08:00
|
|
|
for (unsigned i = 0, e = TI->getNumSuccessors(); i != e; ++i)
|
2012-05-04 18:18:49 +08:00
|
|
|
if (!Blocks.count(TI->getSuccessor(i))) {
|
2004-03-15 07:05:49 +08:00
|
|
|
BasicBlock *OldTarget = TI->getSuccessor(i);
|
|
|
|
// add a new basic block which returns the appropriate value
|
|
|
|
BasicBlock *&NewTarget = ExitBlockMap[OldTarget];
|
|
|
|
if (!NewTarget) {
|
|
|
|
// If we don't already have an exit stub for this non-extracted
|
|
|
|
// destination, create one now!
|
2009-08-14 05:58:54 +08:00
|
|
|
NewTarget = BasicBlock::Create(Context,
|
|
|
|
OldTarget->getName() + ".exitStub",
|
2008-04-07 04:25:17 +08:00
|
|
|
newFunction);
|
2004-05-12 12:14:24 +08:00
|
|
|
unsigned SuccNum = switchVal++;
|
|
|
|
|
2014-04-25 13:29:35 +08:00
|
|
|
Value *brVal = nullptr;
|
2004-05-12 12:14:24 +08:00
|
|
|
switch (NumExitBlocks) {
|
|
|
|
case 0:
|
|
|
|
case 1: break; // No value needed.
|
|
|
|
case 2: // Conditional branch, return a bool
|
2009-08-14 05:58:54 +08:00
|
|
|
brVal = ConstantInt::get(Type::getInt1Ty(Context), !SuccNum);
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
|
|
|
default:
|
2009-08-14 05:58:54 +08:00
|
|
|
brVal = ConstantInt::get(Type::getInt16Ty(Context), SuccNum);
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
|
|
|
}
|
2004-03-15 07:05:49 +08:00
|
|
|
|
[CodeExtractor] Fix multiple bugs under certain shape of extracted region
Summary:
If the extracted region has multiple exported data flows toward the same BB which is not included in the region, correct resotre instructions and PHI nodes won't be generated inside the exitStub. The solution is simply put the restore instructions right after the definition of output values instead of putting in exitStub.
Unittest for this bug is included.
Author: myhsu
Reviewers: chandlerc, davide, lattner, silvas, davidxl, wmi, kuhar
Subscribers: dberlin, kuhar, mgorny, llvm-commits
Differential Revision: https://reviews.llvm.org/D37902
llvm-svn: 315041
2017-10-06 11:37:06 +08:00
|
|
|
ReturnInst::Create(Context, brVal, NewTarget);
|
2004-03-15 07:05:49 +08:00
|
|
|
|
|
|
|
// Update the switch instruction.
|
2009-08-14 05:58:54 +08:00
|
|
|
TheSwitch->addCase(ConstantInt::get(Type::getInt16Ty(Context),
|
|
|
|
SuccNum),
|
2004-05-12 12:14:24 +08:00
|
|
|
OldTarget);
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
2004-03-15 06:34:55 +08:00
|
|
|
|
2004-03-15 07:05:49 +08:00
|
|
|
// rewrite the original branch instruction with this new target
|
|
|
|
TI->setSuccessor(i, NewTarget);
|
|
|
|
}
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
2004-03-15 07:43:24 +08:00
|
|
|
|
2019-02-09 04:48:04 +08:00
|
|
|
// Store the arguments right after the definition of output value.
|
|
|
|
// This should be proceeded after creating exit stubs to be ensure that invoke
|
|
|
|
// result restore will be placed in the outlined function.
|
|
|
|
Function::arg_iterator OAI = OutputArgBegin;
|
|
|
|
for (unsigned i = 0, e = outputs.size(); i != e; ++i) {
|
|
|
|
auto *OutI = dyn_cast<Instruction>(outputs[i]);
|
|
|
|
if (!OutI)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Find proper insertion point.
|
|
|
|
BasicBlock::iterator InsertPt;
|
|
|
|
// In case OutI is an invoke, we insert the store at the beginning in the
|
|
|
|
// 'normal destination' BB. Otherwise we insert the store right after OutI.
|
|
|
|
if (auto *InvokeI = dyn_cast<InvokeInst>(OutI))
|
|
|
|
InsertPt = InvokeI->getNormalDest()->getFirstInsertionPt();
|
|
|
|
else if (auto *Phi = dyn_cast<PHINode>(OutI))
|
|
|
|
InsertPt = Phi->getParent()->getFirstInsertionPt();
|
|
|
|
else
|
|
|
|
InsertPt = std::next(OutI->getIterator());
|
|
|
|
|
|
|
|
Instruction *InsertBefore = &*InsertPt;
|
|
|
|
assert((InsertBefore->getFunction() == newFunction ||
|
|
|
|
Blocks.count(InsertBefore->getParent())) &&
|
|
|
|
"InsertPt should be in new function");
|
|
|
|
assert(OAI != newFunction->arg_end() &&
|
|
|
|
"Number of output arguments should match "
|
|
|
|
"the amount of defined values");
|
|
|
|
if (AggregateArgs) {
|
|
|
|
Value *Idx[2];
|
|
|
|
Idx[0] = Constant::getNullValue(Type::getInt32Ty(Context));
|
|
|
|
Idx[1] = ConstantInt::get(Type::getInt32Ty(Context), FirstOut + i);
|
|
|
|
GetElementPtrInst *GEP = GetElementPtrInst::Create(
|
|
|
|
StructArgTy, &*OAI, Idx, "gep_" + outputs[i]->getName(),
|
|
|
|
InsertBefore);
|
|
|
|
new StoreInst(outputs[i], GEP, InsertBefore);
|
|
|
|
// Since there should be only one struct argument aggregating
|
|
|
|
// all the output values, we shouldn't increment OAI, which always
|
|
|
|
// points to the struct argument, in this case.
|
|
|
|
} else {
|
|
|
|
new StoreInst(outputs[i], &*OAI, InsertBefore);
|
|
|
|
++OAI;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-05-12 11:22:33 +08:00
|
|
|
// Now that we've done the deed, simplify the switch instruction.
|
2011-07-18 12:54:35 +08:00
|
|
|
Type *OldFnRetTy = TheSwitch->getParent()->getParent()->getReturnType();
|
2004-05-12 12:14:24 +08:00
|
|
|
switch (NumExitBlocks) {
|
|
|
|
case 0:
|
2004-08-12 11:17:02 +08:00
|
|
|
// There are no successors (the block containing the switch itself), which
|
2004-04-24 07:54:17 +08:00
|
|
|
// means that previously this was the last part of the function, and hence
|
|
|
|
// this should be rewritten as a `ret'
|
2005-04-22 07:48:37 +08:00
|
|
|
|
2004-04-24 07:54:17 +08:00
|
|
|
// Check if the function should return a value
|
2010-01-05 21:12:22 +08:00
|
|
|
if (OldFnRetTy->isVoidTy()) {
|
2014-04-25 13:29:35 +08:00
|
|
|
ReturnInst::Create(Context, nullptr, TheSwitch); // Return void
|
2004-08-12 11:17:02 +08:00
|
|
|
} else if (OldFnRetTy == TheSwitch->getCondition()->getType()) {
|
2004-04-24 07:54:17 +08:00
|
|
|
// return what we have
|
2009-08-14 05:58:54 +08:00
|
|
|
ReturnInst::Create(Context, TheSwitch->getCondition(), TheSwitch);
|
2004-08-12 11:17:02 +08:00
|
|
|
} else {
|
|
|
|
// Otherwise we must have code extracted an unwind or something, just
|
|
|
|
// return whatever we want.
|
2018-07-31 03:41:25 +08:00
|
|
|
ReturnInst::Create(Context,
|
2009-08-14 05:58:54 +08:00
|
|
|
Constant::getNullValue(OldFnRetTy), TheSwitch);
|
2004-08-12 11:17:02 +08:00
|
|
|
}
|
2004-04-24 07:54:17 +08:00
|
|
|
|
2008-06-22 06:08:46 +08:00
|
|
|
TheSwitch->eraseFromParent();
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
|
|
|
case 1:
|
|
|
|
// Only a single destination, change the switch into an unconditional
|
|
|
|
// branch.
|
2008-04-07 04:25:17 +08:00
|
|
|
BranchInst::Create(TheSwitch->getSuccessor(1), TheSwitch);
|
2008-06-22 06:08:46 +08:00
|
|
|
TheSwitch->eraseFromParent();
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
|
|
|
case 2:
|
2008-04-07 04:25:17 +08:00
|
|
|
BranchInst::Create(TheSwitch->getSuccessor(1), TheSwitch->getSuccessor(2),
|
|
|
|
call, TheSwitch);
|
2008-06-22 06:08:46 +08:00
|
|
|
TheSwitch->eraseFromParent();
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
// Otherwise, make the default destination of the switch instruction be one
|
|
|
|
// of the other successors.
|
SwitchInst refactoring.
The purpose of refactoring is to hide operand roles from SwitchInst user (programmer). If you want to play with operands directly, probably you will need lower level methods than SwitchInst ones (TerminatorInst or may be User). After this patch we can reorganize SwitchInst operands and successors as we want.
What was done:
1. Changed semantics of index inside the getCaseValue method:
getCaseValue(0) means "get first case", not a condition. Use getCondition() if you want to resolve the condition. I propose don't mix SwitchInst case indexing with low level indexing (TI successors indexing, User's operands indexing), since it may be dangerous.
2. By the same reason findCaseValue(ConstantInt*) returns actual number of case value. 0 means first case, not default. If there is no case with given value, ErrorIndex will returned.
3. Added getCaseSuccessor method. I propose to avoid usage of TerminatorInst::getSuccessor if you want to resolve case successor BB. Use getCaseSuccessor instead, since internal SwitchInst organization of operands/successors is hidden and may be changed in any moment.
4. Added resolveSuccessorIndex and resolveCaseIndex. The main purpose of these methods is to see how case successors are really mapped in TerminatorInst.
4.1 "resolveSuccessorIndex" was created if you need to level down from SwitchInst to TerminatorInst. It returns TerminatorInst's successor index for given case successor.
4.2 "resolveCaseIndex" converts low level successors index to case index that curresponds to the given successor.
Note: There are also related compatability fix patches for dragonegg, klee, llvm-gcc-4.0, llvm-gcc-4.2, safecode, clang.
llvm-svn: 149481
2012-02-01 15:49:51 +08:00
|
|
|
TheSwitch->setCondition(call);
|
|
|
|
TheSwitch->setDefaultDest(TheSwitch->getSuccessor(NumExitBlocks));
|
2012-03-08 15:06:20 +08:00
|
|
|
// Remove redundant case
|
Revert patches to add case-range support for PR1255.
The work on this project was left in an unfinished and inconsistent state.
Hopefully someone will eventually get a chance to implement this feature, but
in the meantime, it is better to put things back the way the were. I have
left support in the bitcode reader to handle the case-range bitcode format,
so that we do not lose bitcode compatibility with the llvm 3.3 release.
This reverts the following commits: 155464, 156374, 156377, 156613, 156704,
156757, 156804 156808, 156985, 157046, 157112, 157183, 157315, 157384, 157575,
157576, 157586, 157612, 157810, 157814, 157815, 157880, 157881, 157882, 157884,
157887, 157901, 158979, 157987, 157989, 158986, 158997, 159076, 159101, 159100,
159200, 159201, 159207, 159527, 159532, 159540, 159583, 159618, 159658, 159659,
159660, 159661, 159703, 159704, 160076, 167356, 172025, 186736
llvm-svn: 190328
2013-09-10 03:14:35 +08:00
|
|
|
TheSwitch->removeCase(SwitchInst::CaseIt(TheSwitch, NumExitBlocks-1));
|
2004-05-12 12:14:24 +08:00
|
|
|
break;
|
2004-03-15 07:43:24 +08:00
|
|
|
}
|
2019-01-05 01:43:22 +08:00
|
|
|
|
2019-01-19 10:37:59 +08:00
|
|
|
// Insert lifetime markers around the reloads of any output values. The
|
|
|
|
// allocas output values are stored in are only in-use in the codeRepl block.
|
2019-02-14 03:53:38 +08:00
|
|
|
insertLifetimeMarkersSurroundingCall(M, ReloadOutputs, ReloadOutputs, call);
|
2019-01-19 10:37:59 +08:00
|
|
|
|
2019-01-05 01:43:22 +08:00
|
|
|
return call;
|
2004-02-28 11:26:20 +08:00
|
|
|
}
|
|
|
|
|
2004-05-12 14:01:40 +08:00
|
|
|
void CodeExtractor::moveCodeToFunction(Function *newFunction) {
|
2012-05-04 18:18:49 +08:00
|
|
|
Function *oldFunc = (*Blocks.begin())->getParent();
|
2004-05-12 14:01:40 +08:00
|
|
|
Function::BasicBlockListType &oldBlocks = oldFunc->getBasicBlockList();
|
|
|
|
Function::BasicBlockListType &newBlocks = newFunction->getBasicBlockList();
|
|
|
|
|
2016-06-26 20:28:59 +08:00
|
|
|
for (BasicBlock *Block : Blocks) {
|
2004-05-12 14:01:40 +08:00
|
|
|
// Delete the basic block from the old function, and the list of blocks
|
2016-06-26 20:28:59 +08:00
|
|
|
oldBlocks.remove(Block);
|
2004-05-12 14:01:40 +08:00
|
|
|
|
|
|
|
// Insert this basic block into the new function
|
2016-06-26 20:28:59 +08:00
|
|
|
newBlocks.push_back(Block);
|
2004-05-12 14:01:40 +08:00
|
|
|
}
|
|
|
|
}
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2016-08-02 10:15:45 +08:00
|
|
|
void CodeExtractor::calculateNewCallTerminatorWeights(
|
|
|
|
BasicBlock *CodeReplacer,
|
|
|
|
DenseMap<BasicBlock *, BlockFrequency> &ExitWeights,
|
|
|
|
BranchProbabilityInfo *BPI) {
|
2017-10-12 05:41:43 +08:00
|
|
|
using Distribution = BlockFrequencyInfoImplBase::Distribution;
|
|
|
|
using BlockNode = BlockFrequencyInfoImplBase::BlockNode;
|
2016-08-02 10:15:45 +08:00
|
|
|
|
|
|
|
// Update the branch weights for the exit block.
|
2018-10-15 18:04:59 +08:00
|
|
|
Instruction *TI = CodeReplacer->getTerminator();
|
2016-08-02 10:15:45 +08:00
|
|
|
SmallVector<unsigned, 8> BranchWeights(TI->getNumSuccessors(), 0);
|
|
|
|
|
|
|
|
// Block Frequency distribution with dummy node.
|
|
|
|
Distribution BranchDist;
|
|
|
|
|
|
|
|
// Add each of the frequencies of the successors.
|
|
|
|
for (unsigned i = 0, e = TI->getNumSuccessors(); i < e; ++i) {
|
|
|
|
BlockNode ExitNode(i);
|
|
|
|
uint64_t ExitFreq = ExitWeights[TI->getSuccessor(i)].getFrequency();
|
|
|
|
if (ExitFreq != 0)
|
|
|
|
BranchDist.addExit(ExitNode, ExitFreq);
|
|
|
|
else
|
|
|
|
BPI->setEdgeProbability(CodeReplacer, i, BranchProbability::getZero());
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check for no total weight.
|
|
|
|
if (BranchDist.Total == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
// Normalize the distribution so that they can fit in unsigned.
|
|
|
|
BranchDist.normalize();
|
|
|
|
|
|
|
|
// Create normalized branch weights and set the metadata.
|
|
|
|
for (unsigned I = 0, E = BranchDist.Weights.size(); I < E; ++I) {
|
|
|
|
const auto &Weight = BranchDist.Weights[I];
|
|
|
|
|
|
|
|
// Get the weight and update the current BFI.
|
|
|
|
BranchWeights[Weight.TargetNode.Index] = Weight.Amount;
|
|
|
|
BranchProbability BP(Weight.Amount, BranchDist.Total);
|
|
|
|
BPI->setEdgeProbability(CodeReplacer, Weight.TargetNode.Index, BP);
|
|
|
|
}
|
|
|
|
TI->setMetadata(
|
|
|
|
LLVMContext::MD_prof,
|
|
|
|
MDBuilder(TI->getContext()).createBranchWeights(BranchWeights));
|
|
|
|
}
|
|
|
|
|
2020-01-16 03:26:34 +08:00
|
|
|
/// Erase debug info intrinsics which refer to values in \p F but aren't in
|
|
|
|
/// \p F.
|
|
|
|
static void eraseDebugIntrinsicsWithNonLocalRefs(Function &F) {
|
|
|
|
for (Instruction &I : instructions(F)) {
|
|
|
|
SmallVector<DbgVariableIntrinsic *, 4> DbgUsers;
|
|
|
|
findDbgUsers(DbgUsers, &I);
|
|
|
|
for (DbgVariableIntrinsic *DVI : DbgUsers)
|
|
|
|
if (DVI->getFunction() != &F)
|
|
|
|
DVI->eraseFromParent();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Fix up the debug info in the old and new functions by pointing line
|
|
|
|
/// locations and debug intrinsics to the new subprogram scope, and by deleting
|
|
|
|
/// intrinsics which point to values outside of the new function.
|
|
|
|
static void fixupDebugInfoPostExtraction(Function &OldFunc, Function &NewFunc,
|
|
|
|
CallInst &TheCall) {
|
|
|
|
DISubprogram *OldSP = OldFunc.getSubprogram();
|
|
|
|
LLVMContext &Ctx = OldFunc.getContext();
|
|
|
|
|
2020-02-13 10:52:23 +08:00
|
|
|
if (!OldSP) {
|
2020-01-16 03:26:34 +08:00
|
|
|
// Erase any debug info the new function contains.
|
|
|
|
stripDebugInfo(NewFunc);
|
|
|
|
// Make sure the old function doesn't contain any non-local metadata refs.
|
|
|
|
eraseDebugIntrinsicsWithNonLocalRefs(NewFunc);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a subprogram for the new function. Leave out a description of the
|
|
|
|
// function arguments, as the parameters don't correspond to anything at the
|
|
|
|
// source level.
|
|
|
|
assert(OldSP->getUnit() && "Missing compile unit for subprogram");
|
|
|
|
DIBuilder DIB(*OldFunc.getParent(), /*AllowUnresolvedNodes=*/false,
|
|
|
|
OldSP->getUnit());
|
|
|
|
auto SPType = DIB.createSubroutineType(DIB.getOrCreateTypeArray(None));
|
|
|
|
DISubprogram::DISPFlags SPFlags = DISubprogram::SPFlagDefinition |
|
|
|
|
DISubprogram::SPFlagOptimized |
|
|
|
|
DISubprogram::SPFlagLocalToUnit;
|
|
|
|
auto NewSP = DIB.createFunction(
|
|
|
|
OldSP->getUnit(), NewFunc.getName(), NewFunc.getName(), OldSP->getFile(),
|
|
|
|
/*LineNo=*/0, SPType, /*ScopeLine=*/0, DINode::FlagZero, SPFlags);
|
|
|
|
NewFunc.setSubprogram(NewSP);
|
|
|
|
|
|
|
|
// Debug intrinsics in the new function need to be updated in one of two
|
|
|
|
// ways:
|
|
|
|
// 1) They need to be deleted, because they describe a value in the old
|
|
|
|
// function.
|
|
|
|
// 2) They need to point to fresh metadata, e.g. because they currently
|
|
|
|
// point to a variable in the wrong scope.
|
|
|
|
SmallDenseMap<DINode *, DINode *> RemappedMetadata;
|
|
|
|
SmallVector<Instruction *, 4> DebugIntrinsicsToDelete;
|
|
|
|
for (Instruction &I : instructions(NewFunc)) {
|
|
|
|
auto *DII = dyn_cast<DbgInfoIntrinsic>(&I);
|
|
|
|
if (!DII)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Point the intrinsic to a fresh label within the new function.
|
|
|
|
if (auto *DLI = dyn_cast<DbgLabelInst>(&I)) {
|
|
|
|
DILabel *OldLabel = DLI->getLabel();
|
|
|
|
DINode *&NewLabel = RemappedMetadata[OldLabel];
|
|
|
|
if (!NewLabel)
|
|
|
|
NewLabel = DILabel::get(Ctx, NewSP, OldLabel->getName(),
|
|
|
|
OldLabel->getFile(), OldLabel->getLine());
|
|
|
|
DLI->setArgOperand(0, MetadataAsValue::get(Ctx, NewLabel));
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the location isn't a constant or an instruction, delete the
|
|
|
|
// intrinsic.
|
|
|
|
auto *DVI = cast<DbgVariableIntrinsic>(DII);
|
|
|
|
Value *Location = DVI->getVariableLocation();
|
|
|
|
if (!Location ||
|
|
|
|
(!isa<Constant>(Location) && !isa<Instruction>(Location))) {
|
|
|
|
DebugIntrinsicsToDelete.push_back(DVI);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the variable location is an instruction but isn't in the new
|
|
|
|
// function, delete the intrinsic.
|
|
|
|
Instruction *LocationInst = dyn_cast<Instruction>(Location);
|
|
|
|
if (LocationInst && LocationInst->getFunction() != &NewFunc) {
|
|
|
|
DebugIntrinsicsToDelete.push_back(DVI);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Point the intrinsic to a fresh variable within the new function.
|
|
|
|
DILocalVariable *OldVar = DVI->getVariable();
|
|
|
|
DINode *&NewVar = RemappedMetadata[OldVar];
|
|
|
|
if (!NewVar)
|
|
|
|
NewVar = DIB.createAutoVariable(
|
|
|
|
NewSP, OldVar->getName(), OldVar->getFile(), OldVar->getLine(),
|
|
|
|
OldVar->getType(), /*AlwaysPreserve=*/false, DINode::FlagZero,
|
|
|
|
OldVar->getAlignInBits());
|
|
|
|
DVI->setArgOperand(1, MetadataAsValue::get(Ctx, NewVar));
|
|
|
|
}
|
|
|
|
for (auto *DII : DebugIntrinsicsToDelete)
|
|
|
|
DII->eraseFromParent();
|
|
|
|
DIB.finalizeSubprogram(NewSP);
|
|
|
|
|
|
|
|
// Fix up the scope information attached to the line locations in the new
|
|
|
|
// function.
|
|
|
|
for (Instruction &I : instructions(NewFunc)) {
|
|
|
|
if (const DebugLoc &DL = I.getDebugLoc())
|
|
|
|
I.setDebugLoc(DebugLoc::get(DL.getLine(), DL.getCol(), NewSP));
|
|
|
|
|
|
|
|
// Loop info metadata may contain line locations. Fix them up.
|
|
|
|
auto updateLoopInfoLoc = [&Ctx,
|
|
|
|
NewSP](const DILocation &Loc) -> DILocation * {
|
|
|
|
return DILocation::get(Ctx, Loc.getLine(), Loc.getColumn(), NewSP,
|
|
|
|
nullptr);
|
|
|
|
};
|
|
|
|
updateLoopMetadataDebugLocations(I, updateLoopInfoLoc);
|
|
|
|
}
|
|
|
|
if (!TheCall.getDebugLoc())
|
|
|
|
TheCall.setDebugLoc(DebugLoc::get(0, 0, OldSP));
|
|
|
|
|
|
|
|
eraseDebugIntrinsicsWithNonLocalRefs(NewFunc);
|
|
|
|
}
|
|
|
|
|
2019-10-09 01:17:51 +08:00
|
|
|
Function *
|
|
|
|
CodeExtractor::extractCodeRegion(const CodeExtractorAnalysisCache &CEAC) {
|
2012-05-04 18:18:49 +08:00
|
|
|
if (!isEligible())
|
2014-04-25 13:29:35 +08:00
|
|
|
return nullptr;
|
2004-04-24 07:54:17 +08:00
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
// Assumption: this is a single-entry code region, and the header is the first
|
2004-03-15 09:18:23 +08:00
|
|
|
// block in the region.
|
2012-05-04 18:18:49 +08:00
|
|
|
BasicBlock *header = *Blocks.begin();
|
2017-11-13 18:35:52 +08:00
|
|
|
Function *oldFunction = header->getParent();
|
|
|
|
|
2016-08-02 10:15:45 +08:00
|
|
|
// Calculate the entry frequency of the new function before we change the root
|
|
|
|
// block.
|
|
|
|
BlockFrequency EntryFreq;
|
|
|
|
if (BFI) {
|
|
|
|
assert(BPI && "Both BPI and BFI are required to preserve profile info");
|
|
|
|
for (BasicBlock *Pred : predecessors(header)) {
|
|
|
|
if (Blocks.count(Pred))
|
|
|
|
continue;
|
|
|
|
EntryFreq +=
|
|
|
|
BFI->getBlockFreq(Pred) * BPI->getEdgeProbability(Pred, header);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-29 09:03:39 +08:00
|
|
|
// Remove @llvm.assume calls that will be moved to the new function from the
|
|
|
|
// old function's assumption cache.
|
|
|
|
for (BasicBlock *Block : Blocks) {
|
|
|
|
for (auto It = Block->begin(), End = Block->end(); It != End;) {
|
|
|
|
Instruction *I = &*It;
|
|
|
|
++It;
|
|
|
|
|
|
|
|
if (match(I, m_Intrinsic<Intrinsic::assume>())) {
|
|
|
|
if (AC)
|
|
|
|
AC->unregisterAssumption(cast<CallInst>(I));
|
|
|
|
I->eraseFromParent();
|
|
|
|
}
|
|
|
|
}
|
2019-10-05 06:46:42 +08:00
|
|
|
}
|
|
|
|
|
2004-05-13 00:07:41 +08:00
|
|
|
// If we have any return instructions in the region, split those blocks so
|
|
|
|
// that the return is not in the region.
|
|
|
|
splitReturnBlocks();
|
|
|
|
|
2018-12-04 06:40:21 +08:00
|
|
|
// Calculate the exit blocks for the extracted region and the total exit
|
|
|
|
// weights for each of those blocks.
|
|
|
|
DenseMap<BasicBlock *, BlockFrequency> ExitWeights;
|
|
|
|
SmallPtrSet<BasicBlock *, 1> ExitBlocks;
|
|
|
|
for (BasicBlock *Block : Blocks) {
|
|
|
|
for (succ_iterator SI = succ_begin(Block), SE = succ_end(Block); SI != SE;
|
|
|
|
++SI) {
|
|
|
|
if (!Blocks.count(*SI)) {
|
|
|
|
// Update the branch weight for this successor.
|
|
|
|
if (BFI) {
|
|
|
|
BlockFrequency &BF = ExitWeights[*SI];
|
|
|
|
BF += BFI->getBlockFreq(Block) * BPI->getEdgeProbability(Block, *SI);
|
|
|
|
}
|
|
|
|
ExitBlocks.insert(*SI);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
NumExitBlocks = ExitBlocks.size();
|
|
|
|
|
|
|
|
// If we have to split PHI nodes of the entry or exit blocks, do so now.
|
|
|
|
severSplitPHINodesOfEntry(header);
|
|
|
|
severSplitPHINodesOfExits(ExitBlocks);
|
|
|
|
|
2004-02-28 11:26:20 +08:00
|
|
|
// This takes place of the original loop
|
2018-07-31 03:41:25 +08:00
|
|
|
BasicBlock *codeReplacer = BasicBlock::Create(header->getContext(),
|
2009-08-14 05:58:54 +08:00
|
|
|
"codeRepl", oldFunction,
|
2008-05-15 18:04:30 +08:00
|
|
|
header);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
|
|
|
// The new function needs a root node because other nodes can branch to the
|
2004-05-12 14:01:40 +08:00
|
|
|
// head of the region, but the entry node of a function cannot have preds.
|
2018-07-31 03:41:25 +08:00
|
|
|
BasicBlock *newFuncRoot = BasicBlock::Create(header->getContext(),
|
2009-08-14 05:58:54 +08:00
|
|
|
"newFuncRoot");
|
[CodeExtractor] Add debug locations for new call and branch instrs.
Summary:
If a partially inlined function has debug info, we have to add debug
locations to the call instruction calling the outlined function.
We use the debug location of the first instruction in the outlined
function, as the introduced call transfers control to this statement and
there is no other equivalent line in the source code.
We also use the same debug location for the branch instruction added
to jump from artificial entry block for the outlined function, which just
jumps to the first actual basic block of the outlined function.
Reviewers: davide, aprantl, rriddle, dblaikie, danielcdh, wmi
Reviewed By: aprantl, rriddle, danielcdh
Subscribers: eraman, JDevlieghere, llvm-commits
Differential Revision: https://reviews.llvm.org/D40413
llvm-svn: 320199
2017-12-09 05:49:03 +08:00
|
|
|
auto *BranchI = BranchInst::Create(header);
|
|
|
|
// If the original function has debug info, we have to add a debug location
|
|
|
|
// to the new branch instruction from the artificial entry block.
|
|
|
|
// We use the debug location of the first instruction in the extracted
|
|
|
|
// blocks, as there is no other equivalent line in the source code.
|
|
|
|
if (oldFunction->getSubprogram()) {
|
|
|
|
any_of(Blocks, [&BranchI](const BasicBlock *BB) {
|
|
|
|
return any_of(*BB, [&BranchI](const Instruction &I) {
|
|
|
|
if (!I.getDebugLoc())
|
|
|
|
return false;
|
|
|
|
BranchI->setDebugLoc(I.getDebugLoc());
|
|
|
|
return true;
|
|
|
|
});
|
|
|
|
});
|
|
|
|
}
|
|
|
|
newFuncRoot->getInstList().push_back(BranchI);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2019-10-02 23:36:39 +08:00
|
|
|
ValueSet inputs, outputs, SinkingCands, HoistingCands;
|
|
|
|
BasicBlock *CommonExit = nullptr;
|
2019-10-09 01:17:51 +08:00
|
|
|
findAllocas(CEAC, SinkingCands, HoistingCands, CommonExit);
|
2017-06-12 04:46:05 +08:00
|
|
|
assert(HoistingCands.empty() || CommonExit);
|
2017-05-31 05:22:18 +08:00
|
|
|
|
2004-05-12 14:01:40 +08:00
|
|
|
// Find inputs to, outputs from the code region.
|
2017-05-31 05:22:18 +08:00
|
|
|
findInputsOutputs(inputs, outputs, SinkingCands);
|
|
|
|
|
2019-07-11 00:32:16 +08:00
|
|
|
// Now sink all instructions which only have non-phi uses inside the region.
|
2019-07-11 00:32:20 +08:00
|
|
|
// Group the allocas at the start of the block, so that any bitcast uses of
|
|
|
|
// the allocas are well-defined.
|
|
|
|
AllocaInst *FirstSunkAlloca = nullptr;
|
|
|
|
for (auto *II : SinkingCands) {
|
|
|
|
if (auto *AI = dyn_cast<AllocaInst>(II)) {
|
|
|
|
AI->moveBefore(*newFuncRoot, newFuncRoot->getFirstInsertionPt());
|
|
|
|
if (!FirstSunkAlloca)
|
|
|
|
FirstSunkAlloca = AI;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
assert((SinkingCands.empty() || FirstSunkAlloca) &&
|
|
|
|
"Did not expect a sink candidate without any allocas");
|
|
|
|
for (auto *II : SinkingCands) {
|
|
|
|
if (!isa<AllocaInst>(II)) {
|
|
|
|
cast<Instruction>(II)->moveAfter(FirstSunkAlloca);
|
|
|
|
}
|
|
|
|
}
|
2004-05-12 14:01:40 +08:00
|
|
|
|
2017-06-12 04:46:05 +08:00
|
|
|
if (!HoistingCands.empty()) {
|
|
|
|
auto *HoistToBlock = findOrCreateBlockForHoisting(CommonExit);
|
|
|
|
Instruction *TI = HoistToBlock->getTerminator();
|
|
|
|
for (auto *II : HoistingCands)
|
|
|
|
cast<Instruction>(II)->moveBefore(TI);
|
|
|
|
}
|
|
|
|
|
2019-01-05 01:43:22 +08:00
|
|
|
// Collect objects which are inputs to the extraction region and also
|
2019-02-16 02:46:58 +08:00
|
|
|
// referenced by lifetime start markers within it. The effects of these
|
2019-01-05 01:43:22 +08:00
|
|
|
// markers must be replicated in the calling function to prevent the stack
|
|
|
|
// coloring pass from merging slots which store input objects.
|
2019-02-16 02:46:58 +08:00
|
|
|
ValueSet LifetimesStart;
|
|
|
|
eraseLifetimeMarkersOnInputs(Blocks, SinkingCands, LifetimesStart);
|
2019-01-05 01:43:22 +08:00
|
|
|
|
2004-05-12 14:01:40 +08:00
|
|
|
// Construct new function based on inputs/outputs & add allocas for all defs.
|
2019-01-05 01:43:22 +08:00
|
|
|
Function *newFunction =
|
|
|
|
constructFunction(inputs, outputs, header, newFuncRoot, codeReplacer,
|
|
|
|
oldFunction, oldFunction->getParent());
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2016-08-02 10:15:45 +08:00
|
|
|
// Update the entry count of the function.
|
|
|
|
if (BFI) {
|
2018-01-18 06:24:23 +08:00
|
|
|
auto Count = BFI->getProfileCountFromFreq(EntryFreq.getFrequency());
|
|
|
|
if (Count.hasValue())
|
|
|
|
newFunction->setEntryCount(
|
|
|
|
ProfileCount(Count.getValue(), Function::PCT_Real)); // FIXME
|
2016-08-02 10:15:45 +08:00
|
|
|
BFI->setBlockFreq(codeReplacer, EntryFreq.getFrequency());
|
|
|
|
}
|
|
|
|
|
2019-01-05 01:43:22 +08:00
|
|
|
CallInst *TheCall =
|
|
|
|
emitCallAndSwitchStatement(newFunction, codeReplacer, inputs, outputs);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2004-03-15 06:34:55 +08:00
|
|
|
moveCodeToFunction(newFunction);
|
2019-01-05 01:43:22 +08:00
|
|
|
|
|
|
|
// Replicate the effects of any lifetime start/end markers which referenced
|
|
|
|
// input objects in the extraction region by placing markers around the call.
|
2019-02-16 02:46:58 +08:00
|
|
|
insertLifetimeMarkersSurroundingCall(
|
|
|
|
oldFunction->getParent(), LifetimesStart.getArrayRef(), {}, TheCall);
|
2004-02-28 11:26:20 +08:00
|
|
|
|
2018-05-12 06:49:49 +08:00
|
|
|
// Propagate personality info to the new function if there is one.
|
|
|
|
if (oldFunction->hasPersonalityFn())
|
|
|
|
newFunction->setPersonalityFn(oldFunction->getPersonalityFn());
|
|
|
|
|
2016-08-02 10:15:45 +08:00
|
|
|
// Update the branch weights for the exit block.
|
|
|
|
if (BFI && NumExitBlocks > 1)
|
|
|
|
calculateNewCallTerminatorWeights(codeReplacer, ExitWeights, BPI);
|
|
|
|
|
2018-12-04 06:40:21 +08:00
|
|
|
// Loop over all of the PHI nodes in the header and exit blocks, and change
|
|
|
|
// any references to the old incoming edge to be the new incoming edge.
|
2004-09-16 01:06:42 +08:00
|
|
|
for (BasicBlock::iterator I = header->begin(); isa<PHINode>(I); ++I) {
|
|
|
|
PHINode *PN = cast<PHINode>(I);
|
2004-03-18 13:28:49 +08:00
|
|
|
for (unsigned i = 0, e = PN->getNumIncomingValues(); i != e; ++i)
|
2012-05-04 18:18:49 +08:00
|
|
|
if (!Blocks.count(PN->getIncomingBlock(i)))
|
2004-03-18 13:28:49 +08:00
|
|
|
PN->setIncomingBlock(i, newFuncRoot);
|
2004-09-16 01:06:42 +08:00
|
|
|
}
|
2005-04-22 07:48:37 +08:00
|
|
|
|
2018-12-04 06:40:21 +08:00
|
|
|
for (BasicBlock *ExitBB : ExitBlocks)
|
|
|
|
for (PHINode &PN : ExitBB->phis()) {
|
[HotColdSplitting] Identify larger cold regions using domtree queries
The current splitting algorithm works in three stages:
1) Identify cold blocks, then
2) Use forward/backward propagation to mark hot blocks, then
3) Grow a SESE region of blocks *outside* of the set of hot blocks and
start outlining.
While testing this pass on Apple internal frameworks I noticed that some
kinds of control flow (e.g. loops) are never outlined, even though they
unconditionally lead to / follow cold blocks. I noticed two other issues
related to how cold regions are identified:
- An inconsistency can arise in the internal state of the hotness
propagation stage, as a block may end up in both the ColdBlocks set
and the HotBlocks set. Further inconsistencies can arise as these sets
do not match what's in ProfileSummaryInfo.
- It isn't necessary to limit outlining to single-exit regions.
This patch teaches the splitting algorithm to identify maximal cold
regions and outline them. A maximal cold region is defined as the set of
blocks post-dominated by a cold sink block, or dominated by that sink
block. This approach can successfully outline loops in the cold path. As
a side benefit, it maintains less internal state than the current
approach.
Due to a limitation in CodeExtractor, blocks within the maximal cold
region which aren't dominated by a single entry point (a so-called "max
ancestor") are filtered out.
Results:
- X86 (LNT + -Os + externals): 134KB of TEXT were outlined compared to
47KB pre-patch, or a ~3x improvement. Did not see a performance impact
across two runs.
- AArch64 (LNT + -Os + externals + Apple-internal benchmarks): 149KB
of TEXT were outlined. Ditto re: performance impact.
- Outlining results improve marginally in the internal frameworks I
tested.
Follow-ups:
- Outline more than once per function, outline large single basic
blocks, & try to remove unconditional branches in outlined functions.
Differential Revision: https://reviews.llvm.org/D53627
llvm-svn: 345209
2018-10-25 06:15:41 +08:00
|
|
|
Value *IncomingCodeReplacerVal = nullptr;
|
2018-12-04 06:40:21 +08:00
|
|
|
for (unsigned i = 0, e = PN.getNumIncomingValues(); i != e; ++i) {
|
[HotColdSplitting] Identify larger cold regions using domtree queries
The current splitting algorithm works in three stages:
1) Identify cold blocks, then
2) Use forward/backward propagation to mark hot blocks, then
3) Grow a SESE region of blocks *outside* of the set of hot blocks and
start outlining.
While testing this pass on Apple internal frameworks I noticed that some
kinds of control flow (e.g. loops) are never outlined, even though they
unconditionally lead to / follow cold blocks. I noticed two other issues
related to how cold regions are identified:
- An inconsistency can arise in the internal state of the hotness
propagation stage, as a block may end up in both the ColdBlocks set
and the HotBlocks set. Further inconsistencies can arise as these sets
do not match what's in ProfileSummaryInfo.
- It isn't necessary to limit outlining to single-exit regions.
This patch teaches the splitting algorithm to identify maximal cold
regions and outline them. A maximal cold region is defined as the set of
blocks post-dominated by a cold sink block, or dominated by that sink
block. This approach can successfully outline loops in the cold path. As
a side benefit, it maintains less internal state than the current
approach.
Due to a limitation in CodeExtractor, blocks within the maximal cold
region which aren't dominated by a single entry point (a so-called "max
ancestor") are filtered out.
Results:
- X86 (LNT + -Os + externals): 134KB of TEXT were outlined compared to
47KB pre-patch, or a ~3x improvement. Did not see a performance impact
across two runs.
- AArch64 (LNT + -Os + externals + Apple-internal benchmarks): 149KB
of TEXT were outlined. Ditto re: performance impact.
- Outlining results improve marginally in the internal frameworks I
tested.
Follow-ups:
- Outline more than once per function, outline large single basic
blocks, & try to remove unconditional branches in outlined functions.
Differential Revision: https://reviews.llvm.org/D53627
llvm-svn: 345209
2018-10-25 06:15:41 +08:00
|
|
|
// Ignore incoming values from outside of the extracted region.
|
2018-12-04 06:40:21 +08:00
|
|
|
if (!Blocks.count(PN.getIncomingBlock(i)))
|
[HotColdSplitting] Identify larger cold regions using domtree queries
The current splitting algorithm works in three stages:
1) Identify cold blocks, then
2) Use forward/backward propagation to mark hot blocks, then
3) Grow a SESE region of blocks *outside* of the set of hot blocks and
start outlining.
While testing this pass on Apple internal frameworks I noticed that some
kinds of control flow (e.g. loops) are never outlined, even though they
unconditionally lead to / follow cold blocks. I noticed two other issues
related to how cold regions are identified:
- An inconsistency can arise in the internal state of the hotness
propagation stage, as a block may end up in both the ColdBlocks set
and the HotBlocks set. Further inconsistencies can arise as these sets
do not match what's in ProfileSummaryInfo.
- It isn't necessary to limit outlining to single-exit regions.
This patch teaches the splitting algorithm to identify maximal cold
regions and outline them. A maximal cold region is defined as the set of
blocks post-dominated by a cold sink block, or dominated by that sink
block. This approach can successfully outline loops in the cold path. As
a side benefit, it maintains less internal state than the current
approach.
Due to a limitation in CodeExtractor, blocks within the maximal cold
region which aren't dominated by a single entry point (a so-called "max
ancestor") are filtered out.
Results:
- X86 (LNT + -Os + externals): 134KB of TEXT were outlined compared to
47KB pre-patch, or a ~3x improvement. Did not see a performance impact
across two runs.
- AArch64 (LNT + -Os + externals + Apple-internal benchmarks): 149KB
of TEXT were outlined. Ditto re: performance impact.
- Outlining results improve marginally in the internal frameworks I
tested.
Follow-ups:
- Outline more than once per function, outline large single basic
blocks, & try to remove unconditional branches in outlined functions.
Differential Revision: https://reviews.llvm.org/D53627
llvm-svn: 345209
2018-10-25 06:15:41 +08:00
|
|
|
continue;
|
|
|
|
|
|
|
|
// Ensure that there is only one incoming value from codeReplacer.
|
|
|
|
if (!IncomingCodeReplacerVal) {
|
2018-12-04 06:40:21 +08:00
|
|
|
PN.setIncomingBlock(i, codeReplacer);
|
|
|
|
IncomingCodeReplacerVal = PN.getIncomingValue(i);
|
|
|
|
} else
|
|
|
|
assert(IncomingCodeReplacerVal == PN.getIncomingValue(i) &&
|
[HotColdSplitting] Identify larger cold regions using domtree queries
The current splitting algorithm works in three stages:
1) Identify cold blocks, then
2) Use forward/backward propagation to mark hot blocks, then
3) Grow a SESE region of blocks *outside* of the set of hot blocks and
start outlining.
While testing this pass on Apple internal frameworks I noticed that some
kinds of control flow (e.g. loops) are never outlined, even though they
unconditionally lead to / follow cold blocks. I noticed two other issues
related to how cold regions are identified:
- An inconsistency can arise in the internal state of the hotness
propagation stage, as a block may end up in both the ColdBlocks set
and the HotBlocks set. Further inconsistencies can arise as these sets
do not match what's in ProfileSummaryInfo.
- It isn't necessary to limit outlining to single-exit regions.
This patch teaches the splitting algorithm to identify maximal cold
regions and outline them. A maximal cold region is defined as the set of
blocks post-dominated by a cold sink block, or dominated by that sink
block. This approach can successfully outline loops in the cold path. As
a side benefit, it maintains less internal state than the current
approach.
Due to a limitation in CodeExtractor, blocks within the maximal cold
region which aren't dominated by a single entry point (a so-called "max
ancestor") are filtered out.
Results:
- X86 (LNT + -Os + externals): 134KB of TEXT were outlined compared to
47KB pre-patch, or a ~3x improvement. Did not see a performance impact
across two runs.
- AArch64 (LNT + -Os + externals + Apple-internal benchmarks): 149KB
of TEXT were outlined. Ditto re: performance impact.
- Outlining results improve marginally in the internal frameworks I
tested.
Follow-ups:
- Outline more than once per function, outline large single basic
blocks, & try to remove unconditional branches in outlined functions.
Differential Revision: https://reviews.llvm.org/D53627
llvm-svn: 345209
2018-10-25 06:15:41 +08:00
|
|
|
"PHI has two incompatbile incoming values from codeRepl");
|
|
|
|
}
|
2004-08-13 11:27:07 +08:00
|
|
|
}
|
2005-04-22 07:48:37 +08:00
|
|
|
|
2020-01-16 03:26:34 +08:00
|
|
|
fixupDebugInfoPostExtraction(*oldFunction, *newFunction, *TheCall);
|
2018-10-16 03:22:20 +08:00
|
|
|
|
2018-12-06 03:35:37 +08:00
|
|
|
// Mark the new function `noreturn` if applicable. Terminators which resume
|
|
|
|
// exception propagation are treated as returning instructions. This is to
|
|
|
|
// avoid inserting traps after calls to outlined functions which unwind.
|
2018-11-09 01:57:09 +08:00
|
|
|
bool doesNotReturn = none_of(*newFunction, [](const BasicBlock &BB) {
|
2018-12-06 03:35:37 +08:00
|
|
|
const Instruction *Term = BB.getTerminator();
|
|
|
|
return isa<ReturnInst>(Term) || isa<ResumeInst>(Term);
|
2018-11-09 01:57:09 +08:00
|
|
|
});
|
|
|
|
if (doesNotReturn)
|
|
|
|
newFunction->setDoesNotReturn();
|
|
|
|
|
[CodeExtractor] Store outputs at the first valid insertion point
When CodeExtractor outlines values which are used by the original
function, it must store those values in some in-out parameter. This
store instruction must not be inserted in between a PHI and an EH pad
instruction, as that results in invalid IR.
This fixes the following verifier failure seen while outlining within
ObjC methods with live exit values:
The unwind destination does not have an exception handling instruction!
%call35 = invoke i8* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8* (i8*, i8*)*)(i8* %exn.adjusted, i8* %1)
to label %invoke.cont34 unwind label %lpad33, !dbg !4183
The unwind destination does not have an exception handling instruction!
invoke void @objc_exception_throw(i8* %call35) #12
to label %invoke.cont36 unwind label %lpad33, !dbg !4184
LandingPadInst not the first non-PHI instruction in the block.
%3 = landingpad { i8*, i32 }
catch i8* null, !dbg !1411
rdar://46540815
llvm-svn: 348562
2018-12-07 11:01:54 +08:00
|
|
|
LLVM_DEBUG(if (verifyFunction(*newFunction, &errs())) {
|
|
|
|
newFunction->dump();
|
|
|
|
report_fatal_error("verification of newFunction failed!");
|
|
|
|
});
|
2018-12-04 06:40:21 +08:00
|
|
|
LLVM_DEBUG(if (verifyFunction(*oldFunction))
|
|
|
|
report_fatal_error("verification of oldFunction failed!"));
|
2020-01-29 09:03:39 +08:00
|
|
|
LLVM_DEBUG(if (AC && verifyAssumptionCache(*oldFunction, *newFunction, AC))
|
|
|
|
report_fatal_error("Stale Asumption cache for old Function!"));
|
2004-02-28 11:26:20 +08:00
|
|
|
return newFunction;
|
|
|
|
}
|
2019-10-05 06:46:42 +08:00
|
|
|
|
2020-01-29 09:03:39 +08:00
|
|
|
bool CodeExtractor::verifyAssumptionCache(const Function &OldFunc,
|
|
|
|
const Function &NewFunc,
|
2019-10-05 06:46:42 +08:00
|
|
|
AssumptionCache *AC) {
|
|
|
|
for (auto AssumeVH : AC->assumptions()) {
|
2020-01-29 09:03:39 +08:00
|
|
|
CallInst *I = dyn_cast_or_null<CallInst>(AssumeVH);
|
|
|
|
if (!I)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// There shouldn't be any llvm.assume intrinsics in the new function.
|
|
|
|
if (I->getFunction() != &OldFunc)
|
2019-10-05 06:46:42 +08:00
|
|
|
return true;
|
2020-01-29 09:03:39 +08:00
|
|
|
|
|
|
|
// There shouldn't be any stale affected values in the assumption cache
|
|
|
|
// that were previously in the old function, but that have now been moved
|
|
|
|
// to the new function.
|
|
|
|
for (auto AffectedValVH : AC->assumptionsFor(I->getOperand(0))) {
|
|
|
|
CallInst *AffectedCI = dyn_cast_or_null<CallInst>(AffectedValVH);
|
|
|
|
if (!AffectedCI)
|
|
|
|
continue;
|
|
|
|
if (AffectedCI->getFunction() != &OldFunc)
|
|
|
|
return true;
|
|
|
|
auto *AssumedInst = dyn_cast<Instruction>(AffectedCI->getOperand(0));
|
|
|
|
if (AssumedInst->getFunction() != &OldFunc)
|
|
|
|
return true;
|
|
|
|
}
|
2019-10-05 06:46:42 +08:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|