2009-01-18 20:19:30 +08:00
|
|
|
//===--- CaptureTracking.cpp - Determine whether a pointer is captured ----===//
|
|
|
|
//
|
2019-01-19 16:50:56 +08:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
2009-01-18 20:19:30 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file contains routines that help determine which pointers are captured.
|
|
|
|
// A pointer value is captured if the function makes a copy of any part of the
|
|
|
|
// pointer that outlives the call. Not being captured means, more or less, that
|
|
|
|
// the pointer is only dereferenced and not stored in a global. Returning part
|
|
|
|
// of the pointer as the function return value may or may not count as capturing
|
|
|
|
// the pointer, depending on the context.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2017-06-06 19:49:48 +08:00
|
|
|
#include "llvm/Analysis/CaptureTracking.h"
|
2022-04-06 05:51:53 +08:00
|
|
|
#include "llvm/ADT/SmallPtrSet.h"
|
2012-01-18 06:16:31 +08:00
|
|
|
#include "llvm/ADT/SmallSet.h"
|
|
|
|
#include "llvm/ADT/SmallVector.h"
|
2020-11-07 19:25:55 +08:00
|
|
|
#include "llvm/ADT/Statistic.h"
|
2013-03-10 08:34:01 +08:00
|
|
|
#include "llvm/Analysis/AliasAnalysis.h"
|
2014-07-21 21:15:48 +08:00
|
|
|
#include "llvm/Analysis/CFG.h"
|
2018-05-23 17:16:44 +08:00
|
|
|
#include "llvm/Analysis/ValueTracking.h"
|
2013-03-10 08:34:01 +08:00
|
|
|
#include "llvm/IR/Constants.h"
|
2014-07-21 21:15:48 +08:00
|
|
|
#include "llvm/IR/Dominators.h"
|
2013-03-10 08:34:01 +08:00
|
|
|
#include "llvm/IR/Instructions.h"
|
2016-05-27 01:36:22 +08:00
|
|
|
#include "llvm/IR/IntrinsicInst.h"
|
2020-06-23 19:19:09 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
2013-03-10 08:34:01 +08:00
|
|
|
|
2009-01-18 20:19:30 +08:00
|
|
|
using namespace llvm;
|
|
|
|
|
2020-11-07 19:25:55 +08:00
|
|
|
#define DEBUG_TYPE "capture-tracking"
|
|
|
|
|
|
|
|
STATISTIC(NumCaptured, "Number of pointers maybe captured");
|
|
|
|
STATISTIC(NumNotCaptured, "Number of pointers not captured");
|
|
|
|
STATISTIC(NumCapturedBefore, "Number of pointers maybe captured before");
|
|
|
|
STATISTIC(NumNotCapturedBefore, "Number of pointers not captured before");
|
|
|
|
|
2020-04-22 16:50:04 +08:00
|
|
|
/// The default value for MaxUsesToExplore argument. It's relatively small to
|
|
|
|
/// keep the cost of analysis reasonable for clients like BasicAliasAnalysis,
|
|
|
|
/// where the results can't be cached.
|
|
|
|
/// TODO: we should probably introduce a caching CaptureTracking analysis and
|
|
|
|
/// use it where possible. The caching version can use much higher limit or
|
|
|
|
/// don't have this cap at all.
|
|
|
|
static cl::opt<unsigned>
|
|
|
|
DefaultMaxUsesToExplore("capture-tracking-max-uses-to-explore", cl::Hidden,
|
|
|
|
cl::desc("Maximal number of uses to explore."),
|
|
|
|
cl::init(20));
|
|
|
|
|
|
|
|
unsigned llvm::getDefaultMaxUsesToExploreForCaptureTracking() {
|
|
|
|
return DefaultMaxUsesToExplore;
|
|
|
|
}
|
|
|
|
|
2022-02-07 14:18:35 +08:00
|
|
|
CaptureTracker::~CaptureTracker() = default;
|
2011-11-22 02:32:21 +08:00
|
|
|
|
2014-03-05 18:21:48 +08:00
|
|
|
bool CaptureTracker::shouldExplore(const Use *U) { return true; }
|
2012-10-09 06:12:48 +08:00
|
|
|
|
2019-08-20 05:56:38 +08:00
|
|
|
bool CaptureTracker::isDereferenceableOrNull(Value *O, const DataLayout &DL) {
|
|
|
|
// An inbounds GEP can either be a valid pointer (pointing into
|
|
|
|
// or to the end of an allocation), or be null in the default
|
|
|
|
// address space. So for an inbounds GEP there is no way to let
|
|
|
|
// the pointer escape using clever GEP hacking because doing so
|
|
|
|
// would make the pointer point outside of the allocated object
|
|
|
|
// and thus make the GEP result a poison value. Similarly, other
|
|
|
|
// dereferenceable pointers cannot be manipulated without producing
|
|
|
|
// poison.
|
|
|
|
if (auto *GEP = dyn_cast<GetElementPtrInst>(O))
|
|
|
|
if (GEP->isInBounds())
|
|
|
|
return true;
|
2021-03-20 02:15:29 +08:00
|
|
|
bool CanBeNull, CanBeFreed;
|
|
|
|
return O->getPointerDereferenceableBytes(DL, CanBeNull, CanBeFreed);
|
2019-08-20 05:56:38 +08:00
|
|
|
}
|
|
|
|
|
2011-11-15 06:49:42 +08:00
|
|
|
namespace {
|
2011-11-21 03:37:06 +08:00
|
|
|
struct SimpleCaptureTracker : public CaptureTracker {
|
2022-04-06 05:51:53 +08:00
|
|
|
explicit SimpleCaptureTracker(
|
|
|
|
|
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues, bool ReturnCaptures)
|
|
|
|
: EphValues(EphValues), ReturnCaptures(ReturnCaptures) {}
|
2011-11-15 06:49:42 +08:00
|
|
|
|
2014-03-05 15:30:04 +08:00
|
|
|
void tooManyUses() override { Captured = true; }
|
2011-11-15 06:49:42 +08:00
|
|
|
|
2014-03-05 18:21:48 +08:00
|
|
|
bool captured(const Use *U) override {
|
2011-12-29 07:24:21 +08:00
|
|
|
if (isa<ReturnInst>(U->getUser()) && !ReturnCaptures)
|
2012-05-11 07:38:07 +08:00
|
|
|
return false;
|
2011-11-15 06:49:42 +08:00
|
|
|
|
2022-04-06 05:51:53 +08:00
|
|
|
if (EphValues.contains(U->getUser()))
|
|
|
|
return false;
|
|
|
|
|
2011-11-15 06:49:42 +08:00
|
|
|
Captured = true;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2022-04-06 05:51:53 +08:00
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues;
|
|
|
|
|
2011-11-15 06:49:42 +08:00
|
|
|
bool ReturnCaptures;
|
|
|
|
|
2022-01-24 12:32:56 +08:00
|
|
|
bool Captured = false;
|
2011-11-15 06:49:42 +08:00
|
|
|
};
|
2014-07-21 21:15:48 +08:00
|
|
|
|
|
|
|
/// Only find pointer captures which happen before the given instruction. Uses
|
|
|
|
/// the dominator tree to determine whether one instruction is before another.
|
|
|
|
/// Only support the case where the Value is defined in the same basic block
|
|
|
|
/// as the given instruction and the use.
|
|
|
|
struct CapturesBefore : public CaptureTracker {
|
2015-06-25 01:53:17 +08:00
|
|
|
|
2021-09-20 16:07:34 +08:00
|
|
|
CapturesBefore(bool ReturnCaptures, const Instruction *I,
|
|
|
|
const DominatorTree *DT, bool IncludeI, const LoopInfo *LI)
|
|
|
|
: BeforeHere(I), DT(DT), ReturnCaptures(ReturnCaptures),
|
2022-01-24 12:32:56 +08:00
|
|
|
IncludeI(IncludeI), LI(LI) {}
|
2014-07-21 21:15:48 +08:00
|
|
|
|
|
|
|
void tooManyUses() override { Captured = true; }
|
|
|
|
|
2015-06-25 01:53:17 +08:00
|
|
|
bool isSafeToPrune(Instruction *I) {
|
2021-05-16 01:05:08 +08:00
|
|
|
if (BeforeHere == I)
|
|
|
|
return !IncludeI;
|
2021-05-15 17:57:01 +08:00
|
|
|
|
2014-07-21 21:15:48 +08:00
|
|
|
// We explore this usage only if the usage can reach "BeforeHere".
|
|
|
|
// If use is not reachable from entry, there is no need to explore.
|
2021-05-16 22:04:10 +08:00
|
|
|
if (!DT->isReachableFromEntry(I->getParent()))
|
2015-06-25 01:53:17 +08:00
|
|
|
return true;
|
|
|
|
|
2020-11-06 01:26:36 +08:00
|
|
|
// Check whether there is a path from I to BeforeHere.
|
2021-09-20 16:07:34 +08:00
|
|
|
return !isPotentiallyReachable(I, BeforeHere, nullptr, DT, LI);
|
2015-06-25 01:53:17 +08:00
|
|
|
}
|
|
|
|
|
2021-05-16 01:05:08 +08:00
|
|
|
bool captured(const Use *U) override {
|
2015-06-25 01:53:17 +08:00
|
|
|
Instruction *I = cast<Instruction>(U->getUser());
|
2021-05-16 01:05:08 +08:00
|
|
|
if (isa<ReturnInst>(I) && !ReturnCaptures)
|
|
|
|
return false;
|
2015-06-25 01:53:17 +08:00
|
|
|
|
2021-05-16 01:05:08 +08:00
|
|
|
// Check isSafeToPrune() here rather than in shouldExplore() to avoid
|
|
|
|
// an expensive reachability query for every instruction we look at.
|
|
|
|
// Instead we only do one for actual capturing candidates.
|
|
|
|
if (isSafeToPrune(I))
|
2014-07-21 21:15:48 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
Captured = true;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
const Instruction *BeforeHere;
|
2018-04-25 05:12:45 +08:00
|
|
|
const DominatorTree *DT;
|
2014-07-21 21:15:48 +08:00
|
|
|
|
|
|
|
bool ReturnCaptures;
|
2014-07-22 05:30:22 +08:00
|
|
|
bool IncludeI;
|
2014-07-21 21:15:48 +08:00
|
|
|
|
2022-01-24 12:32:56 +08:00
|
|
|
bool Captured = false;
|
2021-09-20 16:07:34 +08:00
|
|
|
|
|
|
|
const LoopInfo *LI;
|
2014-07-21 21:15:48 +08:00
|
|
|
};
|
2021-09-24 22:41:47 +08:00
|
|
|
|
|
|
|
/// Find the 'earliest' instruction before which the pointer is known not to
|
|
|
|
/// be captured. Here an instruction A is considered earlier than instruction
|
|
|
|
/// B, if A dominates B. If 2 escapes do not dominate each other, the
|
|
|
|
/// terminator of the common dominator is chosen. If not all uses cannot be
|
|
|
|
/// analyzed, the earliest escape is set to the first instruction in the
|
|
|
|
/// function entry block.
|
|
|
|
// NOTE: Users have to make sure instructions compared against the earliest
|
|
|
|
// escape are not in a cycle.
|
|
|
|
struct EarliestCaptures : public CaptureTracker {
|
|
|
|
|
2022-04-08 06:39:53 +08:00
|
|
|
EarliestCaptures(bool ReturnCaptures, Function &F, const DominatorTree &DT,
|
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues)
|
|
|
|
: EphValues(EphValues), DT(DT), ReturnCaptures(ReturnCaptures), F(F) {}
|
2021-09-24 22:41:47 +08:00
|
|
|
|
|
|
|
void tooManyUses() override {
|
|
|
|
Captured = true;
|
|
|
|
EarliestCapture = &*F.getEntryBlock().begin();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool captured(const Use *U) override {
|
|
|
|
Instruction *I = cast<Instruction>(U->getUser());
|
|
|
|
if (isa<ReturnInst>(I) && !ReturnCaptures)
|
|
|
|
return false;
|
|
|
|
|
2022-04-08 06:39:53 +08:00
|
|
|
if (EphValues.contains(I))
|
|
|
|
return false;
|
|
|
|
|
2021-09-24 22:41:47 +08:00
|
|
|
if (!EarliestCapture) {
|
|
|
|
EarliestCapture = I;
|
|
|
|
} else if (EarliestCapture->getParent() == I->getParent()) {
|
|
|
|
if (I->comesBefore(EarliestCapture))
|
|
|
|
EarliestCapture = I;
|
|
|
|
} else {
|
|
|
|
BasicBlock *CurrentBB = I->getParent();
|
|
|
|
BasicBlock *EarliestBB = EarliestCapture->getParent();
|
|
|
|
if (DT.dominates(EarliestBB, CurrentBB)) {
|
|
|
|
// EarliestCapture already comes before the current use.
|
|
|
|
} else if (DT.dominates(CurrentBB, EarliestBB)) {
|
|
|
|
EarliestCapture = I;
|
|
|
|
} else {
|
|
|
|
// Otherwise find the nearest common dominator and use its terminator.
|
|
|
|
auto *NearestCommonDom =
|
|
|
|
DT.findNearestCommonDominator(CurrentBB, EarliestBB);
|
|
|
|
EarliestCapture = NearestCommonDom->getTerminator();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Captured = true;
|
|
|
|
|
|
|
|
// Return false to continue analysis; we need to see all potential
|
|
|
|
// captures.
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2022-04-08 06:39:53 +08:00
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues;
|
|
|
|
|
2021-09-24 22:41:47 +08:00
|
|
|
Instruction *EarliestCapture = nullptr;
|
|
|
|
|
|
|
|
const DominatorTree &DT;
|
|
|
|
|
|
|
|
bool ReturnCaptures;
|
|
|
|
|
2022-01-24 12:32:56 +08:00
|
|
|
bool Captured = false;
|
2021-09-24 22:41:47 +08:00
|
|
|
|
|
|
|
Function &F;
|
|
|
|
};
|
2015-06-23 17:49:53 +08:00
|
|
|
}
|
2009-12-09 07:59:12 +08:00
|
|
|
|
2009-01-18 20:19:30 +08:00
|
|
|
/// PointerMayBeCaptured - Return true if this pointer value may be captured
|
|
|
|
/// by the enclosing function (which is required to exist). This routine can
|
|
|
|
/// be expensive, so consider caching the results. The boolean ReturnCaptures
|
|
|
|
/// specifies whether returning the value (or part of it) from the function
|
2009-11-20 05:57:48 +08:00
|
|
|
/// counts as capturing it or not. The boolean StoreCaptures specified whether
|
|
|
|
/// storing the value (or part of it) into memory anywhere automatically
|
2009-01-18 20:19:30 +08:00
|
|
|
/// counts as capturing it or not.
|
2022-04-06 05:51:53 +08:00
|
|
|
bool llvm::PointerMayBeCaptured(const Value *V, bool ReturnCaptures,
|
|
|
|
bool StoreCaptures, unsigned MaxUsesToExplore) {
|
|
|
|
SmallPtrSet<const Value *, 1> Empty;
|
|
|
|
return PointerMayBeCaptured(V, ReturnCaptures, StoreCaptures, Empty,
|
|
|
|
MaxUsesToExplore);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Variant of the above function which accepts a set of Values that are
|
|
|
|
/// ephemeral and cannot cause pointers to escape.
|
|
|
|
bool llvm::PointerMayBeCaptured(const Value *V, bool ReturnCaptures,
|
|
|
|
bool StoreCaptures,
|
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues,
|
2018-11-30 04:08:12 +08:00
|
|
|
unsigned MaxUsesToExplore) {
|
2011-11-22 03:42:56 +08:00
|
|
|
assert(!isa<GlobalValue>(V) &&
|
|
|
|
"It doesn't make sense to ask whether a global is captured.");
|
|
|
|
|
2011-11-15 06:49:42 +08:00
|
|
|
// TODO: If StoreCaptures is not true, we could do Fancy analysis
|
|
|
|
// to determine whether this store is not actually an escape point.
|
|
|
|
// In that case, BasicAliasAnalysis should be updated as well to
|
|
|
|
// take advantage of this.
|
|
|
|
(void)StoreCaptures;
|
|
|
|
|
2022-04-06 05:51:53 +08:00
|
|
|
SimpleCaptureTracker SCT(EphValues, ReturnCaptures);
|
2018-12-18 11:32:33 +08:00
|
|
|
PointerMayBeCaptured(V, &SCT, MaxUsesToExplore);
|
2020-11-07 19:25:55 +08:00
|
|
|
if (SCT.Captured)
|
|
|
|
++NumCaptured;
|
|
|
|
else
|
|
|
|
++NumNotCaptured;
|
2011-11-15 06:49:42 +08:00
|
|
|
return SCT.Captured;
|
2009-01-18 20:19:30 +08:00
|
|
|
}
|
2011-11-21 03:37:06 +08:00
|
|
|
|
2014-07-21 21:15:48 +08:00
|
|
|
/// PointerMayBeCapturedBefore - Return true if this pointer value may be
|
|
|
|
/// captured by the enclosing function (which is required to exist). If a
|
|
|
|
/// DominatorTree is provided, only captures which happen before the given
|
|
|
|
/// instruction are considered. This routine can be expensive, so consider
|
|
|
|
/// caching the results. The boolean ReturnCaptures specifies whether
|
|
|
|
/// returning the value (or part of it) from the function counts as capturing
|
|
|
|
/// it or not. The boolean StoreCaptures specified whether storing the value
|
|
|
|
/// (or part of it) into memory anywhere automatically counts as capturing it
|
[IR] Lazily number instructions for local dominance queries
Essentially, fold OrderedBasicBlock into BasicBlock, and make it
auto-invalidate the instruction ordering when new instructions are
added. Notably, we don't need to invalidate it when removing
instructions, which is helpful when a pass mostly delete dead
instructions rather than transforming them.
The downside is that Instruction grows from 56 bytes to 64 bytes. The
resulting LLVM code is substantially simpler and automatically handles
invalidation, which makes me think that this is the right speed and size
tradeoff.
The important change is in SymbolTableTraitsImpl.h, where the numbering
is invalidated. Everything else should be straightforward.
We probably want to implement a fancier re-numbering scheme so that
local updates don't invalidate the ordering, but I plan for that to be
future work, maybe for someone else.
Reviewed By: lattner, vsk, fhahn, dexonsmith
Differential Revision: https://reviews.llvm.org/D51664
2020-02-19 06:33:54 +08:00
|
|
|
/// or not.
|
2014-07-21 21:15:48 +08:00
|
|
|
bool llvm::PointerMayBeCapturedBefore(const Value *V, bool ReturnCaptures,
|
|
|
|
bool StoreCaptures, const Instruction *I,
|
2018-04-25 05:12:45 +08:00
|
|
|
const DominatorTree *DT, bool IncludeI,
|
2021-09-20 16:07:34 +08:00
|
|
|
unsigned MaxUsesToExplore,
|
|
|
|
const LoopInfo *LI) {
|
2014-07-21 21:15:48 +08:00
|
|
|
assert(!isa<GlobalValue>(V) &&
|
|
|
|
"It doesn't make sense to ask whether a global is captured.");
|
|
|
|
|
|
|
|
if (!DT)
|
2018-12-18 11:32:33 +08:00
|
|
|
return PointerMayBeCaptured(V, ReturnCaptures, StoreCaptures,
|
|
|
|
MaxUsesToExplore);
|
2014-07-21 21:15:48 +08:00
|
|
|
|
|
|
|
// TODO: See comment in PointerMayBeCaptured regarding what could be done
|
|
|
|
// with StoreCaptures.
|
|
|
|
|
2021-09-20 16:07:34 +08:00
|
|
|
CapturesBefore CB(ReturnCaptures, I, DT, IncludeI, LI);
|
2018-12-18 11:32:33 +08:00
|
|
|
PointerMayBeCaptured(V, &CB, MaxUsesToExplore);
|
2020-11-07 19:25:55 +08:00
|
|
|
if (CB.Captured)
|
|
|
|
++NumCapturedBefore;
|
|
|
|
else
|
|
|
|
++NumNotCapturedBefore;
|
2014-07-21 21:15:48 +08:00
|
|
|
return CB.Captured;
|
|
|
|
}
|
|
|
|
|
2022-04-08 06:39:53 +08:00
|
|
|
Instruction *
|
|
|
|
llvm::FindEarliestCapture(const Value *V, Function &F, bool ReturnCaptures,
|
|
|
|
bool StoreCaptures, const DominatorTree &DT,
|
|
|
|
|
|
|
|
const SmallPtrSetImpl<const Value *> &EphValues,
|
|
|
|
unsigned MaxUsesToExplore) {
|
2021-09-24 22:41:47 +08:00
|
|
|
assert(!isa<GlobalValue>(V) &&
|
|
|
|
"It doesn't make sense to ask whether a global is captured.");
|
|
|
|
|
2022-04-08 06:39:53 +08:00
|
|
|
EarliestCaptures CB(ReturnCaptures, F, DT, EphValues);
|
2021-09-24 22:41:47 +08:00
|
|
|
PointerMayBeCaptured(V, &CB, MaxUsesToExplore);
|
|
|
|
if (CB.Captured)
|
|
|
|
++NumCapturedBefore;
|
|
|
|
else
|
|
|
|
++NumNotCapturedBefore;
|
|
|
|
return CB.EarliestCapture;
|
|
|
|
}
|
|
|
|
|
2022-03-09 05:02:36 +08:00
|
|
|
UseCaptureKind llvm::DetermineUseCaptureKind(
|
|
|
|
const Use &U,
|
|
|
|
function_ref<bool(Value *, const DataLayout &)> IsDereferenceableOrNull) {
|
|
|
|
Instruction *I = cast<Instruction>(U.getUser());
|
|
|
|
|
|
|
|
switch (I->getOpcode()) {
|
|
|
|
case Instruction::Call:
|
|
|
|
case Instruction::Invoke: {
|
|
|
|
auto *Call = cast<CallBase>(I);
|
|
|
|
// Not captured if the callee is readonly, doesn't return a copy through
|
|
|
|
// its return value and doesn't unwind (a readonly function can leak bits
|
|
|
|
// by throwing an exception or not depending on the input value).
|
|
|
|
if (Call->onlyReadsMemory() && Call->doesNotThrow() &&
|
|
|
|
Call->getType()->isVoidTy())
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
|
|
|
|
// The pointer is not captured if returned pointer is not captured.
|
|
|
|
// NOTE: CaptureTracking users should not assume that only functions
|
|
|
|
// marked with nocapture do not capture. This means that places like
|
|
|
|
// getUnderlyingObject in ValueTracking or DecomposeGEPExpression
|
|
|
|
// in BasicAA also need to know about this property.
|
|
|
|
if (isIntrinsicReturningPointerAliasingArgumentWithoutCapturing(Call, true))
|
|
|
|
return UseCaptureKind::PASSTHROUGH;
|
|
|
|
|
|
|
|
// Volatile operations effectively capture the memory location that they
|
|
|
|
// load and store to.
|
|
|
|
if (auto *MI = dyn_cast<MemIntrinsic>(Call))
|
|
|
|
if (MI->isVolatile())
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
|
|
|
|
// Calling a function pointer does not in itself cause the pointer to
|
|
|
|
// be captured. This is a subtle point considering that (for example)
|
|
|
|
// the callee might return its own address. It is analogous to saying
|
|
|
|
// that loading a value from a pointer does not cause the pointer to be
|
|
|
|
// captured, even though the loaded value might be the pointer itself
|
|
|
|
// (think of self-referential objects).
|
|
|
|
if (Call->isCallee(&U))
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
|
|
|
|
// Not captured if only passed via 'nocapture' arguments.
|
|
|
|
if (Call->isDataOperand(&U) &&
|
|
|
|
!Call->doesNotCapture(Call->getDataOperandNo(&U))) {
|
|
|
|
// The parameter is not marked 'nocapture' - captured.
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
}
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
}
|
|
|
|
case Instruction::Load:
|
|
|
|
// Volatile loads make the address observable.
|
|
|
|
if (cast<LoadInst>(I)->isVolatile())
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
case Instruction::VAArg:
|
|
|
|
// "va-arg" from a pointer does not cause it to be captured.
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
case Instruction::Store:
|
|
|
|
// Stored the pointer - conservatively assume it may be captured.
|
|
|
|
// Volatile stores make the address observable.
|
|
|
|
if (U.getOperandNo() == 0 || cast<StoreInst>(I)->isVolatile())
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
case Instruction::AtomicRMW: {
|
|
|
|
// atomicrmw conceptually includes both a load and store from
|
|
|
|
// the same location.
|
|
|
|
// As with a store, the location being accessed is not captured,
|
|
|
|
// but the value being stored is.
|
|
|
|
// Volatile stores make the address observable.
|
|
|
|
auto *ARMWI = cast<AtomicRMWInst>(I);
|
|
|
|
if (U.getOperandNo() == 1 || ARMWI->isVolatile())
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
}
|
|
|
|
case Instruction::AtomicCmpXchg: {
|
|
|
|
// cmpxchg conceptually includes both a load and store from
|
|
|
|
// the same location.
|
|
|
|
// As with a store, the location being accessed is not captured,
|
|
|
|
// but the value being stored is.
|
|
|
|
// Volatile stores make the address observable.
|
|
|
|
auto *ACXI = cast<AtomicCmpXchgInst>(I);
|
|
|
|
if (U.getOperandNo() == 1 || U.getOperandNo() == 2 || ACXI->isVolatile())
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
}
|
|
|
|
case Instruction::BitCast:
|
|
|
|
case Instruction::GetElementPtr:
|
|
|
|
case Instruction::PHI:
|
|
|
|
case Instruction::Select:
|
|
|
|
case Instruction::AddrSpaceCast:
|
|
|
|
// The original value is not captured via this if the new value isn't.
|
|
|
|
return UseCaptureKind::PASSTHROUGH;
|
|
|
|
case Instruction::ICmp: {
|
|
|
|
unsigned Idx = U.getOperandNo();
|
|
|
|
unsigned OtherIdx = 1 - Idx;
|
|
|
|
if (auto *CPN = dyn_cast<ConstantPointerNull>(I->getOperand(OtherIdx))) {
|
|
|
|
// Don't count comparisons of a no-alias return value against null as
|
|
|
|
// captures. This allows us to ignore comparisons of malloc results
|
|
|
|
// with null, for example.
|
|
|
|
if (CPN->getType()->getAddressSpace() == 0)
|
|
|
|
if (isNoAliasCall(U.get()->stripPointerCasts()))
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
if (!I->getFunction()->nullPointerIsDefined()) {
|
|
|
|
auto *O = I->getOperand(Idx)->stripPointerCastsSameRepresentation();
|
|
|
|
// Comparing a dereferenceable_or_null pointer against null cannot
|
|
|
|
// lead to pointer escapes, because if it is not null it must be a
|
|
|
|
// valid (in-bounds) pointer.
|
|
|
|
const DataLayout &DL = I->getModule()->getDataLayout();
|
|
|
|
if (IsDereferenceableOrNull && IsDereferenceableOrNull(O, DL))
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Comparison against value stored in global variable. Given the pointer
|
|
|
|
// does not escape, its value cannot be guessed and stored separately in a
|
|
|
|
// global variable.
|
|
|
|
auto *LI = dyn_cast<LoadInst>(I->getOperand(OtherIdx));
|
|
|
|
if (LI && isa<GlobalVariable>(LI->getPointerOperand()))
|
|
|
|
return UseCaptureKind::NO_CAPTURE;
|
|
|
|
// Otherwise, be conservative. There are crazy ways to capture pointers
|
|
|
|
// using comparisons.
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
// Something else - be conservative and say it is captured.
|
|
|
|
return UseCaptureKind::MAY_CAPTURE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-11-30 04:08:12 +08:00
|
|
|
void llvm::PointerMayBeCaptured(const Value *V, CaptureTracker *Tracker,
|
|
|
|
unsigned MaxUsesToExplore) {
|
2011-11-21 03:37:06 +08:00
|
|
|
assert(V->getType()->isPointerTy() && "Capture is for pointers only!");
|
2020-04-24 00:17:51 +08:00
|
|
|
if (MaxUsesToExplore == 0)
|
|
|
|
MaxUsesToExplore = DefaultMaxUsesToExplore;
|
|
|
|
|
2020-04-22 16:50:04 +08:00
|
|
|
SmallVector<const Use *, 20> Worklist;
|
|
|
|
Worklist.reserve(getDefaultMaxUsesToExploreForCaptureTracking());
|
|
|
|
SmallSet<const Use *, 20> Visited;
|
2011-11-21 03:37:06 +08:00
|
|
|
|
2018-05-05 18:23:27 +08:00
|
|
|
auto AddUses = [&](const Value *V) {
|
2018-11-29 10:15:35 +08:00
|
|
|
unsigned Count = 0;
|
2018-05-05 18:23:27 +08:00
|
|
|
for (const Use &U : V->uses()) {
|
|
|
|
// If there are lots of uses, conservatively say that the value
|
|
|
|
// is captured to avoid taking too much compile time.
|
2020-11-07 18:47:31 +08:00
|
|
|
if (Count++ >= MaxUsesToExplore) {
|
|
|
|
Tracker->tooManyUses();
|
|
|
|
return false;
|
|
|
|
}
|
2018-05-05 18:23:27 +08:00
|
|
|
if (!Visited.insert(&U).second)
|
|
|
|
continue;
|
|
|
|
if (!Tracker->shouldExplore(&U))
|
|
|
|
continue;
|
|
|
|
Worklist.push_back(&U);
|
|
|
|
}
|
2020-11-07 18:47:31 +08:00
|
|
|
return true;
|
2018-05-05 18:23:27 +08:00
|
|
|
};
|
2020-11-07 18:47:31 +08:00
|
|
|
if (!AddUses(V))
|
|
|
|
return;
|
2011-11-21 03:37:06 +08:00
|
|
|
|
2022-03-09 05:02:36 +08:00
|
|
|
auto IsDereferenceableOrNull = [Tracker](Value *V, const DataLayout &DL) {
|
|
|
|
return Tracker->isDereferenceableOrNull(V, DL);
|
|
|
|
};
|
2011-11-21 03:37:06 +08:00
|
|
|
while (!Worklist.empty()) {
|
2014-03-05 18:21:48 +08:00
|
|
|
const Use *U = Worklist.pop_back_val();
|
2022-03-09 05:02:36 +08:00
|
|
|
switch (DetermineUseCaptureKind(*U, IsDereferenceableOrNull)) {
|
|
|
|
case UseCaptureKind::NO_CAPTURE:
|
|
|
|
continue;
|
|
|
|
case UseCaptureKind::MAY_CAPTURE:
|
2011-12-29 07:24:21 +08:00
|
|
|
if (Tracker->captured(U))
|
2011-11-21 03:37:06 +08:00
|
|
|
return;
|
2022-03-09 05:02:36 +08:00
|
|
|
continue;
|
|
|
|
case UseCaptureKind::PASSTHROUGH:
|
|
|
|
if (!AddUses(U->getUser()))
|
2011-11-21 03:37:06 +08:00
|
|
|
return;
|
2022-03-09 05:02:36 +08:00
|
|
|
continue;
|
2011-11-21 03:37:06 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// All uses examined.
|
|
|
|
}
|
2020-10-16 02:37:29 +08:00
|
|
|
|
|
|
|
bool llvm::isNonEscapingLocalObject(
|
|
|
|
const Value *V, SmallDenseMap<const Value *, bool, 8> *IsCapturedCache) {
|
|
|
|
SmallDenseMap<const Value *, bool, 8>::iterator CacheIt;
|
|
|
|
if (IsCapturedCache) {
|
|
|
|
bool Inserted;
|
|
|
|
std::tie(CacheIt, Inserted) = IsCapturedCache->insert({V, false});
|
|
|
|
if (!Inserted)
|
|
|
|
// Found cached result, return it!
|
|
|
|
return CacheIt->second;
|
|
|
|
}
|
|
|
|
|
2021-05-14 05:03:46 +08:00
|
|
|
// If this is an identified function-local object, check to see if it escapes.
|
|
|
|
if (isIdentifiedFunctionLocal(V)) {
|
2020-10-16 02:37:29 +08:00
|
|
|
// Set StoreCaptures to True so that we can assume in our callers that the
|
|
|
|
// pointer is not the result of a load instruction. Currently
|
|
|
|
// PointerMayBeCaptured doesn't have any special analysis for the
|
|
|
|
// StoreCaptures=false case; if it did, our callers could be refined to be
|
|
|
|
// more precise.
|
|
|
|
auto Ret = !PointerMayBeCaptured(V, false, /*StoreCaptures=*/true);
|
|
|
|
if (IsCapturedCache)
|
|
|
|
CacheIt->second = Ret;
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|