[DSE] Make sure that DSE+MSSA can handle masked stores

Differential Revision: https://reviews.llvm.org/D87414
This commit is contained in:
Krzysztof Parzyszek 2020-09-09 15:23:34 -05:00
parent 6aa3fc4a5b
commit f92908cc74
2 changed files with 58 additions and 43 deletions

View File

@ -411,22 +411,53 @@ enum OverwriteResult {
} // end anonymous namespace } // end anonymous namespace
/// Return 'OW_Complete' if a store to the 'Later' location completely /// Check if two instruction are masked stores that completely
/// overwrites a store to the 'Earlier' location. Return OW_MaybePartial /// overwrite one another. More specifically, \p Later has to
/// if \p Later does not completely overwrite \p Earlier, but they both /// overwrite \p Earlier.
/// write to the same underlying object. In that case, use isPartialOverwrite to template <typename AATy>
/// check if \p Later partially overwrites \p Earlier. Returns 'OW_Unknown' if static OverwriteResult isMaskedStoreOverwrite(const Instruction *Later,
/// nothing can be determined. const Instruction *Earlier,
AATy &AA) {
const auto *IIL = dyn_cast<IntrinsicInst>(Later);
const auto *IIE = dyn_cast<IntrinsicInst>(Earlier);
if (IIL == nullptr || IIE == nullptr)
return OW_Unknown;
if (IIL->getIntrinsicID() != Intrinsic::masked_store ||
IIE->getIntrinsicID() != Intrinsic::masked_store)
return OW_Unknown;
// Pointers.
Value *LP = IIL->getArgOperand(1)->stripPointerCasts();
Value *EP = IIE->getArgOperand(1)->stripPointerCasts();
if (LP != EP && !AA.isMustAlias(LP, EP))
return OW_Unknown;
// Masks.
// TODO: check that Later's mask is a superset of the Earlier's mask.
if (IIL->getArgOperand(3) != IIE->getArgOperand(3))
return OW_Unknown;
return OW_Complete;
}
/// Return 'OW_Complete' if a store to the 'Later' location (by \p LaterI
/// instruction) completely overwrites a store to the 'Earlier' location.
/// (by \p EarlierI instruction).
/// Return OW_MaybePartial if \p Later does not completely overwrite
/// \p Earlier, but they both write to the same underlying object. In that
/// case, use isPartialOverwrite to check if \p Later partially overwrites
/// \p Earlier. Returns 'OW_Unknown' if nothing can be determined.
template <typename AATy> template <typename AATy>
static OverwriteResult static OverwriteResult
isOverwrite(const MemoryLocation &Later, const MemoryLocation &Earlier, isOverwrite(const Instruction *LaterI, const Instruction *EarlierI,
const MemoryLocation &Later, const MemoryLocation &Earlier,
const DataLayout &DL, const TargetLibraryInfo &TLI, const DataLayout &DL, const TargetLibraryInfo &TLI,
int64_t &EarlierOff, int64_t &LaterOff, AATy &AA, int64_t &EarlierOff, int64_t &LaterOff, AATy &AA,
const Function *F) { const Function *F) {
// FIXME: Vet that this works for size upper-bounds. Seems unlikely that we'll // FIXME: Vet that this works for size upper-bounds. Seems unlikely that we'll
// get imprecise values here, though (except for unknown sizes). // get imprecise values here, though (except for unknown sizes).
if (!Later.Size.isPrecise() || !Earlier.Size.isPrecise()) if (!Later.Size.isPrecise() || !Earlier.Size.isPrecise()) {
return OW_Unknown; // Masked stores have imprecise locations, but we can reason about them
// to some extent.
return isMaskedStoreOverwrite(LaterI, EarlierI, AA);
}
const uint64_t LaterSize = Later.Size.getValue(); const uint64_t LaterSize = Later.Size.getValue();
const uint64_t EarlierSize = Earlier.Size.getValue(); const uint64_t EarlierSize = Earlier.Size.getValue();
@ -494,24 +525,6 @@ isOverwrite(const MemoryLocation &Later, const MemoryLocation &Earlier,
return OW_MaybePartial; return OW_MaybePartial;
} }
static OverwriteResult isMaskedStoreOverwrite(Instruction *Later,
Instruction *Earlier) {
auto *IIL = dyn_cast<IntrinsicInst>(Later);
auto *IIE = dyn_cast<IntrinsicInst>(Earlier);
if (IIL == nullptr || IIE == nullptr)
return OW_Unknown;
if (IIL->getIntrinsicID() != Intrinsic::masked_store ||
IIE->getIntrinsicID() != Intrinsic::masked_store)
return OW_Unknown;
// Pointers.
if (IIL->getArgOperand(1) != IIE->getArgOperand(1))
return OW_Unknown;
// Masks.
if (IIL->getArgOperand(3) != IIE->getArgOperand(3))
return OW_Unknown;
return OW_Complete;
}
/// Return 'OW_Complete' if a store to the 'Later' location completely /// Return 'OW_Complete' if a store to the 'Later' location completely
/// overwrites a store to the 'Earlier' location, 'OW_End' if the end of the /// overwrites a store to the 'Earlier' location, 'OW_End' if the end of the
/// 'Earlier' location is completely overwritten by 'Later', 'OW_Begin' if the /// 'Earlier' location is completely overwritten by 'Later', 'OW_Begin' if the
@ -1376,13 +1389,9 @@ static bool eliminateDeadStores(BasicBlock &BB, AliasAnalysis *AA,
if (isRemovable(DepWrite) && if (isRemovable(DepWrite) &&
!isPossibleSelfRead(Inst, Loc, DepWrite, *TLI, *AA)) { !isPossibleSelfRead(Inst, Loc, DepWrite, *TLI, *AA)) {
int64_t InstWriteOffset, DepWriteOffset; int64_t InstWriteOffset, DepWriteOffset;
OverwriteResult OR = isOverwrite(Loc, DepLoc, DL, *TLI, DepWriteOffset, OverwriteResult OR = isOverwrite(Inst, DepWrite, Loc, DepLoc, DL, *TLI,
InstWriteOffset, *AA, BB.getParent()); DepWriteOffset, InstWriteOffset, *AA,
if (OR == OW_Unknown) { BB.getParent());
// isOverwrite punts on MemoryLocations with an imprecise size, such
// as masked stores. Handle this here, somwewhat inelegantly.
OR = isMaskedStoreOverwrite(Inst, DepWrite);
}
if (OR == OW_MaybePartial) if (OR == OW_MaybePartial)
OR = isPartialOverwrite(Loc, DepLoc, DepWriteOffset, InstWriteOffset, OR = isPartialOverwrite(Loc, DepLoc, DepWriteOffset, InstWriteOffset,
DepWrite, IOL); DepWrite, IOL);
@ -1707,6 +1716,8 @@ struct DSEState {
switch (CB->getIntrinsicID()) { switch (CB->getIntrinsicID()) {
case Intrinsic::init_trampoline: case Intrinsic::init_trampoline:
return {MemoryLocation(CB->getArgOperand(0))}; return {MemoryLocation(CB->getArgOperand(0))};
case Intrinsic::masked_store:
return {MemoryLocation::getForArgument(CB, 1, TLI)};
default: default:
break; break;
} }
@ -1716,8 +1727,10 @@ struct DSEState {
return MemoryLocation::getOrNone(I); return MemoryLocation::getOrNone(I);
} }
/// Returns true if \p Use completely overwrites \p DefLoc. /// Returns true if \p UseInst completely overwrites \p DefLoc
bool isCompleteOverwrite(MemoryLocation DefLoc, Instruction *UseInst) { /// (stored by \p DefInst).
bool isCompleteOverwrite(MemoryLocation DefLoc, Instruction *DefInst,
Instruction *UseInst) {
// UseInst has a MemoryDef associated in MemorySSA. It's possible for a // UseInst has a MemoryDef associated in MemorySSA. It's possible for a
// MemoryDef to not write to memory, e.g. a volatile load is modeled as a // MemoryDef to not write to memory, e.g. a volatile load is modeled as a
// MemoryDef. // MemoryDef.
@ -1729,9 +1742,10 @@ struct DSEState {
return false; return false;
int64_t InstWriteOffset, DepWriteOffset; int64_t InstWriteOffset, DepWriteOffset;
auto CC = getLocForWriteEx(UseInst); if (auto CC = getLocForWriteEx(UseInst))
return CC && isOverwrite(*CC, DefLoc, DL, TLI, DepWriteOffset, return isOverwrite(UseInst, DefInst, *CC, DefLoc, DL, TLI, DepWriteOffset,
InstWriteOffset, BatchAA, &F) == OW_Complete; InstWriteOffset, BatchAA, &F) == OW_Complete;
return false;
} }
/// Returns true if \p Def is not read before returning from the function. /// Returns true if \p Def is not read before returning from the function.
@ -1977,8 +1991,8 @@ struct DSEState {
continue; continue;
} else { } else {
int64_t InstWriteOffset, DepWriteOffset; int64_t InstWriteOffset, DepWriteOffset;
auto OR = isOverwrite(DefLoc, *CurrentLoc, DL, TLI, DepWriteOffset, auto OR = isOverwrite(KillingI, CurrentI, DefLoc, *CurrentLoc, DL, TLI,
InstWriteOffset, BatchAA, &F); DepWriteOffset, InstWriteOffset, BatchAA, &F);
// If Current does not write to the same object as KillingDef, check // If Current does not write to the same object as KillingDef, check
// the next candidate. // the next candidate.
if (OR == OW_Unknown) { if (OR == OW_Unknown) {
@ -2122,7 +2136,7 @@ struct DSEState {
// 3 = Def(1) ; <---- Current (3, 2) = NoAlias, (3,1) = MayAlias, // 3 = Def(1) ; <---- Current (3, 2) = NoAlias, (3,1) = MayAlias,
// stores [0,1] // stores [0,1]
if (MemoryDef *UseDef = dyn_cast<MemoryDef>(UseAccess)) { if (MemoryDef *UseDef = dyn_cast<MemoryDef>(UseAccess)) {
if (isCompleteOverwrite(DefLoc, UseInst)) { if (isCompleteOverwrite(DefLoc, KillingI, UseInst)) {
if (!isInvisibleToCallerAfterRet(DefUO) && if (!isInvisibleToCallerAfterRet(DefUO) &&
UseAccess != EarlierAccess) { UseAccess != EarlierAccess) {
BasicBlock *MaybeKillingBlock = UseInst->getParent(); BasicBlock *MaybeKillingBlock = UseInst->getParent();
@ -2479,7 +2493,7 @@ bool eliminateDeadStoresMemorySSA(Function &F, AliasAnalysis &AA,
// Check if NI overwrites SI. // Check if NI overwrites SI.
int64_t InstWriteOffset, DepWriteOffset; int64_t InstWriteOffset, DepWriteOffset;
OverwriteResult OR = OverwriteResult OR =
isOverwrite(SILoc, NILoc, State.DL, TLI, DepWriteOffset, isOverwrite(SI, NI, SILoc, NILoc, State.DL, TLI, DepWriteOffset,
InstWriteOffset, State.BatchAA, &F); InstWriteOffset, State.BatchAA, &F);
if (OR == OW_MaybePartial) { if (OR == OW_MaybePartial) {
auto Iter = State.IOLs.insert( auto Iter = State.IOLs.insert(

View File

@ -1,5 +1,6 @@
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt -tbaa -dse -enable-dse-memoryssa=false -S < %s | FileCheck %s ; RUN: opt -tbaa -dse -enable-dse-memoryssa=false -S < %s | FileCheck %s
; RUN: opt -tbaa -dse -enable-dse-memoryssa=true -S < %s | FileCheck %s
target datalayout = "e-m:e-p:32:32:32-a:0-n16:32-i64:64:64-i32:32:32-i16:16:16-i1:8:8-f32:32:32-f64:64:64-v32:32:32-v64:64:64-v512:512:512-v1024:1024:1024-v2048:2048:2048" target datalayout = "e-m:e-p:32:32:32-a:0-n16:32-i64:64:64-i32:32:32-i16:16:16-i1:8:8-f32:32:32-f64:64:64-v32:32:32-v64:64:64-v512:512:512-v1024:1024:1024-v2048:2048:2048"
define dllexport i32 @f0(i8** %a0, i8** %a1, i32 %a2, i32 %a3, i32 %a4, i32 %a5, i32 %a6, i32 %a7) #0 { define dllexport i32 @f0(i8** %a0, i8** %a1, i32 %a2, i32 %a3, i32 %a4, i32 %a5, i32 %a6, i32 %a7) #0 {