forked from OSchip/llvm-project
Another try to commit 323321 (aggressive instruction combine).
llvm-svn: 323416
This commit is contained in:
parent
0027ddfd85
commit
f1f57a3137
|
@ -641,6 +641,21 @@ not library calls are simplified is controlled by the
|
|||
:ref:`-functionattrs <passes-functionattrs>` pass and LLVM's knowledge of
|
||||
library calls on different targets.
|
||||
|
||||
.. _passes-aggressive-instcombine:
|
||||
|
||||
``-aggressive-instcombine``: Combine expression patterns
|
||||
--------------------------------------------------------
|
||||
|
||||
Combine expression patterns to form expressions with fewer, simple instructions.
|
||||
This pass does not modify the CFG.
|
||||
|
||||
For example, this pass reduce width of expressions post-dominated by TruncInst
|
||||
into smaller width when applicable.
|
||||
|
||||
It differs from instcombine pass in that it contains pattern optimization that
|
||||
requires higher complexity than the O(1), thus, it should run fewer times than
|
||||
instcombine pass.
|
||||
|
||||
``-internalize``: Internalize Global Symbols
|
||||
--------------------------------------------
|
||||
|
||||
|
|
|
@ -64,6 +64,7 @@ void initializeADCELegacyPassPass(PassRegistry&);
|
|||
void initializeAddDiscriminatorsLegacyPassPass(PassRegistry&);
|
||||
void initializeAddressSanitizerModulePass(PassRegistry&);
|
||||
void initializeAddressSanitizerPass(PassRegistry&);
|
||||
void initializeAggressiveInstCombinerLegacyPassPass(PassRegistry&);
|
||||
void initializeAliasSetPrinterPass(PassRegistry&);
|
||||
void initializeAlignmentFromAssumptionsPass(PassRegistry&);
|
||||
void initializeAlwaysInlinerLegacyPassPass(PassRegistry&);
|
||||
|
|
|
@ -0,0 +1,34 @@
|
|||
//===- AggressiveInstCombine.h - AggressiveInstCombine pass -----*- C++ -*-===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
/// \file
|
||||
///
|
||||
/// This file provides the primary interface to the aggressive instcombine pass.
|
||||
/// This pass is suitable for use in the new pass manager. For a pass that works
|
||||
/// with the legacy pass manager, please look for
|
||||
/// \c createAggressiveInstCombinerPass() in Scalar.h.
|
||||
///
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
#ifndef LLVM_TRANSFORMS_AGGRESSIVE_INSTCOMBINE_INSTCOMBINE_H
|
||||
#define LLVM_TRANSFORMS_AGGRESSIVE_INSTCOMBINE_INSTCOMBINE_H
|
||||
|
||||
#include "llvm/IR/Function.h"
|
||||
#include "llvm/IR/PassManager.h"
|
||||
|
||||
namespace llvm {
|
||||
|
||||
class AggressiveInstCombinePass
|
||||
: public PassInfoMixin<AggressiveInstCombinePass> {
|
||||
|
||||
public:
|
||||
PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM);
|
||||
};
|
||||
}
|
||||
|
||||
#endif
|
|
@ -140,6 +140,13 @@ Pass *createIndVarSimplifyPass();
|
|||
//
|
||||
FunctionPass *createInstructionCombiningPass(bool ExpensiveCombines = true);
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// AggressiveInstCombiner - Combine expression patterns to form expressions with
|
||||
// fewer, simple instructions. This pass does not modify the CFG.
|
||||
//
|
||||
FunctionPass *createAggressiveInstCombinerPass();
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// LICM - This pass is a loop invariant code motion and memory promotion pass.
|
||||
|
|
|
@ -20,6 +20,7 @@ type = Library
|
|||
name = LTO
|
||||
parent = Libraries
|
||||
required_libraries =
|
||||
AggressiveInstCombine
|
||||
Analysis
|
||||
BitReader
|
||||
BitWriter
|
||||
|
|
|
@ -19,4 +19,4 @@
|
|||
type = Library
|
||||
name = Passes
|
||||
parent = Libraries
|
||||
required_libraries = Analysis CodeGen Core IPO InstCombine Scalar Support Target TransformUtils Vectorize Instrumentation
|
||||
required_libraries = AggressiveInstCombine Analysis CodeGen Core IPO InstCombine Scalar Support Target TransformUtils Vectorize Instrumentation
|
||||
|
|
|
@ -59,6 +59,7 @@
|
|||
#include "llvm/Support/Debug.h"
|
||||
#include "llvm/Support/Regex.h"
|
||||
#include "llvm/Target/TargetMachine.h"
|
||||
#include "llvm/Transforms/AggressiveInstCombine/AggressiveInstCombine.h"
|
||||
#include "llvm/Transforms/GCOVProfiler.h"
|
||||
#include "llvm/Transforms/IPO/AlwaysInliner.h"
|
||||
#include "llvm/Transforms/IPO/ArgumentPromotion.h"
|
||||
|
@ -363,6 +364,8 @@ PassBuilder::buildFunctionSimplificationPipeline(OptimizationLevel Level,
|
|||
FPM.addPass(JumpThreadingPass());
|
||||
FPM.addPass(CorrelatedValuePropagationPass());
|
||||
FPM.addPass(SimplifyCFGPass());
|
||||
if (Level == O3)
|
||||
FPM.addPass(AggressiveInstCombinePass());
|
||||
FPM.addPass(InstCombinePass());
|
||||
|
||||
if (!isOptimizingForSize(Level))
|
||||
|
@ -1010,6 +1013,8 @@ ModulePassManager PassBuilder::buildLTODefaultPipeline(OptimizationLevel Level,
|
|||
// function pointers. When this happens, we often have to resolve varargs
|
||||
// calls, etc, so let instcombine do this.
|
||||
FunctionPassManager PeepholeFPM(DebugLogging);
|
||||
if (Level == O3)
|
||||
PeepholeFPM.addPass(AggressiveInstCombinePass());
|
||||
PeepholeFPM.addPass(InstCombinePass());
|
||||
invokePeepholeEPCallbacks(PeepholeFPM, Level);
|
||||
|
||||
|
|
|
@ -139,6 +139,7 @@ FUNCTION_ALIAS_ANALYSIS("type-based-aa", TypeBasedAA())
|
|||
FUNCTION_PASS("aa-eval", AAEvaluator())
|
||||
FUNCTION_PASS("adce", ADCEPass())
|
||||
FUNCTION_PASS("add-discriminators", AddDiscriminatorsPass())
|
||||
FUNCTION_PASS("aggressive-instcombine", AggressiveInstCombinePass())
|
||||
FUNCTION_PASS("alignment-from-assumptions", AlignmentFromAssumptionsPass())
|
||||
FUNCTION_PASS("bdce", BDCEPass())
|
||||
FUNCTION_PASS("bounds-checking", BoundsCheckingPass())
|
||||
|
|
|
@ -0,0 +1,110 @@
|
|||
//===- AggressiveInstCombine.cpp ------------------------------------------===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// This file implements the aggressive expression pattern combiner classes.
|
||||
// Currently, it handles expression patterns for:
|
||||
// * Truncate instruction
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
#include "llvm/Transforms/AggressiveInstCombine/AggressiveInstCombine.h"
|
||||
#include "AggressiveInstCombineInternal.h"
|
||||
#include "llvm/Analysis/AliasAnalysis.h"
|
||||
#include "llvm/Analysis/BasicAliasAnalysis.h"
|
||||
#include "llvm/Analysis/GlobalsModRef.h"
|
||||
#include "llvm/Analysis/TargetLibraryInfo.h"
|
||||
#include "llvm/IR/DataLayout.h"
|
||||
#include "llvm/Pass.h"
|
||||
#include "llvm/Transforms/Scalar.h"
|
||||
using namespace llvm;
|
||||
|
||||
#define DEBUG_TYPE "aggressive-instcombine"
|
||||
|
||||
namespace {
|
||||
/// Contains expression pattern combiner logic.
|
||||
/// This class provides both the logic to combine expression patterns and
|
||||
/// combine them. It differs from InstCombiner class in that each pattern
|
||||
/// combiner runs only once as opposed to InstCombine's multi-iteration,
|
||||
/// which allows pattern combiner to have higher complexity than the O(1)
|
||||
/// required by the instruction combiner.
|
||||
class AggressiveInstCombinerLegacyPass : public FunctionPass {
|
||||
public:
|
||||
static char ID; // Pass identification, replacement for typeid
|
||||
|
||||
AggressiveInstCombinerLegacyPass() : FunctionPass(ID) {
|
||||
initializeAggressiveInstCombinerLegacyPassPass(
|
||||
*PassRegistry::getPassRegistry());
|
||||
}
|
||||
|
||||
void getAnalysisUsage(AnalysisUsage &AU) const override;
|
||||
|
||||
/// Run all expression pattern optimizations on the given /p F function.
|
||||
///
|
||||
/// \param F function to optimize.
|
||||
/// \returns true if the IR is changed.
|
||||
bool runOnFunction(Function &F) override;
|
||||
};
|
||||
} // namespace
|
||||
|
||||
void AggressiveInstCombinerLegacyPass::getAnalysisUsage(
|
||||
AnalysisUsage &AU) const {
|
||||
AU.setPreservesCFG();
|
||||
AU.addRequired<TargetLibraryInfoWrapperPass>();
|
||||
AU.addPreserved<AAResultsWrapperPass>();
|
||||
AU.addPreserved<BasicAAWrapperPass>();
|
||||
AU.addPreserved<GlobalsAAWrapperPass>();
|
||||
}
|
||||
|
||||
bool AggressiveInstCombinerLegacyPass::runOnFunction(Function &F) {
|
||||
auto &TLI = getAnalysis<TargetLibraryInfoWrapperPass>().getTLI();
|
||||
auto &DL = F.getParent()->getDataLayout();
|
||||
|
||||
bool MadeIRChange = false;
|
||||
|
||||
// Handle TruncInst patterns
|
||||
TruncInstCombine TIC(TLI, DL);
|
||||
MadeIRChange |= TIC.run(F);
|
||||
|
||||
// TODO: add more patterns to handle...
|
||||
|
||||
return MadeIRChange;
|
||||
}
|
||||
|
||||
PreservedAnalyses AggressiveInstCombinePass::run(Function &F,
|
||||
FunctionAnalysisManager &AM) {
|
||||
auto &TLI = AM.getResult<TargetLibraryAnalysis>(F);
|
||||
auto &DL = F.getParent()->getDataLayout();
|
||||
bool MadeIRChange = false;
|
||||
|
||||
// Handle TruncInst patterns
|
||||
TruncInstCombine TIC(TLI, DL);
|
||||
MadeIRChange |= TIC.run(F);
|
||||
if (!MadeIRChange)
|
||||
// No changes, all analyses are preserved.
|
||||
return PreservedAnalyses::all();
|
||||
|
||||
// Mark all the analyses that instcombine updates as preserved.
|
||||
PreservedAnalyses PA;
|
||||
PA.preserveSet<CFGAnalyses>();
|
||||
PA.preserve<AAManager>();
|
||||
PA.preserve<GlobalsAA>();
|
||||
return PA;
|
||||
}
|
||||
|
||||
char AggressiveInstCombinerLegacyPass::ID = 0;
|
||||
INITIALIZE_PASS_BEGIN(AggressiveInstCombinerLegacyPass,
|
||||
"aggressive-instcombine",
|
||||
"Combine pattern based expressions", false, false)
|
||||
INITIALIZE_PASS_DEPENDENCY(TargetLibraryInfoWrapperPass)
|
||||
INITIALIZE_PASS_END(AggressiveInstCombinerLegacyPass, "aggressive-instcombine",
|
||||
"Combine pattern based expressions", false, false)
|
||||
|
||||
FunctionPass *llvm::createAggressiveInstCombinerPass() {
|
||||
return new AggressiveInstCombinerLegacyPass();
|
||||
}
|
|
@ -0,0 +1,119 @@
|
|||
//===- AggressiveInstCombineInternal.h --------------------------*- C++ -*-===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// This file implements the instruction pattern combiner classes.
|
||||
// Currently, it handles pattern expressions for:
|
||||
// * Truncate instruction
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
#include "llvm/ADT/MapVector.h"
|
||||
#include "llvm/ADT/STLExtras.h"
|
||||
#include "llvm/Analysis/AliasAnalysis.h"
|
||||
#include "llvm/Analysis/BasicAliasAnalysis.h"
|
||||
#include "llvm/Analysis/ConstantFolding.h"
|
||||
#include "llvm/Analysis/GlobalsModRef.h"
|
||||
#include "llvm/Analysis/TargetLibraryInfo.h"
|
||||
#include "llvm/IR/DataLayout.h"
|
||||
#include "llvm/Pass.h"
|
||||
#include "llvm/Transforms/Scalar.h"
|
||||
using namespace llvm;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// TruncInstCombine - looks for expression dags dominated by trunc instructions
|
||||
// and for each eligible dag, it will create a reduced bit-width expression and
|
||||
// replace the old expression with this new one and remove the old one.
|
||||
// Eligible expression dag is such that:
|
||||
// 1. Contains only supported instructions.
|
||||
// 2. Supported leaves: ZExtInst, SExtInst, TruncInst and Constant value.
|
||||
// 3. Can be evaluated into type with reduced legal bit-width (or Trunc type).
|
||||
// 4. All instructions in the dag must not have users outside the dag.
|
||||
// Only exception is for {ZExt, SExt}Inst with operand type equal to the
|
||||
// new reduced type chosen in (3).
|
||||
//
|
||||
// The motivation for this optimization is that evaluating and expression using
|
||||
// smaller bit-width is preferable, especially for vectorization where we can
|
||||
// fit more values in one vectorized instruction. In addition, this optimization
|
||||
// may decrease the number of cast instructions, but will not increase it.
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
namespace llvm {
|
||||
class DataLayout;
|
||||
class TargetLibraryInfo;
|
||||
|
||||
class TruncInstCombine {
|
||||
TargetLibraryInfo &TLI;
|
||||
const DataLayout &DL;
|
||||
|
||||
/// List of all TruncInst instructions to be processed.
|
||||
SmallVector<TruncInst *, 4> Worklist;
|
||||
|
||||
/// Current processed TruncInst instruction.
|
||||
TruncInst *CurrentTruncInst;
|
||||
|
||||
/// Information per each instruction in the expression dag.
|
||||
struct Info {
|
||||
/// Number of LSBs that are needed to generate a valid expression.
|
||||
unsigned ValidBitWidth = 0;
|
||||
/// Minimum number of LSBs needed to generate the ValidBitWidth.
|
||||
unsigned MinBitWidth = 0;
|
||||
/// The reduced value generated to replace the old instruction.
|
||||
Value *NewValue = nullptr;
|
||||
};
|
||||
/// An ordered map representing expression dag post-dominated by current
|
||||
/// processed TruncInst. It maps each instruction in the dag to its Info
|
||||
/// structure. The map is ordered such that each instruction appears before
|
||||
/// all other instructions in the dag that uses it.
|
||||
MapVector<Instruction *, Info> InstInfoMap;
|
||||
|
||||
public:
|
||||
TruncInstCombine(TargetLibraryInfo &TLI, const DataLayout &DL)
|
||||
: TLI(TLI), DL(DL), CurrentTruncInst(nullptr) {}
|
||||
|
||||
/// Perform TruncInst pattern optimization on given function.
|
||||
bool run(Function &F);
|
||||
|
||||
private:
|
||||
/// Build expression dag dominated by the /p CurrentTruncInst and append it to
|
||||
/// the InstInfoMap container.
|
||||
///
|
||||
/// \return true only if succeed to generate an eligible sub expression dag.
|
||||
bool buildTruncExpressionDag();
|
||||
|
||||
/// Calculate the minimal allowed bit-width of the chain ending with the
|
||||
/// currently visited truncate's operand.
|
||||
///
|
||||
/// \return minimum number of bits to which the chain ending with the
|
||||
/// truncate's operand can be shrunk to.
|
||||
unsigned getMinBitWidth();
|
||||
|
||||
/// Build an expression dag dominated by the current processed TruncInst and
|
||||
/// Check if it is eligible to be reduced to a smaller type.
|
||||
///
|
||||
/// \return the scalar version of the new type to be used for the reduced
|
||||
/// expression dag, or nullptr if the expression dag is not eligible
|
||||
/// to be reduced.
|
||||
Type *getBestTruncatedType();
|
||||
|
||||
/// Given a \p V value and a \p SclTy scalar type return the generated reduced
|
||||
/// value of \p V based on the type \p SclTy.
|
||||
///
|
||||
/// \param V value to be reduced.
|
||||
/// \param SclTy scalar version of new type to reduce to.
|
||||
/// \return the new reduced value.
|
||||
Value *getReducedOperand(Value *V, Type *SclTy);
|
||||
|
||||
/// Create a new expression dag using the reduced /p SclTy type and replace
|
||||
/// the old expression dag with it. Also erase all instructions in the old
|
||||
/// dag, except those that are still needed outside the dag.
|
||||
///
|
||||
/// \param SclTy scalar version of new type to reduce expression dag into.
|
||||
void ReduceExpressionDag(Type *SclTy);
|
||||
};
|
||||
} // end namespace llvm.
|
|
@ -0,0 +1,11 @@
|
|||
add_llvm_library(LLVMAggressiveInstCombine
|
||||
AggressiveInstCombine.cpp
|
||||
TruncInstCombine.cpp
|
||||
|
||||
ADDITIONAL_HEADER_DIRS
|
||||
${LLVM_MAIN_INCLUDE_DIR}/llvm/Transforms
|
||||
${LLVM_MAIN_INCLUDE_DIR}/llvm/Transforms/AggressiveInstCombine
|
||||
|
||||
DEPENDS
|
||||
intrinsics_gen
|
||||
)
|
|
@ -0,0 +1,22 @@
|
|||
;===- ./lib/Transforms/AggressiveInstCombine/LLVMBuild.txt -----*- Conf -*--===;
|
||||
;
|
||||
; The LLVM Compiler Infrastructure
|
||||
;
|
||||
; This file is distributed under the University of Illinois Open Source
|
||||
; License. See LICENSE.TXT for details.
|
||||
;
|
||||
;===------------------------------------------------------------------------===;
|
||||
;
|
||||
; This is an LLVMBuild description file for the components in this subdirectory.
|
||||
;
|
||||
; For more information on the LLVMBuild system, please see:
|
||||
;
|
||||
; http://llvm.org/docs/LLVMBuild.html
|
||||
;
|
||||
;===------------------------------------------------------------------------===;
|
||||
|
||||
[component_0]
|
||||
type = Library
|
||||
name = AggressiveInstCombine
|
||||
parent = Transforms
|
||||
required_libraries = Analysis Core Support TransformUtils
|
|
@ -0,0 +1,404 @@
|
|||
//===- TruncInstCombine.cpp -----------------------------------------------===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// TruncInstCombine - looks for expression dags post-dominated by TruncInst and
|
||||
// for each eligible dag, it will create a reduced bit-width expression, replace
|
||||
// the old expression with this new one and remove the old expression.
|
||||
// Eligible expression dag is such that:
|
||||
// 1. Contains only supported instructions.
|
||||
// 2. Supported leaves: ZExtInst, SExtInst, TruncInst and Constant value.
|
||||
// 3. Can be evaluated into type with reduced legal bit-width.
|
||||
// 4. All instructions in the dag must not have users outside the dag.
|
||||
// The only exception is for {ZExt, SExt}Inst with operand type equal to
|
||||
// the new reduced type evaluated in (3).
|
||||
//
|
||||
// The motivation for this optimization is that evaluating and expression using
|
||||
// smaller bit-width is preferable, especially for vectorization where we can
|
||||
// fit more values in one vectorized instruction. In addition, this optimization
|
||||
// may decrease the number of cast instructions, but will not increase it.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
#include "AggressiveInstCombineInternal.h"
|
||||
#include "llvm/ADT/MapVector.h"
|
||||
#include "llvm/ADT/STLExtras.h"
|
||||
#include "llvm/Analysis/ConstantFolding.h"
|
||||
#include "llvm/Analysis/TargetLibraryInfo.h"
|
||||
#include "llvm/IR/DataLayout.h"
|
||||
#include "llvm/IR/IRBuilder.h"
|
||||
using namespace llvm;
|
||||
|
||||
#define DEBUG_TYPE "aggressive-instcombine"
|
||||
|
||||
/// Given an instruction and a container, it fills all the relevant operands of
|
||||
/// that instruction, with respect to the Trunc expression dag optimizaton.
|
||||
static void getRelevantOperands(Instruction *I, SmallVectorImpl<Value *> &Ops) {
|
||||
unsigned Opc = I->getOpcode();
|
||||
switch (Opc) {
|
||||
case Instruction::Trunc:
|
||||
case Instruction::ZExt:
|
||||
case Instruction::SExt:
|
||||
// These CastInst are considered leaves of the evaluated expression, thus,
|
||||
// their operands are not relevent.
|
||||
break;
|
||||
case Instruction::Add:
|
||||
case Instruction::Sub:
|
||||
case Instruction::Mul:
|
||||
case Instruction::And:
|
||||
case Instruction::Or:
|
||||
case Instruction::Xor:
|
||||
Ops.push_back(I->getOperand(0));
|
||||
Ops.push_back(I->getOperand(1));
|
||||
break;
|
||||
default:
|
||||
llvm_unreachable("Unreachable!");
|
||||
}
|
||||
}
|
||||
|
||||
bool TruncInstCombine::buildTruncExpressionDag() {
|
||||
SmallVector<Value *, 8> Worklist;
|
||||
SmallVector<Instruction *, 8> Stack;
|
||||
// Clear old expression dag.
|
||||
InstInfoMap.clear();
|
||||
|
||||
Worklist.push_back(CurrentTruncInst->getOperand(0));
|
||||
|
||||
while (!Worklist.empty()) {
|
||||
Value *Curr = Worklist.back();
|
||||
|
||||
if (isa<Constant>(Curr)) {
|
||||
Worklist.pop_back();
|
||||
continue;
|
||||
}
|
||||
|
||||
auto *I = dyn_cast<Instruction>(Curr);
|
||||
if (!I)
|
||||
return false;
|
||||
|
||||
if (!Stack.empty() && Stack.back() == I) {
|
||||
// Already handled all instruction operands, can remove it from both the
|
||||
// Worklist and the Stack, and add it to the instruction info map.
|
||||
Worklist.pop_back();
|
||||
Stack.pop_back();
|
||||
// Insert I to the Info map.
|
||||
InstInfoMap.insert(std::make_pair(I, Info()));
|
||||
continue;
|
||||
}
|
||||
|
||||
if (InstInfoMap.count(I)) {
|
||||
Worklist.pop_back();
|
||||
continue;
|
||||
}
|
||||
|
||||
// Add the instruction to the stack before start handling its operands.
|
||||
Stack.push_back(I);
|
||||
|
||||
unsigned Opc = I->getOpcode();
|
||||
switch (Opc) {
|
||||
case Instruction::Trunc:
|
||||
case Instruction::ZExt:
|
||||
case Instruction::SExt:
|
||||
// trunc(trunc(x)) -> trunc(x)
|
||||
// trunc(ext(x)) -> ext(x) if the source type is smaller than the new dest
|
||||
// trunc(ext(x)) -> trunc(x) if the source type is larger than the new
|
||||
// dest
|
||||
break;
|
||||
case Instruction::Add:
|
||||
case Instruction::Sub:
|
||||
case Instruction::Mul:
|
||||
case Instruction::And:
|
||||
case Instruction::Or:
|
||||
case Instruction::Xor: {
|
||||
SmallVector<Value *, 2> Operands;
|
||||
getRelevantOperands(I, Operands);
|
||||
for (Value *Operand : Operands)
|
||||
Worklist.push_back(Operand);
|
||||
break;
|
||||
}
|
||||
default:
|
||||
// TODO: Can handle more cases here:
|
||||
// 1. select, shufflevector, extractelement, insertelement
|
||||
// 2. udiv, urem
|
||||
// 3. shl, lshr, ashr
|
||||
// 4. phi node(and loop handling)
|
||||
// ...
|
||||
return false;
|
||||
}
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
unsigned TruncInstCombine::getMinBitWidth() {
|
||||
SmallVector<Value *, 8> Worklist;
|
||||
SmallVector<Instruction *, 8> Stack;
|
||||
|
||||
Value *Src = CurrentTruncInst->getOperand(0);
|
||||
Type *DstTy = CurrentTruncInst->getType();
|
||||
unsigned TruncBitWidth = DstTy->getScalarSizeInBits();
|
||||
unsigned OrigBitWidth =
|
||||
CurrentTruncInst->getOperand(0)->getType()->getScalarSizeInBits();
|
||||
|
||||
if (isa<Constant>(Src))
|
||||
return TruncBitWidth;
|
||||
|
||||
Worklist.push_back(Src);
|
||||
InstInfoMap[cast<Instruction>(Src)].ValidBitWidth = TruncBitWidth;
|
||||
|
||||
while (!Worklist.empty()) {
|
||||
Value *Curr = Worklist.back();
|
||||
|
||||
if (isa<Constant>(Curr)) {
|
||||
Worklist.pop_back();
|
||||
continue;
|
||||
}
|
||||
|
||||
// Otherwise, it must be an instruction.
|
||||
auto *I = cast<Instruction>(Curr);
|
||||
|
||||
auto &Info = InstInfoMap[I];
|
||||
|
||||
SmallVector<Value *, 2> Operands;
|
||||
getRelevantOperands(I, Operands);
|
||||
|
||||
if (!Stack.empty() && Stack.back() == I) {
|
||||
// Already handled all instruction operands, can remove it from both, the
|
||||
// Worklist and the Stack, and update MinBitWidth.
|
||||
Worklist.pop_back();
|
||||
Stack.pop_back();
|
||||
for (auto *Operand : Operands)
|
||||
if (auto *IOp = dyn_cast<Instruction>(Operand))
|
||||
Info.MinBitWidth =
|
||||
std::max(Info.MinBitWidth, InstInfoMap[IOp].MinBitWidth);
|
||||
continue;
|
||||
}
|
||||
|
||||
// Add the instruction to the stack before start handling its operands.
|
||||
Stack.push_back(I);
|
||||
unsigned ValidBitWidth = Info.ValidBitWidth;
|
||||
|
||||
// Update minimum bit-width before handling its operands. This is required
|
||||
// when the instruction is part of a loop.
|
||||
Info.MinBitWidth = std::max(Info.MinBitWidth, Info.ValidBitWidth);
|
||||
|
||||
for (auto *Operand : Operands)
|
||||
if (auto *IOp = dyn_cast<Instruction>(Operand)) {
|
||||
// If we already calculated the minimum bit-width for this valid
|
||||
// bit-width, or for a smaller valid bit-width, then just keep the
|
||||
// answer we already calculated.
|
||||
unsigned IOpBitwidth = InstInfoMap.lookup(IOp).ValidBitWidth;
|
||||
if (IOpBitwidth >= ValidBitWidth)
|
||||
continue;
|
||||
InstInfoMap[IOp].ValidBitWidth = std::max(ValidBitWidth, IOpBitwidth);
|
||||
Worklist.push_back(IOp);
|
||||
}
|
||||
}
|
||||
unsigned MinBitWidth = InstInfoMap.lookup(cast<Instruction>(Src)).MinBitWidth;
|
||||
assert(MinBitWidth >= TruncBitWidth);
|
||||
|
||||
if (MinBitWidth > TruncBitWidth) {
|
||||
// In this case reducing expression with vector type might generate a new
|
||||
// vector type, which is not preferable as it might result in generating
|
||||
// sub-optimal code.
|
||||
if (DstTy->isVectorTy())
|
||||
return OrigBitWidth;
|
||||
// Use the smallest integer type in the range [MinBitWidth, OrigBitWidth).
|
||||
Type *Ty = DL.getSmallestLegalIntType(DstTy->getContext(), MinBitWidth);
|
||||
// Update minimum bit-width with the new destination type bit-width if
|
||||
// succeeded to find such, otherwise, with original bit-width.
|
||||
MinBitWidth = Ty ? Ty->getScalarSizeInBits() : OrigBitWidth;
|
||||
} else { // MinBitWidth == TruncBitWidth
|
||||
// In this case the expression can be evaluated with the trunc instruction
|
||||
// destination type, and trunc instruction can be omitted. However, we
|
||||
// should not perform the evaluation if the original type is a legal scalar
|
||||
// type and the target type is illegal.
|
||||
bool FromLegal = MinBitWidth == 1 || DL.isLegalInteger(OrigBitWidth);
|
||||
bool ToLegal = MinBitWidth == 1 || DL.isLegalInteger(MinBitWidth);
|
||||
if (!DstTy->isVectorTy() && FromLegal && !ToLegal)
|
||||
return OrigBitWidth;
|
||||
}
|
||||
return MinBitWidth;
|
||||
}
|
||||
|
||||
Type *TruncInstCombine::getBestTruncatedType() {
|
||||
if (!buildTruncExpressionDag())
|
||||
return nullptr;
|
||||
|
||||
// We don't want to duplicate instructions, which isn't profitable. Thus, we
|
||||
// can't shrink something that has multiple users, unless all users are
|
||||
// post-dominated by the trunc instruction, i.e., were visited during the
|
||||
// expression evaluation.
|
||||
unsigned DesiredBitWidth = 0;
|
||||
for (auto Itr : InstInfoMap) {
|
||||
Instruction *I = Itr.first;
|
||||
if (I->hasOneUse())
|
||||
continue;
|
||||
bool IsExtInst = (isa<ZExtInst>(I) || isa<SExtInst>(I));
|
||||
for (auto *U : I->users())
|
||||
if (auto *UI = dyn_cast<Instruction>(U))
|
||||
if (UI != CurrentTruncInst && !InstInfoMap.count(UI)) {
|
||||
if (!IsExtInst)
|
||||
return nullptr;
|
||||
// If this is an extension from the dest type, we can eliminate it,
|
||||
// even if it has multiple users. Thus, update the DesiredBitWidth and
|
||||
// validate all extension instructions agrees on same DesiredBitWidth.
|
||||
unsigned ExtInstBitWidth =
|
||||
I->getOperand(0)->getType()->getScalarSizeInBits();
|
||||
if (DesiredBitWidth && DesiredBitWidth != ExtInstBitWidth)
|
||||
return nullptr;
|
||||
DesiredBitWidth = ExtInstBitWidth;
|
||||
}
|
||||
}
|
||||
|
||||
unsigned OrigBitWidth =
|
||||
CurrentTruncInst->getOperand(0)->getType()->getScalarSizeInBits();
|
||||
|
||||
// Calculate minimum allowed bit-width allowed for shrinking the currently
|
||||
// visited truncate's operand.
|
||||
unsigned MinBitWidth = getMinBitWidth();
|
||||
|
||||
// Check that we can shrink to smaller bit-width than original one and that
|
||||
// it is similar to the DesiredBitWidth is such exists.
|
||||
if (MinBitWidth >= OrigBitWidth ||
|
||||
(DesiredBitWidth && DesiredBitWidth != MinBitWidth))
|
||||
return nullptr;
|
||||
|
||||
return IntegerType::get(CurrentTruncInst->getContext(), MinBitWidth);
|
||||
}
|
||||
|
||||
/// Given a reduced scalar type \p Ty and a \p V value, return a reduced type
|
||||
/// for \p V, according to its type, if it vector type, return the vector
|
||||
/// version of \p Ty, otherwise return \p Ty.
|
||||
static Type *getReducedType(Value *V, Type *Ty) {
|
||||
assert(Ty && !Ty->isVectorTy() && "Expect Scalar Type");
|
||||
if (auto *VTy = dyn_cast<VectorType>(V->getType()))
|
||||
return VectorType::get(Ty, VTy->getNumElements());
|
||||
return Ty;
|
||||
}
|
||||
|
||||
Value *TruncInstCombine::getReducedOperand(Value *V, Type *SclTy) {
|
||||
Type *Ty = getReducedType(V, SclTy);
|
||||
if (auto *C = dyn_cast<Constant>(V)) {
|
||||
C = ConstantExpr::getIntegerCast(C, Ty, false);
|
||||
// If we got a constantexpr back, try to simplify it with DL info.
|
||||
if (Constant *FoldedC = ConstantFoldConstant(C, DL, &TLI))
|
||||
C = FoldedC;
|
||||
return C;
|
||||
}
|
||||
|
||||
auto *I = cast<Instruction>(V);
|
||||
Info Entry = InstInfoMap.lookup(I);
|
||||
assert(Entry.NewValue);
|
||||
return Entry.NewValue;
|
||||
}
|
||||
|
||||
void TruncInstCombine::ReduceExpressionDag(Type *SclTy) {
|
||||
for (auto &Itr : InstInfoMap) { // Forward
|
||||
Instruction *I = Itr.first;
|
||||
TruncInstCombine::Info &NodeInfo = Itr.second;
|
||||
|
||||
assert(!NodeInfo.NewValue && "Instruction has been evaluated");
|
||||
|
||||
IRBuilder<> Builder(I);
|
||||
Value *Res = nullptr;
|
||||
unsigned Opc = I->getOpcode();
|
||||
switch (Opc) {
|
||||
case Instruction::Trunc:
|
||||
case Instruction::ZExt:
|
||||
case Instruction::SExt: {
|
||||
Type *Ty = getReducedType(I, SclTy);
|
||||
// If the source type of the cast is the type we're trying for then we can
|
||||
// just return the source. There's no need to insert it because it is not
|
||||
// new.
|
||||
if (I->getOperand(0)->getType() == Ty) {
|
||||
NodeInfo.NewValue = I->getOperand(0);
|
||||
continue;
|
||||
}
|
||||
// Otherwise, must be the same type of cast, so just reinsert a new one.
|
||||
// This also handles the case of zext(trunc(x)) -> zext(x).
|
||||
Res = Builder.CreateIntCast(I->getOperand(0), Ty,
|
||||
Opc == Instruction::SExt);
|
||||
|
||||
// Update Worklist entries with new value if needed.
|
||||
if (auto *NewCI = dyn_cast<TruncInst>(Res)) {
|
||||
auto Entry = find(Worklist, I);
|
||||
if (Entry != Worklist.end())
|
||||
*Entry = NewCI;
|
||||
}
|
||||
break;
|
||||
}
|
||||
case Instruction::Add:
|
||||
case Instruction::Sub:
|
||||
case Instruction::Mul:
|
||||
case Instruction::And:
|
||||
case Instruction::Or:
|
||||
case Instruction::Xor: {
|
||||
Value *LHS = getReducedOperand(I->getOperand(0), SclTy);
|
||||
Value *RHS = getReducedOperand(I->getOperand(1), SclTy);
|
||||
Res = Builder.CreateBinOp((Instruction::BinaryOps)Opc, LHS, RHS);
|
||||
break;
|
||||
}
|
||||
default:
|
||||
llvm_unreachable("Unhandled instruction");
|
||||
}
|
||||
|
||||
NodeInfo.NewValue = Res;
|
||||
if (auto *ResI = dyn_cast<Instruction>(Res))
|
||||
ResI->takeName(I);
|
||||
}
|
||||
|
||||
Value *Res = getReducedOperand(CurrentTruncInst->getOperand(0), SclTy);
|
||||
Type *DstTy = CurrentTruncInst->getType();
|
||||
if (Res->getType() != DstTy) {
|
||||
IRBuilder<> Builder(CurrentTruncInst);
|
||||
Res = Builder.CreateIntCast(Res, DstTy, false);
|
||||
if (auto *ResI = dyn_cast<Instruction>(Res))
|
||||
ResI->takeName(CurrentTruncInst);
|
||||
}
|
||||
CurrentTruncInst->replaceAllUsesWith(Res);
|
||||
|
||||
// Erase old expression dag, which was replaced by the reduced expression dag.
|
||||
// We iterate backward, which means we visit the instruction before we visit
|
||||
// any of its operands, this way, when we get to the operand, we already
|
||||
// removed the instructions (from the expression dag) that uses it.
|
||||
CurrentTruncInst->eraseFromParent();
|
||||
for (auto I = InstInfoMap.rbegin(), E = InstInfoMap.rend(); I != E; ++I) {
|
||||
// We still need to check that the instruction has no users before we erase
|
||||
// it, because {SExt, ZExt}Inst Instruction might have other users that was
|
||||
// not reduced, in such case, we need to keep that instruction.
|
||||
if (!I->first->getNumUses())
|
||||
I->first->eraseFromParent();
|
||||
}
|
||||
}
|
||||
|
||||
bool TruncInstCombine::run(Function &F) {
|
||||
bool MadeIRChange = false;
|
||||
|
||||
// Collect all TruncInst in the function into the Worklist for evaluating.
|
||||
for (auto &BB : F)
|
||||
for (auto &I : BB)
|
||||
if (auto *CI = dyn_cast<TruncInst>(&I))
|
||||
Worklist.push_back(CI);
|
||||
|
||||
// Process all TruncInst in the Worklist, for each instruction:
|
||||
// 1. Check if it dominates an eligible expression dag to be reduced.
|
||||
// 2. Create a reduced expression dag and replace the old one with it.
|
||||
while (!Worklist.empty()) {
|
||||
CurrentTruncInst = Worklist.pop_back_val();
|
||||
|
||||
if (Type *NewDstSclTy = getBestTruncatedType()) {
|
||||
DEBUG(dbgs() << "ICE: TruncInstCombine reducing type of expression dag "
|
||||
"dominated by: "
|
||||
<< CurrentTruncInst << '\n');
|
||||
ReduceExpressionDag(NewDstSclTy);
|
||||
MadeIRChange = true;
|
||||
}
|
||||
}
|
||||
|
||||
return MadeIRChange;
|
||||
}
|
|
@ -1,5 +1,6 @@
|
|||
add_subdirectory(Utils)
|
||||
add_subdirectory(Instrumentation)
|
||||
add_subdirectory(AggressiveInstCombine)
|
||||
add_subdirectory(InstCombine)
|
||||
add_subdirectory(Scalar)
|
||||
add_subdirectory(IPO)
|
||||
|
|
|
@ -20,4 +20,4 @@ type = Library
|
|||
name = IPO
|
||||
parent = Transforms
|
||||
library_name = ipo
|
||||
required_libraries = Analysis BitReader BitWriter Core InstCombine IRReader Linker Object ProfileData Scalar Support TransformUtils Vectorize Instrumentation
|
||||
required_libraries = AggressiveInstCombine Analysis BitReader BitWriter Core InstCombine IRReader Linker Object ProfileData Scalar Support TransformUtils Vectorize Instrumentation
|
||||
|
|
|
@ -318,6 +318,8 @@ void PassManagerBuilder::addFunctionSimplificationPasses(
|
|||
MPM.add(createCorrelatedValuePropagationPass()); // Propagate conditionals
|
||||
MPM.add(createCFGSimplificationPass()); // Merge & remove BBs
|
||||
// Combine silly seq's
|
||||
if (OptLevel > 2)
|
||||
MPM.add(createAggressiveInstCombinerPass());
|
||||
addInstructionCombiningPass(MPM);
|
||||
if (SizeLevel == 0 && !DisableLibCallsShrinkWrap)
|
||||
MPM.add(createLibCallsShrinkWrapPass());
|
||||
|
@ -765,6 +767,8 @@ void PassManagerBuilder::addLTOOptimizationPasses(legacy::PassManagerBase &PM) {
|
|||
// simplification opportunities, and both can propagate functions through
|
||||
// function pointers. When this happens, we often have to resolve varargs
|
||||
// calls, etc, so let instcombine do this.
|
||||
if (OptLevel > 2)
|
||||
PM.add(createAggressiveInstCombinerPass());
|
||||
addInstructionCombiningPass(PM);
|
||||
addExtensionsToPM(EP_Peephole, PM);
|
||||
|
||||
|
|
|
@ -16,7 +16,7 @@
|
|||
;===------------------------------------------------------------------------===;
|
||||
|
||||
[common]
|
||||
subdirectories = Coroutines IPO InstCombine Instrumentation Scalar Utils Vectorize ObjCARC
|
||||
subdirectories = AggressiveInstCombine Coroutines IPO InstCombine Instrumentation Scalar Utils Vectorize ObjCARC
|
||||
|
||||
[component_0]
|
||||
type = Group
|
||||
|
|
|
@ -20,4 +20,4 @@ type = Library
|
|||
name = Scalar
|
||||
parent = Transforms
|
||||
library_name = ScalarOpts
|
||||
required_libraries = Analysis Core InstCombine Support TransformUtils
|
||||
required_libraries = AggressiveInstCombine Analysis Core InstCombine Support TransformUtils
|
||||
|
|
|
@ -126,6 +126,7 @@
|
|||
; CHECK-O-NEXT: Running analysis: LazyValueAnalysis
|
||||
; CHECK-O-NEXT: Running pass: CorrelatedValuePropagationPass
|
||||
; CHECK-O-NEXT: Running pass: SimplifyCFGPass
|
||||
; CHECK-O3-NEXT: AggressiveInstCombinePass
|
||||
; CHECK-O-NEXT: Running pass: InstCombinePass
|
||||
; CHECK-O1-NEXT: Running pass: LibCallsShrinkWrapPass
|
||||
; CHECK-O2-NEXT: Running pass: LibCallsShrinkWrapPass
|
||||
|
|
|
@ -10,7 +10,8 @@
|
|||
; RUN: | FileCheck %s --check-prefix=CHECK-O --check-prefix=CHECK-O2
|
||||
; RUN: opt -disable-verify -debug-pass-manager \
|
||||
; RUN: -passes='lto<O3>' -S %s 2>&1 \
|
||||
; RUN: | FileCheck %s --check-prefix=CHECK-O --check-prefix=CHECK-O2
|
||||
; RUN: | FileCheck %s --check-prefix=CHECK-O --check-prefix=CHECK-O2 \
|
||||
; RUN: --check-prefix=CHECK-O3
|
||||
; RUN: opt -disable-verify -debug-pass-manager \
|
||||
; RUN: -passes='lto<Os>' -S %s 2>&1 \
|
||||
; RUN: | FileCheck %s --check-prefix=CHECK-O --check-prefix=CHECK-O2
|
||||
|
@ -20,7 +21,7 @@
|
|||
; RUN: opt -disable-verify -debug-pass-manager \
|
||||
; RUN: -passes='lto<O3>' -S %s -passes-ep-peephole='no-op-function' 2>&1 \
|
||||
; RUN: | FileCheck %s --check-prefix=CHECK-O --check-prefix=CHECK-O2 \
|
||||
; RUN: --check-prefix=CHECK-EP-Peephole
|
||||
; RUN: --check-prefix=CHECK-O3 --check-prefix=CHECK-EP-Peephole
|
||||
|
||||
; CHECK-O: Starting llvm::Module pass manager run.
|
||||
; CHECK-O-NEXT: Running pass: PassManager<{{.*}}Module
|
||||
|
@ -60,6 +61,7 @@
|
|||
; CHECK-O2-NEXT: Running pass: DeadArgumentEliminationPass
|
||||
; CHECK-O2-NEXT: Running pass: ModuleToFunctionPassAdaptor<{{.*}}PassManager{{.*}}>
|
||||
; CHECK-O2-NEXT: Starting llvm::Function pass manager run.
|
||||
; CHECK-O3-NEXT: Running pass: AggressiveInstCombinePass
|
||||
; CHECK-O2-NEXT: Running pass: InstCombinePass
|
||||
; CHECK-EP-Peephole-NEXT: Running pass: NoOpFunctionPass
|
||||
; CHECK-O2-NEXT: Finished llvm::Function pass manager run.
|
||||
|
|
|
@ -111,6 +111,7 @@
|
|||
; CHECK-O-NEXT: Running analysis: LazyValueAnalysis
|
||||
; CHECK-O-NEXT: Running pass: CorrelatedValuePropagationPass
|
||||
; CHECK-O-NEXT: Running pass: SimplifyCFGPass
|
||||
; CHECK-O3-NEXT: Running pass: AggressiveInstCombinePass
|
||||
; CHECK-O-NEXT: Running pass: InstCombinePass
|
||||
; CHECK-O1-NEXT: Running pass: LibCallsShrinkWrapPass
|
||||
; CHECK-O2-NEXT: Running pass: LibCallsShrinkWrapPass
|
||||
|
|
|
@ -0,0 +1,79 @@
|
|||
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
||||
; RUN: opt < %s -aggressive-instcombine -S | FileCheck %s
|
||||
; RUN: opt < %s -passes=aggressive-instcombine -S | FileCheck %s
|
||||
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
|
||||
|
||||
; Aggressive Instcombine should be able to reduce width of these constant
|
||||
; expressions, without crashing.
|
||||
|
||||
declare i32 @use32(i32)
|
||||
declare <2 x i32> @use32_vec(<2 x i32>)
|
||||
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
;; These tests check cases where expression dag post-dominated by TruncInst
|
||||
;; contains instruction, which has more than one usage.
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
|
||||
define void @const_expression_mul() {
|
||||
; CHECK-LABEL: @const_expression_mul(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 242)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A = mul i64 11, 22
|
||||
%T = trunc i64 %A to i32
|
||||
call i32 @use32(i32 %T)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @const_expression_zext() {
|
||||
; CHECK-LABEL: @const_expression_zext(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 33)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A = zext i32 33 to i64
|
||||
%T = trunc i64 %A to i32
|
||||
call i32 @use32(i32 %T)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @const_expression_trunc() {
|
||||
; CHECK-LABEL: @const_expression_trunc(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 44)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%T = trunc i64 44 to i32
|
||||
call i32 @use32(i32 %T)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @const_expression_mul_vec() {
|
||||
; CHECK-LABEL: @const_expression_mul_vec(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> <i32 24531, i32 24864>)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A = mul <2 x i64> <i64 111, i64 112>, <i64 221, i64 222>
|
||||
%T = trunc <2 x i64> %A to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @const_expression_zext_vec() {
|
||||
; CHECK-LABEL: @const_expression_zext_vec(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> <i32 331, i32 332>)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A = zext <2 x i32> <i32 331, i32 332> to <2 x i64>
|
||||
%T = trunc <2 x i64> %A to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @const_expression_trunc_vec() {
|
||||
; CHECK-LABEL: @const_expression_trunc_vec(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> <i32 551, i32 552>)
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%T = trunc <2 x i64> <i64 551, i64 552> to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T)
|
||||
ret void
|
||||
}
|
|
@ -0,0 +1,214 @@
|
|||
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
||||
; RUN: opt < %s -aggressive-instcombine -S | FileCheck %s
|
||||
; RUN: opt < %s -passes=aggressive-instcombine -S | FileCheck %s
|
||||
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
|
||||
|
||||
; Aggressive Instcombine should be able to reduce width of these expressions.
|
||||
|
||||
declare i32 @use32(i32)
|
||||
declare i32 @use64(i64)
|
||||
declare <2 x i32> @use32_vec(<2 x i32>)
|
||||
declare <2 x i32> @use64_vec(<2 x i64>)
|
||||
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
;; These tests check cases where expression dag post-dominated by TruncInst
|
||||
;; contains instruction, which has more than one usage.
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
|
||||
define void @multi_uses_add(i32 %X) {
|
||||
; CHECK-LABEL: @multi_uses_add(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext i32 [[X:%.*]] to i64
|
||||
; CHECK-NEXT: [[B1:%.*]] = add i32 [[X]], 15
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul i32 [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @use64(i64 [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext i32 %X to i64
|
||||
%B1 = add i64 %A1, 15
|
||||
%C1 = mul i64 %B1, %B1
|
||||
%T1 = trunc i64 %C1 to i32
|
||||
call i32 @use32(i32 %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call i32 @use64(i64 %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_uses_or(i32 %X) {
|
||||
; CHECK-LABEL: @multi_uses_or(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext i32 [[X:%.*]] to i64
|
||||
; CHECK-NEXT: [[B1:%.*]] = or i32 [[X]], 15
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul i32 [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @use64(i64 [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext i32 %X to i64
|
||||
%B1 = or i64 %A1, 15
|
||||
%C1 = mul i64 %B1, %B1
|
||||
%T1 = trunc i64 %C1 to i32
|
||||
call i32 @use32(i32 %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call i32 @use64(i64 %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_uses_xor(i32 %X) {
|
||||
; CHECK-LABEL: @multi_uses_xor(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext i32 [[X:%.*]] to i64
|
||||
; CHECK-NEXT: [[B1:%.*]] = xor i32 [[X]], 15
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul i32 [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @use64(i64 [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext i32 %X to i64
|
||||
%B1 = xor i64 %A1, 15
|
||||
%C1 = mul i64 %B1, %B1
|
||||
%T1 = trunc i64 %C1 to i32
|
||||
call i32 @use32(i32 %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call i32 @use64(i64 %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_uses_and(i32 %X) {
|
||||
; CHECK-LABEL: @multi_uses_and(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext i32 [[X:%.*]] to i64
|
||||
; CHECK-NEXT: [[B1:%.*]] = and i32 [[X]], 15
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul i32 [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @use64(i64 [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext i32 %X to i64
|
||||
%B1 = and i64 %A1, 15
|
||||
%C1 = mul i64 %B1, %B1
|
||||
%T1 = trunc i64 %C1 to i32
|
||||
call i32 @use32(i32 %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call i32 @use64(i64 %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_uses_sub(i32 %X, i32 %Y) {
|
||||
; CHECK-LABEL: @multi_uses_sub(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext i32 [[X:%.*]] to i64
|
||||
; CHECK-NEXT: [[A2:%.*]] = zext i32 [[Y:%.*]] to i64
|
||||
; CHECK-NEXT: [[B1:%.*]] = sub i32 [[X]], [[Y]]
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul i32 [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @use32(i32 [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @use64(i64 [[A1]])
|
||||
; CHECK-NEXT: [[TMP3:%.*]] = call i32 @use64(i64 [[A2]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext i32 %X to i64
|
||||
%A2 = zext i32 %Y to i64
|
||||
%B1 = sub i64 %A1, %A2
|
||||
%C1 = mul i64 %B1, %B1
|
||||
%T1 = trunc i64 %C1 to i32
|
||||
call i32 @use32(i32 %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call i32 @use64(i64 %A1)
|
||||
call i32 @use64(i64 %A2)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_use_vec_add(<2 x i32> %X) {
|
||||
; CHECK-LABEL: @multi_use_vec_add(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext <2 x i32> [[X:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[B1:%.*]] = add <2 x i32> [[X]], <i32 15, i32 15>
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul <2 x i32> [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext <2 x i32> %X to <2 x i64>
|
||||
%B1 = add <2 x i64> %A1, <i64 15, i64 15>
|
||||
%C1 = mul <2 x i64> %B1, %B1
|
||||
%T1 = trunc <2 x i64> %C1 to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_use_vec_or(<2 x i32> %X) {
|
||||
; CHECK-LABEL: @multi_use_vec_or(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext <2 x i32> [[X:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[B1:%.*]] = or <2 x i32> [[X]], <i32 15, i32 15>
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul <2 x i32> [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext <2 x i32> %X to <2 x i64>
|
||||
%B1 = or <2 x i64> %A1, <i64 15, i64 15>
|
||||
%C1 = mul <2 x i64> %B1, %B1
|
||||
%T1 = trunc <2 x i64> %C1 to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_use_vec_xor(<2 x i32> %X) {
|
||||
; CHECK-LABEL: @multi_use_vec_xor(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext <2 x i32> [[X:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[B1:%.*]] = xor <2 x i32> [[X]], <i32 15, i32 15>
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul <2 x i32> [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext <2 x i32> %X to <2 x i64>
|
||||
%B1 = xor <2 x i64> %A1, <i64 15, i64 15>
|
||||
%C1 = mul <2 x i64> %B1, %B1
|
||||
%T1 = trunc <2 x i64> %C1 to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_use_vec_and(<2 x i32> %X) {
|
||||
; CHECK-LABEL: @multi_use_vec_and(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext <2 x i32> [[X:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[B1:%.*]] = and <2 x i32> [[X]], <i32 15, i32 15>
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul <2 x i32> [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A1]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext <2 x i32> %X to <2 x i64>
|
||||
%B1 = and <2 x i64> %A1, <i64 15, i64 15>
|
||||
%C1 = mul <2 x i64> %B1, %B1
|
||||
%T1 = trunc <2 x i64> %C1 to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A1)
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @multi_use_vec_sub(<2 x i32> %X, <2 x i32> %Y) {
|
||||
; CHECK-LABEL: @multi_use_vec_sub(
|
||||
; CHECK-NEXT: [[A1:%.*]] = zext <2 x i32> [[X:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[A2:%.*]] = zext <2 x i32> [[Y:%.*]] to <2 x i64>
|
||||
; CHECK-NEXT: [[B1:%.*]] = sub <2 x i32> [[X]], [[Y]]
|
||||
; CHECK-NEXT: [[C1:%.*]] = mul <2 x i32> [[B1]], [[B1]]
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = call <2 x i32> @use32_vec(<2 x i32> [[C1]])
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A1]])
|
||||
; CHECK-NEXT: [[TMP3:%.*]] = call <2 x i32> @use64_vec(<2 x i64> [[A2]])
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%A1 = zext <2 x i32> %X to <2 x i64>
|
||||
%A2 = zext <2 x i32> %Y to <2 x i64>
|
||||
%B1 = sub <2 x i64> %A1, %A2
|
||||
%C1 = mul <2 x i64> %B1, %B1
|
||||
%T1 = trunc <2 x i64> %C1 to <2 x i32>
|
||||
call <2 x i32> @use32_vec(<2 x i32> %T1)
|
||||
; make sure zext have another use that is not post-dominated by the TruncInst.
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A1)
|
||||
call <2 x i32> @use64_vec(<2 x i64> %A2)
|
||||
ret void
|
||||
}
|
|
@ -1,5 +1,6 @@
|
|||
set(LLVM_LINK_COMPONENTS
|
||||
${LLVM_TARGETS_TO_BUILD}
|
||||
AggressiveInstCombine
|
||||
Analysis
|
||||
BitWriter
|
||||
CodeGen
|
||||
|
|
|
@ -395,6 +395,7 @@ int main(int argc, char **argv) {
|
|||
initializeAnalysis(Registry);
|
||||
initializeTransformUtils(Registry);
|
||||
initializeInstCombine(Registry);
|
||||
initializeAggressiveInstCombinerLegacyPassPass(Registry);
|
||||
initializeInstrumentation(Registry);
|
||||
initializeTarget(Registry);
|
||||
// For codegen passes, only passes that do IR to IR transformation are
|
||||
|
|
Loading…
Reference in New Issue