2004-04-18 13:20:17 +08:00
|
|
|
//===-- LoopUnroll.cpp - Loop unroller pass -------------------------------===//
|
2005-04-22 07:48:37 +08:00
|
|
|
//
|
2004-04-18 13:20:17 +08:00
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
2007-12-30 04:36:04 +08:00
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
2005-04-22 07:48:37 +08:00
|
|
|
//
|
2004-04-18 13:20:17 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This pass implements a simple loop unroller. It works best when loops have
|
|
|
|
// been canonicalized by the -indvars pass, allowing it to determine the trip
|
|
|
|
// counts of loops easily.
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2016-07-20 07:54:23 +08:00
|
|
|
#include "llvm/Transforms/Scalar/LoopUnrollPass.h"
|
2015-02-13 11:57:40 +08:00
|
|
|
#include "llvm/ADT/SetVector.h"
|
2016-12-19 16:22:17 +08:00
|
|
|
#include "llvm/Analysis/AssumptionCache.h"
|
2011-01-02 15:35:53 +08:00
|
|
|
#include "llvm/Analysis/CodeMetrics.h"
|
2016-05-05 08:54:54 +08:00
|
|
|
#include "llvm/Analysis/GlobalsModRef.h"
|
2015-03-24 03:32:43 +08:00
|
|
|
#include "llvm/Analysis/InstructionSimplify.h"
|
2012-12-04 00:50:05 +08:00
|
|
|
#include "llvm/Analysis/LoopPass.h"
|
2016-02-09 07:03:59 +08:00
|
|
|
#include "llvm/Analysis/LoopUnrollAnalyzer.h"
|
2016-07-30 03:29:47 +08:00
|
|
|
#include "llvm/Analysis/OptimizationDiagnosticInfo.h"
|
2010-07-27 02:11:16 +08:00
|
|
|
#include "llvm/Analysis/ScalarEvolution.h"
|
2015-02-05 10:34:00 +08:00
|
|
|
#include "llvm/Analysis/ScalarEvolutionExpressions.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/DataLayout.h"
|
2014-01-13 17:26:24 +08:00
|
|
|
#include "llvm/IR/Dominators.h"
|
2015-03-24 03:32:43 +08:00
|
|
|
#include "llvm/IR/InstVisitor.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/IntrinsicInst.h"
|
2014-06-17 07:53:02 +08:00
|
|
|
#include "llvm/IR/Metadata.h"
|
2004-09-02 06:55:40 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
|
|
|
#include "llvm/Support/Debug.h"
|
2009-07-25 08:23:56 +08:00
|
|
|
#include "llvm/Support/raw_ostream.h"
|
2016-05-05 08:54:54 +08:00
|
|
|
#include "llvm/Transforms/Scalar.h"
|
2017-01-11 17:43:56 +08:00
|
|
|
#include "llvm/Transforms/Scalar/LoopPassManager.h"
|
[LPM] Factor all of the loop analysis usage updates into a common helper
routine.
We were getting this wrong in small ways and generally being very
inconsistent about it across loop passes. Instead, let's have a common
place where we do this. One minor downside is that this will require
some analyses like SCEV in more places than they are strictly needed.
However, this seems benign as these analyses are complete no-ops, and
without this consistency we can in many cases end up with the legacy
pass manager scheduling deciding to split up a loop pass pipeline in
order to run the function analysis half-way through. It is very, very
annoying to fix these without just being very pedantic across the board.
The only loop passes I've not updated here are ones that use
AU.setPreservesAll() such as IVUsers (an analysis) and the pass printer.
They seemed less relevant.
With this patch, almost all of the problems in PR24804 around loop pass
pipelines are fixed. The one remaining issue is that we run simplify-cfg
and instcombine in the middle of the loop pass pipeline. We've recently
added some loop variants of these passes that would seem substantially
cleaner to use, but this at least gets us much closer to the previous
state. Notably, the seven loop pass managers is down to three.
I've not updated the loop passes using LoopAccessAnalysis because that
analysis hasn't been fully wired into LoopSimplify/LCSSA, and it isn't
clear that those transforms want to support those forms anyways. They
all run late anyways, so this is harmless. Similarly, LSR is left alone
because it already carefully manages its forms and doesn't need to get
fused into a single loop pass manager with a bunch of other loop passes.
LoopReroll didn't use loop simplified form previously, and I've updated
the test case to match the trivially different output.
Finally, I've also factored all the pass initialization for the passes
that use this technique as well, so that should be done regularly and
reliably.
Thanks to James for the help reviewing and thinking about this stuff,
and Ben for help thinking about it as well!
Differential Revision: http://reviews.llvm.org/D17435
llvm-svn: 261316
2016-02-19 18:45:18 +08:00
|
|
|
#include "llvm/Transforms/Utils/LoopUtils.h"
|
2008-05-14 08:24:14 +08:00
|
|
|
#include "llvm/Transforms/Utils/UnrollLoop.h"
|
2008-05-16 17:30:00 +08:00
|
|
|
#include <climits>
|
2016-05-27 22:27:24 +08:00
|
|
|
#include <utility>
|
2004-04-18 13:20:17 +08:00
|
|
|
|
2008-05-14 08:24:14 +08:00
|
|
|
using namespace llvm;
|
2004-04-18 13:20:17 +08:00
|
|
|
|
2014-04-22 10:55:47 +08:00
|
|
|
#define DEBUG_TYPE "loop-unroll"
|
|
|
|
|
2008-05-13 08:00:25 +08:00
|
|
|
static cl::opt<unsigned>
|
2016-01-12 08:55:26 +08:00
|
|
|
UnrollThreshold("unroll-threshold", cl::Hidden,
|
2017-01-18 07:39:33 +08:00
|
|
|
cl::desc("The cost threshold for loop unrolling"));
|
|
|
|
|
|
|
|
static cl::opt<unsigned> UnrollPartialThreshold(
|
|
|
|
"unroll-partial-threshold", cl::Hidden,
|
|
|
|
cl::desc("The cost threshold for partial loop unrolling"));
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
|
2016-12-30 08:50:28 +08:00
|
|
|
static cl::opt<unsigned> UnrollMaxPercentThresholdBoost(
|
|
|
|
"unroll-max-percent-threshold-boost", cl::init(400), cl::Hidden,
|
|
|
|
cl::desc("The maximum 'boost' (represented as a percentage >= 100) applied "
|
|
|
|
"to the threshold when aggressively unrolling a loop due to the "
|
|
|
|
"dynamic cost savings. If completely unrolling a loop will reduce "
|
|
|
|
"the total runtime from X to Y, we boost the loop unroll "
|
|
|
|
"threshold to DefaultThreshold*std::min(MaxPercentThresholdBoost, "
|
|
|
|
"X/Y). This limit avoids excessive code bloat."));
|
2007-05-12 04:53:41 +08:00
|
|
|
|
2015-02-05 10:34:00 +08:00
|
|
|
static cl::opt<unsigned> UnrollMaxIterationsCountToAnalyze(
|
2016-05-25 07:00:05 +08:00
|
|
|
"unroll-max-iteration-count-to-analyze", cl::init(10), cl::Hidden,
|
2015-02-05 10:34:00 +08:00
|
|
|
cl::desc("Don't allow loop unrolling to simulate more than this number of"
|
|
|
|
"iterations when checking full unroll profitability"));
|
|
|
|
|
2016-05-05 08:54:54 +08:00
|
|
|
static cl::opt<unsigned> UnrollCount(
|
|
|
|
"unroll-count", cl::Hidden,
|
|
|
|
cl::desc("Use this unroll count for all loops including those with "
|
|
|
|
"unroll_count pragma values, for testing purposes"));
|
2004-04-18 13:20:17 +08:00
|
|
|
|
2016-05-05 08:54:54 +08:00
|
|
|
static cl::opt<unsigned> UnrollMaxCount(
|
|
|
|
"unroll-max-count", cl::Hidden,
|
|
|
|
cl::desc("Set the max unroll count for partial and runtime unrolling, for"
|
|
|
|
"testing purposes"));
|
2016-04-07 00:57:25 +08:00
|
|
|
|
2016-05-05 08:54:54 +08:00
|
|
|
static cl::opt<unsigned> UnrollFullMaxCount(
|
|
|
|
"unroll-full-max-count", cl::Hidden,
|
|
|
|
cl::desc(
|
|
|
|
"Set the max unroll count for full unrolling, for testing purposes"));
|
2016-04-07 00:57:25 +08:00
|
|
|
|
2008-07-29 21:21:23 +08:00
|
|
|
static cl::opt<bool>
|
2016-05-05 08:54:54 +08:00
|
|
|
UnrollAllowPartial("unroll-allow-partial", cl::Hidden,
|
|
|
|
cl::desc("Allows loops to be partially unrolled until "
|
|
|
|
"-unroll-threshold loop size is reached."));
|
2008-07-29 21:21:23 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
static cl::opt<bool> UnrollAllowRemainder(
|
|
|
|
"unroll-allow-remainder", cl::Hidden,
|
|
|
|
cl::desc("Allow generation of a loop remainder (extra iterations) "
|
|
|
|
"when unrolling a loop."));
|
|
|
|
|
2011-12-09 14:19:40 +08:00
|
|
|
static cl::opt<bool>
|
2016-05-05 08:54:54 +08:00
|
|
|
UnrollRuntime("unroll-runtime", cl::ZeroOrMore, cl::Hidden,
|
|
|
|
cl::desc("Unroll loops with run-time trip counts"));
|
2014-06-17 07:53:02 +08:00
|
|
|
|
2016-10-13 05:29:38 +08:00
|
|
|
static cl::opt<unsigned> UnrollMaxUpperBound(
|
|
|
|
"unroll-max-upperbound", cl::init(8), cl::Hidden,
|
|
|
|
cl::desc(
|
|
|
|
"The max of trip count upper bound that is considered in unrolling"));
|
|
|
|
|
2016-05-05 08:54:54 +08:00
|
|
|
static cl::opt<unsigned> PragmaUnrollThreshold(
|
|
|
|
"pragma-unroll-threshold", cl::init(16 * 1024), cl::Hidden,
|
|
|
|
cl::desc("Unrolled size limit for loops with an unroll(full) or "
|
|
|
|
"unroll_count pragma."));
|
2016-01-12 08:55:26 +08:00
|
|
|
|
2016-11-17 09:17:02 +08:00
|
|
|
static cl::opt<unsigned> FlatLoopTripCountThreshold(
|
|
|
|
"flat-loop-tripcount-threshold", cl::init(5), cl::Hidden,
|
|
|
|
cl::desc("If the runtime tripcount for the loop is lower than the "
|
|
|
|
"threshold, the loop is considered as flat and will be less "
|
|
|
|
"aggressively unrolled."));
|
|
|
|
|
2016-12-01 05:13:57 +08:00
|
|
|
static cl::opt<bool>
|
|
|
|
UnrollAllowPeeling("unroll-allow-peeling", cl::Hidden,
|
|
|
|
cl::desc("Allows loops to be peeled when the dynamic "
|
|
|
|
"trip count is known to be low."));
|
|
|
|
|
2016-01-12 08:55:26 +08:00
|
|
|
/// A magic value for use with the Threshold parameter to indicate
|
|
|
|
/// that the loop unroll should be performed regardless of how much
|
|
|
|
/// code expansion would result.
|
|
|
|
static const unsigned NoThreshold = UINT_MAX;
|
|
|
|
|
|
|
|
/// Gather the various unrolling parameters based on the defaults, compiler
|
2016-05-28 07:15:06 +08:00
|
|
|
/// flags, TTI overrides and user specified parameters.
|
2016-01-12 08:55:26 +08:00
|
|
|
static TargetTransformInfo::UnrollingPreferences gatherUnrollingPreferences(
|
|
|
|
Loop *L, const TargetTransformInfo &TTI, Optional<unsigned> UserThreshold,
|
|
|
|
Optional<unsigned> UserCount, Optional<bool> UserAllowPartial,
|
2016-10-13 05:29:38 +08:00
|
|
|
Optional<bool> UserRuntime, Optional<bool> UserUpperBound) {
|
2016-01-12 08:55:26 +08:00
|
|
|
TargetTransformInfo::UnrollingPreferences UP;
|
|
|
|
|
|
|
|
// Set up the defaults
|
|
|
|
UP.Threshold = 150;
|
2016-12-30 08:50:28 +08:00
|
|
|
UP.MaxPercentThresholdBoost = 400;
|
2016-05-11 05:45:55 +08:00
|
|
|
UP.OptSizeThreshold = 0;
|
2017-01-18 07:39:33 +08:00
|
|
|
UP.PartialThreshold = 150;
|
2016-05-11 05:45:55 +08:00
|
|
|
UP.PartialOptSizeThreshold = 0;
|
2016-01-12 08:55:26 +08:00
|
|
|
UP.Count = 0;
|
2016-12-01 05:13:57 +08:00
|
|
|
UP.PeelCount = 0;
|
2016-09-28 17:41:38 +08:00
|
|
|
UP.DefaultUnrollRuntimeCount = 8;
|
2016-01-12 08:55:26 +08:00
|
|
|
UP.MaxCount = UINT_MAX;
|
2016-04-07 00:57:25 +08:00
|
|
|
UP.FullUnrollMaxCount = UINT_MAX;
|
2016-11-10 03:56:39 +08:00
|
|
|
UP.BEInsns = 2;
|
2016-01-12 08:55:26 +08:00
|
|
|
UP.Partial = false;
|
|
|
|
UP.Runtime = false;
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.AllowRemainder = true;
|
2016-01-12 08:55:26 +08:00
|
|
|
UP.AllowExpensiveTripCount = false;
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.Force = false;
|
2016-10-13 05:29:38 +08:00
|
|
|
UP.UpperBound = false;
|
2016-12-01 05:13:57 +08:00
|
|
|
UP.AllowPeeling = false;
|
2016-01-12 08:55:26 +08:00
|
|
|
|
|
|
|
// Override with any target specific settings
|
|
|
|
TTI.getUnrollingPreferences(L, UP);
|
|
|
|
|
|
|
|
// Apply size attributes
|
|
|
|
if (L->getHeader()->getParent()->optForSize()) {
|
|
|
|
UP.Threshold = UP.OptSizeThreshold;
|
|
|
|
UP.PartialThreshold = UP.PartialOptSizeThreshold;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Apply any user values specified by cl::opt
|
2017-01-18 07:39:33 +08:00
|
|
|
if (UnrollThreshold.getNumOccurrences() > 0)
|
2016-01-12 08:55:26 +08:00
|
|
|
UP.Threshold = UnrollThreshold;
|
2017-01-18 07:39:33 +08:00
|
|
|
if (UnrollPartialThreshold.getNumOccurrences() > 0)
|
|
|
|
UP.PartialThreshold = UnrollPartialThreshold;
|
2016-12-30 08:50:28 +08:00
|
|
|
if (UnrollMaxPercentThresholdBoost.getNumOccurrences() > 0)
|
|
|
|
UP.MaxPercentThresholdBoost = UnrollMaxPercentThresholdBoost;
|
2016-04-07 00:57:25 +08:00
|
|
|
if (UnrollMaxCount.getNumOccurrences() > 0)
|
|
|
|
UP.MaxCount = UnrollMaxCount;
|
|
|
|
if (UnrollFullMaxCount.getNumOccurrences() > 0)
|
|
|
|
UP.FullUnrollMaxCount = UnrollFullMaxCount;
|
2016-01-12 08:55:26 +08:00
|
|
|
if (UnrollAllowPartial.getNumOccurrences() > 0)
|
|
|
|
UP.Partial = UnrollAllowPartial;
|
2016-05-28 07:15:06 +08:00
|
|
|
if (UnrollAllowRemainder.getNumOccurrences() > 0)
|
|
|
|
UP.AllowRemainder = UnrollAllowRemainder;
|
2016-01-12 08:55:26 +08:00
|
|
|
if (UnrollRuntime.getNumOccurrences() > 0)
|
|
|
|
UP.Runtime = UnrollRuntime;
|
2016-10-13 05:29:38 +08:00
|
|
|
if (UnrollMaxUpperBound == 0)
|
|
|
|
UP.UpperBound = false;
|
2016-12-01 05:13:57 +08:00
|
|
|
if (UnrollAllowPeeling.getNumOccurrences() > 0)
|
|
|
|
UP.AllowPeeling = UnrollAllowPeeling;
|
2016-01-12 08:55:26 +08:00
|
|
|
|
|
|
|
// Apply user values provided by argument
|
|
|
|
if (UserThreshold.hasValue()) {
|
|
|
|
UP.Threshold = *UserThreshold;
|
|
|
|
UP.PartialThreshold = *UserThreshold;
|
|
|
|
}
|
|
|
|
if (UserCount.hasValue())
|
|
|
|
UP.Count = *UserCount;
|
|
|
|
if (UserAllowPartial.hasValue())
|
|
|
|
UP.Partial = *UserAllowPartial;
|
|
|
|
if (UserRuntime.hasValue())
|
|
|
|
UP.Runtime = *UserRuntime;
|
2016-10-13 05:29:38 +08:00
|
|
|
if (UserUpperBound.hasValue())
|
|
|
|
UP.UpperBound = *UserUpperBound;
|
2016-01-12 08:55:26 +08:00
|
|
|
|
|
|
|
return UP;
|
|
|
|
}
|
|
|
|
|
2016-05-14 05:23:25 +08:00
|
|
|
namespace {
|
|
|
|
/// A struct to densely store the state of an instruction after unrolling at
|
|
|
|
/// each iteration.
|
|
|
|
///
|
|
|
|
/// This is designed to work like a tuple of <Instruction *, int> for the
|
|
|
|
/// purposes of hashing and lookup, but to be able to associate two boolean
|
|
|
|
/// states with each key.
|
|
|
|
struct UnrolledInstState {
|
|
|
|
Instruction *I;
|
|
|
|
int Iteration : 30;
|
|
|
|
unsigned IsFree : 1;
|
|
|
|
unsigned IsCounted : 1;
|
|
|
|
};
|
|
|
|
|
|
|
|
/// Hashing and equality testing for a set of the instruction states.
|
|
|
|
struct UnrolledInstStateKeyInfo {
|
|
|
|
typedef DenseMapInfo<Instruction *> PtrInfo;
|
|
|
|
typedef DenseMapInfo<std::pair<Instruction *, int>> PairInfo;
|
|
|
|
static inline UnrolledInstState getEmptyKey() {
|
|
|
|
return {PtrInfo::getEmptyKey(), 0, 0, 0};
|
|
|
|
}
|
|
|
|
static inline UnrolledInstState getTombstoneKey() {
|
|
|
|
return {PtrInfo::getTombstoneKey(), 0, 0, 0};
|
|
|
|
}
|
|
|
|
static inline unsigned getHashValue(const UnrolledInstState &S) {
|
|
|
|
return PairInfo::getHashValue({S.I, S.Iteration});
|
|
|
|
}
|
|
|
|
static inline bool isEqual(const UnrolledInstState &LHS,
|
|
|
|
const UnrolledInstState &RHS) {
|
|
|
|
return PairInfo::isEqual({LHS.I, LHS.Iteration}, {RHS.I, RHS.Iteration});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
namespace {
|
|
|
|
struct EstimatedUnrollCost {
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
/// \brief The estimated cost after unrolling.
|
2016-12-02 11:17:07 +08:00
|
|
|
unsigned UnrolledCost;
|
2015-02-13 10:10:56 +08:00
|
|
|
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
/// \brief The estimated dynamic cost of executing the instructions in the
|
|
|
|
/// rolled form.
|
2016-12-02 11:17:07 +08:00
|
|
|
unsigned RolledDynamicCost;
|
2015-05-23 01:41:35 +08:00
|
|
|
};
|
|
|
|
}
|
2015-05-13 01:20:03 +08:00
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
/// \brief Figure out if the loop is worth full unrolling.
|
|
|
|
///
|
|
|
|
/// Complete loop unrolling can make some loads constant, and we need to know
|
|
|
|
/// if that would expose any further optimization opportunities. This routine
|
2015-06-12 06:17:39 +08:00
|
|
|
/// estimates this optimization. It computes cost of unrolled loop
|
|
|
|
/// (UnrolledCost) and dynamic cost of the original loop (RolledDynamicCost). By
|
|
|
|
/// dynamic cost we mean that we won't count costs of blocks that are known not
|
|
|
|
/// to be executed (i.e. if we have a branch in the loop and we know that at the
|
|
|
|
/// given iteration its condition would be resolved to true, we won't add up the
|
|
|
|
/// cost of the 'false'-block).
|
|
|
|
/// \returns Optional value, holding the RolledDynamicCost and UnrolledCost. If
|
|
|
|
/// the analysis failed (no benefits expected from the unrolling, or the loop is
|
|
|
|
/// too big to analyze), the returned value is None.
|
2015-08-20 17:57:22 +08:00
|
|
|
static Optional<EstimatedUnrollCost>
|
2015-08-04 04:32:27 +08:00
|
|
|
analyzeLoopUnrollCost(const Loop *L, unsigned TripCount, DominatorTree &DT,
|
|
|
|
ScalarEvolution &SE, const TargetTransformInfo &TTI,
|
2016-12-02 11:17:07 +08:00
|
|
|
unsigned MaxUnrolledLoopSize) {
|
2015-05-23 01:41:35 +08:00
|
|
|
// We want to be able to scale offsets by the trip count and add more offsets
|
|
|
|
// to them without checking for overflows, and we already don't want to
|
|
|
|
// analyze *massive* trip counts, so we force the max to be reasonably small.
|
|
|
|
assert(UnrollMaxIterationsCountToAnalyze < (INT_MAX / 2) &&
|
|
|
|
"The unroll iterations max is too large!");
|
|
|
|
|
2016-05-14 05:23:25 +08:00
|
|
|
// Only analyze inner loops. We can't properly estimate cost of nested loops
|
|
|
|
// and we won't visit inner loops again anyway.
|
|
|
|
if (!L->empty())
|
|
|
|
return None;
|
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
// Don't simulate loops with a big or unknown tripcount
|
|
|
|
if (!UnrollMaxIterationsCountToAnalyze || !TripCount ||
|
|
|
|
TripCount > UnrollMaxIterationsCountToAnalyze)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
SmallSetVector<BasicBlock *, 16> BBWorklist;
|
2016-05-14 05:23:25 +08:00
|
|
|
SmallSetVector<std::pair<BasicBlock *, BasicBlock *>, 4> ExitWorklist;
|
2015-05-23 01:41:35 +08:00
|
|
|
DenseMap<Value *, Constant *> SimplifiedValues;
|
2015-08-04 04:32:27 +08:00
|
|
|
SmallVector<std::pair<Value *, Constant *>, 4> SimplifiedInputValues;
|
2015-02-13 10:17:39 +08:00
|
|
|
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
// The estimated cost of the unrolled form of the loop. We try to estimate
|
|
|
|
// this by simplifying as much as we can while computing the estimate.
|
2016-12-02 11:17:07 +08:00
|
|
|
unsigned UnrolledCost = 0;
|
2016-05-14 05:23:25 +08:00
|
|
|
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
// We also track the estimated dynamic (that is, actually executed) cost in
|
|
|
|
// the rolled form. This helps identify cases when the savings from unrolling
|
|
|
|
// aren't just exposing dead control flows, but actual reduced dynamic
|
|
|
|
// instructions due to the simplifications which we expect to occur after
|
|
|
|
// unrolling.
|
2016-12-02 11:17:07 +08:00
|
|
|
unsigned RolledDynamicCost = 0;
|
2015-05-23 01:41:35 +08:00
|
|
|
|
2016-05-14 05:23:25 +08:00
|
|
|
// We track the simplification of each instruction in each iteration. We use
|
|
|
|
// this to recursively merge costs into the unrolled cost on-demand so that
|
|
|
|
// we don't count the cost of any dead code. This is essentially a map from
|
|
|
|
// <instruction, int> to <bool, bool>, but stored as a densely packed struct.
|
|
|
|
DenseSet<UnrolledInstState, UnrolledInstStateKeyInfo> InstCostMap;
|
|
|
|
|
|
|
|
// A small worklist used to accumulate cost of instructions from each
|
|
|
|
// observable and reached root in the loop.
|
|
|
|
SmallVector<Instruction *, 16> CostWorklist;
|
|
|
|
|
|
|
|
// PHI-used worklist used between iterations while accumulating cost.
|
|
|
|
SmallVector<Instruction *, 4> PHIUsedList;
|
|
|
|
|
|
|
|
// Helper function to accumulate cost for instructions in the loop.
|
|
|
|
auto AddCostRecursively = [&](Instruction &RootI, int Iteration) {
|
|
|
|
assert(Iteration >= 0 && "Cannot have a negative iteration!");
|
|
|
|
assert(CostWorklist.empty() && "Must start with an empty cost list");
|
|
|
|
assert(PHIUsedList.empty() && "Must start with an empty phi used list");
|
|
|
|
CostWorklist.push_back(&RootI);
|
|
|
|
for (;; --Iteration) {
|
|
|
|
do {
|
|
|
|
Instruction *I = CostWorklist.pop_back_val();
|
|
|
|
|
|
|
|
// InstCostMap only uses I and Iteration as a key, the other two values
|
|
|
|
// don't matter here.
|
|
|
|
auto CostIter = InstCostMap.find({I, Iteration, 0, 0});
|
|
|
|
if (CostIter == InstCostMap.end())
|
|
|
|
// If an input to a PHI node comes from a dead path through the loop
|
|
|
|
// we may have no cost data for it here. What that actually means is
|
|
|
|
// that it is free.
|
|
|
|
continue;
|
|
|
|
auto &Cost = *CostIter;
|
|
|
|
if (Cost.IsCounted)
|
|
|
|
// Already counted this instruction.
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Mark that we are counting the cost of this instruction now.
|
|
|
|
Cost.IsCounted = true;
|
|
|
|
|
|
|
|
// If this is a PHI node in the loop header, just add it to the PHI set.
|
|
|
|
if (auto *PhiI = dyn_cast<PHINode>(I))
|
|
|
|
if (PhiI->getParent() == L->getHeader()) {
|
|
|
|
assert(Cost.IsFree && "Loop PHIs shouldn't be evaluated as they "
|
|
|
|
"inherently simplify during unrolling.");
|
|
|
|
if (Iteration == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Push the incoming value from the backedge into the PHI used list
|
|
|
|
// if it is an in-loop instruction. We'll use this to populate the
|
|
|
|
// cost worklist for the next iteration (as we count backwards).
|
|
|
|
if (auto *OpI = dyn_cast<Instruction>(
|
|
|
|
PhiI->getIncomingValueForBlock(L->getLoopLatch())))
|
|
|
|
if (L->contains(OpI))
|
|
|
|
PHIUsedList.push_back(OpI);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// First accumulate the cost of this instruction.
|
|
|
|
if (!Cost.IsFree) {
|
|
|
|
UnrolledCost += TTI.getUserCost(I);
|
|
|
|
DEBUG(dbgs() << "Adding cost of instruction (iteration " << Iteration
|
|
|
|
<< "): ");
|
|
|
|
DEBUG(I->dump());
|
|
|
|
}
|
|
|
|
|
|
|
|
// We must count the cost of every operand which is not free,
|
|
|
|
// recursively. If we reach a loop PHI node, simply add it to the set
|
|
|
|
// to be considered on the next iteration (backwards!).
|
|
|
|
for (Value *Op : I->operands()) {
|
|
|
|
// Check whether this operand is free due to being a constant or
|
|
|
|
// outside the loop.
|
|
|
|
auto *OpI = dyn_cast<Instruction>(Op);
|
|
|
|
if (!OpI || !L->contains(OpI))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
// Otherwise accumulate its cost.
|
|
|
|
CostWorklist.push_back(OpI);
|
|
|
|
}
|
|
|
|
} while (!CostWorklist.empty());
|
|
|
|
|
|
|
|
if (PHIUsedList.empty())
|
|
|
|
// We've exhausted the search.
|
|
|
|
break;
|
|
|
|
|
|
|
|
assert(Iteration > 0 &&
|
|
|
|
"Cannot track PHI-used values past the first iteration!");
|
|
|
|
CostWorklist.append(PHIUsedList.begin(), PHIUsedList.end());
|
|
|
|
PHIUsedList.clear();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2015-08-04 04:32:27 +08:00
|
|
|
// Ensure that we don't violate the loop structure invariants relied on by
|
|
|
|
// this analysis.
|
|
|
|
assert(L->isLoopSimplifyForm() && "Must put loop into normal form first.");
|
|
|
|
assert(L->isLCSSAForm(DT) &&
|
|
|
|
"Must have loops in LCSSA form to track live-out values.");
|
|
|
|
|
2015-07-29 04:07:29 +08:00
|
|
|
DEBUG(dbgs() << "Starting LoopUnroll profitability analysis...\n");
|
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
// Simulate execution of each iteration of the loop counting instructions,
|
|
|
|
// which would be simplified.
|
|
|
|
// Since the same load will take different values on different iterations,
|
|
|
|
// we literally have to go through all loop's iterations.
|
|
|
|
for (unsigned Iteration = 0; Iteration < TripCount; ++Iteration) {
|
2015-07-29 04:07:29 +08:00
|
|
|
DEBUG(dbgs() << " Analyzing iteration " << Iteration << "\n");
|
2015-08-04 04:32:27 +08:00
|
|
|
|
|
|
|
// Prepare for the iteration by collecting any simplified entry or backedge
|
|
|
|
// inputs.
|
|
|
|
for (Instruction &I : *L->getHeader()) {
|
|
|
|
auto *PHI = dyn_cast<PHINode>(&I);
|
|
|
|
if (!PHI)
|
|
|
|
break;
|
|
|
|
|
|
|
|
// The loop header PHI nodes must have exactly two input: one from the
|
|
|
|
// loop preheader and one from the loop latch.
|
|
|
|
assert(
|
|
|
|
PHI->getNumIncomingValues() == 2 &&
|
|
|
|
"Must have an incoming value only for the preheader and the latch.");
|
|
|
|
|
|
|
|
Value *V = PHI->getIncomingValueForBlock(
|
|
|
|
Iteration == 0 ? L->getLoopPreheader() : L->getLoopLatch());
|
|
|
|
Constant *C = dyn_cast<Constant>(V);
|
|
|
|
if (Iteration != 0 && !C)
|
|
|
|
C = SimplifiedValues.lookup(V);
|
|
|
|
if (C)
|
|
|
|
SimplifiedInputValues.push_back({PHI, C});
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now clear and re-populate the map for the next iteration.
|
2015-05-23 01:41:35 +08:00
|
|
|
SimplifiedValues.clear();
|
2015-08-04 04:32:27 +08:00
|
|
|
while (!SimplifiedInputValues.empty())
|
|
|
|
SimplifiedValues.insert(SimplifiedInputValues.pop_back_val());
|
|
|
|
|
2016-02-26 10:57:05 +08:00
|
|
|
UnrolledInstAnalyzer Analyzer(Iteration, SimplifiedValues, SE, L);
|
2015-05-23 01:41:35 +08:00
|
|
|
|
|
|
|
BBWorklist.clear();
|
|
|
|
BBWorklist.insert(L->getHeader());
|
|
|
|
// Note that we *must not* cache the size, this loop grows the worklist.
|
|
|
|
for (unsigned Idx = 0; Idx != BBWorklist.size(); ++Idx) {
|
|
|
|
BasicBlock *BB = BBWorklist[Idx];
|
|
|
|
|
|
|
|
// Visit all instructions in the given basic block and try to simplify
|
|
|
|
// it. We don't change the actual IR, just count optimization
|
|
|
|
// opportunities.
|
|
|
|
for (Instruction &I : *BB) {
|
2016-10-01 02:30:04 +08:00
|
|
|
if (isa<DbgInfoIntrinsic>(I))
|
|
|
|
continue;
|
|
|
|
|
2016-05-14 05:23:25 +08:00
|
|
|
// Track this instruction's expected baseline cost when executing the
|
|
|
|
// rolled loop form.
|
|
|
|
RolledDynamicCost += TTI.getUserCost(&I);
|
2015-05-23 01:41:35 +08:00
|
|
|
|
|
|
|
// Visit the instruction to analyze its loop cost after unrolling,
|
2016-05-14 05:23:25 +08:00
|
|
|
// and if the visitor returns true, mark the instruction as free after
|
|
|
|
// unrolling and continue.
|
|
|
|
bool IsFree = Analyzer.visit(I);
|
|
|
|
bool Inserted = InstCostMap.insert({&I, (int)Iteration,
|
|
|
|
(unsigned)IsFree,
|
|
|
|
/*IsCounted*/ false}).second;
|
|
|
|
(void)Inserted;
|
|
|
|
assert(Inserted && "Cannot have a state for an unvisited instruction!");
|
|
|
|
|
|
|
|
if (IsFree)
|
|
|
|
continue;
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
|
2016-05-14 05:23:25 +08:00
|
|
|
// Can't properly model a cost of a call.
|
|
|
|
// FIXME: With a proper cost model we should be able to do it.
|
|
|
|
if(isa<CallInst>(&I))
|
|
|
|
return None;
|
2015-05-23 01:41:35 +08:00
|
|
|
|
2016-08-18 06:42:58 +08:00
|
|
|
// If the instruction might have a side-effect recursively account for
|
|
|
|
// the cost of it and all the instructions leading up to it.
|
|
|
|
if (I.mayHaveSideEffects())
|
|
|
|
AddCostRecursively(I, Iteration);
|
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
// If unrolled body turns out to be too big, bail out.
|
2015-07-29 04:07:29 +08:00
|
|
|
if (UnrolledCost > MaxUnrolledLoopSize) {
|
|
|
|
DEBUG(dbgs() << " Exceeded threshold.. exiting.\n"
|
|
|
|
<< " UnrolledCost: " << UnrolledCost
|
|
|
|
<< ", MaxUnrolledLoopSize: " << MaxUnrolledLoopSize
|
|
|
|
<< "\n");
|
2015-05-23 01:41:35 +08:00
|
|
|
return None;
|
2015-07-29 04:07:29 +08:00
|
|
|
}
|
[unroll] Make the unroll cost analysis terminate deterministically and
reasonably quickly.
I don't have a reduced test case, but for a version of FFMPEG, this
makes the loop unroller start finishing at all (after over 15 minutes of
running, it hadn't terminated for me, no idea if it was a true infloop
or just exponential work).
The key thing here is to check the DeadInstructions set when pulling
things off the worklist. Without this, we would re-walk the user list of
already dead instructions again and again and again. Consider phi nodes
with many, many operands and other patterns.
The other important aspect of this is that because we would keep
re-visiting instructions that were already known dead, we kept adding
their cost savings to this! This would cause our cost savings to be
*insanely* inflated from this.
While I was here, I also rotated the operand walk out of the worklist
loop to make the code easier to read. There is still work to be done to
minimize worklist traffic because we don't de-duplicate operands. This
means we may add the same instruction onto the worklist 1000s of times
if it shows up in 1000s of operansd to a PHI node for example.
Still, with this patch, the ffmpeg testcase I have finishes quickly and
I can't measure the runtime impact of the unroll analysis any more. I'll
probably try to do a few more cleanups to this code, but not sure how
much cleanup I can justify right now.
llvm-svn: 229038
2015-02-13 11:40:58 +08:00
|
|
|
}
|
2015-05-13 01:20:03 +08:00
|
|
|
|
2015-07-24 09:53:04 +08:00
|
|
|
TerminatorInst *TI = BB->getTerminator();
|
|
|
|
|
|
|
|
// Add in the live successors by first checking whether we have terminator
|
|
|
|
// that may be simplified based on the values simplified by this call.
|
2016-05-27 05:42:51 +08:00
|
|
|
BasicBlock *KnownSucc = nullptr;
|
2015-07-24 09:53:04 +08:00
|
|
|
if (BranchInst *BI = dyn_cast<BranchInst>(TI)) {
|
|
|
|
if (BI->isConditional()) {
|
|
|
|
if (Constant *SimpleCond =
|
|
|
|
SimplifiedValues.lookup(BI->getCondition())) {
|
2015-07-30 02:10:29 +08:00
|
|
|
// Just take the first successor if condition is undef
|
|
|
|
if (isa<UndefValue>(SimpleCond))
|
2016-05-27 05:42:51 +08:00
|
|
|
KnownSucc = BI->getSuccessor(0);
|
|
|
|
else if (ConstantInt *SimpleCondVal =
|
|
|
|
dyn_cast<ConstantInt>(SimpleCond))
|
|
|
|
KnownSucc = BI->getSuccessor(SimpleCondVal->isZero() ? 1 : 0);
|
2015-07-24 09:53:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
} else if (SwitchInst *SI = dyn_cast<SwitchInst>(TI)) {
|
|
|
|
if (Constant *SimpleCond =
|
|
|
|
SimplifiedValues.lookup(SI->getCondition())) {
|
2015-07-30 02:10:29 +08:00
|
|
|
// Just take the first successor if condition is undef
|
|
|
|
if (isa<UndefValue>(SimpleCond))
|
2016-05-27 05:42:51 +08:00
|
|
|
KnownSucc = SI->getSuccessor(0);
|
|
|
|
else if (ConstantInt *SimpleCondVal =
|
|
|
|
dyn_cast<ConstantInt>(SimpleCond))
|
|
|
|
KnownSucc = SI->findCaseValue(SimpleCondVal).getCaseSuccessor();
|
2015-07-24 09:53:04 +08:00
|
|
|
}
|
|
|
|
}
|
2016-05-27 05:42:51 +08:00
|
|
|
if (KnownSucc) {
|
|
|
|
if (L->contains(KnownSucc))
|
|
|
|
BBWorklist.insert(KnownSucc);
|
|
|
|
else
|
|
|
|
ExitWorklist.insert({BB, KnownSucc});
|
|
|
|
continue;
|
|
|
|
}
|
2015-07-24 09:53:04 +08:00
|
|
|
|
2015-05-23 01:41:35 +08:00
|
|
|
// Add BB's successors to the worklist.
|
|
|
|
for (BasicBlock *Succ : successors(BB))
|
|
|
|
if (L->contains(Succ))
|
|
|
|
BBWorklist.insert(Succ);
|
2016-05-14 05:23:25 +08:00
|
|
|
else
|
|
|
|
ExitWorklist.insert({BB, Succ});
|
2016-05-19 05:20:12 +08:00
|
|
|
AddCostRecursively(*TI, Iteration);
|
2015-02-05 10:34:00 +08:00
|
|
|
}
|
2015-05-23 01:41:35 +08:00
|
|
|
|
|
|
|
// If we found no optimization opportunities on the first iteration, we
|
|
|
|
// won't find them on later ones too.
|
2015-07-29 04:07:29 +08:00
|
|
|
if (UnrolledCost == RolledDynamicCost) {
|
|
|
|
DEBUG(dbgs() << " No opportunities found.. exiting.\n"
|
|
|
|
<< " UnrolledCost: " << UnrolledCost << "\n");
|
2015-05-23 01:41:35 +08:00
|
|
|
return None;
|
2015-07-29 04:07:29 +08:00
|
|
|
}
|
2015-02-05 10:34:00 +08:00
|
|
|
}
|
2016-05-14 05:23:25 +08:00
|
|
|
|
|
|
|
while (!ExitWorklist.empty()) {
|
|
|
|
BasicBlock *ExitingBB, *ExitBB;
|
|
|
|
std::tie(ExitingBB, ExitBB) = ExitWorklist.pop_back_val();
|
|
|
|
|
|
|
|
for (Instruction &I : *ExitBB) {
|
|
|
|
auto *PN = dyn_cast<PHINode>(&I);
|
|
|
|
if (!PN)
|
|
|
|
break;
|
|
|
|
|
|
|
|
Value *Op = PN->getIncomingValueForBlock(ExitingBB);
|
|
|
|
if (auto *OpI = dyn_cast<Instruction>(Op))
|
|
|
|
if (L->contains(OpI))
|
|
|
|
AddCostRecursively(*OpI, TripCount - 1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-07-29 04:07:29 +08:00
|
|
|
DEBUG(dbgs() << "Analysis finished:\n"
|
|
|
|
<< "UnrolledCost: " << UnrolledCost << ", "
|
|
|
|
<< "RolledDynamicCost: " << RolledDynamicCost << "\n");
|
[Unroll] Rework the naming and structure of the new unroll heuristics.
The new naming is (to me) much easier to understand. Here is a summary
of the new state of the world:
- '*Threshold' is the threshold for full unrolling. It is measured
against the estimated unrolled cost as computed by getUserCost in TTI
(or CodeMetrics, etc). We will exceed this threshold when unrolling
loops where unrolling exposes a significant degree of simplification
of the logic within the loop.
- '*PercentDynamicCostSavedThreshold' is the percentage of the loop's
estimated dynamic execution cost which needs to be saved by unrolling
to apply a discount to the estimated unrolled cost.
- '*DynamicCostSavingsDiscount' is the discount applied to the estimated
unrolling cost when the dynamic savings are expected to be high.
When actually analyzing the loop, we now produce both an estimated
unrolled cost, and an estimated rolled cost. The rolled cost is notably
a dynamic estimate based on our analysis of the expected execution of
each iteration.
While we're still working to build up the infrastructure for making
these estimates, to me it is much more clear *how* to make them better
when they have reasonably descriptive names. For example, we may want to
apply estimated (from heuristics or profiles) dynamic execution weights
to the *dynamic* cost estimates. If we start doing that, we would also
need to track the static unrolled cost and the dynamic unrolled cost, as
only the latter could reasonably be weighted by profile information.
This patch is sadly not without functionality change for the new unroll
analysis logic. Buried in the heuristic management were several things
that surprised me. For example, we never subtracted the optimized
instruction count off when comparing against the unroll heursistics!
I don't know if this just got lost somewhere along the way or what, but
with the new accounting of things, this is much easier to keep track of
and we use the post-simplification cost estimate to compare to the
thresholds, and use the dynamic cost reduction ratio to select whether
we can exceed the baseline threshold.
The old values of these flags also don't necessarily make sense. My
impression is that none of these thresholds or discounts have been tuned
yet, and so they're just arbitrary placehold numbers. As such, I've not
bothered to adjust for the fact that this is now a discount and not
a tow-tier threshold model. We need to tune all these values once the
logic is ready to be enabled.
Differential Revision: http://reviews.llvm.org/D9966
llvm-svn: 239164
2015-06-06 01:01:43 +08:00
|
|
|
return {{UnrolledCost, RolledDynamicCost}};
|
2015-05-23 01:41:35 +08:00
|
|
|
}
|
2015-02-05 10:34:00 +08:00
|
|
|
|
2007-05-08 23:14:19 +08:00
|
|
|
/// ApproximateLoopSize - Approximate the size of the loop.
|
2011-10-01 09:39:05 +08:00
|
|
|
static unsigned ApproximateLoopSize(const Loop *L, unsigned &NumCalls,
|
2016-03-15 07:15:34 +08:00
|
|
|
bool &NotDuplicatable, bool &Convergent,
|
2014-09-07 21:49:57 +08:00
|
|
|
const TargetTransformInfo &TTI,
|
2016-12-19 16:22:17 +08:00
|
|
|
AssumptionCache *AC, unsigned BEInsns) {
|
2014-09-07 21:49:57 +08:00
|
|
|
SmallPtrSet<const Value *, 32> EphValues;
|
2016-12-19 16:22:17 +08:00
|
|
|
CodeMetrics::collectEphemeralValues(L, AC, EphValues);
|
2014-09-07 21:49:57 +08:00
|
|
|
|
2009-10-31 22:54:17 +08:00
|
|
|
CodeMetrics Metrics;
|
2016-03-09 03:06:12 +08:00
|
|
|
for (BasicBlock *BB : L->blocks())
|
|
|
|
Metrics.analyzeBasicBlock(BB, TTI, EphValues);
|
2010-09-10 04:32:23 +08:00
|
|
|
NumCalls = Metrics.NumInlineCandidates;
|
2012-12-21 00:04:27 +08:00
|
|
|
NotDuplicatable = Metrics.notDuplicatable;
|
2016-03-15 07:15:34 +08:00
|
|
|
Convergent = Metrics.convergent;
|
2011-07-23 08:29:16 +08:00
|
|
|
|
2010-09-10 03:07:31 +08:00
|
|
|
unsigned LoopSize = Metrics.NumInsts;
|
2011-07-23 08:29:16 +08:00
|
|
|
|
2010-09-10 03:07:31 +08:00
|
|
|
// Don't allow an estimate of size zero. This would allows unrolling of loops
|
|
|
|
// with huge iteration counts, which is a compile time problem even if it's
|
2015-01-10 08:30:55 +08:00
|
|
|
// not a problem for code quality. Also, the code using this size may assume
|
|
|
|
// that each loop has at least three instructions (likely a conditional
|
|
|
|
// branch, a comparison feeding that branch, and some kind of loop increment
|
|
|
|
// feeding that comparison instruction).
|
2016-11-10 03:56:39 +08:00
|
|
|
LoopSize = std::max(LoopSize, BEInsns + 1);
|
2011-07-23 08:29:16 +08:00
|
|
|
|
2010-09-10 03:07:31 +08:00
|
|
|
return LoopSize;
|
2004-04-18 13:20:17 +08:00
|
|
|
}
|
|
|
|
|
2014-07-24 01:31:37 +08:00
|
|
|
// Returns the loop hint metadata node with the given name (for example,
|
|
|
|
// "llvm.loop.unroll.count"). If no such metadata node exists, then nullptr is
|
|
|
|
// returned.
|
2015-02-03 04:41:11 +08:00
|
|
|
static MDNode *GetUnrollMetadataForLoop(const Loop *L, StringRef Name) {
|
|
|
|
if (MDNode *LoopID = L->getLoopID())
|
|
|
|
return GetUnrollMetadata(LoopID, Name);
|
|
|
|
return nullptr;
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
|
|
|
|
2014-07-24 01:31:37 +08:00
|
|
|
// Returns true if the loop has an unroll(full) pragma.
|
|
|
|
static bool HasUnrollFullPragma(const Loop *L) {
|
2015-02-01 10:27:45 +08:00
|
|
|
return GetUnrollMetadataForLoop(L, "llvm.loop.unroll.full");
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
|
|
|
|
2015-08-11 01:28:08 +08:00
|
|
|
// Returns true if the loop has an unroll(enable) pragma. This metadata is used
|
|
|
|
// for both "#pragma unroll" and "#pragma clang loop unroll(enable)" directives.
|
|
|
|
static bool HasUnrollEnablePragma(const Loop *L) {
|
|
|
|
return GetUnrollMetadataForLoop(L, "llvm.loop.unroll.enable");
|
|
|
|
}
|
|
|
|
|
2014-06-17 07:53:02 +08:00
|
|
|
// Returns true if the loop has an unroll(disable) pragma.
|
|
|
|
static bool HasUnrollDisablePragma(const Loop *L) {
|
2015-02-01 10:27:45 +08:00
|
|
|
return GetUnrollMetadataForLoop(L, "llvm.loop.unroll.disable");
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
|
|
|
|
2015-03-09 14:14:18 +08:00
|
|
|
// Returns true if the loop has an runtime unroll(disable) pragma.
|
|
|
|
static bool HasRuntimeUnrollDisablePragma(const Loop *L) {
|
|
|
|
return GetUnrollMetadataForLoop(L, "llvm.loop.unroll.runtime.disable");
|
|
|
|
}
|
|
|
|
|
2014-06-17 07:53:02 +08:00
|
|
|
// If loop has an unroll_count pragma return the (necessarily
|
|
|
|
// positive) value from the pragma. Otherwise return 0.
|
|
|
|
static unsigned UnrollCountPragmaValue(const Loop *L) {
|
2015-02-03 04:41:11 +08:00
|
|
|
MDNode *MD = GetUnrollMetadataForLoop(L, "llvm.loop.unroll.count");
|
2014-07-24 01:31:37 +08:00
|
|
|
if (MD) {
|
|
|
|
assert(MD->getNumOperands() == 2 &&
|
|
|
|
"Unroll count hint metadata should have two operands.");
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
unsigned Count =
|
|
|
|
mdconst::extract<ConstantInt>(MD->getOperand(1))->getZExtValue();
|
2014-06-17 07:53:02 +08:00
|
|
|
assert(Count >= 1 && "Unroll count must be positive.");
|
|
|
|
return Count;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-07-19 05:04:33 +08:00
|
|
|
// Remove existing unroll metadata and add unroll disable metadata to
|
|
|
|
// indicate the loop has already been unrolled. This prevents a loop
|
|
|
|
// from being unrolled more than is directed by a pragma if the loop
|
|
|
|
// unrolling pass is run more than once (which it generally is).
|
|
|
|
static void SetLoopAlreadyUnrolled(Loop *L) {
|
|
|
|
MDNode *LoopID = L->getLoopID();
|
|
|
|
// First remove any existing loop unrolling metadata.
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
SmallVector<Metadata *, 4> MDs;
|
2014-07-19 05:04:33 +08:00
|
|
|
// Reserve first location for self reference to the LoopID metadata node.
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
MDs.push_back(nullptr);
|
2016-06-09 04:21:24 +08:00
|
|
|
|
|
|
|
if (LoopID) {
|
|
|
|
for (unsigned i = 1, ie = LoopID->getNumOperands(); i < ie; ++i) {
|
|
|
|
bool IsUnrollMetadata = false;
|
|
|
|
MDNode *MD = dyn_cast<MDNode>(LoopID->getOperand(i));
|
|
|
|
if (MD) {
|
|
|
|
const MDString *S = dyn_cast<MDString>(MD->getOperand(0));
|
|
|
|
IsUnrollMetadata = S && S->getString().startswith("llvm.loop.unroll.");
|
|
|
|
}
|
|
|
|
if (!IsUnrollMetadata)
|
|
|
|
MDs.push_back(LoopID->getOperand(i));
|
2014-07-19 05:04:33 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add unroll(disable) metadata to disable future unrolling.
|
|
|
|
LLVMContext &Context = L->getHeader()->getContext();
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
SmallVector<Metadata *, 1> DisableOperands;
|
2014-07-24 01:31:37 +08:00
|
|
|
DisableOperands.push_back(MDString::get(Context, "llvm.loop.unroll.disable"));
|
2014-07-19 05:29:41 +08:00
|
|
|
MDNode *DisableNode = MDNode::get(Context, DisableOperands);
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
MDs.push_back(DisableNode);
|
2014-07-19 05:04:33 +08:00
|
|
|
|
IR: Split Metadata from Value
Split `Metadata` away from the `Value` class hierarchy, as part of
PR21532. Assembly and bitcode changes are in the wings, but this is the
bulk of the change for the IR C++ API.
I have a follow-up patch prepared for `clang`. If this breaks other
sub-projects, I apologize in advance :(. Help me compile it on Darwin
I'll try to fix it. FWIW, the errors should be easy to fix, so it may
be simpler to just fix it yourself.
This breaks the build for all metadata-related code that's out-of-tree.
Rest assured the transition is mechanical and the compiler should catch
almost all of the problems.
Here's a quick guide for updating your code:
- `Metadata` is the root of a class hierarchy with three main classes:
`MDNode`, `MDString`, and `ValueAsMetadata`. It is distinct from
the `Value` class hierarchy. It is typeless -- i.e., instances do
*not* have a `Type`.
- `MDNode`'s operands are all `Metadata *` (instead of `Value *`).
- `TrackingVH<MDNode>` and `WeakVH` referring to metadata can be
replaced with `TrackingMDNodeRef` and `TrackingMDRef`, respectively.
If you're referring solely to resolved `MDNode`s -- post graph
construction -- just use `MDNode*`.
- `MDNode` (and the rest of `Metadata`) have only limited support for
`replaceAllUsesWith()`.
As long as an `MDNode` is pointing at a forward declaration -- the
result of `MDNode::getTemporary()` -- it maintains a side map of its
uses and can RAUW itself. Once the forward declarations are fully
resolved RAUW support is dropped on the ground. This means that
uniquing collisions on changing operands cause nodes to become
"distinct". (This already happened fairly commonly, whenever an
operand went to null.)
If you're constructing complex (non self-reference) `MDNode` cycles,
you need to call `MDNode::resolveCycles()` on each node (or on a
top-level node that somehow references all of the nodes). Also,
don't do that. Metadata cycles (and the RAUW machinery needed to
construct them) are expensive.
- An `MDNode` can only refer to a `Constant` through a bridge called
`ConstantAsMetadata` (one of the subclasses of `ValueAsMetadata`).
As a side effect, accessing an operand of an `MDNode` that is known
to be, e.g., `ConstantInt`, takes three steps: first, cast from
`Metadata` to `ConstantAsMetadata`; second, extract the `Constant`;
third, cast down to `ConstantInt`.
The eventual goal is to introduce `MDInt`/`MDFloat`/etc. and have
metadata schema owners transition away from using `Constant`s when
the type isn't important (and they don't care about referring to
`GlobalValue`s).
In the meantime, I've added transitional API to the `mdconst`
namespace that matches semantics with the old code, in order to
avoid adding the error-prone three-step equivalent to every call
site. If your old code was:
MDNode *N = foo();
bar(isa <ConstantInt>(N->getOperand(0)));
baz(cast <ConstantInt>(N->getOperand(1)));
bak(cast_or_null <ConstantInt>(N->getOperand(2)));
bat(dyn_cast <ConstantInt>(N->getOperand(3)));
bay(dyn_cast_or_null<ConstantInt>(N->getOperand(4)));
you can trivially match its semantics with:
MDNode *N = foo();
bar(mdconst::hasa <ConstantInt>(N->getOperand(0)));
baz(mdconst::extract <ConstantInt>(N->getOperand(1)));
bak(mdconst::extract_or_null <ConstantInt>(N->getOperand(2)));
bat(mdconst::dyn_extract <ConstantInt>(N->getOperand(3)));
bay(mdconst::dyn_extract_or_null<ConstantInt>(N->getOperand(4)));
and when you transition your metadata schema to `MDInt`:
MDNode *N = foo();
bar(isa <MDInt>(N->getOperand(0)));
baz(cast <MDInt>(N->getOperand(1)));
bak(cast_or_null <MDInt>(N->getOperand(2)));
bat(dyn_cast <MDInt>(N->getOperand(3)));
bay(dyn_cast_or_null<MDInt>(N->getOperand(4)));
- A `CallInst` -- specifically, intrinsic instructions -- can refer to
metadata through a bridge called `MetadataAsValue`. This is a
subclass of `Value` where `getType()->isMetadataTy()`.
`MetadataAsValue` is the *only* class that can legally refer to a
`LocalAsMetadata`, which is a bridged form of non-`Constant` values
like `Argument` and `Instruction`. It can also refer to any other
`Metadata` subclass.
(I'll break all your testcases in a follow-up commit, when I propagate
this change to assembly.)
llvm-svn: 223802
2014-12-10 02:38:53 +08:00
|
|
|
MDNode *NewLoopID = MDNode::get(Context, MDs);
|
2014-07-19 05:04:33 +08:00
|
|
|
// Set operand 0 to refer to the loop id itself.
|
|
|
|
NewLoopID->replaceOperandWith(0, NewLoopID);
|
|
|
|
L->setLoopID(NewLoopID);
|
|
|
|
}
|
|
|
|
|
2016-12-30 08:50:28 +08:00
|
|
|
// Computes the boosting factor for complete unrolling.
|
|
|
|
// If fully unrolling the loop would save a lot of RolledDynamicCost, it would
|
|
|
|
// be beneficial to fully unroll the loop even if unrolledcost is large. We
|
|
|
|
// use (RolledDynamicCost / UnrolledCost) to model the unroll benefits to adjust
|
|
|
|
// the unroll threshold.
|
|
|
|
static unsigned getFullUnrollBoostingFactor(const EstimatedUnrollCost &Cost,
|
|
|
|
unsigned MaxPercentThresholdBoost) {
|
|
|
|
if (Cost.RolledDynamicCost >= UINT_MAX / 100)
|
|
|
|
return 100;
|
|
|
|
else if (Cost.UnrolledCost != 0)
|
|
|
|
// The boosting factor is RolledDynamicCost / UnrolledCost
|
|
|
|
return std::min(100 * Cost.RolledDynamicCost / Cost.UnrolledCost,
|
|
|
|
MaxPercentThresholdBoost);
|
|
|
|
else
|
|
|
|
return MaxPercentThresholdBoost;
|
2015-05-13 01:20:03 +08:00
|
|
|
}
|
|
|
|
|
2016-11-10 03:56:39 +08:00
|
|
|
// Returns loop size estimation for unrolled loop.
|
|
|
|
static uint64_t getUnrolledLoopSize(
|
|
|
|
unsigned LoopSize,
|
|
|
|
TargetTransformInfo::UnrollingPreferences &UP) {
|
|
|
|
assert(LoopSize >= UP.BEInsns && "LoopSize should not be less than BEInsns!");
|
|
|
|
return (uint64_t)(LoopSize - UP.BEInsns) * UP.Count + UP.BEInsns;
|
|
|
|
}
|
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// Returns true if unroll count was set explicitly.
|
|
|
|
// Calculates unroll count and writes it to UP.Count.
|
2016-10-13 05:29:38 +08:00
|
|
|
static bool computeUnrollCount(
|
|
|
|
Loop *L, const TargetTransformInfo &TTI, DominatorTree &DT, LoopInfo *LI,
|
|
|
|
ScalarEvolution *SE, OptimizationRemarkEmitter *ORE, unsigned &TripCount,
|
|
|
|
unsigned MaxTripCount, unsigned &TripMultiple, unsigned LoopSize,
|
|
|
|
TargetTransformInfo::UnrollingPreferences &UP, bool &UseUpperBound) {
|
2016-05-28 07:15:06 +08:00
|
|
|
// Check for explicit Count.
|
|
|
|
// 1st priority is unroll count set by "unroll-count" option.
|
|
|
|
bool UserUnrollCount = UnrollCount.getNumOccurrences() > 0;
|
|
|
|
if (UserUnrollCount) {
|
|
|
|
UP.Count = UnrollCount;
|
|
|
|
UP.AllowExpensiveTripCount = true;
|
|
|
|
UP.Force = true;
|
2016-11-10 03:56:39 +08:00
|
|
|
if (UP.AllowRemainder && getUnrolledLoopSize(LoopSize, UP) < UP.Threshold)
|
2016-05-28 07:15:06 +08:00
|
|
|
return true;
|
|
|
|
}
|
2011-07-23 08:29:16 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// 2nd priority is unroll count set by pragma.
|
|
|
|
unsigned PragmaCount = UnrollCountPragmaValue(L);
|
|
|
|
if (PragmaCount > 0) {
|
|
|
|
UP.Count = PragmaCount;
|
|
|
|
UP.Runtime = true;
|
|
|
|
UP.AllowExpensiveTripCount = true;
|
|
|
|
UP.Force = true;
|
|
|
|
if (UP.AllowRemainder &&
|
2016-11-10 03:56:39 +08:00
|
|
|
getUnrolledLoopSize(LoopSize, UP) < PragmaUnrollThreshold)
|
2016-05-28 07:15:06 +08:00
|
|
|
return true;
|
2014-04-02 02:50:30 +08:00
|
|
|
}
|
2014-07-24 01:31:37 +08:00
|
|
|
bool PragmaFullUnroll = HasUnrollFullPragma(L);
|
2016-05-28 07:15:06 +08:00
|
|
|
if (PragmaFullUnroll && TripCount != 0) {
|
|
|
|
UP.Count = TripCount;
|
2016-11-10 03:56:39 +08:00
|
|
|
if (getUnrolledLoopSize(LoopSize, UP) < PragmaUnrollThreshold)
|
2016-05-28 07:15:06 +08:00
|
|
|
return false;
|
2011-08-12 07:36:16 +08:00
|
|
|
}
|
2013-09-12 03:25:43 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
bool PragmaEnableUnroll = HasUnrollEnablePragma(L);
|
|
|
|
bool ExplicitUnroll = PragmaCount > 0 || PragmaFullUnroll ||
|
|
|
|
PragmaEnableUnroll || UserUnrollCount;
|
2014-06-17 07:53:02 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
if (ExplicitUnroll && TripCount != 0) {
|
|
|
|
// If the loop has an unrolling pragma, we want to be more aggressive with
|
|
|
|
// unrolling limits. Set thresholds to at least the PragmaThreshold value
|
|
|
|
// which is larger than the default limits.
|
|
|
|
UP.Threshold = std::max<unsigned>(UP.Threshold, PragmaUnrollThreshold);
|
|
|
|
UP.PartialThreshold =
|
|
|
|
std::max<unsigned>(UP.PartialThreshold, PragmaUnrollThreshold);
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
2011-12-09 14:19:40 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// 3rd priority is full unroll count.
|
2016-10-13 05:29:38 +08:00
|
|
|
// Full unroll makes sense only when TripCount or its upper bound could be
|
|
|
|
// statically calculated.
|
2016-05-28 07:15:06 +08:00
|
|
|
// Also we need to check if we exceed FullUnrollMaxCount.
|
2016-10-13 05:29:38 +08:00
|
|
|
// If using the upper bound to unroll, TripMultiple should be set to 1 because
|
|
|
|
// we do not know when loop may exit.
|
|
|
|
// MaxTripCount and ExactTripCount cannot both be non zero since we only
|
|
|
|
// compute the former when the latter is zero.
|
|
|
|
unsigned ExactTripCount = TripCount;
|
|
|
|
assert((ExactTripCount == 0 || MaxTripCount == 0) &&
|
|
|
|
"ExtractTripCound and MaxTripCount cannot both be non zero.");
|
|
|
|
unsigned FullUnrollTripCount = ExactTripCount ? ExactTripCount : MaxTripCount;
|
2016-11-10 03:56:39 +08:00
|
|
|
UP.Count = FullUnrollTripCount;
|
2016-10-13 05:29:38 +08:00
|
|
|
if (FullUnrollTripCount && FullUnrollTripCount <= UP.FullUnrollMaxCount) {
|
2016-05-28 07:15:06 +08:00
|
|
|
// When computing the unrolled size, note that BEInsns are not replicated
|
|
|
|
// like the rest of the loop body.
|
2016-12-30 08:50:28 +08:00
|
|
|
if (getUnrolledLoopSize(LoopSize, UP) < UP.Threshold) {
|
2016-10-13 05:29:38 +08:00
|
|
|
UseUpperBound = (MaxTripCount == FullUnrollTripCount);
|
|
|
|
TripCount = FullUnrollTripCount;
|
|
|
|
TripMultiple = UP.UpperBound ? 1 : TripMultiple;
|
2016-05-28 07:15:06 +08:00
|
|
|
return ExplicitUnroll;
|
2015-05-13 01:20:03 +08:00
|
|
|
} else {
|
|
|
|
// The loop isn't that small, but we still can fully unroll it if that
|
|
|
|
// helps to remove a significant number of instructions.
|
|
|
|
// To check that, run additional analysis on the loop.
|
2016-01-12 08:55:26 +08:00
|
|
|
if (Optional<EstimatedUnrollCost> Cost = analyzeLoopUnrollCost(
|
2016-10-13 05:29:38 +08:00
|
|
|
L, FullUnrollTripCount, DT, *SE, TTI,
|
2016-12-30 08:50:28 +08:00
|
|
|
UP.Threshold * UP.MaxPercentThresholdBoost / 100)) {
|
|
|
|
unsigned Boost =
|
|
|
|
getFullUnrollBoostingFactor(*Cost, UP.MaxPercentThresholdBoost);
|
|
|
|
if (Cost->UnrolledCost < UP.Threshold * Boost / 100) {
|
2016-10-13 05:29:38 +08:00
|
|
|
UseUpperBound = (MaxTripCount == FullUnrollTripCount);
|
|
|
|
TripCount = FullUnrollTripCount;
|
|
|
|
TripMultiple = UP.UpperBound ? 1 : TripMultiple;
|
2016-05-28 07:15:06 +08:00
|
|
|
return ExplicitUnroll;
|
2015-05-23 01:41:35 +08:00
|
|
|
}
|
2016-12-30 08:50:28 +08:00
|
|
|
}
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
2007-05-12 04:53:41 +08:00
|
|
|
}
|
2007-03-03 07:31:34 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// 4rd priority is partial unrolling.
|
|
|
|
// Try partial unroll only when TripCount could be staticaly calculated.
|
|
|
|
if (TripCount) {
|
|
|
|
UP.Partial |= ExplicitUnroll;
|
|
|
|
if (!UP.Partial) {
|
2014-06-17 07:53:02 +08:00
|
|
|
DEBUG(dbgs() << " will not try to unroll partially because "
|
|
|
|
<< "-unroll-allow-partial not given\n");
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.Count = 0;
|
2012-12-21 00:04:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
2016-10-28 02:40:02 +08:00
|
|
|
if (UP.Count == 0)
|
|
|
|
UP.Count = TripCount;
|
2016-05-28 07:15:06 +08:00
|
|
|
if (UP.PartialThreshold != NoThreshold) {
|
2014-06-17 07:53:02 +08:00
|
|
|
// Reduce unroll count to be modulo of TripCount for partial unrolling.
|
2016-11-10 03:56:39 +08:00
|
|
|
if (getUnrolledLoopSize(LoopSize, UP) > UP.PartialThreshold)
|
|
|
|
UP.Count =
|
|
|
|
(std::max(UP.PartialThreshold, UP.BEInsns + 1) - UP.BEInsns) /
|
|
|
|
(LoopSize - UP.BEInsns);
|
2016-05-28 07:15:06 +08:00
|
|
|
if (UP.Count > UP.MaxCount)
|
|
|
|
UP.Count = UP.MaxCount;
|
|
|
|
while (UP.Count != 0 && TripCount % UP.Count != 0)
|
|
|
|
UP.Count--;
|
|
|
|
if (UP.AllowRemainder && UP.Count <= 1) {
|
2016-04-05 03:24:46 +08:00
|
|
|
// If there is no Count that is modulo of TripCount, set Count to
|
|
|
|
// largest power-of-two factor that satisfies the threshold limit.
|
2016-04-07 00:43:45 +08:00
|
|
|
// As we'll create fixup loop, do the type of unrolling only if
|
2016-05-28 07:15:06 +08:00
|
|
|
// remainder loop is allowed.
|
2016-09-28 17:41:38 +08:00
|
|
|
UP.Count = UP.DefaultUnrollRuntimeCount;
|
2016-11-10 03:56:39 +08:00
|
|
|
while (UP.Count != 0 &&
|
|
|
|
getUnrolledLoopSize(LoopSize, UP) > UP.PartialThreshold)
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.Count >>= 1;
|
2016-04-05 03:24:46 +08:00
|
|
|
}
|
2016-05-28 07:15:06 +08:00
|
|
|
if (UP.Count < 2) {
|
|
|
|
if (PragmaEnableUnroll)
|
2016-09-30 11:44:16 +08:00
|
|
|
ORE->emit(
|
|
|
|
OptimizationRemarkMissed(DEBUG_TYPE, "UnrollAsDirectedTooLarge",
|
|
|
|
L->getStartLoc(), L->getHeader())
|
|
|
|
<< "Unable to unroll loop as directed by unroll(enable) pragma "
|
|
|
|
"because unrolled size is too large.");
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.Count = 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
UP.Count = TripCount;
|
2014-06-17 07:53:02 +08:00
|
|
|
}
|
2016-05-28 07:15:06 +08:00
|
|
|
if ((PragmaFullUnroll || PragmaEnableUnroll) && TripCount &&
|
|
|
|
UP.Count != TripCount)
|
2016-09-30 11:44:16 +08:00
|
|
|
ORE->emit(
|
|
|
|
OptimizationRemarkMissed(DEBUG_TYPE, "FullUnrollAsDirectedTooLarge",
|
|
|
|
L->getStartLoc(), L->getHeader())
|
|
|
|
<< "Unable to fully unroll loop as directed by unroll pragma because "
|
|
|
|
"unrolled size is too large.");
|
2016-05-28 07:15:06 +08:00
|
|
|
return ExplicitUnroll;
|
|
|
|
}
|
|
|
|
assert(TripCount == 0 &&
|
|
|
|
"All cases when TripCount is constant should be covered here.");
|
|
|
|
if (PragmaFullUnroll)
|
2016-09-30 11:44:16 +08:00
|
|
|
ORE->emit(
|
|
|
|
OptimizationRemarkMissed(DEBUG_TYPE,
|
|
|
|
"CantFullUnrollAsDirectedRuntimeTripCount",
|
|
|
|
L->getStartLoc(), L->getHeader())
|
|
|
|
<< "Unable to fully unroll loop as directed by unroll(full) pragma "
|
|
|
|
"because loop has a runtime trip count.");
|
2016-05-28 07:15:06 +08:00
|
|
|
|
2016-12-01 05:13:57 +08:00
|
|
|
// 5th priority is loop peeling
|
|
|
|
computePeelCount(L, LoopSize, UP);
|
|
|
|
if (UP.PeelCount) {
|
|
|
|
UP.Runtime = false;
|
|
|
|
UP.Count = 1;
|
|
|
|
return ExplicitUnroll;
|
|
|
|
}
|
|
|
|
|
|
|
|
// 6th priority is runtime unrolling.
|
2016-05-28 07:15:06 +08:00
|
|
|
// Don't unroll a runtime trip count loop when it is disabled.
|
|
|
|
if (HasRuntimeUnrollDisablePragma(L)) {
|
|
|
|
UP.Count = 0;
|
|
|
|
return false;
|
|
|
|
}
|
2016-12-01 05:13:57 +08:00
|
|
|
|
|
|
|
// Check if the runtime trip count is too small when profile is available.
|
|
|
|
if (L->getHeader()->getParent()->getEntryCount()) {
|
|
|
|
if (auto ProfileTripCount = getLoopEstimatedTripCount(L)) {
|
|
|
|
if (*ProfileTripCount < FlatLoopTripCountThreshold)
|
|
|
|
return false;
|
|
|
|
else
|
|
|
|
UP.AllowExpensiveTripCount = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// Reduce count based on the type of unrolling and the threshold values.
|
|
|
|
UP.Runtime |= PragmaEnableUnroll || PragmaCount > 0 || UserUnrollCount;
|
|
|
|
if (!UP.Runtime) {
|
|
|
|
DEBUG(dbgs() << " will not try to unroll loop with runtime trip count "
|
|
|
|
<< "-unroll-runtime not given\n");
|
|
|
|
UP.Count = 0;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
if (UP.Count == 0)
|
2016-09-28 17:41:38 +08:00
|
|
|
UP.Count = UP.DefaultUnrollRuntimeCount;
|
2016-05-28 07:15:06 +08:00
|
|
|
|
|
|
|
// Reduce unroll count to be the largest power-of-two factor of
|
|
|
|
// the original count which satisfies the threshold limit.
|
2016-11-10 03:56:39 +08:00
|
|
|
while (UP.Count != 0 &&
|
|
|
|
getUnrolledLoopSize(LoopSize, UP) > UP.PartialThreshold)
|
2016-05-28 07:15:06 +08:00
|
|
|
UP.Count >>= 1;
|
2014-06-17 07:53:02 +08:00
|
|
|
|
2016-05-28 08:14:58 +08:00
|
|
|
#ifndef NDEBUG
|
2016-05-28 07:15:06 +08:00
|
|
|
unsigned OrigCount = UP.Count;
|
2016-05-28 08:14:58 +08:00
|
|
|
#endif
|
2016-05-28 07:15:06 +08:00
|
|
|
|
|
|
|
if (!UP.AllowRemainder && UP.Count != 0 && (TripMultiple % UP.Count) != 0) {
|
|
|
|
while (UP.Count != 0 && TripMultiple % UP.Count != 0)
|
|
|
|
UP.Count >>= 1;
|
|
|
|
DEBUG(dbgs() << "Remainder loop is restricted (that could architecture "
|
|
|
|
"specific or because the loop contains a convergent "
|
|
|
|
"instruction), so unroll count must divide the trip "
|
|
|
|
"multiple, "
|
|
|
|
<< TripMultiple << ". Reducing unroll count from "
|
|
|
|
<< OrigCount << " to " << UP.Count << ".\n");
|
2016-09-30 11:44:16 +08:00
|
|
|
using namespace ore;
|
2016-05-28 07:15:06 +08:00
|
|
|
if (PragmaCount > 0 && !UP.AllowRemainder)
|
2016-09-30 11:44:16 +08:00
|
|
|
ORE->emit(
|
|
|
|
OptimizationRemarkMissed(DEBUG_TYPE,
|
|
|
|
"DifferentUnrollCountFromDirected",
|
|
|
|
L->getStartLoc(), L->getHeader())
|
|
|
|
<< "Unable to unroll loop the number of times directed by "
|
|
|
|
"unroll_count pragma because remainder loop is restricted "
|
|
|
|
"(that could architecture specific or because the loop "
|
|
|
|
"contains a convergent instruction) and so must have an unroll "
|
|
|
|
"count that divides the loop trip multiple of "
|
|
|
|
<< NV("TripMultiple", TripMultiple) << ". Unrolling instead "
|
|
|
|
<< NV("UnrollCount", UP.Count) << " time(s).");
|
2004-04-18 13:20:17 +08:00
|
|
|
}
|
2005-04-22 07:48:37 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
if (UP.Count > UP.MaxCount)
|
|
|
|
UP.Count = UP.MaxCount;
|
|
|
|
DEBUG(dbgs() << " partially unrolling with count: " << UP.Count << "\n");
|
|
|
|
if (UP.Count < 2)
|
|
|
|
UP.Count = 0;
|
|
|
|
return ExplicitUnroll;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool tryToUnrollLoop(Loop *L, DominatorTree &DT, LoopInfo *LI,
|
|
|
|
ScalarEvolution *SE, const TargetTransformInfo &TTI,
|
2016-12-19 16:22:17 +08:00
|
|
|
AssumptionCache &AC, OptimizationRemarkEmitter &ORE,
|
2016-07-30 03:29:47 +08:00
|
|
|
bool PreserveLCSSA,
|
2016-05-28 07:15:06 +08:00
|
|
|
Optional<unsigned> ProvidedCount,
|
|
|
|
Optional<unsigned> ProvidedThreshold,
|
|
|
|
Optional<bool> ProvidedAllowPartial,
|
2016-10-13 05:29:38 +08:00
|
|
|
Optional<bool> ProvidedRuntime,
|
|
|
|
Optional<bool> ProvidedUpperBound) {
|
2016-05-28 08:14:58 +08:00
|
|
|
DEBUG(dbgs() << "Loop Unroll: F[" << L->getHeader()->getParent()->getName()
|
|
|
|
<< "] Loop %" << L->getHeader()->getName() << "\n");
|
2016-11-24 03:39:26 +08:00
|
|
|
if (HasUnrollDisablePragma(L))
|
|
|
|
return false;
|
|
|
|
if (!L->isLoopSimplifyForm()) {
|
|
|
|
DEBUG(
|
|
|
|
dbgs() << " Not unrolling loop which is not in loop-simplify form.\n");
|
2014-06-17 07:53:02 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
unsigned NumInlineCandidates;
|
|
|
|
bool NotDuplicatable;
|
|
|
|
bool Convergent;
|
2016-11-10 03:56:39 +08:00
|
|
|
TargetTransformInfo::UnrollingPreferences UP = gatherUnrollingPreferences(
|
|
|
|
L, TTI, ProvidedThreshold, ProvidedCount, ProvidedAllowPartial,
|
|
|
|
ProvidedRuntime, ProvidedUpperBound);
|
2016-11-24 03:39:26 +08:00
|
|
|
// Exit early if unrolling is disabled.
|
|
|
|
if (UP.Threshold == 0 && (!UP.Partial || UP.PartialThreshold == 0))
|
|
|
|
return false;
|
2016-05-28 07:15:06 +08:00
|
|
|
unsigned LoopSize = ApproximateLoopSize(
|
2016-12-19 16:22:17 +08:00
|
|
|
L, NumInlineCandidates, NotDuplicatable, Convergent, TTI, &AC, UP.BEInsns);
|
2016-05-28 07:15:06 +08:00
|
|
|
DEBUG(dbgs() << " Loop Size = " << LoopSize << "\n");
|
|
|
|
if (NotDuplicatable) {
|
|
|
|
DEBUG(dbgs() << " Not unrolling loop which contains non-duplicatable"
|
|
|
|
<< " instructions.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
if (NumInlineCandidates != 0) {
|
|
|
|
DEBUG(dbgs() << " Not unrolling loop with inlinable calls.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Find trip count and trip multiple if count is not available
|
|
|
|
unsigned TripCount = 0;
|
2016-10-13 05:29:38 +08:00
|
|
|
unsigned MaxTripCount = 0;
|
2016-05-28 07:15:06 +08:00
|
|
|
unsigned TripMultiple = 1;
|
|
|
|
// If there are multiple exiting blocks but one of them is the latch, use the
|
|
|
|
// latch for the trip count estimation. Otherwise insist on a single exiting
|
|
|
|
// block for the trip count estimation.
|
|
|
|
BasicBlock *ExitingBlock = L->getLoopLatch();
|
|
|
|
if (!ExitingBlock || !L->isLoopExiting(ExitingBlock))
|
|
|
|
ExitingBlock = L->getExitingBlock();
|
|
|
|
if (ExitingBlock) {
|
|
|
|
TripCount = SE->getSmallConstantTripCount(L, ExitingBlock);
|
|
|
|
TripMultiple = SE->getSmallConstantTripMultiple(L, ExitingBlock);
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the loop contains a convergent operation, the prelude we'd add
|
|
|
|
// to do the first few instructions before we hit the unrolled loop
|
|
|
|
// is unsafe -- it adds a control-flow dependency to the convergent
|
|
|
|
// operation. Therefore restrict remainder loop (try unrollig without).
|
|
|
|
//
|
|
|
|
// TODO: This is quite conservative. In practice, convergent_op()
|
|
|
|
// is likely to be called unconditionally in the loop. In this
|
|
|
|
// case, the program would be ill-formed (on most architectures)
|
|
|
|
// unless n were the same on all threads in a thread group.
|
|
|
|
// Assuming n is the same on all threads, any kind of unrolling is
|
|
|
|
// safe. But currently llvm's notion of convergence isn't powerful
|
|
|
|
// enough to express this.
|
|
|
|
if (Convergent)
|
|
|
|
UP.AllowRemainder = false;
|
|
|
|
|
2016-10-21 19:08:48 +08:00
|
|
|
// Try to find the trip count upper bound if we cannot find the exact trip
|
|
|
|
// count.
|
|
|
|
bool MaxOrZero = false;
|
|
|
|
if (!TripCount) {
|
|
|
|
MaxTripCount = SE->getSmallConstantMaxTripCount(L);
|
|
|
|
MaxOrZero = SE->isBackedgeTakenCountMaxOrZero(L);
|
|
|
|
// We can unroll by the upper bound amount if it's generally allowed or if
|
|
|
|
// we know that the loop is executed either the upper bound or zero times.
|
|
|
|
// (MaxOrZero unrolling keeps only the first loop test, so the number of
|
|
|
|
// loop tests remains the same compared to the non-unrolled version, whereas
|
|
|
|
// the generic upper bound unrolling keeps all but the last loop test so the
|
|
|
|
// number of loop tests goes up which may end up being worse on targets with
|
|
|
|
// constriained branch predictor resources so is controlled by an option.)
|
|
|
|
// In addition we only unroll small upper bounds.
|
|
|
|
if (!(UP.UpperBound || MaxOrZero) || MaxTripCount > UnrollMaxUpperBound) {
|
|
|
|
MaxTripCount = 0;
|
2016-10-13 05:29:38 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// computeUnrollCount() decides whether it is beneficial to use upper bound to
|
|
|
|
// fully unroll the loop.
|
|
|
|
bool UseUpperBound = false;
|
|
|
|
bool IsCountSetExplicitly =
|
|
|
|
computeUnrollCount(L, TTI, DT, LI, SE, &ORE, TripCount, MaxTripCount,
|
|
|
|
TripMultiple, LoopSize, UP, UseUpperBound);
|
2016-05-28 07:15:06 +08:00
|
|
|
if (!UP.Count)
|
|
|
|
return false;
|
|
|
|
// Unroll factor (Count) must be less or equal to TripCount.
|
|
|
|
if (TripCount && UP.Count > TripCount)
|
|
|
|
UP.Count = TripCount;
|
|
|
|
|
2008-05-14 08:24:14 +08:00
|
|
|
// Unroll the loop.
|
2016-05-28 07:15:06 +08:00
|
|
|
if (!UnrollLoop(L, UP.Count, TripCount, UP.Force, UP.Runtime,
|
2016-10-21 19:08:48 +08:00
|
|
|
UP.AllowExpensiveTripCount, UseUpperBound, MaxOrZero,
|
2016-12-19 16:22:17 +08:00
|
|
|
TripMultiple, UP.PeelCount, LI, SE, &DT, &AC, &ORE,
|
2016-12-01 05:13:57 +08:00
|
|
|
PreserveLCSSA))
|
2008-05-14 08:24:14 +08:00
|
|
|
return false;
|
2004-04-18 13:20:17 +08:00
|
|
|
|
2016-05-28 07:15:06 +08:00
|
|
|
// If loop has an unroll count pragma or unrolled by explicitly set count
|
|
|
|
// mark loop as unrolled to prevent unrolling beyond that requested.
|
2016-12-01 05:13:57 +08:00
|
|
|
// If the loop was peeled, we already "used up" the profile information
|
|
|
|
// we had, so we don't want to unroll or peel again.
|
|
|
|
if (IsCountSetExplicitly || UP.PeelCount)
|
2016-03-25 22:24:52 +08:00
|
|
|
SetLoopAlreadyUnrolled(L);
|
2016-12-01 05:13:57 +08:00
|
|
|
|
2004-04-18 13:20:17 +08:00
|
|
|
return true;
|
|
|
|
}
|
2016-01-12 13:21:37 +08:00
|
|
|
|
|
|
|
namespace {
|
|
|
|
class LoopUnroll : public LoopPass {
|
|
|
|
public:
|
|
|
|
static char ID; // Pass ID, replacement for typeid
|
|
|
|
LoopUnroll(Optional<unsigned> Threshold = None,
|
|
|
|
Optional<unsigned> Count = None,
|
2016-10-13 05:29:38 +08:00
|
|
|
Optional<bool> AllowPartial = None, Optional<bool> Runtime = None,
|
|
|
|
Optional<bool> UpperBound = None)
|
2016-05-27 22:27:24 +08:00
|
|
|
: LoopPass(ID), ProvidedCount(std::move(Count)),
|
|
|
|
ProvidedThreshold(Threshold), ProvidedAllowPartial(AllowPartial),
|
2016-10-13 05:29:38 +08:00
|
|
|
ProvidedRuntime(Runtime), ProvidedUpperBound(UpperBound) {
|
2016-01-12 13:21:37 +08:00
|
|
|
initializeLoopUnrollPass(*PassRegistry::getPassRegistry());
|
|
|
|
}
|
|
|
|
|
|
|
|
Optional<unsigned> ProvidedCount;
|
|
|
|
Optional<unsigned> ProvidedThreshold;
|
|
|
|
Optional<bool> ProvidedAllowPartial;
|
|
|
|
Optional<bool> ProvidedRuntime;
|
2016-10-13 05:29:38 +08:00
|
|
|
Optional<bool> ProvidedUpperBound;
|
2016-01-12 13:21:37 +08:00
|
|
|
|
|
|
|
bool runOnLoop(Loop *L, LPPassManager &) override {
|
2016-04-23 06:06:11 +08:00
|
|
|
if (skipLoop(L))
|
2016-01-12 13:21:37 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
Function &F = *L->getHeader()->getParent();
|
|
|
|
|
|
|
|
auto &DT = getAnalysis<DominatorTreeWrapperPass>().getDomTree();
|
|
|
|
LoopInfo *LI = &getAnalysis<LoopInfoWrapperPass>().getLoopInfo();
|
|
|
|
ScalarEvolution *SE = &getAnalysis<ScalarEvolutionWrapperPass>().getSE();
|
|
|
|
const TargetTransformInfo &TTI =
|
|
|
|
getAnalysis<TargetTransformInfoWrapperPass>().getTTI(F);
|
2016-12-19 16:22:17 +08:00
|
|
|
auto &AC = getAnalysis<AssumptionCacheTracker>().getAssumptionCache(F);
|
2016-08-26 23:58:34 +08:00
|
|
|
// For the old PM, we can't use OptimizationRemarkEmitter as an analysis
|
|
|
|
// pass. Function analyses need to be preserved across loop transformations
|
|
|
|
// but ORE cannot be preserved (see comment before the pass definition).
|
|
|
|
OptimizationRemarkEmitter ORE(&F);
|
2016-01-12 13:21:37 +08:00
|
|
|
bool PreserveLCSSA = mustPreserveAnalysisID(LCSSAID);
|
|
|
|
|
2016-12-19 16:22:17 +08:00
|
|
|
return tryToUnrollLoop(L, DT, LI, SE, TTI, AC, ORE, PreserveLCSSA,
|
2016-07-30 03:29:47 +08:00
|
|
|
ProvidedCount, ProvidedThreshold,
|
2016-10-13 05:29:38 +08:00
|
|
|
ProvidedAllowPartial, ProvidedRuntime,
|
|
|
|
ProvidedUpperBound);
|
2016-01-12 13:21:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/// This transformation requires natural loop information & requires that
|
|
|
|
/// loop preheaders be inserted into the CFG...
|
|
|
|
///
|
|
|
|
void getAnalysisUsage(AnalysisUsage &AU) const override {
|
2016-12-19 16:22:17 +08:00
|
|
|
AU.addRequired<AssumptionCacheTracker>();
|
2016-01-12 13:21:37 +08:00
|
|
|
AU.addRequired<TargetTransformInfoWrapperPass>();
|
[LPM] Factor all of the loop analysis usage updates into a common helper
routine.
We were getting this wrong in small ways and generally being very
inconsistent about it across loop passes. Instead, let's have a common
place where we do this. One minor downside is that this will require
some analyses like SCEV in more places than they are strictly needed.
However, this seems benign as these analyses are complete no-ops, and
without this consistency we can in many cases end up with the legacy
pass manager scheduling deciding to split up a loop pass pipeline in
order to run the function analysis half-way through. It is very, very
annoying to fix these without just being very pedantic across the board.
The only loop passes I've not updated here are ones that use
AU.setPreservesAll() such as IVUsers (an analysis) and the pass printer.
They seemed less relevant.
With this patch, almost all of the problems in PR24804 around loop pass
pipelines are fixed. The one remaining issue is that we run simplify-cfg
and instcombine in the middle of the loop pass pipeline. We've recently
added some loop variants of these passes that would seem substantially
cleaner to use, but this at least gets us much closer to the previous
state. Notably, the seven loop pass managers is down to three.
I've not updated the loop passes using LoopAccessAnalysis because that
analysis hasn't been fully wired into LoopSimplify/LCSSA, and it isn't
clear that those transforms want to support those forms anyways. They
all run late anyways, so this is harmless. Similarly, LSR is left alone
because it already carefully manages its forms and doesn't need to get
fused into a single loop pass manager with a bunch of other loop passes.
LoopReroll didn't use loop simplified form previously, and I've updated
the test case to match the trivially different output.
Finally, I've also factored all the pass initialization for the passes
that use this technique as well, so that should be done regularly and
reliably.
Thanks to James for the help reviewing and thinking about this stuff,
and Ben for help thinking about it as well!
Differential Revision: http://reviews.llvm.org/D17435
llvm-svn: 261316
2016-02-19 18:45:18 +08:00
|
|
|
// FIXME: Loop passes are required to preserve domtree, and for now we just
|
|
|
|
// recreate dom info if anything gets unrolled.
|
|
|
|
getLoopAnalysisUsage(AU);
|
2016-01-12 13:21:37 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
char LoopUnroll::ID = 0;
|
|
|
|
INITIALIZE_PASS_BEGIN(LoopUnroll, "loop-unroll", "Unroll loops", false, false)
|
2016-12-19 16:22:17 +08:00
|
|
|
INITIALIZE_PASS_DEPENDENCY(AssumptionCacheTracker)
|
[LPM] Factor all of the loop analysis usage updates into a common helper
routine.
We were getting this wrong in small ways and generally being very
inconsistent about it across loop passes. Instead, let's have a common
place where we do this. One minor downside is that this will require
some analyses like SCEV in more places than they are strictly needed.
However, this seems benign as these analyses are complete no-ops, and
without this consistency we can in many cases end up with the legacy
pass manager scheduling deciding to split up a loop pass pipeline in
order to run the function analysis half-way through. It is very, very
annoying to fix these without just being very pedantic across the board.
The only loop passes I've not updated here are ones that use
AU.setPreservesAll() such as IVUsers (an analysis) and the pass printer.
They seemed less relevant.
With this patch, almost all of the problems in PR24804 around loop pass
pipelines are fixed. The one remaining issue is that we run simplify-cfg
and instcombine in the middle of the loop pass pipeline. We've recently
added some loop variants of these passes that would seem substantially
cleaner to use, but this at least gets us much closer to the previous
state. Notably, the seven loop pass managers is down to three.
I've not updated the loop passes using LoopAccessAnalysis because that
analysis hasn't been fully wired into LoopSimplify/LCSSA, and it isn't
clear that those transforms want to support those forms anyways. They
all run late anyways, so this is harmless. Similarly, LSR is left alone
because it already carefully manages its forms and doesn't need to get
fused into a single loop pass manager with a bunch of other loop passes.
LoopReroll didn't use loop simplified form previously, and I've updated
the test case to match the trivially different output.
Finally, I've also factored all the pass initialization for the passes
that use this technique as well, so that should be done regularly and
reliably.
Thanks to James for the help reviewing and thinking about this stuff,
and Ben for help thinking about it as well!
Differential Revision: http://reviews.llvm.org/D17435
llvm-svn: 261316
2016-02-19 18:45:18 +08:00
|
|
|
INITIALIZE_PASS_DEPENDENCY(LoopPass)
|
|
|
|
INITIALIZE_PASS_DEPENDENCY(TargetTransformInfoWrapperPass)
|
2016-01-12 13:21:37 +08:00
|
|
|
INITIALIZE_PASS_END(LoopUnroll, "loop-unroll", "Unroll loops", false, false)
|
|
|
|
|
|
|
|
Pass *llvm::createLoopUnrollPass(int Threshold, int Count, int AllowPartial,
|
2016-10-13 05:29:38 +08:00
|
|
|
int Runtime, int UpperBound) {
|
2016-01-12 13:21:37 +08:00
|
|
|
// TODO: It would make more sense for this function to take the optionals
|
|
|
|
// directly, but that's dangerous since it would silently break out of tree
|
|
|
|
// callers.
|
|
|
|
return new LoopUnroll(Threshold == -1 ? None : Optional<unsigned>(Threshold),
|
|
|
|
Count == -1 ? None : Optional<unsigned>(Count),
|
|
|
|
AllowPartial == -1 ? None
|
|
|
|
: Optional<bool>(AllowPartial),
|
2016-10-13 05:29:38 +08:00
|
|
|
Runtime == -1 ? None : Optional<bool>(Runtime),
|
|
|
|
UpperBound == -1 ? None : Optional<bool>(UpperBound));
|
2016-01-12 13:21:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
Pass *llvm::createSimpleLoopUnrollPass() {
|
2016-10-13 05:29:38 +08:00
|
|
|
return llvm::createLoopUnrollPass(-1, -1, 0, 0, 0);
|
2016-01-12 13:21:37 +08:00
|
|
|
}
|
2016-07-20 07:54:23 +08:00
|
|
|
|
2017-01-11 14:23:21 +08:00
|
|
|
PreservedAnalyses LoopUnrollPass::run(Loop &L, LoopAnalysisManager &AM,
|
|
|
|
LoopStandardAnalysisResults &AR,
|
|
|
|
LPMUpdater &) {
|
2016-07-20 07:54:23 +08:00
|
|
|
const auto &FAM =
|
2017-01-11 14:23:21 +08:00
|
|
|
AM.getResult<FunctionAnalysisManagerLoopProxy>(L, AR).getManager();
|
2016-07-20 07:54:23 +08:00
|
|
|
Function *F = L.getHeader()->getParent();
|
|
|
|
|
2016-07-30 03:29:47 +08:00
|
|
|
auto *ORE = FAM.getCachedResult<OptimizationRemarkEmitterAnalysis>(*F);
|
2017-01-11 14:23:21 +08:00
|
|
|
// FIXME: This should probably be optional rather than required.
|
2016-07-30 03:29:47 +08:00
|
|
|
if (!ORE)
|
|
|
|
report_fatal_error("LoopUnrollPass: OptimizationRemarkEmitterAnalysis not "
|
|
|
|
"cached at a higher level");
|
2016-07-20 07:54:23 +08:00
|
|
|
|
2017-01-11 14:23:21 +08:00
|
|
|
bool Changed = tryToUnrollLoop(&L, AR.DT, &AR.LI, &AR.SE, AR.TTI, AR.AC, *ORE,
|
|
|
|
/*PreserveLCSSA*/ true, ProvidedCount,
|
|
|
|
ProvidedThreshold, ProvidedAllowPartial,
|
|
|
|
ProvidedRuntime, ProvidedUpperBound);
|
2016-07-20 07:54:23 +08:00
|
|
|
if (!Changed)
|
|
|
|
return PreservedAnalyses::all();
|
2017-01-15 14:32:49 +08:00
|
|
|
|
2016-07-20 07:54:23 +08:00
|
|
|
return getLoopPassPreservedAnalyses();
|
|
|
|
}
|