2017-08-02 05:20:10 +08:00
|
|
|
//===- HexagonTargetTransformInfo.cpp - Hexagon specific TTI pass ---------===//
|
2015-08-06 02:35:37 +08:00
|
|
|
//
|
2019-01-19 16:50:56 +08:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
2015-08-06 02:35:37 +08:00
|
|
|
//
|
|
|
|
/// \file
|
|
|
|
/// This file implements a TargetTransformInfo analysis pass specific to the
|
|
|
|
/// Hexagon target machine. It uses the target's detailed information to provide
|
|
|
|
/// more precise answers to certain TTI queries, while letting the target
|
|
|
|
/// independent and default TTI implementations handle the rest.
|
|
|
|
///
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "HexagonTargetTransformInfo.h"
|
2017-08-02 05:20:10 +08:00
|
|
|
#include "HexagonSubtarget.h"
|
|
|
|
#include "llvm/Analysis/TargetTransformInfo.h"
|
2018-04-14 04:46:50 +08:00
|
|
|
#include "llvm/CodeGen/ValueTypes.h"
|
2017-08-02 05:20:10 +08:00
|
|
|
#include "llvm/IR/InstrTypes.h"
|
2016-08-19 22:22:07 +08:00
|
|
|
#include "llvm/IR/Instructions.h"
|
2017-08-02 05:20:10 +08:00
|
|
|
#include "llvm/IR/User.h"
|
|
|
|
#include "llvm/Support/Casting.h"
|
|
|
|
#include "llvm/Support/CommandLine.h"
|
2018-04-04 06:55:09 +08:00
|
|
|
#include "llvm/Transforms/Utils/UnrollLoop.h"
|
2015-08-06 02:35:37 +08:00
|
|
|
|
|
|
|
using namespace llvm;
|
|
|
|
|
|
|
|
#define DEBUG_TYPE "hexagontti"
|
|
|
|
|
2018-10-27 12:51:12 +08:00
|
|
|
static cl::opt<bool> HexagonAutoHVX("hexagon-autohvx", cl::init(false),
|
2018-03-28 01:07:52 +08:00
|
|
|
cl::Hidden, cl::desc("Enable loop vectorizer for HVX"));
|
|
|
|
|
2017-07-01 04:54:24 +08:00
|
|
|
static cl::opt<bool> EmitLookupTables("hexagon-emit-lookup-tables",
|
|
|
|
cl::init(true), cl::Hidden,
|
|
|
|
cl::desc("Control lookup table emission on Hexagon target"));
|
|
|
|
|
2018-06-12 23:12:50 +08:00
|
|
|
// Constant "cost factor" to make floating point operations more expensive
|
|
|
|
// in terms of vectorization cost. This isn't the best way, but it should
|
|
|
|
// do. Ultimately, the cost should use cycles.
|
|
|
|
static const unsigned FloatFactor = 4;
|
2018-04-14 04:46:50 +08:00
|
|
|
|
|
|
|
bool HexagonTTIImpl::useHVX() const {
|
|
|
|
return ST.useHVXOps() && HexagonAutoHVX;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool HexagonTTIImpl::isTypeForHVX(Type *VecTy) const {
|
|
|
|
assert(VecTy->isVectorTy());
|
2020-04-24 04:53:31 +08:00
|
|
|
if (isa<ScalableVectorType>(VecTy))
|
2019-09-17 18:19:23 +08:00
|
|
|
return false;
|
2018-04-14 04:46:50 +08:00
|
|
|
// Avoid types like <2 x i32*>.
|
|
|
|
if (!cast<VectorType>(VecTy)->getElementType()->isIntegerTy())
|
|
|
|
return false;
|
|
|
|
EVT VecVT = EVT::getEVT(VecTy);
|
|
|
|
if (!VecVT.isSimple() || VecVT.getSizeInBits() <= 64)
|
|
|
|
return false;
|
|
|
|
if (ST.isHVXVectorType(VecVT.getSimpleVT()))
|
|
|
|
return true;
|
2018-11-06 07:26:13 +08:00
|
|
|
auto Action = TLI.getPreferredVectorAction(VecVT.getSimpleVT());
|
2018-04-14 04:46:50 +08:00
|
|
|
return Action == TargetLoweringBase::TypeWidenVector;
|
|
|
|
}
|
|
|
|
|
2018-06-12 23:12:50 +08:00
|
|
|
unsigned HexagonTTIImpl::getTypeNumElements(Type *Ty) const {
|
2020-04-04 02:24:59 +08:00
|
|
|
if (auto *VTy = dyn_cast<VectorType>(Ty))
|
|
|
|
return VTy->getNumElements();
|
2018-06-12 23:12:50 +08:00
|
|
|
assert((Ty->isIntegerTy() || Ty->isFloatingPointTy()) &&
|
|
|
|
"Expecting scalar type");
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-08-06 02:35:37 +08:00
|
|
|
TargetTransformInfo::PopcntSupportKind
|
|
|
|
HexagonTTIImpl::getPopcntSupport(unsigned IntTyWidthInBit) const {
|
2018-04-14 04:46:50 +08:00
|
|
|
// Return fast hardware support as every input < 64 bits will be promoted
|
2015-08-06 02:35:37 +08:00
|
|
|
// to 64 bits.
|
|
|
|
return TargetTransformInfo::PSK_FastHardware;
|
|
|
|
}
|
|
|
|
|
|
|
|
// The Hexagon target can unroll loops with run-time trip counts.
|
[LoopUnroll] Pass SCEV to getUnrollingPreferences hook. NFCI.
Reviewers: sanjoy, anna, reames, apilipenko, igor-laevsky, mkuper
Subscribers: jholewinski, arsenm, mzolotukhin, nemanjai, nhaehnle, javed.absar, mcrosier, llvm-commits
Differential Revision: https://reviews.llvm.org/D34531
llvm-svn: 306554
2017-06-28 23:53:17 +08:00
|
|
|
void HexagonTTIImpl::getUnrollingPreferences(Loop *L, ScalarEvolution &SE,
|
2015-08-06 02:35:37 +08:00
|
|
|
TTI::UnrollingPreferences &UP) {
|
|
|
|
UP.Runtime = UP.Partial = true;
|
2018-04-03 11:39:43 +08:00
|
|
|
// Only try to peel innermost loops with small runtime trip counts.
|
2018-04-04 06:55:09 +08:00
|
|
|
if (L && L->empty() && canPeel(L) &&
|
2018-04-03 11:39:43 +08:00
|
|
|
SE.getSmallConstantTripCount(L) == 0 &&
|
|
|
|
SE.getSmallConstantMaxTripCount(L) > 0 &&
|
|
|
|
SE.getSmallConstantMaxTripCount(L) <= 5) {
|
|
|
|
UP.PeelCount = 2;
|
|
|
|
}
|
2015-08-06 02:35:37 +08:00
|
|
|
}
|
|
|
|
|
2018-03-26 23:32:03 +08:00
|
|
|
bool HexagonTTIImpl::shouldFavorPostInc() const {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-04-14 04:46:50 +08:00
|
|
|
/// --- Vector TTI begin ---
|
|
|
|
|
2018-03-28 01:07:52 +08:00
|
|
|
unsigned HexagonTTIImpl::getNumberOfRegisters(bool Vector) const {
|
|
|
|
if (Vector)
|
2018-04-14 04:46:50 +08:00
|
|
|
return useHVX() ? 32 : 0;
|
2018-03-28 01:07:52 +08:00
|
|
|
return 32;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getMaxInterleaveFactor(unsigned VF) {
|
2018-04-14 04:46:50 +08:00
|
|
|
return useHVX() ? 2 : 0;
|
2018-03-28 01:07:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getRegisterBitWidth(bool Vector) const {
|
|
|
|
return Vector ? getMinVectorRegisterBitWidth() : 32;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getMinVectorRegisterBitWidth() const {
|
2018-04-14 04:46:50 +08:00
|
|
|
return useHVX() ? ST.getVectorLength()*8 : 0;
|
2018-03-28 01:07:52 +08:00
|
|
|
}
|
|
|
|
|
2018-04-14 04:16:32 +08:00
|
|
|
unsigned HexagonTTIImpl::getMinimumVF(unsigned ElemWidth) const {
|
2018-04-14 04:46:50 +08:00
|
|
|
return (8 * ST.getVectorLength()) / ElemWidth;
|
|
|
|
}
|
|
|
|
|
2020-04-29 18:39:13 +08:00
|
|
|
unsigned HexagonTTIImpl::getScalarizationOverhead(Type *Ty,
|
|
|
|
const APInt &DemandedElts,
|
|
|
|
bool Insert, bool Extract) {
|
|
|
|
return BaseT::getScalarizationOverhead(Ty, DemandedElts, Insert, Extract);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getOperandsScalarizationOverhead(
|
|
|
|
ArrayRef<const Value*> Args, unsigned VF) {
|
|
|
|
return BaseT::getOperandsScalarizationOverhead(Args, VF);
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getCallInstrCost(Function *F, Type *RetTy,
|
|
|
|
ArrayRef<Type*> Tys) {
|
|
|
|
return BaseT::getCallInstrCost(F, RetTy, Tys);
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getIntrinsicInstrCost(Intrinsic::ID ID, Type *RetTy,
|
2020-03-11 18:13:11 +08:00
|
|
|
ArrayRef<Value *> Args,
|
|
|
|
FastMathFlags FMF, unsigned VF,
|
|
|
|
const Instruction *I) {
|
|
|
|
return BaseT::getIntrinsicInstrCost(ID, RetTy, Args, FMF, VF, I);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getIntrinsicInstrCost(Intrinsic::ID ID, Type *RetTy,
|
2020-03-11 18:13:11 +08:00
|
|
|
ArrayRef<Type *> Tys,
|
|
|
|
FastMathFlags FMF,
|
|
|
|
unsigned ScalarizationCostPassed,
|
|
|
|
const Instruction *I) {
|
2018-04-14 04:46:50 +08:00
|
|
|
if (ID == Intrinsic::bswap) {
|
|
|
|
std::pair<int, MVT> LT = TLI.getTypeLegalizationCost(DL, RetTy);
|
|
|
|
return LT.first + 2;
|
|
|
|
}
|
|
|
|
return BaseT::getIntrinsicInstrCost(ID, RetTy, Tys, FMF,
|
2020-03-11 18:13:11 +08:00
|
|
|
ScalarizationCostPassed, I);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getAddressComputationCost(Type *Tp,
|
|
|
|
ScalarEvolution *SE, const SCEV *S) {
|
|
|
|
return 0;
|
2018-04-14 04:16:32 +08:00
|
|
|
}
|
|
|
|
|
2018-03-28 01:07:52 +08:00
|
|
|
unsigned HexagonTTIImpl::getMemoryOpCost(unsigned Opcode, Type *Src,
|
2019-10-22 23:16:52 +08:00
|
|
|
MaybeAlign Alignment,
|
|
|
|
unsigned AddressSpace,
|
|
|
|
const Instruction *I) {
|
2018-04-14 04:46:50 +08:00
|
|
|
assert(Opcode == Instruction::Load || Opcode == Instruction::Store);
|
|
|
|
if (Opcode == Instruction::Store)
|
|
|
|
return BaseT::getMemoryOpCost(Opcode, Src, Alignment, AddressSpace, I);
|
|
|
|
|
|
|
|
if (Src->isVectorTy()) {
|
2018-03-28 01:07:52 +08:00
|
|
|
VectorType *VecTy = cast<VectorType>(Src);
|
2020-04-15 01:58:39 +08:00
|
|
|
unsigned VecWidth = VecTy->getPrimitiveSizeInBits().getFixedSize();
|
2018-04-14 04:46:50 +08:00
|
|
|
if (useHVX() && isTypeForHVX(VecTy)) {
|
|
|
|
unsigned RegWidth = getRegisterBitWidth(true);
|
2019-05-22 20:25:46 +08:00
|
|
|
assert(RegWidth && "Non-zero vector register width expected");
|
2018-04-14 04:46:50 +08:00
|
|
|
// Cost of HVX loads.
|
|
|
|
if (VecWidth % RegWidth == 0)
|
|
|
|
return VecWidth / RegWidth;
|
2019-10-22 23:16:52 +08:00
|
|
|
// Cost of constructing HVX vector from scalar loads
|
|
|
|
const Align RegAlign(RegWidth / 8);
|
|
|
|
if (!Alignment || *Alignment > RegAlign)
|
|
|
|
Alignment = RegAlign;
|
|
|
|
assert(Alignment);
|
|
|
|
unsigned AlignWidth = 8 * Alignment->value();
|
2018-03-28 01:07:52 +08:00
|
|
|
unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth;
|
2019-05-22 20:25:46 +08:00
|
|
|
return 3 * NumLoads;
|
2018-03-28 01:07:52 +08:00
|
|
|
}
|
2018-04-14 04:46:50 +08:00
|
|
|
|
|
|
|
// Non-HVX vectors.
|
|
|
|
// Add extra cost for floating point types.
|
2019-10-22 23:16:52 +08:00
|
|
|
unsigned Cost =
|
|
|
|
VecTy->getElementType()->isFloatingPointTy() ? FloatFactor : 1;
|
|
|
|
|
[Alignment][NFC] Deprecate Align::None()
Summary:
This is a follow up on https://reviews.llvm.org/D71473#inline-647262.
There's a caveat here that `Align(1)` relies on the compiler understanding of `Log2_64` implementation to produce good code. One could use `Align()` as a replacement but I believe it is less clear that the alignment is one in that case.
Reviewers: xbolva00, courbet, bollu
Subscribers: arsenm, dylanmckay, sdardis, nemanjai, jvesely, nhaehnle, hiraditya, kbarton, jrtc27, atanasyan, jsji, Jim, kerbowa, cfe-commits, llvm-commits
Tags: #clang, #llvm
Differential Revision: https://reviews.llvm.org/D73099
2020-01-21 22:00:04 +08:00
|
|
|
// At this point unspecified alignment is considered as Align(1).
|
2019-10-22 23:16:52 +08:00
|
|
|
const Align BoundAlignment = std::min(Alignment.valueOrOne(), Align(8));
|
|
|
|
unsigned AlignWidth = 8 * BoundAlignment.value();
|
2018-04-14 04:46:50 +08:00
|
|
|
unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth;
|
2019-10-22 23:16:52 +08:00
|
|
|
if (Alignment == Align(4) || Alignment == Align(8))
|
2018-04-14 04:46:50 +08:00
|
|
|
return Cost * NumLoads;
|
|
|
|
// Loads of less than 32 bits will need extra inserts to compose a vector.
|
2019-10-22 23:16:52 +08:00
|
|
|
assert(BoundAlignment <= Align(8));
|
|
|
|
unsigned LogA = Log2(BoundAlignment);
|
2018-04-14 04:46:50 +08:00
|
|
|
return (3 - LogA) * Cost * NumLoads;
|
2018-03-28 01:07:52 +08:00
|
|
|
}
|
2018-04-14 04:46:50 +08:00
|
|
|
|
2018-03-28 01:07:52 +08:00
|
|
|
return BaseT::getMemoryOpCost(Opcode, Src, Alignment, AddressSpace, I);
|
2015-08-06 02:35:37 +08:00
|
|
|
}
|
2016-07-22 22:22:43 +08:00
|
|
|
|
2018-04-14 04:46:50 +08:00
|
|
|
unsigned HexagonTTIImpl::getMaskedMemoryOpCost(unsigned Opcode,
|
|
|
|
Type *Src, unsigned Alignment, unsigned AddressSpace) {
|
|
|
|
return BaseT::getMaskedMemoryOpCost(Opcode, Src, Alignment, AddressSpace);
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getShuffleCost(TTI::ShuffleKind Kind, Type *Tp,
|
|
|
|
int Index, Type *SubTp) {
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getGatherScatterOpCost(unsigned Opcode, Type *DataTy,
|
2020-03-11 18:13:11 +08:00
|
|
|
Value *Ptr, bool VariableMask,
|
|
|
|
unsigned Alignment,
|
|
|
|
const Instruction *I) {
|
2018-04-14 04:46:50 +08:00
|
|
|
return BaseT::getGatherScatterOpCost(Opcode, DataTy, Ptr, VariableMask,
|
2020-03-11 18:13:11 +08:00
|
|
|
Alignment, I);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getInterleavedMemoryOpCost(unsigned Opcode,
|
|
|
|
Type *VecTy, unsigned Factor, ArrayRef<unsigned> Indices,
|
2018-10-31 17:57:56 +08:00
|
|
|
unsigned Alignment, unsigned AddressSpace, bool UseMaskForCond,
|
|
|
|
bool UseMaskForGaps) {
|
|
|
|
if (Indices.size() != Factor || UseMaskForCond || UseMaskForGaps)
|
2018-08-23 04:15:04 +08:00
|
|
|
return BaseT::getInterleavedMemoryOpCost(Opcode, VecTy, Factor, Indices,
|
2018-10-31 17:57:56 +08:00
|
|
|
Alignment, AddressSpace,
|
|
|
|
UseMaskForCond, UseMaskForGaps);
|
2019-10-22 23:16:52 +08:00
|
|
|
return getMemoryOpCost(Opcode, VecTy, MaybeAlign(Alignment), AddressSpace,
|
|
|
|
nullptr);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getCmpSelInstrCost(unsigned Opcode, Type *ValTy,
|
|
|
|
Type *CondTy, const Instruction *I) {
|
|
|
|
if (ValTy->isVectorTy()) {
|
|
|
|
std::pair<int, MVT> LT = TLI.getTypeLegalizationCost(DL, ValTy);
|
|
|
|
if (Opcode == Instruction::FCmp)
|
2018-06-12 23:12:50 +08:00
|
|
|
return LT.first + FloatFactor * getTypeNumElements(ValTy);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
return BaseT::getCmpSelInstrCost(Opcode, ValTy, CondTy, I);
|
|
|
|
}
|
|
|
|
|
[ARM] Teach the Arm cost model that a Shift can be folded into other instructions
This attempts to teach the cost model in Arm that code such as:
%s = shl i32 %a, 3
%a = and i32 %s, %b
Can under Arm or Thumb2 become:
and r0, r1, r2, lsl #3
So the cost of the shift can essentially be free. To do this without
trying to artificially adjust the cost of the "and" instruction, it
needs to get the users of the shl and check if they are a type of
instruction that the shift can be folded into. And so it needs to have
access to the actual instruction in getArithmeticInstrCost, which if
available is added as an extra parameter much like getCastInstrCost.
We otherwise limit it to shifts with a single user, which should
hopefully handle most of the cases. The list of instruction that the
shift can be folded into include ADC, ADD, AND, BIC, CMP, EOR, MVN, ORR,
ORN, RSB, SBC and SUB. This translates to Add, Sub, And, Or, Xor and
ICmp.
Differential Revision: https://reviews.llvm.org/D70966
2019-12-08 23:33:24 +08:00
|
|
|
unsigned HexagonTTIImpl::getArithmeticInstrCost(
|
|
|
|
unsigned Opcode, Type *Ty, TTI::OperandValueKind Opd1Info,
|
|
|
|
TTI::OperandValueKind Opd2Info, TTI::OperandValueProperties Opd1PropInfo,
|
|
|
|
TTI::OperandValueProperties Opd2PropInfo, ArrayRef<const Value *> Args,
|
|
|
|
const Instruction *CxtI) {
|
2018-06-12 23:12:50 +08:00
|
|
|
if (Ty->isVectorTy()) {
|
|
|
|
std::pair<int, MVT> LT = TLI.getTypeLegalizationCost(DL, Ty);
|
|
|
|
if (LT.second.isFloatingPoint())
|
|
|
|
return LT.first + FloatFactor * getTypeNumElements(Ty);
|
|
|
|
}
|
2018-04-14 04:46:50 +08:00
|
|
|
return BaseT::getArithmeticInstrCost(Opcode, Ty, Opd1Info, Opd2Info,
|
[ARM] Teach the Arm cost model that a Shift can be folded into other instructions
This attempts to teach the cost model in Arm that code such as:
%s = shl i32 %a, 3
%a = and i32 %s, %b
Can under Arm or Thumb2 become:
and r0, r1, r2, lsl #3
So the cost of the shift can essentially be free. To do this without
trying to artificially adjust the cost of the "and" instruction, it
needs to get the users of the shl and check if they are a type of
instruction that the shift can be folded into. And so it needs to have
access to the actual instruction in getArithmeticInstrCost, which if
available is added as an extra parameter much like getCastInstrCost.
We otherwise limit it to shifts with a single user, which should
hopefully handle most of the cases. The list of instruction that the
shift can be folded into include ADC, ADD, AND, BIC, CMP, EOR, MVN, ORR,
ORN, RSB, SBC and SUB. This translates to Add, Sub, And, Or, Xor and
ICmp.
Differential Revision: https://reviews.llvm.org/D70966
2019-12-08 23:33:24 +08:00
|
|
|
Opd1PropInfo, Opd2PropInfo, Args, CxtI);
|
2018-04-14 04:46:50 +08:00
|
|
|
}
|
|
|
|
|
2018-06-12 23:12:50 +08:00
|
|
|
unsigned HexagonTTIImpl::getCastInstrCost(unsigned Opcode, Type *DstTy,
|
|
|
|
Type *SrcTy, const Instruction *I) {
|
|
|
|
if (SrcTy->isFPOrFPVectorTy() || DstTy->isFPOrFPVectorTy()) {
|
|
|
|
unsigned SrcN = SrcTy->isFPOrFPVectorTy() ? getTypeNumElements(SrcTy) : 0;
|
|
|
|
unsigned DstN = DstTy->isFPOrFPVectorTy() ? getTypeNumElements(DstTy) : 0;
|
|
|
|
|
|
|
|
std::pair<int, MVT> SrcLT = TLI.getTypeLegalizationCost(DL, SrcTy);
|
|
|
|
std::pair<int, MVT> DstLT = TLI.getTypeLegalizationCost(DL, DstTy);
|
|
|
|
return std::max(SrcLT.first, DstLT.first) + FloatFactor * (SrcN + DstN);
|
|
|
|
}
|
2018-04-14 04:46:50 +08:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getVectorInstrCost(unsigned Opcode, Type *Val,
|
|
|
|
unsigned Index) {
|
|
|
|
Type *ElemTy = Val->isVectorTy() ? cast<VectorType>(Val)->getElementType()
|
|
|
|
: Val;
|
|
|
|
if (Opcode == Instruction::InsertElement) {
|
|
|
|
// Need two rotations for non-zero index.
|
|
|
|
unsigned Cost = (Index != 0) ? 2 : 0;
|
|
|
|
if (ElemTy->isIntegerTy(32))
|
|
|
|
return Cost;
|
|
|
|
// If it's not a 32-bit value, there will need to be an extract.
|
|
|
|
return Cost + getVectorInstrCost(Instruction::ExtractElement, Val, Index);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (Opcode == Instruction::ExtractElement)
|
|
|
|
return 2;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// --- Vector TTI end ---
|
|
|
|
|
2016-07-22 22:22:43 +08:00
|
|
|
unsigned HexagonTTIImpl::getPrefetchDistance() const {
|
2018-04-14 04:46:50 +08:00
|
|
|
return ST.getL1PrefetchDistance();
|
2016-07-22 22:22:43 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned HexagonTTIImpl::getCacheLineSize() const {
|
2018-04-14 04:46:50 +08:00
|
|
|
return ST.getL1CacheLineSize();
|
2016-07-22 22:22:43 +08:00
|
|
|
}
|
2016-08-19 22:22:07 +08:00
|
|
|
|
2020-04-27 16:02:14 +08:00
|
|
|
int
|
|
|
|
HexagonTTIImpl::getUserCost(const User *U,
|
|
|
|
ArrayRef<const Value *> Operands,
|
|
|
|
TTI::TargetCostKind CostKind) {
|
2018-03-28 01:07:52 +08:00
|
|
|
auto isCastFoldedIntoLoad = [this](const CastInst *CI) -> bool {
|
2016-08-19 22:22:07 +08:00
|
|
|
if (!CI->isIntegerCast())
|
|
|
|
return false;
|
2018-03-28 01:07:52 +08:00
|
|
|
// Only extensions from an integer type shorter than 32-bit to i32
|
|
|
|
// can be folded into the load.
|
|
|
|
const DataLayout &DL = getDataLayout();
|
|
|
|
unsigned SBW = DL.getTypeSizeInBits(CI->getSrcTy());
|
|
|
|
unsigned DBW = DL.getTypeSizeInBits(CI->getDestTy());
|
|
|
|
if (DBW != 32 || SBW >= DBW)
|
|
|
|
return false;
|
|
|
|
|
2016-08-19 22:22:07 +08:00
|
|
|
const LoadInst *LI = dyn_cast<const LoadInst>(CI->getOperand(0));
|
|
|
|
// Technically, this code could allow multiple uses of the load, and
|
|
|
|
// check if all the uses are the same extension operation, but this
|
|
|
|
// should be sufficient for most cases.
|
2018-03-28 01:07:52 +08:00
|
|
|
return LI && LI->hasOneUse();
|
2016-08-19 22:22:07 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
if (const CastInst *CI = dyn_cast<const CastInst>(U))
|
|
|
|
if (isCastFoldedIntoLoad(CI))
|
|
|
|
return TargetTransformInfo::TCC_Free;
|
2020-04-27 16:02:14 +08:00
|
|
|
return BaseT::getUserCost(U, Operands, CostKind);
|
2016-08-19 22:22:07 +08:00
|
|
|
}
|
2017-07-01 04:54:24 +08:00
|
|
|
|
|
|
|
bool HexagonTTIImpl::shouldBuildLookupTables() const {
|
2018-03-26 23:32:03 +08:00
|
|
|
return EmitLookupTables;
|
2017-07-01 04:54:24 +08:00
|
|
|
}
|