forked from OSchip/llvm-project
Update a few calls to getSubtarget<> to either be getSubtargetImpl
when we didn't need the cast to the base class or the cached version off of the subtarget. llvm-svn: 227176
This commit is contained in:
parent
36d9273128
commit
2c63549386
|
@ -225,7 +225,7 @@ void BasicTTI::getUnrollingPreferences(const Function *F, Loop *L,
|
||||||
// until someone finds a case where it matters in practice.
|
// until someone finds a case where it matters in practice.
|
||||||
|
|
||||||
unsigned MaxOps;
|
unsigned MaxOps;
|
||||||
const TargetSubtargetInfo *ST = &TM->getSubtarget<TargetSubtargetInfo>(F);
|
const TargetSubtargetInfo *ST = TM->getSubtargetImpl(F);
|
||||||
if (PartialUnrollingThreshold.getNumOccurrences() > 0)
|
if (PartialUnrollingThreshold.getNumOccurrences() > 0)
|
||||||
MaxOps = PartialUnrollingThreshold;
|
MaxOps = PartialUnrollingThreshold;
|
||||||
else if (ST->getSchedModel().LoopMicroOpBufferSize > 0)
|
else if (ST->getSchedModel().LoopMicroOpBufferSize > 0)
|
||||||
|
|
|
@ -3145,8 +3145,8 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||||
SunkAddr = Builder.CreateBitCast(SunkAddr, Addr->getType());
|
SunkAddr = Builder.CreateBitCast(SunkAddr, Addr->getType());
|
||||||
} else if (AddrSinkUsingGEPs ||
|
} else if (AddrSinkUsingGEPs ||
|
||||||
(!AddrSinkUsingGEPs.getNumOccurrences() && TM &&
|
(!AddrSinkUsingGEPs.getNumOccurrences() && TM &&
|
||||||
TM->getSubtarget<TargetSubtargetInfo>(
|
TM->getSubtargetImpl(*MemoryInst->getParent()->getParent())
|
||||||
MemoryInst->getParent()->getParent()).useAA())) {
|
->useAA())) {
|
||||||
// By default, we use the GEP-based method when AA is used later. This
|
// By default, we use the GEP-based method when AA is used later. This
|
||||||
// prevents new inttoptr/ptrtoint pairs from degrading AA capabilities.
|
// prevents new inttoptr/ptrtoint pairs from degrading AA capabilities.
|
||||||
DEBUG(dbgs() << "CGP: SINKING nonlocal addrmode: " << AddrMode << " for "
|
DEBUG(dbgs() << "CGP: SINKING nonlocal addrmode: " << AddrMode << " for "
|
||||||
|
|
|
@ -249,7 +249,7 @@ TargetPassConfig::TargetPassConfig(TargetMachine *tm, PassManagerBase &pm)
|
||||||
substitutePass(&PostRAMachineLICMID, &MachineLICMID);
|
substitutePass(&PostRAMachineLICMID, &MachineLICMID);
|
||||||
|
|
||||||
// Temporarily disable experimental passes.
|
// Temporarily disable experimental passes.
|
||||||
const TargetSubtargetInfo &ST = TM->getSubtarget<TargetSubtargetInfo>();
|
const TargetSubtargetInfo &ST = *TM->getSubtargetImpl();
|
||||||
if (!ST.useMachineScheduler())
|
if (!ST.useMachineScheduler())
|
||||||
disablePass(&MachineSchedulerID);
|
disablePass(&MachineSchedulerID);
|
||||||
}
|
}
|
||||||
|
|
|
@ -2716,15 +2716,14 @@ bool RegisterCoalescer::runOnMachineFunction(MachineFunction &fn) {
|
||||||
MF = &fn;
|
MF = &fn;
|
||||||
MRI = &fn.getRegInfo();
|
MRI = &fn.getRegInfo();
|
||||||
TM = &fn.getTarget();
|
TM = &fn.getTarget();
|
||||||
TRI = TM->getSubtargetImpl()->getRegisterInfo();
|
const TargetSubtargetInfo &STI = fn.getSubtarget();
|
||||||
TII = TM->getSubtargetImpl()->getInstrInfo();
|
TRI = STI.getRegisterInfo();
|
||||||
|
TII = STI.getInstrInfo();
|
||||||
LIS = &getAnalysis<LiveIntervals>();
|
LIS = &getAnalysis<LiveIntervals>();
|
||||||
AA = &getAnalysis<AliasAnalysis>();
|
AA = &getAnalysis<AliasAnalysis>();
|
||||||
Loops = &getAnalysis<MachineLoopInfo>();
|
Loops = &getAnalysis<MachineLoopInfo>();
|
||||||
|
|
||||||
const TargetSubtargetInfo &ST = TM->getSubtarget<TargetSubtargetInfo>();
|
|
||||||
if (EnableGlobalCopies == cl::BOU_UNSET)
|
if (EnableGlobalCopies == cl::BOU_UNSET)
|
||||||
JoinGlobalCopies = ST.useMachineScheduler();
|
JoinGlobalCopies = STI.useMachineScheduler();
|
||||||
else
|
else
|
||||||
JoinGlobalCopies = (EnableGlobalCopies == cl::BOU_TRUE);
|
JoinGlobalCopies = (EnableGlobalCopies == cl::BOU_TRUE);
|
||||||
|
|
||||||
|
|
|
@ -51,18 +51,17 @@ static cl::opt<bool> UseTBAA("use-tbaa-in-sched-mi", cl::Hidden,
|
||||||
|
|
||||||
ScheduleDAGInstrs::ScheduleDAGInstrs(MachineFunction &mf,
|
ScheduleDAGInstrs::ScheduleDAGInstrs(MachineFunction &mf,
|
||||||
const MachineLoopInfo *mli,
|
const MachineLoopInfo *mli,
|
||||||
bool IsPostRAFlag,
|
bool IsPostRAFlag, bool RemoveKillFlags,
|
||||||
bool RemoveKillFlags,
|
|
||||||
LiveIntervals *lis)
|
LiveIntervals *lis)
|
||||||
: ScheduleDAG(mf), MLI(mli), MFI(mf.getFrameInfo()), LIS(lis),
|
: ScheduleDAG(mf), MLI(mli), MFI(mf.getFrameInfo()), LIS(lis),
|
||||||
IsPostRA(IsPostRAFlag), RemoveKillFlags(RemoveKillFlags),
|
IsPostRA(IsPostRAFlag), RemoveKillFlags(RemoveKillFlags),
|
||||||
CanHandleTerminators(false), FirstDbgValue(nullptr) {
|
CanHandleTerminators(false), FirstDbgValue(nullptr) {
|
||||||
assert((IsPostRA || LIS) && "PreRA scheduling requires LiveIntervals");
|
assert((IsPostRA || LIS) && "PreRA scheduling requires LiveIntervals");
|
||||||
DbgValues.clear();
|
DbgValues.clear();
|
||||||
assert(!(IsPostRA && MRI.getNumVirtRegs()) &&
|
assert(!(IsPostRA && MRI.getNumVirtRegs()) &&
|
||||||
"Virtual registers must be removed prior to PostRA scheduling");
|
"Virtual registers must be removed prior to PostRA scheduling");
|
||||||
|
|
||||||
const TargetSubtargetInfo &ST = TM.getSubtarget<TargetSubtargetInfo>();
|
const TargetSubtargetInfo &ST = mf.getSubtarget();
|
||||||
SchedModel.init(ST.getSchedModel(), &ST, TII);
|
SchedModel.init(ST.getSchedModel(), &ST, TII);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -253,7 +252,7 @@ void ScheduleDAGInstrs::addPhysRegDataDeps(SUnit *SU, unsigned OperIdx) {
|
||||||
assert(MO.isDef() && "expect physreg def");
|
assert(MO.isDef() && "expect physreg def");
|
||||||
|
|
||||||
// Ask the target if address-backscheduling is desirable, and if so how much.
|
// Ask the target if address-backscheduling is desirable, and if so how much.
|
||||||
const TargetSubtargetInfo &ST = TM.getSubtarget<TargetSubtargetInfo>();
|
const TargetSubtargetInfo &ST = MF.getSubtarget();
|
||||||
|
|
||||||
for (MCRegAliasIterator Alias(MO.getReg(), TRI, true);
|
for (MCRegAliasIterator Alias(MO.getReg(), TRI, true);
|
||||||
Alias.isValid(); ++Alias) {
|
Alias.isValid(); ++Alias) {
|
||||||
|
@ -444,7 +443,7 @@ void ScheduleDAGInstrs::addVRegUseDeps(SUnit *SU, unsigned OperIdx) {
|
||||||
int DefOp = Def->findRegisterDefOperandIdx(Reg);
|
int DefOp = Def->findRegisterDefOperandIdx(Reg);
|
||||||
dep.setLatency(SchedModel.computeOperandLatency(Def, DefOp, MI, OperIdx));
|
dep.setLatency(SchedModel.computeOperandLatency(Def, DefOp, MI, OperIdx));
|
||||||
|
|
||||||
const TargetSubtargetInfo &ST = TM.getSubtarget<TargetSubtargetInfo>();
|
const TargetSubtargetInfo &ST = MF.getSubtarget();
|
||||||
ST.adjustSchedDependency(DefSU, SU, const_cast<SDep &>(dep));
|
ST.adjustSchedDependency(DefSU, SU, const_cast<SDep &>(dep));
|
||||||
SU->addPred(dep);
|
SU->addPred(dep);
|
||||||
}
|
}
|
||||||
|
@ -743,7 +742,7 @@ void ScheduleDAGInstrs::initSUnits() {
|
||||||
void ScheduleDAGInstrs::buildSchedGraph(AliasAnalysis *AA,
|
void ScheduleDAGInstrs::buildSchedGraph(AliasAnalysis *AA,
|
||||||
RegPressureTracker *RPTracker,
|
RegPressureTracker *RPTracker,
|
||||||
PressureDiffs *PDiffs) {
|
PressureDiffs *PDiffs) {
|
||||||
const TargetSubtargetInfo &ST = TM.getSubtarget<TargetSubtargetInfo>();
|
const TargetSubtargetInfo &ST = MF.getSubtarget();
|
||||||
bool UseAA = EnableAASchedMI.getNumOccurrences() > 0 ? EnableAASchedMI
|
bool UseAA = EnableAASchedMI.getNumOccurrences() > 0 ? EnableAASchedMI
|
||||||
: ST.useAA();
|
: ST.useAA();
|
||||||
AliasAnalysis *AAForDep = UseAA ? AA : nullptr;
|
AliasAnalysis *AAForDep = UseAA ? AA : nullptr;
|
||||||
|
|
Loading…
Reference in New Issue