2004-07-24 01:56:30 +08:00
|
|
|
//===-- LiveIntervalAnalysis.cpp - Live Interval Analysis -----------------===//
|
2003-11-20 11:32:25 +08:00
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
2007-12-30 04:36:04 +08:00
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
2003-11-20 11:32:25 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file implements the LiveInterval analysis pass which is used
|
|
|
|
// by the Linear Scan Register allocator. This pass linearizes the
|
|
|
|
// basic blocks of the function in DFS order and uses the
|
|
|
|
// LiveVariables pass to conservatively compute live intervals for
|
|
|
|
// each virtual and physical register.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#define DEBUG_TYPE "liveintervals"
|
2005-09-21 12:19:09 +08:00
|
|
|
#include "llvm/CodeGen/LiveIntervalAnalysis.h"
|
2004-09-04 02:25:53 +08:00
|
|
|
#include "VirtRegMap.h"
|
2004-05-02 05:24:39 +08:00
|
|
|
#include "llvm/Value.h"
|
2003-11-20 11:32:25 +08:00
|
|
|
#include "llvm/CodeGen/LiveVariables.h"
|
|
|
|
#include "llvm/CodeGen/MachineFrameInfo.h"
|
|
|
|
#include "llvm/CodeGen/MachineInstr.h"
|
2007-12-11 10:09:15 +08:00
|
|
|
#include "llvm/CodeGen/MachineLoopInfo.h"
|
2007-12-31 12:13:23 +08:00
|
|
|
#include "llvm/CodeGen/MachineRegisterInfo.h"
|
2003-11-20 11:32:25 +08:00
|
|
|
#include "llvm/CodeGen/Passes.h"
|
2008-02-11 02:45:23 +08:00
|
|
|
#include "llvm/Target/TargetRegisterInfo.h"
|
2003-11-20 11:32:25 +08:00
|
|
|
#include "llvm/Target/TargetInstrInfo.h"
|
|
|
|
#include "llvm/Target/TargetMachine.h"
|
2004-09-02 06:55:40 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
|
|
|
#include "llvm/Support/Debug.h"
|
|
|
|
#include "llvm/ADT/Statistic.h"
|
|
|
|
#include "llvm/ADT/STLExtras.h"
|
2004-09-04 02:19:51 +08:00
|
|
|
#include <algorithm>
|
2006-12-02 10:22:01 +08:00
|
|
|
#include <cmath>
|
2003-11-20 11:32:25 +08:00
|
|
|
using namespace llvm;
|
|
|
|
|
2007-08-16 15:24:22 +08:00
|
|
|
namespace {
|
|
|
|
// Hidden options for help debugging.
|
|
|
|
cl::opt<bool> DisableReMat("disable-rematerialization",
|
|
|
|
cl::init(false), cl::Hidden);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
|
|
|
|
cl::opt<bool> SplitAtBB("split-intervals-at-bb",
|
2007-12-06 16:54:31 +08:00
|
|
|
cl::init(true), cl::Hidden);
|
2007-11-29 09:06:25 +08:00
|
|
|
cl::opt<int> SplitLimit("split-limit",
|
|
|
|
cl::init(-1), cl::Hidden);
|
2007-08-16 15:24:22 +08:00
|
|
|
}
|
|
|
|
|
2006-12-20 06:41:21 +08:00
|
|
|
STATISTIC(numIntervals, "Number of original intervals");
|
|
|
|
STATISTIC(numIntervalsAfter, "Number of intervals after coalescing");
|
2007-11-29 09:06:25 +08:00
|
|
|
STATISTIC(numFolds , "Number of loads/stores folded into instructions");
|
|
|
|
STATISTIC(numSplits , "Number of intervals split");
|
2006-12-20 06:41:21 +08:00
|
|
|
|
2007-05-03 09:11:54 +08:00
|
|
|
char LiveIntervals::ID = 0;
|
2003-11-20 11:32:25 +08:00
|
|
|
namespace {
|
2006-08-28 06:30:17 +08:00
|
|
|
RegisterPass<LiveIntervals> X("liveintervals", "Live Interval Analysis");
|
2006-05-25 01:04:05 +08:00
|
|
|
}
|
2003-11-20 11:32:25 +08:00
|
|
|
|
2006-08-25 06:43:55 +08:00
|
|
|
void LiveIntervals::getAnalysisUsage(AnalysisUsage &AU) const {
|
2007-06-09 01:18:56 +08:00
|
|
|
AU.addPreserved<LiveVariables>();
|
2004-08-04 17:46:26 +08:00
|
|
|
AU.addRequired<LiveVariables>();
|
2008-01-05 04:54:55 +08:00
|
|
|
AU.addPreservedID(MachineLoopInfoID);
|
|
|
|
AU.addPreservedID(MachineDominatorsID);
|
2004-08-04 17:46:26 +08:00
|
|
|
AU.addPreservedID(PHIEliminationID);
|
|
|
|
AU.addRequiredID(PHIEliminationID);
|
|
|
|
AU.addRequiredID(TwoAddressInstructionPassID);
|
|
|
|
MachineFunctionPass::getAnalysisUsage(AU);
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
|
|
|
|
2006-08-25 06:43:55 +08:00
|
|
|
void LiveIntervals::releaseMemory() {
|
2007-10-17 10:10:22 +08:00
|
|
|
Idx2MBBMap.clear();
|
2004-08-04 17:46:26 +08:00
|
|
|
mi2iMap_.clear();
|
|
|
|
i2miMap_.clear();
|
|
|
|
r2iMap_.clear();
|
2007-09-06 09:07:24 +08:00
|
|
|
// Release VNInfo memroy regions after all VNInfo objects are dtor'd.
|
|
|
|
VNInfoAllocator.Reset();
|
2007-08-14 07:45:17 +08:00
|
|
|
for (unsigned i = 0, e = ClonedMIs.size(); i != e; ++i)
|
|
|
|
delete ClonedMIs[i];
|
2006-05-11 15:29:24 +08:00
|
|
|
}
|
|
|
|
|
2003-11-20 11:32:25 +08:00
|
|
|
/// runOnMachineFunction - Register allocate the whole function
|
|
|
|
///
|
|
|
|
bool LiveIntervals::runOnMachineFunction(MachineFunction &fn) {
|
2004-08-04 17:46:26 +08:00
|
|
|
mf_ = &fn;
|
|
|
|
tm_ = &fn.getTarget();
|
2008-02-11 02:45:23 +08:00
|
|
|
tri_ = tm_->getRegisterInfo();
|
Allow the live interval analysis pass to be a bit more aggressive about
numbering values in live ranges for physical registers.
The alpha backend currently generates code that looks like this:
vreg = preg
...
preg = vreg
use preg
...
preg = vreg
use preg
etc. Because vreg contains the value of preg coming in, each of the
copies back into preg contain that initial value as well.
In the case of the Alpha, this allows this testcase:
void "foo"(int %blah) {
store int 5, int *%MyVar
store int 12, int* %MyVar2
ret void
}
to compile to:
foo:
ldgp $29, 0($27)
ldiq $0,5
stl $0,MyVar
ldiq $0,12
stl $0,MyVar2
ret $31,($26),1
instead of:
foo:
ldgp $29, 0($27)
bis $29,$29,$0
ldiq $1,5
bis $0,$0,$29
stl $1,MyVar
ldiq $1,12
bis $0,$0,$29
stl $1,MyVar2
ret $31,($26),1
This does not seem to have any noticable effect on X86 code.
This fixes PR535.
llvm-svn: 20536
2005-03-10 07:05:19 +08:00
|
|
|
tii_ = tm_->getInstrInfo();
|
2004-08-04 17:46:26 +08:00
|
|
|
lv_ = &getAnalysis<LiveVariables>();
|
2008-02-11 02:45:23 +08:00
|
|
|
allocatableRegs_ = tri_->getAllocatableSet(fn);
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2006-09-15 11:57:23 +08:00
|
|
|
// Number MachineInstrs and MachineBasicBlocks.
|
|
|
|
// Initialize MBB indexes to a sentinal.
|
2007-08-14 07:45:17 +08:00
|
|
|
MBB2IdxMap.resize(mf_->getNumBlockIDs(), std::make_pair(~0U,~0U));
|
2006-09-15 11:57:23 +08:00
|
|
|
|
|
|
|
unsigned MIIndex = 0;
|
|
|
|
for (MachineFunction::iterator MBB = mf_->begin(), E = mf_->end();
|
|
|
|
MBB != E; ++MBB) {
|
2007-08-14 07:45:17 +08:00
|
|
|
unsigned StartIdx = MIIndex;
|
2007-02-13 09:30:55 +08:00
|
|
|
|
2006-09-15 11:57:23 +08:00
|
|
|
for (MachineBasicBlock::iterator I = MBB->begin(), E = MBB->end();
|
|
|
|
I != E; ++I) {
|
|
|
|
bool inserted = mi2iMap_.insert(std::make_pair(I, MIIndex)).second;
|
2004-08-04 17:46:26 +08:00
|
|
|
assert(inserted && "multiple MachineInstr -> index mappings");
|
2006-09-15 11:57:23 +08:00
|
|
|
i2miMap_.push_back(I);
|
|
|
|
MIIndex += InstrSlots::NUM;
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2007-08-14 07:45:17 +08:00
|
|
|
|
|
|
|
// Set the MBB2IdxMap entry for this MBB.
|
|
|
|
MBB2IdxMap[MBB->getNumber()] = std::make_pair(StartIdx, MIIndex - 1);
|
2007-10-17 10:10:22 +08:00
|
|
|
Idx2MBBMap.push_back(std::make_pair(StartIdx, MBB));
|
2006-09-15 11:57:23 +08:00
|
|
|
}
|
2007-10-17 10:10:22 +08:00
|
|
|
std::sort(Idx2MBBMap.begin(), Idx2MBBMap.end(), Idx2MBBCompare());
|
2003-11-20 11:32:25 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
computeIntervals();
|
2003-11-20 11:32:25 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
numIntervals += getNumIntervals();
|
2004-02-15 18:24:21 +08:00
|
|
|
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "********** INTERVALS **********\n";
|
|
|
|
for (iterator I = begin(), E = end(); I != E; ++I) {
|
2008-02-11 02:45:23 +08:00
|
|
|
I->second.print(DOUT, tri_);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "\n";
|
|
|
|
}
|
2004-07-24 10:59:07 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
numIntervalsAfter += getNumIntervals();
|
2004-09-30 23:59:17 +08:00
|
|
|
DEBUG(dump());
|
2004-08-04 17:46:26 +08:00
|
|
|
return true;
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
|
|
|
|
2004-09-30 23:59:17 +08:00
|
|
|
/// print - Implement the dump method.
|
2004-12-07 12:03:45 +08:00
|
|
|
void LiveIntervals::print(std::ostream &O, const Module* ) const {
|
2004-09-30 23:59:17 +08:00
|
|
|
O << "********** INTERVALS **********\n";
|
2005-07-28 07:03:38 +08:00
|
|
|
for (const_iterator I = begin(), E = end(); I != E; ++I) {
|
2008-02-11 02:45:23 +08:00
|
|
|
I->second.print(DOUT, tri_);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "\n";
|
2005-07-28 07:03:38 +08:00
|
|
|
}
|
2004-09-30 23:59:17 +08:00
|
|
|
|
|
|
|
O << "********** MACHINEINSTRS **********\n";
|
|
|
|
for (MachineFunction::iterator mbbi = mf_->begin(), mbbe = mf_->end();
|
|
|
|
mbbi != mbbe; ++mbbi) {
|
|
|
|
O << ((Value*)mbbi->getBasicBlock())->getName() << ":\n";
|
|
|
|
for (MachineBasicBlock::iterator mii = mbbi->begin(),
|
|
|
|
mie = mbbi->end(); mii != mie; ++mii) {
|
2004-10-01 00:10:45 +08:00
|
|
|
O << getInstructionIndex(mii) << '\t' << *mii;
|
2004-09-30 23:59:17 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-11-03 15:20:12 +08:00
|
|
|
/// conflictsWithPhysRegDef - Returns true if the specified register
|
|
|
|
/// is defined during the duration of the specified interval.
|
|
|
|
bool LiveIntervals::conflictsWithPhysRegDef(const LiveInterval &li,
|
|
|
|
VirtRegMap &vrm, unsigned reg) {
|
|
|
|
for (LiveInterval::Ranges::const_iterator
|
|
|
|
I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) {
|
|
|
|
for (unsigned index = getBaseIndex(I->start),
|
|
|
|
end = getBaseIndex(I->end-1) + InstrSlots::NUM; index != end;
|
|
|
|
index += InstrSlots::NUM) {
|
|
|
|
// skip deleted instructions
|
|
|
|
while (index != end && !getInstructionFromIndex(index))
|
|
|
|
index += InstrSlots::NUM;
|
|
|
|
if (index == end) break;
|
|
|
|
|
|
|
|
MachineInstr *MI = getInstructionFromIndex(index);
|
2007-11-15 16:13:29 +08:00
|
|
|
unsigned SrcReg, DstReg;
|
|
|
|
if (tii_->isMoveInstr(*MI, SrcReg, DstReg))
|
|
|
|
if (SrcReg == li.reg || DstReg == li.reg)
|
|
|
|
continue;
|
2007-11-03 15:20:12 +08:00
|
|
|
for (unsigned i = 0; i != MI->getNumOperands(); ++i) {
|
|
|
|
MachineOperand& mop = MI->getOperand(i);
|
2007-11-15 16:13:29 +08:00
|
|
|
if (!mop.isRegister())
|
2007-11-03 15:20:12 +08:00
|
|
|
continue;
|
|
|
|
unsigned PhysReg = mop.getReg();
|
2007-11-15 16:13:29 +08:00
|
|
|
if (PhysReg == 0 || PhysReg == li.reg)
|
2007-11-03 15:20:12 +08:00
|
|
|
continue;
|
2008-02-11 02:45:23 +08:00
|
|
|
if (TargetRegisterInfo::isVirtualRegister(PhysReg)) {
|
2007-11-15 16:13:29 +08:00
|
|
|
if (!vrm.hasPhys(PhysReg))
|
|
|
|
continue;
|
2007-11-03 15:20:12 +08:00
|
|
|
PhysReg = vrm.getPhys(PhysReg);
|
2007-11-15 16:13:29 +08:00
|
|
|
}
|
2008-02-11 02:45:23 +08:00
|
|
|
if (PhysReg && tri_->regsOverlap(PhysReg, reg))
|
2007-11-03 15:20:12 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2006-08-23 02:19:46 +08:00
|
|
|
void LiveIntervals::printRegName(unsigned reg) const {
|
2008-02-11 02:45:23 +08:00
|
|
|
if (TargetRegisterInfo::isPhysicalRegister(reg))
|
|
|
|
cerr << tri_->getName(reg);
|
2004-08-04 17:46:26 +08:00
|
|
|
else
|
2006-12-07 09:30:32 +08:00
|
|
|
cerr << "%reg" << reg;
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
|
|
|
|
2006-08-23 02:19:46 +08:00
|
|
|
void LiveIntervals::handleVirtualRegisterDef(MachineBasicBlock *mbb,
|
2003-11-20 11:32:25 +08:00
|
|
|
MachineBasicBlock::iterator mi,
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned MIIdx,
|
2006-08-23 02:19:46 +08:00
|
|
|
LiveInterval &interval) {
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "\t\tregister: "; DEBUG(printRegName(interval.reg));
|
2004-08-04 17:46:26 +08:00
|
|
|
LiveVariables::VarInfo& vi = lv_->getVarInfo(interval.reg);
|
|
|
|
|
2004-08-04 17:46:56 +08:00
|
|
|
// Virtual registers may be defined multiple times (due to phi
|
|
|
|
// elimination and 2-addr elimination). Much of what we do only has to be
|
|
|
|
// done once for the vreg. We use an empty interval to detect the first
|
2004-08-04 17:46:26 +08:00
|
|
|
// time we see a vreg.
|
|
|
|
if (interval.empty()) {
|
|
|
|
// Get the Idx of the defining instructions.
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned defIndex = getDefIndex(MIIdx);
|
2007-08-30 04:45:00 +08:00
|
|
|
VNInfo *ValNo;
|
2008-02-16 02:24:29 +08:00
|
|
|
MachineInstr *CopyMI = NULL;
|
2006-08-31 13:54:43 +08:00
|
|
|
unsigned SrcReg, DstReg;
|
2008-02-16 02:24:29 +08:00
|
|
|
if (mi->getOpcode() == TargetInstrInfo::EXTRACT_SUBREG ||
|
|
|
|
tii_->isMoveInstr(*mi, SrcReg, DstReg))
|
|
|
|
CopyMI = mi;
|
|
|
|
ValNo = interval.getNextValue(defIndex, CopyMI, VNInfoAllocator);
|
2007-08-30 04:45:00 +08:00
|
|
|
|
|
|
|
assert(ValNo->id == 0 && "First value in interval is not 0?");
|
2004-08-04 17:46:26 +08:00
|
|
|
|
|
|
|
// Loop over all of the blocks that the vreg is defined in. There are
|
|
|
|
// two cases we have to handle here. The most common case is a vreg
|
|
|
|
// whose lifetime is contained within a basic block. In this case there
|
|
|
|
// will be a single kill, in MBB, which comes after the definition.
|
|
|
|
if (vi.Kills.size() == 1 && vi.Kills[0]->getParent() == mbb) {
|
|
|
|
// FIXME: what about dead vars?
|
|
|
|
unsigned killIdx;
|
|
|
|
if (vi.Kills[0] != mi)
|
|
|
|
killIdx = getUseIndex(getInstructionIndex(vi.Kills[0]))+1;
|
|
|
|
else
|
|
|
|
killIdx = defIndex+1;
|
|
|
|
|
|
|
|
// If the kill happens after the definition, we have an intra-block
|
|
|
|
// live range.
|
|
|
|
if (killIdx > defIndex) {
|
2007-02-15 13:59:24 +08:00
|
|
|
assert(vi.AliveBlocks.none() &&
|
2004-08-04 17:46:26 +08:00
|
|
|
"Shouldn't be alive across any blocks!");
|
2007-08-30 04:45:00 +08:00
|
|
|
LiveRange LR(defIndex, killIdx, ValNo);
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << LR << "\n";
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(ValNo, killIdx);
|
2004-08-04 17:46:26 +08:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
2004-07-19 10:15:56 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
// The other case we handle is when a virtual register lives to the end
|
|
|
|
// of the defining block, potentially live across some blocks, then is
|
|
|
|
// live into some number of blocks, but gets killed. Start by adding a
|
|
|
|
// range that goes from this definition to the end of the defining block.
|
2004-09-01 01:39:15 +08:00
|
|
|
LiveRange NewLR(defIndex,
|
|
|
|
getInstructionIndex(&mbb->back()) + InstrSlots::NUM,
|
2007-08-30 04:45:00 +08:00
|
|
|
ValNo);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << NewLR;
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(NewLR);
|
|
|
|
|
|
|
|
// Iterate over all of the blocks that the variable is completely
|
|
|
|
// live in, adding [insrtIndex(begin), instrIndex(end)+4) to the
|
|
|
|
// live interval.
|
|
|
|
for (unsigned i = 0, e = vi.AliveBlocks.size(); i != e; ++i) {
|
|
|
|
if (vi.AliveBlocks[i]) {
|
2006-09-15 11:57:23 +08:00
|
|
|
MachineBasicBlock *MBB = mf_->getBlockNumbered(i);
|
|
|
|
if (!MBB->empty()) {
|
|
|
|
LiveRange LR(getMBBStartIdx(i),
|
|
|
|
getInstructionIndex(&MBB->back()) + InstrSlots::NUM,
|
2007-08-30 04:45:00 +08:00
|
|
|
ValNo);
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << LR;
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, this virtual register is live from the start of any killing
|
|
|
|
// block to the 'use' slot of the killing instruction.
|
|
|
|
for (unsigned i = 0, e = vi.Kills.size(); i != e; ++i) {
|
|
|
|
MachineInstr *Kill = vi.Kills[i];
|
2007-08-08 11:00:28 +08:00
|
|
|
unsigned killIdx = getUseIndex(getInstructionIndex(Kill))+1;
|
2006-09-15 11:57:23 +08:00
|
|
|
LiveRange LR(getMBBStartIdx(Kill->getParent()),
|
2007-08-30 04:45:00 +08:00
|
|
|
killIdx, ValNo);
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(ValNo, killIdx);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << LR;
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
} else {
|
|
|
|
// If this is the second time we see a virtual register definition, it
|
|
|
|
// must be due to phi elimination or two addr elimination. If this is
|
2006-11-03 11:04:46 +08:00
|
|
|
// the result of two address elimination, then the vreg is one of the
|
|
|
|
// def-and-use register operand.
|
2007-10-12 16:50:34 +08:00
|
|
|
if (mi->isRegReDefinedByTwoAddr(interval.reg)) {
|
2004-08-04 17:46:26 +08:00
|
|
|
// If this is a two-address definition, then we have already processed
|
|
|
|
// the live range. The only problem is that we didn't realize there
|
|
|
|
// are actually two values in the live interval. Because of this we
|
|
|
|
// need to take the LiveRegion that defines this register and split it
|
|
|
|
// into two values.
|
2008-01-10 16:22:10 +08:00
|
|
|
assert(interval.containsOneValue());
|
|
|
|
unsigned DefIndex = getDefIndex(interval.getValNumInfo(0)->def);
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned RedefIndex = getDefIndex(MIIdx);
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2007-08-11 08:59:19 +08:00
|
|
|
const LiveRange *OldLR = interval.getLiveRangeContaining(RedefIndex-1);
|
2007-08-30 04:45:00 +08:00
|
|
|
VNInfo *OldValNo = OldLR->valno;
|
2007-08-11 08:59:19 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
// Delete the initial value, which should be short and continuous,
|
2006-08-23 02:19:46 +08:00
|
|
|
// because the 2-addr copy must be in the same MBB as the redef.
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.removeRange(DefIndex, RedefIndex);
|
2004-08-04 17:46:56 +08:00
|
|
|
|
2006-08-23 02:19:46 +08:00
|
|
|
// Two-address vregs should always only be redefined once. This means
|
|
|
|
// that at this point, there should be exactly one value number in it.
|
|
|
|
assert(interval.containsOneValue() && "Unexpected 2-addr liveint!");
|
|
|
|
|
2006-08-31 13:54:43 +08:00
|
|
|
// The new value number (#1) is defined by the instruction we claimed
|
|
|
|
// defined value #0.
|
2008-02-16 02:24:29 +08:00
|
|
|
VNInfo *ValNo = interval.getNextValue(OldValNo->def, OldValNo->copy,
|
|
|
|
VNInfoAllocator);
|
2006-08-23 02:19:46 +08:00
|
|
|
|
2006-08-31 13:54:43 +08:00
|
|
|
// Value#0 is now defined by the 2-addr instruction.
|
2008-02-16 02:24:29 +08:00
|
|
|
OldValNo->def = RedefIndex;
|
|
|
|
OldValNo->copy = 0;
|
2006-08-23 02:19:46 +08:00
|
|
|
|
|
|
|
// Add the new live interval which replaces the range for the input copy.
|
|
|
|
LiveRange LR(DefIndex, RedefIndex, ValNo);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " replace range with " << LR;
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(ValNo, RedefIndex);
|
2004-08-04 17:46:26 +08:00
|
|
|
|
|
|
|
// If this redefinition is dead, we need to add a dummy unit live
|
|
|
|
// range covering the def slot.
|
2005-08-24 06:51:41 +08:00
|
|
|
if (lv_->RegisterDefIsDead(mi, interval.reg))
|
2007-08-30 04:45:00 +08:00
|
|
|
interval.addRange(LiveRange(RedefIndex, RedefIndex+1, OldValNo));
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2007-03-16 05:19:28 +08:00
|
|
|
DOUT << " RESULT: ";
|
2008-02-11 02:45:23 +08:00
|
|
|
interval.print(DOUT, tri_);
|
2004-08-04 17:46:26 +08:00
|
|
|
|
|
|
|
} else {
|
|
|
|
// Otherwise, this must be because of phi elimination. If this is the
|
|
|
|
// first redefinition of the vreg that we have seen, go back and change
|
|
|
|
// the live range in the PHI block to be a different value number.
|
|
|
|
if (interval.containsOneValue()) {
|
|
|
|
assert(vi.Kills.size() == 1 &&
|
|
|
|
"PHI elimination vreg should have one kill, the PHI itself!");
|
|
|
|
|
|
|
|
// Remove the old range that we now know has an incorrect number.
|
2007-09-06 05:46:51 +08:00
|
|
|
VNInfo *VNI = interval.getValNumInfo(0);
|
2004-08-04 17:46:26 +08:00
|
|
|
MachineInstr *Killer = vi.Kills[0];
|
2006-09-15 11:57:23 +08:00
|
|
|
unsigned Start = getMBBStartIdx(Killer->getParent());
|
2004-08-04 17:46:26 +08:00
|
|
|
unsigned End = getUseIndex(getInstructionIndex(Killer))+1;
|
2007-03-16 05:19:28 +08:00
|
|
|
DOUT << " Removing [" << Start << "," << End << "] from: ";
|
2008-02-11 02:45:23 +08:00
|
|
|
interval.print(DOUT, tri_); DOUT << "\n";
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.removeRange(Start, End);
|
2007-11-29 17:49:23 +08:00
|
|
|
VNI->hasPHIKill = true;
|
2008-02-11 02:45:23 +08:00
|
|
|
DOUT << " RESULT: "; interval.print(DOUT, tri_);
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2006-08-23 02:19:46 +08:00
|
|
|
// Replace the interval with one of a NEW value number. Note that this
|
|
|
|
// value number isn't actually defined by an instruction, weird huh? :)
|
2007-09-06 05:46:51 +08:00
|
|
|
LiveRange LR(Start, End, interval.getNextValue(~0, 0, VNInfoAllocator));
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " replace range with " << LR;
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(LR.valno, End);
|
2008-02-11 02:45:23 +08:00
|
|
|
DOUT << " RESULT: "; interval.print(DOUT, tri_);
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// In the case of PHI elimination, each variable definition is only
|
|
|
|
// live until the end of the block. We've already taken care of the
|
|
|
|
// rest of the live range.
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned defIndex = getDefIndex(MIIdx);
|
2006-08-31 13:54:43 +08:00
|
|
|
|
2007-08-30 04:45:00 +08:00
|
|
|
VNInfo *ValNo;
|
2008-02-16 02:24:29 +08:00
|
|
|
MachineInstr *CopyMI = NULL;
|
2006-08-31 13:54:43 +08:00
|
|
|
unsigned SrcReg, DstReg;
|
2008-02-16 02:24:29 +08:00
|
|
|
if (mi->getOpcode() == TargetInstrInfo::EXTRACT_SUBREG ||
|
|
|
|
tii_->isMoveInstr(*mi, SrcReg, DstReg))
|
|
|
|
CopyMI = mi;
|
|
|
|
ValNo = interval.getNextValue(defIndex, CopyMI, VNInfoAllocator);
|
2006-08-31 13:54:43 +08:00
|
|
|
|
2007-08-08 15:03:29 +08:00
|
|
|
unsigned killIndex = getInstructionIndex(&mbb->back()) + InstrSlots::NUM;
|
2007-08-30 04:45:00 +08:00
|
|
|
LiveRange LR(defIndex, killIndex, ValNo);
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2007-11-29 17:49:23 +08:00
|
|
|
interval.addKill(ValNo, killIndex);
|
|
|
|
ValNo->hasPHIKill = true;
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << LR;
|
2003-12-18 16:48:48 +08:00
|
|
|
}
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2003-11-20 11:32:25 +08:00
|
|
|
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << '\n';
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
|
|
|
|
2004-07-24 05:24:19 +08:00
|
|
|
void LiveIntervals::handlePhysicalRegisterDef(MachineBasicBlock *MBB,
|
2003-11-20 11:32:25 +08:00
|
|
|
MachineBasicBlock::iterator mi,
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned MIIdx,
|
2006-08-31 13:54:43 +08:00
|
|
|
LiveInterval &interval,
|
2008-02-16 02:24:29 +08:00
|
|
|
MachineInstr *CopyMI) {
|
2004-08-04 17:46:26 +08:00
|
|
|
// A physical register cannot be live across basic block, so its
|
|
|
|
// lifetime must end somewhere in its defining basic block.
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "\t\tregister: "; DEBUG(printRegName(interval.reg));
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned baseIndex = MIIdx;
|
2004-08-04 17:46:26 +08:00
|
|
|
unsigned start = getDefIndex(baseIndex);
|
|
|
|
unsigned end = start;
|
|
|
|
|
|
|
|
// If it is not used after definition, it is considered dead at
|
|
|
|
// the instruction defining it. Hence its interval is:
|
|
|
|
// [defSlot(def), defSlot(def)+1)
|
2005-08-24 06:51:41 +08:00
|
|
|
if (lv_->RegisterDefIsDead(mi, interval.reg)) {
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " dead";
|
2005-08-24 06:51:41 +08:00
|
|
|
end = getDefIndex(start) + 1;
|
|
|
|
goto exit;
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2003-11-20 11:32:25 +08:00
|
|
|
|
2004-08-04 17:46:26 +08:00
|
|
|
// If it is not dead on definition, it must be killed by a
|
|
|
|
// subsequent instruction. Hence its interval is:
|
|
|
|
// [defSlot(def), useSlot(kill)+1)
|
2005-09-02 08:20:32 +08:00
|
|
|
while (++mi != MBB->end()) {
|
2004-08-04 17:46:26 +08:00
|
|
|
baseIndex += InstrSlots::NUM;
|
2005-08-24 06:51:41 +08:00
|
|
|
if (lv_->KillsRegister(mi, interval.reg)) {
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " killed";
|
2005-08-24 06:51:41 +08:00
|
|
|
end = getUseIndex(baseIndex) + 1;
|
|
|
|
goto exit;
|
2006-11-16 04:54:11 +08:00
|
|
|
} else if (lv_->ModifiesRegister(mi, interval.reg)) {
|
|
|
|
// Another instruction redefines the register before it is ever read.
|
|
|
|
// Then the register is essentially dead at the instruction that defines
|
|
|
|
// it. Hence its interval is:
|
|
|
|
// [defSlot(def), defSlot(def)+1)
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " dead";
|
2006-11-16 04:54:11 +08:00
|
|
|
end = getDefIndex(start) + 1;
|
|
|
|
goto exit;
|
2004-07-24 05:24:19 +08:00
|
|
|
}
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2005-09-02 08:20:32 +08:00
|
|
|
|
|
|
|
// The only case we should have a dead physreg here without a killing or
|
|
|
|
// instruction where we know it's dead is if it is live-in to the function
|
|
|
|
// and never used.
|
2008-02-16 02:24:29 +08:00
|
|
|
assert(!CopyMI && "physreg was not killed in defining block!");
|
2005-09-02 08:20:32 +08:00
|
|
|
end = getDefIndex(start) + 1; // It's dead.
|
2004-02-01 07:13:30 +08:00
|
|
|
|
2003-11-20 11:32:25 +08:00
|
|
|
exit:
|
2004-08-04 17:46:26 +08:00
|
|
|
assert(start < end && "did not find end of interval?");
|
Allow the live interval analysis pass to be a bit more aggressive about
numbering values in live ranges for physical registers.
The alpha backend currently generates code that looks like this:
vreg = preg
...
preg = vreg
use preg
...
preg = vreg
use preg
etc. Because vreg contains the value of preg coming in, each of the
copies back into preg contain that initial value as well.
In the case of the Alpha, this allows this testcase:
void "foo"(int %blah) {
store int 5, int *%MyVar
store int 12, int* %MyVar2
ret void
}
to compile to:
foo:
ldgp $29, 0($27)
ldiq $0,5
stl $0,MyVar
ldiq $0,12
stl $0,MyVar2
ret $31,($26),1
instead of:
foo:
ldgp $29, 0($27)
bis $29,$29,$0
ldiq $1,5
bis $0,$0,$29
stl $1,MyVar
ldiq $1,12
bis $0,$0,$29
stl $1,MyVar2
ret $31,($26),1
This does not seem to have any noticable effect on X86 code.
This fixes PR535.
llvm-svn: 20536
2005-03-10 07:05:19 +08:00
|
|
|
|
2007-04-25 15:30:23 +08:00
|
|
|
// Already exists? Extend old live interval.
|
|
|
|
LiveInterval::iterator OldLR = interval.FindLiveRangeContaining(start);
|
2007-08-30 04:45:00 +08:00
|
|
|
VNInfo *ValNo = (OldLR != interval.end())
|
2008-02-16 02:24:29 +08:00
|
|
|
? OldLR->valno : interval.getNextValue(start, CopyMI, VNInfoAllocator);
|
2007-08-30 04:45:00 +08:00
|
|
|
LiveRange LR(start, end, ValNo);
|
2004-08-04 17:46:26 +08:00
|
|
|
interval.addRange(LR);
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(LR.valno, end);
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << " +" << LR << '\n';
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
|
|
|
|
2004-07-24 05:24:19 +08:00
|
|
|
void LiveIntervals::handleRegisterDef(MachineBasicBlock *MBB,
|
|
|
|
MachineBasicBlock::iterator MI,
|
2006-09-03 16:07:11 +08:00
|
|
|
unsigned MIIdx,
|
2004-07-24 05:24:19 +08:00
|
|
|
unsigned reg) {
|
2008-02-11 02:45:23 +08:00
|
|
|
if (TargetRegisterInfo::isVirtualRegister(reg))
|
2006-09-03 16:07:11 +08:00
|
|
|
handleVirtualRegisterDef(MBB, MI, MIIdx, getOrCreateInterval(reg));
|
2004-08-27 06:22:38 +08:00
|
|
|
else if (allocatableRegs_[reg]) {
|
2008-02-16 02:24:29 +08:00
|
|
|
MachineInstr *CopyMI = NULL;
|
2006-08-31 13:54:43 +08:00
|
|
|
unsigned SrcReg, DstReg;
|
2008-02-16 02:24:29 +08:00
|
|
|
if (MI->getOpcode() == TargetInstrInfo::EXTRACT_SUBREG ||
|
|
|
|
tii_->isMoveInstr(*MI, SrcReg, DstReg))
|
|
|
|
CopyMI = MI;
|
|
|
|
handlePhysicalRegisterDef(MBB, MI, MIIdx, getOrCreateInterval(reg), CopyMI);
|
2007-04-25 15:30:23 +08:00
|
|
|
// Def of a register also defines its sub-registers.
|
2008-02-11 02:45:23 +08:00
|
|
|
for (const unsigned* AS = tri_->getSubRegisters(reg); *AS; ++AS)
|
2007-04-25 15:30:23 +08:00
|
|
|
// Avoid processing some defs more than once.
|
|
|
|
if (!MI->findRegisterDefOperand(*AS))
|
|
|
|
handlePhysicalRegisterDef(MBB, MI, MIIdx, getOrCreateInterval(*AS), 0);
|
2004-07-24 05:24:19 +08:00
|
|
|
}
|
2004-01-31 22:37:41 +08:00
|
|
|
}
|
|
|
|
|
2007-02-20 05:49:54 +08:00
|
|
|
void LiveIntervals::handleLiveInRegister(MachineBasicBlock *MBB,
|
2007-02-22 06:41:17 +08:00
|
|
|
unsigned MIIdx,
|
2007-04-25 15:30:23 +08:00
|
|
|
LiveInterval &interval, bool isAlias) {
|
2007-02-20 05:49:54 +08:00
|
|
|
DOUT << "\t\tlivein register: "; DEBUG(printRegName(interval.reg));
|
|
|
|
|
|
|
|
// Look for kills, if it reaches a def before it's killed, then it shouldn't
|
|
|
|
// be considered a livein.
|
|
|
|
MachineBasicBlock::iterator mi = MBB->begin();
|
2007-02-22 06:41:17 +08:00
|
|
|
unsigned baseIndex = MIIdx;
|
|
|
|
unsigned start = baseIndex;
|
2007-02-20 05:49:54 +08:00
|
|
|
unsigned end = start;
|
|
|
|
while (mi != MBB->end()) {
|
|
|
|
if (lv_->KillsRegister(mi, interval.reg)) {
|
|
|
|
DOUT << " killed";
|
|
|
|
end = getUseIndex(baseIndex) + 1;
|
|
|
|
goto exit;
|
|
|
|
} else if (lv_->ModifiesRegister(mi, interval.reg)) {
|
|
|
|
// Another instruction redefines the register before it is ever read.
|
|
|
|
// Then the register is essentially dead at the instruction that defines
|
|
|
|
// it. Hence its interval is:
|
|
|
|
// [defSlot(def), defSlot(def)+1)
|
|
|
|
DOUT << " dead";
|
|
|
|
end = getDefIndex(start) + 1;
|
|
|
|
goto exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
baseIndex += InstrSlots::NUM;
|
|
|
|
++mi;
|
|
|
|
}
|
|
|
|
|
|
|
|
exit:
|
2007-06-27 09:16:36 +08:00
|
|
|
// Live-in register might not be used at all.
|
|
|
|
if (end == MIIdx) {
|
2007-06-28 02:47:28 +08:00
|
|
|
if (isAlias) {
|
|
|
|
DOUT << " dead";
|
2007-06-27 09:16:36 +08:00
|
|
|
end = getDefIndex(MIIdx) + 1;
|
2007-06-28 02:47:28 +08:00
|
|
|
} else {
|
|
|
|
DOUT << " live through";
|
|
|
|
end = baseIndex;
|
|
|
|
}
|
2007-04-25 15:30:23 +08:00
|
|
|
}
|
|
|
|
|
2007-09-06 05:46:51 +08:00
|
|
|
LiveRange LR(start, end, interval.getNextValue(start, 0, VNInfoAllocator));
|
2007-02-22 06:41:17 +08:00
|
|
|
interval.addRange(LR);
|
2007-09-06 05:46:51 +08:00
|
|
|
interval.addKill(LR.valno, end);
|
2007-08-08 15:03:29 +08:00
|
|
|
DOUT << " +" << LR << '\n';
|
2007-02-20 05:49:54 +08:00
|
|
|
}
|
|
|
|
|
2003-11-20 11:32:25 +08:00
|
|
|
/// computeIntervals - computes the live intervals for virtual
|
2004-01-31 22:37:41 +08:00
|
|
|
/// registers. for some ordering of the machine instructions [1,N] a
|
2004-02-01 03:59:32 +08:00
|
|
|
/// live interval is an interval [i, j) where 1 <= i <= j < N for
|
2003-11-20 11:32:25 +08:00
|
|
|
/// which a variable is live
|
2006-08-25 06:43:55 +08:00
|
|
|
void LiveIntervals::computeIntervals() {
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << "********** COMPUTING LIVE INTERVALS **********\n"
|
|
|
|
<< "********** Function: "
|
|
|
|
<< ((Value*)mf_->getFunction())->getName() << '\n';
|
2006-09-03 16:07:11 +08:00
|
|
|
// Track the index of the current machine instr.
|
|
|
|
unsigned MIIndex = 0;
|
2006-09-15 11:57:23 +08:00
|
|
|
for (MachineFunction::iterator MBBI = mf_->begin(), E = mf_->end();
|
|
|
|
MBBI != E; ++MBBI) {
|
|
|
|
MachineBasicBlock *MBB = MBBI;
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << ((Value*)MBB->getBasicBlock())->getName() << ":\n";
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2006-09-15 11:57:23 +08:00
|
|
|
MachineBasicBlock::iterator MI = MBB->begin(), miEnd = MBB->end();
|
2007-02-13 09:30:55 +08:00
|
|
|
|
2007-10-04 03:26:29 +08:00
|
|
|
// Create intervals for live-ins to this BB first.
|
|
|
|
for (MachineBasicBlock::const_livein_iterator LI = MBB->livein_begin(),
|
|
|
|
LE = MBB->livein_end(); LI != LE; ++LI) {
|
|
|
|
handleLiveInRegister(MBB, MIIndex, getOrCreateInterval(*LI));
|
|
|
|
// Multiple live-ins can alias the same register.
|
2008-02-11 02:45:23 +08:00
|
|
|
for (const unsigned* AS = tri_->getSubRegisters(*LI); *AS; ++AS)
|
2007-10-04 03:26:29 +08:00
|
|
|
if (!hasInterval(*AS))
|
|
|
|
handleLiveInRegister(MBB, MIIndex, getOrCreateInterval(*AS),
|
|
|
|
true);
|
2006-09-05 02:27:40 +08:00
|
|
|
}
|
|
|
|
|
2006-09-15 11:57:23 +08:00
|
|
|
for (; MI != miEnd; ++MI) {
|
2006-11-29 08:39:47 +08:00
|
|
|
DOUT << MIIndex << "\t" << *MI;
|
2004-08-04 17:46:26 +08:00
|
|
|
|
2006-11-10 16:43:01 +08:00
|
|
|
// Handle defs.
|
2006-09-15 11:57:23 +08:00
|
|
|
for (int i = MI->getNumOperands() - 1; i >= 0; --i) {
|
|
|
|
MachineOperand &MO = MI->getOperand(i);
|
2004-08-04 17:46:26 +08:00
|
|
|
// handle register defs - build intervals
|
2006-09-15 11:57:23 +08:00
|
|
|
if (MO.isRegister() && MO.getReg() && MO.isDef())
|
|
|
|
handleRegisterDef(MBB, MI, MIIndex, MO.getReg());
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2006-09-03 16:07:11 +08:00
|
|
|
|
|
|
|
MIIndex += InstrSlots::NUM;
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
2004-08-04 17:46:26 +08:00
|
|
|
}
|
2003-11-20 11:32:25 +08:00
|
|
|
}
|
2003-12-05 18:38:28 +08:00
|
|
|
|
2007-10-17 10:10:22 +08:00
|
|
|
bool LiveIntervals::findLiveInMBBs(const LiveRange &LR,
|
2007-10-17 14:53:44 +08:00
|
|
|
SmallVectorImpl<MachineBasicBlock*> &MBBs) const {
|
2007-10-17 10:10:22 +08:00
|
|
|
std::vector<IdxMBBPair>::const_iterator I =
|
|
|
|
std::lower_bound(Idx2MBBMap.begin(), Idx2MBBMap.end(), LR.start);
|
|
|
|
|
|
|
|
bool ResVal = false;
|
|
|
|
while (I != Idx2MBBMap.end()) {
|
|
|
|
if (LR.end <= I->first)
|
|
|
|
break;
|
|
|
|
MBBs.push_back(I->second);
|
|
|
|
ResVal = true;
|
|
|
|
++I;
|
|
|
|
}
|
|
|
|
return ResVal;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-07-24 19:44:15 +08:00
|
|
|
LiveInterval LiveIntervals::createInterval(unsigned reg) {
|
2008-02-11 02:45:23 +08:00
|
|
|
float Weight = TargetRegisterInfo::isPhysicalRegister(reg) ?
|
2006-11-07 20:25:45 +08:00
|
|
|
HUGE_VALF : 0.0F;
|
2004-07-24 19:44:15 +08:00
|
|
|
return LiveInterval(reg, Weight);
|
2004-04-10 02:07:57 +08:00
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
|
2008-02-16 02:24:29 +08:00
|
|
|
/// getVNInfoSourceReg - Helper function that parses the specified VNInfo
|
|
|
|
/// copy field and returns the source register that defines it.
|
|
|
|
unsigned LiveIntervals::getVNInfoSourceReg(const VNInfo *VNI) const {
|
|
|
|
if (!VNI->copy)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (VNI->copy->getOpcode() == TargetInstrInfo::EXTRACT_SUBREG)
|
|
|
|
return VNI->copy->getOperand(1).getReg();
|
|
|
|
unsigned SrcReg, DstReg;
|
|
|
|
if (tii_->isMoveInstr(*VNI->copy, SrcReg, DstReg))
|
|
|
|
return SrcReg;
|
|
|
|
assert(0 && "Unrecognized copy instruction!");
|
|
|
|
return 0;
|
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Register allocator hooks.
|
|
|
|
//
|
|
|
|
|
|
|
|
/// isReMaterializable - Returns true if the definition MI of the specified
|
|
|
|
/// val# of the specified interval is re-materializable.
|
|
|
|
bool LiveIntervals::isReMaterializable(const LiveInterval &li,
|
2007-12-06 08:01:56 +08:00
|
|
|
const VNInfo *ValNo, MachineInstr *MI,
|
|
|
|
bool &isLoad) {
|
2007-11-12 14:35:08 +08:00
|
|
|
if (DisableReMat)
|
|
|
|
return false;
|
|
|
|
|
2007-12-06 08:01:56 +08:00
|
|
|
isLoad = false;
|
2008-01-07 15:27:27 +08:00
|
|
|
const TargetInstrDesc &TID = MI->getDesc();
|
|
|
|
if (TID.isImplicitDef() || tii_->isTriviallyReMaterializable(MI)) {
|
|
|
|
isLoad = TID.isSimpleLoad();
|
2007-11-12 14:35:08 +08:00
|
|
|
return true;
|
2007-12-06 08:01:56 +08:00
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
|
|
|
|
int FrameIdx = 0;
|
|
|
|
if (!tii_->isLoadFromStackSlot(MI, FrameIdx) ||
|
2008-01-10 16:24:38 +08:00
|
|
|
!mf_->getFrameInfo()->isImmutableObjectIndex(FrameIdx))
|
2007-11-12 14:35:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// This is a load from fixed stack slot. It can be rematerialized unless it's
|
|
|
|
// re-defined by a two-address instruction.
|
2007-12-06 08:01:56 +08:00
|
|
|
isLoad = true;
|
2007-11-12 14:35:08 +08:00
|
|
|
for (LiveInterval::const_vni_iterator i = li.vni_begin(), e = li.vni_end();
|
|
|
|
i != e; ++i) {
|
|
|
|
const VNInfo *VNI = *i;
|
|
|
|
if (VNI == ValNo)
|
|
|
|
continue;
|
|
|
|
unsigned DefIdx = VNI->def;
|
|
|
|
if (DefIdx == ~1U)
|
|
|
|
continue; // Dead val#.
|
|
|
|
MachineInstr *DefMI = (DefIdx == ~0u)
|
|
|
|
? NULL : getInstructionFromIndex(DefIdx);
|
2007-12-06 08:01:56 +08:00
|
|
|
if (DefMI && DefMI->isRegReDefinedByTwoAddr(li.reg)) {
|
|
|
|
isLoad = false;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// isReMaterializable - Returns true if every definition of MI of every
|
|
|
|
/// val# of the specified interval is re-materializable.
|
|
|
|
bool LiveIntervals::isReMaterializable(const LiveInterval &li, bool &isLoad) {
|
|
|
|
isLoad = false;
|
|
|
|
for (LiveInterval::const_vni_iterator i = li.vni_begin(), e = li.vni_end();
|
|
|
|
i != e; ++i) {
|
|
|
|
const VNInfo *VNI = *i;
|
|
|
|
unsigned DefIdx = VNI->def;
|
|
|
|
if (DefIdx == ~1U)
|
|
|
|
continue; // Dead val#.
|
|
|
|
// Is the def for the val# rematerializable?
|
|
|
|
if (DefIdx == ~0u)
|
|
|
|
return false;
|
|
|
|
MachineInstr *ReMatDefMI = getInstructionFromIndex(DefIdx);
|
|
|
|
bool DefIsLoad = false;
|
|
|
|
if (!ReMatDefMI || !isReMaterializable(li, VNI, ReMatDefMI, DefIsLoad))
|
2007-11-12 14:35:08 +08:00
|
|
|
return false;
|
2007-12-06 08:01:56 +08:00
|
|
|
isLoad |= DefIsLoad;
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// tryFoldMemoryOperand - Attempts to fold either a spill / restore from
|
|
|
|
/// slot / to reg or any rematerialized load into ith operand of specified
|
|
|
|
/// MI. If it is successul, MI is updated with the newly created MI and
|
|
|
|
/// returns true.
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool LiveIntervals::tryFoldMemoryOperand(MachineInstr* &MI,
|
2007-12-01 05:23:43 +08:00
|
|
|
VirtRegMap &vrm, MachineInstr *DefMI,
|
2007-12-02 16:30:39 +08:00
|
|
|
unsigned InstrIdx,
|
|
|
|
SmallVector<unsigned, 2> &Ops,
|
2007-12-01 05:23:43 +08:00
|
|
|
bool isSS, int Slot, unsigned Reg) {
|
2007-12-02 16:30:39 +08:00
|
|
|
unsigned MRInfo = 0;
|
2008-01-07 15:27:27 +08:00
|
|
|
const TargetInstrDesc &TID = MI->getDesc();
|
2007-12-13 07:12:09 +08:00
|
|
|
// If it is an implicit def instruction, just delete it.
|
2008-01-07 15:27:27 +08:00
|
|
|
if (TID.isImplicitDef()) {
|
2007-12-13 07:12:09 +08:00
|
|
|
RemoveMachineInstrFromMaps(MI);
|
|
|
|
vrm.RemoveMachineInstrFromMaps(MI);
|
|
|
|
MI->eraseFromParent();
|
|
|
|
++numFolds;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2007-12-02 16:30:39 +08:00
|
|
|
SmallVector<unsigned, 2> FoldOps;
|
|
|
|
for (unsigned i = 0, e = Ops.size(); i != e; ++i) {
|
|
|
|
unsigned OpIdx = Ops[i];
|
|
|
|
// FIXME: fold subreg use.
|
|
|
|
if (MI->getOperand(OpIdx).getSubReg())
|
2007-12-01 10:07:52 +08:00
|
|
|
return false;
|
2007-12-02 16:30:39 +08:00
|
|
|
if (MI->getOperand(OpIdx).isDef())
|
|
|
|
MRInfo |= (unsigned)VirtRegMap::isMod;
|
|
|
|
else {
|
|
|
|
// Filter out two-address use operand(s).
|
2008-01-07 15:27:27 +08:00
|
|
|
if (TID.getOperandConstraint(OpIdx, TOI::TIED_TO) != -1) {
|
2007-12-02 16:30:39 +08:00
|
|
|
MRInfo = VirtRegMap::isModRef;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
MRInfo |= (unsigned)VirtRegMap::isRef;
|
|
|
|
}
|
|
|
|
FoldOps.push_back(OpIdx);
|
2007-12-01 10:07:52 +08:00
|
|
|
}
|
2007-12-01 05:23:43 +08:00
|
|
|
|
2008-02-09 06:05:27 +08:00
|
|
|
MachineInstr *fmi = isSS ? tii_->foldMemoryOperand(*mf_, MI, FoldOps, Slot)
|
|
|
|
: tii_->foldMemoryOperand(*mf_, MI, FoldOps, DefMI);
|
2007-11-12 14:35:08 +08:00
|
|
|
if (fmi) {
|
|
|
|
// Attempt to fold the memory reference into the instruction. If
|
|
|
|
// we can do this, we don't need to insert spill code.
|
|
|
|
if (lv_)
|
|
|
|
lv_->instructionChanged(MI, fmi);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
else
|
2008-02-11 02:45:23 +08:00
|
|
|
fmi->copyKillDeadInfo(MI, tri_);
|
2007-11-12 14:35:08 +08:00
|
|
|
MachineBasicBlock &MBB = *MI->getParent();
|
2008-01-10 16:24:38 +08:00
|
|
|
if (isSS && !mf_->getFrameInfo()->isImmutableObjectIndex(Slot))
|
2007-12-02 16:30:39 +08:00
|
|
|
vrm.virtFolded(Reg, MI, fmi, (VirtRegMap::ModRef)MRInfo);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
vrm.transferSpillPts(MI, fmi);
|
2007-11-29 09:06:25 +08:00
|
|
|
vrm.transferRestorePts(MI, fmi);
|
2007-11-12 14:35:08 +08:00
|
|
|
mi2iMap_.erase(MI);
|
2007-12-01 05:23:43 +08:00
|
|
|
i2miMap_[InstrIdx /InstrSlots::NUM] = fmi;
|
|
|
|
mi2iMap_[fmi] = InstrIdx;
|
2007-11-12 14:35:08 +08:00
|
|
|
MI = MBB.insert(MBB.erase(MI), fmi);
|
2007-11-29 09:06:25 +08:00
|
|
|
++numFolds;
|
2007-11-12 14:35:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2007-12-05 11:22:34 +08:00
|
|
|
/// canFoldMemoryOperand - Returns true if the specified load / store
|
|
|
|
/// folding is possible.
|
|
|
|
bool LiveIntervals::canFoldMemoryOperand(MachineInstr *MI,
|
|
|
|
SmallVector<unsigned, 2> &Ops) const {
|
|
|
|
SmallVector<unsigned, 2> FoldOps;
|
|
|
|
for (unsigned i = 0, e = Ops.size(); i != e; ++i) {
|
|
|
|
unsigned OpIdx = Ops[i];
|
|
|
|
// FIXME: fold subreg use.
|
|
|
|
if (MI->getOperand(OpIdx).getSubReg())
|
|
|
|
return false;
|
|
|
|
FoldOps.push_back(OpIdx);
|
|
|
|
}
|
|
|
|
|
2008-01-07 09:35:56 +08:00
|
|
|
return tii_->canFoldMemoryOperand(MI, FoldOps);
|
2007-12-05 11:22:34 +08:00
|
|
|
}
|
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool LiveIntervals::intervalIsInOneMBB(const LiveInterval &li) const {
|
|
|
|
SmallPtrSet<MachineBasicBlock*, 4> MBBs;
|
|
|
|
for (LiveInterval::Ranges::const_iterator
|
|
|
|
I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) {
|
|
|
|
std::vector<IdxMBBPair>::const_iterator II =
|
|
|
|
std::lower_bound(Idx2MBBMap.begin(), Idx2MBBMap.end(), I->start);
|
|
|
|
if (II == Idx2MBBMap.end())
|
|
|
|
continue;
|
|
|
|
if (I->end > II->first) // crossing a MBB.
|
|
|
|
return false;
|
|
|
|
MBBs.insert(II->second);
|
|
|
|
if (MBBs.size() > 1)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2007-11-12 14:35:08 +08:00
|
|
|
/// rewriteInstructionForSpills, rewriteInstructionsForSpills - Helper functions
|
|
|
|
/// for addIntervalsForSpills to rewrite uses / defs for the given live range.
|
2007-12-05 11:22:34 +08:00
|
|
|
bool LiveIntervals::
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
rewriteInstructionForSpills(const LiveInterval &li, bool TrySplit,
|
|
|
|
unsigned id, unsigned index, unsigned end, MachineInstr *MI,
|
|
|
|
MachineInstr *ReMatOrigDefMI, MachineInstr *ReMatDefMI,
|
2007-11-12 14:35:08 +08:00
|
|
|
unsigned Slot, int LdSlot,
|
|
|
|
bool isLoad, bool isLoadSS, bool DefIsReMat, bool CanDelete,
|
2007-12-31 12:13:23 +08:00
|
|
|
VirtRegMap &vrm, MachineRegisterInfo &RegInfo,
|
2007-11-12 14:35:08 +08:00
|
|
|
const TargetRegisterClass* rc,
|
|
|
|
SmallVector<int, 4> &ReMatIds,
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
unsigned &NewVReg, bool &HasDef, bool &HasUse,
|
2007-12-11 10:09:15 +08:00
|
|
|
const MachineLoopInfo *loopInfo,
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned,unsigned> &MBBVRegsMap,
|
2007-11-12 14:35:08 +08:00
|
|
|
std::vector<LiveInterval*> &NewLIs) {
|
2007-12-05 11:22:34 +08:00
|
|
|
bool CanFold = false;
|
2007-11-12 14:35:08 +08:00
|
|
|
RestartInstruction:
|
|
|
|
for (unsigned i = 0; i != MI->getNumOperands(); ++i) {
|
|
|
|
MachineOperand& mop = MI->getOperand(i);
|
|
|
|
if (!mop.isRegister())
|
|
|
|
continue;
|
|
|
|
unsigned Reg = mop.getReg();
|
|
|
|
unsigned RegI = Reg;
|
2008-02-11 02:45:23 +08:00
|
|
|
if (Reg == 0 || TargetRegisterInfo::isPhysicalRegister(Reg))
|
2007-11-12 14:35:08 +08:00
|
|
|
continue;
|
|
|
|
if (Reg != li.reg)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
bool TryFold = !DefIsReMat;
|
2007-11-30 07:02:50 +08:00
|
|
|
bool FoldSS = true; // Default behavior unless it's a remat.
|
2007-11-12 14:35:08 +08:00
|
|
|
int FoldSlot = Slot;
|
|
|
|
if (DefIsReMat) {
|
|
|
|
// If this is the rematerializable definition MI itself and
|
|
|
|
// all of its uses are rematerialized, simply delete it.
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (MI == ReMatOrigDefMI && CanDelete) {
|
2007-12-01 05:23:43 +08:00
|
|
|
DOUT << "\t\t\t\tErasing re-materlizable def: ";
|
|
|
|
DOUT << MI << '\n';
|
2007-11-12 14:35:08 +08:00
|
|
|
RemoveMachineInstrFromMaps(MI);
|
2007-11-28 09:28:46 +08:00
|
|
|
vrm.RemoveMachineInstrFromMaps(MI);
|
2007-11-12 14:35:08 +08:00
|
|
|
MI->eraseFromParent();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If def for this use can't be rematerialized, then try folding.
|
2007-11-29 09:06:25 +08:00
|
|
|
// If def is rematerializable and it's a load, also try folding.
|
2007-11-30 07:02:50 +08:00
|
|
|
TryFold = !ReMatDefMI || (ReMatDefMI && (MI == ReMatOrigDefMI || isLoad));
|
2007-11-12 14:35:08 +08:00
|
|
|
if (isLoad) {
|
|
|
|
// Try fold loads (from stack slot, constant pool, etc.) into uses.
|
|
|
|
FoldSS = isLoadSS;
|
|
|
|
FoldSlot = LdSlot;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Scan all of the operands of this instruction rewriting operands
|
|
|
|
// to use NewVReg instead of li.reg as appropriate. We do this for
|
|
|
|
// two reasons:
|
|
|
|
//
|
|
|
|
// 1. If the instr reads the same spilled vreg multiple times, we
|
|
|
|
// want to reuse the NewVReg.
|
|
|
|
// 2. If the instr is a two-addr instruction, we are required to
|
|
|
|
// keep the src/dst regs pinned.
|
|
|
|
//
|
|
|
|
// Keep track of whether we replace a use and/or def so that we can
|
|
|
|
// create the spill interval with the appropriate range.
|
2007-12-01 05:23:43 +08:00
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
HasUse = mop.isUse();
|
|
|
|
HasDef = mop.isDef();
|
2007-12-02 16:30:39 +08:00
|
|
|
SmallVector<unsigned, 2> Ops;
|
|
|
|
Ops.push_back(i);
|
2007-11-12 14:35:08 +08:00
|
|
|
for (unsigned j = i+1, e = MI->getNumOperands(); j != e; ++j) {
|
2007-12-02 16:30:39 +08:00
|
|
|
const MachineOperand &MOj = MI->getOperand(j);
|
|
|
|
if (!MOj.isRegister())
|
2007-11-12 14:35:08 +08:00
|
|
|
continue;
|
2007-12-02 16:30:39 +08:00
|
|
|
unsigned RegJ = MOj.getReg();
|
2008-02-11 02:45:23 +08:00
|
|
|
if (RegJ == 0 || TargetRegisterInfo::isPhysicalRegister(RegJ))
|
2007-11-12 14:35:08 +08:00
|
|
|
continue;
|
|
|
|
if (RegJ == RegI) {
|
2007-12-02 16:30:39 +08:00
|
|
|
Ops.push_back(j);
|
|
|
|
HasUse |= MOj.isUse();
|
|
|
|
HasDef |= MOj.isDef();
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-12-05 11:22:34 +08:00
|
|
|
if (TryFold) {
|
|
|
|
// Do not fold load / store here if we are splitting. We'll find an
|
|
|
|
// optimal point to insert a load / store later.
|
|
|
|
if (!TrySplit) {
|
|
|
|
if (tryFoldMemoryOperand(MI, vrm, ReMatDefMI, index,
|
|
|
|
Ops, FoldSS, FoldSlot, Reg)) {
|
|
|
|
// Folding the load/store can completely change the instruction in
|
|
|
|
// unpredictable ways, rescan it from the beginning.
|
|
|
|
HasUse = false;
|
|
|
|
HasDef = false;
|
|
|
|
CanFold = false;
|
|
|
|
goto RestartInstruction;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
CanFold = canFoldMemoryOperand(MI, Ops);
|
|
|
|
}
|
2007-12-13 07:12:09 +08:00
|
|
|
} else
|
|
|
|
CanFold = false;
|
2007-12-01 05:23:43 +08:00
|
|
|
|
|
|
|
// Create a new virtual register for the spill interval.
|
|
|
|
bool CreatedNewVReg = false;
|
|
|
|
if (NewVReg == 0) {
|
2007-12-31 12:13:23 +08:00
|
|
|
NewVReg = RegInfo.createVirtualRegister(rc);
|
2007-12-01 05:23:43 +08:00
|
|
|
vrm.grow();
|
|
|
|
CreatedNewVReg = true;
|
|
|
|
}
|
|
|
|
mop.setReg(NewVReg);
|
|
|
|
|
|
|
|
// Reuse NewVReg for other reads.
|
2007-12-02 16:30:39 +08:00
|
|
|
for (unsigned j = 0, e = Ops.size(); j != e; ++j)
|
|
|
|
MI->getOperand(Ops[j]).setReg(NewVReg);
|
2007-12-01 05:23:43 +08:00
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (CreatedNewVReg) {
|
|
|
|
if (DefIsReMat) {
|
|
|
|
vrm.setVirtIsReMaterialized(NewVReg, ReMatDefMI/*, CanDelete*/);
|
|
|
|
if (ReMatIds[id] == VirtRegMap::MAX_STACK_SLOT) {
|
|
|
|
// Each valnum may have its own remat id.
|
|
|
|
ReMatIds[id] = vrm.assignVirtReMatId(NewVReg);
|
|
|
|
} else {
|
|
|
|
vrm.assignVirtReMatId(NewVReg, ReMatIds[id]);
|
|
|
|
}
|
|
|
|
if (!CanDelete || (HasUse && HasDef)) {
|
|
|
|
// If this is a two-addr instruction then its use operands are
|
|
|
|
// rematerializable but its def is not. It should be assigned a
|
|
|
|
// stack slot.
|
|
|
|
vrm.assignVirt2StackSlot(NewVReg, Slot);
|
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
} else {
|
|
|
|
vrm.assignVirt2StackSlot(NewVReg, Slot);
|
|
|
|
}
|
2007-11-30 07:02:50 +08:00
|
|
|
} else if (HasUse && HasDef &&
|
|
|
|
vrm.getStackSlot(NewVReg) == VirtRegMap::NO_STACK_SLOT) {
|
|
|
|
// If this interval hasn't been assigned a stack slot (because earlier
|
|
|
|
// def is a deleted remat def), do it now.
|
|
|
|
assert(Slot != VirtRegMap::NO_STACK_SLOT);
|
|
|
|
vrm.assignVirt2StackSlot(NewVReg, Slot);
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// create a new register interval for this spill / remat.
|
|
|
|
LiveInterval &nI = getOrCreateInterval(NewVReg);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (CreatedNewVReg) {
|
|
|
|
NewLIs.push_back(&nI);
|
2007-11-29 18:12:14 +08:00
|
|
|
MBBVRegsMap.insert(std::make_pair(MI->getParent()->getNumber(), NewVReg));
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (TrySplit)
|
|
|
|
vrm.setIsSplitFromReg(NewVReg, li.reg);
|
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
|
|
|
|
if (HasUse) {
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (CreatedNewVReg) {
|
|
|
|
LiveRange LR(getLoadIndex(index), getUseIndex(index)+1,
|
|
|
|
nI.getNextValue(~0U, 0, VNInfoAllocator));
|
|
|
|
DOUT << " +" << LR;
|
|
|
|
nI.addRange(LR);
|
|
|
|
} else {
|
|
|
|
// Extend the split live interval to this def / use.
|
|
|
|
unsigned End = getUseIndex(index)+1;
|
|
|
|
LiveRange LR(nI.ranges[nI.ranges.size()-1].end, End,
|
|
|
|
nI.getValNumInfo(nI.getNumValNums()-1));
|
|
|
|
DOUT << " +" << LR;
|
|
|
|
nI.addRange(LR);
|
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
if (HasDef) {
|
|
|
|
LiveRange LR(getDefIndex(index), getStoreIndex(index),
|
|
|
|
nI.getNextValue(~0U, 0, VNInfoAllocator));
|
|
|
|
DOUT << " +" << LR;
|
|
|
|
nI.addRange(LR);
|
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
|
2007-11-12 14:35:08 +08:00
|
|
|
DOUT << "\t\t\t\tAdded new interval: ";
|
2008-02-11 02:45:23 +08:00
|
|
|
nI.print(DOUT, tri_);
|
2007-11-12 14:35:08 +08:00
|
|
|
DOUT << '\n';
|
|
|
|
}
|
2007-12-05 11:22:34 +08:00
|
|
|
return CanFold;
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool LiveIntervals::anyKillInMBBAfterIdx(const LiveInterval &li,
|
2007-11-29 09:06:25 +08:00
|
|
|
const VNInfo *VNI,
|
|
|
|
MachineBasicBlock *MBB, unsigned Idx) const {
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
unsigned End = getMBBEndIdx(MBB);
|
2007-11-29 09:06:25 +08:00
|
|
|
for (unsigned j = 0, ee = VNI->kills.size(); j != ee; ++j) {
|
|
|
|
unsigned KillIdx = VNI->kills[j];
|
|
|
|
if (KillIdx > Idx && KillIdx < End)
|
|
|
|
return true;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2007-11-29 18:12:14 +08:00
|
|
|
static const VNInfo *findDefinedVNInfo(const LiveInterval &li, unsigned DefIdx) {
|
|
|
|
const VNInfo *VNI = NULL;
|
|
|
|
for (LiveInterval::const_vni_iterator i = li.vni_begin(),
|
|
|
|
e = li.vni_end(); i != e; ++i)
|
|
|
|
if ((*i)->def == DefIdx) {
|
|
|
|
VNI = *i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return VNI;
|
|
|
|
}
|
|
|
|
|
2007-11-12 14:35:08 +08:00
|
|
|
void LiveIntervals::
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
rewriteInstructionsForSpills(const LiveInterval &li, bool TrySplit,
|
2007-11-12 14:35:08 +08:00
|
|
|
LiveInterval::Ranges::const_iterator &I,
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
MachineInstr *ReMatOrigDefMI, MachineInstr *ReMatDefMI,
|
2007-11-12 14:35:08 +08:00
|
|
|
unsigned Slot, int LdSlot,
|
|
|
|
bool isLoad, bool isLoadSS, bool DefIsReMat, bool CanDelete,
|
2007-12-31 12:13:23 +08:00
|
|
|
VirtRegMap &vrm, MachineRegisterInfo &RegInfo,
|
2007-11-12 14:35:08 +08:00
|
|
|
const TargetRegisterClass* rc,
|
|
|
|
SmallVector<int, 4> &ReMatIds,
|
2007-12-11 10:09:15 +08:00
|
|
|
const MachineLoopInfo *loopInfo,
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
BitVector &SpillMBBs,
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> > &SpillIdxes,
|
2007-11-29 09:06:25 +08:00
|
|
|
BitVector &RestoreMBBs,
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> > &RestoreIdxes,
|
|
|
|
std::map<unsigned,unsigned> &MBBVRegsMap,
|
2007-11-12 14:35:08 +08:00
|
|
|
std::vector<LiveInterval*> &NewLIs) {
|
2007-12-05 11:22:34 +08:00
|
|
|
bool AllCanFold = true;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
unsigned NewVReg = 0;
|
2007-11-12 14:35:08 +08:00
|
|
|
unsigned index = getBaseIndex(I->start);
|
|
|
|
unsigned end = getBaseIndex(I->end-1) + InstrSlots::NUM;
|
|
|
|
for (; index != end; index += InstrSlots::NUM) {
|
|
|
|
// skip deleted instructions
|
|
|
|
while (index != end && !getInstructionFromIndex(index))
|
|
|
|
index += InstrSlots::NUM;
|
|
|
|
if (index == end) break;
|
|
|
|
|
|
|
|
MachineInstr *MI = getInstructionFromIndex(index);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
2007-12-05 11:22:34 +08:00
|
|
|
unsigned ThisVReg = 0;
|
2007-12-03 17:58:48 +08:00
|
|
|
if (TrySplit) {
|
2007-11-28 09:28:46 +08:00
|
|
|
std::map<unsigned,unsigned>::const_iterator NVI =
|
2007-11-29 18:12:14 +08:00
|
|
|
MBBVRegsMap.find(MBB->getNumber());
|
|
|
|
if (NVI != MBBVRegsMap.end()) {
|
2007-12-05 11:22:34 +08:00
|
|
|
ThisVReg = NVI->second;
|
2007-11-29 18:12:14 +08:00
|
|
|
// One common case:
|
|
|
|
// x = use
|
|
|
|
// ...
|
|
|
|
// ...
|
|
|
|
// def = ...
|
|
|
|
// = use
|
|
|
|
// It's better to start a new interval to avoid artifically
|
|
|
|
// extend the new interval.
|
|
|
|
// FIXME: Too slow? Can we fix it after rewriteInstructionsForSpills?
|
|
|
|
bool MIHasUse = false;
|
|
|
|
bool MIHasDef = false;
|
|
|
|
for (unsigned i = 0; i != MI->getNumOperands(); ++i) {
|
|
|
|
MachineOperand& mop = MI->getOperand(i);
|
|
|
|
if (!mop.isRegister() || mop.getReg() != li.reg)
|
|
|
|
continue;
|
|
|
|
if (mop.isUse())
|
|
|
|
MIHasUse = true;
|
|
|
|
else
|
|
|
|
MIHasDef = true;
|
|
|
|
}
|
|
|
|
if (MIHasDef && !MIHasUse) {
|
|
|
|
MBBVRegsMap.erase(MBB->getNumber());
|
2007-12-05 11:22:34 +08:00
|
|
|
ThisVReg = 0;
|
2007-11-29 18:12:14 +08:00
|
|
|
}
|
|
|
|
}
|
2007-11-28 09:28:46 +08:00
|
|
|
}
|
2007-12-05 11:22:34 +08:00
|
|
|
|
|
|
|
bool IsNew = ThisVReg == 0;
|
|
|
|
if (IsNew) {
|
|
|
|
// This ends the previous live interval. If all of its def / use
|
|
|
|
// can be folded, give it a low spill weight.
|
|
|
|
if (NewVReg && TrySplit && AllCanFold) {
|
|
|
|
LiveInterval &nI = getOrCreateInterval(NewVReg);
|
|
|
|
nI.weight /= 10.0F;
|
|
|
|
}
|
|
|
|
AllCanFold = true;
|
|
|
|
}
|
|
|
|
NewVReg = ThisVReg;
|
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool HasDef = false;
|
|
|
|
bool HasUse = false;
|
2007-12-05 11:22:34 +08:00
|
|
|
bool CanFold = rewriteInstructionForSpills(li, TrySplit, I->valno->id,
|
|
|
|
index, end, MI, ReMatOrigDefMI, ReMatDefMI,
|
|
|
|
Slot, LdSlot, isLoad, isLoadSS, DefIsReMat,
|
2007-12-31 12:13:23 +08:00
|
|
|
CanDelete, vrm, RegInfo, rc, ReMatIds, NewVReg,
|
2007-12-05 11:22:34 +08:00
|
|
|
HasDef, HasUse, loopInfo, MBBVRegsMap, NewLIs);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (!HasDef && !HasUse)
|
|
|
|
continue;
|
|
|
|
|
2007-12-05 11:22:34 +08:00
|
|
|
AllCanFold &= CanFold;
|
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
// Update weight of spill interval.
|
|
|
|
LiveInterval &nI = getOrCreateInterval(NewVReg);
|
2007-12-03 17:58:48 +08:00
|
|
|
if (!TrySplit) {
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
// The spill weight is now infinity as it cannot be spilled again.
|
|
|
|
nI.weight = HUGE_VALF;
|
2007-11-29 09:06:25 +08:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Keep track of the last def and first use in each MBB.
|
|
|
|
unsigned MBBId = MBB->getNumber();
|
|
|
|
if (HasDef) {
|
|
|
|
if (MI != ReMatOrigDefMI || !CanDelete) {
|
|
|
|
bool HasKill = false;
|
|
|
|
if (!HasUse)
|
|
|
|
HasKill = anyKillInMBBAfterIdx(li, I->valno, MBB, getDefIndex(index));
|
|
|
|
else {
|
2007-11-29 18:12:14 +08:00
|
|
|
// If this is a two-address code, then this index starts a new VNInfo.
|
|
|
|
const VNInfo *VNI = findDefinedVNInfo(li, getDefIndex(index));
|
2007-11-29 09:06:25 +08:00
|
|
|
if (VNI)
|
|
|
|
HasKill = anyKillInMBBAfterIdx(li, VNI, MBB, getDefIndex(index));
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
2007-12-01 12:42:39 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> >::iterator SII =
|
|
|
|
SpillIdxes.find(MBBId);
|
2007-11-29 09:06:25 +08:00
|
|
|
if (!HasKill) {
|
2007-11-29 18:12:14 +08:00
|
|
|
if (SII == SpillIdxes.end()) {
|
|
|
|
std::vector<SRInfo> S;
|
|
|
|
S.push_back(SRInfo(index, NewVReg, true));
|
|
|
|
SpillIdxes.insert(std::make_pair(MBBId, S));
|
|
|
|
} else if (SII->second.back().vreg != NewVReg) {
|
|
|
|
SII->second.push_back(SRInfo(index, NewVReg, true));
|
|
|
|
} else if ((int)index > SII->second.back().index) {
|
2007-11-29 09:06:25 +08:00
|
|
|
// If there is an earlier def and this is a two-address
|
|
|
|
// instruction, then it's not possible to fold the store (which
|
|
|
|
// would also fold the load).
|
2007-11-29 18:12:14 +08:00
|
|
|
SRInfo &Info = SII->second.back();
|
|
|
|
Info.index = index;
|
|
|
|
Info.canFold = !HasUse;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
SpillMBBs.set(MBBId);
|
2007-12-01 12:42:39 +08:00
|
|
|
} else if (SII != SpillIdxes.end() &&
|
|
|
|
SII->second.back().vreg == NewVReg &&
|
|
|
|
(int)index > SII->second.back().index) {
|
|
|
|
// There is an earlier def that's not killed (must be two-address).
|
|
|
|
// The spill is no longer needed.
|
|
|
|
SII->second.pop_back();
|
|
|
|
if (SII->second.empty()) {
|
|
|
|
SpillIdxes.erase(MBBId);
|
|
|
|
SpillMBBs.reset(MBBId);
|
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
|
2007-11-29 09:06:25 +08:00
|
|
|
if (HasUse) {
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> >::iterator SII =
|
2007-11-29 09:06:25 +08:00
|
|
|
SpillIdxes.find(MBBId);
|
2007-11-29 18:12:14 +08:00
|
|
|
if (SII != SpillIdxes.end() &&
|
|
|
|
SII->second.back().vreg == NewVReg &&
|
|
|
|
(int)index > SII->second.back().index)
|
2007-11-29 09:06:25 +08:00
|
|
|
// Use(s) following the last def, it's not safe to fold the spill.
|
2007-11-29 18:12:14 +08:00
|
|
|
SII->second.back().canFold = false;
|
|
|
|
std::map<unsigned, std::vector<SRInfo> >::iterator RII =
|
2007-11-29 09:06:25 +08:00
|
|
|
RestoreIdxes.find(MBBId);
|
2007-11-29 18:12:14 +08:00
|
|
|
if (RII != RestoreIdxes.end() && RII->second.back().vreg == NewVReg)
|
2007-11-29 09:06:25 +08:00
|
|
|
// If we are splitting live intervals, only fold if it's the first
|
|
|
|
// use and there isn't another use later in the MBB.
|
2007-11-29 18:12:14 +08:00
|
|
|
RII->second.back().canFold = false;
|
2007-11-29 09:06:25 +08:00
|
|
|
else if (IsNew) {
|
|
|
|
// Only need a reload if there isn't an earlier def / use.
|
2007-11-29 18:12:14 +08:00
|
|
|
if (RII == RestoreIdxes.end()) {
|
|
|
|
std::vector<SRInfo> Infos;
|
|
|
|
Infos.push_back(SRInfo(index, NewVReg, true));
|
|
|
|
RestoreIdxes.insert(std::make_pair(MBBId, Infos));
|
|
|
|
} else {
|
|
|
|
RII->second.push_back(SRInfo(index, NewVReg, true));
|
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
RestoreMBBs.set(MBBId);
|
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
|
|
|
|
// Update spill weight.
|
2007-12-11 10:09:15 +08:00
|
|
|
unsigned loopDepth = loopInfo->getLoopDepth(MBB);
|
2007-11-29 09:06:25 +08:00
|
|
|
nI.weight += getSpillWeight(HasDef, HasUse, loopDepth);
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
2007-12-05 11:22:34 +08:00
|
|
|
|
|
|
|
if (NewVReg && TrySplit && AllCanFold) {
|
|
|
|
// If all of its def / use can be folded, give it a low spill weight.
|
|
|
|
LiveInterval &nI = getOrCreateInterval(NewVReg);
|
|
|
|
nI.weight /= 10.0F;
|
|
|
|
}
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
|
2007-11-29 18:12:14 +08:00
|
|
|
bool LiveIntervals::alsoFoldARestore(int Id, int index, unsigned vr,
|
|
|
|
BitVector &RestoreMBBs,
|
|
|
|
std::map<unsigned,std::vector<SRInfo> > &RestoreIdxes) {
|
|
|
|
if (!RestoreMBBs[Id])
|
|
|
|
return false;
|
|
|
|
std::vector<SRInfo> &Restores = RestoreIdxes[Id];
|
|
|
|
for (unsigned i = 0, e = Restores.size(); i != e; ++i)
|
|
|
|
if (Restores[i].index == index &&
|
|
|
|
Restores[i].vreg == vr &&
|
|
|
|
Restores[i].canFold)
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void LiveIntervals::eraseRestoreInfo(int Id, int index, unsigned vr,
|
|
|
|
BitVector &RestoreMBBs,
|
|
|
|
std::map<unsigned,std::vector<SRInfo> > &RestoreIdxes) {
|
|
|
|
if (!RestoreMBBs[Id])
|
|
|
|
return;
|
|
|
|
std::vector<SRInfo> &Restores = RestoreIdxes[Id];
|
|
|
|
for (unsigned i = 0, e = Restores.size(); i != e; ++i)
|
|
|
|
if (Restores[i].index == index && Restores[i].vreg)
|
|
|
|
Restores[i].index = -1;
|
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
|
|
|
|
|
2007-11-12 14:35:08 +08:00
|
|
|
std::vector<LiveInterval*> LiveIntervals::
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
addIntervalsForSpills(const LiveInterval &li,
|
2007-12-11 10:09:15 +08:00
|
|
|
const MachineLoopInfo *loopInfo, VirtRegMap &vrm) {
|
2007-11-12 14:35:08 +08:00
|
|
|
// Since this is called after the analysis is done we don't know if
|
|
|
|
// LiveVariables is available
|
|
|
|
lv_ = getAnalysisToUpdate<LiveVariables>();
|
|
|
|
|
|
|
|
assert(li.weight != HUGE_VALF &&
|
|
|
|
"attempt to spill already spilled interval!");
|
|
|
|
|
|
|
|
DOUT << "\t\t\t\tadding intervals for spills for interval: ";
|
2008-02-11 02:45:23 +08:00
|
|
|
li.print(DOUT, tri_);
|
2007-11-12 14:35:08 +08:00
|
|
|
DOUT << '\n';
|
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
// Each bit specify whether it a spill is required in the MBB.
|
|
|
|
BitVector SpillMBBs(mf_->getNumBlockIDs());
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> > SpillIdxes;
|
2007-11-29 09:06:25 +08:00
|
|
|
BitVector RestoreMBBs(mf_->getNumBlockIDs());
|
2007-11-29 18:12:14 +08:00
|
|
|
std::map<unsigned, std::vector<SRInfo> > RestoreIdxes;
|
|
|
|
std::map<unsigned,unsigned> MBBVRegsMap;
|
2007-11-12 14:35:08 +08:00
|
|
|
std::vector<LiveInterval*> NewLIs;
|
2007-12-31 12:13:23 +08:00
|
|
|
MachineRegisterInfo &RegInfo = mf_->getRegInfo();
|
|
|
|
const TargetRegisterClass* rc = RegInfo.getRegClass(li.reg);
|
2007-11-12 14:35:08 +08:00
|
|
|
|
|
|
|
unsigned NumValNums = li.getNumValNums();
|
|
|
|
SmallVector<MachineInstr*, 4> ReMatDefs;
|
|
|
|
ReMatDefs.resize(NumValNums, NULL);
|
|
|
|
SmallVector<MachineInstr*, 4> ReMatOrigDefs;
|
|
|
|
ReMatOrigDefs.resize(NumValNums, NULL);
|
|
|
|
SmallVector<int, 4> ReMatIds;
|
|
|
|
ReMatIds.resize(NumValNums, VirtRegMap::MAX_STACK_SLOT);
|
|
|
|
BitVector ReMatDelete(NumValNums);
|
|
|
|
unsigned Slot = VirtRegMap::MAX_STACK_SLOT;
|
|
|
|
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
// Spilling a split live interval. It cannot be split any further. Also,
|
|
|
|
// it's also guaranteed to be a single val# / range interval.
|
|
|
|
if (vrm.getPreSplitReg(li.reg)) {
|
|
|
|
vrm.setIsSplitFromReg(li.reg, 0);
|
2007-12-05 18:24:35 +08:00
|
|
|
// Unset the split kill marker on the last use.
|
|
|
|
unsigned KillIdx = vrm.getKillPoint(li.reg);
|
|
|
|
if (KillIdx) {
|
|
|
|
MachineInstr *KillMI = getInstructionFromIndex(KillIdx);
|
|
|
|
assert(KillMI && "Last use disappeared?");
|
|
|
|
int KillOp = KillMI->findRegisterUseOperandIdx(li.reg, true);
|
|
|
|
assert(KillOp != -1 && "Last use disappeared?");
|
2007-12-31 05:56:09 +08:00
|
|
|
KillMI->getOperand(KillOp).setIsKill(false);
|
2007-12-05 18:24:35 +08:00
|
|
|
}
|
2007-12-05 17:51:10 +08:00
|
|
|
vrm.removeKillPoint(li.reg);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool DefIsReMat = vrm.isReMaterialized(li.reg);
|
|
|
|
Slot = vrm.getStackSlot(li.reg);
|
|
|
|
assert(Slot != VirtRegMap::MAX_STACK_SLOT);
|
|
|
|
MachineInstr *ReMatDefMI = DefIsReMat ?
|
|
|
|
vrm.getReMaterializedMI(li.reg) : NULL;
|
|
|
|
int LdSlot = 0;
|
|
|
|
bool isLoadSS = DefIsReMat && tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot);
|
|
|
|
bool isLoad = isLoadSS ||
|
2008-01-07 15:27:27 +08:00
|
|
|
(DefIsReMat && (ReMatDefMI->getDesc().isSimpleLoad()));
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool IsFirstRange = true;
|
|
|
|
for (LiveInterval::Ranges::const_iterator
|
|
|
|
I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) {
|
|
|
|
// If this is a split live interval with multiple ranges, it means there
|
|
|
|
// are two-address instructions that re-defined the value. Only the
|
|
|
|
// first def can be rematerialized!
|
|
|
|
if (IsFirstRange) {
|
2007-11-30 07:02:50 +08:00
|
|
|
// Note ReMatOrigDefMI has already been deleted.
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
rewriteInstructionsForSpills(li, false, I, NULL, ReMatDefMI,
|
|
|
|
Slot, LdSlot, isLoad, isLoadSS, DefIsReMat,
|
2007-12-31 12:13:23 +08:00
|
|
|
false, vrm, RegInfo, rc, ReMatIds, loopInfo,
|
2007-11-29 09:06:25 +08:00
|
|
|
SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes,
|
2007-11-29 18:12:14 +08:00
|
|
|
MBBVRegsMap, NewLIs);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
} else {
|
|
|
|
rewriteInstructionsForSpills(li, false, I, NULL, 0,
|
|
|
|
Slot, 0, false, false, false,
|
2007-12-31 12:13:23 +08:00
|
|
|
false, vrm, RegInfo, rc, ReMatIds, loopInfo,
|
2007-11-29 09:06:25 +08:00
|
|
|
SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes,
|
2007-11-29 18:12:14 +08:00
|
|
|
MBBVRegsMap, NewLIs);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
|
|
|
IsFirstRange = false;
|
|
|
|
}
|
|
|
|
return NewLIs;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool TrySplit = SplitAtBB && !intervalIsInOneMBB(li);
|
2007-11-29 09:06:25 +08:00
|
|
|
if (SplitLimit != -1 && (int)numSplits >= SplitLimit)
|
|
|
|
TrySplit = false;
|
|
|
|
if (TrySplit)
|
|
|
|
++numSplits;
|
2007-11-12 14:35:08 +08:00
|
|
|
bool NeedStackSlot = false;
|
|
|
|
for (LiveInterval::const_vni_iterator i = li.vni_begin(), e = li.vni_end();
|
|
|
|
i != e; ++i) {
|
|
|
|
const VNInfo *VNI = *i;
|
|
|
|
unsigned VN = VNI->id;
|
|
|
|
unsigned DefIdx = VNI->def;
|
|
|
|
if (DefIdx == ~1U)
|
|
|
|
continue; // Dead val#.
|
|
|
|
// Is the def for the val# rematerializable?
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
MachineInstr *ReMatDefMI = (DefIdx == ~0u)
|
|
|
|
? 0 : getInstructionFromIndex(DefIdx);
|
2007-12-06 08:01:56 +08:00
|
|
|
bool dummy;
|
|
|
|
if (ReMatDefMI && isReMaterializable(li, VNI, ReMatDefMI, dummy)) {
|
2007-11-12 14:35:08 +08:00
|
|
|
// Remember how to remat the def of this val#.
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
ReMatOrigDefs[VN] = ReMatDefMI;
|
2007-11-12 14:35:08 +08:00
|
|
|
// Original def may be modified so we have to make a copy here. vrm must
|
|
|
|
// delete these!
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
ReMatDefs[VN] = ReMatDefMI = ReMatDefMI->clone();
|
2007-11-12 14:35:08 +08:00
|
|
|
|
|
|
|
bool CanDelete = true;
|
2007-11-29 17:49:23 +08:00
|
|
|
if (VNI->hasPHIKill) {
|
|
|
|
// A kill is a phi node, not all of its uses can be rematerialized.
|
2007-11-12 14:35:08 +08:00
|
|
|
// It must not be deleted.
|
2007-11-29 17:49:23 +08:00
|
|
|
CanDelete = false;
|
|
|
|
// Need a stack slot if there is any live range where uses cannot be
|
|
|
|
// rematerialized.
|
|
|
|
NeedStackSlot = true;
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
if (CanDelete)
|
|
|
|
ReMatDelete.set(VN);
|
|
|
|
} else {
|
|
|
|
// Need a stack slot if there is any live range where uses cannot be
|
|
|
|
// rematerialized.
|
|
|
|
NeedStackSlot = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// One stack slot per live interval.
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
if (NeedStackSlot && vrm.getPreSplitReg(li.reg) == 0)
|
2007-11-12 14:35:08 +08:00
|
|
|
Slot = vrm.assignVirt2StackSlot(li.reg);
|
|
|
|
|
|
|
|
// Create new intervals and rewrite defs and uses.
|
|
|
|
for (LiveInterval::Ranges::const_iterator
|
|
|
|
I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) {
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
MachineInstr *ReMatDefMI = ReMatDefs[I->valno->id];
|
|
|
|
MachineInstr *ReMatOrigDefMI = ReMatOrigDefs[I->valno->id];
|
|
|
|
bool DefIsReMat = ReMatDefMI != NULL;
|
2007-11-12 14:35:08 +08:00
|
|
|
bool CanDelete = ReMatDelete[I->valno->id];
|
|
|
|
int LdSlot = 0;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
bool isLoadSS = DefIsReMat && tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot);
|
2007-11-12 14:35:08 +08:00
|
|
|
bool isLoad = isLoadSS ||
|
2008-01-07 15:27:27 +08:00
|
|
|
(DefIsReMat && ReMatDefMI->getDesc().isSimpleLoad());
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
rewriteInstructionsForSpills(li, TrySplit, I, ReMatOrigDefMI, ReMatDefMI,
|
2007-11-29 09:06:25 +08:00
|
|
|
Slot, LdSlot, isLoad, isLoadSS, DefIsReMat,
|
2007-12-31 12:13:23 +08:00
|
|
|
CanDelete, vrm, RegInfo, rc, ReMatIds, loopInfo,
|
2007-11-29 09:06:25 +08:00
|
|
|
SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes,
|
2007-11-29 18:12:14 +08:00
|
|
|
MBBVRegsMap, NewLIs);
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|
|
|
|
|
2007-11-29 09:06:25 +08:00
|
|
|
// Insert spills / restores if we are splitting.
|
2007-11-29 18:12:14 +08:00
|
|
|
if (!TrySplit)
|
|
|
|
return NewLIs;
|
|
|
|
|
2007-12-05 16:16:32 +08:00
|
|
|
SmallPtrSet<LiveInterval*, 4> AddedKill;
|
2007-12-02 16:30:39 +08:00
|
|
|
SmallVector<unsigned, 2> Ops;
|
2007-11-29 18:12:14 +08:00
|
|
|
if (NeedStackSlot) {
|
|
|
|
int Id = SpillMBBs.find_first();
|
|
|
|
while (Id != -1) {
|
|
|
|
std::vector<SRInfo> &spills = SpillIdxes[Id];
|
|
|
|
for (unsigned i = 0, e = spills.size(); i != e; ++i) {
|
|
|
|
int index = spills[i].index;
|
|
|
|
unsigned VReg = spills[i].vreg;
|
2007-12-04 08:32:23 +08:00
|
|
|
LiveInterval &nI = getOrCreateInterval(VReg);
|
2007-11-29 09:06:25 +08:00
|
|
|
bool isReMat = vrm.isReMaterialized(VReg);
|
|
|
|
MachineInstr *MI = getInstructionFromIndex(index);
|
2007-12-02 16:30:39 +08:00
|
|
|
bool CanFold = false;
|
|
|
|
bool FoundUse = false;
|
|
|
|
Ops.clear();
|
2007-12-01 05:23:43 +08:00
|
|
|
if (spills[i].canFold) {
|
2007-12-02 16:30:39 +08:00
|
|
|
CanFold = true;
|
2007-11-29 09:06:25 +08:00
|
|
|
for (unsigned j = 0, ee = MI->getNumOperands(); j != ee; ++j) {
|
|
|
|
MachineOperand &MO = MI->getOperand(j);
|
|
|
|
if (!MO.isRegister() || MO.getReg() != VReg)
|
|
|
|
continue;
|
2007-12-02 16:30:39 +08:00
|
|
|
|
|
|
|
Ops.push_back(j);
|
|
|
|
if (MO.isDef())
|
2007-12-01 05:23:43 +08:00
|
|
|
continue;
|
2007-12-02 16:30:39 +08:00
|
|
|
if (isReMat ||
|
|
|
|
(!FoundUse && !alsoFoldARestore(Id, index, VReg,
|
|
|
|
RestoreMBBs, RestoreIdxes))) {
|
|
|
|
// MI has two-address uses of the same register. If the use
|
|
|
|
// isn't the first and only use in the BB, then we can't fold
|
|
|
|
// it. FIXME: Move this to rewriteInstructionsForSpills.
|
|
|
|
CanFold = false;
|
2007-12-01 05:23:43 +08:00
|
|
|
break;
|
|
|
|
}
|
2007-12-02 16:30:39 +08:00
|
|
|
FoundUse = true;
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
// Fold the store into the def if possible.
|
2007-12-01 05:23:43 +08:00
|
|
|
bool Folded = false;
|
2007-12-02 16:30:39 +08:00
|
|
|
if (CanFold && !Ops.empty()) {
|
|
|
|
if (tryFoldMemoryOperand(MI, vrm, NULL, index, Ops, true, Slot,VReg)){
|
2007-12-01 05:23:43 +08:00
|
|
|
Folded = true;
|
2007-12-05 17:05:34 +08:00
|
|
|
if (FoundUse > 0) {
|
2007-12-02 16:30:39 +08:00
|
|
|
// Also folded uses, do not issue a load.
|
|
|
|
eraseRestoreInfo(Id, index, VReg, RestoreMBBs, RestoreIdxes);
|
2007-12-05 17:05:34 +08:00
|
|
|
nI.removeRange(getLoadIndex(index), getUseIndex(index)+1);
|
|
|
|
}
|
2007-12-04 08:32:23 +08:00
|
|
|
nI.removeRange(getDefIndex(index), getStoreIndex(index));
|
2007-12-01 05:23:43 +08:00
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
|
|
|
|
2007-12-02 16:30:39 +08:00
|
|
|
// Else tell the spiller to issue a spill.
|
2007-12-05 16:16:32 +08:00
|
|
|
if (!Folded) {
|
|
|
|
LiveRange *LR = &nI.ranges[nI.ranges.size()-1];
|
|
|
|
bool isKill = LR->end == getStoreIndex(index);
|
|
|
|
vrm.addSpillPoint(VReg, isKill, MI);
|
|
|
|
if (isKill)
|
|
|
|
AddedKill.insert(&nI);
|
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
2007-11-29 18:12:14 +08:00
|
|
|
Id = SpillMBBs.find_next(Id);
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
2007-11-29 18:12:14 +08:00
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
|
2007-11-29 18:12:14 +08:00
|
|
|
int Id = RestoreMBBs.find_first();
|
|
|
|
while (Id != -1) {
|
|
|
|
std::vector<SRInfo> &restores = RestoreIdxes[Id];
|
|
|
|
for (unsigned i = 0, e = restores.size(); i != e; ++i) {
|
|
|
|
int index = restores[i].index;
|
|
|
|
if (index == -1)
|
|
|
|
continue;
|
|
|
|
unsigned VReg = restores[i].vreg;
|
2007-12-04 08:32:23 +08:00
|
|
|
LiveInterval &nI = getOrCreateInterval(VReg);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
MachineInstr *MI = getInstructionFromIndex(index);
|
2007-12-02 16:30:39 +08:00
|
|
|
bool CanFold = false;
|
|
|
|
Ops.clear();
|
2007-12-01 05:23:43 +08:00
|
|
|
if (restores[i].canFold) {
|
2007-12-02 16:30:39 +08:00
|
|
|
CanFold = true;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
for (unsigned j = 0, ee = MI->getNumOperands(); j != ee; ++j) {
|
|
|
|
MachineOperand &MO = MI->getOperand(j);
|
|
|
|
if (!MO.isRegister() || MO.getReg() != VReg)
|
|
|
|
continue;
|
2007-12-02 16:30:39 +08:00
|
|
|
|
2007-11-29 09:06:25 +08:00
|
|
|
if (MO.isDef()) {
|
2007-12-02 16:30:39 +08:00
|
|
|
// If this restore were to be folded, it would have been folded
|
|
|
|
// already.
|
|
|
|
CanFold = false;
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
break;
|
|
|
|
}
|
2007-12-02 16:30:39 +08:00
|
|
|
Ops.push_back(j);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
|
|
|
|
// Fold the load into the use if possible.
|
2007-12-01 05:23:43 +08:00
|
|
|
bool Folded = false;
|
2007-12-02 16:30:39 +08:00
|
|
|
if (CanFold && !Ops.empty()) {
|
|
|
|
if (!vrm.isReMaterialized(VReg))
|
|
|
|
Folded = tryFoldMemoryOperand(MI, vrm, NULL,index,Ops,true,Slot,VReg);
|
|
|
|
else {
|
2007-11-29 09:06:25 +08:00
|
|
|
MachineInstr *ReMatDefMI = vrm.getReMaterializedMI(VReg);
|
|
|
|
int LdSlot = 0;
|
|
|
|
bool isLoadSS = tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot);
|
|
|
|
// If the rematerializable def is a load, also try to fold it.
|
2008-01-07 15:27:27 +08:00
|
|
|
if (isLoadSS || ReMatDefMI->getDesc().isSimpleLoad())
|
2007-12-02 16:30:39 +08:00
|
|
|
Folded = tryFoldMemoryOperand(MI, vrm, ReMatDefMI, index,
|
|
|
|
Ops, isLoadSS, LdSlot, VReg);
|
|
|
|
}
|
2007-11-29 09:06:25 +08:00
|
|
|
}
|
|
|
|
// If folding is not possible / failed, then tell the spiller to issue a
|
|
|
|
// load / rematerialization for us.
|
2007-12-04 08:32:23 +08:00
|
|
|
if (Folded)
|
|
|
|
nI.removeRange(getLoadIndex(index), getUseIndex(index)+1);
|
2007-12-05 16:16:32 +08:00
|
|
|
else
|
2007-11-29 09:06:25 +08:00
|
|
|
vrm.addRestorePoint(VReg, MI);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
2007-11-29 18:12:14 +08:00
|
|
|
Id = RestoreMBBs.find_next(Id);
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
}
|
|
|
|
|
2007-12-05 16:16:32 +08:00
|
|
|
// Finalize intervals: add kills, finalize spill weights, and filter out
|
|
|
|
// dead intervals.
|
2007-12-04 08:32:23 +08:00
|
|
|
std::vector<LiveInterval*> RetNewLIs;
|
|
|
|
for (unsigned i = 0, e = NewLIs.size(); i != e; ++i) {
|
|
|
|
LiveInterval *LI = NewLIs[i];
|
|
|
|
if (!LI->empty()) {
|
|
|
|
LI->weight /= LI->getSize();
|
2007-12-05 16:16:32 +08:00
|
|
|
if (!AddedKill.count(LI)) {
|
|
|
|
LiveRange *LR = &LI->ranges[LI->ranges.size()-1];
|
2007-12-05 18:24:35 +08:00
|
|
|
unsigned LastUseIdx = getBaseIndex(LR->end);
|
|
|
|
MachineInstr *LastUse = getInstructionFromIndex(LastUseIdx);
|
2007-12-05 16:16:32 +08:00
|
|
|
int UseIdx = LastUse->findRegisterUseOperandIdx(LI->reg);
|
|
|
|
assert(UseIdx != -1);
|
2008-01-07 15:27:27 +08:00
|
|
|
if (LastUse->getDesc().getOperandConstraint(UseIdx, TOI::TIED_TO) ==
|
2008-01-07 09:56:04 +08:00
|
|
|
-1) {
|
2007-12-05 16:16:32 +08:00
|
|
|
LastUse->getOperand(UseIdx).setIsKill();
|
2007-12-05 18:24:35 +08:00
|
|
|
vrm.addKillPoint(LI->reg, LastUseIdx);
|
2007-12-05 17:51:10 +08:00
|
|
|
}
|
2007-12-05 16:16:32 +08:00
|
|
|
}
|
2007-12-04 08:32:23 +08:00
|
|
|
RetNewLIs.push_back(LI);
|
|
|
|
}
|
|
|
|
}
|
Live interval splitting:
When a live interval is being spilled, rather than creating short, non-spillable
intervals for every def / use, split the interval at BB boundaries. That is, for
every BB where the live interval is defined or used, create a new interval that
covers all the defs and uses in the BB.
This is designed to eliminate one common problem: multiple reloads of the same
value in a single basic block. Note, it does *not* decrease the number of spills
since no copies are inserted so the split intervals are *connected* through
spill and reloads (or rematerialization). The newly created intervals can be
spilled again, in that case, since it does not span multiple basic blocks, it's
spilled in the usual manner. However, it can reuse the same stack slot as the
previously split interval.
This is currently controlled by -split-intervals-at-bb.
llvm-svn: 44198
2007-11-17 08:40:40 +08:00
|
|
|
|
2007-12-04 08:32:23 +08:00
|
|
|
return RetNewLIs;
|
2007-11-12 14:35:08 +08:00
|
|
|
}
|