forked from OSchip/llvm-project
384 lines
15 KiB
C++
384 lines
15 KiB
C++
//===- PipelineDataTransfer.cpp --- Pass for pipelining data movement ---*-===//
|
|
//
|
|
// Copyright 2019 The MLIR Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
// =============================================================================
|
|
//
|
|
// This file implements a pass to pipeline data transfers.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "mlir/Transforms/Passes.h"
|
|
|
|
#include "mlir/AffineOps/AffineOps.h"
|
|
#include "mlir/Analysis/AffineAnalysis.h"
|
|
#include "mlir/Analysis/LoopAnalysis.h"
|
|
#include "mlir/Analysis/Utils.h"
|
|
#include "mlir/IR/Builders.h"
|
|
#include "mlir/Pass/Pass.h"
|
|
#include "mlir/StandardOps/Ops.h"
|
|
#include "mlir/Transforms/LoopUtils.h"
|
|
#include "mlir/Transforms/Utils.h"
|
|
#include "llvm/ADT/DenseMap.h"
|
|
#include "llvm/Support/Debug.h"
|
|
#define DEBUG_TYPE "pipeline-data-transfer"
|
|
|
|
using namespace mlir;
|
|
|
|
namespace {
|
|
|
|
struct PipelineDataTransfer : public FunctionPass<PipelineDataTransfer> {
|
|
void runOnFunction() override;
|
|
void runOnAffineForOp(AffineForOp forOp);
|
|
|
|
std::vector<AffineForOp> forOps;
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
/// Creates a pass to pipeline explicit movement of data across levels of the
|
|
/// memory hierarchy.
|
|
FunctionPassBase *mlir::createPipelineDataTransferPass() {
|
|
return new PipelineDataTransfer();
|
|
}
|
|
|
|
// Returns the position of the tag memref operand given a DMA instruction.
|
|
// Temporary utility: will be replaced when DmaStart/DmaFinish abstract op's are
|
|
// added. TODO(b/117228571)
|
|
static unsigned getTagMemRefPos(Instruction &dmaInst) {
|
|
assert(dmaInst.isa<DmaStartOp>() || dmaInst.isa<DmaWaitOp>());
|
|
if (dmaInst.isa<DmaStartOp>()) {
|
|
// Second to last operand.
|
|
return dmaInst.getNumOperands() - 2;
|
|
}
|
|
// First operand for a dma finish instruction.
|
|
return 0;
|
|
}
|
|
|
|
/// Doubles the buffer of the supplied memref on the specified 'affine.for'
|
|
/// instruction by adding a leading dimension of size two to the memref.
|
|
/// Replaces all uses of the old memref by the new one while indexing the newly
|
|
/// added dimension by the loop IV of the specified 'affine.for' instruction
|
|
/// modulo 2. Returns false if such a replacement cannot be performed.
|
|
static bool doubleBuffer(Value *oldMemRef, AffineForOp forOp) {
|
|
auto *forBody = forOp.getBody();
|
|
FuncBuilder bInner(forBody, forBody->begin());
|
|
bInner.setInsertionPoint(forBody, forBody->begin());
|
|
|
|
// Doubles the shape with a leading dimension extent of 2.
|
|
auto doubleShape = [&](MemRefType oldMemRefType) -> MemRefType {
|
|
// Add the leading dimension in the shape for the double buffer.
|
|
ArrayRef<int64_t> oldShape = oldMemRefType.getShape();
|
|
SmallVector<int64_t, 4> newShape(1 + oldMemRefType.getRank());
|
|
newShape[0] = 2;
|
|
std::copy(oldShape.begin(), oldShape.end(), newShape.begin() + 1);
|
|
auto newMemRefType =
|
|
bInner.getMemRefType(newShape, oldMemRefType.getElementType(), {},
|
|
oldMemRefType.getMemorySpace());
|
|
return newMemRefType;
|
|
};
|
|
|
|
auto oldMemRefType = oldMemRef->getType().cast<MemRefType>();
|
|
auto newMemRefType = doubleShape(oldMemRefType);
|
|
|
|
// The double buffer is allocated right before 'forInst'.
|
|
auto *forInst = forOp.getInstruction();
|
|
FuncBuilder bOuter(forInst);
|
|
// Put together alloc operands for any dynamic dimensions of the memref.
|
|
SmallVector<Value *, 4> allocOperands;
|
|
unsigned dynamicDimCount = 0;
|
|
for (auto dimSize : oldMemRefType.getShape()) {
|
|
if (dimSize == -1)
|
|
allocOperands.push_back(bOuter.create<DimOp>(forInst->getLoc(), oldMemRef,
|
|
dynamicDimCount++));
|
|
}
|
|
|
|
// Create and place the alloc right before the 'affine.for' instruction.
|
|
Value *newMemRef =
|
|
bOuter.create<AllocOp>(forInst->getLoc(), newMemRefType, allocOperands);
|
|
|
|
// Create 'iv mod 2' value to index the leading dimension.
|
|
auto d0 = bInner.getAffineDimExpr(0);
|
|
int64_t step = forOp.getStep();
|
|
auto modTwoMap = bInner.getAffineMap(/*dimCount=*/1, /*symbolCount=*/0,
|
|
{d0.floorDiv(step) % 2}, {});
|
|
auto ivModTwoOp = bInner.create<AffineApplyOp>(forOp.getLoc(), modTwoMap,
|
|
forOp.getInductionVar());
|
|
|
|
// replaceAllMemRefUsesWith will always succeed unless the forOp body has
|
|
// non-deferencing uses of the memref (dealloc's are fine though).
|
|
if (!replaceAllMemRefUsesWith(oldMemRef, newMemRef,
|
|
/*extraIndices=*/{ivModTwoOp},
|
|
/*indexRemap=*/AffineMap(),
|
|
/*extraOperands=*/{},
|
|
/*domInstFilter=*/&*forOp.getBody()->begin())) {
|
|
LLVM_DEBUG(
|
|
forOp.emitError("memref replacement for double buffering failed"));
|
|
ivModTwoOp->getInstruction()->erase();
|
|
return false;
|
|
}
|
|
// Insert the dealloc op right after the for loop.
|
|
bOuter.setInsertionPoint(forInst->getBlock(),
|
|
std::next(Block::iterator(forInst)));
|
|
bOuter.create<DeallocOp>(forInst->getLoc(), newMemRef);
|
|
|
|
return true;
|
|
}
|
|
|
|
/// Returns success if the IR is in a valid state.
|
|
void PipelineDataTransfer::runOnFunction() {
|
|
// Do a post order walk so that inner loop DMAs are processed first. This is
|
|
// necessary since 'affine.for' instructions nested within would otherwise
|
|
// become invalid (erased) when the outer loop is pipelined (the pipelined one
|
|
// gets deleted and replaced by a prologue, a new steady-state loop and an
|
|
// epilogue).
|
|
forOps.clear();
|
|
getFunction()->walkPostOrder<AffineForOp>(
|
|
[&](AffineForOp forOp) { forOps.push_back(forOp); });
|
|
for (auto forOp : forOps)
|
|
runOnAffineForOp(forOp);
|
|
}
|
|
|
|
// Check if tags of the dma start op and dma wait op match.
|
|
static bool checkTagMatch(DmaStartOp startOp, DmaWaitOp waitOp) {
|
|
if (startOp->getTagMemRef() != waitOp->getTagMemRef())
|
|
return false;
|
|
auto startIndices = startOp->getTagIndices();
|
|
auto waitIndices = waitOp->getTagIndices();
|
|
// Both of these have the same number of indices since they correspond to the
|
|
// same tag memref.
|
|
for (auto it = startIndices.begin(), wIt = waitIndices.begin(),
|
|
e = startIndices.end();
|
|
it != e; ++it, ++wIt) {
|
|
// Keep it simple for now, just checking if indices match.
|
|
// TODO(mlir-team): this would in general need to check if there is no
|
|
// intervening write writing to the same tag location, i.e., memory last
|
|
// write/data flow analysis. This is however sufficient/powerful enough for
|
|
// now since the DMA generation pass or the input for it will always have
|
|
// start/wait with matching tags (same SSA operand indices).
|
|
if (*it != *wIt)
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// Identify matching DMA start/finish instructions to overlap computation with.
|
|
static void findMatchingStartFinishInsts(
|
|
AffineForOp forOp,
|
|
SmallVectorImpl<std::pair<Instruction *, Instruction *>> &startWaitPairs) {
|
|
|
|
// Collect outgoing DMA instructions - needed to check for dependences below.
|
|
SmallVector<DmaStartOp, 4> outgoingDmaOps;
|
|
for (auto &inst : *forOp.getBody()) {
|
|
auto dmaStartOp = inst.dyn_cast<DmaStartOp>();
|
|
if (dmaStartOp && dmaStartOp->isSrcMemorySpaceFaster())
|
|
outgoingDmaOps.push_back(dmaStartOp);
|
|
}
|
|
|
|
SmallVector<Instruction *, 4> dmaStartInsts, dmaFinishInsts;
|
|
for (auto &inst : *forOp.getBody()) {
|
|
// Collect DMA finish instructions.
|
|
if (inst.isa<DmaWaitOp>()) {
|
|
dmaFinishInsts.push_back(&inst);
|
|
continue;
|
|
}
|
|
auto dmaStartOp = inst.dyn_cast<DmaStartOp>();
|
|
if (!dmaStartOp)
|
|
continue;
|
|
|
|
// Only DMAs incoming into higher memory spaces are pipelined for now.
|
|
// TODO(bondhugula): handle outgoing DMA pipelining.
|
|
if (!dmaStartOp->isDestMemorySpaceFaster())
|
|
continue;
|
|
|
|
// Check for dependence with outgoing DMAs. Doing this conservatively.
|
|
// TODO(andydavis,bondhugula): use the dependence analysis to check for
|
|
// dependences between an incoming and outgoing DMA in the same iteration.
|
|
auto it = outgoingDmaOps.begin();
|
|
for (; it != outgoingDmaOps.end(); ++it) {
|
|
if ((*it)->getDstMemRef() == dmaStartOp->getSrcMemRef())
|
|
break;
|
|
}
|
|
if (it != outgoingDmaOps.end())
|
|
continue;
|
|
|
|
// We only double buffer if the buffer is not live out of loop.
|
|
auto *memref = dmaStartOp->getOperand(dmaStartOp->getFasterMemPos());
|
|
bool escapingUses = false;
|
|
for (const auto &use : memref->getUses()) {
|
|
// We can double buffer regardless of dealloc's outside the loop.
|
|
if (use.getOwner()->isa<DeallocOp>())
|
|
continue;
|
|
if (!forOp.getBody()->findAncestorInstInBlock(*use.getOwner())) {
|
|
LLVM_DEBUG(llvm::dbgs()
|
|
<< "can't pipeline: buffer is live out of loop\n";);
|
|
escapingUses = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!escapingUses)
|
|
dmaStartInsts.push_back(&inst);
|
|
}
|
|
|
|
// For each start instruction, we look for a matching finish instruction.
|
|
for (auto *dmaStartInst : dmaStartInsts) {
|
|
for (auto *dmaFinishInst : dmaFinishInsts) {
|
|
if (checkTagMatch(dmaStartInst->cast<DmaStartOp>(),
|
|
dmaFinishInst->cast<DmaWaitOp>())) {
|
|
startWaitPairs.push_back({dmaStartInst, dmaFinishInst});
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Overlap DMA transfers with computation in this loop. If successful,
|
|
/// 'forOp' is deleted, and a prologue, a new pipelined loop, and epilogue are
|
|
/// inserted right before where it was.
|
|
void PipelineDataTransfer::runOnAffineForOp(AffineForOp forOp) {
|
|
auto mayBeConstTripCount = getConstantTripCount(forOp);
|
|
if (!mayBeConstTripCount.hasValue()) {
|
|
LLVM_DEBUG(forOp.emitNote("won't pipeline due to unknown trip count loop"));
|
|
return;
|
|
}
|
|
|
|
SmallVector<std::pair<Instruction *, Instruction *>, 4> startWaitPairs;
|
|
findMatchingStartFinishInsts(forOp, startWaitPairs);
|
|
|
|
if (startWaitPairs.empty()) {
|
|
LLVM_DEBUG(forOp.emitNote("No dma start/finish pairs\n"));
|
|
return;
|
|
}
|
|
|
|
// Double the buffers for the higher memory space memref's.
|
|
// Identify memref's to replace by scanning through all DMA start
|
|
// instructions. A DMA start instruction has two memref's - the one from the
|
|
// higher level of memory hierarchy is the one to double buffer.
|
|
// TODO(bondhugula): check whether double-buffering is even necessary.
|
|
// TODO(bondhugula): make this work with different layouts: assuming here that
|
|
// the dimension we are adding here for the double buffering is the outermost
|
|
// dimension.
|
|
for (auto &pair : startWaitPairs) {
|
|
auto *dmaStartInst = pair.first;
|
|
Value *oldMemRef = dmaStartInst->getOperand(
|
|
dmaStartInst->cast<DmaStartOp>()->getFasterMemPos());
|
|
if (!doubleBuffer(oldMemRef, forOp)) {
|
|
// Normally, double buffering should not fail because we already checked
|
|
// that there are no uses outside.
|
|
LLVM_DEBUG(llvm::dbgs() << "double buffering failed for: \n";);
|
|
LLVM_DEBUG(dmaStartInst->dump());
|
|
// IR still in a valid state.
|
|
return;
|
|
}
|
|
// If the old memref has no more uses, remove its 'dead' alloc if it was
|
|
// alloc'ed. (note: DMA buffers are rarely function live-in; but a 'dim'
|
|
// operation could have been used on it if it was dynamically shaped in
|
|
// order to create the double buffer above.)
|
|
// '-canonicalize' does this in a more general way, but we'll anyway do the
|
|
// simple/common case so that the output / test cases looks clear.
|
|
if (auto *allocInst = oldMemRef->getDefiningInst()) {
|
|
if (oldMemRef->use_empty()) {
|
|
allocInst->erase();
|
|
} else if (oldMemRef->hasOneUse()) {
|
|
auto *singleUse = oldMemRef->use_begin()->getOwner();
|
|
if (singleUse->isa<DeallocOp>()) {
|
|
singleUse->erase();
|
|
oldMemRef->getDefiningInst()->erase();
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Double the buffers for tag memrefs.
|
|
for (auto &pair : startWaitPairs) {
|
|
auto *dmaFinishInst = pair.second;
|
|
Value *oldTagMemRef =
|
|
dmaFinishInst->getOperand(getTagMemRefPos(*dmaFinishInst));
|
|
if (!doubleBuffer(oldTagMemRef, forOp)) {
|
|
LLVM_DEBUG(llvm::dbgs() << "tag double buffering failed\n";);
|
|
return;
|
|
}
|
|
// If the old tag has no more uses, remove its 'dead' alloc if it was
|
|
// alloc'ed.
|
|
if (oldTagMemRef->use_empty())
|
|
if (auto *allocInst = oldTagMemRef->getDefiningInst())
|
|
allocInst->erase();
|
|
}
|
|
|
|
// Double buffering would have invalidated all the old DMA start/wait insts.
|
|
startWaitPairs.clear();
|
|
findMatchingStartFinishInsts(forOp, startWaitPairs);
|
|
|
|
// Store shift for instruction for later lookup for AffineApplyOp's.
|
|
DenseMap<Instruction *, unsigned> instShiftMap;
|
|
for (auto &pair : startWaitPairs) {
|
|
auto *dmaStartInst = pair.first;
|
|
assert(dmaStartInst->isa<DmaStartOp>());
|
|
instShiftMap[dmaStartInst] = 0;
|
|
// Set shifts for DMA start inst's affine operand computation slices to 0.
|
|
SmallVector<AffineApplyOp, 4> sliceOps;
|
|
mlir::createAffineComputationSlice(dmaStartInst, &sliceOps);
|
|
if (!sliceOps.empty()) {
|
|
for (auto sliceOp : sliceOps) {
|
|
instShiftMap[sliceOp.getInstruction()] = 0;
|
|
}
|
|
} else {
|
|
// If a slice wasn't created, the reachable affine.apply op's from its
|
|
// operands are the ones that go with it.
|
|
SmallVector<Instruction *, 4> affineApplyInsts;
|
|
SmallVector<Value *, 4> operands(dmaStartInst->getOperands());
|
|
getReachableAffineApplyOps(operands, affineApplyInsts);
|
|
for (auto *inst : affineApplyInsts) {
|
|
instShiftMap[inst] = 0;
|
|
}
|
|
}
|
|
}
|
|
// Everything else (including compute ops and dma finish) are shifted by one.
|
|
for (auto &inst : *forOp.getBody()) {
|
|
if (instShiftMap.find(&inst) == instShiftMap.end()) {
|
|
instShiftMap[&inst] = 1;
|
|
}
|
|
}
|
|
|
|
// Get shifts stored in map.
|
|
std::vector<uint64_t> shifts(forOp.getBody()->getInstructions().size());
|
|
unsigned s = 0;
|
|
for (auto &inst : *forOp.getBody()) {
|
|
assert(instShiftMap.find(&inst) != instShiftMap.end());
|
|
shifts[s++] = instShiftMap[&inst];
|
|
|
|
// Tagging instructions with shifts for debugging purposes.
|
|
LLVM_DEBUG({
|
|
FuncBuilder b(&inst);
|
|
inst.setAttr("shift", b.getI64IntegerAttr(shifts[s - 1]));
|
|
});
|
|
}
|
|
|
|
if (!isInstwiseShiftValid(forOp, shifts)) {
|
|
// Violates dependences.
|
|
LLVM_DEBUG(llvm::dbgs() << "Shifts invalid - unexpected\n";);
|
|
return;
|
|
}
|
|
|
|
if (failed(instBodySkew(forOp, shifts))) {
|
|
LLVM_DEBUG(llvm::dbgs() << "inst body skewing failed - unexpected\n";);
|
|
return;
|
|
}
|
|
}
|
|
|
|
static PassRegistration<PipelineDataTransfer> pass(
|
|
"pipeline-data-transfer",
|
|
"Pipeline non-blocking data transfers between explicitly managed levels of "
|
|
"the memory hierarchy");
|