forked from OSchip/llvm-project
[mlir] NFC: fix trivial typos
fix typos in comments and documents Reviewed By: jpienaar Differential Revision: https://reviews.llvm.org/D90089
This commit is contained in:
parent
220293da53
commit
41b09f4eff
|
@ -797,7 +797,7 @@ A custom directive has two main parts: The `UserDirective` and the `Params`. A
|
||||||
custom directive is transformed into a call to a `print*` and a `parse*` method
|
custom directive is transformed into a call to a `print*` and a `parse*` method
|
||||||
when generating the C++ code for the format. The `UserDirective` is an
|
when generating the C++ code for the format. The `UserDirective` is an
|
||||||
identifier used as a suffix to these two calls, i.e., `custom<MyDirective>(...)`
|
identifier used as a suffix to these two calls, i.e., `custom<MyDirective>(...)`
|
||||||
would result in calls to `parseMyDirective` and `printMyDirective` wihtin the
|
would result in calls to `parseMyDirective` and `printMyDirective` within the
|
||||||
parser and printer respectively. `Params` may be any combination of variables
|
parser and printer respectively. `Params` may be any combination of variables
|
||||||
(i.e. Attribute, Operand, Successor, etc.), type directives, and `attr-dict`.
|
(i.e. Attribute, Operand, Successor, etc.), type directives, and `attr-dict`.
|
||||||
The type directives must refer to a variable, but that variable need not also
|
The type directives must refer to a variable, but that variable need not also
|
||||||
|
@ -1515,7 +1515,7 @@ def IntegerType : Test_Type<"TestInteger"> {
|
||||||
|
|
||||||
The name of the C++ class which gets generated defaults to
|
The name of the C++ class which gets generated defaults to
|
||||||
`<classParamName>Type` (e.g. `TestIntegerType` in the above example). This
|
`<classParamName>Type` (e.g. `TestIntegerType` in the above example). This
|
||||||
can be overridden via the the `cppClassName` field. The field `mnemonic` is
|
can be overridden via the `cppClassName` field. The field `mnemonic` is
|
||||||
to specify the asm name for parsing. It is optional and not specifying it
|
to specify the asm name for parsing. It is optional and not specifying it
|
||||||
will imply that no parser or printer methods are attached to this class.
|
will imply that no parser or printer methods are attached to this class.
|
||||||
|
|
||||||
|
@ -1554,7 +1554,7 @@ The default storage constructor blindly copies fields by value. It does not
|
||||||
know anything about the types. In this case, the ArrayRef<int> requires
|
know anything about the types. In this case, the ArrayRef<int> requires
|
||||||
allocation with `dims = allocator.copyInto(dims)`.
|
allocation with `dims = allocator.copyInto(dims)`.
|
||||||
|
|
||||||
You can specify the necessary constuctor by specializing the `TypeParameter`
|
You can specify the necessary constructor by specializing the `TypeParameter`
|
||||||
tblgen class:
|
tblgen class:
|
||||||
|
|
||||||
```tablegen
|
```tablegen
|
||||||
|
|
|
@ -21,7 +21,7 @@
|
||||||
extern "C" {
|
extern "C" {
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/** An opaque reference to a dignostic, always owned by the diagnostics engine
|
/** An opaque reference to a diagnostic, always owned by the diagnostics engine
|
||||||
* (context). Must not be stored outside of the diagnostic handler. */
|
* (context). Must not be stored outside of the diagnostic handler. */
|
||||||
struct MlirDiagnostic {
|
struct MlirDiagnostic {
|
||||||
void *ptr;
|
void *ptr;
|
||||||
|
|
|
@ -203,7 +203,7 @@ MlirAttribute mlirSymbolRefAttrGet(MlirContext ctx, intptr_t length,
|
||||||
* live as long as the context in which the attribute lives. */
|
* live as long as the context in which the attribute lives. */
|
||||||
MlirStringRef mlirSymbolRefAttrGetRootReference(MlirAttribute attr);
|
MlirStringRef mlirSymbolRefAttrGetRootReference(MlirAttribute attr);
|
||||||
|
|
||||||
/** Returns the stirng reference to the leaf referenced symbol. The data remains
|
/** Returns the string reference to the leaf referenced symbol. The data remains
|
||||||
* live as long as the context in which the attribute lives. */
|
* live as long as the context in which the attribute lives. */
|
||||||
MlirStringRef mlirSymbolRefAttrGetLeafReference(MlirAttribute attr);
|
MlirStringRef mlirSymbolRefAttrGetLeafReference(MlirAttribute attr);
|
||||||
|
|
||||||
|
|
|
@ -43,7 +43,7 @@ def Async_ExecuteOp :
|
||||||
(`async.token` or `async.value`).
|
(`async.token` or `async.value`).
|
||||||
|
|
||||||
`async.execute` operation takes `async.token` dependencies and `async.value`
|
`async.execute` operation takes `async.token` dependencies and `async.value`
|
||||||
operands separatly, and starts execution of the attached body region only
|
operands separately, and starts execution of the attached body region only
|
||||||
when all tokens and values become ready.
|
when all tokens and values become ready.
|
||||||
|
|
||||||
Example:
|
Example:
|
||||||
|
|
|
@ -913,7 +913,7 @@ class SignlessIntegerAttrBase<I attrValType, string descr> :
|
||||||
let returnType = [{ ::llvm::APInt }];
|
let returnType = [{ ::llvm::APInt }];
|
||||||
}
|
}
|
||||||
// Base class for signless integer attributes of fixed width that have a
|
// Base class for signless integer attributes of fixed width that have a
|
||||||
// correpsonding C++ type.
|
// corresponding C++ type.
|
||||||
class TypedSignlessIntegerAttrBase<I attrValType, string retType, string descr>
|
class TypedSignlessIntegerAttrBase<I attrValType, string retType, string descr>
|
||||||
: SignlessIntegerAttrBase<attrValType, descr> {
|
: SignlessIntegerAttrBase<attrValType, descr> {
|
||||||
let returnType = retType;
|
let returnType = retType;
|
||||||
|
@ -942,7 +942,7 @@ class SignedIntegerAttrBase<SI attrValType, string descr> :
|
||||||
let returnType = [{ ::llvm::APInt }];
|
let returnType = [{ ::llvm::APInt }];
|
||||||
}
|
}
|
||||||
// Base class for signed integer attributes of fixed width that have a
|
// Base class for signed integer attributes of fixed width that have a
|
||||||
// correpsonding C++ type.
|
// corresponding C++ type.
|
||||||
class TypedSignedIntegerAttrBase<SI attrValType, string retType, string descr>
|
class TypedSignedIntegerAttrBase<SI attrValType, string retType, string descr>
|
||||||
: SignedIntegerAttrBase<attrValType, descr> {
|
: SignedIntegerAttrBase<attrValType, descr> {
|
||||||
let returnType = retType;
|
let returnType = retType;
|
||||||
|
@ -971,7 +971,7 @@ class UnsignedIntegerAttrBase<UI attrValType, string descr> :
|
||||||
let returnType = [{ ::llvm::APInt }];
|
let returnType = [{ ::llvm::APInt }];
|
||||||
}
|
}
|
||||||
// Base class for unsigned integer attributes of fixed width that have a
|
// Base class for unsigned integer attributes of fixed width that have a
|
||||||
// correpsonding C++ type.
|
// corresponding C++ type.
|
||||||
class TypedUnsignedIntegerAttrBase<UI attrValType, string retType, string descr>
|
class TypedUnsignedIntegerAttrBase<UI attrValType, string retType, string descr>
|
||||||
: UnsignedIntegerAttrBase<attrValType, descr> {
|
: UnsignedIntegerAttrBase<attrValType, descr> {
|
||||||
let returnType = retType;
|
let returnType = retType;
|
||||||
|
|
|
@ -8,7 +8,7 @@
|
||||||
//
|
//
|
||||||
// This file defines the OpReducer class. It defines a variant generator method
|
// This file defines the OpReducer class. It defines a variant generator method
|
||||||
// with the purpose of producing different variants by eliminating a
|
// with the purpose of producing different variants by eliminating a
|
||||||
// parametarizable type of operations from the parent module.
|
// parameterizable type of operations from the parent module.
|
||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
|
||||||
|
|
|
@ -51,7 +51,7 @@ public:
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Re-indents by removing the leading whitespace from the first non-empty
|
/// Re-indents by removing the leading whitespace from the first non-empty
|
||||||
/// line from every line of the the string, skipping over empty lines at the
|
/// line from every line of the string, skipping over empty lines at the
|
||||||
/// start.
|
/// start.
|
||||||
raw_indented_ostream &reindent(StringRef str);
|
raw_indented_ostream &reindent(StringRef str);
|
||||||
|
|
||||||
|
|
|
@ -4,7 +4,7 @@
|
||||||
// RUN: FileCheck %s
|
// RUN: FileCheck %s
|
||||||
|
|
||||||
// NOTE: This is similar to test-create-mask.mlir, but with a different length,
|
// NOTE: This is similar to test-create-mask.mlir, but with a different length,
|
||||||
// because the v4i1 vector specifially exposed bugs in the LLVM backend.
|
// because the v4i1 vector specifically exposed bugs in the LLVM backend.
|
||||||
|
|
||||||
func @entry() {
|
func @entry() {
|
||||||
%c0 = constant 0 : index
|
%c0 = constant 0 : index
|
||||||
|
|
|
@ -351,7 +351,7 @@ void Simplex::markEmpty() {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Add an inequality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
|
/// Add an inequality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
|
||||||
/// is the curent number of variables, then the corresponding inequality is
|
/// is the current number of variables, then the corresponding inequality is
|
||||||
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} >= 0.
|
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} >= 0.
|
||||||
///
|
///
|
||||||
/// We add the inequality and mark it as restricted. We then try to make its
|
/// We add the inequality and mark it as restricted. We then try to make its
|
||||||
|
@ -367,7 +367,7 @@ void Simplex::addInequality(ArrayRef<int64_t> coeffs) {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Add an equality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
|
/// Add an equality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
|
||||||
/// is the curent number of variables, then the corresponding equality is
|
/// is the current number of variables, then the corresponding equality is
|
||||||
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} == 0.
|
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} == 0.
|
||||||
///
|
///
|
||||||
/// We simply add two opposing inequalities, which force the expression to
|
/// We simply add two opposing inequalities, which force the expression to
|
||||||
|
@ -383,7 +383,7 @@ void Simplex::addEquality(ArrayRef<int64_t> coeffs) {
|
||||||
unsigned Simplex::numVariables() const { return var.size(); }
|
unsigned Simplex::numVariables() const { return var.size(); }
|
||||||
unsigned Simplex::numConstraints() const { return con.size(); }
|
unsigned Simplex::numConstraints() const { return con.size(); }
|
||||||
|
|
||||||
/// Return a snapshot of the curent state. This is just the current size of the
|
/// Return a snapshot of the current state. This is just the current size of the
|
||||||
/// undo log.
|
/// undo log.
|
||||||
unsigned Simplex::getSnapshot() const { return undoLog.size(); }
|
unsigned Simplex::getSnapshot() const { return undoLog.size(); }
|
||||||
|
|
||||||
|
|
|
@ -326,7 +326,7 @@ private:
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Blocks are exposed by the C-API as a forward-only linked list. In Python,
|
/// Blocks are exposed by the C-API as a forward-only linked list. In Python,
|
||||||
/// we present them as a more full-featured list-like container but optimzie
|
/// we present them as a more full-featured list-like container but optimize
|
||||||
/// it for forward iteration. Blocks are always owned by a region.
|
/// it for forward iteration. Blocks are always owned by a region.
|
||||||
class PyBlockList {
|
class PyBlockList {
|
||||||
public:
|
public:
|
||||||
|
@ -424,7 +424,7 @@ private:
|
||||||
|
|
||||||
/// Operations are exposed by the C-API as a forward-only linked list. In
|
/// Operations are exposed by the C-API as a forward-only linked list. In
|
||||||
/// Python, we present them as a more full-featured list-like container but
|
/// Python, we present them as a more full-featured list-like container but
|
||||||
/// optimzie it for forward iteration. Iterable operations are always owned
|
/// optimize it for forward iteration. Iterable operations are always owned
|
||||||
/// by a block.
|
/// by a block.
|
||||||
class PyOperationList {
|
class PyOperationList {
|
||||||
public:
|
public:
|
||||||
|
|
|
@ -34,7 +34,7 @@ public:
|
||||||
ConversionPatternRewriter &rewriter) const override;
|
ConversionPatternRewriter &rewriter) const override;
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Pattern lowering subgoup size/id to loading SPIR-V invocation
|
/// Pattern lowering subgroup size/id to loading SPIR-V invocation
|
||||||
/// builtin variables.
|
/// builtin variables.
|
||||||
template <typename SourceOp, spirv::BuiltIn builtin>
|
template <typename SourceOp, spirv::BuiltIn builtin>
|
||||||
class SingleDimLaunchConfigConversion : public SPIRVOpLowering<SourceOp> {
|
class SingleDimLaunchConfigConversion : public SPIRVOpLowering<SourceOp> {
|
||||||
|
|
|
@ -108,7 +108,7 @@ static Value createFPConstant(Location loc, Type srcType, Type dstType,
|
||||||
loc, dstType, rewriter.getFloatAttr(floatType, value));
|
loc, dstType, rewriter.getFloatAttr(floatType, value));
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Utility function for bitfiled ops:
|
/// Utility function for bitfield ops:
|
||||||
/// - `BitFieldInsert`
|
/// - `BitFieldInsert`
|
||||||
/// - `BitFieldSExtract`
|
/// - `BitFieldSExtract`
|
||||||
/// - `BitFieldUExtract`
|
/// - `BitFieldUExtract`
|
||||||
|
@ -163,7 +163,7 @@ static Value optionallyBroadcast(Location loc, Value value, Type srcType,
|
||||||
return value;
|
return value;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Utility function for bitfiled ops: `BitFieldInsert`, `BitFieldSExtract` and
|
/// Utility function for bitfield ops: `BitFieldInsert`, `BitFieldSExtract` and
|
||||||
/// `BitFieldUExtract`.
|
/// `BitFieldUExtract`.
|
||||||
/// Broadcast `Offset` and `Count` to match the type of `Base`. If `Base` is of
|
/// Broadcast `Offset` and `Count` to match the type of `Base`. If `Base` is of
|
||||||
/// a vector type, construct a vector that has:
|
/// a vector type, construct a vector that has:
|
||||||
|
@ -971,8 +971,8 @@ public:
|
||||||
|
|
||||||
Location loc = loopOp.getLoc();
|
Location loc = loopOp.getLoc();
|
||||||
|
|
||||||
// Split the current block after `spv.loop`. The remaing ops will be used in
|
// Split the current block after `spv.loop`. The remaining ops will be used
|
||||||
// `endBlock`.
|
// in `endBlock`.
|
||||||
Block *currentBlock = rewriter.getBlock();
|
Block *currentBlock = rewriter.getBlock();
|
||||||
auto position = Block::iterator(loopOp);
|
auto position = Block::iterator(loopOp);
|
||||||
Block *endBlock = rewriter.splitBlock(currentBlock, position);
|
Block *endBlock = rewriter.splitBlock(currentBlock, position);
|
||||||
|
@ -1028,7 +1028,7 @@ public:
|
||||||
|
|
||||||
Location loc = op.getLoc();
|
Location loc = op.getLoc();
|
||||||
|
|
||||||
// Split the current block after `spv.selection`. The remaing ops will be
|
// Split the current block after `spv.selection`. The remaining ops will be
|
||||||
// used in `continueBlock`.
|
// used in `continueBlock`.
|
||||||
auto *currentBlock = rewriter.getInsertionBlock();
|
auto *currentBlock = rewriter.getInsertionBlock();
|
||||||
rewriter.setInsertionPointAfter(op);
|
rewriter.setInsertionPointAfter(op);
|
||||||
|
@ -1387,7 +1387,7 @@ void mlir::populateSPIRVToLLVMConversionPatterns(
|
||||||
|
|
||||||
// Entry points and execution mode
|
// Entry points and execution mode
|
||||||
// Module generated from SPIR-V could have other "internal" functions, so
|
// Module generated from SPIR-V could have other "internal" functions, so
|
||||||
// having entry point and execution mode metadat can be useful. For now,
|
// having entry point and execution mode metadata can be useful. For now,
|
||||||
// simply remove them.
|
// simply remove them.
|
||||||
// TODO: Support EntryPoint/ExecutionMode properly.
|
// TODO: Support EntryPoint/ExecutionMode properly.
|
||||||
ErasePattern<spirv::EntryPointOp>, ErasePattern<spirv::ExecutionModeOp>,
|
ErasePattern<spirv::EntryPointOp>, ErasePattern<spirv::ExecutionModeOp>,
|
||||||
|
|
|
@ -132,7 +132,7 @@ LogicalResult BroadcastOpConverter::matchAndRewrite(
|
||||||
},
|
},
|
||||||
[&](OpBuilder &b, Location loc) {
|
[&](OpBuilder &b, Location loc) {
|
||||||
// The broadcasting logic is:
|
// The broadcasting logic is:
|
||||||
// - if one extent (here we arbitrariliy choose the extent from
|
// - if one extent (here we arbitrarily choose the extent from
|
||||||
// the greater-rank operand) is equal to 1, then take the extent
|
// the greater-rank operand) is equal to 1, then take the extent
|
||||||
// from the other operand
|
// from the other operand
|
||||||
// - otherwise, take the extent as-is.
|
// - otherwise, take the extent as-is.
|
||||||
|
|
|
@ -249,7 +249,7 @@ SmallVector<Type, 2> LLVMTypeConverter::convertUnrankedMemRefSignature() {
|
||||||
LLVM::LLVMType LLVMTypeConverter::convertFunctionSignature(
|
LLVM::LLVMType LLVMTypeConverter::convertFunctionSignature(
|
||||||
FunctionType funcTy, bool isVariadic,
|
FunctionType funcTy, bool isVariadic,
|
||||||
LLVMTypeConverter::SignatureConversion &result) {
|
LLVMTypeConverter::SignatureConversion &result) {
|
||||||
// Select the argument converter depending on the calling convetion.
|
// Select the argument converter depending on the calling convention.
|
||||||
auto funcArgConverter = options.useBarePtrCallConv
|
auto funcArgConverter = options.useBarePtrCallConv
|
||||||
? barePtrFuncArgTypeConverter
|
? barePtrFuncArgTypeConverter
|
||||||
: structFuncArgTypeConverter;
|
: structFuncArgTypeConverter;
|
||||||
|
|
|
@ -414,7 +414,7 @@ public:
|
||||||
ConversionPatternRewriter &rewriter) const override;
|
ConversionPatternRewriter &rewriter) const override;
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Converts integer compare operation on i1 type opearnds to SPIR-V ops.
|
/// Converts integer compare operation on i1 type operands to SPIR-V ops.
|
||||||
class BoolCmpIOpPattern final : public SPIRVOpLowering<CmpIOp> {
|
class BoolCmpIOpPattern final : public SPIRVOpLowering<CmpIOp> {
|
||||||
public:
|
public:
|
||||||
using SPIRVOpLowering<CmpIOp>::SPIRVOpLowering;
|
using SPIRVOpLowering<CmpIOp>::SPIRVOpLowering;
|
||||||
|
|
|
@ -100,7 +100,7 @@ public:
|
||||||
return failure();
|
return failure();
|
||||||
|
|
||||||
// Note that the dataPtr starts at the offset address specified by
|
// Note that the dataPtr starts at the offset address specified by
|
||||||
// indices, so no need to calculat offset size in bytes again in
|
// indices, so no need to calculate offset size in bytes again in
|
||||||
// the MUBUF instruction.
|
// the MUBUF instruction.
|
||||||
Value dataPtr = getDataPtr(loc, memRefType, adaptor.memref(),
|
Value dataPtr = getDataPtr(loc, memRefType, adaptor.memref(),
|
||||||
adaptor.indices(), rewriter);
|
adaptor.indices(), rewriter);
|
||||||
|
|
|
@ -84,7 +84,7 @@ extractBeneficiaryOps(Operation *op,
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
for (Value operand : op->getOperands()) {
|
for (Value operand : op->getOperands()) {
|
||||||
// It is already visisble in the kernel, keep going.
|
// It is already visible in the kernel, keep going.
|
||||||
if (availableValues.count(operand))
|
if (availableValues.count(operand))
|
||||||
continue;
|
continue;
|
||||||
// Else check whether it can be made available via sinking or already is a
|
// Else check whether it can be made available via sinking or already is a
|
||||||
|
|
|
@ -79,7 +79,7 @@ public:
|
||||||
bool isIdentified() const { return identified; }
|
bool isIdentified() const { return identified; }
|
||||||
bool isPacked() const {
|
bool isPacked() const {
|
||||||
assert(!isIdentified() &&
|
assert(!isIdentified() &&
|
||||||
"'packed' bit is not part of the key for identified stucts");
|
"'packed' bit is not part of the key for identified structs");
|
||||||
return packed;
|
return packed;
|
||||||
}
|
}
|
||||||
bool isOpaque() const {
|
bool isOpaque() const {
|
||||||
|
@ -196,7 +196,7 @@ public:
|
||||||
/// Constructs the storage from the given key. This sets up the uniquing key
|
/// Constructs the storage from the given key. This sets up the uniquing key
|
||||||
/// components and optionally the mutable component if they construction key
|
/// components and optionally the mutable component if they construction key
|
||||||
/// has the relevant information. In the latter case, the struct is considered
|
/// has the relevant information. In the latter case, the struct is considered
|
||||||
/// as initalized and can no longer be mutated.
|
/// as initialized and can no longer be mutated.
|
||||||
LLVMStructTypeStorage(const KeyTy &key) {
|
LLVMStructTypeStorage(const KeyTy &key) {
|
||||||
if (!key.isIdentified()) {
|
if (!key.isIdentified()) {
|
||||||
ArrayRef<LLVMType> types = key.getTypeList();
|
ArrayRef<LLVMType> types = key.getTypeList();
|
||||||
|
|
|
@ -711,10 +711,10 @@ static SmallVector<SmallVector<AffineExpr, 2>, 2>
|
||||||
convertReassociationIndicesToMaps(
|
convertReassociationIndicesToMaps(
|
||||||
OpBuilder &b, ArrayRef<ReassociationIndices> reassociationIndices) {
|
OpBuilder &b, ArrayRef<ReassociationIndices> reassociationIndices) {
|
||||||
SmallVector<SmallVector<AffineExpr, 2>, 2> reassociationMaps;
|
SmallVector<SmallVector<AffineExpr, 2>, 2> reassociationMaps;
|
||||||
for (const auto &indicies : reassociationIndices) {
|
for (const auto &indices : reassociationIndices) {
|
||||||
SmallVector<AffineExpr, 2> reassociationMap;
|
SmallVector<AffineExpr, 2> reassociationMap;
|
||||||
reassociationMap.reserve(indicies.size());
|
reassociationMap.reserve(indices.size());
|
||||||
for (int64_t index : indicies)
|
for (int64_t index : indices)
|
||||||
reassociationMap.push_back(b.getAffineDimExpr(index));
|
reassociationMap.push_back(b.getAffineDimExpr(index));
|
||||||
reassociationMaps.push_back(std::move(reassociationMap));
|
reassociationMaps.push_back(std::move(reassociationMap));
|
||||||
}
|
}
|
||||||
|
|
|
@ -147,7 +147,7 @@ static void generateFusedTensorOpRegion(PatternRewriter &rewriter,
|
||||||
}
|
}
|
||||||
|
|
||||||
// If consumer is an indexed_generic op, map the indices to the block
|
// If consumer is an indexed_generic op, map the indices to the block
|
||||||
// arguments directly. Otherwise, add the same type of arugment and map to
|
// arguments directly. Otherwise, add the same type of argument and map to
|
||||||
// it.
|
// it.
|
||||||
if (consumerArg.index() < numConsumerIndices) {
|
if (consumerArg.index() < numConsumerIndices) {
|
||||||
mapper.map(consumerArg.value(),
|
mapper.map(consumerArg.value(),
|
||||||
|
|
|
@ -80,7 +80,7 @@ void mlir::linalg::hoistViewAllocOps(FuncOp func) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Return true if we can prove that the transfer operations access dijoint
|
/// Return true if we can prove that the transfer operations access disjoint
|
||||||
/// memory.
|
/// memory.
|
||||||
static bool isDisjoint(VectorTransferOpInterface transferA,
|
static bool isDisjoint(VectorTransferOpInterface transferA,
|
||||||
VectorTransferOpInterface transferB) {
|
VectorTransferOpInterface transferB) {
|
||||||
|
|
|
@ -161,7 +161,7 @@ struct LinalgOpInstancePromotionOptions {
|
||||||
CopyCallbackFn copyInFn;
|
CopyCallbackFn copyInFn;
|
||||||
CopyCallbackFn copyOutFn;
|
CopyCallbackFn copyOutFn;
|
||||||
|
|
||||||
/// Allow the use of dynamicaly-sized buffers.
|
/// Allow the use of dynamically-sized buffers.
|
||||||
bool dynamicBuffers;
|
bool dynamicBuffers;
|
||||||
/// Alignment of promoted buffer.
|
/// Alignment of promoted buffer.
|
||||||
Optional<unsigned> alignment;
|
Optional<unsigned> alignment;
|
||||||
|
|
|
@ -302,7 +302,7 @@ static Type parseArrayType(SPIRVDialect const &dialect,
|
||||||
}
|
}
|
||||||
|
|
||||||
// cooperative-matrix-type ::= `!spv.coopmatrix` `<` element-type ',' scope ','
|
// cooperative-matrix-type ::= `!spv.coopmatrix` `<` element-type ',' scope ','
|
||||||
// rows ',' coloumns>`
|
// rows ',' columns>`
|
||||||
static Type parseCooperativeMatrixType(SPIRVDialect const &dialect,
|
static Type parseCooperativeMatrixType(SPIRVDialect const &dialect,
|
||||||
DialectAsmParser &parser) {
|
DialectAsmParser &parser) {
|
||||||
if (parser.parseLess())
|
if (parser.parseLess())
|
||||||
|
@ -621,7 +621,7 @@ static Type parseStructType(SPIRVDialect const &dialect,
|
||||||
|
|
||||||
StringRef identifier;
|
StringRef identifier;
|
||||||
|
|
||||||
// Check if this is an idenitifed struct type.
|
// Check if this is an identified struct type.
|
||||||
if (succeeded(parser.parseOptionalKeyword(&identifier))) {
|
if (succeeded(parser.parseOptionalKeyword(&identifier))) {
|
||||||
// Check if this is a possible recursive reference.
|
// Check if this is a possible recursive reference.
|
||||||
if (succeeded(parser.parseOptionalGreater())) {
|
if (succeeded(parser.parseOptionalGreater())) {
|
||||||
|
|
|
@ -217,9 +217,9 @@ static ParseResult parseMemoryAccessAttributes(OpAsmParser &parser,
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO Make sure to merge this and the previous function into one template
|
// TODO Make sure to merge this and the previous function into one template
|
||||||
// parameterized by memroy access attribute name and alignment. Doing so now
|
// parameterized by memory access attribute name and alignment. Doing so now
|
||||||
// results in VS2017 in producing an internal error (at the call site) that's
|
// results in VS2017 in producing an internal error (at the call site) that's
|
||||||
// not detailed enough to understand what is happenning.
|
// not detailed enough to understand what is happening.
|
||||||
static ParseResult parseSourceMemoryAccessAttributes(OpAsmParser &parser,
|
static ParseResult parseSourceMemoryAccessAttributes(OpAsmParser &parser,
|
||||||
OperationState &state) {
|
OperationState &state) {
|
||||||
// Parse an optional list of attributes staring with '['
|
// Parse an optional list of attributes staring with '['
|
||||||
|
@ -274,9 +274,9 @@ static void printMemoryAccessAttribute(
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO Make sure to merge this and the previous function into one template
|
// TODO Make sure to merge this and the previous function into one template
|
||||||
// parameterized by memroy access attribute name and alignment. Doing so now
|
// parameterized by memory access attribute name and alignment. Doing so now
|
||||||
// results in VS2017 in producing an internal error (at the call site) that's
|
// results in VS2017 in producing an internal error (at the call site) that's
|
||||||
// not detailed enough to understand what is happenning.
|
// not detailed enough to understand what is happening.
|
||||||
template <typename MemoryOpTy>
|
template <typename MemoryOpTy>
|
||||||
static void printSourceMemoryAccessAttribute(
|
static void printSourceMemoryAccessAttribute(
|
||||||
MemoryOpTy memoryOp, OpAsmPrinter &printer,
|
MemoryOpTy memoryOp, OpAsmPrinter &printer,
|
||||||
|
@ -393,9 +393,9 @@ static LogicalResult verifyMemoryAccessAttribute(MemoryOpTy memoryOp) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO Make sure to merge this and the previous function into one template
|
// TODO Make sure to merge this and the previous function into one template
|
||||||
// parameterized by memroy access attribute name and alignment. Doing so now
|
// parameterized by memory access attribute name and alignment. Doing so now
|
||||||
// results in VS2017 in producing an internal error (at the call site) that's
|
// results in VS2017 in producing an internal error (at the call site) that's
|
||||||
// not detailed enough to understand what is happenning.
|
// not detailed enough to understand what is happening.
|
||||||
template <typename MemoryOpTy>
|
template <typename MemoryOpTy>
|
||||||
static LogicalResult verifySourceMemoryAccessAttribute(MemoryOpTy memoryOp) {
|
static LogicalResult verifySourceMemoryAccessAttribute(MemoryOpTy memoryOp) {
|
||||||
// ODS checks for attributes values. Just need to verify that if the
|
// ODS checks for attributes values. Just need to verify that if the
|
||||||
|
|
|
@ -810,7 +810,7 @@ struct spirv::detail::StructTypeStorage : public TypeStorage {
|
||||||
std::tuple<StringRef, ArrayRef<Type>, ArrayRef<StructType::OffsetInfo>,
|
std::tuple<StringRef, ArrayRef<Type>, ArrayRef<StructType::OffsetInfo>,
|
||||||
ArrayRef<StructType::MemberDecorationInfo>>;
|
ArrayRef<StructType::MemberDecorationInfo>>;
|
||||||
|
|
||||||
/// For idetified structs, return true if the given key contains the same
|
/// For identified structs, return true if the given key contains the same
|
||||||
/// identifier.
|
/// identifier.
|
||||||
///
|
///
|
||||||
/// For literal structs, return true if the given key contains a matching list
|
/// For literal structs, return true if the given key contains a matching list
|
||||||
|
|
|
@ -1033,7 +1033,7 @@ bool Serializer::isInterfaceStructPtrType(Type type) const {
|
||||||
LogicalResult Serializer::processType(Location loc, Type type,
|
LogicalResult Serializer::processType(Location loc, Type type,
|
||||||
uint32_t &typeID) {
|
uint32_t &typeID) {
|
||||||
// Maintains a set of names for nested identified struct types. This is used
|
// Maintains a set of names for nested identified struct types. This is used
|
||||||
// to properly seialize resursive references.
|
// to properly serialize resursive references.
|
||||||
llvm::SetVector<StringRef> serializationCtx;
|
llvm::SetVector<StringRef> serializationCtx;
|
||||||
return processTypeImpl(loc, type, typeID, serializationCtx);
|
return processTypeImpl(loc, type, typeID, serializationCtx);
|
||||||
}
|
}
|
||||||
|
@ -1170,7 +1170,7 @@ LogicalResult Serializer::prepareBasicType(
|
||||||
spirv::Opcode::OpTypeForwardPointer,
|
spirv::Opcode::OpTypeForwardPointer,
|
||||||
forwardPtrOperands);
|
forwardPtrOperands);
|
||||||
|
|
||||||
// 2. Find the the pointee (enclosing) struct.
|
// 2. Find the pointee (enclosing) struct.
|
||||||
auto structType = spirv::StructType::getIdentified(
|
auto structType = spirv::StructType::getIdentified(
|
||||||
module.getContext(), pointeeStruct.getIdentifier());
|
module.getContext(), pointeeStruct.getIdentifier());
|
||||||
|
|
||||||
|
|
|
@ -878,7 +878,7 @@ static Value foldExtractFromShapeCast(ExtractOp extractOp) {
|
||||||
}
|
}
|
||||||
|
|
||||||
int64_t position = linearize(extractedPos, strides);
|
int64_t position = linearize(extractedPos, strides);
|
||||||
// Then extract the strides assoociated to the shapeCast op vector source and
|
// Then extract the strides associated to the shapeCast op vector source and
|
||||||
// delinearize the position using those strides.
|
// delinearize the position using those strides.
|
||||||
SmallVector<int64_t, 4> newStrides;
|
SmallVector<int64_t, 4> newStrides;
|
||||||
int64_t numDimension =
|
int64_t numDimension =
|
||||||
|
|
|
@ -2178,7 +2178,7 @@ LogicalResult mlir::vector::splitFullAndPartialTransferPrecondition(
|
||||||
/// 2. else return a new MemRefType obtained by iterating over the shape and
|
/// 2. else return a new MemRefType obtained by iterating over the shape and
|
||||||
/// strides and:
|
/// strides and:
|
||||||
/// a. keeping the ones that are static and equal across `aT` and `bT`.
|
/// a. keeping the ones that are static and equal across `aT` and `bT`.
|
||||||
/// b. using a dynamic shape and/or stride for the dimeniosns that don't
|
/// b. using a dynamic shape and/or stride for the dimensions that don't
|
||||||
/// agree.
|
/// agree.
|
||||||
static MemRefType getCastCompatibleMemRefType(MemRefType aT, MemRefType bT) {
|
static MemRefType getCastCompatibleMemRefType(MemRefType aT, MemRefType bT) {
|
||||||
if (MemRefCastOp::areCastCompatible(aT, bT))
|
if (MemRefCastOp::areCastCompatible(aT, bT))
|
||||||
|
|
|
@ -292,8 +292,8 @@ public:
|
||||||
/// operations.
|
/// operations.
|
||||||
llvm::StringMap<AbstractOperation> registeredOperations;
|
llvm::StringMap<AbstractOperation> registeredOperations;
|
||||||
|
|
||||||
/// Identifers are uniqued by string value and use the internal string set for
|
/// Identifiers are uniqued by string value and use the internal string set
|
||||||
/// storage.
|
/// for storage.
|
||||||
llvm::StringSet<llvm::BumpPtrAllocator &> identifiers;
|
llvm::StringSet<llvm::BumpPtrAllocator &> identifiers;
|
||||||
/// A thread local cache of identifiers to reduce lock contention.
|
/// A thread local cache of identifiers to reduce lock contention.
|
||||||
ThreadLocalCache<llvm::StringMap<llvm::StringMapEntry<llvm::NoneType> *>>
|
ThreadLocalCache<llvm::StringMap<llvm::StringMapEntry<llvm::NoneType> *>>
|
||||||
|
|
|
@ -140,7 +140,7 @@ void PatternRewriter::mergeBlockBefore(Block *source, Operation *op,
|
||||||
assert(source->hasNoSuccessors() &&
|
assert(source->hasNoSuccessors() &&
|
||||||
"expected 'source' to have no successors");
|
"expected 'source' to have no successors");
|
||||||
|
|
||||||
// Split the block containing 'op' into two, one containg all operations
|
// Split the block containing 'op' into two, one containing all operations
|
||||||
// before 'op' (prologue) and another (epilogue) containing 'op' and all
|
// before 'op' (prologue) and another (epilogue) containing 'op' and all
|
||||||
// operations after it.
|
// operations after it.
|
||||||
Block *prologue = op->getBlock();
|
Block *prologue = op->getBlock();
|
||||||
|
|
|
@ -167,7 +167,7 @@ LogicalResult detail::verifyTypesAlongControlFlowEdges(Operation *op) {
|
||||||
for (unsigned regionNo : llvm::seq(0U, op->getNumRegions())) {
|
for (unsigned regionNo : llvm::seq(0U, op->getNumRegions())) {
|
||||||
Region ®ion = op->getRegion(regionNo);
|
Region ®ion = op->getRegion(regionNo);
|
||||||
|
|
||||||
// Since the interface cannnot distinguish between different ReturnLike
|
// Since the interface cannot distinguish between different ReturnLike
|
||||||
// ops within the region branching to different successors, all ReturnLike
|
// ops within the region branching to different successors, all ReturnLike
|
||||||
// ops in this region should have the same operand types. We will then use
|
// ops in this region should have the same operand types. We will then use
|
||||||
// one of them as the representative for type matching.
|
// one of them as the representative for type matching.
|
||||||
|
|
|
@ -14,5 +14,5 @@ using namespace mlir;
|
||||||
// VectorUnroll Interfaces
|
// VectorUnroll Interfaces
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
|
||||||
/// Include the definitions of the VectorUntoll interfaces.
|
/// Include the definitions of the VectorUnroll interfaces.
|
||||||
#include "mlir/Interfaces/VectorInterfaces.cpp.inc"
|
#include "mlir/Interfaces/VectorInterfaces.cpp.inc"
|
||||||
|
|
|
@ -9,7 +9,7 @@
|
||||||
// This file defines the Tester class used in the MLIR Reduce tool.
|
// This file defines the Tester class used in the MLIR Reduce tool.
|
||||||
//
|
//
|
||||||
// A Tester object is passed as an argument to the reduction passes and it is
|
// A Tester object is passed as an argument to the reduction passes and it is
|
||||||
// used to run the interestigness testing script on the different generated
|
// used to run the interestingness testing script on the different generated
|
||||||
// reduced variants of the test case.
|
// reduced variants of the test case.
|
||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
|
|
@ -5,7 +5,7 @@
|
||||||
# component dependencies, LLVM_LINK_LLVM_DYLIB tends to introduce a
|
# component dependencies, LLVM_LINK_LLVM_DYLIB tends to introduce a
|
||||||
# dependence on libLLVM.so) However, it must also be linkable against
|
# dependence on libLLVM.so) However, it must also be linkable against
|
||||||
# libMLIR.so in some contexts (see unittests/Tablegen, for instance, which
|
# libMLIR.so in some contexts (see unittests/Tablegen, for instance, which
|
||||||
# has a dependance on MLIRIR, which must depend on libLLVM.so). This works
|
# has a dependence on MLIRIR, which must depend on libLLVM.so). This works
|
||||||
# in this special case because this library is static.
|
# in this special case because this library is static.
|
||||||
|
|
||||||
llvm_add_library(MLIRTableGen STATIC
|
llvm_add_library(MLIRTableGen STATIC
|
||||||
|
|
|
@ -487,7 +487,7 @@ ModuleTranslation::convertOmpOperation(Operation &opInst,
|
||||||
return success();
|
return success();
|
||||||
})
|
})
|
||||||
.Case([&](omp::FlushOp) {
|
.Case([&](omp::FlushOp) {
|
||||||
// No support in Openmp runtime funciton (__kmpc_flush) to accept
|
// No support in Openmp runtime function (__kmpc_flush) to accept
|
||||||
// the argument list.
|
// the argument list.
|
||||||
// OpenMP standard states the following:
|
// OpenMP standard states the following:
|
||||||
// "An implementation may implement a flush with a list by ignoring
|
// "An implementation may implement a flush with a list by ignoring
|
||||||
|
|
|
@ -148,7 +148,7 @@ public:
|
||||||
}
|
}
|
||||||
|
|
||||||
/// This method returns the mapping values list. The unknown result values
|
/// This method returns the mapping values list. The unknown result values
|
||||||
/// that only their indicies are available are replaced with their values.
|
/// that only their indices are available are replaced with their values.
|
||||||
void getMappingValues(ValueRange valuesToReplaceIndices,
|
void getMappingValues(ValueRange valuesToReplaceIndices,
|
||||||
SmallVectorImpl<Value> &values) {
|
SmallVectorImpl<Value> &values) {
|
||||||
// Append available values to the list.
|
// Append available values to the list.
|
||||||
|
|
|
@ -9,7 +9,7 @@
|
||||||
// This transformation pass performs a sparse conditional constant propagation
|
// This transformation pass performs a sparse conditional constant propagation
|
||||||
// in MLIR. It identifies values known to be constant, propagates that
|
// in MLIR. It identifies values known to be constant, propagates that
|
||||||
// information throughout the IR, and replaces them. This is done with an
|
// information throughout the IR, and replaces them. This is done with an
|
||||||
// optimisitic dataflow analysis that assumes that all values are constant until
|
// optimistic dataflow analysis that assumes that all values are constant until
|
||||||
// proven otherwise.
|
// proven otherwise.
|
||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
@ -27,7 +27,7 @@ using namespace mlir;
|
||||||
|
|
||||||
namespace {
|
namespace {
|
||||||
/// This class represents a single lattice value. A lattive value corresponds to
|
/// This class represents a single lattice value. A lattive value corresponds to
|
||||||
/// the various different states that a value in the SCCP dataflow anaylsis can
|
/// the various different states that a value in the SCCP dataflow analysis can
|
||||||
/// take. See 'Kind' below for more details on the different states a value can
|
/// take. See 'Kind' below for more details on the different states a value can
|
||||||
/// take.
|
/// take.
|
||||||
class LatticeValue {
|
class LatticeValue {
|
||||||
|
|
|
@ -522,7 +522,7 @@ void ArgConverter::insertConversion(Block *newBlock,
|
||||||
}
|
}
|
||||||
|
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
// Rewriter and Transation State
|
// Rewriter and Translation State
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
namespace {
|
namespace {
|
||||||
/// This class contains a snapshot of the current conversion rewriter state.
|
/// This class contains a snapshot of the current conversion rewriter state.
|
||||||
|
|
|
@ -694,13 +694,13 @@ static void setInterTileBoundsParametric(OpBuilder &b, AffineForOp origLoop,
|
||||||
AffineForOp newLoop, Value tileSize) {
|
AffineForOp newLoop, Value tileSize) {
|
||||||
OperandRange newLbOperands = origLoop.getLowerBoundOperands();
|
OperandRange newLbOperands = origLoop.getLowerBoundOperands();
|
||||||
|
|
||||||
// The lower bounds for inter-tile loops are same as the correspondig lower
|
// The lower bounds for inter-tile loops are same as the corresponding lower
|
||||||
// bounds of original loops.
|
// bounds of original loops.
|
||||||
newLoop.setLowerBound(newLbOperands, origLoop.getLowerBoundMap());
|
newLoop.setLowerBound(newLbOperands, origLoop.getLowerBoundMap());
|
||||||
|
|
||||||
// The new upper bound map for inter-tile loops, assuming constant lower
|
// The new upper bound map for inter-tile loops, assuming constant lower
|
||||||
// bounds, are now originalLowerBound + ceildiv((orignalUpperBound -
|
// bounds, are now originalLowerBound + ceildiv((originalUpperBound -
|
||||||
// originalLowerBound), tiling paramter); where tiling parameter is the
|
// originalLowerBound), tiling parameter); where tiling parameter is the
|
||||||
// respective tile size for that loop. For e.g. if the original ubmap was
|
// respective tile size for that loop. For e.g. if the original ubmap was
|
||||||
// ()->(1024), the new map will be
|
// ()->(1024), the new map will be
|
||||||
// ()[s0]->(ceildiv((1024 -lb) % s0)), where s0 is the tiling parameter.
|
// ()[s0]->(ceildiv((1024 -lb) % s0)), where s0 is the tiling parameter.
|
||||||
|
@ -757,7 +757,7 @@ static void setInterTileBoundsParametric(OpBuilder &b, AffineForOp origLoop,
|
||||||
// ubmap has only one result expression. For e.g.
|
// ubmap has only one result expression. For e.g.
|
||||||
// affine.for %i = 5 to %ub
|
// affine.for %i = 5 to %ub
|
||||||
//
|
//
|
||||||
// A symbol operand is added which represents the tiling paramater. The
|
// A symbol operand is added which represents the tiling parameter. The
|
||||||
// new loop bounds here will be like ()[s0, s1] -> ((s0 - 5) ceildiv s1 + 5)
|
// new loop bounds here will be like ()[s0, s1] -> ((s0 - 5) ceildiv s1 + 5)
|
||||||
// where 's0' is the original upper bound and 's1' is the tiling
|
// where 's0' is the original upper bound and 's1' is the tiling
|
||||||
// parameter. 2.) When ubMap has more than one result expression. For e.g.
|
// parameter. 2.) When ubMap has more than one result expression. For e.g.
|
||||||
|
|
|
@ -981,7 +981,7 @@ int printAffineExpr(MlirContext ctx) {
|
||||||
!mlirAffineExprIsFunctionOfDim(affineCeilDivExpr, 5))
|
!mlirAffineExprIsFunctionOfDim(affineCeilDivExpr, 5))
|
||||||
return 8;
|
return 8;
|
||||||
|
|
||||||
// Tests 'IsA' methods of affine binary operaion expression.
|
// Tests 'IsA' methods of affine binary operation expression.
|
||||||
if (!mlirAffineExprIsAAdd(affineAddExpr))
|
if (!mlirAffineExprIsAAdd(affineAddExpr))
|
||||||
return 9;
|
return 9;
|
||||||
|
|
||||||
|
|
|
@ -111,9 +111,9 @@ func @caller(%arg0: tensor<5xf32>) -> tensor<5xf32> {
|
||||||
|
|
||||||
// -----
|
// -----
|
||||||
|
|
||||||
// Test case: Testing BufferAssginmnetCallOpConverter to see if it matches with the
|
// Test case: Testing BufferAssignmentCallOpConverter to see if it matches with the
|
||||||
// signature of the new signature of the callee function when there are tuple typed
|
// signature of the new signature of the callee function when there are tuple typed
|
||||||
// args and results. BufferAssginmentTypeConverter is set to flatten tuple typed
|
// args and results. BufferAssignmentTypeConverter is set to flatten tuple typed
|
||||||
// arguments. The tuple typed values should be decomposed and composed using
|
// arguments. The tuple typed values should be decomposed and composed using
|
||||||
// get_tuple_element and make_tuple operations of test dialect. Tensor types are
|
// get_tuple_element and make_tuple operations of test dialect. Tensor types are
|
||||||
// converted to Memref. Memref typed function results remain as function results.
|
// converted to Memref. Memref typed function results remain as function results.
|
||||||
|
@ -158,10 +158,10 @@ func @caller(%arg0: tuple<tensor<2xf32>,i1, tensor<5xf32>>) -> tuple<tensor<2xf3
|
||||||
|
|
||||||
// -----
|
// -----
|
||||||
|
|
||||||
// Test case: Testing BufferAssginmnetFuncOpConverter and
|
// Test case: Testing BufferAssignmentFuncOpConverter and
|
||||||
// BufferAssginmentReturnOpConverter to see if the return operation matches with
|
// BufferAssignmentReturnOpConverter to see if the return operation matches with
|
||||||
// the new function signature when there are tuple typed args and results.
|
// the new function signature when there are tuple typed args and results.
|
||||||
// BufferAssginmentTypeConverter is set to flatten tuple typed arguments. The tuple
|
// BufferAssignmentTypeConverter is set to flatten tuple typed arguments. The tuple
|
||||||
// typed values should be decomposed and composed using get_tuple_element and
|
// typed values should be decomposed and composed using get_tuple_element and
|
||||||
// make_tuple operations of test dialect. Tensor types are converted to Memref.
|
// make_tuple operations of test dialect. Tensor types are converted to Memref.
|
||||||
// Memref typed function results remain as function results.
|
// Memref typed function results remain as function results.
|
||||||
|
|
|
@ -302,9 +302,9 @@ func @func_with_unranked_arg(%arg0: tensor<*xf32>) {
|
||||||
|
|
||||||
// -----
|
// -----
|
||||||
|
|
||||||
// Test case: Testing BufferAssginmnetCallOpConverter to see if it matches with the
|
// Test case: Testing BufferAssignmentCallOpConverter to see if it matches with the
|
||||||
// signature of the new signature of the callee function when there are tuple typed
|
// signature of the new signature of the callee function when there are tuple typed
|
||||||
// args and results. BufferAssginmentTypeConverter is set to flatten tuple typed
|
// args and results. BufferAssignmentTypeConverter is set to flatten tuple typed
|
||||||
// arguments. The tuple typed values should be decomposed and composed using
|
// arguments. The tuple typed values should be decomposed and composed using
|
||||||
// get_tuple_element and make_tuple operations of test dialect. Tensor types are
|
// get_tuple_element and make_tuple operations of test dialect. Tensor types are
|
||||||
// converted to Memref. Memref typed function results are appended to the function
|
// converted to Memref. Memref typed function results are appended to the function
|
||||||
|
@ -359,10 +359,10 @@ func @caller(%arg0: tuple<tensor<2xf32>,i1, tensor<5xf32>>) -> tuple<tensor<2xf3
|
||||||
|
|
||||||
// -----
|
// -----
|
||||||
|
|
||||||
// Test case: Testing BufferAssginmnetFuncOpConverter and
|
// Test case: Testing BufferAssignmentFuncOpConverter and
|
||||||
// BufferAssginmentReturnOpConverter to see if the return operation matches with
|
// BufferAssignmentReturnOpConverter to see if the return operation matches with
|
||||||
// the new function signature when there are tuple typed args and results.
|
// the new function signature when there are tuple typed args and results.
|
||||||
// BufferAssginmentTypeConverter is set to flatten tuple typed arguments. The tuple
|
// BufferAssignmentTypeConverter is set to flatten tuple typed arguments. The tuple
|
||||||
// typed values should be decomposed and composed using get_tuple_element and
|
// typed values should be decomposed and composed using get_tuple_element and
|
||||||
// make_tuple operations of test dialect. Tensor types are converted to Memref.
|
// make_tuple operations of test dialect. Tensor types are converted to Memref.
|
||||||
// Memref typed function results are appended to the function arguments list.
|
// Memref typed function results are appended to the function arguments list.
|
||||||
|
|
|
@ -26,7 +26,7 @@ func @test_norm(%arg0 : memref<1x16x14x14xf32, #map0>) -> () {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Same test with op_nonnorm, with maps in the argmentets and the operations in the function.
|
// Same test with op_nonnorm, with maps in the arguments and the operations in the function.
|
||||||
|
|
||||||
// CHECK-LABEL: test_nonnorm
|
// CHECK-LABEL: test_nonnorm
|
||||||
// CHECK-SAME: (%[[ARG0:[a-z0-9]*]]: memref<1x16x14x14xf32, #map0>)
|
// CHECK-SAME: (%[[ARG0:[a-z0-9]*]]: memref<1x16x14x14xf32, #map0>)
|
||||||
|
|
|
@ -90,7 +90,7 @@ def E_IntegerType : TestType<"Integer"> {
|
||||||
let parameters = (
|
let parameters = (
|
||||||
ins
|
ins
|
||||||
"SignednessSemantics":$signedness,
|
"SignednessSemantics":$signedness,
|
||||||
TypeParameter<"unsigned", "Bitwdith of integer">:$width
|
TypeParameter<"unsigned", "Bitwidth of integer">:$width
|
||||||
);
|
);
|
||||||
|
|
||||||
// DECL-LABEL: IntegerType: public ::mlir::Type
|
// DECL-LABEL: IntegerType: public ::mlir::Type
|
||||||
|
|
|
@ -8,7 +8,7 @@
|
||||||
//
|
//
|
||||||
// This file defines the OpReducer class. It defines a variant generator method
|
// This file defines the OpReducer class. It defines a variant generator method
|
||||||
// with the purpose of producing different variants by eliminating a
|
// with the purpose of producing different variants by eliminating a
|
||||||
// parametarizable type of operations from the parent module.
|
// parameterizable type of operations from the parent module.
|
||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
#include "mlir/Reducer/Passes/OpReducer.h"
|
#include "mlir/Reducer/Passes/OpReducer.h"
|
||||||
|
|
|
@ -39,7 +39,7 @@ void ReductionNode::measureAndTest(const Tester &test) {
|
||||||
SmallString<128> filepath;
|
SmallString<128> filepath;
|
||||||
int fd;
|
int fd;
|
||||||
|
|
||||||
// Print module to temprary file.
|
// Print module to temporary file.
|
||||||
std::error_code ec =
|
std::error_code ec =
|
||||||
llvm::sys::fs::createTemporaryFile("mlir-reduce", "mlir", fd, filepath);
|
llvm::sys::fs::createTemporaryFile("mlir-reduce", "mlir", fd, filepath);
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,7 @@ void ReductionTreeUtils::updateGoldenModule(ModuleOp &golden,
|
||||||
reduced.getBody()->getOperations());
|
reduced.getBody()->getOperations());
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Update the the smallest node traversed so far in the reduction tree and
|
/// Update the smallest node traversed so far in the reduction tree and
|
||||||
/// print the debugging information for the currNode being traversed.
|
/// print the debugging information for the currNode being traversed.
|
||||||
void ReductionTreeUtils::updateSmallestNode(ReductionNode *currNode,
|
void ReductionTreeUtils::updateSmallestNode(ReductionNode *currNode,
|
||||||
ReductionNode *&smallestNode,
|
ReductionNode *&smallestNode,
|
||||||
|
@ -95,7 +95,7 @@ static std::vector<std::tuple<int, int>> getRanges(std::vector<bool> tSpace,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create the specified number of variants by applying the transform method
|
/// Create the specified number of variants by applying the transform method
|
||||||
/// to different ranges of indices in the parent module. The isDeletion bolean
|
/// to different ranges of indices in the parent module. The isDeletion boolean
|
||||||
/// specifies if the transformation is the deletion of indices.
|
/// specifies if the transformation is the deletion of indices.
|
||||||
void ReductionTreeUtils::createVariants(
|
void ReductionTreeUtils::createVariants(
|
||||||
ReductionNode *parent, const Tester &test, int numVariants,
|
ReductionNode *parent, const Tester &test, int numVariants,
|
||||||
|
|
|
@ -325,7 +325,7 @@ private:
|
||||||
// attribute (the generated function call returns an Attribute);
|
// attribute (the generated function call returns an Attribute);
|
||||||
// - operandGet corresponds to the name of the function with which to retrieve
|
// - operandGet corresponds to the name of the function with which to retrieve
|
||||||
// an operand (the generated function call returns an OperandRange);
|
// an operand (the generated function call returns an OperandRange);
|
||||||
// - reultGet corresponds to the name of the function to get an result (the
|
// - resultGet corresponds to the name of the function to get an result (the
|
||||||
// generated function call returns a ValueRange);
|
// generated function call returns a ValueRange);
|
||||||
static void populateSubstitutions(const Operator &op, const char *attrGet,
|
static void populateSubstitutions(const Operator &op, const char *attrGet,
|
||||||
const char *operandGet, const char *resultGet,
|
const char *operandGet, const char *resultGet,
|
||||||
|
@ -570,7 +570,7 @@ void OpEmitter::genAttrGetters() {
|
||||||
PrintWarning(
|
PrintWarning(
|
||||||
op.getLoc(),
|
op.getLoc(),
|
||||||
formatv(
|
formatv(
|
||||||
"op has non-materialzable derived attributes '{0}', skipping",
|
"op has non-materializable derived attributes '{0}', skipping",
|
||||||
os.str()));
|
os.str()));
|
||||||
body << formatv(" emitOpError(\"op has non-materializable derived "
|
body << formatv(" emitOpError(\"op has non-materializable derived "
|
||||||
"attributes '{0}'\");\n",
|
"attributes '{0}'\");\n",
|
||||||
|
@ -965,7 +965,7 @@ void OpEmitter::genSeparateArgParamBuilder() {
|
||||||
llvm_unreachable("unhandled TypeParamKind");
|
llvm_unreachable("unhandled TypeParamKind");
|
||||||
};
|
};
|
||||||
|
|
||||||
// Some of the build methods generated here may be amiguous, but TableGen's
|
// Some of the build methods generated here may be ambiguous, but TableGen's
|
||||||
// ambiguous function detection will elide those ones.
|
// ambiguous function detection will elide those ones.
|
||||||
for (auto attrType : attrBuilderType) {
|
for (auto attrType : attrBuilderType) {
|
||||||
emit(attrType, TypeParamKind::Separate, /*inferType=*/false);
|
emit(attrType, TypeParamKind::Separate, /*inferType=*/false);
|
||||||
|
|
Loading…
Reference in New Issue