[mlir] NFC: fix trivial typos

fix typos in comments and documents

Reviewed By: jpienaar

Differential Revision: https://reviews.llvm.org/D90089
This commit is contained in:
Kazuaki Ishizaki 2020-10-29 04:03:15 +09:00
parent 220293da53
commit 41b09f4eff
48 changed files with 86 additions and 86 deletions

View File

@ -797,7 +797,7 @@ A custom directive has two main parts: The `UserDirective` and the `Params`. A
custom directive is transformed into a call to a `print*` and a `parse*` method custom directive is transformed into a call to a `print*` and a `parse*` method
when generating the C++ code for the format. The `UserDirective` is an when generating the C++ code for the format. The `UserDirective` is an
identifier used as a suffix to these two calls, i.e., `custom<MyDirective>(...)` identifier used as a suffix to these two calls, i.e., `custom<MyDirective>(...)`
would result in calls to `parseMyDirective` and `printMyDirective` wihtin the would result in calls to `parseMyDirective` and `printMyDirective` within the
parser and printer respectively. `Params` may be any combination of variables parser and printer respectively. `Params` may be any combination of variables
(i.e. Attribute, Operand, Successor, etc.), type directives, and `attr-dict`. (i.e. Attribute, Operand, Successor, etc.), type directives, and `attr-dict`.
The type directives must refer to a variable, but that variable need not also The type directives must refer to a variable, but that variable need not also
@ -1515,7 +1515,7 @@ def IntegerType : Test_Type<"TestInteger"> {
The name of the C++ class which gets generated defaults to The name of the C++ class which gets generated defaults to
`<classParamName>Type` (e.g. `TestIntegerType` in the above example). This `<classParamName>Type` (e.g. `TestIntegerType` in the above example). This
can be overridden via the the `cppClassName` field. The field `mnemonic` is can be overridden via the `cppClassName` field. The field `mnemonic` is
to specify the asm name for parsing. It is optional and not specifying it to specify the asm name for parsing. It is optional and not specifying it
will imply that no parser or printer methods are attached to this class. will imply that no parser or printer methods are attached to this class.
@ -1554,7 +1554,7 @@ The default storage constructor blindly copies fields by value. It does not
know anything about the types. In this case, the ArrayRef<int> requires know anything about the types. In this case, the ArrayRef<int> requires
allocation with `dims = allocator.copyInto(dims)`. allocation with `dims = allocator.copyInto(dims)`.
You can specify the necessary constuctor by specializing the `TypeParameter` You can specify the necessary constructor by specializing the `TypeParameter`
tblgen class: tblgen class:
```tablegen ```tablegen

View File

@ -21,7 +21,7 @@
extern "C" { extern "C" {
#endif #endif
/** An opaque reference to a dignostic, always owned by the diagnostics engine /** An opaque reference to a diagnostic, always owned by the diagnostics engine
* (context). Must not be stored outside of the diagnostic handler. */ * (context). Must not be stored outside of the diagnostic handler. */
struct MlirDiagnostic { struct MlirDiagnostic {
void *ptr; void *ptr;

View File

@ -203,7 +203,7 @@ MlirAttribute mlirSymbolRefAttrGet(MlirContext ctx, intptr_t length,
* live as long as the context in which the attribute lives. */ * live as long as the context in which the attribute lives. */
MlirStringRef mlirSymbolRefAttrGetRootReference(MlirAttribute attr); MlirStringRef mlirSymbolRefAttrGetRootReference(MlirAttribute attr);
/** Returns the stirng reference to the leaf referenced symbol. The data remains /** Returns the string reference to the leaf referenced symbol. The data remains
* live as long as the context in which the attribute lives. */ * live as long as the context in which the attribute lives. */
MlirStringRef mlirSymbolRefAttrGetLeafReference(MlirAttribute attr); MlirStringRef mlirSymbolRefAttrGetLeafReference(MlirAttribute attr);

View File

@ -43,7 +43,7 @@ def Async_ExecuteOp :
(`async.token` or `async.value`). (`async.token` or `async.value`).
`async.execute` operation takes `async.token` dependencies and `async.value` `async.execute` operation takes `async.token` dependencies and `async.value`
operands separatly, and starts execution of the attached body region only operands separately, and starts execution of the attached body region only
when all tokens and values become ready. when all tokens and values become ready.
Example: Example:

View File

@ -913,7 +913,7 @@ class SignlessIntegerAttrBase<I attrValType, string descr> :
let returnType = [{ ::llvm::APInt }]; let returnType = [{ ::llvm::APInt }];
} }
// Base class for signless integer attributes of fixed width that have a // Base class for signless integer attributes of fixed width that have a
// correpsonding C++ type. // corresponding C++ type.
class TypedSignlessIntegerAttrBase<I attrValType, string retType, string descr> class TypedSignlessIntegerAttrBase<I attrValType, string retType, string descr>
: SignlessIntegerAttrBase<attrValType, descr> { : SignlessIntegerAttrBase<attrValType, descr> {
let returnType = retType; let returnType = retType;
@ -942,7 +942,7 @@ class SignedIntegerAttrBase<SI attrValType, string descr> :
let returnType = [{ ::llvm::APInt }]; let returnType = [{ ::llvm::APInt }];
} }
// Base class for signed integer attributes of fixed width that have a // Base class for signed integer attributes of fixed width that have a
// correpsonding C++ type. // corresponding C++ type.
class TypedSignedIntegerAttrBase<SI attrValType, string retType, string descr> class TypedSignedIntegerAttrBase<SI attrValType, string retType, string descr>
: SignedIntegerAttrBase<attrValType, descr> { : SignedIntegerAttrBase<attrValType, descr> {
let returnType = retType; let returnType = retType;
@ -971,7 +971,7 @@ class UnsignedIntegerAttrBase<UI attrValType, string descr> :
let returnType = [{ ::llvm::APInt }]; let returnType = [{ ::llvm::APInt }];
} }
// Base class for unsigned integer attributes of fixed width that have a // Base class for unsigned integer attributes of fixed width that have a
// correpsonding C++ type. // corresponding C++ type.
class TypedUnsignedIntegerAttrBase<UI attrValType, string retType, string descr> class TypedUnsignedIntegerAttrBase<UI attrValType, string retType, string descr>
: UnsignedIntegerAttrBase<attrValType, descr> { : UnsignedIntegerAttrBase<attrValType, descr> {
let returnType = retType; let returnType = retType;

View File

@ -8,7 +8,7 @@
// //
// This file defines the OpReducer class. It defines a variant generator method // This file defines the OpReducer class. It defines a variant generator method
// with the purpose of producing different variants by eliminating a // with the purpose of producing different variants by eliminating a
// parametarizable type of operations from the parent module. // parameterizable type of operations from the parent module.
// //
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//

View File

@ -51,7 +51,7 @@ public:
} }
/// Re-indents by removing the leading whitespace from the first non-empty /// Re-indents by removing the leading whitespace from the first non-empty
/// line from every line of the the string, skipping over empty lines at the /// line from every line of the string, skipping over empty lines at the
/// start. /// start.
raw_indented_ostream &reindent(StringRef str); raw_indented_ostream &reindent(StringRef str);

View File

@ -4,7 +4,7 @@
// RUN: FileCheck %s // RUN: FileCheck %s
// NOTE: This is similar to test-create-mask.mlir, but with a different length, // NOTE: This is similar to test-create-mask.mlir, but with a different length,
// because the v4i1 vector specifially exposed bugs in the LLVM backend. // because the v4i1 vector specifically exposed bugs in the LLVM backend.
func @entry() { func @entry() {
%c0 = constant 0 : index %c0 = constant 0 : index

View File

@ -351,7 +351,7 @@ void Simplex::markEmpty() {
} }
/// Add an inequality to the tableau. If coeffs is c_0, c_1, ... c_n, where n /// Add an inequality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
/// is the curent number of variables, then the corresponding inequality is /// is the current number of variables, then the corresponding inequality is
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} >= 0. /// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} >= 0.
/// ///
/// We add the inequality and mark it as restricted. We then try to make its /// We add the inequality and mark it as restricted. We then try to make its
@ -367,7 +367,7 @@ void Simplex::addInequality(ArrayRef<int64_t> coeffs) {
} }
/// Add an equality to the tableau. If coeffs is c_0, c_1, ... c_n, where n /// Add an equality to the tableau. If coeffs is c_0, c_1, ... c_n, where n
/// is the curent number of variables, then the corresponding equality is /// is the current number of variables, then the corresponding equality is
/// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} == 0. /// c_n + c_0*x_0 + c_1*x_1 + ... + c_{n-1}*x_{n-1} == 0.
/// ///
/// We simply add two opposing inequalities, which force the expression to /// We simply add two opposing inequalities, which force the expression to
@ -383,7 +383,7 @@ void Simplex::addEquality(ArrayRef<int64_t> coeffs) {
unsigned Simplex::numVariables() const { return var.size(); } unsigned Simplex::numVariables() const { return var.size(); }
unsigned Simplex::numConstraints() const { return con.size(); } unsigned Simplex::numConstraints() const { return con.size(); }
/// Return a snapshot of the curent state. This is just the current size of the /// Return a snapshot of the current state. This is just the current size of the
/// undo log. /// undo log.
unsigned Simplex::getSnapshot() const { return undoLog.size(); } unsigned Simplex::getSnapshot() const { return undoLog.size(); }

View File

@ -326,7 +326,7 @@ private:
}; };
/// Blocks are exposed by the C-API as a forward-only linked list. In Python, /// Blocks are exposed by the C-API as a forward-only linked list. In Python,
/// we present them as a more full-featured list-like container but optimzie /// we present them as a more full-featured list-like container but optimize
/// it for forward iteration. Blocks are always owned by a region. /// it for forward iteration. Blocks are always owned by a region.
class PyBlockList { class PyBlockList {
public: public:
@ -424,7 +424,7 @@ private:
/// Operations are exposed by the C-API as a forward-only linked list. In /// Operations are exposed by the C-API as a forward-only linked list. In
/// Python, we present them as a more full-featured list-like container but /// Python, we present them as a more full-featured list-like container but
/// optimzie it for forward iteration. Iterable operations are always owned /// optimize it for forward iteration. Iterable operations are always owned
/// by a block. /// by a block.
class PyOperationList { class PyOperationList {
public: public:

View File

@ -34,7 +34,7 @@ public:
ConversionPatternRewriter &rewriter) const override; ConversionPatternRewriter &rewriter) const override;
}; };
/// Pattern lowering subgoup size/id to loading SPIR-V invocation /// Pattern lowering subgroup size/id to loading SPIR-V invocation
/// builtin variables. /// builtin variables.
template <typename SourceOp, spirv::BuiltIn builtin> template <typename SourceOp, spirv::BuiltIn builtin>
class SingleDimLaunchConfigConversion : public SPIRVOpLowering<SourceOp> { class SingleDimLaunchConfigConversion : public SPIRVOpLowering<SourceOp> {

View File

@ -108,7 +108,7 @@ static Value createFPConstant(Location loc, Type srcType, Type dstType,
loc, dstType, rewriter.getFloatAttr(floatType, value)); loc, dstType, rewriter.getFloatAttr(floatType, value));
} }
/// Utility function for bitfiled ops: /// Utility function for bitfield ops:
/// - `BitFieldInsert` /// - `BitFieldInsert`
/// - `BitFieldSExtract` /// - `BitFieldSExtract`
/// - `BitFieldUExtract` /// - `BitFieldUExtract`
@ -163,7 +163,7 @@ static Value optionallyBroadcast(Location loc, Value value, Type srcType,
return value; return value;
} }
/// Utility function for bitfiled ops: `BitFieldInsert`, `BitFieldSExtract` and /// Utility function for bitfield ops: `BitFieldInsert`, `BitFieldSExtract` and
/// `BitFieldUExtract`. /// `BitFieldUExtract`.
/// Broadcast `Offset` and `Count` to match the type of `Base`. If `Base` is of /// Broadcast `Offset` and `Count` to match the type of `Base`. If `Base` is of
/// a vector type, construct a vector that has: /// a vector type, construct a vector that has:
@ -971,8 +971,8 @@ public:
Location loc = loopOp.getLoc(); Location loc = loopOp.getLoc();
// Split the current block after `spv.loop`. The remaing ops will be used in // Split the current block after `spv.loop`. The remaining ops will be used
// `endBlock`. // in `endBlock`.
Block *currentBlock = rewriter.getBlock(); Block *currentBlock = rewriter.getBlock();
auto position = Block::iterator(loopOp); auto position = Block::iterator(loopOp);
Block *endBlock = rewriter.splitBlock(currentBlock, position); Block *endBlock = rewriter.splitBlock(currentBlock, position);
@ -1028,7 +1028,7 @@ public:
Location loc = op.getLoc(); Location loc = op.getLoc();
// Split the current block after `spv.selection`. The remaing ops will be // Split the current block after `spv.selection`. The remaining ops will be
// used in `continueBlock`. // used in `continueBlock`.
auto *currentBlock = rewriter.getInsertionBlock(); auto *currentBlock = rewriter.getInsertionBlock();
rewriter.setInsertionPointAfter(op); rewriter.setInsertionPointAfter(op);
@ -1387,7 +1387,7 @@ void mlir::populateSPIRVToLLVMConversionPatterns(
// Entry points and execution mode // Entry points and execution mode
// Module generated from SPIR-V could have other "internal" functions, so // Module generated from SPIR-V could have other "internal" functions, so
// having entry point and execution mode metadat can be useful. For now, // having entry point and execution mode metadata can be useful. For now,
// simply remove them. // simply remove them.
// TODO: Support EntryPoint/ExecutionMode properly. // TODO: Support EntryPoint/ExecutionMode properly.
ErasePattern<spirv::EntryPointOp>, ErasePattern<spirv::ExecutionModeOp>, ErasePattern<spirv::EntryPointOp>, ErasePattern<spirv::ExecutionModeOp>,

View File

@ -132,7 +132,7 @@ LogicalResult BroadcastOpConverter::matchAndRewrite(
}, },
[&](OpBuilder &b, Location loc) { [&](OpBuilder &b, Location loc) {
// The broadcasting logic is: // The broadcasting logic is:
// - if one extent (here we arbitrariliy choose the extent from // - if one extent (here we arbitrarily choose the extent from
// the greater-rank operand) is equal to 1, then take the extent // the greater-rank operand) is equal to 1, then take the extent
// from the other operand // from the other operand
// - otherwise, take the extent as-is. // - otherwise, take the extent as-is.

View File

@ -249,7 +249,7 @@ SmallVector<Type, 2> LLVMTypeConverter::convertUnrankedMemRefSignature() {
LLVM::LLVMType LLVMTypeConverter::convertFunctionSignature( LLVM::LLVMType LLVMTypeConverter::convertFunctionSignature(
FunctionType funcTy, bool isVariadic, FunctionType funcTy, bool isVariadic,
LLVMTypeConverter::SignatureConversion &result) { LLVMTypeConverter::SignatureConversion &result) {
// Select the argument converter depending on the calling convetion. // Select the argument converter depending on the calling convention.
auto funcArgConverter = options.useBarePtrCallConv auto funcArgConverter = options.useBarePtrCallConv
? barePtrFuncArgTypeConverter ? barePtrFuncArgTypeConverter
: structFuncArgTypeConverter; : structFuncArgTypeConverter;

View File

@ -414,7 +414,7 @@ public:
ConversionPatternRewriter &rewriter) const override; ConversionPatternRewriter &rewriter) const override;
}; };
/// Converts integer compare operation on i1 type opearnds to SPIR-V ops. /// Converts integer compare operation on i1 type operands to SPIR-V ops.
class BoolCmpIOpPattern final : public SPIRVOpLowering<CmpIOp> { class BoolCmpIOpPattern final : public SPIRVOpLowering<CmpIOp> {
public: public:
using SPIRVOpLowering<CmpIOp>::SPIRVOpLowering; using SPIRVOpLowering<CmpIOp>::SPIRVOpLowering;

View File

@ -100,7 +100,7 @@ public:
return failure(); return failure();
// Note that the dataPtr starts at the offset address specified by // Note that the dataPtr starts at the offset address specified by
// indices, so no need to calculat offset size in bytes again in // indices, so no need to calculate offset size in bytes again in
// the MUBUF instruction. // the MUBUF instruction.
Value dataPtr = getDataPtr(loc, memRefType, adaptor.memref(), Value dataPtr = getDataPtr(loc, memRefType, adaptor.memref(),
adaptor.indices(), rewriter); adaptor.indices(), rewriter);

View File

@ -84,7 +84,7 @@ extractBeneficiaryOps(Operation *op,
return false; return false;
for (Value operand : op->getOperands()) { for (Value operand : op->getOperands()) {
// It is already visisble in the kernel, keep going. // It is already visible in the kernel, keep going.
if (availableValues.count(operand)) if (availableValues.count(operand))
continue; continue;
// Else check whether it can be made available via sinking or already is a // Else check whether it can be made available via sinking or already is a

View File

@ -79,7 +79,7 @@ public:
bool isIdentified() const { return identified; } bool isIdentified() const { return identified; }
bool isPacked() const { bool isPacked() const {
assert(!isIdentified() && assert(!isIdentified() &&
"'packed' bit is not part of the key for identified stucts"); "'packed' bit is not part of the key for identified structs");
return packed; return packed;
} }
bool isOpaque() const { bool isOpaque() const {
@ -196,7 +196,7 @@ public:
/// Constructs the storage from the given key. This sets up the uniquing key /// Constructs the storage from the given key. This sets up the uniquing key
/// components and optionally the mutable component if they construction key /// components and optionally the mutable component if they construction key
/// has the relevant information. In the latter case, the struct is considered /// has the relevant information. In the latter case, the struct is considered
/// as initalized and can no longer be mutated. /// as initialized and can no longer be mutated.
LLVMStructTypeStorage(const KeyTy &key) { LLVMStructTypeStorage(const KeyTy &key) {
if (!key.isIdentified()) { if (!key.isIdentified()) {
ArrayRef<LLVMType> types = key.getTypeList(); ArrayRef<LLVMType> types = key.getTypeList();

View File

@ -711,10 +711,10 @@ static SmallVector<SmallVector<AffineExpr, 2>, 2>
convertReassociationIndicesToMaps( convertReassociationIndicesToMaps(
OpBuilder &b, ArrayRef<ReassociationIndices> reassociationIndices) { OpBuilder &b, ArrayRef<ReassociationIndices> reassociationIndices) {
SmallVector<SmallVector<AffineExpr, 2>, 2> reassociationMaps; SmallVector<SmallVector<AffineExpr, 2>, 2> reassociationMaps;
for (const auto &indicies : reassociationIndices) { for (const auto &indices : reassociationIndices) {
SmallVector<AffineExpr, 2> reassociationMap; SmallVector<AffineExpr, 2> reassociationMap;
reassociationMap.reserve(indicies.size()); reassociationMap.reserve(indices.size());
for (int64_t index : indicies) for (int64_t index : indices)
reassociationMap.push_back(b.getAffineDimExpr(index)); reassociationMap.push_back(b.getAffineDimExpr(index));
reassociationMaps.push_back(std::move(reassociationMap)); reassociationMaps.push_back(std::move(reassociationMap));
} }

View File

@ -147,7 +147,7 @@ static void generateFusedTensorOpRegion(PatternRewriter &rewriter,
} }
// If consumer is an indexed_generic op, map the indices to the block // If consumer is an indexed_generic op, map the indices to the block
// arguments directly. Otherwise, add the same type of arugment and map to // arguments directly. Otherwise, add the same type of argument and map to
// it. // it.
if (consumerArg.index() < numConsumerIndices) { if (consumerArg.index() < numConsumerIndices) {
mapper.map(consumerArg.value(), mapper.map(consumerArg.value(),

View File

@ -80,7 +80,7 @@ void mlir::linalg::hoistViewAllocOps(FuncOp func) {
} }
} }
/// Return true if we can prove that the transfer operations access dijoint /// Return true if we can prove that the transfer operations access disjoint
/// memory. /// memory.
static bool isDisjoint(VectorTransferOpInterface transferA, static bool isDisjoint(VectorTransferOpInterface transferA,
VectorTransferOpInterface transferB) { VectorTransferOpInterface transferB) {

View File

@ -161,7 +161,7 @@ struct LinalgOpInstancePromotionOptions {
CopyCallbackFn copyInFn; CopyCallbackFn copyInFn;
CopyCallbackFn copyOutFn; CopyCallbackFn copyOutFn;
/// Allow the use of dynamicaly-sized buffers. /// Allow the use of dynamically-sized buffers.
bool dynamicBuffers; bool dynamicBuffers;
/// Alignment of promoted buffer. /// Alignment of promoted buffer.
Optional<unsigned> alignment; Optional<unsigned> alignment;

View File

@ -302,7 +302,7 @@ static Type parseArrayType(SPIRVDialect const &dialect,
} }
// cooperative-matrix-type ::= `!spv.coopmatrix` `<` element-type ',' scope ',' // cooperative-matrix-type ::= `!spv.coopmatrix` `<` element-type ',' scope ','
// rows ',' coloumns>` // rows ',' columns>`
static Type parseCooperativeMatrixType(SPIRVDialect const &dialect, static Type parseCooperativeMatrixType(SPIRVDialect const &dialect,
DialectAsmParser &parser) { DialectAsmParser &parser) {
if (parser.parseLess()) if (parser.parseLess())
@ -621,7 +621,7 @@ static Type parseStructType(SPIRVDialect const &dialect,
StringRef identifier; StringRef identifier;
// Check if this is an idenitifed struct type. // Check if this is an identified struct type.
if (succeeded(parser.parseOptionalKeyword(&identifier))) { if (succeeded(parser.parseOptionalKeyword(&identifier))) {
// Check if this is a possible recursive reference. // Check if this is a possible recursive reference.
if (succeeded(parser.parseOptionalGreater())) { if (succeeded(parser.parseOptionalGreater())) {

View File

@ -217,9 +217,9 @@ static ParseResult parseMemoryAccessAttributes(OpAsmParser &parser,
} }
// TODO Make sure to merge this and the previous function into one template // TODO Make sure to merge this and the previous function into one template
// parameterized by memroy access attribute name and alignment. Doing so now // parameterized by memory access attribute name and alignment. Doing so now
// results in VS2017 in producing an internal error (at the call site) that's // results in VS2017 in producing an internal error (at the call site) that's
// not detailed enough to understand what is happenning. // not detailed enough to understand what is happening.
static ParseResult parseSourceMemoryAccessAttributes(OpAsmParser &parser, static ParseResult parseSourceMemoryAccessAttributes(OpAsmParser &parser,
OperationState &state) { OperationState &state) {
// Parse an optional list of attributes staring with '[' // Parse an optional list of attributes staring with '['
@ -274,9 +274,9 @@ static void printMemoryAccessAttribute(
} }
// TODO Make sure to merge this and the previous function into one template // TODO Make sure to merge this and the previous function into one template
// parameterized by memroy access attribute name and alignment. Doing so now // parameterized by memory access attribute name and alignment. Doing so now
// results in VS2017 in producing an internal error (at the call site) that's // results in VS2017 in producing an internal error (at the call site) that's
// not detailed enough to understand what is happenning. // not detailed enough to understand what is happening.
template <typename MemoryOpTy> template <typename MemoryOpTy>
static void printSourceMemoryAccessAttribute( static void printSourceMemoryAccessAttribute(
MemoryOpTy memoryOp, OpAsmPrinter &printer, MemoryOpTy memoryOp, OpAsmPrinter &printer,
@ -393,9 +393,9 @@ static LogicalResult verifyMemoryAccessAttribute(MemoryOpTy memoryOp) {
} }
// TODO Make sure to merge this and the previous function into one template // TODO Make sure to merge this and the previous function into one template
// parameterized by memroy access attribute name and alignment. Doing so now // parameterized by memory access attribute name and alignment. Doing so now
// results in VS2017 in producing an internal error (at the call site) that's // results in VS2017 in producing an internal error (at the call site) that's
// not detailed enough to understand what is happenning. // not detailed enough to understand what is happening.
template <typename MemoryOpTy> template <typename MemoryOpTy>
static LogicalResult verifySourceMemoryAccessAttribute(MemoryOpTy memoryOp) { static LogicalResult verifySourceMemoryAccessAttribute(MemoryOpTy memoryOp) {
// ODS checks for attributes values. Just need to verify that if the // ODS checks for attributes values. Just need to verify that if the

View File

@ -810,7 +810,7 @@ struct spirv::detail::StructTypeStorage : public TypeStorage {
std::tuple<StringRef, ArrayRef<Type>, ArrayRef<StructType::OffsetInfo>, std::tuple<StringRef, ArrayRef<Type>, ArrayRef<StructType::OffsetInfo>,
ArrayRef<StructType::MemberDecorationInfo>>; ArrayRef<StructType::MemberDecorationInfo>>;
/// For idetified structs, return true if the given key contains the same /// For identified structs, return true if the given key contains the same
/// identifier. /// identifier.
/// ///
/// For literal structs, return true if the given key contains a matching list /// For literal structs, return true if the given key contains a matching list

View File

@ -1033,7 +1033,7 @@ bool Serializer::isInterfaceStructPtrType(Type type) const {
LogicalResult Serializer::processType(Location loc, Type type, LogicalResult Serializer::processType(Location loc, Type type,
uint32_t &typeID) { uint32_t &typeID) {
// Maintains a set of names for nested identified struct types. This is used // Maintains a set of names for nested identified struct types. This is used
// to properly seialize resursive references. // to properly serialize resursive references.
llvm::SetVector<StringRef> serializationCtx; llvm::SetVector<StringRef> serializationCtx;
return processTypeImpl(loc, type, typeID, serializationCtx); return processTypeImpl(loc, type, typeID, serializationCtx);
} }
@ -1170,7 +1170,7 @@ LogicalResult Serializer::prepareBasicType(
spirv::Opcode::OpTypeForwardPointer, spirv::Opcode::OpTypeForwardPointer,
forwardPtrOperands); forwardPtrOperands);
// 2. Find the the pointee (enclosing) struct. // 2. Find the pointee (enclosing) struct.
auto structType = spirv::StructType::getIdentified( auto structType = spirv::StructType::getIdentified(
module.getContext(), pointeeStruct.getIdentifier()); module.getContext(), pointeeStruct.getIdentifier());

View File

@ -878,7 +878,7 @@ static Value foldExtractFromShapeCast(ExtractOp extractOp) {
} }
int64_t position = linearize(extractedPos, strides); int64_t position = linearize(extractedPos, strides);
// Then extract the strides assoociated to the shapeCast op vector source and // Then extract the strides associated to the shapeCast op vector source and
// delinearize the position using those strides. // delinearize the position using those strides.
SmallVector<int64_t, 4> newStrides; SmallVector<int64_t, 4> newStrides;
int64_t numDimension = int64_t numDimension =

View File

@ -2178,7 +2178,7 @@ LogicalResult mlir::vector::splitFullAndPartialTransferPrecondition(
/// 2. else return a new MemRefType obtained by iterating over the shape and /// 2. else return a new MemRefType obtained by iterating over the shape and
/// strides and: /// strides and:
/// a. keeping the ones that are static and equal across `aT` and `bT`. /// a. keeping the ones that are static and equal across `aT` and `bT`.
/// b. using a dynamic shape and/or stride for the dimeniosns that don't /// b. using a dynamic shape and/or stride for the dimensions that don't
/// agree. /// agree.
static MemRefType getCastCompatibleMemRefType(MemRefType aT, MemRefType bT) { static MemRefType getCastCompatibleMemRefType(MemRefType aT, MemRefType bT) {
if (MemRefCastOp::areCastCompatible(aT, bT)) if (MemRefCastOp::areCastCompatible(aT, bT))

View File

@ -292,8 +292,8 @@ public:
/// operations. /// operations.
llvm::StringMap<AbstractOperation> registeredOperations; llvm::StringMap<AbstractOperation> registeredOperations;
/// Identifers are uniqued by string value and use the internal string set for /// Identifiers are uniqued by string value and use the internal string set
/// storage. /// for storage.
llvm::StringSet<llvm::BumpPtrAllocator &> identifiers; llvm::StringSet<llvm::BumpPtrAllocator &> identifiers;
/// A thread local cache of identifiers to reduce lock contention. /// A thread local cache of identifiers to reduce lock contention.
ThreadLocalCache<llvm::StringMap<llvm::StringMapEntry<llvm::NoneType> *>> ThreadLocalCache<llvm::StringMap<llvm::StringMapEntry<llvm::NoneType> *>>

View File

@ -140,7 +140,7 @@ void PatternRewriter::mergeBlockBefore(Block *source, Operation *op,
assert(source->hasNoSuccessors() && assert(source->hasNoSuccessors() &&
"expected 'source' to have no successors"); "expected 'source' to have no successors");
// Split the block containing 'op' into two, one containg all operations // Split the block containing 'op' into two, one containing all operations
// before 'op' (prologue) and another (epilogue) containing 'op' and all // before 'op' (prologue) and another (epilogue) containing 'op' and all
// operations after it. // operations after it.
Block *prologue = op->getBlock(); Block *prologue = op->getBlock();

View File

@ -167,7 +167,7 @@ LogicalResult detail::verifyTypesAlongControlFlowEdges(Operation *op) {
for (unsigned regionNo : llvm::seq(0U, op->getNumRegions())) { for (unsigned regionNo : llvm::seq(0U, op->getNumRegions())) {
Region &region = op->getRegion(regionNo); Region &region = op->getRegion(regionNo);
// Since the interface cannnot distinguish between different ReturnLike // Since the interface cannot distinguish between different ReturnLike
// ops within the region branching to different successors, all ReturnLike // ops within the region branching to different successors, all ReturnLike
// ops in this region should have the same operand types. We will then use // ops in this region should have the same operand types. We will then use
// one of them as the representative for type matching. // one of them as the representative for type matching.

View File

@ -14,5 +14,5 @@ using namespace mlir;
// VectorUnroll Interfaces // VectorUnroll Interfaces
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//
/// Include the definitions of the VectorUntoll interfaces. /// Include the definitions of the VectorUnroll interfaces.
#include "mlir/Interfaces/VectorInterfaces.cpp.inc" #include "mlir/Interfaces/VectorInterfaces.cpp.inc"

View File

@ -9,7 +9,7 @@
// This file defines the Tester class used in the MLIR Reduce tool. // This file defines the Tester class used in the MLIR Reduce tool.
// //
// A Tester object is passed as an argument to the reduction passes and it is // A Tester object is passed as an argument to the reduction passes and it is
// used to run the interestigness testing script on the different generated // used to run the interestingness testing script on the different generated
// reduced variants of the test case. // reduced variants of the test case.
// //
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//

View File

@ -5,7 +5,7 @@
# component dependencies, LLVM_LINK_LLVM_DYLIB tends to introduce a # component dependencies, LLVM_LINK_LLVM_DYLIB tends to introduce a
# dependence on libLLVM.so) However, it must also be linkable against # dependence on libLLVM.so) However, it must also be linkable against
# libMLIR.so in some contexts (see unittests/Tablegen, for instance, which # libMLIR.so in some contexts (see unittests/Tablegen, for instance, which
# has a dependance on MLIRIR, which must depend on libLLVM.so). This works # has a dependence on MLIRIR, which must depend on libLLVM.so). This works
# in this special case because this library is static. # in this special case because this library is static.
llvm_add_library(MLIRTableGen STATIC llvm_add_library(MLIRTableGen STATIC

View File

@ -487,7 +487,7 @@ ModuleTranslation::convertOmpOperation(Operation &opInst,
return success(); return success();
}) })
.Case([&](omp::FlushOp) { .Case([&](omp::FlushOp) {
// No support in Openmp runtime funciton (__kmpc_flush) to accept // No support in Openmp runtime function (__kmpc_flush) to accept
// the argument list. // the argument list.
// OpenMP standard states the following: // OpenMP standard states the following:
// "An implementation may implement a flush with a list by ignoring // "An implementation may implement a flush with a list by ignoring

View File

@ -148,7 +148,7 @@ public:
} }
/// This method returns the mapping values list. The unknown result values /// This method returns the mapping values list. The unknown result values
/// that only their indicies are available are replaced with their values. /// that only their indices are available are replaced with their values.
void getMappingValues(ValueRange valuesToReplaceIndices, void getMappingValues(ValueRange valuesToReplaceIndices,
SmallVectorImpl<Value> &values) { SmallVectorImpl<Value> &values) {
// Append available values to the list. // Append available values to the list.

View File

@ -9,7 +9,7 @@
// This transformation pass performs a sparse conditional constant propagation // This transformation pass performs a sparse conditional constant propagation
// in MLIR. It identifies values known to be constant, propagates that // in MLIR. It identifies values known to be constant, propagates that
// information throughout the IR, and replaces them. This is done with an // information throughout the IR, and replaces them. This is done with an
// optimisitic dataflow analysis that assumes that all values are constant until // optimistic dataflow analysis that assumes that all values are constant until
// proven otherwise. // proven otherwise.
// //
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//
@ -27,7 +27,7 @@ using namespace mlir;
namespace { namespace {
/// This class represents a single lattice value. A lattive value corresponds to /// This class represents a single lattice value. A lattive value corresponds to
/// the various different states that a value in the SCCP dataflow anaylsis can /// the various different states that a value in the SCCP dataflow analysis can
/// take. See 'Kind' below for more details on the different states a value can /// take. See 'Kind' below for more details on the different states a value can
/// take. /// take.
class LatticeValue { class LatticeValue {

View File

@ -522,7 +522,7 @@ void ArgConverter::insertConversion(Block *newBlock,
} }
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//
// Rewriter and Transation State // Rewriter and Translation State
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//
namespace { namespace {
/// This class contains a snapshot of the current conversion rewriter state. /// This class contains a snapshot of the current conversion rewriter state.

View File

@ -694,13 +694,13 @@ static void setInterTileBoundsParametric(OpBuilder &b, AffineForOp origLoop,
AffineForOp newLoop, Value tileSize) { AffineForOp newLoop, Value tileSize) {
OperandRange newLbOperands = origLoop.getLowerBoundOperands(); OperandRange newLbOperands = origLoop.getLowerBoundOperands();
// The lower bounds for inter-tile loops are same as the correspondig lower // The lower bounds for inter-tile loops are same as the corresponding lower
// bounds of original loops. // bounds of original loops.
newLoop.setLowerBound(newLbOperands, origLoop.getLowerBoundMap()); newLoop.setLowerBound(newLbOperands, origLoop.getLowerBoundMap());
// The new upper bound map for inter-tile loops, assuming constant lower // The new upper bound map for inter-tile loops, assuming constant lower
// bounds, are now originalLowerBound + ceildiv((orignalUpperBound - // bounds, are now originalLowerBound + ceildiv((originalUpperBound -
// originalLowerBound), tiling paramter); where tiling parameter is the // originalLowerBound), tiling parameter); where tiling parameter is the
// respective tile size for that loop. For e.g. if the original ubmap was // respective tile size for that loop. For e.g. if the original ubmap was
// ()->(1024), the new map will be // ()->(1024), the new map will be
// ()[s0]->(ceildiv((1024 -lb) % s0)), where s0 is the tiling parameter. // ()[s0]->(ceildiv((1024 -lb) % s0)), where s0 is the tiling parameter.
@ -757,7 +757,7 @@ static void setInterTileBoundsParametric(OpBuilder &b, AffineForOp origLoop,
// ubmap has only one result expression. For e.g. // ubmap has only one result expression. For e.g.
// affine.for %i = 5 to %ub // affine.for %i = 5 to %ub
// //
// A symbol operand is added which represents the tiling paramater. The // A symbol operand is added which represents the tiling parameter. The
// new loop bounds here will be like ()[s0, s1] -> ((s0 - 5) ceildiv s1 + 5) // new loop bounds here will be like ()[s0, s1] -> ((s0 - 5) ceildiv s1 + 5)
// where 's0' is the original upper bound and 's1' is the tiling // where 's0' is the original upper bound and 's1' is the tiling
// parameter. 2.) When ubMap has more than one result expression. For e.g. // parameter. 2.) When ubMap has more than one result expression. For e.g.

View File

@ -981,7 +981,7 @@ int printAffineExpr(MlirContext ctx) {
!mlirAffineExprIsFunctionOfDim(affineCeilDivExpr, 5)) !mlirAffineExprIsFunctionOfDim(affineCeilDivExpr, 5))
return 8; return 8;
// Tests 'IsA' methods of affine binary operaion expression. // Tests 'IsA' methods of affine binary operation expression.
if (!mlirAffineExprIsAAdd(affineAddExpr)) if (!mlirAffineExprIsAAdd(affineAddExpr))
return 9; return 9;

View File

@ -111,9 +111,9 @@ func @caller(%arg0: tensor<5xf32>) -> tensor<5xf32> {
// ----- // -----
// Test case: Testing BufferAssginmnetCallOpConverter to see if it matches with the // Test case: Testing BufferAssignmentCallOpConverter to see if it matches with the
// signature of the new signature of the callee function when there are tuple typed // signature of the new signature of the callee function when there are tuple typed
// args and results. BufferAssginmentTypeConverter is set to flatten tuple typed // args and results. BufferAssignmentTypeConverter is set to flatten tuple typed
// arguments. The tuple typed values should be decomposed and composed using // arguments. The tuple typed values should be decomposed and composed using
// get_tuple_element and make_tuple operations of test dialect. Tensor types are // get_tuple_element and make_tuple operations of test dialect. Tensor types are
// converted to Memref. Memref typed function results remain as function results. // converted to Memref. Memref typed function results remain as function results.
@ -158,10 +158,10 @@ func @caller(%arg0: tuple<tensor<2xf32>,i1, tensor<5xf32>>) -> tuple<tensor<2xf3
// ----- // -----
// Test case: Testing BufferAssginmnetFuncOpConverter and // Test case: Testing BufferAssignmentFuncOpConverter and
// BufferAssginmentReturnOpConverter to see if the return operation matches with // BufferAssignmentReturnOpConverter to see if the return operation matches with
// the new function signature when there are tuple typed args and results. // the new function signature when there are tuple typed args and results.
// BufferAssginmentTypeConverter is set to flatten tuple typed arguments. The tuple // BufferAssignmentTypeConverter is set to flatten tuple typed arguments. The tuple
// typed values should be decomposed and composed using get_tuple_element and // typed values should be decomposed and composed using get_tuple_element and
// make_tuple operations of test dialect. Tensor types are converted to Memref. // make_tuple operations of test dialect. Tensor types are converted to Memref.
// Memref typed function results remain as function results. // Memref typed function results remain as function results.

View File

@ -302,9 +302,9 @@ func @func_with_unranked_arg(%arg0: tensor<*xf32>) {
// ----- // -----
// Test case: Testing BufferAssginmnetCallOpConverter to see if it matches with the // Test case: Testing BufferAssignmentCallOpConverter to see if it matches with the
// signature of the new signature of the callee function when there are tuple typed // signature of the new signature of the callee function when there are tuple typed
// args and results. BufferAssginmentTypeConverter is set to flatten tuple typed // args and results. BufferAssignmentTypeConverter is set to flatten tuple typed
// arguments. The tuple typed values should be decomposed and composed using // arguments. The tuple typed values should be decomposed and composed using
// get_tuple_element and make_tuple operations of test dialect. Tensor types are // get_tuple_element and make_tuple operations of test dialect. Tensor types are
// converted to Memref. Memref typed function results are appended to the function // converted to Memref. Memref typed function results are appended to the function
@ -359,10 +359,10 @@ func @caller(%arg0: tuple<tensor<2xf32>,i1, tensor<5xf32>>) -> tuple<tensor<2xf3
// ----- // -----
// Test case: Testing BufferAssginmnetFuncOpConverter and // Test case: Testing BufferAssignmentFuncOpConverter and
// BufferAssginmentReturnOpConverter to see if the return operation matches with // BufferAssignmentReturnOpConverter to see if the return operation matches with
// the new function signature when there are tuple typed args and results. // the new function signature when there are tuple typed args and results.
// BufferAssginmentTypeConverter is set to flatten tuple typed arguments. The tuple // BufferAssignmentTypeConverter is set to flatten tuple typed arguments. The tuple
// typed values should be decomposed and composed using get_tuple_element and // typed values should be decomposed and composed using get_tuple_element and
// make_tuple operations of test dialect. Tensor types are converted to Memref. // make_tuple operations of test dialect. Tensor types are converted to Memref.
// Memref typed function results are appended to the function arguments list. // Memref typed function results are appended to the function arguments list.

View File

@ -26,7 +26,7 @@ func @test_norm(%arg0 : memref<1x16x14x14xf32, #map0>) -> () {
return return
} }
// Same test with op_nonnorm, with maps in the argmentets and the operations in the function. // Same test with op_nonnorm, with maps in the arguments and the operations in the function.
// CHECK-LABEL: test_nonnorm // CHECK-LABEL: test_nonnorm
// CHECK-SAME: (%[[ARG0:[a-z0-9]*]]: memref<1x16x14x14xf32, #map0>) // CHECK-SAME: (%[[ARG0:[a-z0-9]*]]: memref<1x16x14x14xf32, #map0>)

View File

@ -90,7 +90,7 @@ def E_IntegerType : TestType<"Integer"> {
let parameters = ( let parameters = (
ins ins
"SignednessSemantics":$signedness, "SignednessSemantics":$signedness,
TypeParameter<"unsigned", "Bitwdith of integer">:$width TypeParameter<"unsigned", "Bitwidth of integer">:$width
); );
// DECL-LABEL: IntegerType: public ::mlir::Type // DECL-LABEL: IntegerType: public ::mlir::Type

View File

@ -8,7 +8,7 @@
// //
// This file defines the OpReducer class. It defines a variant generator method // This file defines the OpReducer class. It defines a variant generator method
// with the purpose of producing different variants by eliminating a // with the purpose of producing different variants by eliminating a
// parametarizable type of operations from the parent module. // parameterizable type of operations from the parent module.
// //
//===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===//
#include "mlir/Reducer/Passes/OpReducer.h" #include "mlir/Reducer/Passes/OpReducer.h"

View File

@ -39,7 +39,7 @@ void ReductionNode::measureAndTest(const Tester &test) {
SmallString<128> filepath; SmallString<128> filepath;
int fd; int fd;
// Print module to temprary file. // Print module to temporary file.
std::error_code ec = std::error_code ec =
llvm::sys::fs::createTemporaryFile("mlir-reduce", "mlir", fd, filepath); llvm::sys::fs::createTemporaryFile("mlir-reduce", "mlir", fd, filepath);

View File

@ -26,7 +26,7 @@ void ReductionTreeUtils::updateGoldenModule(ModuleOp &golden,
reduced.getBody()->getOperations()); reduced.getBody()->getOperations());
} }
/// Update the the smallest node traversed so far in the reduction tree and /// Update the smallest node traversed so far in the reduction tree and
/// print the debugging information for the currNode being traversed. /// print the debugging information for the currNode being traversed.
void ReductionTreeUtils::updateSmallestNode(ReductionNode *currNode, void ReductionTreeUtils::updateSmallestNode(ReductionNode *currNode,
ReductionNode *&smallestNode, ReductionNode *&smallestNode,
@ -95,7 +95,7 @@ static std::vector<std::tuple<int, int>> getRanges(std::vector<bool> tSpace,
} }
/// Create the specified number of variants by applying the transform method /// Create the specified number of variants by applying the transform method
/// to different ranges of indices in the parent module. The isDeletion bolean /// to different ranges of indices in the parent module. The isDeletion boolean
/// specifies if the transformation is the deletion of indices. /// specifies if the transformation is the deletion of indices.
void ReductionTreeUtils::createVariants( void ReductionTreeUtils::createVariants(
ReductionNode *parent, const Tester &test, int numVariants, ReductionNode *parent, const Tester &test, int numVariants,

View File

@ -325,7 +325,7 @@ private:
// attribute (the generated function call returns an Attribute); // attribute (the generated function call returns an Attribute);
// - operandGet corresponds to the name of the function with which to retrieve // - operandGet corresponds to the name of the function with which to retrieve
// an operand (the generated function call returns an OperandRange); // an operand (the generated function call returns an OperandRange);
// - reultGet corresponds to the name of the function to get an result (the // - resultGet corresponds to the name of the function to get an result (the
// generated function call returns a ValueRange); // generated function call returns a ValueRange);
static void populateSubstitutions(const Operator &op, const char *attrGet, static void populateSubstitutions(const Operator &op, const char *attrGet,
const char *operandGet, const char *resultGet, const char *operandGet, const char *resultGet,
@ -570,7 +570,7 @@ void OpEmitter::genAttrGetters() {
PrintWarning( PrintWarning(
op.getLoc(), op.getLoc(),
formatv( formatv(
"op has non-materialzable derived attributes '{0}', skipping", "op has non-materializable derived attributes '{0}', skipping",
os.str())); os.str()));
body << formatv(" emitOpError(\"op has non-materializable derived " body << formatv(" emitOpError(\"op has non-materializable derived "
"attributes '{0}'\");\n", "attributes '{0}'\");\n",
@ -965,7 +965,7 @@ void OpEmitter::genSeparateArgParamBuilder() {
llvm_unreachable("unhandled TypeParamKind"); llvm_unreachable("unhandled TypeParamKind");
}; };
// Some of the build methods generated here may be amiguous, but TableGen's // Some of the build methods generated here may be ambiguous, but TableGen's
// ambiguous function detection will elide those ones. // ambiguous function detection will elide those ones.
for (auto attrType : attrBuilderType) { for (auto attrType : attrBuilderType) {
emit(attrType, TypeParamKind::Separate, /*inferType=*/false); emit(attrType, TypeParamKind::Separate, /*inferType=*/false);