Fix a few spelling mistakes in comments. NFCI.

This commit is contained in:
Simon Pilgrim 2020-02-02 18:15:44 +00:00
parent 9288bb8d20
commit 5d86ac82a6
1 changed files with 7 additions and 7 deletions

View File

@ -164,7 +164,7 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
// If we don't have cmpxchg8b(meaing this is a 386/486), limit atomic size to // If we don't have cmpxchg8b(meaing this is a 386/486), limit atomic size to
// 32 bits so the AtomicExpandPass will expand it so we don't need cmpxchg8b. // 32 bits so the AtomicExpandPass will expand it so we don't need cmpxchg8b.
// FIXME: Should we be limitting the atomic size on other configs? Default is // FIXME: Should we be limiting the atomic size on other configs? Default is
// 1024. // 1024.
if (!Subtarget.hasCmpxchg8b()) if (!Subtarget.hasCmpxchg8b())
setMaxAtomicSizeInBitsSupported(32); setMaxAtomicSizeInBitsSupported(32);
@ -4335,7 +4335,7 @@ X86TargetLowering::LowerCall(TargetLowering::CallLoweringInfo &CLI,
// (within module) calls are supported at the moment. // (within module) calls are supported at the moment.
// To keep the stack aligned according to platform abi the function // To keep the stack aligned according to platform abi the function
// GetAlignedArgumentStackSize ensures that argument delta is always multiples // GetAlignedArgumentStackSize ensures that argument delta is always multiples
// of stack alignment. (Dynamic linkers need this - darwin's dyld for example) // of stack alignment. (Dynamic linkers need this - Darwin's dyld for example)
// If a tail called function callee has more arguments than the caller the // If a tail called function callee has more arguments than the caller the
// caller needs to make sure that there is room to move the RETADDR to. This is // caller needs to make sure that there is room to move the RETADDR to. This is
// achieved by reserving an area the size of the argument delta right after the // achieved by reserving an area the size of the argument delta right after the
@ -5723,7 +5723,7 @@ static SDValue widenSubVector(SDValue Vec, bool ZeroNewElements,
return widenSubVector(VT, Vec, ZeroNewElements, Subtarget, DAG, dl); return widenSubVector(VT, Vec, ZeroNewElements, Subtarget, DAG, dl);
} }
// Helper function to collect subvector ops that are concated together, // Helper function to collect subvector ops that are concatenated together,
// either by ISD::CONCAT_VECTORS or a ISD::INSERT_SUBVECTOR series. // either by ISD::CONCAT_VECTORS or a ISD::INSERT_SUBVECTOR series.
// The subvectors in Ops are guaranteed to be the same type. // The subvectors in Ops are guaranteed to be the same type.
static bool collectConcatOps(SDNode *N, SmallVectorImpl<SDValue> &Ops) { static bool collectConcatOps(SDNode *N, SmallVectorImpl<SDValue> &Ops) {
@ -5898,7 +5898,7 @@ static SDValue insert1BitVector(SDValue Op, SelectionDAG &DAG,
DAG.getTargetConstant(IdxVal, dl, MVT::i8)); DAG.getTargetConstant(IdxVal, dl, MVT::i8));
if (SubVecNumElems * 2 == NumElems) { if (SubVecNumElems * 2 == NumElems) {
// Special case, use legal zero extending insert_subvector. This allows // Special case, use legal zero extending insert_subvector. This allows
// isel to opimitize when bits are known zero. // isel to optimize when bits are known zero.
Vec = DAG.getNode(ISD::EXTRACT_SUBVECTOR, dl, SubVecVT, Vec, ZeroIdx); Vec = DAG.getNode(ISD::EXTRACT_SUBVECTOR, dl, SubVecVT, Vec, ZeroIdx);
Vec = DAG.getNode(ISD::INSERT_SUBVECTOR, dl, WideOpVT, Vec = DAG.getNode(ISD::INSERT_SUBVECTOR, dl, WideOpVT,
DAG.getConstant(0, dl, WideOpVT), DAG.getConstant(0, dl, WideOpVT),
@ -7581,7 +7581,7 @@ static bool getTargetShuffleInputs(SDValue Op, SmallVectorImpl<SDValue> &Inputs,
KnownZero, DAG, Depth, ResolveKnownElts); KnownZero, DAG, Depth, ResolveKnownElts);
} }
/// Returns the scalar element that will make up the ith /// Returns the scalar element that will make up the i'th
/// element of the result of the vector shuffle. /// element of the result of the vector shuffle.
static SDValue getShuffleScalarElt(SDNode *N, unsigned Index, SelectionDAG &DAG, static SDValue getShuffleScalarElt(SDNode *N, unsigned Index, SelectionDAG &DAG,
unsigned Depth) { unsigned Depth) {
@ -8437,7 +8437,7 @@ static SDValue lowerBuildVectorAsBroadcast(BuildVectorSDNode *BVOp,
SDValue Ld = BVOp->getSplatValue(&UndefElements); SDValue Ld = BVOp->getSplatValue(&UndefElements);
// Attempt to use VBROADCASTM // Attempt to use VBROADCASTM
// From this paterrn: // From this pattern:
// a. t0 = (zext_i64 (bitcast_i8 v2i1 X)) // a. t0 = (zext_i64 (bitcast_i8 v2i1 X))
// b. t1 = (build_vector t0 t0) // b. t1 = (build_vector t0 t0)
// //
@ -36636,7 +36636,7 @@ static SDValue combineBitcast(SDNode *N, SelectionDAG &DAG,
if (SDValue V = combineBitcastvxi1(DAG, VT, N0, dl, Subtarget)) if (SDValue V = combineBitcastvxi1(DAG, VT, N0, dl, Subtarget))
return V; return V;
// Recognize the IR pattern for the movmsk intrinsic under SSE1 befoer type // Recognize the IR pattern for the movmsk intrinsic under SSE1 before type
// legalization destroys the v4i32 type. // legalization destroys the v4i32 type.
if (Subtarget.hasSSE1() && !Subtarget.hasSSE2() && SrcVT == MVT::v4i1 && if (Subtarget.hasSSE1() && !Subtarget.hasSSE2() && SrcVT == MVT::v4i1 &&
VT.isScalarInteger() && N0.getOpcode() == ISD::SETCC && VT.isScalarInteger() && N0.getOpcode() == ISD::SETCC &&