forked from OSchip/llvm-project
[RISCV] Implement RISCVTTIImpl::getMaxVScale correctly
The comments in the existing code appear to pre-exist the standardization of the +v extension. In particular, the specification *does* provide a bound on the maximum value VLEN can take. From what I can tell, the LMUL comment was simply a misunderstanding of what this API returns. This API returns the maximum value that vscale can take at runtime. This is used in the vectorizer to bound the largest scalable VF (e.g. LMUL in RISCV terms) which can be used without violating memory dependence. Differential Revision: https://reviews.llvm.org/D128538
This commit is contained in:
parent
27aca975b6
commit
4710e78974
|
@ -132,16 +132,8 @@ bool RISCVTTIImpl::shouldExpandReduction(const IntrinsicInst *II) const {
|
|||
}
|
||||
|
||||
Optional<unsigned> RISCVTTIImpl::getMaxVScale() const {
|
||||
// There is no assumption of the maximum vector length in V specification.
|
||||
// We use the value specified by users as the maximum vector length.
|
||||
// This function will use the assumed maximum vector length to get the
|
||||
// maximum vscale for LoopVectorizer.
|
||||
// If users do not specify the maximum vector length, we have no way to
|
||||
// know whether the LoopVectorizer is safe to do or not.
|
||||
// We only consider to use single vector register (LMUL = 1) to vectorize.
|
||||
unsigned MaxVectorSizeInBits = ST->getMaxRVVVectorSizeInBits();
|
||||
if (ST->hasVInstructions() && MaxVectorSizeInBits != 0)
|
||||
return MaxVectorSizeInBits / RISCV::RVVBitsPerBlock;
|
||||
if (ST->hasVInstructions())
|
||||
return ST->getRealMaxVLen() / RISCV::RVVBitsPerBlock;
|
||||
return BaseT::getMaxVScale();
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue