llvm-project/llvm/test/Transforms/LoopDistribute/symbolic-stride.ll

66 lines
2.0 KiB
LLVM

; RUN: opt -basicaa -loop-distribute -S < %s | \
; RUN: FileCheck %s --check-prefix=ALL --check-prefix=STRIDE_SPEC
; RUN: opt -basicaa -loop-distribute -S -enable-mem-access-versioning=0 < %s | \
; RUN: FileCheck %s --check-prefix=ALL --check-prefix=NO_STRIDE_SPEC
; If we don't speculate stride for 1 we can't distribute along the line
; because we could have a backward dependence:
;
; for (i = 0; i < n; i++) {
; A[i + 1] = A[i] * B[i];
; =======================
; C[i] = D[i] * A[stride * i];
; }
target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128"
target triple = "x86_64-apple-macosx10.10.0"
; ALL-LABEL: @f(
define void @f(i32* noalias %a,
i32* noalias %b,
i32* noalias %c,
i32* noalias %d,
i64 %stride) {
entry:
br label %for.body
; STRIDE_SPEC: %ident.check = icmp ne i64 %stride, 1
; STRIDE_SPEC: for.body.ldist1:
; NO_STRIDE_SPEC-NOT: for.body.ldist1:
for.body: ; preds = %for.body, %entry
%ind = phi i64 [ 0, %entry ], [ %add, %for.body ]
%arrayidxA = getelementptr inbounds i32, i32* %a, i64 %ind
%loadA = load i32, i32* %arrayidxA, align 4
%arrayidxB = getelementptr inbounds i32, i32* %b, i64 %ind
%loadB = load i32, i32* %arrayidxB, align 4
%mulA = mul i32 %loadB, %loadA
%add = add nuw nsw i64 %ind, 1
%arrayidxA_plus_4 = getelementptr inbounds i32, i32* %a, i64 %add
store i32 %mulA, i32* %arrayidxA_plus_4, align 4
%arrayidxD = getelementptr inbounds i32, i32* %d, i64 %ind
%loadD = load i32, i32* %arrayidxD, align 4
%mul = mul i64 %ind, %stride
%arrayidxStridedA = getelementptr inbounds i32, i32* %a, i64 %mul
%loadStridedA = load i32, i32* %arrayidxStridedA, align 4
%mulC = mul i32 %loadD, %loadStridedA
%arrayidxC = getelementptr inbounds i32, i32* %c, i64 %ind
store i32 %mulC, i32* %arrayidxC, align 4
%exitcond = icmp eq i64 %add, 20
br i1 %exitcond, label %for.end, label %for.body
for.end: ; preds = %for.body
ret void
}