2015-05-14 20:05:18 +08:00
|
|
|
; RUN: opt -basicaa -loop-distribute -verify-loop-info -verify-dom-info -S \
|
|
|
|
; RUN: < %s | FileCheck %s
|
|
|
|
|
|
|
|
; RUN: opt -basicaa -loop-distribute -loop-vectorize -force-vector-width=4 \
|
|
|
|
; RUN: -verify-loop-info -verify-dom-info -S < %s | \
|
|
|
|
; RUN: FileCheck --check-prefix=VECTORIZE %s
|
|
|
|
|
|
|
|
; The memcheck version of basic.ll. We should distribute and vectorize the
|
|
|
|
; second part of this loop with 5 memchecks (A+1 x {C, D, E} + C x {A, B})
|
|
|
|
;
|
|
|
|
; for (i = 0; i < n; i++) {
|
|
|
|
; A[i + 1] = A[i] * B[i];
|
|
|
|
; -------------------------------
|
|
|
|
; C[i] = D[i] * E[i];
|
|
|
|
; }
|
|
|
|
|
|
|
|
target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128"
|
|
|
|
target triple = "x86_64-apple-macosx10.10.0"
|
|
|
|
|
|
|
|
@B = common global i32* null, align 8
|
|
|
|
@A = common global i32* null, align 8
|
|
|
|
@C = common global i32* null, align 8
|
|
|
|
@D = common global i32* null, align 8
|
|
|
|
@E = common global i32* null, align 8
|
|
|
|
|
|
|
|
define void @f() {
|
|
|
|
entry:
|
|
|
|
%a = load i32*, i32** @A, align 8
|
|
|
|
%b = load i32*, i32** @B, align 8
|
|
|
|
%c = load i32*, i32** @C, align 8
|
|
|
|
%d = load i32*, i32** @D, align 8
|
|
|
|
%e = load i32*, i32** @E, align 8
|
|
|
|
br label %for.body
|
|
|
|
|
[LAA] Merge memchecks for accesses separated by a constant offset
Summary:
Often filter-like loops will do memory accesses that are
separated by constant offsets. In these cases it is
common that we will exceed the threshold for the
allowable number of checks.
However, it should be possible to merge such checks,
sice a check of any interval againt two other intervals separated
by a constant offset (a,b), (a+c, b+c) will be equivalent with
a check againt (a, b+c), as long as (a,b) and (a+c, b+c) overlap.
Assuming the loop will be executed for a sufficient number of
iterations, this will be true. If not true, checking against
(a, b+c) is still safe (although not equivalent).
As long as there are no dependencies between two accesses,
we can merge their checks into a single one. We use this
technique to construct groups of accesses, and then check
the intervals associated with the groups instead of
checking the accesses directly.
Reviewers: anemet
Subscribers: llvm-commits
Differential Revision: http://reviews.llvm.org/D10386
llvm-svn: 241673
2015-07-08 17:16:33 +08:00
|
|
|
; We have two compares for each array overlap check.
|
|
|
|
; Since the checks to A and A + 4 get merged, this will give us a
|
|
|
|
; total of 8 compares.
|
2015-05-14 20:05:18 +08:00
|
|
|
;
|
2015-11-09 21:26:09 +08:00
|
|
|
; CHECK: for.body.lver.check:
|
2015-05-14 20:05:18 +08:00
|
|
|
; CHECK: = icmp
|
|
|
|
; CHECK: = icmp
|
|
|
|
|
|
|
|
; CHECK: = icmp
|
|
|
|
; CHECK: = icmp
|
|
|
|
|
|
|
|
; CHECK: = icmp
|
|
|
|
; CHECK: = icmp
|
|
|
|
|
|
|
|
; CHECK: = icmp
|
|
|
|
; CHECK: = icmp
|
|
|
|
|
|
|
|
; CHECK-NOT: = icmp
|
2015-06-23 06:59:40 +08:00
|
|
|
; CHECK: br i1 %memcheck.conflict, label %for.body.ph.lver.orig, label %for.body.ph.ldist1
|
2015-05-14 20:05:18 +08:00
|
|
|
|
|
|
|
; The non-distributed loop that the memchecks fall back on.
|
|
|
|
|
2015-06-23 06:59:40 +08:00
|
|
|
; CHECK: for.body.ph.lver.orig:
|
|
|
|
; CHECK: br label %for.body.lver.orig
|
|
|
|
; CHECK: for.body.lver.orig:
|
|
|
|
; CHECK: br i1 %exitcond.lver.orig, label %for.end, label %for.body.lver.orig
|
2015-05-14 20:05:18 +08:00
|
|
|
|
|
|
|
; Verify the two distributed loops.
|
|
|
|
|
|
|
|
; CHECK: for.body.ph.ldist1:
|
|
|
|
; CHECK: br label %for.body.ldist1
|
|
|
|
; CHECK: for.body.ldist1:
|
|
|
|
; CHECK: %mulA.ldist1 = mul i32 %loadB.ldist1, %loadA.ldist1
|
|
|
|
; CHECK: br i1 %exitcond.ldist1, label %for.body.ph, label %for.body.ldist1
|
|
|
|
|
|
|
|
; CHECK: for.body.ph:
|
|
|
|
; CHECK: br label %for.body
|
|
|
|
; CHECK: for.body:
|
|
|
|
; CHECK: %mulC = mul i32 %loadD, %loadE
|
|
|
|
; CHECK: for.end:
|
|
|
|
|
|
|
|
|
|
|
|
; VECTORIZE: mul <4 x i32>
|
|
|
|
|
|
|
|
for.body: ; preds = %for.body, %entry
|
|
|
|
%ind = phi i64 [ 0, %entry ], [ %add, %for.body ]
|
|
|
|
|
|
|
|
%arrayidxA = getelementptr inbounds i32, i32* %a, i64 %ind
|
|
|
|
%loadA = load i32, i32* %arrayidxA, align 4
|
|
|
|
|
|
|
|
%arrayidxB = getelementptr inbounds i32, i32* %b, i64 %ind
|
|
|
|
%loadB = load i32, i32* %arrayidxB, align 4
|
|
|
|
|
|
|
|
%mulA = mul i32 %loadB, %loadA
|
|
|
|
|
|
|
|
%add = add nuw nsw i64 %ind, 1
|
|
|
|
%arrayidxA_plus_4 = getelementptr inbounds i32, i32* %a, i64 %add
|
|
|
|
store i32 %mulA, i32* %arrayidxA_plus_4, align 4
|
|
|
|
|
|
|
|
%arrayidxD = getelementptr inbounds i32, i32* %d, i64 %ind
|
|
|
|
%loadD = load i32, i32* %arrayidxD, align 4
|
|
|
|
|
|
|
|
%arrayidxE = getelementptr inbounds i32, i32* %e, i64 %ind
|
|
|
|
%loadE = load i32, i32* %arrayidxE, align 4
|
|
|
|
|
|
|
|
%mulC = mul i32 %loadD, %loadE
|
|
|
|
|
|
|
|
%arrayidxC = getelementptr inbounds i32, i32* %c, i64 %ind
|
|
|
|
store i32 %mulC, i32* %arrayidxC, align 4
|
|
|
|
|
|
|
|
%exitcond = icmp eq i64 %add, 20
|
|
|
|
br i1 %exitcond, label %for.end, label %for.body
|
|
|
|
|
|
|
|
for.end: ; preds = %for.body
|
|
|
|
ret void
|
|
|
|
}
|