2018-12-18 01:57:14 +08:00
|
|
|
// RUN: mlir-opt %s -loop-fusion -split-input-file -verify | FileCheck %s
|
|
|
|
|
|
|
|
// TODO(andydavis) Add more tests:
|
|
|
|
// *) Add nested fusion test cases when non-constant loop bound support is
|
|
|
|
// added to iteration domain in dependence check.
|
|
|
|
// *) Add a test w/ floordiv/ceildiv/mod when supported in dependence check.
|
|
|
|
// *) Add tests which check fused computation slice indexing and loop bounds.
|
2019-01-03 02:20:00 +08:00
|
|
|
// TODO(andydavis) Test clean up: move memref allocs to func args.
|
2018-12-18 01:57:14 +08:00
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_raw_dep_for_locality() {
|
|
|
|
func @should_fuse_raw_dep_for_locality() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = affine_apply [[MAP0]](%i0)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_reduction_to_pointwise() {
|
|
|
|
func @should_fuse_reduction_to_pointwise() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10x10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%c = alloc() : memref<10xf32>
|
|
|
|
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
for %i1 = 0 to 10 {
|
2018-12-20 12:42:55 +08:00
|
|
|
%v0 = load %b[%i0] : memref<10xf32>
|
2018-12-18 01:57:14 +08:00
|
|
|
%v1 = load %a[%i0, %i1] : memref<10x10xf32>
|
|
|
|
%v3 = addf %v0, %v1 : f32
|
|
|
|
store %v3, %b[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v4 = load %b[%i2] : memref<10xf32>
|
|
|
|
store %v4, %c[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
// Should fuse in entire inner loop on %i1 from source loop nest, as %i1
|
|
|
|
// is not used in the access function of the store/load on %b.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
2018-12-20 12:42:55 +08:00
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP0]](%i0)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: for %i1 = 0 to 10 {
|
2018-12-20 12:42:55 +08:00
|
|
|
// CHECK-NEXT: %4 = load %1[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %5 = load %0[%3, %i1] : memref<10x10xf32>
|
|
|
|
// CHECK-NEXT: %6 = addf %4, %5 : f32
|
|
|
|
// CHECK-NEXT: store %6, %1[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: %7 = load %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %7, %2[%i0] : memref<10xf32>
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK: [[MAP_SHIFT_MINUS_ONE:#map[0-9]+]] = (d0) -> (d0 - 1)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK: [[MAP_SHIFT_BY_ONE:#map[0-9]+]] = (d0, d1) -> (d0 + 1, d1 + 1)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_loop_nests_with_shifts() {
|
|
|
|
func @should_fuse_loop_nests_with_shifts() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10x10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%a0 = affine_apply (d0, d1) -> (d0 + 1, d1 + 1) (%i0, %i1)
|
|
|
|
store %cf7, %a[%a0#0, %a0#1] : memref<10x10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
for %i3 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i2, %i3] : memref<10x10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-17 01:55:02 +08:00
|
|
|
// The cost of fusing the src loop nest at dst loop depth 1 is less expensive
|
|
|
|
// than fusing at dst loop depth 2, because at dst loop depth 1, we are
|
|
|
|
// able to reduce the trip count around the %i1 loop by one (because the
|
|
|
|
// dst loop never reads the last element written by the src loop).
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
2019-01-17 01:55:02 +08:00
|
|
|
// CHECK-NEXT: %1 = affine_apply [[MAP_SHIFT_MINUS_ONE]](%i0)
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 9 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply [[MAP_SHIFT_BY_ONE]](%1, %i1)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%2#0, %2#1] : memref<10x10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %3 = load %0[%i0, %i2] : memref<10x10xf32>
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: }
|
2018-11-01 22:26:00 +08:00
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
2018-12-18 01:57:14 +08:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-DAG: [[MAP_ID:#map[0-9]+]] = (d0) -> (d0)
|
2018-12-18 01:57:14 +08:00
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_loop_nest() {
|
|
|
|
func @should_fuse_loop_nest() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10x10xf32>
|
|
|
|
%b = alloc() : memref<10x10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %a[%i0, %i1] : memref<10x10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
for %i3 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i3, %i2] : memref<10x10xf32>
|
|
|
|
store %v0, %b[%i2, %i3] : memref<10x10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i4 = 0 to 10 {
|
|
|
|
for %i5 = 0 to 10 {
|
|
|
|
%v1 = load %b[%i4, %i5] : memref<10x10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 10 {
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-NEXT: %2 = affine_apply [[MAP_ID]](%i1)
|
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP_ID]](%i0)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: store %cst, %0[%2, %3] : memref<10x10xf32>
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-NEXT: %4 = affine_apply [[MAP_ID]](%i0)
|
|
|
|
// CHECK-NEXT: %5 = affine_apply [[MAP_ID]](%i1)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: %6 = load %0[%5, %4] : memref<10x10xf32>
|
|
|
|
// CHECK-NEXT: store %6, %1[%4, %5] : memref<10x10xf32>
|
|
|
|
// CHECK-NEXT: %7 = load %1[%i0, %i1] : memref<10x10xf32>
|
2018-11-01 22:26:00 +08:00
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
2018-12-18 01:57:14 +08:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_across_intermediate_loop_with_no_deps() {
|
|
|
|
func @should_fuse_across_intermediate_loop_with_no_deps() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%c = alloc() : memref<10xf32>
|
|
|
|
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i0] : memref<10xf32>
|
|
|
|
store %v0, %b[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %c[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v1 = load %b[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
// Should fuse first loop (past second loop with no dependences) into third.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %2[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP0]](%i1)
|
|
|
|
// CHECK-NEXT: %4 = load %0[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %4, %1[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %5 = load %1[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_all_loops() {
|
|
|
|
func @should_fuse_all_loops() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
// Set up flow dependences from first and second loops to third.
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %a[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %b[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i2] : memref<10xf32>
|
|
|
|
%v1 = load %b[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
// Should fuse first and second loops into third.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply [[MAP0]](%i0)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP0]](%i0)
|
|
|
|
// CHECK-NEXT: store %cst, %1[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %4 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %5 = load %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_first_and_second_loops() {
|
|
|
|
func @should_fuse_first_and_second_loops() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%c = alloc() : memref<10xf32>
|
|
|
|
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %a[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i1] : memref<10xf32>
|
|
|
|
store %cf7, %b[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v1 = load %c[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
// Should fuse first loop into the second (last loop should not be fused).
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP0]](%i0)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%3] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %4 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %cst, %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %5 = load %2[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_would_create_cycle() {
|
|
|
|
func @should_not_fuse_would_create_cycle() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%c = alloc() : memref<10xf32>
|
|
|
|
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
// Set up the following dependences:
|
|
|
|
// 1) loop0 -> loop1 on memref '%a'
|
|
|
|
// 2) loop0 -> loop2 on memref '%b'
|
|
|
|
// 3) loop1 -> loop2 on memref '%c'
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i0] : memref<10xf32>
|
|
|
|
store %cf7, %b[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %a[%i1] : memref<10xf32>
|
|
|
|
%v1 = load %c[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v2 = load %b[%i2] : memref<10xf32>
|
|
|
|
store %cf7, %c[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// Should not fuse: fusing loop first loop into last would create a cycle.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %3 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %cst, %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %4 = load %2[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %5 = load %1[%i2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %cst, %2[%i2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_raw_dep_would_be_violated() {
|
|
|
|
func @should_not_fuse_raw_dep_would_be_violated() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v1 = load %m[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// Fusing loop %i0 to %i2 would violate the RAW dependence between %i0 and %i1
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_waw_dep_would_be_violated() {
|
|
|
|
func @should_not_fuse_waw_dep_would_be_violated() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v1 = load %m[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// Fusing loop %i0 to %i2 would violate the WAW dependence between %i0 and %i1
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_war_dep_would_be_violated() {
|
|
|
|
func @should_not_fuse_war_dep_would_be_violated() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%a = alloc() : memref<10xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i0] : memref<10xf32>
|
|
|
|
store %v0, %b[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
store %cf7, %a[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v1 = load %b[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// Fusing loop %i0 to %i2 would violate the WAR dependence between %i0 and %i1
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: store %2, %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %3 = load %1[%i2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_if_top_level_access() {
|
|
|
|
func @should_not_fuse_if_top_level_access() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
%c0 = constant 4 : index
|
|
|
|
%v1 = load %m[%c0] : memref<10xf32>
|
|
|
|
// Top-level load to '%m' should prevent fusion.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_fuse_no_top_level_access() {
|
|
|
|
func @should_fuse_no_top_level_access() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
#set0 = (d0) : (1 == 0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_if_inst_at_top_level() {
|
|
|
|
func @should_not_fuse_if_inst_at_top_level() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
|
|
|
}
|
|
|
|
%c0 = constant 4 : index
|
|
|
|
if #set0(%c0) {
|
|
|
|
}
|
2018-12-29 08:05:35 +08:00
|
|
|
// Top-level IfInst should prevent fusion.
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
#set0 = (d0) : (1 == 0)
|
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @should_not_fuse_if_inst_in_loop_nest() {
|
|
|
|
func @should_not_fuse_if_inst_in_loop_nest() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10xf32>
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
%c4 = constant 4 : index
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cf7, %m[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
if #set0(%c4) {
|
2018-11-01 22:26:00 +08:00
|
|
|
}
|
2018-12-18 01:57:14 +08:00
|
|
|
%v0 = load %m[%i1] : memref<10xf32>
|
2018-11-01 22:26:00 +08:00
|
|
|
}
|
|
|
|
|
2018-12-29 08:05:35 +08:00
|
|
|
// IfInst in ForInst should prevent fusion.
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: if #set0(%c4) {
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK: [[MAP0:#map[0-9]+]] = (d0) -> (d0)
|
|
|
|
// CHECK: [[MAP1:#map[0-9]+]] = (d0, d1, d2) -> (d0, d1, d2)
|
2019-01-08 09:34:26 +08:00
|
|
|
// CHECK: [[MAP_PERMUTE:#map[0-9]+]] = (d0, d1, d2) -> (d1, d2, d0)
|
2018-12-18 01:57:14 +08:00
|
|
|
|
2019-01-03 02:20:00 +08:00
|
|
|
// CHECK-LABEL: func @remap_ivs() {
|
|
|
|
func @remap_ivs() {
|
2018-12-18 01:57:14 +08:00
|
|
|
%m = alloc() : memref<10x20x30xf32>
|
|
|
|
|
|
|
|
%cf7 = constant 7.0 : f32
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
for %i1 = 0 to 20 {
|
|
|
|
for %i2 = 0 to 30 {
|
|
|
|
%a0 = affine_apply (d0, d1, d2) -> (d0, d1, d2) (%i0, %i1, %i2)
|
|
|
|
store %cf7, %m[%a0#0, %a0#1, %a0#2] : memref<10x20x30xf32>
|
|
|
|
}
|
2018-11-01 22:26:00 +08:00
|
|
|
}
|
|
|
|
}
|
2018-12-18 01:57:14 +08:00
|
|
|
for %i3 = 0 to 30 {
|
|
|
|
for %i4 = 0 to 10 {
|
|
|
|
for %i5 = 0 to 20 {
|
|
|
|
%a1 = affine_apply (d0, d1, d2) -> (d1, d2, d0) (%i3, %i4, %i5)
|
|
|
|
%v0 = load %m[%a1#0, %a1#1, %a1#2] : memref<10x20x30xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 30 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 20 {
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-NEXT: %1 = affine_apply [[MAP0]](%i1)
|
|
|
|
// CHECK-NEXT: %2 = affine_apply [[MAP0]](%i2)
|
|
|
|
// CHECK-NEXT: %3 = affine_apply [[MAP0]](%i0)
|
|
|
|
// CHECK-NEXT: %4 = affine_apply [[MAP1]](%1, %2, %3)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: store %cst, %0[%4#0, %4#1, %4#2] : memref<10x20x30xf32>
|
2019-01-08 09:34:26 +08:00
|
|
|
// CHECK-NEXT: %5 = affine_apply [[MAP_PERMUTE]](%i0, %i1, %i2)
|
2018-12-18 01:57:14 +08:00
|
|
|
// CHECK-NEXT: %6 = load %0[%5#0, %5#1, %5#2] : memref<10x20x30xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
2018-11-01 22:26:00 +08:00
|
|
|
|
|
|
|
return
|
2018-12-18 01:57:14 +08:00
|
|
|
}
|
2018-12-22 03:06:23 +08:00
|
|
|
|
|
|
|
// -----
|
|
|
|
|
2019-01-05 06:52:21 +08:00
|
|
|
// CHECK-DAG: #map0 = (d0, d1) -> (d0 * 4 + d1)
|
|
|
|
// CHECK-DAG: #map1 = (d0) -> (d0 floordiv 4, d0 mod 4)
|
|
|
|
|
|
|
|
// Reshape from a 64 x f32 to 16 x 4 x f32.
|
|
|
|
// CHECK-LABEL: func @fuse_reshape_64_16_4
|
|
|
|
func @fuse_reshape_64_16_4(%in : memref<64xf32>) {
|
|
|
|
%out = alloc() : memref<16x4xf32>
|
|
|
|
|
|
|
|
for %i0 = 0 to 64 {
|
|
|
|
%v = load %in[%i0] : memref<64xf32>
|
|
|
|
%idx = affine_apply (d0) -> (d0 floordiv 4, d0 mod 4) (%i0)
|
|
|
|
store %v, %out[%idx#0, %idx#1] : memref<16x4xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i1 = 0 to 16 {
|
|
|
|
for %i2 = 0 to 4 {
|
|
|
|
%w = load %out[%i1, %i2] : memref<16x4xf32>
|
|
|
|
"foo"(%w) : (f32) -> ()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
// CHECK: for %i0 =
|
|
|
|
// CHECK-NEXT: for %i1 =
|
|
|
|
// CHECK-NOT: for
|
|
|
|
// CHECK: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
}
|
2019-01-08 09:34:26 +08:00
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0) -> (d0 floordiv 4)
|
|
|
|
// CHECK: #map1 = (d0) -> (d0 mod 4)
|
|
|
|
|
|
|
|
// Reshape a 16x4xf32 to 64xf32.
|
|
|
|
// CHECK-LABEL: func @fuse_reshape_16_4_64
|
|
|
|
func @fuse_reshape_16_4_64() {
|
|
|
|
%in = alloc() : memref<16x4xf32>
|
|
|
|
%out = alloc() : memref<64xf32>
|
|
|
|
|
|
|
|
for %i0 = 0 to 16 {
|
|
|
|
for %i1 = 0 to 4 {
|
|
|
|
%v = load %in[%i0, %i1] : memref<16x4xf32>
|
|
|
|
%idx = affine_apply (d0, d1) -> (4*d0 + d1) (%i0, %i1)
|
|
|
|
store %v, %out[%idx] : memref<64xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i2 = 0 to 64 {
|
|
|
|
%w = load %out[%i2] : memref<64xf32>
|
|
|
|
"foo"(%w) : (f32) -> ()
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 64 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: %3 = affine_apply #map1(%i0)
|
|
|
|
// CHECK-NEXT: %4 = load %0[%2, %3] : memref<16x4xf32>
|
|
|
|
// CHECK-NEXT: %5 = affine_apply #map2(%2, %3)
|
|
|
|
// CHECK-NEXT: store %4, %1[%5] : memref<64xf32>
|
|
|
|
// CHECK-NEXT: %6 = load %1[%i0] : memref<64xf32>
|
|
|
|
// CHECK-NEXT: "foo"(%6) : (f32) -> ()
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// All three loop nests below (6-d one, 2-d one, 2-d one is fused into a single
|
|
|
|
// 2-d loop nest).
|
|
|
|
// CHECK-LABEL: func @R6_to_R2_reshape
|
|
|
|
func @R6_to_R2_reshape_square() -> memref<64x9xi32> {
|
|
|
|
%in = alloc() : memref<2x2x3x3x16x1xi32>
|
|
|
|
%out = alloc() : memref<64x9xi32>
|
|
|
|
|
2019-01-18 04:51:48 +08:00
|
|
|
// Initialize input.
|
2019-01-08 09:34:26 +08:00
|
|
|
for %i0 = 0 to 2 {
|
|
|
|
for %i1 = 0 to 2 {
|
|
|
|
for %i2 = 0 to 3 {
|
|
|
|
for %i3 = 0 to 3 {
|
|
|
|
for %i4 = 0 to 16 {
|
|
|
|
for %i5 = 0 to 1 {
|
|
|
|
%val = "foo"(%i0, %i1, %i2, %i3, %i4, %i5) : (index, index, index, index, index, index) -> i32
|
|
|
|
store %val, %in[%i0, %i1, %i2, %i3, %i4, %i5] : memref<2x2x3x3x16x1xi32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %ii = 0 to 64 {
|
|
|
|
for %jj = 0 to 9 {
|
|
|
|
// Convert output coordinates to linear index.
|
|
|
|
%a0 = affine_apply (d0, d1) -> (d0 * 9 + d1) (%ii, %jj)
|
|
|
|
%a1 = affine_apply (d0) -> (
|
|
|
|
d0 floordiv (2 * 3 * 3 * 16 * 1),
|
|
|
|
(d0 mod 288) floordiv (3 * 3 * 16 * 1),
|
|
|
|
((d0 mod 288) mod 144) floordiv 48,
|
|
|
|
(((d0 mod 288) mod 144) mod 48) floordiv 16,
|
|
|
|
((((d0 mod 288) mod 144) mod 48) mod 16),
|
|
|
|
(((d0 mod 144) mod 144) mod 48) mod 16
|
|
|
|
) (%a0)
|
|
|
|
%v = load %in[%a1#0, %a1#1, %a1#3, %a1#4, %a1#2, %a1#5]
|
|
|
|
: memref<2x2x3x3x16x1xi32>
|
|
|
|
store %v, %out[%ii, %jj] : memref<64x9xi32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i = 0 to 64 {
|
|
|
|
for %j = 0 to 9 {
|
|
|
|
%a = load %out[%i, %j] : memref<64x9xi32>
|
|
|
|
%b = muli %a, %a : i32
|
|
|
|
store %b, %out[%i, %j] : memref<64x9xi32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return %out : memref<64x9xi32>
|
|
|
|
}
|
|
|
|
// Everything above is fused to a single 2-d loop nest, and the 6-d tensor %in
|
|
|
|
// is eliminated if -memref-dataflow-opt is also supplied.
|
|
|
|
//
|
|
|
|
// CHECK: for %i0 = 0 to 64 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 9 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %3 = affine_apply #map1(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %4 = affine_apply #map2(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %5 = affine_apply #map3(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %6 = affine_apply #map4(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %7 = "foo"(%2, %3, %4, %5, %6, %c0) : (index, index, index, index, index, index) -> i32
|
|
|
|
// CHECK-NEXT: store %7, %0[%2, %3, %4, %5, %6, %c0] : memref<2x2x3x3x16x1xi32>
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-NEXT: %8 = affine_apply #map5(%i0)
|
|
|
|
// CHECK-NEXT: %9 = affine_apply #map5(%i1)
|
|
|
|
// CHECK-NEXT: %10 = affine_apply #map6(%8, %9)
|
|
|
|
// CHECK-NEXT: %11 = affine_apply #map7(%10)
|
2019-01-08 09:34:26 +08:00
|
|
|
// CHECK-NEXT: %12 = load %0[%11#0, %11#1, %11#3, %11#4, %11#2, %11#5] : memref<2x2x3x3x16x1xi32>
|
|
|
|
// CHECK-NEXT: store %12, %1[%8, %9] : memref<64x9xi32>
|
|
|
|
// CHECK-NEXT: %13 = load %1[%i0, %i1] : memref<64x9xi32>
|
|
|
|
// CHECK-NEXT: %14 = muli %13, %13 : i32
|
|
|
|
// CHECK-NEXT: store %14, %1[%i0, %i1] : memref<64x9xi32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return %1 : memref<64x9xi32>
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @fuse_symbolic_bounds
|
|
|
|
func @fuse_symbolic_bounds(%M : index, %N : index) {
|
|
|
|
%m = alloc() : memref<800x800xf32>
|
|
|
|
|
|
|
|
%c0 = constant 0.0 : f32
|
|
|
|
%s = constant 5 : index
|
|
|
|
|
|
|
|
for %i0 = 0 to %M {
|
|
|
|
for %i1 = 0 to (d0) -> (d0 + 5) (%N) {
|
|
|
|
store %c0, %m[%i0, %i1] : memref<800 x 800 x f32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i2 = 0 to %M {
|
|
|
|
for %i3 = 0 to %N {
|
|
|
|
%idx = affine_apply (d0, d1)[s0] -> (d0, d1 + s0) (%i2, %i3)[%s]
|
|
|
|
%v = load %m[%idx#0, %idx#1] : memref<800 x 800 x f32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|
2019-01-15 03:26:25 +08:00
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0) -> (d0)
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_reduction_at_depth1
|
|
|
|
func @should_fuse_reduction_at_depth1() {
|
|
|
|
%a = alloc() : memref<10x100xf32>
|
|
|
|
%b = alloc() : memref<10xf32>
|
|
|
|
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
for %i1 = 0 to 100 {
|
|
|
|
%v0 = load %b[%i0] : memref<10xf32>
|
|
|
|
%v1 = load %a[%i0, %i1] : memref<10x100xf32>
|
|
|
|
%v2 = "maxf"(%v0, %v1) : (f32, f32) -> f32
|
|
|
|
store %v2, %b[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
for %i3 = 0 to 100 {
|
|
|
|
%v3 = load %b[%i2] : memref<10xf32>
|
|
|
|
%v4 = load %a[%i2, %i3] : memref<10x100xf32>
|
|
|
|
%v5 = subf %v4, %v3 : f32
|
|
|
|
store %v5, %b[%i2] : memref<10xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// This test should fuse the src reduction loop at depth 1 in the destination
|
|
|
|
// loop nest, which improves locality and enables subsequence passes to
|
|
|
|
// decrease the reduction memref size and possibly place it in a faster
|
|
|
|
// memory space.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 100 {
|
|
|
|
// CHECK-NEXT: %3 = load %1[%2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %4 = load %0[%2, %i1] : memref<10x100xf32>
|
|
|
|
// CHECK-NEXT: %5 = "maxf"(%3, %4) : (f32, f32) -> f32
|
|
|
|
// CHECK-NEXT: store %5, %1[%2] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 100 {
|
|
|
|
// CHECK-NEXT: %6 = load %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %7 = load %0[%i0, %i2] : memref<10x100xf32>
|
|
|
|
// CHECK-NEXT: %8 = subf %7, %6 : f32
|
|
|
|
// CHECK-NEXT: store %8, %1[%i0] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0) -> (d0)
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_at_src_depth1_and_dst_depth1
|
|
|
|
func @should_fuse_at_src_depth1_and_dst_depth1() {
|
|
|
|
%a = alloc() : memref<100x16xf32>
|
|
|
|
%b = alloc() : memref<100x16xf32>
|
|
|
|
|
|
|
|
for %i0 = 0 to 100 {
|
|
|
|
for %i1 = 0 to 16 {
|
|
|
|
%v0 = load %a[%i0, %i1] : memref<100x16xf32>
|
|
|
|
"op0"(%v0) : (f32) -> ()
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 16 {
|
|
|
|
%v1 = "op1"() : () -> (f32)
|
|
|
|
store %v1, %b[%i0, %i2] : memref<100x16xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i3 = 0 to 100 {
|
|
|
|
for %i4 = 0 to 16 {
|
|
|
|
%v2 = load %b[%i3, %i4] : memref<100x16xf32>
|
|
|
|
"op2"(%v2) : (f32) -> ()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// We can slice iterations of the '%i0' and '%i1' loops in the the source
|
|
|
|
// loop nest, but slicing at depth 2 and inserting the slice in the
|
|
|
|
// destination loop nest at depth2 causes extra computation. Instead,
|
|
|
|
// the fusion algorithm should detect that the source loop should be sliced
|
|
|
|
// at depth 1 and the slice should be inserted at depth 1.
|
|
|
|
// CHECK: for %i0 = 0 to 100 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: %3 = load %0[%2, %i1] : memref<100x16xf32>
|
|
|
|
// CHECK-NEXT: "op0"(%3) : (f32) -> ()
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: %4 = "op1"() : () -> f32
|
|
|
|
// CHECK-NEXT: store %4, %1[%2, %i2] : memref<100x16xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i3 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: %5 = load %1[%i0, %i3] : memref<100x16xf32>
|
|
|
|
// CHECK-NEXT: "op2"(%5) : (f32) -> ()
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0, d1) -> (d0 * 10 + d1)
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_src_depth1_at_dst_depth2
|
|
|
|
func @should_fuse_src_depth1_at_dst_depth2() {
|
|
|
|
%a = alloc() : memref<100xf32>
|
|
|
|
%c0 = constant 0.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 100 {
|
|
|
|
store %c0, %a[%i0] : memref<100xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%a0 = affine_apply (d0, d1) -> (d0 * 10 + d1) (%i1, %i2)
|
|
|
|
%v0 = load %a[%a0] : memref<100xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// The source loop nest slice loop bound is a function of both destination
|
|
|
|
// loop IVs, so we should slice at depth 1 and insert the slice at depth 2.
|
|
|
|
// CHECK: for %i0 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = affine_apply #map0(%i0, %i1)
|
|
|
|
// CHECK-NEXT: store %cst, %0[%1] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0, %i1)
|
|
|
|
// CHECK-NEXT: %3 = load %0[%2] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
2019-01-17 01:55:02 +08:00
|
|
|
// CHECK: #map0 = ()[s0] -> (s0)
|
2019-01-15 03:26:25 +08:00
|
|
|
|
|
|
|
// CHECK-LABEL: func @fusion_at_depth0_not_currently_supported
|
|
|
|
func @fusion_at_depth0_not_currently_supported() {
|
|
|
|
%0 = alloc() : memref<10xf32>
|
|
|
|
%c0 = constant 0 : index
|
|
|
|
%cst = constant 0.000000e+00 : f32
|
|
|
|
for %i0 = 0 to 10 {
|
|
|
|
store %cst, %0[%i0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
for %i1 = 0 to 10 {
|
|
|
|
%1 = load %0[%c0] : memref<10xf32>
|
|
|
|
}
|
|
|
|
// CHECK:for %i0 = 0 to 10 {
|
2019-01-17 01:55:02 +08:00
|
|
|
// CHECK-NEXT: %1 = affine_apply #map0()[%c0]
|
|
|
|
// CHECK-NEXT: store %cst, %0[%1] : memref<10xf32>
|
|
|
|
// CHECK-NEXT: %2 = load %0[%c0] : memref<10xf32>
|
2019-01-15 03:26:25 +08:00
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0) -> (d0)
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_deep_loop_nests
|
|
|
|
func @should_fuse_deep_loop_nests() {
|
|
|
|
%0 = alloc() : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
%1 = alloc() : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
%2 = alloc() : memref<3x3x3x3x16x10xf32, 2>
|
|
|
|
%c0 = constant 0 : index
|
|
|
|
%c1 = constant 1 : index
|
|
|
|
%c1_0 = constant 1 : index
|
|
|
|
%cst = constant 0.000000e+00 : f32
|
|
|
|
for %i0 = 0 to 2 {
|
|
|
|
for %i1 = 0 to 2 {
|
|
|
|
for %i2 = 0 to 3 {
|
|
|
|
for %i3 = 0 to 3 {
|
|
|
|
for %i4 = 0 to 16 {
|
|
|
|
for %i5 = 0 to 10 {
|
|
|
|
%3 = load %0[%i0, %i1, %i2, %i3, %i4, %i5]
|
|
|
|
: memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i6 = 0 to 16 {
|
|
|
|
for %i7 = 0 to 10 {
|
|
|
|
store %cst, %1[%i0, %i1, %i2, %i3, %i6, %i7]
|
|
|
|
: memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i8 = 0 to 3 {
|
|
|
|
for %i9 = 0 to 3 {
|
|
|
|
for %i10 = 0 to 2 {
|
|
|
|
for %i11 = 0 to 2 {
|
|
|
|
for %i12 = 0 to 3 {
|
|
|
|
for %i13 = 0 to 3 {
|
|
|
|
for %i14 = 0 to 2 {
|
|
|
|
for %i15 = 0 to 2 {
|
|
|
|
for %i16 = 0 to 16 {
|
|
|
|
for %i17 = 0 to 10 {
|
|
|
|
%5 = load %0[%i14, %i15, %i12, %i13, %i16, %i17]
|
|
|
|
: memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for %i18 = 0 to 16 {
|
|
|
|
for %i19 = 0 to 10 {
|
|
|
|
%6 = load %1[%i10, %i11, %i8, %i9, %i18, %i19]
|
|
|
|
: memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// The first four loops of the source loop nest can be sliced with iteration
|
|
|
|
// bounds which are a function of the first four loops of destination loop nest,
|
|
|
|
// where the destination loops nests have been interchanged.
|
|
|
|
// CHECK: for %i0 = 0 to 3 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 3 {
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 2 {
|
|
|
|
// CHECK-NEXT: for %i3 = 0 to 2 {
|
|
|
|
// CHECK-NEXT: %3 = affine_apply #map0(%i2)
|
|
|
|
// CHECK-NEXT: %4 = affine_apply #map0(%i3)
|
|
|
|
// CHECK-NEXT: %5 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: %6 = affine_apply #map0(%i1)
|
|
|
|
// CHECK-NEXT: for %i4 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: for %i5 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %7 = load %0[%3, %4, %5, %6, %i4, %i5] : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i6 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: for %i7 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: store %cst, %1[%3, %4, %5, %6, %i6, %i7] : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i8 = 0 to 3 {
|
|
|
|
// CHECK-NEXT: for %i9 = 0 to 3 {
|
|
|
|
// CHECK-NEXT: for %i10 = 0 to 2 {
|
|
|
|
// CHECK-NEXT: for %i11 = 0 to 2 {
|
|
|
|
// CHECK-NEXT: for %i12 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: for %i13 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %8 = load %0[%i10, %i11, %i8, %i9, %i12, %i13] : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i14 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: for %i15 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %9 = load %1[%i2, %i3, %i0, %i1, %i14, %i15] : memref<2x2x3x3x16x10xf32, 2>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
2019-01-17 01:55:02 +08:00
|
|
|
|
|
|
|
// -----
|
|
|
|
// CHECK: #map0 = (d0) -> (d0)
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_at_depth1_and_reduce_slice_trip_count
|
|
|
|
func @should_fuse_at_depth1_and_reduce_slice_trip_count() {
|
|
|
|
%a = alloc() : memref<4x256xf32>
|
|
|
|
%b = alloc() : memref<4x256xf32>
|
|
|
|
|
|
|
|
%c0 = constant 0 : index
|
|
|
|
%cf0 = constant 0.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 4 {
|
|
|
|
for %i1 = 0 to 256 {
|
|
|
|
%v0 = load %b[%i0, %i1] : memref<4x256xf32>
|
|
|
|
}
|
|
|
|
for %i2 = 0 to 256 {
|
|
|
|
store %cf0, %a[%i0, %i2] : memref<4x256xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for %d0 = 0 to 4 {
|
|
|
|
for %d1 = 0 to 16 {
|
|
|
|
%v1 = load %a[%d0, %d1] : memref<4x256xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// The cost of fusing at depth 2 is greater than the cost of fusing at depth 1
|
|
|
|
// for two reasons:
|
|
|
|
// 1) Inserting the unsliceable src loop %i1 to a higher depth removes
|
|
|
|
// redundant computation and reduces costs.
|
|
|
|
// 2) Inserting the sliceable src loop %i2 at depth 1, we can still reduce
|
|
|
|
// its trip count to 16 (from 256) reducing costs.
|
|
|
|
// CHECK: for %i0 = 0 to 4 {
|
|
|
|
// CHECK-NEXT: %2 = affine_apply #map0(%i0)
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 256 {
|
|
|
|
// CHECK-NEXT: %3 = load %1[%2, %i1] : memref<4x256xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%2, %i2] : memref<4x256xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i3 = 0 to 16 {
|
|
|
|
// CHECK-NEXT: %4 = load %0[%i0, %i3] : memref<4x256xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_at_depth1_with_trip_count_20
|
|
|
|
func @should_fuse_at_depth1_with_trip_count_20() {
|
|
|
|
%a = alloc() : memref<100xf32>
|
|
|
|
%c0 = constant 0 : index
|
|
|
|
%cf0 = constant 0.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 100 {
|
|
|
|
store %cf0, %a[%i0]: memref<100xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i1 = 0 to 5 {
|
|
|
|
for %i2 = 0 to 10 {
|
|
|
|
%v0 = load %a[%i2]: memref<100xf32>
|
|
|
|
}
|
|
|
|
for %i3 = 0 to 10 {
|
|
|
|
for %i4 = 0 to 20 {
|
|
|
|
%v1 = load %a[%i4]: memref<100xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 5 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 20 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i1] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i2] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i3 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: for %i4 = 0 to 20 {
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i4] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----
|
|
|
|
|
|
|
|
// CHECK-LABEL: func @should_fuse_at_depth1_with_trip_count_19
|
|
|
|
func @should_fuse_at_depth1_with_trip_count_19() {
|
|
|
|
%a = alloc() : memref<100xf32>
|
|
|
|
%c0 = constant 0 : index
|
|
|
|
%cf0 = constant 0.0 : f32
|
|
|
|
|
|
|
|
for %i0 = 0 to 100 {
|
|
|
|
store %cf0, %a[%i0]: memref<100xf32>
|
|
|
|
}
|
|
|
|
|
|
|
|
for %i1 = 0 to 5 {
|
|
|
|
for %i2 = 0 to 19 {
|
|
|
|
%v0 = load %a[%i2]: memref<100xf32>
|
|
|
|
}
|
|
|
|
for %i3 = 0 to 10 {
|
|
|
|
for %i4 = 0 to 10 {
|
|
|
|
%v1 = load %a[%i4]: memref<100xf32>
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// CHECK: for %i0 = 0 to 5 {
|
|
|
|
// CHECK-NEXT: for %i1 = 0 to 19 {
|
|
|
|
// CHECK-NEXT: store %cst, %0[%i1] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i2 = 0 to 19 {
|
|
|
|
// CHECK-NEXT: %1 = load %0[%i2] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: for %i3 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: for %i4 = 0 to 10 {
|
|
|
|
// CHECK-NEXT: %2 = load %0[%i4] : memref<100xf32>
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: }
|
|
|
|
// CHECK-NEXT: return
|
|
|
|
return
|
|
|
|
}
|