2014-09-22 08:32:15 +08:00
|
|
|
; RUN: llc < %s -mcpu=x86-64 -mattr=+avx -x86-experimental-vector-shuffle-lowering | FileCheck %s --check-prefix=ALL --check-prefix=AVX --check-prefix=AVX1
|
|
|
|
; RUN: llc < %s -mcpu=x86-64 -mattr=+avx2 -x86-experimental-vector-shuffle-lowering | FileCheck %s --check-prefix=ALL --check-prefix=AVX --check-prefix=AVX2
|
2014-08-14 20:13:59 +08:00
|
|
|
|
|
|
|
target triple = "x86_64-unknown-unknown"
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0000(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4f64_0000:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_0000:
|
2014-09-30 10:32:36 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
|
2014-09-30 10:32:36 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0001(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4f64_0001:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm1 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_0001:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,0,1]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 0, i32 0, i32 1>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0020(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_0020:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_0020:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,2,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 0, i32 2, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0300(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_0300:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm1 = ymm1[0,1,2,2]
|
2014-10-04 06:43:17 +08:00
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_0300:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,3,0,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 3, i32 0, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1000(<4 x double> %a, <4 x double> %b) {
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_1000:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} xmm1 = xmm0[1,0]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_1000:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,0,0,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_2200(<4 x double> %a, <4 x double> %b) {
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_2200:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4f64_2200:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,0,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 2, i32 2, i32 0, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_3330(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_3330:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm1 = ymm1[1,1,2,2]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,0,3,2]
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2],ymm1[3]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_3330:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 3, i32 3, i32 3, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_3210(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4f64_3210:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_3210:
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,2,1,0]
|
2014-09-25 19:03:55 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0023(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_0023:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,0,2,3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 0, i32 2, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0022(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; ALL-LABEL: shuffle_v4f64_0022:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
|
|
|
|
; ALL-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 0, i32 2, i32 2>
|
|
|
|
ret <4 x double> %shuffle
|
2014-08-15 19:01:40 +08:00
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1032(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1032:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 3, i32 2>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1133(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1133:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,3,3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 1, i32 3, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1023(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1023:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 2, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1022(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1022:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,2]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 2, i32 2>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0423(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4f64_0423:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} ymm1 = ymm1[0,0,2,2]
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4f64_0423:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vbroadcastsd %xmm1, %ymm1
|
|
|
|
; AVX2-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
|
|
|
|
; AVX2-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 4, i32 2, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0462(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; ALL-LABEL: shuffle_v4f64_0462:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vmovddup {{.*#+}} ymm1 = ymm1[0,0,2,2]
|
|
|
|
; ALL-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
|
|
|
|
; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2],ymm0[3]
|
|
|
|
; ALL-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 4, i32 6, i32 2>
|
2014-08-15 19:01:40 +08:00
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0426(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_0426:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[2]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 4, i32 2, i32 6>
|
|
|
|
ret <4 x double> %shuffle
|
2014-08-16 01:42:00 +08:00
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-16 01:42:00 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1537(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1537:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vunpckhpd {{.*#+}} ymm0 = ymm0[1],ymm1[1],ymm0[3],ymm1[3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-16 01:42:00 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 5, i32 3, i32 7>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-16 01:42:00 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_4062(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_4062:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-16 01:42:00 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 4, i32 0, i32 6, i32 2>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-16 01:42:00 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_5173(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_5173:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vunpckhpd {{.*#+}} ymm0 = ymm1[1],ymm0[1],ymm1[3],ymm0[3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-16 01:42:00 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 5, i32 1, i32 7, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
2014-08-15 19:01:40 +08:00
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-15 19:01:40 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_5163(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_5163:
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vshufpd {{.*#+}} ymm0 = ymm1[1],ymm0[1],ymm1[2],ymm0[3]
|
2014-09-21 19:07:41 +08:00
|
|
|
; ALL-NEXT: retq
|
2014-08-15 19:01:40 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 5, i32 1, i32 6, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-08-14 20:13:59 +08:00
|
|
|
|
2014-09-21 19:12:19 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0527(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_0527:
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 5, i32 2, i32 7>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_4163(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_4163:
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3]
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 4, i32 1, i32 6, i32 3>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_0145(<4 x double> %a, <4 x double> %b) {
|
2014-10-05 19:41:36 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_0145:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; ALL-NEXT: retq
|
2014-09-21 19:12:19 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 1, i32 4, i32 5>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_4501(<4 x double> %a, <4 x double> %b) {
|
2014-10-05 19:41:36 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_4501:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; ALL-NEXT: retq
|
2014-09-21 19:12:19 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 4, i32 5, i32 0, i32 1>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_0167(<4 x double> %a, <4 x double> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_0167:
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
|
2014-09-21 19:12:19 +08:00
|
|
|
; ALL-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 1, i32 6, i32 7>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
2014-11-21 20:17:50 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_1054(<4 x double> %a, <4 x double> %b) {
|
2014-11-21 21:56:05 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1054:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
|
|
|
; ALL-NEXT: retq
|
2014-11-21 20:17:50 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 5, i32 4>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_3254(<4 x double> %a, <4 x double> %b) {
|
2014-11-21 21:56:05 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_3254:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
|
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
|
|
|
; ALL-NEXT: retq
|
2014-11-21 20:17:50 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 3, i32 2, i32 5, i32 4>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_3276(<4 x double> %a, <4 x double> %b) {
|
2014-11-21 21:56:05 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_3276:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
|
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
|
|
|
; ALL-NEXT: retq
|
2014-11-21 20:17:50 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 3, i32 2, i32 7, i32 6>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @shuffle_v4f64_1076(<4 x double> %a, <4 x double> %b) {
|
2014-11-21 22:33:24 +08:00
|
|
|
; ALL-LABEL: shuffle_v4f64_1076:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
|
|
|
|
; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
|
|
|
; ALL-NEXT: retq
|
2014-11-21 20:17:50 +08:00
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 1, i32 0, i32 7, i32 6>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
2014-11-22 13:44:43 +08:00
|
|
|
define <4 x double> @shuffle_v4f64_0415(<4 x double> %a, <4 x double> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4f64_0415:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vunpckhpd {{.*#+}} xmm2 = xmm0[1],xmm1[1]
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4f64_0415:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm1 = ymm1[0,0,2,1]
|
|
|
|
; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
|
|
|
|
; AVX2-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
|
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 4, i32 1, i32 5>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0000(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_0000:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0000:
|
2014-09-30 10:32:36 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
|
2014-09-30 10:32:36 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0001(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_0001:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm1 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0001:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,0,1]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 0, i32 0, i32 1>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_0020(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0020:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0020:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,2,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 0, i32 2, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_0112(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0112:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vshufpd {{.*#+}} xmm1 = xmm0[1],xmm1[0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0112:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 1, i32 1, i32 2>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_0300(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0300:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm1 = ymm1[0,1,2,2]
|
2014-10-04 06:43:17 +08:00
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0300:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,3,0,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 3, i32 0, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_1000(<4 x i64> %a, <4 x i64> %b) {
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_1000:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} xmm1 = xmm0[1,0]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_1000:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,0,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 1, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_2200(<4 x i64> %a, <4 x i64> %b) {
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_2200:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_2200:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,0,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 2, i32 2, i32 0, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_3330(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_3330:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm1 = ymm1[1,1,2,2]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,0,3,2]
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2],ymm1[3]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_3330:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 3, i32 3, i32 3, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_3210(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_3210:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_3210:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,2,1,0]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-22 11:05:23 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0124(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_0124:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm1 = xmm1[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm1
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1,2],ymm1[3]
|
|
|
|
; AVX1-NEXT: retq
|
2014-09-24 06:39:02 +08:00
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0124:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX2-NEXT: vpbroadcastq %xmm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5],ymm1[6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 1, i32 2, i32 4>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0142(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0142:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-05 19:41:36 +08:00
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,1,2,2]
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2],ymm0[3]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0142:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-05 19:41:36 +08:00
|
|
|
; AVX2-NEXT: vinserti128 $1, %xmm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,2,2]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 1, i32 4, i32 2>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0412(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0412:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
|
|
|
|
; AVX1-NEXT: vshufpd {{.*#+}} xmm2 = xmm0[1],xmm2[0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} ymm1 = ymm1[0,0,2,2]
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0412:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX2-NEXT: vpbroadcastq %xmm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 4, i32 1, i32 2>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_4012(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_4012:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
|
|
|
|
; AVX1-NEXT: vshufpd {{.*#+}} xmm2 = xmm0[1],xmm2[0]
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3]
|
|
|
|
; AVX1-NEXT: retq
|
2014-09-24 06:39:02 +08:00
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_4012:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,2]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5,6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 4, i32 0, i32 1, i32 2>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0145(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-05 19:41:36 +08:00
|
|
|
; ALL-LABEL: shuffle_v4i64_0145:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
|
|
|
; ALL-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 1, i32 4, i32 5>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0451(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_0451:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-11-13 12:06:10 +08:00
|
|
|
; AVX1-NEXT: vunpckhpd {{.*#+}} xmm2 = xmm1[1],xmm0[1]
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_0451:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-05 19:41:36 +08:00
|
|
|
; AVX2-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm1 = ymm1[0,0,1,3]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5],ymm0[6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 4, i32 5, i32 1>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_4501(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-05 19:41:36 +08:00
|
|
|
; ALL-LABEL: shuffle_v4i64_4501:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
|
|
|
|
; ALL-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 4, i32 5, i32 0, i32 1>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-09-21 19:07:41 +08:00
|
|
|
|
2014-08-14 20:13:59 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_4015(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_4015:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-11-13 12:06:10 +08:00
|
|
|
; AVX1-NEXT: vunpckhpd {{.*#+}} xmm2 = xmm0[1],xmm1[1]
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm1[0],xmm0[0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_4015:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-05 19:41:36 +08:00
|
|
|
; AVX2-NEXT: vinserti128 $1, %xmm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,3]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5],ymm1[6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-08-14 20:13:59 +08:00
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 4, i32 0, i32 1, i32 5>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
[x86] Start fixing a really subtle and terrible form of miscompile in
these DAG combines.
The DAG auto-CSE thing is truly terrible. Due to it, when RAUW-ing
a node with its operand, you can cause its uses to CSE to itself, which
then causes their uses to become your uses which causes them to be
picked up by the RAUW. For nodes that are determined to be "no-ops",
this is "fine". But if the RAUW is one of several steps to enact
a transformation, this causes the DAG to really silently eat an discard
nodes that you would never expect. It took days for me to actually
pinpoint a test case triggering this and a really frustrating amount of
time to even comprehend the bug because I never even thought about the
ability of RAUW to iteratively consume nodes due to CSE-ing them into
itself.
To fix this, we have to build up a brand-new chain of operations any
time we are combining across (potentially) intervening nodes. But once
the logic is added to do this, another issue surfaces: CombineTo eagerly
deletes the one node combined, *but no others*. This is... really
frustrating. If deleting it makes its operands become dead, those
operand nodes often won't go onto the worklist in the
order you would want -- they're already on it and not near the top. That
means things higher on the worklist will get combined prior to these
dead nodes being GCed out of the worklist, and if the chain is long, the
immediate users won't be enough to re-detect where the root of the chain
is that became single-use again after deleting the dead nodes. The
better way to do this is to never immediately delete nodes, and instead
to just enqueue them so we can recursively delete them. The
combined-from node is typically not on the worklist anyways by virtue of
having been popped off.... But that in turn breaks other tests that
*require* CombineTo to delete unused nodes. :: sigh ::
Fortunately, there is a better way. This whole routine should have been
returning the replacement rather than using CombineTo which is quite
hacky. Switch to that, and all the pieces fall together.
I suspect the same kind of miscompile is possible in the half-shuffle
folding code, and potentially the recursive folding code. I'll be
switching those over to a pattern more like this one for safety's sake
even though I don't immediately have any test cases for them. Note that
the only way I got a test case for this instance was with *heavily* DAG
combined 256-bit shuffle sequences generated by my fuzzer. ;]
llvm-svn: 216319
2014-08-23 18:25:15 +08:00
|
|
|
|
2014-09-27 01:11:02 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_2u35(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_2u35:
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
|
2014-11-13 12:06:10 +08:00
|
|
|
; AVX1-NEXT: vunpckhpd {{.*#+}} xmm1 = xmm0[1],xmm1[1]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_2u35:
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-05 19:41:36 +08:00
|
|
|
; AVX2-NEXT: vinserti128 $1, %xmm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,1,3,3]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5],ymm1[6,7]
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 2, i32 undef, i32 3, i32 5>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_1251(<4 x i64> %a, <4 x i64> %b) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: shuffle_v4i64_1251:
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm2 = ymm0[2,3,0,1]
|
|
|
|
; AVX1-NEXT: vshufpd {{.*#+}} ymm0 = ymm0[1],ymm2[0],ymm0[2],ymm2[3]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} xmm1 = xmm1[1,0]
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2],ymm0[3]
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: shuffle_v4i64_1251:
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm1 = ymm1[0,1,1,3]
|
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,2,2,1]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
|
2014-09-27 01:11:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 1, i32 2, i32 5, i32 1>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
2014-11-21 20:17:50 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_1054(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_1054:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_1054:
|
|
|
|
; AVX2: # BB#0:
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX2-NEXT: vinserti128 $1, %xmm1, %ymm0, %ymm0
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
|
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 1, i32 0, i32 5, i32 4>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_3254(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_3254:
|
|
|
|
; AVX1: # BB#0:
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_3254:
|
|
|
|
; AVX2: # BB#0:
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
|
|
|
|
; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 3, i32 2, i32 5, i32 4>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_3276(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_3276:
|
|
|
|
; AVX1: # BB#0:
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_3276:
|
|
|
|
; AVX2: # BB#0:
|
2014-11-21 21:56:05 +08:00
|
|
|
; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
|
|
|
|
; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 3, i32 2, i32 7, i32 6>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @shuffle_v4i64_1076(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_1076:
|
|
|
|
; AVX1: # BB#0:
|
2014-11-21 22:33:24 +08:00
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
|
|
|
|
; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_1076:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5,6,7]
|
2014-11-21 22:33:24 +08:00
|
|
|
; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
|
2014-11-21 20:17:50 +08:00
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 1, i32 0, i32 7, i32 6>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
2014-11-22 13:44:43 +08:00
|
|
|
define <4 x i64> @shuffle_v4i64_0415(<4 x i64> %a, <4 x i64> %b) {
|
|
|
|
; AVX1-LABEL: shuffle_v4i64_0415:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vunpckhpd {{.*#+}} xmm2 = xmm0[1],xmm1[1]
|
|
|
|
; AVX1-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: shuffle_v4i64_0415:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm1 = ymm1[0,0,2,1]
|
|
|
|
; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
|
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
|
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> <i32 0, i32 4, i32 1, i32 5>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
[x86] Start fixing a really subtle and terrible form of miscompile in
these DAG combines.
The DAG auto-CSE thing is truly terrible. Due to it, when RAUW-ing
a node with its operand, you can cause its uses to CSE to itself, which
then causes their uses to become your uses which causes them to be
picked up by the RAUW. For nodes that are determined to be "no-ops",
this is "fine". But if the RAUW is one of several steps to enact
a transformation, this causes the DAG to really silently eat an discard
nodes that you would never expect. It took days for me to actually
pinpoint a test case triggering this and a really frustrating amount of
time to even comprehend the bug because I never even thought about the
ability of RAUW to iteratively consume nodes due to CSE-ing them into
itself.
To fix this, we have to build up a brand-new chain of operations any
time we are combining across (potentially) intervening nodes. But once
the logic is added to do this, another issue surfaces: CombineTo eagerly
deletes the one node combined, *but no others*. This is... really
frustrating. If deleting it makes its operands become dead, those
operand nodes often won't go onto the worklist in the
order you would want -- they're already on it and not near the top. That
means things higher on the worklist will get combined prior to these
dead nodes being GCed out of the worklist, and if the chain is long, the
immediate users won't be enough to re-detect where the root of the chain
is that became single-use again after deleting the dead nodes. The
better way to do this is to never immediately delete nodes, and instead
to just enqueue them so we can recursively delete them. The
combined-from node is typically not on the worklist anyways by virtue of
having been popped off.... But that in turn breaks other tests that
*require* CombineTo to delete unused nodes. :: sigh ::
Fortunately, there is a better way. This whole routine should have been
returning the replacement rather than using CombineTo which is quite
hacky. Switch to that, and all the pieces fall together.
I suspect the same kind of miscompile is possible in the half-shuffle
folding code, and potentially the recursive folding code. I'll be
switching those over to a pattern more like this one for safety's sake
even though I don't immediately have any test cases for them. Note that
the only way I got a test case for this instance was with *heavily* DAG
combined 256-bit shuffle sequences generated by my fuzzer. ;]
llvm-svn: 216319
2014-08-23 18:25:15 +08:00
|
|
|
define <4 x i64> @stress_test1(<4 x i64> %a, <4 x i64> %b) {
|
2015-02-04 18:46:53 +08:00
|
|
|
; ALL-LABEL: stress_test1:
|
|
|
|
; ALL: retq
|
[x86] Start fixing a really subtle and terrible form of miscompile in
these DAG combines.
The DAG auto-CSE thing is truly terrible. Due to it, when RAUW-ing
a node with its operand, you can cause its uses to CSE to itself, which
then causes their uses to become your uses which causes them to be
picked up by the RAUW. For nodes that are determined to be "no-ops",
this is "fine". But if the RAUW is one of several steps to enact
a transformation, this causes the DAG to really silently eat an discard
nodes that you would never expect. It took days for me to actually
pinpoint a test case triggering this and a really frustrating amount of
time to even comprehend the bug because I never even thought about the
ability of RAUW to iteratively consume nodes due to CSE-ing them into
itself.
To fix this, we have to build up a brand-new chain of operations any
time we are combining across (potentially) intervening nodes. But once
the logic is added to do this, another issue surfaces: CombineTo eagerly
deletes the one node combined, *but no others*. This is... really
frustrating. If deleting it makes its operands become dead, those
operand nodes often won't go onto the worklist in the
order you would want -- they're already on it and not near the top. That
means things higher on the worklist will get combined prior to these
dead nodes being GCed out of the worklist, and if the chain is long, the
immediate users won't be enough to re-detect where the root of the chain
is that became single-use again after deleting the dead nodes. The
better way to do this is to never immediately delete nodes, and instead
to just enqueue them so we can recursively delete them. The
combined-from node is typically not on the worklist anyways by virtue of
having been popped off.... But that in turn breaks other tests that
*require* CombineTo to delete unused nodes. :: sigh ::
Fortunately, there is a better way. This whole routine should have been
returning the replacement rather than using CombineTo which is quite
hacky. Switch to that, and all the pieces fall together.
I suspect the same kind of miscompile is possible in the half-shuffle
folding code, and potentially the recursive folding code. I'll be
switching those over to a pattern more like this one for safety's sake
even though I don't immediately have any test cases for them. Note that
the only way I got a test case for this instance was with *heavily* DAG
combined 256-bit shuffle sequences generated by my fuzzer. ;]
llvm-svn: 216319
2014-08-23 18:25:15 +08:00
|
|
|
%c = shufflevector <4 x i64> %b, <4 x i64> undef, <4 x i32> <i32 3, i32 1, i32 1, i32 0>
|
|
|
|
%d = shufflevector <4 x i64> %c, <4 x i64> undef, <4 x i32> <i32 3, i32 undef, i32 2, i32 undef>
|
|
|
|
%e = shufflevector <4 x i64> %b, <4 x i64> undef, <4 x i32> <i32 3, i32 3, i32 1, i32 undef>
|
|
|
|
%f = shufflevector <4 x i64> %d, <4 x i64> %e, <4 x i32> <i32 5, i32 1, i32 1, i32 0>
|
|
|
|
|
|
|
|
ret <4 x i64> %f
|
|
|
|
}
|
2014-09-21 20:49:46 +08:00
|
|
|
|
|
|
|
define <4 x i64> @insert_reg_and_zero_v4i64(i64 %a) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: insert_reg_and_zero_v4i64:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovq %rdi, %xmm0
|
2014-10-04 05:38:49 +08:00
|
|
|
; AVX1-NEXT: vxorpd %ymm1, %ymm1, %ymm1
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: insert_reg_and_zero_v4i64:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vmovq %rdi, %xmm0
|
|
|
|
; AVX2-NEXT: vpxor %ymm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5,6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-21 20:49:46 +08:00
|
|
|
%v = insertelement <4 x i64> undef, i64 %a, i64 0
|
|
|
|
%shuffle = shufflevector <4 x i64> %v, <4 x i64> zeroinitializer, <4 x i32> <i32 0, i32 5, i32 6, i32 7>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @insert_mem_and_zero_v4i64(i64* %ptr) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX1-LABEL: insert_mem_and_zero_v4i64:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1: # BB#0:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX1-NEXT: vmovq {{.*#+}} xmm0 = mem[0],zero
|
2014-10-04 05:38:49 +08:00
|
|
|
; AVX1-NEXT: vxorpd %ymm1, %ymm1, %ymm1
|
|
|
|
; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-LABEL: insert_mem_and_zero_v4i64:
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2: # BB#0:
|
2015-01-31 22:09:36 +08:00
|
|
|
; AVX2-NEXT: vmovq {{.*#+}} xmm0 = mem[0],zero
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: vpxor %ymm1, %ymm1, %ymm1
|
2014-10-01 06:04:45 +08:00
|
|
|
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5,6,7]
|
2014-09-24 06:39:02 +08:00
|
|
|
; AVX2-NEXT: retq
|
2014-09-21 20:49:46 +08:00
|
|
|
%a = load i64* %ptr
|
|
|
|
%v = insertelement <4 x i64> undef, i64 %a, i64 0
|
|
|
|
%shuffle = shufflevector <4 x i64> %v, <4 x i64> zeroinitializer, <4 x i32> <i32 0, i32 5, i32 6, i32 7>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @insert_reg_and_zero_v4f64(double %a) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: insert_reg_and_zero_v4f64:
|
2014-09-21 20:49:46 +08:00
|
|
|
; ALL: # BB#0:
|
2015-02-04 18:58:53 +08:00
|
|
|
; ALL-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
2015-01-31 22:09:36 +08:00
|
|
|
; ALL-NEXT: vmovsd {{.*#+}} xmm0 = xmm0[0],xmm1[1]
|
2014-09-21 20:49:46 +08:00
|
|
|
; ALL-NEXT: retq
|
|
|
|
%v = insertelement <4 x double> undef, double %a, i32 0
|
|
|
|
%shuffle = shufflevector <4 x double> %v, <4 x double> zeroinitializer, <4 x i32> <i32 0, i32 5, i32 6, i32 7>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x double> @insert_mem_and_zero_v4f64(double* %ptr) {
|
2014-10-01 06:04:45 +08:00
|
|
|
; ALL-LABEL: insert_mem_and_zero_v4f64:
|
2014-09-21 20:49:46 +08:00
|
|
|
; ALL: # BB#0:
|
2015-01-31 22:09:36 +08:00
|
|
|
; ALL-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero
|
2014-09-21 20:49:46 +08:00
|
|
|
; ALL-NEXT: retq
|
|
|
|
%a = load double* %ptr
|
|
|
|
%v = insertelement <4 x double> undef, double %a, i32 0
|
|
|
|
%shuffle = shufflevector <4 x double> %v, <4 x double> zeroinitializer, <4 x i32> <i32 0, i32 5, i32 6, i32 7>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
2014-10-01 08:41:21 +08:00
|
|
|
|
|
|
|
define <4 x double> @splat_mem_v4f64(double* %ptr) {
|
|
|
|
; ALL-LABEL: splat_mem_v4f64:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vbroadcastsd (%rdi), %ymm0
|
|
|
|
; ALL-NEXT: retq
|
|
|
|
%a = load double* %ptr
|
|
|
|
%v = insertelement <4 x double> undef, double %a, i32 0
|
|
|
|
%shuffle = shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x double> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i64> @splat_mem_v4i64(i64* %ptr) {
|
|
|
|
; AVX1-LABEL: splat_mem_v4i64:
|
|
|
|
; AVX1: # BB#0:
|
2015-02-04 18:46:53 +08:00
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0]
|
2014-10-01 08:41:21 +08:00
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
|
|
|
; AVX2-LABEL: splat_mem_v4i64:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vbroadcastsd (%rdi), %ymm0
|
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%a = load i64* %ptr
|
|
|
|
%v = insertelement <4 x i64> undef, i64 %a, i64 0
|
|
|
|
%shuffle = shufflevector <4 x i64> %v, <4 x i64> undef, <4 x i32> <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i64> %shuffle
|
|
|
|
}
|
2014-10-14 00:16:16 +08:00
|
|
|
|
|
|
|
define <4 x double> @splat_mem_v4f64_2(double* %p) {
|
|
|
|
; ALL-LABEL: splat_mem_v4f64_2:
|
|
|
|
; ALL: # BB#0:
|
|
|
|
; ALL-NEXT: vbroadcastsd (%rdi), %ymm0
|
|
|
|
; ALL-NEXT: retq
|
|
|
|
%1 = load double* %p
|
|
|
|
%2 = insertelement <2 x double> undef, double %1, i32 0
|
|
|
|
%3 = shufflevector <2 x double> %2, <2 x double> undef, <4 x i32> zeroinitializer
|
|
|
|
ret <4 x double> %3
|
|
|
|
}
|
|
|
|
|
2015-01-31 22:09:36 +08:00
|
|
|
define <4 x double> @splat_v4f64(<2 x double> %r) {
|
|
|
|
; AVX1-LABEL: splat_v4f64:
|
|
|
|
; AVX1: # BB#0:
|
|
|
|
; AVX1-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
|
|
|
|
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
|
|
|
|
; AVX1-NEXT: retq
|
|
|
|
;
|
2014-10-14 00:16:16 +08:00
|
|
|
; AVX2-LABEL: splat_v4f64:
|
|
|
|
; AVX2: # BB#0:
|
|
|
|
; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
|
|
|
|
; AVX2-NEXT: retq
|
|
|
|
%1 = shufflevector <2 x double> %r, <2 x double> undef, <4 x i32> zeroinitializer
|
|
|
|
ret <4 x double> %1
|
|
|
|
}
|