diff --git a/llvm/test/Transforms/SLPVectorizer/X86/hadd.ll b/llvm/test/Transforms/SLPVectorizer/X86/hadd.ll new file mode 100644 index 000000000000..6cf2b72baed8 --- /dev/null +++ b/llvm/test/Transforms/SLPVectorizer/X86/hadd.ll @@ -0,0 +1,547 @@ +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py +; RUN: opt < %s -mtriple=x86_64-unknown -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=SSE +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=slm -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=SLM +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=corei7-avx -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX --check-prefix=AVX1 +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=core-avx2 -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX --check-prefix=AVX2 +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=knl -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512 --check-prefix=AVX512F +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=skx -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512 --check-prefix=AVX512BW + +; +; 128-bit vectors +; + +define <2 x double> @test_v2f64(<2 x double> %a, <2 x double> %b) { +; SSE-LABEL: @test_v2f64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; SSE-NEXT: ret <2 x double> [[TMP3]] +; +; SLM-LABEL: @test_v2f64( +; SLM-NEXT: [[A0:%.*]] = extractelement <2 x double> [[A:%.*]], i32 0 +; SLM-NEXT: [[A1:%.*]] = extractelement <2 x double> [[A]], i32 1 +; SLM-NEXT: [[B0:%.*]] = extractelement <2 x double> [[B:%.*]], i32 0 +; SLM-NEXT: [[B1:%.*]] = extractelement <2 x double> [[B]], i32 1 +; SLM-NEXT: [[R0:%.*]] = fadd double [[A0]], [[A1]] +; SLM-NEXT: [[R1:%.*]] = fadd double [[B0]], [[B1]] +; SLM-NEXT: [[R00:%.*]] = insertelement <2 x double> undef, double [[R0]], i32 0 +; SLM-NEXT: [[R01:%.*]] = insertelement <2 x double> [[R00]], double [[R1]], i32 1 +; SLM-NEXT: ret <2 x double> [[R01]] +; +; AVX-LABEL: @test_v2f64( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; AVX-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <2 x double> [[TMP3]] +; +; AVX512-LABEL: @test_v2f64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <2 x double> [[TMP3]] +; + %a0 = extractelement <2 x double> %a, i32 0 + %a1 = extractelement <2 x double> %a, i32 1 + %b0 = extractelement <2 x double> %b, i32 0 + %b1 = extractelement <2 x double> %b, i32 1 + %r0 = fadd double %a0, %a1 + %r1 = fadd double %b0, %b1 + %r00 = insertelement <2 x double> undef, double %r0, i32 0 + %r01 = insertelement <2 x double> %r00, double %r1, i32 1 + ret <2 x double> %r01 +} + +define <4 x float> @test_v4f32(<4 x float> %a, <4 x float> %b) { +; CHECK-LABEL: @test_v4f32( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <4 x float> [[A:%.*]], <4 x float> [[B:%.*]], <4 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x float> [[A]], <4 x float> [[B]], <4 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = fadd <4 x float> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <4 x float> [[TMP3]] +; + %a0 = extractelement <4 x float> %a, i32 0 + %a1 = extractelement <4 x float> %a, i32 1 + %a2 = extractelement <4 x float> %a, i32 2 + %a3 = extractelement <4 x float> %a, i32 3 + %b0 = extractelement <4 x float> %b, i32 0 + %b1 = extractelement <4 x float> %b, i32 1 + %b2 = extractelement <4 x float> %b, i32 2 + %b3 = extractelement <4 x float> %b, i32 3 + %r0 = fadd float %a0, %a1 + %r1 = fadd float %a2, %a3 + %r2 = fadd float %b0, %b1 + %r3 = fadd float %b2, %b3 + %r00 = insertelement <4 x float> undef, float %r0, i32 0 + %r01 = insertelement <4 x float> %r00, float %r1, i32 1 + %r02 = insertelement <4 x float> %r01, float %r2, i32 2 + %r03 = insertelement <4 x float> %r02, float %r3, i32 3 + ret <4 x float> %r03 +} + +define <2 x i64> @test_v2i64(<2 x i64> %a, <2 x i64> %b) { +; SSE-LABEL: @test_v2i64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; SSE-NEXT: ret <2 x i64> [[TMP3]] +; +; SLM-LABEL: @test_v2i64( +; SLM-NEXT: [[A0:%.*]] = extractelement <2 x i64> [[A:%.*]], i32 0 +; SLM-NEXT: [[A1:%.*]] = extractelement <2 x i64> [[A]], i32 1 +; SLM-NEXT: [[B0:%.*]] = extractelement <2 x i64> [[B:%.*]], i32 0 +; SLM-NEXT: [[B1:%.*]] = extractelement <2 x i64> [[B]], i32 1 +; SLM-NEXT: [[R0:%.*]] = add i64 [[A0]], [[A1]] +; SLM-NEXT: [[R1:%.*]] = add i64 [[B0]], [[B1]] +; SLM-NEXT: [[R00:%.*]] = insertelement <2 x i64> undef, i64 [[R0]], i32 0 +; SLM-NEXT: [[R01:%.*]] = insertelement <2 x i64> [[R00]], i64 [[R1]], i32 1 +; SLM-NEXT: ret <2 x i64> [[R01]] +; +; AVX-LABEL: @test_v2i64( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; AVX-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <2 x i64> [[TMP3]] +; +; AVX512-LABEL: @test_v2i64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <2 x i64> [[TMP3]] +; + %a0 = extractelement <2 x i64> %a, i32 0 + %a1 = extractelement <2 x i64> %a, i32 1 + %b0 = extractelement <2 x i64> %b, i32 0 + %b1 = extractelement <2 x i64> %b, i32 1 + %r0 = add i64 %a0, %a1 + %r1 = add i64 %b0, %b1 + %r00 = insertelement <2 x i64> undef, i64 %r0, i32 0 + %r01 = insertelement <2 x i64> %r00, i64 %r1, i32 1 + ret <2 x i64> %r01 +} + +define <4 x i32> @test_v4i32(<4 x i32> %a, <4 x i32> %b) { +; CHECK-LABEL: @test_v4i32( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]], <4 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x i32> [[A]], <4 x i32> [[B]], <4 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = add <4 x i32> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <4 x i32> [[TMP3]] +; + %a0 = extractelement <4 x i32> %a, i32 0 + %a1 = extractelement <4 x i32> %a, i32 1 + %a2 = extractelement <4 x i32> %a, i32 2 + %a3 = extractelement <4 x i32> %a, i32 3 + %b0 = extractelement <4 x i32> %b, i32 0 + %b1 = extractelement <4 x i32> %b, i32 1 + %b2 = extractelement <4 x i32> %b, i32 2 + %b3 = extractelement <4 x i32> %b, i32 3 + %r0 = add i32 %a0, %a1 + %r1 = add i32 %a2, %a3 + %r2 = add i32 %b0, %b1 + %r3 = add i32 %b2, %b3 + %r00 = insertelement <4 x i32> undef, i32 %r0, i32 0 + %r01 = insertelement <4 x i32> %r00, i32 %r1, i32 1 + %r02 = insertelement <4 x i32> %r01, i32 %r2, i32 2 + %r03 = insertelement <4 x i32> %r02, i32 %r3, i32 3 + ret <4 x i32> %r03 +} + +define <8 x i16> @test_v8i16(<8 x i16> %a, <8 x i16> %b) { +; CHECK-LABEL: @test_v8i16( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <8 x i16> [[A:%.*]], <8 x i16> [[B:%.*]], <8 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <8 x i16> [[A]], <8 x i16> [[B]], <8 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = add <8 x i16> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <8 x i16> [[TMP3]] +; + %a0 = extractelement <8 x i16> %a, i32 0 + %a1 = extractelement <8 x i16> %a, i32 1 + %a2 = extractelement <8 x i16> %a, i32 2 + %a3 = extractelement <8 x i16> %a, i32 3 + %a4 = extractelement <8 x i16> %a, i32 4 + %a5 = extractelement <8 x i16> %a, i32 5 + %a6 = extractelement <8 x i16> %a, i32 6 + %a7 = extractelement <8 x i16> %a, i32 7 + %b0 = extractelement <8 x i16> %b, i32 0 + %b1 = extractelement <8 x i16> %b, i32 1 + %b2 = extractelement <8 x i16> %b, i32 2 + %b3 = extractelement <8 x i16> %b, i32 3 + %b4 = extractelement <8 x i16> %b, i32 4 + %b5 = extractelement <8 x i16> %b, i32 5 + %b6 = extractelement <8 x i16> %b, i32 6 + %b7 = extractelement <8 x i16> %b, i32 7 + %r0 = add i16 %a0, %a1 + %r1 = add i16 %a2, %a3 + %r2 = add i16 %a4, %a5 + %r3 = add i16 %a6, %a7 + %r4 = add i16 %b0, %b1 + %r5 = add i16 %b2, %b3 + %r6 = add i16 %b4, %b5 + %r7 = add i16 %b6, %b7 + %r00 = insertelement <8 x i16> undef, i16 %r0, i32 0 + %r01 = insertelement <8 x i16> %r00, i16 %r1, i32 1 + %r02 = insertelement <8 x i16> %r01, i16 %r2, i32 2 + %r03 = insertelement <8 x i16> %r02, i16 %r3, i32 3 + %r04 = insertelement <8 x i16> %r03, i16 %r4, i32 4 + %r05 = insertelement <8 x i16> %r04, i16 %r5, i32 5 + %r06 = insertelement <8 x i16> %r05, i16 %r6, i32 6 + %r07 = insertelement <8 x i16> %r06, i16 %r7, i32 7 + ret <8 x i16> %r07 +} + +; +; 256-bit vectors +; + +define <4 x double> @test_v4f64(<4 x double> %a, <4 x double> %b) { +; SSE-LABEL: @test_v4f64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP6:%.*]] = fadd <2 x double> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; SSE-NEXT: ret <4 x double> [[R03]] +; +; SLM-LABEL: @test_v4f64( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP6:%.*]] = fadd <2 x double> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; SLM-NEXT: ret <4 x double> [[R03]] +; +; AVX1-LABEL: @test_v4f64( +; AVX1-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; AVX1-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP3:%.*]] = fadd <2 x double> [[TMP1]], [[TMP2]] +; AVX1-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP6:%.*]] = fadd <2 x double> [[TMP4]], [[TMP5]] +; AVX1-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; AVX1-NEXT: ret <4 x double> [[R03]] +; +; AVX2-LABEL: @test_v4f64( +; AVX2-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <4 x i32> +; AVX2-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <4 x i32> +; AVX2-NEXT: [[TMP3:%.*]] = fadd <4 x double> [[TMP1]], [[TMP2]] +; AVX2-NEXT: ret <4 x double> [[TMP3]] +; +; AVX512-LABEL: @test_v4f64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <4 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <4 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fadd <4 x double> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <4 x double> [[TMP3]] +; + %a0 = extractelement <4 x double> %a, i32 0 + %a1 = extractelement <4 x double> %a, i32 1 + %a2 = extractelement <4 x double> %a, i32 2 + %a3 = extractelement <4 x double> %a, i32 3 + %b0 = extractelement <4 x double> %b, i32 0 + %b1 = extractelement <4 x double> %b, i32 1 + %b2 = extractelement <4 x double> %b, i32 2 + %b3 = extractelement <4 x double> %b, i32 3 + %r0 = fadd double %a0, %a1 + %r1 = fadd double %b0, %b1 + %r2 = fadd double %a2, %a3 + %r3 = fadd double %b2, %b3 + %r00 = insertelement <4 x double> undef, double %r0, i32 0 + %r01 = insertelement <4 x double> %r00, double %r1, i32 1 + %r02 = insertelement <4 x double> %r01, double %r2, i32 2 + %r03 = insertelement <4 x double> %r02, double %r3, i32 3 + ret <4 x double> %r03 +} + +define <8 x float> @test_v8f32(<8 x float> %a, <8 x float> %b) { +; SSE-LABEL: @test_v8f32( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <4 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fadd <4 x float> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP6:%.*]] = fadd <4 x float> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R07:%.*]] = shufflevector <4 x float> [[TMP3]], <4 x float> [[TMP6]], <8 x i32> +; SSE-NEXT: ret <8 x float> [[R07]] +; +; SLM-LABEL: @test_v8f32( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <4 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP3:%.*]] = fadd <4 x float> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP6:%.*]] = fadd <4 x float> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R07:%.*]] = shufflevector <4 x float> [[TMP3]], <4 x float> [[TMP6]], <8 x i32> +; SLM-NEXT: ret <8 x float> [[R07]] +; +; AVX-LABEL: @test_v8f32( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <8 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <8 x i32> +; AVX-NEXT: [[TMP3:%.*]] = fadd <8 x float> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <8 x float> [[TMP3]] +; +; AVX512-LABEL: @test_v8f32( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <8 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <8 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fadd <8 x float> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <8 x float> [[TMP3]] +; + %a0 = extractelement <8 x float> %a, i32 0 + %a1 = extractelement <8 x float> %a, i32 1 + %a2 = extractelement <8 x float> %a, i32 2 + %a3 = extractelement <8 x float> %a, i32 3 + %a4 = extractelement <8 x float> %a, i32 4 + %a5 = extractelement <8 x float> %a, i32 5 + %a6 = extractelement <8 x float> %a, i32 6 + %a7 = extractelement <8 x float> %a, i32 7 + %b0 = extractelement <8 x float> %b, i32 0 + %b1 = extractelement <8 x float> %b, i32 1 + %b2 = extractelement <8 x float> %b, i32 2 + %b3 = extractelement <8 x float> %b, i32 3 + %b4 = extractelement <8 x float> %b, i32 4 + %b5 = extractelement <8 x float> %b, i32 5 + %b6 = extractelement <8 x float> %b, i32 6 + %b7 = extractelement <8 x float> %b, i32 7 + %r0 = fadd float %a0, %a1 + %r1 = fadd float %a2, %a3 + %r2 = fadd float %b0, %b1 + %r3 = fadd float %b2, %b3 + %r4 = fadd float %a4, %a5 + %r5 = fadd float %a6, %a7 + %r6 = fadd float %b4, %b5 + %r7 = fadd float %b6, %b7 + %r00 = insertelement <8 x float> undef, float %r0, i32 0 + %r01 = insertelement <8 x float> %r00, float %r1, i32 1 + %r02 = insertelement <8 x float> %r01, float %r2, i32 2 + %r03 = insertelement <8 x float> %r02, float %r3, i32 3 + %r04 = insertelement <8 x float> %r03, float %r4, i32 4 + %r05 = insertelement <8 x float> %r04, float %r5, i32 5 + %r06 = insertelement <8 x float> %r05, float %r6, i32 6 + %r07 = insertelement <8 x float> %r06, float %r7, i32 7 + ret <8 x float> %r07 +} + +define <4 x i64> @test_v4i64(<4 x i64> %a, <4 x i64> %b) { +; SSE-LABEL: @test_v4i64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP6:%.*]] = add <2 x i64> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; SSE-NEXT: ret <4 x i64> [[R03]] +; +; SLM-LABEL: @test_v4i64( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP6:%.*]] = add <2 x i64> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; SLM-NEXT: ret <4 x i64> [[R03]] +; +; AVX1-LABEL: @test_v4i64( +; AVX1-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; AVX1-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP3:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] +; AVX1-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP6:%.*]] = add <2 x i64> [[TMP4]], [[TMP5]] +; AVX1-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; AVX1-NEXT: ret <4 x i64> [[R03]] +; +; AVX2-LABEL: @test_v4i64( +; AVX2-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <4 x i32> +; AVX2-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <4 x i32> +; AVX2-NEXT: [[TMP3:%.*]] = add <4 x i64> [[TMP1]], [[TMP2]] +; AVX2-NEXT: ret <4 x i64> [[TMP3]] +; +; AVX512-LABEL: @test_v4i64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <4 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <4 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = add <4 x i64> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <4 x i64> [[TMP3]] +; + %a0 = extractelement <4 x i64> %a, i32 0 + %a1 = extractelement <4 x i64> %a, i32 1 + %a2 = extractelement <4 x i64> %a, i32 2 + %a3 = extractelement <4 x i64> %a, i32 3 + %b0 = extractelement <4 x i64> %b, i32 0 + %b1 = extractelement <4 x i64> %b, i32 1 + %b2 = extractelement <4 x i64> %b, i32 2 + %b3 = extractelement <4 x i64> %b, i32 3 + %r0 = add i64 %a0, %a1 + %r1 = add i64 %b0, %b1 + %r2 = add i64 %a2, %a3 + %r3 = add i64 %b2, %b3 + %r00 = insertelement <4 x i64> undef, i64 %r0, i32 0 + %r01 = insertelement <4 x i64> %r00, i64 %r1, i32 1 + %r02 = insertelement <4 x i64> %r01, i64 %r2, i32 2 + %r03 = insertelement <4 x i64> %r02, i64 %r3, i32 3 + ret <4 x i64> %r03 +} + +define <8 x i32> @test_v8i32(<8 x i32> %a, <8 x i32> %b) { +; SSE-LABEL: @test_v8i32( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <4 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP3:%.*]] = add <4 x i32> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP6:%.*]] = add <4 x i32> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R07:%.*]] = shufflevector <4 x i32> [[TMP3]], <4 x i32> [[TMP6]], <8 x i32> +; SSE-NEXT: ret <8 x i32> [[R07]] +; +; SLM-LABEL: @test_v8i32( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <4 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP3:%.*]] = add <4 x i32> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP6:%.*]] = add <4 x i32> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R07:%.*]] = shufflevector <4 x i32> [[TMP3]], <4 x i32> [[TMP6]], <8 x i32> +; SLM-NEXT: ret <8 x i32> [[R07]] +; +; AVX-LABEL: @test_v8i32( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <8 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <8 x i32> +; AVX-NEXT: [[TMP3:%.*]] = add <8 x i32> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <8 x i32> [[TMP3]] +; +; AVX512-LABEL: @test_v8i32( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <8 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <8 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = add <8 x i32> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <8 x i32> [[TMP3]] +; + %a0 = extractelement <8 x i32> %a, i32 0 + %a1 = extractelement <8 x i32> %a, i32 1 + %a2 = extractelement <8 x i32> %a, i32 2 + %a3 = extractelement <8 x i32> %a, i32 3 + %a4 = extractelement <8 x i32> %a, i32 4 + %a5 = extractelement <8 x i32> %a, i32 5 + %a6 = extractelement <8 x i32> %a, i32 6 + %a7 = extractelement <8 x i32> %a, i32 7 + %b0 = extractelement <8 x i32> %b, i32 0 + %b1 = extractelement <8 x i32> %b, i32 1 + %b2 = extractelement <8 x i32> %b, i32 2 + %b3 = extractelement <8 x i32> %b, i32 3 + %b4 = extractelement <8 x i32> %b, i32 4 + %b5 = extractelement <8 x i32> %b, i32 5 + %b6 = extractelement <8 x i32> %b, i32 6 + %b7 = extractelement <8 x i32> %b, i32 7 + %r0 = add i32 %a0, %a1 + %r1 = add i32 %a2, %a3 + %r2 = add i32 %b0, %b1 + %r3 = add i32 %b2, %b3 + %r4 = add i32 %a4, %a5 + %r5 = add i32 %a6, %a7 + %r6 = add i32 %b4, %b5 + %r7 = add i32 %b6, %b7 + %r00 = insertelement <8 x i32> undef, i32 %r0, i32 0 + %r01 = insertelement <8 x i32> %r00, i32 %r1, i32 1 + %r02 = insertelement <8 x i32> %r01, i32 %r2, i32 2 + %r03 = insertelement <8 x i32> %r02, i32 %r3, i32 3 + %r04 = insertelement <8 x i32> %r03, i32 %r4, i32 4 + %r05 = insertelement <8 x i32> %r04, i32 %r5, i32 5 + %r06 = insertelement <8 x i32> %r05, i32 %r6, i32 6 + %r07 = insertelement <8 x i32> %r06, i32 %r7, i32 7 + ret <8 x i32> %r07 +} + +define <16 x i16> @test_v16i16(<16 x i16> %a, <16 x i16> %b) { +; SSE-LABEL: @test_v16i16( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <8 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP3:%.*]] = add <8 x i16> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP6:%.*]] = add <8 x i16> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[RV15:%.*]] = shufflevector <8 x i16> [[TMP3]], <8 x i16> [[TMP6]], <16 x i32> +; SSE-NEXT: ret <16 x i16> [[RV15]] +; +; SLM-LABEL: @test_v16i16( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; SLM-NEXT: [[TMP3:%.*]] = add <16 x i16> [[TMP1]], [[TMP2]] +; SLM-NEXT: ret <16 x i16> [[TMP3]] +; +; AVX-LABEL: @test_v16i16( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; AVX-NEXT: [[TMP3:%.*]] = add <16 x i16> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <16 x i16> [[TMP3]] +; +; AVX512-LABEL: @test_v16i16( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = add <16 x i16> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <16 x i16> [[TMP3]] +; + %a0 = extractelement <16 x i16> %a, i32 0 + %a1 = extractelement <16 x i16> %a, i32 1 + %a2 = extractelement <16 x i16> %a, i32 2 + %a3 = extractelement <16 x i16> %a, i32 3 + %a4 = extractelement <16 x i16> %a, i32 4 + %a5 = extractelement <16 x i16> %a, i32 5 + %a6 = extractelement <16 x i16> %a, i32 6 + %a7 = extractelement <16 x i16> %a, i32 7 + %a8 = extractelement <16 x i16> %a, i32 8 + %a9 = extractelement <16 x i16> %a, i32 9 + %a10 = extractelement <16 x i16> %a, i32 10 + %a11 = extractelement <16 x i16> %a, i32 11 + %a12 = extractelement <16 x i16> %a, i32 12 + %a13 = extractelement <16 x i16> %a, i32 13 + %a14 = extractelement <16 x i16> %a, i32 14 + %a15 = extractelement <16 x i16> %a, i32 15 + %b0 = extractelement <16 x i16> %b, i32 0 + %b1 = extractelement <16 x i16> %b, i32 1 + %b2 = extractelement <16 x i16> %b, i32 2 + %b3 = extractelement <16 x i16> %b, i32 3 + %b4 = extractelement <16 x i16> %b, i32 4 + %b5 = extractelement <16 x i16> %b, i32 5 + %b6 = extractelement <16 x i16> %b, i32 6 + %b7 = extractelement <16 x i16> %b, i32 7 + %b8 = extractelement <16 x i16> %b, i32 8 + %b9 = extractelement <16 x i16> %b, i32 9 + %b10 = extractelement <16 x i16> %b, i32 10 + %b11 = extractelement <16 x i16> %b, i32 11 + %b12 = extractelement <16 x i16> %b, i32 12 + %b13 = extractelement <16 x i16> %b, i32 13 + %b14 = extractelement <16 x i16> %b, i32 14 + %b15 = extractelement <16 x i16> %b, i32 15 + %r0 = add i16 %a0 , %a1 + %r1 = add i16 %a2 , %a3 + %r2 = add i16 %a4 , %a5 + %r3 = add i16 %a6 , %a7 + %r4 = add i16 %b0 , %b1 + %r5 = add i16 %b2 , %b3 + %r6 = add i16 %b4 , %b5 + %r7 = add i16 %b6 , %b7 + %r8 = add i16 %a8 , %a9 + %r9 = add i16 %a10, %a11 + %r10 = add i16 %a12, %a13 + %r11 = add i16 %a14, %a15 + %r12 = add i16 %b8 , %b9 + %r13 = add i16 %b10, %b11 + %r14 = add i16 %b12, %b13 + %r15 = add i16 %b14, %b15 + %rv0 = insertelement <16 x i16> undef, i16 %r0 , i32 0 + %rv1 = insertelement <16 x i16> %rv0 , i16 %r1 , i32 1 + %rv2 = insertelement <16 x i16> %rv1 , i16 %r2 , i32 2 + %rv3 = insertelement <16 x i16> %rv2 , i16 %r3 , i32 3 + %rv4 = insertelement <16 x i16> %rv3 , i16 %r4 , i32 4 + %rv5 = insertelement <16 x i16> %rv4 , i16 %r5 , i32 5 + %rv6 = insertelement <16 x i16> %rv5 , i16 %r6 , i32 6 + %rv7 = insertelement <16 x i16> %rv6 , i16 %r7 , i32 7 + %rv8 = insertelement <16 x i16> %rv7 , i16 %r8 , i32 8 + %rv9 = insertelement <16 x i16> %rv8 , i16 %r9 , i32 9 + %rv10 = insertelement <16 x i16> %rv9 , i16 %r10, i32 10 + %rv11 = insertelement <16 x i16> %rv10, i16 %r11, i32 11 + %rv12 = insertelement <16 x i16> %rv11, i16 %r12, i32 12 + %rv13 = insertelement <16 x i16> %rv12, i16 %r13, i32 13 + %rv14 = insertelement <16 x i16> %rv13, i16 %r14, i32 14 + %rv15 = insertelement <16 x i16> %rv14, i16 %r15, i32 15 + ret <16 x i16> %rv15 +} diff --git a/llvm/test/Transforms/SLPVectorizer/X86/hsub.ll b/llvm/test/Transforms/SLPVectorizer/X86/hsub.ll new file mode 100644 index 000000000000..e936dc29d452 --- /dev/null +++ b/llvm/test/Transforms/SLPVectorizer/X86/hsub.ll @@ -0,0 +1,547 @@ +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py +; RUN: opt < %s -mtriple=x86_64-unknown -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=SSE +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=slm -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=SLM +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=corei7-avx -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX --check-prefix=AVX1 +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=core-avx2 -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX --check-prefix=AVX2 +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=knl -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512 --check-prefix=AVX512F +; RUN: opt < %s -mtriple=x86_64-unknown -mcpu=skx -basicaa -slp-vectorizer -instcombine -S | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512 --check-prefix=AVX512BW + +; +; 128-bit vectors +; + +define <2 x double> @test_v2f64(<2 x double> %a, <2 x double> %b) { +; SSE-LABEL: @test_v2f64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; SSE-NEXT: ret <2 x double> [[TMP3]] +; +; SLM-LABEL: @test_v2f64( +; SLM-NEXT: [[A0:%.*]] = extractelement <2 x double> [[A:%.*]], i32 0 +; SLM-NEXT: [[A1:%.*]] = extractelement <2 x double> [[A]], i32 1 +; SLM-NEXT: [[B0:%.*]] = extractelement <2 x double> [[B:%.*]], i32 0 +; SLM-NEXT: [[B1:%.*]] = extractelement <2 x double> [[B]], i32 1 +; SLM-NEXT: [[R0:%.*]] = fsub double [[A0]], [[A1]] +; SLM-NEXT: [[R1:%.*]] = fsub double [[B0]], [[B1]] +; SLM-NEXT: [[R00:%.*]] = insertelement <2 x double> undef, double [[R0]], i32 0 +; SLM-NEXT: [[R01:%.*]] = insertelement <2 x double> [[R00]], double [[R1]], i32 1 +; SLM-NEXT: ret <2 x double> [[R01]] +; +; AVX-LABEL: @test_v2f64( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; AVX-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <2 x double> [[TMP3]] +; +; AVX512-LABEL: @test_v2f64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <2 x double> [[A:%.*]], <2 x double> [[B:%.*]], <2 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <2 x double> [[A]], <2 x double> [[B]], <2 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <2 x double> [[TMP3]] +; + %a0 = extractelement <2 x double> %a, i32 0 + %a1 = extractelement <2 x double> %a, i32 1 + %b0 = extractelement <2 x double> %b, i32 0 + %b1 = extractelement <2 x double> %b, i32 1 + %r0 = fsub double %a0, %a1 + %r1 = fsub double %b0, %b1 + %r00 = insertelement <2 x double> undef, double %r0, i32 0 + %r01 = insertelement <2 x double> %r00, double %r1, i32 1 + ret <2 x double> %r01 +} + +define <4 x float> @test_v4f32(<4 x float> %a, <4 x float> %b) { +; CHECK-LABEL: @test_v4f32( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <4 x float> [[A:%.*]], <4 x float> [[B:%.*]], <4 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x float> [[A]], <4 x float> [[B]], <4 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = fsub <4 x float> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <4 x float> [[TMP3]] +; + %a0 = extractelement <4 x float> %a, i32 0 + %a1 = extractelement <4 x float> %a, i32 1 + %a2 = extractelement <4 x float> %a, i32 2 + %a3 = extractelement <4 x float> %a, i32 3 + %b0 = extractelement <4 x float> %b, i32 0 + %b1 = extractelement <4 x float> %b, i32 1 + %b2 = extractelement <4 x float> %b, i32 2 + %b3 = extractelement <4 x float> %b, i32 3 + %r0 = fsub float %a0, %a1 + %r1 = fsub float %a2, %a3 + %r2 = fsub float %b0, %b1 + %r3 = fsub float %b2, %b3 + %r00 = insertelement <4 x float> undef, float %r0, i32 0 + %r01 = insertelement <4 x float> %r00, float %r1, i32 1 + %r02 = insertelement <4 x float> %r01, float %r2, i32 2 + %r03 = insertelement <4 x float> %r02, float %r3, i32 3 + ret <4 x float> %r03 +} + +define <2 x i64> @test_v2i64(<2 x i64> %a, <2 x i64> %b) { +; SSE-LABEL: @test_v2i64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; SSE-NEXT: ret <2 x i64> [[TMP3]] +; +; SLM-LABEL: @test_v2i64( +; SLM-NEXT: [[A0:%.*]] = extractelement <2 x i64> [[A:%.*]], i32 0 +; SLM-NEXT: [[A1:%.*]] = extractelement <2 x i64> [[A]], i32 1 +; SLM-NEXT: [[B0:%.*]] = extractelement <2 x i64> [[B:%.*]], i32 0 +; SLM-NEXT: [[B1:%.*]] = extractelement <2 x i64> [[B]], i32 1 +; SLM-NEXT: [[R0:%.*]] = sub i64 [[A0]], [[A1]] +; SLM-NEXT: [[R1:%.*]] = sub i64 [[B0]], [[B1]] +; SLM-NEXT: [[R00:%.*]] = insertelement <2 x i64> undef, i64 [[R0]], i32 0 +; SLM-NEXT: [[R01:%.*]] = insertelement <2 x i64> [[R00]], i64 [[R1]], i32 1 +; SLM-NEXT: ret <2 x i64> [[R01]] +; +; AVX-LABEL: @test_v2i64( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; AVX-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <2 x i64> [[TMP3]] +; +; AVX512-LABEL: @test_v2i64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <2 x i64> [[A:%.*]], <2 x i64> [[B:%.*]], <2 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <2 x i64> [[A]], <2 x i64> [[B]], <2 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <2 x i64> [[TMP3]] +; + %a0 = extractelement <2 x i64> %a, i32 0 + %a1 = extractelement <2 x i64> %a, i32 1 + %b0 = extractelement <2 x i64> %b, i32 0 + %b1 = extractelement <2 x i64> %b, i32 1 + %r0 = sub i64 %a0, %a1 + %r1 = sub i64 %b0, %b1 + %r00 = insertelement <2 x i64> undef, i64 %r0, i32 0 + %r01 = insertelement <2 x i64> %r00, i64 %r1, i32 1 + ret <2 x i64> %r01 +} + +define <4 x i32> @test_v4i32(<4 x i32> %a, <4 x i32> %b) { +; CHECK-LABEL: @test_v4i32( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]], <4 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x i32> [[A]], <4 x i32> [[B]], <4 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = sub <4 x i32> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <4 x i32> [[TMP3]] +; + %a0 = extractelement <4 x i32> %a, i32 0 + %a1 = extractelement <4 x i32> %a, i32 1 + %a2 = extractelement <4 x i32> %a, i32 2 + %a3 = extractelement <4 x i32> %a, i32 3 + %b0 = extractelement <4 x i32> %b, i32 0 + %b1 = extractelement <4 x i32> %b, i32 1 + %b2 = extractelement <4 x i32> %b, i32 2 + %b3 = extractelement <4 x i32> %b, i32 3 + %r0 = sub i32 %a0, %a1 + %r1 = sub i32 %a2, %a3 + %r2 = sub i32 %b0, %b1 + %r3 = sub i32 %b2, %b3 + %r00 = insertelement <4 x i32> undef, i32 %r0, i32 0 + %r01 = insertelement <4 x i32> %r00, i32 %r1, i32 1 + %r02 = insertelement <4 x i32> %r01, i32 %r2, i32 2 + %r03 = insertelement <4 x i32> %r02, i32 %r3, i32 3 + ret <4 x i32> %r03 +} + +define <8 x i16> @test_v8i16(<8 x i16> %a, <8 x i16> %b) { +; CHECK-LABEL: @test_v8i16( +; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <8 x i16> [[A:%.*]], <8 x i16> [[B:%.*]], <8 x i32> +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <8 x i16> [[A]], <8 x i16> [[B]], <8 x i32> +; CHECK-NEXT: [[TMP3:%.*]] = sub <8 x i16> [[TMP1]], [[TMP2]] +; CHECK-NEXT: ret <8 x i16> [[TMP3]] +; + %a0 = extractelement <8 x i16> %a, i32 0 + %a1 = extractelement <8 x i16> %a, i32 1 + %a2 = extractelement <8 x i16> %a, i32 2 + %a3 = extractelement <8 x i16> %a, i32 3 + %a4 = extractelement <8 x i16> %a, i32 4 + %a5 = extractelement <8 x i16> %a, i32 5 + %a6 = extractelement <8 x i16> %a, i32 6 + %a7 = extractelement <8 x i16> %a, i32 7 + %b0 = extractelement <8 x i16> %b, i32 0 + %b1 = extractelement <8 x i16> %b, i32 1 + %b2 = extractelement <8 x i16> %b, i32 2 + %b3 = extractelement <8 x i16> %b, i32 3 + %b4 = extractelement <8 x i16> %b, i32 4 + %b5 = extractelement <8 x i16> %b, i32 5 + %b6 = extractelement <8 x i16> %b, i32 6 + %b7 = extractelement <8 x i16> %b, i32 7 + %r0 = sub i16 %a0, %a1 + %r1 = sub i16 %a2, %a3 + %r2 = sub i16 %a4, %a5 + %r3 = sub i16 %a6, %a7 + %r4 = sub i16 %b0, %b1 + %r5 = sub i16 %b2, %b3 + %r6 = sub i16 %b4, %b5 + %r7 = sub i16 %b6, %b7 + %r00 = insertelement <8 x i16> undef, i16 %r0, i32 0 + %r01 = insertelement <8 x i16> %r00, i16 %r1, i32 1 + %r02 = insertelement <8 x i16> %r01, i16 %r2, i32 2 + %r03 = insertelement <8 x i16> %r02, i16 %r3, i32 3 + %r04 = insertelement <8 x i16> %r03, i16 %r4, i32 4 + %r05 = insertelement <8 x i16> %r04, i16 %r5, i32 5 + %r06 = insertelement <8 x i16> %r05, i16 %r6, i32 6 + %r07 = insertelement <8 x i16> %r06, i16 %r7, i32 7 + ret <8 x i16> %r07 +} + +; +; 256-bit vectors +; + +define <4 x double> @test_v4f64(<4 x double> %a, <4 x double> %b) { +; SSE-LABEL: @test_v4f64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SSE-NEXT: [[TMP6:%.*]] = fsub <2 x double> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; SSE-NEXT: ret <4 x double> [[R03]] +; +; SLM-LABEL: @test_v4f64( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; SLM-NEXT: [[TMP6:%.*]] = fsub <2 x double> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; SLM-NEXT: ret <4 x double> [[R03]] +; +; AVX1-LABEL: @test_v4f64( +; AVX1-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <2 x i32> +; AVX1-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP3:%.*]] = fsub <2 x double> [[TMP1]], [[TMP2]] +; AVX1-NEXT: [[TMP4:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP5:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP6:%.*]] = fsub <2 x double> [[TMP4]], [[TMP5]] +; AVX1-NEXT: [[R03:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> [[TMP6]], <4 x i32> +; AVX1-NEXT: ret <4 x double> [[R03]] +; +; AVX2-LABEL: @test_v4f64( +; AVX2-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <4 x i32> +; AVX2-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <4 x i32> +; AVX2-NEXT: [[TMP3:%.*]] = fsub <4 x double> [[TMP1]], [[TMP2]] +; AVX2-NEXT: ret <4 x double> [[TMP3]] +; +; AVX512-LABEL: @test_v4f64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <4 x double> [[A:%.*]], <4 x double> [[B:%.*]], <4 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <4 x double> [[A]], <4 x double> [[B]], <4 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fsub <4 x double> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <4 x double> [[TMP3]] +; + %a0 = extractelement <4 x double> %a, i32 0 + %a1 = extractelement <4 x double> %a, i32 1 + %a2 = extractelement <4 x double> %a, i32 2 + %a3 = extractelement <4 x double> %a, i32 3 + %b0 = extractelement <4 x double> %b, i32 0 + %b1 = extractelement <4 x double> %b, i32 1 + %b2 = extractelement <4 x double> %b, i32 2 + %b3 = extractelement <4 x double> %b, i32 3 + %r0 = fsub double %a0, %a1 + %r1 = fsub double %b0, %b1 + %r2 = fsub double %a2, %a3 + %r3 = fsub double %b2, %b3 + %r00 = insertelement <4 x double> undef, double %r0, i32 0 + %r01 = insertelement <4 x double> %r00, double %r1, i32 1 + %r02 = insertelement <4 x double> %r01, double %r2, i32 2 + %r03 = insertelement <4 x double> %r02, double %r3, i32 3 + ret <4 x double> %r03 +} + +define <8 x float> @test_v8f32(<8 x float> %a, <8 x float> %b) { +; SSE-LABEL: @test_v8f32( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <4 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP3:%.*]] = fsub <4 x float> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SSE-NEXT: [[TMP6:%.*]] = fsub <4 x float> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R07:%.*]] = shufflevector <4 x float> [[TMP3]], <4 x float> [[TMP6]], <8 x i32> +; SSE-NEXT: ret <8 x float> [[R07]] +; +; SLM-LABEL: @test_v8f32( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <4 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP3:%.*]] = fsub <4 x float> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <4 x i32> +; SLM-NEXT: [[TMP6:%.*]] = fsub <4 x float> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R07:%.*]] = shufflevector <4 x float> [[TMP3]], <4 x float> [[TMP6]], <8 x i32> +; SLM-NEXT: ret <8 x float> [[R07]] +; +; AVX-LABEL: @test_v8f32( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <8 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <8 x i32> +; AVX-NEXT: [[TMP3:%.*]] = fsub <8 x float> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <8 x float> [[TMP3]] +; +; AVX512-LABEL: @test_v8f32( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <8 x float> [[A:%.*]], <8 x float> [[B:%.*]], <8 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <8 x float> [[A]], <8 x float> [[B]], <8 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = fsub <8 x float> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <8 x float> [[TMP3]] +; + %a0 = extractelement <8 x float> %a, i32 0 + %a1 = extractelement <8 x float> %a, i32 1 + %a2 = extractelement <8 x float> %a, i32 2 + %a3 = extractelement <8 x float> %a, i32 3 + %a4 = extractelement <8 x float> %a, i32 4 + %a5 = extractelement <8 x float> %a, i32 5 + %a6 = extractelement <8 x float> %a, i32 6 + %a7 = extractelement <8 x float> %a, i32 7 + %b0 = extractelement <8 x float> %b, i32 0 + %b1 = extractelement <8 x float> %b, i32 1 + %b2 = extractelement <8 x float> %b, i32 2 + %b3 = extractelement <8 x float> %b, i32 3 + %b4 = extractelement <8 x float> %b, i32 4 + %b5 = extractelement <8 x float> %b, i32 5 + %b6 = extractelement <8 x float> %b, i32 6 + %b7 = extractelement <8 x float> %b, i32 7 + %r0 = fsub float %a0, %a1 + %r1 = fsub float %a2, %a3 + %r2 = fsub float %b0, %b1 + %r3 = fsub float %b2, %b3 + %r4 = fsub float %a4, %a5 + %r5 = fsub float %a6, %a7 + %r6 = fsub float %b4, %b5 + %r7 = fsub float %b6, %b7 + %r00 = insertelement <8 x float> undef, float %r0, i32 0 + %r01 = insertelement <8 x float> %r00, float %r1, i32 1 + %r02 = insertelement <8 x float> %r01, float %r2, i32 2 + %r03 = insertelement <8 x float> %r02, float %r3, i32 3 + %r04 = insertelement <8 x float> %r03, float %r4, i32 4 + %r05 = insertelement <8 x float> %r04, float %r5, i32 5 + %r06 = insertelement <8 x float> %r05, float %r6, i32 6 + %r07 = insertelement <8 x float> %r06, float %r7, i32 7 + ret <8 x float> %r07 +} + +define <4 x i64> @test_v4i64(<4 x i64> %a, <4 x i64> %b) { +; SSE-LABEL: @test_v4i64( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SSE-NEXT: [[TMP6:%.*]] = sub <2 x i64> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; SSE-NEXT: ret <4 x i64> [[R03]] +; +; SLM-LABEL: @test_v4i64( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; SLM-NEXT: [[TMP6:%.*]] = sub <2 x i64> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; SLM-NEXT: ret <4 x i64> [[R03]] +; +; AVX1-LABEL: @test_v4i64( +; AVX1-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <2 x i32> +; AVX1-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP3:%.*]] = sub <2 x i64> [[TMP1]], [[TMP2]] +; AVX1-NEXT: [[TMP4:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP5:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <2 x i32> +; AVX1-NEXT: [[TMP6:%.*]] = sub <2 x i64> [[TMP4]], [[TMP5]] +; AVX1-NEXT: [[R03:%.*]] = shufflevector <2 x i64> [[TMP3]], <2 x i64> [[TMP6]], <4 x i32> +; AVX1-NEXT: ret <4 x i64> [[R03]] +; +; AVX2-LABEL: @test_v4i64( +; AVX2-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <4 x i32> +; AVX2-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <4 x i32> +; AVX2-NEXT: [[TMP3:%.*]] = sub <4 x i64> [[TMP1]], [[TMP2]] +; AVX2-NEXT: ret <4 x i64> [[TMP3]] +; +; AVX512-LABEL: @test_v4i64( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <4 x i64> [[A:%.*]], <4 x i64> [[B:%.*]], <4 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[A]], <4 x i64> [[B]], <4 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = sub <4 x i64> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <4 x i64> [[TMP3]] +; + %a0 = extractelement <4 x i64> %a, i32 0 + %a1 = extractelement <4 x i64> %a, i32 1 + %a2 = extractelement <4 x i64> %a, i32 2 + %a3 = extractelement <4 x i64> %a, i32 3 + %b0 = extractelement <4 x i64> %b, i32 0 + %b1 = extractelement <4 x i64> %b, i32 1 + %b2 = extractelement <4 x i64> %b, i32 2 + %b3 = extractelement <4 x i64> %b, i32 3 + %r0 = sub i64 %a0, %a1 + %r1 = sub i64 %b0, %b1 + %r2 = sub i64 %a2, %a3 + %r3 = sub i64 %b2, %b3 + %r00 = insertelement <4 x i64> undef, i64 %r0, i32 0 + %r01 = insertelement <4 x i64> %r00, i64 %r1, i32 1 + %r02 = insertelement <4 x i64> %r01, i64 %r2, i32 2 + %r03 = insertelement <4 x i64> %r02, i64 %r3, i32 3 + ret <4 x i64> %r03 +} + +define <8 x i32> @test_v8i32(<8 x i32> %a, <8 x i32> %b) { +; SSE-LABEL: @test_v8i32( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <4 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP3:%.*]] = sub <4 x i32> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SSE-NEXT: [[TMP6:%.*]] = sub <4 x i32> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[R07:%.*]] = shufflevector <4 x i32> [[TMP3]], <4 x i32> [[TMP6]], <8 x i32> +; SSE-NEXT: ret <8 x i32> [[R07]] +; +; SLM-LABEL: @test_v8i32( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <4 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP3:%.*]] = sub <4 x i32> [[TMP1]], [[TMP2]] +; SLM-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP5:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <4 x i32> +; SLM-NEXT: [[TMP6:%.*]] = sub <4 x i32> [[TMP4]], [[TMP5]] +; SLM-NEXT: [[R07:%.*]] = shufflevector <4 x i32> [[TMP3]], <4 x i32> [[TMP6]], <8 x i32> +; SLM-NEXT: ret <8 x i32> [[R07]] +; +; AVX-LABEL: @test_v8i32( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <8 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <8 x i32> +; AVX-NEXT: [[TMP3:%.*]] = sub <8 x i32> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <8 x i32> [[TMP3]] +; +; AVX512-LABEL: @test_v8i32( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[A:%.*]], <8 x i32> [[B:%.*]], <8 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[A]], <8 x i32> [[B]], <8 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = sub <8 x i32> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <8 x i32> [[TMP3]] +; + %a0 = extractelement <8 x i32> %a, i32 0 + %a1 = extractelement <8 x i32> %a, i32 1 + %a2 = extractelement <8 x i32> %a, i32 2 + %a3 = extractelement <8 x i32> %a, i32 3 + %a4 = extractelement <8 x i32> %a, i32 4 + %a5 = extractelement <8 x i32> %a, i32 5 + %a6 = extractelement <8 x i32> %a, i32 6 + %a7 = extractelement <8 x i32> %a, i32 7 + %b0 = extractelement <8 x i32> %b, i32 0 + %b1 = extractelement <8 x i32> %b, i32 1 + %b2 = extractelement <8 x i32> %b, i32 2 + %b3 = extractelement <8 x i32> %b, i32 3 + %b4 = extractelement <8 x i32> %b, i32 4 + %b5 = extractelement <8 x i32> %b, i32 5 + %b6 = extractelement <8 x i32> %b, i32 6 + %b7 = extractelement <8 x i32> %b, i32 7 + %r0 = sub i32 %a0, %a1 + %r1 = sub i32 %a2, %a3 + %r2 = sub i32 %b0, %b1 + %r3 = sub i32 %b2, %b3 + %r4 = sub i32 %a4, %a5 + %r5 = sub i32 %a6, %a7 + %r6 = sub i32 %b4, %b5 + %r7 = sub i32 %b6, %b7 + %r00 = insertelement <8 x i32> undef, i32 %r0, i32 0 + %r01 = insertelement <8 x i32> %r00, i32 %r1, i32 1 + %r02 = insertelement <8 x i32> %r01, i32 %r2, i32 2 + %r03 = insertelement <8 x i32> %r02, i32 %r3, i32 3 + %r04 = insertelement <8 x i32> %r03, i32 %r4, i32 4 + %r05 = insertelement <8 x i32> %r04, i32 %r5, i32 5 + %r06 = insertelement <8 x i32> %r05, i32 %r6, i32 6 + %r07 = insertelement <8 x i32> %r06, i32 %r7, i32 7 + ret <8 x i32> %r07 +} + +define <16 x i16> @test_v16i16(<16 x i16> %a, <16 x i16> %b) { +; SSE-LABEL: @test_v16i16( +; SSE-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <8 x i32> +; SSE-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP3:%.*]] = sub <8 x i16> [[TMP1]], [[TMP2]] +; SSE-NEXT: [[TMP4:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP5:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <8 x i32> +; SSE-NEXT: [[TMP6:%.*]] = sub <8 x i16> [[TMP4]], [[TMP5]] +; SSE-NEXT: [[RV15:%.*]] = shufflevector <8 x i16> [[TMP3]], <8 x i16> [[TMP6]], <16 x i32> +; SSE-NEXT: ret <16 x i16> [[RV15]] +; +; SLM-LABEL: @test_v16i16( +; SLM-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; SLM-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; SLM-NEXT: [[TMP3:%.*]] = sub <16 x i16> [[TMP1]], [[TMP2]] +; SLM-NEXT: ret <16 x i16> [[TMP3]] +; +; AVX-LABEL: @test_v16i16( +; AVX-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; AVX-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; AVX-NEXT: [[TMP3:%.*]] = sub <16 x i16> [[TMP1]], [[TMP2]] +; AVX-NEXT: ret <16 x i16> [[TMP3]] +; +; AVX512-LABEL: @test_v16i16( +; AVX512-NEXT: [[TMP1:%.*]] = shufflevector <16 x i16> [[A:%.*]], <16 x i16> [[B:%.*]], <16 x i32> +; AVX512-NEXT: [[TMP2:%.*]] = shufflevector <16 x i16> [[A]], <16 x i16> [[B]], <16 x i32> +; AVX512-NEXT: [[TMP3:%.*]] = sub <16 x i16> [[TMP1]], [[TMP2]] +; AVX512-NEXT: ret <16 x i16> [[TMP3]] +; + %a0 = extractelement <16 x i16> %a, i32 0 + %a1 = extractelement <16 x i16> %a, i32 1 + %a2 = extractelement <16 x i16> %a, i32 2 + %a3 = extractelement <16 x i16> %a, i32 3 + %a4 = extractelement <16 x i16> %a, i32 4 + %a5 = extractelement <16 x i16> %a, i32 5 + %a6 = extractelement <16 x i16> %a, i32 6 + %a7 = extractelement <16 x i16> %a, i32 7 + %a8 = extractelement <16 x i16> %a, i32 8 + %a9 = extractelement <16 x i16> %a, i32 9 + %a10 = extractelement <16 x i16> %a, i32 10 + %a11 = extractelement <16 x i16> %a, i32 11 + %a12 = extractelement <16 x i16> %a, i32 12 + %a13 = extractelement <16 x i16> %a, i32 13 + %a14 = extractelement <16 x i16> %a, i32 14 + %a15 = extractelement <16 x i16> %a, i32 15 + %b0 = extractelement <16 x i16> %b, i32 0 + %b1 = extractelement <16 x i16> %b, i32 1 + %b2 = extractelement <16 x i16> %b, i32 2 + %b3 = extractelement <16 x i16> %b, i32 3 + %b4 = extractelement <16 x i16> %b, i32 4 + %b5 = extractelement <16 x i16> %b, i32 5 + %b6 = extractelement <16 x i16> %b, i32 6 + %b7 = extractelement <16 x i16> %b, i32 7 + %b8 = extractelement <16 x i16> %b, i32 8 + %b9 = extractelement <16 x i16> %b, i32 9 + %b10 = extractelement <16 x i16> %b, i32 10 + %b11 = extractelement <16 x i16> %b, i32 11 + %b12 = extractelement <16 x i16> %b, i32 12 + %b13 = extractelement <16 x i16> %b, i32 13 + %b14 = extractelement <16 x i16> %b, i32 14 + %b15 = extractelement <16 x i16> %b, i32 15 + %r0 = sub i16 %a0 , %a1 + %r1 = sub i16 %a2 , %a3 + %r2 = sub i16 %a4 , %a5 + %r3 = sub i16 %a6 , %a7 + %r4 = sub i16 %b0 , %b1 + %r5 = sub i16 %b2 , %b3 + %r6 = sub i16 %b4 , %b5 + %r7 = sub i16 %b6 , %b7 + %r8 = sub i16 %a8 , %a9 + %r9 = sub i16 %a10, %a11 + %r10 = sub i16 %a12, %a13 + %r11 = sub i16 %a14, %a15 + %r12 = sub i16 %b8 , %b9 + %r13 = sub i16 %b10, %b11 + %r14 = sub i16 %b12, %b13 + %r15 = sub i16 %b14, %b15 + %rv0 = insertelement <16 x i16> undef, i16 %r0 , i32 0 + %rv1 = insertelement <16 x i16> %rv0 , i16 %r1 , i32 1 + %rv2 = insertelement <16 x i16> %rv1 , i16 %r2 , i32 2 + %rv3 = insertelement <16 x i16> %rv2 , i16 %r3 , i32 3 + %rv4 = insertelement <16 x i16> %rv3 , i16 %r4 , i32 4 + %rv5 = insertelement <16 x i16> %rv4 , i16 %r5 , i32 5 + %rv6 = insertelement <16 x i16> %rv5 , i16 %r6 , i32 6 + %rv7 = insertelement <16 x i16> %rv6 , i16 %r7 , i32 7 + %rv8 = insertelement <16 x i16> %rv7 , i16 %r8 , i32 8 + %rv9 = insertelement <16 x i16> %rv8 , i16 %r9 , i32 9 + %rv10 = insertelement <16 x i16> %rv9 , i16 %r10, i32 10 + %rv11 = insertelement <16 x i16> %rv10, i16 %r11, i32 11 + %rv12 = insertelement <16 x i16> %rv11, i16 %r12, i32 12 + %rv13 = insertelement <16 x i16> %rv12, i16 %r13, i32 13 + %rv14 = insertelement <16 x i16> %rv13, i16 %r14, i32 14 + %rv15 = insertelement <16 x i16> %rv14, i16 %r15, i32 15 + ret <16 x i16> %rv15 +}