2016-03-02 05:38:41 +08:00
|
|
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2018-10-16 05:51:26 +08:00
|
|
|
; RUN: llc < %s -disable-peephole -mtriple=i686-apple-darwin -mattr=avx2 -show-mc-encoding | FileCheck %s --check-prefix=CHECK --check-prefix=AVX2 --check-prefix=X86 --check-prefix=X86-AVX
|
|
|
|
; RUN: llc < %s -disable-peephole -mtriple=i686-apple-darwin -mattr=+avx512f,+avx512bw,+avx512vl,+avx512dq -show-mc-encoding | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512VL --check-prefix=X86 --check-prefix=X86-AVX512VL
|
|
|
|
; RUN: llc < %s -disable-peephole -mtriple=x86_64-apple-darwin -mattr=avx2 -show-mc-encoding | FileCheck %s --check-prefix=CHECK --check-prefix=AVX2 --check-prefix=X64 --check-prefix=X64-AVX
|
|
|
|
; RUN: llc < %s -disable-peephole -mtriple=x86_64-apple-darwin -mattr=+avx512f,+avx512bw,+avx512vl,+avx512dq -show-mc-encoding | FileCheck %s --check-prefix=CHECK --check-prefix=AVX512VL --check-prefix=X64 --check-prefix=X64-AVX512VL
|
2011-10-31 10:15:10 +08:00
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_packssdw(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_packssdw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpackssdw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x6b,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_packssdw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpackssdw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x6b,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.packssdw(<8 x i32> %a0, <8 x i32> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.packssdw(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
2017-10-02 23:43:26 +08:00
|
|
|
define <16 x i16> @test_x86_avx2_packssdw_fold() {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_packssdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,0,0,0,255,32767,32767,65535,0,0,0,0,32769,32768,0,65280]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI1_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_packssdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovaps LCPI1_0, %ymm0 ## EVEX TO VEX Compression ymm0 = [0,0,0,0,255,32767,32767,65535,0,0,0,0,32769,32768,0,65280]
|
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI1_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_packssdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,0,0,0,255,32767,32767,65535,0,0,0,0,32769,32768,0,65280]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI1_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_packssdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovaps {{.*}}(%rip), %ymm0 ## EVEX TO VEX Compression ymm0 = [0,0,0,0,255,32767,32767,65535,0,0,0,0,32769,32768,0,65280]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI1_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2017-10-02 23:43:26 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.packssdw(<8 x i32> zeroinitializer, <8 x i32> <i32 255, i32 32767, i32 65535, i32 -1, i32 -32767, i32 -65535, i32 0, i32 -256>)
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_packsswb(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_packsswb:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpacksswb %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x63,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_packsswb:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpacksswb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x63,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.packsswb(<16 x i16> %a0, <16 x i16> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.packsswb(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2017-10-02 23:43:26 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_packsswb_fold() {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_packsswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0,0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI3_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_packsswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovaps LCPI3_0, %ymm0 ## EVEX TO VEX Compression ymm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0,0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0]
|
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI3_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_packsswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0,0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI3_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_packsswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovaps {{.*}}(%rip), %ymm0 ## EVEX TO VEX Compression ymm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0,0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI3_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2017-10-02 23:43:26 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.packsswb(<16 x i16> <i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678, i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678>, <16 x i16> zeroinitializer)
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_packuswb(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_packuswb:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpackuswb %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x67,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_packuswb:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpackuswb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x67,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.packuswb(<16 x i16> %a0, <16 x i16> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.packuswb(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2017-10-02 23:43:26 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_packuswb_fold() {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_packuswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI5_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_packuswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovaps LCPI5_0, %ymm0 ## EVEX TO VEX Compression ymm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI5_0, kind: FK_Data_4
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_packuswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI5_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_packuswb_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovaps {{.*}}(%rip), %ymm0 ## EVEX TO VEX Compression ymm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI5_0-4, kind: reloc_riprel_4byte
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2017-10-02 23:43:26 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.packuswb(<16 x i16> <i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678, i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678>, <16 x i16> zeroinitializer)
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2018-04-27 05:46:01 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_padds_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_padds_b:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpaddsb %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xec,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_padds_b:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpaddsb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xec,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2018-04-27 05:46:01 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.padds.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.padds.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_padds_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_padds_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpaddsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xed,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_padds_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpaddsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xed,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2018-04-27 05:46:01 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.padds.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.padds.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <8 x i32> @test_x86_avx2_pmadd_wd(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmadd_wd:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaddwd %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xf5,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmadd_wd:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaddwd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xf5,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pmadd.wd(<16 x i16> %a0, <16 x i16> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pmadd.wd(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmaxs_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxs_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xee,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxs_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xee,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmaxs.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmaxs.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pmaxu_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxu_b:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxub %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xde,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxu_b:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxub %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xde,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pmaxu.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pmaxu.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmins_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmins_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xea,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmins_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xea,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmins.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmins.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pminu_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pminu_b:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminub %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xda,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pminu_b:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminub %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xda,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pminu.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pminu.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
2011-11-02 12:42:13 +08:00
|
|
|
define i32 @test_x86_avx2_pmovmskb(<32 x i8> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_pmovmskb:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpmovmskb %ymm0, %eax ## encoding: [0xc5,0xfd,0xd7,0xc0]
|
|
|
|
; CHECK-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call i32 @llvm.x86.avx2.pmovmskb(<32 x i8> %a0) ; <i32> [#uses=1]
|
|
|
|
ret i32 %res
|
|
|
|
}
|
|
|
|
declare i32 @llvm.x86.avx2.pmovmskb(<32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <16 x i16> @test_x86_avx2_pmulh_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmulh_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmulhw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe5,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmulh_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmulhw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe5,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmulh.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmulh.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmulhu_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmulhu_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe4,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmulhu_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe4,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmulhu.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmulhu.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_psad_bw(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psad_bw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsadbw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xf6,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psad_bw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsadbw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xf6,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psad.bw(<32 x i8> %a0, <32 x i8> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psad.bw(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psll_d(<8 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psll_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpslld %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xf2,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psll_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpslld %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xf2,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32> %a0, <4 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32>, <4 x i32>) nounwind readnone
|
2015-01-11 09:36:20 +08:00
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <4 x i64> @test_x86_avx2_psll_q(<4 x i64> %a0, <2 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psll_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllq %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xf3,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psll_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllq %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xf3,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psll.q(<4 x i64> %a0, <2 x i64> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psll.q(<4 x i64>, <2 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psll_w(<16 x i16> %a0, <8 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psll_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllw %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xf1,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psll_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllw %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xf1,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16> %a0, <8 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16>, <8 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pslli_d(<8 x i32> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pslli_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpslld $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x72,0xf0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pslli_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpslld $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x72,0xf0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32> %a0, i32 7) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_pslli_q(<4 x i64> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pslli_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllq $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x73,0xf0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pslli_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllq $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x73,0xf0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.pslli.q(<4 x i64> %a0, i32 7) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.pslli.q(<4 x i64>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pslli_w(<16 x i16> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pslli_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllw $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x71,0xf0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pslli_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllw $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x71,0xf0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16> %a0, i32 7) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psra_d(<8 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psra_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrad %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe2,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psra_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrad %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe2,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %a0, <4 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32>, <4 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psra_w(<16 x i16> %a0, <8 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psra_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsraw %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe1,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psra_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsraw %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe1,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %a0, <8 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16>, <8 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psrai_d(<8 x i32> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrai_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrad $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x72,0xe0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrai_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrad $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x72,0xe0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %a0, i32 7) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psrai_w(<16 x i16> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrai_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsraw $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x71,0xe0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrai_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsraw $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x71,0xe0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %a0, i32 7) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psrl_d(<8 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrl_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrld %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xd2,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrl_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrld %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xd2,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32> %a0, <4 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32>, <4 x i32>) nounwind readnone
|
2015-01-11 09:36:20 +08:00
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <4 x i64> @test_x86_avx2_psrl_q(<4 x i64> %a0, <2 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrl_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlq %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xd3,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrl_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlq %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xd3,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psrl.q(<4 x i64> %a0, <2 x i64> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psrl.q(<4 x i64>, <2 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psrl_w(<16 x i16> %a0, <8 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrl_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlw %xmm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xd1,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrl_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlw %xmm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xd1,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16> %a0, <8 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16>, <8 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2018-10-16 05:51:29 +08:00
|
|
|
define <16 x i16> @test_x86_avx2_psrl_w_load(<16 x i16> %a0, <8 x i16>* %p) {
|
|
|
|
; X86-AVX-LABEL: test_x86_avx2_psrl_w_load:
|
|
|
|
; X86-AVX: ## %bb.0:
|
|
|
|
; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-AVX-NEXT: vpsrlw (%eax), %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xd1,0x00]
|
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_psrl_w_load:
|
|
|
|
; X86-AVX512VL: ## %bb.0:
|
|
|
|
; X86-AVX512VL-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
2018-10-16 05:51:32 +08:00
|
|
|
; X86-AVX512VL-NEXT: vpsrlw (%eax), %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xd1,0x00]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_psrl_w_load:
|
|
|
|
; X64-AVX: ## %bb.0:
|
|
|
|
; X64-AVX-NEXT: vpsrlw (%rdi), %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xd1,0x07]
|
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_psrl_w_load:
|
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2018-10-16 05:51:32 +08:00
|
|
|
; X64-AVX512VL-NEXT: vpsrlw (%rdi), %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xd1,0x07]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
|
|
|
%a1 = load <8 x i16>, <8 x i16>* %p
|
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16> %a0, <8 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2011-10-31 10:15:10 +08:00
|
|
|
define <8 x i32> @test_x86_avx2_psrli_d(<8 x i32> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrli_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrld $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x72,0xd0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrli_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrld $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x72,0xd0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32> %a0, i32 7) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_psrli_q(<4 x i64> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrli_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlq $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x73,0xd0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrli_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlq $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x73,0xd0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psrli.q(<4 x i64> %a0, i32 7) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psrli.q(<4 x i64>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psrli_w(<16 x i16> %a0) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrli_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlw $7, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0x71,0xd0,0x07]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrli_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlw $7, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x71,0xd0,0x07]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-10-31 10:15:10 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16> %a0, i32 7) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16>, i32) nounwind readnone
|
|
|
|
|
|
|
|
|
2018-04-27 05:46:01 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_psubs_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psubs_b:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsubsb %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe8,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psubs_b:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsubsb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe8,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2018-04-27 05:46:01 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.psubs.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.psubs.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psubs_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psubs_w:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsubsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc5,0xfd,0xe9,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psubs_w:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsubsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0xe9,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2018-04-27 05:46:01 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psubs.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psubs.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
2011-11-02 12:42:13 +08:00
|
|
|
define <8 x i32> @test_x86_avx2_phadd_d(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phadd_d:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphaddd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x02,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.phadd.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.phadd.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_phadd_sw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phadd_sw:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphaddsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x03,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.phadd.sw(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.phadd.sw(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_phadd_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phadd_w:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphaddw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x01,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.phadd.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.phadd.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_phsub_d(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phsub_d:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphsubd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x06,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.phsub.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.phsub.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_phsub_sw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phsub_sw:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphsubsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x07,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.phsub.sw(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.phsub.sw(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_phsub_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_phsub_w:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vphsubw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x05,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.phsub.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.phsub.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmadd_ub_sw(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmadd_ub_sw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaddubsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x04,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmadd_ub_sw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaddubsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x04,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmadd.ub.sw(<32 x i8> %a0, <32 x i8> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmadd.ub.sw(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
2016-12-03 13:35:38 +08:00
|
|
|
; Make sure we don't commute this operation.
|
|
|
|
define <16 x i16> @test_x86_avx2_pmadd_ub_sw_load_op0(<32 x i8>* %ptr, <32 x i8> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_pmadd_ub_sw_load_op0:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-AVX-NEXT: vmovdqa (%eax), %ymm1 ## encoding: [0xc5,0xfd,0x6f,0x08]
|
|
|
|
; X86-AVX-NEXT: vpmaddubsw %ymm0, %ymm1, %ymm0 ## encoding: [0xc4,0xe2,0x75,0x04,0xc0]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_pmadd_ub_sw_load_op0:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-AVX512VL-NEXT: vmovdqa (%eax), %ymm1 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x6f,0x08]
|
|
|
|
; X86-AVX512VL-NEXT: vpmaddubsw %ymm0, %ymm1, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x75,0x04,0xc0]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_pmadd_ub_sw_load_op0:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovdqa (%rdi), %ymm1 ## encoding: [0xc5,0xfd,0x6f,0x0f]
|
|
|
|
; X64-AVX-NEXT: vpmaddubsw %ymm0, %ymm1, %ymm0 ## encoding: [0xc4,0xe2,0x75,0x04,0xc0]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_pmadd_ub_sw_load_op0:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovdqa (%rdi), %ymm1 ## EVEX TO VEX Compression encoding: [0xc5,0xfd,0x6f,0x0f]
|
|
|
|
; X64-AVX512VL-NEXT: vpmaddubsw %ymm0, %ymm1, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x75,0x04,0xc0]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2016-12-03 13:35:38 +08:00
|
|
|
%a0 = load <32 x i8>, <32 x i8>* %ptr
|
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmadd.ub.sw(<32 x i8> %a0, <32 x i8> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
2011-11-02 12:42:13 +08:00
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmul_hr_sw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmul_hr_sw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmulhrsw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x0b,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmul_hr_sw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmulhrsw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x0b,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmul.hr.sw(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmul.hr.sw(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pshuf_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pshuf_b:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpshufb %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x00,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pshuf_b:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpshufb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x00,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pshuf.b(<32 x i8> %a0, <32 x i8> %a1) ; <<16 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pshuf.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_psign_b(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_psign_b:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpsignb %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x08,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.psign.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.psign.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psign_d(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_psign_d:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpsignd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x0a,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psign.d(<8 x i32> %a0, <8 x i32> %a1) ; <<4 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psign.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_psign_w(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_psign_w:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpsignw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x09,0xc1]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 12:42:13 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.psign.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.psign.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2011-11-02 14:54:17 +08:00
|
|
|
define <16 x i16> @test_x86_avx2_mpsadbw(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_mpsadbw:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vmpsadbw $7, %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe3,0x7d,0x42,0xc1,0x07]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.mpsadbw(<32 x i8> %a0, <32 x i8> %a1, i8 7) ; <<16 x i16>> [#uses=1]
|
2011-11-02 14:54:17 +08:00
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
declare <16 x i16> @llvm.x86.avx2.mpsadbw(<32 x i8>, <32 x i8>, i8) nounwind readnone
|
2011-11-02 14:54:17 +08:00
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_packusdw(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_packusdw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpackusdw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x2b,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_packusdw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpackusdw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x2b,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.packusdw(<8 x i32> %a0, <8 x i32> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.packusdw(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
2017-10-02 23:43:26 +08:00
|
|
|
define <16 x i16> @test_x86_avx2_packusdw_fold() {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_packusdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,0,0,0,255,32767,65535,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI51_0, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_packusdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovaps LCPI51_0, %ymm0 ## EVEX TO VEX Compression ymm0 = [0,0,0,0,255,32767,65535,0,0,0,0,0,0,0,0,0]
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI51_0, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_packusdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [0,0,0,0,255,32767,65535,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI51_0-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_packusdw_fold:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovaps {{.*}}(%rip), %ymm0 ## EVEX TO VEX Compression ymm0 = [0,0,0,0,255,32767,65535,0,0,0,0,0,0,0,0,0]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xfc,0x28,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI51_0-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2017-10-02 23:43:26 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.packusdw(<8 x i32> zeroinitializer, <8 x i32> <i32 255, i32 32767, i32 65535, i32 -1, i32 -32767, i32 -65535, i32 0, i32 -256>)
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2011-11-02 14:54:17 +08:00
|
|
|
define <32 x i8> @test_x86_avx2_pblendvb(<32 x i8> %a0, <32 x i8> %a1, <32 x i8> %a2) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_pblendvb:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpblendvb %ymm2, %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe3,0x7d,0x4c,0xc1,0x20]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8> %a0, <32 x i8> %a1, <32 x i8> %a2) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8>, <32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pblendw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_pblendw:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vpblendw $7, %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe3,0x7d,0x0e,0xc1,0x07]
|
|
|
|
; CHECK-NEXT: ## ymm0 = ymm1[0,1,2],ymm0[3,4,5,6,7],ymm1[8,9,10],ymm0[11,12,13,14,15]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16> %a0, <16 x i16> %a1, i8 7) ; <<16 x i16>> [#uses=1]
|
2011-11-02 14:54:17 +08:00
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
declare <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16>, <16 x i16>, i8) nounwind readnone
|
2011-11-02 14:54:17 +08:00
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pmaxsb(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxsb:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxsb %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3c,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxsb:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxsb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3c,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pmaxs.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pmaxs.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pmaxsd(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxsd:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxsd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3d,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxsd:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxsd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3d,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pmaxs.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pmaxs.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pmaxud(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxud:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxud %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3f,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxud:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxud %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3f,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pmaxu.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pmaxu.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pmaxuw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pmaxuw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpmaxuw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3e,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pmaxuw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpmaxuw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3e,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pmaxu.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pmaxu.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pminsb(<32 x i8> %a0, <32 x i8> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pminsb:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminsb %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x38,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pminsb:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminsb %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x38,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <32 x i8> @llvm.x86.avx2.pmins.b(<32 x i8> %a0, <32 x i8> %a1) ; <<32 x i8>> [#uses=1]
|
|
|
|
ret <32 x i8> %res
|
|
|
|
}
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pmins.b(<32 x i8>, <32 x i8>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pminsd(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pminsd:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminsd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x39,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pminsd:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminsd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x39,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pmins.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pmins.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pminud(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pminud:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminud %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3b,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pminud:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminud %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3b,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pminu.d(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.pminu.d(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pminuw(<16 x i16> %a0, <16 x i16> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_pminuw:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpminuw %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x3a,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_pminuw:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpminuw %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x3a,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-02 14:54:17 +08:00
|
|
|
%res = call <16 x i16> @llvm.x86.avx2.pminu.w(<16 x i16> %a0, <16 x i16> %a1) ; <<16 x i16>> [#uses=1]
|
|
|
|
ret <16 x i16> %res
|
|
|
|
}
|
|
|
|
declare <16 x i16> @llvm.x86.avx2.pminu.w(<16 x i16>, <16 x i16>) nounwind readnone
|
|
|
|
|
|
|
|
|
2011-11-06 14:12:20 +08:00
|
|
|
define <4 x i32> @test_x86_avx2_pblendd_128(<4 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_pblendd_128:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vblendps $8, %xmm0, %xmm1, %xmm0 ## encoding: [0xc4,0xe3,0x71,0x0c,0xc0,0x08]
|
|
|
|
; CHECK-NEXT: ## xmm0 = xmm1[0,1,2],xmm0[3]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32> %a0, <4 x i32> %a1, i8 7) ; <<4 x i32>> [#uses=1]
|
2011-11-06 14:12:20 +08:00
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
declare <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32>, <4 x i32>, i8) nounwind readnone
|
2011-11-06 14:12:20 +08:00
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pblendd_256(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; CHECK-LABEL: test_x86_avx2_pblendd_256:
|
|
|
|
; CHECK: ## %bb.0:
|
|
|
|
; CHECK-NEXT: vblendps $7, %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe3,0x7d,0x0c,0xc1,0x07]
|
|
|
|
; CHECK-NEXT: ## ymm0 = ymm1[0,1,2],ymm0[3,4,5,6,7]
|
|
|
|
; CHECK-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32> %a0, <8 x i32> %a1, i8 7) ; <<8 x i32>> [#uses=1]
|
2011-11-06 14:12:20 +08:00
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
[x86] Fix a pretty horrible bug and inconsistency in the x86 asm
parsing (and latent bug in the instruction definitions).
This is effectively a revert of r136287 which tried to address
a specific and narrow case of immediate operands failing to be accepted
by x86 instructions with a pretty heavy hammer: it introduced a new kind
of operand that behaved differently. All of that is removed with this
commit, but the test cases are both preserved and enhanced.
The core problem that r136287 and this commit are trying to handle is
that gas accepts both of the following instructions:
insertps $192, %xmm0, %xmm1
insertps $-64, %xmm0, %xmm1
These will encode to the same byte sequence, with the immediate
occupying an 8-bit entry. The first form was fixed by r136287 but that
broke the prior handling of the second form! =[ Ironically, we would
still emit the second form in some cases and then be unable to
re-assemble the output.
The reason why the first instruction failed to be handled is because
prior to r136287 the operands ere marked 'i32i8imm' which forces them to
be sign-extenable. Clearly, that won't work for 192 in a single byte.
However, making thim zero-extended or "unsigned" doesn't really address
the core issue either because it breaks negative immediates. The correct
fix is to make these operands 'i8imm' reflecting that they can be either
signed or unsigned but must be 8-bit immediates. This patch backs out
r136287 and then changes those places as well as some others to use
'i8imm' rather than one of the extended variants.
Naturally, this broke something else. The custom DAG nodes had to be
updated to have a much more accurate type constraint of an i8 node, and
a bunch of Pat immediates needed to be specified as i8 values.
The fallout didn't end there though. We also then ceased to be able to
match the instruction-specific intrinsics to the instructions so
modified. Digging, this is because they too used i32 rather than i8 in
their signature. So I've also switched those intrinsics to i8 arguments
in line with the instructions.
In order to make the intrinsic adjustments of course, I also had to add
auto upgrading for the intrinsics.
I suspect that the intrinsic argument types may have led everything down
this rabbit hole. Pretty happy with the result.
llvm-svn: 217310
2014-09-06 18:00:01 +08:00
|
|
|
declare <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32>, <8 x i32>, i8) nounwind readnone
|
2011-11-06 14:12:20 +08:00
|
|
|
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
; Check that the arguments are swapped between the intrinsic definition
|
|
|
|
; and its lowering. Indeed, the offsets are the first source in
|
|
|
|
; the instruction.
|
2011-11-07 07:04:08 +08:00
|
|
|
define <8 x i32> @test_x86_avx2_permd(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_permd:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpermps %ymm0, %ymm1, %ymm0 ## encoding: [0xc4,0xe2,0x75,0x16,0xc0]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_permd:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpermps %ymm0, %ymm1, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x75,0x16,0xc0]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 07:04:08 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.permd(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.permd(<8 x i32>, <8 x i32>) nounwind readonly
|
|
|
|
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
; Check that the arguments are swapped between the intrinsic definition
|
|
|
|
; and its lowering. Indeed, the offsets are the first source in
|
|
|
|
; the instruction.
|
2015-11-30 06:53:22 +08:00
|
|
|
define <8 x float> @test_x86_avx2_permps(<8 x float> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_permps:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpermps %ymm0, %ymm1, %ymm0 ## encoding: [0xc4,0xe2,0x75,0x16,0xc0]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_permps:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpermps %ymm0, %ymm1, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x75,0x16,0xc0]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2015-11-30 06:53:22 +08:00
|
|
|
%res = call <8 x float> @llvm.x86.avx2.permps(<8 x float> %a0, <8 x i32> %a1) ; <<8 x float>> [#uses=1]
|
2011-11-07 07:04:08 +08:00
|
|
|
ret <8 x float> %res
|
|
|
|
}
|
2015-11-30 06:53:22 +08:00
|
|
|
declare <8 x float> @llvm.x86.avx2.permps(<8 x float>, <8 x i32>) nounwind readonly
|
2011-11-07 07:04:08 +08:00
|
|
|
|
|
|
|
|
2011-11-07 11:20:35 +08:00
|
|
|
define <2 x i64> @test_x86_avx2_maskload_q(i8* %a0, <2 x i64> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskload_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovq (%eax), %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0xf9,0x8c,0x00]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskload_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovq (%rdi), %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0xf9,0x8c,0x07]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
%res = call <2 x i64> @llvm.x86.avx2.maskload.q(i8* %a0, <2 x i64> %a1) ; <<2 x i64>> [#uses=1]
|
|
|
|
ret <2 x i64> %res
|
|
|
|
}
|
|
|
|
declare <2 x i64> @llvm.x86.avx2.maskload.q(i8*, <2 x i64>) nounwind readonly
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_maskload_q_256(i8* %a0, <4 x i64> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskload_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovq (%eax), %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0xfd,0x8c,0x00]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskload_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovq (%rdi), %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0xfd,0x8c,0x07]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.maskload.q.256(i8* %a0, <4 x i64> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.maskload.q.256(i8*, <4 x i64>) nounwind readonly
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i32> @test_x86_avx2_maskload_d(i8* %a0, <4 x i32> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskload_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovd (%eax), %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x8c,0x00]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskload_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovd (%rdi), %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x8c,0x07]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.maskload.d(i8* %a0, <4 x i32> %a1) ; <<4 x i32>> [#uses=1]
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.maskload.d(i8*, <4 x i32>) nounwind readonly
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_maskload_d_256(i8* %a0, <8 x i32> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskload_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovd (%eax), %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x8c,0x00]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskload_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovd (%rdi), %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x8c,0x07]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.maskload.d.256(i8* %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.maskload.d.256(i8*, <8 x i32>) nounwind readonly
|
|
|
|
|
|
|
|
|
|
|
|
define void @test_x86_avx2_maskstore_q(i8* %a0, <2 x i64> %a1, <2 x i64> %a2) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskstore_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovq %xmm1, %xmm0, (%eax) ## encoding: [0xc4,0xe2,0xf9,0x8e,0x08]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskstore_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovq %xmm1, %xmm0, (%rdi) ## encoding: [0xc4,0xe2,0xf9,0x8e,0x0f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
call void @llvm.x86.avx2.maskstore.q(i8* %a0, <2 x i64> %a1, <2 x i64> %a2)
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
declare void @llvm.x86.avx2.maskstore.q(i8*, <2 x i64>, <2 x i64>) nounwind
|
|
|
|
|
|
|
|
|
|
|
|
define void @test_x86_avx2_maskstore_q_256(i8* %a0, <4 x i64> %a1, <4 x i64> %a2) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskstore_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovq %ymm1, %ymm0, (%eax) ## encoding: [0xc4,0xe2,0xfd,0x8e,0x08]
|
|
|
|
; X86-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskstore_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovq %ymm1, %ymm0, (%rdi) ## encoding: [0xc4,0xe2,0xfd,0x8e,0x0f]
|
|
|
|
; X64-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
call void @llvm.x86.avx2.maskstore.q.256(i8* %a0, <4 x i64> %a1, <4 x i64> %a2)
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
declare void @llvm.x86.avx2.maskstore.q.256(i8*, <4 x i64>, <4 x i64>) nounwind
|
|
|
|
|
|
|
|
|
|
|
|
define void @test_x86_avx2_maskstore_d(i8* %a0, <4 x i32> %a1, <4 x i32> %a2) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskstore_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovd %xmm1, %xmm0, (%eax) ## encoding: [0xc4,0xe2,0x79,0x8e,0x08]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskstore_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovd %xmm1, %xmm0, (%rdi) ## encoding: [0xc4,0xe2,0x79,0x8e,0x0f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
call void @llvm.x86.avx2.maskstore.d(i8* %a0, <4 x i32> %a1, <4 x i32> %a2)
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
declare void @llvm.x86.avx2.maskstore.d(i8*, <4 x i32>, <4 x i32>) nounwind
|
|
|
|
|
|
|
|
|
|
|
|
define void @test_x86_avx2_maskstore_d_256(i8* %a0, <8 x i32> %a1, <8 x i32> %a2) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_maskstore_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpmaskmovd %ymm1, %ymm0, (%eax) ## encoding: [0xc4,0xe2,0x7d,0x8e,0x08]
|
|
|
|
; X86-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_maskstore_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpmaskmovd %ymm1, %ymm0, (%rdi) ## encoding: [0xc4,0xe2,0x7d,0x8e,0x0f]
|
|
|
|
; X64-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2011-11-07 11:20:35 +08:00
|
|
|
call void @llvm.x86.avx2.maskstore.d.256(i8* %a0, <8 x i32> %a1, <8 x i32> %a2)
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
declare void @llvm.x86.avx2.maskstore.d.256(i8*, <8 x i32>, <8 x i32>) nounwind
|
2011-11-07 16:26:24 +08:00
|
|
|
|
|
|
|
|
|
|
|
define <4 x i32> @test_x86_avx2_psllv_d(<4 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psllv_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllvd %xmm1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x47,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psllv_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllvd %xmm1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x79,0x47,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.psllv.d(<4 x i32> %a0, <4 x i32> %a1) ; <<4 x i32>> [#uses=1]
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.psllv.d(<4 x i32>, <4 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psllv_d_256(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psllv_d_256:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllvd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x47,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psllv_d_256:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllvd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x47,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psllv.d.256(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psllv.d.256(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <2 x i64> @test_x86_avx2_psllv_q(<2 x i64> %a0, <2 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psllv_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllvq %xmm1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0xf9,0x47,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psllv_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllvq %xmm1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0xf9,0x47,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <2 x i64> @llvm.x86.avx2.psllv.q(<2 x i64> %a0, <2 x i64> %a1) ; <<2 x i64>> [#uses=1]
|
|
|
|
ret <2 x i64> %res
|
|
|
|
}
|
|
|
|
declare <2 x i64> @llvm.x86.avx2.psllv.q(<2 x i64>, <2 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_psllv_q_256(<4 x i64> %a0, <4 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psllv_q_256:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsllvq %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0xfd,0x47,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psllv_q_256:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsllvq %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0xfd,0x47,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psllv.q.256(<4 x i64> %a0, <4 x i64> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psllv.q.256(<4 x i64>, <4 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i32> @test_x86_avx2_psrlv_d(<4 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrlv_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlvd %xmm1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x45,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrlv_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlvd %xmm1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x79,0x45,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.psrlv.d(<4 x i32> %a0, <4 x i32> %a1) ; <<4 x i32>> [#uses=1]
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.psrlv.d(<4 x i32>, <4 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psrlv_d_256(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrlv_d_256:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlvd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x45,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrlv_d_256:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlvd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x45,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrlv.d.256(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.psrlv.d.256(<8 x i32>, <8 x i32>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <2 x i64> @test_x86_avx2_psrlv_q(<2 x i64> %a0, <2 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrlv_q:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlvq %xmm1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0xf9,0x45,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrlv_q:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlvq %xmm1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0xf9,0x45,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <2 x i64> @llvm.x86.avx2.psrlv.q(<2 x i64> %a0, <2 x i64> %a1) ; <<2 x i64>> [#uses=1]
|
|
|
|
ret <2 x i64> %res
|
|
|
|
}
|
|
|
|
declare <2 x i64> @llvm.x86.avx2.psrlv.q(<2 x i64>, <2 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i64> @test_x86_avx2_psrlv_q_256(<4 x i64> %a0, <4 x i64> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrlv_q_256:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsrlvq %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0xfd,0x45,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrlv_q_256:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsrlvq %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0xfd,0x45,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.psrlv.q.256(<4 x i64> %a0, <4 x i64> %a1) ; <<4 x i64>> [#uses=1]
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.psrlv.q.256(<4 x i64>, <4 x i64>) nounwind readnone
|
|
|
|
|
|
|
|
|
|
|
|
define <4 x i32> @test_x86_avx2_psrav_d(<4 x i32> %a0, <4 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrav_d:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsravd %xmm1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x46,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrav_d:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsravd %xmm1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x79,0x46,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.psrav.d(<4 x i32> %a0, <4 x i32> %a1) ; <<4 x i32>> [#uses=1]
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
|
2016-06-20 15:05:43 +08:00
|
|
|
define <4 x i32> @test_x86_avx2_psrav_d_const(<4 x i32> %a0, <4 x i32> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_psrav_d_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovdqa {{.*#+}} xmm0 = [2,9,4294967284,23]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xf9,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI83_0, kind: FK_Data_4
|
|
|
|
; X86-AVX-NEXT: vpsravd LCPI83_1, %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x46,0x05,A,A,A,A]
|
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 5, value: LCPI83_1, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_psrav_d_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovdqa LCPI83_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [2,9,4294967284,23]
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xf9,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI83_0, kind: FK_Data_4
|
|
|
|
; X86-AVX512VL-NEXT: vpsravd LCPI83_1, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x79,0x46,0x05,A,A,A,A]
|
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 5, value: LCPI83_1, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_psrav_d_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovdqa {{.*#+}} xmm0 = [2,9,4294967284,23]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xf9,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI83_0-4, kind: reloc_riprel_4byte
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vpsravd {{.*}}(%rip), %xmm0, %xmm0 ## encoding: [0xc4,0xe2,0x79,0x46,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 5, value: LCPI83_1-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_psrav_d_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovdqa {{.*}}(%rip), %xmm0 ## EVEX TO VEX Compression xmm0 = [2,9,4294967284,23]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xf9,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI83_0-4, kind: reloc_riprel_4byte
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vpsravd {{.*}}(%rip), %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x79,0x46,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 5, value: LCPI83_1-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2016-06-20 15:05:43 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.psrav.d(<4 x i32> <i32 2, i32 9, i32 -12, i32 23>, <4 x i32> <i32 1, i32 18, i32 35, i32 52>)
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.psrav.d(<4 x i32>, <4 x i32>) nounwind readnone
|
2011-11-07 16:26:24 +08:00
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psrav_d_256(<8 x i32> %a0, <8 x i32> %a1) {
|
2018-10-16 05:51:22 +08:00
|
|
|
; AVX2-LABEL: test_x86_avx2_psrav_d_256:
|
|
|
|
; AVX2: ## %bb.0:
|
|
|
|
; AVX2-NEXT: vpsravd %ymm1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x46,0xc1]
|
|
|
|
; AVX2-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
|
|
|
;
|
|
|
|
; AVX512VL-LABEL: test_x86_avx2_psrav_d_256:
|
|
|
|
; AVX512VL: ## %bb.0:
|
|
|
|
; AVX512VL-NEXT: vpsravd %ymm1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x46,0xc1]
|
|
|
|
; AVX512VL-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
2011-11-07 16:26:24 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrav.d.256(<8 x i32> %a0, <8 x i32> %a1) ; <<8 x i32>> [#uses=1]
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
2016-06-20 15:05:43 +08:00
|
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_psrav_d_256_const(<8 x i32> %a0, <8 x i32> %a1) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_x86_avx2_psrav_d_256_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: vmovdqa {{.*#+}} ymm0 = [2,9,4294967284,23,4294967270,37,4294967256,51]
|
|
|
|
; X86-AVX-NEXT: ## encoding: [0xc5,0xfd,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 4, value: LCPI85_0, kind: FK_Data_4
|
|
|
|
; X86-AVX-NEXT: vpsravd LCPI85_1, %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x46,0x05,A,A,A,A]
|
|
|
|
; X86-AVX-NEXT: ## fixup A - offset: 5, value: LCPI85_1, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_x86_avx2_psrav_d_256_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: vmovdqa LCPI85_0, %ymm0 ## EVEX TO VEX Compression ymm0 = [2,9,4294967284,23,4294967270,37,4294967256,51]
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## encoding: [0xc5,0xfd,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI85_0, kind: FK_Data_4
|
|
|
|
; X86-AVX512VL-NEXT: vpsravd LCPI85_1, %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x46,0x05,A,A,A,A]
|
|
|
|
; X86-AVX512VL-NEXT: ## fixup A - offset: 5, value: LCPI85_1, kind: FK_Data_4
|
2018-04-27 05:46:01 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_x86_avx2_psrav_d_256_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovdqa {{.*#+}} ymm0 = [2,9,4294967284,23,4294967270,37,4294967256,51]
|
|
|
|
; X64-AVX-NEXT: ## encoding: [0xc5,0xfd,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 4, value: LCPI85_0-4, kind: reloc_riprel_4byte
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vpsravd {{.*}}(%rip), %ymm0, %ymm0 ## encoding: [0xc4,0xe2,0x7d,0x46,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX-NEXT: ## fixup A - offset: 5, value: LCPI85_1-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_x86_avx2_psrav_d_256_const:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovdqa {{.*}}(%rip), %ymm0 ## EVEX TO VEX Compression ymm0 = [2,9,4294967284,23,4294967270,37,4294967256,51]
|
|
|
|
; X64-AVX512VL-NEXT: ## encoding: [0xc5,0xfd,0x6f,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 4, value: LCPI85_0-4, kind: reloc_riprel_4byte
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vpsravd {{.*}}(%rip), %ymm0, %ymm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe2,0x7d,0x46,0x05,A,A,A,A]
|
2018-10-16 05:51:29 +08:00
|
|
|
; X64-AVX512VL-NEXT: ## fixup A - offset: 5, value: LCPI85_1-4, kind: reloc_riprel_4byte
|
2018-04-27 05:46:01 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2016-06-20 15:05:43 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.psrav.d.256(<8 x i32> <i32 2, i32 9, i32 -12, i32 23, i32 -26, i32 37, i32 -40, i32 51>, <8 x i32> <i32 1, i32 18, i32 35, i32 52, i32 69, i32 15, i32 32, i32 49>)
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
2011-11-07 16:26:24 +08:00
|
|
|
declare <8 x i32> @llvm.x86.avx2.psrav.d.256(<8 x i32>, <8 x i32>) nounwind readnone
|
2011-11-29 13:37:58 +08:00
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <2 x double> @test_x86_avx2_gather_d_pd(<2 x double> %a0, i8* %a1, <4 x i32> %idx, <2 x double> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_pd:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherdpd %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x92,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_pd:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x92,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> %a0,
|
|
|
|
i8* %a1, <4 x i32> %idx, <2 x double> %mask, i8 2) ;
|
|
|
|
ret <2 x double> %res
|
|
|
|
}
|
|
|
|
declare <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double>, i8*,
|
|
|
|
<4 x i32>, <2 x double>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x double> @test_x86_avx2_gather_d_pd_256(<4 x double> %a0, i8* %a1, <4 x i32> %idx, <4 x double> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_pd_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherdpd %ymm2, (%eax,%xmm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x92,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_pd_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherdpd %ymm2, (%rdi,%xmm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x92,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> %a0,
|
2012-06-29 08:54:20 +08:00
|
|
|
i8* %a1, <4 x i32> %idx, <4 x double> %mask, i8 2) ;
|
2012-06-27 03:47:59 +08:00
|
|
|
ret <4 x double> %res
|
|
|
|
}
|
|
|
|
declare <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double>, i8*,
|
2012-06-29 08:54:20 +08:00
|
|
|
<4 x i32>, <4 x double>, i8) nounwind readonly
|
2012-06-27 03:47:59 +08:00
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <2 x double> @test_x86_avx2_gather_q_pd(<2 x double> %a0, i8* %a1, <2 x i64> %idx, <2 x double> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_pd:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherqpd %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x93,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_pd:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherqpd %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x93,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double> %a0,
|
|
|
|
i8* %a1, <2 x i64> %idx, <2 x double> %mask, i8 2) ;
|
|
|
|
ret <2 x double> %res
|
|
|
|
}
|
|
|
|
declare <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double>, i8*,
|
|
|
|
<2 x i64>, <2 x double>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x double> @test_x86_avx2_gather_q_pd_256(<4 x double> %a0, i8* %a1, <4 x i64> %idx, <4 x double> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_pd_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherqpd %ymm2, (%eax,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x93,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_pd_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherqpd %ymm2, (%rdi,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x93,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> %a0,
|
|
|
|
i8* %a1, <4 x i64> %idx, <4 x double> %mask, i8 2) ;
|
|
|
|
ret <4 x double> %res
|
|
|
|
}
|
|
|
|
declare <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double>, i8*,
|
|
|
|
<4 x i64>, <4 x double>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x float> @test_x86_avx2_gather_d_ps(<4 x float> %a0, i8* %a1, <4 x i32> %idx, <4 x float> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_ps:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherdps %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x92,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_ps:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherdps %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x92,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> %a0,
|
|
|
|
i8* %a1, <4 x i32> %idx, <4 x float> %mask, i8 2) ;
|
|
|
|
ret <4 x float> %res
|
|
|
|
}
|
|
|
|
declare <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float>, i8*,
|
|
|
|
<4 x i32>, <4 x float>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <8 x float> @test_x86_avx2_gather_d_ps_256(<8 x float> %a0, i8* %a1, <8 x i32> %idx, <8 x float> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_ps_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherdps %ymm2, (%eax,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0x6d,0x92,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_ps_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherdps %ymm2, (%rdi,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0x6d,0x92,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> %a0,
|
|
|
|
i8* %a1, <8 x i32> %idx, <8 x float> %mask, i8 2) ;
|
|
|
|
ret <8 x float> %res
|
|
|
|
}
|
|
|
|
declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*,
|
|
|
|
<8 x i32>, <8 x float>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x float> @test_x86_avx2_gather_q_ps(<4 x float> %a0, i8* %a1, <2 x i64> %idx, <4 x float> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_ps:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x93,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_ps:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x93,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-27 03:47:59 +08:00
|
|
|
%res = call <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float> %a0,
|
|
|
|
i8* %a1, <2 x i64> %idx, <4 x float> %mask, i8 2) ;
|
|
|
|
ret <4 x float> %res
|
|
|
|
}
|
|
|
|
declare <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float>, i8*,
|
|
|
|
<2 x i64>, <4 x float>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x float> @test_x86_avx2_gather_q_ps_256(<4 x float> %a0, i8* %a1, <4 x i64> %idx, <4 x float> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_ps_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%ymm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x6d,0x93,0x04,0x48]
|
|
|
|
; X86-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_ps_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%ymm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x6d,0x93,0x04,0x4f]
|
|
|
|
; X64-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x float> @llvm.x86.avx2.gather.q.ps.256(<4 x float> %a0,
|
|
|
|
i8* %a1, <4 x i64> %idx, <4 x float> %mask, i8 2) ;
|
|
|
|
ret <4 x float> %res
|
|
|
|
}
|
|
|
|
declare <4 x float> @llvm.x86.avx2.gather.q.ps.256(<4 x float>, i8*,
|
|
|
|
<4 x i64>, <4 x float>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <2 x i64> @test_x86_avx2_gather_d_q(<2 x i64> %a0, i8* %a1, <4 x i32> %idx, <2 x i64> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherdq %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x90,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherdq %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x90,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> %a0,
|
|
|
|
i8* %a1, <4 x i32> %idx, <2 x i64> %mask, i8 2) ;
|
|
|
|
ret <2 x i64> %res
|
|
|
|
}
|
|
|
|
declare <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64>, i8*,
|
|
|
|
<4 x i32>, <2 x i64>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x i64> @test_x86_avx2_gather_d_q_256(<4 x i64> %a0, i8* %a1, <4 x i32> %idx, <4 x i64> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherdq %ymm2, (%eax,%xmm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x90,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherdq %ymm2, (%rdi,%xmm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x90,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> %a0,
|
|
|
|
i8* %a1, <4 x i32> %idx, <4 x i64> %mask, i8 2) ;
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64>, i8*,
|
|
|
|
<4 x i32>, <4 x i64>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <2 x i64> @test_x86_avx2_gather_q_q(<2 x i64> %a0, i8* %a1, <2 x i64> %idx, <2 x i64> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherqq %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x91,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_q:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherqq %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0xe9,0x91,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> %a0,
|
|
|
|
i8* %a1, <2 x i64> %idx, <2 x i64> %mask, i8 2) ;
|
|
|
|
ret <2 x i64> %res
|
|
|
|
}
|
|
|
|
declare <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64>, i8*,
|
|
|
|
<2 x i64>, <2 x i64>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x i64> @test_x86_avx2_gather_q_q_256(<4 x i64> %a0, i8* %a1, <4 x i64> %idx, <4 x i64> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherqq %ymm2, (%eax,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x91,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_q_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherqq %ymm2, (%rdi,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0xed,0x91,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> %a0,
|
|
|
|
i8* %a1, <4 x i64> %idx, <4 x i64> %mask, i8 2) ;
|
|
|
|
ret <4 x i64> %res
|
|
|
|
}
|
|
|
|
declare <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64>, i8*,
|
|
|
|
<4 x i64>, <4 x i64>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x i32> @test_x86_avx2_gather_d_d(<4 x i32> %a0, i8* %a1, <4 x i32> %idx, <4 x i32> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherdd %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x90,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x90,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32> %a0,
|
|
|
|
i8* %a1, <4 x i32> %idx, <4 x i32> %mask, i8 2) ;
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32>, i8*,
|
|
|
|
<4 x i32>, <4 x i32>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <8 x i32> @test_x86_avx2_gather_d_d_256(<8 x i32> %a0, i8* %a1, <8 x i32> %idx, <8 x i32> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_d_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherdd %ymm2, (%eax,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0x6d,0x90,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_d_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherdd %ymm2, (%rdi,%ymm1,2), %ymm0 ## encoding: [0xc4,0xe2,0x6d,0x90,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <8 x i32> @llvm.x86.avx2.gather.d.d.256(<8 x i32> %a0,
|
|
|
|
i8* %a1, <8 x i32> %idx, <8 x i32> %mask, i8 2) ;
|
|
|
|
ret <8 x i32> %res
|
|
|
|
}
|
|
|
|
declare <8 x i32> @llvm.x86.avx2.gather.d.d.256(<8 x i32>, i8*,
|
|
|
|
<8 x i32>, <8 x i32>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x i32> @test_x86_avx2_gather_q_d(<4 x i32> %a0, i8* %a1, <2 x i64> %idx, <4 x i32> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x91,0x04,0x48]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_d:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%xmm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x69,0x91,0x04,0x4f]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32> %a0,
|
|
|
|
i8* %a1, <2 x i64> %idx, <4 x i32> %mask, i8 2) ;
|
|
|
|
ret <4 x i32> %res
|
|
|
|
}
|
|
|
|
declare <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32>, i8*,
|
|
|
|
<2 x i64>, <4 x i32>, i8) nounwind readonly
|
|
|
|
|
2016-03-02 05:38:41 +08:00
|
|
|
define <4 x i32> @test_x86_avx2_gather_q_d_256(<4 x i32> %a0, i8* %a1, <4 x i64> %idx, <4 x i32> %mask) {
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-LABEL: test_x86_avx2_gather_q_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
|
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%ymm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x6d,0x91,0x04,0x48]
|
|
|
|
; X86-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-LABEL: test_x86_avx2_gather_q_d_256:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%ymm1,2), %xmm0 ## encoding: [0xc4,0xe2,0x6d,0x91,0x04,0x4f]
|
|
|
|
; X64-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-NEXT: retq ## encoding: [0xc3]
|
2012-06-29 08:54:20 +08:00
|
|
|
%res = call <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32> %a0,
|
|
|
|
i8* %a1, <4 x i64> %idx, <4 x i32> %mask, i8 2) ;
|
|
|
|
ret <4 x i32> %res
|
2012-06-27 03:47:59 +08:00
|
|
|
}
|
2012-06-29 08:54:20 +08:00
|
|
|
declare <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32>, i8*,
|
|
|
|
<4 x i64>, <4 x i32>, i8) nounwind readonly
|
2012-07-12 14:52:41 +08:00
|
|
|
|
|
|
|
; PR13298
|
2016-03-02 05:38:41 +08:00
|
|
|
define <8 x float> @test_gather_mask(<8 x float> %a0, float* %a, <8 x i32> %idx, <8 x float> %mask, float* nocapture %out) {
|
2012-07-12 14:52:41 +08:00
|
|
|
;; gather with mask
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-LABEL: test_gather_mask:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x08]
|
|
|
|
; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx ## encoding: [0x8b,0x4c,0x24,0x04]
|
|
|
|
; X86-AVX-NEXT: vmovaps %ymm2, %ymm3 ## encoding: [0xc5,0xfc,0x28,0xda]
|
|
|
|
; X86-AVX-NEXT: vgatherdps %ymm3, (%ecx,%ymm1,4), %ymm0 ## encoding: [0xc4,0xe2,0x65,0x92,0x04,0x89]
|
|
|
|
; X86-AVX-NEXT: vmovups %ymm2, (%eax) ## encoding: [0xc5,0xfc,0x11,0x10]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X86-AVX512VL-LABEL: test_gather_mask:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X86-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X86-AVX512VL-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x08]
|
|
|
|
; X86-AVX512VL-NEXT: movl {{[0-9]+}}(%esp), %ecx ## encoding: [0x8b,0x4c,0x24,0x04]
|
|
|
|
; X86-AVX512VL-NEXT: vmovaps %ymm2, %ymm3 ## EVEX TO VEX Compression encoding: [0xc5,0xfc,0x28,0xda]
|
|
|
|
; X86-AVX512VL-NEXT: vgatherdps %ymm3, (%ecx,%ymm1,4), %ymm0 ## encoding: [0xc4,0xe2,0x65,0x92,0x04,0x89]
|
|
|
|
; X86-AVX512VL-NEXT: vmovups %ymm2, (%eax) ## EVEX TO VEX Compression encoding: [0xc5,0xfc,0x11,0x10]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X86-AVX512VL-NEXT: retl ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX-LABEL: test_gather_mask:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX-NEXT: vmovaps %ymm2, %ymm3 ## encoding: [0xc5,0xfc,0x28,0xda]
|
|
|
|
; X64-AVX-NEXT: vgatherdps %ymm3, (%rdi,%ymm1,4), %ymm0 ## encoding: [0xc4,0xe2,0x65,0x92,0x04,0x8f]
|
|
|
|
; X64-AVX-NEXT: vmovups %ymm2, (%rsi) ## encoding: [0xc5,0xfc,0x11,0x16]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX-NEXT: retq ## encoding: [0xc3]
|
2017-10-23 22:19:46 +08:00
|
|
|
;
|
|
|
|
; X64-AVX512VL-LABEL: test_gather_mask:
|
2017-12-05 01:18:51 +08:00
|
|
|
; X64-AVX512VL: ## %bb.0:
|
2017-10-23 22:19:46 +08:00
|
|
|
; X64-AVX512VL-NEXT: vmovaps %ymm2, %ymm3 ## EVEX TO VEX Compression encoding: [0xc5,0xfc,0x28,0xda]
|
|
|
|
; X64-AVX512VL-NEXT: vgatherdps %ymm3, (%rdi,%ymm1,4), %ymm0 ## encoding: [0xc4,0xe2,0x65,0x92,0x04,0x8f]
|
|
|
|
; X64-AVX512VL-NEXT: vmovups %ymm2, (%rsi) ## EVEX TO VEX Compression encoding: [0xc5,0xfc,0x11,0x16]
|
2018-04-13 14:07:18 +08:00
|
|
|
; X64-AVX512VL-NEXT: retq ## encoding: [0xc3]
|
2012-07-12 14:52:41 +08:00
|
|
|
%a_i8 = bitcast float* %a to i8*
|
|
|
|
%res = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> %a0,
|
|
|
|
i8* %a_i8, <8 x i32> %idx, <8 x float> %mask, i8 4) ;
|
|
|
|
|
|
|
|
;; for debugging, we'll just dump out the mask
|
|
|
|
%out_ptr = bitcast float * %out to <8 x float> *
|
|
|
|
store <8 x float> %mask, <8 x float> * %out_ptr, align 4
|
|
|
|
|
|
|
|
ret <8 x float> %res
|
|
|
|
}
|