forked from OSchip/llvm-project
[X86] Add AVX512 equivalents of some isel patterns so we get EVEX instructions.
These are the patterns for matching fceil, ffloor, and sqrt to intrinsic instructions if they have a MOVSS/SD. llvm-svn: 336951
This commit is contained in:
parent
b01a355354
commit
3b837b6b63
|
@ -11525,6 +11525,37 @@ defm : AVX512_scalar_math_fp_patterns<fsub, "SUBSD", X86Movsd, v2f64x_info, fp64
|
||||||
defm : AVX512_scalar_math_fp_patterns<fmul, "MULSD", X86Movsd, v2f64x_info, fp64imm0>;
|
defm : AVX512_scalar_math_fp_patterns<fmul, "MULSD", X86Movsd, v2f64x_info, fp64imm0>;
|
||||||
defm : AVX512_scalar_math_fp_patterns<fdiv, "DIVSD", X86Movsd, v2f64x_info, fp64imm0>;
|
defm : AVX512_scalar_math_fp_patterns<fdiv, "DIVSD", X86Movsd, v2f64x_info, fp64imm0>;
|
||||||
|
|
||||||
|
multiclass AVX512_scalar_unary_math_patterns<SDNode OpNode, string OpcPrefix,
|
||||||
|
SDNode Move, X86VectorVTInfo _> {
|
||||||
|
let Predicates = [HasAVX512] in {
|
||||||
|
def : Pat<(_.VT (Move _.VT:$dst,
|
||||||
|
(scalar_to_vector (OpNode (extractelt _.VT:$src, 0))))),
|
||||||
|
(!cast<Instruction>("V"#OpcPrefix#Zr_Int) _.VT:$dst, _.VT:$src)>;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
defm : AVX512_scalar_unary_math_patterns<fsqrt, "SQRTSS", X86Movss, v4f32x_info>;
|
||||||
|
defm : AVX512_scalar_unary_math_patterns<fsqrt, "SQRTSD", X86Movsd, v2f64x_info>;
|
||||||
|
|
||||||
|
multiclass AVX512_scalar_unary_math_imm_patterns<SDNode OpNode, string OpcPrefix,
|
||||||
|
SDNode Move, X86VectorVTInfo _,
|
||||||
|
bits<8> ImmV> {
|
||||||
|
let Predicates = [HasAVX512] in {
|
||||||
|
def : Pat<(_.VT (Move _.VT:$dst,
|
||||||
|
(scalar_to_vector (OpNode (extractelt _.VT:$src, 0))))),
|
||||||
|
(!cast<Instruction>("V"#OpcPrefix#Zr_Int) _.VT:$dst, _.VT:$src,
|
||||||
|
(i32 ImmV))>;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
defm : AVX512_scalar_unary_math_imm_patterns<ffloor, "RNDSCALESS", X86Movss,
|
||||||
|
v4f32x_info, 0x01>;
|
||||||
|
defm : AVX512_scalar_unary_math_imm_patterns<fceil, "RNDSCALESS", X86Movss,
|
||||||
|
v4f32x_info, 0x02>;
|
||||||
|
defm : AVX512_scalar_unary_math_imm_patterns<ffloor, "RNDSCALESD", X86Movsd,
|
||||||
|
v2f64x_info, 0x01>;
|
||||||
|
defm : AVX512_scalar_unary_math_imm_patterns<fceil, "RNDSCALESD", X86Movsd,
|
||||||
|
v2f64x_info, 0x02>;
|
||||||
|
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
// AES instructions
|
// AES instructions
|
||||||
|
|
|
@ -2558,10 +2558,15 @@ define <4 x float> @test_mm_sqrt_ss(<4 x float> %a0) {
|
||||||
; SSE-NEXT: sqrtss %xmm0, %xmm0 # encoding: [0xf3,0x0f,0x51,0xc0]
|
; SSE-NEXT: sqrtss %xmm0, %xmm0 # encoding: [0xf3,0x0f,0x51,0xc0]
|
||||||
; SSE-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
; SSE-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
;
|
;
|
||||||
; AVX-LABEL: test_mm_sqrt_ss:
|
; AVX1-LABEL: test_mm_sqrt_ss:
|
||||||
; AVX: # %bb.0:
|
; AVX1: # %bb.0:
|
||||||
; AVX-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 # encoding: [0xc5,0xfa,0x51,0xc0]
|
; AVX1-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 # encoding: [0xc5,0xfa,0x51,0xc0]
|
||||||
; AVX-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
; AVX1-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
|
;
|
||||||
|
; AVX512-LABEL: test_mm_sqrt_ss:
|
||||||
|
; AVX512: # %bb.0:
|
||||||
|
; AVX512-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x51,0xc0]
|
||||||
|
; AVX512-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
%ext = extractelement <4 x float> %a0, i32 0
|
%ext = extractelement <4 x float> %a0, i32 0
|
||||||
%sqrt = call float @llvm.sqrt.f32(float %ext)
|
%sqrt = call float @llvm.sqrt.f32(float %ext)
|
||||||
%ins = insertelement <4 x float> %a0, float %sqrt, i32 0
|
%ins = insertelement <4 x float> %a0, float %sqrt, i32 0
|
||||||
|
|
|
@ -34,10 +34,15 @@ define <4 x float> @test_x86_sse_sqrt_ss(<4 x float> %a0) {
|
||||||
; SSE-NEXT: sqrtss %xmm0, %xmm0 ## encoding: [0xf3,0x0f,0x51,0xc0]
|
; SSE-NEXT: sqrtss %xmm0, %xmm0 ## encoding: [0xf3,0x0f,0x51,0xc0]
|
||||||
; SSE-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
; SSE-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
;
|
;
|
||||||
; AVX-LABEL: test_x86_sse_sqrt_ss:
|
; AVX1-LABEL: test_x86_sse_sqrt_ss:
|
||||||
; AVX: ## %bb.0:
|
; AVX1: ## %bb.0:
|
||||||
; AVX-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfa,0x51,0xc0]
|
; AVX1-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfa,0x51,0xc0]
|
||||||
; AVX-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
; AVX1-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
|
;
|
||||||
|
; AVX512-LABEL: test_x86_sse_sqrt_ss:
|
||||||
|
; AVX512: ## %bb.0:
|
||||||
|
; AVX512-NEXT: vsqrtss %xmm0, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfa,0x51,0xc0]
|
||||||
|
; AVX512-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
%res = call <4 x float> @llvm.x86.sse.sqrt.ss(<4 x float> %a0) ; <<4 x float>> [#uses=1]
|
%res = call <4 x float> @llvm.x86.sse.sqrt.ss(<4 x float> %a0) ; <<4 x float>> [#uses=1]
|
||||||
ret <4 x float> %res
|
ret <4 x float> %res
|
||||||
}
|
}
|
||||||
|
|
|
@ -4896,10 +4896,15 @@ define <2 x double> @test_mm_sqrt_sd(<2 x double> %a0, <2 x double> %a1) nounwin
|
||||||
; SSE-NEXT: movapd %xmm1, %xmm0 # encoding: [0x66,0x0f,0x28,0xc1]
|
; SSE-NEXT: movapd %xmm1, %xmm0 # encoding: [0x66,0x0f,0x28,0xc1]
|
||||||
; SSE-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
; SSE-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
;
|
;
|
||||||
; AVX-LABEL: test_mm_sqrt_sd:
|
; AVX1-LABEL: test_mm_sqrt_sd:
|
||||||
; AVX: # %bb.0:
|
; AVX1: # %bb.0:
|
||||||
; AVX-NEXT: vsqrtsd %xmm0, %xmm1, %xmm0 # encoding: [0xc5,0xf3,0x51,0xc0]
|
; AVX1-NEXT: vsqrtsd %xmm0, %xmm1, %xmm0 # encoding: [0xc5,0xf3,0x51,0xc0]
|
||||||
; AVX-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
; AVX1-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
|
;
|
||||||
|
; AVX512-LABEL: test_mm_sqrt_sd:
|
||||||
|
; AVX512: # %bb.0:
|
||||||
|
; AVX512-NEXT: vsqrtsd %xmm0, %xmm1, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xf3,0x51,0xc0]
|
||||||
|
; AVX512-NEXT: ret{{[l|q]}} # encoding: [0xc3]
|
||||||
%ext = extractelement <2 x double> %a0, i32 0
|
%ext = extractelement <2 x double> %a0, i32 0
|
||||||
%sqrt = call double @llvm.sqrt.f64(double %ext)
|
%sqrt = call double @llvm.sqrt.f64(double %ext)
|
||||||
%ins = insertelement <2 x double> %a1, double %sqrt, i32 0
|
%ins = insertelement <2 x double> %a1, double %sqrt, i32 0
|
||||||
|
|
|
@ -34,10 +34,15 @@ define <2 x double> @test_x86_sse2_sqrt_sd(<2 x double> %a0) {
|
||||||
; SSE-NEXT: sqrtsd %xmm0, %xmm0 ## encoding: [0xf2,0x0f,0x51,0xc0]
|
; SSE-NEXT: sqrtsd %xmm0, %xmm0 ## encoding: [0xf2,0x0f,0x51,0xc0]
|
||||||
; SSE-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
; SSE-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
;
|
;
|
||||||
; AVX-LABEL: test_x86_sse2_sqrt_sd:
|
; AVX1-LABEL: test_x86_sse2_sqrt_sd:
|
||||||
; AVX: ## %bb.0:
|
; AVX1: ## %bb.0:
|
||||||
; AVX-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x51,0xc0]
|
; AVX1-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x51,0xc0]
|
||||||
; AVX-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
; AVX1-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
|
;
|
||||||
|
; AVX512-LABEL: test_x86_sse2_sqrt_sd:
|
||||||
|
; AVX512: ## %bb.0:
|
||||||
|
; AVX512-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfb,0x51,0xc0]
|
||||||
|
; AVX512-NEXT: ret{{[l|q]}} ## encoding: [0xc3]
|
||||||
%res = call <2 x double> @llvm.x86.sse2.sqrt.sd(<2 x double> %a0) ; <<2 x double>> [#uses=1]
|
%res = call <2 x double> @llvm.x86.sse2.sqrt.sd(<2 x double> %a0) ; <<2 x double>> [#uses=1]
|
||||||
ret <2 x double> %res
|
ret <2 x double> %res
|
||||||
}
|
}
|
||||||
|
@ -63,7 +68,7 @@ define <2 x double> @test_x86_sse2_sqrt_sd_vec_load(<2 x double>* %a0) {
|
||||||
; X86-AVX512: ## %bb.0:
|
; X86-AVX512: ## %bb.0:
|
||||||
; X86-AVX512-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
; X86-AVX512-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04]
|
||||||
; X86-AVX512-NEXT: vmovapd (%eax), %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xf9,0x28,0x00]
|
; X86-AVX512-NEXT: vmovapd (%eax), %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xf9,0x28,0x00]
|
||||||
; X86-AVX512-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x51,0xc0]
|
; X86-AVX512-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfb,0x51,0xc0]
|
||||||
; X86-AVX512-NEXT: retl ## encoding: [0xc3]
|
; X86-AVX512-NEXT: retl ## encoding: [0xc3]
|
||||||
;
|
;
|
||||||
; X64-SSE-LABEL: test_x86_sse2_sqrt_sd_vec_load:
|
; X64-SSE-LABEL: test_x86_sse2_sqrt_sd_vec_load:
|
||||||
|
@ -81,7 +86,7 @@ define <2 x double> @test_x86_sse2_sqrt_sd_vec_load(<2 x double>* %a0) {
|
||||||
; X64-AVX512-LABEL: test_x86_sse2_sqrt_sd_vec_load:
|
; X64-AVX512-LABEL: test_x86_sse2_sqrt_sd_vec_load:
|
||||||
; X64-AVX512: ## %bb.0:
|
; X64-AVX512: ## %bb.0:
|
||||||
; X64-AVX512-NEXT: vmovapd (%rdi), %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xf9,0x28,0x07]
|
; X64-AVX512-NEXT: vmovapd (%rdi), %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xf9,0x28,0x07]
|
||||||
; X64-AVX512-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x51,0xc0]
|
; X64-AVX512-NEXT: vsqrtsd %xmm0, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfb,0x51,0xc0]
|
||||||
; X64-AVX512-NEXT: retq ## encoding: [0xc3]
|
; X64-AVX512-NEXT: retq ## encoding: [0xc3]
|
||||||
%a1 = load <2 x double>, <2 x double>* %a0, align 16
|
%a1 = load <2 x double>, <2 x double>* %a0, align 16
|
||||||
%res = call <2 x double> @llvm.x86.sse2.sqrt.sd(<2 x double> %a1) ; <<2 x double>> [#uses=1]
|
%res = call <2 x double> @llvm.x86.sse2.sqrt.sd(<2 x double> %a1) ; <<2 x double>> [#uses=1]
|
||||||
|
|
Loading…
Reference in New Issue