forked from OSchip/llvm-project
Add strict fp support for instructions fadd/fsub/fmul/fdiv
Differential Revision: https://reviews.llvm.org/D68757
This commit is contained in:
parent
1c918ff2c1
commit
3041434450
|
@ -5219,10 +5219,8 @@ void X86DAGToDAGISel::Select(SDNode *Node) {
|
||||||
SelectCode(Res.getNode());
|
SelectCode(Res.getNode());
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
case ISD::STRICT_FADD:
|
|
||||||
case ISD::STRICT_FSUB:
|
|
||||||
case ISD::STRICT_FP_ROUND: {
|
case ISD::STRICT_FP_ROUND: {
|
||||||
// X87 instructions has enabled these strict fp operation.
|
// X87 instructions has enabled this strict fp operation.
|
||||||
bool UsingFp80 = Node->getSimpleValueType(0) == MVT::f80 ||
|
bool UsingFp80 = Node->getSimpleValueType(0) == MVT::f80 ||
|
||||||
Node->getOperand(1).getSimpleValueType() == MVT::f80;
|
Node->getOperand(1).getSimpleValueType() == MVT::f80;
|
||||||
if (UsingFp80 || (!Subtarget->hasSSE1() && Subtarget->hasX87()))
|
if (UsingFp80 || (!Subtarget->hasSSE1() && Subtarget->hasX87()))
|
||||||
|
|
|
@ -593,8 +593,6 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
|
||||||
setOperationAction(ISD::FSINCOS, VT, Expand);
|
setOperationAction(ISD::FSINCOS, VT, Expand);
|
||||||
|
|
||||||
// Handle constrained floating-point operations of scalar.
|
// Handle constrained floating-point operations of scalar.
|
||||||
setOperationAction(ISD::STRICT_FMUL , VT, Legal);
|
|
||||||
setOperationAction(ISD::STRICT_FDIV , VT, Legal);
|
|
||||||
setOperationAction(ISD::STRICT_FSQRT , VT, Legal);
|
setOperationAction(ISD::STRICT_FSQRT , VT, Legal);
|
||||||
setOperationAction(ISD::STRICT_FP_EXTEND, VT, Legal);
|
setOperationAction(ISD::STRICT_FP_EXTEND, VT, Legal);
|
||||||
// FIXME: When the target is 64-bit, STRICT_FP_ROUND will be overwritten
|
// FIXME: When the target is 64-bit, STRICT_FP_ROUND will be overwritten
|
||||||
|
@ -623,13 +621,15 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
|
||||||
} else // SSE immediates.
|
} else // SSE immediates.
|
||||||
addLegalFPImmediate(APFloat(+0.0)); // xorpd
|
addLegalFPImmediate(APFloat(+0.0)); // xorpd
|
||||||
}
|
}
|
||||||
|
// Handle constrained floating-point operations of scalar.
|
||||||
// FIXME: Mark these legal to prevent them from being expanded to a
|
|
||||||
// libcall in LegalizeDAG. They'll be mutated by X86ISelDAGToDAG::Select.
|
|
||||||
setOperationAction(ISD::STRICT_FADD, MVT::f32, Legal);
|
setOperationAction(ISD::STRICT_FADD, MVT::f32, Legal);
|
||||||
setOperationAction(ISD::STRICT_FADD, MVT::f64, Legal);
|
setOperationAction(ISD::STRICT_FADD, MVT::f64, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::f32, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::f32, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::f64, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::f64, Legal);
|
||||||
|
|
||||||
// We don't support FMA.
|
// We don't support FMA.
|
||||||
setOperationAction(ISD::FMA, MVT::f64, Expand);
|
setOperationAction(ISD::FMA, MVT::f64, Expand);
|
||||||
|
@ -864,6 +864,10 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
|
||||||
setOperationAction(ISD::STRICT_FADD, MVT::v2f64, Legal);
|
setOperationAction(ISD::STRICT_FADD, MVT::v2f64, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v4f32, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v4f32, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v2f64, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v2f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v4f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v2f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v4f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v2f64, Legal);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!Subtarget.useSoftFloat() && Subtarget.hasSSE2()) {
|
if (!Subtarget.useSoftFloat() && Subtarget.hasSSE2()) {
|
||||||
|
@ -1160,6 +1164,10 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
|
||||||
setOperationAction(ISD::STRICT_FADD, MVT::v4f64, Legal);
|
setOperationAction(ISD::STRICT_FADD, MVT::v4f64, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v8f32, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v8f32, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v4f64, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v4f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v8f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v4f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v8f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v4f64, Legal);
|
||||||
|
|
||||||
if (!Subtarget.hasAVX512())
|
if (!Subtarget.hasAVX512())
|
||||||
setOperationAction(ISD::BITCAST, MVT::v32i1, Custom);
|
setOperationAction(ISD::BITCAST, MVT::v32i1, Custom);
|
||||||
|
@ -1429,6 +1437,10 @@ X86TargetLowering::X86TargetLowering(const X86TargetMachine &TM,
|
||||||
setOperationAction(ISD::STRICT_FADD, MVT::v8f64, Legal);
|
setOperationAction(ISD::STRICT_FADD, MVT::v8f64, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v16f32, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v16f32, Legal);
|
||||||
setOperationAction(ISD::STRICT_FSUB, MVT::v8f64, Legal);
|
setOperationAction(ISD::STRICT_FSUB, MVT::v8f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v16f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FMUL, MVT::v8f64, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v16f32, Legal);
|
||||||
|
setOperationAction(ISD::STRICT_FDIV, MVT::v8f64, Legal);
|
||||||
|
|
||||||
setTruncStoreAction(MVT::v8i64, MVT::v8i8, Legal);
|
setTruncStoreAction(MVT::v8i64, MVT::v8i8, Legal);
|
||||||
setTruncStoreAction(MVT::v8i64, MVT::v8i16, Legal);
|
setTruncStoreAction(MVT::v8i64, MVT::v8i16, Legal);
|
||||||
|
|
|
@ -5397,13 +5397,13 @@ multiclass avx512_binop_s_sae<bits<8> opc, string OpcodeStr, SDNode OpNode,
|
||||||
NAME#"SD">,
|
NAME#"SD">,
|
||||||
XD, VEX_W, EVEX_4V, VEX_LIG, EVEX_CD8<64, CD8VT1>;
|
XD, VEX_W, EVEX_4V, VEX_LIG, EVEX_CD8<64, CD8VT1>;
|
||||||
}
|
}
|
||||||
defm VADD : avx512_binop_s_round<0x58, "vadd", fadd, X86fadds, X86faddRnds,
|
defm VADD : avx512_binop_s_round<0x58, "vadd", any_fadd, X86fadds, X86faddRnds,
|
||||||
SchedWriteFAddSizes, 1>;
|
SchedWriteFAddSizes, 1>;
|
||||||
defm VMUL : avx512_binop_s_round<0x59, "vmul", fmul, X86fmuls, X86fmulRnds,
|
defm VMUL : avx512_binop_s_round<0x59, "vmul", any_fmul, X86fmuls, X86fmulRnds,
|
||||||
SchedWriteFMulSizes, 1>;
|
SchedWriteFMulSizes, 1>;
|
||||||
defm VSUB : avx512_binop_s_round<0x5C, "vsub", fsub, X86fsubs, X86fsubRnds,
|
defm VSUB : avx512_binop_s_round<0x5C, "vsub", any_fsub, X86fsubs, X86fsubRnds,
|
||||||
SchedWriteFAddSizes, 0>;
|
SchedWriteFAddSizes, 0>;
|
||||||
defm VDIV : avx512_binop_s_round<0x5E, "vdiv", fdiv, X86fdivs, X86fdivRnds,
|
defm VDIV : avx512_binop_s_round<0x5E, "vdiv", any_fdiv, X86fdivs, X86fdivRnds,
|
||||||
SchedWriteFDivSizes, 0>;
|
SchedWriteFDivSizes, 0>;
|
||||||
defm VMIN : avx512_binop_s_sae<0x5D, "vmin", X86fmin, X86fmins, X86fminSAEs,
|
defm VMIN : avx512_binop_s_sae<0x5D, "vmin", X86fmin, X86fmins, X86fminSAEs,
|
||||||
SchedWriteFCmpSizes, 0>;
|
SchedWriteFCmpSizes, 0>;
|
||||||
|
@ -5555,16 +5555,16 @@ multiclass avx512_fp_binop_p_sae<bits<8> opc, string OpcodeStr, SDNode OpNodeRnd
|
||||||
EVEX_V512, PD, VEX_W,EVEX_CD8<64, CD8VF>;
|
EVEX_V512, PD, VEX_W,EVEX_CD8<64, CD8VF>;
|
||||||
}
|
}
|
||||||
|
|
||||||
defm VADD : avx512_fp_binop_p<0x58, "vadd", fadd, HasAVX512,
|
defm VADD : avx512_fp_binop_p<0x58, "vadd", any_fadd, HasAVX512,
|
||||||
SchedWriteFAddSizes, 1>,
|
SchedWriteFAddSizes, 1>,
|
||||||
avx512_fp_binop_p_round<0x58, "vadd", X86faddRnd, SchedWriteFAddSizes>;
|
avx512_fp_binop_p_round<0x58, "vadd", X86faddRnd, SchedWriteFAddSizes>;
|
||||||
defm VMUL : avx512_fp_binop_p<0x59, "vmul", fmul, HasAVX512,
|
defm VMUL : avx512_fp_binop_p<0x59, "vmul", any_fmul, HasAVX512,
|
||||||
SchedWriteFMulSizes, 1>,
|
SchedWriteFMulSizes, 1>,
|
||||||
avx512_fp_binop_p_round<0x59, "vmul", X86fmulRnd, SchedWriteFMulSizes>;
|
avx512_fp_binop_p_round<0x59, "vmul", X86fmulRnd, SchedWriteFMulSizes>;
|
||||||
defm VSUB : avx512_fp_binop_p<0x5C, "vsub", fsub, HasAVX512,
|
defm VSUB : avx512_fp_binop_p<0x5C, "vsub", any_fsub, HasAVX512,
|
||||||
SchedWriteFAddSizes>,
|
SchedWriteFAddSizes>,
|
||||||
avx512_fp_binop_p_round<0x5C, "vsub", X86fsubRnd, SchedWriteFAddSizes>;
|
avx512_fp_binop_p_round<0x5C, "vsub", X86fsubRnd, SchedWriteFAddSizes>;
|
||||||
defm VDIV : avx512_fp_binop_p<0x5E, "vdiv", fdiv, HasAVX512,
|
defm VDIV : avx512_fp_binop_p<0x5E, "vdiv", any_fdiv, HasAVX512,
|
||||||
SchedWriteFDivSizes>,
|
SchedWriteFDivSizes>,
|
||||||
avx512_fp_binop_p_round<0x5E, "vdiv", X86fdivRnd, SchedWriteFDivSizes>;
|
avx512_fp_binop_p_round<0x5E, "vdiv", X86fdivRnd, SchedWriteFDivSizes>;
|
||||||
defm VMIN : avx512_fp_binop_p<0x5D, "vmin", X86fmin, HasAVX512,
|
defm VMIN : avx512_fp_binop_p<0x5D, "vmin", X86fmin, HasAVX512,
|
||||||
|
|
|
@ -2667,18 +2667,18 @@ let Uses = [MXCSR], mayRaiseFPException = 1 in {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Binary Arithmetic instructions
|
// Binary Arithmetic instructions
|
||||||
defm ADD : basic_sse12_fp_binop_p<0x58, "add", fadd, SchedWriteFAddSizes>,
|
defm ADD : basic_sse12_fp_binop_p<0x58, "add", any_fadd, SchedWriteFAddSizes>,
|
||||||
basic_sse12_fp_binop_s<0x58, "add", fadd, SchedWriteFAddSizes>,
|
basic_sse12_fp_binop_s<0x58, "add", any_fadd, SchedWriteFAddSizes>,
|
||||||
basic_sse12_fp_binop_s_int<0x58, "add", null_frag, SchedWriteFAddSizes>;
|
basic_sse12_fp_binop_s_int<0x58, "add", null_frag, SchedWriteFAddSizes>;
|
||||||
defm MUL : basic_sse12_fp_binop_p<0x59, "mul", fmul, SchedWriteFMulSizes>,
|
defm MUL : basic_sse12_fp_binop_p<0x59, "mul", any_fmul, SchedWriteFMulSizes>,
|
||||||
basic_sse12_fp_binop_s<0x59, "mul", fmul, SchedWriteFMulSizes>,
|
basic_sse12_fp_binop_s<0x59, "mul", any_fmul, SchedWriteFMulSizes>,
|
||||||
basic_sse12_fp_binop_s_int<0x59, "mul", null_frag, SchedWriteFMulSizes>;
|
basic_sse12_fp_binop_s_int<0x59, "mul", null_frag, SchedWriteFMulSizes>;
|
||||||
let isCommutable = 0 in {
|
let isCommutable = 0 in {
|
||||||
defm SUB : basic_sse12_fp_binop_p<0x5C, "sub", fsub, SchedWriteFAddSizes>,
|
defm SUB : basic_sse12_fp_binop_p<0x5C, "sub", any_fsub, SchedWriteFAddSizes>,
|
||||||
basic_sse12_fp_binop_s<0x5C, "sub", fsub, SchedWriteFAddSizes>,
|
basic_sse12_fp_binop_s<0x5C, "sub", any_fsub, SchedWriteFAddSizes>,
|
||||||
basic_sse12_fp_binop_s_int<0x5C, "sub", null_frag, SchedWriteFAddSizes>;
|
basic_sse12_fp_binop_s_int<0x5C, "sub", null_frag, SchedWriteFAddSizes>;
|
||||||
defm DIV : basic_sse12_fp_binop_p<0x5E, "div", fdiv, SchedWriteFDivSizes>,
|
defm DIV : basic_sse12_fp_binop_p<0x5E, "div", any_fdiv, SchedWriteFDivSizes>,
|
||||||
basic_sse12_fp_binop_s<0x5E, "div", fdiv, SchedWriteFDivSizes>,
|
basic_sse12_fp_binop_s<0x5E, "div", any_fdiv, SchedWriteFDivSizes>,
|
||||||
basic_sse12_fp_binop_s_int<0x5E, "div", null_frag, SchedWriteFDivSizes>;
|
basic_sse12_fp_binop_s_int<0x5E, "div", null_frag, SchedWriteFDivSizes>;
|
||||||
defm MAX : basic_sse12_fp_binop_p<0x5F, "max", X86fmax, SchedWriteFCmpSizes>,
|
defm MAX : basic_sse12_fp_binop_p<0x5F, "max", X86fmax, SchedWriteFCmpSizes>,
|
||||||
basic_sse12_fp_binop_s<0x5F, "max", X86fmax, SchedWriteFCmpSizes>,
|
basic_sse12_fp_binop_s<0x5F, "max", X86fmax, SchedWriteFCmpSizes>,
|
||||||
|
@ -2773,15 +2773,15 @@ multiclass scalar_math_patterns<SDNode Op, string OpcPrefix, SDNode Move,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
defm : scalar_math_patterns<fadd, "ADDSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
defm : scalar_math_patterns<any_fadd, "ADDSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
||||||
defm : scalar_math_patterns<fsub, "SUBSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
defm : scalar_math_patterns<any_fsub, "SUBSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
||||||
defm : scalar_math_patterns<fmul, "MULSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
defm : scalar_math_patterns<any_fmul, "MULSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
||||||
defm : scalar_math_patterns<fdiv, "DIVSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
defm : scalar_math_patterns<any_fdiv, "DIVSS", X86Movss, v4f32, f32, FR32, loadf32, UseSSE1>;
|
||||||
|
|
||||||
defm : scalar_math_patterns<fadd, "ADDSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
defm : scalar_math_patterns<any_fadd, "ADDSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
||||||
defm : scalar_math_patterns<fsub, "SUBSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
defm : scalar_math_patterns<any_fsub, "SUBSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
||||||
defm : scalar_math_patterns<fmul, "MULSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
defm : scalar_math_patterns<any_fmul, "MULSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
||||||
defm : scalar_math_patterns<fdiv, "DIVSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
defm : scalar_math_patterns<any_fdiv, "DIVSD", X86Movsd, v2f64, f64, FR64, loadf64, UseSSE2>;
|
||||||
|
|
||||||
/// Unop Arithmetic
|
/// Unop Arithmetic
|
||||||
/// In addition, we also have a special variant of the scalar form here to
|
/// In addition, we also have a special variant of the scalar form here to
|
||||||
|
|
|
@ -1,10 +1,10 @@
|
||||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+sse2 -O3 | FileCheck %s --check-prefixes=CHECK,SSE
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+sse2 -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,SSE
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse2 -O3 | FileCheck %s --check-prefixes=CHECK,SSE
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse2 -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,SSE
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx -O3 | FileCheck %s --check-prefixes=CHECK,AVX
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,AVX
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx -O3 | FileCheck %s --check-prefixes=CHECK,AVX
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,AVX
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s --check-prefixes=CHECK,AVX
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,AVX
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s --check-prefixes=CHECK,AVX
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s --check-prefixes=CHECK,AVX
|
||||||
|
|
||||||
declare <2 x double> @llvm.experimental.constrained.fadd.v2f64(<2 x double>, <2 x double>, metadata, metadata)
|
declare <2 x double> @llvm.experimental.constrained.fadd.v2f64(<2 x double>, <2 x double>, metadata, metadata)
|
||||||
declare <4 x float> @llvm.experimental.constrained.fadd.v4f32(<4 x float>, <4 x float>, metadata, metadata)
|
declare <4 x float> @llvm.experimental.constrained.fadd.v4f32(<4 x float>, <4 x float>, metadata, metadata)
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
|
|
||||||
declare <4 x double> @llvm.experimental.constrained.fadd.v4f64(<4 x double>, <4 x double>, metadata, metadata)
|
declare <4 x double> @llvm.experimental.constrained.fadd.v4f64(<4 x double>, <4 x double>, metadata, metadata)
|
||||||
declare <8 x float> @llvm.experimental.constrained.fadd.v8f32(<8 x float>, <8 x float>, metadata, metadata)
|
declare <8 x float> @llvm.experimental.constrained.fadd.v8f32(<8 x float>, <8 x float>, metadata, metadata)
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||||
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 | FileCheck %s
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -mattr=+avx512vl -O3 -disable-strictnode-mutation | FileCheck %s
|
||||||
|
|
||||||
declare <8 x double> @llvm.experimental.constrained.fadd.v8f64(<8 x double>, <8 x double>, metadata, metadata)
|
declare <8 x double> @llvm.experimental.constrained.fadd.v8f64(<8 x double>, <8 x double>, metadata, metadata)
|
||||||
declare <16 x float> @llvm.experimental.constrained.fadd.v16f32(<16 x float>, <16 x float>, metadata, metadata)
|
declare <16 x float> @llvm.experimental.constrained.fadd.v16f32(<16 x float>, <16 x float>, metadata, metadata)
|
||||||
|
|
|
@ -115,10 +115,10 @@ define <4 x double> @constrained_vector_fdiv_v4f64() #0 {
|
||||||
; CHECK-LABEL: constrained_vector_fdiv_v4f64:
|
; CHECK-LABEL: constrained_vector_fdiv_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm2 = [1.0E+1,1.0E+1]
|
; CHECK-NEXT: movapd {{.*#+}} xmm2 = [1.0E+1,1.0E+1]
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [1.0E+0,2.0E+0]
|
|
||||||
; CHECK-NEXT: divpd %xmm2, %xmm0
|
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [3.0E+0,4.0E+0]
|
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [3.0E+0,4.0E+0]
|
||||||
; CHECK-NEXT: divpd %xmm2, %xmm1
|
; CHECK-NEXT: divpd %xmm2, %xmm1
|
||||||
|
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [1.0E+0,2.0E+0]
|
||||||
|
; CHECK-NEXT: divpd %xmm2, %xmm0
|
||||||
; CHECK-NEXT: retq
|
; CHECK-NEXT: retq
|
||||||
;
|
;
|
||||||
; AVX1-LABEL: constrained_vector_fdiv_v4f64:
|
; AVX1-LABEL: constrained_vector_fdiv_v4f64:
|
||||||
|
@ -507,10 +507,10 @@ entry:
|
||||||
define <4 x double> @constrained_vector_fmul_v4f64() #0 {
|
define <4 x double> @constrained_vector_fmul_v4f64() #0 {
|
||||||
; CHECK-LABEL: constrained_vector_fmul_v4f64:
|
; CHECK-LABEL: constrained_vector_fmul_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [1.7976931348623157E+308,1.7976931348623157E+308]
|
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [1.7976931348623157E+308,1.7976931348623157E+308]
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [2.0E+0,3.0E+0]
|
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [4.0E+0,5.0E+0]
|
||||||
; CHECK-NEXT: mulpd %xmm1, %xmm0
|
; CHECK-NEXT: mulpd %xmm0, %xmm1
|
||||||
; CHECK-NEXT: mulpd {{.*}}(%rip), %xmm1
|
; CHECK-NEXT: mulpd {{.*}}(%rip), %xmm0
|
||||||
; CHECK-NEXT: retq
|
; CHECK-NEXT: retq
|
||||||
;
|
;
|
||||||
; AVX1-LABEL: constrained_vector_fmul_v4f64:
|
; AVX1-LABEL: constrained_vector_fmul_v4f64:
|
||||||
|
@ -644,10 +644,10 @@ entry:
|
||||||
define <4 x double> @constrained_vector_fadd_v4f64() #0 {
|
define <4 x double> @constrained_vector_fadd_v4f64() #0 {
|
||||||
; CHECK-LABEL: constrained_vector_fadd_v4f64:
|
; CHECK-LABEL: constrained_vector_fadd_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [1.7976931348623157E+308,1.7976931348623157E+308]
|
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [1.7976931348623157E+308,1.7976931348623157E+308]
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [1.0E+0,1.0000000000000001E-1]
|
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [2.0E+0,2.0000000000000001E-1]
|
||||||
; CHECK-NEXT: addpd %xmm1, %xmm0
|
; CHECK-NEXT: addpd %xmm0, %xmm1
|
||||||
; CHECK-NEXT: addpd {{.*}}(%rip), %xmm1
|
; CHECK-NEXT: addpd {{.*}}(%rip), %xmm0
|
||||||
; CHECK-NEXT: retq
|
; CHECK-NEXT: retq
|
||||||
;
|
;
|
||||||
; AVX1-LABEL: constrained_vector_fadd_v4f64:
|
; AVX1-LABEL: constrained_vector_fadd_v4f64:
|
||||||
|
@ -784,10 +784,10 @@ entry:
|
||||||
define <4 x double> @constrained_vector_fsub_v4f64() #0 {
|
define <4 x double> @constrained_vector_fsub_v4f64() #0 {
|
||||||
; CHECK-LABEL: constrained_vector_fsub_v4f64:
|
; CHECK-LABEL: constrained_vector_fsub_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: movapd {{.*#+}} xmm1 = [-1.7976931348623157E+308,-1.7976931348623157E+308]
|
; CHECK-NEXT: movapd {{.*#+}} xmm0 = [-1.7976931348623157E+308,-1.7976931348623157E+308]
|
||||||
; CHECK-NEXT: movapd %xmm1, %xmm0
|
; CHECK-NEXT: movapd %xmm0, %xmm1
|
||||||
; CHECK-NEXT: subpd {{.*}}(%rip), %xmm0
|
|
||||||
; CHECK-NEXT: subpd {{.*}}(%rip), %xmm1
|
; CHECK-NEXT: subpd {{.*}}(%rip), %xmm1
|
||||||
|
; CHECK-NEXT: subpd {{.*}}(%rip), %xmm0
|
||||||
; CHECK-NEXT: retq
|
; CHECK-NEXT: retq
|
||||||
;
|
;
|
||||||
; AVX1-LABEL: constrained_vector_fsub_v4f64:
|
; AVX1-LABEL: constrained_vector_fsub_v4f64:
|
||||||
|
|
Loading…
Reference in New Issue