[X86] Add avx and avx512f RUN lines to fp128-cast.ll

llvm-svn: 371356
This commit is contained in:
Craig Topper 2019-09-09 01:34:55 +00:00
parent debac75dea
commit 861d343949
1 changed files with 395 additions and 186 deletions

View File

@ -1,7 +1,11 @@
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -O2 -mtriple=x86_64-linux-android -mattr=+sse | FileCheck %s --check-prefix=X64
; RUN: llc < %s -O2 -mtriple=x86_64-linux-gnu -mattr=+sse | FileCheck %s --check-prefix=X64
; RUN: llc < %s -O2 -mtriple=x86_64-linux-android -mattr=+sse | FileCheck %s --check-prefixes=X64,X64-SSE
; RUN: llc < %s -O2 -mtriple=x86_64-linux-gnu -mattr=+sse | FileCheck %s --check-prefixes=X64,X64-SSE
; RUN: llc < %s -O2 -mtriple=i686-linux-gnu -mattr=+mmx | FileCheck %s --check-prefix=X32
; RUN: llc < %s -O2 -mtriple=x86_64-linux-android -mattr=+avx | FileCheck %s --check-prefixes=X64,X64-AVX
; RUN: llc < %s -O2 -mtriple=x86_64-linux-gnu -mattr=+avx | FileCheck %s --check-prefixes=X64,X64-AVX
; RUN: llc < %s -O2 -mtriple=x86_64-linux-android -mattr=+avx512f | FileCheck %s --check-prefixes=X64,X64-AVX
; RUN: llc < %s -O2 -mtriple=x86_64-linux-gnu -mattr=+avx512f | FileCheck %s --check-prefixes=X64,X64-AVX
; Check soft floating point conversion function calls.
@ -15,14 +19,14 @@
@vf128 = common global fp128 0xL00000000000000000000000000000000, align 16
define void @TestFPExtF32_F128() nounwind {
; X64-LABEL: TestFPExtF32_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
; X64-NEXT: callq __extendsftf2
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPExtF32_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
; X64-SSE-NEXT: callq __extendsftf2
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPExtF32_F128:
; X32: # %bb.0: # %entry
@ -45,6 +49,15 @@ define void @TestFPExtF32_F128() nounwind {
; X32-NEXT: addl $24, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPExtF32_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero
; X64-AVX-NEXT: callq __extendsftf2
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load float, float* @vf32, align 4
%conv = fpext float %0 to fp128
@ -53,14 +66,14 @@ entry:
}
define void @TestFPExtF64_F128() nounwind {
; X64-LABEL: TestFPExtF64_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero
; X64-NEXT: callq __extenddftf2
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPExtF64_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero
; X64-SSE-NEXT: callq __extenddftf2
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPExtF64_F128:
; X32: # %bb.0: # %entry
@ -83,6 +96,15 @@ define void @TestFPExtF64_F128() nounwind {
; X32-NEXT: addl $40, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPExtF64_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero
; X64-AVX-NEXT: callq __extenddftf2
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load double, double* @vf64, align 8
%conv = fpext double %0 to fp128
@ -91,15 +113,15 @@ entry:
}
define void @TestFPExtF80_F128() nounwind {
; X64-LABEL: TestFPExtF80_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: subq $24, %rsp
; X64-NEXT: fldt {{.*}}(%rip)
; X64-NEXT: fstpt (%rsp)
; X64-NEXT: callq __extendxftf2
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: addq $24, %rsp
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPExtF80_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: subq $24, %rsp
; X64-SSE-NEXT: fldt {{.*}}(%rip)
; X64-SSE-NEXT: fstpt (%rsp)
; X64-SSE-NEXT: callq __extendxftf2
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: addq $24, %rsp
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPExtF80_F128:
; X32: # %bb.0: # %entry
@ -122,6 +144,16 @@ define void @TestFPExtF80_F128() nounwind {
; X32-NEXT: addl $40, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPExtF80_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: subq $24, %rsp
; X64-AVX-NEXT: fldt {{.*}}(%rip)
; X64-AVX-NEXT: fstpt (%rsp)
; X64-AVX-NEXT: callq __extendxftf2
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: addq $24, %rsp
; X64-AVX-NEXT: retq
entry:
%0 = load x86_fp80, x86_fp80* @vf80, align 8
%conv = fpext x86_fp80 %0 to fp128
@ -130,14 +162,14 @@ entry:
}
define void @TestFPToSIF128_I32() nounwind {
; X64-LABEL: TestFPToSIF128_I32:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __fixtfsi
; X64-NEXT: movl %eax, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPToSIF128_I32:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __fixtfsi
; X64-SSE-NEXT: movl %eax, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPToSIF128_I32:
; X32: # %bb.0: # %entry
@ -151,6 +183,15 @@ define void @TestFPToSIF128_I32() nounwind {
; X32-NEXT: movl %eax, vi32
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPToSIF128_I32:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __fixtfsi
; X64-AVX-NEXT: movl %eax, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptosi fp128 %0 to i32
@ -159,14 +200,14 @@ entry:
}
define void @TestFPToUIF128_U32() nounwind {
; X64-LABEL: TestFPToUIF128_U32:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __fixunstfsi
; X64-NEXT: movl %eax, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPToUIF128_U32:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __fixunstfsi
; X64-SSE-NEXT: movl %eax, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPToUIF128_U32:
; X32: # %bb.0: # %entry
@ -180,6 +221,15 @@ define void @TestFPToUIF128_U32() nounwind {
; X32-NEXT: movl %eax, vu32
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPToUIF128_U32:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __fixunstfsi
; X64-AVX-NEXT: movl %eax, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptoui fp128 %0 to i32
@ -188,15 +238,15 @@ entry:
}
define void @TestFPToSIF128_I64() nounwind {
; X64-LABEL: TestFPToSIF128_I64:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __fixtfsi
; X64-NEXT: cltq
; X64-NEXT: movq %rax, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPToSIF128_I64:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __fixtfsi
; X64-SSE-NEXT: cltq
; X64-SSE-NEXT: movq %rax, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPToSIF128_I64:
; X32: # %bb.0: # %entry
@ -212,6 +262,16 @@ define void @TestFPToSIF128_I64() nounwind {
; X32-NEXT: movl %eax, vi64+4
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPToSIF128_I64:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __fixtfsi
; X64-AVX-NEXT: cltq
; X64-AVX-NEXT: movq %rax, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptosi fp128 %0 to i32
@ -221,15 +281,15 @@ entry:
}
define void @TestFPToUIF128_U64() nounwind {
; X64-LABEL: TestFPToUIF128_U64:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __fixunstfsi
; X64-NEXT: movl %eax, %eax
; X64-NEXT: movq %rax, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPToUIF128_U64:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __fixunstfsi
; X64-SSE-NEXT: movl %eax, %eax
; X64-SSE-NEXT: movq %rax, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPToUIF128_U64:
; X32: # %bb.0: # %entry
@ -244,6 +304,16 @@ define void @TestFPToUIF128_U64() nounwind {
; X32-NEXT: movl $0, vu64+4
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPToUIF128_U64:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __fixunstfsi
; X64-AVX-NEXT: movl %eax, %eax
; X64-AVX-NEXT: movq %rax, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptoui fp128 %0 to i32
@ -253,14 +323,14 @@ entry:
}
define void @TestFPTruncF128_F32() nounwind {
; X64-LABEL: TestFPTruncF128_F32:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __trunctfsf2
; X64-NEXT: movss %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPTruncF128_F32:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __trunctfsf2
; X64-SSE-NEXT: movss %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPTruncF128_F32:
; X32: # %bb.0: # %entry
@ -274,6 +344,15 @@ define void @TestFPTruncF128_F32() nounwind {
; X32-NEXT: fstps vf32
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPTruncF128_F32:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __trunctfsf2
; X64-AVX-NEXT: vmovss %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptrunc fp128 %0 to float
@ -282,14 +361,14 @@ entry:
}
define void @TestFPTruncF128_F64() nounwind {
; X64-LABEL: TestFPTruncF128_F64:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __trunctfdf2
; X64-NEXT: movsd %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPTruncF128_F64:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __trunctfdf2
; X64-SSE-NEXT: movsd %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPTruncF128_F64:
; X32: # %bb.0: # %entry
@ -303,6 +382,15 @@ define void @TestFPTruncF128_F64() nounwind {
; X32-NEXT: fstpl vf64
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPTruncF128_F64:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __trunctfdf2
; X64-AVX-NEXT: vmovsd %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptrunc fp128 %0 to double
@ -311,18 +399,18 @@ entry:
}
define void @TestFPTruncF128_F80() nounwind {
; X64-LABEL: TestFPTruncF128_F80:
; X64: # %bb.0: # %entry
; X64-NEXT: subq $24, %rsp
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: callq __trunctfxf2
; X64-NEXT: fstpt (%rsp)
; X64-NEXT: movq (%rsp), %rax
; X64-NEXT: movq %rax, {{.*}}(%rip)
; X64-NEXT: movl {{[0-9]+}}(%rsp), %eax
; X64-NEXT: movw %ax, vf80+{{.*}}(%rip)
; X64-NEXT: addq $24, %rsp
; X64-NEXT: retq
; X64-SSE-LABEL: TestFPTruncF128_F80:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: subq $24, %rsp
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: callq __trunctfxf2
; X64-SSE-NEXT: fstpt (%rsp)
; X64-SSE-NEXT: movq (%rsp), %rax
; X64-SSE-NEXT: movq %rax, {{.*}}(%rip)
; X64-SSE-NEXT: movl {{[0-9]+}}(%rsp), %eax
; X64-SSE-NEXT: movw %ax, vf80+{{.*}}(%rip)
; X64-SSE-NEXT: addq $24, %rsp
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestFPTruncF128_F80:
; X32: # %bb.0: # %entry
@ -336,6 +424,19 @@ define void @TestFPTruncF128_F80() nounwind {
; X32-NEXT: fstpt vf80
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestFPTruncF128_F80:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: subq $24, %rsp
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: callq __trunctfxf2
; X64-AVX-NEXT: fstpt (%rsp)
; X64-AVX-NEXT: movq (%rsp), %rax
; X64-AVX-NEXT: movq %rax, {{.*}}(%rip)
; X64-AVX-NEXT: movl {{[0-9]+}}(%rsp), %eax
; X64-AVX-NEXT: movw %ax, vf80+{{.*}}(%rip)
; X64-AVX-NEXT: addq $24, %rsp
; X64-AVX-NEXT: retq
entry:
%0 = load fp128, fp128* @vf128, align 16
%conv = fptrunc fp128 %0 to x86_fp80
@ -344,14 +445,14 @@ entry:
}
define void @TestSIToFPI32_F128() nounwind {
; X64-LABEL: TestSIToFPI32_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movl {{.*}}(%rip), %edi
; X64-NEXT: callq __floatsitf
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestSIToFPI32_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movl {{.*}}(%rip), %edi
; X64-SSE-NEXT: callq __floatsitf
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestSIToFPI32_F128:
; X32: # %bb.0: # %entry
@ -373,6 +474,15 @@ define void @TestSIToFPI32_F128() nounwind {
; X32-NEXT: addl $24, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestSIToFPI32_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: movl {{.*}}(%rip), %edi
; X64-AVX-NEXT: callq __floatsitf
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load i32, i32* @vi32, align 4
%conv = sitofp i32 %0 to fp128
@ -381,14 +491,14 @@ entry:
}
define void @TestUIToFPU32_F128() #2 {
; X64-LABEL: TestUIToFPU32_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movl {{.*}}(%rip), %edi
; X64-NEXT: callq __floatunsitf
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestUIToFPU32_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movl {{.*}}(%rip), %edi
; X64-SSE-NEXT: callq __floatunsitf
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestUIToFPU32_F128:
; X32: # %bb.0: # %entry
@ -410,6 +520,15 @@ define void @TestUIToFPU32_F128() #2 {
; X32-NEXT: addl $24, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestUIToFPU32_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: movl {{.*}}(%rip), %edi
; X64-AVX-NEXT: callq __floatunsitf
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load i32, i32* @vu32, align 4
%conv = uitofp i32 %0 to fp128
@ -418,14 +537,14 @@ entry:
}
define void @TestSIToFPI64_F128() nounwind {
; X64-LABEL: TestSIToFPI64_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movq {{.*}}(%rip), %rdi
; X64-NEXT: callq __floatditf
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestSIToFPI64_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movq {{.*}}(%rip), %rdi
; X64-SSE-NEXT: callq __floatditf
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestSIToFPI64_F128:
; X32: # %bb.0: # %entry
@ -448,6 +567,15 @@ define void @TestSIToFPI64_F128() nounwind {
; X32-NEXT: addl $24, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestSIToFPI64_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: movq {{.*}}(%rip), %rdi
; X64-AVX-NEXT: callq __floatditf
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load i64, i64* @vi64, align 8
%conv = sitofp i64 %0 to fp128
@ -456,14 +584,14 @@ entry:
}
define void @TestUIToFPU64_F128() #2 {
; X64-LABEL: TestUIToFPU64_F128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movq {{.*}}(%rip), %rdi
; X64-NEXT: callq __floatunditf
; X64-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-NEXT: popq %rax
; X64-NEXT: retq
; X64-SSE-LABEL: TestUIToFPU64_F128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movq {{.*}}(%rip), %rdi
; X64-SSE-NEXT: callq __floatunditf
; X64-SSE-NEXT: movaps %xmm0, {{.*}}(%rip)
; X64-SSE-NEXT: popq %rax
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestUIToFPU64_F128:
; X32: # %bb.0: # %entry
@ -486,6 +614,15 @@ define void @TestUIToFPU64_F128() #2 {
; X32-NEXT: addl $24, %esp
; X32-NEXT: popl %esi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestUIToFPU64_F128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: movq {{.*}}(%rip), %rdi
; X64-AVX-NEXT: callq __floatunditf
; X64-AVX-NEXT: vmovaps %xmm0, {{.*}}(%rip)
; X64-AVX-NEXT: popq %rax
; X64-AVX-NEXT: retq
entry:
%0 = load i64, i64* @vu64, align 8
%conv = uitofp i64 %0 to fp128
@ -494,17 +631,17 @@ entry:
}
define i32 @TestConst128(fp128 %v) nounwind {
; X64-LABEL: TestConst128:
; X64: # %bb.0: # %entry
; X64-NEXT: pushq %rax
; X64-NEXT: movaps {{.*}}(%rip), %xmm1
; X64-NEXT: callq __gttf2
; X64-NEXT: xorl %ecx, %ecx
; X64-NEXT: testl %eax, %eax
; X64-NEXT: setg %cl
; X64-NEXT: movl %ecx, %eax
; X64-NEXT: popq %rcx
; X64-NEXT: retq
; X64-SSE-LABEL: TestConst128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm1
; X64-SSE-NEXT: callq __gttf2
; X64-SSE-NEXT: xorl %ecx, %ecx
; X64-SSE-NEXT: testl %eax, %eax
; X64-SSE-NEXT: setg %cl
; X64-SSE-NEXT: movl %ecx, %eax
; X64-SSE-NEXT: popq %rcx
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestConst128:
; X32: # %bb.0: # %entry
@ -525,6 +662,18 @@ define i32 @TestConst128(fp128 %v) nounwind {
; X32-NEXT: movl %ecx, %eax
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestConst128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm1
; X64-AVX-NEXT: callq __gttf2
; X64-AVX-NEXT: xorl %ecx, %ecx
; X64-AVX-NEXT: testl %eax, %eax
; X64-AVX-NEXT: setg %cl
; X64-AVX-NEXT: movl %ecx, %eax
; X64-AVX-NEXT: popq %rcx
; X64-AVX-NEXT: retq
entry:
%cmp = fcmp ogt fp128 %v, 0xL00000000000000003FFF000000000000
%conv = zext i1 %cmp to i32
@ -546,20 +695,20 @@ entry:
; return ((u.bits.v1 | u.bits.v2) == 0);
; }
define i32 @TestBits128(fp128 %ld) nounwind {
; X64-LABEL: TestBits128:
; X64: # %bb.0: # %entry
; X64-NEXT: subq $24, %rsp
; X64-NEXT: movaps %xmm0, %xmm1
; X64-NEXT: callq __multf3
; X64-NEXT: movaps %xmm0, (%rsp)
; X64-NEXT: movq (%rsp), %rcx
; X64-NEXT: movq %rcx, %rdx
; X64-NEXT: shrq $32, %rdx
; X64-NEXT: xorl %eax, %eax
; X64-NEXT: orl %ecx, %edx
; X64-NEXT: sete %al
; X64-NEXT: addq $24, %rsp
; X64-NEXT: retq
; X64-SSE-LABEL: TestBits128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: subq $24, %rsp
; X64-SSE-NEXT: movaps %xmm0, %xmm1
; X64-SSE-NEXT: callq __multf3
; X64-SSE-NEXT: movaps %xmm0, (%rsp)
; X64-SSE-NEXT: movq (%rsp), %rcx
; X64-SSE-NEXT: movq %rcx, %rdx
; X64-SSE-NEXT: shrq $32, %rdx
; X64-SSE-NEXT: xorl %eax, %eax
; X64-SSE-NEXT: orl %ecx, %edx
; X64-SSE-NEXT: sete %al
; X64-SSE-NEXT: addq $24, %rsp
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestBits128:
; X32: # %bb.0: # %entry
@ -591,6 +740,21 @@ define i32 @TestBits128(fp128 %ld) nounwind {
; X32-NEXT: popl %esi
; X32-NEXT: popl %edi
; X32-NEXT: retl
;
; X64-AVX-LABEL: TestBits128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: subq $24, %rsp
; X64-AVX-NEXT: vmovaps %xmm0, %xmm1
; X64-AVX-NEXT: callq __multf3
; X64-AVX-NEXT: vmovaps %xmm0, (%rsp)
; X64-AVX-NEXT: movq (%rsp), %rcx
; X64-AVX-NEXT: movq %rcx, %rdx
; X64-AVX-NEXT: shrq $32, %rdx
; X64-AVX-NEXT: xorl %eax, %eax
; X64-AVX-NEXT: orl %ecx, %edx
; X64-AVX-NEXT: sete %al
; X64-AVX-NEXT: addq $24, %rsp
; X64-AVX-NEXT: retq
entry:
%mul = fmul fp128 %ld, %ld
%0 = bitcast fp128 %mul to i128
@ -617,14 +781,14 @@ entry:
; return *(__float128*)&n;
; }
define fp128 @TestPair128(i64 %a, i64 %b) nounwind {
; X64-LABEL: TestPair128:
; X64: # %bb.0: # %entry
; X64-NEXT: addq $3, %rsi
; X64-NEXT: movq %rsi, -{{[0-9]+}}(%rsp)
; X64-NEXT: adcq $0, %rdi
; X64-NEXT: movq %rdi, -{{[0-9]+}}(%rsp)
; X64-NEXT: movaps -{{[0-9]+}}(%rsp), %xmm0
; X64-NEXT: retq
; X64-SSE-LABEL: TestPair128:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: addq $3, %rsi
; X64-SSE-NEXT: movq %rsi, -{{[0-9]+}}(%rsp)
; X64-SSE-NEXT: adcq $0, %rdi
; X64-SSE-NEXT: movq %rdi, -{{[0-9]+}}(%rsp)
; X64-SSE-NEXT: movaps -{{[0-9]+}}(%rsp), %xmm0
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestPair128:
; X32: # %bb.0: # %entry
@ -646,6 +810,15 @@ define fp128 @TestPair128(i64 %a, i64 %b) nounwind {
; X32-NEXT: popl %esi
; X32-NEXT: popl %edi
; X32-NEXT: retl $4
;
; X64-AVX-LABEL: TestPair128:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: addq $3, %rsi
; X64-AVX-NEXT: movq %rsi, -{{[0-9]+}}(%rsp)
; X64-AVX-NEXT: adcq $0, %rdi
; X64-AVX-NEXT: movq %rdi, -{{[0-9]+}}(%rsp)
; X64-AVX-NEXT: vmovaps -{{[0-9]+}}(%rsp), %xmm0
; X64-AVX-NEXT: retq
entry:
%conv = zext i64 %a to i128
%shl = shl nuw i128 %conv, 64
@ -657,20 +830,20 @@ entry:
}
define fp128 @TestTruncCopysign(fp128 %x, i32 %n) nounwind {
; X64-LABEL: TestTruncCopysign:
; X64: # %bb.0: # %entry
; X64-NEXT: cmpl $50001, %edi # imm = 0xC351
; X64-NEXT: jl .LBB17_2
; X64-NEXT: # %bb.1: # %if.then
; X64-NEXT: pushq %rax
; X64-NEXT: callq __trunctfdf2
; X64-NEXT: andps {{.*}}(%rip), %xmm0
; X64-NEXT: movsd {{.*#+}} xmm1 = mem[0],zero
; X64-NEXT: orps %xmm1, %xmm0
; X64-NEXT: callq __extenddftf2
; X64-NEXT: addq $8, %rsp
; X64-NEXT: .LBB17_2: # %cleanup
; X64-NEXT: retq
; X64-SSE-LABEL: TestTruncCopysign:
; X64-SSE: # %bb.0: # %entry
; X64-SSE-NEXT: cmpl $50001, %edi # imm = 0xC351
; X64-SSE-NEXT: jl .LBB17_2
; X64-SSE-NEXT: # %bb.1: # %if.then
; X64-SSE-NEXT: pushq %rax
; X64-SSE-NEXT: callq __trunctfdf2
; X64-SSE-NEXT: andps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: movsd {{.*#+}} xmm1 = mem[0],zero
; X64-SSE-NEXT: orps %xmm1, %xmm0
; X64-SSE-NEXT: callq __extenddftf2
; X64-SSE-NEXT: addq $8, %rsp
; X64-SSE-NEXT: .LBB17_2: # %cleanup
; X64-SSE-NEXT: retq
;
; X32-LABEL: TestTruncCopysign:
; X32: # %bb.0: # %entry
@ -721,6 +894,22 @@ define fp128 @TestTruncCopysign(fp128 %x, i32 %n) nounwind {
; X32-NEXT: popl %esi
; X32-NEXT: popl %edi
; X32-NEXT: retl $4
;
; X64-AVX-LABEL: TestTruncCopysign:
; X64-AVX: # %bb.0: # %entry
; X64-AVX-NEXT: cmpl $50001, %edi # imm = 0xC351
; X64-AVX-NEXT: jl .LBB17_2
; X64-AVX-NEXT: # %bb.1: # %if.then
; X64-AVX-NEXT: pushq %rax
; X64-AVX-NEXT: callq __trunctfdf2
; X64-AVX-NEXT: vandps {{.*}}(%rip), %xmm0, %xmm0
; X64-AVX-NEXT: vmovddup {{.*#+}} xmm1 = [+Inf,+Inf]
; X64-AVX-NEXT: # xmm1 = mem[0,0]
; X64-AVX-NEXT: vorps %xmm0, %xmm1, %xmm0
; X64-AVX-NEXT: callq __extenddftf2
; X64-AVX-NEXT: addq $8, %rsp
; X64-AVX-NEXT: .LBB17_2: # %cleanup
; X64-AVX-NEXT: retq
entry:
%cmp = icmp sgt i32 %n, 50000
br i1 %cmp, label %if.then, label %cleanup
@ -737,15 +926,15 @@ cleanup: ; preds = %entry, %if.then
}
define i1 @PR34866(i128 %x) nounwind {
; X64-LABEL: PR34866:
; X64: # %bb.0:
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: movaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-NEXT: orq %rsi, %rdi
; X64-NEXT: sete %al
; X64-NEXT: retq
; X64-SSE-LABEL: PR34866:
; X64-SSE: # %bb.0:
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: movaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-SSE-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-SSE-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-SSE-NEXT: orq %rsi, %rdi
; X64-SSE-NEXT: sete %al
; X64-SSE-NEXT: retq
;
; X32-LABEL: PR34866:
; X32: # %bb.0:
@ -756,21 +945,31 @@ define i1 @PR34866(i128 %x) nounwind {
; X32-NEXT: orl %ecx, %eax
; X32-NEXT: sete %al
; X32-NEXT: retl
;
; X64-AVX-LABEL: PR34866:
; X64-AVX: # %bb.0:
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-AVX-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-AVX-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-AVX-NEXT: orq %rsi, %rdi
; X64-AVX-NEXT: sete %al
; X64-AVX-NEXT: retq
%bc_mmx = bitcast fp128 0xL00000000000000000000000000000000 to i128
%cmp = icmp eq i128 %bc_mmx, %x
ret i1 %cmp
}
define i1 @PR34866_commute(i128 %x) nounwind {
; X64-LABEL: PR34866_commute:
; X64: # %bb.0:
; X64-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-NEXT: movaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-NEXT: orq %rsi, %rdi
; X64-NEXT: sete %al
; X64-NEXT: retq
; X64-SSE-LABEL: PR34866_commute:
; X64-SSE: # %bb.0:
; X64-SSE-NEXT: movaps {{.*}}(%rip), %xmm0
; X64-SSE-NEXT: movaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-SSE-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-SSE-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-SSE-NEXT: orq %rsi, %rdi
; X64-SSE-NEXT: sete %al
; X64-SSE-NEXT: retq
;
; X32-LABEL: PR34866_commute:
; X32: # %bb.0:
@ -781,6 +980,16 @@ define i1 @PR34866_commute(i128 %x) nounwind {
; X32-NEXT: orl %ecx, %eax
; X32-NEXT: sete %al
; X32-NEXT: retl
;
; X64-AVX-LABEL: PR34866_commute:
; X64-AVX: # %bb.0:
; X64-AVX-NEXT: vmovaps {{.*}}(%rip), %xmm0
; X64-AVX-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp)
; X64-AVX-NEXT: xorq -{{[0-9]+}}(%rsp), %rsi
; X64-AVX-NEXT: xorq -{{[0-9]+}}(%rsp), %rdi
; X64-AVX-NEXT: orq %rsi, %rdi
; X64-AVX-NEXT: sete %al
; X64-AVX-NEXT: retq
%bc_mmx = bitcast fp128 0xL00000000000000000000000000000000 to i128
%cmp = icmp eq i128 %x, %bc_mmx
ret i1 %cmp