diff --git a/llvm/test/CodeGen/X86/sse-fcopysign.ll b/llvm/test/CodeGen/X86/sse-fcopysign.ll index 8a5462bea82d..b01bbac77a47 100644 --- a/llvm/test/CodeGen/X86/sse-fcopysign.ll +++ b/llvm/test/CodeGen/X86/sse-fcopysign.ll @@ -1,3 +1,4 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse2 | FileCheck %s --check-prefix=ALL --check-prefix=X32 ; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse2 | FileCheck %s --check-prefix=ALL --check-prefix=X64 @@ -5,41 +6,47 @@ ; Library Functions ; -define float @tst1(float %a, float %b) { -; X32-LABEL: @tst1 -; X32: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero -; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero -; X32-NEXT: movss %xmm1, 4(%esp) -; X32-NEXT: movss %xmm0, (%esp) -; X32-NEXT: calll copysignf -; X32-NEXT: addl $8, %esp -; X32-NEXT: retl +define float @tst1(float %a, float %b) nounwind { +; X32-LABEL: tst1: +; X32: # BB#0: +; X32-NEXT: subl $8, %esp +; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero +; X32-NEXT: movss %xmm1, {{[0-9]+}}(%esp) +; X32-NEXT: movss %xmm0, (%esp) +; X32-NEXT: calll copysignf +; X32-NEXT: addl $8, %esp +; X32-NEXT: retl ; -; X64-LABEL: @tst1 -; X64: movaps %xmm0, %xmm2 -; X64-NEXT: movaps %xmm1, %xmm0 -; X64-NEXT: movaps %xmm2, %xmm1 -; X64-NEXT: jmp copysignf +; X64-LABEL: tst1: +; X64: # BB#0: +; X64-NEXT: movaps %xmm0, %xmm2 +; X64-NEXT: movaps %xmm1, %xmm0 +; X64-NEXT: movaps %xmm2, %xmm1 +; X64-NEXT: jmp copysignf # TAILCALL %tmp = tail call float @copysignf( float %b, float %a ) ret float %tmp } -define double @tst2(double %a, float %b, float %c) { -; X32-LABEL: @tst2 -; X32: movsd {{.*#+}} xmm0 = mem[0],zero -; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero -; X32-NEXT: addss 32(%esp), %xmm1 -; X32-NEXT: cvtss2sd %xmm1, %xmm1 -; X32-NEXT: movsd %xmm0, (%esp) -; X32-NEXT: movsd %xmm1, 8(%esp) -; X32-NEXT: calll copysign -; X32-NEXT: addl $16, %esp -; X32-NEXT: retl +define double @tst2(double %a, float %b, float %c) nounwind { +; X32-LABEL: tst2: +; X32: # BB#0: +; X32-NEXT: subl $16, %esp +; X32-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero +; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero +; X32-NEXT: addss {{[0-9]+}}(%esp), %xmm1 +; X32-NEXT: cvtss2sd %xmm1, %xmm1 +; X32-NEXT: movsd %xmm0, (%esp) +; X32-NEXT: movsd %xmm1, {{[0-9]+}}(%esp) +; X32-NEXT: calll copysign +; X32-NEXT: addl $16, %esp +; X32-NEXT: retl ; -; X64-LABEL: @tst2 -; X64: addss %xmm2, %xmm1 -; X64-NEXT: cvtss2sd %xmm1, %xmm1 -; X64-NEXT: jmp copysign +; X64-LABEL: tst2: +; X64: # BB#0: +; X64-NEXT: addss %xmm2, %xmm1 +; X64-NEXT: cvtss2sd %xmm1, %xmm1 +; X64-NEXT: jmp copysign # TAILCALL %tmp1 = fadd float %b, %c %tmp2 = fpext float %tmp1 to double %tmp = tail call double @copysign( double %a, double %tmp2 ) @@ -53,77 +60,90 @@ declare double @copysign(double, double) ; LLVM Intrinsic ; -define float @int1(float %a, float %b) { -; X32-LABEL: @int1 -; X32: movss 8(%esp), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero -; X32-NEXT: andps .LCPI2_0, %xmm0 -; X32-NEXT: movss 12(%esp), %xmm1 {{.*#+}} xmm1 = mem[0],zero,zero,zero -; X32-NEXT: andps .LCPI2_1, %xmm1 -; X32-NEXT: orps %xmm0, %xmm1 -; X32-NEXT: movss %xmm1, (%esp) -; X32-NEXT: flds (%esp) -; X32-NEXT: popl %eax -; X32-NEXT: retl +define float @int1(float %a, float %b) nounwind { +; X32-LABEL: int1: +; X32: # BB#0: +; X32-NEXT: pushl %eax +; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; X32-NEXT: andps {{\.LCPI.*}}, %xmm0 +; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero +; X32-NEXT: andps {{\.LCPI.*}}, %xmm1 +; X32-NEXT: orps %xmm0, %xmm1 +; X32-NEXT: movss %xmm1, (%esp) +; X32-NEXT: flds (%esp) +; X32-NEXT: popl %eax +; X32-NEXT: retl ; -; X64-LABEL: @int1 -; X64: andps .LCPI2_0(%rip), %xmm0 -; X64-NEXT: andps .LCPI2_1(%rip), %xmm1 -; X64-NEXT: orps %xmm1, %xmm0 -; X64-NEXT: retq +; X64-LABEL: int1: +; X64: # BB#0: +; X64-NEXT: andps {{.*}}(%rip), %xmm0 +; X64-NEXT: andps {{.*}}(%rip), %xmm1 +; X64-NEXT: orps %xmm1, %xmm0 +; X64-NEXT: retq %tmp = tail call float @llvm.copysign.f32( float %b, float %a ) ret float %tmp } -define double @int2(double %a, float %b, float %c) { -; X32-LABEL: @int2 -; X32: movss 16(%ebp), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero -; X32-NEXT: addss 20(%ebp), %xmm0 -; X32-NEXT: movsd 8(%ebp), %xmm1 {{.*#+}} xmm1 = mem[0],zero -; X32-NEXT: andpd .LCPI3_0, %xmm1 -; X32-NEXT: cvtss2sd %xmm0, %xmm0 -; X32-NEXT: andpd .LCPI3_1, %xmm0 -; X32-NEXT: orpd %xmm1, %xmm0 -; X32-NEXT: movlpd %xmm0, (%esp) -; X32-NEXT: fldl (%esp) -; X32-NEXT: movl %ebp, %esp -; X32-NEXT: popl %ebp -; X32-NEXT: retl +define double @int2(double %a, float %b, float %c) nounwind { +; X32-LABEL: int2: +; X32: # BB#0: +; X32-NEXT: pushl %ebp +; X32-NEXT: movl %esp, %ebp +; X32-NEXT: andl $-8, %esp +; X32-NEXT: subl $8, %esp +; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; X32-NEXT: addss 20(%ebp), %xmm0 +; X32-NEXT: movsd {{.*#+}} xmm1 = mem[0],zero +; X32-NEXT: andpd {{\.LCPI.*}}, %xmm1 +; X32-NEXT: cvtss2sd %xmm0, %xmm0 +; X32-NEXT: andpd {{\.LCPI.*}}, %xmm0 +; X32-NEXT: orpd %xmm1, %xmm0 +; X32-NEXT: movlpd %xmm0, (%esp) +; X32-NEXT: fldl (%esp) +; X32-NEXT: movl %ebp, %esp +; X32-NEXT: popl %ebp +; X32-NEXT: retl ; -; X64-LABEL: @int2 -; X64: addss %xmm2, %xmm1 -; X64-NEXT: cvtss2sd %xmm1, %xmm1 -; X64-NEXT: andpd .LCPI3_0(%rip), %xmm1 -; X64-NEXT: andpd .LCPI3_1(%rip), %xmm0 -; X64-NEXT: orpd %xmm1, %xmm0 -; X64-NEXT: retq +; X64-LABEL: int2: +; X64: # BB#0: +; X64-NEXT: addss %xmm2, %xmm1 +; X64-NEXT: cvtss2sd %xmm1, %xmm1 +; X64-NEXT: andpd {{.*}}(%rip), %xmm1 +; X64-NEXT: andpd {{.*}}(%rip), %xmm0 +; X64-NEXT: orpd %xmm1, %xmm0 +; X64-NEXT: retq %tmp1 = fadd float %b, %c %tmp2 = fpext float %tmp1 to double %tmp = tail call double @llvm.copysign.f64( double %a, double %tmp2 ) ret double %tmp } -define float @cst1() { -; X32-LABEL: @cst1 -; X32: fld1 -; X32-NEXT: fchs -; X32-NEXT: retl +define float @cst1() nounwind { +; X32-LABEL: cst1: +; X32: # BB#0: +; X32-NEXT: fld1 +; X32-NEXT: fchs +; X32-NEXT: retl ; -; X64-LABEL: @cst1 -; X64: movss .LCPI4_0(%rip), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero -; X64-NEXT: retq +; X64-LABEL: cst1: +; X64: # BB#0: +; X64-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; X64-NEXT: retq %tmp = tail call float @llvm.copysign.f32( float 1.0, float -2.0 ) ret float %tmp } -define double @cst2() { -; X32-LABEL: @cst2 -; X32: fldz -; X32-NEXT: fchs -; X32-NEXT: retl +define double @cst2() nounwind { +; X32-LABEL: cst2: +; X32: # BB#0: +; X32-NEXT: fldz +; X32-NEXT: fchs +; X32-NEXT: retl ; -; X64-LABEL: @cst2 -; X64: movsd .LCPI5_0(%rip), %xmm0 {{.*#+}} xmm0 = mem[0],zero -; X64-NEXT: retq +; X64-LABEL: cst2: +; X64: # BB#0: +; X64-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero +; X64-NEXT: retq %tmp1 = fadd float -1.0, -1.0 %tmp2 = fpext float %tmp1 to double %tmp = tail call double @llvm.copysign.f64( double 0.0, double %tmp2 )