X86: Constant fold converting vector setcc results to float.
Since the result of a SETCC for X86 is 0 or -1 in each lane, we can
move unary operations, in this case [su]int_to_fp through the mask
operation and constant fold the operation away. Generally speaking:
UNARYOP(AND(VECTOR_CMP(x,y), constant))
--> AND(VECTOR_CMP(x,y), constant2)
where constant2 is UNARYOP(constant).
This implements the transform where UNARYOP is [su]int_to_fp.
For example, consider the simple function:
define <4 x float> @foo(<4 x float> %val, <4 x float> %test) nounwind {
%cmp = fcmp oeq <4 x float> %val, %test
%ext = zext <4 x i1> %cmp to <4 x i32>
%result = sitofp <4 x i32> %ext to <4 x float>
ret <4 x float> %result
}
Before this change, the SSE code is generated as:
LCPI0_0:
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
cvtdq2ps %xmm0, %xmm0
retq
After, the code is improved to:
LCPI0_0:
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
retq
The cvtdq2ps has been constant folded away and the floating point 1.0f
vector lanes are materialized directly via the ModRM operand of andps.
llvm-svn: 213342
2014-07-18 08:40:56 +08:00
|
|
|
; RUN: llc < %s -mtriple=x86_64-apple-darwin | FileCheck %s
|
|
|
|
|
|
|
|
define <4 x float> @foo(<4 x float> %val, <4 x float> %test) nounwind {
|
2014-07-24 04:41:31 +08:00
|
|
|
; CHECK-LABEL: LCPI0_0:
|
2014-07-25 06:15:28 +08:00
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
X86: Constant fold converting vector setcc results to float.
Since the result of a SETCC for X86 is 0 or -1 in each lane, we can
move unary operations, in this case [su]int_to_fp through the mask
operation and constant fold the operation away. Generally speaking:
UNARYOP(AND(VECTOR_CMP(x,y), constant))
--> AND(VECTOR_CMP(x,y), constant2)
where constant2 is UNARYOP(constant).
This implements the transform where UNARYOP is [su]int_to_fp.
For example, consider the simple function:
define <4 x float> @foo(<4 x float> %val, <4 x float> %test) nounwind {
%cmp = fcmp oeq <4 x float> %val, %test
%ext = zext <4 x i1> %cmp to <4 x i32>
%result = sitofp <4 x i32> %ext to <4 x float>
ret <4 x float> %result
}
Before this change, the SSE code is generated as:
LCPI0_0:
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
cvtdq2ps %xmm0, %xmm0
retq
After, the code is improved to:
LCPI0_0:
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
retq
The cvtdq2ps has been constant folded away and the floating point 1.0f
vector lanes are materialized directly via the ModRM operand of andps.
llvm-svn: 213342
2014-07-18 08:40:56 +08:00
|
|
|
; CHECK-LABEL: foo:
|
2017-12-05 01:18:51 +08:00
|
|
|
; CHECK: ## %bb.0:
|
2016-09-05 01:16:01 +08:00
|
|
|
; CHECK-NEXT: cmpeqps %xmm1, %xmm0
|
|
|
|
; CHECK-NEXT: andps {{.*}}(%rip), %xmm0
|
|
|
|
; CHECK-NEXT: retq
|
X86: Constant fold converting vector setcc results to float.
Since the result of a SETCC for X86 is 0 or -1 in each lane, we can
move unary operations, in this case [su]int_to_fp through the mask
operation and constant fold the operation away. Generally speaking:
UNARYOP(AND(VECTOR_CMP(x,y), constant))
--> AND(VECTOR_CMP(x,y), constant2)
where constant2 is UNARYOP(constant).
This implements the transform where UNARYOP is [su]int_to_fp.
For example, consider the simple function:
define <4 x float> @foo(<4 x float> %val, <4 x float> %test) nounwind {
%cmp = fcmp oeq <4 x float> %val, %test
%ext = zext <4 x i1> %cmp to <4 x i32>
%result = sitofp <4 x i32> %ext to <4 x float>
ret <4 x float> %result
}
Before this change, the SSE code is generated as:
LCPI0_0:
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.long 1 ## 0x1
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
cvtdq2ps %xmm0, %xmm0
retq
After, the code is improved to:
LCPI0_0:
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.long 1065353216 ## float 1.000000e+00
.section __TEXT,__text,regular,pure_instructions
.globl _foo
.align 4, 0x90
_foo: ## @foo
cmpeqps %xmm1, %xmm0
andps LCPI0_0(%rip), %xmm0
retq
The cvtdq2ps has been constant folded away and the floating point 1.0f
vector lanes are materialized directly via the ModRM operand of andps.
llvm-svn: 213342
2014-07-18 08:40:56 +08:00
|
|
|
%cmp = fcmp oeq <4 x float> %val, %test
|
|
|
|
%ext = zext <4 x i1> %cmp to <4 x i32>
|
|
|
|
%result = sitofp <4 x i32> %ext to <4 x float>
|
|
|
|
ret <4 x float> %result
|
|
|
|
}
|
2014-07-24 04:41:31 +08:00
|
|
|
|
2014-07-24 04:41:38 +08:00
|
|
|
; Make sure the operation doesn't try to get folded when the sizes don't match,
|
|
|
|
; as that ends up crashing later when trying to form a bitcast operation for
|
|
|
|
; the folded nodes.
|
|
|
|
define void @foo1(<4 x float> %val, <4 x float> %test, <4 x double>* %p) nounwind {
|
2014-07-24 04:41:31 +08:00
|
|
|
; CHECK-LABEL: LCPI1_0:
|
2014-07-24 04:41:38 +08:00
|
|
|
; CHECK-NEXT: .long 1 ## 0x1
|
|
|
|
; CHECK-NEXT: .long 1 ## 0x1
|
|
|
|
; CHECK-NEXT: .long 1 ## 0x1
|
|
|
|
; CHECK-NEXT: .long 1 ## 0x1
|
|
|
|
; CHECK-LABEL: foo1:
|
2017-12-05 01:18:51 +08:00
|
|
|
; CHECK: ## %bb.0:
|
2016-09-05 01:16:01 +08:00
|
|
|
; CHECK-NEXT: cmpeqps %xmm1, %xmm0
|
|
|
|
; CHECK-NEXT: andps {{.*}}(%rip), %xmm0
|
|
|
|
; CHECK-NEXT: pshufd {{.*#+}} xmm1 = xmm0[2,3,0,1]
|
|
|
|
; CHECK-NEXT: cvtdq2pd %xmm1, %xmm1
|
|
|
|
; CHECK-NEXT: cvtdq2pd %xmm0, %xmm0
|
|
|
|
; CHECK-NEXT: movaps %xmm0, (%rdi)
|
|
|
|
; CHECK-NEXT: movaps %xmm1, 16(%rdi)
|
|
|
|
; CHECK-NEXT: retq
|
2014-07-24 04:41:38 +08:00
|
|
|
%cmp = fcmp oeq <4 x float> %val, %test
|
|
|
|
%ext = zext <4 x i1> %cmp to <4 x i32>
|
|
|
|
%result = sitofp <4 x i32> %ext to <4 x double>
|
|
|
|
store <4 x double> %result, <4 x double>* %p
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; Also test the general purpose constant folding of int->fp.
|
|
|
|
define void @foo2(<4 x float>* noalias %result) nounwind {
|
|
|
|
; CHECK-LABEL: LCPI2_0:
|
2015-12-08 10:37:48 +08:00
|
|
|
; CHECK-NEXT: .long 1082130432 ## float 4
|
|
|
|
; CHECK-NEXT: .long 1084227584 ## float 5
|
|
|
|
; CHECK-NEXT: .long 1086324736 ## float 6
|
|
|
|
; CHECK-NEXT: .long 1088421888 ## float 7
|
2014-07-24 04:41:38 +08:00
|
|
|
; CHECK-LABEL: foo2:
|
2017-12-05 01:18:51 +08:00
|
|
|
; CHECK: ## %bb.0:
|
2016-09-05 01:16:01 +08:00
|
|
|
; CHECK-NEXT: movaps {{.*#+}} xmm0 = [4.000000e+00,5.000000e+00,6.000000e+00,7.000000e+00]
|
|
|
|
; CHECK-NEXT: movaps %xmm0, (%rdi)
|
|
|
|
; CHECK-NEXT: retq
|
2014-07-24 04:41:31 +08:00
|
|
|
%val = uitofp <4 x i32> <i32 4, i32 5, i32 6, i32 7> to <4 x float>
|
|
|
|
store <4 x float> %val, <4 x float>* %result
|
|
|
|
ret void
|
|
|
|
}
|
2014-07-24 04:41:43 +08:00
|
|
|
|
|
|
|
; Fold explicit AND operations when the constant isn't a splat of a single
|
|
|
|
; scalar value like what the zext creates.
|
|
|
|
define <4 x float> @foo3(<4 x float> %val, <4 x float> %test) nounwind {
|
|
|
|
; CHECK-LABEL: LCPI3_0:
|
2014-07-25 06:15:28 +08:00
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
|
|
|
; CHECK-NEXT: .long 0 ## 0x0
|
|
|
|
; CHECK-NEXT: .long 1065353216 ## 0x3f800000
|
|
|
|
; CHECK-NEXT: .long 0 ## 0x0
|
2014-07-24 04:41:43 +08:00
|
|
|
; CHECK-LABEL: foo3:
|
2017-12-05 01:18:51 +08:00
|
|
|
; CHECK: ## %bb.0:
|
2016-09-05 01:16:01 +08:00
|
|
|
; CHECK-NEXT: cmpeqps %xmm1, %xmm0
|
|
|
|
; CHECK-NEXT: andps {{.*}}(%rip), %xmm0
|
|
|
|
; CHECK-NEXT: retq
|
2014-07-24 04:41:43 +08:00
|
|
|
%cmp = fcmp oeq <4 x float> %val, %test
|
|
|
|
%ext = zext <4 x i1> %cmp to <4 x i32>
|
|
|
|
%and = and <4 x i32> %ext, <i32 255, i32 256, i32 257, i32 258>
|
|
|
|
%result = sitofp <4 x i32> %and to <4 x float>
|
|
|
|
ret <4 x float> %result
|
|
|
|
}
|
2015-03-24 06:44:55 +08:00
|
|
|
|
|
|
|
; Test the general purpose constant folding of uint->fp.
|
|
|
|
define void @foo4(<4 x float>* noalias %result) nounwind {
|
|
|
|
; CHECK-LABEL: LCPI4_0:
|
2015-12-08 10:37:48 +08:00
|
|
|
; CHECK-NEXT: .long 1065353216 ## float 1
|
|
|
|
; CHECK-NEXT: .long 1123942400 ## float 127
|
|
|
|
; CHECK-NEXT: .long 1124073472 ## float 128
|
|
|
|
; CHECK-NEXT: .long 1132396544 ## float 255
|
2015-03-24 06:44:55 +08:00
|
|
|
; CHECK-LABEL: foo4:
|
2017-12-05 01:18:51 +08:00
|
|
|
; CHECK: ## %bb.0:
|
2016-09-05 01:16:01 +08:00
|
|
|
; CHECK-NEXT: movaps {{.*#+}} xmm0 = [1.000000e+00,1.270000e+02,1.280000e+02,2.550000e+02]
|
|
|
|
; CHECK-NEXT: movaps %xmm0, (%rdi)
|
|
|
|
; CHECK-NEXT: retq
|
2015-03-24 06:44:55 +08:00
|
|
|
%val = uitofp <4 x i8> <i8 1, i8 127, i8 -128, i8 -1> to <4 x float>
|
|
|
|
store <4 x float> %val, <4 x float>* %result
|
|
|
|
ret void
|
|
|
|
}
|