[InstCombine] Add tests to show missing fold opportunity for "icmp and shift" (nfc).
Summary:
For icmp pred (and (sh X, Y), C), 0
When C is signbit, expect to fold (X << Y) & signbit ==/!= 0 into (X << Y) >=/< 0,
rather than (X & (signbit >> Y)) != 0.
When C+1 is power of 2, expect to fold (X << Y) & ~C ==/!= 0 into (X << Y) </>= C+1,
rather than (X & (~C >> Y)) == 0.
For icmp pred (and X, (sh signbit, Y)), 0
Expect to fold (X & (signbit l>> Y)) ==/!= 0 into (X << Y) >=/< 0
Expect to fold (X & (signbit << Y)) ==/!= 0 into (X l>> Y) >=/< 0
Reviewers: lebedev.ri, efriedma, spatel, craig.topper
Reviewed By: lebedev.ri
Subscribers: llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D63025
llvm-svn: 363479
2019-06-15 08:33:41 +08:00
|
|
|
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
|
|
|
; RUN: opt %s -instcombine -S | FileCheck %s
|
|
|
|
|
|
|
|
; For pattern (X & (signbit l>> Y)) ==/!= 0
|
|
|
|
; it may be optimal to fold into (X << Y) >=/< 0
|
|
|
|
|
|
|
|
; Scalar tests
|
|
|
|
|
|
|
|
define i1 @scalar_i8_signbit_lshr_and_eq(i8 %x, i8 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i8_signbit_lshr_and_eq(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i8 -128, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i8 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i8 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i8 128, %y
|
|
|
|
%and = and i8 %lshr, %x
|
|
|
|
%r = icmp eq i8 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @scalar_i16_signbit_lshr_and_eq(i16 %x, i16 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i16_signbit_lshr_and_eq(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i16 -32768, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i16 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i16 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i16 32768, %y
|
|
|
|
%and = and i16 %lshr, %x
|
|
|
|
%r = icmp eq i16 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq(i32 %x, i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @scalar_i64_signbit_lshr_and_eq(i64 %x, i64 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i64_signbit_lshr_and_eq(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 -9223372036854775808, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i64 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i64 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i64 9223372036854775808, %y
|
|
|
|
%and = and i64 %lshr, %x
|
|
|
|
%r = icmp eq i64 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_ne(i32 %x, i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_ne(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp ne i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%r = icmp ne i32 %and, 0 ; check 'ne' predicate
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; Vector tests
|
|
|
|
|
|
|
|
define <4 x i1> @vec_4xi32_signbit_lshr_and_eq(<4 x i32> %x, <4 x i32> %y) {
|
|
|
|
; CHECK-LABEL: @vec_4xi32_signbit_lshr_and_eq(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr <4 x i32> <i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 -2147483648>, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and <4 x i32> [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq <4 x i32> [[AND]], zeroinitializer
|
|
|
|
; CHECK-NEXT: ret <4 x i1> [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr <4 x i32> <i32 2147483648, i32 2147483648, i32 2147483648, i32 2147483648>, %y
|
|
|
|
%and = and <4 x i32> %lshr, %x
|
|
|
|
%r = icmp eq <4 x i32> %and, <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i1> %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i1> @vec_4xi32_signbit_lshr_and_eq_undef1(<4 x i32> %x, <4 x i32> %y) {
|
|
|
|
; CHECK-LABEL: @vec_4xi32_signbit_lshr_and_eq_undef1(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr <4 x i32> <i32 -2147483648, i32 undef, i32 -2147483648, i32 2147473648>, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and <4 x i32> [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq <4 x i32> [[AND]], zeroinitializer
|
|
|
|
; CHECK-NEXT: ret <4 x i1> [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr <4 x i32> <i32 2147483648, i32 undef, i32 2147483648, i32 2147473648>, %y
|
|
|
|
%and = and <4 x i32> %lshr, %x
|
|
|
|
%r = icmp eq <4 x i32> %and, <i32 0, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i1> %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i1> @vec_4xi32_signbit_lshr_and_eq_undef2(<4 x i32> %x, <4 x i32> %y) {
|
|
|
|
; CHECK-LABEL: @vec_4xi32_signbit_lshr_and_eq_undef2(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr <4 x i32> <i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 2147473648>, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and <4 x i32> [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq <4 x i32> [[AND]], <i32 0, i32 0, i32 0, i32 undef>
|
|
|
|
; CHECK-NEXT: ret <4 x i1> [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr <4 x i32> <i32 2147483648, i32 2147483648, i32 2147483648, i32 2147473648>, %y
|
|
|
|
%and = and <4 x i32> %lshr, %x
|
|
|
|
%r = icmp eq <4 x i32> %and, <i32 0, i32 0, i32 0, i32 undef>
|
|
|
|
ret <4 x i1> %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i1> @vec_4xi32_signbit_lshr_and_eq_undef3(<4 x i32> %x, <4 x i32> %y) {
|
|
|
|
; CHECK-LABEL: @vec_4xi32_signbit_lshr_and_eq_undef3(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr <4 x i32> <i32 -2147483648, i32 undef, i32 -2147483648, i32 2147473648>, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and <4 x i32> [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq <4 x i32> [[AND]], <i32 undef, i32 0, i32 0, i32 0>
|
|
|
|
; CHECK-NEXT: ret <4 x i1> [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr <4 x i32> <i32 2147483648, i32 undef, i32 2147483648, i32 2147473648>, %y
|
|
|
|
%and = and <4 x i32> %lshr, %x
|
|
|
|
%r = icmp eq <4 x i32> %and, <i32 undef, i32 0, i32 0, i32 0>
|
|
|
|
ret <4 x i1> %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; Extra use
|
|
|
|
|
|
|
|
; Fold happened
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_extra_use_lshr(i32 %x, i32 %y, i32 %z, i32* %p) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_extra_use_lshr(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[XOR:%.*]] = xor i32 [[LSHR]], [[Z:%.*]]
|
|
|
|
; CHECK-NEXT: store i32 [[XOR]], i32* [[P:%.*]], align 4
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%xor = xor i32 %lshr, %z ; extra use of lshr
|
|
|
|
store i32 %xor, i32* %p
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; Not fold
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_extra_use_and(i32 %x, i32 %y, i32 %z, i32* %p) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_extra_use_and(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[MUL:%.*]] = mul i32 [[AND]], [[Z:%.*]]
|
|
|
|
; CHECK-NEXT: store i32 [[MUL]], i32* [[P:%.*]], align 4
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%mul = mul i32 %and, %z ; extra use of and
|
|
|
|
store i32 %mul, i32* %p
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; Not fold
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_extra_use_lshr_and(i32 %x, i32 %y, i32 %z, i32* %p, i32* %q) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_extra_use_lshr_and(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: store i32 [[AND]], i32* [[P:%.*]], align 4
|
|
|
|
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[LSHR]], [[Z:%.*]]
|
|
|
|
; CHECK-NEXT: store i32 [[ADD]], i32* [[Q:%.*]], align 4
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
store i32 %and, i32* %p ; extra use of and
|
|
|
|
%add = add i32 %lshr, %z ; extra use of lshr
|
|
|
|
store i32 %add, i32* %q
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; X is constant
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_X_is_constant1(i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_X_is_constant1(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], 12345
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, 12345
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_X_is_constant2(i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_X_is_constant2(
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp ne i32 [[Y:%.*]], 31
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, 1
|
|
|
|
%r = icmp eq i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
2019-06-27 13:46:06 +08:00
|
|
|
; Negative tests
|
|
|
|
|
[InstCombine] Add tests to show missing fold opportunity for "icmp and shift" (nfc).
Summary:
For icmp pred (and (sh X, Y), C), 0
When C is signbit, expect to fold (X << Y) & signbit ==/!= 0 into (X << Y) >=/< 0,
rather than (X & (signbit >> Y)) != 0.
When C+1 is power of 2, expect to fold (X << Y) & ~C ==/!= 0 into (X << Y) </>= C+1,
rather than (X & (~C >> Y)) == 0.
For icmp pred (and X, (sh signbit, Y)), 0
Expect to fold (X & (signbit l>> Y)) ==/!= 0 into (X << Y) >=/< 0
Expect to fold (X & (signbit << Y)) ==/!= 0 into (X l>> Y) >=/< 0
Reviewers: lebedev.ri, efriedma, spatel, craig.topper
Reviewed By: lebedev.ri
Subscribers: llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D63025
llvm-svn: 363479
2019-06-15 08:33:41 +08:00
|
|
|
; Check 'slt' predicate
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_slt(i32 %x, i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_slt(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp slt i32 [[AND]], 0
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%r = icmp slt i32 %and, 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|
|
|
|
|
|
|
|
; Compare with nonzero
|
|
|
|
|
|
|
|
define i1 @scalar_i32_signbit_lshr_and_eq_nonzero(i32 %x, i32 %y) {
|
|
|
|
; CHECK-LABEL: @scalar_i32_signbit_lshr_and_eq_nonzero(
|
|
|
|
; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 -2147483648, [[Y:%.*]]
|
|
|
|
; CHECK-NEXT: [[AND:%.*]] = and i32 [[LSHR]], [[X:%.*]]
|
|
|
|
; CHECK-NEXT: [[R:%.*]] = icmp eq i32 [[AND]], 1
|
|
|
|
; CHECK-NEXT: ret i1 [[R]]
|
|
|
|
;
|
|
|
|
%lshr = lshr i32 2147483648, %y
|
|
|
|
%and = and i32 %lshr, %x
|
|
|
|
%r = icmp eq i32 %and, 1 ; should be comparing with 0
|
|
|
|
ret i1 %r
|
|
|
|
}
|