diff --git a/llvm/test/Transforms/InstSimplify/result-of-usub-by-nonzero-is-non-zero-and-no-overflow.ll b/llvm/test/Transforms/InstSimplify/result-of-usub-by-nonzero-is-non-zero-and-no-overflow.ll index 5a64daf11a47..38538ff3736b 100644 --- a/llvm/test/Transforms/InstSimplify/result-of-usub-by-nonzero-is-non-zero-and-no-overflow.ll +++ b/llvm/test/Transforms/InstSimplify/result-of-usub-by-nonzero-is-non-zero-and-no-overflow.ll @@ -5,80 +5,90 @@ ; that the result is non-zero. This can be simplified just to a comparison ; between the base and offset. -declare void @llvm.assume(i1) - -define i1 @t0(i8 %base, i8 %offset) { +define i1 @t0(i64 %base, i64* nonnull %offsetptr) { ; CHECK-LABEL: @t0( -; CHECK-NEXT: [[CMP:%.*]] = icmp ne i8 [[OFFSET:%.*]], 0 -; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]]) -; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i8 [[BASE:%.*]], [[OFFSET]] -; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp uge i8 [[ADJUSTED]], [[BASE]] -; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i8 [[ADJUSTED]], 0 +; CHECK-NEXT: [[OFFSET:%.*]] = ptrtoint i64* [[OFFSETPTR:%.*]] to i64 +; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i64 [[BASE:%.*]], [[OFFSET]] +; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp uge i64 [[ADJUSTED]], [[BASE]] +; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i64 [[ADJUSTED]], 0 ; CHECK-NEXT: [[R:%.*]] = and i1 [[NOT_NULL]], [[NO_UNDERFLOW]] ; CHECK-NEXT: ret i1 [[R]] ; - %cmp = icmp ne i8 %offset, 0 - call void @llvm.assume(i1 %cmp) + %offset = ptrtoint i64* %offsetptr to i64 - %adjusted = sub i8 %base, %offset - %no_underflow = icmp uge i8 %adjusted, %base - %not_null = icmp ne i8 %adjusted, 0 + %adjusted = sub i64 %base, %offset + %no_underflow = icmp uge i64 %adjusted, %base + %not_null = icmp ne i64 %adjusted, 0 %r = and i1 %not_null, %no_underflow ret i1 %r } -define i1 @t1(i8 %base, i8 %offset) { +define i1 @t1(i64 %base, i64* nonnull %offsetptr) { ; CHECK-LABEL: @t1( -; CHECK-NEXT: [[CMP:%.*]] = icmp ne i8 [[OFFSET:%.*]], 0 -; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]]) -; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i8 [[BASE:%.*]], [[OFFSET]] -; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp ult i8 [[ADJUSTED]], [[BASE]] -; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp eq i8 [[ADJUSTED]], 0 +; CHECK-NEXT: [[OFFSET:%.*]] = ptrtoint i64* [[OFFSETPTR:%.*]] to i64 +; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i64 [[BASE:%.*]], [[OFFSET]] +; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp ult i64 [[ADJUSTED]], [[BASE]] +; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp eq i64 [[ADJUSTED]], 0 ; CHECK-NEXT: [[R:%.*]] = or i1 [[NOT_NULL]], [[NO_UNDERFLOW]] ; CHECK-NEXT: ret i1 [[R]] ; - %cmp = icmp ne i8 %offset, 0 - call void @llvm.assume(i1 %cmp) + %offset = ptrtoint i64* %offsetptr to i64 - %adjusted = sub i8 %base, %offset - %no_underflow = icmp ult i8 %adjusted, %base - %not_null = icmp eq i8 %adjusted, 0 + %adjusted = sub i64 %base, %offset + %no_underflow = icmp ult i64 %adjusted, %base + %not_null = icmp eq i64 %adjusted, 0 %r = or i1 %not_null, %no_underflow ret i1 %r } -define i1 @t2_commutative(i8 %base, i8 %offset) { +define i1 @t2_commutative(i64 %base, i64* nonnull %offsetptr) { ; CHECK-LABEL: @t2_commutative( -; CHECK-NEXT: [[CMP:%.*]] = icmp ne i8 [[OFFSET:%.*]], 0 -; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]]) -; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i8 [[BASE:%.*]], [[OFFSET]] -; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp ule i8 [[BASE]], [[ADJUSTED]] -; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i8 [[ADJUSTED]], 0 +; CHECK-NEXT: [[OFFSET:%.*]] = ptrtoint i64* [[OFFSETPTR:%.*]] to i64 +; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i64 [[BASE:%.*]], [[OFFSET]] +; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp ule i64 [[BASE]], [[ADJUSTED]] +; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i64 [[ADJUSTED]], 0 ; CHECK-NEXT: [[R:%.*]] = and i1 [[NOT_NULL]], [[NO_UNDERFLOW]] ; CHECK-NEXT: ret i1 [[R]] ; - %cmp = icmp ne i8 %offset, 0 - call void @llvm.assume(i1 %cmp) + %offset = ptrtoint i64* %offsetptr to i64 - %adjusted = sub i8 %base, %offset - %no_underflow = icmp ule i8 %base, %adjusted - %not_null = icmp ne i8 %adjusted, 0 + %adjusted = sub i64 %base, %offset + %no_underflow = icmp ule i64 %base, %adjusted + %not_null = icmp ne i64 %adjusted, 0 %r = and i1 %not_null, %no_underflow ret i1 %r } +define i1 @t3_commutative(i64 %base, i64* nonnull %offsetptr) { +; CHECK-LABEL: @t3_commutative( +; CHECK-NEXT: [[OFFSET:%.*]] = ptrtoint i64* [[OFFSETPTR:%.*]] to i64 +; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i64 [[BASE:%.*]], [[OFFSET]] +; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp ugt i64 [[BASE]], [[ADJUSTED]] +; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp eq i64 [[ADJUSTED]], 0 +; CHECK-NEXT: [[R:%.*]] = or i1 [[NOT_NULL]], [[NO_UNDERFLOW]] +; CHECK-NEXT: ret i1 [[R]] +; + %offset = ptrtoint i64* %offsetptr to i64 + + %adjusted = sub i64 %base, %offset + %no_underflow = icmp ugt i64 %base, %adjusted + %not_null = icmp eq i64 %adjusted, 0 + %r = or i1 %not_null, %no_underflow + ret i1 %r +} + ; We don't know that offset is non-zero, so we can't fold. -define i1 @t3_bad(i8 %base, i8 %offset) { -; CHECK-LABEL: @t3_bad( -; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i8 [[BASE:%.*]], [[OFFSET:%.*]] -; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp uge i8 [[ADJUSTED]], [[BASE]] -; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i8 [[ADJUSTED]], 0 +define i1 @t4_bad(i64 %base, i64 %offset) { +; CHECK-LABEL: @t4_bad( +; CHECK-NEXT: [[ADJUSTED:%.*]] = sub i64 [[BASE:%.*]], [[OFFSET:%.*]] +; CHECK-NEXT: [[NO_UNDERFLOW:%.*]] = icmp uge i64 [[ADJUSTED]], [[BASE]] +; CHECK-NEXT: [[NOT_NULL:%.*]] = icmp ne i64 [[ADJUSTED]], 0 ; CHECK-NEXT: [[R:%.*]] = and i1 [[NOT_NULL]], [[NO_UNDERFLOW]] ; CHECK-NEXT: ret i1 [[R]] ; - %adjusted = sub i8 %base, %offset - %no_underflow = icmp uge i8 %adjusted, %base - %not_null = icmp ne i8 %adjusted, 0 + %adjusted = sub i64 %base, %offset + %no_underflow = icmp uge i64 %adjusted, %base + %not_null = icmp ne i64 %adjusted, 0 %r = and i1 %not_null, %no_underflow ret i1 %r }