2021-04-28 10:25:30 +08:00
|
|
|
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
2022-02-04 20:39:52 +08:00
|
|
|
; RUN: opt %s -passes=instcombine -S | FileCheck %s
|
2021-04-28 10:25:30 +08:00
|
|
|
|
|
|
|
declare { i4, i1 } @llvm.smul.with.overflow.i4(i4, i4) #1
|
|
|
|
|
|
|
|
define i1 @t0_umul(i4 %size, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @t0_umul(
|
[InstCombine] Fold overflow bit of [u|s]mul.with.overflow in a poison-safe way
As discussed in D101191, this patch adds a poison-safe folding of overflow bit check:
```
%Op0 = icmp ne i4 %X, 0
%Agg = call { i4, i1 } @llvm.[us]mul.with.overflow.i4(i4 %X, i4 %Y)
%Op1 = extractvalue { i4, i1 } %Agg, 1
%ret = select i1 %Op0, i1 %Op1, i1 false
=>
%Y.fr = freeze %Y
%Agg = call { i4, i1 } @llvm.[us]mul.with.overflow.i4(i4 %X, i4 %Y.fr)
%Op1 = extractvalue { i4, i1 } %Agg, 1
%ret = %Op1
```
https://alive2.llvm.org/ce/z/zgPUGT
https://alive2.llvm.org/ce/z/h2gZ_6
Note that there are cases where inserting freeze is not necessary: e.g. %Y is `noundef`.
In this case, LLVM is already good because `%ret` is already successfully folded into `and`,
triggering the pre-existing optimization in InstSimplify: https://godbolt.org/z/v6qena15K
Differential Revision: https://reviews.llvm.org/D101423
2021-04-28 10:34:45 +08:00
|
|
|
; CHECK-NEXT: [[NMEMB_FR:%.*]] = freeze i4 [[NMEMB:%.*]]
|
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE:%.*]], i4 [[NMEMB_FR]])
|
2021-04-28 10:25:30 +08:00
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
[InstCombine] Fold overflow bit of [u|s]mul.with.overflow in a poison-safe way
As discussed in D101191, this patch adds a poison-safe folding of overflow bit check:
```
%Op0 = icmp ne i4 %X, 0
%Agg = call { i4, i1 } @llvm.[us]mul.with.overflow.i4(i4 %X, i4 %Y)
%Op1 = extractvalue { i4, i1 } %Agg, 1
%ret = select i1 %Op0, i1 %Op1, i1 false
=>
%Y.fr = freeze %Y
%Agg = call { i4, i1 } @llvm.[us]mul.with.overflow.i4(i4 %X, i4 %Y.fr)
%Op1 = extractvalue { i4, i1 } %Agg, 1
%ret = %Op1
```
https://alive2.llvm.org/ce/z/zgPUGT
https://alive2.llvm.org/ce/z/h2gZ_6
Note that there are cases where inserting freeze is not necessary: e.g. %Y is `noundef`.
In this case, LLVM is already good because `%ret` is already successfully folded into `and`,
triggering the pre-existing optimization in InstSimplify: https://godbolt.org/z/v6qena15K
Differential Revision: https://reviews.llvm.org/D101423
2021-04-28 10:34:45 +08:00
|
|
|
; CHECK-NEXT: ret i1 [[PHITMP]]
|
2021-04-28 10:25:30 +08:00
|
|
|
;
|
|
|
|
%cmp = icmp eq i4 %size, 0
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %cmp, i1 true, i1 %phitmp
|
|
|
|
ret i1 %or
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @t1_commutative(i4 %size, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @t1_commutative(
|
2021-05-02 11:02:02 +08:00
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE:%.*]], i4 [[NMEMB:%.*]])
|
2021-04-28 10:25:30 +08:00
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
2021-05-02 11:02:02 +08:00
|
|
|
; CHECK-NEXT: ret i1 [[PHITMP]]
|
2021-04-28 10:25:30 +08:00
|
|
|
;
|
|
|
|
%cmp = icmp eq i4 %size, 0
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %phitmp, i1 true, i1 %cmp ; swapped
|
|
|
|
ret i1 %or
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @n2_wrong_size(i4 %size0, i4 %size1, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @n2_wrong_size(
|
|
|
|
; CHECK-NEXT: [[CMP:%.*]] = icmp eq i4 [[SIZE1:%.*]], 0
|
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE0:%.*]], i4 [[NMEMB:%.*]])
|
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
|
|
|
; CHECK-NEXT: [[OR:%.*]] = select i1 [[CMP]], i1 true, i1 [[PHITMP]]
|
|
|
|
; CHECK-NEXT: ret i1 [[OR]]
|
|
|
|
;
|
|
|
|
%cmp = icmp eq i4 %size1, 0 ; not %size0
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size0, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %cmp, i1 true, i1 %phitmp
|
|
|
|
ret i1 %or
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @n3_wrong_pred(i4 %size, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @n3_wrong_pred(
|
|
|
|
; CHECK-NEXT: [[CMP:%.*]] = icmp ne i4 [[SIZE:%.*]], 0
|
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE]], i4 [[NMEMB:%.*]])
|
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
|
|
|
; CHECK-NEXT: [[OR:%.*]] = select i1 [[CMP]], i1 true, i1 [[PHITMP]]
|
|
|
|
; CHECK-NEXT: ret i1 [[OR]]
|
|
|
|
;
|
|
|
|
%cmp = icmp ne i4 %size, 0 ; not 'eq'
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %cmp, i1 true, i1 %phitmp
|
|
|
|
ret i1 %or
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @n4_not_and(i4 %size, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @n4_not_and(
|
|
|
|
; CHECK-NEXT: [[CMP:%.*]] = icmp eq i4 [[SIZE:%.*]], 0
|
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE]], i4 [[NMEMB:%.*]])
|
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
|
|
|
; CHECK-NEXT: [[OR:%.*]] = select i1 [[CMP]], i1 [[PHITMP]], i1 false
|
|
|
|
; CHECK-NEXT: ret i1 [[OR]]
|
|
|
|
;
|
|
|
|
%cmp = icmp eq i4 %size, 0
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %cmp, i1 %phitmp, i1 false ; not 'or'
|
|
|
|
ret i1 %or
|
|
|
|
}
|
|
|
|
|
|
|
|
define i1 @n5_not_zero(i4 %size, i4 %nmemb) {
|
|
|
|
; CHECK-LABEL: @n5_not_zero(
|
|
|
|
; CHECK-NEXT: [[CMP:%.*]] = icmp eq i4 [[SIZE:%.*]], 1
|
|
|
|
; CHECK-NEXT: [[SMUL:%.*]] = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 [[SIZE]], i4 [[NMEMB:%.*]])
|
|
|
|
; CHECK-NEXT: [[SMUL_OV:%.*]] = extractvalue { i4, i1 } [[SMUL]], 1
|
|
|
|
; CHECK-NEXT: [[PHITMP:%.*]] = xor i1 [[SMUL_OV]], true
|
|
|
|
; CHECK-NEXT: [[OR:%.*]] = select i1 [[CMP]], i1 true, i1 [[PHITMP]]
|
|
|
|
; CHECK-NEXT: ret i1 [[OR]]
|
|
|
|
;
|
|
|
|
%cmp = icmp eq i4 %size, 1 ; should be '0'
|
|
|
|
%smul = tail call { i4, i1 } @llvm.smul.with.overflow.i4(i4 %size, i4 %nmemb)
|
|
|
|
%smul.ov = extractvalue { i4, i1 } %smul, 1
|
|
|
|
%phitmp = xor i1 %smul.ov, true
|
|
|
|
%or = select i1 %cmp, i1 true, i1 %phitmp
|
|
|
|
ret i1 %or
|
|
|
|
}
|