forked from OSchip/llvm-project
parent
d064cc93b2
commit
72f259b8eb
|
@ -141,6 +141,9 @@ private:
|
|||
void inferAddressSpaces(const std::vector<Value *> &Postorder,
|
||||
ValueToAddrSpaceMapTy *InferredAddrSpace) const;
|
||||
|
||||
bool handleComplexPtrUse(User &U, Value *OldV, Value *NewV) const;
|
||||
bool isSafeToCastConstAddrSpace(Constant *C, unsigned NewAS) const;
|
||||
|
||||
// Changes the flat address expressions in function F to point to specific
|
||||
// address spaces if InferredAddrSpace says so. Postorder is the postorder of
|
||||
// all flat expressions in the use-def graph of function F.
|
||||
|
@ -307,6 +310,13 @@ InferAddressSpaces::collectFlatAddressExpressions(Function &F) const {
|
|||
PushPtrOperand(MTI->getRawSource());
|
||||
} else if (auto *II = dyn_cast<IntrinsicInst>(&I))
|
||||
collectRewritableIntrinsicOperands(II, &PostorderStack, &Visited);
|
||||
else if (ICmpInst *Cmp = dyn_cast<ICmpInst>(&I)) {
|
||||
// FIXME: Handle vectors of pointers
|
||||
if (Cmp->getOperand(0)->getType()->isPointerTy()) {
|
||||
PushPtrOperand(Cmp->getOperand(0));
|
||||
PushPtrOperand(Cmp->getOperand(1));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
std::vector<Value *> Postorder; // The resultant postorder.
|
||||
|
@ -662,6 +672,29 @@ static bool handleMemIntrinsicPtrUse(MemIntrinsic *MI,
|
|||
return true;
|
||||
}
|
||||
|
||||
// \p returns true if it is OK to change the address space of constant \p C with
|
||||
// a ConstantExpr addrspacecast.
|
||||
bool InferAddressSpaces::isSafeToCastConstAddrSpace(Constant *C, unsigned NewAS) const {
|
||||
if (C->getType()->getPointerAddressSpace() == NewAS)
|
||||
return true;
|
||||
|
||||
if (isa<UndefValue>(C) || isa<ConstantPointerNull>(C))
|
||||
return true;
|
||||
|
||||
if (auto *Op = dyn_cast<Operator>(C)) {
|
||||
// If we already have a constant addrspacecast, it should be safe to cast it
|
||||
// off.
|
||||
if (Op->getOpcode() == Instruction::AddrSpaceCast)
|
||||
return isSafeToCastConstAddrSpace(cast<Constant>(Op->getOperand(0)), NewAS);
|
||||
|
||||
if (Op->getOpcode() == Instruction::IntToPtr &&
|
||||
Op->getType()->getPointerAddressSpace() == FlatAddrSpace)
|
||||
return true;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
static Value::use_iterator skipToNextUser(Value::use_iterator I,
|
||||
Value::use_iterator End) {
|
||||
User *CurUser = I->getUser();
|
||||
|
@ -740,15 +773,38 @@ bool InferAddressSpaces::rewriteWithNewAddressSpaces(
|
|||
}
|
||||
|
||||
if (isa<Instruction>(CurUser)) {
|
||||
if (ICmpInst *Cmp = dyn_cast<ICmpInst>(CurUser)) {
|
||||
// If we can infer that both pointers are in the same addrspace,
|
||||
// transform e.g.
|
||||
// %cmp = icmp eq float* %p, %q
|
||||
// into
|
||||
// %cmp = icmp eq float addrspace(3)* %new_p, %new_q
|
||||
|
||||
unsigned NewAS = NewV->getType()->getPointerAddressSpace();
|
||||
int SrcIdx = U.getOperandNo();
|
||||
int OtherIdx = (SrcIdx == 0) ? 1 : 0;
|
||||
Value *OtherSrc = Cmp->getOperand(OtherIdx);
|
||||
|
||||
if (Value *OtherNewV = ValueWithNewAddrSpace.lookup(OtherSrc)) {
|
||||
if (OtherNewV->getType()->getPointerAddressSpace() == NewAS) {
|
||||
Cmp->setOperand(OtherIdx, OtherNewV);
|
||||
Cmp->setOperand(SrcIdx, NewV);
|
||||
continue;
|
||||
}
|
||||
}
|
||||
|
||||
// Even if the type mismatches, we can cast the constant.
|
||||
if (auto *KOtherSrc = dyn_cast<Constant>(OtherSrc)) {
|
||||
if (isSafeToCastConstAddrSpace(KOtherSrc, NewAS)) {
|
||||
Cmp->setOperand(SrcIdx, NewV);
|
||||
Cmp->setOperand(OtherIdx,
|
||||
ConstantExpr::getAddrSpaceCast(KOtherSrc, NewV->getType()));
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Otherwise, replaces the use with flat(NewV).
|
||||
// TODO: Some optimization opportunities are missed. For example, in
|
||||
// %0 = icmp eq float* %p, %q
|
||||
// if both p and q are inferred to be shared, we can rewrite %0 as
|
||||
// %0 = icmp eq float addrspace(3)* %new_p, %new_q
|
||||
// instead of currently
|
||||
// %flat_p = addrspacecast float addrspace(3)* %new_p to float*
|
||||
// %flat_q = addrspacecast float addrspace(3)* %new_q to float*
|
||||
// %0 = icmp eq float* %flat_p, %flat_q
|
||||
if (Instruction *I = dyn_cast<Instruction>(V)) {
|
||||
BasicBlock::iterator InsertPos = std::next(I->getIterator());
|
||||
while (isa<PHINode>(InsertPos))
|
||||
|
|
|
@ -0,0 +1,142 @@
|
|||
; RUN: opt -S -mtriple=amdgcn-amd-amdhsa -infer-address-spaces %s | FileCheck %s
|
||||
|
||||
; CHECK-LABEL: @icmp_flat_cmp_self(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, %group.ptr.0
|
||||
define i1 @icmp_flat_cmp_self(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, %cast0
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_flat_flat_from_group(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, %group.ptr.1
|
||||
define i1 @icmp_flat_flat_from_group(i32 addrspace(3)* %group.ptr.0, i32 addrspace(3)* %group.ptr.1) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cast1 = addrspacecast i32 addrspace(3)* %group.ptr.1 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, %cast1
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_mismatch_flat_from_group_private(
|
||||
; CHECK: %1 = addrspacecast i32* %private.ptr.0 to i32 addrspace(4)*
|
||||
; CHECK: %2 = addrspacecast i32 addrspace(3)* %group.ptr.1 to i32 addrspace(4)*
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(4)* %1, %2
|
||||
define i1 @icmp_mismatch_flat_from_group_private(i32* %private.ptr.0, i32 addrspace(3)* %group.ptr.1) #0 {
|
||||
%cast0 = addrspacecast i32* %private.ptr.0 to i32 addrspace(4)*
|
||||
%cast1 = addrspacecast i32 addrspace(3)* %group.ptr.1 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, %cast1
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_flat_group_flat(
|
||||
; CHECK: %1 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(4)* %1, %flat.ptr.1
|
||||
define i1 @icmp_flat_group_flat(i32 addrspace(3)* %group.ptr.0, i32 addrspace(4)* %flat.ptr.1) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, %flat.ptr.1
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_flat_flat_group(
|
||||
; CHECK: %1 = addrspacecast i32 addrspace(3)* %group.ptr.1 to i32 addrspace(4)*
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(4)* %flat.ptr.0, %1
|
||||
define i1 @icmp_flat_flat_group(i32 addrspace(4)* %flat.ptr.0, i32 addrspace(3)* %group.ptr.1) #0 {
|
||||
%cast1 = addrspacecast i32 addrspace(3)* %group.ptr.1 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %flat.ptr.0, %cast1
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; Keeping as cmp addrspace(3)* is better
|
||||
; CHECK-LABEL: @icmp_flat_to_group_cmp(
|
||||
; CHECK: %cast0 = addrspacecast i32 addrspace(4)* %flat.ptr.0 to i32 addrspace(3)*
|
||||
; CHECK: %cast1 = addrspacecast i32 addrspace(4)* %flat.ptr.1 to i32 addrspace(3)*
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %cast0, %cast1
|
||||
define i1 @icmp_flat_to_group_cmp(i32 addrspace(4)* %flat.ptr.0, i32 addrspace(4)* %flat.ptr.1) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(4)* %flat.ptr.0 to i32 addrspace(3)*
|
||||
%cast1 = addrspacecast i32 addrspace(4)* %flat.ptr.1 to i32 addrspace(3)*
|
||||
%cmp = icmp eq i32 addrspace(3)* %cast0, %cast1
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; FIXME: Should be able to ask target about how to constant fold the
|
||||
; constant cast if this is OK to change if 0 is a valid pointer.
|
||||
|
||||
; CHECK-LABEL: @icmp_group_flat_cmp_null(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, addrspacecast (i32 addrspace(4)* null to i32 addrspace(3)*)
|
||||
define i1 @icmp_group_flat_cmp_null(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, null
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_group_flat_cmp_constant_inttoptr(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, addrspacecast (i32 addrspace(4)* inttoptr (i64 400 to i32 addrspace(4)*) to i32 addrspace(3)*)
|
||||
define i1 @icmp_group_flat_cmp_constant_inttoptr(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, inttoptr (i64 400 to i32 addrspace(4)*)
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_mismatch_flat_group_private_cmp_null(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, addrspacecast (i32* null to i32 addrspace(3)*)
|
||||
define i1 @icmp_mismatch_flat_group_private_cmp_null(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, addrspacecast (i32* null to i32 addrspace(4)*)
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
@lds0 = internal addrspace(3) global i32 0, align 4
|
||||
@global0 = internal addrspace(1) global i32 0, align 4
|
||||
|
||||
; CHECK-LABEL: @icmp_mismatch_flat_group_global_cmp_gv(
|
||||
; CHECK: %1 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(4)* %1, addrspacecast (i32 addrspace(1)* @global0 to i32 addrspace(4)*)
|
||||
define i1 @icmp_mismatch_flat_group_global_cmp_gv(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, addrspacecast (i32 addrspace(1)* @global0 to i32 addrspace(4)*)
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_mismatch_group_global_cmp_gv_gv(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds0 to i32 addrspace(4)*), addrspacecast (i32 addrspace(1)* @global0 to i32 addrspace(4)*)
|
||||
define i1 @icmp_mismatch_group_global_cmp_gv_gv(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cmp = icmp eq i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds0 to i32 addrspace(4)*), addrspacecast (i32 addrspace(1)* @global0 to i32 addrspace(4)*)
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_group_flat_cmp_undef(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* %group.ptr.0, undef
|
||||
define i1 @icmp_group_flat_cmp_undef(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* %cast0, undef
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; Test non-canonical orders
|
||||
; CHECK-LABEL: @icmp_mismatch_flat_group_private_cmp_null_swap(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* addrspacecast (i32* null to i32 addrspace(3)*), %group.ptr.0
|
||||
define i1 @icmp_mismatch_flat_group_private_cmp_null_swap(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* addrspacecast (i32* null to i32 addrspace(4)*), %cast0
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; CHECK-LABEL: @icmp_group_flat_cmp_undef_swap(
|
||||
; CHECK: %cmp = icmp eq i32 addrspace(3)* undef, %group.ptr.0
|
||||
define i1 @icmp_group_flat_cmp_undef_swap(i32 addrspace(3)* %group.ptr.0) #0 {
|
||||
%cast0 = addrspacecast i32 addrspace(3)* %group.ptr.0 to i32 addrspace(4)*
|
||||
%cmp = icmp eq i32 addrspace(4)* undef, %cast0
|
||||
ret i1 %cmp
|
||||
}
|
||||
|
||||
; TODO: Should be handled
|
||||
; CHECK-LABEL: @icmp_flat_flat_from_group_vector(
|
||||
; CHECK: %cmp = icmp eq <2 x i32 addrspace(4)*> %cast0, %cast1
|
||||
define <2 x i1> @icmp_flat_flat_from_group_vector(<2 x i32 addrspace(3)*> %group.ptr.0, <2 x i32 addrspace(3)*> %group.ptr.1) #0 {
|
||||
%cast0 = addrspacecast <2 x i32 addrspace(3)*> %group.ptr.0 to <2 x i32 addrspace(4)*>
|
||||
%cast1 = addrspacecast <2 x i32 addrspace(3)*> %group.ptr.1 to <2 x i32 addrspace(4)*>
|
||||
%cmp = icmp eq <2 x i32 addrspace(4)*> %cast0, %cast1
|
||||
ret <2 x i1> %cmp
|
||||
}
|
||||
|
||||
attributes #0 = { nounwind }
|
|
@ -106,8 +106,7 @@ bb:
|
|||
; CHECK-LABEL: @loop(
|
||||
|
||||
; CHECK: %p = bitcast [10 x float] addrspace(3)* @array to float addrspace(3)*
|
||||
; CHECK: %0 = addrspacecast float addrspace(3)* %p to float addrspace(4)*
|
||||
; CHECK: %end = getelementptr float, float addrspace(4)* %0, i64 10
|
||||
; CHECK: %end = getelementptr float, float addrspace(3)* %p, i64 10
|
||||
; CHECK: br label %loop
|
||||
|
||||
; CHECK: loop: ; preds = %loop, %entry
|
||||
|
@ -115,8 +114,8 @@ bb:
|
|||
; CHECK: %v = load float, float addrspace(3)* %i
|
||||
; CHECK: call void @use(float %v)
|
||||
; CHECK: %i2 = getelementptr float, float addrspace(3)* %i, i64 1
|
||||
; CHECK: %1 = addrspacecast float addrspace(3)* %i2 to float addrspace(4)*
|
||||
; CHECK: %exit_cond = icmp eq float addrspace(4)* %1, %end
|
||||
; CHECK: %exit_cond = icmp eq float addrspace(3)* %i2, %end
|
||||
|
||||
; CHECK: br i1 %exit_cond, label %exit, label %loop
|
||||
define void @loop() #0 {
|
||||
entry:
|
||||
|
|
Loading…
Reference in New Issue