forked from OSchip/llvm-project
AMDGPU: Annotate features from x work item/group IDs.
This wasn't necessary before since they are always enabled for kernels, but this is necessary if they need to be forwarded to a callable function. llvm-svn: 308226
This commit is contained in:
parent
e59dd00038
commit
e15855d9e3
|
@ -129,8 +129,16 @@ bool AMDGPUAnnotateKernelFeatures::visitConstantExprsRecursively(
|
||||||
//
|
//
|
||||||
// TODO: We should not add the attributes if the known compile time workgroup
|
// TODO: We should not add the attributes if the known compile time workgroup
|
||||||
// size is 1 for y/z.
|
// size is 1 for y/z.
|
||||||
static StringRef intrinsicToAttrName(Intrinsic::ID ID, bool &IsQueuePtr) {
|
static StringRef intrinsicToAttrName(Intrinsic::ID ID,
|
||||||
|
bool &NonKernelOnly,
|
||||||
|
bool &IsQueuePtr) {
|
||||||
switch (ID) {
|
switch (ID) {
|
||||||
|
case Intrinsic::amdgcn_workitem_id_x:
|
||||||
|
NonKernelOnly = true;
|
||||||
|
return "amdgpu-work-item-id-x";
|
||||||
|
case Intrinsic::amdgcn_workgroup_id_x:
|
||||||
|
NonKernelOnly = true;
|
||||||
|
return "amdgpu-work-group-id-x";
|
||||||
case Intrinsic::amdgcn_workitem_id_y:
|
case Intrinsic::amdgcn_workitem_id_y:
|
||||||
case Intrinsic::r600_read_tidig_y:
|
case Intrinsic::r600_read_tidig_y:
|
||||||
return "amdgpu-work-item-id-y";
|
return "amdgpu-work-item-id-y";
|
||||||
|
@ -172,12 +180,12 @@ static bool handleAttr(Function &Parent, const Function &Callee,
|
||||||
|
|
||||||
static void copyFeaturesToFunction(Function &Parent, const Function &Callee,
|
static void copyFeaturesToFunction(Function &Parent, const Function &Callee,
|
||||||
bool &NeedQueuePtr) {
|
bool &NeedQueuePtr) {
|
||||||
|
// X ids unnecessarily propagated to kernels.
|
||||||
static const StringRef AttrNames[] = {
|
static const StringRef AttrNames[] = {
|
||||||
// .x omitted
|
{ "amdgpu-work-item-id-x" },
|
||||||
{ "amdgpu-work-item-id-y" },
|
{ "amdgpu-work-item-id-y" },
|
||||||
{ "amdgpu-work-item-id-z" },
|
{ "amdgpu-work-item-id-z" },
|
||||||
// .x omitted
|
{ "amdgpu-work-group-id-x" },
|
||||||
{ "amdgpu-work-group-id-y" },
|
{ "amdgpu-work-group-id-y" },
|
||||||
{ "amdgpu-work-group-id-z" },
|
{ "amdgpu-work-group-id-z" },
|
||||||
{ "amdgpu-dispatch-ptr" },
|
{ "amdgpu-dispatch-ptr" },
|
||||||
|
@ -198,6 +206,7 @@ bool AMDGPUAnnotateKernelFeatures::addFeatureAttributes(Function &F) {
|
||||||
|
|
||||||
bool Changed = false;
|
bool Changed = false;
|
||||||
bool NeedQueuePtr = false;
|
bool NeedQueuePtr = false;
|
||||||
|
bool IsFunc = !AMDGPU::isEntryFunctionCC(F.getCallingConv());
|
||||||
|
|
||||||
for (BasicBlock &BB : F) {
|
for (BasicBlock &BB : F) {
|
||||||
for (Instruction &I : BB) {
|
for (Instruction &I : BB) {
|
||||||
|
@ -214,8 +223,10 @@ bool AMDGPUAnnotateKernelFeatures::addFeatureAttributes(Function &F) {
|
||||||
copyFeaturesToFunction(F, *Callee, NeedQueuePtr);
|
copyFeaturesToFunction(F, *Callee, NeedQueuePtr);
|
||||||
Changed = true;
|
Changed = true;
|
||||||
} else {
|
} else {
|
||||||
StringRef AttrName = intrinsicToAttrName(IID, NeedQueuePtr);
|
bool NonKernelOnly = false;
|
||||||
if (!AttrName.empty()) {
|
StringRef AttrName = intrinsicToAttrName(IID,
|
||||||
|
NonKernelOnly, NeedQueuePtr);
|
||||||
|
if (!AttrName.empty() && (IsFunc || !NonKernelOnly)) {
|
||||||
F.addFnAttr(AttrName);
|
F.addFnAttr(AttrName);
|
||||||
Changed = true;
|
Changed = true;
|
||||||
}
|
}
|
||||||
|
|
|
@ -42,6 +42,9 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
|
||||||
WorkGroupIDZSystemSGPR(AMDGPU::NoRegister),
|
WorkGroupIDZSystemSGPR(AMDGPU::NoRegister),
|
||||||
WorkGroupInfoSystemSGPR(AMDGPU::NoRegister),
|
WorkGroupInfoSystemSGPR(AMDGPU::NoRegister),
|
||||||
PrivateSegmentWaveByteOffsetSystemSGPR(AMDGPU::NoRegister),
|
PrivateSegmentWaveByteOffsetSystemSGPR(AMDGPU::NoRegister),
|
||||||
|
WorkItemIDXVGPR(AMDGPU::NoRegister),
|
||||||
|
WorkItemIDYVGPR(AMDGPU::NoRegister),
|
||||||
|
WorkItemIDZVGPR(AMDGPU::NoRegister),
|
||||||
PSInputAddr(0),
|
PSInputAddr(0),
|
||||||
PSInputEnable(0),
|
PSInputEnable(0),
|
||||||
ReturnsVoid(true),
|
ReturnsVoid(true),
|
||||||
|
@ -87,7 +90,6 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
|
||||||
ScratchWaveOffsetReg = AMDGPU::SGPR4;
|
ScratchWaveOffsetReg = AMDGPU::SGPR4;
|
||||||
FrameOffsetReg = AMDGPU::SGPR5;
|
FrameOffsetReg = AMDGPU::SGPR5;
|
||||||
StackPtrOffsetReg = AMDGPU::SGPR32;
|
StackPtrOffsetReg = AMDGPU::SGPR32;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
CallingConv::ID CC = F->getCallingConv();
|
CallingConv::ID CC = F->getCallingConv();
|
||||||
|
@ -101,17 +103,25 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
|
||||||
|
|
||||||
if (ST.debuggerEmitPrologue()) {
|
if (ST.debuggerEmitPrologue()) {
|
||||||
// Enable everything.
|
// Enable everything.
|
||||||
|
WorkGroupIDX = true;
|
||||||
WorkGroupIDY = true;
|
WorkGroupIDY = true;
|
||||||
WorkGroupIDZ = true;
|
WorkGroupIDZ = true;
|
||||||
|
WorkItemIDX = true;
|
||||||
WorkItemIDY = true;
|
WorkItemIDY = true;
|
||||||
WorkItemIDZ = true;
|
WorkItemIDZ = true;
|
||||||
} else {
|
} else {
|
||||||
|
if (F->hasFnAttribute("amdgpu-work-group-id-x"))
|
||||||
|
WorkGroupIDX = true;
|
||||||
|
|
||||||
if (F->hasFnAttribute("amdgpu-work-group-id-y"))
|
if (F->hasFnAttribute("amdgpu-work-group-id-y"))
|
||||||
WorkGroupIDY = true;
|
WorkGroupIDY = true;
|
||||||
|
|
||||||
if (F->hasFnAttribute("amdgpu-work-group-id-z"))
|
if (F->hasFnAttribute("amdgpu-work-group-id-z"))
|
||||||
WorkGroupIDZ = true;
|
WorkGroupIDZ = true;
|
||||||
|
|
||||||
|
if (F->hasFnAttribute("amdgpu-work-item-id-x"))
|
||||||
|
WorkItemIDX = true;
|
||||||
|
|
||||||
if (F->hasFnAttribute("amdgpu-work-item-id-y"))
|
if (F->hasFnAttribute("amdgpu-work-item-id-y"))
|
||||||
WorkItemIDY = true;
|
WorkItemIDY = true;
|
||||||
|
|
||||||
|
@ -119,22 +129,24 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
|
||||||
WorkItemIDZ = true;
|
WorkItemIDZ = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
// X, XY, and XYZ are the only supported combinations, so make sure Y is
|
|
||||||
// enabled if Z is.
|
|
||||||
if (WorkItemIDZ)
|
|
||||||
WorkItemIDY = true;
|
|
||||||
|
|
||||||
const MachineFrameInfo &FrameInfo = MF.getFrameInfo();
|
const MachineFrameInfo &FrameInfo = MF.getFrameInfo();
|
||||||
bool MaySpill = ST.isVGPRSpillingEnabled(*F);
|
bool MaySpill = ST.isVGPRSpillingEnabled(*F);
|
||||||
bool HasStackObjects = FrameInfo.hasStackObjects() || FrameInfo.hasCalls();
|
bool HasStackObjects = FrameInfo.hasStackObjects() || FrameInfo.hasCalls();
|
||||||
|
|
||||||
if (HasStackObjects || MaySpill) {
|
if (isEntryFunction()) {
|
||||||
PrivateSegmentWaveByteOffset = true;
|
// X, XY, and XYZ are the only supported combinations, so make sure Y is
|
||||||
|
// enabled if Z is.
|
||||||
|
if (WorkItemIDZ)
|
||||||
|
WorkItemIDY = true;
|
||||||
|
|
||||||
// HS and GS always have the scratch wave offset in SGPR5 on GFX9.
|
if (HasStackObjects || MaySpill) {
|
||||||
if (ST.getGeneration() >= AMDGPUSubtarget::GFX9 &&
|
PrivateSegmentWaveByteOffset = true;
|
||||||
(CC == CallingConv::AMDGPU_HS || CC == CallingConv::AMDGPU_GS))
|
|
||||||
PrivateSegmentWaveByteOffsetSystemSGPR = AMDGPU::SGPR5;
|
// HS and GS always have the scratch wave offset in SGPR5 on GFX9.
|
||||||
|
if (ST.getGeneration() >= AMDGPUSubtarget::GFX9 &&
|
||||||
|
(CC == CallingConv::AMDGPU_HS || CC == CallingConv::AMDGPU_GS))
|
||||||
|
PrivateSegmentWaveByteOffsetSystemSGPR = AMDGPU::SGPR5;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ST.isAmdCodeObjectV2(MF)) {
|
if (ST.isAmdCodeObjectV2(MF)) {
|
||||||
|
@ -160,7 +172,8 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
|
||||||
// We don't need to worry about accessing spills with flat instructions.
|
// We don't need to worry about accessing spills with flat instructions.
|
||||||
// TODO: On VI where we must use flat for global, we should be able to omit
|
// TODO: On VI where we must use flat for global, we should be able to omit
|
||||||
// this if it is never used for generic access.
|
// this if it is never used for generic access.
|
||||||
if (HasStackObjects && ST.hasFlatAddressSpace() && ST.isAmdHsaOS())
|
if (HasStackObjects && ST.hasFlatAddressSpace() && ST.isAmdHsaOS() &&
|
||||||
|
isEntryFunction())
|
||||||
FlatScratchInit = true;
|
FlatScratchInit = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -119,6 +119,11 @@ class SIMachineFunctionInfo final : public AMDGPUMachineFunction {
|
||||||
unsigned WorkGroupInfoSystemSGPR;
|
unsigned WorkGroupInfoSystemSGPR;
|
||||||
unsigned PrivateSegmentWaveByteOffsetSystemSGPR;
|
unsigned PrivateSegmentWaveByteOffsetSystemSGPR;
|
||||||
|
|
||||||
|
// VGPR inputs. These are always v0, v1 and v2 for entry functions.
|
||||||
|
unsigned WorkItemIDXVGPR;
|
||||||
|
unsigned WorkItemIDYVGPR;
|
||||||
|
unsigned WorkItemIDZVGPR;
|
||||||
|
|
||||||
// Graphics info.
|
// Graphics info.
|
||||||
unsigned PSInputAddr;
|
unsigned PSInputAddr;
|
||||||
unsigned PSInputEnable;
|
unsigned PSInputEnable;
|
||||||
|
|
|
@ -1,8 +1,10 @@
|
||||||
; RUN: opt -S -mtriple=amdgcn-unknown-amdhsa -amdgpu-annotate-kernel-features %s | FileCheck -check-prefix=HSA %s
|
; RUN: opt -S -mtriple=amdgcn-unknown-amdhsa -amdgpu-annotate-kernel-features %s | FileCheck -check-prefix=HSA %s
|
||||||
|
|
||||||
|
declare i32 @llvm.amdgcn.workgroup.id.x() #0
|
||||||
declare i32 @llvm.amdgcn.workgroup.id.y() #0
|
declare i32 @llvm.amdgcn.workgroup.id.y() #0
|
||||||
declare i32 @llvm.amdgcn.workgroup.id.z() #0
|
declare i32 @llvm.amdgcn.workgroup.id.z() #0
|
||||||
|
|
||||||
|
declare i32 @llvm.amdgcn.workitem.id.x() #0
|
||||||
declare i32 @llvm.amdgcn.workitem.id.y() #0
|
declare i32 @llvm.amdgcn.workitem.id.y() #0
|
||||||
declare i32 @llvm.amdgcn.workitem.id.z() #0
|
declare i32 @llvm.amdgcn.workitem.id.z() #0
|
||||||
|
|
||||||
|
@ -12,56 +14,70 @@ declare i8 addrspace(2)* @llvm.amdgcn.kernarg.segment.ptr() #0
|
||||||
declare i8 addrspace(2)* @llvm.amdgcn.implicitarg.ptr() #0
|
declare i8 addrspace(2)* @llvm.amdgcn.implicitarg.ptr() #0
|
||||||
declare i64 @llvm.amdgcn.dispatch.id() #0
|
declare i64 @llvm.amdgcn.dispatch.id() #0
|
||||||
|
|
||||||
; HSA: define void @use_workitem_id_y() #1 {
|
; HSA: define void @use_workitem_id_x() #1 {
|
||||||
|
define void @use_workitem_id_x() #1 {
|
||||||
|
%val = call i32 @llvm.amdgcn.workitem.id.x()
|
||||||
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @use_workitem_id_y() #2 {
|
||||||
define void @use_workitem_id_y() #1 {
|
define void @use_workitem_id_y() #1 {
|
||||||
%val = call i32 @llvm.amdgcn.workitem.id.y()
|
%val = call i32 @llvm.amdgcn.workitem.id.y()
|
||||||
store volatile i32 %val, i32 addrspace(1)* undef
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_workitem_id_z() #2 {
|
; HSA: define void @use_workitem_id_z() #3 {
|
||||||
define void @use_workitem_id_z() #1 {
|
define void @use_workitem_id_z() #1 {
|
||||||
%val = call i32 @llvm.amdgcn.workitem.id.z()
|
%val = call i32 @llvm.amdgcn.workitem.id.z()
|
||||||
store volatile i32 %val, i32 addrspace(1)* undef
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_workgroup_id_y() #3 {
|
; HSA: define void @use_workgroup_id_x() #4 {
|
||||||
|
define void @use_workgroup_id_x() #1 {
|
||||||
|
%val = call i32 @llvm.amdgcn.workgroup.id.x()
|
||||||
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @use_workgroup_id_y() #5 {
|
||||||
define void @use_workgroup_id_y() #1 {
|
define void @use_workgroup_id_y() #1 {
|
||||||
%val = call i32 @llvm.amdgcn.workgroup.id.y()
|
%val = call i32 @llvm.amdgcn.workgroup.id.y()
|
||||||
store volatile i32 %val, i32 addrspace(1)* undef
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_workgroup_id_z() #4 {
|
; HSA: define void @use_workgroup_id_z() #6 {
|
||||||
define void @use_workgroup_id_z() #1 {
|
define void @use_workgroup_id_z() #1 {
|
||||||
%val = call i32 @llvm.amdgcn.workgroup.id.z()
|
%val = call i32 @llvm.amdgcn.workgroup.id.z()
|
||||||
store volatile i32 %val, i32 addrspace(1)* undef
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_dispatch_ptr() #5 {
|
; HSA: define void @use_dispatch_ptr() #7 {
|
||||||
define void @use_dispatch_ptr() #1 {
|
define void @use_dispatch_ptr() #1 {
|
||||||
%dispatch.ptr = call i8 addrspace(2)* @llvm.amdgcn.dispatch.ptr()
|
%dispatch.ptr = call i8 addrspace(2)* @llvm.amdgcn.dispatch.ptr()
|
||||||
store volatile i8 addrspace(2)* %dispatch.ptr, i8 addrspace(2)* addrspace(1)* undef
|
store volatile i8 addrspace(2)* %dispatch.ptr, i8 addrspace(2)* addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_queue_ptr() #6 {
|
; HSA: define void @use_queue_ptr() #8 {
|
||||||
define void @use_queue_ptr() #1 {
|
define void @use_queue_ptr() #1 {
|
||||||
%queue.ptr = call i8 addrspace(2)* @llvm.amdgcn.queue.ptr()
|
%queue.ptr = call i8 addrspace(2)* @llvm.amdgcn.queue.ptr()
|
||||||
store volatile i8 addrspace(2)* %queue.ptr, i8 addrspace(2)* addrspace(1)* undef
|
store volatile i8 addrspace(2)* %queue.ptr, i8 addrspace(2)* addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_dispatch_id() #7 {
|
; HSA: define void @use_dispatch_id() #9 {
|
||||||
define void @use_dispatch_id() #1 {
|
define void @use_dispatch_id() #1 {
|
||||||
%val = call i64 @llvm.amdgcn.dispatch.id()
|
%val = call i64 @llvm.amdgcn.dispatch.id()
|
||||||
store volatile i64 %val, i64 addrspace(1)* undef
|
store volatile i64 %val, i64 addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_workgroup_id_y_workgroup_id_z() #8 {
|
; HSA: define void @use_workgroup_id_y_workgroup_id_z() #10 {
|
||||||
define void @use_workgroup_id_y_workgroup_id_z() #1 {
|
define void @use_workgroup_id_y_workgroup_id_z() #1 {
|
||||||
%val0 = call i32 @llvm.amdgcn.workgroup.id.y()
|
%val0 = call i32 @llvm.amdgcn.workgroup.id.y()
|
||||||
%val1 = call i32 @llvm.amdgcn.workgroup.id.z()
|
%val1 = call i32 @llvm.amdgcn.workgroup.id.z()
|
||||||
|
@ -70,67 +86,91 @@ define void @use_workgroup_id_y_workgroup_id_z() #1 {
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_workitem_id_y() #1 {
|
; HSA: define void @func_indirect_use_workitem_id_x() #1 {
|
||||||
|
define void @func_indirect_use_workitem_id_x() #1 {
|
||||||
|
call void @use_workitem_id_x()
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @kernel_indirect_use_workitem_id_x() #1 {
|
||||||
|
define void @kernel_indirect_use_workitem_id_x() #1 {
|
||||||
|
call void @use_workitem_id_x()
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @func_indirect_use_workitem_id_y() #2 {
|
||||||
define void @func_indirect_use_workitem_id_y() #1 {
|
define void @func_indirect_use_workitem_id_y() #1 {
|
||||||
call void @use_workitem_id_y()
|
call void @use_workitem_id_y()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_workitem_id_z() #2 {
|
; HSA: define void @func_indirect_use_workitem_id_z() #3 {
|
||||||
define void @func_indirect_use_workitem_id_z() #1 {
|
define void @func_indirect_use_workitem_id_z() #1 {
|
||||||
call void @use_workitem_id_z()
|
call void @use_workitem_id_z()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_workgroup_id_y() #3 {
|
; HSA: define void @func_indirect_use_workgroup_id_x() #4 {
|
||||||
|
define void @func_indirect_use_workgroup_id_x() #1 {
|
||||||
|
call void @use_workgroup_id_x()
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @kernel_indirect_use_workgroup_id_x() #4 {
|
||||||
|
define void @kernel_indirect_use_workgroup_id_x() #1 {
|
||||||
|
call void @use_workgroup_id_x()
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
; HSA: define void @func_indirect_use_workgroup_id_y() #5 {
|
||||||
define void @func_indirect_use_workgroup_id_y() #1 {
|
define void @func_indirect_use_workgroup_id_y() #1 {
|
||||||
call void @use_workgroup_id_y()
|
call void @use_workgroup_id_y()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_workgroup_id_z() #4 {
|
; HSA: define void @func_indirect_use_workgroup_id_z() #6 {
|
||||||
define void @func_indirect_use_workgroup_id_z() #1 {
|
define void @func_indirect_use_workgroup_id_z() #1 {
|
||||||
call void @use_workgroup_id_z()
|
call void @use_workgroup_id_z()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_indirect_use_workgroup_id_y() #3 {
|
; HSA: define void @func_indirect_indirect_use_workgroup_id_y() #5 {
|
||||||
define void @func_indirect_indirect_use_workgroup_id_y() #1 {
|
define void @func_indirect_indirect_use_workgroup_id_y() #1 {
|
||||||
call void @func_indirect_use_workgroup_id_y()
|
call void @func_indirect_use_workgroup_id_y()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @indirect_x2_use_workgroup_id_y() #3 {
|
; HSA: define void @indirect_x2_use_workgroup_id_y() #5 {
|
||||||
define void @indirect_x2_use_workgroup_id_y() #1 {
|
define void @indirect_x2_use_workgroup_id_y() #1 {
|
||||||
call void @func_indirect_indirect_use_workgroup_id_y()
|
call void @func_indirect_indirect_use_workgroup_id_y()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_dispatch_ptr() #5 {
|
; HSA: define void @func_indirect_use_dispatch_ptr() #7 {
|
||||||
define void @func_indirect_use_dispatch_ptr() #1 {
|
define void @func_indirect_use_dispatch_ptr() #1 {
|
||||||
call void @use_dispatch_ptr()
|
call void @use_dispatch_ptr()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_queue_ptr() #6 {
|
; HSA: define void @func_indirect_use_queue_ptr() #8 {
|
||||||
define void @func_indirect_use_queue_ptr() #1 {
|
define void @func_indirect_use_queue_ptr() #1 {
|
||||||
call void @use_queue_ptr()
|
call void @use_queue_ptr()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_dispatch_id() #7 {
|
; HSA: define void @func_indirect_use_dispatch_id() #9 {
|
||||||
define void @func_indirect_use_dispatch_id() #1 {
|
define void @func_indirect_use_dispatch_id() #1 {
|
||||||
call void @use_dispatch_id()
|
call void @use_dispatch_id()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_workgroup_id_y_workgroup_id_z() #9 {
|
; HSA: define void @func_indirect_use_workgroup_id_y_workgroup_id_z() #11 {
|
||||||
define void @func_indirect_use_workgroup_id_y_workgroup_id_z() #1 {
|
define void @func_indirect_use_workgroup_id_y_workgroup_id_z() #1 {
|
||||||
call void @func_indirect_use_workgroup_id_y_workgroup_id_z()
|
call void @func_indirect_use_workgroup_id_y_workgroup_id_z()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @recursive_use_workitem_id_y() #1 {
|
; HSA: define void @recursive_use_workitem_id_y() #2 {
|
||||||
define void @recursive_use_workitem_id_y() #1 {
|
define void @recursive_use_workitem_id_y() #1 {
|
||||||
%val = call i32 @llvm.amdgcn.workitem.id.y()
|
%val = call i32 @llvm.amdgcn.workitem.id.y()
|
||||||
store volatile i32 %val, i32 addrspace(1)* undef
|
store volatile i32 %val, i32 addrspace(1)* undef
|
||||||
|
@ -138,27 +178,27 @@ define void @recursive_use_workitem_id_y() #1 {
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @call_recursive_use_workitem_id_y() #1 {
|
; HSA: define void @call_recursive_use_workitem_id_y() #2 {
|
||||||
define void @call_recursive_use_workitem_id_y() #1 {
|
define void @call_recursive_use_workitem_id_y() #1 {
|
||||||
call void @recursive_use_workitem_id_y()
|
call void @recursive_use_workitem_id_y()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_group_to_flat_addrspacecast(i32 addrspace(3)* %ptr) #6 {
|
; HSA: define void @use_group_to_flat_addrspacecast(i32 addrspace(3)* %ptr) #8 {
|
||||||
define void @use_group_to_flat_addrspacecast(i32 addrspace(3)* %ptr) #1 {
|
define void @use_group_to_flat_addrspacecast(i32 addrspace(3)* %ptr) #1 {
|
||||||
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
||||||
store volatile i32 0, i32 addrspace(4)* %stof
|
store volatile i32 0, i32 addrspace(4)* %stof
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* %ptr) #10 {
|
; HSA: define void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* %ptr) #12 {
|
||||||
define void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* %ptr) #2 {
|
define void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* %ptr) #2 {
|
||||||
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
||||||
store volatile i32 0, i32 addrspace(4)* %stof
|
store volatile i32 0, i32 addrspace(4)* %stof
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* %ptr) #11 {
|
; HSA: define void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* %ptr) #13 {
|
||||||
define void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* %ptr) #2 {
|
define void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* %ptr) #2 {
|
||||||
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
%stof = addrspacecast i32 addrspace(3)* %ptr to i32 addrspace(4)*
|
||||||
store volatile i32 0, i32 addrspace(4)* %stof
|
store volatile i32 0, i32 addrspace(4)* %stof
|
||||||
|
@ -166,45 +206,45 @@ define void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* %p
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @indirect_use_group_to_flat_addrspacecast() #6 {
|
; HSA: define void @indirect_use_group_to_flat_addrspacecast() #8 {
|
||||||
define void @indirect_use_group_to_flat_addrspacecast() #1 {
|
define void @indirect_use_group_to_flat_addrspacecast() #1 {
|
||||||
call void @use_group_to_flat_addrspacecast(i32 addrspace(3)* null)
|
call void @use_group_to_flat_addrspacecast(i32 addrspace(3)* null)
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @indirect_use_group_to_flat_addrspacecast_gfx9() #9 {
|
; HSA: define void @indirect_use_group_to_flat_addrspacecast_gfx9() #11 {
|
||||||
define void @indirect_use_group_to_flat_addrspacecast_gfx9() #1 {
|
define void @indirect_use_group_to_flat_addrspacecast_gfx9() #1 {
|
||||||
call void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* null)
|
call void @use_group_to_flat_addrspacecast_gfx9(i32 addrspace(3)* null)
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @indirect_use_group_to_flat_addrspacecast_queue_ptr_gfx9() #6 {
|
; HSA: define void @indirect_use_group_to_flat_addrspacecast_queue_ptr_gfx9() #8 {
|
||||||
define void @indirect_use_group_to_flat_addrspacecast_queue_ptr_gfx9() #1 {
|
define void @indirect_use_group_to_flat_addrspacecast_queue_ptr_gfx9() #1 {
|
||||||
call void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* null)
|
call void @use_group_to_flat_addrspacecast_queue_ptr_gfx9(i32 addrspace(3)* null)
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_kernarg_segment_ptr() #12 {
|
; HSA: define void @use_kernarg_segment_ptr() #14 {
|
||||||
define void @use_kernarg_segment_ptr() #1 {
|
define void @use_kernarg_segment_ptr() #1 {
|
||||||
%kernarg.segment.ptr = call i8 addrspace(2)* @llvm.amdgcn.kernarg.segment.ptr()
|
%kernarg.segment.ptr = call i8 addrspace(2)* @llvm.amdgcn.kernarg.segment.ptr()
|
||||||
store volatile i8 addrspace(2)* %kernarg.segment.ptr, i8 addrspace(2)* addrspace(1)* undef
|
store volatile i8 addrspace(2)* %kernarg.segment.ptr, i8 addrspace(2)* addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_kernarg_segment_ptr() #12 {
|
; HSA: define void @func_indirect_use_kernarg_segment_ptr() #14 {
|
||||||
define void @func_indirect_use_kernarg_segment_ptr() #1 {
|
define void @func_indirect_use_kernarg_segment_ptr() #1 {
|
||||||
call void @use_kernarg_segment_ptr()
|
call void @use_kernarg_segment_ptr()
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @use_implicitarg_ptr() #12 {
|
; HSA: define void @use_implicitarg_ptr() #14 {
|
||||||
define void @use_implicitarg_ptr() #1 {
|
define void @use_implicitarg_ptr() #1 {
|
||||||
%implicitarg.ptr = call i8 addrspace(2)* @llvm.amdgcn.implicitarg.ptr()
|
%implicitarg.ptr = call i8 addrspace(2)* @llvm.amdgcn.implicitarg.ptr()
|
||||||
store volatile i8 addrspace(2)* %implicitarg.ptr, i8 addrspace(2)* addrspace(1)* undef
|
store volatile i8 addrspace(2)* %implicitarg.ptr, i8 addrspace(2)* addrspace(1)* undef
|
||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
; HSA: define void @func_indirect_use_implicitarg_ptr() #12 {
|
; HSA: define void @func_indirect_use_implicitarg_ptr() #14 {
|
||||||
define void @func_indirect_use_implicitarg_ptr() #1 {
|
define void @func_indirect_use_implicitarg_ptr() #1 {
|
||||||
call void @use_implicitarg_ptr()
|
call void @use_implicitarg_ptr()
|
||||||
ret void
|
ret void
|
||||||
|
@ -215,15 +255,17 @@ attributes #1 = { nounwind "target-cpu"="fiji" }
|
||||||
attributes #2 = { nounwind "target-cpu"="gfx900" }
|
attributes #2 = { nounwind "target-cpu"="gfx900" }
|
||||||
|
|
||||||
; HSA: attributes #0 = { nounwind readnone speculatable }
|
; HSA: attributes #0 = { nounwind readnone speculatable }
|
||||||
; HSA: attributes #1 = { nounwind "amdgpu-work-item-id-y" "target-cpu"="fiji" }
|
; HSA: attributes #1 = { nounwind "amdgpu-work-item-id-x" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #2 = { nounwind "amdgpu-work-item-id-z" "target-cpu"="fiji" }
|
; HSA: attributes #2 = { nounwind "amdgpu-work-item-id-y" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #3 = { nounwind "amdgpu-work-group-id-y" "target-cpu"="fiji" }
|
; HSA: attributes #3 = { nounwind "amdgpu-work-item-id-z" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #4 = { nounwind "amdgpu-work-group-id-z" "target-cpu"="fiji" }
|
; HSA: attributes #4 = { nounwind "amdgpu-work-group-id-x" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #5 = { nounwind "amdgpu-dispatch-ptr" "target-cpu"="fiji" }
|
; HSA: attributes #5 = { nounwind "amdgpu-work-group-id-y" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #6 = { nounwind "amdgpu-queue-ptr" "target-cpu"="fiji" }
|
; HSA: attributes #6 = { nounwind "amdgpu-work-group-id-z" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #7 = { nounwind "amdgpu-dispatch-id" "target-cpu"="fiji" }
|
; HSA: attributes #7 = { nounwind "amdgpu-dispatch-ptr" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #8 = { nounwind "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "target-cpu"="fiji" }
|
; HSA: attributes #8 = { nounwind "amdgpu-queue-ptr" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #9 = { nounwind "target-cpu"="fiji" }
|
; HSA: attributes #9 = { nounwind "amdgpu-dispatch-id" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #10 = { nounwind "target-cpu"="gfx900" }
|
; HSA: attributes #10 = { nounwind "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "target-cpu"="fiji" }
|
||||||
; HSA: attributes #11 = { nounwind "amdgpu-queue-ptr" "target-cpu"="gfx900" }
|
; HSA: attributes #11 = { nounwind "target-cpu"="fiji" }
|
||||||
; HSA: attributes #12 = { nounwind "amdgpu-kernarg-segment-ptr" "target-cpu"="fiji" }
|
; HSA: attributes #12 = { nounwind "target-cpu"="gfx900" }
|
||||||
|
; HSA: attributes #13 = { nounwind "amdgpu-queue-ptr" "target-cpu"="gfx900" }
|
||||||
|
; HSA: attributes #14 = { nounwind "amdgpu-kernarg-segment-ptr" "target-cpu"="fiji" }
|
||||||
|
|
Loading…
Reference in New Issue