From 8871683d60ce03a271e4ac9c09746a558db8e38f Mon Sep 17 00:00:00 2001 From: Matt Arsenault Date: Tue, 10 Jan 2017 19:08:15 +0000 Subject: [PATCH] AMDGPU: Add tests for HasMultipleConditionRegisters This was enabled without many specific tests or the comment. llvm-svn: 291586 --- llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp | 7 + llvm/test/CodeGen/AMDGPU/select-opt.ll | 161 ++++++++++++++++++ 2 files changed, 168 insertions(+) create mode 100644 llvm/test/CodeGen/AMDGPU/select-opt.ll diff --git a/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp b/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp index 730bcdcf7afa..a700aa85929a 100644 --- a/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp +++ b/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp @@ -434,6 +434,13 @@ AMDGPUTargetLowering::AMDGPUTargetLowering(const TargetMachine &TM, setSchedulingPreference(Sched::RegPressure); setJumpIsExpensive(true); + + // FIXME: This is only partially true. If we have to do vector compares, any + // SGPR pair can be a condition register. If we have a uniform condition, we + // are better off doing SALU operations, where there is only one SCC. For now, + // we don't have a way of knowing during instruction selection if a condition + // will be uniform and we always use vector compares. Assume we are using + // vector compares until that is fixed. setHasMultipleConditionRegisters(true); // SI at least has hardware support for floating point exceptions, but no way diff --git a/llvm/test/CodeGen/AMDGPU/select-opt.ll b/llvm/test/CodeGen/AMDGPU/select-opt.ll new file mode 100644 index 000000000000..ad358d33c405 --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/select-opt.ll @@ -0,0 +1,161 @@ +; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s + +; Make sure to test with f32 and i32 compares. If we have to use float +; compares, we always have multiple condition registers. If we can do +; scalar compares, we don't want to use multiple condition registers. + +; GCN-LABEL: {{^}}opt_select_i32_and_cmp_i32: +; GCN-DAG: v_cmp_ne_u32_e32 vcc, +; GCN-DAG: v_cmp_ne_u32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_and_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 [[RESULT:v[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN-NOT: [[RESULT]] +; GCN: buffer_store_dword [[RESULT]] +define void @opt_select_i32_and_cmp_i32(i32 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i32 %x, i32 %y) #0 { + %icmp0 = icmp ne i32 %a, %b + %icmp1 = icmp ne i32 %a, %c + %and = and i1 %icmp0, %icmp1 + %select = select i1 %and, i32 %x, i32 %y + store i32 %select, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i32_and_cmp_f32: +; GCN-DAG: v_cmp_lg_f32_e32 vcc +; GCN-DAG: v_cmp_lg_f32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_and_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 [[RESULT:v[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN-NOT: [[RESULT]] +; GCN: buffer_store_dword [[RESULT]] +define void @opt_select_i32_and_cmp_f32(i32 addrspace(1)* %out, float %a, float %b, float %c, i32 %x, i32 %y) #0 { + %fcmp0 = fcmp one float %a, %b + %fcmp1 = fcmp one float %a, %c + %and = and i1 %fcmp0, %fcmp1 + %select = select i1 %and, i32 %x, i32 %y + store i32 %select, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i64_and_cmp_i32: +; GCN-DAG: v_cmp_ne_u32_e32 vcc, +; GCN-DAG: v_cmp_ne_u32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_and_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 v[[RESULT1:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: v_cndmask_b32_e32 v[[RESULT0:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: buffer_store_dwordx2 v{{\[}}[[RESULT0]]:[[RESULT1]]{{\]}} +define void @opt_select_i64_and_cmp_i32(i64 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i64 %x, i64 %y) #0 { + %icmp0 = icmp ne i32 %a, %b + %icmp1 = icmp ne i32 %a, %c + %and = and i1 %icmp0, %icmp1 + %select = select i1 %and, i64 %x, i64 %y + store i64 %select, i64 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i64_and_cmp_f32: +; GCN-DAG: v_cmp_lg_f32_e32 vcc, +; GCN-DAG: v_cmp_lg_f32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_and_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 v[[RESULT1:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: v_cndmask_b32_e32 v[[RESULT0:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: buffer_store_dwordx2 v{{\[}}[[RESULT0]]:[[RESULT1]]{{\]}} +define void @opt_select_i64_and_cmp_f32(i64 addrspace(1)* %out, float %a, float %b, float %c, i64 %x, i64 %y) #0 { + %fcmp0 = fcmp one float %a, %b + %fcmp1 = fcmp one float %a, %c + %and = and i1 %fcmp0, %fcmp1 + %select = select i1 %and, i64 %x, i64 %y + store i64 %select, i64 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i32_or_cmp_i32: +; GCN-DAG: v_cmp_ne_u32_e32 vcc, +; GCN-DAG: v_cmp_ne_u32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_or_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 [[RESULT:v[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN-NOT: [[RESULT]] +; GCN: buffer_store_dword [[RESULT]] +; GCN: s_endpgm +define void @opt_select_i32_or_cmp_i32(i32 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i32 %x, i32 %y) #0 { + %icmp0 = icmp ne i32 %a, %b + %icmp1 = icmp ne i32 %a, %c + %or = or i1 %icmp0, %icmp1 + %select = select i1 %or, i32 %x, i32 %y + store i32 %select, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i32_or_cmp_f32: +; GCN-DAG: v_cmp_lg_f32_e32 vcc +; GCN-DAG: v_cmp_lg_f32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_or_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 [[RESULT:v[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN-NOT: [[RESULT]] +; GCN: buffer_store_dword [[RESULT]] +define void @opt_select_i32_or_cmp_f32(i32 addrspace(1)* %out, float %a, float %b, float %c, i32 %x, i32 %y) #0 { + %fcmp0 = fcmp one float %a, %b + %fcmp1 = fcmp one float %a, %c + %or = or i1 %fcmp0, %fcmp1 + %select = select i1 %or, i32 %x, i32 %y + store i32 %select, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i64_or_cmp_i32: +; GCN-DAG: v_cmp_ne_u32_e32 vcc, +; GCN-DAG: v_cmp_ne_u32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_or_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 v[[RESULT1:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: v_cndmask_b32_e32 v[[RESULT0:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: buffer_store_dwordx2 v{{\[}}[[RESULT0]]:[[RESULT1]]{{\]}} +define void @opt_select_i64_or_cmp_i32(i64 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i64 %x, i64 %y) #0 { + %icmp0 = icmp ne i32 %a, %b + %icmp1 = icmp ne i32 %a, %c + %or = or i1 %icmp0, %icmp1 + %select = select i1 %or, i64 %x, i64 %y + store i64 %select, i64 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}opt_select_i64_or_cmp_f32: +; GCN-DAG: v_cmp_lg_f32_e32 vcc, +; GCN-DAG: v_cmp_lg_f32_e64 [[CMP1:s\[[0-9]+:[0-9]+\]]] +; GCN: s_or_b64 vcc, vcc, [[CMP1]] +; GCN: v_cndmask_b32_e32 v[[RESULT1:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: v_cndmask_b32_e32 v[[RESULT0:[0-9]+]], {{v[0-9]+}}, {{v[0-9]+}}, vcc +; GCN: buffer_store_dwordx2 v{{\[}}[[RESULT0]]:[[RESULT1]]{{\]}} +define void @opt_select_i64_or_cmp_f32(i64 addrspace(1)* %out, float %a, float %b, float %c, i64 %x, i64 %y) #0 { + %fcmp0 = fcmp one float %a, %b + %fcmp1 = fcmp one float %a, %c + %or = or i1 %fcmp0, %fcmp1 + %select = select i1 %or, i64 %x, i64 %y + store i64 %select, i64 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}regression: +; GCN: v_cmp_neq_f32_e64 vcc +; GCN: v_cmp_neq_f32_e64 vcc, s{{[0-9]+}}, 0 +; GCN: v_cmp_ne_u32_e32 vcc, 0, v{{[0-9]+}} + +define void @regression(float addrspace(1)* %out, float %c0, float %c1) #0 { +entry: + %cmp0 = fcmp oeq float %c0, 1.0 + br i1 %cmp0, label %if0, label %endif + +if0: + %cmp1 = fcmp oeq float %c1, 0.0 + br i1 %cmp1, label %if1, label %endif + +if1: + %cmp2 = xor i1 %cmp1, true + br label %endif + +endif: + %tmp0 = phi i1 [ true, %entry ], [ %cmp2, %if1 ], [ false, %if0 ] + %tmp2 = select i1 %tmp0, float 4.0, float 0.0 + store float %tmp2, float addrspace(1)* %out + ret void +} + +attributes #0 = { nounwind }