2016-06-21 01:51:32 +08:00
|
|
|
; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck %s
|
2017-01-25 06:02:15 +08:00
|
|
|
; RUN: llc -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck %s
|
2014-12-03 12:08:00 +08:00
|
|
|
|
2016-06-21 01:51:32 +08:00
|
|
|
; CHECK-LABEL: {{^}}inline_asm:
|
2014-12-03 12:08:00 +08:00
|
|
|
; CHECK: s_endpgm
|
|
|
|
; CHECK: s_endpgm
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @inline_asm(i32 addrspace(1)* %out) {
|
2014-12-03 12:08:00 +08:00
|
|
|
entry:
|
|
|
|
store i32 5, i32 addrspace(1)* %out
|
|
|
|
call void asm sideeffect "s_endpgm", ""()
|
|
|
|
ret void
|
|
|
|
}
|
2016-01-07 06:01:04 +08:00
|
|
|
|
2016-06-21 01:51:32 +08:00
|
|
|
; CHECK-LABEL: {{^}}inline_asm_shader:
|
2016-01-07 06:01:04 +08:00
|
|
|
; CHECK: s_endpgm
|
|
|
|
; CHECK: s_endpgm
|
2016-04-07 03:40:20 +08:00
|
|
|
define amdgpu_ps void @inline_asm_shader() {
|
2016-01-07 06:01:04 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "s_endpgm", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2016-02-13 07:45:29 +08:00
|
|
|
|
|
|
|
; CHECK: {{^}}branch_on_asm:
|
|
|
|
; Make sure inline assembly is treted as divergent.
|
|
|
|
; CHECK: s_mov_b32 s{{[0-9]+}}, 0
|
|
|
|
; CHECK: s_and_saveexec_b64
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @branch_on_asm(i32 addrspace(1)* %out) {
|
2016-02-13 07:45:29 +08:00
|
|
|
%zero = call i32 asm "s_mov_b32 $0, 0", "=s"()
|
|
|
|
%cmp = icmp eq i32 %zero, 0
|
|
|
|
br i1 %cmp, label %if, label %endif
|
|
|
|
|
|
|
|
if:
|
|
|
|
store i32 0, i32 addrspace(1)* %out
|
|
|
|
br label %endif
|
|
|
|
|
|
|
|
endif:
|
|
|
|
ret void
|
|
|
|
}
|
2016-03-10 00:02:52 +08:00
|
|
|
|
2016-06-21 01:51:32 +08:00
|
|
|
; CHECK-LABEL: {{^}}v_cmp_asm:
|
2016-03-10 00:02:52 +08:00
|
|
|
; CHECK: v_mov_b32_e32 [[SRC:v[0-9]+]], s{{[0-9]+}}
|
2016-09-30 09:50:20 +08:00
|
|
|
; CHECK: v_cmp_ne_u32_e64 s{{\[}}[[MASK_LO:[0-9]+]]:[[MASK_HI:[0-9]+]]{{\]}}, 0, [[SRC]]
|
2016-03-10 00:02:52 +08:00
|
|
|
; CHECK-DAG: v_mov_b32_e32 v[[V_LO:[0-9]+]], s[[MASK_LO]]
|
|
|
|
; CHECK-DAG: v_mov_b32_e32 v[[V_HI:[0-9]+]], s[[MASK_HI]]
|
|
|
|
; CHECK: buffer_store_dwordx2 v{{\[}}[[V_LO]]:[[V_HI]]{{\]}}
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @v_cmp_asm(i64 addrspace(1)* %out, i32 %in) {
|
2016-09-30 09:50:20 +08:00
|
|
|
%sgpr = tail call i64 asm "v_cmp_ne_u32_e64 $0, 0, $1", "=s,v"(i32 %in)
|
2016-03-10 00:02:52 +08:00
|
|
|
store i64 %sgpr, i64 addrspace(1)* %out
|
|
|
|
ret void
|
|
|
|
}
|
2016-06-21 01:51:32 +08:00
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm:
|
|
|
|
; CHECK: codeLenInByte = 12
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm(i32 addrspace(1)* %out) {
|
2016-06-21 01:51:32 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "v_nop_e64", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; All inlineasm instructions are assumed to be the maximum size
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_small_inst:
|
|
|
|
; CHECK: codeLenInByte = 12
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_small_inst(i32 addrspace(1)* %out) {
|
2016-06-21 01:51:32 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "v_nop_e32", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst:
|
|
|
|
; CHECK: codeLenInByte = 20
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_2_inst(i32 addrspace(1)* %out) {
|
2016-06-21 01:51:32 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "
|
|
|
|
v_nop_e64
|
|
|
|
v_nop_e64
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
2016-07-02 07:26:50 +08:00
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst_extra_newline:
|
|
|
|
; CHECK: codeLenInByte = 20
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_2_inst_extra_newline(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "
|
|
|
|
v_nop_e64
|
|
|
|
|
|
|
|
v_nop_e64
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_0_inst:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_0_inst(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_1_comment(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; comment", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_newline_1_comment:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_newline_1_comment(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "
|
|
|
|
; comment", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment_newline:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_1_comment_newline(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; comment
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_2_comments_line(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; first comment ; second comment", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line_nospace:
|
|
|
|
; CHECK: codeLenInByte = 4
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_2_comments_line_nospace(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; first comment;second comment", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments0:
|
|
|
|
; CHECK: codeLenInByte = 20
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_mixed_comments0(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; comment
|
|
|
|
v_nop_e64 ; inline comment
|
|
|
|
; separate comment
|
|
|
|
v_nop_e64
|
|
|
|
|
|
|
|
; trailing comment
|
|
|
|
; extra comment
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments1:
|
|
|
|
; CHECK: codeLenInByte = 20
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_mixed_comments1(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "v_nop_e64 ; inline comment
|
|
|
|
; separate comment
|
|
|
|
v_nop_e64
|
|
|
|
|
|
|
|
; trailing comment
|
|
|
|
; extra comment
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments_operands:
|
|
|
|
; CHECK: codeLenInByte = 20
|
2017-03-22 05:39:51 +08:00
|
|
|
define amdgpu_kernel void @code_size_inline_asm_mixed_comments_operands(i32 addrspace(1)* %out) {
|
2016-07-02 07:26:50 +08:00
|
|
|
entry:
|
|
|
|
call void asm sideeffect "; comment
|
|
|
|
v_add_i32_e32 v0, vcc, v1, v2 ; inline comment
|
|
|
|
; separate comment
|
|
|
|
v_bfrev_b32_e32 v0, 1
|
|
|
|
|
|
|
|
; trailing comment
|
|
|
|
; extra comment
|
|
|
|
", ""()
|
|
|
|
ret void
|
|
|
|
}
|
2017-04-12 06:29:19 +08:00
|
|
|
|
|
|
|
; FIXME: Should not have intermediate sgprs
|
|
|
|
; CHECK-LABEL: {{^}}i64_imm_input_phys_vgpr:
|
2018-10-30 23:04:40 +08:00
|
|
|
; CHECK-DAG: s_mov_b32 s1, 0
|
|
|
|
; CHECK-DAG: s_mov_b32 s0, 0x1e240
|
2017-04-12 06:29:19 +08:00
|
|
|
; CHECK: v_mov_b32_e32 v0, s0
|
|
|
|
; CHECK: v_mov_b32_e32 v1, s1
|
|
|
|
; CHECK: use v[0:1]
|
2017-05-18 05:56:25 +08:00
|
|
|
define amdgpu_kernel void @i64_imm_input_phys_vgpr() {
|
2017-04-12 06:29:19 +08:00
|
|
|
entry:
|
2017-06-09 03:03:20 +08:00
|
|
|
call void asm sideeffect "; use $0 ", "{v[0:1]}"(i64 123456)
|
2017-04-12 06:29:19 +08:00
|
|
|
ret void
|
|
|
|
}
|
2017-04-13 05:58:23 +08:00
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}i1_imm_input_phys_vgpr:
|
|
|
|
; CHECK: v_mov_b32_e32 v0, -1{{$}}
|
|
|
|
; CHECK: ; use v0
|
|
|
|
define amdgpu_kernel void @i1_imm_input_phys_vgpr() {
|
|
|
|
entry:
|
2017-06-09 03:03:20 +08:00
|
|
|
call void asm sideeffect "; use $0 ", "{v0}"(i1 true)
|
2017-04-13 05:58:23 +08:00
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}i1_input_phys_vgpr:
|
|
|
|
; CHECK: {{buffer|flat}}_load_ubyte [[LOAD:v[0-9]+]]
|
|
|
|
; CHECK: v_and_b32_e32 [[LOAD]], 1, [[LOAD]]
|
|
|
|
; CHECK-NEXT: v_cmp_eq_u32_e32 vcc, 1, [[LOAD]]
|
|
|
|
; CHECK-NEXT: v_cndmask_b32_e64 v0, 0, -1, vcc
|
|
|
|
; CHECK: ; use v0
|
|
|
|
define amdgpu_kernel void @i1_input_phys_vgpr() {
|
|
|
|
entry:
|
|
|
|
%val = load i1, i1 addrspace(1)* undef
|
2017-06-09 03:03:20 +08:00
|
|
|
call void asm sideeffect "; use $0 ", "{v0}"(i1 %val)
|
2017-04-13 05:58:23 +08:00
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; FIXME: Should be scheduled to shrink vcc
|
|
|
|
; CHECK-LABEL: {{^}}i1_input_phys_vgpr_x2:
|
|
|
|
; CHECK: v_cmp_eq_u32_e32 vcc, 1, v0
|
|
|
|
; CHECK: v_cndmask_b32_e64 v0, 0, -1, vcc
|
2017-07-07 04:57:05 +08:00
|
|
|
; CHECK: v_cmp_eq_u32_e32 vcc, 1, v1
|
|
|
|
; CHECK: v_cndmask_b32_e64 v1, 0, -1, vcc
|
2017-04-13 05:58:23 +08:00
|
|
|
define amdgpu_kernel void @i1_input_phys_vgpr_x2() {
|
|
|
|
entry:
|
|
|
|
%val0 = load volatile i1, i1 addrspace(1)* undef
|
|
|
|
%val1 = load volatile i1, i1 addrspace(1)* undef
|
2017-06-09 03:03:20 +08:00
|
|
|
call void asm sideeffect "; use $0 $1 ", "{v0}, {v1}"(i1 %val0, i1 %val1)
|
2017-04-13 05:58:23 +08:00
|
|
|
ret void
|
|
|
|
}
|
2017-04-29 09:26:34 +08:00
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}muliple_def_phys_vgpr:
|
|
|
|
; CHECK: ; def v0
|
|
|
|
; CHECK: v_mov_b32_e32 v1, v0
|
|
|
|
; CHECK: ; def v0
|
|
|
|
; CHECK: v_lshlrev_b32_e32 v{{[0-9]+}}, v0, v1
|
|
|
|
define amdgpu_kernel void @muliple_def_phys_vgpr() {
|
|
|
|
entry:
|
2017-06-09 03:03:20 +08:00
|
|
|
%def0 = call i32 asm sideeffect "; def $0 ", "={v0}"()
|
|
|
|
%def1 = call i32 asm sideeffect "; def $0 ", "={v0}"()
|
2017-04-29 09:26:34 +08:00
|
|
|
%add = shl i32 %def0, %def1
|
|
|
|
store i32 %add, i32 addrspace(1)* undef
|
|
|
|
ret void
|
|
|
|
}
|
2017-08-10 04:09:35 +08:00
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}asm_constraint_c_n:
|
|
|
|
; CHECK: s_trap 10{{$}}
|
|
|
|
define amdgpu_kernel void @asm_constraint_c_n() {
|
|
|
|
entry:
|
|
|
|
tail call void asm sideeffect "s_trap ${0:c}", "n"(i32 10) #1
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
; CHECK-LABEL: {{^}}asm_constraint_n_n:
|
|
|
|
; CHECK: s_trap -10{{$}}
|
|
|
|
define amdgpu_kernel void @asm_constraint_n_n() {
|
|
|
|
entry:
|
|
|
|
tail call void asm sideeffect "s_trap ${0:n}", "n"(i32 10) #1
|
|
|
|
ret void
|
|
|
|
}
|