blob: 85eb163383e143ec4f451bc9b41f165ce471d356 [file] [log] [blame]
Matt Arsenaulta9720c62016-06-20 17:51:32 +00001; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck %s
Matt Arsenault7aad8fd2017-01-24 22:02:15 +00002; RUN: llc -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck %s
Tom Stellard36930802014-12-03 04:08:00 +00003
Matt Arsenaulta9720c62016-06-20 17:51:32 +00004; CHECK-LABEL: {{^}}inline_asm:
Tom Stellard36930802014-12-03 04:08:00 +00005; CHECK: s_endpgm
6; CHECK: s_endpgm
Matt Arsenault3dbeefa2017-03-21 21:39:51 +00007define amdgpu_kernel void @inline_asm(i32 addrspace(1)* %out) {
Tom Stellard36930802014-12-03 04:08:00 +00008entry:
9 store i32 5, i32 addrspace(1)* %out
10 call void asm sideeffect "s_endpgm", ""()
11 ret void
12}
Nicolai Haehnlea61e5a82016-01-06 22:01:04 +000013
Matt Arsenaulta9720c62016-06-20 17:51:32 +000014; CHECK-LABEL: {{^}}inline_asm_shader:
Nicolai Haehnlea61e5a82016-01-06 22:01:04 +000015; CHECK: s_endpgm
16; CHECK: s_endpgm
Nicolai Haehnledf3a20c2016-04-06 19:40:20 +000017define amdgpu_ps void @inline_asm_shader() {
Nicolai Haehnlea61e5a82016-01-06 22:01:04 +000018entry:
19 call void asm sideeffect "s_endpgm", ""()
20 ret void
21}
22
Tom Stellardbc4497b2016-02-12 23:45:29 +000023
24; CHECK: {{^}}branch_on_asm:
25; Make sure inline assembly is treted as divergent.
26; CHECK: s_mov_b32 s{{[0-9]+}}, 0
27; CHECK: s_and_saveexec_b64
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000028define amdgpu_kernel void @branch_on_asm(i32 addrspace(1)* %out) {
Tom Stellardbc4497b2016-02-12 23:45:29 +000029 %zero = call i32 asm "s_mov_b32 $0, 0", "=s"()
30 %cmp = icmp eq i32 %zero, 0
31 br i1 %cmp, label %if, label %endif
32
33if:
34 store i32 0, i32 addrspace(1)* %out
35 br label %endif
36
37endif:
38 ret void
39}
Tom Stellard9f2e00d2016-03-09 16:02:52 +000040
Matt Arsenaulta9720c62016-06-20 17:51:32 +000041; CHECK-LABEL: {{^}}v_cmp_asm:
Tom Stellard9f2e00d2016-03-09 16:02:52 +000042; CHECK: v_mov_b32_e32 [[SRC:v[0-9]+]], s{{[0-9]+}}
Matt Arsenault5d8eb252016-09-30 01:50:20 +000043; CHECK: v_cmp_ne_u32_e64 s{{\[}}[[MASK_LO:[0-9]+]]:[[MASK_HI:[0-9]+]]{{\]}}, 0, [[SRC]]
Tom Stellard9f2e00d2016-03-09 16:02:52 +000044; CHECK-DAG: v_mov_b32_e32 v[[V_LO:[0-9]+]], s[[MASK_LO]]
45; CHECK-DAG: v_mov_b32_e32 v[[V_HI:[0-9]+]], s[[MASK_HI]]
46; CHECK: buffer_store_dwordx2 v{{\[}}[[V_LO]]:[[V_HI]]{{\]}}
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000047define amdgpu_kernel void @v_cmp_asm(i64 addrspace(1)* %out, i32 %in) {
Matt Arsenault5d8eb252016-09-30 01:50:20 +000048 %sgpr = tail call i64 asm "v_cmp_ne_u32_e64 $0, 0, $1", "=s,v"(i32 %in)
Tom Stellard9f2e00d2016-03-09 16:02:52 +000049 store i64 %sgpr, i64 addrspace(1)* %out
50 ret void
51}
Matt Arsenaulta9720c62016-06-20 17:51:32 +000052
53; CHECK-LABEL: {{^}}code_size_inline_asm:
54; CHECK: codeLenInByte = 12
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000055define amdgpu_kernel void @code_size_inline_asm(i32 addrspace(1)* %out) {
Matt Arsenaulta9720c62016-06-20 17:51:32 +000056entry:
57 call void asm sideeffect "v_nop_e64", ""()
58 ret void
59}
60
61; All inlineasm instructions are assumed to be the maximum size
62; CHECK-LABEL: {{^}}code_size_inline_asm_small_inst:
63; CHECK: codeLenInByte = 12
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000064define amdgpu_kernel void @code_size_inline_asm_small_inst(i32 addrspace(1)* %out) {
Matt Arsenaulta9720c62016-06-20 17:51:32 +000065entry:
66 call void asm sideeffect "v_nop_e32", ""()
67 ret void
68}
69
70; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst:
71; CHECK: codeLenInByte = 20
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000072define amdgpu_kernel void @code_size_inline_asm_2_inst(i32 addrspace(1)* %out) {
Matt Arsenaulta9720c62016-06-20 17:51:32 +000073entry:
74 call void asm sideeffect "
75 v_nop_e64
76 v_nop_e64
77 ", ""()
78 ret void
79}
Matt Arsenaultaccddac2016-07-01 23:26:50 +000080
81; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst_extra_newline:
82; CHECK: codeLenInByte = 20
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000083define amdgpu_kernel void @code_size_inline_asm_2_inst_extra_newline(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +000084entry:
85 call void asm sideeffect "
86 v_nop_e64
87
88 v_nop_e64
89 ", ""()
90 ret void
91}
92
93; CHECK-LABEL: {{^}}code_size_inline_asm_0_inst:
94; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +000095define amdgpu_kernel void @code_size_inline_asm_0_inst(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +000096entry:
97 call void asm sideeffect "", ""()
98 ret void
99}
100
101; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment:
102; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000103define amdgpu_kernel void @code_size_inline_asm_1_comment(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000104entry:
105 call void asm sideeffect "; comment", ""()
106 ret void
107}
108
109; CHECK-LABEL: {{^}}code_size_inline_asm_newline_1_comment:
110; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000111define amdgpu_kernel void @code_size_inline_asm_newline_1_comment(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000112entry:
113 call void asm sideeffect "
114; comment", ""()
115 ret void
116}
117
118; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment_newline:
119; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000120define amdgpu_kernel void @code_size_inline_asm_1_comment_newline(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000121entry:
122 call void asm sideeffect "; comment
123", ""()
124 ret void
125}
126
127; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line:
128; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000129define amdgpu_kernel void @code_size_inline_asm_2_comments_line(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000130entry:
131 call void asm sideeffect "; first comment ; second comment", ""()
132 ret void
133}
134
135; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line_nospace:
136; CHECK: codeLenInByte = 4
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000137define amdgpu_kernel void @code_size_inline_asm_2_comments_line_nospace(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000138entry:
139 call void asm sideeffect "; first comment;second comment", ""()
140 ret void
141}
142
143; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments0:
144; CHECK: codeLenInByte = 20
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000145define amdgpu_kernel void @code_size_inline_asm_mixed_comments0(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000146entry:
147 call void asm sideeffect "; comment
148 v_nop_e64 ; inline comment
149; separate comment
150 v_nop_e64
151
152 ; trailing comment
153 ; extra comment
154 ", ""()
155 ret void
156}
157
158; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments1:
159; CHECK: codeLenInByte = 20
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000160define amdgpu_kernel void @code_size_inline_asm_mixed_comments1(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000161entry:
162 call void asm sideeffect "v_nop_e64 ; inline comment
163; separate comment
164 v_nop_e64
165
166 ; trailing comment
167 ; extra comment
168 ", ""()
169 ret void
170}
171
172; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments_operands:
173; CHECK: codeLenInByte = 20
Matt Arsenault3dbeefa2017-03-21 21:39:51 +0000174define amdgpu_kernel void @code_size_inline_asm_mixed_comments_operands(i32 addrspace(1)* %out) {
Matt Arsenaultaccddac2016-07-01 23:26:50 +0000175entry:
176 call void asm sideeffect "; comment
177 v_add_i32_e32 v0, vcc, v1, v2 ; inline comment
178; separate comment
179 v_bfrev_b32_e32 v0, 1
180
181 ; trailing comment
182 ; extra comment
183 ", ""()
184 ret void
185}