blob: 457695e5a0e383609209ea1e43af5d6e3bcc3047 [file] [log] [blame]
Tom Stellard5cd09ad2016-01-05 03:40:16 +00001; RUN: llc -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=SI %s
2; RUN: llc -march=amdgcn -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=CI %s
3; RUN: llc -mtriple=amdgcn--amdhsa -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=CI --check-prefix=GCN-HSA %s
Matt Arsenaultf3c91f52015-09-28 20:54:32 +00004
Matt Arsenault9c47dd52016-02-11 06:02:01 +00005declare i32 @llvm.amdgcn.workitem.id.x() #0
6declare i32 @llvm.amdgcn.workitem.id.y() #0
Tom Stellarde0387202014-03-21 15:51:54 +00007
8; In this test both the pointer and the offset operands to the
9; BUFFER_LOAD instructions end up being stored in vgprs. This
10; requires us to add the pointer and offset together, store the
11; result in the offset operand (vaddr), and then store 0 in an
12; sgpr register pair and use that for the pointer operand
13; (low 64-bits of srsrc).
14
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000015; GCN-LABEL: {{^}}mubuf:
Tom Stellard15834092014-03-21 15:51:57 +000016
Tom Stellard326d6ec2014-11-05 14:50:53 +000017; Make sure we aren't using VGPRs for the source operand of s_mov_b64
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000018; GCN-NOT: s_mov_b64 s[{{[0-9]+:[0-9]+}}], v
Tom Stellard15834092014-03-21 15:51:57 +000019
20; Make sure we aren't using VGPR's for the srsrc operand of BUFFER_LOAD_*
21; instructions
Tom Stellard5cd09ad2016-01-05 03:40:16 +000022; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
23; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
24; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
25; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000026
27define void @mubuf(i32 addrspace(1)* %out, i8 addrspace(1)* %in) #1 {
Tom Stellarde0387202014-03-21 15:51:54 +000028entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +000029 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
30 %tmp1 = call i32 @llvm.amdgcn.workitem.id.y()
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000031 %tmp2 = sext i32 %tmp to i64
32 %tmp3 = sext i32 %tmp1 to i64
Tom Stellarde0387202014-03-21 15:51:54 +000033 br label %loop
34
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000035loop: ; preds = %loop, %entry
36 %tmp4 = phi i64 [ 0, %entry ], [ %tmp5, %loop ]
37 %tmp5 = add i64 %tmp2, %tmp4
38 %tmp6 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp5
39 %tmp7 = load i8, i8 addrspace(1)* %tmp6, align 1
40 %tmp8 = or i64 %tmp5, 1
41 %tmp9 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp8
42 %tmp10 = load i8, i8 addrspace(1)* %tmp9, align 1
43 %tmp11 = add i8 %tmp7, %tmp10
44 %tmp12 = sext i8 %tmp11 to i32
45 store i32 %tmp12, i32 addrspace(1)* %out
46 %tmp13 = icmp slt i64 %tmp5, 10
47 br i1 %tmp13, label %loop, label %done
Tom Stellarde0387202014-03-21 15:51:54 +000048
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000049done: ; preds = %loop
Tom Stellarde0387202014-03-21 15:51:54 +000050 ret void
51}
52
Tom Stellard0c354f22014-04-30 15:31:29 +000053; Test moving an SMRD instruction to the VALU
54
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000055; GCN-LABEL: {{^}}smrd_valu:
Tom Stellard467b5b92016-02-20 00:37:25 +000056; SI: s_movk_i32 [[OFFSET:s[0-9]+]], 0x2ee0
57; GCN: v_readfirstlane_b32 s[[PTR_LO:[0-9]+]], v{{[0-9]+}}
58; GCN: v_readfirstlane_b32 s[[PTR_HI:[0-9]+]], v{{[0-9]+}}
Tom Stellardab1d3a92016-04-12 18:40:43 +000059; SI-NEXT: s_nop
Tom Stellard467b5b92016-02-20 00:37:25 +000060; SI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, [[OFFSET]]
61; CI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, 0xbb8
62; GCN: v_mov_b32_e32 [[V_OUT:v[0-9]+]], [[OUT]]
63; GCN-NOHSA: buffer_store_dword [[V_OUT]]
64; GCN-HSA: flat_store_dword {{.*}}, [[V_OUT]]
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000065define void @smrd_valu(i32 addrspace(2)* addrspace(1)* %in, i32 %a, i32 %b, i32 addrspace(1)* %out) #1 {
Tom Stellard0c354f22014-04-30 15:31:29 +000066entry:
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000067 %tmp = icmp ne i32 %a, 0
68 br i1 %tmp, label %if, label %else
Tom Stellard0c354f22014-04-30 15:31:29 +000069
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000070if: ; preds = %entry
71 %tmp1 = load i32 addrspace(2)*, i32 addrspace(2)* addrspace(1)* %in
Tom Stellard0c354f22014-04-30 15:31:29 +000072 br label %endif
73
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000074else: ; preds = %entry
75 %tmp2 = getelementptr i32 addrspace(2)*, i32 addrspace(2)* addrspace(1)* %in
76 %tmp3 = load i32 addrspace(2)*, i32 addrspace(2)* addrspace(1)* %tmp2
Tom Stellard0c354f22014-04-30 15:31:29 +000077 br label %endif
78
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000079endif: ; preds = %else, %if
80 %tmp4 = phi i32 addrspace(2)* [ %tmp1, %if ], [ %tmp3, %else ]
81 %tmp5 = getelementptr i32, i32 addrspace(2)* %tmp4, i32 3000
82 %tmp6 = load i32, i32 addrspace(2)* %tmp5
83 store i32 %tmp6, i32 addrspace(1)* %out
Tom Stellard0c354f22014-04-30 15:31:29 +000084 ret void
85}
Tom Stellard4c00b522014-05-09 16:42:22 +000086
Matt Arsenault711b3902015-08-07 20:18:34 +000087; Test moving an SMRD with an immediate offset to the VALU
Tom Stellard4c00b522014-05-09 16:42:22 +000088
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000089; GCN-LABEL: {{^}}smrd_valu2:
Tom Stellard5cd09ad2016-01-05 03:40:16 +000090; GCN-NOHSA-NOT: v_add
91; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:16{{$}}
92; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000093define void @smrd_valu2(i32 addrspace(1)* %out, [8 x i32] addrspace(2)* %in) #1 {
Tom Stellard4c00b522014-05-09 16:42:22 +000094entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +000095 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultf3c91f52015-09-28 20:54:32 +000096 %tmp1 = add i32 %tmp, 4
97 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(2)* %in, i32 %tmp, i32 4
98 %tmp3 = load i32, i32 addrspace(2)* %tmp2
99 store i32 %tmp3, i32 addrspace(1)* %out
Tom Stellard4c00b522014-05-09 16:42:22 +0000100 ret void
101}
Tom Stellard745f2ed2014-08-21 20:41:00 +0000102
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000103; Use a big offset that will use the SMRD literal offset on CI
104; GCN-LABEL: {{^}}smrd_valu_ci_offset:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000105; GCN-NOHSA-NOT: v_add
106; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4e20{{$}}
107; GCN-NOHSA-NOT: v_add
108; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
109; GCN-NOHSA: v_add_i32_e32
110; GCN-NOHSA: buffer_store_dword
111; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
Tom Stellard46937ca2016-02-12 17:57:54 +0000112; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], v{{[0-9]+}}
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000113define void @smrd_valu_ci_offset(i32 addrspace(1)* %out, i32 addrspace(2)* %in, i32 %c) #1 {
114entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000115 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000116 %tmp2 = getelementptr i32, i32 addrspace(2)* %in, i32 %tmp
117 %tmp3 = getelementptr i32, i32 addrspace(2)* %tmp2, i32 5000
118 %tmp4 = load i32, i32 addrspace(2)* %tmp3
119 %tmp5 = add i32 %tmp4, %c
120 store i32 %tmp5, i32 addrspace(1)* %out
121 ret void
122}
123
124; GCN-LABEL: {{^}}smrd_valu_ci_offset_x2:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000125; GCN-NOHSA-NOT: v_add
126; GCN-NOHSA: s_mov_b32 [[OFFSET:s[0-9]+]], 0x9c40{{$}}
127; GCN-NOHSA-NOT: v_add
128; GCN-NOHSA: buffer_load_dwordx2 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
129; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
130; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
131; GCN-NOHSA: buffer_store_dwordx2
132; GCN-HSA: flat_load_dwordx2 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000133define void @smrd_valu_ci_offset_x2(i64 addrspace(1)* %out, i64 addrspace(2)* %in, i64 %c) #1 {
134entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000135 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000136 %tmp2 = getelementptr i64, i64 addrspace(2)* %in, i32 %tmp
137 %tmp3 = getelementptr i64, i64 addrspace(2)* %tmp2, i32 5000
138 %tmp4 = load i64, i64 addrspace(2)* %tmp3
139 %tmp5 = or i64 %tmp4, %c
140 store i64 %tmp5, i64 addrspace(1)* %out
141 ret void
142}
143
144; GCN-LABEL: {{^}}smrd_valu_ci_offset_x4:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000145; GCN-NOHSA-NOT: v_add
146; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4d20{{$}}
147; GCN-NOHSA-NOT: v_add
148; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
149; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
150; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
151; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
152; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
153; GCN-NOHSA: buffer_store_dwordx4
154; GCN-HSA: flat_load_dwordx4 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000155define void @smrd_valu_ci_offset_x4(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(2)* %in, <4 x i32> %c) #1 {
156entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000157 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000158 %tmp2 = getelementptr <4 x i32>, <4 x i32> addrspace(2)* %in, i32 %tmp
159 %tmp3 = getelementptr <4 x i32>, <4 x i32> addrspace(2)* %tmp2, i32 1234
160 %tmp4 = load <4 x i32>, <4 x i32> addrspace(2)* %tmp3
161 %tmp5 = or <4 x i32> %tmp4, %c
162 store <4 x i32> %tmp5, <4 x i32> addrspace(1)* %out
163 ret void
164}
165
166; Original scalar load uses SGPR offset on SI and 32-bit literal on
167; CI.
168
169; GCN-LABEL: {{^}}smrd_valu_ci_offset_x8:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000170; GCN-NOHSA-NOT: v_add
171; GCN-NOHSA: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x9a40{{$}}
172; GCN-NOHSA-NOT: v_add
173; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
174; GCN-NOHSA-NOT: v_add
175; GCN-NOHSA: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x9a50{{$}}
176; GCN-NOHSA-NOT: v_add
177; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000178
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000179; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
180; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
181; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
182; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
183; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
184; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
185; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
186; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
187; GCN-NOHSA: buffer_store_dwordx4
188; GCN-NOHSA: buffer_store_dwordx4
189; GCN-HSA: flat_load_dwordx4
190; GCN-HSA: flat_load_dwordx4
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000191define void @smrd_valu_ci_offset_x8(<8 x i32> addrspace(1)* %out, <8 x i32> addrspace(2)* %in, <8 x i32> %c) #1 {
192entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000193 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaulte5d042c2015-09-28 20:54:46 +0000194 %tmp2 = getelementptr <8 x i32>, <8 x i32> addrspace(2)* %in, i32 %tmp
195 %tmp3 = getelementptr <8 x i32>, <8 x i32> addrspace(2)* %tmp2, i32 1234
196 %tmp4 = load <8 x i32>, <8 x i32> addrspace(2)* %tmp3
197 %tmp5 = or <8 x i32> %tmp4, %c
198 store <8 x i32> %tmp5, <8 x i32> addrspace(1)* %out
199 ret void
200}
201
Matt Arsenault73aa8f62015-09-28 20:54:52 +0000202; GCN-LABEL: {{^}}smrd_valu_ci_offset_x16:
203
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000204; GCN-NOHSA-NOT: v_add
205; GCN-NOHSA: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x13480{{$}}
206; GCN-NOHSA-NOT: v_add
207; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
208; GCN-NOHSA-NOT: v_add
209; GCN-NOHSA: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x13490{{$}}
210; GCN-NOHSA-NOT: v_add
211; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
212; GCN-NOHSA-NOT: v_add
213; GCN-NOHSA: s_mov_b32 [[OFFSET2:s[0-9]+]], 0x134a0{{$}}
214; GCN-NOHSA-NOT: v_add
215; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET2]] addr64{{$}}
216; GCN-NOHSA-NOT: v_add
217; GCN-NOHSA: s_mov_b32 [[OFFSET3:s[0-9]+]], 0x134b0{{$}}
218; GCN-NOHSA-NOT: v_add
219; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET3]] addr64{{$}}
Matt Arsenault4d801cd2015-11-24 12:05:03 +0000220
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000221; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
222; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
223; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
224; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
225; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
226; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
227; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
228; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
229; GCN-NOHSA: buffer_store_dwordx4
230; GCN-NOHSA: buffer_store_dwordx4
231; GCN-NOHSA: buffer_store_dwordx4
232; GCN-NOHSA: buffer_store_dwordx4
233
234; GCN-HSA: flat_load_dwordx4
235; GCN-HSA: flat_load_dwordx4
236; GCN-HSA: flat_load_dwordx4
237; GCN-HSA: flat_load_dwordx4
Matt Arsenault4d801cd2015-11-24 12:05:03 +0000238
239; GCN: s_endpgm
Matt Arsenault73aa8f62015-09-28 20:54:52 +0000240define void @smrd_valu_ci_offset_x16(<16 x i32> addrspace(1)* %out, <16 x i32> addrspace(2)* %in, <16 x i32> %c) #1 {
241entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000242 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenault73aa8f62015-09-28 20:54:52 +0000243 %tmp2 = getelementptr <16 x i32>, <16 x i32> addrspace(2)* %in, i32 %tmp
244 %tmp3 = getelementptr <16 x i32>, <16 x i32> addrspace(2)* %tmp2, i32 1234
245 %tmp4 = load <16 x i32>, <16 x i32> addrspace(2)* %tmp3
246 %tmp5 = or <16 x i32> %tmp4, %c
247 store <16 x i32> %tmp5, <16 x i32> addrspace(1)* %out
248 ret void
249}
250
Matt Arsenaultb378f072015-09-28 20:54:38 +0000251; GCN-LABEL: {{^}}smrd_valu2_salu_user:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000252; GCN-NOHSA: buffer_load_dword [[MOVED:v[0-9]+]], v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:16{{$}}
253; GCN-HSA: flat_load_dword [[MOVED:v[0-9]+]], v[{{[0-9+:[0-9]+}}]
Matt Arsenaultb378f072015-09-28 20:54:38 +0000254; GCN: v_add_i32_e32 [[ADD:v[0-9]+]], vcc, s{{[0-9]+}}, [[MOVED]]
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000255; GCN-NOHSA: buffer_store_dword [[ADD]]
Tom Stellard46937ca2016-02-12 17:57:54 +0000256; GCN-HSA: flat_store_dword {{.*}}, [[ADD]]
Matt Arsenaultb378f072015-09-28 20:54:38 +0000257define void @smrd_valu2_salu_user(i32 addrspace(1)* %out, [8 x i32] addrspace(2)* %in, i32 %a) #1 {
258entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000259 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultb378f072015-09-28 20:54:38 +0000260 %tmp1 = add i32 %tmp, 4
261 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(2)* %in, i32 %tmp, i32 4
262 %tmp3 = load i32, i32 addrspace(2)* %tmp2
263 %tmp4 = add i32 %tmp3, %a
264 store i32 %tmp4, i32 addrspace(1)* %out
265 ret void
266}
267
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000268; GCN-LABEL: {{^}}smrd_valu2_max_smrd_offset:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000269; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1020{{$}}
270; GCN-HSA flat_load_dword v{{[0-9]}}, v{{[0-9]+:[0-9]+}}
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000271define void @smrd_valu2_max_smrd_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(2)* %in) #1 {
Matt Arsenault711b3902015-08-07 20:18:34 +0000272entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000273 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000274 %tmp1 = add i32 %tmp, 4
275 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(2)* %in, i32 %tmp, i32 255
276 %tmp3 = load i32, i32 addrspace(2)* %tmp2
277 store i32 %tmp3, i32 addrspace(1)* %out
Matt Arsenault711b3902015-08-07 20:18:34 +0000278 ret void
279}
280
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000281; GCN-LABEL: {{^}}smrd_valu2_mubuf_offset:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000282; GCN-NOHSA-NOT: v_add
283; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1024{{$}}
284; GCN-HSA: flat_load_dword v{{[0-9]}}, v[{{[0-9]+:[0-9]+}}]
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000285define void @smrd_valu2_mubuf_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(2)* %in) #1 {
Matt Arsenault711b3902015-08-07 20:18:34 +0000286entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000287 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000288 %tmp1 = add i32 %tmp, 4
289 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(2)* %in, i32 %tmp, i32 256
290 %tmp3 = load i32, i32 addrspace(2)* %tmp2
291 store i32 %tmp3, i32 addrspace(1)* %out
Matt Arsenault711b3902015-08-07 20:18:34 +0000292 ret void
293}
294
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000295; GCN-LABEL: {{^}}s_load_imm_v8i32:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000296; GCN-NOHSA: buffer_load_dwordx4
297; GCN-NOHSA: buffer_load_dwordx4
298; GCN-HSA: flat_load_dwordx4
299; GCN-HSA: flat_load_dwordx4
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000300define void @s_load_imm_v8i32(<8 x i32> addrspace(1)* %out, i32 addrspace(2)* nocapture readonly %in) #1 {
Tom Stellard745f2ed2014-08-21 20:41:00 +0000301entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000302 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
David Blaikie79e6c742015-02-27 19:29:02 +0000303 %tmp1 = getelementptr inbounds i32, i32 addrspace(2)* %in, i32 %tmp0
Tom Stellard745f2ed2014-08-21 20:41:00 +0000304 %tmp2 = bitcast i32 addrspace(2)* %tmp1 to <8 x i32> addrspace(2)*
David Blaikiea79ac142015-02-27 21:17:42 +0000305 %tmp3 = load <8 x i32>, <8 x i32> addrspace(2)* %tmp2, align 4
Tom Stellard745f2ed2014-08-21 20:41:00 +0000306 store <8 x i32> %tmp3, <8 x i32> addrspace(1)* %out, align 32
307 ret void
308}
309
Matt Arsenaultb378f072015-09-28 20:54:38 +0000310; GCN-LABEL: {{^}}s_load_imm_v8i32_salu_user:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000311; GCN-NOHSA: buffer_load_dwordx4
312; GCN-NOHSA: buffer_load_dwordx4
313; GCN-NOHSA: v_add_i32_e32
314; GCN-NOHSA: v_add_i32_e32
315; GCN-NOHSA: v_add_i32_e32
316; GCN-NOHSA: v_add_i32_e32
317; GCN-NOHSA: v_add_i32_e32
318; GCN-NOHSA: v_add_i32_e32
319; GCN-NOHSA: v_add_i32_e32
320; GCN-NOHSA: buffer_store_dword
321; GCN-HSA: flat_load_dwordx4
322; GCN-HSA: flat_load_dwordx4
Matt Arsenaultb378f072015-09-28 20:54:38 +0000323define void @s_load_imm_v8i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(2)* nocapture readonly %in) #1 {
324entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000325 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultb378f072015-09-28 20:54:38 +0000326 %tmp1 = getelementptr inbounds i32, i32 addrspace(2)* %in, i32 %tmp0
327 %tmp2 = bitcast i32 addrspace(2)* %tmp1 to <8 x i32> addrspace(2)*
328 %tmp3 = load <8 x i32>, <8 x i32> addrspace(2)* %tmp2, align 4
329
330 %elt0 = extractelement <8 x i32> %tmp3, i32 0
331 %elt1 = extractelement <8 x i32> %tmp3, i32 1
332 %elt2 = extractelement <8 x i32> %tmp3, i32 2
333 %elt3 = extractelement <8 x i32> %tmp3, i32 3
334 %elt4 = extractelement <8 x i32> %tmp3, i32 4
335 %elt5 = extractelement <8 x i32> %tmp3, i32 5
336 %elt6 = extractelement <8 x i32> %tmp3, i32 6
337 %elt7 = extractelement <8 x i32> %tmp3, i32 7
338
339 %add0 = add i32 %elt0, %elt1
340 %add1 = add i32 %add0, %elt2
341 %add2 = add i32 %add1, %elt3
342 %add3 = add i32 %add2, %elt4
343 %add4 = add i32 %add3, %elt5
344 %add5 = add i32 %add4, %elt6
345 %add6 = add i32 %add5, %elt7
346
347 store i32 %add6, i32 addrspace(1)* %out
348 ret void
349}
350
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000351; GCN-LABEL: {{^}}s_load_imm_v16i32:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000352; GCN-NOHSA: buffer_load_dwordx4
353; GCN-NOHSA: buffer_load_dwordx4
354; GCN-NOHSA: buffer_load_dwordx4
355; GCN-NOHSA: buffer_load_dwordx4
356; GCN-HSA: flat_load_dwordx4
357; GCN-HSA: flat_load_dwordx4
358; GCN-HSA: flat_load_dwordx4
359; GCN-HSA: flat_load_dwordx4
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000360define void @s_load_imm_v16i32(<16 x i32> addrspace(1)* %out, i32 addrspace(2)* nocapture readonly %in) #1 {
Tom Stellard745f2ed2014-08-21 20:41:00 +0000361entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000362 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
David Blaikie79e6c742015-02-27 19:29:02 +0000363 %tmp1 = getelementptr inbounds i32, i32 addrspace(2)* %in, i32 %tmp0
Tom Stellard745f2ed2014-08-21 20:41:00 +0000364 %tmp2 = bitcast i32 addrspace(2)* %tmp1 to <16 x i32> addrspace(2)*
David Blaikiea79ac142015-02-27 21:17:42 +0000365 %tmp3 = load <16 x i32>, <16 x i32> addrspace(2)* %tmp2, align 4
Tom Stellard745f2ed2014-08-21 20:41:00 +0000366 store <16 x i32> %tmp3, <16 x i32> addrspace(1)* %out, align 32
367 ret void
368}
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000369
Matt Arsenaultb378f072015-09-28 20:54:38 +0000370; GCN-LABEL: {{^}}s_load_imm_v16i32_salu_user:
Tom Stellard5cd09ad2016-01-05 03:40:16 +0000371; GCN-NOHSA: buffer_load_dwordx4
372; GCN-NOHSA: buffer_load_dwordx4
373; GCN-NOHSA: buffer_load_dwordx4
374; GCN-NOHSA: buffer_load_dwordx4
375; GCN-NOHSA: v_add_i32_e32
376; GCN-NOHSA: v_add_i32_e32
377; GCN-NOHSA: v_add_i32_e32
378; GCN-NOHSA: v_add_i32_e32
379; GCN-NOHSA: v_add_i32_e32
380; GCN-NOHSA: v_add_i32_e32
381; GCN-NOHSA: v_add_i32_e32
382; GCN-NOHSA: v_add_i32_e32
383; GCN-NOHSA: v_add_i32_e32
384; GCN-NOHSA: v_add_i32_e32
385; GCN-NOHSA: v_add_i32_e32
386; GCN-NOHSA: v_add_i32_e32
387; GCN-NOHSA: v_add_i32_e32
388; GCN-NOHSA: v_add_i32_e32
389; GCN-NOHSA: v_add_i32_e32
390; GCN-NOHSA: buffer_store_dword
391; GCN-HSA: flat_load_dwordx4
392; GCN-HSA: flat_load_dwordx4
393; GCN-HSA: flat_load_dwordx4
394; GCN-HSA: flat_load_dwordx4
Matt Arsenaultb378f072015-09-28 20:54:38 +0000395define void @s_load_imm_v16i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(2)* nocapture readonly %in) #1 {
396entry:
Matt Arsenault9c47dd52016-02-11 06:02:01 +0000397 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
Matt Arsenaultb378f072015-09-28 20:54:38 +0000398 %tmp1 = getelementptr inbounds i32, i32 addrspace(2)* %in, i32 %tmp0
399 %tmp2 = bitcast i32 addrspace(2)* %tmp1 to <16 x i32> addrspace(2)*
400 %tmp3 = load <16 x i32>, <16 x i32> addrspace(2)* %tmp2, align 4
401
402 %elt0 = extractelement <16 x i32> %tmp3, i32 0
403 %elt1 = extractelement <16 x i32> %tmp3, i32 1
404 %elt2 = extractelement <16 x i32> %tmp3, i32 2
405 %elt3 = extractelement <16 x i32> %tmp3, i32 3
406 %elt4 = extractelement <16 x i32> %tmp3, i32 4
407 %elt5 = extractelement <16 x i32> %tmp3, i32 5
408 %elt6 = extractelement <16 x i32> %tmp3, i32 6
409 %elt7 = extractelement <16 x i32> %tmp3, i32 7
410 %elt8 = extractelement <16 x i32> %tmp3, i32 8
411 %elt9 = extractelement <16 x i32> %tmp3, i32 9
412 %elt10 = extractelement <16 x i32> %tmp3, i32 10
413 %elt11 = extractelement <16 x i32> %tmp3, i32 11
414 %elt12 = extractelement <16 x i32> %tmp3, i32 12
415 %elt13 = extractelement <16 x i32> %tmp3, i32 13
416 %elt14 = extractelement <16 x i32> %tmp3, i32 14
417 %elt15 = extractelement <16 x i32> %tmp3, i32 15
418
419 %add0 = add i32 %elt0, %elt1
420 %add1 = add i32 %add0, %elt2
421 %add2 = add i32 %add1, %elt3
422 %add3 = add i32 %add2, %elt4
423 %add4 = add i32 %add3, %elt5
424 %add5 = add i32 %add4, %elt6
425 %add6 = add i32 %add5, %elt7
426 %add7 = add i32 %add6, %elt8
427 %add8 = add i32 %add7, %elt9
428 %add9 = add i32 %add8, %elt10
429 %add10 = add i32 %add9, %elt11
430 %add11 = add i32 %add10, %elt12
431 %add12 = add i32 %add11, %elt13
432 %add13 = add i32 %add12, %elt14
433 %add14 = add i32 %add13, %elt15
434
435 store i32 %add14, i32 addrspace(1)* %out
436 ret void
437}
438
Tom Stellardbc4497b2016-02-12 23:45:29 +0000439; Make sure we legalize vopc operands after moving an sopc to the value.
440
441; {{^}}sopc_vopc_legalize_bug:
442; GCN: s_load_dword [[SGPR:s[0-9]+]]
443; GCN: v_cmp_le_u32_e32 vcc, [[SGPR]], v{{[0-9]+}}
444; GCN: s_and_b64 vcc, exec, vcc
445; GCN: s_cbranch_vccnz [[EXIT:[A-Z0-9_]+]]
446; GCN: v_mov_b32_e32 [[ONE:v[0-9]+]], 1
447; GCN-NOHSA: buffer_store_dword [[ONE]]
448; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], [[ONE]]
449; GCN; {{^}}[[EXIT]]:
450; GCN: s_endpgm
451define void @sopc_vopc_legalize_bug(i32 %cond, i32 addrspace(1)* %out, i32 addrspace(1)* %in) {
452bb3: ; preds = %bb2
453 %tmp0 = bitcast i32 %cond to float
454 %tmp1 = fadd float %tmp0, 2.500000e-01
455 %tmp2 = bitcast float %tmp1 to i32
456 %tmp3 = icmp ult i32 %tmp2, %cond
457 br i1 %tmp3, label %bb6, label %bb7
458
459bb6:
460 store i32 1, i32 addrspace(1)* %out
461 br label %bb7
462
463bb7: ; preds = %bb3
464 ret void
465}
466
Matt Arsenaultf3c91f52015-09-28 20:54:32 +0000467attributes #0 = { nounwind readnone }
468attributes #1 = { nounwind }