1 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=SI %s
2 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=CI %s
3 ; RUN: llc -amdgpu-scalarize-global-loads=false -mtriple=amdgcn--amdhsa -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=CI --check-prefix=GCN-HSA %s
5 declare i32 @llvm.amdgcn.workitem.id.x() #0
6 declare i32 @llvm.amdgcn.workitem.id.y() #0
8 ; In this test both the pointer and the offset operands to the
9 ; BUFFER_LOAD instructions end up being stored in vgprs. This
10 ; requires us to add the pointer and offset together, store the
11 ; result in the offset operand (vaddr), and then store 0 in an
12 ; sgpr register pair and use that for the pointer operand
13 ; (low 64-bits of srsrc).
15 ; GCN-LABEL: {{^}}mubuf:
17 ; Make sure we aren't using VGPRs for the source operand of s_mov_b64
18 ; GCN-NOT: s_mov_b64 s[{{[0-9]+:[0-9]+}}], v
20 ; Make sure we aren't using VGPR's for the srsrc operand of BUFFER_LOAD_*
22 ; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
23 ; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
24 ; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
25 ; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
27 define amdgpu_kernel void @mubuf(i32 addrspace(1)* %out, i8 addrspace(1)* %in) #1 {
29 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
30 %tmp1 = call i32 @llvm.amdgcn.workitem.id.y()
31 %tmp2 = sext i32 %tmp to i64
32 %tmp3 = sext i32 %tmp1 to i64
35 loop: ; preds = %loop, %entry
36 %tmp4 = phi i64 [ 0, %entry ], [ %tmp5, %loop ]
37 %tmp5 = add i64 %tmp2, %tmp4
38 %tmp6 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp5
39 %tmp7 = load i8, i8 addrspace(1)* %tmp6, align 1
40 %tmp8 = or i64 %tmp5, 1
41 %tmp9 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp8
42 %tmp10 = load i8, i8 addrspace(1)* %tmp9, align 1
43 %tmp11 = add i8 %tmp7, %tmp10
44 %tmp12 = sext i8 %tmp11 to i32
45 store i32 %tmp12, i32 addrspace(1)* %out
46 %tmp13 = icmp slt i64 %tmp5, 10
47 br i1 %tmp13, label %loop, label %done
53 ; Test moving an SMRD instruction to the VALU
54 ; FIXME: movs can be moved before nop to reduce count
56 ; GCN-LABEL: {{^}}smrd_valu:
57 ; SI: s_movk_i32 [[OFFSET:s[0-9]+]], 0x2ee0
59 ; GCN: v_readfirstlane_b32 s[[PTR_LO:[0-9]+]], v{{[0-9]+}}
60 ; GCN: v_readfirstlane_b32 s[[PTR_HI:[0-9]+]], v{{[0-9]+}}
62 ; SI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, [[OFFSET]]
64 ; CI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, 0xbb8
65 ; GCN: v_mov_b32_e32 [[V_OUT:v[0-9]+]], [[OUT]]
66 ; GCN-NOHSA: buffer_store_dword [[V_OUT]]
67 ; GCN-HSA: flat_store_dword {{.*}}, [[V_OUT]]
68 define amdgpu_kernel void @smrd_valu(i32 addrspace(4)* addrspace(1)* %in, i32 %a, i32 %b, i32 addrspace(1)* %out) #1 {
70 %tmp = icmp ne i32 %a, 0
71 br i1 %tmp, label %if, label %else
74 %tmp1 = load i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %in
77 else: ; preds = %entry
78 %tmp2 = getelementptr i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %in
79 %tmp3 = load i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %tmp2
82 endif: ; preds = %else, %if
83 %tmp4 = phi i32 addrspace(4)* [ %tmp1, %if ], [ %tmp3, %else ]
84 %tmp5 = getelementptr i32, i32 addrspace(4)* %tmp4, i32 3000
85 %tmp6 = load i32, i32 addrspace(4)* %tmp5
86 store i32 %tmp6, i32 addrspace(1)* %out
90 ; Test moving an SMRD with an immediate offset to the VALU
92 ; GCN-LABEL: {{^}}smrd_valu2:
93 ; GCN-NOHSA-NOT: v_add
94 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:16{{$}}
95 ; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
96 define amdgpu_kernel void @smrd_valu2(i32 addrspace(1)* %out, [8 x i32] addrspace(4)* %in) #1 {
98 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
99 %tmp1 = add i32 %tmp, 4
100 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(4)* %in, i32 %tmp, i32 4
101 %tmp3 = load i32, i32 addrspace(4)* %tmp2
102 store i32 %tmp3, i32 addrspace(1)* %out
106 ; Use a big offset that will use the SMRD literal offset on CI
107 ; GCN-LABEL: {{^}}smrd_valu_ci_offset:
108 ; GCN-NOHSA-NOT: v_add
109 ; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4e20{{$}}
110 ; GCN-NOHSA-NOT: v_add
111 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
112 ; GCN-NOHSA: v_add_i32_e32
113 ; GCN-NOHSA: buffer_store_dword
114 ; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
115 ; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], v{{[0-9]+}}
116 define amdgpu_kernel void @smrd_valu_ci_offset(i32 addrspace(1)* %out, i32 addrspace(4)* %in, i32 %c) #1 {
118 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
119 %tmp2 = getelementptr i32, i32 addrspace(4)* %in, i32 %tmp
120 %tmp3 = getelementptr i32, i32 addrspace(4)* %tmp2, i32 5000
121 %tmp4 = load i32, i32 addrspace(4)* %tmp3
122 %tmp5 = add i32 %tmp4, %c
123 store i32 %tmp5, i32 addrspace(1)* %out
127 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x2:
128 ; GCN-NOHSA-NOT: v_add
129 ; GCN-NOHSA: s_mov_b32 [[OFFSET:s[0-9]+]], 0x9c40{{$}}
130 ; GCN-NOHSA-NOT: v_add
131 ; GCN-NOHSA: buffer_load_dwordx2 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
132 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
133 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
134 ; GCN-NOHSA: buffer_store_dwordx2
135 ; GCN-HSA: flat_load_dwordx2 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
136 define amdgpu_kernel void @smrd_valu_ci_offset_x2(i64 addrspace(1)* %out, i64 addrspace(4)* %in, i64 %c) #1 {
138 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
139 %tmp2 = getelementptr i64, i64 addrspace(4)* %in, i32 %tmp
140 %tmp3 = getelementptr i64, i64 addrspace(4)* %tmp2, i32 5000
141 %tmp4 = load i64, i64 addrspace(4)* %tmp3
142 %tmp5 = or i64 %tmp4, %c
143 store i64 %tmp5, i64 addrspace(1)* %out
147 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x4:
148 ; GCN-NOHSA-NOT: v_add
149 ; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4d20{{$}}
150 ; GCN-NOHSA-NOT: v_add
151 ; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
152 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
153 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
154 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
155 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
156 ; GCN-NOHSA: buffer_store_dwordx4
157 ; GCN-HSA: flat_load_dwordx4 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
158 define amdgpu_kernel void @smrd_valu_ci_offset_x4(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(4)* %in, <4 x i32> %c) #1 {
160 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
161 %tmp2 = getelementptr <4 x i32>, <4 x i32> addrspace(4)* %in, i32 %tmp
162 %tmp3 = getelementptr <4 x i32>, <4 x i32> addrspace(4)* %tmp2, i32 1234
163 %tmp4 = load <4 x i32>, <4 x i32> addrspace(4)* %tmp3
164 %tmp5 = or <4 x i32> %tmp4, %c
165 store <4 x i32> %tmp5, <4 x i32> addrspace(1)* %out
169 ; Original scalar load uses SGPR offset on SI and 32-bit literal on
172 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x8:
173 ; GCN-NOHSA: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x9a40{{$}}
174 ; GCN-NOHSA-NOT: v_add
175 ; GCN-NOHSA: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x9a50{{$}}
176 ; GCN-NOHSA-NOT: v_add
177 ; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
178 ; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
180 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
181 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
182 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
183 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
184 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
185 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
186 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
187 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
188 ; GCN-NOHSA: buffer_store_dwordx4
189 ; GCN-NOHSA: buffer_store_dwordx4
190 ; GCN-HSA: flat_load_dwordx4
191 ; GCN-HSA: flat_load_dwordx4
192 define amdgpu_kernel void @smrd_valu_ci_offset_x8(<8 x i32> addrspace(1)* %out, <8 x i32> addrspace(4)* %in, <8 x i32> %c) #1 {
194 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
195 %tmp2 = getelementptr <8 x i32>, <8 x i32> addrspace(4)* %in, i32 %tmp
196 %tmp3 = getelementptr <8 x i32>, <8 x i32> addrspace(4)* %tmp2, i32 1234
197 %tmp4 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp3
198 %tmp5 = or <8 x i32> %tmp4, %c
199 store <8 x i32> %tmp5, <8 x i32> addrspace(1)* %out
203 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x16:
205 ; GCN-NOHSA-DAG: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x13480{{$}}
206 ; GCN-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
207 ; GCN-NOHSA-DAG: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x13490{{$}}
208 ; GCN-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
209 ; GCN-NOHSA-DAG: s_mov_b32 [[OFFSET2:s[0-9]+]], 0x134a0{{$}}
210 ; GCN-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET2]] addr64{{$}}
211 ; GCN-NOHSA-DAG: s_mov_b32 [[OFFSET3:s[0-9]+]], 0x134b0{{$}}
212 ; GCN-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET3]] addr64{{$}}
214 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
215 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
216 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
217 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
218 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
219 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
220 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
221 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
222 ; GCN-NOHSA: buffer_store_dwordx4
223 ; GCN-NOHSA: buffer_store_dwordx4
224 ; GCN-NOHSA: buffer_store_dwordx4
225 ; GCN-NOHSA: buffer_store_dwordx4
227 ; GCN-HSA: flat_load_dwordx4
228 ; GCN-HSA: flat_load_dwordx4
229 ; GCN-HSA: flat_load_dwordx4
230 ; GCN-HSA: flat_load_dwordx4
233 define amdgpu_kernel void @smrd_valu_ci_offset_x16(<16 x i32> addrspace(1)* %out, <16 x i32> addrspace(4)* %in, <16 x i32> %c) #1 {
235 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
236 %tmp2 = getelementptr <16 x i32>, <16 x i32> addrspace(4)* %in, i32 %tmp
237 %tmp3 = getelementptr <16 x i32>, <16 x i32> addrspace(4)* %tmp2, i32 1234
238 %tmp4 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp3
239 %tmp5 = or <16 x i32> %tmp4, %c
240 store <16 x i32> %tmp5, <16 x i32> addrspace(1)* %out
244 ; GCN-LABEL: {{^}}smrd_valu2_salu_user:
245 ; GCN-NOHSA: buffer_load_dword [[MOVED:v[0-9]+]], v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:16{{$}}
246 ; GCN-HSA: flat_load_dword [[MOVED:v[0-9]+]], v[{{[0-9+:[0-9]+}}]
247 ; GCN: v_add_i32_e32 [[ADD:v[0-9]+]], vcc, s{{[0-9]+}}, [[MOVED]]
248 ; GCN-NOHSA: buffer_store_dword [[ADD]]
249 ; GCN-HSA: flat_store_dword {{.*}}, [[ADD]]
250 define amdgpu_kernel void @smrd_valu2_salu_user(i32 addrspace(1)* %out, [8 x i32] addrspace(4)* %in, i32 %a) #1 {
252 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
253 %tmp1 = add i32 %tmp, 4
254 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(4)* %in, i32 %tmp, i32 4
255 %tmp3 = load i32, i32 addrspace(4)* %tmp2
256 %tmp4 = add i32 %tmp3, %a
257 store i32 %tmp4, i32 addrspace(1)* %out
261 ; GCN-LABEL: {{^}}smrd_valu2_max_smrd_offset:
262 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1020{{$}}
263 ; GCN-HSA flat_load_dword v{{[0-9]}}, v{{[0-9]+:[0-9]+}}
264 define amdgpu_kernel void @smrd_valu2_max_smrd_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(4)* %in) #1 {
266 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
267 %tmp1 = add i32 %tmp, 4
268 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(4)* %in, i32 %tmp, i32 255
269 %tmp3 = load i32, i32 addrspace(4)* %tmp2
270 store i32 %tmp3, i32 addrspace(1)* %out
274 ; GCN-LABEL: {{^}}smrd_valu2_mubuf_offset:
275 ; GCN-NOHSA-NOT: v_add
276 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1024{{$}}
277 ; GCN-HSA: flat_load_dword v{{[0-9]}}, v[{{[0-9]+:[0-9]+}}]
278 define amdgpu_kernel void @smrd_valu2_mubuf_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(4)* %in) #1 {
280 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
281 %tmp1 = add i32 %tmp, 4
282 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(4)* %in, i32 %tmp, i32 256
283 %tmp3 = load i32, i32 addrspace(4)* %tmp2
284 store i32 %tmp3, i32 addrspace(1)* %out
288 ; GCN-LABEL: {{^}}s_load_imm_v8i32:
289 ; GCN-NOHSA: buffer_load_dwordx4
290 ; GCN-NOHSA: buffer_load_dwordx4
291 ; GCN-HSA: flat_load_dwordx4
292 ; GCN-HSA: flat_load_dwordx4
293 define amdgpu_kernel void @s_load_imm_v8i32(<8 x i32> addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
295 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
296 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
297 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <8 x i32> addrspace(4)*
298 %tmp3 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp2, align 4
299 store <8 x i32> %tmp3, <8 x i32> addrspace(1)* %out, align 32
303 ; GCN-LABEL: {{^}}s_load_imm_v8i32_salu_user:
304 ; GCN-NOHSA: buffer_load_dwordx4
305 ; GCN-NOHSA: buffer_load_dwordx4
306 ; GCN-NOHSA: v_add_i32_e32
307 ; GCN-NOHSA: v_add_i32_e32
308 ; GCN-NOHSA: v_add_i32_e32
309 ; GCN-NOHSA: v_add_i32_e32
310 ; GCN-NOHSA: v_add_i32_e32
311 ; GCN-NOHSA: v_add_i32_e32
312 ; GCN-NOHSA: v_add_i32_e32
313 ; GCN-NOHSA: buffer_store_dword
314 ; GCN-HSA: flat_load_dwordx4
315 ; GCN-HSA: flat_load_dwordx4
316 define amdgpu_kernel void @s_load_imm_v8i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
318 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
319 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
320 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <8 x i32> addrspace(4)*
321 %tmp3 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp2, align 4
323 %elt0 = extractelement <8 x i32> %tmp3, i32 0
324 %elt1 = extractelement <8 x i32> %tmp3, i32 1
325 %elt2 = extractelement <8 x i32> %tmp3, i32 2
326 %elt3 = extractelement <8 x i32> %tmp3, i32 3
327 %elt4 = extractelement <8 x i32> %tmp3, i32 4
328 %elt5 = extractelement <8 x i32> %tmp3, i32 5
329 %elt6 = extractelement <8 x i32> %tmp3, i32 6
330 %elt7 = extractelement <8 x i32> %tmp3, i32 7
332 %add0 = add i32 %elt0, %elt1
333 %add1 = add i32 %add0, %elt2
334 %add2 = add i32 %add1, %elt3
335 %add3 = add i32 %add2, %elt4
336 %add4 = add i32 %add3, %elt5
337 %add5 = add i32 %add4, %elt6
338 %add6 = add i32 %add5, %elt7
340 store i32 %add6, i32 addrspace(1)* %out
344 ; GCN-LABEL: {{^}}s_load_imm_v16i32:
345 ; GCN-NOHSA: buffer_load_dwordx4
346 ; GCN-NOHSA: buffer_load_dwordx4
347 ; GCN-NOHSA: buffer_load_dwordx4
348 ; GCN-NOHSA: buffer_load_dwordx4
349 ; GCN-HSA: flat_load_dwordx4
350 ; GCN-HSA: flat_load_dwordx4
351 ; GCN-HSA: flat_load_dwordx4
352 ; GCN-HSA: flat_load_dwordx4
353 define amdgpu_kernel void @s_load_imm_v16i32(<16 x i32> addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
355 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
356 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
357 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <16 x i32> addrspace(4)*
358 %tmp3 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp2, align 4
359 store <16 x i32> %tmp3, <16 x i32> addrspace(1)* %out, align 32
363 ; GCN-LABEL: {{^}}s_load_imm_v16i32_salu_user:
364 ; GCN-NOHSA: buffer_load_dwordx4
365 ; GCN-NOHSA: buffer_load_dwordx4
366 ; GCN-NOHSA: buffer_load_dwordx4
367 ; GCN-NOHSA: buffer_load_dwordx4
368 ; GCN-NOHSA: v_add_i32_e32
369 ; GCN-NOHSA: v_add_i32_e32
370 ; GCN-NOHSA: v_add_i32_e32
371 ; GCN-NOHSA: v_add_i32_e32
372 ; GCN-NOHSA: v_add_i32_e32
373 ; GCN-NOHSA: v_add_i32_e32
374 ; GCN-NOHSA: v_add_i32_e32
375 ; GCN-NOHSA: v_add_i32_e32
376 ; GCN-NOHSA: v_add_i32_e32
377 ; GCN-NOHSA: v_add_i32_e32
378 ; GCN-NOHSA: v_add_i32_e32
379 ; GCN-NOHSA: v_add_i32_e32
380 ; GCN-NOHSA: v_add_i32_e32
381 ; GCN-NOHSA: v_add_i32_e32
382 ; GCN-NOHSA: v_add_i32_e32
383 ; GCN-NOHSA: buffer_store_dword
384 ; GCN-HSA: flat_load_dwordx4
385 ; GCN-HSA: flat_load_dwordx4
386 ; GCN-HSA: flat_load_dwordx4
387 ; GCN-HSA: flat_load_dwordx4
388 define amdgpu_kernel void @s_load_imm_v16i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
390 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
391 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
392 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <16 x i32> addrspace(4)*
393 %tmp3 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp2, align 4
395 %elt0 = extractelement <16 x i32> %tmp3, i32 0
396 %elt1 = extractelement <16 x i32> %tmp3, i32 1
397 %elt2 = extractelement <16 x i32> %tmp3, i32 2
398 %elt3 = extractelement <16 x i32> %tmp3, i32 3
399 %elt4 = extractelement <16 x i32> %tmp3, i32 4
400 %elt5 = extractelement <16 x i32> %tmp3, i32 5
401 %elt6 = extractelement <16 x i32> %tmp3, i32 6
402 %elt7 = extractelement <16 x i32> %tmp3, i32 7
403 %elt8 = extractelement <16 x i32> %tmp3, i32 8
404 %elt9 = extractelement <16 x i32> %tmp3, i32 9
405 %elt10 = extractelement <16 x i32> %tmp3, i32 10
406 %elt11 = extractelement <16 x i32> %tmp3, i32 11
407 %elt12 = extractelement <16 x i32> %tmp3, i32 12
408 %elt13 = extractelement <16 x i32> %tmp3, i32 13
409 %elt14 = extractelement <16 x i32> %tmp3, i32 14
410 %elt15 = extractelement <16 x i32> %tmp3, i32 15
412 %add0 = add i32 %elt0, %elt1
413 %add1 = add i32 %add0, %elt2
414 %add2 = add i32 %add1, %elt3
415 %add3 = add i32 %add2, %elt4
416 %add4 = add i32 %add3, %elt5
417 %add5 = add i32 %add4, %elt6
418 %add6 = add i32 %add5, %elt7
419 %add7 = add i32 %add6, %elt8
420 %add8 = add i32 %add7, %elt9
421 %add9 = add i32 %add8, %elt10
422 %add10 = add i32 %add9, %elt11
423 %add11 = add i32 %add10, %elt12
424 %add12 = add i32 %add11, %elt13
425 %add13 = add i32 %add12, %elt14
426 %add14 = add i32 %add13, %elt15
428 store i32 %add14, i32 addrspace(1)* %out
432 ; Make sure we legalize vopc operands after moving an sopc to the value.
434 ; {{^}}sopc_vopc_legalize_bug:
435 ; GCN: s_load_dword [[SGPR:s[0-9]+]]
436 ; GCN: v_cmp_le_u32_e32 vcc, [[SGPR]], v{{[0-9]+}}
437 ; GCN: s_and_b64 vcc, exec, vcc
438 ; GCN: s_cbranch_vccnz [[EXIT:[A-Z0-9_]+]]
439 ; GCN: v_mov_b32_e32 [[ONE:v[0-9]+]], 1
440 ; GCN-NOHSA: buffer_store_dword [[ONE]]
441 ; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], [[ONE]]
442 ; GCN: {{^}}[[EXIT]]:
444 define amdgpu_kernel void @sopc_vopc_legalize_bug(i32 %cond, i32 addrspace(1)* %out, i32 addrspace(1)* %in) {
446 %tmp0 = bitcast i32 %cond to float
447 %tmp1 = fadd float %tmp0, 2.500000e-01
448 %tmp2 = bitcast float %tmp1 to i32
449 %tmp3 = icmp ult i32 %tmp2, %cond
450 br i1 %tmp3, label %bb6, label %bb7
453 store i32 1, i32 addrspace(1)* %out
460 ; GCN-LABEL: {{^}}phi_visit_order:
461 ; GCN: v_add_i32_e32 v{{[0-9]+}}, vcc, 1, v{{[0-9]+}}
462 define amdgpu_kernel void @phi_visit_order() {
467 %tmp = phi i32 [ 0, %bb ], [ %tmp5, %bb4 ]
468 %tid = call i32 @llvm.amdgcn.workitem.id.x()
469 %cnd = icmp eq i32 %tid, 0
470 br i1 %cnd, label %bb4, label %bb2
473 %tmp3 = add nsw i32 %tmp, 1
477 %tmp5 = phi i32 [ %tmp3, %bb2 ], [ %tmp, %bb1 ]
481 ; GCN-LABEL: {{^}}phi_imm_in_sgprs
482 ; GCN: s_movk_i32 [[A:s[0-9]+]], 0x400
483 ; GCN: s_movk_i32 [[B:s[0-9]+]], 0x400
484 ; GCN: [[LOOP_LABEL:[0-9a-zA-Z_]+]]:
485 ; GCN: s_xor_b32 [[B]], [[B]], [[A]]
486 ; GCN: s_cbranch_scc{{[01]}} [[LOOP_LABEL]]
487 define amdgpu_kernel void @phi_imm_in_sgprs(i32 addrspace(3)* %out, i32 %cond) {
492 %i = phi i32 [0, %entry], [%i.add, %loop]
493 %offset = phi i32 [1024, %entry], [%offset.xor, %loop]
494 %offset.xor = xor i32 %offset, 1024
495 %offset.i = add i32 %offset.xor, %i
496 %ptr = getelementptr i32, i32 addrspace(3)* %out, i32 %offset.i
497 store i32 0, i32 addrspace(3)* %ptr
498 %i.add = add i32 %i, 1
499 %cmp = icmp ult i32 %i.add, %cond
500 br i1 %cmp, label %loop, label %exit
506 attributes #0 = { nounwind readnone }
507 attributes #1 = { nounwind }