1 ; RUN: llc -march=amdgcn -mcpu=tahiti -verify-machineinstrs -amdgpu-s-branch-bits=4 < %s | FileCheck -enable-var-scope -check-prefix=GCN %s
4 ; FIXME: We should use llvm-mc for this, but we can't even parse our own output.
6 ; RUN: llc -march=amdgcn -verify-machineinstrs -amdgpu-s-branch-bits=4 -o %t.o -filetype=obj %s
7 ; RUN: llvm-readobj -r %t.o | FileCheck --check-prefix=OBJ %s
12 ; Restrict maximum branch to between +7 and -8 dwords
14 ; Used to emit an always 4 byte instruction. Inline asm always assumes
15 ; each instruction is the maximum size.
16 declare void @llvm.amdgcn.s.sleep(i32) #0
18 declare i32 @llvm.amdgcn.workitem.id.x() #1
21 ; GCN-LABEL: {{^}}uniform_conditional_max_short_forward_branch:
22 ; GCN: s_load_dword [[CND:s[0-9]+]]
23 ; GCN: s_cmp_eq_u32 [[CND]], 0
24 ; GCN-NEXT: s_cbranch_scc1 [[BB3:BB[0-9]+_[0-9]+]]
27 ; GCN-NEXT: ; %bb.1: ; %bb2
28 ; GCN-NEXT: ;;#ASMSTART
35 ; GCN-NEXT: [[BB3]]: ; %bb3
36 ; GCN: v_mov_b32_e32 [[V_CND:v[0-9]+]], [[CND]]
37 ; GCN: buffer_store_dword [[V_CND]]
39 define amdgpu_kernel void @uniform_conditional_max_short_forward_branch(i32 addrspace(1)* %arg, i32 %cnd) #0 {
41 %cmp = icmp eq i32 %cnd, 0
42 br i1 %cmp, label %bb3, label %bb2 ; +8 dword branch
46 call void asm sideeffect
50 call void @llvm.amdgcn.s.sleep(i32 0)
54 store volatile i32 %cnd, i32 addrspace(1)* %arg
58 ; GCN-LABEL: {{^}}uniform_conditional_min_long_forward_branch:
59 ; GCN: s_load_dword [[CND:s[0-9]+]]
60 ; GCN: s_cmp_eq_u32 [[CND]], 0
61 ; GCN-NEXT: s_cbranch_scc0 [[LONGBB:BB[0-9]+_[0-9]+]]
63 ; GCN-NEXT: [[LONG_JUMP:BB[0-9]+_[0-9]+]]: ; %bb0
64 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
65 ; GCN-NEXT: s_add_u32 s[[PC_LO]], s[[PC_LO]], [[ENDBB:BB[0-9]+_[0-9]+]]-([[LONG_JUMP]]+4)
66 ; GCN-NEXT: s_addc_u32 s[[PC_HI]], s[[PC_HI]], 0
67 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
69 ; GCN-NEXT: [[LONGBB]]:
70 ; GCN-NEXT: ;;#ASMSTART
77 ; GCN-NEXT: [[ENDBB]]:
78 ; GCN: v_mov_b32_e32 [[V_CND:v[0-9]+]], [[CND]]
79 ; GCN: buffer_store_dword [[V_CND]]
81 define amdgpu_kernel void @uniform_conditional_min_long_forward_branch(i32 addrspace(1)* %arg, i32 %cnd) #0 {
83 %cmp = icmp eq i32 %cnd, 0
84 br i1 %cmp, label %bb3, label %bb2 ; +9 dword branch
88 call void asm sideeffect
96 store volatile i32 %cnd, i32 addrspace(1)* %arg
100 ; GCN-LABEL: {{^}}uniform_conditional_min_long_forward_vcnd_branch:
101 ; GCN: s_load_dword [[CND:s[0-9]+]]
103 ; GCN-DAG: v_cmp_eq_f32_e64 [[UNMASKED:s\[[0-9]+:[0-9]+\]]], [[CND]], 0
104 ; GCN-DAG: s_and_b64 vcc, exec, [[UNMASKED]]
105 ; GCN: s_cbranch_vccz [[LONGBB:BB[0-9]+_[0-9]+]]
107 ; GCN-NEXT: [[LONG_JUMP:BB[0-9]+_[0-9]+]]: ; %bb0
108 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
109 ; GCN-NEXT: s_add_u32 s[[PC_LO]], s[[PC_LO]], [[ENDBB:BB[0-9]+_[0-9]+]]-([[LONG_JUMP]]+4)
110 ; GCN-NEXT: s_addc_u32 s[[PC_HI]], s[[PC_HI]], 0
111 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
113 ; GCN-NEXT: [[LONGBB]]:
120 ; GCN: v_mov_b32_e32 [[V_CND:v[0-9]+]], [[CND]]
121 ; GCN: buffer_store_dword [[V_CND]]
123 define amdgpu_kernel void @uniform_conditional_min_long_forward_vcnd_branch(float addrspace(1)* %arg, float %cnd) #0 {
125 %cmp = fcmp oeq float %cnd, 0.0
126 br i1 %cmp, label %bb3, label %bb2 ; + 8 dword branch
129 call void asm sideeffect " ; 32 bytes
137 store volatile float %cnd, float addrspace(1)* %arg
141 ; GCN-LABEL: {{^}}min_long_forward_vbranch:
143 ; GCN: buffer_load_dword
144 ; GCN: v_cmp_ne_u32_e32 vcc, 0, v{{[0-9]+}}
145 ; GCN: s_and_saveexec_b64 [[SAVE:s\[[0-9]+:[0-9]+\]]], vcc
152 ; GCN: s_or_b64 exec, exec, [[SAVE]]
153 ; GCN: buffer_store_dword
155 define amdgpu_kernel void @min_long_forward_vbranch(i32 addrspace(1)* %arg) #0 {
157 %tid = call i32 @llvm.amdgcn.workitem.id.x()
158 %tid.ext = zext i32 %tid to i64
159 %gep = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tid.ext
160 %load = load volatile i32, i32 addrspace(1)* %gep
161 %cmp = icmp eq i32 %load, 0
162 br i1 %cmp, label %bb3, label %bb2 ; + 8 dword branch
165 call void asm sideeffect " ; 32 bytes
173 store volatile i32 %load, i32 addrspace(1)* %gep
177 ; GCN-LABEL: {{^}}long_backward_sbranch:
178 ; GCN: s_mov_b32 [[LOOPIDX:s[0-9]+]], 0{{$}}
180 ; GCN: [[LOOPBB:BB[0-9]+_[0-9]+]]: ; %bb2
181 ; GCN-NEXT: ; =>This Inner Loop Header: Depth=1
182 ; GCN-NEXT: s_add_i32 [[INC:s[0-9]+]], [[LOOPIDX]], 1
183 ; GCN-NEXT: s_cmp_lt_i32 [[INC]], 10
185 ; GCN-NEXT: ;;#ASMSTART
186 ; GCN-NEXT: v_nop_e64
187 ; GCN-NEXT: v_nop_e64
188 ; GCN-NEXT: v_nop_e64
189 ; GCN-NEXT: ;;#ASMEND
191 ; GCN-NEXT: s_cbranch_scc0 [[ENDBB:BB[0-9]+_[0-9]+]]
193 ; GCN-NEXT: [[LONG_JUMP:BB[0-9]+_[0-9]+]]: ; %bb2
194 ; GCN-NEXT: ; in Loop: Header=[[LOOPBB]] Depth=1
196 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
197 ; GCN-NEXT: s_sub_u32 s[[PC_LO]], s[[PC_LO]], ([[LONG_JUMP]]+4)-[[LOOPBB]]
198 ; GCN-NEXT: s_subb_u32 s[[PC_HI]], s[[PC_HI]], 0
199 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
201 ; GCN-NEXT: [[ENDBB]]:
203 define amdgpu_kernel void @long_backward_sbranch(i32 addrspace(1)* %arg) #0 {
208 %loop.idx = phi i32 [ 0, %bb ], [ %inc, %bb2 ]
210 call void asm sideeffect
214 %inc = add nsw i32 %loop.idx, 1 ; add cost 4
215 %cmp = icmp slt i32 %inc, 10 ; condition cost = 8
216 br i1 %cmp, label %bb2, label %bb3 ; -
222 ; Requires expansion of unconditional branch from %bb2 to %bb4 (and
223 ; expansion of conditional branch from %bb to %bb3.
225 ; GCN-LABEL: {{^}}uniform_unconditional_min_long_forward_branch:
227 ; GCN-NEXT: s_cbranch_scc0 [[BB2:BB[0-9]+_[0-9]+]]
229 ; GCN-NEXT: [[LONG_JUMP0:BB[0-9]+_[0-9]+]]: ; %bb0
230 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC0_LO:[0-9]+]]:[[PC0_HI:[0-9]+]]{{\]}}
231 ; GCN-NEXT: s_add_u32 s[[PC0_LO]], s[[PC0_LO]], [[BB3:BB[0-9]_[0-9]+]]-([[LONG_JUMP0]]+4)
232 ; GCN-NEXT: s_addc_u32 s[[PC0_HI]], s[[PC0_HI]], 0{{$}}
233 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC0_LO]]:[[PC0_HI]]{{\]}}
235 ; GCN-NEXT: [[BB2]]: ; %bb2
236 ; GCN: v_mov_b32_e32 [[BB2_K:v[0-9]+]], 17
237 ; GCN: buffer_store_dword [[BB2_K]]
239 ; GCN-NEXT: [[LONG_JUMP1:BB[0-9]+_[0-9]+]]: ; %bb2
240 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC1_LO:[0-9]+]]:[[PC1_HI:[0-9]+]]{{\]}}
241 ; GCN-NEXT: s_add_u32 s[[PC1_LO]], s[[PC1_LO]], [[BB4:BB[0-9]_[0-9]+]]-([[LONG_JUMP1]]+4)
242 ; GCN-NEXT: s_addc_u32 s[[PC1_HI]], s[[PC1_HI]], 0{{$}}
243 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC1_LO]]:[[PC1_HI]]{{\]}}
245 ; GCN: [[BB3]]: ; %bb3
252 ; GCN-NEXT: [[BB4]]: ; %bb4
253 ; GCN: v_mov_b32_e32 [[BB4_K:v[0-9]+]], 63
254 ; GCN: buffer_store_dword [[BB4_K]]
256 ; GCN-NEXT: .Lfunc_end{{[0-9]+}}:
257 define amdgpu_kernel void @uniform_unconditional_min_long_forward_branch(i32 addrspace(1)* %arg, i32 %arg1) {
259 %tmp = icmp ne i32 %arg1, 0
260 br i1 %tmp, label %bb2, label %bb3
263 store volatile i32 17, i32 addrspace(1)* undef
268 call void asm sideeffect
276 store volatile i32 63, i32 addrspace(1)* %arg
280 ; GCN-LABEL: {{^}}uniform_unconditional_min_long_backward_branch:
281 ; GCN-NEXT: ; %bb.0: ; %entry
283 ; GCN-NEXT: [[LOOP:BB[0-9]_[0-9]+]]: ; %loop
284 ; GCN-NEXT: ; =>This Inner Loop Header: Depth=1
285 ; GCN-NEXT: ;;#ASMSTART
286 ; GCN-NEXT: v_nop_e64
287 ; GCN-NEXT: v_nop_e64
288 ; GCN-NEXT: v_nop_e64
289 ; GCN-NEXT: v_nop_e64
290 ; GCN-NEXT: ;;#ASMEND
292 ; GCN-NEXT: [[LONGBB:BB[0-9]+_[0-9]+]]: ; %loop
293 ; GCN-NEXT: ; in Loop: Header=[[LOOP]] Depth=1
295 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
296 ; GCN-NEXT: s_sub_u32 s[[PC_LO]], s[[PC_LO]], ([[LONGBB]]+4)-[[LOOP]]
297 ; GCN-NEXT: s_subb_u32 s[[PC_HI]], s[[PC_HI]], 0{{$}}
298 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
299 ; GCN-NEXT .Lfunc_end{{[0-9]+}}:
300 define amdgpu_kernel void @uniform_unconditional_min_long_backward_branch(i32 addrspace(1)* %arg, i32 %arg1) {
306 call void asm sideeffect
314 ; Expansion of branch from %bb1 to %bb3 introduces need to expand
315 ; branch from %bb0 to %bb2
317 ; GCN-LABEL: {{^}}expand_requires_expand:
318 ; GCN-NEXT: ; %bb.0: ; %bb0
320 ; GCN: s_cmp_lt_i32 s{{[0-9]+}}, 0{{$}}
321 ; GCN-NEXT: s_cbranch_scc0 [[BB1:BB[0-9]+_[0-9]+]]
323 ; GCN-NEXT: [[LONGBB0:BB[0-9]+_[0-9]+]]: ; %bb0
325 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC0_LO:[0-9]+]]:[[PC0_HI:[0-9]+]]{{\]}}
326 ; GCN-NEXT: s_add_u32 s[[PC0_LO]], s[[PC0_LO]], [[BB2:BB[0-9]_[0-9]+]]-([[LONGBB0]]+4)
327 ; GCN-NEXT: s_addc_u32 s[[PC0_HI]], s[[PC0_HI]], 0{{$}}
328 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC0_LO]]:[[PC0_HI]]{{\]}}
330 ; GCN-NEXT: [[BB1]]: ; %bb1
331 ; GCN-NEXT: s_load_dword
332 ; GCN-NEXT: s_waitcnt lgkmcnt(0)
333 ; GCN-NEXT: s_cmp_eq_u32 s{{[0-9]+}}, 3{{$}}
334 ; GCN-NEXT: s_cbranch_scc0 [[BB2:BB[0-9]_[0-9]+]]
336 ; GCN-NEXT: [[LONGBB1:BB[0-9]+_[0-9]+]]: ; %bb1
337 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC1_LO:[0-9]+]]:[[PC1_HI:[0-9]+]]{{\]}}
338 ; GCN-NEXT: s_add_u32 s[[PC1_LO]], s[[PC1_LO]], [[BB3:BB[0-9]+_[0-9]+]]-([[LONGBB1]]+4)
339 ; GCN-NEXT: s_addc_u32 s[[PC1_HI]], s[[PC1_HI]], 0{{$}}
340 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC1_LO]]:[[PC1_HI]]{{\]}}
342 ; GCN-NEXT: [[BB2]]: ; %bb2
343 ; GCN-NEXT: ;;#ASMSTART
344 ; GCN-NEXT: v_nop_e64
345 ; GCN-NEXT: v_nop_e64
346 ; GCN-NEXT: v_nop_e64
347 ; GCN-NEXT: v_nop_e64
348 ; GCN-NEXT: ;;#ASMEND
350 ; GCN-NEXT: [[BB3]]: ; %bb3
351 ; GCN-NEXT: ;;#ASMSTART
352 ; GCN-NEXT: v_nop_e64
353 ; GCN-NEXT: ;;#ASMEND
354 ; GCN-NEXT: ;;#ASMSTART
355 ; GCN-NEXT: v_nop_e64
356 ; GCN-NEXT: ;;#ASMEND
358 define amdgpu_kernel void @expand_requires_expand(i32 %cond0) #0 {
360 %tmp = tail call i32 @llvm.amdgcn.workitem.id.x() #0
361 %cmp0 = icmp slt i32 %cond0, 0
362 br i1 %cmp0, label %bb2, label %bb1
365 %val = load volatile i32, i32 addrspace(4)* undef
366 %cmp1 = icmp eq i32 %val, 3
367 br i1 %cmp1, label %bb3, label %bb2
370 call void asm sideeffect
378 ; These NOPs prevent tail-duplication-based outlining
379 ; from firing, which defeats the need to expand the branches and this test.
380 call void asm sideeffect
382 call void asm sideeffect
387 ; Requires expanding of required skip branch.
389 ; GCN-LABEL: {{^}}uniform_inside_divergent:
390 ; GCN: v_cmp_gt_u32_e32 vcc, 16, v{{[0-9]+}}
391 ; GCN-NEXT: s_and_saveexec_b64 [[MASK:s\[[0-9]+:[0-9]+\]]], vcc
392 ; GCN-NEXT: ; mask branch [[ENDIF:BB[0-9]+_[0-9]+]]
393 ; GCN-NEXT: s_cbranch_execnz [[IF:BB[0-9]+_[0-9]+]]
395 ; GCN-NEXT: [[LONGBB:BB[0-9]+_[0-9]+]]: ; %entry
396 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
397 ; GCN-NEXT: s_add_u32 s[[PC_LO]], s[[PC_LO]], [[BB2:BB[0-9]_[0-9]+]]-([[LONGBB]]+4)
398 ; GCN-NEXT: s_addc_u32 s[[PC_HI]], s[[PC_HI]], 0{{$}}
399 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
401 ; GCN-NEXT: [[IF]]: ; %if
402 ; GCN: buffer_store_dword
404 ; GCN: s_cbranch_scc1 [[ENDIF]]
406 ; GCN-NEXT: ; %bb.2: ; %if_uniform
407 ; GCN: buffer_store_dword
409 ; GCN-NEXT: [[ENDIF]]: ; %endif
410 ; GCN-NEXT: s_or_b64 exec, exec, [[MASK]]
411 ; GCN-NEXT: s_sleep 5
413 define amdgpu_kernel void @uniform_inside_divergent(i32 addrspace(1)* %out, i32 %cond) #0 {
415 %tid = call i32 @llvm.amdgcn.workitem.id.x()
416 %d_cmp = icmp ult i32 %tid, 16
417 br i1 %d_cmp, label %if, label %endif
420 store i32 0, i32 addrspace(1)* %out
421 %u_cmp = icmp eq i32 %cond, 0
422 br i1 %u_cmp, label %if_uniform, label %endif
425 store i32 1, i32 addrspace(1)* %out
429 ; layout can remove the split branch if it can copy the return block.
430 ; This call makes the return block long enough that it doesn't get copied.
431 call void @llvm.amdgcn.s.sleep(i32 5);
437 ; GCN-LABEL: {{^}}analyze_mask_branch:
438 ; GCN: v_cmp_nlt_f32_e32 vcc
439 ; GCN-NEXT: s_and_saveexec_b64 [[TEMP_MASK:s\[[0-9]+:[0-9]+\]]], vcc
440 ; GCN-NEXT: s_xor_b64 [[MASK:s\[[0-9]+:[0-9]+\]]], exec, [[TEMP_MASK]]
441 ; GCN-NEXT: ; mask branch [[FLOW:BB[0-9]+_[0-9]+]]
443 ; GCN: [[FLOW]]: ; %Flow
444 ; GCN-NEXT: s_or_saveexec_b64 [[TEMP_MASK1:s\[[0-9]+:[0-9]+\]]], [[MASK]]
445 ; GCN-NEXT: s_xor_b64 exec, exec, [[TEMP_MASK1]]
446 ; GCN-NEXT: ; mask branch [[RET:BB[0-9]+_[0-9]+]]
448 ; GCN: [[LOOP_BODY:BB[0-9]+_[0-9]+]]: ; %loop{{$}}
457 ; GCN: s_cbranch_vccz [[RET]]
459 ; GCN-NEXT: [[LONGBB:BB[0-9]+_[0-9]+]]: ; %loop
460 ; GCN-NEXT: ; in Loop: Header=[[LOOP_BODY]] Depth=1
461 ; GCN-NEXT: s_getpc_b64 s{{\[}}[[PC_LO:[0-9]+]]:[[PC_HI:[0-9]+]]{{\]}}
462 ; GCN-NEXT: s_sub_u32 s[[PC_LO]], s[[PC_LO]], ([[LONGBB]]+4)-[[LOOP_BODY]]
463 ; GCN-NEXT: s_subb_u32 s[[PC_HI]], s[[PC_HI]], 0
464 ; GCN-NEXT: s_setpc_b64 s{{\[}}[[PC_LO]]:[[PC_HI]]{{\]}}
466 ; GCN-NEXT: [[RET]]: ; %UnifiedReturnBlock
468 define amdgpu_kernel void @analyze_mask_branch() #0 {
470 %reg = call float asm sideeffect "v_mov_b32_e64 $0, 0", "=v"()
471 %cmp0 = fcmp ogt float %reg, 0.000000e+00
472 br i1 %cmp0, label %loop, label %ret
475 %phi = phi float [ 0.000000e+00, %loop_body ], [ 1.000000e+00, %entry ]
476 call void asm sideeffect
479 %cmp1 = fcmp olt float %phi, 8.0
480 br i1 %cmp1, label %loop_body, label %ret
483 call void asm sideeffect
491 store volatile i32 7, i32 addrspace(1)* undef
495 ; GCN-LABEL: {{^}}long_branch_hang:
496 ; GCN: s_cmp_lt_i32 s{{[0-9]+}}, 6
497 ; GCN: s_cbranch_scc1 {{BB[0-9]+_[0-9]+}}
498 ; GCN-NEXT: s_branch [[LONG_BR_0:BB[0-9]+_[0-9]+]]
499 ; GCN-NEXT: BB{{[0-9]+_[0-9]+}}:
501 ; GCN: s_add_u32 s{{[0-9]+}}, s{{[0-9]+}}, [[LONG_BR_DEST0:BB[0-9]+_[0-9]+]]-(
502 ; GCN-NEXT: s_addc_u32
503 ; GCN-NEXT: s_setpc_b64
505 ; GCN-NEXT: [[LONG_BR_0]]:
506 ; GCN-DAG: v_cmp_lt_i32
507 ; GCN-DAG: v_cmp_gt_i32
508 ; GCN: s_cbranch_vccnz
513 ; GCN: [[LONG_BR_DEST0]]
514 ; GCN: s_cbranch_vccz
518 define amdgpu_kernel void @long_branch_hang(i32 addrspace(1)* nocapture %arg, i32 %arg1, i32 %arg2, i32 %arg3, i32 %arg4, i64 %arg5) #0 {
520 %tmp = icmp slt i32 %arg2, 9
521 %tmp6 = icmp eq i32 %arg1, 0
522 %tmp7 = icmp sgt i32 %arg4, 0
523 %tmp8 = icmp sgt i32 %arg4, 5
524 br i1 %tmp8, label %bb9, label %bb13
527 %tmp10 = and i1 %tmp7, %tmp
528 %tmp11 = icmp slt i32 %arg3, %arg4
529 %tmp12 = or i1 %tmp11, %tmp7
530 br i1 %tmp12, label %bb19, label %bb14
533 call void asm sideeffect
538 br i1 %tmp6, label %bb19, label %bb14
540 bb14: ; preds = %bb13, %bb9
541 %tmp15 = icmp slt i32 %arg3, %arg4
542 %tmp16 = or i1 %tmp15, %tmp
543 %tmp17 = and i1 %tmp6, %tmp16
544 %tmp18 = zext i1 %tmp17 to i32
547 bb19: ; preds = %bb14, %bb13, %bb9
548 %tmp20 = phi i32 [ undef, %bb9 ], [ undef, %bb13 ], [ %tmp18, %bb14 ]
549 %tmp21 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %arg5
550 store i32 %tmp20, i32 addrspace(1)* %tmp21, align 4
554 attributes #0 = { nounwind }
555 attributes #1 = { nounwind readnone }