1 # NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py
2 # RUN: llc -mtriple=thumbv8.1m.main -mattr=+mve -run-pass=arm-low-overhead-loops %s -o - --verify-machineinstrs | FileCheck %s
5 define dso_local <4 x i32> @exit_liveout(i16* nocapture readonly %a, i16* nocapture readonly %b, i32* %c, i32 %N, <4 x i32> %pass) {
7 %cmp9 = icmp eq i32 %N, 0
9 %tmp1 = lshr i32 %tmp, 2
10 %tmp2 = shl nuw i32 %tmp1, 2
11 %tmp3 = add i32 %tmp2, -4
12 %tmp4 = lshr i32 %tmp3, 2
13 %tmp5 = add nuw nsw i32 %tmp4, 1
14 br i1 %cmp9, label %exit, label %vector.ph
16 vector.ph: ; preds = %entry
17 %start = call i32 @llvm.start.loop.iterations.i32(i32 %tmp5)
20 vector.body: ; preds = %vector.body, %vector.ph
21 %lsr.iv1 = phi i32 [ %lsr.iv.next, %vector.body ], [ %start, %vector.ph ]
22 %lsr.iv18 = phi i16* [ %scevgep19, %vector.body ], [ %b, %vector.ph ]
23 %lsr.iv20 = phi i32* [ %scevgep20, %vector.body ], [ %c, %vector.ph ]
24 %lsr.iv = phi i16* [ %scevgep, %vector.body ], [ %a, %vector.ph ]
25 %vec.phi = phi <4 x i32> [ zeroinitializer, %vector.ph ], [ %tmp13, %vector.body ]
26 %tmp7 = phi i32 [ %N, %vector.ph ], [ %tmp9, %vector.body ]
27 %lsr.iv17 = bitcast i16* %lsr.iv to <4 x i16>*
28 %lsr.store = bitcast i32* %lsr.iv20 to <4 x i32>*
29 %lsr.iv1820 = bitcast i16* %lsr.iv18 to <4 x i16>*
30 %tmp8 = call <4 x i1> @llvm.arm.mve.vctp32(i32 %tmp7)
31 %tmp9 = sub i32 %tmp7, 4
32 %wide.masked.load = call <4 x i16> @llvm.masked.load.v4i16.p0v4i16(<4 x i16>* %lsr.iv17, i32 2, <4 x i1> %tmp8, <4 x i16> undef)
33 %tmp10 = sext <4 x i16> %wide.masked.load to <4 x i32>
34 %wide.masked.load14 = call <4 x i16> @llvm.masked.load.v4i16.p0v4i16(<4 x i16>* %lsr.iv1820, i32 2, <4 x i1> %tmp8, <4 x i16> undef)
35 %tmp11 = sext <4 x i16> %wide.masked.load14 to <4 x i32>
36 %tmp12 = mul nsw <4 x i32> %tmp11, %tmp10
37 %tmp13 = add <4 x i32> %tmp12, %vec.phi
38 call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %tmp13, <4 x i32>* %lsr.store, i32 4, <4 x i1> %tmp8)
39 %scevgep = getelementptr i16, i16* %lsr.iv, i32 4
40 %scevgep19 = getelementptr i16, i16* %lsr.iv18, i32 4
41 %scevgep20 = getelementptr i32, i32* %lsr.iv20, i32 4
42 %tmp14 = call i32 @llvm.loop.decrement.reg.i32.i32.i32(i32 %lsr.iv1, i32 1)
43 %tmp15 = icmp ne i32 %tmp14, 0
44 %lsr.iv.next = add nsw i32 %lsr.iv1, -1
45 br i1 %tmp15, label %vector.body, label %exit
47 exit: ; preds = %vector.body, %entry
51 declare <4 x i16> @llvm.masked.load.v4i16.p0v4i16(<4 x i16>*, i32 immarg, <4 x i1>, <4 x i16>)
52 declare void @llvm.masked.store.v4i32.p0v4i32(<4 x i32>, <4 x i32>*, i32 immarg, <4 x i1>)
53 declare i32 @llvm.start.loop.iterations.i32(i32)
54 declare i32 @llvm.loop.decrement.reg.i32.i32.i32(i32, i32)
55 declare <4 x i1> @llvm.arm.mve.vctp32(i32)
61 tracksRegLiveness: true
64 - { reg: '$r0', virtual-reg: '' }
65 - { reg: '$r1', virtual-reg: '' }
66 - { reg: '$r2', virtual-reg: '' }
67 - { reg: '$r3', virtual-reg: '' }
73 - { id: 0, type: default, offset: 0, size: 16, alignment: 8, stack-id: default,
74 isImmutable: true, isAliased: false, callee-saved-register: '', callee-saved-restored: true,
75 debug-info-variable: '', debug-info-expression: '', debug-info-location: '' }
77 - { id: 0, name: '', type: spill-slot, offset: -4, size: 4, alignment: 4,
78 stack-id: default, callee-saved-register: '$lr', callee-saved-restored: false,
79 debug-info-variable: '', debug-info-expression: '', debug-info-location: '' }
80 - { id: 1, name: '', type: spill-slot, offset: -8, size: 4, alignment: 4,
81 stack-id: default, callee-saved-register: '$r4', callee-saved-restored: true,
82 debug-info-variable: '', debug-info-expression: '', debug-info-location: '' }
85 machineFunctionInfo: {}
87 ; CHECK-LABEL: name: exit_liveout
89 ; CHECK: successors: %bb.3(0x30000000), %bb.1(0x50000000)
90 ; CHECK: liveins: $lr, $r0, $r1, $r2, $r3, $r4
91 ; CHECK: frame-setup tPUSH 14 /* CC::al */, $noreg, killed $r4, killed $lr, implicit-def $sp, implicit $sp
92 ; CHECK: frame-setup CFI_INSTRUCTION def_cfa_offset 8
93 ; CHECK: frame-setup CFI_INSTRUCTION offset $lr, -4
94 ; CHECK: frame-setup CFI_INSTRUCTION offset $r4, -8
95 ; CHECK: renamable $r12 = t2ADDri $sp, 8, 14 /* CC::al */, $noreg, $noreg
96 ; CHECK: renamable $q0 = MVE_VLDRWU32 killed renamable $r12, 0, 0, $noreg :: (load (s128) from %fixed-stack.0, align 8)
97 ; CHECK: tCBZ $r3, %bb.3
98 ; CHECK: bb.1.vector.ph:
99 ; CHECK: successors: %bb.2(0x80000000)
100 ; CHECK: liveins: $q0, $r0, $r1, $r2, $r3
101 ; CHECK: renamable $q1 = MVE_VMOVimmi32 0, 0, $noreg, undef renamable $q1
102 ; CHECK: $lr = MVE_DLSTP_32 killed renamable $r3
103 ; CHECK: bb.2.vector.body:
104 ; CHECK: successors: %bb.2(0x7c000000), %bb.3(0x04000000)
105 ; CHECK: liveins: $lr, $q0, $q1, $r0, $r1, $r2
106 ; CHECK: renamable $r0, renamable $q2 = MVE_VLDRHS32_post killed renamable $r0, 8, 0, $noreg :: (load (s64) from %ir.lsr.iv17, align 2)
107 ; CHECK: renamable $r1, renamable $q3 = MVE_VLDRHS32_post killed renamable $r1, 8, 0, $noreg :: (load (s64) from %ir.lsr.iv1820, align 2)
108 ; CHECK: renamable $q2 = nsw MVE_VMULi32 killed renamable $q3, killed renamable $q2, 0, $noreg, undef renamable $q2
109 ; CHECK: renamable $q1 = MVE_VADDi32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
110 ; CHECK: renamable $r2 = MVE_VSTRWU32_post renamable $q1, killed renamable $r2, 16, 0, killed $noreg :: (store (s128) into %ir.lsr.store, align 4)
111 ; CHECK: $lr = MVE_LETP killed renamable $lr, %bb.2
113 ; CHECK: liveins: $q0
114 ; CHECK: renamable $r0, renamable $r1 = VMOVRRD renamable $d0, 14 /* CC::al */, $noreg
115 ; CHECK: renamable $r2, renamable $r3 = VMOVRRD killed renamable $d1, 14 /* CC::al */, $noreg, implicit killed $q0
116 ; CHECK: tPOP_RET 14 /* CC::al */, $noreg, def $r4, def $pc, implicit killed $r0, implicit killed $r1, implicit killed $r2, implicit killed $r3
118 successors: %bb.3(0x30000000), %bb.1(0x50000000)
119 liveins: $r0, $r1, $r2, $r3, $r4, $lr
121 frame-setup tPUSH 14, $noreg, killed $r4, killed $lr, implicit-def $sp, implicit $sp
122 frame-setup CFI_INSTRUCTION def_cfa_offset 8
123 frame-setup CFI_INSTRUCTION offset $lr, -4
124 frame-setup CFI_INSTRUCTION offset $r4, -8
125 renamable $r12 = t2ADDri $sp, 8, 14, $noreg, $noreg
126 renamable $q0 = MVE_VLDRWU32 killed renamable $r12, 0, 0, $noreg :: (load (s128) from %fixed-stack.0, align 8)
130 successors: %bb.2(0x80000000)
131 liveins: $q0, $r0, $r1, $r2, $r3
133 renamable $r12 = t2ADDri renamable $r3, 3, 14, $noreg, $noreg
134 renamable $lr = t2MOVi 1, 14, $noreg, $noreg
135 renamable $r12 = t2BICri killed renamable $r12, 3, 14, $noreg, $noreg
136 renamable $q1 = MVE_VMOVimmi32 0, 0, $noreg, undef renamable $q1
137 renamable $r12 = t2SUBri killed renamable $r12, 4, 14, $noreg, $noreg
138 renamable $r4 = nuw nsw t2ADDrs killed renamable $lr, killed renamable $r12, 19, 14, $noreg, $noreg
139 $lr = t2DoLoopStart renamable $r4
140 $r12 = tMOVr killed $r4, 14, $noreg
143 successors: %bb.2(0x7c000000), %bb.3(0x04000000)
144 liveins: $q0, $q1, $r0, $r1, $r2, $r3, $r12
146 renamable $vpr = MVE_VCTP32 renamable $r3, 0, $noreg
147 MVE_VPST 4, implicit $vpr
148 renamable $r0, renamable $q2 = MVE_VLDRHS32_post killed renamable $r0, 8, 1, renamable $vpr :: (load (s64) from %ir.lsr.iv17, align 2)
149 renamable $r1, renamable $q3 = MVE_VLDRHS32_post killed renamable $r1, 8, 1, renamable $vpr :: (load (s64) from %ir.lsr.iv1820, align 2)
150 $lr = tMOVr $r12, 14, $noreg
151 renamable $q2 = nsw MVE_VMULi32 killed renamable $q3, killed renamable $q2, 0, $noreg, undef renamable $q2
152 renamable $r12 = nsw t2SUBri killed $r12, 1, 14, $noreg, $noreg
153 renamable $r3, dead $cpsr = tSUBi8 killed renamable $r3, 4, 14, $noreg
154 renamable $q1 = MVE_VADDi32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
155 renamable $lr = t2LoopDec killed renamable $lr, 1
156 MVE_VPST 8, implicit $vpr
157 renamable $r2 = MVE_VSTRWU32_post renamable $q1, killed renamable $r2, 16, 1, killed renamable $vpr :: (store (s128) into %ir.lsr.store, align 4)
158 t2LoopEnd killed renamable $lr, %bb.2, implicit-def dead $cpsr
164 renamable $r0, renamable $r1 = VMOVRRD renamable $d0, 14, $noreg
165 renamable $r2, renamable $r3 = VMOVRRD killed renamable $d1, 14, $noreg, implicit $q0
166 tPOP_RET 14, $noreg, def $r4, def $pc, implicit killed $r0, implicit killed $r1, implicit killed $r2, implicit killed $r3