1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 3
2 ; RUN: llc < %s -mtriple=riscv64 -mattr=+m,+v,+xsfvcp \
3 ; RUN: -riscv-use-rematerializable-movimm=false | FileCheck %s --check-prefix=NOREMAT
4 ; RUN: llc < %s -mtriple=riscv64 -mattr=+m,+v,+xsfvcp \
5 ; RUN: --riscv-use-rematerializable-movimm=true | FileCheck %s --check-prefix=REMAT
7 define void @test(ptr %0, ptr %1, i64 %2) {
10 ; NOREMAT-NEXT: addi sp, sp, -752
11 ; NOREMAT-NEXT: .cfi_def_cfa_offset 752
12 ; NOREMAT-NEXT: sd ra, 744(sp) # 8-byte Folded Spill
13 ; NOREMAT-NEXT: sd s0, 736(sp) # 8-byte Folded Spill
14 ; NOREMAT-NEXT: sd s1, 728(sp) # 8-byte Folded Spill
15 ; NOREMAT-NEXT: sd s2, 720(sp) # 8-byte Folded Spill
16 ; NOREMAT-NEXT: sd s3, 712(sp) # 8-byte Folded Spill
17 ; NOREMAT-NEXT: sd s4, 704(sp) # 8-byte Folded Spill
18 ; NOREMAT-NEXT: sd s5, 696(sp) # 8-byte Folded Spill
19 ; NOREMAT-NEXT: sd s6, 688(sp) # 8-byte Folded Spill
20 ; NOREMAT-NEXT: sd s7, 680(sp) # 8-byte Folded Spill
21 ; NOREMAT-NEXT: sd s8, 672(sp) # 8-byte Folded Spill
22 ; NOREMAT-NEXT: sd s9, 664(sp) # 8-byte Folded Spill
23 ; NOREMAT-NEXT: sd s10, 656(sp) # 8-byte Folded Spill
24 ; NOREMAT-NEXT: sd s11, 648(sp) # 8-byte Folded Spill
25 ; NOREMAT-NEXT: .cfi_offset ra, -8
26 ; NOREMAT-NEXT: .cfi_offset s0, -16
27 ; NOREMAT-NEXT: .cfi_offset s1, -24
28 ; NOREMAT-NEXT: .cfi_offset s2, -32
29 ; NOREMAT-NEXT: .cfi_offset s3, -40
30 ; NOREMAT-NEXT: .cfi_offset s4, -48
31 ; NOREMAT-NEXT: .cfi_offset s5, -56
32 ; NOREMAT-NEXT: .cfi_offset s6, -64
33 ; NOREMAT-NEXT: .cfi_offset s7, -72
34 ; NOREMAT-NEXT: .cfi_offset s8, -80
35 ; NOREMAT-NEXT: .cfi_offset s9, -88
36 ; NOREMAT-NEXT: .cfi_offset s10, -96
37 ; NOREMAT-NEXT: .cfi_offset s11, -104
38 ; NOREMAT-NEXT: csrr a2, vlenb
39 ; NOREMAT-NEXT: slli a2, a2, 1
40 ; NOREMAT-NEXT: sub sp, sp, a2
41 ; NOREMAT-NEXT: .cfi_escape 0x0f, 0x0e, 0x72, 0x00, 0x11, 0xf0, 0x05, 0x22, 0x11, 0x02, 0x92, 0xa2, 0x38, 0x00, 0x1e, 0x22 # sp + 752 + 2 * vlenb
42 ; NOREMAT-NEXT: mv a7, a0
43 ; NOREMAT-NEXT: li a0, 32
44 ; NOREMAT-NEXT: addi a5, a7, 512
45 ; NOREMAT-NEXT: addi a4, a7, 1024
46 ; NOREMAT-NEXT: addi a6, a7, 1536
47 ; NOREMAT-NEXT: li t4, 1
48 ; NOREMAT-NEXT: li a2, 5
49 ; NOREMAT-NEXT: li t1, 3
50 ; NOREMAT-NEXT: li t0, 7
51 ; NOREMAT-NEXT: lui t5, 1
52 ; NOREMAT-NEXT: li s4, 9
53 ; NOREMAT-NEXT: li s6, 11
54 ; NOREMAT-NEXT: li s9, 13
55 ; NOREMAT-NEXT: li ra, 15
56 ; NOREMAT-NEXT: lui t2, 2
57 ; NOREMAT-NEXT: lui s1, 3
58 ; NOREMAT-NEXT: lui t3, 4
59 ; NOREMAT-NEXT: lui s0, 5
60 ; NOREMAT-NEXT: lui s3, 6
61 ; NOREMAT-NEXT: lui s7, 7
62 ; NOREMAT-NEXT: vsetvli zero, a0, e32, m2, ta, ma
63 ; NOREMAT-NEXT: slli t4, t4, 11
64 ; NOREMAT-NEXT: sd t4, 512(sp) # 8-byte Folded Spill
65 ; NOREMAT-NEXT: slli a3, a2, 9
66 ; NOREMAT-NEXT: sd a3, 504(sp) # 8-byte Folded Spill
67 ; NOREMAT-NEXT: slli t6, t1, 10
68 ; NOREMAT-NEXT: slli s2, t0, 9
69 ; NOREMAT-NEXT: add a0, a7, t5
70 ; NOREMAT-NEXT: lui s11, 1
71 ; NOREMAT-NEXT: slli s4, s4, 9
72 ; NOREMAT-NEXT: slli s5, a2, 10
73 ; NOREMAT-NEXT: slli s6, s6, 9
74 ; NOREMAT-NEXT: slli s8, t1, 11
75 ; NOREMAT-NEXT: vle32.v v8, (a5)
76 ; NOREMAT-NEXT: slli s9, s9, 9
77 ; NOREMAT-NEXT: li t5, 13
78 ; NOREMAT-NEXT: vle32.v v10, (a4)
79 ; NOREMAT-NEXT: vle32.v v2, (a4)
80 ; NOREMAT-NEXT: slli s10, t0, 10
81 ; NOREMAT-NEXT: vle32.v v0, (a6)
82 ; NOREMAT-NEXT: vle32.v v12, (a6)
83 ; NOREMAT-NEXT: slli ra, ra, 9
84 ; NOREMAT-NEXT: vle32.v v4, (a0)
85 ; NOREMAT-NEXT: vle32.v v20, (a0)
86 ; NOREMAT-NEXT: add a4, a7, t2
87 ; NOREMAT-NEXT: vle32.v v6, (a4)
88 ; NOREMAT-NEXT: vle32.v v30, (a4)
89 ; NOREMAT-NEXT: add a4, a7, s1
90 ; NOREMAT-NEXT: vle32.v v28, (a4)
91 ; NOREMAT-NEXT: vle32.v v26, (a4)
92 ; NOREMAT-NEXT: add a4, a7, t3
93 ; NOREMAT-NEXT: vle32.v v24, (a4)
94 ; NOREMAT-NEXT: vle32.v v22, (a4)
95 ; NOREMAT-NEXT: add a4, a7, s0
96 ; NOREMAT-NEXT: vle32.v v14, (a7)
97 ; NOREMAT-NEXT: vle32.v v18, (a4)
98 ; NOREMAT-NEXT: vle32.v v16, (a4)
99 ; NOREMAT-NEXT: add a4, a7, s3
100 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v14, v8
101 ; NOREMAT-NEXT: vle32.v v14, (a4)
102 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v8, v10
103 ; NOREMAT-NEXT: vle32.v v8, (a4)
104 ; NOREMAT-NEXT: addi a0, sp, 640
105 ; NOREMAT-NEXT: vs2r.v v8, (a0) # Unknown-size Folded Spill
106 ; NOREMAT-NEXT: add a4, a7, t4
107 ; NOREMAT-NEXT: vle32.v v10, (a4)
108 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v2, v0
109 ; NOREMAT-NEXT: vle32.v v2, (a4)
110 ; NOREMAT-NEXT: add a4, a7, a3
111 ; NOREMAT-NEXT: vle32.v v0, (a4)
112 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v10
113 ; NOREMAT-NEXT: vle32.v v10, (a4)
114 ; NOREMAT-NEXT: add a4, a7, t6
115 ; NOREMAT-NEXT: vle32.v v12, (a4)
116 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v2, v0
117 ; NOREMAT-NEXT: vle32.v v2, (a4)
118 ; NOREMAT-NEXT: add a4, a7, s2
119 ; NOREMAT-NEXT: vle32.v v8, (a4)
120 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v10, v12
121 ; NOREMAT-NEXT: vle32.v v12, (a4)
122 ; NOREMAT-NEXT: add a4, a7, s7
123 ; NOREMAT-NEXT: vle32.v v0, (a4)
124 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v2, v8
125 ; NOREMAT-NEXT: vle32.v v10, (a4)
126 ; NOREMAT-NEXT: add a4, a7, s4
127 ; NOREMAT-NEXT: vle32.v v8, (a4)
128 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v4
129 ; NOREMAT-NEXT: vle32.v v12, (a4)
130 ; NOREMAT-NEXT: add a4, a7, s5
131 ; NOREMAT-NEXT: vle32.v v4, (a4)
132 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v20, v8
133 ; NOREMAT-NEXT: vle32.v v8, (a4)
134 ; NOREMAT-NEXT: add a4, a7, s6
135 ; NOREMAT-NEXT: vle32.v v20, (a4)
136 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v4
137 ; NOREMAT-NEXT: vle32.v v12, (a4)
138 ; NOREMAT-NEXT: add a4, a7, s8
139 ; NOREMAT-NEXT: vle32.v v4, (a4)
140 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v8, v20
141 ; NOREMAT-NEXT: vle32.v v8, (a4)
142 ; NOREMAT-NEXT: add a4, a7, s9
143 ; NOREMAT-NEXT: vle32.v v20, (a4)
144 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v4
145 ; NOREMAT-NEXT: vle32.v v12, (a4)
146 ; NOREMAT-NEXT: add a4, a7, s10
147 ; NOREMAT-NEXT: vle32.v v4, (a4)
148 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v8, v20
149 ; NOREMAT-NEXT: vle32.v v8, (a4)
150 ; NOREMAT-NEXT: add a4, a7, ra
151 ; NOREMAT-NEXT: vle32.v v2, (a4)
152 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v4
153 ; NOREMAT-NEXT: lui t4, 8
154 ; NOREMAT-NEXT: add a5, a7, t4
155 ; NOREMAT-NEXT: vle32.v v20, (a5)
156 ; NOREMAT-NEXT: vle32.v v12, (a4)
157 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v8, v2
158 ; NOREMAT-NEXT: li a4, 17
159 ; NOREMAT-NEXT: slli a4, a4, 9
160 ; NOREMAT-NEXT: li s1, 17
161 ; NOREMAT-NEXT: sd a4, 624(sp) # 8-byte Folded Spill
162 ; NOREMAT-NEXT: add a4, a7, a4
163 ; NOREMAT-NEXT: vle32.v v8, (a4)
164 ; NOREMAT-NEXT: vle32.v v4, (a4)
165 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v6
166 ; NOREMAT-NEXT: li a5, 9
167 ; NOREMAT-NEXT: slli a4, a5, 10
168 ; NOREMAT-NEXT: sd a4, 616(sp) # 8-byte Folded Spill
169 ; NOREMAT-NEXT: add a4, a7, a4
170 ; NOREMAT-NEXT: vle32.v v12, (a4)
171 ; NOREMAT-NEXT: vle32.v v6, (a4)
172 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v8
173 ; NOREMAT-NEXT: li a4, 19
174 ; NOREMAT-NEXT: slli a4, a4, 9
175 ; NOREMAT-NEXT: li t2, 19
176 ; NOREMAT-NEXT: sd a4, 608(sp) # 8-byte Folded Spill
177 ; NOREMAT-NEXT: add a4, a7, a4
178 ; NOREMAT-NEXT: vle32.v v8, (a4)
179 ; NOREMAT-NEXT: vle32.v v30, (a4)
180 ; NOREMAT-NEXT: slli a3, a2, 11
181 ; NOREMAT-NEXT: sd a3, 600(sp) # 8-byte Folded Spill
182 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v12
183 ; NOREMAT-NEXT: add a3, a7, a3
184 ; NOREMAT-NEXT: vle32.v v12, (a3)
185 ; NOREMAT-NEXT: vle32.v v4, (a3)
186 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v8
187 ; NOREMAT-NEXT: li s7, 21
188 ; NOREMAT-NEXT: slli a3, s7, 9
189 ; NOREMAT-NEXT: sd a3, 592(sp) # 8-byte Folded Spill
190 ; NOREMAT-NEXT: add a3, a7, a3
191 ; NOREMAT-NEXT: vle32.v v8, (a3)
192 ; NOREMAT-NEXT: vle32.v v6, (a3)
193 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v12
194 ; NOREMAT-NEXT: li a6, 11
195 ; NOREMAT-NEXT: slli a3, a6, 10
196 ; NOREMAT-NEXT: sd a3, 584(sp) # 8-byte Folded Spill
197 ; NOREMAT-NEXT: add a3, a7, a3
198 ; NOREMAT-NEXT: vle32.v v12, (a3)
199 ; NOREMAT-NEXT: vle32.v v30, (a3)
200 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v8
201 ; NOREMAT-NEXT: li s3, 23
202 ; NOREMAT-NEXT: slli a3, s3, 9
203 ; NOREMAT-NEXT: sd a3, 576(sp) # 8-byte Folded Spill
204 ; NOREMAT-NEXT: add a3, a7, a3
205 ; NOREMAT-NEXT: vle32.v v8, (a3)
206 ; NOREMAT-NEXT: vle32.v v4, (a3)
207 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v12
208 ; NOREMAT-NEXT: li s0, 25
209 ; NOREMAT-NEXT: slli a3, s0, 9
210 ; NOREMAT-NEXT: sd a3, 568(sp) # 8-byte Folded Spill
211 ; NOREMAT-NEXT: add a3, a7, a3
212 ; NOREMAT-NEXT: vle32.v v12, (a3)
213 ; NOREMAT-NEXT: vle32.v v6, (a3)
214 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v8
215 ; NOREMAT-NEXT: slli a3, t5, 10
216 ; NOREMAT-NEXT: sd a3, 560(sp) # 8-byte Folded Spill
217 ; NOREMAT-NEXT: add a3, a7, a3
218 ; NOREMAT-NEXT: vle32.v v8, (a3)
219 ; NOREMAT-NEXT: vle32.v v30, (a3)
220 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v28
221 ; NOREMAT-NEXT: li t3, 27
222 ; NOREMAT-NEXT: slli a3, t3, 9
223 ; NOREMAT-NEXT: sd a3, 552(sp) # 8-byte Folded Spill
224 ; NOREMAT-NEXT: add a3, a7, a3
225 ; NOREMAT-NEXT: vle32.v v28, (a3)
226 ; NOREMAT-NEXT: vle32.v v4, (a3)
227 ; NOREMAT-NEXT: slli a2, t0, 11
228 ; NOREMAT-NEXT: sd a2, 544(sp) # 8-byte Folded Spill
229 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v26, v12
230 ; NOREMAT-NEXT: add a2, a7, a2
231 ; NOREMAT-NEXT: vle32.v v12, (a2)
232 ; NOREMAT-NEXT: vle32.v v26, (a2)
233 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v8
234 ; NOREMAT-NEXT: li t0, 29
235 ; NOREMAT-NEXT: slli a2, t0, 9
236 ; NOREMAT-NEXT: sd a2, 536(sp) # 8-byte Folded Spill
237 ; NOREMAT-NEXT: add a2, a7, a2
238 ; NOREMAT-NEXT: vle32.v v8, (a2)
239 ; NOREMAT-NEXT: vle32.v v6, (a2)
240 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v28
241 ; NOREMAT-NEXT: li a3, 15
242 ; NOREMAT-NEXT: slli a2, a3, 10
243 ; NOREMAT-NEXT: sd a2, 528(sp) # 8-byte Folded Spill
244 ; NOREMAT-NEXT: add a2, a7, a2
245 ; NOREMAT-NEXT: vle32.v v28, (a2)
246 ; NOREMAT-NEXT: vle32.v v30, (a2)
247 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v12
248 ; NOREMAT-NEXT: li t1, 31
249 ; NOREMAT-NEXT: slli a2, t1, 9
250 ; NOREMAT-NEXT: sd a2, 520(sp) # 8-byte Folded Spill
251 ; NOREMAT-NEXT: add a2, a7, a2
252 ; NOREMAT-NEXT: vle32.v v12, (a2)
253 ; NOREMAT-NEXT: vle32.v v4, (a2)
254 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v26, v8
255 ; NOREMAT-NEXT: lui a4, 4
256 ; NOREMAT-NEXT: addiw a0, a4, 512
257 ; NOREMAT-NEXT: sd a0, 496(sp) # 8-byte Folded Spill
258 ; NOREMAT-NEXT: add a0, a7, a0
259 ; NOREMAT-NEXT: vle32.v v8, (a0)
260 ; NOREMAT-NEXT: vle32.v v26, (a0)
261 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v28
262 ; NOREMAT-NEXT: slli a2, s1, 10
263 ; NOREMAT-NEXT: sd a2, 488(sp) # 8-byte Folded Spill
264 ; NOREMAT-NEXT: add a2, a7, a2
265 ; NOREMAT-NEXT: vle32.v v28, (a2)
266 ; NOREMAT-NEXT: vle32.v v6, (a2)
267 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v12
268 ; NOREMAT-NEXT: addiw a2, a4, 1536
269 ; NOREMAT-NEXT: sd a2, 480(sp) # 8-byte Folded Spill
270 ; NOREMAT-NEXT: add a2, a7, a2
271 ; NOREMAT-NEXT: vle32.v v12, (a2)
272 ; NOREMAT-NEXT: vle32.v v30, (a2)
273 ; NOREMAT-NEXT: slli a2, a5, 11
274 ; NOREMAT-NEXT: sd a2, 472(sp) # 8-byte Folded Spill
275 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v24
276 ; NOREMAT-NEXT: add a2, a7, a2
277 ; NOREMAT-NEXT: vle32.v v24, (a2)
278 ; NOREMAT-NEXT: vle32.v v4, (a2)
279 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v22, v8
280 ; NOREMAT-NEXT: lui a5, 5
281 ; NOREMAT-NEXT: addiw a2, a5, -1536
282 ; NOREMAT-NEXT: sd a2, 464(sp) # 8-byte Folded Spill
283 ; NOREMAT-NEXT: add a2, a7, a2
284 ; NOREMAT-NEXT: vle32.v v8, (a2)
285 ; NOREMAT-NEXT: vle32.v v22, (a2)
286 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v26, v28
287 ; NOREMAT-NEXT: slli a2, t2, 10
288 ; NOREMAT-NEXT: sd a2, 456(sp) # 8-byte Folded Spill
289 ; NOREMAT-NEXT: li t2, 19
290 ; NOREMAT-NEXT: add a2, a7, a2
291 ; NOREMAT-NEXT: vle32.v v26, (a2)
292 ; NOREMAT-NEXT: vle32.v v28, (a2)
293 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v12
294 ; NOREMAT-NEXT: addiw a2, a5, -512
295 ; NOREMAT-NEXT: sd a2, 448(sp) # 8-byte Folded Spill
296 ; NOREMAT-NEXT: add a2, a7, a2
297 ; NOREMAT-NEXT: vle32.v v12, (a2)
298 ; NOREMAT-NEXT: vle32.v v6, (a2)
299 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v24
300 ; NOREMAT-NEXT: addiw a2, a5, 512
301 ; NOREMAT-NEXT: sd a2, 440(sp) # 8-byte Folded Spill
302 ; NOREMAT-NEXT: add a2, a7, a2
303 ; NOREMAT-NEXT: vle32.v v24, (a2)
304 ; NOREMAT-NEXT: vle32.v v30, (a2)
305 ; NOREMAT-NEXT: slli a2, s7, 10
306 ; NOREMAT-NEXT: sd a2, 432(sp) # 8-byte Folded Spill
307 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v8
308 ; NOREMAT-NEXT: add a2, a7, a2
309 ; NOREMAT-NEXT: vle32.v v8, (a2)
310 ; NOREMAT-NEXT: vle32.v v4, (a2)
311 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v22, v26
312 ; NOREMAT-NEXT: addiw a2, a5, 1536
313 ; NOREMAT-NEXT: sd a2, 424(sp) # 8-byte Folded Spill
314 ; NOREMAT-NEXT: add a2, a7, a2
315 ; NOREMAT-NEXT: vle32.v v22, (a2)
316 ; NOREMAT-NEXT: vle32.v v26, (a2)
317 ; NOREMAT-NEXT: slli a2, a6, 11
318 ; NOREMAT-NEXT: sd a2, 416(sp) # 8-byte Folded Spill
319 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v28, v12
320 ; NOREMAT-NEXT: add a2, a7, a2
321 ; NOREMAT-NEXT: vle32.v v12, (a2)
322 ; NOREMAT-NEXT: vle32.v v28, (a2)
323 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v18
324 ; NOREMAT-NEXT: lui a6, 6
325 ; NOREMAT-NEXT: addiw a2, a6, -1536
326 ; NOREMAT-NEXT: sd a2, 408(sp) # 8-byte Folded Spill
327 ; NOREMAT-NEXT: add a2, a7, a2
328 ; NOREMAT-NEXT: vle32.v v18, (a2)
329 ; NOREMAT-NEXT: vle32.v v6, (a2)
330 ; NOREMAT-NEXT: slli a2, s3, 10
331 ; NOREMAT-NEXT: sd a2, 400(sp) # 8-byte Folded Spill
332 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v16, v24
333 ; NOREMAT-NEXT: add a2, a7, a2
334 ; NOREMAT-NEXT: vle32.v v16, (a2)
335 ; NOREMAT-NEXT: vle32.v v24, (a2)
336 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v8
337 ; NOREMAT-NEXT: addiw a2, a6, -512
338 ; NOREMAT-NEXT: sd a2, 392(sp) # 8-byte Folded Spill
339 ; NOREMAT-NEXT: add a2, a7, a2
340 ; NOREMAT-NEXT: vle32.v v8, (a2)
341 ; NOREMAT-NEXT: vle32.v v30, (a2)
342 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v22
343 ; NOREMAT-NEXT: addiw a2, a6, 512
344 ; NOREMAT-NEXT: sd a2, 384(sp) # 8-byte Folded Spill
345 ; NOREMAT-NEXT: add a2, a7, a2
346 ; NOREMAT-NEXT: vle32.v v22, (a2)
347 ; NOREMAT-NEXT: vle32.v v4, (a2)
348 ; NOREMAT-NEXT: slli a2, s0, 10
349 ; NOREMAT-NEXT: sd a2, 376(sp) # 8-byte Folded Spill
350 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v26, v12
351 ; NOREMAT-NEXT: add a2, a7, a2
352 ; NOREMAT-NEXT: vle32.v v26, (a2)
353 ; NOREMAT-NEXT: vle32.v v2, (a2)
354 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v28, v18
355 ; NOREMAT-NEXT: addiw a2, a6, 1536
356 ; NOREMAT-NEXT: sd a2, 368(sp) # 8-byte Folded Spill
357 ; NOREMAT-NEXT: add a2, a7, a2
358 ; NOREMAT-NEXT: vle32.v v18, (a2)
359 ; NOREMAT-NEXT: vle32.v v28, (a2)
360 ; NOREMAT-NEXT: slli a2, t5, 11
361 ; NOREMAT-NEXT: sd a2, 360(sp) # 8-byte Folded Spill
362 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v16
363 ; NOREMAT-NEXT: add a2, a7, a2
364 ; NOREMAT-NEXT: vle32.v v16, (a2)
365 ; NOREMAT-NEXT: vle32.v v6, (a2)
366 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v24, v8
367 ; NOREMAT-NEXT: lui s0, 7
368 ; NOREMAT-NEXT: addiw a2, s0, -1536
369 ; NOREMAT-NEXT: sd a2, 352(sp) # 8-byte Folded Spill
370 ; NOREMAT-NEXT: add a2, a7, a2
371 ; NOREMAT-NEXT: vle32.v v8, (a2)
372 ; NOREMAT-NEXT: vle32.v v24, (a2)
373 ; NOREMAT-NEXT: slli a2, t3, 10
374 ; NOREMAT-NEXT: sd a2, 344(sp) # 8-byte Folded Spill
375 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v14
376 ; NOREMAT-NEXT: add a2, a7, a2
377 ; NOREMAT-NEXT: vle32.v v14, (a2)
378 ; NOREMAT-NEXT: vle32.v v30, (a2)
379 ; NOREMAT-NEXT: addi a0, sp, 640
380 ; NOREMAT-NEXT: vl2r.v v12, (a0) # Unknown-size Folded Reload
381 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v22
382 ; NOREMAT-NEXT: addiw a2, s0, -512
383 ; NOREMAT-NEXT: sd a2, 336(sp) # 8-byte Folded Spill
384 ; NOREMAT-NEXT: add a2, a7, a2
385 ; NOREMAT-NEXT: vle32.v v22, (a2)
386 ; NOREMAT-NEXT: vle32.v v12, (a2)
387 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v26
388 ; NOREMAT-NEXT: addiw a2, s0, 512
389 ; NOREMAT-NEXT: sd a2, 328(sp) # 8-byte Folded Spill
390 ; NOREMAT-NEXT: lui t3, 7
391 ; NOREMAT-NEXT: add a2, a7, a2
392 ; NOREMAT-NEXT: vle32.v v26, (a2)
393 ; NOREMAT-NEXT: vle32.v v4, (a2)
394 ; NOREMAT-NEXT: slli a2, t0, 10
395 ; NOREMAT-NEXT: sd a2, 320(sp) # 8-byte Folded Spill
396 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v2, v18
397 ; NOREMAT-NEXT: add a2, a7, a2
398 ; NOREMAT-NEXT: vle32.v v18, (a2)
399 ; NOREMAT-NEXT: vle32.v v2, (a2)
400 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v28, v16
401 ; NOREMAT-NEXT: addiw a2, t3, 1536
402 ; NOREMAT-NEXT: sd a2, 312(sp) # 8-byte Folded Spill
403 ; NOREMAT-NEXT: add a2, a7, a2
404 ; NOREMAT-NEXT: vle32.v v16, (a2)
405 ; NOREMAT-NEXT: vle32.v v28, (a2)
406 ; NOREMAT-NEXT: slli a2, a3, 11
407 ; NOREMAT-NEXT: sd a2, 304(sp) # 8-byte Folded Spill
408 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v8
409 ; NOREMAT-NEXT: add a2, a7, a2
410 ; NOREMAT-NEXT: vle32.v v8, (a2)
411 ; NOREMAT-NEXT: vle32.v v6, (a2)
412 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v24, v14
413 ; NOREMAT-NEXT: addiw a2, t4, -1536
414 ; NOREMAT-NEXT: sd a2, 296(sp) # 8-byte Folded Spill
415 ; NOREMAT-NEXT: add a2, a7, a2
416 ; NOREMAT-NEXT: vle32.v v14, (a2)
417 ; NOREMAT-NEXT: vle32.v v24, (a2)
418 ; NOREMAT-NEXT: slli a2, t1, 10
419 ; NOREMAT-NEXT: sd a2, 288(sp) # 8-byte Folded Spill
420 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v22
421 ; NOREMAT-NEXT: add a2, a7, a2
422 ; NOREMAT-NEXT: vle32.v v22, (a2)
423 ; NOREMAT-NEXT: vle32.v v30, (a2)
424 ; NOREMAT-NEXT: addiw a0, t4, -512
425 ; NOREMAT-NEXT: sd a0, 280(sp) # 8-byte Folded Spill
426 ; NOREMAT-NEXT: add a0, a7, a0
427 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v12, v0
428 ; NOREMAT-NEXT: vle32.v v12, (a0)
429 ; NOREMAT-NEXT: vle32.v v0, (a0)
430 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v10, v26
431 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v4, v18
432 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v2, v16
433 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v28, v8
434 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v6, v14
435 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v24, v22
436 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v30, v12
437 ; NOREMAT-NEXT: sf.vc.vv 3, 0, v0, v20
438 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
439 ; NOREMAT-NEXT: addi a0, a1, 1024
440 ; NOREMAT-NEXT: vse32.v v8, (a0)
441 ; NOREMAT-NEXT: add s11, a1, s11
442 ; NOREMAT-NEXT: sd s11, 272(sp) # 8-byte Folded Spill
443 ; NOREMAT-NEXT: lui a0, 2
444 ; NOREMAT-NEXT: add a0, a1, a0
445 ; NOREMAT-NEXT: sd a0, 264(sp) # 8-byte Folded Spill
446 ; NOREMAT-NEXT: lui a0, 3
447 ; NOREMAT-NEXT: add a0, a1, a0
448 ; NOREMAT-NEXT: sd a0, 256(sp) # 8-byte Folded Spill
449 ; NOREMAT-NEXT: add a4, a1, a4
450 ; NOREMAT-NEXT: sd a4, 248(sp) # 8-byte Folded Spill
451 ; NOREMAT-NEXT: add a5, a1, a5
452 ; NOREMAT-NEXT: sd a5, 240(sp) # 8-byte Folded Spill
453 ; NOREMAT-NEXT: add a6, a1, a6
454 ; NOREMAT-NEXT: sd a6, 232(sp) # 8-byte Folded Spill
455 ; NOREMAT-NEXT: add t3, a1, t3
456 ; NOREMAT-NEXT: sd t3, 224(sp) # 8-byte Folded Spill
457 ; NOREMAT-NEXT: add a0, a1, t4
458 ; NOREMAT-NEXT: sd a0, 216(sp) # 8-byte Folded Spill
459 ; NOREMAT-NEXT: addiw a0, t4, 512
460 ; NOREMAT-NEXT: sd a0, 192(sp) # 8-byte Folded Spill
461 ; NOREMAT-NEXT: addiw a0, t4, 1024
462 ; NOREMAT-NEXT: sd a0, 176(sp) # 8-byte Folded Spill
463 ; NOREMAT-NEXT: addiw a0, t4, 1536
464 ; NOREMAT-NEXT: sd a0, 160(sp) # 8-byte Folded Spill
465 ; NOREMAT-NEXT: slli s1, s1, 11
466 ; NOREMAT-NEXT: sd s1, 128(sp) # 8-byte Folded Spill
467 ; NOREMAT-NEXT: lui a0, 9
468 ; NOREMAT-NEXT: addiw a2, a0, -1536
469 ; NOREMAT-NEXT: sd a2, 88(sp) # 8-byte Folded Spill
470 ; NOREMAT-NEXT: addiw a2, a0, -1024
471 ; NOREMAT-NEXT: sd a2, 72(sp) # 8-byte Folded Spill
472 ; NOREMAT-NEXT: addiw a2, a0, -512
473 ; NOREMAT-NEXT: sd a2, 40(sp) # 8-byte Folded Spill
474 ; NOREMAT-NEXT: add a2, a1, a0
475 ; NOREMAT-NEXT: sd a2, 208(sp) # 8-byte Folded Spill
476 ; NOREMAT-NEXT: addiw s11, a0, 512
477 ; NOREMAT-NEXT: addiw s7, a0, 1024
478 ; NOREMAT-NEXT: addiw s3, a0, 1536
479 ; NOREMAT-NEXT: slli s1, t2, 11
480 ; NOREMAT-NEXT: lui a0, 10
481 ; NOREMAT-NEXT: addiw t2, a0, -1536
482 ; NOREMAT-NEXT: addiw a7, a0, -1024
483 ; NOREMAT-NEXT: addiw a4, a0, -512
484 ; NOREMAT-NEXT: add a2, a1, a0
485 ; NOREMAT-NEXT: sd a2, 200(sp) # 8-byte Folded Spill
486 ; NOREMAT-NEXT: addiw a0, a0, 512
487 ; NOREMAT-NEXT: ld a2, 512(sp) # 8-byte Folded Reload
488 ; NOREMAT-NEXT: add a2, a1, a2
489 ; NOREMAT-NEXT: ld a3, 504(sp) # 8-byte Folded Reload
490 ; NOREMAT-NEXT: add a3, a1, a3
491 ; NOREMAT-NEXT: add a5, a1, t6
492 ; NOREMAT-NEXT: add a6, a1, s2
493 ; NOREMAT-NEXT: add t0, a1, s4
494 ; NOREMAT-NEXT: add t1, a1, s5
495 ; NOREMAT-NEXT: add t3, a1, s6
496 ; NOREMAT-NEXT: add t4, a1, s8
497 ; NOREMAT-NEXT: add t5, a1, s9
498 ; NOREMAT-NEXT: add t6, a1, s10
499 ; NOREMAT-NEXT: add s0, a1, ra
500 ; NOREMAT-NEXT: ld s2, 624(sp) # 8-byte Folded Reload
501 ; NOREMAT-NEXT: add s2, a1, s2
502 ; NOREMAT-NEXT: ld s4, 616(sp) # 8-byte Folded Reload
503 ; NOREMAT-NEXT: add s4, a1, s4
504 ; NOREMAT-NEXT: ld s5, 608(sp) # 8-byte Folded Reload
505 ; NOREMAT-NEXT: add s5, a1, s5
506 ; NOREMAT-NEXT: ld s6, 600(sp) # 8-byte Folded Reload
507 ; NOREMAT-NEXT: add s6, a1, s6
508 ; NOREMAT-NEXT: ld s8, 592(sp) # 8-byte Folded Reload
509 ; NOREMAT-NEXT: add s8, a1, s8
510 ; NOREMAT-NEXT: ld s9, 584(sp) # 8-byte Folded Reload
511 ; NOREMAT-NEXT: add s9, a1, s9
512 ; NOREMAT-NEXT: ld s10, 576(sp) # 8-byte Folded Reload
513 ; NOREMAT-NEXT: add s10, a1, s10
514 ; NOREMAT-NEXT: ld ra, 568(sp) # 8-byte Folded Reload
515 ; NOREMAT-NEXT: add ra, a1, ra
516 ; NOREMAT-NEXT: sd ra, 16(sp) # 8-byte Folded Spill
517 ; NOREMAT-NEXT: ld ra, 560(sp) # 8-byte Folded Reload
518 ; NOREMAT-NEXT: add ra, a1, ra
519 ; NOREMAT-NEXT: sd ra, 24(sp) # 8-byte Folded Spill
520 ; NOREMAT-NEXT: ld ra, 552(sp) # 8-byte Folded Reload
521 ; NOREMAT-NEXT: add ra, a1, ra
522 ; NOREMAT-NEXT: sd ra, 32(sp) # 8-byte Folded Spill
523 ; NOREMAT-NEXT: ld ra, 544(sp) # 8-byte Folded Reload
524 ; NOREMAT-NEXT: add ra, a1, ra
525 ; NOREMAT-NEXT: sd ra, 48(sp) # 8-byte Folded Spill
526 ; NOREMAT-NEXT: ld ra, 536(sp) # 8-byte Folded Reload
527 ; NOREMAT-NEXT: add ra, a1, ra
528 ; NOREMAT-NEXT: sd ra, 56(sp) # 8-byte Folded Spill
529 ; NOREMAT-NEXT: ld ra, 528(sp) # 8-byte Folded Reload
530 ; NOREMAT-NEXT: add ra, a1, ra
531 ; NOREMAT-NEXT: sd ra, 64(sp) # 8-byte Folded Spill
532 ; NOREMAT-NEXT: ld ra, 520(sp) # 8-byte Folded Reload
533 ; NOREMAT-NEXT: add ra, a1, ra
534 ; NOREMAT-NEXT: sd ra, 80(sp) # 8-byte Folded Spill
535 ; NOREMAT-NEXT: ld ra, 496(sp) # 8-byte Folded Reload
536 ; NOREMAT-NEXT: add ra, a1, ra
537 ; NOREMAT-NEXT: sd ra, 96(sp) # 8-byte Folded Spill
538 ; NOREMAT-NEXT: ld ra, 488(sp) # 8-byte Folded Reload
539 ; NOREMAT-NEXT: add ra, a1, ra
540 ; NOREMAT-NEXT: sd ra, 104(sp) # 8-byte Folded Spill
541 ; NOREMAT-NEXT: ld ra, 480(sp) # 8-byte Folded Reload
542 ; NOREMAT-NEXT: add ra, a1, ra
543 ; NOREMAT-NEXT: sd ra, 112(sp) # 8-byte Folded Spill
544 ; NOREMAT-NEXT: ld ra, 472(sp) # 8-byte Folded Reload
545 ; NOREMAT-NEXT: add ra, a1, ra
546 ; NOREMAT-NEXT: sd ra, 120(sp) # 8-byte Folded Spill
547 ; NOREMAT-NEXT: ld ra, 464(sp) # 8-byte Folded Reload
548 ; NOREMAT-NEXT: add ra, a1, ra
549 ; NOREMAT-NEXT: sd ra, 136(sp) # 8-byte Folded Spill
550 ; NOREMAT-NEXT: ld ra, 456(sp) # 8-byte Folded Reload
551 ; NOREMAT-NEXT: add ra, a1, ra
552 ; NOREMAT-NEXT: sd ra, 144(sp) # 8-byte Folded Spill
553 ; NOREMAT-NEXT: ld ra, 448(sp) # 8-byte Folded Reload
554 ; NOREMAT-NEXT: add ra, a1, ra
555 ; NOREMAT-NEXT: sd ra, 152(sp) # 8-byte Folded Spill
556 ; NOREMAT-NEXT: ld ra, 440(sp) # 8-byte Folded Reload
557 ; NOREMAT-NEXT: add ra, a1, ra
558 ; NOREMAT-NEXT: sd ra, 168(sp) # 8-byte Folded Spill
559 ; NOREMAT-NEXT: ld ra, 432(sp) # 8-byte Folded Reload
560 ; NOREMAT-NEXT: add ra, a1, ra
561 ; NOREMAT-NEXT: sd ra, 184(sp) # 8-byte Folded Spill
562 ; NOREMAT-NEXT: ld ra, 424(sp) # 8-byte Folded Reload
563 ; NOREMAT-NEXT: add ra, a1, ra
564 ; NOREMAT-NEXT: sd ra, 424(sp) # 8-byte Folded Spill
565 ; NOREMAT-NEXT: ld ra, 416(sp) # 8-byte Folded Reload
566 ; NOREMAT-NEXT: add ra, a1, ra
567 ; NOREMAT-NEXT: sd ra, 432(sp) # 8-byte Folded Spill
568 ; NOREMAT-NEXT: ld ra, 408(sp) # 8-byte Folded Reload
569 ; NOREMAT-NEXT: add ra, a1, ra
570 ; NOREMAT-NEXT: sd ra, 440(sp) # 8-byte Folded Spill
571 ; NOREMAT-NEXT: ld ra, 400(sp) # 8-byte Folded Reload
572 ; NOREMAT-NEXT: add ra, a1, ra
573 ; NOREMAT-NEXT: sd ra, 448(sp) # 8-byte Folded Spill
574 ; NOREMAT-NEXT: ld ra, 392(sp) # 8-byte Folded Reload
575 ; NOREMAT-NEXT: add ra, a1, ra
576 ; NOREMAT-NEXT: sd ra, 456(sp) # 8-byte Folded Spill
577 ; NOREMAT-NEXT: ld ra, 384(sp) # 8-byte Folded Reload
578 ; NOREMAT-NEXT: add ra, a1, ra
579 ; NOREMAT-NEXT: sd ra, 464(sp) # 8-byte Folded Spill
580 ; NOREMAT-NEXT: ld ra, 376(sp) # 8-byte Folded Reload
581 ; NOREMAT-NEXT: add ra, a1, ra
582 ; NOREMAT-NEXT: sd ra, 472(sp) # 8-byte Folded Spill
583 ; NOREMAT-NEXT: ld ra, 368(sp) # 8-byte Folded Reload
584 ; NOREMAT-NEXT: add ra, a1, ra
585 ; NOREMAT-NEXT: sd ra, 480(sp) # 8-byte Folded Spill
586 ; NOREMAT-NEXT: ld ra, 360(sp) # 8-byte Folded Reload
587 ; NOREMAT-NEXT: add ra, a1, ra
588 ; NOREMAT-NEXT: sd ra, 488(sp) # 8-byte Folded Spill
589 ; NOREMAT-NEXT: ld ra, 352(sp) # 8-byte Folded Reload
590 ; NOREMAT-NEXT: add ra, a1, ra
591 ; NOREMAT-NEXT: sd ra, 496(sp) # 8-byte Folded Spill
592 ; NOREMAT-NEXT: ld ra, 344(sp) # 8-byte Folded Reload
593 ; NOREMAT-NEXT: add ra, a1, ra
594 ; NOREMAT-NEXT: sd ra, 504(sp) # 8-byte Folded Spill
595 ; NOREMAT-NEXT: ld ra, 336(sp) # 8-byte Folded Reload
596 ; NOREMAT-NEXT: add ra, a1, ra
597 ; NOREMAT-NEXT: sd ra, 512(sp) # 8-byte Folded Spill
598 ; NOREMAT-NEXT: ld ra, 328(sp) # 8-byte Folded Reload
599 ; NOREMAT-NEXT: add ra, a1, ra
600 ; NOREMAT-NEXT: sd ra, 520(sp) # 8-byte Folded Spill
601 ; NOREMAT-NEXT: ld ra, 320(sp) # 8-byte Folded Reload
602 ; NOREMAT-NEXT: add ra, a1, ra
603 ; NOREMAT-NEXT: sd ra, 528(sp) # 8-byte Folded Spill
604 ; NOREMAT-NEXT: ld ra, 312(sp) # 8-byte Folded Reload
605 ; NOREMAT-NEXT: add ra, a1, ra
606 ; NOREMAT-NEXT: sd ra, 536(sp) # 8-byte Folded Spill
607 ; NOREMAT-NEXT: ld ra, 304(sp) # 8-byte Folded Reload
608 ; NOREMAT-NEXT: add ra, a1, ra
609 ; NOREMAT-NEXT: sd ra, 544(sp) # 8-byte Folded Spill
610 ; NOREMAT-NEXT: ld ra, 296(sp) # 8-byte Folded Reload
611 ; NOREMAT-NEXT: add ra, a1, ra
612 ; NOREMAT-NEXT: sd ra, 552(sp) # 8-byte Folded Spill
613 ; NOREMAT-NEXT: ld ra, 288(sp) # 8-byte Folded Reload
614 ; NOREMAT-NEXT: add ra, a1, ra
615 ; NOREMAT-NEXT: sd ra, 560(sp) # 8-byte Folded Spill
616 ; NOREMAT-NEXT: ld ra, 280(sp) # 8-byte Folded Reload
617 ; NOREMAT-NEXT: add ra, a1, ra
618 ; NOREMAT-NEXT: sd ra, 568(sp) # 8-byte Folded Spill
619 ; NOREMAT-NEXT: ld ra, 192(sp) # 8-byte Folded Reload
620 ; NOREMAT-NEXT: add ra, a1, ra
621 ; NOREMAT-NEXT: sd ra, 576(sp) # 8-byte Folded Spill
622 ; NOREMAT-NEXT: ld ra, 176(sp) # 8-byte Folded Reload
623 ; NOREMAT-NEXT: add ra, a1, ra
624 ; NOREMAT-NEXT: sd ra, 584(sp) # 8-byte Folded Spill
625 ; NOREMAT-NEXT: ld ra, 160(sp) # 8-byte Folded Reload
626 ; NOREMAT-NEXT: add ra, a1, ra
627 ; NOREMAT-NEXT: sd ra, 592(sp) # 8-byte Folded Spill
628 ; NOREMAT-NEXT: ld ra, 128(sp) # 8-byte Folded Reload
629 ; NOREMAT-NEXT: add ra, a1, ra
630 ; NOREMAT-NEXT: sd ra, 600(sp) # 8-byte Folded Spill
631 ; NOREMAT-NEXT: ld ra, 88(sp) # 8-byte Folded Reload
632 ; NOREMAT-NEXT: add ra, a1, ra
633 ; NOREMAT-NEXT: sd ra, 608(sp) # 8-byte Folded Spill
634 ; NOREMAT-NEXT: ld ra, 72(sp) # 8-byte Folded Reload
635 ; NOREMAT-NEXT: add ra, a1, ra
636 ; NOREMAT-NEXT: sd ra, 616(sp) # 8-byte Folded Spill
637 ; NOREMAT-NEXT: ld ra, 40(sp) # 8-byte Folded Reload
638 ; NOREMAT-NEXT: add ra, a1, ra
639 ; NOREMAT-NEXT: sd ra, 624(sp) # 8-byte Folded Spill
640 ; NOREMAT-NEXT: add ra, a1, s11
641 ; NOREMAT-NEXT: add s11, a1, s7
642 ; NOREMAT-NEXT: add s7, a1, s3
643 ; NOREMAT-NEXT: add s3, a1, s1
644 ; NOREMAT-NEXT: add s1, a1, t2
645 ; NOREMAT-NEXT: add t2, a1, a7
646 ; NOREMAT-NEXT: add a7, a1, a4
647 ; NOREMAT-NEXT: add a4, a1, a0
648 ; NOREMAT-NEXT: addi a0, a1, 1536
649 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
650 ; NOREMAT-NEXT: vse32.v v8, (a0)
651 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
652 ; NOREMAT-NEXT: vse32.v v8, (a2)
653 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
654 ; NOREMAT-NEXT: vse32.v v8, (a3)
655 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
656 ; NOREMAT-NEXT: vse32.v v8, (a5)
657 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
658 ; NOREMAT-NEXT: vse32.v v8, (a6)
659 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
660 ; NOREMAT-NEXT: ld a0, 272(sp) # 8-byte Folded Reload
661 ; NOREMAT-NEXT: vse32.v v8, (a0)
662 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
663 ; NOREMAT-NEXT: vse32.v v8, (t0)
664 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
665 ; NOREMAT-NEXT: vse32.v v8, (t1)
666 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
667 ; NOREMAT-NEXT: vse32.v v8, (t3)
668 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
669 ; NOREMAT-NEXT: vse32.v v8, (t4)
670 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
671 ; NOREMAT-NEXT: vse32.v v8, (t5)
672 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
673 ; NOREMAT-NEXT: vse32.v v8, (t6)
674 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
675 ; NOREMAT-NEXT: vse32.v v8, (s0)
676 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
677 ; NOREMAT-NEXT: ld a0, 264(sp) # 8-byte Folded Reload
678 ; NOREMAT-NEXT: vse32.v v8, (a0)
679 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
680 ; NOREMAT-NEXT: vse32.v v8, (s2)
681 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
682 ; NOREMAT-NEXT: vse32.v v8, (s4)
683 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
684 ; NOREMAT-NEXT: vse32.v v8, (s5)
685 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
686 ; NOREMAT-NEXT: vse32.v v8, (s6)
687 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
688 ; NOREMAT-NEXT: vse32.v v8, (s8)
689 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
690 ; NOREMAT-NEXT: vse32.v v8, (s9)
691 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
692 ; NOREMAT-NEXT: vse32.v v8, (s10)
693 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
694 ; NOREMAT-NEXT: ld a0, 256(sp) # 8-byte Folded Reload
695 ; NOREMAT-NEXT: vse32.v v8, (a0)
696 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
697 ; NOREMAT-NEXT: ld a0, 16(sp) # 8-byte Folded Reload
698 ; NOREMAT-NEXT: vse32.v v8, (a0)
699 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
700 ; NOREMAT-NEXT: ld a0, 24(sp) # 8-byte Folded Reload
701 ; NOREMAT-NEXT: vse32.v v8, (a0)
702 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
703 ; NOREMAT-NEXT: ld a0, 32(sp) # 8-byte Folded Reload
704 ; NOREMAT-NEXT: vse32.v v8, (a0)
705 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
706 ; NOREMAT-NEXT: ld a0, 48(sp) # 8-byte Folded Reload
707 ; NOREMAT-NEXT: vse32.v v8, (a0)
708 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
709 ; NOREMAT-NEXT: ld a0, 56(sp) # 8-byte Folded Reload
710 ; NOREMAT-NEXT: vse32.v v8, (a0)
711 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
712 ; NOREMAT-NEXT: ld a0, 64(sp) # 8-byte Folded Reload
713 ; NOREMAT-NEXT: vse32.v v8, (a0)
714 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
715 ; NOREMAT-NEXT: ld a0, 80(sp) # 8-byte Folded Reload
716 ; NOREMAT-NEXT: vse32.v v8, (a0)
717 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
718 ; NOREMAT-NEXT: ld a0, 248(sp) # 8-byte Folded Reload
719 ; NOREMAT-NEXT: vse32.v v8, (a0)
720 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
721 ; NOREMAT-NEXT: ld a0, 96(sp) # 8-byte Folded Reload
722 ; NOREMAT-NEXT: vse32.v v8, (a0)
723 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
724 ; NOREMAT-NEXT: ld a0, 104(sp) # 8-byte Folded Reload
725 ; NOREMAT-NEXT: vse32.v v8, (a0)
726 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
727 ; NOREMAT-NEXT: ld a0, 112(sp) # 8-byte Folded Reload
728 ; NOREMAT-NEXT: vse32.v v8, (a0)
729 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
730 ; NOREMAT-NEXT: ld a0, 120(sp) # 8-byte Folded Reload
731 ; NOREMAT-NEXT: vse32.v v8, (a0)
732 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
733 ; NOREMAT-NEXT: ld a0, 136(sp) # 8-byte Folded Reload
734 ; NOREMAT-NEXT: vse32.v v8, (a0)
735 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
736 ; NOREMAT-NEXT: ld a0, 144(sp) # 8-byte Folded Reload
737 ; NOREMAT-NEXT: vse32.v v8, (a0)
738 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
739 ; NOREMAT-NEXT: ld a0, 152(sp) # 8-byte Folded Reload
740 ; NOREMAT-NEXT: vse32.v v8, (a0)
741 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
742 ; NOREMAT-NEXT: ld a0, 240(sp) # 8-byte Folded Reload
743 ; NOREMAT-NEXT: vse32.v v8, (a0)
744 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
745 ; NOREMAT-NEXT: ld a0, 168(sp) # 8-byte Folded Reload
746 ; NOREMAT-NEXT: vse32.v v8, (a0)
747 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
748 ; NOREMAT-NEXT: ld a0, 184(sp) # 8-byte Folded Reload
749 ; NOREMAT-NEXT: vse32.v v8, (a0)
750 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
751 ; NOREMAT-NEXT: ld a0, 424(sp) # 8-byte Folded Reload
752 ; NOREMAT-NEXT: vse32.v v8, (a0)
753 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
754 ; NOREMAT-NEXT: ld a0, 432(sp) # 8-byte Folded Reload
755 ; NOREMAT-NEXT: vse32.v v8, (a0)
756 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
757 ; NOREMAT-NEXT: ld a0, 440(sp) # 8-byte Folded Reload
758 ; NOREMAT-NEXT: vse32.v v8, (a0)
759 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
760 ; NOREMAT-NEXT: ld a0, 448(sp) # 8-byte Folded Reload
761 ; NOREMAT-NEXT: vse32.v v8, (a0)
762 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
763 ; NOREMAT-NEXT: ld a0, 456(sp) # 8-byte Folded Reload
764 ; NOREMAT-NEXT: vse32.v v8, (a0)
765 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
766 ; NOREMAT-NEXT: ld a0, 232(sp) # 8-byte Folded Reload
767 ; NOREMAT-NEXT: vse32.v v8, (a0)
768 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
769 ; NOREMAT-NEXT: ld a0, 464(sp) # 8-byte Folded Reload
770 ; NOREMAT-NEXT: vse32.v v8, (a0)
771 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
772 ; NOREMAT-NEXT: ld a0, 472(sp) # 8-byte Folded Reload
773 ; NOREMAT-NEXT: vse32.v v8, (a0)
774 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
775 ; NOREMAT-NEXT: ld a0, 480(sp) # 8-byte Folded Reload
776 ; NOREMAT-NEXT: vse32.v v8, (a0)
777 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
778 ; NOREMAT-NEXT: ld a0, 488(sp) # 8-byte Folded Reload
779 ; NOREMAT-NEXT: vse32.v v8, (a0)
780 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
781 ; NOREMAT-NEXT: ld a0, 496(sp) # 8-byte Folded Reload
782 ; NOREMAT-NEXT: vse32.v v8, (a0)
783 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
784 ; NOREMAT-NEXT: ld a0, 504(sp) # 8-byte Folded Reload
785 ; NOREMAT-NEXT: vse32.v v8, (a0)
786 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
787 ; NOREMAT-NEXT: ld a0, 512(sp) # 8-byte Folded Reload
788 ; NOREMAT-NEXT: vse32.v v8, (a0)
789 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
790 ; NOREMAT-NEXT: ld a0, 224(sp) # 8-byte Folded Reload
791 ; NOREMAT-NEXT: vse32.v v8, (a0)
792 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
793 ; NOREMAT-NEXT: ld a0, 520(sp) # 8-byte Folded Reload
794 ; NOREMAT-NEXT: vse32.v v8, (a0)
795 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
796 ; NOREMAT-NEXT: ld a0, 528(sp) # 8-byte Folded Reload
797 ; NOREMAT-NEXT: vse32.v v8, (a0)
798 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
799 ; NOREMAT-NEXT: ld a0, 536(sp) # 8-byte Folded Reload
800 ; NOREMAT-NEXT: vse32.v v8, (a0)
801 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
802 ; NOREMAT-NEXT: ld a0, 544(sp) # 8-byte Folded Reload
803 ; NOREMAT-NEXT: vse32.v v8, (a0)
804 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
805 ; NOREMAT-NEXT: ld a0, 552(sp) # 8-byte Folded Reload
806 ; NOREMAT-NEXT: vse32.v v8, (a0)
807 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
808 ; NOREMAT-NEXT: ld a0, 560(sp) # 8-byte Folded Reload
809 ; NOREMAT-NEXT: vse32.v v8, (a0)
810 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
811 ; NOREMAT-NEXT: ld a0, 568(sp) # 8-byte Folded Reload
812 ; NOREMAT-NEXT: vse32.v v8, (a0)
813 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
814 ; NOREMAT-NEXT: ld a0, 216(sp) # 8-byte Folded Reload
815 ; NOREMAT-NEXT: vse32.v v8, (a0)
816 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
817 ; NOREMAT-NEXT: ld a0, 576(sp) # 8-byte Folded Reload
818 ; NOREMAT-NEXT: vse32.v v8, (a0)
819 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
820 ; NOREMAT-NEXT: ld a0, 584(sp) # 8-byte Folded Reload
821 ; NOREMAT-NEXT: vse32.v v8, (a0)
822 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
823 ; NOREMAT-NEXT: ld a0, 592(sp) # 8-byte Folded Reload
824 ; NOREMAT-NEXT: vse32.v v8, (a0)
825 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
826 ; NOREMAT-NEXT: ld a0, 600(sp) # 8-byte Folded Reload
827 ; NOREMAT-NEXT: vse32.v v8, (a0)
828 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
829 ; NOREMAT-NEXT: ld a0, 608(sp) # 8-byte Folded Reload
830 ; NOREMAT-NEXT: vse32.v v8, (a0)
831 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
832 ; NOREMAT-NEXT: ld a0, 616(sp) # 8-byte Folded Reload
833 ; NOREMAT-NEXT: vse32.v v8, (a0)
834 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
835 ; NOREMAT-NEXT: ld a0, 624(sp) # 8-byte Folded Reload
836 ; NOREMAT-NEXT: vse32.v v8, (a0)
837 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
838 ; NOREMAT-NEXT: ld a0, 208(sp) # 8-byte Folded Reload
839 ; NOREMAT-NEXT: vse32.v v8, (a0)
840 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
841 ; NOREMAT-NEXT: vse32.v v8, (ra)
842 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
843 ; NOREMAT-NEXT: vse32.v v8, (s11)
844 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
845 ; NOREMAT-NEXT: vse32.v v8, (s7)
846 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
847 ; NOREMAT-NEXT: vse32.v v8, (s3)
848 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
849 ; NOREMAT-NEXT: vse32.v v8, (s1)
850 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
851 ; NOREMAT-NEXT: vse32.v v8, (t2)
852 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
853 ; NOREMAT-NEXT: vse32.v v8, (a7)
854 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
855 ; NOREMAT-NEXT: ld a0, 200(sp) # 8-byte Folded Reload
856 ; NOREMAT-NEXT: vse32.v v8, (a0)
857 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
858 ; NOREMAT-NEXT: vse32.v v8, (a4)
859 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
860 ; NOREMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
861 ; NOREMAT-NEXT: csrr a0, vlenb
862 ; NOREMAT-NEXT: slli a0, a0, 1
863 ; NOREMAT-NEXT: add sp, sp, a0
864 ; NOREMAT-NEXT: .cfi_def_cfa sp, 752
865 ; NOREMAT-NEXT: ld ra, 744(sp) # 8-byte Folded Reload
866 ; NOREMAT-NEXT: ld s0, 736(sp) # 8-byte Folded Reload
867 ; NOREMAT-NEXT: ld s1, 728(sp) # 8-byte Folded Reload
868 ; NOREMAT-NEXT: ld s2, 720(sp) # 8-byte Folded Reload
869 ; NOREMAT-NEXT: ld s3, 712(sp) # 8-byte Folded Reload
870 ; NOREMAT-NEXT: ld s4, 704(sp) # 8-byte Folded Reload
871 ; NOREMAT-NEXT: ld s5, 696(sp) # 8-byte Folded Reload
872 ; NOREMAT-NEXT: ld s6, 688(sp) # 8-byte Folded Reload
873 ; NOREMAT-NEXT: ld s7, 680(sp) # 8-byte Folded Reload
874 ; NOREMAT-NEXT: ld s8, 672(sp) # 8-byte Folded Reload
875 ; NOREMAT-NEXT: ld s9, 664(sp) # 8-byte Folded Reload
876 ; NOREMAT-NEXT: ld s10, 656(sp) # 8-byte Folded Reload
877 ; NOREMAT-NEXT: ld s11, 648(sp) # 8-byte Folded Reload
878 ; NOREMAT-NEXT: .cfi_restore ra
879 ; NOREMAT-NEXT: .cfi_restore s0
880 ; NOREMAT-NEXT: .cfi_restore s1
881 ; NOREMAT-NEXT: .cfi_restore s2
882 ; NOREMAT-NEXT: .cfi_restore s3
883 ; NOREMAT-NEXT: .cfi_restore s4
884 ; NOREMAT-NEXT: .cfi_restore s5
885 ; NOREMAT-NEXT: .cfi_restore s6
886 ; NOREMAT-NEXT: .cfi_restore s7
887 ; NOREMAT-NEXT: .cfi_restore s8
888 ; NOREMAT-NEXT: .cfi_restore s9
889 ; NOREMAT-NEXT: .cfi_restore s10
890 ; NOREMAT-NEXT: .cfi_restore s11
891 ; NOREMAT-NEXT: addi sp, sp, 752
892 ; NOREMAT-NEXT: .cfi_def_cfa_offset 0
897 ; REMAT-NEXT: addi sp, sp, -544
898 ; REMAT-NEXT: .cfi_def_cfa_offset 544
899 ; REMAT-NEXT: sd ra, 536(sp) # 8-byte Folded Spill
900 ; REMAT-NEXT: sd s0, 528(sp) # 8-byte Folded Spill
901 ; REMAT-NEXT: sd s1, 520(sp) # 8-byte Folded Spill
902 ; REMAT-NEXT: sd s2, 512(sp) # 8-byte Folded Spill
903 ; REMAT-NEXT: sd s3, 504(sp) # 8-byte Folded Spill
904 ; REMAT-NEXT: sd s4, 496(sp) # 8-byte Folded Spill
905 ; REMAT-NEXT: sd s5, 488(sp) # 8-byte Folded Spill
906 ; REMAT-NEXT: sd s6, 480(sp) # 8-byte Folded Spill
907 ; REMAT-NEXT: sd s7, 472(sp) # 8-byte Folded Spill
908 ; REMAT-NEXT: sd s8, 464(sp) # 8-byte Folded Spill
909 ; REMAT-NEXT: sd s9, 456(sp) # 8-byte Folded Spill
910 ; REMAT-NEXT: sd s10, 448(sp) # 8-byte Folded Spill
911 ; REMAT-NEXT: sd s11, 440(sp) # 8-byte Folded Spill
912 ; REMAT-NEXT: .cfi_offset ra, -8
913 ; REMAT-NEXT: .cfi_offset s0, -16
914 ; REMAT-NEXT: .cfi_offset s1, -24
915 ; REMAT-NEXT: .cfi_offset s2, -32
916 ; REMAT-NEXT: .cfi_offset s3, -40
917 ; REMAT-NEXT: .cfi_offset s4, -48
918 ; REMAT-NEXT: .cfi_offset s5, -56
919 ; REMAT-NEXT: .cfi_offset s6, -64
920 ; REMAT-NEXT: .cfi_offset s7, -72
921 ; REMAT-NEXT: .cfi_offset s8, -80
922 ; REMAT-NEXT: .cfi_offset s9, -88
923 ; REMAT-NEXT: .cfi_offset s10, -96
924 ; REMAT-NEXT: .cfi_offset s11, -104
925 ; REMAT-NEXT: csrr a2, vlenb
926 ; REMAT-NEXT: li a3, 18
927 ; REMAT-NEXT: mul a2, a2, a3
928 ; REMAT-NEXT: sub sp, sp, a2
929 ; REMAT-NEXT: .cfi_escape 0x0f, 0x0e, 0x72, 0x00, 0x11, 0xa0, 0x04, 0x22, 0x11, 0x12, 0x92, 0xa2, 0x38, 0x00, 0x1e, 0x22 # sp + 544 + 18 * vlenb
930 ; REMAT-NEXT: li a4, 32
931 ; REMAT-NEXT: addi a5, a0, 512
932 ; REMAT-NEXT: addi a3, a0, 1024
933 ; REMAT-NEXT: addi a2, a0, 1536
934 ; REMAT-NEXT: li a6, 1
935 ; REMAT-NEXT: slli a6, a6, 11
936 ; REMAT-NEXT: li a7, 5
937 ; REMAT-NEXT: slli a7, a7, 9
938 ; REMAT-NEXT: li t0, 3
939 ; REMAT-NEXT: slli t0, t0, 10
940 ; REMAT-NEXT: li t1, 7
941 ; REMAT-NEXT: slli t1, t1, 9
942 ; REMAT-NEXT: lui t2, 1
943 ; REMAT-NEXT: li t3, 9
944 ; REMAT-NEXT: slli t3, t3, 9
945 ; REMAT-NEXT: li t4, 5
946 ; REMAT-NEXT: slli t4, t4, 10
947 ; REMAT-NEXT: li t5, 11
948 ; REMAT-NEXT: slli t5, t5, 9
949 ; REMAT-NEXT: li t6, 3
950 ; REMAT-NEXT: slli t6, t6, 11
951 ; REMAT-NEXT: li s0, 13
952 ; REMAT-NEXT: slli s0, s0, 9
953 ; REMAT-NEXT: li s1, 7
954 ; REMAT-NEXT: slli s1, s1, 10
955 ; REMAT-NEXT: li s2, 15
956 ; REMAT-NEXT: slli s2, s2, 9
957 ; REMAT-NEXT: lui s3, 2
958 ; REMAT-NEXT: li s4, 17
959 ; REMAT-NEXT: slli s4, s4, 9
960 ; REMAT-NEXT: li s5, 9
961 ; REMAT-NEXT: slli s5, s5, 10
962 ; REMAT-NEXT: li s6, 19
963 ; REMAT-NEXT: slli s6, s6, 9
964 ; REMAT-NEXT: li s7, 5
965 ; REMAT-NEXT: slli s7, s7, 11
966 ; REMAT-NEXT: li s8, 21
967 ; REMAT-NEXT: slli s8, s8, 9
968 ; REMAT-NEXT: li s9, 11
969 ; REMAT-NEXT: slli s9, s9, 10
970 ; REMAT-NEXT: li s10, 23
971 ; REMAT-NEXT: slli s10, s10, 9
972 ; REMAT-NEXT: lui s11, 3
973 ; REMAT-NEXT: li ra, 25
974 ; REMAT-NEXT: slli ra, ra, 9
975 ; REMAT-NEXT: vsetvli zero, a4, e32, m2, ta, ma
976 ; REMAT-NEXT: vle32.v v8, (a5)
977 ; REMAT-NEXT: li a4, 13
978 ; REMAT-NEXT: slli a4, a4, 10
979 ; REMAT-NEXT: vle32.v v10, (a3)
980 ; REMAT-NEXT: vle32.v v12, (a3)
981 ; REMAT-NEXT: li a3, 27
982 ; REMAT-NEXT: slli a3, a3, 9
983 ; REMAT-NEXT: vle32.v v14, (a2)
984 ; REMAT-NEXT: vle32.v v16, (a2)
985 ; REMAT-NEXT: add a2, a0, a6
986 ; REMAT-NEXT: vle32.v v18, (a2)
987 ; REMAT-NEXT: vle32.v v20, (a2)
988 ; REMAT-NEXT: add a2, a0, a7
989 ; REMAT-NEXT: vle32.v v22, (a2)
990 ; REMAT-NEXT: vle32.v v24, (a2)
991 ; REMAT-NEXT: add a2, a0, t0
992 ; REMAT-NEXT: vle32.v v26, (a2)
993 ; REMAT-NEXT: vle32.v v28, (a2)
994 ; REMAT-NEXT: add a2, a0, t1
995 ; REMAT-NEXT: vle32.v v30, (a2)
996 ; REMAT-NEXT: vle32.v v6, (a2)
997 ; REMAT-NEXT: csrr a2, vlenb
998 ; REMAT-NEXT: slli a2, a2, 4
999 ; REMAT-NEXT: add a2, sp, a2
1000 ; REMAT-NEXT: addi a2, a2, 432
1001 ; REMAT-NEXT: vs2r.v v6, (a2) # Unknown-size Folded Spill
1002 ; REMAT-NEXT: add a2, a0, t2
1003 ; REMAT-NEXT: vle32.v v4, (a0)
1004 ; REMAT-NEXT: vle32.v v2, (a2)
1005 ; REMAT-NEXT: vle32.v v6, (a2)
1006 ; REMAT-NEXT: csrr a2, vlenb
1007 ; REMAT-NEXT: li a5, 14
1008 ; REMAT-NEXT: mul a2, a2, a5
1009 ; REMAT-NEXT: add a2, sp, a2
1010 ; REMAT-NEXT: addi a2, a2, 432
1011 ; REMAT-NEXT: vs2r.v v6, (a2) # Unknown-size Folded Spill
1012 ; REMAT-NEXT: add a2, a0, t3
1013 ; REMAT-NEXT: sf.vc.vv 3, 0, v4, v8
1014 ; REMAT-NEXT: vle32.v v4, (a2)
1015 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v10
1016 ; REMAT-NEXT: vle32.v v6, (a2)
1017 ; REMAT-NEXT: add a2, a0, t4
1018 ; REMAT-NEXT: vle32.v v10, (a2)
1019 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v14
1020 ; REMAT-NEXT: vle32.v v0, (a2)
1021 ; REMAT-NEXT: add a2, a0, t5
1022 ; REMAT-NEXT: vle32.v v14, (a2)
1023 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v18
1024 ; REMAT-NEXT: vle32.v v8, (a2)
1025 ; REMAT-NEXT: csrr a2, vlenb
1026 ; REMAT-NEXT: li a5, 12
1027 ; REMAT-NEXT: mul a2, a2, a5
1028 ; REMAT-NEXT: add a2, sp, a2
1029 ; REMAT-NEXT: addi a2, a2, 432
1030 ; REMAT-NEXT: vs2r.v v8, (a2) # Unknown-size Folded Spill
1031 ; REMAT-NEXT: add a2, a0, t6
1032 ; REMAT-NEXT: vle32.v v18, (a2)
1033 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v22
1034 ; REMAT-NEXT: vle32.v v20, (a2)
1035 ; REMAT-NEXT: add a2, a0, s0
1036 ; REMAT-NEXT: vle32.v v22, (a2)
1037 ; REMAT-NEXT: sf.vc.vv 3, 0, v24, v26
1038 ; REMAT-NEXT: vle32.v v24, (a2)
1039 ; REMAT-NEXT: add a2, a0, s1
1040 ; REMAT-NEXT: vle32.v v26, (a2)
1041 ; REMAT-NEXT: sf.vc.vv 3, 0, v28, v30
1042 ; REMAT-NEXT: vle32.v v28, (a2)
1043 ; REMAT-NEXT: add a2, a0, s2
1044 ; REMAT-NEXT: vle32.v v8, (a2)
1045 ; REMAT-NEXT: csrr a5, vlenb
1046 ; REMAT-NEXT: slli a5, a5, 4
1047 ; REMAT-NEXT: add a5, sp, a5
1048 ; REMAT-NEXT: addi a5, a5, 432
1049 ; REMAT-NEXT: vl2r.v v12, (a5) # Unknown-size Folded Reload
1050 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v2
1051 ; REMAT-NEXT: vle32.v v2, (a2)
1052 ; REMAT-NEXT: add a2, a0, s3
1053 ; REMAT-NEXT: vle32.v v12, (a2)
1054 ; REMAT-NEXT: csrr a5, vlenb
1055 ; REMAT-NEXT: li a6, 14
1056 ; REMAT-NEXT: mul a5, a5, a6
1057 ; REMAT-NEXT: add a5, sp, a5
1058 ; REMAT-NEXT: addi a5, a5, 432
1059 ; REMAT-NEXT: vl2r.v v16, (a5) # Unknown-size Folded Reload
1060 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1061 ; REMAT-NEXT: vle32.v v30, (a2)
1062 ; REMAT-NEXT: add a2, a0, s4
1063 ; REMAT-NEXT: vle32.v v16, (a2)
1064 ; REMAT-NEXT: sf.vc.vv 3, 0, v6, v10
1065 ; REMAT-NEXT: vle32.v v6, (a2)
1066 ; REMAT-NEXT: add a2, a0, s5
1067 ; REMAT-NEXT: vle32.v v10, (a2)
1068 ; REMAT-NEXT: sf.vc.vv 3, 0, v0, v14
1069 ; REMAT-NEXT: vle32.v v4, (a2)
1070 ; REMAT-NEXT: add a2, a0, s6
1071 ; REMAT-NEXT: vle32.v v14, (a2)
1072 ; REMAT-NEXT: csrr a5, vlenb
1073 ; REMAT-NEXT: li a6, 12
1074 ; REMAT-NEXT: mul a5, a5, a6
1075 ; REMAT-NEXT: add a5, sp, a5
1076 ; REMAT-NEXT: addi a5, a5, 432
1077 ; REMAT-NEXT: vl2r.v v0, (a5) # Unknown-size Folded Reload
1078 ; REMAT-NEXT: sf.vc.vv 3, 0, v0, v18
1079 ; REMAT-NEXT: vle32.v v0, (a2)
1080 ; REMAT-NEXT: add a2, a0, s7
1081 ; REMAT-NEXT: vle32.v v18, (a2)
1082 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v22
1083 ; REMAT-NEXT: vle32.v v22, (a2)
1084 ; REMAT-NEXT: add a2, a0, s8
1085 ; REMAT-NEXT: vle32.v v20, (a2)
1086 ; REMAT-NEXT: sf.vc.vv 3, 0, v24, v26
1087 ; REMAT-NEXT: vle32.v v26, (a2)
1088 ; REMAT-NEXT: add a2, a0, s9
1089 ; REMAT-NEXT: vle32.v v24, (a2)
1090 ; REMAT-NEXT: sf.vc.vv 3, 0, v28, v8
1091 ; REMAT-NEXT: vle32.v v28, (a2)
1092 ; REMAT-NEXT: add a2, a0, s10
1093 ; REMAT-NEXT: vle32.v v8, (a2)
1094 ; REMAT-NEXT: sf.vc.vv 3, 0, v2, v12
1095 ; REMAT-NEXT: vle32.v v12, (a2)
1096 ; REMAT-NEXT: csrr a2, vlenb
1097 ; REMAT-NEXT: slli a2, a2, 3
1098 ; REMAT-NEXT: add a2, sp, a2
1099 ; REMAT-NEXT: addi a2, a2, 432
1100 ; REMAT-NEXT: vs2r.v v12, (a2) # Unknown-size Folded Spill
1101 ; REMAT-NEXT: add a2, a0, s11
1102 ; REMAT-NEXT: vle32.v v12, (a2)
1103 ; REMAT-NEXT: sf.vc.vv 3, 0, v30, v16
1104 ; REMAT-NEXT: vle32.v v16, (a2)
1105 ; REMAT-NEXT: add a2, a0, ra
1106 ; REMAT-NEXT: vle32.v v2, (a2)
1107 ; REMAT-NEXT: sf.vc.vv 3, 0, v6, v10
1108 ; REMAT-NEXT: vle32.v v10, (a2)
1109 ; REMAT-NEXT: csrr a2, vlenb
1110 ; REMAT-NEXT: slli a2, a2, 1
1111 ; REMAT-NEXT: add a2, sp, a2
1112 ; REMAT-NEXT: addi a2, a2, 432
1113 ; REMAT-NEXT: vs2r.v v10, (a2) # Unknown-size Folded Spill
1114 ; REMAT-NEXT: add a2, a0, a4
1115 ; REMAT-NEXT: vle32.v v10, (a2)
1116 ; REMAT-NEXT: sf.vc.vv 3, 0, v4, v14
1117 ; REMAT-NEXT: vle32.v v14, (a2)
1118 ; REMAT-NEXT: csrr a2, vlenb
1119 ; REMAT-NEXT: slli a2, a2, 2
1120 ; REMAT-NEXT: add a2, sp, a2
1121 ; REMAT-NEXT: addi a2, a2, 432
1122 ; REMAT-NEXT: vs2r.v v14, (a2) # Unknown-size Folded Spill
1123 ; REMAT-NEXT: add a2, a0, a3
1124 ; REMAT-NEXT: vle32.v v14, (a2)
1125 ; REMAT-NEXT: sf.vc.vv 3, 0, v0, v18
1126 ; REMAT-NEXT: vle32.v v18, (a2)
1127 ; REMAT-NEXT: csrr a2, vlenb
1128 ; REMAT-NEXT: slli a2, a2, 4
1129 ; REMAT-NEXT: add a2, sp, a2
1130 ; REMAT-NEXT: addi a2, a2, 432
1131 ; REMAT-NEXT: vs2r.v v18, (a2) # Unknown-size Folded Spill
1132 ; REMAT-NEXT: li a5, 7
1133 ; REMAT-NEXT: slli a5, a5, 11
1134 ; REMAT-NEXT: add a2, a0, a5
1135 ; REMAT-NEXT: vle32.v v18, (a2)
1136 ; REMAT-NEXT: addi a3, sp, 432
1137 ; REMAT-NEXT: vs2r.v v18, (a3) # Unknown-size Folded Spill
1138 ; REMAT-NEXT: sf.vc.vv 3, 0, v22, v20
1139 ; REMAT-NEXT: vle32.v v18, (a2)
1140 ; REMAT-NEXT: csrr a2, vlenb
1141 ; REMAT-NEXT: li a3, 14
1142 ; REMAT-NEXT: mul a2, a2, a3
1143 ; REMAT-NEXT: add a2, sp, a2
1144 ; REMAT-NEXT: addi a2, a2, 432
1145 ; REMAT-NEXT: vs2r.v v18, (a2) # Unknown-size Folded Spill
1146 ; REMAT-NEXT: li a2, 29
1147 ; REMAT-NEXT: slli a2, a2, 9
1148 ; REMAT-NEXT: add a2, a0, a2
1149 ; REMAT-NEXT: vle32.v v18, (a2)
1150 ; REMAT-NEXT: sf.vc.vv 3, 0, v26, v24
1151 ; REMAT-NEXT: vle32.v v20, (a2)
1152 ; REMAT-NEXT: csrr a2, vlenb
1153 ; REMAT-NEXT: li a3, 12
1154 ; REMAT-NEXT: mul a2, a2, a3
1155 ; REMAT-NEXT: add a2, sp, a2
1156 ; REMAT-NEXT: addi a2, a2, 432
1157 ; REMAT-NEXT: vs2r.v v20, (a2) # Unknown-size Folded Spill
1158 ; REMAT-NEXT: li a2, 15
1159 ; REMAT-NEXT: slli a2, a2, 10
1160 ; REMAT-NEXT: add a2, a0, a2
1161 ; REMAT-NEXT: vle32.v v30, (a2)
1162 ; REMAT-NEXT: sf.vc.vv 3, 0, v28, v8
1163 ; REMAT-NEXT: vle32.v v8, (a2)
1164 ; REMAT-NEXT: csrr a2, vlenb
1165 ; REMAT-NEXT: li a3, 10
1166 ; REMAT-NEXT: mul a2, a2, a3
1167 ; REMAT-NEXT: add a2, sp, a2
1168 ; REMAT-NEXT: addi a2, a2, 432
1169 ; REMAT-NEXT: vs2r.v v8, (a2) # Unknown-size Folded Spill
1170 ; REMAT-NEXT: li a2, 31
1171 ; REMAT-NEXT: slli a2, a2, 9
1172 ; REMAT-NEXT: add a2, a0, a2
1173 ; REMAT-NEXT: vle32.v v6, (a2)
1174 ; REMAT-NEXT: csrr a3, vlenb
1175 ; REMAT-NEXT: slli a3, a3, 3
1176 ; REMAT-NEXT: add a3, sp, a3
1177 ; REMAT-NEXT: addi a3, a3, 432
1178 ; REMAT-NEXT: vl2r.v v8, (a3) # Unknown-size Folded Reload
1179 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v12
1180 ; REMAT-NEXT: vle32.v v8, (a2)
1181 ; REMAT-NEXT: csrr a2, vlenb
1182 ; REMAT-NEXT: slli a2, a2, 3
1183 ; REMAT-NEXT: add a2, sp, a2
1184 ; REMAT-NEXT: addi a2, a2, 432
1185 ; REMAT-NEXT: vs2r.v v8, (a2) # Unknown-size Folded Spill
1186 ; REMAT-NEXT: lui a2, 4
1187 ; REMAT-NEXT: add a2, a0, a2
1188 ; REMAT-NEXT: vle32.v v4, (a2)
1189 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v2
1190 ; REMAT-NEXT: vle32.v v8, (a2)
1191 ; REMAT-NEXT: csrr a2, vlenb
1192 ; REMAT-NEXT: li a3, 6
1193 ; REMAT-NEXT: mul a2, a2, a3
1194 ; REMAT-NEXT: add a2, sp, a2
1195 ; REMAT-NEXT: addi a2, a2, 432
1196 ; REMAT-NEXT: vs2r.v v8, (a2) # Unknown-size Folded Spill
1197 ; REMAT-NEXT: lui a2, 4
1198 ; REMAT-NEXT: addiw a2, a2, 512
1199 ; REMAT-NEXT: add a2, a0, a2
1200 ; REMAT-NEXT: vle32.v v2, (a2)
1201 ; REMAT-NEXT: csrr a3, vlenb
1202 ; REMAT-NEXT: slli a3, a3, 1
1203 ; REMAT-NEXT: add a3, sp, a3
1204 ; REMAT-NEXT: addi a3, a3, 432
1205 ; REMAT-NEXT: vl2r.v v8, (a3) # Unknown-size Folded Reload
1206 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v10
1207 ; REMAT-NEXT: vle32.v v20, (a2)
1208 ; REMAT-NEXT: li a2, 17
1209 ; REMAT-NEXT: slli a2, a2, 10
1210 ; REMAT-NEXT: add a2, a0, a2
1211 ; REMAT-NEXT: vle32.v v0, (a2)
1212 ; REMAT-NEXT: csrr a3, vlenb
1213 ; REMAT-NEXT: slli a3, a3, 2
1214 ; REMAT-NEXT: add a3, sp, a3
1215 ; REMAT-NEXT: addi a3, a3, 432
1216 ; REMAT-NEXT: vl2r.v v8, (a3) # Unknown-size Folded Reload
1217 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v14
1218 ; REMAT-NEXT: vle32.v v22, (a2)
1219 ; REMAT-NEXT: lui a2, 4
1220 ; REMAT-NEXT: addiw a2, a2, 1536
1221 ; REMAT-NEXT: add a2, a0, a2
1222 ; REMAT-NEXT: vle32.v v24, (a2)
1223 ; REMAT-NEXT: csrr a3, vlenb
1224 ; REMAT-NEXT: slli a3, a3, 4
1225 ; REMAT-NEXT: add a3, sp, a3
1226 ; REMAT-NEXT: addi a3, a3, 432
1227 ; REMAT-NEXT: vl2r.v v8, (a3) # Unknown-size Folded Reload
1228 ; REMAT-NEXT: addi a3, sp, 432
1229 ; REMAT-NEXT: vl2r.v v10, (a3) # Unknown-size Folded Reload
1230 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v10
1231 ; REMAT-NEXT: vle32.v v8, (a2)
1232 ; REMAT-NEXT: li a2, 9
1233 ; REMAT-NEXT: slli a2, a2, 11
1234 ; REMAT-NEXT: add a2, a0, a2
1235 ; REMAT-NEXT: vle32.v v26, (a2)
1236 ; REMAT-NEXT: csrr a3, vlenb
1237 ; REMAT-NEXT: li a4, 14
1238 ; REMAT-NEXT: mul a3, a3, a4
1239 ; REMAT-NEXT: add a3, sp, a3
1240 ; REMAT-NEXT: addi a3, a3, 432
1241 ; REMAT-NEXT: vl2r.v v10, (a3) # Unknown-size Folded Reload
1242 ; REMAT-NEXT: sf.vc.vv 3, 0, v10, v18
1243 ; REMAT-NEXT: vle32.v v10, (a2)
1244 ; REMAT-NEXT: lui a2, 5
1245 ; REMAT-NEXT: addiw a2, a2, -1536
1246 ; REMAT-NEXT: add a2, a0, a2
1247 ; REMAT-NEXT: vle32.v v28, (a2)
1248 ; REMAT-NEXT: csrr a3, vlenb
1249 ; REMAT-NEXT: li a4, 12
1250 ; REMAT-NEXT: mul a3, a3, a4
1251 ; REMAT-NEXT: add a3, sp, a3
1252 ; REMAT-NEXT: addi a3, a3, 432
1253 ; REMAT-NEXT: vl2r.v v12, (a3) # Unknown-size Folded Reload
1254 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v30
1255 ; REMAT-NEXT: vle32.v v12, (a2)
1256 ; REMAT-NEXT: li a2, 19
1257 ; REMAT-NEXT: slli a2, a2, 10
1258 ; REMAT-NEXT: add a2, a0, a2
1259 ; REMAT-NEXT: vle32.v v30, (a2)
1260 ; REMAT-NEXT: csrr a3, vlenb
1261 ; REMAT-NEXT: li a4, 10
1262 ; REMAT-NEXT: mul a3, a3, a4
1263 ; REMAT-NEXT: add a3, sp, a3
1264 ; REMAT-NEXT: addi a3, a3, 432
1265 ; REMAT-NEXT: vl2r.v v14, (a3) # Unknown-size Folded Reload
1266 ; REMAT-NEXT: sf.vc.vv 3, 0, v14, v6
1267 ; REMAT-NEXT: vle32.v v14, (a2)
1268 ; REMAT-NEXT: lui a2, 5
1269 ; REMAT-NEXT: addiw a2, a2, -512
1270 ; REMAT-NEXT: add a2, a0, a2
1271 ; REMAT-NEXT: vle32.v v6, (a2)
1272 ; REMAT-NEXT: csrr a3, vlenb
1273 ; REMAT-NEXT: slli a3, a3, 3
1274 ; REMAT-NEXT: add a3, sp, a3
1275 ; REMAT-NEXT: addi a3, a3, 432
1276 ; REMAT-NEXT: vl2r.v v16, (a3) # Unknown-size Folded Reload
1277 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1278 ; REMAT-NEXT: vle32.v v16, (a2)
1279 ; REMAT-NEXT: lui a2, 5
1280 ; REMAT-NEXT: add a2, a0, a2
1281 ; REMAT-NEXT: vle32.v v4, (a2)
1282 ; REMAT-NEXT: csrr a3, vlenb
1283 ; REMAT-NEXT: li a4, 6
1284 ; REMAT-NEXT: mul a3, a3, a4
1285 ; REMAT-NEXT: add a3, sp, a3
1286 ; REMAT-NEXT: addi a3, a3, 432
1287 ; REMAT-NEXT: vl2r.v v18, (a3) # Unknown-size Folded Reload
1288 ; REMAT-NEXT: sf.vc.vv 3, 0, v18, v2
1289 ; REMAT-NEXT: vle32.v v18, (a2)
1290 ; REMAT-NEXT: lui a2, 5
1291 ; REMAT-NEXT: addiw a2, a2, 512
1292 ; REMAT-NEXT: add a2, a0, a2
1293 ; REMAT-NEXT: vle32.v v2, (a2)
1294 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v0
1295 ; REMAT-NEXT: vle32.v v20, (a2)
1296 ; REMAT-NEXT: li s7, 21
1297 ; REMAT-NEXT: slli s7, s7, 10
1298 ; REMAT-NEXT: add a2, a0, s7
1299 ; REMAT-NEXT: vle32.v v0, (a2)
1300 ; REMAT-NEXT: sf.vc.vv 3, 0, v22, v24
1301 ; REMAT-NEXT: vle32.v v22, (a2)
1302 ; REMAT-NEXT: lui s4, 5
1303 ; REMAT-NEXT: addiw s4, s4, 1536
1304 ; REMAT-NEXT: add a2, a0, s4
1305 ; REMAT-NEXT: vle32.v v24, (a2)
1306 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v26
1307 ; REMAT-NEXT: vle32.v v8, (a2)
1308 ; REMAT-NEXT: li a2, 11
1309 ; REMAT-NEXT: slli a2, a2, 11
1310 ; REMAT-NEXT: add a2, a0, a2
1311 ; REMAT-NEXT: vle32.v v26, (a2)
1312 ; REMAT-NEXT: sf.vc.vv 3, 0, v10, v28
1313 ; REMAT-NEXT: vle32.v v10, (a2)
1314 ; REMAT-NEXT: lui s3, 6
1315 ; REMAT-NEXT: addiw s3, s3, -1536
1316 ; REMAT-NEXT: add a2, a0, s3
1317 ; REMAT-NEXT: vle32.v v28, (a2)
1318 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v30
1319 ; REMAT-NEXT: vle32.v v12, (a2)
1320 ; REMAT-NEXT: li s2, 23
1321 ; REMAT-NEXT: slli s2, s2, 10
1322 ; REMAT-NEXT: add a2, a0, s2
1323 ; REMAT-NEXT: vle32.v v30, (a2)
1324 ; REMAT-NEXT: sf.vc.vv 3, 0, v14, v6
1325 ; REMAT-NEXT: vle32.v v14, (a2)
1326 ; REMAT-NEXT: lui a2, 6
1327 ; REMAT-NEXT: addiw a2, a2, -512
1328 ; REMAT-NEXT: add a2, a0, a2
1329 ; REMAT-NEXT: vle32.v v6, (a2)
1330 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1331 ; REMAT-NEXT: vle32.v v16, (a2)
1332 ; REMAT-NEXT: lui a2, 6
1333 ; REMAT-NEXT: add a2, a0, a2
1334 ; REMAT-NEXT: lui s1, 6
1335 ; REMAT-NEXT: vle32.v v4, (a2)
1336 ; REMAT-NEXT: sf.vc.vv 3, 0, v18, v2
1337 ; REMAT-NEXT: vle32.v v18, (a2)
1338 ; REMAT-NEXT: lui s0, 6
1339 ; REMAT-NEXT: addiw s0, s0, 512
1340 ; REMAT-NEXT: add a2, a0, s0
1341 ; REMAT-NEXT: vle32.v v2, (a2)
1342 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v0
1343 ; REMAT-NEXT: vle32.v v20, (a2)
1344 ; REMAT-NEXT: li a2, 25
1345 ; REMAT-NEXT: slli a2, a2, 10
1346 ; REMAT-NEXT: add a2, a0, a2
1347 ; REMAT-NEXT: vle32.v v0, (a2)
1348 ; REMAT-NEXT: sf.vc.vv 3, 0, v22, v24
1349 ; REMAT-NEXT: vle32.v v22, (a2)
1350 ; REMAT-NEXT: lui t6, 6
1351 ; REMAT-NEXT: addiw t6, t6, 1536
1352 ; REMAT-NEXT: add a2, a0, t6
1353 ; REMAT-NEXT: vle32.v v24, (a2)
1354 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v26
1355 ; REMAT-NEXT: vle32.v v8, (a2)
1356 ; REMAT-NEXT: li t5, 13
1357 ; REMAT-NEXT: slli t5, t5, 11
1358 ; REMAT-NEXT: add a2, a0, t5
1359 ; REMAT-NEXT: vle32.v v26, (a2)
1360 ; REMAT-NEXT: sf.vc.vv 3, 0, v10, v28
1361 ; REMAT-NEXT: vle32.v v10, (a2)
1362 ; REMAT-NEXT: lui a2, 7
1363 ; REMAT-NEXT: addiw a2, a2, -1536
1364 ; REMAT-NEXT: add a2, a0, a2
1365 ; REMAT-NEXT: vle32.v v28, (a2)
1366 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v30
1367 ; REMAT-NEXT: vle32.v v12, (a2)
1368 ; REMAT-NEXT: li t4, 27
1369 ; REMAT-NEXT: slli t4, t4, 10
1370 ; REMAT-NEXT: add a2, a0, t4
1371 ; REMAT-NEXT: vle32.v v30, (a2)
1372 ; REMAT-NEXT: sf.vc.vv 3, 0, v14, v6
1373 ; REMAT-NEXT: vle32.v v14, (a2)
1374 ; REMAT-NEXT: lui a2, 7
1375 ; REMAT-NEXT: addiw a2, a2, -512
1376 ; REMAT-NEXT: add a2, a0, a2
1377 ; REMAT-NEXT: vle32.v v6, (a2)
1378 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1379 ; REMAT-NEXT: vle32.v v16, (a2)
1380 ; REMAT-NEXT: lui a2, 7
1381 ; REMAT-NEXT: add a2, a0, a2
1382 ; REMAT-NEXT: lui t3, 7
1383 ; REMAT-NEXT: vle32.v v4, (a2)
1384 ; REMAT-NEXT: sf.vc.vv 3, 0, v18, v2
1385 ; REMAT-NEXT: vle32.v v18, (a2)
1386 ; REMAT-NEXT: lui t2, 7
1387 ; REMAT-NEXT: addiw t2, t2, 512
1388 ; REMAT-NEXT: add a2, a0, t2
1389 ; REMAT-NEXT: vle32.v v2, (a2)
1390 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v0
1391 ; REMAT-NEXT: vle32.v v20, (a2)
1392 ; REMAT-NEXT: li t1, 29
1393 ; REMAT-NEXT: slli t1, t1, 10
1394 ; REMAT-NEXT: add a2, a0, t1
1395 ; REMAT-NEXT: vle32.v v0, (a2)
1396 ; REMAT-NEXT: sf.vc.vv 3, 0, v22, v24
1397 ; REMAT-NEXT: vle32.v v22, (a2)
1398 ; REMAT-NEXT: lui t0, 7
1399 ; REMAT-NEXT: addiw t0, t0, 1536
1400 ; REMAT-NEXT: add a2, a0, t0
1401 ; REMAT-NEXT: vle32.v v24, (a2)
1402 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v26
1403 ; REMAT-NEXT: vle32.v v8, (a2)
1404 ; REMAT-NEXT: li a7, 15
1405 ; REMAT-NEXT: slli a7, a7, 11
1406 ; REMAT-NEXT: add a2, a0, a7
1407 ; REMAT-NEXT: vle32.v v26, (a2)
1408 ; REMAT-NEXT: sf.vc.vv 3, 0, v10, v28
1409 ; REMAT-NEXT: vle32.v v10, (a2)
1410 ; REMAT-NEXT: lui a6, 8
1411 ; REMAT-NEXT: addiw a6, a6, -1536
1412 ; REMAT-NEXT: add a2, a0, a6
1413 ; REMAT-NEXT: vle32.v v28, (a2)
1414 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v30
1415 ; REMAT-NEXT: vle32.v v12, (a2)
1416 ; REMAT-NEXT: li a4, 31
1417 ; REMAT-NEXT: slli a4, a4, 10
1418 ; REMAT-NEXT: add a2, a0, a4
1419 ; REMAT-NEXT: vle32.v v30, (a2)
1420 ; REMAT-NEXT: sf.vc.vv 3, 0, v14, v6
1421 ; REMAT-NEXT: vle32.v v14, (a2)
1422 ; REMAT-NEXT: lui a3, 8
1423 ; REMAT-NEXT: addiw a3, a3, -512
1424 ; REMAT-NEXT: add a2, a0, a3
1425 ; REMAT-NEXT: vle32.v v6, (a2)
1426 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1427 ; REMAT-NEXT: vle32.v v16, (a2)
1428 ; REMAT-NEXT: lui a2, 8
1429 ; REMAT-NEXT: add a0, a0, a2
1430 ; REMAT-NEXT: vle32.v v4, (a0)
1431 ; REMAT-NEXT: sf.vc.vv 3, 0, v18, v2
1432 ; REMAT-NEXT: sf.vc.vv 3, 0, v20, v0
1433 ; REMAT-NEXT: sf.vc.vv 3, 0, v22, v24
1434 ; REMAT-NEXT: sf.vc.vv 3, 0, v8, v26
1435 ; REMAT-NEXT: sf.vc.vv 3, 0, v10, v28
1436 ; REMAT-NEXT: sf.vc.vv 3, 0, v12, v30
1437 ; REMAT-NEXT: sf.vc.vv 3, 0, v14, v6
1438 ; REMAT-NEXT: sf.vc.vv 3, 0, v16, v4
1439 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1440 ; REMAT-NEXT: addi a0, a1, 1024
1441 ; REMAT-NEXT: vse32.v v8, (a0)
1442 ; REMAT-NEXT: li a0, 1
1443 ; REMAT-NEXT: slli a0, a0, 11
1444 ; REMAT-NEXT: add a0, a1, a0
1445 ; REMAT-NEXT: sd a0, 416(sp) # 8-byte Folded Spill
1446 ; REMAT-NEXT: li a0, 5
1447 ; REMAT-NEXT: slli a0, a0, 9
1448 ; REMAT-NEXT: add a0, a1, a0
1449 ; REMAT-NEXT: sd a0, 408(sp) # 8-byte Folded Spill
1450 ; REMAT-NEXT: li a0, 3
1451 ; REMAT-NEXT: slli a0, a0, 10
1452 ; REMAT-NEXT: add a0, a1, a0
1453 ; REMAT-NEXT: sd a0, 400(sp) # 8-byte Folded Spill
1454 ; REMAT-NEXT: li a0, 7
1455 ; REMAT-NEXT: slli a0, a0, 9
1456 ; REMAT-NEXT: add a0, a1, a0
1457 ; REMAT-NEXT: sd a0, 392(sp) # 8-byte Folded Spill
1458 ; REMAT-NEXT: lui a0, 1
1459 ; REMAT-NEXT: add a0, a1, a0
1460 ; REMAT-NEXT: sd a0, 384(sp) # 8-byte Folded Spill
1461 ; REMAT-NEXT: li a0, 9
1462 ; REMAT-NEXT: slli a0, a0, 9
1463 ; REMAT-NEXT: add a0, a1, a0
1464 ; REMAT-NEXT: sd a0, 376(sp) # 8-byte Folded Spill
1465 ; REMAT-NEXT: li a0, 5
1466 ; REMAT-NEXT: slli a0, a0, 10
1467 ; REMAT-NEXT: add a0, a1, a0
1468 ; REMAT-NEXT: sd a0, 368(sp) # 8-byte Folded Spill
1469 ; REMAT-NEXT: li a0, 11
1470 ; REMAT-NEXT: slli a0, a0, 9
1471 ; REMAT-NEXT: add a0, a1, a0
1472 ; REMAT-NEXT: sd a0, 360(sp) # 8-byte Folded Spill
1473 ; REMAT-NEXT: li a0, 3
1474 ; REMAT-NEXT: slli a0, a0, 11
1475 ; REMAT-NEXT: add a0, a1, a0
1476 ; REMAT-NEXT: sd a0, 352(sp) # 8-byte Folded Spill
1477 ; REMAT-NEXT: li a0, 13
1478 ; REMAT-NEXT: slli a0, a0, 9
1479 ; REMAT-NEXT: add a0, a1, a0
1480 ; REMAT-NEXT: sd a0, 344(sp) # 8-byte Folded Spill
1481 ; REMAT-NEXT: li a0, 7
1482 ; REMAT-NEXT: slli a0, a0, 10
1483 ; REMAT-NEXT: add a0, a1, a0
1484 ; REMAT-NEXT: sd a0, 336(sp) # 8-byte Folded Spill
1485 ; REMAT-NEXT: li a0, 15
1486 ; REMAT-NEXT: slli a0, a0, 9
1487 ; REMAT-NEXT: add a0, a1, a0
1488 ; REMAT-NEXT: sd a0, 328(sp) # 8-byte Folded Spill
1489 ; REMAT-NEXT: lui a0, 2
1490 ; REMAT-NEXT: add a0, a1, a0
1491 ; REMAT-NEXT: sd a0, 320(sp) # 8-byte Folded Spill
1492 ; REMAT-NEXT: li a0, 17
1493 ; REMAT-NEXT: slli a0, a0, 9
1494 ; REMAT-NEXT: add a0, a1, a0
1495 ; REMAT-NEXT: sd a0, 312(sp) # 8-byte Folded Spill
1496 ; REMAT-NEXT: add s5, a1, s5
1497 ; REMAT-NEXT: sd s5, 304(sp) # 8-byte Folded Spill
1498 ; REMAT-NEXT: add s6, a1, s6
1499 ; REMAT-NEXT: sd s6, 296(sp) # 8-byte Folded Spill
1500 ; REMAT-NEXT: li a0, 5
1501 ; REMAT-NEXT: slli a0, a0, 11
1502 ; REMAT-NEXT: add a0, a1, a0
1503 ; REMAT-NEXT: sd a0, 288(sp) # 8-byte Folded Spill
1504 ; REMAT-NEXT: add s8, a1, s8
1505 ; REMAT-NEXT: sd s8, 280(sp) # 8-byte Folded Spill
1506 ; REMAT-NEXT: add s9, a1, s9
1507 ; REMAT-NEXT: sd s9, 272(sp) # 8-byte Folded Spill
1508 ; REMAT-NEXT: add s10, a1, s10
1509 ; REMAT-NEXT: sd s10, 264(sp) # 8-byte Folded Spill
1510 ; REMAT-NEXT: add s11, a1, s11
1511 ; REMAT-NEXT: sd s11, 256(sp) # 8-byte Folded Spill
1512 ; REMAT-NEXT: add ra, a1, ra
1513 ; REMAT-NEXT: sd ra, 248(sp) # 8-byte Folded Spill
1514 ; REMAT-NEXT: li a0, 13
1515 ; REMAT-NEXT: slli a0, a0, 10
1516 ; REMAT-NEXT: add a0, a1, a0
1517 ; REMAT-NEXT: sd a0, 240(sp) # 8-byte Folded Spill
1518 ; REMAT-NEXT: li a0, 27
1519 ; REMAT-NEXT: slli a0, a0, 9
1520 ; REMAT-NEXT: add a0, a1, a0
1521 ; REMAT-NEXT: sd a0, 232(sp) # 8-byte Folded Spill
1522 ; REMAT-NEXT: add a5, a1, a5
1523 ; REMAT-NEXT: sd a5, 224(sp) # 8-byte Folded Spill
1524 ; REMAT-NEXT: li a0, 29
1525 ; REMAT-NEXT: slli a0, a0, 9
1526 ; REMAT-NEXT: add a0, a1, a0
1527 ; REMAT-NEXT: sd a0, 216(sp) # 8-byte Folded Spill
1528 ; REMAT-NEXT: li a0, 15
1529 ; REMAT-NEXT: slli a0, a0, 10
1530 ; REMAT-NEXT: add a0, a1, a0
1531 ; REMAT-NEXT: sd a0, 208(sp) # 8-byte Folded Spill
1532 ; REMAT-NEXT: li a0, 31
1533 ; REMAT-NEXT: slli a0, a0, 9
1534 ; REMAT-NEXT: add a0, a1, a0
1535 ; REMAT-NEXT: sd a0, 200(sp) # 8-byte Folded Spill
1536 ; REMAT-NEXT: lui a0, 4
1537 ; REMAT-NEXT: add a0, a1, a0
1538 ; REMAT-NEXT: sd a0, 192(sp) # 8-byte Folded Spill
1539 ; REMAT-NEXT: lui a0, 4
1540 ; REMAT-NEXT: addiw a0, a0, 512
1541 ; REMAT-NEXT: add a0, a1, a0
1542 ; REMAT-NEXT: sd a0, 184(sp) # 8-byte Folded Spill
1543 ; REMAT-NEXT: li a0, 17
1544 ; REMAT-NEXT: slli a0, a0, 10
1545 ; REMAT-NEXT: add a0, a1, a0
1546 ; REMAT-NEXT: sd a0, 176(sp) # 8-byte Folded Spill
1547 ; REMAT-NEXT: lui a0, 4
1548 ; REMAT-NEXT: addiw a0, a0, 1536
1549 ; REMAT-NEXT: add a0, a1, a0
1550 ; REMAT-NEXT: sd a0, 168(sp) # 8-byte Folded Spill
1551 ; REMAT-NEXT: li a0, 9
1552 ; REMAT-NEXT: slli a0, a0, 11
1553 ; REMAT-NEXT: add a0, a1, a0
1554 ; REMAT-NEXT: sd a0, 160(sp) # 8-byte Folded Spill
1555 ; REMAT-NEXT: lui a0, 5
1556 ; REMAT-NEXT: addiw a0, a0, -1536
1557 ; REMAT-NEXT: add a0, a1, a0
1558 ; REMAT-NEXT: sd a0, 152(sp) # 8-byte Folded Spill
1559 ; REMAT-NEXT: li a0, 19
1560 ; REMAT-NEXT: slli a0, a0, 10
1561 ; REMAT-NEXT: add a0, a1, a0
1562 ; REMAT-NEXT: sd a0, 144(sp) # 8-byte Folded Spill
1563 ; REMAT-NEXT: lui a0, 5
1564 ; REMAT-NEXT: addiw a0, a0, -512
1565 ; REMAT-NEXT: add a0, a1, a0
1566 ; REMAT-NEXT: sd a0, 136(sp) # 8-byte Folded Spill
1567 ; REMAT-NEXT: lui a0, 5
1568 ; REMAT-NEXT: add a0, a1, a0
1569 ; REMAT-NEXT: sd a0, 128(sp) # 8-byte Folded Spill
1570 ; REMAT-NEXT: lui a0, 5
1571 ; REMAT-NEXT: addiw a0, a0, 512
1572 ; REMAT-NEXT: add a0, a1, a0
1573 ; REMAT-NEXT: sd a0, 120(sp) # 8-byte Folded Spill
1574 ; REMAT-NEXT: add s7, a1, s7
1575 ; REMAT-NEXT: sd s7, 112(sp) # 8-byte Folded Spill
1576 ; REMAT-NEXT: add s4, a1, s4
1577 ; REMAT-NEXT: sd s4, 104(sp) # 8-byte Folded Spill
1578 ; REMAT-NEXT: li a0, 11
1579 ; REMAT-NEXT: slli a0, a0, 11
1580 ; REMAT-NEXT: add a0, a1, a0
1581 ; REMAT-NEXT: sd a0, 96(sp) # 8-byte Folded Spill
1582 ; REMAT-NEXT: add s3, a1, s3
1583 ; REMAT-NEXT: sd s3, 88(sp) # 8-byte Folded Spill
1584 ; REMAT-NEXT: add s2, a1, s2
1585 ; REMAT-NEXT: sd s2, 80(sp) # 8-byte Folded Spill
1586 ; REMAT-NEXT: lui a0, 6
1587 ; REMAT-NEXT: addiw a0, a0, -512
1588 ; REMAT-NEXT: add a0, a1, a0
1589 ; REMAT-NEXT: sd a0, 72(sp) # 8-byte Folded Spill
1590 ; REMAT-NEXT: add s1, a1, s1
1591 ; REMAT-NEXT: sd s1, 64(sp) # 8-byte Folded Spill
1592 ; REMAT-NEXT: add s0, a1, s0
1593 ; REMAT-NEXT: sd s0, 56(sp) # 8-byte Folded Spill
1594 ; REMAT-NEXT: li a0, 25
1595 ; REMAT-NEXT: slli a0, a0, 10
1596 ; REMAT-NEXT: add a0, a1, a0
1597 ; REMAT-NEXT: sd a0, 48(sp) # 8-byte Folded Spill
1598 ; REMAT-NEXT: add t6, a1, t6
1599 ; REMAT-NEXT: sd t6, 40(sp) # 8-byte Folded Spill
1600 ; REMAT-NEXT: add t5, a1, t5
1601 ; REMAT-NEXT: sd t5, 32(sp) # 8-byte Folded Spill
1602 ; REMAT-NEXT: lui a0, 7
1603 ; REMAT-NEXT: addiw a0, a0, -1536
1604 ; REMAT-NEXT: add a0, a1, a0
1605 ; REMAT-NEXT: sd a0, 24(sp) # 8-byte Folded Spill
1606 ; REMAT-NEXT: add t4, a1, t4
1607 ; REMAT-NEXT: sd t4, 16(sp) # 8-byte Folded Spill
1608 ; REMAT-NEXT: lui ra, 7
1609 ; REMAT-NEXT: addiw ra, ra, -512
1610 ; REMAT-NEXT: add ra, a1, ra
1611 ; REMAT-NEXT: add s11, a1, t3
1612 ; REMAT-NEXT: add s10, a1, t2
1613 ; REMAT-NEXT: add s9, a1, t1
1614 ; REMAT-NEXT: add s8, a1, t0
1615 ; REMAT-NEXT: add s7, a1, a7
1616 ; REMAT-NEXT: add s6, a1, a6
1617 ; REMAT-NEXT: add s5, a1, a4
1618 ; REMAT-NEXT: add s4, a1, a3
1619 ; REMAT-NEXT: add s3, a1, a2
1620 ; REMAT-NEXT: lui s2, 8
1621 ; REMAT-NEXT: addiw s2, s2, 512
1622 ; REMAT-NEXT: add s2, a1, s2
1623 ; REMAT-NEXT: lui s1, 8
1624 ; REMAT-NEXT: addiw s1, s1, 1024
1625 ; REMAT-NEXT: add s1, a1, s1
1626 ; REMAT-NEXT: lui s0, 8
1627 ; REMAT-NEXT: addiw s0, s0, 1536
1628 ; REMAT-NEXT: add s0, a1, s0
1629 ; REMAT-NEXT: li t6, 17
1630 ; REMAT-NEXT: slli t6, t6, 11
1631 ; REMAT-NEXT: add t6, a1, t6
1632 ; REMAT-NEXT: lui t5, 9
1633 ; REMAT-NEXT: addiw t5, t5, -1536
1634 ; REMAT-NEXT: add t5, a1, t5
1635 ; REMAT-NEXT: lui t4, 9
1636 ; REMAT-NEXT: addiw t4, t4, -1024
1637 ; REMAT-NEXT: add t4, a1, t4
1638 ; REMAT-NEXT: lui t3, 9
1639 ; REMAT-NEXT: addiw t3, t3, -512
1640 ; REMAT-NEXT: add t3, a1, t3
1641 ; REMAT-NEXT: lui t2, 9
1642 ; REMAT-NEXT: add t2, a1, t2
1643 ; REMAT-NEXT: lui t1, 9
1644 ; REMAT-NEXT: addiw t1, t1, 512
1645 ; REMAT-NEXT: add t1, a1, t1
1646 ; REMAT-NEXT: lui t0, 9
1647 ; REMAT-NEXT: addiw t0, t0, 1024
1648 ; REMAT-NEXT: add t0, a1, t0
1649 ; REMAT-NEXT: lui a7, 9
1650 ; REMAT-NEXT: addiw a7, a7, 1536
1651 ; REMAT-NEXT: add a7, a1, a7
1652 ; REMAT-NEXT: li a6, 19
1653 ; REMAT-NEXT: slli a6, a6, 11
1654 ; REMAT-NEXT: add a6, a1, a6
1655 ; REMAT-NEXT: lui a5, 10
1656 ; REMAT-NEXT: addiw a5, a5, -1536
1657 ; REMAT-NEXT: add a5, a1, a5
1658 ; REMAT-NEXT: lui a4, 10
1659 ; REMAT-NEXT: addiw a4, a4, -1024
1660 ; REMAT-NEXT: add a4, a1, a4
1661 ; REMAT-NEXT: lui a3, 10
1662 ; REMAT-NEXT: addiw a3, a3, -512
1663 ; REMAT-NEXT: add a3, a1, a3
1664 ; REMAT-NEXT: lui a2, 10
1665 ; REMAT-NEXT: add a2, a1, a2
1666 ; REMAT-NEXT: lui a0, 10
1667 ; REMAT-NEXT: addiw a0, a0, 512
1668 ; REMAT-NEXT: add a0, a1, a0
1669 ; REMAT-NEXT: addi a1, a1, 1536
1670 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1671 ; REMAT-NEXT: vse32.v v8, (a1)
1672 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1673 ; REMAT-NEXT: ld a1, 416(sp) # 8-byte Folded Reload
1674 ; REMAT-NEXT: vse32.v v8, (a1)
1675 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1676 ; REMAT-NEXT: ld a1, 408(sp) # 8-byte Folded Reload
1677 ; REMAT-NEXT: vse32.v v8, (a1)
1678 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1679 ; REMAT-NEXT: ld a1, 400(sp) # 8-byte Folded Reload
1680 ; REMAT-NEXT: vse32.v v8, (a1)
1681 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1682 ; REMAT-NEXT: ld a1, 392(sp) # 8-byte Folded Reload
1683 ; REMAT-NEXT: vse32.v v8, (a1)
1684 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1685 ; REMAT-NEXT: ld a1, 384(sp) # 8-byte Folded Reload
1686 ; REMAT-NEXT: vse32.v v8, (a1)
1687 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1688 ; REMAT-NEXT: ld a1, 376(sp) # 8-byte Folded Reload
1689 ; REMAT-NEXT: vse32.v v8, (a1)
1690 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1691 ; REMAT-NEXT: ld a1, 368(sp) # 8-byte Folded Reload
1692 ; REMAT-NEXT: vse32.v v8, (a1)
1693 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1694 ; REMAT-NEXT: ld a1, 360(sp) # 8-byte Folded Reload
1695 ; REMAT-NEXT: vse32.v v8, (a1)
1696 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1697 ; REMAT-NEXT: ld a1, 352(sp) # 8-byte Folded Reload
1698 ; REMAT-NEXT: vse32.v v8, (a1)
1699 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1700 ; REMAT-NEXT: ld a1, 344(sp) # 8-byte Folded Reload
1701 ; REMAT-NEXT: vse32.v v8, (a1)
1702 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1703 ; REMAT-NEXT: ld a1, 336(sp) # 8-byte Folded Reload
1704 ; REMAT-NEXT: vse32.v v8, (a1)
1705 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1706 ; REMAT-NEXT: ld a1, 328(sp) # 8-byte Folded Reload
1707 ; REMAT-NEXT: vse32.v v8, (a1)
1708 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1709 ; REMAT-NEXT: ld a1, 320(sp) # 8-byte Folded Reload
1710 ; REMAT-NEXT: vse32.v v8, (a1)
1711 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1712 ; REMAT-NEXT: ld a1, 312(sp) # 8-byte Folded Reload
1713 ; REMAT-NEXT: vse32.v v8, (a1)
1714 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1715 ; REMAT-NEXT: ld a1, 304(sp) # 8-byte Folded Reload
1716 ; REMAT-NEXT: vse32.v v8, (a1)
1717 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1718 ; REMAT-NEXT: ld a1, 296(sp) # 8-byte Folded Reload
1719 ; REMAT-NEXT: vse32.v v8, (a1)
1720 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1721 ; REMAT-NEXT: ld a1, 288(sp) # 8-byte Folded Reload
1722 ; REMAT-NEXT: vse32.v v8, (a1)
1723 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1724 ; REMAT-NEXT: ld a1, 280(sp) # 8-byte Folded Reload
1725 ; REMAT-NEXT: vse32.v v8, (a1)
1726 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1727 ; REMAT-NEXT: ld a1, 272(sp) # 8-byte Folded Reload
1728 ; REMAT-NEXT: vse32.v v8, (a1)
1729 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1730 ; REMAT-NEXT: ld a1, 264(sp) # 8-byte Folded Reload
1731 ; REMAT-NEXT: vse32.v v8, (a1)
1732 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1733 ; REMAT-NEXT: ld a1, 256(sp) # 8-byte Folded Reload
1734 ; REMAT-NEXT: vse32.v v8, (a1)
1735 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1736 ; REMAT-NEXT: ld a1, 248(sp) # 8-byte Folded Reload
1737 ; REMAT-NEXT: vse32.v v8, (a1)
1738 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1739 ; REMAT-NEXT: ld a1, 240(sp) # 8-byte Folded Reload
1740 ; REMAT-NEXT: vse32.v v8, (a1)
1741 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1742 ; REMAT-NEXT: ld a1, 232(sp) # 8-byte Folded Reload
1743 ; REMAT-NEXT: vse32.v v8, (a1)
1744 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1745 ; REMAT-NEXT: ld a1, 224(sp) # 8-byte Folded Reload
1746 ; REMAT-NEXT: vse32.v v8, (a1)
1747 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1748 ; REMAT-NEXT: ld a1, 216(sp) # 8-byte Folded Reload
1749 ; REMAT-NEXT: vse32.v v8, (a1)
1750 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1751 ; REMAT-NEXT: ld a1, 208(sp) # 8-byte Folded Reload
1752 ; REMAT-NEXT: vse32.v v8, (a1)
1753 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1754 ; REMAT-NEXT: ld a1, 200(sp) # 8-byte Folded Reload
1755 ; REMAT-NEXT: vse32.v v8, (a1)
1756 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1757 ; REMAT-NEXT: ld a1, 192(sp) # 8-byte Folded Reload
1758 ; REMAT-NEXT: vse32.v v8, (a1)
1759 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1760 ; REMAT-NEXT: ld a1, 184(sp) # 8-byte Folded Reload
1761 ; REMAT-NEXT: vse32.v v8, (a1)
1762 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1763 ; REMAT-NEXT: ld a1, 176(sp) # 8-byte Folded Reload
1764 ; REMAT-NEXT: vse32.v v8, (a1)
1765 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1766 ; REMAT-NEXT: ld a1, 168(sp) # 8-byte Folded Reload
1767 ; REMAT-NEXT: vse32.v v8, (a1)
1768 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1769 ; REMAT-NEXT: ld a1, 160(sp) # 8-byte Folded Reload
1770 ; REMAT-NEXT: vse32.v v8, (a1)
1771 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1772 ; REMAT-NEXT: ld a1, 152(sp) # 8-byte Folded Reload
1773 ; REMAT-NEXT: vse32.v v8, (a1)
1774 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1775 ; REMAT-NEXT: ld a1, 144(sp) # 8-byte Folded Reload
1776 ; REMAT-NEXT: vse32.v v8, (a1)
1777 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1778 ; REMAT-NEXT: ld a1, 136(sp) # 8-byte Folded Reload
1779 ; REMAT-NEXT: vse32.v v8, (a1)
1780 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1781 ; REMAT-NEXT: ld a1, 128(sp) # 8-byte Folded Reload
1782 ; REMAT-NEXT: vse32.v v8, (a1)
1783 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1784 ; REMAT-NEXT: ld a1, 120(sp) # 8-byte Folded Reload
1785 ; REMAT-NEXT: vse32.v v8, (a1)
1786 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1787 ; REMAT-NEXT: ld a1, 112(sp) # 8-byte Folded Reload
1788 ; REMAT-NEXT: vse32.v v8, (a1)
1789 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1790 ; REMAT-NEXT: ld a1, 104(sp) # 8-byte Folded Reload
1791 ; REMAT-NEXT: vse32.v v8, (a1)
1792 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1793 ; REMAT-NEXT: ld a1, 96(sp) # 8-byte Folded Reload
1794 ; REMAT-NEXT: vse32.v v8, (a1)
1795 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1796 ; REMAT-NEXT: ld a1, 88(sp) # 8-byte Folded Reload
1797 ; REMAT-NEXT: vse32.v v8, (a1)
1798 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1799 ; REMAT-NEXT: ld a1, 80(sp) # 8-byte Folded Reload
1800 ; REMAT-NEXT: vse32.v v8, (a1)
1801 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1802 ; REMAT-NEXT: ld a1, 72(sp) # 8-byte Folded Reload
1803 ; REMAT-NEXT: vse32.v v8, (a1)
1804 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1805 ; REMAT-NEXT: ld a1, 64(sp) # 8-byte Folded Reload
1806 ; REMAT-NEXT: vse32.v v8, (a1)
1807 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1808 ; REMAT-NEXT: ld a1, 56(sp) # 8-byte Folded Reload
1809 ; REMAT-NEXT: vse32.v v8, (a1)
1810 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1811 ; REMAT-NEXT: ld a1, 48(sp) # 8-byte Folded Reload
1812 ; REMAT-NEXT: vse32.v v8, (a1)
1813 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1814 ; REMAT-NEXT: ld a1, 40(sp) # 8-byte Folded Reload
1815 ; REMAT-NEXT: vse32.v v8, (a1)
1816 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1817 ; REMAT-NEXT: ld a1, 32(sp) # 8-byte Folded Reload
1818 ; REMAT-NEXT: vse32.v v8, (a1)
1819 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1820 ; REMAT-NEXT: ld a1, 24(sp) # 8-byte Folded Reload
1821 ; REMAT-NEXT: vse32.v v8, (a1)
1822 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1823 ; REMAT-NEXT: ld a1, 16(sp) # 8-byte Folded Reload
1824 ; REMAT-NEXT: vse32.v v8, (a1)
1825 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1826 ; REMAT-NEXT: vse32.v v8, (ra)
1827 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1828 ; REMAT-NEXT: vse32.v v8, (s11)
1829 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1830 ; REMAT-NEXT: vse32.v v8, (s10)
1831 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1832 ; REMAT-NEXT: vse32.v v8, (s9)
1833 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1834 ; REMAT-NEXT: vse32.v v8, (s8)
1835 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1836 ; REMAT-NEXT: vse32.v v8, (s7)
1837 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1838 ; REMAT-NEXT: vse32.v v8, (s6)
1839 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1840 ; REMAT-NEXT: vse32.v v8, (s5)
1841 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1842 ; REMAT-NEXT: vse32.v v8, (s4)
1843 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1844 ; REMAT-NEXT: vse32.v v8, (s3)
1845 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1846 ; REMAT-NEXT: vse32.v v8, (s2)
1847 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1848 ; REMAT-NEXT: vse32.v v8, (s1)
1849 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1850 ; REMAT-NEXT: vse32.v v8, (s0)
1851 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1852 ; REMAT-NEXT: vse32.v v8, (t6)
1853 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1854 ; REMAT-NEXT: vse32.v v8, (t5)
1855 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1856 ; REMAT-NEXT: vse32.v v8, (t4)
1857 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1858 ; REMAT-NEXT: vse32.v v8, (t3)
1859 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1860 ; REMAT-NEXT: vse32.v v8, (t2)
1861 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1862 ; REMAT-NEXT: vse32.v v8, (t1)
1863 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1864 ; REMAT-NEXT: vse32.v v8, (t0)
1865 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1866 ; REMAT-NEXT: vse32.v v8, (a7)
1867 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1868 ; REMAT-NEXT: vse32.v v8, (a6)
1869 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1870 ; REMAT-NEXT: vse32.v v8, (a5)
1871 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1872 ; REMAT-NEXT: vse32.v v8, (a4)
1873 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1874 ; REMAT-NEXT: vse32.v v8, (a3)
1875 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1876 ; REMAT-NEXT: vse32.v v8, (a2)
1877 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1878 ; REMAT-NEXT: vse32.v v8, (a0)
1879 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1880 ; REMAT-NEXT: sf.vc.v.i 2, 0, v8, 0
1881 ; REMAT-NEXT: csrr a0, vlenb
1882 ; REMAT-NEXT: li a1, 18
1883 ; REMAT-NEXT: mul a0, a0, a1
1884 ; REMAT-NEXT: add sp, sp, a0
1885 ; REMAT-NEXT: .cfi_def_cfa sp, 544
1886 ; REMAT-NEXT: ld ra, 536(sp) # 8-byte Folded Reload
1887 ; REMAT-NEXT: ld s0, 528(sp) # 8-byte Folded Reload
1888 ; REMAT-NEXT: ld s1, 520(sp) # 8-byte Folded Reload
1889 ; REMAT-NEXT: ld s2, 512(sp) # 8-byte Folded Reload
1890 ; REMAT-NEXT: ld s3, 504(sp) # 8-byte Folded Reload
1891 ; REMAT-NEXT: ld s4, 496(sp) # 8-byte Folded Reload
1892 ; REMAT-NEXT: ld s5, 488(sp) # 8-byte Folded Reload
1893 ; REMAT-NEXT: ld s6, 480(sp) # 8-byte Folded Reload
1894 ; REMAT-NEXT: ld s7, 472(sp) # 8-byte Folded Reload
1895 ; REMAT-NEXT: ld s8, 464(sp) # 8-byte Folded Reload
1896 ; REMAT-NEXT: ld s9, 456(sp) # 8-byte Folded Reload
1897 ; REMAT-NEXT: ld s10, 448(sp) # 8-byte Folded Reload
1898 ; REMAT-NEXT: ld s11, 440(sp) # 8-byte Folded Reload
1899 ; REMAT-NEXT: .cfi_restore ra
1900 ; REMAT-NEXT: .cfi_restore s0
1901 ; REMAT-NEXT: .cfi_restore s1
1902 ; REMAT-NEXT: .cfi_restore s2
1903 ; REMAT-NEXT: .cfi_restore s3
1904 ; REMAT-NEXT: .cfi_restore s4
1905 ; REMAT-NEXT: .cfi_restore s5
1906 ; REMAT-NEXT: .cfi_restore s6
1907 ; REMAT-NEXT: .cfi_restore s7
1908 ; REMAT-NEXT: .cfi_restore s8
1909 ; REMAT-NEXT: .cfi_restore s9
1910 ; REMAT-NEXT: .cfi_restore s10
1911 ; REMAT-NEXT: .cfi_restore s11
1912 ; REMAT-NEXT: addi sp, sp, 544
1913 ; REMAT-NEXT: .cfi_def_cfa_offset 0
1915 %4 = tail call i64 @llvm.riscv.vsetvli.i64(i64 32, i64 2, i64 1)
1916 %5 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %0, i64 %4)
1917 %6 = getelementptr inbounds i32, ptr %0, i64 128
1918 %7 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %6, i64 %4)
1919 %8 = getelementptr inbounds i32, ptr %0, i64 256
1920 %9 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %8, i64 %4)
1921 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %5, <vscale x 4 x i32> %7, i64 %4)
1922 %10 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %8, i64 %4)
1923 %11 = getelementptr inbounds i32, ptr %0, i64 384
1924 %12 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %11, i64 %4)
1925 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %7, <vscale x 4 x i32> %9, i64 %4)
1926 %13 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %11, i64 %4)
1927 %14 = getelementptr inbounds i32, ptr %0, i64 512
1928 %15 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %14, i64 %4)
1929 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %10, <vscale x 4 x i32> %12, i64 %4)
1930 %16 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %14, i64 %4)
1931 %17 = getelementptr inbounds i32, ptr %0, i64 640
1932 %18 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %17, i64 %4)
1933 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %13, <vscale x 4 x i32> %15, i64 %4)
1934 %19 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %17, i64 %4)
1935 %20 = getelementptr inbounds i32, ptr %0, i64 768
1936 %21 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %20, i64 %4)
1937 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %16, <vscale x 4 x i32> %18, i64 %4)
1938 %22 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %20, i64 %4)
1939 %23 = getelementptr inbounds i32, ptr %0, i64 896
1940 %24 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %23, i64 %4)
1941 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %19, <vscale x 4 x i32> %21, i64 %4)
1942 %25 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %23, i64 %4)
1943 %26 = getelementptr inbounds i32, ptr %0, i64 1024
1944 %27 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %26, i64 %4)
1945 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %22, <vscale x 4 x i32> %24, i64 %4)
1946 %28 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %26, i64 %4)
1947 %29 = getelementptr inbounds i32, ptr %0, i64 1152
1948 %30 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %29, i64 %4)
1949 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %25, <vscale x 4 x i32> %27, i64 %4)
1950 %31 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %29, i64 %4)
1951 %32 = getelementptr inbounds i32, ptr %0, i64 1280
1952 %33 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %32, i64 %4)
1953 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %28, <vscale x 4 x i32> %30, i64 %4)
1954 %34 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %32, i64 %4)
1955 %35 = getelementptr inbounds i32, ptr %0, i64 1408
1956 %36 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %35, i64 %4)
1957 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %31, <vscale x 4 x i32> %33, i64 %4)
1958 %37 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %35, i64 %4)
1959 %38 = getelementptr inbounds i32, ptr %0, i64 1536
1960 %39 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %38, i64 %4)
1961 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %34, <vscale x 4 x i32> %36, i64 %4)
1962 %40 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %38, i64 %4)
1963 %41 = getelementptr inbounds i32, ptr %0, i64 1664
1964 %42 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %41, i64 %4)
1965 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %37, <vscale x 4 x i32> %39, i64 %4)
1966 %43 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %41, i64 %4)
1967 %44 = getelementptr inbounds i32, ptr %0, i64 1792
1968 %45 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %44, i64 %4)
1969 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %40, <vscale x 4 x i32> %42, i64 %4)
1970 %46 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %44, i64 %4)
1971 %47 = getelementptr inbounds i32, ptr %0, i64 1920
1972 %48 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %47, i64 %4)
1973 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %43, <vscale x 4 x i32> %45, i64 %4)
1974 %49 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %47, i64 %4)
1975 %50 = getelementptr inbounds i32, ptr %0, i64 2048
1976 %51 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %50, i64 %4)
1977 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %46, <vscale x 4 x i32> %48, i64 %4)
1978 %52 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %50, i64 %4)
1979 %53 = getelementptr inbounds i32, ptr %0, i64 2176
1980 %54 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %53, i64 %4)
1981 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %49, <vscale x 4 x i32> %51, i64 %4)
1982 %55 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %53, i64 %4)
1983 %56 = getelementptr inbounds i32, ptr %0, i64 2304
1984 %57 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %56, i64 %4)
1985 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %52, <vscale x 4 x i32> %54, i64 %4)
1986 %58 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %56, i64 %4)
1987 %59 = getelementptr inbounds i32, ptr %0, i64 2432
1988 %60 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %59, i64 %4)
1989 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %55, <vscale x 4 x i32> %57, i64 %4)
1990 %61 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %59, i64 %4)
1991 %62 = getelementptr inbounds i32, ptr %0, i64 2560
1992 %63 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %62, i64 %4)
1993 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %58, <vscale x 4 x i32> %60, i64 %4)
1994 %64 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %62, i64 %4)
1995 %65 = getelementptr inbounds i32, ptr %0, i64 2688
1996 %66 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %65, i64 %4)
1997 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %61, <vscale x 4 x i32> %63, i64 %4)
1998 %67 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %65, i64 %4)
1999 %68 = getelementptr inbounds i32, ptr %0, i64 2816
2000 %69 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %68, i64 %4)
2001 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %64, <vscale x 4 x i32> %66, i64 %4)
2002 %70 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %68, i64 %4)
2003 %71 = getelementptr inbounds i32, ptr %0, i64 2944
2004 %72 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %71, i64 %4)
2005 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %67, <vscale x 4 x i32> %69, i64 %4)
2006 %73 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %71, i64 %4)
2007 %74 = getelementptr inbounds i32, ptr %0, i64 3072
2008 %75 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %74, i64 %4)
2009 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %70, <vscale x 4 x i32> %72, i64 %4)
2010 %76 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %74, i64 %4)
2011 %77 = getelementptr inbounds i32, ptr %0, i64 3200
2012 %78 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %77, i64 %4)
2013 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %73, <vscale x 4 x i32> %75, i64 %4)
2014 %79 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %77, i64 %4)
2015 %80 = getelementptr inbounds i32, ptr %0, i64 3328
2016 %81 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %80, i64 %4)
2017 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %76, <vscale x 4 x i32> %78, i64 %4)
2018 %82 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %80, i64 %4)
2019 %83 = getelementptr inbounds i32, ptr %0, i64 3456
2020 %84 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %83, i64 %4)
2021 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %79, <vscale x 4 x i32> %81, i64 %4)
2022 %85 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %83, i64 %4)
2023 %86 = getelementptr inbounds i32, ptr %0, i64 3584
2024 %87 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %86, i64 %4)
2025 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %82, <vscale x 4 x i32> %84, i64 %4)
2026 %88 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %86, i64 %4)
2027 %89 = getelementptr inbounds i32, ptr %0, i64 3712
2028 %90 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %89, i64 %4)
2029 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %85, <vscale x 4 x i32> %87, i64 %4)
2030 %91 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %89, i64 %4)
2031 %92 = getelementptr inbounds i32, ptr %0, i64 3840
2032 %93 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %92, i64 %4)
2033 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %88, <vscale x 4 x i32> %90, i64 %4)
2034 %94 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %92, i64 %4)
2035 %95 = getelementptr inbounds i32, ptr %0, i64 3968
2036 %96 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %95, i64 %4)
2037 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %91, <vscale x 4 x i32> %93, i64 %4)
2038 %97 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %95, i64 %4)
2039 %98 = getelementptr inbounds i32, ptr %0, i64 4096
2040 %99 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %98, i64 %4)
2041 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %94, <vscale x 4 x i32> %96, i64 %4)
2042 %100 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %98, i64 %4)
2043 %101 = getelementptr inbounds i32, ptr %0, i64 4224
2044 %102 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %101, i64 %4)
2045 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %97, <vscale x 4 x i32> %99, i64 %4)
2046 %103 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %101, i64 %4)
2047 %104 = getelementptr inbounds i32, ptr %0, i64 4352
2048 %105 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %104, i64 %4)
2049 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %100, <vscale x 4 x i32> %102, i64 %4)
2050 %106 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %104, i64 %4)
2051 %107 = getelementptr inbounds i32, ptr %0, i64 4480
2052 %108 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %107, i64 %4)
2053 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %103, <vscale x 4 x i32> %105, i64 %4)
2054 %109 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %107, i64 %4)
2055 %110 = getelementptr inbounds i32, ptr %0, i64 4608
2056 %111 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %110, i64 %4)
2057 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %106, <vscale x 4 x i32> %108, i64 %4)
2058 %112 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %110, i64 %4)
2059 %113 = getelementptr inbounds i32, ptr %0, i64 4736
2060 %114 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %113, i64 %4)
2061 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %109, <vscale x 4 x i32> %111, i64 %4)
2062 %115 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %113, i64 %4)
2063 %116 = getelementptr inbounds i32, ptr %0, i64 4864
2064 %117 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %116, i64 %4)
2065 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %112, <vscale x 4 x i32> %114, i64 %4)
2066 %118 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %116, i64 %4)
2067 %119 = getelementptr inbounds i32, ptr %0, i64 4992
2068 %120 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %119, i64 %4)
2069 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %115, <vscale x 4 x i32> %117, i64 %4)
2070 %121 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %119, i64 %4)
2071 %122 = getelementptr inbounds i32, ptr %0, i64 5120
2072 %123 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %122, i64 %4)
2073 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %118, <vscale x 4 x i32> %120, i64 %4)
2074 %124 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %122, i64 %4)
2075 %125 = getelementptr inbounds i32, ptr %0, i64 5248
2076 %126 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %125, i64 %4)
2077 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %121, <vscale x 4 x i32> %123, i64 %4)
2078 %127 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %125, i64 %4)
2079 %128 = getelementptr inbounds i32, ptr %0, i64 5376
2080 %129 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %128, i64 %4)
2081 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %124, <vscale x 4 x i32> %126, i64 %4)
2082 %130 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %128, i64 %4)
2083 %131 = getelementptr inbounds i32, ptr %0, i64 5504
2084 %132 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %131, i64 %4)
2085 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %127, <vscale x 4 x i32> %129, i64 %4)
2086 %133 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %131, i64 %4)
2087 %134 = getelementptr inbounds i32, ptr %0, i64 5632
2088 %135 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %134, i64 %4)
2089 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %130, <vscale x 4 x i32> %132, i64 %4)
2090 %136 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %134, i64 %4)
2091 %137 = getelementptr inbounds i32, ptr %0, i64 5760
2092 %138 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %137, i64 %4)
2093 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %133, <vscale x 4 x i32> %135, i64 %4)
2094 %139 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %137, i64 %4)
2095 %140 = getelementptr inbounds i32, ptr %0, i64 5888
2096 %141 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %140, i64 %4)
2097 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %136, <vscale x 4 x i32> %138, i64 %4)
2098 %142 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %140, i64 %4)
2099 %143 = getelementptr inbounds i32, ptr %0, i64 6016
2100 %144 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %143, i64 %4)
2101 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %139, <vscale x 4 x i32> %141, i64 %4)
2102 %145 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %143, i64 %4)
2103 %146 = getelementptr inbounds i32, ptr %0, i64 6144
2104 %147 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %146, i64 %4)
2105 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %142, <vscale x 4 x i32> %144, i64 %4)
2106 %148 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %146, i64 %4)
2107 %149 = getelementptr inbounds i32, ptr %0, i64 6272
2108 %150 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %149, i64 %4)
2109 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %145, <vscale x 4 x i32> %147, i64 %4)
2110 %151 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %149, i64 %4)
2111 %152 = getelementptr inbounds i32, ptr %0, i64 6400
2112 %153 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %152, i64 %4)
2113 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %148, <vscale x 4 x i32> %150, i64 %4)
2114 %154 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %152, i64 %4)
2115 %155 = getelementptr inbounds i32, ptr %0, i64 6528
2116 %156 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %155, i64 %4)
2117 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %151, <vscale x 4 x i32> %153, i64 %4)
2118 %157 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %155, i64 %4)
2119 %158 = getelementptr inbounds i32, ptr %0, i64 6656
2120 %159 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %158, i64 %4)
2121 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %154, <vscale x 4 x i32> %156, i64 %4)
2122 %160 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %158, i64 %4)
2123 %161 = getelementptr inbounds i32, ptr %0, i64 6784
2124 %162 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %161, i64 %4)
2125 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %157, <vscale x 4 x i32> %159, i64 %4)
2126 %163 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %161, i64 %4)
2127 %164 = getelementptr inbounds i32, ptr %0, i64 6912
2128 %165 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %164, i64 %4)
2129 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %160, <vscale x 4 x i32> %162, i64 %4)
2130 %166 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %164, i64 %4)
2131 %167 = getelementptr inbounds i32, ptr %0, i64 7040
2132 %168 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %167, i64 %4)
2133 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %163, <vscale x 4 x i32> %165, i64 %4)
2134 %169 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %167, i64 %4)
2135 %170 = getelementptr inbounds i32, ptr %0, i64 7168
2136 %171 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %170, i64 %4)
2137 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %166, <vscale x 4 x i32> %168, i64 %4)
2138 %172 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %170, i64 %4)
2139 %173 = getelementptr inbounds i32, ptr %0, i64 7296
2140 %174 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %173, i64 %4)
2141 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %169, <vscale x 4 x i32> %171, i64 %4)
2142 %175 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %173, i64 %4)
2143 %176 = getelementptr inbounds i32, ptr %0, i64 7424
2144 %177 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %176, i64 %4)
2145 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %172, <vscale x 4 x i32> %174, i64 %4)
2146 %178 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %176, i64 %4)
2147 %179 = getelementptr inbounds i32, ptr %0, i64 7552
2148 %180 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %179, i64 %4)
2149 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %175, <vscale x 4 x i32> %177, i64 %4)
2150 %181 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %179, i64 %4)
2151 %182 = getelementptr inbounds i32, ptr %0, i64 7680
2152 %183 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %182, i64 %4)
2153 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %178, <vscale x 4 x i32> %180, i64 %4)
2154 %184 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %182, i64 %4)
2155 %185 = getelementptr inbounds i32, ptr %0, i64 7808
2156 %186 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %185, i64 %4)
2157 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %181, <vscale x 4 x i32> %183, i64 %4)
2158 %187 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %185, i64 %4)
2159 %188 = getelementptr inbounds i32, ptr %0, i64 7936
2160 %189 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %188, i64 %4)
2161 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %184, <vscale x 4 x i32> %186, i64 %4)
2162 %190 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %188, i64 %4)
2163 %191 = getelementptr inbounds i32, ptr %0, i64 8064
2164 %192 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %191, i64 %4)
2165 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %187, <vscale x 4 x i32> %189, i64 %4)
2166 %193 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %191, i64 %4)
2167 %194 = getelementptr inbounds i32, ptr %0, i64 8192
2168 %195 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %194, i64 %4)
2169 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %190, <vscale x 4 x i32> %192, i64 %4)
2170 tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %193, <vscale x 4 x i32> %195, i64 %4)
2171 %196 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2172 %197 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2173 %198 = getelementptr inbounds i32, ptr %1, i64 256
2174 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %196, ptr %198, i64 %4)
2175 %199 = getelementptr inbounds i32, ptr %1, i64 384
2176 %200 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2177 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %197, ptr %199, i64 %4)
2178 %201 = getelementptr inbounds i32, ptr %1, i64 512
2179 %202 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2180 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %200, ptr %201, i64 %4)
2181 %203 = getelementptr inbounds i32, ptr %1, i64 640
2182 %204 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2183 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %202, ptr %203, i64 %4)
2184 %205 = getelementptr inbounds i32, ptr %1, i64 768
2185 %206 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2186 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %204, ptr %205, i64 %4)
2187 %207 = getelementptr inbounds i32, ptr %1, i64 896
2188 %208 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2189 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %206, ptr %207, i64 %4)
2190 %209 = getelementptr inbounds i32, ptr %1, i64 1024
2191 %210 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2192 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %208, ptr %209, i64 %4)
2193 %211 = getelementptr inbounds i32, ptr %1, i64 1152
2194 %212 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2195 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %210, ptr %211, i64 %4)
2196 %213 = getelementptr inbounds i32, ptr %1, i64 1280
2197 %214 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2198 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %212, ptr %213, i64 %4)
2199 %215 = getelementptr inbounds i32, ptr %1, i64 1408
2200 %216 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2201 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %214, ptr %215, i64 %4)
2202 %217 = getelementptr inbounds i32, ptr %1, i64 1536
2203 %218 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2204 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %216, ptr %217, i64 %4)
2205 %219 = getelementptr inbounds i32, ptr %1, i64 1664
2206 %220 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2207 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %218, ptr %219, i64 %4)
2208 %221 = getelementptr inbounds i32, ptr %1, i64 1792
2209 %222 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2210 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %220, ptr %221, i64 %4)
2211 %223 = getelementptr inbounds i32, ptr %1, i64 1920
2212 %224 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2213 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %222, ptr %223, i64 %4)
2214 %225 = getelementptr inbounds i32, ptr %1, i64 2048
2215 %226 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2216 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %224, ptr %225, i64 %4)
2217 %227 = getelementptr inbounds i32, ptr %1, i64 2176
2218 %228 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2219 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %226, ptr %227, i64 %4)
2220 %229 = getelementptr inbounds i32, ptr %1, i64 2304
2221 %230 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2222 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %228, ptr %229, i64 %4)
2223 %231 = getelementptr inbounds i32, ptr %1, i64 2432
2224 %232 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2225 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %230, ptr %231, i64 %4)
2226 %233 = getelementptr inbounds i32, ptr %1, i64 2560
2227 %234 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2228 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %232, ptr %233, i64 %4)
2229 %235 = getelementptr inbounds i32, ptr %1, i64 2688
2230 %236 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2231 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %234, ptr %235, i64 %4)
2232 %237 = getelementptr inbounds i32, ptr %1, i64 2816
2233 %238 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2234 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %236, ptr %237, i64 %4)
2235 %239 = getelementptr inbounds i32, ptr %1, i64 2944
2236 %240 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2237 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %238, ptr %239, i64 %4)
2238 %241 = getelementptr inbounds i32, ptr %1, i64 3072
2239 %242 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2240 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %240, ptr %241, i64 %4)
2241 %243 = getelementptr inbounds i32, ptr %1, i64 3200
2242 %244 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2243 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %242, ptr %243, i64 %4)
2244 %245 = getelementptr inbounds i32, ptr %1, i64 3328
2245 %246 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2246 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %244, ptr %245, i64 %4)
2247 %247 = getelementptr inbounds i32, ptr %1, i64 3456
2248 %248 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2249 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %246, ptr %247, i64 %4)
2250 %249 = getelementptr inbounds i32, ptr %1, i64 3584
2251 %250 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2252 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %248, ptr %249, i64 %4)
2253 %251 = getelementptr inbounds i32, ptr %1, i64 3712
2254 %252 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2255 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %250, ptr %251, i64 %4)
2256 %253 = getelementptr inbounds i32, ptr %1, i64 3840
2257 %254 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2258 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %252, ptr %253, i64 %4)
2259 %255 = getelementptr inbounds i32, ptr %1, i64 3968
2260 %256 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2261 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %254, ptr %255, i64 %4)
2262 %257 = getelementptr inbounds i32, ptr %1, i64 4096
2263 %258 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2264 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %256, ptr %257, i64 %4)
2265 %259 = getelementptr inbounds i32, ptr %1, i64 4224
2266 %260 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2267 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %258, ptr %259, i64 %4)
2268 %261 = getelementptr inbounds i32, ptr %1, i64 4352
2269 %262 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2270 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %260, ptr %261, i64 %4)
2271 %263 = getelementptr inbounds i32, ptr %1, i64 4480
2272 %264 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2273 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %262, ptr %263, i64 %4)
2274 %265 = getelementptr inbounds i32, ptr %1, i64 4608
2275 %266 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2276 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %264, ptr %265, i64 %4)
2277 %267 = getelementptr inbounds i32, ptr %1, i64 4736
2278 %268 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2279 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %266, ptr %267, i64 %4)
2280 %269 = getelementptr inbounds i32, ptr %1, i64 4864
2281 %270 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2282 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %268, ptr %269, i64 %4)
2283 %271 = getelementptr inbounds i32, ptr %1, i64 4992
2284 %272 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2285 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %270, ptr %271, i64 %4)
2286 %273 = getelementptr inbounds i32, ptr %1, i64 5120
2287 %274 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2288 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %272, ptr %273, i64 %4)
2289 %275 = getelementptr inbounds i32, ptr %1, i64 5248
2290 %276 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2291 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %274, ptr %275, i64 %4)
2292 %277 = getelementptr inbounds i32, ptr %1, i64 5376
2293 %278 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2294 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %276, ptr %277, i64 %4)
2295 %279 = getelementptr inbounds i32, ptr %1, i64 5504
2296 %280 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2297 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %278, ptr %279, i64 %4)
2298 %281 = getelementptr inbounds i32, ptr %1, i64 5632
2299 %282 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2300 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %280, ptr %281, i64 %4)
2301 %283 = getelementptr inbounds i32, ptr %1, i64 5760
2302 %284 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2303 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %282, ptr %283, i64 %4)
2304 %285 = getelementptr inbounds i32, ptr %1, i64 5888
2305 %286 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2306 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %284, ptr %285, i64 %4)
2307 %287 = getelementptr inbounds i32, ptr %1, i64 6016
2308 %288 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2309 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %286, ptr %287, i64 %4)
2310 %289 = getelementptr inbounds i32, ptr %1, i64 6144
2311 %290 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2312 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %288, ptr %289, i64 %4)
2313 %291 = getelementptr inbounds i32, ptr %1, i64 6272
2314 %292 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2315 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %290, ptr %291, i64 %4)
2316 %293 = getelementptr inbounds i32, ptr %1, i64 6400
2317 %294 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2318 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %292, ptr %293, i64 %4)
2319 %295 = getelementptr inbounds i32, ptr %1, i64 6528
2320 %296 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2321 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %294, ptr %295, i64 %4)
2322 %297 = getelementptr inbounds i32, ptr %1, i64 6656
2323 %298 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2324 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %296, ptr %297, i64 %4)
2325 %299 = getelementptr inbounds i32, ptr %1, i64 6784
2326 %300 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2327 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %298, ptr %299, i64 %4)
2328 %301 = getelementptr inbounds i32, ptr %1, i64 6912
2329 %302 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2330 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %300, ptr %301, i64 %4)
2331 %303 = getelementptr inbounds i32, ptr %1, i64 7040
2332 %304 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2333 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %302, ptr %303, i64 %4)
2334 %305 = getelementptr inbounds i32, ptr %1, i64 7168
2335 %306 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2336 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %304, ptr %305, i64 %4)
2337 %307 = getelementptr inbounds i32, ptr %1, i64 7296
2338 %308 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2339 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %306, ptr %307, i64 %4)
2340 %309 = getelementptr inbounds i32, ptr %1, i64 7424
2341 %310 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2342 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %308, ptr %309, i64 %4)
2343 %311 = getelementptr inbounds i32, ptr %1, i64 7552
2344 %312 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2345 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %310, ptr %311, i64 %4)
2346 %313 = getelementptr inbounds i32, ptr %1, i64 7680
2347 %314 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2348 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %312, ptr %313, i64 %4)
2349 %315 = getelementptr inbounds i32, ptr %1, i64 7808
2350 %316 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2351 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %314, ptr %315, i64 %4)
2352 %317 = getelementptr inbounds i32, ptr %1, i64 7936
2353 %318 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2354 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %316, ptr %317, i64 %4)
2355 %319 = getelementptr inbounds i32, ptr %1, i64 8064
2356 %320 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2357 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %318, ptr %319, i64 %4)
2358 %321 = getelementptr inbounds i32, ptr %1, i64 8192
2359 %322 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2360 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %320, ptr %321, i64 %4)
2361 %323 = getelementptr inbounds i32, ptr %1, i64 8320
2362 %324 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2363 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %322, ptr %323, i64 %4)
2364 %325 = getelementptr inbounds i32, ptr %1, i64 8448
2365 %326 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2366 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %324, ptr %325, i64 %4)
2367 %327 = getelementptr inbounds i32, ptr %1, i64 8576
2368 %328 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2369 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %326, ptr %327, i64 %4)
2370 %329 = getelementptr inbounds i32, ptr %1, i64 8704
2371 %330 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2372 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %328, ptr %329, i64 %4)
2373 %331 = getelementptr inbounds i32, ptr %1, i64 8832
2374 %332 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2375 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %330, ptr %331, i64 %4)
2376 %333 = getelementptr inbounds i32, ptr %1, i64 8960
2377 %334 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2378 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %332, ptr %333, i64 %4)
2379 %335 = getelementptr inbounds i32, ptr %1, i64 9088
2380 %336 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2381 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %334, ptr %335, i64 %4)
2382 %337 = getelementptr inbounds i32, ptr %1, i64 9216
2383 %338 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2384 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %336, ptr %337, i64 %4)
2385 %339 = getelementptr inbounds i32, ptr %1, i64 9344
2386 %340 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2387 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %338, ptr %339, i64 %4)
2388 %341 = getelementptr inbounds i32, ptr %1, i64 9472
2389 %342 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2390 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %340, ptr %341, i64 %4)
2391 %343 = getelementptr inbounds i32, ptr %1, i64 9600
2392 %344 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2393 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %342, ptr %343, i64 %4)
2394 %345 = getelementptr inbounds i32, ptr %1, i64 9728
2395 %346 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2396 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %344, ptr %345, i64 %4)
2397 %347 = getelementptr inbounds i32, ptr %1, i64 9856
2398 %348 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2399 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %346, ptr %347, i64 %4)
2400 %349 = getelementptr inbounds i32, ptr %1, i64 9984
2401 %350 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2402 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %348, ptr %349, i64 %4)
2403 %351 = getelementptr inbounds i32, ptr %1, i64 10112
2404 %352 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2405 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %350, ptr %351, i64 %4)
2406 %353 = getelementptr inbounds i32, ptr %1, i64 10240
2407 %354 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2408 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %352, ptr %353, i64 %4)
2409 %355 = getelementptr inbounds i32, ptr %1, i64 10368
2410 %356 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2411 tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %354, ptr %355, i64 %4)
2412 %357 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2416 declare i64 @llvm.riscv.vsetvli.i64(i64, i64, i64)
2417 declare <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32>, ptr, i64)
2418 declare void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64, i64, <vscale x 4 x i32>, <vscale x 4 x i32>, i64)
2419 declare <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64, i64, i64, i64)
2420 declare void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32>, ptr, i64)