[clang] Implement lifetime analysis for lifetime_capture_by(X) (#115921)
[llvm-project.git] / clang / test / OpenMP / for_simd_scan_codegen.cpp
blob829f2656042fb45651d9169a0681fde73b11a802
1 // RUN: %clang_cc1 -verify -fopenmp -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck %s
2 // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple x86_64-unknown-unknown -emit-pch -o %t %s
3 // RUN: %clang_cc1 -fopenmp -x c++ -triple x86_64-unknown-unknown -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s
5 // RUN: %clang_cc1 -verify -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck --check-prefix SIMD-ONLY0 %s
6 // RUN: %clang_cc1 -fopenmp-simd -x c++ -std=c++11 -triple x86_64-unknown-unknown -emit-pch -o %t %s
7 // RUN: %clang_cc1 -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck --check-prefix SIMD-ONLY0 %s
8 // SIMD-ONLY0-NOT: {{__kmpc|__tgt}}
9 // expected-no-diagnostics
10 #ifndef HEADER
11 #define HEADER
13 void foo();
14 void bar();
16 // CHECK: define{{.*}} void @{{.*}}baz{{.*}}(i32 noundef %n)
17 void baz(int n) {
18 static float a[10];
19 static double b;
20 // CHECK: call ptr @llvm.stacksave.p0()
21 // CHECK: [[A_BUF_SIZE:%.+]] = mul nuw i64 10, [[NUM_ELEMS:%[^,]+]]
23 // float a_buffer[10][n];
24 // CHECK: [[A_BUF:%.+]] = alloca float, i64 [[A_BUF_SIZE]],
26 // double b_buffer[10];
27 // CHECK: [[B_BUF:%.+]] = alloca double, i64 10,
28 #pragma omp for simd reduction(inscan, +:a[:n], b)
29 for (int i = 0; i < 10; ++i) {
30 // CHECK: call void @__kmpc_for_static_init_4(
31 // CHECK: call ptr @llvm.stacksave.p0()
32 // CHECK: store float 0.000000e+00, ptr %
33 // CHECK: store double 0.000000e+00, ptr [[B_PRIV_ADDR:%.+]],
34 // CHECK: br label %[[DISPATCH:[^,]+]]
35 // CHECK: [[INPUT_PHASE:.+]]:
36 // CHECK: call void @{{.+}}foo{{.+}}()
38 // a_buffer[i][0..n] = a_priv[[0..n];
39 // CHECK: [[BASE_IDX_I:%.+]] = load i32, ptr [[IV_ADDR:%.+]],
40 // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
41 // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
42 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
43 // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds nuw [10 x float], ptr [[A_PRIV_ADDR:%.+]], i64 0, i64 0
44 // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
45 // CHECK: call void @llvm.memcpy.p0.p0.i64(ptr {{.*}}[[A_BUF_IDX]], ptr {{.*}}[[A_PRIV]], i64 [[BYTES]], i1 false)
47 // b_buffer[i] = b_priv;
48 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[BASE_IDX]]
49 // CHECK: [[B_PRIV:%.+]] = load double, ptr [[B_PRIV_ADDR]],
50 // CHECK: store double [[B_PRIV]], ptr [[B_BUF_IDX]],
51 // CHECK: br label %[[LOOP_CONTINUE:.+]]
53 // CHECK: [[DISPATCH]]:
54 // CHECK: br label %[[INPUT_PHASE]]
55 // CHECK: [[LOOP_CONTINUE]]:
56 // CHECK: call void @llvm.stackrestore.p0(ptr %
57 // CHECK: call void @__kmpc_for_static_fini(
58 // CHECK: call void @__kmpc_barrier(
59 foo();
60 #pragma omp scan inclusive(a[:n], b)
61 // CHECK: [[LOG2_10:%.+]] = call double @llvm.log2.f64(double 1.000000e+01)
62 // CHECK: [[CEIL_LOG2_10:%.+]] = call double @llvm.ceil.f64(double [[LOG2_10]])
63 // CHECK: [[CEIL_LOG2_10_INT:%.+]] = fptoui double [[CEIL_LOG2_10]] to i32
64 // CHECK: br label %[[OUTER_BODY:[^,]+]]
65 // CHECK: [[OUTER_BODY]]:
66 // CHECK: [[K:%.+]] = phi i32 [ 0, %{{.+}} ], [ [[K_NEXT:%.+]], %{{.+}} ]
67 // CHECK: [[K2POW:%.+]] = phi i64 [ 1, %{{.+}} ], [ [[K2POW_NEXT:%.+]], %{{.+}} ]
68 // CHECK: [[CMP:%.+]] = icmp uge i64 9, [[K2POW]]
69 // CHECK: br i1 [[CMP]], label %[[INNER_BODY:[^,]+]], label %[[INNER_EXIT:[^,]+]]
70 // CHECK: [[INNER_BODY]]:
71 // CHECK: [[I:%.+]] = phi i64 [ 9, %[[OUTER_BODY]] ], [ [[I_PREV:%.+]], %{{.+}} ]
73 // a_buffer[i] += a_buffer[i-pow(2, k)];
74 // CHECK: [[IDX:%.+]] = mul nsw i64 [[I]], [[NUM_ELEMS]]
75 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
76 // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
77 // CHECK: [[IDX:%.+]] = mul nsw i64 [[IDX_SUB_K2POW]], [[NUM_ELEMS]]
78 // CHECK: [[A_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
79 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[I]]
80 // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
81 // CHECK: [[B_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[IDX_SUB_K2POW]]
82 // CHECK: [[A_BUF_END:%.+]] = getelementptr float, ptr [[A_BUF_IDX]], i64 [[NUM_ELEMS]]
83 // CHECK: [[ISEMPTY:%.+]] = icmp eq ptr [[A_BUF_IDX]], [[A_BUF_END]]
84 // CHECK: br i1 [[ISEMPTY]], label %[[RED_DONE:[^,]+]], label %[[RED_BODY:[^,]+]]
85 // CHECK: [[RED_BODY]]:
86 // CHECK: [[A_BUF_IDX_SUB_K2POW_ELEM:%.+]] = phi ptr [ [[A_BUF_IDX_SUB_K2POW]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_SUB_K2POW_NEXT:%.+]], %[[RED_BODY]] ]
87 // CHECK: [[A_BUF_IDX_ELEM:%.+]] = phi ptr [ [[A_BUF_IDX]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_NEXT:%.+]], %[[RED_BODY]] ]
88 // CHECK: [[A_BUF_IDX_VAL:%.+]] = load float, ptr [[A_BUF_IDX_ELEM]],
89 // CHECK: [[A_BUF_IDX_SUB_K2POW_VAL:%.+]] = load float, ptr [[A_BUF_IDX_SUB_K2POW_ELEM]],
90 // CHECK: [[RED:%.+]] = fadd float [[A_BUF_IDX_VAL]], [[A_BUF_IDX_SUB_K2POW_VAL]]
91 // CHECK: store float [[RED]], ptr [[A_BUF_IDX_ELEM]],
92 // CHECK: [[A_BUF_IDX_NEXT]] = getelementptr float, ptr [[A_BUF_IDX_ELEM]], i32 1
93 // CHECK: [[A_BUF_IDX_SUB_K2POW_NEXT]] = getelementptr float, ptr [[A_BUF_IDX_SUB_K2POW_ELEM]], i32 1
94 // CHECK: [[DONE:%.+]] = icmp eq ptr [[A_BUF_IDX_NEXT]], [[A_BUF_END]]
95 // CHECK: br i1 [[DONE]], label %[[RED_DONE]], label %[[RED_BODY]]
96 // CHECK: [[RED_DONE]]:
98 // b_buffer[i] += b_buffer[i-pow(2, k)];
99 // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, ptr [[B_BUF_IDX]],
100 // CHECK: [[B_BUF_IDX_SUB_K2POW_VAL:%.+]] = load double, ptr [[B_BUF_IDX_SUB_K2POW]],
101 // CHECK: [[RED:%.+]] = fadd double [[B_BUF_IDX_VAL]], [[B_BUF_IDX_SUB_K2POW_VAL]]
102 // CHECK: store double [[RED]], ptr [[B_BUF_IDX]],
104 // --i;
105 // CHECK: [[I_PREV:%.+]] = sub nuw i64 [[I]], 1
106 // CHECK: [[CMP:%.+]] = icmp uge i64 [[I_PREV]], [[K2POW]]
107 // CHECK: br i1 [[CMP]], label %[[INNER_BODY]], label %[[INNER_EXIT]]
108 // CHECK: [[INNER_EXIT]]:
110 // ++k;
111 // CHECK: [[K_NEXT]] = add nuw i32 [[K]], 1
112 // k2pow <<= 1;
113 // CHECK: [[K2POW_NEXT]] = shl nuw i64 [[K2POW]], 1
114 // CHECK: [[CMP:%.+]] = icmp ne i32 [[K_NEXT]], [[CEIL_LOG2_10_INT]]
115 // CHECK: br i1 [[CMP]], label %[[OUTER_BODY]], label %[[OUTER_EXIT:[^,]+]]
116 // CHECK: [[OUTER_EXIT]]:
117 bar();
118 // CHECK: call void @__kmpc_for_static_init_4(
119 // CHECK: call ptr @llvm.stacksave.p0()
120 // CHECK: store float 0.000000e+00, ptr %
121 // CHECK: store double 0.000000e+00, ptr [[B_PRIV_ADDR:%.+]],
122 // CHECK: br label %[[DISPATCH:[^,]+]]
124 // Skip the before scan body.
125 // CHECK: call void @{{.+}}foo{{.+}}()
127 // CHECK: [[EXIT_INSCAN:[^,]+]]:
128 // CHECK: br label %[[LOOP_CONTINUE:[^,]+]]
130 // CHECK: [[DISPATCH]]:
131 // a_priv[[0..n] = a_buffer[i][0..n];
132 // CHECK: [[BASE_IDX_I:%.+]] = load i32, ptr [[IV_ADDR:%.+]],
133 // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
134 // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
135 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
136 // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds nuw [10 x float], ptr [[A_PRIV_ADDR:%.+]], i64 0, i64 0
137 // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
138 // CHECK: call void @llvm.memcpy.p0.p0.i64(ptr {{.*}}[[A_PRIV]], ptr {{.*}}[[A_BUF_IDX]], i64 [[BYTES]], i1 false)
140 // b_priv = b_buffer[i];
141 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[BASE_IDX]]
142 // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, ptr [[B_BUF_IDX]],
143 // CHECK: store double [[B_BUF_IDX_VAL]], ptr [[B_PRIV_ADDR]],
144 // CHECK: br label %[[SCAN_PHASE:[^,]+]]
146 // CHECK: [[SCAN_PHASE]]:
147 // CHECK: call void @{{.+}}bar{{.+}}()
148 // CHECK: br label %[[EXIT_INSCAN]]
150 // CHECK: [[LOOP_CONTINUE]]:
151 // CHECK: call void @llvm.stackrestore.p0(ptr %
152 // CHECK: call void @__kmpc_for_static_fini(
153 // CHECK: call void @llvm.stackrestore.p0(ptr
154 // CHECK: call void @__kmpc_barrier(
157 // CHECK: call ptr @llvm.stacksave.p0()
158 // CHECK: [[A_BUF_SIZE:%.+]] = mul nuw i64 10, [[NUM_ELEMS:%[^,]+]]
160 // float a_buffer[10][n];
161 // CHECK: [[A_BUF:%.+]] = alloca float, i64 [[A_BUF_SIZE]],
163 // double b_buffer[10];
164 // CHECK: [[B_BUF:%.+]] = alloca double, i64 10,
165 #pragma omp for simd reduction(inscan, +:a[:n], b)
166 for (int i = 0; i < 10; ++i) {
167 // CHECK: call void @__kmpc_for_static_init_4(
168 // CHECK: call ptr @llvm.stacksave.p0()
169 // CHECK: store float 0.000000e+00, ptr %
170 // CHECK: store double 0.000000e+00, ptr [[B_PRIV_ADDR:%.+]],
171 // CHECK: br label %[[DISPATCH:[^,]+]]
173 // Skip the before scan body.
174 // CHECK: call void @{{.+}}foo{{.+}}()
176 // CHECK: [[EXIT_INSCAN:[^,]+]]:
178 // a_buffer[i][0..n] = a_priv[[0..n];
179 // CHECK: [[BASE_IDX_I:%.+]] = load i32, ptr [[IV_ADDR:%.+]],
180 // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
181 // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
182 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
183 // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds nuw [10 x float], ptr [[A_PRIV_ADDR:%.+]], i64 0, i64 0
184 // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
185 // CHECK: call void @llvm.memcpy.p0.p0.i64(ptr {{.*}}[[A_BUF_IDX]], ptr {{.*}}[[A_PRIV]], i64 [[BYTES]], i1 false)
187 // b_buffer[i] = b_priv;
188 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[BASE_IDX]]
189 // CHECK: [[B_PRIV:%.+]] = load double, ptr [[B_PRIV_ADDR]],
190 // CHECK: store double [[B_PRIV]], ptr [[B_BUF_IDX]],
191 // CHECK: br label %[[LOOP_CONTINUE:[^,]+]]
193 // CHECK: [[DISPATCH]]:
194 // CHECK: br label %[[INPUT_PHASE:[^,]+]]
196 // CHECK: [[INPUT_PHASE]]:
197 // CHECK: call void @{{.+}}bar{{.+}}()
198 // CHECK: br label %[[EXIT_INSCAN]]
200 // CHECK: [[LOOP_CONTINUE]]:
201 // CHECK: call void @llvm.stackrestore.p0(ptr %
202 // CHECK: call void @__kmpc_for_static_fini(
203 // CHECK: call void @__kmpc_barrier(
204 foo();
205 #pragma omp scan exclusive(a[:n], b)
206 // CHECK: [[LOG2_10:%.+]] = call double @llvm.log2.f64(double 1.000000e+01)
207 // CHECK: [[CEIL_LOG2_10:%.+]] = call double @llvm.ceil.f64(double [[LOG2_10]])
208 // CHECK: [[CEIL_LOG2_10_INT:%.+]] = fptoui double [[CEIL_LOG2_10]] to i32
209 // CHECK: br label %[[OUTER_BODY:[^,]+]]
210 // CHECK: [[OUTER_BODY]]:
211 // CHECK: [[K:%.+]] = phi i32 [ 0, %{{.+}} ], [ [[K_NEXT:%.+]], %{{.+}} ]
212 // CHECK: [[K2POW:%.+]] = phi i64 [ 1, %{{.+}} ], [ [[K2POW_NEXT:%.+]], %{{.+}} ]
213 // CHECK: [[CMP:%.+]] = icmp uge i64 9, [[K2POW]]
214 // CHECK: br i1 [[CMP]], label %[[INNER_BODY:[^,]+]], label %[[INNER_EXIT:[^,]+]]
215 // CHECK: [[INNER_BODY]]:
216 // CHECK: [[I:%.+]] = phi i64 [ 9, %[[OUTER_BODY]] ], [ [[I_PREV:%.+]], %{{.+}} ]
218 // a_buffer[i] += a_buffer[i-pow(2, k)];
219 // CHECK: [[IDX:%.+]] = mul nsw i64 [[I]], [[NUM_ELEMS]]
220 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
221 // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
222 // CHECK: [[IDX:%.+]] = mul nsw i64 [[IDX_SUB_K2POW]], [[NUM_ELEMS]]
223 // CHECK: [[A_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
224 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[I]]
225 // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
226 // CHECK: [[B_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[IDX_SUB_K2POW]]
227 // CHECK: [[A_BUF_END:%.+]] = getelementptr float, ptr [[A_BUF_IDX]], i64 [[NUM_ELEMS]]
228 // CHECK: [[ISEMPTY:%.+]] = icmp eq ptr [[A_BUF_IDX]], [[A_BUF_END]]
229 // CHECK: br i1 [[ISEMPTY]], label %[[RED_DONE:[^,]+]], label %[[RED_BODY:[^,]+]]
230 // CHECK: [[RED_BODY]]:
231 // CHECK: [[A_BUF_IDX_SUB_K2POW_ELEM:%.+]] = phi ptr [ [[A_BUF_IDX_SUB_K2POW]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_SUB_K2POW_NEXT:%.+]], %[[RED_BODY]] ]
232 // CHECK: [[A_BUF_IDX_ELEM:%.+]] = phi ptr [ [[A_BUF_IDX]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_NEXT:%.+]], %[[RED_BODY]] ]
233 // CHECK: [[A_BUF_IDX_VAL:%.+]] = load float, ptr [[A_BUF_IDX_ELEM]],
234 // CHECK: [[A_BUF_IDX_SUB_K2POW_VAL:%.+]] = load float, ptr [[A_BUF_IDX_SUB_K2POW_ELEM]],
235 // CHECK: [[RED:%.+]] = fadd float [[A_BUF_IDX_VAL]], [[A_BUF_IDX_SUB_K2POW_VAL]]
236 // CHECK: store float [[RED]], ptr [[A_BUF_IDX_ELEM]],
237 // CHECK: [[A_BUF_IDX_NEXT]] = getelementptr float, ptr [[A_BUF_IDX_ELEM]], i32 1
238 // CHECK: [[A_BUF_IDX_SUB_K2POW_NEXT]] = getelementptr float, ptr [[A_BUF_IDX_SUB_K2POW_ELEM]], i32 1
239 // CHECK: [[DONE:%.+]] = icmp eq ptr [[A_BUF_IDX_NEXT]], [[A_BUF_END]]
240 // CHECK: br i1 [[DONE]], label %[[RED_DONE]], label %[[RED_BODY]]
241 // CHECK: [[RED_DONE]]:
243 // b_buffer[i] += b_buffer[i-pow(2, k)];
244 // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, ptr [[B_BUF_IDX]],
245 // CHECK: [[B_BUF_IDX_SUB_K2POW_VAL:%.+]] = load double, ptr [[B_BUF_IDX_SUB_K2POW]],
246 // CHECK: [[RED:%.+]] = fadd double [[B_BUF_IDX_VAL]], [[B_BUF_IDX_SUB_K2POW_VAL]]
247 // CHECK: store double [[RED]], ptr [[B_BUF_IDX]],
249 // --i;
250 // CHECK: [[I_PREV:%.+]] = sub nuw i64 [[I]], 1
251 // CHECK: [[CMP:%.+]] = icmp uge i64 [[I_PREV]], [[K2POW]]
252 // CHECK: br i1 [[CMP]], label %[[INNER_BODY]], label %[[INNER_EXIT]]
253 // CHECK: [[INNER_EXIT]]:
255 // ++k;
256 // CHECK: [[K_NEXT]] = add nuw i32 [[K]], 1
257 // k2pow <<= 1;
258 // CHECK: [[K2POW_NEXT]] = shl nuw i64 [[K2POW]], 1
259 // CHECK: [[CMP:%.+]] = icmp ne i32 [[K_NEXT]], [[CEIL_LOG2_10_INT]]
260 // CHECK: br i1 [[CMP]], label %[[OUTER_BODY]], label %[[OUTER_EXIT:[^,]+]]
261 // CHECK: [[OUTER_EXIT]]:
262 bar();
263 // CHECK: call void @__kmpc_for_static_init_4(
264 // CHECK: call ptr @llvm.stacksave.p0()
265 // CHECK: store float 0.000000e+00, ptr %
266 // CHECK: store double 0.000000e+00, ptr [[B_PRIV_ADDR:%.+]],
267 // CHECK: br label %[[DISPATCH:[^,]+]]
269 // CHECK: [[SCAN_PHASE:.+]]:
270 // CHECK: call void @{{.+}}foo{{.+}}()
271 // CHECK: br label %[[LOOP_CONTINUE:.+]]
273 // CHECK: [[DISPATCH]]:
274 // if (i >0)
275 // a_priv[[0..n] = a_buffer[i-1][0..n];
276 // CHECK: [[BASE_IDX_I:%.+]] = load i32, ptr [[IV_ADDR:%.+]],
277 // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
278 // CHECK: [[CMP:%.+]] = icmp eq i64 [[BASE_IDX]], 0
279 // CHECK: br i1 [[CMP]], label %[[IF_DONE:[^,]+]], label %[[IF_THEN:[^,]+]]
280 // CHECK: [[IF_THEN]]:
281 // CHECK: [[BASE_IDX_SUB_1:%.+]] = sub nuw i64 [[BASE_IDX]], 1
282 // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX_SUB_1]], [[NUM_ELEMS]]
283 // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds nuw float, ptr [[A_BUF]], i64 [[IDX]]
284 // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds nuw [10 x float], ptr [[A_PRIV_ADDR:%.+]], i64 0, i64 0
285 // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
286 // CHECK: call void @llvm.memcpy.p0.p0.i64(ptr {{.*}}[[A_PRIV]], ptr {{.*}}[[A_BUF_IDX]], i64 [[BYTES]], i1 false)
288 // b_priv = b_buffer[i];
289 // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds nuw double, ptr [[B_BUF]], i64 [[BASE_IDX_SUB_1]]
290 // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, ptr [[B_BUF_IDX]],
291 // CHECK: store double [[B_BUF_IDX_VAL]], ptr [[B_PRIV_ADDR]],
292 // CHECK: br label %[[SCAN_PHASE]]
294 // CHECK: [[LOOP_CONTINUE]]:
295 // CHECK: call void @llvm.stackrestore.p0(ptr %
296 // CHECK: call void @__kmpc_for_static_fini(
297 // CHECK: call void @llvm.stackrestore.p0(ptr
298 // CHECK: call void @__kmpc_barrier(
302 #endif
303 // CHECK: !{!"llvm.loop.vectorize.enable", i1 true}