1 ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --version 5
2 ; RUN: opt < %s -passes=loop-vectorize -force-vector-width=4 -force-vector-interleave=1 -S | FileCheck %s
4 ; Check that we correctly handle the use of %start2 in the exit block, and do
7 define i64 @foo(ptr %p1, ptr %p2, i64 %start, i64 %end) {
8 ; CHECK-LABEL: define i64 @foo(
9 ; CHECK-SAME: ptr [[P1:%.*]], ptr [[P2:%.*]], i64 [[START:%.*]], i64 [[END:%.*]]) {
10 ; CHECK-NEXT: [[ENTRY:.*]]:
11 ; CHECK-NEXT: [[START2:%.*]] = and i64 [[START]], 12345
12 ; CHECK-NEXT: [[TMP0:%.*]] = sub i64 [[END]], [[START2]]
13 ; CHECK-NEXT: [[MIN_ITERS_CHECK:%.*]] = icmp ult i64 [[TMP0]], 4
14 ; CHECK-NEXT: br i1 [[MIN_ITERS_CHECK]], label %[[SCALAR_PH:.*]], label %[[VECTOR_PH:.*]]
15 ; CHECK: [[VECTOR_PH]]:
16 ; CHECK-NEXT: [[N_MOD_VF:%.*]] = urem i64 [[TMP0]], 4
17 ; CHECK-NEXT: [[N_VEC:%.*]] = sub i64 [[TMP0]], [[N_MOD_VF]]
18 ; CHECK-NEXT: [[IND_END:%.*]] = add i64 [[START2]], [[N_VEC]]
19 ; CHECK-NEXT: br label %[[VECTOR_BODY:.*]]
20 ; CHECK: [[VECTOR_BODY]]:
21 ; CHECK-NEXT: [[INDEX:%.*]] = phi i64 [ 0, %[[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], %[[VECTOR_BODY]] ]
22 ; CHECK-NEXT: [[INDEX_NEXT]] = add nuw i64 [[INDEX]], 4
23 ; CHECK-NEXT: [[TMP6:%.*]] = icmp eq i64 [[INDEX_NEXT]], [[N_VEC]]
24 ; CHECK-NEXT: br i1 [[TMP6]], label %[[MIDDLE_BLOCK:.*]], label %[[VECTOR_BODY]], !llvm.loop [[LOOP0:![0-9]+]]
25 ; CHECK: [[MIDDLE_BLOCK]]:
26 ; CHECK-NEXT: [[CMP_N:%.*]] = icmp eq i64 [[TMP0]], [[N_VEC]]
27 ; CHECK-NEXT: br i1 [[CMP_N]], label %[[EXIT:.*]], label %[[SCALAR_PH]]
28 ; CHECK: [[SCALAR_PH]]:
29 ; CHECK-NEXT: [[BC_RESUME_VAL:%.*]] = phi i64 [ [[IND_END]], %[[MIDDLE_BLOCK]] ], [ [[START2]], %[[ENTRY]] ]
30 ; CHECK-NEXT: br label %[[FOR_BODY:.*]]
31 ; CHECK: [[FOR_BODY]]:
32 ; CHECK-NEXT: [[IND1:%.*]] = phi i64 [ [[BC_RESUME_VAL]], %[[SCALAR_PH]] ], [ [[IND_NEXT1:%.*]], %[[FOR_BODY]] ]
33 ; CHECK-NEXT: [[ARRAYIDX3:%.*]] = getelementptr inbounds i32, ptr [[P1]], i64 [[IND1]]
34 ; CHECK-NEXT: [[TMP7:%.*]] = load i32, ptr [[ARRAYIDX3]], align 4
35 ; CHECK-NEXT: [[ARRAYIDX2:%.*]] = getelementptr inbounds i32, ptr [[P2]], i64 [[IND1]]
36 ; CHECK-NEXT: [[TMP8:%.*]] = load i32, ptr [[ARRAYIDX2]], align 4
37 ; CHECK-NEXT: [[IND_NEXT1]] = add i64 [[IND1]], 1
38 ; CHECK-NEXT: [[CMP:%.*]] = icmp ne i64 [[IND_NEXT1]], [[END]]
39 ; CHECK-NEXT: br i1 [[CMP]], label %[[FOR_BODY]], label %[[EXIT]], !llvm.loop [[LOOP3:![0-9]+]]
41 ; CHECK-NEXT: [[USE:%.*]] = phi i64 [ [[START2]], %[[FOR_BODY]] ], [ [[START2]], %[[MIDDLE_BLOCK]] ]
42 ; CHECK-NEXT: ret i64 [[USE]]
45 %start2 = and i64 %start, 12345
49 %ind = phi i64 [ %start2, %entry ], [ %ind.next, %for.body ]
50 %arrayidx1 = getelementptr inbounds i32, ptr %p1, i64 %ind
51 %0 = load i32, ptr %arrayidx1, align 4
52 %arrayidx2 = getelementptr inbounds i32, ptr %p2, i64 %ind
53 %1 = load i32, ptr %arrayidx2, align 4
54 %ind.next = add i64 %ind, 1
55 %cmp = icmp ne i64 %ind.next, %end
56 br i1 %cmp, label %for.body, label %exit
59 %use = phi i64 [ %start2, %for.body ]
64 ; CHECK: [[LOOP0]] = distinct !{[[LOOP0]], [[META1:![0-9]+]], [[META2:![0-9]+]]}
65 ; CHECK: [[META1]] = !{!"llvm.loop.isvectorized", i32 1}
66 ; CHECK: [[META2]] = !{!"llvm.loop.unroll.runtime.disable"}
67 ; CHECK: [[LOOP3]] = distinct !{[[LOOP3]], [[META2]], [[META1]]}