1 ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --version 2
3 ; RUN: opt < %s -hoist-runtime-checks -p 'loop-vectorize' -force-vector-interleave=1 -S \
4 ; RUN: -force-vector-width=4 -debug-only=loop-accesses,loop-vectorize,loop-utils 2> %t | FileCheck %s
5 ; RUN: cat %t | FileCheck %s --check-prefix=DEBUG
7 target datalayout = "e-m:e-p:32:32-Fi8-i64:64-v128:64:128-a:0:32-n32-S64"
8 target triple = "thumbv8.1m.main-none-unknown-eabi"
10 ; Equivalent example in C:
11 ; void diff_checks(int32_t *dst, int32_t *src, int m, int n) {
12 ; for (int i = 0; i < m; i++) {
13 ; for (int j = 0; j < n; j++) {
14 ; dst[(i * (n + 1)) + j] = src[(i * n) + j];
18 ; NOTE: The strides of the starting address values in the inner loop differ, i.e.
19 ; '(i * (n + 1))' vs '(i * n)'.
21 ; DEBUG-LABEL: 'diff_checks'
22 ; DEBUG: LAA: Found an analyzable loop: inner.loop
23 ; DEBUG: LAA: Not creating diff runtime check, since these cannot be hoisted out of the outer loop
24 ; DEBUG: LAA: Adding RT check for range:
25 ; DEBUG-NEXT: LAA: Expanded RT check for range to include outer loop in order to permit hoisting
26 ; DEBUG-NEXT: LAA: ... but need to check stride is positive: (4 + (4 * %n))
27 ; DEBUG-NEXT: Start: %dst End: ((4 * %n) + ((4 + (4 * %n)) * (-1 + %m)) + %dst)
28 ; DEBUG-NEXT: LAA: Adding RT check for range:
29 ; DEBUG-NEXT: LAA: Expanded RT check for range to include outer loop in order to permit hoisting
30 ; DEBUG-NEXT: LAA: ... but need to check stride is positive: (4 * %n)
31 ; DEBUG-NEXT: Start: %src End: ((4 * %m * %n) + %src)
33 define void @diff_checks(ptr nocapture noundef writeonly %dst, ptr nocapture noundef readonly %src, i32 noundef %m, i32 noundef %n) #0 {
34 ; CHECK-LABEL: define void @diff_checks
35 ; CHECK-SAME: (ptr noundef writeonly captures(none) [[DST:%.*]], ptr noundef readonly captures(none) [[SRC:%.*]], i32 noundef [[M:%.*]], i32 noundef [[N:%.*]]) #[[ATTR0:[0-9]+]] {
37 ; CHECK-NEXT: [[ADD5:%.*]] = add nsw i32 [[N]], 1
38 ; CHECK-NEXT: [[TMP0:%.*]] = add i32 [[M]], -1
39 ; CHECK-NEXT: [[TMP1:%.*]] = shl i32 [[N]], 2
40 ; CHECK-NEXT: [[TMP2:%.*]] = add i32 [[TMP1]], 4
41 ; CHECK-NEXT: [[TMP3:%.*]] = mul i32 [[TMP0]], [[TMP2]]
42 ; CHECK-NEXT: [[TMP4:%.*]] = add i32 [[TMP3]], [[TMP1]]
43 ; CHECK-NEXT: [[SCEVGEP:%.*]] = getelementptr i8, ptr [[DST]], i32 [[TMP4]]
44 ; CHECK-NEXT: [[TMP5:%.*]] = mul i32 [[N]], [[M]]
45 ; CHECK-NEXT: [[TMP6:%.*]] = shl i32 [[TMP5]], 2
46 ; CHECK-NEXT: [[SCEVGEP1:%.*]] = getelementptr i8, ptr [[SRC]], i32 [[TMP6]]
47 ; CHECK-NEXT: br label [[OUTER_LOOP:%.*]]
49 ; CHECK-NEXT: [[I_023_US:%.*]] = phi i32 [ [[INC10_US:%.*]], [[INNER_LOOP_EXIT:%.*]] ], [ 0, [[ENTRY:%.*]] ]
50 ; CHECK-NEXT: [[MUL_US:%.*]] = mul nsw i32 [[I_023_US]], [[N]]
51 ; CHECK-NEXT: [[TMP7:%.*]] = getelementptr i32, ptr [[SRC]], i32 [[MUL_US]]
52 ; CHECK-NEXT: [[MUL6_US:%.*]] = mul nsw i32 [[I_023_US]], [[ADD5]]
53 ; CHECK-NEXT: [[TMP8:%.*]] = getelementptr i32, ptr [[DST]], i32 [[MUL6_US]]
54 ; CHECK-NEXT: br i1 false, label [[SCALAR_PH:%.*]], label [[VECTOR_MEMCHECK:%.*]]
55 ; CHECK: vector.memcheck:
56 ; CHECK-NEXT: [[BOUND0:%.*]] = icmp ult ptr [[DST]], [[SCEVGEP1]]
57 ; CHECK-NEXT: [[BOUND1:%.*]] = icmp ult ptr [[SRC]], [[SCEVGEP]]
58 ; CHECK-NEXT: [[FOUND_CONFLICT:%.*]] = and i1 [[BOUND0]], [[BOUND1]]
59 ; CHECK-NEXT: [[STRIDE_CHECK:%.*]] = icmp slt i32 [[TMP2]], 0
60 ; CHECK-NEXT: [[TMP9:%.*]] = or i1 [[FOUND_CONFLICT]], [[STRIDE_CHECK]]
61 ; CHECK-NEXT: [[STRIDE_CHECK2:%.*]] = icmp slt i32 [[TMP1]], 0
62 ; CHECK-NEXT: [[TMP10:%.*]] = or i1 [[TMP9]], [[STRIDE_CHECK2]]
63 ; CHECK-NEXT: br i1 [[TMP10]], label [[SCALAR_PH]], label [[VECTOR_PH:%.*]]
65 ; CHECK-NEXT: [[N_RND_UP:%.*]] = add i32 [[N]], 3
66 ; CHECK-NEXT: [[N_MOD_VF:%.*]] = urem i32 [[N_RND_UP]], 4
67 ; CHECK-NEXT: [[N_VEC:%.*]] = sub i32 [[N_RND_UP]], [[N_MOD_VF]]
68 ; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
70 ; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, [[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], [[VECTOR_BODY]] ]
71 ; CHECK-NEXT: [[TMP11:%.*]] = add i32 [[INDEX]], 0
72 ; CHECK-NEXT: [[ACTIVE_LANE_MASK:%.*]] = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 [[TMP11]], i32 [[N]])
73 ; CHECK-NEXT: [[TMP12:%.*]] = getelementptr i32, ptr [[TMP7]], i32 [[TMP11]]
74 ; CHECK-NEXT: [[TMP13:%.*]] = getelementptr i32, ptr [[TMP12]], i32 0
75 ; CHECK-NEXT: [[WIDE_MASKED_LOAD:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[TMP13]], i32 4, <4 x i1> [[ACTIVE_LANE_MASK]], <4 x i32> poison), !alias.scope [[META0:![0-9]+]]
76 ; CHECK-NEXT: [[TMP14:%.*]] = getelementptr i32, ptr [[TMP8]], i32 [[TMP11]]
77 ; CHECK-NEXT: [[TMP15:%.*]] = getelementptr i32, ptr [[TMP14]], i32 0
78 ; CHECK-NEXT: call void @llvm.masked.store.v4i32.p0(<4 x i32> [[WIDE_MASKED_LOAD]], ptr [[TMP15]], i32 4, <4 x i1> [[ACTIVE_LANE_MASK]]), !alias.scope [[META3:![0-9]+]], !noalias [[META0]]
79 ; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 4
80 ; CHECK-NEXT: [[TMP16:%.*]] = icmp eq i32 [[INDEX_NEXT]], [[N_VEC]]
81 ; CHECK-NEXT: br i1 [[TMP16]], label [[MIDDLE_BLOCK:%.*]], label [[VECTOR_BODY]], !llvm.loop [[LOOP5:![0-9]+]]
82 ; CHECK: middle.block:
83 ; CHECK-NEXT: br i1 true, label [[INNER_LOOP_EXIT]], label [[SCALAR_PH]]
85 ; CHECK-NEXT: [[BC_RESUME_VAL:%.*]] = phi i32 [ [[N_VEC]], [[MIDDLE_BLOCK]] ], [ 0, [[OUTER_LOOP]] ], [ 0, [[VECTOR_MEMCHECK]] ]
86 ; CHECK-NEXT: br label [[INNER_LOOP:%.*]]
88 ; CHECK-NEXT: [[J_021_US:%.*]] = phi i32 [ [[BC_RESUME_VAL]], [[SCALAR_PH]] ], [ [[INC_US:%.*]], [[INNER_LOOP]] ]
89 ; CHECK-NEXT: [[ARRAYIDX_US:%.*]] = getelementptr i32, ptr [[TMP7]], i32 [[J_021_US]]
90 ; CHECK-NEXT: [[TMP17:%.*]] = load i32, ptr [[ARRAYIDX_US]], align 4
91 ; CHECK-NEXT: [[ARRAYIDX8_US:%.*]] = getelementptr i32, ptr [[TMP8]], i32 [[J_021_US]]
92 ; CHECK-NEXT: store i32 [[TMP17]], ptr [[ARRAYIDX8_US]], align 4
93 ; CHECK-NEXT: [[INC_US]] = add nuw nsw i32 [[J_021_US]], 1
94 ; CHECK-NEXT: [[EXITCOND_NOT:%.*]] = icmp eq i32 [[INC_US]], [[N]]
95 ; CHECK-NEXT: br i1 [[EXITCOND_NOT]], label [[INNER_LOOP_EXIT]], label [[INNER_LOOP]], !llvm.loop [[LOOP8:![0-9]+]]
96 ; CHECK: inner.loop.exit:
97 ; CHECK-NEXT: [[INC10_US]] = add nuw nsw i32 [[I_023_US]], 1
98 ; CHECK-NEXT: [[EXITCOND26_NOT:%.*]] = icmp eq i32 [[INC10_US]], [[M]]
99 ; CHECK-NEXT: br i1 [[EXITCOND26_NOT]], label [[EXIT:%.*]], label [[OUTER_LOOP]]
101 ; CHECK-NEXT: ret void
104 %add5 = add nsw i32 %n, 1
108 %i.023.us = phi i32 [ %inc10.us, %inner.loop.exit ], [ 0, %entry ]
109 %mul.us = mul nsw i32 %i.023.us, %n
110 %0 = getelementptr i32, ptr %src, i32 %mul.us
111 %mul6.us = mul nsw i32 %i.023.us, %add5
112 %1 = getelementptr i32, ptr %dst, i32 %mul6.us
116 %j.021.us = phi i32 [ 0, %outer.loop ], [ %inc.us, %inner.loop ]
117 %arrayidx.us = getelementptr i32, ptr %0, i32 %j.021.us
118 %2 = load i32, ptr %arrayidx.us, align 4
119 %arrayidx8.us = getelementptr i32, ptr %1, i32 %j.021.us
120 store i32 %2, ptr %arrayidx8.us, align 4
121 %inc.us = add nuw nsw i32 %j.021.us, 1
122 %exitcond.not = icmp eq i32 %inc.us, %n
123 br i1 %exitcond.not, label %inner.loop.exit, label %inner.loop
126 %inc10.us = add nuw nsw i32 %i.023.us, 1
127 %exitcond26.not = icmp eq i32 %inc10.us, %m
128 br i1 %exitcond26.not, label %exit, label %outer.loop
134 attributes #0 = { "target-cpu"="cortex-m55" }