1 ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
2 ; RUN: opt < %s -passes=instcombine -S | FileCheck %s
6 define i8 @shl_and(i8 %x, i8 %y) {
7 ; CHECK-LABEL: @shl_and(
8 ; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[X:%.*]], 5
9 ; CHECK-NEXT: [[TMP2:%.*]] = shl i8 [[Y:%.*]], 2
10 ; CHECK-NEXT: [[SH1:%.*]] = and i8 [[TMP1]], [[TMP2]]
11 ; CHECK-NEXT: ret i8 [[SH1]]
19 define <2 x i8> @shl_and_nonuniform(<2 x i8> %x, <2 x i8> %y) {
20 ; CHECK-LABEL: @shl_and_nonuniform(
21 ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i8> [[X:%.*]], <i8 5, i8 4>
22 ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i8> [[Y:%.*]], <i8 2, i8 0>
23 ; CHECK-NEXT: [[SH1:%.*]] = and <2 x i8> [[TMP1]], [[TMP2]]
24 ; CHECK-NEXT: ret <2 x i8> [[SH1]]
26 %sh0 = shl <2 x i8> %x, <i8 3, i8 4>
27 %r = and <2 x i8> %sh0, %y
28 %sh1 = shl <2 x i8> %r, <i8 2, i8 0>
32 define i16 @shl_or(i16 %x, i16 %py) {
33 ; CHECK-LABEL: @shl_or(
34 ; CHECK-NEXT: [[Y:%.*]] = srem i16 [[PY:%.*]], 42
35 ; CHECK-NEXT: [[TMP1:%.*]] = shl i16 [[X:%.*]], 12
36 ; CHECK-NEXT: [[TMP2:%.*]] = shl nsw i16 [[Y]], 7
37 ; CHECK-NEXT: [[SH1:%.*]] = or i16 [[TMP1]], [[TMP2]]
38 ; CHECK-NEXT: ret i16 [[SH1]]
40 %y = srem i16 %py, 42 ; thwart complexity-based canonicalization
47 define <2 x i16> @shl_or_undef(<2 x i16> %x, <2 x i16> %py) {
48 ; CHECK-LABEL: @shl_or_undef(
49 ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i16> [[PY:%.*]], <i16 42, i16 42>
50 ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i16> [[X:%.*]], <i16 12, i16 undef>
51 ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i16> [[Y]], <i16 7, i16 undef>
52 ; CHECK-NEXT: [[SH1:%.*]] = or <2 x i16> [[TMP1]], [[TMP2]]
53 ; CHECK-NEXT: ret <2 x i16> [[SH1]]
55 %y = srem <2 x i16> %py, <i16 42, i16 42> ; thwart complexity-based canonicalization
56 %sh0 = shl <2 x i16> %x, <i16 5, i16 undef>
57 %r = or <2 x i16> %y, %sh0
58 %sh1 = shl <2 x i16> %r, <i16 7, i16 undef>
62 define i32 @shl_xor(i32 %x, i32 %y) {
63 ; CHECK-LABEL: @shl_xor(
64 ; CHECK-NEXT: [[TMP1:%.*]] = shl i32 [[X:%.*]], 12
65 ; CHECK-NEXT: [[TMP2:%.*]] = shl i32 [[Y:%.*]], 7
66 ; CHECK-NEXT: [[SH1:%.*]] = xor i32 [[TMP1]], [[TMP2]]
67 ; CHECK-NEXT: ret i32 [[SH1]]
75 define <2 x i32> @shl_xor_nonuniform(<2 x i32> %x, <2 x i32> %y) {
76 ; CHECK-LABEL: @shl_xor_nonuniform(
77 ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i32> [[X:%.*]], <i32 12, i32 14>
78 ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 7, i32 8>
79 ; CHECK-NEXT: [[SH1:%.*]] = xor <2 x i32> [[TMP1]], [[TMP2]]
80 ; CHECK-NEXT: ret <2 x i32> [[SH1]]
82 %sh0 = shl <2 x i32> %x, <i32 5, i32 6>
83 %r = xor <2 x i32> %sh0, %y
84 %sh1 = shl <2 x i32> %r, <i32 7, i32 8>
88 define i64 @lshr_and(i64 %x, i64 %py) {
89 ; CHECK-LABEL: @lshr_and(
90 ; CHECK-NEXT: [[Y:%.*]] = srem i64 [[PY:%.*]], 42
91 ; CHECK-NEXT: [[TMP1:%.*]] = lshr i64 [[X:%.*]], 12
92 ; CHECK-NEXT: [[TMP2:%.*]] = lshr i64 [[Y]], 7
93 ; CHECK-NEXT: [[SH1:%.*]] = and i64 [[TMP1]], [[TMP2]]
94 ; CHECK-NEXT: ret i64 [[SH1]]
96 %y = srem i64 %py, 42 ; thwart complexity-based canonicalization
103 define <2 x i64> @lshr_and_undef(<2 x i64> %x, <2 x i64> %py) {
104 ; CHECK-LABEL: @lshr_and_undef(
105 ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42>
106 ; CHECK-NEXT: [[TMP1:%.*]] = lshr <2 x i64> [[X:%.*]], <i64 12, i64 undef>
107 ; CHECK-NEXT: [[TMP2:%.*]] = lshr <2 x i64> [[Y]], <i64 7, i64 undef>
108 ; CHECK-NEXT: [[SH1:%.*]] = and <2 x i64> [[TMP1]], [[TMP2]]
109 ; CHECK-NEXT: ret <2 x i64> [[SH1]]
111 %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization
112 %sh0 = lshr <2 x i64> %x, <i64 5, i64 undef>
113 %r = and <2 x i64> %y, %sh0
114 %sh1 = lshr <2 x i64> %r, <i64 7, i64 undef>
118 define <4 x i32> @lshr_or(<4 x i32> %x, <4 x i32> %y) {
119 ; CHECK-LABEL: @lshr_or(
120 ; CHECK-NEXT: [[TMP1:%.*]] = lshr <4 x i32> [[X:%.*]], <i32 12, i32 12, i32 12, i32 12>
121 ; CHECK-NEXT: [[TMP2:%.*]] = lshr <4 x i32> [[Y:%.*]], <i32 7, i32 7, i32 7, i32 7>
122 ; CHECK-NEXT: [[SH1:%.*]] = or <4 x i32> [[TMP1]], [[TMP2]]
123 ; CHECK-NEXT: ret <4 x i32> [[SH1]]
125 %sh0 = lshr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5>
126 %r = or <4 x i32> %sh0, %y
127 %sh1 = lshr <4 x i32> %r, <i32 7, i32 7, i32 7, i32 7>
131 define <8 x i16> @lshr_xor(<8 x i16> %x, <8 x i16> %py) {
132 ; CHECK-LABEL: @lshr_xor(
133 ; CHECK-NEXT: [[Y:%.*]] = srem <8 x i16> [[PY:%.*]], <i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42>
134 ; CHECK-NEXT: [[TMP1:%.*]] = lshr <8 x i16> [[X:%.*]], <i16 12, i16 12, i16 12, i16 12, i16 12, i16 12, i16 12, i16 12>
135 ; CHECK-NEXT: [[TMP2:%.*]] = lshr <8 x i16> [[Y]], <i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7>
136 ; CHECK-NEXT: [[SH1:%.*]] = xor <8 x i16> [[TMP1]], [[TMP2]]
137 ; CHECK-NEXT: ret <8 x i16> [[SH1]]
139 %y = srem <8 x i16> %py, <i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 -42> ; thwart complexity-based canonicalization
140 %sh0 = lshr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5>
141 %r = xor <8 x i16> %y, %sh0
142 %sh1 = lshr <8 x i16> %r, <i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7>
146 define <16 x i8> @ashr_and(<16 x i8> %x, <16 x i8> %py, <16 x i8> %pz) {
147 ; CHECK-LABEL: @ashr_and(
148 ; CHECK-NEXT: [[Y:%.*]] = srem <16 x i8> [[PY:%.*]], [[PZ:%.*]]
149 ; CHECK-NEXT: [[TMP1:%.*]] = ashr <16 x i8> [[X:%.*]], <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5>
150 ; CHECK-NEXT: [[TMP2:%.*]] = ashr <16 x i8> [[Y]], <i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2>
151 ; CHECK-NEXT: [[SH1:%.*]] = and <16 x i8> [[TMP1]], [[TMP2]]
152 ; CHECK-NEXT: ret <16 x i8> [[SH1]]
154 %y = srem <16 x i8> %py, %pz ; thwart complexity-based canonicalization
155 %sh0 = ashr <16 x i8> %x, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
156 %r = and <16 x i8> %y, %sh0
157 %sh1 = ashr <16 x i8> %r, <i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2>
161 define <2 x i64> @ashr_or(<2 x i64> %x, <2 x i64> %y) {
162 ; CHECK-LABEL: @ashr_or(
163 ; CHECK-NEXT: [[TMP1:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 12, i64 12>
164 ; CHECK-NEXT: [[TMP2:%.*]] = ashr <2 x i64> [[Y:%.*]], <i64 7, i64 7>
165 ; CHECK-NEXT: [[SH1:%.*]] = or <2 x i64> [[TMP1]], [[TMP2]]
166 ; CHECK-NEXT: ret <2 x i64> [[SH1]]
168 %sh0 = ashr <2 x i64> %x, <i64 5, i64 5>
169 %r = or <2 x i64> %sh0, %y
170 %sh1 = ashr <2 x i64> %r, <i64 7, i64 7>
174 define i32 @ashr_xor(i32 %x, i32 %py) {
175 ; CHECK-LABEL: @ashr_xor(
176 ; CHECK-NEXT: [[Y:%.*]] = srem i32 [[PY:%.*]], 42
177 ; CHECK-NEXT: [[TMP1:%.*]] = ashr i32 [[X:%.*]], 12
178 ; CHECK-NEXT: [[TMP2:%.*]] = ashr i32 [[Y]], 7
179 ; CHECK-NEXT: [[SH1:%.*]] = xor i32 [[TMP1]], [[TMP2]]
180 ; CHECK-NEXT: ret i32 [[SH1]]
182 %y = srem i32 %py, 42 ; thwart complexity-based canonicalization
183 %sh0 = ashr i32 %x, 5
184 %r = xor i32 %y, %sh0
185 %sh1 = ashr i32 %r, 7
189 define i32 @shr_mismatch_xor(i32 %x, i32 %y) {
190 ; CHECK-LABEL: @shr_mismatch_xor(
191 ; CHECK-NEXT: [[SH0:%.*]] = ashr i32 [[X:%.*]], 5
192 ; CHECK-NEXT: [[R:%.*]] = xor i32 [[SH0]], [[Y:%.*]]
193 ; CHECK-NEXT: [[SH1:%.*]] = lshr i32 [[R]], 7
194 ; CHECK-NEXT: ret i32 [[SH1]]
196 %sh0 = ashr i32 %x, 5
197 %r = xor i32 %y, %sh0
198 %sh1 = lshr i32 %r, 7
202 define i32 @ashr_overshift_xor(i32 %x, i32 %y) {
203 ; CHECK-LABEL: @ashr_overshift_xor(
204 ; CHECK-NEXT: [[SH0:%.*]] = ashr i32 [[X:%.*]], 15
205 ; CHECK-NEXT: [[R:%.*]] = xor i32 [[SH0]], [[Y:%.*]]
206 ; CHECK-NEXT: [[SH1:%.*]] = ashr i32 [[R]], 17
207 ; CHECK-NEXT: ret i32 [[SH1]]
209 %sh0 = ashr i32 %x, 15
210 %r = xor i32 %y, %sh0
211 %sh1 = ashr i32 %r, 17
215 define <2 x i32> @ashr_undef_undef_xor(<2 x i32> %x, <2 x i32> %y) {
216 ; CHECK-LABEL: @ashr_undef_undef_xor(
217 ; CHECK-NEXT: [[SH0:%.*]] = ashr <2 x i32> [[X:%.*]], <i32 15, i32 undef>
218 ; CHECK-NEXT: [[R:%.*]] = xor <2 x i32> [[SH0]], [[Y:%.*]]
219 ; CHECK-NEXT: [[SH1:%.*]] = ashr <2 x i32> [[R]], <i32 undef, i32 17>
220 ; CHECK-NEXT: ret <2 x i32> [[SH1]]
222 %sh0 = ashr <2 x i32> %x, <i32 15, i32 undef>
223 %r = xor <2 x i32> %y, %sh0
224 %sh1 = ashr <2 x i32> %r, <i32 undef, i32 17>
228 define i32 @lshr_or_extra_use(i32 %x, i32 %y, i32* %p) {
229 ; CHECK-LABEL: @lshr_or_extra_use(
230 ; CHECK-NEXT: [[SH0:%.*]] = lshr i32 [[X:%.*]], 5
231 ; CHECK-NEXT: [[R:%.*]] = or i32 [[SH0]], [[Y:%.*]]
232 ; CHECK-NEXT: store i32 [[R]], i32* [[P:%.*]], align 4
233 ; CHECK-NEXT: [[SH1:%.*]] = lshr i32 [[R]], 7
234 ; CHECK-NEXT: ret i32 [[SH1]]
236 %sh0 = lshr i32 %x, 5
238 store i32 %r, i32* %p
239 %sh1 = lshr i32 %r, 7
243 ; Avoid crashing on constant expressions.
245 @g = external global i32
247 define i32 @PR44028(i32 %x) {
248 ; CHECK-LABEL: @PR44028(
249 ; CHECK-NEXT: [[SH1:%.*]] = ashr exact i32 [[X:%.*]], 16
250 ; CHECK-NEXT: [[T0:%.*]] = xor i32 [[SH1]], shl (i32 ptrtoint (i32* @g to i32), i32 16)
251 ; CHECK-NEXT: [[T27:%.*]] = ashr exact i32 [[T0]], 16
252 ; CHECK-NEXT: ret i32 [[T27]]
254 %sh1 = ashr exact i32 %x, 16
255 %t0 = xor i32 %sh1, shl (i32 ptrtoint (i32* @g to i32), i32 16)
256 %t27 = ashr exact i32 %t0, 16
260 define i64 @lshr_mul(i64 %0) {
261 ; CHECK-LABEL: @lshr_mul(
262 ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 13
263 ; CHECK-NEXT: ret i64 [[TMP2]]
265 %2 = mul nuw i64 %0, 52
270 define i64 @lshr_mul_nuw_nsw(i64 %0) {
271 ; CHECK-LABEL: @lshr_mul_nuw_nsw(
272 ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw nsw i64 [[TMP0:%.*]], 13
273 ; CHECK-NEXT: ret i64 [[TMP2]]
275 %2 = mul nuw nsw i64 %0, 52
280 define <4 x i32> @lshr_mul_vector(<4 x i32> %0) {
281 ; CHECK-LABEL: @lshr_mul_vector(
282 ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw <4 x i32> [[TMP0:%.*]], <i32 13, i32 13, i32 13, i32 13>
283 ; CHECK-NEXT: ret <4 x i32> [[TMP2]]
285 %2 = mul nuw <4 x i32> %0, <i32 52, i32 52, i32 52, i32 52>
286 %3 = lshr <4 x i32> %2, <i32 2, i32 2, i32 2, i32 2>
290 define i64 @lshr_mul_negative_noexact(i64 %0) {
291 ; CHECK-LABEL: @lshr_mul_negative_noexact(
292 ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 53
293 ; CHECK-NEXT: [[TMP3:%.*]] = lshr i64 [[TMP2]], 2
294 ; CHECK-NEXT: ret i64 [[TMP3]]
296 %2 = mul nuw i64 %0, 53
301 define i64 @lshr_mul_negative_oneuse(i64 %0) {
302 ; CHECK-LABEL: @lshr_mul_negative_oneuse(
303 ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 52
304 ; CHECK-NEXT: call void @use(i64 [[TMP2]])
305 ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2
306 ; CHECK-NEXT: ret i64 [[TMP3]]
308 %2 = mul nuw i64 %0, 52
309 call void @use(i64 %2)
314 define i64 @lshr_mul_negative_nonuw(i64 %0) {
315 ; CHECK-LABEL: @lshr_mul_negative_nonuw(
316 ; CHECK-NEXT: [[TMP2:%.*]] = mul i64 [[TMP0:%.*]], 52
317 ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2
318 ; CHECK-NEXT: ret i64 [[TMP3]]
325 define i64 @lshr_mul_negative_nsw(i64 %0) {
326 ; CHECK-LABEL: @lshr_mul_negative_nsw(
327 ; CHECK-NEXT: [[TMP2:%.*]] = mul nsw i64 [[TMP0:%.*]], 52
328 ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2
329 ; CHECK-NEXT: ret i64 [[TMP3]]
331 %2 = mul nsw i64 %0, 52