1 ; RUN: llvm-as < %s | opt -gvn -enable-load-pre | llvm-dis > %t
2 ; RUN: %prcontext bb3.backedge: 2 < %t | grep phi
3 ; RUN: %prcontext bb3.backedge: 2 < %t | not grep load
5 ; Make sure the load in bb3.backedge is removed and moved into bb1 after the
6 ; call. This makes the non-call case faster.
8 ; This test is derived from this C++ code (GCC PR 37810):
12 ; A& operator++(void) { ++n; if (n == m) g(); return *this; }
13 ; A() : n(0), m(0) { }
14 ; friend bool operator!=(A const& a1, A const& a2) { return a1.n != a2.n; }
16 ; void testfunction(A& iter) { A const end; while (iter != end) ++iter; }
18 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128"
19 target triple = "i386-apple-darwin7"
20 %struct.A = type { i32, i32 }
22 define void @_Z12testfunctionR1A(%struct.A* %iter) {
24 %0 = getelementptr %struct.A* %iter, i32 0, i32 0 ; <i32*> [#uses=3]
25 %1 = load i32* %0, align 4 ; <i32> [#uses=2]
26 %2 = icmp eq i32 %1, 0 ; <i1> [#uses=1]
27 br i1 %2, label %return, label %bb.nph
29 bb.nph: ; preds = %entry
30 %3 = getelementptr %struct.A* %iter, i32 0, i32 1 ; <i32*> [#uses=1]
33 bb: ; preds = %bb3.backedge, %bb.nph
34 %.rle = phi i32 [ %1, %bb.nph ], [ %7, %bb3.backedge ] ; <i32> [#uses=1]
35 %4 = add i32 %.rle, 1 ; <i32> [#uses=2]
36 store i32 %4, i32* %0, align 4
37 %5 = load i32* %3, align 4 ; <i32> [#uses=1]
38 %6 = icmp eq i32 %4, %5 ; <i1> [#uses=1]
39 br i1 %6, label %bb1, label %bb3.backedge
42 tail call void @_Z1gv()
43 br label %bb3.backedge
45 bb3.backedge: ; preds = %bb, %bb1
46 %7 = load i32* %0, align 4 ; <i32> [#uses=2]
47 %8 = icmp eq i32 %7, 0 ; <i1> [#uses=1]
48 br i1 %8, label %return, label %bb
50 return: ; preds = %bb3.backedge, %entry