Revert "[llvm] Improve llvm.objectsize computation by computing GEP, alloca and mallo...
[llvm-project.git] / clang / test / CodeGen / builtins-multiprecision.c
blob66e833319f82ccfef906d4090da5ca49f4aeab17
1 // RUN: %clang_cc1 -triple "i686-unknown-unknown" -emit-llvm -x c %s -o - -O3 | FileCheck %s
2 // RUN: %clang_cc1 -triple "x86_64-unknown-unknown" -emit-llvm -x c %s -o - -O3 | FileCheck %s
3 // RUN: %clang_cc1 -triple "x86_64-mingw32" -emit-llvm -x c %s -o - -O3 | FileCheck %s
5 unsigned char test_addcb(unsigned char x, unsigned char y,
6 unsigned char carryin, unsigned char *z) {
7 // CHECK: @test_addcb
8 // CHECK: %{{.+}} = {{.*}} call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %x, i8 %y)
9 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 1
10 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 0
11 // CHECK: %{{.+}} = {{.*}} call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %{{.+}}, i8 %carryin)
12 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 1
13 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 0
14 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
15 // CHECK: %{{.+}} = zext i1 %{{.+}} to i8
16 // CHECK: store i8 %{{.+}}, ptr %z, align 1
18 unsigned char carryout;
19 *z = __builtin_addcb(x, y, carryin, &carryout);
21 return carryout;
24 unsigned short test_addcs(unsigned short x, unsigned short y,
25 unsigned short carryin, unsigned short *z) {
26 // CHECK: @test_addcs
27 // CHECK: %{{.+}} = {{.*}} call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %x, i16 %y)
28 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 1
29 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 0
30 // CHECK: %{{.+}} = {{.*}} call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %{{.+}}, i16 %carryin)
31 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 1
32 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 0
33 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
34 // CHECK: %{{.+}} = zext i1 %{{.+}} to i16
35 // CHECK: store i16 %{{.+}}, ptr %z, align 2
37 unsigned short carryout;
38 *z = __builtin_addcs(x, y, carryin, &carryout);
40 return carryout;
43 unsigned test_addc(unsigned x, unsigned y, unsigned carryin, unsigned *z) {
44 // CHECK: @test_addc
45 // CHECK: %{{.+}} = {{.*}} call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %x, i32 %y)
46 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 1
47 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 0
48 // CHECK: %{{.+}} = {{.*}} call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %{{.+}}, i32 %carryin)
49 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 1
50 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 0
51 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
52 // CHECK: %{{.+}} = zext i1 %{{.+}} to i32
53 // CHECK: store i32 %{{.+}}, ptr %z, align 4
54 unsigned carryout;
55 *z = __builtin_addc(x, y, carryin, &carryout);
57 return carryout;
60 unsigned long test_addcl(unsigned long x, unsigned long y,
61 unsigned long carryin, unsigned long *z) {
62 // long is i32 on i686, i64 on x86_64.
63 // CHECK: @test_addcl([[UL:i32|i64]] noundef %x
64 // CHECK: %{{.+}} = {{.*}} call { [[UL]], i1 } @llvm.uadd.with.overflow.[[UL]]([[UL]] %x, [[UL]] %y)
65 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 1
66 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 0
67 // CHECK: %{{.+}} = {{.*}} call { [[UL]], i1 } @llvm.uadd.with.overflow.[[UL]]([[UL]] %{{.+}}, [[UL]] %carryin)
68 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 1
69 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 0
70 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
71 // CHECK: %{{.+}} = zext i1 %{{.+}} to [[UL]]
72 // CHECK: store [[UL]] %{{.+}}, ptr %z
73 unsigned long carryout;
74 *z = __builtin_addcl(x, y, carryin, &carryout);
76 return carryout;
79 unsigned long long test_addcll(unsigned long long x, unsigned long long y,
80 unsigned long long carryin,
81 unsigned long long *z) {
82 // CHECK: @test_addcll
83 // CHECK: %{{.+}} = {{.*}} call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %x, i64 %y)
84 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 1
85 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 0
86 // CHECK: %{{.+}} = {{.*}} call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %{{.+}}, i64 %carryin)
87 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 1
88 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 0
89 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
90 // CHECK: %{{.+}} = zext i1 %{{.+}} to i64
91 // CHECK: store i64 %{{.+}}, ptr %z
92 unsigned long long carryout;
93 *z = __builtin_addcll(x, y, carryin, &carryout);
95 return carryout;
98 unsigned char test_subcb(unsigned char x, unsigned char y,
99 unsigned char carryin, unsigned char *z) {
100 // CHECK: @test_subcb
101 // CHECK: %{{.+}} = {{.*}} call { i8, i1 } @llvm.usub.with.overflow.i8(i8 %x, i8 %y)
102 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 1
103 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 0
104 // CHECK: %{{.+}} = {{.*}} call { i8, i1 } @llvm.usub.with.overflow.i8(i8 %{{.+}}, i8 %carryin)
105 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 1
106 // CHECK: %{{.+}} = extractvalue { i8, i1 } %{{.+}}, 0
107 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
108 // CHECK: %{{.+}} = zext i1 %{{.+}} to i8
109 // CHECK: store i8 %{{.+}}, ptr %z, align 1
111 unsigned char carryout;
112 *z = __builtin_subcb(x, y, carryin, &carryout);
114 return carryout;
117 unsigned short test_subcs(unsigned short x, unsigned short y,
118 unsigned short carryin, unsigned short *z) {
119 // CHECK: @test_subcs
120 // CHECK: %{{.+}} = {{.*}} call { i16, i1 } @llvm.usub.with.overflow.i16(i16 %x, i16 %y)
121 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 1
122 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 0
123 // CHECK: %{{.+}} = {{.*}} call { i16, i1 } @llvm.usub.with.overflow.i16(i16 %{{.+}}, i16 %carryin)
124 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 1
125 // CHECK: %{{.+}} = extractvalue { i16, i1 } %{{.+}}, 0
126 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
127 // CHECK: %{{.+}} = zext i1 %{{.+}} to i16
128 // CHECK: store i16 %{{.+}}, ptr %z, align 2
130 unsigned short carryout;
131 *z = __builtin_subcs(x, y, carryin, &carryout);
133 return carryout;
136 unsigned test_subc(unsigned x, unsigned y, unsigned carryin, unsigned *z) {
137 // CHECK: @test_subc
138 // CHECK: %{{.+}} = {{.*}} call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %x, i32 %y)
139 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 1
140 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 0
141 // CHECK: %{{.+}} = {{.*}} call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %{{.+}}, i32 %carryin)
142 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 1
143 // CHECK: %{{.+}} = extractvalue { i32, i1 } %{{.+}}, 0
144 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
145 // CHECK: %{{.+}} = zext i1 %{{.+}} to i32
146 // CHECK: store i32 %{{.+}}, ptr %z, align 4
147 unsigned carryout;
148 *z = __builtin_subc(x, y, carryin, &carryout);
150 return carryout;
153 unsigned long test_subcl(unsigned long x, unsigned long y,
154 unsigned long carryin, unsigned long *z) {
155 // CHECK: @test_subcl([[UL:i32|i64]] noundef %x
156 // CHECK: %{{.+}} = {{.*}} call { [[UL]], i1 } @llvm.usub.with.overflow.[[UL]]([[UL]] %x, [[UL]] %y)
157 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 1
158 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 0
159 // CHECK: %{{.+}} = {{.*}} call { [[UL]], i1 } @llvm.usub.with.overflow.[[UL]]([[UL]] %{{.+}}, [[UL]] %carryin)
160 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 1
161 // CHECK: %{{.+}} = extractvalue { [[UL]], i1 } %{{.+}}, 0
162 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
163 // CHECK: %{{.+}} = zext i1 %{{.+}} to [[UL]]
164 // CHECK: store [[UL]] %{{.+}}, ptr %z
165 unsigned long carryout;
166 *z = __builtin_subcl(x, y, carryin, &carryout);
168 return carryout;
171 unsigned long long test_subcll(unsigned long long x, unsigned long long y,
172 unsigned long long carryin,
173 unsigned long long *z) {
174 // CHECK: @test_subcll
175 // CHECK: %{{.+}} = {{.*}} call { i64, i1 } @llvm.usub.with.overflow.i64(i64 %x, i64 %y)
176 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 1
177 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 0
178 // CHECK: %{{.+}} = {{.*}} call { i64, i1 } @llvm.usub.with.overflow.i64(i64 %{{.+}}, i64 %carryin)
179 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 1
180 // CHECK: %{{.+}} = extractvalue { i64, i1 } %{{.+}}, 0
181 // CHECK: %{{.+}} = or i1 %{{.+}}, %{{.+}}
182 // CHECK: %{{.+}} = zext i1 %{{.+}} to i64
183 // CHECK: store i64 %{{.+}}, ptr %z
184 unsigned long long carryout;
185 *z = __builtin_subcll(x, y, carryin, &carryout);
187 return carryout;