1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
2 ; RUN: llc < %s -mtriple=i686-unknown-linux-gnu | FileCheck %s --check-prefix=X86
3 ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu | FileCheck %s --check-prefix=X64
5 define void @i24_or(i24* %a) {
8 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
9 ; X86-NEXT: movzwl (%ecx), %edx
10 ; X86-NEXT: movzbl 2(%ecx), %eax
11 ; X86-NEXT: movb %al, 2(%ecx)
12 ; X86-NEXT: shll $16, %eax
13 ; X86-NEXT: orl %edx, %eax
14 ; X86-NEXT: orl $384, %eax # imm = 0x180
15 ; X86-NEXT: movw %ax, (%ecx)
20 ; X64-NEXT: movzwl (%rdi), %eax
21 ; X64-NEXT: movzbl 2(%rdi), %ecx
22 ; X64-NEXT: movb %cl, 2(%rdi)
23 ; X64-NEXT: shll $16, %ecx
24 ; X64-NEXT: orl %eax, %ecx
25 ; X64-NEXT: orl $384, %ecx # imm = 0x180
26 ; X64-NEXT: movw %cx, (%rdi)
28 %aa = load i24, i24* %a, align 1
30 store i24 %b, i24* %a, align 1
34 define void @i24_and_or(i24* %a) {
35 ; X86-LABEL: i24_and_or:
37 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
38 ; X86-NEXT: movzwl (%ecx), %edx
39 ; X86-NEXT: movzbl 2(%ecx), %eax
40 ; X86-NEXT: movb %al, 2(%ecx)
41 ; X86-NEXT: shll $16, %eax
42 ; X86-NEXT: orl %edx, %eax
43 ; X86-NEXT: orl $384, %eax # imm = 0x180
44 ; X86-NEXT: andl $16777088, %eax # imm = 0xFFFF80
45 ; X86-NEXT: movw %ax, (%ecx)
48 ; X64-LABEL: i24_and_or:
50 ; X64-NEXT: movzwl (%rdi), %eax
51 ; X64-NEXT: movzbl 2(%rdi), %ecx
52 ; X64-NEXT: movb %cl, 2(%rdi)
53 ; X64-NEXT: shll $16, %ecx
54 ; X64-NEXT: orl %eax, %ecx
55 ; X64-NEXT: orl $384, %ecx # imm = 0x180
56 ; X64-NEXT: andl $16777088, %ecx # imm = 0xFFFF80
57 ; X64-NEXT: movw %cx, (%rdi)
59 %b = load i24, i24* %a, align 1
62 store i24 %d, i24* %a, align 1
66 define void @i24_insert_bit(i24* %a, i1 zeroext %bit) {
67 ; X86-LABEL: i24_insert_bit:
69 ; X86-NEXT: pushl %esi
70 ; X86-NEXT: .cfi_def_cfa_offset 8
71 ; X86-NEXT: .cfi_offset %esi, -8
72 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
73 ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %edx
74 ; X86-NEXT: movzwl (%ecx), %esi
75 ; X86-NEXT: movzbl 2(%ecx), %eax
76 ; X86-NEXT: movb %al, 2(%ecx)
77 ; X86-NEXT: shll $16, %eax
78 ; X86-NEXT: orl %esi, %eax
79 ; X86-NEXT: shll $13, %edx
80 ; X86-NEXT: andl $16769023, %eax # imm = 0xFFDFFF
81 ; X86-NEXT: orl %edx, %eax
82 ; X86-NEXT: movw %ax, (%ecx)
86 ; X64-LABEL: i24_insert_bit:
88 ; X64-NEXT: movzwl (%rdi), %eax
89 ; X64-NEXT: movzbl 2(%rdi), %ecx
90 ; X64-NEXT: movb %cl, 2(%rdi)
91 ; X64-NEXT: shll $16, %ecx
92 ; X64-NEXT: orl %eax, %ecx
93 ; X64-NEXT: shll $13, %esi
94 ; X64-NEXT: andl $16769023, %ecx # imm = 0xFFDFFF
95 ; X64-NEXT: orl %esi, %ecx
96 ; X64-NEXT: movw %cx, (%rdi)
98 %extbit = zext i1 %bit to i24
99 %b = load i24, i24* %a, align 1
100 %extbit.shl = shl nuw nsw i24 %extbit, 13
101 %c = and i24 %b, -8193
102 %d = or i24 %c, %extbit.shl
103 store i24 %d, i24* %a, align 1
107 define void @i56_or(i56* %a) {
110 ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
111 ; X86-NEXT: orl $384, (%eax) # imm = 0x180
116 ; X64-NEXT: movzwl 4(%rdi), %eax
117 ; X64-NEXT: movzbl 6(%rdi), %ecx
118 ; X64-NEXT: movb %cl, 6(%rdi)
119 ; X64-NEXT: # kill: %ecx<def> %ecx<kill> %rcx<kill> %rcx<def>
120 ; X64-NEXT: shll $16, %ecx
121 ; X64-NEXT: orl %eax, %ecx
122 ; X64-NEXT: shlq $32, %rcx
123 ; X64-NEXT: movl (%rdi), %eax
124 ; X64-NEXT: orq %rcx, %rax
125 ; X64-NEXT: orq $384, %rax # imm = 0x180
126 ; X64-NEXT: movl %eax, (%rdi)
127 ; X64-NEXT: shrq $32, %rax
128 ; X64-NEXT: movw %ax, 4(%rdi)
130 %aa = load i56, i56* %a, align 1
132 store i56 %b, i56* %a, align 1
136 define void @i56_and_or(i56* %a) {
137 ; X86-LABEL: i56_and_or:
139 ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
140 ; X86-NEXT: movl $384, %ecx # imm = 0x180
141 ; X86-NEXT: orl (%eax), %ecx
142 ; X86-NEXT: andl $-128, %ecx
143 ; X86-NEXT: movl %ecx, (%eax)
146 ; X64-LABEL: i56_and_or:
148 ; X64-NEXT: movzwl 4(%rdi), %eax
149 ; X64-NEXT: movzbl 6(%rdi), %ecx
150 ; X64-NEXT: movb %cl, 6(%rdi)
151 ; X64-NEXT: # kill: %ecx<def> %ecx<kill> %rcx<kill> %rcx<def>
152 ; X64-NEXT: shll $16, %ecx
153 ; X64-NEXT: orl %eax, %ecx
154 ; X64-NEXT: shlq $32, %rcx
155 ; X64-NEXT: movl (%rdi), %eax
156 ; X64-NEXT: orq %rcx, %rax
157 ; X64-NEXT: orq $384, %rax # imm = 0x180
158 ; X64-NEXT: movabsq $72057594037927808, %rcx # imm = 0xFFFFFFFFFFFF80
159 ; X64-NEXT: andq %rax, %rcx
160 ; X64-NEXT: movl %ecx, (%rdi)
161 ; X64-NEXT: shrq $32, %rcx
162 ; X64-NEXT: movw %cx, 4(%rdi)
164 %b = load i56, i56* %a, align 1
165 %c = and i56 %b, -128
167 store i56 %d, i56* %a, align 1
171 define void @i56_insert_bit(i56* %a, i1 zeroext %bit) {
172 ; X86-LABEL: i56_insert_bit:
174 ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
175 ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
176 ; X86-NEXT: shll $13, %ecx
177 ; X86-NEXT: movl $-8193, %edx # imm = 0xDFFF
178 ; X86-NEXT: andl (%eax), %edx
179 ; X86-NEXT: orl %ecx, %edx
180 ; X86-NEXT: movl %edx, (%eax)
183 ; X64-LABEL: i56_insert_bit:
185 ; X64-NEXT: movl %esi, %eax
186 ; X64-NEXT: movzwl 4(%rdi), %ecx
187 ; X64-NEXT: movzbl 6(%rdi), %edx
188 ; X64-NEXT: movb %dl, 6(%rdi)
189 ; X64-NEXT: # kill: %edx<def> %edx<kill> %rdx<kill> %rdx<def>
190 ; X64-NEXT: shll $16, %edx
191 ; X64-NEXT: orl %ecx, %edx
192 ; X64-NEXT: shlq $32, %rdx
193 ; X64-NEXT: movl (%rdi), %ecx
194 ; X64-NEXT: orq %rdx, %rcx
195 ; X64-NEXT: shlq $13, %rax
196 ; X64-NEXT: movabsq $72057594037919743, %rdx # imm = 0xFFFFFFFFFFDFFF
197 ; X64-NEXT: andq %rcx, %rdx
198 ; X64-NEXT: orq %rax, %rdx
199 ; X64-NEXT: movl %edx, (%rdi)
200 ; X64-NEXT: shrq $32, %rdx
201 ; X64-NEXT: movw %dx, 4(%rdi)
203 %extbit = zext i1 %bit to i56
204 %b = load i56, i56* %a, align 1
205 %extbit.shl = shl nuw nsw i56 %extbit, 13
206 %c = and i56 %b, -8193
207 %d = or i56 %c, %extbit.shl
208 store i56 %d, i56* %a, align 1