4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
23 * Copyright (c) 2009 Intel Corporation
24 * All Rights Reserved.
27 * Copyright 2009 Sun Microsystems, Inc. All rights reserved.
28 * Use is subject to license terms.
32 * Accelerated GHASH implementation with Intel PCLMULQDQ-NI
33 * instructions. This file contains an accelerated
34 * Galois Field Multiplication implementation.
36 * PCLMULQDQ is used to accelerate the most time-consuming part of GHASH,
37 * carry-less multiplication. More information about PCLMULQDQ can be
39 * http://software.intel.com/en-us/articles/
40 * carry-less-multiplication-and-its-usage-for-computing-the-gcm-mode/
45 * ====================================================================
46 * OpenSolaris OS modifications
48 * This source originates as file galois_hash_asm.c from
49 * Intel Corporation dated September 21, 2009.
51 * This OpenSolaris version has these major changes from the original source:
53 * 1. Added OpenSolaris ENTRY_NP/SET_SIZE macros from
54 * /usr/include/sys/asm_linkage.h, lint(1B) guards, and a dummy C function
55 * definition for lint.
57 * 2. Formatted code, added comments, and added #includes and #defines.
59 * 3. If bit CR0.TS is set, clear and set the TS bit, after and before
60 * calling kpreempt_disable() and kpreempt_enable().
61 * If the TS bit is not set, Save and restore %xmm registers at the beginning
62 * and end of function calls (%xmm* registers are not saved and restored by
63 * during kernel thread preemption).
65 * 4. Removed code to perform hashing. This is already done with C macro
66 * GHASH in gcm.c. For better performance, this removed code should be
67 * reintegrated in the future to replace the C GHASH macro.
69 * 5. Added code to byte swap 16-byte input and output.
71 * 6. Folded in comments from the original C source with embedded assembly
74 * 7. Renamed function and reordered parameters to match OpenSolaris:
76 * void galois_hash_asm(unsigned char *hk, unsigned char *s,
77 * unsigned char *d, int length)
78 * OpenSolaris OS interface:
79 * void gcm_mul_pclmulqdq(uint64_t *x_in, uint64_t *y, uint64_t *res);
80 * ====================================================================
84 #if defined(lint) || defined(__lint)
86 #include <sys/types.h>
90 gcm_mul_pclmulqdq
(uint64_t
*x_in
, uint64_t
*y
, uint64_t
*res
) {
95 #include <sys/asm_linkage.h>
96 #include <sys/controlregs.h>
98 #include <sys/machprivregs.h>
103 * Note: the CLTS macro clobbers P2 (%rsi) under i86xpv. That is,
104 * it calls HYPERVISOR_fpu_taskswitch() which modifies %rsi when it
105 * uses it to pass P2 to syscall.
106 * This also occurs with the STTS macro, but we don't care if
107 * P2 (%rsi) is modified just before function exit.
108 * The CLTS and STTS macros push and pop P1 (%rdi) already.
111 #define PROTECTED_CLTS \
116 #define PROTECTED_CLTS \
121 * If CR0_TS is not set, align stack (with push %rbp) and push
122 * %xmm0 - %xmm10 on stack, otherwise clear CR0_TS
124 #define CLEAR_TS_OR_PUSH_XMM_REGISTERS(tmpreg) \
128 testq $CR0_TS
, tmpreg; \
130 and $
-XMM_ALIGN
, %rsp; \
131 sub $
[XMM_SIZE
* 11], %rsp; \
132 movaps
%xmm0
, 160(%rsp
); \
133 movaps
%xmm1
, 144(%rsp
); \
134 movaps
%xmm2
, 128(%rsp
); \
135 movaps
%xmm3
, 112(%rsp
); \
136 movaps
%xmm4
, 96(%rsp
); \
137 movaps
%xmm5
, 80(%rsp
); \
138 movaps
%xmm6
, 64(%rsp
); \
139 movaps
%xmm7
, 48(%rsp
); \
140 movaps
%xmm8
, 32(%rsp
); \
141 movaps
%xmm9
, 16(%rsp
); \
142 movaps
%xmm10
, (%rsp
); \
150 * If CR0_TS was not set above, pop %xmm0 - %xmm10 off stack,
151 * otherwise set CR0_TS.
153 #define SET_TS_OR_POP_XMM_REGISTERS(tmpreg) \
154 testq $CR0_TS
, tmpreg; \
156 movaps
(%rsp
), %xmm10; \
157 movaps
16(%rsp
), %xmm9; \
158 movaps
32(%rsp
), %xmm8; \
159 movaps
48(%rsp
), %xmm7; \
160 movaps
64(%rsp
), %xmm6; \
161 movaps
80(%rsp
), %xmm5; \
162 movaps
96(%rsp
), %xmm4; \
163 movaps
112(%rsp
), %xmm3; \
164 movaps
128(%rsp
), %xmm2; \
165 movaps
144(%rsp
), %xmm1; \
166 movaps
160(%rsp
), %xmm0; \
176 #define PROTECTED_CLTS
177 #define CLEAR_TS_OR_PUSH_XMM_REGISTERS(tmpreg)
178 #define SET_TS_OR_POP_XMM_REGISTERS(tmpreg)
182 * Use this mask to byte-swap a 16-byte integer with the pshufb instruction
185 // static uint8_t byte_swap16_mask
[] = {
186 // 15, 14, 13, 12, 11, 10, 9, 8, 7, 6 ,5, 4, 3, 2, 1, 0 };
190 .byte 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0
195 * void gcm_mul_pclmulqdq(uint64_t *x_in, uint64_t *y, uint64_t *res);
197 * Perform a carry-less multiplication (that is, use XOR instead of the
198 * multiply operator) on P1 and P2 and place the result in P3.
200 * Byte swap the input and the output.
202 * Note: x_in, y, and res all point to a block of 20-byte numbers
203 * (an array of two 64-bit integers).
205 * Note2: For kernel code, caller is responsible for ensuring
206 * kpreempt_disable() has been called. This is because %xmm registers are
207 * not saved/restored. Clear and set the CR0.TS bit on entry and exit,
208 * respectively, if TS is set on entry. Otherwise, if TS is not set,
209 * save and restore %xmm registers on the stack.
211 * Note3: Original Intel definition:
212 * void galois_hash_asm(unsigned char *hk, unsigned char *s,
213 * unsigned char *d, int length)
215 * Note4: Register/parameter mapping:
217 * Parameter 1: %rcx (copied to %xmm0) hk or x_in
218 * Parameter 2: %rdx (copied to %xmm1) s or y
219 * Parameter 3: %rdi (result) d or res
221 * Parameter 1: %rdi (copied to %xmm0) x_in
222 * Parameter 2: %rsi (copied to %xmm1) y
223 * Parameter 3: %rdx (result) res
226 ENTRY_NP
(gcm_mul_pclmulqdq
)
227 CLEAR_TS_OR_PUSH_XMM_REGISTERS
(%r10)
232 movdqu
(%rdi
), %xmm0
// P1
233 movdqu
(%rsi
), %xmm1
// P2
236 // Byte swap
16-byte input
238 lea
.Lbyte_swap16_mask(%rip), %rax
239 movaps
(%rax
), %xmm10
245 // Multiply with the hash key
248 pclmulqdq $
0, %xmm1
, %xmm3
// xmm3 holds a0
*b0
251 pclmulqdq $
16, %xmm1
, %xmm4
// xmm4 holds a0
*b1
254 pclmulqdq $
1, %xmm1
, %xmm5
// xmm5 holds a1
*b0
256 pclmulqdq $
17, %xmm1
, %xmm6
// xmm6 holds a1
*b1
258 pxor
%xmm5
, %xmm4
// xmm4 holds a0
*b1
+ a1
*b0
260 movdqu
%xmm4
, %xmm5
// move the contents of xmm4 to xmm5
261 psrldq $
8, %xmm4
// shift by xmm4
64 bits to the right
262 pslldq $
8, %xmm5
// shift by xmm5
64 bits to the left
264 pxor
%xmm4
, %xmm6
// Register pair
<xmm6
:xmm3
> holds the result
265 // of the carry-less multiplication of
268 // We shift the result of the multiplication by one bit position
269 // to the left to cope for the fact that the bits are reversed.
285 // First phase of the reduction
287 // Move xmm3 into xmm7
, xmm8
, xmm9 in order to perform the shifts
292 pslld $
31, %xmm7
// packed right shift shifting
<< 31
293 pslld $
30, %xmm8
// packed right shift shifting
<< 30
294 pslld $
25, %xmm9
// packed right shift shifting
<< 25
295 pxor
%xmm8
, %xmm7
// xor the shifted versions
300 pxor
%xmm7
, %xmm3
// first phase of the reduction complete
303 // Second phase of the reduction
305 // Make
3 copies of xmm3 in xmm2
, xmm4
, xmm5 for doing these
308 movdqu
%xmm3
, %xmm4
// packed left shifting
>> 1
311 psrld $
2, %xmm4
// packed left shifting
>> 2
312 psrld $
7, %xmm5
// packed left shifting
>> 7
313 pxor
%xmm4
, %xmm2
// xor the shifted versions
317 pxor
%xmm3
, %xmm6
// the result is in xmm6
320 // Byte swap
16-byte result
322 pshufb
%xmm10
, %xmm6
// %xmm10 has the swap mask
327 movdqu
%xmm6
, (%rdx
) // P3
331 // Cleanup
and Return
333 SET_TS_OR_POP_XMM_REGISTERS
(%r10)
335 SET_SIZE
(gcm_mul_pclmulqdq
)
337 #endif /* lint || __lint */