1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright 2016 Broadcom
7 * This file works with the SPU2 version of the SPU. SPU2 has different message
8 * formats than the previous version of the SPU. All SPU message format
9 * differences should be hidden in the spux.c,h files.
12 #include <linux/kernel.h>
13 #include <linux/string.h>
19 #define SPU2_TX_STATUS_LEN 0 /* SPU2 has no STATUS in input packet */
22 * Controlled by pkt_stat_cnt field in CRYPTO_SS_SPU0_CORE_SPU2_CONTROL0
23 * register. Defaults to 2.
25 #define SPU2_RX_STATUS_LEN 2
29 SPU2_MACSEC_SECTAG8_ECB
= 1,
30 SPU2_MACSEC_SECTAG8_SCB
= 2,
31 SPU2_MACSEC_SECTAG16
= 3,
32 SPU2_MACSEC_SECTAG16_8_XPN
= 4,
41 static char *spu2_cipher_type_names
[] = { "None", "AES128", "AES192", "AES256",
45 static char *spu2_cipher_mode_names
[] = { "ECB", "CBC", "CTR", "CFB", "OFB",
49 static char *spu2_hash_type_names
[] = { "None", "AES128", "AES192", "AES256",
50 "Reserved", "Reserved", "MD5", "SHA1", "SHA224", "SHA256", "SHA384",
51 "SHA512", "SHA512/224", "SHA512/256", "SHA3-224", "SHA3-256",
52 "SHA3-384", "SHA3-512"
55 static char *spu2_hash_mode_names
[] = { "CMAC", "CBC-MAC", "XCBC-MAC", "HMAC",
56 "Rabin", "CCM", "GCM", "Reserved"
59 static char *spu2_ciph_type_name(enum spu2_cipher_type cipher_type
)
61 if (cipher_type
>= SPU2_CIPHER_TYPE_LAST
)
63 return spu2_cipher_type_names
[cipher_type
];
66 static char *spu2_ciph_mode_name(enum spu2_cipher_mode cipher_mode
)
68 if (cipher_mode
>= SPU2_CIPHER_MODE_LAST
)
70 return spu2_cipher_mode_names
[cipher_mode
];
73 static char *spu2_hash_type_name(enum spu2_hash_type hash_type
)
75 if (hash_type
>= SPU2_HASH_TYPE_LAST
)
77 return spu2_hash_type_names
[hash_type
];
80 static char *spu2_hash_mode_name(enum spu2_hash_mode hash_mode
)
82 if (hash_mode
>= SPU2_HASH_MODE_LAST
)
84 return spu2_hash_mode_names
[hash_mode
];
88 * Convert from a software cipher mode value to the corresponding value
91 static int spu2_cipher_mode_xlate(enum spu_cipher_mode cipher_mode
,
92 enum spu2_cipher_mode
*spu2_mode
)
94 switch (cipher_mode
) {
96 *spu2_mode
= SPU2_CIPHER_MODE_ECB
;
99 *spu2_mode
= SPU2_CIPHER_MODE_CBC
;
101 case CIPHER_MODE_OFB
:
102 *spu2_mode
= SPU2_CIPHER_MODE_OFB
;
104 case CIPHER_MODE_CFB
:
105 *spu2_mode
= SPU2_CIPHER_MODE_CFB
;
107 case CIPHER_MODE_CTR
:
108 *spu2_mode
= SPU2_CIPHER_MODE_CTR
;
110 case CIPHER_MODE_CCM
:
111 *spu2_mode
= SPU2_CIPHER_MODE_CCM
;
113 case CIPHER_MODE_GCM
:
114 *spu2_mode
= SPU2_CIPHER_MODE_GCM
;
116 case CIPHER_MODE_XTS
:
117 *spu2_mode
= SPU2_CIPHER_MODE_XTS
;
126 * spu2_cipher_xlate() - Convert a cipher {alg/mode/type} triple to a SPU2
127 * cipher type and mode.
128 * @cipher_alg: [in] cipher algorithm value from software enumeration
129 * @cipher_mode: [in] cipher mode value from software enumeration
130 * @cipher_type: [in] cipher type value from software enumeration
131 * @spu2_type: [out] cipher type value used by spu2 hardware
132 * @spu2_mode: [out] cipher mode value used by spu2 hardware
134 * Return: 0 if successful
136 static int spu2_cipher_xlate(enum spu_cipher_alg cipher_alg
,
137 enum spu_cipher_mode cipher_mode
,
138 enum spu_cipher_type cipher_type
,
139 enum spu2_cipher_type
*spu2_type
,
140 enum spu2_cipher_mode
*spu2_mode
)
144 err
= spu2_cipher_mode_xlate(cipher_mode
, spu2_mode
);
146 flow_log("Invalid cipher mode %d\n", cipher_mode
);
150 switch (cipher_alg
) {
151 case CIPHER_ALG_NONE
:
152 *spu2_type
= SPU2_CIPHER_TYPE_NONE
;
155 /* SPU2 does not support RC4 */
157 *spu2_type
= SPU2_CIPHER_TYPE_NONE
;
160 *spu2_type
= SPU2_CIPHER_TYPE_DES
;
162 case CIPHER_ALG_3DES
:
163 *spu2_type
= SPU2_CIPHER_TYPE_3DES
;
166 switch (cipher_type
) {
167 case CIPHER_TYPE_AES128
:
168 *spu2_type
= SPU2_CIPHER_TYPE_AES128
;
170 case CIPHER_TYPE_AES192
:
171 *spu2_type
= SPU2_CIPHER_TYPE_AES192
;
173 case CIPHER_TYPE_AES256
:
174 *spu2_type
= SPU2_CIPHER_TYPE_AES256
;
180 case CIPHER_ALG_LAST
:
187 flow_log("Invalid cipher alg %d or type %d\n",
188 cipher_alg
, cipher_type
);
193 * Convert from a software hash mode value to the corresponding value
194 * for SPU2. Note that HASH_MODE_NONE and HASH_MODE_XCBC have the same value.
196 static int spu2_hash_mode_xlate(enum hash_mode hash_mode
,
197 enum spu2_hash_mode
*spu2_mode
)
201 *spu2_mode
= SPU2_HASH_MODE_XCBC_MAC
;
204 *spu2_mode
= SPU2_HASH_MODE_CMAC
;
207 *spu2_mode
= SPU2_HASH_MODE_HMAC
;
210 *spu2_mode
= SPU2_HASH_MODE_CCM
;
213 *spu2_mode
= SPU2_HASH_MODE_GCM
;
222 * spu2_hash_xlate() - Convert a hash {alg/mode/type} triple to a SPU2 hash type
224 * @hash_alg: [in] hash algorithm value from software enumeration
225 * @hash_mode: [in] hash mode value from software enumeration
226 * @hash_type: [in] hash type value from software enumeration
227 * @ciph_type: [in] cipher type value from software enumeration
228 * @spu2_type: [out] hash type value used by SPU2 hardware
229 * @spu2_mode: [out] hash mode value used by SPU2 hardware
231 * Return: 0 if successful
234 spu2_hash_xlate(enum hash_alg hash_alg
, enum hash_mode hash_mode
,
235 enum hash_type hash_type
, enum spu_cipher_type ciph_type
,
236 enum spu2_hash_type
*spu2_type
, enum spu2_hash_mode
*spu2_mode
)
240 err
= spu2_hash_mode_xlate(hash_mode
, spu2_mode
);
242 flow_log("Invalid hash mode %d\n", hash_mode
);
248 *spu2_type
= SPU2_HASH_TYPE_NONE
;
251 *spu2_type
= SPU2_HASH_TYPE_MD5
;
254 *spu2_type
= SPU2_HASH_TYPE_SHA1
;
256 case HASH_ALG_SHA224
:
257 *spu2_type
= SPU2_HASH_TYPE_SHA224
;
259 case HASH_ALG_SHA256
:
260 *spu2_type
= SPU2_HASH_TYPE_SHA256
;
262 case HASH_ALG_SHA384
:
263 *spu2_type
= SPU2_HASH_TYPE_SHA384
;
265 case HASH_ALG_SHA512
:
266 *spu2_type
= SPU2_HASH_TYPE_SHA512
;
270 case CIPHER_TYPE_AES128
:
271 *spu2_type
= SPU2_HASH_TYPE_AES128
;
273 case CIPHER_TYPE_AES192
:
274 *spu2_type
= SPU2_HASH_TYPE_AES192
;
276 case CIPHER_TYPE_AES256
:
277 *spu2_type
= SPU2_HASH_TYPE_AES256
;
283 case HASH_ALG_SHA3_224
:
284 *spu2_type
= SPU2_HASH_TYPE_SHA3_224
;
286 case HASH_ALG_SHA3_256
:
287 *spu2_type
= SPU2_HASH_TYPE_SHA3_256
;
289 case HASH_ALG_SHA3_384
:
290 *spu2_type
= SPU2_HASH_TYPE_SHA3_384
;
292 case HASH_ALG_SHA3_512
:
293 *spu2_type
= SPU2_HASH_TYPE_SHA3_512
;
302 flow_log("Invalid hash alg %d or type %d\n",
303 hash_alg
, hash_type
);
307 /* Dump FMD ctrl0. The ctrl0 input is in host byte order */
308 static void spu2_dump_fmd_ctrl0(u64 ctrl0
)
310 enum spu2_cipher_type ciph_type
;
311 enum spu2_cipher_mode ciph_mode
;
312 enum spu2_hash_type hash_type
;
313 enum spu2_hash_mode hash_mode
;
315 char *ciph_mode_name
;
317 char *hash_mode_name
;
321 packet_log(" FMD CTRL0 %#16llx\n", ctrl0
);
322 if (ctrl0
& SPU2_CIPH_ENCRYPT_EN
)
323 packet_log(" encrypt\n");
325 packet_log(" decrypt\n");
327 ciph_type
= (ctrl0
& SPU2_CIPH_TYPE
) >> SPU2_CIPH_TYPE_SHIFT
;
328 ciph_name
= spu2_ciph_type_name(ciph_type
);
329 packet_log(" Cipher type: %s\n", ciph_name
);
331 if (ciph_type
!= SPU2_CIPHER_TYPE_NONE
) {
332 ciph_mode
= (ctrl0
& SPU2_CIPH_MODE
) >> SPU2_CIPH_MODE_SHIFT
;
333 ciph_mode_name
= spu2_ciph_mode_name(ciph_mode
);
334 packet_log(" Cipher mode: %s\n", ciph_mode_name
);
337 cfb
= (ctrl0
& SPU2_CFB_MASK
) >> SPU2_CFB_MASK_SHIFT
;
338 packet_log(" CFB %#x\n", cfb
);
340 proto
= (ctrl0
& SPU2_PROTO_SEL
) >> SPU2_PROTO_SEL_SHIFT
;
341 packet_log(" protocol %#x\n", proto
);
343 if (ctrl0
& SPU2_HASH_FIRST
)
344 packet_log(" hash first\n");
346 packet_log(" cipher first\n");
348 if (ctrl0
& SPU2_CHK_TAG
)
349 packet_log(" check tag\n");
351 hash_type
= (ctrl0
& SPU2_HASH_TYPE
) >> SPU2_HASH_TYPE_SHIFT
;
352 hash_name
= spu2_hash_type_name(hash_type
);
353 packet_log(" Hash type: %s\n", hash_name
);
355 if (hash_type
!= SPU2_HASH_TYPE_NONE
) {
356 hash_mode
= (ctrl0
& SPU2_HASH_MODE
) >> SPU2_HASH_MODE_SHIFT
;
357 hash_mode_name
= spu2_hash_mode_name(hash_mode
);
358 packet_log(" Hash mode: %s\n", hash_mode_name
);
361 if (ctrl0
& SPU2_CIPH_PAD_EN
) {
362 packet_log(" Cipher pad: %#2llx\n",
363 (ctrl0
& SPU2_CIPH_PAD
) >> SPU2_CIPH_PAD_SHIFT
);
367 /* Dump FMD ctrl1. The ctrl1 input is in host byte order */
368 static void spu2_dump_fmd_ctrl1(u64 ctrl1
)
378 packet_log(" FMD CTRL1 %#16llx\n", ctrl1
);
379 if (ctrl1
& SPU2_TAG_LOC
)
380 packet_log(" Tag after payload\n");
382 packet_log(" Msg includes ");
383 if (ctrl1
& SPU2_HAS_FR_DATA
)
385 if (ctrl1
& SPU2_HAS_AAD1
)
387 if (ctrl1
& SPU2_HAS_NAAD
)
389 if (ctrl1
& SPU2_HAS_AAD2
)
391 if (ctrl1
& SPU2_HAS_ESN
)
395 hash_key_len
= (ctrl1
& SPU2_HASH_KEY_LEN
) >> SPU2_HASH_KEY_LEN_SHIFT
;
396 packet_log(" Hash key len %u\n", hash_key_len
);
398 ciph_key_len
= (ctrl1
& SPU2_CIPH_KEY_LEN
) >> SPU2_CIPH_KEY_LEN_SHIFT
;
399 packet_log(" Cipher key len %u\n", ciph_key_len
);
401 if (ctrl1
& SPU2_GENIV
)
402 packet_log(" Generate IV\n");
404 if (ctrl1
& SPU2_HASH_IV
)
405 packet_log(" IV included in hash\n");
407 if (ctrl1
& SPU2_RET_IV
)
408 packet_log(" Return IV in output before payload\n");
410 ret_iv_len
= (ctrl1
& SPU2_RET_IV_LEN
) >> SPU2_RET_IV_LEN_SHIFT
;
411 packet_log(" Length of returned IV %u bytes\n",
412 ret_iv_len
? ret_iv_len
: 16);
414 iv_offset
= (ctrl1
& SPU2_IV_OFFSET
) >> SPU2_IV_OFFSET_SHIFT
;
415 packet_log(" IV offset %u\n", iv_offset
);
417 iv_len
= (ctrl1
& SPU2_IV_LEN
) >> SPU2_IV_LEN_SHIFT
;
418 packet_log(" Input IV len %u bytes\n", iv_len
);
420 hash_tag_len
= (ctrl1
& SPU2_HASH_TAG_LEN
) >> SPU2_HASH_TAG_LEN_SHIFT
;
421 packet_log(" Hash tag length %u bytes\n", hash_tag_len
);
423 packet_log(" Return ");
424 ret_md
= (ctrl1
& SPU2_RETURN_MD
) >> SPU2_RETURN_MD_SHIFT
;
427 if (ret_md
== SPU2_RET_FMD_OMD
)
429 else if (ret_md
== SPU2_RET_FMD_OMD_IV
)
430 packet_log("OMD IV ");
431 if (ctrl1
& SPU2_RETURN_FD
)
433 if (ctrl1
& SPU2_RETURN_AAD1
)
435 if (ctrl1
& SPU2_RETURN_NAAD
)
437 if (ctrl1
& SPU2_RETURN_AAD2
)
439 if (ctrl1
& SPU2_RETURN_PAY
)
440 packet_log("Payload");
444 /* Dump FMD ctrl2. The ctrl2 input is in host byte order */
445 static void spu2_dump_fmd_ctrl2(u64 ctrl2
)
447 packet_log(" FMD CTRL2 %#16llx\n", ctrl2
);
449 packet_log(" AAD1 offset %llu length %llu bytes\n",
450 ctrl2
& SPU2_AAD1_OFFSET
,
451 (ctrl2
& SPU2_AAD1_LEN
) >> SPU2_AAD1_LEN_SHIFT
);
452 packet_log(" AAD2 offset %llu\n",
453 (ctrl2
& SPU2_AAD2_OFFSET
) >> SPU2_AAD2_OFFSET_SHIFT
);
454 packet_log(" Payload offset %llu\n",
455 (ctrl2
& SPU2_PL_OFFSET
) >> SPU2_PL_OFFSET_SHIFT
);
458 /* Dump FMD ctrl3. The ctrl3 input is in host byte order */
459 static void spu2_dump_fmd_ctrl3(u64 ctrl3
)
461 packet_log(" FMD CTRL3 %#16llx\n", ctrl3
);
463 packet_log(" Payload length %llu bytes\n", ctrl3
& SPU2_PL_LEN
);
464 packet_log(" TLS length %llu bytes\n",
465 (ctrl3
& SPU2_TLS_LEN
) >> SPU2_TLS_LEN_SHIFT
);
468 static void spu2_dump_fmd(struct SPU2_FMD
*fmd
)
470 spu2_dump_fmd_ctrl0(le64_to_cpu(fmd
->ctrl0
));
471 spu2_dump_fmd_ctrl1(le64_to_cpu(fmd
->ctrl1
));
472 spu2_dump_fmd_ctrl2(le64_to_cpu(fmd
->ctrl2
));
473 spu2_dump_fmd_ctrl3(le64_to_cpu(fmd
->ctrl3
));
476 static void spu2_dump_omd(u8
*omd
, u16 hash_key_len
, u16 ciph_key_len
,
477 u16 hash_iv_len
, u16 ciph_iv_len
)
481 packet_log(" OMD:\n");
484 packet_log(" Hash Key Length %u bytes\n", hash_key_len
);
485 packet_dump(" KEY: ", ptr
, hash_key_len
);
490 packet_log(" Cipher Key Length %u bytes\n", ciph_key_len
);
491 packet_dump(" KEY: ", ptr
, ciph_key_len
);
496 packet_log(" Hash IV Length %u bytes\n", hash_iv_len
);
497 packet_dump(" hash IV: ", ptr
, hash_iv_len
);
502 packet_log(" Cipher IV Length %u bytes\n", ciph_iv_len
);
503 packet_dump(" cipher IV: ", ptr
, ciph_iv_len
);
507 /* Dump a SPU2 header for debug */
508 void spu2_dump_msg_hdr(u8
*buf
, unsigned int buf_len
)
510 struct SPU2_FMD
*fmd
= (struct SPU2_FMD
*)buf
;
520 packet_log("SPU2 message header %p len: %u\n", buf
, buf_len
);
523 omd
= (u8
*)(fmd
+ 1);
525 ctrl1
= le64_to_cpu(fmd
->ctrl1
);
526 hash_key_len
= (ctrl1
& SPU2_HASH_KEY_LEN
) >> SPU2_HASH_KEY_LEN_SHIFT
;
527 ciph_key_len
= (ctrl1
& SPU2_CIPH_KEY_LEN
) >> SPU2_CIPH_KEY_LEN_SHIFT
;
529 ciph_iv_len
= (ctrl1
& SPU2_IV_LEN
) >> SPU2_IV_LEN_SHIFT
;
530 spu2_dump_omd(omd
, hash_key_len
, ciph_key_len
, hash_iv_len
,
533 /* Double check sanity */
534 omd_len
= hash_key_len
+ ciph_key_len
+ hash_iv_len
+ ciph_iv_len
;
535 if (FMD_SIZE
+ omd_len
!= buf_len
) {
537 (" Packet parsed incorrectly. buf_len %u, sum of MD %zu\n",
538 buf_len
, FMD_SIZE
+ omd_len
);
544 * spu2_fmd_init() - At setkey time, initialize the fixed meta data for
545 * subsequent skcipher requests for this context.
546 * @fmd: Start of FMD field to be written
547 * @spu2_type: Cipher algorithm
548 * @spu2_mode: Cipher mode
549 * @cipher_key_len: Length of cipher key, in bytes
550 * @cipher_iv_len: Length of cipher initialization vector, in bytes
552 * Return: 0 (success)
554 static int spu2_fmd_init(struct SPU2_FMD
*fmd
,
555 enum spu2_cipher_type spu2_type
,
556 enum spu2_cipher_mode spu2_mode
,
557 u32 cipher_key_len
, u32 cipher_iv_len
)
568 ctrl0
= (spu2_type
<< SPU2_CIPH_TYPE_SHIFT
) |
569 (spu2_mode
<< SPU2_CIPH_MODE_SHIFT
);
571 ctrl1
= (cipher_key_len
<< SPU2_CIPH_KEY_LEN_SHIFT
) |
572 ((u64
)cipher_iv_len
<< SPU2_IV_LEN_SHIFT
) |
573 ((u64
)SPU2_RET_FMD_ONLY
<< SPU2_RETURN_MD_SHIFT
) | SPU2_RETURN_PAY
;
576 * AAD1 offset is from start of FD. FD length is always 0 for this
577 * driver. So AAD1_offset is always 0.
580 aad2_offset
= aad1_offset
;
582 ctrl2
= aad1_offset
|
583 (aad1_len
<< SPU2_AAD1_LEN_SHIFT
) |
584 (aad2_offset
<< SPU2_AAD2_OFFSET_SHIFT
) |
585 (payload_offset
<< SPU2_PL_OFFSET_SHIFT
);
589 fmd
->ctrl0
= cpu_to_le64(ctrl0
);
590 fmd
->ctrl1
= cpu_to_le64(ctrl1
);
591 fmd
->ctrl2
= cpu_to_le64(ctrl2
);
592 fmd
->ctrl3
= cpu_to_le64(ctrl3
);
598 * spu2_fmd_ctrl0_write() - Write ctrl0 field in fixed metadata (FMD) field of
599 * SPU request packet.
600 * @fmd: Start of FMD field to be written
601 * @is_inbound: true if decrypting. false if encrypting.
602 * @auth_first: true if alg authenticates before encrypting
603 * @protocol: protocol selector
604 * @cipher_type: cipher algorithm
605 * @cipher_mode: cipher mode
606 * @auth_type: authentication type
607 * @auth_mode: authentication mode
609 static void spu2_fmd_ctrl0_write(struct SPU2_FMD
*fmd
,
610 bool is_inbound
, bool auth_first
,
611 enum spu2_proto_sel protocol
,
612 enum spu2_cipher_type cipher_type
,
613 enum spu2_cipher_mode cipher_mode
,
614 enum spu2_hash_type auth_type
,
615 enum spu2_hash_mode auth_mode
)
619 if ((cipher_type
!= SPU2_CIPHER_TYPE_NONE
) && !is_inbound
)
620 ctrl0
|= SPU2_CIPH_ENCRYPT_EN
;
622 ctrl0
|= ((u64
)cipher_type
<< SPU2_CIPH_TYPE_SHIFT
) |
623 ((u64
)cipher_mode
<< SPU2_CIPH_MODE_SHIFT
);
626 ctrl0
|= (u64
)protocol
<< SPU2_PROTO_SEL_SHIFT
;
629 ctrl0
|= SPU2_HASH_FIRST
;
631 if (is_inbound
&& (auth_type
!= SPU2_HASH_TYPE_NONE
))
632 ctrl0
|= SPU2_CHK_TAG
;
634 ctrl0
|= (((u64
)auth_type
<< SPU2_HASH_TYPE_SHIFT
) |
635 ((u64
)auth_mode
<< SPU2_HASH_MODE_SHIFT
));
637 fmd
->ctrl0
= cpu_to_le64(ctrl0
);
641 * spu2_fmd_ctrl1_write() - Write ctrl1 field in fixed metadata (FMD) field of
642 * SPU request packet.
643 * @fmd: Start of FMD field to be written
644 * @is_inbound: true if decrypting. false if encrypting.
645 * @assoc_size: Length of additional associated data, in bytes
646 * @auth_key_len: Length of authentication key, in bytes
647 * @cipher_key_len: Length of cipher key, in bytes
648 * @gen_iv: If true, hw generates IV and returns in response
649 * @hash_iv: IV participates in hash. Used for IPSEC and TLS.
650 * @return_iv: Return IV in output packet before payload
651 * @ret_iv_len: Length of IV returned from SPU, in bytes
652 * @ret_iv_offset: Offset into full IV of start of returned IV
653 * @cipher_iv_len: Length of input cipher IV, in bytes
654 * @digest_size: Length of digest (aka, hash tag or ICV), in bytes
655 * @return_payload: Return payload in SPU response
656 * @return_md : return metadata in SPU response
658 * Packet can have AAD2 w/o AAD1. For algorithms currently supported,
659 * associated data goes in AAD2.
661 static void spu2_fmd_ctrl1_write(struct SPU2_FMD
*fmd
, bool is_inbound
,
663 u64 auth_key_len
, u64 cipher_key_len
,
664 bool gen_iv
, bool hash_iv
, bool return_iv
,
665 u64 ret_iv_len
, u64 ret_iv_offset
,
666 u64 cipher_iv_len
, u64 digest_size
,
667 bool return_payload
, bool return_md
)
671 if (is_inbound
&& digest_size
)
672 ctrl1
|= SPU2_TAG_LOC
;
675 ctrl1
|= SPU2_HAS_AAD2
;
676 ctrl1
|= SPU2_RETURN_AAD2
; /* need aad2 for gcm aes esp */
680 ctrl1
|= ((auth_key_len
<< SPU2_HASH_KEY_LEN_SHIFT
) &
684 ctrl1
|= ((cipher_key_len
<< SPU2_CIPH_KEY_LEN_SHIFT
) &
691 ctrl1
|= SPU2_HASH_IV
;
694 ctrl1
|= SPU2_RET_IV
;
695 ctrl1
|= ret_iv_len
<< SPU2_RET_IV_LEN_SHIFT
;
696 ctrl1
|= ret_iv_offset
<< SPU2_IV_OFFSET_SHIFT
;
699 ctrl1
|= ((cipher_iv_len
<< SPU2_IV_LEN_SHIFT
) & SPU2_IV_LEN
);
702 ctrl1
|= ((digest_size
<< SPU2_HASH_TAG_LEN_SHIFT
) &
705 /* Let's ask for the output pkt to include FMD, but don't need to
706 * get keys and IVs back in OMD.
709 ctrl1
|= ((u64
)SPU2_RET_FMD_ONLY
<< SPU2_RETURN_MD_SHIFT
);
711 ctrl1
|= ((u64
)SPU2_RET_NO_MD
<< SPU2_RETURN_MD_SHIFT
);
713 /* Crypto API does not get assoc data back. So no need for AAD2. */
716 ctrl1
|= SPU2_RETURN_PAY
;
718 fmd
->ctrl1
= cpu_to_le64(ctrl1
);
722 * spu2_fmd_ctrl2_write() - Set the ctrl2 field in the fixed metadata field of
724 * @fmd: Start of FMD field to be written
725 * @cipher_offset: Number of bytes from Start of Packet (end of FD field) where
726 * data to be encrypted or decrypted begins
727 * @auth_key_len: Length of authentication key, in bytes
728 * @auth_iv_len: Length of authentication initialization vector, in bytes
729 * @cipher_key_len: Length of cipher key, in bytes
730 * @cipher_iv_len: Length of cipher IV, in bytes
732 static void spu2_fmd_ctrl2_write(struct SPU2_FMD
*fmd
, u64 cipher_offset
,
733 u64 auth_key_len
, u64 auth_iv_len
,
734 u64 cipher_key_len
, u64 cipher_iv_len
)
742 /* AAD1 offset is from start of FD. FD length always 0. */
745 aad2_offset
= aad1_offset
;
746 payload_offset
= cipher_offset
;
747 ctrl2
= aad1_offset
|
748 (aad1_len
<< SPU2_AAD1_LEN_SHIFT
) |
749 (aad2_offset
<< SPU2_AAD2_OFFSET_SHIFT
) |
750 (payload_offset
<< SPU2_PL_OFFSET_SHIFT
);
752 fmd
->ctrl2
= cpu_to_le64(ctrl2
);
756 * spu2_fmd_ctrl3_write() - Set the ctrl3 field in FMD
757 * @fmd: Fixed meta data. First field in SPU2 msg header.
758 * @payload_len: Length of payload, in bytes
760 static void spu2_fmd_ctrl3_write(struct SPU2_FMD
*fmd
, u64 payload_len
)
764 ctrl3
= payload_len
& SPU2_PL_LEN
;
766 fmd
->ctrl3
= cpu_to_le64(ctrl3
);
770 * spu2_ctx_max_payload() - Determine the maximum length of the payload for a
771 * SPU message for a given cipher and hash alg context.
772 * @cipher_alg: The cipher algorithm
773 * @cipher_mode: The cipher mode
774 * @blocksize: The size of a block of data for this algo
776 * For SPU2, the hardware generally ignores the PayloadLen field in ctrl3 of
777 * FMD and just keeps computing until it receives a DMA descriptor with the EOF
778 * flag set. So we consider the max payload to be infinite. AES CCM is an
781 * Return: Max payload length in bytes
783 u32
spu2_ctx_max_payload(enum spu_cipher_alg cipher_alg
,
784 enum spu_cipher_mode cipher_mode
,
785 unsigned int blocksize
)
787 if ((cipher_alg
== CIPHER_ALG_AES
) &&
788 (cipher_mode
== CIPHER_MODE_CCM
)) {
789 u32 excess
= SPU2_MAX_PAYLOAD
% blocksize
;
791 return SPU2_MAX_PAYLOAD
- excess
;
793 return SPU_MAX_PAYLOAD_INF
;
798 * spu2_payload_length() - Given a SPU2 message header, extract the payload
800 * @spu_hdr: Start of SPU message header (FMD)
802 * Return: payload length, in bytes
804 u32
spu2_payload_length(u8
*spu_hdr
)
806 struct SPU2_FMD
*fmd
= (struct SPU2_FMD
*)spu_hdr
;
810 ctrl3
= le64_to_cpu(fmd
->ctrl3
);
811 pl_len
= ctrl3
& SPU2_PL_LEN
;
817 * spu2_response_hdr_len() - Determine the expected length of a SPU response
819 * @auth_key_len: Length of authentication key, in bytes
820 * @enc_key_len: Length of encryption key, in bytes
823 * For SPU2, includes just FMD. OMD is never requested.
825 * Return: Length of FMD, in bytes
827 u16
spu2_response_hdr_len(u16 auth_key_len
, u16 enc_key_len
, bool is_hash
)
833 * spu2_hash_pad_len() - Calculate the length of hash padding required to extend
834 * data to a full block size.
835 * @hash_alg: hash algorithm
836 * @hash_mode: hash mode
837 * @chunksize: length of data, in bytes
838 * @hash_block_size: size of a hash block, in bytes
840 * SPU2 hardware does all hash padding
842 * Return: length of hash pad in bytes
844 u16
spu2_hash_pad_len(enum hash_alg hash_alg
, enum hash_mode hash_mode
,
845 u32 chunksize
, u16 hash_block_size
)
851 * spu2_gcm_ccm_pad_len() - Determine the length of GCM/CCM padding for either
852 * the AAD field or the data.
853 * @cipher_mode: Unused
856 * Return: 0. Unlike SPU-M, SPU2 hardware does any GCM/CCM padding required.
858 u32
spu2_gcm_ccm_pad_len(enum spu_cipher_mode cipher_mode
,
859 unsigned int data_size
)
865 * spu2_assoc_resp_len() - Determine the size of the AAD2 buffer needed to catch
866 * associated data in a SPU2 output packet.
867 * @cipher_mode: cipher mode
868 * @assoc_len: length of additional associated data, in bytes
869 * @iv_len: length of initialization vector, in bytes
870 * @is_encrypt: true if encrypting. false if decrypt.
872 * Return: Length of buffer to catch associated data in response
874 u32
spu2_assoc_resp_len(enum spu_cipher_mode cipher_mode
,
875 unsigned int assoc_len
, unsigned int iv_len
,
878 u32 resp_len
= assoc_len
;
881 /* gcm aes esp has to write 8-byte IV in response */
887 * spu2_aead_ivlen() - Calculate the length of the AEAD IV to be included
888 * in a SPU request after the AAD and before the payload.
889 * @cipher_mode: cipher mode
890 * @iv_len: initialization vector length in bytes
892 * For SPU2, AEAD IV is included in OMD and does not need to be repeated
893 * prior to the payload.
895 * Return: Length of AEAD IV in bytes
897 u8
spu2_aead_ivlen(enum spu_cipher_mode cipher_mode
, u16 iv_len
)
903 * spu2_hash_type() - Determine the type of hash operation.
904 * @src_sent: The number of bytes in the current request that have already
905 * been sent to the SPU to be hashed.
907 * SPU2 always does a FULL hash operation
909 enum hash_type
spu2_hash_type(u32 src_sent
)
911 return HASH_TYPE_FULL
;
915 * spu2_digest_size() - Determine the size of a hash digest to expect the SPU to
917 * @alg_digest_size: Number of bytes in the final digest for the given algo
918 * @alg: The hash algorithm
919 * @htype: Type of hash operation (init, update, full, etc)
922 u32
spu2_digest_size(u32 alg_digest_size
, enum hash_alg alg
,
923 enum hash_type htype
)
925 return alg_digest_size
;
929 * spu2_create_request() - Build a SPU2 request message header, includint FMD and
931 * @spu_hdr: Start of buffer where SPU request header is to be written
932 * @req_opts: SPU request message options
933 * @cipher_parms: Parameters related to cipher algorithm
934 * @hash_parms: Parameters related to hash algorithm
935 * @aead_parms: Parameters related to AEAD operation
936 * @data_size: Length of data to be encrypted or authenticated. If AEAD, does
937 * not include length of AAD.
939 * Construct the message starting at spu_hdr. Caller should allocate this buffer
940 * in DMA-able memory at least SPU_HEADER_ALLOC_LEN bytes long.
942 * Return: the length of the SPU header in bytes. 0 if an error occurs.
944 u32
spu2_create_request(u8
*spu_hdr
,
945 struct spu_request_opts
*req_opts
,
946 struct spu_cipher_parms
*cipher_parms
,
947 struct spu_hash_parms
*hash_parms
,
948 struct spu_aead_parms
*aead_parms
,
949 unsigned int data_size
)
951 struct SPU2_FMD
*fmd
;
953 unsigned int buf_len
;
955 enum spu2_cipher_type spu2_ciph_type
= SPU2_CIPHER_TYPE_NONE
;
956 enum spu2_cipher_mode spu2_ciph_mode
;
957 enum spu2_hash_type spu2_auth_type
= SPU2_HASH_TYPE_NONE
;
958 enum spu2_hash_mode spu2_auth_mode
;
959 bool return_md
= true;
960 enum spu2_proto_sel proto
= SPU2_PROTO_RESV
;
962 /* size of the payload */
963 unsigned int payload_len
=
964 hash_parms
->prebuf_len
+ data_size
+ hash_parms
->pad_len
-
965 ((req_opts
->is_aead
&& req_opts
->is_inbound
) ?
966 hash_parms
->digestsize
: 0);
968 /* offset of prebuf or data from start of AAD2 */
969 unsigned int cipher_offset
= aead_parms
->assoc_size
+
970 aead_parms
->aad_pad_len
+ aead_parms
->iv_len
;
972 /* total size of the data following OMD (without STAT word padding) */
973 unsigned int real_db_size
= spu_real_db_size(aead_parms
->assoc_size
,
975 hash_parms
->prebuf_len
,
977 aead_parms
->aad_pad_len
,
978 aead_parms
->data_pad_len
,
979 hash_parms
->pad_len
);
980 unsigned int assoc_size
= aead_parms
->assoc_size
;
982 if (req_opts
->is_aead
&&
983 (cipher_parms
->alg
== CIPHER_ALG_AES
) &&
984 (cipher_parms
->mode
== CIPHER_MODE_GCM
))
986 * On SPU 2, aes gcm cipher first on encrypt, auth first on
989 req_opts
->auth_first
= req_opts
->is_inbound
;
991 /* and do opposite for ccm (auth 1st on encrypt) */
992 if (req_opts
->is_aead
&&
993 (cipher_parms
->alg
== CIPHER_ALG_AES
) &&
994 (cipher_parms
->mode
== CIPHER_MODE_CCM
))
995 req_opts
->auth_first
= !req_opts
->is_inbound
;
997 flow_log("%s()\n", __func__
);
998 flow_log(" in:%u authFirst:%u\n",
999 req_opts
->is_inbound
, req_opts
->auth_first
);
1000 flow_log(" cipher alg:%u mode:%u type %u\n", cipher_parms
->alg
,
1001 cipher_parms
->mode
, cipher_parms
->type
);
1002 flow_log(" is_esp: %s\n", req_opts
->is_esp
? "yes" : "no");
1003 flow_log(" key: %d\n", cipher_parms
->key_len
);
1004 flow_dump(" key: ", cipher_parms
->key_buf
, cipher_parms
->key_len
);
1005 flow_log(" iv: %d\n", cipher_parms
->iv_len
);
1006 flow_dump(" iv: ", cipher_parms
->iv_buf
, cipher_parms
->iv_len
);
1007 flow_log(" auth alg:%u mode:%u type %u\n",
1008 hash_parms
->alg
, hash_parms
->mode
, hash_parms
->type
);
1009 flow_log(" digestsize: %u\n", hash_parms
->digestsize
);
1010 flow_log(" authkey: %d\n", hash_parms
->key_len
);
1011 flow_dump(" authkey: ", hash_parms
->key_buf
, hash_parms
->key_len
);
1012 flow_log(" assoc_size:%u\n", assoc_size
);
1013 flow_log(" prebuf_len:%u\n", hash_parms
->prebuf_len
);
1014 flow_log(" data_size:%u\n", data_size
);
1015 flow_log(" hash_pad_len:%u\n", hash_parms
->pad_len
);
1016 flow_log(" real_db_size:%u\n", real_db_size
);
1017 flow_log(" cipher_offset:%u payload_len:%u\n",
1018 cipher_offset
, payload_len
);
1019 flow_log(" aead_iv: %u\n", aead_parms
->iv_len
);
1021 /* Convert to spu2 values for cipher alg, hash alg */
1022 err
= spu2_cipher_xlate(cipher_parms
->alg
, cipher_parms
->mode
,
1024 &spu2_ciph_type
, &spu2_ciph_mode
);
1026 /* If we are doing GCM hashing only - either via rfc4543 transform
1027 * or because we happen to do GCM with AAD only and no payload - we
1028 * need to configure hardware to use hash key rather than cipher key
1029 * and put data into payload. This is because unlike SPU-M, running
1030 * GCM cipher with 0 size payload is not permitted.
1032 if ((req_opts
->is_rfc4543
) ||
1033 ((spu2_ciph_mode
== SPU2_CIPHER_MODE_GCM
) &&
1034 (payload_len
== 0))) {
1035 /* Use hashing (only) and set up hash key */
1036 spu2_ciph_type
= SPU2_CIPHER_TYPE_NONE
;
1037 hash_parms
->key_len
= cipher_parms
->key_len
;
1038 memcpy(hash_parms
->key_buf
, cipher_parms
->key_buf
,
1039 cipher_parms
->key_len
);
1040 cipher_parms
->key_len
= 0;
1042 if (req_opts
->is_rfc4543
)
1043 payload_len
+= assoc_size
;
1045 payload_len
= assoc_size
;
1053 flow_log("spu2 cipher type %s, cipher mode %s\n",
1054 spu2_ciph_type_name(spu2_ciph_type
),
1055 spu2_ciph_mode_name(spu2_ciph_mode
));
1057 err
= spu2_hash_xlate(hash_parms
->alg
, hash_parms
->mode
,
1060 &spu2_auth_type
, &spu2_auth_mode
);
1064 flow_log("spu2 hash type %s, hash mode %s\n",
1065 spu2_hash_type_name(spu2_auth_type
),
1066 spu2_hash_mode_name(spu2_auth_mode
));
1068 fmd
= (struct SPU2_FMD
*)spu_hdr
;
1070 spu2_fmd_ctrl0_write(fmd
, req_opts
->is_inbound
, req_opts
->auth_first
,
1071 proto
, spu2_ciph_type
, spu2_ciph_mode
,
1072 spu2_auth_type
, spu2_auth_mode
);
1074 spu2_fmd_ctrl1_write(fmd
, req_opts
->is_inbound
, assoc_size
,
1075 hash_parms
->key_len
, cipher_parms
->key_len
,
1077 aead_parms
->return_iv
, aead_parms
->ret_iv_len
,
1078 aead_parms
->ret_iv_off
,
1079 cipher_parms
->iv_len
, hash_parms
->digestsize
,
1080 !req_opts
->bd_suppress
, return_md
);
1082 spu2_fmd_ctrl2_write(fmd
, cipher_offset
, hash_parms
->key_len
, 0,
1083 cipher_parms
->key_len
, cipher_parms
->iv_len
);
1085 spu2_fmd_ctrl3_write(fmd
, payload_len
);
1087 ptr
= (u8
*)(fmd
+ 1);
1088 buf_len
= sizeof(struct SPU2_FMD
);
1091 if (hash_parms
->key_len
) {
1092 memcpy(ptr
, hash_parms
->key_buf
, hash_parms
->key_len
);
1093 ptr
+= hash_parms
->key_len
;
1094 buf_len
+= hash_parms
->key_len
;
1096 if (cipher_parms
->key_len
) {
1097 memcpy(ptr
, cipher_parms
->key_buf
, cipher_parms
->key_len
);
1098 ptr
+= cipher_parms
->key_len
;
1099 buf_len
+= cipher_parms
->key_len
;
1101 if (cipher_parms
->iv_len
) {
1102 memcpy(ptr
, cipher_parms
->iv_buf
, cipher_parms
->iv_len
);
1103 ptr
+= cipher_parms
->iv_len
;
1104 buf_len
+= cipher_parms
->iv_len
;
1107 packet_dump(" SPU request header: ", spu_hdr
, buf_len
);
1113 * spu2_cipher_req_init() - Build an skcipher SPU2 request message header,
1114 * including FMD and OMD.
1115 * @spu_hdr: Location of start of SPU request (FMD field)
1116 * @cipher_parms: Parameters describing cipher request
1118 * Called at setkey time to initialize a msg header that can be reused for all
1119 * subsequent skcipher requests. Construct the message starting at spu_hdr.
1120 * Caller should allocate this buffer in DMA-able memory at least
1121 * SPU_HEADER_ALLOC_LEN bytes long.
1123 * Return: the total length of the SPU header (FMD and OMD) in bytes. 0 if an
1126 u16
spu2_cipher_req_init(u8
*spu_hdr
, struct spu_cipher_parms
*cipher_parms
)
1128 struct SPU2_FMD
*fmd
;
1130 enum spu2_cipher_type spu2_type
= SPU2_CIPHER_TYPE_NONE
;
1131 enum spu2_cipher_mode spu2_mode
;
1134 flow_log("%s()\n", __func__
);
1135 flow_log(" cipher alg:%u mode:%u type %u\n", cipher_parms
->alg
,
1136 cipher_parms
->mode
, cipher_parms
->type
);
1137 flow_log(" cipher_iv_len: %u\n", cipher_parms
->iv_len
);
1138 flow_log(" key: %d\n", cipher_parms
->key_len
);
1139 flow_dump(" key: ", cipher_parms
->key_buf
, cipher_parms
->key_len
);
1141 /* Convert to spu2 values */
1142 err
= spu2_cipher_xlate(cipher_parms
->alg
, cipher_parms
->mode
,
1143 cipher_parms
->type
, &spu2_type
, &spu2_mode
);
1147 flow_log("spu2 cipher type %s, cipher mode %s\n",
1148 spu2_ciph_type_name(spu2_type
),
1149 spu2_ciph_mode_name(spu2_mode
));
1151 /* Construct the FMD header */
1152 fmd
= (struct SPU2_FMD
*)spu_hdr
;
1153 err
= spu2_fmd_init(fmd
, spu2_type
, spu2_mode
, cipher_parms
->key_len
,
1154 cipher_parms
->iv_len
);
1158 /* Write cipher key to OMD */
1159 omd
= (u8
*)(fmd
+ 1);
1160 if (cipher_parms
->key_buf
&& cipher_parms
->key_len
)
1161 memcpy(omd
, cipher_parms
->key_buf
, cipher_parms
->key_len
);
1163 packet_dump(" SPU request header: ", spu_hdr
,
1164 FMD_SIZE
+ cipher_parms
->key_len
+ cipher_parms
->iv_len
);
1166 return FMD_SIZE
+ cipher_parms
->key_len
+ cipher_parms
->iv_len
;
1170 * spu2_cipher_req_finish() - Finish building a SPU request message header for a
1171 * block cipher request.
1172 * @spu_hdr: Start of the request message header (MH field)
1173 * @spu_req_hdr_len: Length in bytes of the SPU request header
1174 * @is_inbound: 0 encrypt, 1 decrypt
1175 * @cipher_parms: Parameters describing cipher operation to be performed
1176 * @data_size: Length of the data in the BD field
1178 * Assumes much of the header was already filled in at setkey() time in
1179 * spu_cipher_req_init().
1180 * spu_cipher_req_init() fills in the encryption key.
1182 void spu2_cipher_req_finish(u8
*spu_hdr
,
1183 u16 spu_req_hdr_len
,
1184 unsigned int is_inbound
,
1185 struct spu_cipher_parms
*cipher_parms
,
1186 unsigned int data_size
)
1188 struct SPU2_FMD
*fmd
;
1189 u8
*omd
; /* start of optional metadata */
1193 flow_log("%s()\n", __func__
);
1194 flow_log(" in: %u\n", is_inbound
);
1195 flow_log(" cipher alg: %u, cipher_type: %u\n", cipher_parms
->alg
,
1196 cipher_parms
->type
);
1197 flow_log(" iv len: %d\n", cipher_parms
->iv_len
);
1198 flow_dump(" iv: ", cipher_parms
->iv_buf
, cipher_parms
->iv_len
);
1199 flow_log(" data_size: %u\n", data_size
);
1201 fmd
= (struct SPU2_FMD
*)spu_hdr
;
1202 omd
= (u8
*)(fmd
+ 1);
1205 * FMD ctrl0 was initialized at setkey time. update it to indicate
1206 * whether we are encrypting or decrypting.
1208 ctrl0
= le64_to_cpu(fmd
->ctrl0
);
1210 ctrl0
&= ~SPU2_CIPH_ENCRYPT_EN
; /* decrypt */
1212 ctrl0
|= SPU2_CIPH_ENCRYPT_EN
; /* encrypt */
1213 fmd
->ctrl0
= cpu_to_le64(ctrl0
);
1215 if (cipher_parms
->alg
&& cipher_parms
->iv_buf
&& cipher_parms
->iv_len
) {
1216 /* cipher iv provided so put it in here */
1217 memcpy(omd
+ cipher_parms
->key_len
, cipher_parms
->iv_buf
,
1218 cipher_parms
->iv_len
);
1221 ctrl3
= le64_to_cpu(fmd
->ctrl3
);
1222 data_size
&= SPU2_PL_LEN
;
1224 fmd
->ctrl3
= cpu_to_le64(ctrl3
);
1226 packet_dump(" SPU request header: ", spu_hdr
, spu_req_hdr_len
);
1230 * spu2_request_pad() - Create pad bytes at the end of the data.
1231 * @pad_start: Start of buffer where pad bytes are to be written
1232 * @gcm_padding: Length of GCM padding, in bytes
1233 * @hash_pad_len: Number of bytes of padding extend data to full block
1234 * @auth_alg: Authentication algorithm
1235 * @auth_mode: Authentication mode
1236 * @total_sent: Length inserted at end of hash pad
1237 * @status_padding: Number of bytes of padding to align STATUS word
1239 * There may be three forms of pad:
1240 * 1. GCM pad - for GCM mode ciphers, pad to 16-byte alignment
1241 * 2. hash pad - pad to a block length, with 0x80 data terminator and
1243 * 3. STAT pad - to ensure the STAT field is 4-byte aligned
1245 void spu2_request_pad(u8
*pad_start
, u32 gcm_padding
, u32 hash_pad_len
,
1246 enum hash_alg auth_alg
, enum hash_mode auth_mode
,
1247 unsigned int total_sent
, u32 status_padding
)
1249 u8
*ptr
= pad_start
;
1251 /* fix data alignent for GCM */
1252 if (gcm_padding
> 0) {
1253 flow_log(" GCM: padding to 16 byte alignment: %u bytes\n",
1255 memset(ptr
, 0, gcm_padding
);
1259 if (hash_pad_len
> 0) {
1260 /* clear the padding section */
1261 memset(ptr
, 0, hash_pad_len
);
1263 /* terminate the data */
1265 ptr
+= (hash_pad_len
- sizeof(u64
));
1267 /* add the size at the end as required per alg */
1268 if (auth_alg
== HASH_ALG_MD5
)
1269 *(__le64
*)ptr
= cpu_to_le64(total_sent
* 8ull);
1270 else /* SHA1, SHA2-224, SHA2-256 */
1271 *(__be64
*)ptr
= cpu_to_be64(total_sent
* 8ull);
1275 /* pad to a 4byte alignment for STAT */
1276 if (status_padding
> 0) {
1277 flow_log(" STAT: padding to 4 byte alignment: %u bytes\n",
1280 memset(ptr
, 0, status_padding
);
1281 ptr
+= status_padding
;
1286 * spu2_xts_tweak_in_payload() - Indicate that SPU2 does NOT place the XTS
1287 * tweak field in the packet payload (it uses IV instead)
1291 u8
spu2_xts_tweak_in_payload(void)
1297 * spu2_tx_status_len() - Return the length of the STATUS field in a SPU
1300 * Return: Length of STATUS field in bytes.
1302 u8
spu2_tx_status_len(void)
1304 return SPU2_TX_STATUS_LEN
;
1308 * spu2_rx_status_len() - Return the length of the STATUS field in a SPU
1311 * Return: Length of STATUS field in bytes.
1313 u8
spu2_rx_status_len(void)
1315 return SPU2_RX_STATUS_LEN
;
1319 * spu2_status_process() - Process the status from a SPU response message.
1320 * @statp: start of STATUS word
1322 * Return: 0 - if status is good and response should be processed
1323 * !0 - status indicates an error and response is invalid
1325 int spu2_status_process(u8
*statp
)
1327 /* SPU2 status is 2 bytes by default - SPU_RX_STATUS_LEN */
1328 u16 status
= le16_to_cpu(*(__le16
*)statp
);
1333 flow_log("rx status is %#x\n", status
);
1334 if (status
== SPU2_INVALID_ICV
)
1335 return SPU_INVALID_ICV
;
1341 * spu2_ccm_update_iv() - Update the IV as per the requirements for CCM mode.
1343 * @digestsize: Digest size of this request
1344 * @cipher_parms: (pointer to) cipher parmaeters, includes IV buf & IV len
1345 * @assoclen: Length of AAD data
1346 * @chunksize: length of input data to be sent in this req
1347 * @is_encrypt: true if this is an output/encrypt operation
1348 * @is_esp: true if this is an ESP / RFC4309 operation
1351 void spu2_ccm_update_iv(unsigned int digestsize
,
1352 struct spu_cipher_parms
*cipher_parms
,
1353 unsigned int assoclen
, unsigned int chunksize
,
1354 bool is_encrypt
, bool is_esp
)
1356 int L
; /* size of length field, in bytes */
1359 * In RFC4309 mode, L is fixed at 4 bytes; otherwise, IV from
1360 * testmgr contains (L-1) in bottom 3 bits of first byte,
1364 L
= CCM_ESP_L_VALUE
;
1366 L
= ((cipher_parms
->iv_buf
[0] & CCM_B0_L_PRIME
) >>
1367 CCM_B0_L_PRIME_SHIFT
) + 1;
1369 /* SPU2 doesn't want these length bytes nor the first byte... */
1370 cipher_parms
->iv_len
-= (1 + L
);
1371 memmove(cipher_parms
->iv_buf
, &cipher_parms
->iv_buf
[1],
1372 cipher_parms
->iv_len
);
1376 * spu2_wordalign_padlen() - SPU2 does not require padding.
1377 * @data_size: length of data field in bytes
1379 * Return: length of status field padding, in bytes (always 0 on SPU2)
1381 u32
spu2_wordalign_padlen(u32 data_size
)