2 * AES CTR routines supporting VMX instructions on the Power 8
4 * Copyright (C) 2015 International Business Machines Inc.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; version 2 only.
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
15 * You should have received a copy of the GNU General Public License
16 * along with this program; if not, write to the Free Software
17 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
19 * Author: Marcelo Henrique Cerri <mhcerri@br.ibm.com>
22 #include <linux/types.h>
23 #include <linux/err.h>
24 #include <linux/crypto.h>
25 #include <linux/delay.h>
26 #include <linux/hardirq.h>
27 #include <asm/switch_to.h>
28 #include <crypto/aes.h>
29 #include <crypto/scatterwalk.h>
30 #include <crypto/skcipher.h>
32 #include "aesp8-ppc.h"
34 struct p8_aes_ctr_ctx
{
35 struct crypto_skcipher
*fallback
;
36 struct aes_key enc_key
;
39 static int p8_aes_ctr_init(struct crypto_tfm
*tfm
)
41 const char *alg
= crypto_tfm_alg_name(tfm
);
42 struct crypto_skcipher
*fallback
;
43 struct p8_aes_ctr_ctx
*ctx
= crypto_tfm_ctx(tfm
);
45 fallback
= crypto_alloc_skcipher(alg
, 0,
46 CRYPTO_ALG_ASYNC
| CRYPTO_ALG_NEED_FALLBACK
);
47 if (IS_ERR(fallback
)) {
49 "Failed to allocate transformation for '%s': %ld\n",
50 alg
, PTR_ERR(fallback
));
51 return PTR_ERR(fallback
);
54 crypto_skcipher_set_flags(
56 crypto_skcipher_get_flags((struct crypto_skcipher
*)tfm
));
57 ctx
->fallback
= fallback
;
62 static void p8_aes_ctr_exit(struct crypto_tfm
*tfm
)
64 struct p8_aes_ctr_ctx
*ctx
= crypto_tfm_ctx(tfm
);
67 crypto_free_skcipher(ctx
->fallback
);
72 static int p8_aes_ctr_setkey(struct crypto_tfm
*tfm
, const u8
*key
,
76 struct p8_aes_ctr_ctx
*ctx
= crypto_tfm_ctx(tfm
);
81 ret
= aes_p8_set_encrypt_key(key
, keylen
* 8, &ctx
->enc_key
);
86 ret
+= crypto_skcipher_setkey(ctx
->fallback
, key
, keylen
);
90 static void p8_aes_ctr_final(struct p8_aes_ctr_ctx
*ctx
,
91 struct blkcipher_walk
*walk
)
93 u8
*ctrblk
= walk
->iv
;
94 u8 keystream
[AES_BLOCK_SIZE
];
95 u8
*src
= walk
->src
.virt
.addr
;
96 u8
*dst
= walk
->dst
.virt
.addr
;
97 unsigned int nbytes
= walk
->nbytes
;
102 aes_p8_encrypt(ctrblk
, keystream
, &ctx
->enc_key
);
103 disable_kernel_vsx();
107 crypto_xor_cpy(dst
, keystream
, src
, nbytes
);
108 crypto_inc(ctrblk
, AES_BLOCK_SIZE
);
111 static int p8_aes_ctr_crypt(struct blkcipher_desc
*desc
,
112 struct scatterlist
*dst
,
113 struct scatterlist
*src
, unsigned int nbytes
)
117 struct blkcipher_walk walk
;
118 struct p8_aes_ctr_ctx
*ctx
=
119 crypto_tfm_ctx(crypto_blkcipher_tfm(desc
->tfm
));
121 if (in_interrupt()) {
122 SKCIPHER_REQUEST_ON_STACK(req
, ctx
->fallback
);
123 skcipher_request_set_tfm(req
, ctx
->fallback
);
124 skcipher_request_set_callback(req
, desc
->flags
, NULL
, NULL
);
125 skcipher_request_set_crypt(req
, src
, dst
, nbytes
, desc
->info
);
126 ret
= crypto_skcipher_encrypt(req
);
127 skcipher_request_zero(req
);
129 blkcipher_walk_init(&walk
, dst
, src
, nbytes
);
130 ret
= blkcipher_walk_virt_block(desc
, &walk
, AES_BLOCK_SIZE
);
131 while ((nbytes
= walk
.nbytes
) >= AES_BLOCK_SIZE
) {
135 aes_p8_ctr32_encrypt_blocks(walk
.src
.virt
.addr
,
142 disable_kernel_vsx();
146 /* We need to update IV mostly for last bytes/round */
147 inc
= (nbytes
& AES_BLOCK_MASK
) / AES_BLOCK_SIZE
;
150 crypto_inc(walk
.iv
, AES_BLOCK_SIZE
);
152 nbytes
&= AES_BLOCK_SIZE
- 1;
153 ret
= blkcipher_walk_done(desc
, &walk
, nbytes
);
156 p8_aes_ctr_final(ctx
, &walk
);
157 ret
= blkcipher_walk_done(desc
, &walk
, 0);
164 struct crypto_alg p8_aes_ctr_alg
= {
165 .cra_name
= "ctr(aes)",
166 .cra_driver_name
= "p8_aes_ctr",
167 .cra_module
= THIS_MODULE
,
168 .cra_priority
= 2000,
169 .cra_type
= &crypto_blkcipher_type
,
170 .cra_flags
= CRYPTO_ALG_TYPE_BLKCIPHER
| CRYPTO_ALG_NEED_FALLBACK
,
173 .cra_ctxsize
= sizeof(struct p8_aes_ctr_ctx
),
174 .cra_init
= p8_aes_ctr_init
,
175 .cra_exit
= p8_aes_ctr_exit
,
177 .ivsize
= AES_BLOCK_SIZE
,
178 .min_keysize
= AES_MIN_KEY_SIZE
,
179 .max_keysize
= AES_MAX_KEY_SIZE
,
180 .setkey
= p8_aes_ctr_setkey
,
181 .encrypt
= p8_aes_ctr_crypt
,
182 .decrypt
= p8_aes_ctr_crypt
,