1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * seqiv: Sequence Number IV Generator
5 * This generator generates an IV based on a sequence number by xoring it
6 * with a salt. This algorithm is mainly useful for CTR and similar modes.
8 * Copyright (c) 2007 Herbert Xu <herbert@gondor.apana.org.au>
11 #include <crypto/internal/geniv.h>
12 #include <crypto/scatterwalk.h>
13 #include <crypto/skcipher.h>
14 #include <linux/err.h>
15 #include <linux/init.h>
16 #include <linux/kernel.h>
17 #include <linux/module.h>
18 #include <linux/slab.h>
19 #include <linux/string.h>
21 static void seqiv_free(struct crypto_instance
*inst
);
23 static void seqiv_aead_encrypt_complete2(struct aead_request
*req
, int err
)
25 struct aead_request
*subreq
= aead_request_ctx(req
);
26 struct crypto_aead
*geniv
;
28 if (err
== -EINPROGRESS
)
34 geniv
= crypto_aead_reqtfm(req
);
35 memcpy(req
->iv
, subreq
->iv
, crypto_aead_ivsize(geniv
));
41 static void seqiv_aead_encrypt_complete(struct crypto_async_request
*base
,
44 struct aead_request
*req
= base
->data
;
46 seqiv_aead_encrypt_complete2(req
, err
);
47 aead_request_complete(req
, err
);
50 static int seqiv_aead_encrypt(struct aead_request
*req
)
52 struct crypto_aead
*geniv
= crypto_aead_reqtfm(req
);
53 struct aead_geniv_ctx
*ctx
= crypto_aead_ctx(geniv
);
54 struct aead_request
*subreq
= aead_request_ctx(req
);
55 crypto_completion_t
compl;
58 unsigned int ivsize
= 8;
61 if (req
->cryptlen
< ivsize
)
64 aead_request_set_tfm(subreq
, ctx
->child
);
66 compl = req
->base
.complete
;
67 data
= req
->base
.data
;
70 if (req
->src
!= req
->dst
) {
71 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq
, ctx
->sknull
);
73 skcipher_request_set_sync_tfm(nreq
, ctx
->sknull
);
74 skcipher_request_set_callback(nreq
, req
->base
.flags
,
76 skcipher_request_set_crypt(nreq
, req
->src
, req
->dst
,
77 req
->assoclen
+ req
->cryptlen
,
80 err
= crypto_skcipher_encrypt(nreq
);
85 if (unlikely(!IS_ALIGNED((unsigned long)info
,
86 crypto_aead_alignmask(geniv
) + 1))) {
87 info
= kmemdup(req
->iv
, ivsize
, req
->base
.flags
&
88 CRYPTO_TFM_REQ_MAY_SLEEP
? GFP_KERNEL
:
93 compl = seqiv_aead_encrypt_complete
;
97 aead_request_set_callback(subreq
, req
->base
.flags
, compl, data
);
98 aead_request_set_crypt(subreq
, req
->dst
, req
->dst
,
99 req
->cryptlen
- ivsize
, info
);
100 aead_request_set_ad(subreq
, req
->assoclen
+ ivsize
);
102 crypto_xor(info
, ctx
->salt
, ivsize
);
103 scatterwalk_map_and_copy(info
, req
->dst
, req
->assoclen
, ivsize
, 1);
105 err
= crypto_aead_encrypt(subreq
);
106 if (unlikely(info
!= req
->iv
))
107 seqiv_aead_encrypt_complete2(req
, err
);
111 static int seqiv_aead_decrypt(struct aead_request
*req
)
113 struct crypto_aead
*geniv
= crypto_aead_reqtfm(req
);
114 struct aead_geniv_ctx
*ctx
= crypto_aead_ctx(geniv
);
115 struct aead_request
*subreq
= aead_request_ctx(req
);
116 crypto_completion_t
compl;
118 unsigned int ivsize
= 8;
120 if (req
->cryptlen
< ivsize
+ crypto_aead_authsize(geniv
))
123 aead_request_set_tfm(subreq
, ctx
->child
);
125 compl = req
->base
.complete
;
126 data
= req
->base
.data
;
128 aead_request_set_callback(subreq
, req
->base
.flags
, compl, data
);
129 aead_request_set_crypt(subreq
, req
->src
, req
->dst
,
130 req
->cryptlen
- ivsize
, req
->iv
);
131 aead_request_set_ad(subreq
, req
->assoclen
+ ivsize
);
133 scatterwalk_map_and_copy(req
->iv
, req
->src
, req
->assoclen
, ivsize
, 0);
135 return crypto_aead_decrypt(subreq
);
138 static int seqiv_aead_create(struct crypto_template
*tmpl
, struct rtattr
**tb
)
140 struct aead_instance
*inst
;
143 inst
= aead_geniv_alloc(tmpl
, tb
, 0, 0);
146 return PTR_ERR(inst
);
149 if (inst
->alg
.ivsize
!= sizeof(u64
))
152 inst
->alg
.encrypt
= seqiv_aead_encrypt
;
153 inst
->alg
.decrypt
= seqiv_aead_decrypt
;
155 inst
->alg
.init
= aead_init_geniv
;
156 inst
->alg
.exit
= aead_exit_geniv
;
158 inst
->alg
.base
.cra_ctxsize
= sizeof(struct aead_geniv_ctx
);
159 inst
->alg
.base
.cra_ctxsize
+= inst
->alg
.ivsize
;
161 err
= aead_register_instance(tmpl
, inst
);
169 aead_geniv_free(inst
);
173 static int seqiv_create(struct crypto_template
*tmpl
, struct rtattr
**tb
)
175 struct crypto_attr_type
*algt
;
177 algt
= crypto_get_attr_type(tb
);
179 return PTR_ERR(algt
);
181 if ((algt
->type
^ CRYPTO_ALG_TYPE_AEAD
) & CRYPTO_ALG_TYPE_MASK
)
184 return seqiv_aead_create(tmpl
, tb
);
187 static void seqiv_free(struct crypto_instance
*inst
)
189 aead_geniv_free(aead_instance(inst
));
192 static struct crypto_template seqiv_tmpl
= {
194 .create
= seqiv_create
,
196 .module
= THIS_MODULE
,
199 static int __init
seqiv_module_init(void)
201 return crypto_register_template(&seqiv_tmpl
);
204 static void __exit
seqiv_module_exit(void)
206 crypto_unregister_template(&seqiv_tmpl
);
209 subsys_initcall(seqiv_module_init
);
210 module_exit(seqiv_module_exit
);
212 MODULE_LICENSE("GPL");
213 MODULE_DESCRIPTION("Sequence Number IV Generator");
214 MODULE_ALIAS_CRYPTO("seqiv");