2 * Crypto user configuration API.
4 * Copyright (C) 2011 secunet Security Networks AG
5 * Copyright (C) 2011 Steffen Klassert <steffen.klassert@secunet.com>
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms and conditions of the GNU General Public License,
9 * version 2, as published by the Free Software Foundation.
11 * This program is distributed in the hope it will be useful, but WITHOUT
12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
16 * You should have received a copy of the GNU General Public License along with
17 * this program; if not, write to the Free Software Foundation, Inc.,
18 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21 #include <linux/module.h>
22 #include <linux/crypto.h>
23 #include <linux/cryptouser.h>
24 #include <linux/sched.h>
25 #include <net/netlink.h>
26 #include <linux/security.h>
27 #include <net/net_namespace.h>
28 #include <crypto/internal/aead.h>
29 #include <crypto/internal/skcipher.h>
30 #include <crypto/internal/rng.h>
31 #include <crypto/akcipher.h>
35 #define null_terminated(x) (strnlen(x, sizeof(x)) < sizeof(x))
37 static DEFINE_MUTEX(crypto_cfg_mutex
);
39 /* The crypto netlink socket */
40 static struct sock
*crypto_nlsk
;
42 struct crypto_dump_info
{
43 struct sk_buff
*in_skb
;
44 struct sk_buff
*out_skb
;
49 static struct crypto_alg
*crypto_alg_match(struct crypto_user_alg
*p
, int exact
)
51 struct crypto_alg
*q
, *alg
= NULL
;
53 down_read(&crypto_alg_sem
);
55 list_for_each_entry(q
, &crypto_alg_list
, cra_list
) {
58 if ((q
->cra_flags
^ p
->cru_type
) & p
->cru_mask
)
61 if (strlen(p
->cru_driver_name
))
62 match
= !strcmp(q
->cra_driver_name
,
65 match
= !strcmp(q
->cra_name
, p
->cru_name
);
70 if (unlikely(!crypto_mod_get(q
)))
77 up_read(&crypto_alg_sem
);
82 static int crypto_report_cipher(struct sk_buff
*skb
, struct crypto_alg
*alg
)
84 struct crypto_report_cipher rcipher
;
86 strncpy(rcipher
.type
, "cipher", sizeof(rcipher
.type
));
88 rcipher
.blocksize
= alg
->cra_blocksize
;
89 rcipher
.min_keysize
= alg
->cra_cipher
.cia_min_keysize
;
90 rcipher
.max_keysize
= alg
->cra_cipher
.cia_max_keysize
;
92 if (nla_put(skb
, CRYPTOCFGA_REPORT_CIPHER
,
93 sizeof(struct crypto_report_cipher
), &rcipher
))
101 static int crypto_report_comp(struct sk_buff
*skb
, struct crypto_alg
*alg
)
103 struct crypto_report_comp rcomp
;
105 strncpy(rcomp
.type
, "compression", sizeof(rcomp
.type
));
106 if (nla_put(skb
, CRYPTOCFGA_REPORT_COMPRESS
,
107 sizeof(struct crypto_report_comp
), &rcomp
))
108 goto nla_put_failure
;
115 static int crypto_report_akcipher(struct sk_buff
*skb
, struct crypto_alg
*alg
)
117 struct crypto_report_akcipher rakcipher
;
119 strncpy(rakcipher
.type
, "akcipher", sizeof(rakcipher
.type
));
121 if (nla_put(skb
, CRYPTOCFGA_REPORT_AKCIPHER
,
122 sizeof(struct crypto_report_akcipher
), &rakcipher
))
123 goto nla_put_failure
;
130 static int crypto_report_one(struct crypto_alg
*alg
,
131 struct crypto_user_alg
*ualg
, struct sk_buff
*skb
)
133 strncpy(ualg
->cru_name
, alg
->cra_name
, sizeof(ualg
->cru_name
));
134 strncpy(ualg
->cru_driver_name
, alg
->cra_driver_name
,
135 sizeof(ualg
->cru_driver_name
));
136 strncpy(ualg
->cru_module_name
, module_name(alg
->cra_module
),
137 sizeof(ualg
->cru_module_name
));
141 ualg
->cru_flags
= alg
->cra_flags
;
142 ualg
->cru_refcnt
= atomic_read(&alg
->cra_refcnt
);
144 if (nla_put_u32(skb
, CRYPTOCFGA_PRIORITY_VAL
, alg
->cra_priority
))
145 goto nla_put_failure
;
146 if (alg
->cra_flags
& CRYPTO_ALG_LARVAL
) {
147 struct crypto_report_larval rl
;
149 strncpy(rl
.type
, "larval", sizeof(rl
.type
));
150 if (nla_put(skb
, CRYPTOCFGA_REPORT_LARVAL
,
151 sizeof(struct crypto_report_larval
), &rl
))
152 goto nla_put_failure
;
156 if (alg
->cra_type
&& alg
->cra_type
->report
) {
157 if (alg
->cra_type
->report(skb
, alg
))
158 goto nla_put_failure
;
163 switch (alg
->cra_flags
& (CRYPTO_ALG_TYPE_MASK
| CRYPTO_ALG_LARVAL
)) {
164 case CRYPTO_ALG_TYPE_CIPHER
:
165 if (crypto_report_cipher(skb
, alg
))
166 goto nla_put_failure
;
169 case CRYPTO_ALG_TYPE_COMPRESS
:
170 if (crypto_report_comp(skb
, alg
))
171 goto nla_put_failure
;
175 case CRYPTO_ALG_TYPE_AKCIPHER
:
176 if (crypto_report_akcipher(skb
, alg
))
177 goto nla_put_failure
;
189 static int crypto_report_alg(struct crypto_alg
*alg
,
190 struct crypto_dump_info
*info
)
192 struct sk_buff
*in_skb
= info
->in_skb
;
193 struct sk_buff
*skb
= info
->out_skb
;
194 struct nlmsghdr
*nlh
;
195 struct crypto_user_alg
*ualg
;
198 nlh
= nlmsg_put(skb
, NETLINK_CB(in_skb
).portid
, info
->nlmsg_seq
,
199 CRYPTO_MSG_GETALG
, sizeof(*ualg
), info
->nlmsg_flags
);
205 ualg
= nlmsg_data(nlh
);
207 err
= crypto_report_one(alg
, ualg
, skb
);
209 nlmsg_cancel(skb
, nlh
);
219 static int crypto_report(struct sk_buff
*in_skb
, struct nlmsghdr
*in_nlh
,
220 struct nlattr
**attrs
)
222 struct crypto_user_alg
*p
= nlmsg_data(in_nlh
);
223 struct crypto_alg
*alg
;
225 struct crypto_dump_info info
;
228 if (!null_terminated(p
->cru_name
) || !null_terminated(p
->cru_driver_name
))
231 alg
= crypto_alg_match(p
, 0);
236 skb
= nlmsg_new(NLMSG_DEFAULT_SIZE
, GFP_ATOMIC
);
240 info
.in_skb
= in_skb
;
242 info
.nlmsg_seq
= in_nlh
->nlmsg_seq
;
243 info
.nlmsg_flags
= 0;
245 err
= crypto_report_alg(alg
, &info
);
253 return nlmsg_unicast(crypto_nlsk
, skb
, NETLINK_CB(in_skb
).portid
);
256 static int crypto_dump_report(struct sk_buff
*skb
, struct netlink_callback
*cb
)
258 struct crypto_alg
*alg
;
259 struct crypto_dump_info info
;
267 info
.in_skb
= cb
->skb
;
269 info
.nlmsg_seq
= cb
->nlh
->nlmsg_seq
;
270 info
.nlmsg_flags
= NLM_F_MULTI
;
272 list_for_each_entry(alg
, &crypto_alg_list
, cra_list
) {
273 err
= crypto_report_alg(alg
, &info
);
284 static int crypto_dump_report_done(struct netlink_callback
*cb
)
289 static int crypto_update_alg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
,
290 struct nlattr
**attrs
)
292 struct crypto_alg
*alg
;
293 struct crypto_user_alg
*p
= nlmsg_data(nlh
);
294 struct nlattr
*priority
= attrs
[CRYPTOCFGA_PRIORITY_VAL
];
297 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
300 if (!null_terminated(p
->cru_name
) || !null_terminated(p
->cru_driver_name
))
303 if (priority
&& !strlen(p
->cru_driver_name
))
306 alg
= crypto_alg_match(p
, 1);
310 down_write(&crypto_alg_sem
);
312 crypto_remove_spawns(alg
, &list
, NULL
);
315 alg
->cra_priority
= nla_get_u32(priority
);
317 up_write(&crypto_alg_sem
);
320 crypto_remove_final(&list
);
325 static int crypto_del_alg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
,
326 struct nlattr
**attrs
)
328 struct crypto_alg
*alg
;
329 struct crypto_user_alg
*p
= nlmsg_data(nlh
);
332 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
335 if (!null_terminated(p
->cru_name
) || !null_terminated(p
->cru_driver_name
))
338 alg
= crypto_alg_match(p
, 1);
342 /* We can not unregister core algorithms such as aes-generic.
343 * We would loose the reference in the crypto_alg_list to this algorithm
344 * if we try to unregister. Unregistering such an algorithm without
345 * removing the module is not possible, so we restrict to crypto
346 * instances that are build from templates. */
348 if (!(alg
->cra_flags
& CRYPTO_ALG_INSTANCE
))
352 if (atomic_read(&alg
->cra_refcnt
) > 2)
355 err
= crypto_unregister_instance((struct crypto_instance
*)alg
);
362 static struct crypto_alg
*crypto_user_skcipher_alg(const char *name
, u32 type
,
366 struct crypto_alg
*alg
;
368 type
= crypto_skcipher_type(type
);
369 mask
= crypto_skcipher_mask(mask
);
372 alg
= crypto_lookup_skcipher(name
, type
, mask
);
379 if (signal_pending(current
)) {
388 static struct crypto_alg
*crypto_user_aead_alg(const char *name
, u32 type
,
392 struct crypto_alg
*alg
;
394 type
&= ~(CRYPTO_ALG_TYPE_MASK
| CRYPTO_ALG_GENIV
);
395 type
|= CRYPTO_ALG_TYPE_AEAD
;
396 mask
&= ~(CRYPTO_ALG_TYPE_MASK
| CRYPTO_ALG_GENIV
);
397 mask
|= CRYPTO_ALG_TYPE_MASK
;
400 alg
= crypto_lookup_aead(name
, type
, mask
);
407 if (signal_pending(current
)) {
416 static int crypto_add_alg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
,
417 struct nlattr
**attrs
)
421 struct crypto_alg
*alg
;
422 struct crypto_user_alg
*p
= nlmsg_data(nlh
);
423 struct nlattr
*priority
= attrs
[CRYPTOCFGA_PRIORITY_VAL
];
425 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
428 if (!null_terminated(p
->cru_name
) || !null_terminated(p
->cru_driver_name
))
431 if (strlen(p
->cru_driver_name
))
434 if (priority
&& !exact
)
437 alg
= crypto_alg_match(p
, exact
);
443 if (strlen(p
->cru_driver_name
))
444 name
= p
->cru_driver_name
;
448 switch (p
->cru_type
& p
->cru_mask
& CRYPTO_ALG_TYPE_MASK
) {
449 case CRYPTO_ALG_TYPE_AEAD
:
450 alg
= crypto_user_aead_alg(name
, p
->cru_type
, p
->cru_mask
);
452 case CRYPTO_ALG_TYPE_GIVCIPHER
:
453 case CRYPTO_ALG_TYPE_BLKCIPHER
:
454 case CRYPTO_ALG_TYPE_ABLKCIPHER
:
455 alg
= crypto_user_skcipher_alg(name
, p
->cru_type
, p
->cru_mask
);
458 alg
= crypto_alg_mod_lookup(name
, p
->cru_type
, p
->cru_mask
);
464 down_write(&crypto_alg_sem
);
467 alg
->cra_priority
= nla_get_u32(priority
);
469 up_write(&crypto_alg_sem
);
476 static int crypto_del_rng(struct sk_buff
*skb
, struct nlmsghdr
*nlh
,
477 struct nlattr
**attrs
)
479 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
481 return crypto_del_default_rng();
484 #define MSGSIZE(type) sizeof(struct type)
486 static const int crypto_msg_min
[CRYPTO_NR_MSGTYPES
] = {
487 [CRYPTO_MSG_NEWALG
- CRYPTO_MSG_BASE
] = MSGSIZE(crypto_user_alg
),
488 [CRYPTO_MSG_DELALG
- CRYPTO_MSG_BASE
] = MSGSIZE(crypto_user_alg
),
489 [CRYPTO_MSG_UPDATEALG
- CRYPTO_MSG_BASE
] = MSGSIZE(crypto_user_alg
),
490 [CRYPTO_MSG_DELRNG
- CRYPTO_MSG_BASE
] = 0,
493 static const struct nla_policy crypto_policy
[CRYPTOCFGA_MAX
+1] = {
494 [CRYPTOCFGA_PRIORITY_VAL
] = { .type
= NLA_U32
},
499 static const struct crypto_link
{
500 int (*doit
)(struct sk_buff
*, struct nlmsghdr
*, struct nlattr
**);
501 int (*dump
)(struct sk_buff
*, struct netlink_callback
*);
502 int (*done
)(struct netlink_callback
*);
503 } crypto_dispatch
[CRYPTO_NR_MSGTYPES
] = {
504 [CRYPTO_MSG_NEWALG
- CRYPTO_MSG_BASE
] = { .doit
= crypto_add_alg
},
505 [CRYPTO_MSG_DELALG
- CRYPTO_MSG_BASE
] = { .doit
= crypto_del_alg
},
506 [CRYPTO_MSG_UPDATEALG
- CRYPTO_MSG_BASE
] = { .doit
= crypto_update_alg
},
507 [CRYPTO_MSG_GETALG
- CRYPTO_MSG_BASE
] = { .doit
= crypto_report
,
508 .dump
= crypto_dump_report
,
509 .done
= crypto_dump_report_done
},
510 [CRYPTO_MSG_DELRNG
- CRYPTO_MSG_BASE
] = { .doit
= crypto_del_rng
},
513 static int crypto_user_rcv_msg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
515 struct nlattr
*attrs
[CRYPTOCFGA_MAX
+1];
516 const struct crypto_link
*link
;
519 type
= nlh
->nlmsg_type
;
520 if (type
> CRYPTO_MSG_MAX
)
523 type
-= CRYPTO_MSG_BASE
;
524 link
= &crypto_dispatch
[type
];
526 if ((type
== (CRYPTO_MSG_GETALG
- CRYPTO_MSG_BASE
) &&
527 (nlh
->nlmsg_flags
& NLM_F_DUMP
))) {
528 struct crypto_alg
*alg
;
531 if (link
->dump
== NULL
)
534 list_for_each_entry(alg
, &crypto_alg_list
, cra_list
)
535 dump_alloc
+= CRYPTO_REPORT_MAXSIZE
;
538 struct netlink_dump_control c
= {
541 .min_dump_alloc
= dump_alloc
,
543 return netlink_dump_start(crypto_nlsk
, skb
, nlh
, &c
);
547 err
= nlmsg_parse(nlh
, crypto_msg_min
[type
], attrs
, CRYPTOCFGA_MAX
,
552 if (link
->doit
== NULL
)
555 return link
->doit(skb
, nlh
, attrs
);
558 static void crypto_netlink_rcv(struct sk_buff
*skb
)
560 mutex_lock(&crypto_cfg_mutex
);
561 netlink_rcv_skb(skb
, &crypto_user_rcv_msg
);
562 mutex_unlock(&crypto_cfg_mutex
);
565 static int __init
crypto_user_init(void)
567 struct netlink_kernel_cfg cfg
= {
568 .input
= crypto_netlink_rcv
,
571 crypto_nlsk
= netlink_kernel_create(&init_net
, NETLINK_CRYPTO
, &cfg
);
578 static void __exit
crypto_user_exit(void)
580 netlink_kernel_release(crypto_nlsk
);
583 module_init(crypto_user_init
);
584 module_exit(crypto_user_exit
);
585 MODULE_LICENSE("GPL");
586 MODULE_AUTHOR("Steffen Klassert <steffen.klassert@secunet.com>");
587 MODULE_DESCRIPTION("Crypto userspace configuration API");
588 MODULE_ALIAS("net-pf-16-proto-21");