Linux 4.14.212
[linux/fpc-iii.git] / crypto / algapi.c
blob603d2d6372091d199b6f45af7e1417936a7eef76
1 /*
2 * Cryptographic API for algorithms (i.e., low-level API).
4 * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au>
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License as published by the Free
8 * Software Foundation; either version 2 of the License, or (at your option)
9 * any later version.
13 #include <linux/err.h>
14 #include <linux/errno.h>
15 #include <linux/fips.h>
16 #include <linux/init.h>
17 #include <linux/kernel.h>
18 #include <linux/list.h>
19 #include <linux/module.h>
20 #include <linux/rtnetlink.h>
21 #include <linux/slab.h>
22 #include <linux/string.h>
24 #include "internal.h"
26 static LIST_HEAD(crypto_template_list);
28 static inline int crypto_set_driver_name(struct crypto_alg *alg)
30 static const char suffix[] = "-generic";
31 char *driver_name = alg->cra_driver_name;
32 int len;
34 if (*driver_name)
35 return 0;
37 len = strlcpy(driver_name, alg->cra_name, CRYPTO_MAX_ALG_NAME);
38 if (len + sizeof(suffix) > CRYPTO_MAX_ALG_NAME)
39 return -ENAMETOOLONG;
41 memcpy(driver_name + len, suffix, sizeof(suffix));
42 return 0;
45 static inline void crypto_check_module_sig(struct module *mod)
47 if (fips_enabled && mod && !module_sig_ok(mod))
48 panic("Module %s signature verification failed in FIPS mode\n",
49 module_name(mod));
52 static int crypto_check_alg(struct crypto_alg *alg)
54 crypto_check_module_sig(alg->cra_module);
56 if (alg->cra_alignmask & (alg->cra_alignmask + 1))
57 return -EINVAL;
59 if (alg->cra_blocksize > PAGE_SIZE / 8)
60 return -EINVAL;
62 if (alg->cra_priority < 0)
63 return -EINVAL;
65 atomic_set(&alg->cra_refcnt, 1);
67 return crypto_set_driver_name(alg);
70 static void crypto_free_instance(struct crypto_instance *inst)
72 if (!inst->alg.cra_type->free) {
73 inst->tmpl->free(inst);
74 return;
77 inst->alg.cra_type->free(inst);
80 static void crypto_destroy_instance(struct crypto_alg *alg)
82 struct crypto_instance *inst = (void *)alg;
83 struct crypto_template *tmpl = inst->tmpl;
85 crypto_free_instance(inst);
86 crypto_tmpl_put(tmpl);
89 static struct list_head *crypto_more_spawns(struct crypto_alg *alg,
90 struct list_head *stack,
91 struct list_head *top,
92 struct list_head *secondary_spawns)
94 struct crypto_spawn *spawn, *n;
96 spawn = list_first_entry_or_null(stack, struct crypto_spawn, list);
97 if (!spawn)
98 return NULL;
100 n = list_next_entry(spawn, list);
102 if (spawn->alg && &n->list != stack && !n->alg)
103 n->alg = (n->list.next == stack) ? alg :
104 &list_next_entry(n, list)->inst->alg;
106 list_move(&spawn->list, secondary_spawns);
108 return &n->list == stack ? top : &n->inst->alg.cra_users;
111 static void crypto_remove_instance(struct crypto_instance *inst,
112 struct list_head *list)
114 struct crypto_template *tmpl = inst->tmpl;
116 if (crypto_is_dead(&inst->alg))
117 return;
119 inst->alg.cra_flags |= CRYPTO_ALG_DEAD;
120 if (hlist_unhashed(&inst->list))
121 return;
123 if (!tmpl || !crypto_tmpl_get(tmpl))
124 return;
126 crypto_notify(CRYPTO_MSG_ALG_UNREGISTER, &inst->alg);
127 list_move(&inst->alg.cra_list, list);
128 hlist_del(&inst->list);
129 inst->alg.cra_destroy = crypto_destroy_instance;
131 BUG_ON(!list_empty(&inst->alg.cra_users));
134 void crypto_remove_spawns(struct crypto_alg *alg, struct list_head *list,
135 struct crypto_alg *nalg)
137 u32 new_type = (nalg ?: alg)->cra_flags;
138 struct crypto_spawn *spawn, *n;
139 LIST_HEAD(secondary_spawns);
140 struct list_head *spawns;
141 LIST_HEAD(stack);
142 LIST_HEAD(top);
144 spawns = &alg->cra_users;
145 list_for_each_entry_safe(spawn, n, spawns, list) {
146 if ((spawn->alg->cra_flags ^ new_type) & spawn->mask)
147 continue;
149 list_move(&spawn->list, &top);
152 spawns = &top;
153 do {
154 while (!list_empty(spawns)) {
155 struct crypto_instance *inst;
157 spawn = list_first_entry(spawns, struct crypto_spawn,
158 list);
159 inst = spawn->inst;
161 BUG_ON(&inst->alg == alg);
163 list_move(&spawn->list, &stack);
165 if (&inst->alg == nalg)
166 break;
168 spawn->alg = NULL;
169 spawns = &inst->alg.cra_users;
172 * We may encounter an unregistered instance here, since
173 * an instance's spawns are set up prior to the instance
174 * being registered. An unregistered instance will have
175 * NULL ->cra_users.next, since ->cra_users isn't
176 * properly initialized until registration. But an
177 * unregistered instance cannot have any users, so treat
178 * it the same as ->cra_users being empty.
180 if (spawns->next == NULL)
181 break;
183 } while ((spawns = crypto_more_spawns(alg, &stack, &top,
184 &secondary_spawns)));
186 list_for_each_entry_safe(spawn, n, &secondary_spawns, list) {
187 if (spawn->alg)
188 list_move(&spawn->list, &spawn->alg->cra_users);
189 else
190 crypto_remove_instance(spawn->inst, list);
193 EXPORT_SYMBOL_GPL(crypto_remove_spawns);
195 static struct crypto_larval *__crypto_register_alg(struct crypto_alg *alg)
197 struct crypto_alg *q;
198 struct crypto_larval *larval;
199 int ret = -EAGAIN;
201 if (crypto_is_dead(alg))
202 goto err;
204 INIT_LIST_HEAD(&alg->cra_users);
206 /* No cheating! */
207 alg->cra_flags &= ~CRYPTO_ALG_TESTED;
209 ret = -EEXIST;
211 list_for_each_entry(q, &crypto_alg_list, cra_list) {
212 if (q == alg)
213 goto err;
215 if (crypto_is_moribund(q))
216 continue;
218 if (crypto_is_larval(q)) {
219 if (!strcmp(alg->cra_driver_name, q->cra_driver_name))
220 goto err;
221 continue;
224 if (!strcmp(q->cra_driver_name, alg->cra_name) ||
225 !strcmp(q->cra_name, alg->cra_driver_name))
226 goto err;
229 larval = crypto_larval_alloc(alg->cra_name,
230 alg->cra_flags | CRYPTO_ALG_TESTED, 0);
231 if (IS_ERR(larval))
232 goto out;
234 ret = -ENOENT;
235 larval->adult = crypto_mod_get(alg);
236 if (!larval->adult)
237 goto free_larval;
239 atomic_set(&larval->alg.cra_refcnt, 1);
240 memcpy(larval->alg.cra_driver_name, alg->cra_driver_name,
241 CRYPTO_MAX_ALG_NAME);
242 larval->alg.cra_priority = alg->cra_priority;
244 list_add(&alg->cra_list, &crypto_alg_list);
245 list_add(&larval->alg.cra_list, &crypto_alg_list);
247 out:
248 return larval;
250 free_larval:
251 kfree(larval);
252 err:
253 larval = ERR_PTR(ret);
254 goto out;
257 void crypto_alg_tested(const char *name, int err)
259 struct crypto_larval *test;
260 struct crypto_alg *alg;
261 struct crypto_alg *q;
262 LIST_HEAD(list);
264 down_write(&crypto_alg_sem);
265 list_for_each_entry(q, &crypto_alg_list, cra_list) {
266 if (crypto_is_moribund(q) || !crypto_is_larval(q))
267 continue;
269 test = (struct crypto_larval *)q;
271 if (!strcmp(q->cra_driver_name, name))
272 goto found;
275 pr_err("alg: Unexpected test result for %s: %d\n", name, err);
276 goto unlock;
278 found:
279 q->cra_flags |= CRYPTO_ALG_DEAD;
280 alg = test->adult;
281 if (err || list_empty(&alg->cra_list))
282 goto complete;
284 alg->cra_flags |= CRYPTO_ALG_TESTED;
286 list_for_each_entry(q, &crypto_alg_list, cra_list) {
287 if (q == alg)
288 continue;
290 if (crypto_is_moribund(q))
291 continue;
293 if (crypto_is_larval(q)) {
294 struct crypto_larval *larval = (void *)q;
297 * Check to see if either our generic name or
298 * specific name can satisfy the name requested
299 * by the larval entry q.
301 if (strcmp(alg->cra_name, q->cra_name) &&
302 strcmp(alg->cra_driver_name, q->cra_name))
303 continue;
305 if (larval->adult)
306 continue;
307 if ((q->cra_flags ^ alg->cra_flags) & larval->mask)
308 continue;
309 if (!crypto_mod_get(alg))
310 continue;
312 larval->adult = alg;
313 continue;
316 if (strcmp(alg->cra_name, q->cra_name))
317 continue;
319 if (strcmp(alg->cra_driver_name, q->cra_driver_name) &&
320 q->cra_priority > alg->cra_priority)
321 continue;
323 crypto_remove_spawns(q, &list, alg);
326 complete:
327 complete_all(&test->completion);
329 unlock:
330 up_write(&crypto_alg_sem);
332 crypto_remove_final(&list);
334 EXPORT_SYMBOL_GPL(crypto_alg_tested);
336 void crypto_remove_final(struct list_head *list)
338 struct crypto_alg *alg;
339 struct crypto_alg *n;
341 list_for_each_entry_safe(alg, n, list, cra_list) {
342 list_del_init(&alg->cra_list);
343 crypto_alg_put(alg);
346 EXPORT_SYMBOL_GPL(crypto_remove_final);
348 static void crypto_wait_for_test(struct crypto_larval *larval)
350 int err;
352 err = crypto_probing_notify(CRYPTO_MSG_ALG_REGISTER, larval->adult);
353 if (err != NOTIFY_STOP) {
354 if (WARN_ON(err != NOTIFY_DONE))
355 goto out;
356 crypto_alg_tested(larval->alg.cra_driver_name, 0);
359 err = wait_for_completion_killable(&larval->completion);
360 WARN_ON(err);
362 out:
363 crypto_larval_kill(&larval->alg);
366 int crypto_register_alg(struct crypto_alg *alg)
368 struct crypto_larval *larval;
369 int err;
371 alg->cra_flags &= ~CRYPTO_ALG_DEAD;
372 err = crypto_check_alg(alg);
373 if (err)
374 return err;
376 down_write(&crypto_alg_sem);
377 larval = __crypto_register_alg(alg);
378 up_write(&crypto_alg_sem);
380 if (IS_ERR(larval))
381 return PTR_ERR(larval);
383 crypto_wait_for_test(larval);
384 return 0;
386 EXPORT_SYMBOL_GPL(crypto_register_alg);
388 static int crypto_remove_alg(struct crypto_alg *alg, struct list_head *list)
390 if (unlikely(list_empty(&alg->cra_list)))
391 return -ENOENT;
393 alg->cra_flags |= CRYPTO_ALG_DEAD;
395 crypto_notify(CRYPTO_MSG_ALG_UNREGISTER, alg);
396 list_del_init(&alg->cra_list);
397 crypto_remove_spawns(alg, list, NULL);
399 return 0;
402 int crypto_unregister_alg(struct crypto_alg *alg)
404 int ret;
405 LIST_HEAD(list);
407 down_write(&crypto_alg_sem);
408 ret = crypto_remove_alg(alg, &list);
409 up_write(&crypto_alg_sem);
411 if (ret)
412 return ret;
414 BUG_ON(atomic_read(&alg->cra_refcnt) != 1);
415 if (alg->cra_destroy)
416 alg->cra_destroy(alg);
418 crypto_remove_final(&list);
419 return 0;
421 EXPORT_SYMBOL_GPL(crypto_unregister_alg);
423 int crypto_register_algs(struct crypto_alg *algs, int count)
425 int i, ret;
427 for (i = 0; i < count; i++) {
428 ret = crypto_register_alg(&algs[i]);
429 if (ret)
430 goto err;
433 return 0;
435 err:
436 for (--i; i >= 0; --i)
437 crypto_unregister_alg(&algs[i]);
439 return ret;
441 EXPORT_SYMBOL_GPL(crypto_register_algs);
443 int crypto_unregister_algs(struct crypto_alg *algs, int count)
445 int i, ret;
447 for (i = 0; i < count; i++) {
448 ret = crypto_unregister_alg(&algs[i]);
449 if (ret)
450 pr_err("Failed to unregister %s %s: %d\n",
451 algs[i].cra_driver_name, algs[i].cra_name, ret);
454 return 0;
456 EXPORT_SYMBOL_GPL(crypto_unregister_algs);
458 int crypto_register_template(struct crypto_template *tmpl)
460 struct crypto_template *q;
461 int err = -EEXIST;
463 down_write(&crypto_alg_sem);
465 crypto_check_module_sig(tmpl->module);
467 list_for_each_entry(q, &crypto_template_list, list) {
468 if (q == tmpl)
469 goto out;
472 list_add(&tmpl->list, &crypto_template_list);
473 crypto_notify(CRYPTO_MSG_TMPL_REGISTER, tmpl);
474 err = 0;
475 out:
476 up_write(&crypto_alg_sem);
477 return err;
479 EXPORT_SYMBOL_GPL(crypto_register_template);
481 void crypto_unregister_template(struct crypto_template *tmpl)
483 struct crypto_instance *inst;
484 struct hlist_node *n;
485 struct hlist_head *list;
486 LIST_HEAD(users);
488 down_write(&crypto_alg_sem);
490 BUG_ON(list_empty(&tmpl->list));
491 list_del_init(&tmpl->list);
493 list = &tmpl->instances;
494 hlist_for_each_entry(inst, list, list) {
495 int err = crypto_remove_alg(&inst->alg, &users);
497 BUG_ON(err);
500 crypto_notify(CRYPTO_MSG_TMPL_UNREGISTER, tmpl);
502 up_write(&crypto_alg_sem);
504 hlist_for_each_entry_safe(inst, n, list, list) {
505 BUG_ON(atomic_read(&inst->alg.cra_refcnt) != 1);
506 crypto_free_instance(inst);
508 crypto_remove_final(&users);
510 EXPORT_SYMBOL_GPL(crypto_unregister_template);
512 static struct crypto_template *__crypto_lookup_template(const char *name)
514 struct crypto_template *q, *tmpl = NULL;
516 down_read(&crypto_alg_sem);
517 list_for_each_entry(q, &crypto_template_list, list) {
518 if (strcmp(q->name, name))
519 continue;
520 if (unlikely(!crypto_tmpl_get(q)))
521 continue;
523 tmpl = q;
524 break;
526 up_read(&crypto_alg_sem);
528 return tmpl;
531 struct crypto_template *crypto_lookup_template(const char *name)
533 return try_then_request_module(__crypto_lookup_template(name),
534 "crypto-%s", name);
536 EXPORT_SYMBOL_GPL(crypto_lookup_template);
538 int crypto_register_instance(struct crypto_template *tmpl,
539 struct crypto_instance *inst)
541 struct crypto_larval *larval;
542 int err;
544 err = crypto_check_alg(&inst->alg);
545 if (err)
546 return err;
548 inst->alg.cra_module = tmpl->module;
549 inst->alg.cra_flags |= CRYPTO_ALG_INSTANCE;
551 if (unlikely(!crypto_mod_get(&inst->alg)))
552 return -EAGAIN;
554 down_write(&crypto_alg_sem);
556 larval = __crypto_register_alg(&inst->alg);
557 if (IS_ERR(larval))
558 goto unlock;
560 hlist_add_head(&inst->list, &tmpl->instances);
561 inst->tmpl = tmpl;
563 unlock:
564 up_write(&crypto_alg_sem);
566 err = PTR_ERR(larval);
567 if (IS_ERR(larval))
568 goto err;
570 crypto_wait_for_test(larval);
572 /* Remove instance if test failed */
573 if (!(inst->alg.cra_flags & CRYPTO_ALG_TESTED))
574 crypto_unregister_instance(inst);
575 err = 0;
577 err:
578 crypto_mod_put(&inst->alg);
579 return err;
581 EXPORT_SYMBOL_GPL(crypto_register_instance);
583 int crypto_unregister_instance(struct crypto_instance *inst)
585 LIST_HEAD(list);
587 down_write(&crypto_alg_sem);
589 crypto_remove_spawns(&inst->alg, &list, NULL);
590 crypto_remove_instance(inst, &list);
592 up_write(&crypto_alg_sem);
594 crypto_remove_final(&list);
596 return 0;
598 EXPORT_SYMBOL_GPL(crypto_unregister_instance);
600 int crypto_init_spawn(struct crypto_spawn *spawn, struct crypto_alg *alg,
601 struct crypto_instance *inst, u32 mask)
603 int err = -EAGAIN;
605 spawn->inst = inst;
606 spawn->mask = mask;
608 down_write(&crypto_alg_sem);
609 if (!crypto_is_moribund(alg)) {
610 list_add(&spawn->list, &alg->cra_users);
611 spawn->alg = alg;
612 err = 0;
614 up_write(&crypto_alg_sem);
616 return err;
618 EXPORT_SYMBOL_GPL(crypto_init_spawn);
620 int crypto_init_spawn2(struct crypto_spawn *spawn, struct crypto_alg *alg,
621 struct crypto_instance *inst,
622 const struct crypto_type *frontend)
624 int err = -EINVAL;
626 if ((alg->cra_flags ^ frontend->type) & frontend->maskset)
627 goto out;
629 spawn->frontend = frontend;
630 err = crypto_init_spawn(spawn, alg, inst, frontend->maskset);
632 out:
633 return err;
635 EXPORT_SYMBOL_GPL(crypto_init_spawn2);
637 int crypto_grab_spawn(struct crypto_spawn *spawn, const char *name,
638 u32 type, u32 mask)
640 struct crypto_alg *alg;
641 int err;
643 alg = crypto_find_alg(name, spawn->frontend, type, mask);
644 if (IS_ERR(alg))
645 return PTR_ERR(alg);
647 err = crypto_init_spawn(spawn, alg, spawn->inst, mask);
648 crypto_mod_put(alg);
649 return err;
651 EXPORT_SYMBOL_GPL(crypto_grab_spawn);
653 void crypto_drop_spawn(struct crypto_spawn *spawn)
655 down_write(&crypto_alg_sem);
656 if (spawn->alg)
657 list_del(&spawn->list);
658 up_write(&crypto_alg_sem);
660 EXPORT_SYMBOL_GPL(crypto_drop_spawn);
662 static struct crypto_alg *crypto_spawn_alg(struct crypto_spawn *spawn)
664 struct crypto_alg *alg;
666 down_read(&crypto_alg_sem);
667 alg = spawn->alg;
668 if (alg && !crypto_mod_get(alg)) {
669 alg->cra_flags |= CRYPTO_ALG_DYING;
670 alg = NULL;
672 up_read(&crypto_alg_sem);
674 return alg ?: ERR_PTR(-EAGAIN);
677 struct crypto_tfm *crypto_spawn_tfm(struct crypto_spawn *spawn, u32 type,
678 u32 mask)
680 struct crypto_alg *alg;
681 struct crypto_tfm *tfm;
683 alg = crypto_spawn_alg(spawn);
684 if (IS_ERR(alg))
685 return ERR_CAST(alg);
687 tfm = ERR_PTR(-EINVAL);
688 if (unlikely((alg->cra_flags ^ type) & mask))
689 goto out_put_alg;
691 tfm = __crypto_alloc_tfm(alg, type, mask);
692 if (IS_ERR(tfm))
693 goto out_put_alg;
695 return tfm;
697 out_put_alg:
698 crypto_mod_put(alg);
699 return tfm;
701 EXPORT_SYMBOL_GPL(crypto_spawn_tfm);
703 void *crypto_spawn_tfm2(struct crypto_spawn *spawn)
705 struct crypto_alg *alg;
706 struct crypto_tfm *tfm;
708 alg = crypto_spawn_alg(spawn);
709 if (IS_ERR(alg))
710 return ERR_CAST(alg);
712 tfm = crypto_create_tfm(alg, spawn->frontend);
713 if (IS_ERR(tfm))
714 goto out_put_alg;
716 return tfm;
718 out_put_alg:
719 crypto_mod_put(alg);
720 return tfm;
722 EXPORT_SYMBOL_GPL(crypto_spawn_tfm2);
724 int crypto_register_notifier(struct notifier_block *nb)
726 return blocking_notifier_chain_register(&crypto_chain, nb);
728 EXPORT_SYMBOL_GPL(crypto_register_notifier);
730 int crypto_unregister_notifier(struct notifier_block *nb)
732 return blocking_notifier_chain_unregister(&crypto_chain, nb);
734 EXPORT_SYMBOL_GPL(crypto_unregister_notifier);
736 struct crypto_attr_type *crypto_get_attr_type(struct rtattr **tb)
738 struct rtattr *rta = tb[0];
739 struct crypto_attr_type *algt;
741 if (!rta)
742 return ERR_PTR(-ENOENT);
743 if (RTA_PAYLOAD(rta) < sizeof(*algt))
744 return ERR_PTR(-EINVAL);
745 if (rta->rta_type != CRYPTOA_TYPE)
746 return ERR_PTR(-EINVAL);
748 algt = RTA_DATA(rta);
750 return algt;
752 EXPORT_SYMBOL_GPL(crypto_get_attr_type);
754 int crypto_check_attr_type(struct rtattr **tb, u32 type)
756 struct crypto_attr_type *algt;
758 algt = crypto_get_attr_type(tb);
759 if (IS_ERR(algt))
760 return PTR_ERR(algt);
762 if ((algt->type ^ type) & algt->mask)
763 return -EINVAL;
765 return 0;
767 EXPORT_SYMBOL_GPL(crypto_check_attr_type);
769 const char *crypto_attr_alg_name(struct rtattr *rta)
771 struct crypto_attr_alg *alga;
773 if (!rta)
774 return ERR_PTR(-ENOENT);
775 if (RTA_PAYLOAD(rta) < sizeof(*alga))
776 return ERR_PTR(-EINVAL);
777 if (rta->rta_type != CRYPTOA_ALG)
778 return ERR_PTR(-EINVAL);
780 alga = RTA_DATA(rta);
781 alga->name[CRYPTO_MAX_ALG_NAME - 1] = 0;
783 return alga->name;
785 EXPORT_SYMBOL_GPL(crypto_attr_alg_name);
787 struct crypto_alg *crypto_attr_alg2(struct rtattr *rta,
788 const struct crypto_type *frontend,
789 u32 type, u32 mask)
791 const char *name;
793 name = crypto_attr_alg_name(rta);
794 if (IS_ERR(name))
795 return ERR_CAST(name);
797 return crypto_find_alg(name, frontend, type, mask);
799 EXPORT_SYMBOL_GPL(crypto_attr_alg2);
801 int crypto_attr_u32(struct rtattr *rta, u32 *num)
803 struct crypto_attr_u32 *nu32;
805 if (!rta)
806 return -ENOENT;
807 if (RTA_PAYLOAD(rta) < sizeof(*nu32))
808 return -EINVAL;
809 if (rta->rta_type != CRYPTOA_U32)
810 return -EINVAL;
812 nu32 = RTA_DATA(rta);
813 *num = nu32->num;
815 return 0;
817 EXPORT_SYMBOL_GPL(crypto_attr_u32);
819 int crypto_inst_setname(struct crypto_instance *inst, const char *name,
820 struct crypto_alg *alg)
822 if (snprintf(inst->alg.cra_name, CRYPTO_MAX_ALG_NAME, "%s(%s)", name,
823 alg->cra_name) >= CRYPTO_MAX_ALG_NAME)
824 return -ENAMETOOLONG;
826 if (snprintf(inst->alg.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s(%s)",
827 name, alg->cra_driver_name) >= CRYPTO_MAX_ALG_NAME)
828 return -ENAMETOOLONG;
830 return 0;
832 EXPORT_SYMBOL_GPL(crypto_inst_setname);
834 void *crypto_alloc_instance2(const char *name, struct crypto_alg *alg,
835 unsigned int head)
837 struct crypto_instance *inst;
838 char *p;
839 int err;
841 p = kzalloc(head + sizeof(*inst) + sizeof(struct crypto_spawn),
842 GFP_KERNEL);
843 if (!p)
844 return ERR_PTR(-ENOMEM);
846 inst = (void *)(p + head);
848 err = crypto_inst_setname(inst, name, alg);
849 if (err)
850 goto err_free_inst;
852 return p;
854 err_free_inst:
855 kfree(p);
856 return ERR_PTR(err);
858 EXPORT_SYMBOL_GPL(crypto_alloc_instance2);
860 struct crypto_instance *crypto_alloc_instance(const char *name,
861 struct crypto_alg *alg)
863 struct crypto_instance *inst;
864 struct crypto_spawn *spawn;
865 int err;
867 inst = crypto_alloc_instance2(name, alg, 0);
868 if (IS_ERR(inst))
869 goto out;
871 spawn = crypto_instance_ctx(inst);
872 err = crypto_init_spawn(spawn, alg, inst,
873 CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
875 if (err)
876 goto err_free_inst;
878 return inst;
880 err_free_inst:
881 kfree(inst);
882 inst = ERR_PTR(err);
884 out:
885 return inst;
887 EXPORT_SYMBOL_GPL(crypto_alloc_instance);
889 void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen)
891 INIT_LIST_HEAD(&queue->list);
892 queue->backlog = &queue->list;
893 queue->qlen = 0;
894 queue->max_qlen = max_qlen;
896 EXPORT_SYMBOL_GPL(crypto_init_queue);
898 int crypto_enqueue_request(struct crypto_queue *queue,
899 struct crypto_async_request *request)
901 int err = -EINPROGRESS;
903 if (unlikely(queue->qlen >= queue->max_qlen)) {
904 err = -EBUSY;
905 if (!(request->flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
906 goto out;
907 if (queue->backlog == &queue->list)
908 queue->backlog = &request->list;
911 queue->qlen++;
912 list_add_tail(&request->list, &queue->list);
914 out:
915 return err;
917 EXPORT_SYMBOL_GPL(crypto_enqueue_request);
919 struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue)
921 struct list_head *request;
923 if (unlikely(!queue->qlen))
924 return NULL;
926 queue->qlen--;
928 if (queue->backlog != &queue->list)
929 queue->backlog = queue->backlog->next;
931 request = queue->list.next;
932 list_del(request);
934 return list_entry(request, struct crypto_async_request, list);
936 EXPORT_SYMBOL_GPL(crypto_dequeue_request);
938 int crypto_tfm_in_queue(struct crypto_queue *queue, struct crypto_tfm *tfm)
940 struct crypto_async_request *req;
942 list_for_each_entry(req, &queue->list, list) {
943 if (req->tfm == tfm)
944 return 1;
947 return 0;
949 EXPORT_SYMBOL_GPL(crypto_tfm_in_queue);
951 static inline void crypto_inc_byte(u8 *a, unsigned int size)
953 u8 *b = (a + size);
954 u8 c;
956 for (; size; size--) {
957 c = *--b + 1;
958 *b = c;
959 if (c)
960 break;
964 void crypto_inc(u8 *a, unsigned int size)
966 __be32 *b = (__be32 *)(a + size);
967 u32 c;
969 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) ||
970 IS_ALIGNED((unsigned long)b, __alignof__(*b)))
971 for (; size >= 4; size -= 4) {
972 c = be32_to_cpu(*--b) + 1;
973 *b = cpu_to_be32(c);
974 if (likely(c))
975 return;
978 crypto_inc_byte(a, size);
980 EXPORT_SYMBOL_GPL(crypto_inc);
982 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len)
984 int relalign = 0;
986 if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) {
987 int size = sizeof(unsigned long);
988 int d = (((unsigned long)dst ^ (unsigned long)src1) |
989 ((unsigned long)dst ^ (unsigned long)src2)) &
990 (size - 1);
992 relalign = d ? 1 << __ffs(d) : size;
995 * If we care about alignment, process as many bytes as
996 * needed to advance dst and src to values whose alignments
997 * equal their relative alignment. This will allow us to
998 * process the remainder of the input using optimal strides.
1000 while (((unsigned long)dst & (relalign - 1)) && len > 0) {
1001 *dst++ = *src1++ ^ *src2++;
1002 len--;
1006 while (IS_ENABLED(CONFIG_64BIT) && len >= 8 && !(relalign & 7)) {
1007 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2;
1008 dst += 8;
1009 src1 += 8;
1010 src2 += 8;
1011 len -= 8;
1014 while (len >= 4 && !(relalign & 3)) {
1015 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2;
1016 dst += 4;
1017 src1 += 4;
1018 src2 += 4;
1019 len -= 4;
1022 while (len >= 2 && !(relalign & 1)) {
1023 *(u16 *)dst = *(u16 *)src1 ^ *(u16 *)src2;
1024 dst += 2;
1025 src1 += 2;
1026 src2 += 2;
1027 len -= 2;
1030 while (len--)
1031 *dst++ = *src1++ ^ *src2++;
1033 EXPORT_SYMBOL_GPL(__crypto_xor);
1035 unsigned int crypto_alg_extsize(struct crypto_alg *alg)
1037 return alg->cra_ctxsize +
1038 (alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1));
1040 EXPORT_SYMBOL_GPL(crypto_alg_extsize);
1042 int crypto_type_has_alg(const char *name, const struct crypto_type *frontend,
1043 u32 type, u32 mask)
1045 int ret = 0;
1046 struct crypto_alg *alg = crypto_find_alg(name, frontend, type, mask);
1048 if (!IS_ERR(alg)) {
1049 crypto_mod_put(alg);
1050 ret = 1;
1053 return ret;
1055 EXPORT_SYMBOL_GPL(crypto_type_has_alg);
1057 static int __init crypto_algapi_init(void)
1059 crypto_init_proc();
1060 return 0;
1063 static void __exit crypto_algapi_exit(void)
1065 crypto_exit_proc();
1068 module_init(crypto_algapi_init);
1069 module_exit(crypto_algapi_exit);
1071 MODULE_LICENSE("GPL");
1072 MODULE_DESCRIPTION("Cryptographic algorithms API");