4 * Glue code for the SHA256 Secure Hash Algorithm assembler
5 * implementation using supplemental SSE3 / AVX / AVX2 instructions.
7 * This file is based on sha256_generic.c
9 * Copyright (C) 2013 Intel Corporation.
12 * Tim Chen <tim.c.chen@linux.intel.com>
14 * This program is free software; you can redistribute it and/or modify it
15 * under the terms of the GNU General Public License as published by the Free
16 * Software Foundation; either version 2 of the License, or (at your option)
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
20 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
21 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
22 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
23 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
24 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
25 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
32 #include <crypto/internal/hash.h>
33 #include <linux/init.h>
34 #include <linux/module.h>
36 #include <linux/cryptohash.h>
37 #include <linux/types.h>
38 #include <crypto/sha.h>
39 #include <crypto/sha256_base.h>
42 #include <asm/xsave.h>
43 #include <linux/string.h>
45 asmlinkage
void sha256_transform_ssse3(u32
*digest
, const char *data
,
48 asmlinkage
void sha256_transform_avx(u32
*digest
, const char *data
,
52 asmlinkage
void sha256_transform_rorx(u32
*digest
, const char *data
,
56 static void (*sha256_transform_asm
)(u32
*, const char *, u64
);
58 static int sha256_ssse3_update(struct shash_desc
*desc
, const u8
*data
,
61 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
63 if (!irq_fpu_usable() ||
64 (sctx
->count
% SHA256_BLOCK_SIZE
) + len
< SHA256_BLOCK_SIZE
)
65 return crypto_sha256_update(desc
, data
, len
);
67 /* make sure casting to sha256_block_fn() is safe */
68 BUILD_BUG_ON(offsetof(struct sha256_state
, state
) != 0);
71 sha256_base_do_update(desc
, data
, len
,
72 (sha256_block_fn
*)sha256_transform_asm
);
78 static int sha256_ssse3_finup(struct shash_desc
*desc
, const u8
*data
,
79 unsigned int len
, u8
*out
)
81 if (!irq_fpu_usable())
82 return crypto_sha256_finup(desc
, data
, len
, out
);
86 sha256_base_do_update(desc
, data
, len
,
87 (sha256_block_fn
*)sha256_transform_asm
);
88 sha256_base_do_finalize(desc
, (sha256_block_fn
*)sha256_transform_asm
);
91 return sha256_base_finish(desc
, out
);
94 /* Add padding and return the message digest. */
95 static int sha256_ssse3_final(struct shash_desc
*desc
, u8
*out
)
97 return sha256_ssse3_finup(desc
, NULL
, 0, out
);
100 static struct shash_alg algs
[] = { {
101 .digestsize
= SHA256_DIGEST_SIZE
,
102 .init
= sha256_base_init
,
103 .update
= sha256_ssse3_update
,
104 .final
= sha256_ssse3_final
,
105 .finup
= sha256_ssse3_finup
,
106 .descsize
= sizeof(struct sha256_state
),
108 .cra_name
= "sha256",
109 .cra_driver_name
= "sha256-ssse3",
111 .cra_flags
= CRYPTO_ALG_TYPE_SHASH
,
112 .cra_blocksize
= SHA256_BLOCK_SIZE
,
113 .cra_module
= THIS_MODULE
,
116 .digestsize
= SHA224_DIGEST_SIZE
,
117 .init
= sha224_base_init
,
118 .update
= sha256_ssse3_update
,
119 .final
= sha256_ssse3_final
,
120 .finup
= sha256_ssse3_finup
,
121 .descsize
= sizeof(struct sha256_state
),
123 .cra_name
= "sha224",
124 .cra_driver_name
= "sha224-ssse3",
126 .cra_flags
= CRYPTO_ALG_TYPE_SHASH
,
127 .cra_blocksize
= SHA224_BLOCK_SIZE
,
128 .cra_module
= THIS_MODULE
,
133 static bool __init
avx_usable(void)
137 if (!cpu_has_avx
|| !cpu_has_osxsave
)
140 xcr0
= xgetbv(XCR_XFEATURE_ENABLED_MASK
);
141 if ((xcr0
& (XSTATE_SSE
| XSTATE_YMM
)) != (XSTATE_SSE
| XSTATE_YMM
)) {
142 pr_info("AVX detected but unusable.\n");
151 static int __init
sha256_ssse3_mod_init(void)
153 /* test for SSSE3 first */
155 sha256_transform_asm
= sha256_transform_ssse3
;
158 /* allow AVX to override SSSE3, it's a little faster */
160 #ifdef CONFIG_AS_AVX2
161 if (boot_cpu_has(X86_FEATURE_AVX2
) && boot_cpu_has(X86_FEATURE_BMI2
))
162 sha256_transform_asm
= sha256_transform_rorx
;
165 sha256_transform_asm
= sha256_transform_avx
;
169 if (sha256_transform_asm
) {
171 if (sha256_transform_asm
== sha256_transform_avx
)
172 pr_info("Using AVX optimized SHA-256 implementation\n");
173 #ifdef CONFIG_AS_AVX2
174 else if (sha256_transform_asm
== sha256_transform_rorx
)
175 pr_info("Using AVX2 optimized SHA-256 implementation\n");
179 pr_info("Using SSSE3 optimized SHA-256 implementation\n");
180 return crypto_register_shashes(algs
, ARRAY_SIZE(algs
));
182 pr_info("Neither AVX nor SSSE3 is available/usable.\n");
187 static void __exit
sha256_ssse3_mod_fini(void)
189 crypto_unregister_shashes(algs
, ARRAY_SIZE(algs
));
192 module_init(sha256_ssse3_mod_init
);
193 module_exit(sha256_ssse3_mod_fini
);
195 MODULE_LICENSE("GPL");
196 MODULE_DESCRIPTION("SHA256 Secure Hash Algorithm, Supplemental SSE3 accelerated");
198 MODULE_ALIAS_CRYPTO("sha256");
199 MODULE_ALIAS_CRYPTO("sha224");