1 // SPDX-License-Identifier: GPL-2.0-or-later
5 * SHA-256, as specified in
6 * http://csrc.nist.gov/groups/STM/cavp/documents/shs/sha256-384-512.pdf
8 * SHA-256 code by Jean-Luc Cooke <jlcooke@certainkey.com>.
10 * Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
11 * Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
12 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
13 * SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com>
15 #include <crypto/internal/hash.h>
16 #include <linux/init.h>
17 #include <linux/module.h>
19 #include <linux/types.h>
20 #include <crypto/sha.h>
21 #include <crypto/sha256_base.h>
22 #include <asm/byteorder.h>
23 #include <asm/unaligned.h>
25 const u8 sha224_zero_message_hash
[SHA224_DIGEST_SIZE
] = {
26 0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47,
27 0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2,
28 0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4,
31 EXPORT_SYMBOL_GPL(sha224_zero_message_hash
);
33 const u8 sha256_zero_message_hash
[SHA256_DIGEST_SIZE
] = {
34 0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14,
35 0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24,
36 0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c,
37 0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55
39 EXPORT_SYMBOL_GPL(sha256_zero_message_hash
);
41 static inline u32
Ch(u32 x
, u32 y
, u32 z
)
43 return z
^ (x
& (y
^ z
));
46 static inline u32
Maj(u32 x
, u32 y
, u32 z
)
48 return (x
& y
) | (z
& (x
| y
));
51 #define e0(x) (ror32(x, 2) ^ ror32(x,13) ^ ror32(x,22))
52 #define e1(x) (ror32(x, 6) ^ ror32(x,11) ^ ror32(x,25))
53 #define s0(x) (ror32(x, 7) ^ ror32(x,18) ^ (x >> 3))
54 #define s1(x) (ror32(x,17) ^ ror32(x,19) ^ (x >> 10))
56 static inline void LOAD_OP(int I
, u32
*W
, const u8
*input
)
58 W
[I
] = get_unaligned_be32((__u32
*)input
+ I
);
61 static inline void BLEND_OP(int I
, u32
*W
)
63 W
[I
] = s1(W
[I
-2]) + W
[I
-7] + s0(W
[I
-15]) + W
[I
-16];
66 static void sha256_transform(u32
*state
, const u8
*input
)
68 u32 a
, b
, c
, d
, e
, f
, g
, h
, t1
, t2
;
73 for (i
= 0; i
< 16; i
++)
77 for (i
= 16; i
< 64; i
++)
80 /* load the state into our registers */
81 a
=state
[0]; b
=state
[1]; c
=state
[2]; d
=state
[3];
82 e
=state
[4]; f
=state
[5]; g
=state
[6]; h
=state
[7];
85 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x428a2f98 + W
[ 0];
86 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
87 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x71374491 + W
[ 1];
88 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
89 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xb5c0fbcf + W
[ 2];
90 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
91 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xe9b5dba5 + W
[ 3];
92 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
93 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x3956c25b + W
[ 4];
94 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
95 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x59f111f1 + W
[ 5];
96 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
97 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x923f82a4 + W
[ 6];
98 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
99 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xab1c5ed5 + W
[ 7];
100 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
102 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xd807aa98 + W
[ 8];
103 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
104 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x12835b01 + W
[ 9];
105 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
106 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x243185be + W
[10];
107 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
108 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x550c7dc3 + W
[11];
109 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
110 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x72be5d74 + W
[12];
111 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
112 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x80deb1fe + W
[13];
113 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
114 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x9bdc06a7 + W
[14];
115 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
116 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xc19bf174 + W
[15];
117 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
119 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xe49b69c1 + W
[16];
120 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
121 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xefbe4786 + W
[17];
122 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
123 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x0fc19dc6 + W
[18];
124 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
125 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x240ca1cc + W
[19];
126 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
127 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x2de92c6f + W
[20];
128 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
129 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x4a7484aa + W
[21];
130 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
131 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x5cb0a9dc + W
[22];
132 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
133 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x76f988da + W
[23];
134 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
136 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x983e5152 + W
[24];
137 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
138 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xa831c66d + W
[25];
139 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
140 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xb00327c8 + W
[26];
141 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
142 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xbf597fc7 + W
[27];
143 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
144 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0xc6e00bf3 + W
[28];
145 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
146 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xd5a79147 + W
[29];
147 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
148 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x06ca6351 + W
[30];
149 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
150 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x14292967 + W
[31];
151 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
153 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x27b70a85 + W
[32];
154 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
155 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x2e1b2138 + W
[33];
156 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
157 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x4d2c6dfc + W
[34];
158 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
159 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x53380d13 + W
[35];
160 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
161 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x650a7354 + W
[36];
162 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
163 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x766a0abb + W
[37];
164 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
165 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x81c2c92e + W
[38];
166 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
167 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x92722c85 + W
[39];
168 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
170 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xa2bfe8a1 + W
[40];
171 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
172 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xa81a664b + W
[41];
173 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
174 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xc24b8b70 + W
[42];
175 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
176 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xc76c51a3 + W
[43];
177 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
178 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0xd192e819 + W
[44];
179 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
180 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xd6990624 + W
[45];
181 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
182 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0xf40e3585 + W
[46];
183 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
184 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x106aa070 + W
[47];
185 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
187 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x19a4c116 + W
[48];
188 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
189 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x1e376c08 + W
[49];
190 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
191 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x2748774c + W
[50];
192 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
193 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x34b0bcb5 + W
[51];
194 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
195 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x391c0cb3 + W
[52];
196 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
197 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x4ed8aa4a + W
[53];
198 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
199 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x5b9cca4f + W
[54];
200 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
201 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x682e6ff3 + W
[55];
202 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
204 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x748f82ee + W
[56];
205 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
206 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x78a5636f + W
[57];
207 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
208 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x84c87814 + W
[58];
209 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
210 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x8cc70208 + W
[59];
211 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
212 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x90befffa + W
[60];
213 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
214 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xa4506ceb + W
[61];
215 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
216 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0xbef9a3f7 + W
[62];
217 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
218 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xc67178f2 + W
[63];
219 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
221 state
[0] += a
; state
[1] += b
; state
[2] += c
; state
[3] += d
;
222 state
[4] += e
; state
[5] += f
; state
[6] += g
; state
[7] += h
;
224 /* clear any sensitive info... */
225 a
= b
= c
= d
= e
= f
= g
= h
= t1
= t2
= 0;
226 memzero_explicit(W
, 64 * sizeof(u32
));
229 static void sha256_generic_block_fn(struct sha256_state
*sst
, u8
const *src
,
233 sha256_transform(sst
->state
, src
);
234 src
+= SHA256_BLOCK_SIZE
;
238 int crypto_sha256_update(struct shash_desc
*desc
, const u8
*data
,
241 return sha256_base_do_update(desc
, data
, len
, sha256_generic_block_fn
);
243 EXPORT_SYMBOL(crypto_sha256_update
);
245 static int sha256_final(struct shash_desc
*desc
, u8
*out
)
247 sha256_base_do_finalize(desc
, sha256_generic_block_fn
);
248 return sha256_base_finish(desc
, out
);
251 int crypto_sha256_finup(struct shash_desc
*desc
, const u8
*data
,
252 unsigned int len
, u8
*hash
)
254 sha256_base_do_update(desc
, data
, len
, sha256_generic_block_fn
);
255 return sha256_final(desc
, hash
);
257 EXPORT_SYMBOL(crypto_sha256_finup
);
259 static struct shash_alg sha256_algs
[2] = { {
260 .digestsize
= SHA256_DIGEST_SIZE
,
261 .init
= sha256_base_init
,
262 .update
= crypto_sha256_update
,
263 .final
= sha256_final
,
264 .finup
= crypto_sha256_finup
,
265 .descsize
= sizeof(struct sha256_state
),
267 .cra_name
= "sha256",
268 .cra_driver_name
= "sha256-generic",
270 .cra_blocksize
= SHA256_BLOCK_SIZE
,
271 .cra_module
= THIS_MODULE
,
274 .digestsize
= SHA224_DIGEST_SIZE
,
275 .init
= sha224_base_init
,
276 .update
= crypto_sha256_update
,
277 .final
= sha256_final
,
278 .finup
= crypto_sha256_finup
,
279 .descsize
= sizeof(struct sha256_state
),
281 .cra_name
= "sha224",
282 .cra_driver_name
= "sha224-generic",
284 .cra_blocksize
= SHA224_BLOCK_SIZE
,
285 .cra_module
= THIS_MODULE
,
289 static int __init
sha256_generic_mod_init(void)
291 return crypto_register_shashes(sha256_algs
, ARRAY_SIZE(sha256_algs
));
294 static void __exit
sha256_generic_mod_fini(void)
296 crypto_unregister_shashes(sha256_algs
, ARRAY_SIZE(sha256_algs
));
299 subsys_initcall(sha256_generic_mod_init
);
300 module_exit(sha256_generic_mod_fini
);
302 MODULE_LICENSE("GPL");
303 MODULE_DESCRIPTION("SHA-224 and SHA-256 Secure Hash Algorithm");
305 MODULE_ALIAS_CRYPTO("sha224");
306 MODULE_ALIAS_CRYPTO("sha224-generic");
307 MODULE_ALIAS_CRYPTO("sha256");
308 MODULE_ALIAS_CRYPTO("sha256-generic");