4 * SHA-256, as specified in
5 * http://csrc.nist.gov/groups/STM/cavp/documents/shs/sha256-384-512.pdf
7 * SHA-256 code by Jean-Luc Cooke <jlcooke@certainkey.com>.
9 * Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
10 * Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
11 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
12 * SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com>
14 * This program is free software; you can redistribute it and/or modify it
15 * under the terms of the GNU General Public License as published by the Free
16 * Software Foundation; either version 2 of the License, or (at your option)
20 #include <crypto/internal/hash.h>
21 #include <linux/init.h>
22 #include <linux/module.h>
24 #include <linux/types.h>
25 #include <crypto/sha.h>
26 #include <asm/byteorder.h>
27 #include <asm/unaligned.h>
29 static inline u32
Ch(u32 x
, u32 y
, u32 z
)
31 return z
^ (x
& (y
^ z
));
34 static inline u32
Maj(u32 x
, u32 y
, u32 z
)
36 return (x
& y
) | (z
& (x
| y
));
39 #define e0(x) (ror32(x, 2) ^ ror32(x,13) ^ ror32(x,22))
40 #define e1(x) (ror32(x, 6) ^ ror32(x,11) ^ ror32(x,25))
41 #define s0(x) (ror32(x, 7) ^ ror32(x,18) ^ (x >> 3))
42 #define s1(x) (ror32(x,17) ^ ror32(x,19) ^ (x >> 10))
44 static inline void LOAD_OP(int I
, u32
*W
, const u8
*input
)
46 W
[I
] = get_unaligned_be32((__u32
*)input
+ I
);
49 static inline void BLEND_OP(int I
, u32
*W
)
51 W
[I
] = s1(W
[I
-2]) + W
[I
-7] + s0(W
[I
-15]) + W
[I
-16];
54 static void sha256_transform(u32
*state
, const u8
*input
)
56 u32 a
, b
, c
, d
, e
, f
, g
, h
, t1
, t2
;
61 for (i
= 0; i
< 16; i
++)
65 for (i
= 16; i
< 64; i
++)
68 /* load the state into our registers */
69 a
=state
[0]; b
=state
[1]; c
=state
[2]; d
=state
[3];
70 e
=state
[4]; f
=state
[5]; g
=state
[6]; h
=state
[7];
73 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x428a2f98 + W
[ 0];
74 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
75 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x71374491 + W
[ 1];
76 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
77 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xb5c0fbcf + W
[ 2];
78 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
79 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xe9b5dba5 + W
[ 3];
80 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
81 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x3956c25b + W
[ 4];
82 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
83 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x59f111f1 + W
[ 5];
84 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
85 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x923f82a4 + W
[ 6];
86 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
87 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xab1c5ed5 + W
[ 7];
88 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
90 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xd807aa98 + W
[ 8];
91 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
92 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x12835b01 + W
[ 9];
93 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
94 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x243185be + W
[10];
95 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
96 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x550c7dc3 + W
[11];
97 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
98 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x72be5d74 + W
[12];
99 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
100 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x80deb1fe + W
[13];
101 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
102 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x9bdc06a7 + W
[14];
103 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
104 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xc19bf174 + W
[15];
105 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
107 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xe49b69c1 + W
[16];
108 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
109 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xefbe4786 + W
[17];
110 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
111 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x0fc19dc6 + W
[18];
112 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
113 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x240ca1cc + W
[19];
114 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
115 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x2de92c6f + W
[20];
116 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
117 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x4a7484aa + W
[21];
118 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
119 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x5cb0a9dc + W
[22];
120 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
121 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x76f988da + W
[23];
122 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
124 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x983e5152 + W
[24];
125 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
126 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xa831c66d + W
[25];
127 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
128 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xb00327c8 + W
[26];
129 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
130 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xbf597fc7 + W
[27];
131 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
132 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0xc6e00bf3 + W
[28];
133 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
134 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xd5a79147 + W
[29];
135 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
136 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x06ca6351 + W
[30];
137 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
138 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x14292967 + W
[31];
139 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
141 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x27b70a85 + W
[32];
142 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
143 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x2e1b2138 + W
[33];
144 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
145 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x4d2c6dfc + W
[34];
146 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
147 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x53380d13 + W
[35];
148 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
149 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x650a7354 + W
[36];
150 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
151 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x766a0abb + W
[37];
152 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
153 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x81c2c92e + W
[38];
154 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
155 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x92722c85 + W
[39];
156 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
158 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0xa2bfe8a1 + W
[40];
159 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
160 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0xa81a664b + W
[41];
161 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
162 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0xc24b8b70 + W
[42];
163 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
164 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0xc76c51a3 + W
[43];
165 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
166 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0xd192e819 + W
[44];
167 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
168 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xd6990624 + W
[45];
169 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
170 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0xf40e3585 + W
[46];
171 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
172 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x106aa070 + W
[47];
173 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
175 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x19a4c116 + W
[48];
176 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
177 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x1e376c08 + W
[49];
178 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
179 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x2748774c + W
[50];
180 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
181 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x34b0bcb5 + W
[51];
182 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
183 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x391c0cb3 + W
[52];
184 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
185 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0x4ed8aa4a + W
[53];
186 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
187 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0x5b9cca4f + W
[54];
188 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
189 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0x682e6ff3 + W
[55];
190 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
192 t1
= h
+ e1(e
) + Ch(e
,f
,g
) + 0x748f82ee + W
[56];
193 t2
= e0(a
) + Maj(a
,b
,c
); d
+=t1
; h
=t1
+t2
;
194 t1
= g
+ e1(d
) + Ch(d
,e
,f
) + 0x78a5636f + W
[57];
195 t2
= e0(h
) + Maj(h
,a
,b
); c
+=t1
; g
=t1
+t2
;
196 t1
= f
+ e1(c
) + Ch(c
,d
,e
) + 0x84c87814 + W
[58];
197 t2
= e0(g
) + Maj(g
,h
,a
); b
+=t1
; f
=t1
+t2
;
198 t1
= e
+ e1(b
) + Ch(b
,c
,d
) + 0x8cc70208 + W
[59];
199 t2
= e0(f
) + Maj(f
,g
,h
); a
+=t1
; e
=t1
+t2
;
200 t1
= d
+ e1(a
) + Ch(a
,b
,c
) + 0x90befffa + W
[60];
201 t2
= e0(e
) + Maj(e
,f
,g
); h
+=t1
; d
=t1
+t2
;
202 t1
= c
+ e1(h
) + Ch(h
,a
,b
) + 0xa4506ceb + W
[61];
203 t2
= e0(d
) + Maj(d
,e
,f
); g
+=t1
; c
=t1
+t2
;
204 t1
= b
+ e1(g
) + Ch(g
,h
,a
) + 0xbef9a3f7 + W
[62];
205 t2
= e0(c
) + Maj(c
,d
,e
); f
+=t1
; b
=t1
+t2
;
206 t1
= a
+ e1(f
) + Ch(f
,g
,h
) + 0xc67178f2 + W
[63];
207 t2
= e0(b
) + Maj(b
,c
,d
); e
+=t1
; a
=t1
+t2
;
209 state
[0] += a
; state
[1] += b
; state
[2] += c
; state
[3] += d
;
210 state
[4] += e
; state
[5] += f
; state
[6] += g
; state
[7] += h
;
212 /* clear any sensitive info... */
213 a
= b
= c
= d
= e
= f
= g
= h
= t1
= t2
= 0;
214 memzero_explicit(W
, 64 * sizeof(u32
));
217 static int sha224_init(struct shash_desc
*desc
)
219 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
220 sctx
->state
[0] = SHA224_H0
;
221 sctx
->state
[1] = SHA224_H1
;
222 sctx
->state
[2] = SHA224_H2
;
223 sctx
->state
[3] = SHA224_H3
;
224 sctx
->state
[4] = SHA224_H4
;
225 sctx
->state
[5] = SHA224_H5
;
226 sctx
->state
[6] = SHA224_H6
;
227 sctx
->state
[7] = SHA224_H7
;
233 static int sha256_init(struct shash_desc
*desc
)
235 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
236 sctx
->state
[0] = SHA256_H0
;
237 sctx
->state
[1] = SHA256_H1
;
238 sctx
->state
[2] = SHA256_H2
;
239 sctx
->state
[3] = SHA256_H3
;
240 sctx
->state
[4] = SHA256_H4
;
241 sctx
->state
[5] = SHA256_H5
;
242 sctx
->state
[6] = SHA256_H6
;
243 sctx
->state
[7] = SHA256_H7
;
249 int crypto_sha256_update(struct shash_desc
*desc
, const u8
*data
,
252 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
253 unsigned int partial
, done
;
256 partial
= sctx
->count
& 0x3f;
261 if ((partial
+ len
) > 63) {
264 memcpy(sctx
->buf
+ partial
, data
, done
+ 64);
269 sha256_transform(sctx
->state
, src
);
272 } while (done
+ 63 < len
);
276 memcpy(sctx
->buf
+ partial
, src
, len
- done
);
280 EXPORT_SYMBOL(crypto_sha256_update
);
282 static int sha256_final(struct shash_desc
*desc
, u8
*out
)
284 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
285 __be32
*dst
= (__be32
*)out
;
287 unsigned int index
, pad_len
;
289 static const u8 padding
[64] = { 0x80, };
291 /* Save number of bits */
292 bits
= cpu_to_be64(sctx
->count
<< 3);
294 /* Pad out to 56 mod 64. */
295 index
= sctx
->count
& 0x3f;
296 pad_len
= (index
< 56) ? (56 - index
) : ((64+56) - index
);
297 crypto_sha256_update(desc
, padding
, pad_len
);
299 /* Append length (before padding) */
300 crypto_sha256_update(desc
, (const u8
*)&bits
, sizeof(bits
));
302 /* Store state in digest */
303 for (i
= 0; i
< 8; i
++)
304 dst
[i
] = cpu_to_be32(sctx
->state
[i
]);
306 /* Zeroize sensitive information. */
307 memset(sctx
, 0, sizeof(*sctx
));
312 static int sha224_final(struct shash_desc
*desc
, u8
*hash
)
314 u8 D
[SHA256_DIGEST_SIZE
];
316 sha256_final(desc
, D
);
318 memcpy(hash
, D
, SHA224_DIGEST_SIZE
);
319 memzero_explicit(D
, SHA256_DIGEST_SIZE
);
324 static int sha256_export(struct shash_desc
*desc
, void *out
)
326 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
328 memcpy(out
, sctx
, sizeof(*sctx
));
332 static int sha256_import(struct shash_desc
*desc
, const void *in
)
334 struct sha256_state
*sctx
= shash_desc_ctx(desc
);
336 memcpy(sctx
, in
, sizeof(*sctx
));
340 static struct shash_alg sha256_algs
[2] = { {
341 .digestsize
= SHA256_DIGEST_SIZE
,
343 .update
= crypto_sha256_update
,
344 .final
= sha256_final
,
345 .export
= sha256_export
,
346 .import
= sha256_import
,
347 .descsize
= sizeof(struct sha256_state
),
348 .statesize
= sizeof(struct sha256_state
),
350 .cra_name
= "sha256",
351 .cra_driver_name
= "sha256-generic",
352 .cra_flags
= CRYPTO_ALG_TYPE_SHASH
,
353 .cra_blocksize
= SHA256_BLOCK_SIZE
,
354 .cra_module
= THIS_MODULE
,
357 .digestsize
= SHA224_DIGEST_SIZE
,
359 .update
= crypto_sha256_update
,
360 .final
= sha224_final
,
361 .descsize
= sizeof(struct sha256_state
),
363 .cra_name
= "sha224",
364 .cra_driver_name
= "sha224-generic",
365 .cra_flags
= CRYPTO_ALG_TYPE_SHASH
,
366 .cra_blocksize
= SHA224_BLOCK_SIZE
,
367 .cra_module
= THIS_MODULE
,
371 static int __init
sha256_generic_mod_init(void)
373 return crypto_register_shashes(sha256_algs
, ARRAY_SIZE(sha256_algs
));
376 static void __exit
sha256_generic_mod_fini(void)
378 crypto_unregister_shashes(sha256_algs
, ARRAY_SIZE(sha256_algs
));
381 module_init(sha256_generic_mod_init
);
382 module_exit(sha256_generic_mod_fini
);
384 MODULE_LICENSE("GPL");
385 MODULE_DESCRIPTION("SHA-224 and SHA-256 Secure Hash Algorithm");
387 MODULE_ALIAS_CRYPTO("sha224");
388 MODULE_ALIAS_CRYPTO("sha224-generic");
389 MODULE_ALIAS_CRYPTO("sha256");
390 MODULE_ALIAS_CRYPTO("sha256-generic");