RDMA/srp: Rework SCSI device reset handling
[linux/fpc-iii.git] / net / ceph / crypto.c
blob42e8649c6e79b4950531f0027f2e67820377c066
2 #include <linux/ceph/ceph_debug.h>
4 #include <linux/err.h>
5 #include <linux/scatterlist.h>
6 #include <linux/slab.h>
7 #include <crypto/hash.h>
8 #include <linux/key-type.h>
10 #include <keys/ceph-type.h>
11 #include <keys/user-type.h>
12 #include <linux/ceph/decode.h>
13 #include "crypto.h"
15 int ceph_crypto_key_clone(struct ceph_crypto_key *dst,
16 const struct ceph_crypto_key *src)
18 memcpy(dst, src, sizeof(struct ceph_crypto_key));
19 dst->key = kmemdup(src->key, src->len, GFP_NOFS);
20 if (!dst->key)
21 return -ENOMEM;
22 return 0;
25 int ceph_crypto_key_encode(struct ceph_crypto_key *key, void **p, void *end)
27 if (*p + sizeof(u16) + sizeof(key->created) +
28 sizeof(u16) + key->len > end)
29 return -ERANGE;
30 ceph_encode_16(p, key->type);
31 ceph_encode_copy(p, &key->created, sizeof(key->created));
32 ceph_encode_16(p, key->len);
33 ceph_encode_copy(p, key->key, key->len);
34 return 0;
37 int ceph_crypto_key_decode(struct ceph_crypto_key *key, void **p, void *end)
39 ceph_decode_need(p, end, 2*sizeof(u16) + sizeof(key->created), bad);
40 key->type = ceph_decode_16(p);
41 ceph_decode_copy(p, &key->created, sizeof(key->created));
42 key->len = ceph_decode_16(p);
43 ceph_decode_need(p, end, key->len, bad);
44 key->key = kmalloc(key->len, GFP_NOFS);
45 if (!key->key)
46 return -ENOMEM;
47 ceph_decode_copy(p, key->key, key->len);
48 return 0;
50 bad:
51 dout("failed to decode crypto key\n");
52 return -EINVAL;
55 int ceph_crypto_key_unarmor(struct ceph_crypto_key *key, const char *inkey)
57 int inlen = strlen(inkey);
58 int blen = inlen * 3 / 4;
59 void *buf, *p;
60 int ret;
62 dout("crypto_key_unarmor %s\n", inkey);
63 buf = kmalloc(blen, GFP_NOFS);
64 if (!buf)
65 return -ENOMEM;
66 blen = ceph_unarmor(buf, inkey, inkey+inlen);
67 if (blen < 0) {
68 kfree(buf);
69 return blen;
72 p = buf;
73 ret = ceph_crypto_key_decode(key, &p, p + blen);
74 kfree(buf);
75 if (ret)
76 return ret;
77 dout("crypto_key_unarmor key %p type %d len %d\n", key,
78 key->type, key->len);
79 return 0;
82 static struct crypto_blkcipher *ceph_crypto_alloc_cipher(void)
84 return crypto_alloc_blkcipher("cbc(aes)", 0, CRYPTO_ALG_ASYNC);
87 static const u8 *aes_iv = (u8 *)CEPH_AES_IV;
90 * Should be used for buffers allocated with ceph_kvmalloc().
91 * Currently these are encrypt out-buffer (ceph_buffer) and decrypt
92 * in-buffer (msg front).
94 * Dispose of @sgt with teardown_sgtable().
96 * @prealloc_sg is to avoid memory allocation inside sg_alloc_table()
97 * in cases where a single sg is sufficient. No attempt to reduce the
98 * number of sgs by squeezing physically contiguous pages together is
99 * made though, for simplicity.
101 static int setup_sgtable(struct sg_table *sgt, struct scatterlist *prealloc_sg,
102 const void *buf, unsigned int buf_len)
104 struct scatterlist *sg;
105 const bool is_vmalloc = is_vmalloc_addr(buf);
106 unsigned int off = offset_in_page(buf);
107 unsigned int chunk_cnt = 1;
108 unsigned int chunk_len = PAGE_ALIGN(off + buf_len);
109 int i;
110 int ret;
112 if (buf_len == 0) {
113 memset(sgt, 0, sizeof(*sgt));
114 return -EINVAL;
117 if (is_vmalloc) {
118 chunk_cnt = chunk_len >> PAGE_SHIFT;
119 chunk_len = PAGE_SIZE;
122 if (chunk_cnt > 1) {
123 ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS);
124 if (ret)
125 return ret;
126 } else {
127 WARN_ON(chunk_cnt != 1);
128 sg_init_table(prealloc_sg, 1);
129 sgt->sgl = prealloc_sg;
130 sgt->nents = sgt->orig_nents = 1;
133 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) {
134 struct page *page;
135 unsigned int len = min(chunk_len - off, buf_len);
137 if (is_vmalloc)
138 page = vmalloc_to_page(buf);
139 else
140 page = virt_to_page(buf);
142 sg_set_page(sg, page, len, off);
144 off = 0;
145 buf += len;
146 buf_len -= len;
148 WARN_ON(buf_len != 0);
150 return 0;
153 static void teardown_sgtable(struct sg_table *sgt)
155 if (sgt->orig_nents > 1)
156 sg_free_table(sgt);
159 static int ceph_aes_encrypt(const void *key, int key_len,
160 void *dst, size_t *dst_len,
161 const void *src, size_t src_len)
163 struct scatterlist sg_in[2], prealloc_sg;
164 struct sg_table sg_out;
165 struct crypto_blkcipher *tfm = ceph_crypto_alloc_cipher();
166 struct blkcipher_desc desc = { .tfm = tfm, .flags = 0 };
167 int ret;
168 void *iv;
169 int ivsize;
170 size_t zero_padding = (0x10 - (src_len & 0x0f));
171 char pad[16];
173 if (IS_ERR(tfm))
174 return PTR_ERR(tfm);
176 memset(pad, zero_padding, zero_padding);
178 *dst_len = src_len + zero_padding;
180 sg_init_table(sg_in, 2);
181 sg_set_buf(&sg_in[0], src, src_len);
182 sg_set_buf(&sg_in[1], pad, zero_padding);
183 ret = setup_sgtable(&sg_out, &prealloc_sg, dst, *dst_len);
184 if (ret)
185 goto out_tfm;
187 crypto_blkcipher_setkey((void *)tfm, key, key_len);
188 iv = crypto_blkcipher_crt(tfm)->iv;
189 ivsize = crypto_blkcipher_ivsize(tfm);
190 memcpy(iv, aes_iv, ivsize);
193 print_hex_dump(KERN_ERR, "enc key: ", DUMP_PREFIX_NONE, 16, 1,
194 key, key_len, 1);
195 print_hex_dump(KERN_ERR, "enc src: ", DUMP_PREFIX_NONE, 16, 1,
196 src, src_len, 1);
197 print_hex_dump(KERN_ERR, "enc pad: ", DUMP_PREFIX_NONE, 16, 1,
198 pad, zero_padding, 1);
200 ret = crypto_blkcipher_encrypt(&desc, sg_out.sgl, sg_in,
201 src_len + zero_padding);
202 if (ret < 0) {
203 pr_err("ceph_aes_crypt failed %d\n", ret);
204 goto out_sg;
207 print_hex_dump(KERN_ERR, "enc out: ", DUMP_PREFIX_NONE, 16, 1,
208 dst, *dst_len, 1);
211 out_sg:
212 teardown_sgtable(&sg_out);
213 out_tfm:
214 crypto_free_blkcipher(tfm);
215 return ret;
218 static int ceph_aes_encrypt2(const void *key, int key_len, void *dst,
219 size_t *dst_len,
220 const void *src1, size_t src1_len,
221 const void *src2, size_t src2_len)
223 struct scatterlist sg_in[3], prealloc_sg;
224 struct sg_table sg_out;
225 struct crypto_blkcipher *tfm = ceph_crypto_alloc_cipher();
226 struct blkcipher_desc desc = { .tfm = tfm, .flags = 0 };
227 int ret;
228 void *iv;
229 int ivsize;
230 size_t zero_padding = (0x10 - ((src1_len + src2_len) & 0x0f));
231 char pad[16];
233 if (IS_ERR(tfm))
234 return PTR_ERR(tfm);
236 memset(pad, zero_padding, zero_padding);
238 *dst_len = src1_len + src2_len + zero_padding;
240 sg_init_table(sg_in, 3);
241 sg_set_buf(&sg_in[0], src1, src1_len);
242 sg_set_buf(&sg_in[1], src2, src2_len);
243 sg_set_buf(&sg_in[2], pad, zero_padding);
244 ret = setup_sgtable(&sg_out, &prealloc_sg, dst, *dst_len);
245 if (ret)
246 goto out_tfm;
248 crypto_blkcipher_setkey((void *)tfm, key, key_len);
249 iv = crypto_blkcipher_crt(tfm)->iv;
250 ivsize = crypto_blkcipher_ivsize(tfm);
251 memcpy(iv, aes_iv, ivsize);
254 print_hex_dump(KERN_ERR, "enc key: ", DUMP_PREFIX_NONE, 16, 1,
255 key, key_len, 1);
256 print_hex_dump(KERN_ERR, "enc src1: ", DUMP_PREFIX_NONE, 16, 1,
257 src1, src1_len, 1);
258 print_hex_dump(KERN_ERR, "enc src2: ", DUMP_PREFIX_NONE, 16, 1,
259 src2, src2_len, 1);
260 print_hex_dump(KERN_ERR, "enc pad: ", DUMP_PREFIX_NONE, 16, 1,
261 pad, zero_padding, 1);
263 ret = crypto_blkcipher_encrypt(&desc, sg_out.sgl, sg_in,
264 src1_len + src2_len + zero_padding);
265 if (ret < 0) {
266 pr_err("ceph_aes_crypt2 failed %d\n", ret);
267 goto out_sg;
270 print_hex_dump(KERN_ERR, "enc out: ", DUMP_PREFIX_NONE, 16, 1,
271 dst, *dst_len, 1);
274 out_sg:
275 teardown_sgtable(&sg_out);
276 out_tfm:
277 crypto_free_blkcipher(tfm);
278 return ret;
281 static int ceph_aes_decrypt(const void *key, int key_len,
282 void *dst, size_t *dst_len,
283 const void *src, size_t src_len)
285 struct sg_table sg_in;
286 struct scatterlist sg_out[2], prealloc_sg;
287 struct crypto_blkcipher *tfm = ceph_crypto_alloc_cipher();
288 struct blkcipher_desc desc = { .tfm = tfm };
289 char pad[16];
290 void *iv;
291 int ivsize;
292 int ret;
293 int last_byte;
295 if (IS_ERR(tfm))
296 return PTR_ERR(tfm);
298 sg_init_table(sg_out, 2);
299 sg_set_buf(&sg_out[0], dst, *dst_len);
300 sg_set_buf(&sg_out[1], pad, sizeof(pad));
301 ret = setup_sgtable(&sg_in, &prealloc_sg, src, src_len);
302 if (ret)
303 goto out_tfm;
305 crypto_blkcipher_setkey((void *)tfm, key, key_len);
306 iv = crypto_blkcipher_crt(tfm)->iv;
307 ivsize = crypto_blkcipher_ivsize(tfm);
308 memcpy(iv, aes_iv, ivsize);
311 print_hex_dump(KERN_ERR, "dec key: ", DUMP_PREFIX_NONE, 16, 1,
312 key, key_len, 1);
313 print_hex_dump(KERN_ERR, "dec in: ", DUMP_PREFIX_NONE, 16, 1,
314 src, src_len, 1);
316 ret = crypto_blkcipher_decrypt(&desc, sg_out, sg_in.sgl, src_len);
317 if (ret < 0) {
318 pr_err("ceph_aes_decrypt failed %d\n", ret);
319 goto out_sg;
322 if (src_len <= *dst_len)
323 last_byte = ((char *)dst)[src_len - 1];
324 else
325 last_byte = pad[src_len - *dst_len - 1];
326 if (last_byte <= 16 && src_len >= last_byte) {
327 *dst_len = src_len - last_byte;
328 } else {
329 pr_err("ceph_aes_decrypt got bad padding %d on src len %d\n",
330 last_byte, (int)src_len);
331 return -EPERM; /* bad padding */
334 print_hex_dump(KERN_ERR, "dec out: ", DUMP_PREFIX_NONE, 16, 1,
335 dst, *dst_len, 1);
338 out_sg:
339 teardown_sgtable(&sg_in);
340 out_tfm:
341 crypto_free_blkcipher(tfm);
342 return ret;
345 static int ceph_aes_decrypt2(const void *key, int key_len,
346 void *dst1, size_t *dst1_len,
347 void *dst2, size_t *dst2_len,
348 const void *src, size_t src_len)
350 struct sg_table sg_in;
351 struct scatterlist sg_out[3], prealloc_sg;
352 struct crypto_blkcipher *tfm = ceph_crypto_alloc_cipher();
353 struct blkcipher_desc desc = { .tfm = tfm };
354 char pad[16];
355 void *iv;
356 int ivsize;
357 int ret;
358 int last_byte;
360 if (IS_ERR(tfm))
361 return PTR_ERR(tfm);
363 sg_init_table(sg_out, 3);
364 sg_set_buf(&sg_out[0], dst1, *dst1_len);
365 sg_set_buf(&sg_out[1], dst2, *dst2_len);
366 sg_set_buf(&sg_out[2], pad, sizeof(pad));
367 ret = setup_sgtable(&sg_in, &prealloc_sg, src, src_len);
368 if (ret)
369 goto out_tfm;
371 crypto_blkcipher_setkey((void *)tfm, key, key_len);
372 iv = crypto_blkcipher_crt(tfm)->iv;
373 ivsize = crypto_blkcipher_ivsize(tfm);
374 memcpy(iv, aes_iv, ivsize);
377 print_hex_dump(KERN_ERR, "dec key: ", DUMP_PREFIX_NONE, 16, 1,
378 key, key_len, 1);
379 print_hex_dump(KERN_ERR, "dec in: ", DUMP_PREFIX_NONE, 16, 1,
380 src, src_len, 1);
382 ret = crypto_blkcipher_decrypt(&desc, sg_out, sg_in.sgl, src_len);
383 if (ret < 0) {
384 pr_err("ceph_aes_decrypt failed %d\n", ret);
385 goto out_sg;
388 if (src_len <= *dst1_len)
389 last_byte = ((char *)dst1)[src_len - 1];
390 else if (src_len <= *dst1_len + *dst2_len)
391 last_byte = ((char *)dst2)[src_len - *dst1_len - 1];
392 else
393 last_byte = pad[src_len - *dst1_len - *dst2_len - 1];
394 if (last_byte <= 16 && src_len >= last_byte) {
395 src_len -= last_byte;
396 } else {
397 pr_err("ceph_aes_decrypt got bad padding %d on src len %d\n",
398 last_byte, (int)src_len);
399 return -EPERM; /* bad padding */
402 if (src_len < *dst1_len) {
403 *dst1_len = src_len;
404 *dst2_len = 0;
405 } else {
406 *dst2_len = src_len - *dst1_len;
409 print_hex_dump(KERN_ERR, "dec out1: ", DUMP_PREFIX_NONE, 16, 1,
410 dst1, *dst1_len, 1);
411 print_hex_dump(KERN_ERR, "dec out2: ", DUMP_PREFIX_NONE, 16, 1,
412 dst2, *dst2_len, 1);
415 out_sg:
416 teardown_sgtable(&sg_in);
417 out_tfm:
418 crypto_free_blkcipher(tfm);
419 return ret;
423 int ceph_decrypt(struct ceph_crypto_key *secret, void *dst, size_t *dst_len,
424 const void *src, size_t src_len)
426 switch (secret->type) {
427 case CEPH_CRYPTO_NONE:
428 if (*dst_len < src_len)
429 return -ERANGE;
430 memcpy(dst, src, src_len);
431 *dst_len = src_len;
432 return 0;
434 case CEPH_CRYPTO_AES:
435 return ceph_aes_decrypt(secret->key, secret->len, dst,
436 dst_len, src, src_len);
438 default:
439 return -EINVAL;
443 int ceph_decrypt2(struct ceph_crypto_key *secret,
444 void *dst1, size_t *dst1_len,
445 void *dst2, size_t *dst2_len,
446 const void *src, size_t src_len)
448 size_t t;
450 switch (secret->type) {
451 case CEPH_CRYPTO_NONE:
452 if (*dst1_len + *dst2_len < src_len)
453 return -ERANGE;
454 t = min(*dst1_len, src_len);
455 memcpy(dst1, src, t);
456 *dst1_len = t;
457 src += t;
458 src_len -= t;
459 if (src_len) {
460 t = min(*dst2_len, src_len);
461 memcpy(dst2, src, t);
462 *dst2_len = t;
464 return 0;
466 case CEPH_CRYPTO_AES:
467 return ceph_aes_decrypt2(secret->key, secret->len,
468 dst1, dst1_len, dst2, dst2_len,
469 src, src_len);
471 default:
472 return -EINVAL;
476 int ceph_encrypt(struct ceph_crypto_key *secret, void *dst, size_t *dst_len,
477 const void *src, size_t src_len)
479 switch (secret->type) {
480 case CEPH_CRYPTO_NONE:
481 if (*dst_len < src_len)
482 return -ERANGE;
483 memcpy(dst, src, src_len);
484 *dst_len = src_len;
485 return 0;
487 case CEPH_CRYPTO_AES:
488 return ceph_aes_encrypt(secret->key, secret->len, dst,
489 dst_len, src, src_len);
491 default:
492 return -EINVAL;
496 int ceph_encrypt2(struct ceph_crypto_key *secret, void *dst, size_t *dst_len,
497 const void *src1, size_t src1_len,
498 const void *src2, size_t src2_len)
500 switch (secret->type) {
501 case CEPH_CRYPTO_NONE:
502 if (*dst_len < src1_len + src2_len)
503 return -ERANGE;
504 memcpy(dst, src1, src1_len);
505 memcpy(dst + src1_len, src2, src2_len);
506 *dst_len = src1_len + src2_len;
507 return 0;
509 case CEPH_CRYPTO_AES:
510 return ceph_aes_encrypt2(secret->key, secret->len, dst, dst_len,
511 src1, src1_len, src2, src2_len);
513 default:
514 return -EINVAL;
518 static int ceph_key_preparse(struct key_preparsed_payload *prep)
520 struct ceph_crypto_key *ckey;
521 size_t datalen = prep->datalen;
522 int ret;
523 void *p;
525 ret = -EINVAL;
526 if (datalen <= 0 || datalen > 32767 || !prep->data)
527 goto err;
529 ret = -ENOMEM;
530 ckey = kmalloc(sizeof(*ckey), GFP_KERNEL);
531 if (!ckey)
532 goto err;
534 /* TODO ceph_crypto_key_decode should really take const input */
535 p = (void *)prep->data;
536 ret = ceph_crypto_key_decode(ckey, &p, (char*)prep->data+datalen);
537 if (ret < 0)
538 goto err_ckey;
540 prep->payload.data[0] = ckey;
541 prep->quotalen = datalen;
542 return 0;
544 err_ckey:
545 kfree(ckey);
546 err:
547 return ret;
550 static void ceph_key_free_preparse(struct key_preparsed_payload *prep)
552 struct ceph_crypto_key *ckey = prep->payload.data[0];
553 ceph_crypto_key_destroy(ckey);
554 kfree(ckey);
557 static void ceph_key_destroy(struct key *key)
559 struct ceph_crypto_key *ckey = key->payload.data[0];
561 ceph_crypto_key_destroy(ckey);
562 kfree(ckey);
565 struct key_type key_type_ceph = {
566 .name = "ceph",
567 .preparse = ceph_key_preparse,
568 .free_preparse = ceph_key_free_preparse,
569 .instantiate = generic_key_instantiate,
570 .destroy = ceph_key_destroy,
573 int ceph_crypto_init(void) {
574 return register_key_type(&key_type_ceph);
577 void ceph_crypto_shutdown(void) {
578 unregister_key_type(&key_type_ceph);