Merge tag 'trace-printf-v6.13' of git://git.kernel.org/pub/scm/linux/kernel/git/trace...
[drm/drm-misc.git] / drivers / tee / tee_core.c
blobd113679b1e2d7a6159b18e3c2044346bd37e4566
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Copyright (c) 2015-2016, Linaro Limited
4 */
6 #define pr_fmt(fmt) "%s: " fmt, __func__
8 #include <linux/cdev.h>
9 #include <linux/cred.h>
10 #include <linux/fs.h>
11 #include <linux/idr.h>
12 #include <linux/module.h>
13 #include <linux/slab.h>
14 #include <linux/tee_core.h>
15 #include <linux/uaccess.h>
16 #include <crypto/hash.h>
17 #include <crypto/sha1.h>
18 #include "tee_private.h"
20 #define TEE_NUM_DEVICES 32
22 #define TEE_IOCTL_PARAM_SIZE(x) (sizeof(struct tee_param) * (x))
24 #define TEE_UUID_NS_NAME_SIZE 128
27 * TEE Client UUID name space identifier (UUIDv4)
29 * Value here is random UUID that is allocated as name space identifier for
30 * forming Client UUID's for TEE environment using UUIDv5 scheme.
32 static const uuid_t tee_client_uuid_ns = UUID_INIT(0x58ac9ca0, 0x2086, 0x4683,
33 0xa1, 0xb8, 0xec, 0x4b,
34 0xc0, 0x8e, 0x01, 0xb6);
37 * Unprivileged devices in the lower half range and privileged devices in
38 * the upper half range.
40 static DECLARE_BITMAP(dev_mask, TEE_NUM_DEVICES);
41 static DEFINE_SPINLOCK(driver_lock);
43 static const struct class tee_class;
44 static dev_t tee_devt;
46 struct tee_context *teedev_open(struct tee_device *teedev)
48 int rc;
49 struct tee_context *ctx;
51 if (!tee_device_get(teedev))
52 return ERR_PTR(-EINVAL);
54 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
55 if (!ctx) {
56 rc = -ENOMEM;
57 goto err;
60 kref_init(&ctx->refcount);
61 ctx->teedev = teedev;
62 rc = teedev->desc->ops->open(ctx);
63 if (rc)
64 goto err;
66 return ctx;
67 err:
68 kfree(ctx);
69 tee_device_put(teedev);
70 return ERR_PTR(rc);
73 EXPORT_SYMBOL_GPL(teedev_open);
75 void teedev_ctx_get(struct tee_context *ctx)
77 if (ctx->releasing)
78 return;
80 kref_get(&ctx->refcount);
83 static void teedev_ctx_release(struct kref *ref)
85 struct tee_context *ctx = container_of(ref, struct tee_context,
86 refcount);
87 ctx->releasing = true;
88 ctx->teedev->desc->ops->release(ctx);
89 kfree(ctx);
92 void teedev_ctx_put(struct tee_context *ctx)
94 if (ctx->releasing)
95 return;
97 kref_put(&ctx->refcount, teedev_ctx_release);
100 void teedev_close_context(struct tee_context *ctx)
102 struct tee_device *teedev = ctx->teedev;
104 teedev_ctx_put(ctx);
105 tee_device_put(teedev);
107 EXPORT_SYMBOL_GPL(teedev_close_context);
109 static int tee_open(struct inode *inode, struct file *filp)
111 struct tee_context *ctx;
113 ctx = teedev_open(container_of(inode->i_cdev, struct tee_device, cdev));
114 if (IS_ERR(ctx))
115 return PTR_ERR(ctx);
118 * Default user-space behaviour is to wait for tee-supplicant
119 * if not present for any requests in this context.
121 ctx->supp_nowait = false;
122 filp->private_data = ctx;
123 return 0;
126 static int tee_release(struct inode *inode, struct file *filp)
128 teedev_close_context(filp->private_data);
129 return 0;
133 * uuid_v5() - Calculate UUIDv5
134 * @uuid: Resulting UUID
135 * @ns: Name space ID for UUIDv5 function
136 * @name: Name for UUIDv5 function
137 * @size: Size of name
139 * UUIDv5 is specific in RFC 4122.
141 * This implements section (for SHA-1):
142 * 4.3. Algorithm for Creating a Name-Based UUID
144 static int uuid_v5(uuid_t *uuid, const uuid_t *ns, const void *name,
145 size_t size)
147 unsigned char hash[SHA1_DIGEST_SIZE];
148 struct crypto_shash *shash = NULL;
149 struct shash_desc *desc = NULL;
150 int rc;
152 shash = crypto_alloc_shash("sha1", 0, 0);
153 if (IS_ERR(shash)) {
154 rc = PTR_ERR(shash);
155 pr_err("shash(sha1) allocation failed\n");
156 return rc;
159 desc = kzalloc(sizeof(*desc) + crypto_shash_descsize(shash),
160 GFP_KERNEL);
161 if (!desc) {
162 rc = -ENOMEM;
163 goto out_free_shash;
166 desc->tfm = shash;
168 rc = crypto_shash_init(desc);
169 if (rc < 0)
170 goto out_free_desc;
172 rc = crypto_shash_update(desc, (const u8 *)ns, sizeof(*ns));
173 if (rc < 0)
174 goto out_free_desc;
176 rc = crypto_shash_update(desc, (const u8 *)name, size);
177 if (rc < 0)
178 goto out_free_desc;
180 rc = crypto_shash_final(desc, hash);
181 if (rc < 0)
182 goto out_free_desc;
184 memcpy(uuid->b, hash, UUID_SIZE);
186 /* Tag for version 5 */
187 uuid->b[6] = (hash[6] & 0x0F) | 0x50;
188 uuid->b[8] = (hash[8] & 0x3F) | 0x80;
190 out_free_desc:
191 kfree(desc);
193 out_free_shash:
194 crypto_free_shash(shash);
195 return rc;
198 int tee_session_calc_client_uuid(uuid_t *uuid, u32 connection_method,
199 const u8 connection_data[TEE_IOCTL_UUID_LEN])
201 gid_t ns_grp = (gid_t)-1;
202 kgid_t grp = INVALID_GID;
203 char *name = NULL;
204 int name_len;
205 int rc;
207 if (connection_method == TEE_IOCTL_LOGIN_PUBLIC ||
208 connection_method == TEE_IOCTL_LOGIN_REE_KERNEL) {
209 /* Nil UUID to be passed to TEE environment */
210 uuid_copy(uuid, &uuid_null);
211 return 0;
215 * In Linux environment client UUID is based on UUIDv5.
217 * Determine client UUID with following semantics for 'name':
219 * For TEEC_LOGIN_USER:
220 * uid=<uid>
222 * For TEEC_LOGIN_GROUP:
223 * gid=<gid>
227 name = kzalloc(TEE_UUID_NS_NAME_SIZE, GFP_KERNEL);
228 if (!name)
229 return -ENOMEM;
231 switch (connection_method) {
232 case TEE_IOCTL_LOGIN_USER:
233 name_len = snprintf(name, TEE_UUID_NS_NAME_SIZE, "uid=%x",
234 current_euid().val);
235 if (name_len >= TEE_UUID_NS_NAME_SIZE) {
236 rc = -E2BIG;
237 goto out_free_name;
239 break;
241 case TEE_IOCTL_LOGIN_GROUP:
242 memcpy(&ns_grp, connection_data, sizeof(gid_t));
243 grp = make_kgid(current_user_ns(), ns_grp);
244 if (!gid_valid(grp) || !in_egroup_p(grp)) {
245 rc = -EPERM;
246 goto out_free_name;
249 name_len = snprintf(name, TEE_UUID_NS_NAME_SIZE, "gid=%x",
250 grp.val);
251 if (name_len >= TEE_UUID_NS_NAME_SIZE) {
252 rc = -E2BIG;
253 goto out_free_name;
255 break;
257 default:
258 rc = -EINVAL;
259 goto out_free_name;
262 rc = uuid_v5(uuid, &tee_client_uuid_ns, name, name_len);
263 out_free_name:
264 kfree(name);
266 return rc;
268 EXPORT_SYMBOL_GPL(tee_session_calc_client_uuid);
270 static int tee_ioctl_version(struct tee_context *ctx,
271 struct tee_ioctl_version_data __user *uvers)
273 struct tee_ioctl_version_data vers;
275 ctx->teedev->desc->ops->get_version(ctx->teedev, &vers);
277 if (ctx->teedev->desc->flags & TEE_DESC_PRIVILEGED)
278 vers.gen_caps |= TEE_GEN_CAP_PRIVILEGED;
280 if (copy_to_user(uvers, &vers, sizeof(vers)))
281 return -EFAULT;
283 return 0;
286 static int tee_ioctl_shm_alloc(struct tee_context *ctx,
287 struct tee_ioctl_shm_alloc_data __user *udata)
289 long ret;
290 struct tee_ioctl_shm_alloc_data data;
291 struct tee_shm *shm;
293 if (copy_from_user(&data, udata, sizeof(data)))
294 return -EFAULT;
296 /* Currently no input flags are supported */
297 if (data.flags)
298 return -EINVAL;
300 shm = tee_shm_alloc_user_buf(ctx, data.size);
301 if (IS_ERR(shm))
302 return PTR_ERR(shm);
304 data.id = shm->id;
305 data.size = shm->size;
307 if (copy_to_user(udata, &data, sizeof(data)))
308 ret = -EFAULT;
309 else
310 ret = tee_shm_get_fd(shm);
313 * When user space closes the file descriptor the shared memory
314 * should be freed or if tee_shm_get_fd() failed then it will
315 * be freed immediately.
317 tee_shm_put(shm);
318 return ret;
321 static int
322 tee_ioctl_shm_register(struct tee_context *ctx,
323 struct tee_ioctl_shm_register_data __user *udata)
325 long ret;
326 struct tee_ioctl_shm_register_data data;
327 struct tee_shm *shm;
329 if (copy_from_user(&data, udata, sizeof(data)))
330 return -EFAULT;
332 /* Currently no input flags are supported */
333 if (data.flags)
334 return -EINVAL;
336 shm = tee_shm_register_user_buf(ctx, data.addr, data.length);
337 if (IS_ERR(shm))
338 return PTR_ERR(shm);
340 data.id = shm->id;
341 data.length = shm->size;
343 if (copy_to_user(udata, &data, sizeof(data)))
344 ret = -EFAULT;
345 else
346 ret = tee_shm_get_fd(shm);
348 * When user space closes the file descriptor the shared memory
349 * should be freed or if tee_shm_get_fd() failed then it will
350 * be freed immediately.
352 tee_shm_put(shm);
353 return ret;
356 static int params_from_user(struct tee_context *ctx, struct tee_param *params,
357 size_t num_params,
358 struct tee_ioctl_param __user *uparams)
360 size_t n;
362 for (n = 0; n < num_params; n++) {
363 struct tee_shm *shm;
364 struct tee_ioctl_param ip;
366 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
367 return -EFAULT;
369 /* All unused attribute bits has to be zero */
370 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
371 return -EINVAL;
373 params[n].attr = ip.attr;
374 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
375 case TEE_IOCTL_PARAM_ATTR_TYPE_NONE:
376 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
377 break;
378 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
379 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
380 params[n].u.value.a = ip.a;
381 params[n].u.value.b = ip.b;
382 params[n].u.value.c = ip.c;
383 break;
384 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
385 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
386 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
388 * If a NULL pointer is passed to a TA in the TEE,
389 * the ip.c IOCTL parameters is set to TEE_MEMREF_NULL
390 * indicating a NULL memory reference.
392 if (ip.c != TEE_MEMREF_NULL) {
394 * If we fail to get a pointer to a shared
395 * memory object (and increase the ref count)
396 * from an identifier we return an error. All
397 * pointers that has been added in params have
398 * an increased ref count. It's the callers
399 * responibility to do tee_shm_put() on all
400 * resolved pointers.
402 shm = tee_shm_get_from_id(ctx, ip.c);
403 if (IS_ERR(shm))
404 return PTR_ERR(shm);
407 * Ensure offset + size does not overflow
408 * offset and does not overflow the size of
409 * the referred shared memory object.
411 if ((ip.a + ip.b) < ip.a ||
412 (ip.a + ip.b) > shm->size) {
413 tee_shm_put(shm);
414 return -EINVAL;
416 } else if (ctx->cap_memref_null) {
417 /* Pass NULL pointer to OP-TEE */
418 shm = NULL;
419 } else {
420 return -EINVAL;
423 params[n].u.memref.shm_offs = ip.a;
424 params[n].u.memref.size = ip.b;
425 params[n].u.memref.shm = shm;
426 break;
427 default:
428 /* Unknown attribute */
429 return -EINVAL;
432 return 0;
435 static int params_to_user(struct tee_ioctl_param __user *uparams,
436 size_t num_params, struct tee_param *params)
438 size_t n;
440 for (n = 0; n < num_params; n++) {
441 struct tee_ioctl_param __user *up = uparams + n;
442 struct tee_param *p = params + n;
444 switch (p->attr) {
445 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
446 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
447 if (put_user(p->u.value.a, &up->a) ||
448 put_user(p->u.value.b, &up->b) ||
449 put_user(p->u.value.c, &up->c))
450 return -EFAULT;
451 break;
452 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
453 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
454 if (put_user((u64)p->u.memref.size, &up->b))
455 return -EFAULT;
456 break;
457 default:
458 break;
461 return 0;
464 static int tee_ioctl_open_session(struct tee_context *ctx,
465 struct tee_ioctl_buf_data __user *ubuf)
467 int rc;
468 size_t n;
469 struct tee_ioctl_buf_data buf;
470 struct tee_ioctl_open_session_arg __user *uarg;
471 struct tee_ioctl_open_session_arg arg;
472 struct tee_ioctl_param __user *uparams = NULL;
473 struct tee_param *params = NULL;
474 bool have_session = false;
476 if (!ctx->teedev->desc->ops->open_session)
477 return -EINVAL;
479 if (copy_from_user(&buf, ubuf, sizeof(buf)))
480 return -EFAULT;
482 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
483 buf.buf_len < sizeof(struct tee_ioctl_open_session_arg))
484 return -EINVAL;
486 uarg = u64_to_user_ptr(buf.buf_ptr);
487 if (copy_from_user(&arg, uarg, sizeof(arg)))
488 return -EFAULT;
490 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
491 return -EINVAL;
493 if (arg.num_params) {
494 params = kcalloc(arg.num_params, sizeof(struct tee_param),
495 GFP_KERNEL);
496 if (!params)
497 return -ENOMEM;
498 uparams = uarg->params;
499 rc = params_from_user(ctx, params, arg.num_params, uparams);
500 if (rc)
501 goto out;
504 if (arg.clnt_login >= TEE_IOCTL_LOGIN_REE_KERNEL_MIN &&
505 arg.clnt_login <= TEE_IOCTL_LOGIN_REE_KERNEL_MAX) {
506 pr_debug("login method not allowed for user-space client\n");
507 rc = -EPERM;
508 goto out;
511 rc = ctx->teedev->desc->ops->open_session(ctx, &arg, params);
512 if (rc)
513 goto out;
514 have_session = true;
516 if (put_user(arg.session, &uarg->session) ||
517 put_user(arg.ret, &uarg->ret) ||
518 put_user(arg.ret_origin, &uarg->ret_origin)) {
519 rc = -EFAULT;
520 goto out;
522 rc = params_to_user(uparams, arg.num_params, params);
523 out:
525 * If we've succeeded to open the session but failed to communicate
526 * it back to user space, close the session again to avoid leakage.
528 if (rc && have_session && ctx->teedev->desc->ops->close_session)
529 ctx->teedev->desc->ops->close_session(ctx, arg.session);
531 if (params) {
532 /* Decrease ref count for all valid shared memory pointers */
533 for (n = 0; n < arg.num_params; n++)
534 if (tee_param_is_memref(params + n) &&
535 params[n].u.memref.shm)
536 tee_shm_put(params[n].u.memref.shm);
537 kfree(params);
540 return rc;
543 static int tee_ioctl_invoke(struct tee_context *ctx,
544 struct tee_ioctl_buf_data __user *ubuf)
546 int rc;
547 size_t n;
548 struct tee_ioctl_buf_data buf;
549 struct tee_ioctl_invoke_arg __user *uarg;
550 struct tee_ioctl_invoke_arg arg;
551 struct tee_ioctl_param __user *uparams = NULL;
552 struct tee_param *params = NULL;
554 if (!ctx->teedev->desc->ops->invoke_func)
555 return -EINVAL;
557 if (copy_from_user(&buf, ubuf, sizeof(buf)))
558 return -EFAULT;
560 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
561 buf.buf_len < sizeof(struct tee_ioctl_invoke_arg))
562 return -EINVAL;
564 uarg = u64_to_user_ptr(buf.buf_ptr);
565 if (copy_from_user(&arg, uarg, sizeof(arg)))
566 return -EFAULT;
568 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
569 return -EINVAL;
571 if (arg.num_params) {
572 params = kcalloc(arg.num_params, sizeof(struct tee_param),
573 GFP_KERNEL);
574 if (!params)
575 return -ENOMEM;
576 uparams = uarg->params;
577 rc = params_from_user(ctx, params, arg.num_params, uparams);
578 if (rc)
579 goto out;
582 rc = ctx->teedev->desc->ops->invoke_func(ctx, &arg, params);
583 if (rc)
584 goto out;
586 if (put_user(arg.ret, &uarg->ret) ||
587 put_user(arg.ret_origin, &uarg->ret_origin)) {
588 rc = -EFAULT;
589 goto out;
591 rc = params_to_user(uparams, arg.num_params, params);
592 out:
593 if (params) {
594 /* Decrease ref count for all valid shared memory pointers */
595 for (n = 0; n < arg.num_params; n++)
596 if (tee_param_is_memref(params + n) &&
597 params[n].u.memref.shm)
598 tee_shm_put(params[n].u.memref.shm);
599 kfree(params);
601 return rc;
604 static int tee_ioctl_cancel(struct tee_context *ctx,
605 struct tee_ioctl_cancel_arg __user *uarg)
607 struct tee_ioctl_cancel_arg arg;
609 if (!ctx->teedev->desc->ops->cancel_req)
610 return -EINVAL;
612 if (copy_from_user(&arg, uarg, sizeof(arg)))
613 return -EFAULT;
615 return ctx->teedev->desc->ops->cancel_req(ctx, arg.cancel_id,
616 arg.session);
619 static int
620 tee_ioctl_close_session(struct tee_context *ctx,
621 struct tee_ioctl_close_session_arg __user *uarg)
623 struct tee_ioctl_close_session_arg arg;
625 if (!ctx->teedev->desc->ops->close_session)
626 return -EINVAL;
628 if (copy_from_user(&arg, uarg, sizeof(arg)))
629 return -EFAULT;
631 return ctx->teedev->desc->ops->close_session(ctx, arg.session);
634 static int params_to_supp(struct tee_context *ctx,
635 struct tee_ioctl_param __user *uparams,
636 size_t num_params, struct tee_param *params)
638 size_t n;
640 for (n = 0; n < num_params; n++) {
641 struct tee_ioctl_param ip;
642 struct tee_param *p = params + n;
644 ip.attr = p->attr;
645 switch (p->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
646 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
647 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
648 ip.a = p->u.value.a;
649 ip.b = p->u.value.b;
650 ip.c = p->u.value.c;
651 break;
652 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
653 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
654 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
655 ip.b = p->u.memref.size;
656 if (!p->u.memref.shm) {
657 ip.a = 0;
658 ip.c = (u64)-1; /* invalid shm id */
659 break;
661 ip.a = p->u.memref.shm_offs;
662 ip.c = p->u.memref.shm->id;
663 break;
664 default:
665 ip.a = 0;
666 ip.b = 0;
667 ip.c = 0;
668 break;
671 if (copy_to_user(uparams + n, &ip, sizeof(ip)))
672 return -EFAULT;
675 return 0;
678 static int tee_ioctl_supp_recv(struct tee_context *ctx,
679 struct tee_ioctl_buf_data __user *ubuf)
681 int rc;
682 struct tee_ioctl_buf_data buf;
683 struct tee_iocl_supp_recv_arg __user *uarg;
684 struct tee_param *params;
685 u32 num_params;
686 u32 func;
688 if (!ctx->teedev->desc->ops->supp_recv)
689 return -EINVAL;
691 if (copy_from_user(&buf, ubuf, sizeof(buf)))
692 return -EFAULT;
694 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
695 buf.buf_len < sizeof(struct tee_iocl_supp_recv_arg))
696 return -EINVAL;
698 uarg = u64_to_user_ptr(buf.buf_ptr);
699 if (get_user(num_params, &uarg->num_params))
700 return -EFAULT;
702 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) != buf.buf_len)
703 return -EINVAL;
705 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
706 if (!params)
707 return -ENOMEM;
709 rc = params_from_user(ctx, params, num_params, uarg->params);
710 if (rc)
711 goto out;
713 rc = ctx->teedev->desc->ops->supp_recv(ctx, &func, &num_params, params);
714 if (rc)
715 goto out;
717 if (put_user(func, &uarg->func) ||
718 put_user(num_params, &uarg->num_params)) {
719 rc = -EFAULT;
720 goto out;
723 rc = params_to_supp(ctx, uarg->params, num_params, params);
724 out:
725 kfree(params);
726 return rc;
729 static int params_from_supp(struct tee_param *params, size_t num_params,
730 struct tee_ioctl_param __user *uparams)
732 size_t n;
734 for (n = 0; n < num_params; n++) {
735 struct tee_param *p = params + n;
736 struct tee_ioctl_param ip;
738 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
739 return -EFAULT;
741 /* All unused attribute bits has to be zero */
742 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
743 return -EINVAL;
745 p->attr = ip.attr;
746 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
747 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
748 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
749 /* Only out and in/out values can be updated */
750 p->u.value.a = ip.a;
751 p->u.value.b = ip.b;
752 p->u.value.c = ip.c;
753 break;
754 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
755 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
757 * Only the size of the memref can be updated.
758 * Since we don't have access to the original
759 * parameters here, only store the supplied size.
760 * The driver will copy the updated size into the
761 * original parameters.
763 p->u.memref.shm = NULL;
764 p->u.memref.shm_offs = 0;
765 p->u.memref.size = ip.b;
766 break;
767 default:
768 memset(&p->u, 0, sizeof(p->u));
769 break;
772 return 0;
775 static int tee_ioctl_supp_send(struct tee_context *ctx,
776 struct tee_ioctl_buf_data __user *ubuf)
778 long rc;
779 struct tee_ioctl_buf_data buf;
780 struct tee_iocl_supp_send_arg __user *uarg;
781 struct tee_param *params;
782 u32 num_params;
783 u32 ret;
785 /* Not valid for this driver */
786 if (!ctx->teedev->desc->ops->supp_send)
787 return -EINVAL;
789 if (copy_from_user(&buf, ubuf, sizeof(buf)))
790 return -EFAULT;
792 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
793 buf.buf_len < sizeof(struct tee_iocl_supp_send_arg))
794 return -EINVAL;
796 uarg = u64_to_user_ptr(buf.buf_ptr);
797 if (get_user(ret, &uarg->ret) ||
798 get_user(num_params, &uarg->num_params))
799 return -EFAULT;
801 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) > buf.buf_len)
802 return -EINVAL;
804 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
805 if (!params)
806 return -ENOMEM;
808 rc = params_from_supp(params, num_params, uarg->params);
809 if (rc)
810 goto out;
812 rc = ctx->teedev->desc->ops->supp_send(ctx, ret, num_params, params);
813 out:
814 kfree(params);
815 return rc;
818 static long tee_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
820 struct tee_context *ctx = filp->private_data;
821 void __user *uarg = (void __user *)arg;
823 switch (cmd) {
824 case TEE_IOC_VERSION:
825 return tee_ioctl_version(ctx, uarg);
826 case TEE_IOC_SHM_ALLOC:
827 return tee_ioctl_shm_alloc(ctx, uarg);
828 case TEE_IOC_SHM_REGISTER:
829 return tee_ioctl_shm_register(ctx, uarg);
830 case TEE_IOC_OPEN_SESSION:
831 return tee_ioctl_open_session(ctx, uarg);
832 case TEE_IOC_INVOKE:
833 return tee_ioctl_invoke(ctx, uarg);
834 case TEE_IOC_CANCEL:
835 return tee_ioctl_cancel(ctx, uarg);
836 case TEE_IOC_CLOSE_SESSION:
837 return tee_ioctl_close_session(ctx, uarg);
838 case TEE_IOC_SUPPL_RECV:
839 return tee_ioctl_supp_recv(ctx, uarg);
840 case TEE_IOC_SUPPL_SEND:
841 return tee_ioctl_supp_send(ctx, uarg);
842 default:
843 return -EINVAL;
847 static const struct file_operations tee_fops = {
848 .owner = THIS_MODULE,
849 .open = tee_open,
850 .release = tee_release,
851 .unlocked_ioctl = tee_ioctl,
852 .compat_ioctl = compat_ptr_ioctl,
855 static void tee_release_device(struct device *dev)
857 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
859 spin_lock(&driver_lock);
860 clear_bit(teedev->id, dev_mask);
861 spin_unlock(&driver_lock);
862 mutex_destroy(&teedev->mutex);
863 idr_destroy(&teedev->idr);
864 kfree(teedev);
868 * tee_device_alloc() - Allocate a new struct tee_device instance
869 * @teedesc: Descriptor for this driver
870 * @dev: Parent device for this device
871 * @pool: Shared memory pool, NULL if not used
872 * @driver_data: Private driver data for this device
874 * Allocates a new struct tee_device instance. The device is
875 * removed by tee_device_unregister().
877 * @returns a pointer to a 'struct tee_device' or an ERR_PTR on failure
879 struct tee_device *tee_device_alloc(const struct tee_desc *teedesc,
880 struct device *dev,
881 struct tee_shm_pool *pool,
882 void *driver_data)
884 struct tee_device *teedev;
885 void *ret;
886 int rc, max_id;
887 int offs = 0;
889 if (!teedesc || !teedesc->name || !teedesc->ops ||
890 !teedesc->ops->get_version || !teedesc->ops->open ||
891 !teedesc->ops->release || !pool)
892 return ERR_PTR(-EINVAL);
894 teedev = kzalloc(sizeof(*teedev), GFP_KERNEL);
895 if (!teedev) {
896 ret = ERR_PTR(-ENOMEM);
897 goto err;
900 max_id = TEE_NUM_DEVICES / 2;
902 if (teedesc->flags & TEE_DESC_PRIVILEGED) {
903 offs = TEE_NUM_DEVICES / 2;
904 max_id = TEE_NUM_DEVICES;
907 spin_lock(&driver_lock);
908 teedev->id = find_next_zero_bit(dev_mask, max_id, offs);
909 if (teedev->id < max_id)
910 set_bit(teedev->id, dev_mask);
911 spin_unlock(&driver_lock);
913 if (teedev->id >= max_id) {
914 ret = ERR_PTR(-ENOMEM);
915 goto err;
918 snprintf(teedev->name, sizeof(teedev->name), "tee%s%d",
919 teedesc->flags & TEE_DESC_PRIVILEGED ? "priv" : "",
920 teedev->id - offs);
922 teedev->dev.class = &tee_class;
923 teedev->dev.release = tee_release_device;
924 teedev->dev.parent = dev;
926 teedev->dev.devt = MKDEV(MAJOR(tee_devt), teedev->id);
928 rc = dev_set_name(&teedev->dev, "%s", teedev->name);
929 if (rc) {
930 ret = ERR_PTR(rc);
931 goto err_devt;
934 cdev_init(&teedev->cdev, &tee_fops);
935 teedev->cdev.owner = teedesc->owner;
937 dev_set_drvdata(&teedev->dev, driver_data);
938 device_initialize(&teedev->dev);
940 /* 1 as tee_device_unregister() does one final tee_device_put() */
941 teedev->num_users = 1;
942 init_completion(&teedev->c_no_users);
943 mutex_init(&teedev->mutex);
944 idr_init(&teedev->idr);
946 teedev->desc = teedesc;
947 teedev->pool = pool;
949 return teedev;
950 err_devt:
951 unregister_chrdev_region(teedev->dev.devt, 1);
952 err:
953 pr_err("could not register %s driver\n",
954 teedesc->flags & TEE_DESC_PRIVILEGED ? "privileged" : "client");
955 if (teedev && teedev->id < TEE_NUM_DEVICES) {
956 spin_lock(&driver_lock);
957 clear_bit(teedev->id, dev_mask);
958 spin_unlock(&driver_lock);
960 kfree(teedev);
961 return ret;
963 EXPORT_SYMBOL_GPL(tee_device_alloc);
965 void tee_device_set_dev_groups(struct tee_device *teedev,
966 const struct attribute_group **dev_groups)
968 teedev->dev.groups = dev_groups;
970 EXPORT_SYMBOL_GPL(tee_device_set_dev_groups);
972 static ssize_t implementation_id_show(struct device *dev,
973 struct device_attribute *attr, char *buf)
975 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
976 struct tee_ioctl_version_data vers;
978 teedev->desc->ops->get_version(teedev, &vers);
979 return scnprintf(buf, PAGE_SIZE, "%d\n", vers.impl_id);
981 static DEVICE_ATTR_RO(implementation_id);
983 static struct attribute *tee_dev_attrs[] = {
984 &dev_attr_implementation_id.attr,
985 NULL
988 ATTRIBUTE_GROUPS(tee_dev);
990 static const struct class tee_class = {
991 .name = "tee",
992 .dev_groups = tee_dev_groups,
996 * tee_device_register() - Registers a TEE device
997 * @teedev: Device to register
999 * tee_device_unregister() need to be called to remove the @teedev if
1000 * this function fails.
1002 * @returns < 0 on failure
1004 int tee_device_register(struct tee_device *teedev)
1006 int rc;
1008 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
1009 dev_err(&teedev->dev, "attempt to register twice\n");
1010 return -EINVAL;
1013 rc = cdev_device_add(&teedev->cdev, &teedev->dev);
1014 if (rc) {
1015 dev_err(&teedev->dev,
1016 "unable to cdev_device_add() %s, major %d, minor %d, err=%d\n",
1017 teedev->name, MAJOR(teedev->dev.devt),
1018 MINOR(teedev->dev.devt), rc);
1019 return rc;
1022 teedev->flags |= TEE_DEVICE_FLAG_REGISTERED;
1023 return 0;
1025 EXPORT_SYMBOL_GPL(tee_device_register);
1027 void tee_device_put(struct tee_device *teedev)
1029 mutex_lock(&teedev->mutex);
1030 /* Shouldn't put in this state */
1031 if (!WARN_ON(!teedev->desc)) {
1032 teedev->num_users--;
1033 if (!teedev->num_users) {
1034 teedev->desc = NULL;
1035 complete(&teedev->c_no_users);
1038 mutex_unlock(&teedev->mutex);
1041 bool tee_device_get(struct tee_device *teedev)
1043 mutex_lock(&teedev->mutex);
1044 if (!teedev->desc) {
1045 mutex_unlock(&teedev->mutex);
1046 return false;
1048 teedev->num_users++;
1049 mutex_unlock(&teedev->mutex);
1050 return true;
1054 * tee_device_unregister() - Removes a TEE device
1055 * @teedev: Device to unregister
1057 * This function should be called to remove the @teedev even if
1058 * tee_device_register() hasn't been called yet. Does nothing if
1059 * @teedev is NULL.
1061 void tee_device_unregister(struct tee_device *teedev)
1063 if (!teedev)
1064 return;
1066 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED)
1067 cdev_device_del(&teedev->cdev, &teedev->dev);
1069 tee_device_put(teedev);
1070 wait_for_completion(&teedev->c_no_users);
1073 * No need to take a mutex any longer now since teedev->desc was
1074 * set to NULL before teedev->c_no_users was completed.
1077 teedev->pool = NULL;
1079 put_device(&teedev->dev);
1081 EXPORT_SYMBOL_GPL(tee_device_unregister);
1084 * tee_get_drvdata() - Return driver_data pointer
1085 * @teedev: Device containing the driver_data pointer
1086 * @returns the driver_data pointer supplied to tee_device_alloc().
1088 void *tee_get_drvdata(struct tee_device *teedev)
1090 return dev_get_drvdata(&teedev->dev);
1092 EXPORT_SYMBOL_GPL(tee_get_drvdata);
1094 struct match_dev_data {
1095 struct tee_ioctl_version_data *vers;
1096 const void *data;
1097 int (*match)(struct tee_ioctl_version_data *, const void *);
1100 static int match_dev(struct device *dev, const void *data)
1102 const struct match_dev_data *match_data = data;
1103 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
1105 teedev->desc->ops->get_version(teedev, match_data->vers);
1106 return match_data->match(match_data->vers, match_data->data);
1109 struct tee_context *
1110 tee_client_open_context(struct tee_context *start,
1111 int (*match)(struct tee_ioctl_version_data *,
1112 const void *),
1113 const void *data, struct tee_ioctl_version_data *vers)
1115 struct device *dev = NULL;
1116 struct device *put_dev = NULL;
1117 struct tee_context *ctx = NULL;
1118 struct tee_ioctl_version_data v;
1119 struct match_dev_data match_data = { vers ? vers : &v, data, match };
1121 if (start)
1122 dev = &start->teedev->dev;
1124 do {
1125 dev = class_find_device(&tee_class, dev, &match_data, match_dev);
1126 if (!dev) {
1127 ctx = ERR_PTR(-ENOENT);
1128 break;
1131 put_device(put_dev);
1132 put_dev = dev;
1134 ctx = teedev_open(container_of(dev, struct tee_device, dev));
1135 } while (IS_ERR(ctx) && PTR_ERR(ctx) != -ENOMEM);
1137 put_device(put_dev);
1139 * Default behaviour for in kernel client is to not wait for
1140 * tee-supplicant if not present for any requests in this context.
1141 * Also this flag could be configured again before call to
1142 * tee_client_open_session() if any in kernel client requires
1143 * different behaviour.
1145 if (!IS_ERR(ctx))
1146 ctx->supp_nowait = true;
1148 return ctx;
1150 EXPORT_SYMBOL_GPL(tee_client_open_context);
1152 void tee_client_close_context(struct tee_context *ctx)
1154 teedev_close_context(ctx);
1156 EXPORT_SYMBOL_GPL(tee_client_close_context);
1158 void tee_client_get_version(struct tee_context *ctx,
1159 struct tee_ioctl_version_data *vers)
1161 ctx->teedev->desc->ops->get_version(ctx->teedev, vers);
1163 EXPORT_SYMBOL_GPL(tee_client_get_version);
1165 int tee_client_open_session(struct tee_context *ctx,
1166 struct tee_ioctl_open_session_arg *arg,
1167 struct tee_param *param)
1169 if (!ctx->teedev->desc->ops->open_session)
1170 return -EINVAL;
1171 return ctx->teedev->desc->ops->open_session(ctx, arg, param);
1173 EXPORT_SYMBOL_GPL(tee_client_open_session);
1175 int tee_client_close_session(struct tee_context *ctx, u32 session)
1177 if (!ctx->teedev->desc->ops->close_session)
1178 return -EINVAL;
1179 return ctx->teedev->desc->ops->close_session(ctx, session);
1181 EXPORT_SYMBOL_GPL(tee_client_close_session);
1183 int tee_client_system_session(struct tee_context *ctx, u32 session)
1185 if (!ctx->teedev->desc->ops->system_session)
1186 return -EINVAL;
1187 return ctx->teedev->desc->ops->system_session(ctx, session);
1189 EXPORT_SYMBOL_GPL(tee_client_system_session);
1191 int tee_client_invoke_func(struct tee_context *ctx,
1192 struct tee_ioctl_invoke_arg *arg,
1193 struct tee_param *param)
1195 if (!ctx->teedev->desc->ops->invoke_func)
1196 return -EINVAL;
1197 return ctx->teedev->desc->ops->invoke_func(ctx, arg, param);
1199 EXPORT_SYMBOL_GPL(tee_client_invoke_func);
1201 int tee_client_cancel_req(struct tee_context *ctx,
1202 struct tee_ioctl_cancel_arg *arg)
1204 if (!ctx->teedev->desc->ops->cancel_req)
1205 return -EINVAL;
1206 return ctx->teedev->desc->ops->cancel_req(ctx, arg->cancel_id,
1207 arg->session);
1210 static int tee_client_device_match(struct device *dev,
1211 const struct device_driver *drv)
1213 const struct tee_client_device_id *id_table;
1214 struct tee_client_device *tee_device;
1216 id_table = to_tee_client_driver(drv)->id_table;
1217 tee_device = to_tee_client_device(dev);
1219 while (!uuid_is_null(&id_table->uuid)) {
1220 if (uuid_equal(&tee_device->id.uuid, &id_table->uuid))
1221 return 1;
1222 id_table++;
1225 return 0;
1228 static int tee_client_device_uevent(const struct device *dev,
1229 struct kobj_uevent_env *env)
1231 uuid_t *dev_id = &to_tee_client_device(dev)->id.uuid;
1233 return add_uevent_var(env, "MODALIAS=tee:%pUb", dev_id);
1236 const struct bus_type tee_bus_type = {
1237 .name = "tee",
1238 .match = tee_client_device_match,
1239 .uevent = tee_client_device_uevent,
1241 EXPORT_SYMBOL_GPL(tee_bus_type);
1243 static int __init tee_init(void)
1245 int rc;
1247 rc = class_register(&tee_class);
1248 if (rc) {
1249 pr_err("couldn't create class\n");
1250 return rc;
1253 rc = alloc_chrdev_region(&tee_devt, 0, TEE_NUM_DEVICES, "tee");
1254 if (rc) {
1255 pr_err("failed to allocate char dev region\n");
1256 goto out_unreg_class;
1259 rc = bus_register(&tee_bus_type);
1260 if (rc) {
1261 pr_err("failed to register tee bus\n");
1262 goto out_unreg_chrdev;
1265 return 0;
1267 out_unreg_chrdev:
1268 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1269 out_unreg_class:
1270 class_unregister(&tee_class);
1272 return rc;
1275 static void __exit tee_exit(void)
1277 bus_unregister(&tee_bus_type);
1278 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1279 class_unregister(&tee_class);
1282 subsys_initcall(tee_init);
1283 module_exit(tee_exit);
1285 MODULE_AUTHOR("Linaro");
1286 MODULE_DESCRIPTION("TEE Driver");
1287 MODULE_VERSION("1.0");
1288 MODULE_LICENSE("GPL v2");