2 * Copyright (C) 2013 Red Hat
3 * Author: Rob Clark <robdclark@gmail.com>
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
18 #include <linux/sync_file.h>
25 * Cmdstream submission:
28 /* make sure these don't conflict w/ MSM_SUBMIT_BO_x */
29 #define BO_VALID 0x8000 /* is current addr in cmdstream correct/valid? */
30 #define BO_LOCKED 0x4000
31 #define BO_PINNED 0x2000
33 static struct msm_gem_submit
*submit_create(struct drm_device
*dev
,
34 struct msm_gpu
*gpu
, struct msm_gpu_submitqueue
*queue
,
35 uint32_t nr_bos
, uint32_t nr_cmds
)
37 struct msm_gem_submit
*submit
;
38 uint64_t sz
= sizeof(*submit
) + ((u64
)nr_bos
* sizeof(submit
->bos
[0])) +
39 ((u64
)nr_cmds
* sizeof(submit
->cmd
[0]));
44 submit
= kmalloc(sz
, GFP_KERNEL
| __GFP_NOWARN
| __GFP_NORETRY
);
51 submit
->pid
= get_pid(task_pid(current
));
52 submit
->cmd
= (void *)&submit
->bos
[nr_bos
];
53 submit
->queue
= queue
;
54 submit
->ring
= gpu
->rb
[queue
->prio
];
56 /* initially, until copy_from_user() and bo lookup succeeds: */
60 INIT_LIST_HEAD(&submit
->node
);
61 INIT_LIST_HEAD(&submit
->bo_list
);
62 ww_acquire_init(&submit
->ticket
, &reservation_ww_class
);
67 void msm_gem_submit_free(struct msm_gem_submit
*submit
)
69 dma_fence_put(submit
->fence
);
70 list_del(&submit
->node
);
72 msm_submitqueue_put(submit
->queue
);
77 static inline unsigned long __must_check
78 copy_from_user_inatomic(void *to
, const void __user
*from
, unsigned long n
)
80 if (access_ok(VERIFY_READ
, from
, n
))
81 return __copy_from_user_inatomic(to
, from
, n
);
85 static int submit_lookup_objects(struct msm_gem_submit
*submit
,
86 struct drm_msm_gem_submit
*args
, struct drm_file
*file
)
91 spin_lock(&file
->table_lock
);
94 for (i
= 0; i
< args
->nr_bos
; i
++) {
95 struct drm_msm_gem_submit_bo submit_bo
;
96 struct drm_gem_object
*obj
;
97 struct msm_gem_object
*msm_obj
;
98 void __user
*userptr
=
99 u64_to_user_ptr(args
->bos
+ (i
* sizeof(submit_bo
)));
101 /* make sure we don't have garbage flags, in case we hit
102 * error path before flags is initialized:
104 submit
->bos
[i
].flags
= 0;
106 if (copy_from_user_inatomic(&submit_bo
, userptr
, sizeof(submit_bo
))) {
108 spin_unlock(&file
->table_lock
);
109 if (copy_from_user(&submit_bo
, userptr
, sizeof(submit_bo
))) {
113 spin_lock(&file
->table_lock
);
117 if ((submit_bo
.flags
& ~MSM_SUBMIT_BO_FLAGS
) ||
118 !(submit_bo
.flags
& MSM_SUBMIT_BO_FLAGS
)) {
119 DRM_ERROR("invalid flags: %x\n", submit_bo
.flags
);
124 submit
->bos
[i
].flags
= submit_bo
.flags
;
125 /* in validate_objects() we figure out if this is true: */
126 submit
->bos
[i
].iova
= submit_bo
.presumed
;
128 /* normally use drm_gem_object_lookup(), but for bulk lookup
129 * all under single table_lock just hit object_idr directly:
131 obj
= idr_find(&file
->object_idr
, submit_bo
.handle
);
133 DRM_ERROR("invalid handle %u at index %u\n", submit_bo
.handle
, i
);
138 msm_obj
= to_msm_bo(obj
);
140 if (!list_empty(&msm_obj
->submit_entry
)) {
141 DRM_ERROR("handle %u at index %u already on submit list\n",
142 submit_bo
.handle
, i
);
147 drm_gem_object_reference(obj
);
149 submit
->bos
[i
].obj
= msm_obj
;
151 list_add_tail(&msm_obj
->submit_entry
, &submit
->bo_list
);
156 spin_unlock(&file
->table_lock
);
164 static void submit_unlock_unpin_bo(struct msm_gem_submit
*submit
,
167 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
169 if (submit
->bos
[i
].flags
& BO_PINNED
)
170 msm_gem_put_iova(&msm_obj
->base
, submit
->gpu
->aspace
);
172 if (submit
->bos
[i
].flags
& BO_LOCKED
)
173 ww_mutex_unlock(&msm_obj
->resv
->lock
);
175 if (backoff
&& !(submit
->bos
[i
].flags
& BO_VALID
))
176 submit
->bos
[i
].iova
= 0;
178 submit
->bos
[i
].flags
&= ~(BO_LOCKED
| BO_PINNED
);
181 /* This is where we make sure all the bo's are reserved and pin'd: */
182 static int submit_lock_objects(struct msm_gem_submit
*submit
)
184 int contended
, slow_locked
= -1, i
, ret
= 0;
187 for (i
= 0; i
< submit
->nr_bos
; i
++) {
188 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
190 if (slow_locked
== i
)
195 if (!(submit
->bos
[i
].flags
& BO_LOCKED
)) {
196 ret
= ww_mutex_lock_interruptible(&msm_obj
->resv
->lock
,
200 submit
->bos
[i
].flags
|= BO_LOCKED
;
204 ww_acquire_done(&submit
->ticket
);
210 submit_unlock_unpin_bo(submit
, i
, true);
213 submit_unlock_unpin_bo(submit
, slow_locked
, true);
215 if (ret
== -EDEADLK
) {
216 struct msm_gem_object
*msm_obj
= submit
->bos
[contended
].obj
;
217 /* we lost out in a seqno race, lock and retry.. */
218 ret
= ww_mutex_lock_slow_interruptible(&msm_obj
->resv
->lock
,
221 submit
->bos
[contended
].flags
|= BO_LOCKED
;
222 slow_locked
= contended
;
230 static int submit_fence_sync(struct msm_gem_submit
*submit
, bool no_implicit
)
234 for (i
= 0; i
< submit
->nr_bos
; i
++) {
235 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
236 bool write
= submit
->bos
[i
].flags
& MSM_SUBMIT_BO_WRITE
;
239 /* NOTE: _reserve_shared() must happen before
240 * _add_shared_fence(), which makes this a slightly
241 * strange place to call it. OTOH this is a
242 * convenient can-fail point to hook it in.
244 ret
= reservation_object_reserve_shared(msm_obj
->resv
);
252 ret
= msm_gem_sync_object(&msm_obj
->base
, submit
->ring
->fctx
,
261 static int submit_pin_objects(struct msm_gem_submit
*submit
)
265 submit
->valid
= true;
267 for (i
= 0; i
< submit
->nr_bos
; i
++) {
268 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
271 /* if locking succeeded, pin bo: */
272 ret
= msm_gem_get_iova(&msm_obj
->base
,
273 submit
->gpu
->aspace
, &iova
);
278 submit
->bos
[i
].flags
|= BO_PINNED
;
280 if (iova
== submit
->bos
[i
].iova
) {
281 submit
->bos
[i
].flags
|= BO_VALID
;
283 submit
->bos
[i
].iova
= iova
;
284 /* iova changed, so address in cmdstream is not valid: */
285 submit
->bos
[i
].flags
&= ~BO_VALID
;
286 submit
->valid
= false;
293 static int submit_bo(struct msm_gem_submit
*submit
, uint32_t idx
,
294 struct msm_gem_object
**obj
, uint64_t *iova
, bool *valid
)
296 if (idx
>= submit
->nr_bos
) {
297 DRM_ERROR("invalid buffer index: %u (out of %u)\n",
298 idx
, submit
->nr_bos
);
303 *obj
= submit
->bos
[idx
].obj
;
305 *iova
= submit
->bos
[idx
].iova
;
307 *valid
= !!(submit
->bos
[idx
].flags
& BO_VALID
);
312 /* process the reloc's and patch up the cmdstream as needed: */
313 static int submit_reloc(struct msm_gem_submit
*submit
, struct msm_gem_object
*obj
,
314 uint32_t offset
, uint32_t nr_relocs
, uint64_t relocs
)
316 uint32_t i
, last_offset
= 0;
321 DRM_ERROR("non-aligned cmdstream buffer: %u\n", offset
);
325 /* For now, just map the entire thing. Eventually we probably
326 * to do it page-by-page, w/ kmap() if not vmap()d..
328 ptr
= msm_gem_get_vaddr(&obj
->base
);
332 DBG("failed to map: %d", ret
);
336 for (i
= 0; i
< nr_relocs
; i
++) {
337 struct drm_msm_gem_submit_reloc submit_reloc
;
338 void __user
*userptr
=
339 u64_to_user_ptr(relocs
+ (i
* sizeof(submit_reloc
)));
344 if (copy_from_user(&submit_reloc
, userptr
, sizeof(submit_reloc
))) {
349 if (submit_reloc
.submit_offset
% 4) {
350 DRM_ERROR("non-aligned reloc offset: %u\n",
351 submit_reloc
.submit_offset
);
356 /* offset in dwords: */
357 off
= submit_reloc
.submit_offset
/ 4;
359 if ((off
>= (obj
->base
.size
/ 4)) ||
360 (off
< last_offset
)) {
361 DRM_ERROR("invalid offset %u at reloc %u\n", off
, i
);
366 ret
= submit_bo(submit
, submit_reloc
.reloc_idx
, NULL
, &iova
, &valid
);
373 iova
+= submit_reloc
.reloc_offset
;
375 if (submit_reloc
.shift
< 0)
376 iova
>>= -submit_reloc
.shift
;
378 iova
<<= submit_reloc
.shift
;
380 ptr
[off
] = iova
| submit_reloc
.or;
386 msm_gem_put_vaddr(&obj
->base
);
391 static void submit_cleanup(struct msm_gem_submit
*submit
)
395 for (i
= 0; i
< submit
->nr_bos
; i
++) {
396 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
397 submit_unlock_unpin_bo(submit
, i
, false);
398 list_del_init(&msm_obj
->submit_entry
);
399 drm_gem_object_unreference(&msm_obj
->base
);
402 ww_acquire_fini(&submit
->ticket
);
405 int msm_ioctl_gem_submit(struct drm_device
*dev
, void *data
,
406 struct drm_file
*file
)
408 struct msm_drm_private
*priv
= dev
->dev_private
;
409 struct drm_msm_gem_submit
*args
= data
;
410 struct msm_file_private
*ctx
= file
->driver_priv
;
411 struct msm_gem_submit
*submit
;
412 struct msm_gpu
*gpu
= priv
->gpu
;
413 struct dma_fence
*in_fence
= NULL
;
414 struct sync_file
*sync_file
= NULL
;
415 struct msm_gpu_submitqueue
*queue
;
416 struct msm_ringbuffer
*ring
;
417 int out_fence_fd
= -1;
424 /* for now, we just have 3d pipe.. eventually this would need to
425 * be more clever to dispatch to appropriate gpu module:
427 if (MSM_PIPE_ID(args
->flags
) != MSM_PIPE_3D0
)
430 if (MSM_PIPE_FLAGS(args
->flags
) & ~MSM_SUBMIT_FLAGS
)
433 queue
= msm_submitqueue_get(ctx
, args
->queueid
);
437 ring
= gpu
->rb
[queue
->prio
];
439 if (args
->flags
& MSM_SUBMIT_FENCE_FD_IN
) {
440 in_fence
= sync_file_get_fence(args
->fence_fd
);
446 * Wait if the fence is from a foreign context, or if the fence
447 * array contains any fence from a foreign context.
449 if (!dma_fence_match_context(in_fence
, ring
->fctx
->context
)) {
450 ret
= dma_fence_wait(in_fence
, true);
456 ret
= mutex_lock_interruptible(&dev
->struct_mutex
);
460 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
461 out_fence_fd
= get_unused_fd_flags(O_CLOEXEC
);
462 if (out_fence_fd
< 0) {
468 submit
= submit_create(dev
, gpu
, queue
, args
->nr_bos
, args
->nr_cmds
);
474 ret
= submit_lookup_objects(submit
, args
, file
);
478 ret
= submit_lock_objects(submit
);
482 ret
= submit_fence_sync(submit
, !!(args
->flags
& MSM_SUBMIT_NO_IMPLICIT
));
486 ret
= submit_pin_objects(submit
);
490 for (i
= 0; i
< args
->nr_cmds
; i
++) {
491 struct drm_msm_gem_submit_cmd submit_cmd
;
492 void __user
*userptr
=
493 u64_to_user_ptr(args
->cmds
+ (i
* sizeof(submit_cmd
)));
494 struct msm_gem_object
*msm_obj
;
497 ret
= copy_from_user(&submit_cmd
, userptr
, sizeof(submit_cmd
));
503 /* validate input from userspace: */
504 switch (submit_cmd
.type
) {
505 case MSM_SUBMIT_CMD_BUF
:
506 case MSM_SUBMIT_CMD_IB_TARGET_BUF
:
507 case MSM_SUBMIT_CMD_CTX_RESTORE_BUF
:
510 DRM_ERROR("invalid type: %08x\n", submit_cmd
.type
);
515 ret
= submit_bo(submit
, submit_cmd
.submit_idx
,
516 &msm_obj
, &iova
, NULL
);
520 if (submit_cmd
.size
% 4) {
521 DRM_ERROR("non-aligned cmdstream buffer size: %u\n",
527 if (!submit_cmd
.size
||
528 ((submit_cmd
.size
+ submit_cmd
.submit_offset
) >
529 msm_obj
->base
.size
)) {
530 DRM_ERROR("invalid cmdstream size: %u\n", submit_cmd
.size
);
535 submit
->cmd
[i
].type
= submit_cmd
.type
;
536 submit
->cmd
[i
].size
= submit_cmd
.size
/ 4;
537 submit
->cmd
[i
].iova
= iova
+ submit_cmd
.submit_offset
;
538 submit
->cmd
[i
].idx
= submit_cmd
.submit_idx
;
543 ret
= submit_reloc(submit
, msm_obj
, submit_cmd
.submit_offset
,
544 submit_cmd
.nr_relocs
, submit_cmd
.relocs
);
551 submit
->fence
= msm_fence_alloc(ring
->fctx
);
552 if (IS_ERR(submit
->fence
)) {
553 ret
= PTR_ERR(submit
->fence
);
554 submit
->fence
= NULL
;
558 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
559 sync_file
= sync_file_create(submit
->fence
);
566 msm_gpu_submit(gpu
, submit
, ctx
);
568 args
->fence
= submit
->fence
->seqno
;
570 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
571 fd_install(out_fence_fd
, sync_file
->file
);
572 args
->fence_fd
= out_fence_fd
;
577 dma_fence_put(in_fence
);
578 submit_cleanup(submit
);
580 msm_gem_submit_free(submit
);
582 if (ret
&& (out_fence_fd
>= 0))
583 put_unused_fd(out_fence_fd
);
584 mutex_unlock(&dev
->struct_mutex
);