2 * Copyright (C) 2013 Red Hat
3 * Author: Rob Clark <robdclark@gmail.com>
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
18 #include <linux/sync_file.h>
25 * Cmdstream submission:
28 /* make sure these don't conflict w/ MSM_SUBMIT_BO_x */
29 #define BO_VALID 0x8000 /* is current addr in cmdstream correct/valid? */
30 #define BO_LOCKED 0x4000
31 #define BO_PINNED 0x2000
33 static struct msm_gem_submit
*submit_create(struct drm_device
*dev
,
34 struct msm_gpu
*gpu
, int nr_bos
, int nr_cmds
)
36 struct msm_gem_submit
*submit
;
37 int sz
= sizeof(*submit
) + (nr_bos
* sizeof(submit
->bos
[0])) +
38 (nr_cmds
* sizeof(*submit
->cmd
));
40 submit
= kmalloc(sz
, GFP_TEMPORARY
| __GFP_NOWARN
| __GFP_NORETRY
);
47 submit
->pid
= get_pid(task_pid(current
));
48 submit
->cmd
= (void *)&submit
->bos
[nr_bos
];
50 /* initially, until copy_from_user() and bo lookup succeeds: */
54 INIT_LIST_HEAD(&submit
->node
);
55 INIT_LIST_HEAD(&submit
->bo_list
);
56 ww_acquire_init(&submit
->ticket
, &reservation_ww_class
);
61 void msm_gem_submit_free(struct msm_gem_submit
*submit
)
63 dma_fence_put(submit
->fence
);
64 list_del(&submit
->node
);
69 static inline unsigned long __must_check
70 copy_from_user_inatomic(void *to
, const void __user
*from
, unsigned long n
)
72 if (access_ok(VERIFY_READ
, from
, n
))
73 return __copy_from_user_inatomic(to
, from
, n
);
77 static int submit_lookup_objects(struct msm_gem_submit
*submit
,
78 struct drm_msm_gem_submit
*args
, struct drm_file
*file
)
83 spin_lock(&file
->table_lock
);
86 for (i
= 0; i
< args
->nr_bos
; i
++) {
87 struct drm_msm_gem_submit_bo submit_bo
;
88 struct drm_gem_object
*obj
;
89 struct msm_gem_object
*msm_obj
;
90 void __user
*userptr
=
91 u64_to_user_ptr(args
->bos
+ (i
* sizeof(submit_bo
)));
93 /* make sure we don't have garbage flags, in case we hit
94 * error path before flags is initialized:
96 submit
->bos
[i
].flags
= 0;
98 ret
= copy_from_user_inatomic(&submit_bo
, userptr
, sizeof(submit_bo
));
101 spin_unlock(&file
->table_lock
);
102 ret
= copy_from_user(&submit_bo
, userptr
, sizeof(submit_bo
));
105 spin_lock(&file
->table_lock
);
109 if ((submit_bo
.flags
& ~MSM_SUBMIT_BO_FLAGS
) ||
110 !(submit_bo
.flags
& MSM_SUBMIT_BO_FLAGS
)) {
111 DRM_ERROR("invalid flags: %x\n", submit_bo
.flags
);
116 submit
->bos
[i
].flags
= submit_bo
.flags
;
117 /* in validate_objects() we figure out if this is true: */
118 submit
->bos
[i
].iova
= submit_bo
.presumed
;
120 /* normally use drm_gem_object_lookup(), but for bulk lookup
121 * all under single table_lock just hit object_idr directly:
123 obj
= idr_find(&file
->object_idr
, submit_bo
.handle
);
125 DRM_ERROR("invalid handle %u at index %u\n", submit_bo
.handle
, i
);
130 msm_obj
= to_msm_bo(obj
);
132 if (!list_empty(&msm_obj
->submit_entry
)) {
133 DRM_ERROR("handle %u at index %u already on submit list\n",
134 submit_bo
.handle
, i
);
139 drm_gem_object_reference(obj
);
141 submit
->bos
[i
].obj
= msm_obj
;
143 list_add_tail(&msm_obj
->submit_entry
, &submit
->bo_list
);
148 spin_unlock(&file
->table_lock
);
156 static void submit_unlock_unpin_bo(struct msm_gem_submit
*submit
, int i
)
158 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
160 if (submit
->bos
[i
].flags
& BO_PINNED
)
161 msm_gem_put_iova(&msm_obj
->base
, submit
->gpu
->id
);
163 if (submit
->bos
[i
].flags
& BO_LOCKED
)
164 ww_mutex_unlock(&msm_obj
->resv
->lock
);
166 if (!(submit
->bos
[i
].flags
& BO_VALID
))
167 submit
->bos
[i
].iova
= 0;
169 submit
->bos
[i
].flags
&= ~(BO_LOCKED
| BO_PINNED
);
172 /* This is where we make sure all the bo's are reserved and pin'd: */
173 static int submit_lock_objects(struct msm_gem_submit
*submit
)
175 int contended
, slow_locked
= -1, i
, ret
= 0;
178 for (i
= 0; i
< submit
->nr_bos
; i
++) {
179 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
181 if (slow_locked
== i
)
186 if (!(submit
->bos
[i
].flags
& BO_LOCKED
)) {
187 ret
= ww_mutex_lock_interruptible(&msm_obj
->resv
->lock
,
191 submit
->bos
[i
].flags
|= BO_LOCKED
;
195 ww_acquire_done(&submit
->ticket
);
201 submit_unlock_unpin_bo(submit
, i
);
204 submit_unlock_unpin_bo(submit
, slow_locked
);
206 if (ret
== -EDEADLK
) {
207 struct msm_gem_object
*msm_obj
= submit
->bos
[contended
].obj
;
208 /* we lost out in a seqno race, lock and retry.. */
209 ret
= ww_mutex_lock_slow_interruptible(&msm_obj
->resv
->lock
,
212 submit
->bos
[contended
].flags
|= BO_LOCKED
;
213 slow_locked
= contended
;
221 static int submit_fence_sync(struct msm_gem_submit
*submit
)
225 for (i
= 0; i
< submit
->nr_bos
; i
++) {
226 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
227 bool write
= submit
->bos
[i
].flags
& MSM_SUBMIT_BO_WRITE
;
229 ret
= msm_gem_sync_object(&msm_obj
->base
, submit
->gpu
->fctx
, write
);
237 static int submit_pin_objects(struct msm_gem_submit
*submit
)
241 submit
->valid
= true;
243 for (i
= 0; i
< submit
->nr_bos
; i
++) {
244 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
247 /* if locking succeeded, pin bo: */
248 ret
= msm_gem_get_iova_locked(&msm_obj
->base
,
249 submit
->gpu
->id
, &iova
);
254 submit
->bos
[i
].flags
|= BO_PINNED
;
256 if (iova
== submit
->bos
[i
].iova
) {
257 submit
->bos
[i
].flags
|= BO_VALID
;
259 submit
->bos
[i
].iova
= iova
;
260 /* iova changed, so address in cmdstream is not valid: */
261 submit
->bos
[i
].flags
&= ~BO_VALID
;
262 submit
->valid
= false;
269 static int submit_bo(struct msm_gem_submit
*submit
, uint32_t idx
,
270 struct msm_gem_object
**obj
, uint64_t *iova
, bool *valid
)
272 if (idx
>= submit
->nr_bos
) {
273 DRM_ERROR("invalid buffer index: %u (out of %u)\n",
274 idx
, submit
->nr_bos
);
279 *obj
= submit
->bos
[idx
].obj
;
281 *iova
= submit
->bos
[idx
].iova
;
283 *valid
= !!(submit
->bos
[idx
].flags
& BO_VALID
);
288 /* process the reloc's and patch up the cmdstream as needed: */
289 static int submit_reloc(struct msm_gem_submit
*submit
, struct msm_gem_object
*obj
,
290 uint32_t offset
, uint32_t nr_relocs
, uint64_t relocs
)
292 uint32_t i
, last_offset
= 0;
297 DRM_ERROR("non-aligned cmdstream buffer: %u\n", offset
);
301 /* For now, just map the entire thing. Eventually we probably
302 * to do it page-by-page, w/ kmap() if not vmap()d..
304 ptr
= msm_gem_get_vaddr_locked(&obj
->base
);
308 DBG("failed to map: %d", ret
);
312 for (i
= 0; i
< nr_relocs
; i
++) {
313 struct drm_msm_gem_submit_reloc submit_reloc
;
314 void __user
*userptr
=
315 u64_to_user_ptr(relocs
+ (i
* sizeof(submit_reloc
)));
320 ret
= copy_from_user(&submit_reloc
, userptr
, sizeof(submit_reloc
));
324 if (submit_reloc
.submit_offset
% 4) {
325 DRM_ERROR("non-aligned reloc offset: %u\n",
326 submit_reloc
.submit_offset
);
331 /* offset in dwords: */
332 off
= submit_reloc
.submit_offset
/ 4;
334 if ((off
>= (obj
->base
.size
/ 4)) ||
335 (off
< last_offset
)) {
336 DRM_ERROR("invalid offset %u at reloc %u\n", off
, i
);
341 ret
= submit_bo(submit
, submit_reloc
.reloc_idx
, NULL
, &iova
, &valid
);
348 iova
+= submit_reloc
.reloc_offset
;
350 if (submit_reloc
.shift
< 0)
351 iova
>>= -submit_reloc
.shift
;
353 iova
<<= submit_reloc
.shift
;
355 ptr
[off
] = iova
| submit_reloc
.or;
361 msm_gem_put_vaddr_locked(&obj
->base
);
366 static void submit_cleanup(struct msm_gem_submit
*submit
)
370 for (i
= 0; i
< submit
->nr_bos
; i
++) {
371 struct msm_gem_object
*msm_obj
= submit
->bos
[i
].obj
;
372 submit_unlock_unpin_bo(submit
, i
);
373 list_del_init(&msm_obj
->submit_entry
);
374 drm_gem_object_unreference(&msm_obj
->base
);
377 ww_acquire_fini(&submit
->ticket
);
380 int msm_ioctl_gem_submit(struct drm_device
*dev
, void *data
,
381 struct drm_file
*file
)
383 struct msm_drm_private
*priv
= dev
->dev_private
;
384 struct drm_msm_gem_submit
*args
= data
;
385 struct msm_file_private
*ctx
= file
->driver_priv
;
386 struct msm_gem_submit
*submit
;
387 struct msm_gpu
*gpu
= priv
->gpu
;
388 struct dma_fence
*in_fence
= NULL
;
389 struct sync_file
*sync_file
= NULL
;
390 int out_fence_fd
= -1;
397 /* for now, we just have 3d pipe.. eventually this would need to
398 * be more clever to dispatch to appropriate gpu module:
400 if (MSM_PIPE_ID(args
->flags
) != MSM_PIPE_3D0
)
403 if (MSM_PIPE_FLAGS(args
->flags
) & ~MSM_SUBMIT_FLAGS
)
406 ret
= mutex_lock_interruptible(&dev
->struct_mutex
);
410 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
411 out_fence_fd
= get_unused_fd_flags(O_CLOEXEC
);
412 if (out_fence_fd
< 0) {
417 priv
->struct_mutex_task
= current
;
419 submit
= submit_create(dev
, gpu
, args
->nr_bos
, args
->nr_cmds
);
425 ret
= submit_lookup_objects(submit
, args
, file
);
429 ret
= submit_lock_objects(submit
);
433 if (args
->flags
& MSM_SUBMIT_FENCE_FD_IN
) {
434 in_fence
= sync_file_get_fence(args
->fence_fd
);
441 /* TODO if we get an array-fence due to userspace merging multiple
442 * fences, we need a way to determine if all the backing fences
443 * are from our own context..
446 if (in_fence
->context
!= gpu
->fctx
->context
) {
447 ret
= dma_fence_wait(in_fence
, true);
454 if (!(args
->fence
& MSM_SUBMIT_NO_IMPLICIT
)) {
455 ret
= submit_fence_sync(submit
);
460 ret
= submit_pin_objects(submit
);
464 for (i
= 0; i
< args
->nr_cmds
; i
++) {
465 struct drm_msm_gem_submit_cmd submit_cmd
;
466 void __user
*userptr
=
467 u64_to_user_ptr(args
->cmds
+ (i
* sizeof(submit_cmd
)));
468 struct msm_gem_object
*msm_obj
;
471 ret
= copy_from_user(&submit_cmd
, userptr
, sizeof(submit_cmd
));
477 /* validate input from userspace: */
478 switch (submit_cmd
.type
) {
479 case MSM_SUBMIT_CMD_BUF
:
480 case MSM_SUBMIT_CMD_IB_TARGET_BUF
:
481 case MSM_SUBMIT_CMD_CTX_RESTORE_BUF
:
484 DRM_ERROR("invalid type: %08x\n", submit_cmd
.type
);
489 ret
= submit_bo(submit
, submit_cmd
.submit_idx
,
490 &msm_obj
, &iova
, NULL
);
494 if (submit_cmd
.size
% 4) {
495 DRM_ERROR("non-aligned cmdstream buffer size: %u\n",
501 if ((submit_cmd
.size
+ submit_cmd
.submit_offset
) >=
502 msm_obj
->base
.size
) {
503 DRM_ERROR("invalid cmdstream size: %u\n", submit_cmd
.size
);
508 submit
->cmd
[i
].type
= submit_cmd
.type
;
509 submit
->cmd
[i
].size
= submit_cmd
.size
/ 4;
510 submit
->cmd
[i
].iova
= iova
+ submit_cmd
.submit_offset
;
511 submit
->cmd
[i
].idx
= submit_cmd
.submit_idx
;
516 ret
= submit_reloc(submit
, msm_obj
, submit_cmd
.submit_offset
,
517 submit_cmd
.nr_relocs
, submit_cmd
.relocs
);
524 submit
->fence
= msm_fence_alloc(gpu
->fctx
);
525 if (IS_ERR(submit
->fence
)) {
526 ret
= PTR_ERR(submit
->fence
);
527 submit
->fence
= NULL
;
531 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
532 sync_file
= sync_file_create(submit
->fence
);
539 msm_gpu_submit(gpu
, submit
, ctx
);
541 args
->fence
= submit
->fence
->seqno
;
543 if (args
->flags
& MSM_SUBMIT_FENCE_FD_OUT
) {
544 fd_install(out_fence_fd
, sync_file
->file
);
545 args
->fence_fd
= out_fence_fd
;
550 dma_fence_put(in_fence
);
551 submit_cleanup(submit
);
553 msm_gem_submit_free(submit
);
555 if (ret
&& (out_fence_fd
>= 0))
556 put_unused_fd(out_fence_fd
);
557 priv
->struct_mutex_task
= NULL
;
558 mutex_unlock(&dev
->struct_mutex
);