2 * Copyright 2012 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
24 #include <nvif/client.h>
25 #include <nvif/driver.h>
26 #include <nvif/ioctl.h>
27 #include <nvif/class.h>
29 #include "nouveau_drm.h"
30 #include "nouveau_dma.h"
31 #include "nouveau_gem.h"
32 #include "nouveau_chan.h"
33 #include "nouveau_abi16.h"
35 struct nouveau_abi16
*
36 nouveau_abi16_get(struct drm_file
*file_priv
, struct drm_device
*dev
)
38 struct nouveau_cli
*cli
= nouveau_cli(file_priv
);
39 mutex_lock(&cli
->mutex
);
41 struct nouveau_abi16
*abi16
;
42 cli
->abi16
= abi16
= kzalloc(sizeof(*abi16
), GFP_KERNEL
);
44 struct nv_device_v0 args
= {
48 INIT_LIST_HEAD(&abi16
->channels
);
50 /* allocate device object targeting client's default
51 * device (ie. the one that belongs to the fd it
54 if (nvif_device_init(&cli
->base
.base
, NULL
,
55 NOUVEAU_ABI16_DEVICE
, NV_DEVICE
,
64 mutex_unlock(&cli
->mutex
);
70 nouveau_abi16_put(struct nouveau_abi16
*abi16
, int ret
)
72 struct nouveau_cli
*cli
= (void *)nvif_client(&abi16
->device
.base
);
73 mutex_unlock(&cli
->mutex
);
78 nouveau_abi16_swclass(struct nouveau_drm
*drm
)
80 switch (drm
->device
.info
.family
) {
81 case NV_DEVICE_INFO_V0_TNT
:
83 case NV_DEVICE_INFO_V0_CELSIUS
:
84 case NV_DEVICE_INFO_V0_KELVIN
:
85 case NV_DEVICE_INFO_V0_RANKINE
:
86 case NV_DEVICE_INFO_V0_CURIE
:
88 case NV_DEVICE_INFO_V0_TESLA
:
90 case NV_DEVICE_INFO_V0_FERMI
:
91 case NV_DEVICE_INFO_V0_KEPLER
:
92 case NV_DEVICE_INFO_V0_MAXWELL
:
100 nouveau_abi16_ntfy_fini(struct nouveau_abi16_chan
*chan
,
101 struct nouveau_abi16_ntfy
*ntfy
)
103 nvkm_mm_free(&chan
->heap
, &ntfy
->node
);
104 list_del(&ntfy
->head
);
109 nouveau_abi16_chan_fini(struct nouveau_abi16
*abi16
,
110 struct nouveau_abi16_chan
*chan
)
112 struct nouveau_abi16_ntfy
*ntfy
, *temp
;
114 /* wait for all activity to stop before releasing notify object, which
115 * may be still in use */
116 if (chan
->chan
&& chan
->ntfy
)
117 nouveau_channel_idle(chan
->chan
);
119 /* cleanup notifier state */
120 list_for_each_entry_safe(ntfy
, temp
, &chan
->notifiers
, head
) {
121 nouveau_abi16_ntfy_fini(chan
, ntfy
);
125 nouveau_bo_vma_del(chan
->ntfy
, &chan
->ntfy_vma
);
126 nouveau_bo_unpin(chan
->ntfy
);
127 drm_gem_object_unreference_unlocked(&chan
->ntfy
->gem
);
130 if (chan
->heap
.block_size
)
131 nvkm_mm_fini(&chan
->heap
);
133 /* destroy channel object, all children will be killed too */
135 abi16
->handles
&= ~(1ULL << (chan
->chan
->object
->handle
& 0xffff));
136 nouveau_channel_del(&chan
->chan
);
139 list_del(&chan
->head
);
144 nouveau_abi16_fini(struct nouveau_abi16
*abi16
)
146 struct nouveau_cli
*cli
= (void *)nvif_client(&abi16
->device
.base
);
147 struct nouveau_abi16_chan
*chan
, *temp
;
149 /* cleanup channels */
150 list_for_each_entry_safe(chan
, temp
, &abi16
->channels
, head
) {
151 nouveau_abi16_chan_fini(abi16
, chan
);
154 /* destroy the device object */
155 nvif_device_fini(&abi16
->device
);
162 nouveau_abi16_ioctl_getparam(ABI16_IOCTL_ARGS
)
164 struct nouveau_cli
*cli
= nouveau_cli(file_priv
);
165 struct nouveau_drm
*drm
= nouveau_drm(dev
);
166 struct nvif_device
*device
= &drm
->device
;
167 struct nvkm_timer
*ptimer
= nvxx_timer(device
);
168 struct nvkm_gr
*gr
= nvxx_gr(device
);
169 struct drm_nouveau_getparam
*getparam
= data
;
171 switch (getparam
->param
) {
172 case NOUVEAU_GETPARAM_CHIPSET_ID
:
173 getparam
->value
= device
->info
.chipset
;
175 case NOUVEAU_GETPARAM_PCI_VENDOR
:
176 if (nv_device_is_pci(nvxx_device(device
)))
177 getparam
->value
= dev
->pdev
->vendor
;
181 case NOUVEAU_GETPARAM_PCI_DEVICE
:
182 if (nv_device_is_pci(nvxx_device(device
)))
183 getparam
->value
= dev
->pdev
->device
;
187 case NOUVEAU_GETPARAM_BUS_TYPE
:
188 if (!nv_device_is_pci(nvxx_device(device
)))
191 if (drm_pci_device_is_agp(dev
))
194 if (!pci_is_pcie(dev
->pdev
))
199 case NOUVEAU_GETPARAM_FB_SIZE
:
200 getparam
->value
= drm
->gem
.vram_available
;
202 case NOUVEAU_GETPARAM_AGP_SIZE
:
203 getparam
->value
= drm
->gem
.gart_available
;
205 case NOUVEAU_GETPARAM_VM_VRAM_BASE
:
206 getparam
->value
= 0; /* deprecated */
208 case NOUVEAU_GETPARAM_PTIMER_TIME
:
209 getparam
->value
= ptimer
->read(ptimer
);
211 case NOUVEAU_GETPARAM_HAS_BO_USAGE
:
214 case NOUVEAU_GETPARAM_HAS_PAGEFLIP
:
217 case NOUVEAU_GETPARAM_GRAPH_UNITS
:
218 getparam
->value
= gr
->units
? gr
->units(gr
) : 0;
221 NV_PRINTK(debug
, cli
, "unknown parameter %lld\n", getparam
->param
);
229 nouveau_abi16_ioctl_setparam(ABI16_IOCTL_ARGS
)
235 nouveau_abi16_ioctl_channel_alloc(ABI16_IOCTL_ARGS
)
237 struct drm_nouveau_channel_alloc
*init
= data
;
238 struct nouveau_cli
*cli
= nouveau_cli(file_priv
);
239 struct nouveau_drm
*drm
= nouveau_drm(dev
);
240 struct nouveau_abi16
*abi16
= nouveau_abi16_get(file_priv
, dev
);
241 struct nouveau_abi16_chan
*chan
;
242 struct nvif_device
*device
;
245 if (unlikely(!abi16
))
249 return nouveau_abi16_put(abi16
, -ENODEV
);
251 device
= &abi16
->device
;
253 /* hack to allow channel engine type specification on kepler */
254 if (device
->info
.family
>= NV_DEVICE_INFO_V0_KEPLER
) {
255 if (init
->fb_ctxdma_handle
!= ~0)
256 init
->fb_ctxdma_handle
= KEPLER_CHANNEL_GPFIFO_A_V0_ENGINE_GR
;
258 init
->fb_ctxdma_handle
= init
->tt_ctxdma_handle
;
260 /* allow flips to be executed if this is a graphics channel */
261 init
->tt_ctxdma_handle
= 0;
262 if (init
->fb_ctxdma_handle
== KEPLER_CHANNEL_GPFIFO_A_V0_ENGINE_GR
)
263 init
->tt_ctxdma_handle
= 1;
266 if (init
->fb_ctxdma_handle
== ~0 || init
->tt_ctxdma_handle
== ~0)
267 return nouveau_abi16_put(abi16
, -EINVAL
);
269 /* allocate "abi16 channel" data and make up a handle for it */
270 init
->channel
= __ffs64(~abi16
->handles
);
271 if (~abi16
->handles
== 0)
272 return nouveau_abi16_put(abi16
, -ENOSPC
);
274 chan
= kzalloc(sizeof(*chan
), GFP_KERNEL
);
276 return nouveau_abi16_put(abi16
, -ENOMEM
);
278 INIT_LIST_HEAD(&chan
->notifiers
);
279 list_add(&chan
->head
, &abi16
->channels
);
280 abi16
->handles
|= (1ULL << init
->channel
);
282 /* create channel object and initialise dma and fence management */
283 ret
= nouveau_channel_new(drm
, device
,
284 NOUVEAU_ABI16_CHAN(init
->channel
),
285 init
->fb_ctxdma_handle
,
286 init
->tt_ctxdma_handle
, &chan
->chan
);
290 if (device
->info
.family
>= NV_DEVICE_INFO_V0_TESLA
)
291 init
->pushbuf_domains
= NOUVEAU_GEM_DOMAIN_VRAM
|
292 NOUVEAU_GEM_DOMAIN_GART
;
294 if (chan
->chan
->push
.buffer
->bo
.mem
.mem_type
== TTM_PL_VRAM
)
295 init
->pushbuf_domains
= NOUVEAU_GEM_DOMAIN_VRAM
;
297 init
->pushbuf_domains
= NOUVEAU_GEM_DOMAIN_GART
;
299 if (device
->info
.family
< NV_DEVICE_INFO_V0_CELSIUS
) {
300 init
->subchan
[0].handle
= 0x00000000;
301 init
->subchan
[0].grclass
= 0x0000;
302 init
->subchan
[1].handle
= chan
->chan
->nvsw
.handle
;
303 init
->subchan
[1].grclass
= 0x506e;
304 init
->nr_subchan
= 2;
307 /* Named memory object area */
308 ret
= nouveau_gem_new(dev
, PAGE_SIZE
, 0, NOUVEAU_GEM_DOMAIN_GART
,
311 ret
= nouveau_bo_pin(chan
->ntfy
, TTM_PL_FLAG_TT
, false);
315 if (device
->info
.family
>= NV_DEVICE_INFO_V0_TESLA
) {
316 ret
= nouveau_bo_vma_add(chan
->ntfy
, cli
->vm
,
322 ret
= drm_gem_handle_create(file_priv
, &chan
->ntfy
->gem
,
323 &init
->notifier_handle
);
327 ret
= nvkm_mm_init(&chan
->heap
, 0, PAGE_SIZE
, 1);
330 nouveau_abi16_chan_fini(abi16
, chan
);
331 return nouveau_abi16_put(abi16
, ret
);
334 static struct nouveau_abi16_chan
*
335 nouveau_abi16_chan(struct nouveau_abi16
*abi16
, int channel
)
337 struct nouveau_abi16_chan
*chan
;
339 list_for_each_entry(chan
, &abi16
->channels
, head
) {
340 if (chan
->chan
->object
->handle
== NOUVEAU_ABI16_CHAN(channel
))
348 nouveau_abi16_ioctl_channel_free(ABI16_IOCTL_ARGS
)
350 struct drm_nouveau_channel_free
*req
= data
;
351 struct nouveau_abi16
*abi16
= nouveau_abi16_get(file_priv
, dev
);
352 struct nouveau_abi16_chan
*chan
;
354 if (unlikely(!abi16
))
357 chan
= nouveau_abi16_chan(abi16
, req
->channel
);
359 return nouveau_abi16_put(abi16
, -ENOENT
);
360 nouveau_abi16_chan_fini(abi16
, chan
);
361 return nouveau_abi16_put(abi16
, 0);
365 nouveau_abi16_ioctl_grobj_alloc(ABI16_IOCTL_ARGS
)
367 struct drm_nouveau_grobj_alloc
*init
= data
;
369 struct nvif_ioctl_v0 ioctl
;
370 struct nvif_ioctl_new_v0
new;
372 .ioctl
.owner
= NVIF_IOCTL_V0_OWNER_ANY
,
373 .ioctl
.type
= NVIF_IOCTL_V0_NEW
,
375 .ioctl
.path
[2] = NOUVEAU_ABI16_CLIENT
,
376 .ioctl
.path
[1] = NOUVEAU_ABI16_DEVICE
,
377 .ioctl
.path
[0] = NOUVEAU_ABI16_CHAN(init
->channel
),
378 .new.route
= NVDRM_OBJECT_ABI16
,
379 .new.handle
= init
->handle
,
380 .new.oclass
= init
->class,
382 struct nouveau_abi16
*abi16
= nouveau_abi16_get(file_priv
, dev
);
383 struct nouveau_drm
*drm
= nouveau_drm(dev
);
384 struct nvif_client
*client
;
387 if (unlikely(!abi16
))
390 if (init
->handle
== ~0)
391 return nouveau_abi16_put(abi16
, -EINVAL
);
392 client
= nvif_client(nvif_object(&abi16
->device
));
394 /* compatibility with userspace that assumes 506e for all chipsets */
395 if (init
->class == 0x506e) {
396 init
->class = nouveau_abi16_swclass(drm
);
397 if (init
->class == 0x906e)
398 return nouveau_abi16_put(abi16
, 0);
401 ret
= nvif_client_ioctl(client
, &args
, sizeof(args
));
402 return nouveau_abi16_put(abi16
, ret
);
406 nouveau_abi16_ioctl_notifierobj_alloc(ABI16_IOCTL_ARGS
)
408 struct drm_nouveau_notifierobj_alloc
*info
= data
;
410 struct nvif_ioctl_v0 ioctl
;
411 struct nvif_ioctl_new_v0
new;
412 struct nv_dma_v0 ctxdma
;
414 .ioctl
.owner
= NVIF_IOCTL_V0_OWNER_ANY
,
415 .ioctl
.type
= NVIF_IOCTL_V0_NEW
,
417 .ioctl
.path
[2] = NOUVEAU_ABI16_CLIENT
,
418 .ioctl
.path
[1] = NOUVEAU_ABI16_DEVICE
,
419 .ioctl
.path
[0] = NOUVEAU_ABI16_CHAN(info
->channel
),
420 .new.route
= NVDRM_OBJECT_ABI16
,
421 .new.handle
= info
->handle
,
422 .new.oclass
= NV_DMA_IN_MEMORY
,
424 struct nouveau_drm
*drm
= nouveau_drm(dev
);
425 struct nouveau_abi16
*abi16
= nouveau_abi16_get(file_priv
, dev
);
426 struct nouveau_abi16_chan
*chan
;
427 struct nouveau_abi16_ntfy
*ntfy
;
428 struct nvif_device
*device
= &abi16
->device
;
429 struct nvif_client
*client
;
432 if (unlikely(!abi16
))
435 /* completely unnecessary for these chipsets... */
436 if (unlikely(device
->info
.family
>= NV_DEVICE_INFO_V0_FERMI
))
437 return nouveau_abi16_put(abi16
, -EINVAL
);
438 client
= nvif_client(nvif_object(&abi16
->device
));
440 chan
= nouveau_abi16_chan(abi16
, info
->channel
);
442 return nouveau_abi16_put(abi16
, -ENOENT
);
444 ntfy
= kzalloc(sizeof(*ntfy
), GFP_KERNEL
);
446 return nouveau_abi16_put(abi16
, -ENOMEM
);
448 list_add(&ntfy
->head
, &chan
->notifiers
);
449 ntfy
->handle
= info
->handle
;
451 ret
= nvkm_mm_head(&chan
->heap
, 0, 1, info
->size
, info
->size
, 1,
456 args
.ctxdma
.start
= ntfy
->node
->offset
;
457 args
.ctxdma
.limit
= ntfy
->node
->offset
+ ntfy
->node
->length
- 1;
458 if (device
->info
.family
>= NV_DEVICE_INFO_V0_TESLA
) {
459 args
.ctxdma
.target
= NV_DMA_V0_TARGET_VM
;
460 args
.ctxdma
.access
= NV_DMA_V0_ACCESS_VM
;
461 args
.ctxdma
.start
+= chan
->ntfy_vma
.offset
;
462 args
.ctxdma
.limit
+= chan
->ntfy_vma
.offset
;
464 if (drm
->agp
.stat
== ENABLED
) {
465 args
.ctxdma
.target
= NV_DMA_V0_TARGET_AGP
;
466 args
.ctxdma
.access
= NV_DMA_V0_ACCESS_RDWR
;
467 args
.ctxdma
.start
+= drm
->agp
.base
+ chan
->ntfy
->bo
.offset
;
468 args
.ctxdma
.limit
+= drm
->agp
.base
+ chan
->ntfy
->bo
.offset
;
469 client
->super
= true;
471 args
.ctxdma
.target
= NV_DMA_V0_TARGET_VM
;
472 args
.ctxdma
.access
= NV_DMA_V0_ACCESS_RDWR
;
473 args
.ctxdma
.start
+= chan
->ntfy
->bo
.offset
;
474 args
.ctxdma
.limit
+= chan
->ntfy
->bo
.offset
;
477 ret
= nvif_client_ioctl(client
, &args
, sizeof(args
));
478 client
->super
= false;
482 info
->offset
= ntfy
->node
->offset
;
486 nouveau_abi16_ntfy_fini(chan
, ntfy
);
487 return nouveau_abi16_put(abi16
, ret
);
491 nouveau_abi16_ioctl_gpuobj_free(ABI16_IOCTL_ARGS
)
493 struct drm_nouveau_gpuobj_free
*fini
= data
;
495 struct nvif_ioctl_v0 ioctl
;
496 struct nvif_ioctl_del del
;
498 .ioctl
.owner
= NVDRM_OBJECT_ABI16
,
499 .ioctl
.type
= NVIF_IOCTL_V0_DEL
,
501 .ioctl
.path
[3] = NOUVEAU_ABI16_CLIENT
,
502 .ioctl
.path
[2] = NOUVEAU_ABI16_DEVICE
,
503 .ioctl
.path
[1] = NOUVEAU_ABI16_CHAN(fini
->channel
),
504 .ioctl
.path
[0] = fini
->handle
,
506 struct nouveau_abi16
*abi16
= nouveau_abi16_get(file_priv
, dev
);
507 struct nouveau_abi16_chan
*chan
;
508 struct nouveau_abi16_ntfy
*ntfy
;
509 struct nvif_client
*client
;
512 if (unlikely(!abi16
))
515 chan
= nouveau_abi16_chan(abi16
, fini
->channel
);
517 return nouveau_abi16_put(abi16
, -ENOENT
);
518 client
= nvif_client(nvif_object(&abi16
->device
));
520 /* synchronize with the user channel and destroy the gpu object */
521 nouveau_channel_idle(chan
->chan
);
523 ret
= nvif_client_ioctl(client
, &args
, sizeof(args
));
525 return nouveau_abi16_put(abi16
, ret
);
527 /* cleanup extra state if this object was a notifier */
528 list_for_each_entry(ntfy
, &chan
->notifiers
, head
) {
529 if (ntfy
->handle
== fini
->handle
) {
530 nvkm_mm_free(&chan
->heap
, &ntfy
->node
);
531 list_del(&ntfy
->head
);
536 return nouveau_abi16_put(abi16
, 0);