2 * Copyright 2009 Jerome Glisse.
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the
7 * "Software"), to deal in the Software without restriction, including
8 * without limitation the rights to use, copy, modify, merge, publish,
9 * distribute, sub license, and/or sell copies of the Software, and to
10 * permit persons to whom the Software is furnished to do so, subject to
11 * the following conditions:
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
16 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
17 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
18 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
19 * USE OR OTHER DEALINGS IN THE SOFTWARE.
21 * The above copyright notice and this permission notice (including the
22 * next paragraph) shall be included in all copies or substantial portions
28 * Jerome Glisse <glisse@freedesktop.org>
29 * Thomas Hellstrom <thomas-at-tungstengraphics-dot-com>
32 #include <ttm/ttm_bo_api.h>
33 #include <ttm/ttm_bo_driver.h>
34 #include <ttm/ttm_placement.h>
35 #include <ttm/ttm_module.h>
37 #include <drm/radeon_drm.h>
38 #include "radeon_reg.h"
41 #define DRM_FILE_PAGE_OFFSET (0x100000000ULL >> PAGE_SHIFT)
43 static struct radeon_device
*radeon_get_rdev(struct ttm_bo_device
*bdev
)
45 struct radeon_mman
*mman
;
46 struct radeon_device
*rdev
;
48 mman
= container_of(bdev
, struct radeon_mman
, bdev
);
49 rdev
= container_of(mman
, struct radeon_device
, mman
);
57 static int radeon_ttm_mem_global_init(struct ttm_global_reference
*ref
)
59 return ttm_mem_global_init(ref
->object
);
62 static void radeon_ttm_mem_global_release(struct ttm_global_reference
*ref
)
64 ttm_mem_global_release(ref
->object
);
67 static int radeon_ttm_global_init(struct radeon_device
*rdev
)
69 struct ttm_global_reference
*global_ref
;
72 rdev
->mman
.mem_global_referenced
= false;
73 global_ref
= &rdev
->mman
.mem_global_ref
;
74 global_ref
->global_type
= TTM_GLOBAL_TTM_MEM
;
75 global_ref
->size
= sizeof(struct ttm_mem_global
);
76 global_ref
->init
= &radeon_ttm_mem_global_init
;
77 global_ref
->release
= &radeon_ttm_mem_global_release
;
78 r
= ttm_global_item_ref(global_ref
);
80 DRM_ERROR("Failed referencing a global TTM memory object.\n");
83 rdev
->mman
.mem_global_referenced
= true;
87 static void radeon_ttm_global_fini(struct radeon_device
*rdev
)
89 if (rdev
->mman
.mem_global_referenced
) {
90 ttm_global_item_unref(&rdev
->mman
.mem_global_ref
);
91 rdev
->mman
.mem_global_referenced
= false;
95 struct ttm_backend
*radeon_ttm_backend_create(struct radeon_device
*rdev
);
97 static struct ttm_backend
*
98 radeon_create_ttm_backend_entry(struct ttm_bo_device
*bdev
)
100 struct radeon_device
*rdev
;
102 rdev
= radeon_get_rdev(bdev
);
104 if (rdev
->flags
& RADEON_IS_AGP
) {
105 return ttm_agp_backend_init(bdev
, rdev
->ddev
->agp
->bridge
);
109 return radeon_ttm_backend_create(rdev
);
113 static int radeon_invalidate_caches(struct ttm_bo_device
*bdev
, uint32_t flags
)
118 static int radeon_init_mem_type(struct ttm_bo_device
*bdev
, uint32_t type
,
119 struct ttm_mem_type_manager
*man
)
121 struct radeon_device
*rdev
;
123 rdev
= radeon_get_rdev(bdev
);
128 man
->flags
= TTM_MEMTYPE_FLAG_MAPPABLE
;
129 man
->available_caching
= TTM_PL_MASK_CACHING
;
130 man
->default_caching
= TTM_PL_FLAG_CACHED
;
134 man
->available_caching
= TTM_PL_MASK_CACHING
;
135 man
->default_caching
= TTM_PL_FLAG_CACHED
;
136 man
->flags
= TTM_MEMTYPE_FLAG_MAPPABLE
| TTM_MEMTYPE_FLAG_CMA
;
138 if (rdev
->flags
& RADEON_IS_AGP
) {
139 if (!(drm_core_has_AGP(rdev
->ddev
) && rdev
->ddev
->agp
)) {
140 DRM_ERROR("AGP is not enabled for memory type %u\n",
144 man
->io_offset
= rdev
->mc
.agp_base
;
145 man
->io_size
= rdev
->mc
.gtt_size
;
147 if (!rdev
->ddev
->agp
->cant_use_aperture
)
148 man
->flags
= TTM_MEMTYPE_FLAG_NEEDS_IOREMAP
|
149 TTM_MEMTYPE_FLAG_MAPPABLE
;
150 man
->available_caching
= TTM_PL_FLAG_UNCACHED
|
152 man
->default_caching
= TTM_PL_FLAG_WC
;
162 /* "On-card" video ram */
164 man
->flags
= TTM_MEMTYPE_FLAG_FIXED
|
165 TTM_MEMTYPE_FLAG_NEEDS_IOREMAP
|
166 TTM_MEMTYPE_FLAG_MAPPABLE
;
167 man
->available_caching
= TTM_PL_FLAG_UNCACHED
| TTM_PL_FLAG_WC
;
168 man
->default_caching
= TTM_PL_FLAG_WC
;
170 man
->io_offset
= rdev
->mc
.aper_base
;
171 man
->io_size
= rdev
->mc
.aper_size
;
174 DRM_ERROR("Unsupported memory type %u\n", (unsigned)type
);
180 static uint32_t radeon_evict_flags(struct ttm_buffer_object
*bo
)
182 uint32_t cur_placement
= bo
->mem
.placement
& ~TTM_PL_MASK_MEMTYPE
;
184 switch (bo
->mem
.mem_type
) {
186 return (cur_placement
& ~TTM_PL_MASK_CACHING
) |
192 static int radeon_verify_access(struct ttm_buffer_object
*bo
, struct file
*filp
)
197 static void radeon_move_null(struct ttm_buffer_object
*bo
,
198 struct ttm_mem_reg
*new_mem
)
200 struct ttm_mem_reg
*old_mem
= &bo
->mem
;
202 BUG_ON(old_mem
->mm_node
!= NULL
);
204 new_mem
->mm_node
= NULL
;
207 static int radeon_move_blit(struct ttm_buffer_object
*bo
,
208 bool evict
, int no_wait
,
209 struct ttm_mem_reg
*new_mem
,
210 struct ttm_mem_reg
*old_mem
)
212 struct radeon_device
*rdev
;
213 uint64_t old_start
, new_start
;
214 struct radeon_fence
*fence
;
217 rdev
= radeon_get_rdev(bo
->bdev
);
218 r
= radeon_fence_create(rdev
, &fence
);
222 old_start
= old_mem
->mm_node
->start
<< PAGE_SHIFT
;
223 new_start
= new_mem
->mm_node
->start
<< PAGE_SHIFT
;
225 switch (old_mem
->mem_type
) {
227 old_start
+= rdev
->mc
.vram_location
;
230 old_start
+= rdev
->mc
.gtt_location
;
233 DRM_ERROR("Unknown placement %d\n", old_mem
->mem_type
);
236 switch (new_mem
->mem_type
) {
238 new_start
+= rdev
->mc
.vram_location
;
241 new_start
+= rdev
->mc
.gtt_location
;
244 DRM_ERROR("Unknown placement %d\n", old_mem
->mem_type
);
247 if (!rdev
->cp
.ready
) {
248 DRM_ERROR("Trying to move memory with CP turned off.\n");
251 r
= radeon_copy(rdev
, old_start
, new_start
, new_mem
->num_pages
, fence
);
252 /* FIXME: handle copy error */
253 r
= ttm_bo_move_accel_cleanup(bo
, (void *)fence
, NULL
,
254 evict
, no_wait
, new_mem
);
255 radeon_fence_unref(&fence
);
259 static int radeon_move_vram_ram(struct ttm_buffer_object
*bo
,
260 bool evict
, bool interruptible
, bool no_wait
,
261 struct ttm_mem_reg
*new_mem
)
263 struct radeon_device
*rdev
;
264 struct ttm_mem_reg
*old_mem
= &bo
->mem
;
265 struct ttm_mem_reg tmp_mem
;
266 uint32_t proposed_placement
;
269 rdev
= radeon_get_rdev(bo
->bdev
);
271 tmp_mem
.mm_node
= NULL
;
272 proposed_placement
= TTM_PL_FLAG_TT
| TTM_PL_MASK_CACHING
;
273 r
= ttm_bo_mem_space(bo
, proposed_placement
, &tmp_mem
,
274 interruptible
, no_wait
);
278 r
= ttm_tt_bind(bo
->ttm
, &tmp_mem
);
282 r
= radeon_move_blit(bo
, true, no_wait
, &tmp_mem
, old_mem
);
286 r
= ttm_bo_move_ttm(bo
, true, no_wait
, new_mem
);
288 if (tmp_mem
.mm_node
) {
289 spin_lock(&rdev
->mman
.bdev
.lru_lock
);
290 drm_mm_put_block(tmp_mem
.mm_node
);
291 spin_unlock(&rdev
->mman
.bdev
.lru_lock
);
297 static int radeon_move_ram_vram(struct ttm_buffer_object
*bo
,
298 bool evict
, bool interruptible
, bool no_wait
,
299 struct ttm_mem_reg
*new_mem
)
301 struct radeon_device
*rdev
;
302 struct ttm_mem_reg
*old_mem
= &bo
->mem
;
303 struct ttm_mem_reg tmp_mem
;
304 uint32_t proposed_flags
;
307 rdev
= radeon_get_rdev(bo
->bdev
);
309 tmp_mem
.mm_node
= NULL
;
310 proposed_flags
= TTM_PL_FLAG_TT
| TTM_PL_MASK_CACHING
;
311 r
= ttm_bo_mem_space(bo
, proposed_flags
, &tmp_mem
,
312 interruptible
, no_wait
);
316 r
= ttm_bo_move_ttm(bo
, true, no_wait
, &tmp_mem
);
320 r
= radeon_move_blit(bo
, true, no_wait
, new_mem
, old_mem
);
325 if (tmp_mem
.mm_node
) {
326 spin_lock(&rdev
->mman
.bdev
.lru_lock
);
327 drm_mm_put_block(tmp_mem
.mm_node
);
328 spin_unlock(&rdev
->mman
.bdev
.lru_lock
);
334 static int radeon_bo_move(struct ttm_buffer_object
*bo
,
335 bool evict
, bool interruptible
, bool no_wait
,
336 struct ttm_mem_reg
*new_mem
)
338 struct radeon_device
*rdev
;
339 struct ttm_mem_reg
*old_mem
= &bo
->mem
;
342 rdev
= radeon_get_rdev(bo
->bdev
);
343 if (old_mem
->mem_type
== TTM_PL_SYSTEM
&& bo
->ttm
== NULL
) {
344 radeon_move_null(bo
, new_mem
);
347 if ((old_mem
->mem_type
== TTM_PL_TT
&&
348 new_mem
->mem_type
== TTM_PL_SYSTEM
) ||
349 (old_mem
->mem_type
== TTM_PL_SYSTEM
&&
350 new_mem
->mem_type
== TTM_PL_TT
)) {
351 /* bind is enought */
352 radeon_move_null(bo
, new_mem
);
355 if (!rdev
->cp
.ready
) {
357 DRM_ERROR("CP is not ready use memcpy.\n");
361 if (old_mem
->mem_type
== TTM_PL_VRAM
&&
362 new_mem
->mem_type
== TTM_PL_SYSTEM
) {
363 r
= radeon_move_vram_ram(bo
, evict
, interruptible
,
365 } else if (old_mem
->mem_type
== TTM_PL_SYSTEM
&&
366 new_mem
->mem_type
== TTM_PL_VRAM
) {
367 r
= radeon_move_ram_vram(bo
, evict
, interruptible
,
370 r
= radeon_move_blit(bo
, evict
, no_wait
, new_mem
, old_mem
);
375 r
= ttm_bo_move_memcpy(bo
, evict
, no_wait
, new_mem
);
381 const uint32_t radeon_mem_prios
[] = {
387 const uint32_t radeon_busy_prios
[] = {
393 static int radeon_sync_obj_wait(void *sync_obj
, void *sync_arg
,
394 bool lazy
, bool interruptible
)
396 return radeon_fence_wait((struct radeon_fence
*)sync_obj
, interruptible
);
399 static int radeon_sync_obj_flush(void *sync_obj
, void *sync_arg
)
404 static void radeon_sync_obj_unref(void **sync_obj
)
406 radeon_fence_unref((struct radeon_fence
**)sync_obj
);
409 static void *radeon_sync_obj_ref(void *sync_obj
)
411 return radeon_fence_ref((struct radeon_fence
*)sync_obj
);
414 static bool radeon_sync_obj_signaled(void *sync_obj
, void *sync_arg
)
416 return radeon_fence_signaled((struct radeon_fence
*)sync_obj
);
419 static struct ttm_bo_driver radeon_bo_driver
= {
420 .mem_type_prio
= radeon_mem_prios
,
421 .mem_busy_prio
= radeon_busy_prios
,
422 .num_mem_type_prio
= ARRAY_SIZE(radeon_mem_prios
),
423 .num_mem_busy_prio
= ARRAY_SIZE(radeon_busy_prios
),
424 .create_ttm_backend_entry
= &radeon_create_ttm_backend_entry
,
425 .invalidate_caches
= &radeon_invalidate_caches
,
426 .init_mem_type
= &radeon_init_mem_type
,
427 .evict_flags
= &radeon_evict_flags
,
428 .move
= &radeon_bo_move
,
429 .verify_access
= &radeon_verify_access
,
430 .sync_obj_signaled
= &radeon_sync_obj_signaled
,
431 .sync_obj_wait
= &radeon_sync_obj_wait
,
432 .sync_obj_flush
= &radeon_sync_obj_flush
,
433 .sync_obj_unref
= &radeon_sync_obj_unref
,
434 .sync_obj_ref
= &radeon_sync_obj_ref
,
435 .move_notify
= &radeon_bo_move_notify
,
436 .fault_reserve_notify
= &radeon_bo_fault_reserve_notify
,
439 int radeon_ttm_init(struct radeon_device
*rdev
)
443 r
= radeon_ttm_global_init(rdev
);
447 /* No others user of address space so set it to 0 */
448 r
= ttm_bo_device_init(&rdev
->mman
.bdev
,
449 rdev
->mman
.mem_global_ref
.object
,
450 &radeon_bo_driver
, DRM_FILE_PAGE_OFFSET
,
453 DRM_ERROR("failed initializing buffer object driver(%d).\n", r
);
456 r
= ttm_bo_init_mm(&rdev
->mman
.bdev
, TTM_PL_VRAM
, 0,
457 ((rdev
->mc
.real_vram_size
) >> PAGE_SHIFT
));
459 DRM_ERROR("Failed initializing VRAM heap.\n");
462 r
= radeon_object_create(rdev
, NULL
, 256 * 1024, true,
463 RADEON_GEM_DOMAIN_VRAM
, false,
464 &rdev
->stollen_vga_memory
);
468 r
= radeon_object_pin(rdev
->stollen_vga_memory
, RADEON_GEM_DOMAIN_VRAM
, NULL
);
470 radeon_object_unref(&rdev
->stollen_vga_memory
);
473 DRM_INFO("radeon: %uM of VRAM memory ready\n",
474 rdev
->mc
.real_vram_size
/ (1024 * 1024));
475 r
= ttm_bo_init_mm(&rdev
->mman
.bdev
, TTM_PL_TT
, 0,
476 ((rdev
->mc
.gtt_size
) >> PAGE_SHIFT
));
478 DRM_ERROR("Failed initializing GTT heap.\n");
481 DRM_INFO("radeon: %uM of GTT memory ready.\n",
482 rdev
->mc
.gtt_size
/ (1024 * 1024));
483 if (unlikely(rdev
->mman
.bdev
.dev_mapping
== NULL
)) {
484 rdev
->mman
.bdev
.dev_mapping
= rdev
->ddev
->dev_mapping
;
489 void radeon_ttm_fini(struct radeon_device
*rdev
)
491 if (rdev
->stollen_vga_memory
) {
492 radeon_object_unpin(rdev
->stollen_vga_memory
);
493 radeon_object_unref(&rdev
->stollen_vga_memory
);
495 ttm_bo_clean_mm(&rdev
->mman
.bdev
, TTM_PL_VRAM
);
496 ttm_bo_clean_mm(&rdev
->mman
.bdev
, TTM_PL_TT
);
497 ttm_bo_device_release(&rdev
->mman
.bdev
);
498 radeon_gart_fini(rdev
);
499 radeon_ttm_global_fini(rdev
);
500 DRM_INFO("radeon: ttm finalized\n");
503 static struct vm_operations_struct radeon_ttm_vm_ops
;
504 static struct vm_operations_struct
*ttm_vm_ops
= NULL
;
506 static int radeon_ttm_fault(struct vm_area_struct
*vma
, struct vm_fault
*vmf
)
508 struct ttm_buffer_object
*bo
;
511 bo
= (struct ttm_buffer_object
*)vma
->vm_private_data
;
513 return VM_FAULT_NOPAGE
;
515 r
= ttm_vm_ops
->fault(vma
, vmf
);
519 int radeon_mmap(struct file
*filp
, struct vm_area_struct
*vma
)
521 struct drm_file
*file_priv
;
522 struct radeon_device
*rdev
;
525 if (unlikely(vma
->vm_pgoff
< DRM_FILE_PAGE_OFFSET
)) {
526 return drm_mmap(filp
, vma
);
529 file_priv
= (struct drm_file
*)filp
->private_data
;
530 rdev
= file_priv
->minor
->dev
->dev_private
;
534 r
= ttm_bo_mmap(filp
, vma
, &rdev
->mman
.bdev
);
535 if (unlikely(r
!= 0)) {
538 if (unlikely(ttm_vm_ops
== NULL
)) {
539 ttm_vm_ops
= vma
->vm_ops
;
540 radeon_ttm_vm_ops
= *ttm_vm_ops
;
541 radeon_ttm_vm_ops
.fault
= &radeon_ttm_fault
;
543 vma
->vm_ops
= &radeon_ttm_vm_ops
;
549 * TTM backend functions.
551 struct radeon_ttm_backend
{
552 struct ttm_backend backend
;
553 struct radeon_device
*rdev
;
554 unsigned long num_pages
;
556 struct page
*dummy_read_page
;
562 static int radeon_ttm_backend_populate(struct ttm_backend
*backend
,
563 unsigned long num_pages
,
565 struct page
*dummy_read_page
)
567 struct radeon_ttm_backend
*gtt
;
569 gtt
= container_of(backend
, struct radeon_ttm_backend
, backend
);
571 gtt
->num_pages
= num_pages
;
572 gtt
->dummy_read_page
= dummy_read_page
;
573 gtt
->populated
= true;
577 static void radeon_ttm_backend_clear(struct ttm_backend
*backend
)
579 struct radeon_ttm_backend
*gtt
;
581 gtt
= container_of(backend
, struct radeon_ttm_backend
, backend
);
584 gtt
->dummy_read_page
= NULL
;
585 gtt
->populated
= false;
590 static int radeon_ttm_backend_bind(struct ttm_backend
*backend
,
591 struct ttm_mem_reg
*bo_mem
)
593 struct radeon_ttm_backend
*gtt
;
596 gtt
= container_of(backend
, struct radeon_ttm_backend
, backend
);
597 gtt
->offset
= bo_mem
->mm_node
->start
<< PAGE_SHIFT
;
598 if (!gtt
->num_pages
) {
599 WARN(1, "nothing to bind %lu pages for mreg %p back %p!\n", gtt
->num_pages
, bo_mem
, backend
);
601 r
= radeon_gart_bind(gtt
->rdev
, gtt
->offset
,
602 gtt
->num_pages
, gtt
->pages
);
604 DRM_ERROR("failed to bind %lu pages at 0x%08X\n",
605 gtt
->num_pages
, gtt
->offset
);
612 static int radeon_ttm_backend_unbind(struct ttm_backend
*backend
)
614 struct radeon_ttm_backend
*gtt
;
616 gtt
= container_of(backend
, struct radeon_ttm_backend
, backend
);
617 radeon_gart_unbind(gtt
->rdev
, gtt
->offset
, gtt
->num_pages
);
622 static void radeon_ttm_backend_destroy(struct ttm_backend
*backend
)
624 struct radeon_ttm_backend
*gtt
;
626 gtt
= container_of(backend
, struct radeon_ttm_backend
, backend
);
628 radeon_ttm_backend_unbind(backend
);
633 static struct ttm_backend_func radeon_backend_func
= {
634 .populate
= &radeon_ttm_backend_populate
,
635 .clear
= &radeon_ttm_backend_clear
,
636 .bind
= &radeon_ttm_backend_bind
,
637 .unbind
= &radeon_ttm_backend_unbind
,
638 .destroy
= &radeon_ttm_backend_destroy
,
641 struct ttm_backend
*radeon_ttm_backend_create(struct radeon_device
*rdev
)
643 struct radeon_ttm_backend
*gtt
;
645 gtt
= kzalloc(sizeof(struct radeon_ttm_backend
), GFP_KERNEL
);
649 gtt
->backend
.bdev
= &rdev
->mman
.bdev
;
650 gtt
->backend
.flags
= 0;
651 gtt
->backend
.func
= &radeon_backend_func
;
655 gtt
->dummy_read_page
= NULL
;
656 gtt
->populated
= false;
658 return >t
->backend
;