Lines Matching defs:tbo
63 static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo)
67 bo = container_of(tbo, struct radeon_bo, tbo);
156 r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type,
190 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
213 struct ttm_buffer_object *tbo;
219 tbo = &((*bo)->tbo);
221 ttm_bo_unref(&tbo);
223 if (tbo == NULL)
244 r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false, false);
268 r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false, false);
353 r = ttm_bo_validate(&bo->tbo, &bo->placement,
372 return ttm_fbdev_mmap(vma, &bo->tbo);
383 BUG_ON(!atomic_read(&bo->tbo.reserved));
415 ttm_bo_unmap_virtual(&old_object->tbo);
425 bo->tbo.mem.start << PAGE_SHIFT,
426 bo->tbo.num_pages << PAGE_SHIFT);
463 BUG_ON(!atomic_read(&bo->tbo.reserved));
473 BUG_ON(!atomic_read(&bo->tbo.reserved));
483 if (bo->tbo.mem.mem_type != TTM_PL_VRAM) {
504 rbo = container_of(bo, struct radeon_bo, tbo);
518 rbo = container_of(bo, struct radeon_bo, tbo);
544 r = ttm_bo_reserve(&bo->tbo, true, no_wait, false, 0);
547 spin_lock(&bo->tbo.bdev->fence_lock);
549 *mem_type = bo->tbo.mem.mem_type;
550 if (bo->tbo.sync_obj)
551 r = ttm_bo_wait(&bo->tbo, true, true, no_wait);
552 spin_unlock(&bo->tbo.bdev->fence_lock);
553 ttm_bo_unreserve(&bo->tbo);
572 r = ttm_bo_reserve(&bo->tbo, true, no_wait, false, 0);