Lines Matching refs:bo

45 	struct ttm_buffer_object *bo = &nvbo->bo;
56 ttm_bo_unref(&bo);
73 ret = ttm_bo_reserve(&nvbo->bo, false, false, false, 0);
95 ttm_bo_unreserve(&nvbo->bo);
110 ret = ttm_bo_reserve(&nvbo->bo, false, false, false, 0);
121 ttm_bo_unreserve(&nvbo->bo);
156 nvbo->gem = drm_gem_object_alloc(dev, nvbo->bo.mem.size);
162 nvbo->bo.persistent_swap_storage = nvbo->gem->filp;
175 if (nvbo->bo.mem.mem_type == TTM_PL_TT)
180 rep->offset = nvbo->bo.offset;
189 rep->size = nvbo->bo.mem.num_pages << PAGE_SHIFT;
190 rep->map_handle = nvbo->bo.addr_space_offset;
236 struct ttm_buffer_object *bo = &nvbo->bo;
251 bo->mem.mem_type == TTM_PL_VRAM)
255 bo->mem.mem_type == TTM_PL_TT)
293 ttm_bo_unreserve(&nvbo->bo);
344 ret = ttm_bo_reserve(&nvbo->bo, true, false, true, sequence);
348 ret = ttm_bo_wait_unreserved(&nvbo->bo, true);
388 spin_lock(&nvbo->bo.bdev->fence_lock);
389 if (nvbo->bo.sync_obj)
390 fence = nouveau_fence_ref(nvbo->bo.sync_obj);
391 spin_unlock(&nvbo->bo.bdev->fence_lock);
443 if (nvbo->bo.offset == b->presumed.offset &&
444 ((nvbo->bo.mem.mem_type == TTM_PL_VRAM &&
446 (nvbo->bo.mem.mem_type == TTM_PL_TT &&
450 if (nvbo->bo.mem.mem_type == TTM_PL_TT)
454 b->presumed.offset = nvbo->bo.offset;
543 struct drm_nouveau_gem_pushbuf_bo *bo)
560 NV_ERROR(dev, "reloc bo index invalid\n");
565 b = &bo[r->bo_index];
570 NV_ERROR(dev, "reloc container bo index invalid\n");
574 nvbo = (void *)(unsigned long)bo[r->reloc_bo_index].user_priv;
577 nvbo->bo.mem.num_pages << PAGE_SHIFT)) {
578 NV_ERROR(dev, "reloc outside of bo\n");
584 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages,
608 spin_lock(&nvbo->bo.bdev->fence_lock);
609 ret = ttm_bo_wait(&nvbo->bo, false, false, false);
610 spin_unlock(&nvbo->bo.bdev->fence_lock);
630 struct drm_nouveau_gem_pushbuf_bo *bo;
653 NV_ERROR(dev, "pushbuf bo count exceeds limit: %d max %d\n",
672 bo = u_memcpya(req->buffers, req->nr_buffers, sizeof(*bo));
673 if (IS_ERR(bo)) {
676 return PTR_ERR(bo);
689 ret = nouveau_gem_pushbuf_validate(chan, file_priv, bo, req->buffers,
699 ret = nouveau_gem_pushbuf_reloc_apply(dev, req, bo);
715 bo[push[i].bo_index].user_priv;
730 bo[push[i].bo_index].user_priv;
731 struct drm_mm_node *mem = nvbo->bo.mem.mm_node;
746 bo[push[i].bo_index].user_priv;
747 struct drm_mm_node *mem = nvbo->bo.mem.mm_node;
754 ret = ttm_bo_kmap(&nvbo->bo, 0,
755 nvbo->bo.mem.
789 kfree(bo);
838 spin_lock(&nvbo->bo.bdev->fence_lock);
839 ret = ttm_bo_wait(&nvbo->bo, true, true, no_wait);
840 spin_unlock(&nvbo->bo.bdev->fence_lock);