Lines Matching refs:tbo

57 	u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT;
75 static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo)
79 bo = container_of(tbo, struct radeon_bo, tbo);
81 radeon_update_memory_usage(bo, bo->tbo.mem.mem_type, -1);
88 if (bo->tbo.base.import_attach)
89 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg);
90 drm_gem_object_release(&bo->tbo.base);
212 drm_gem_private_object_init(rdev->ddev, &bo->tbo.base, size);
262 r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type,
287 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
313 ttm_bo_get(&bo->tbo);
319 struct ttm_buffer_object *tbo;
323 tbo = &((*bo)->tbo);
324 ttm_bo_put(tbo);
334 if (radeon_ttm_tt_has_userptr(bo->rdev, bo->tbo.ttm))
374 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
410 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
412 if (bo->tbo.mem.mem_type == TTM_PL_VRAM)
445 &bo->tbo.base, bo, (unsigned long)bo->tbo.base.size,
446 *((unsigned long *)&bo->tbo.base.refcount));
451 drm_gem_object_put(&bo->tbo.base);
556 radeon_mem_type_to_domain(bo->tbo.mem.mem_type);
579 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
613 dma_resv_assert_held(bo->tbo.base.resv);
645 ttm_bo_unmap_virtual(&old_object->tbo);
655 bo->tbo.mem.start << PAGE_SHIFT,
656 bo->tbo.num_pages << PAGE_SHIFT);
739 dma_resv_assert_held(bo->tbo.base.resv);
751 dma_resv_assert_held(bo->tbo.base.resv);
761 if (bo->tbo.mem.mem_type != TTM_PL_VRAM) {
785 rbo = container_of(bo, struct radeon_bo, tbo);
807 rbo = container_of(bo, struct radeon_bo, tbo);
851 r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
855 *mem_type = bo->tbo.mem.mem_type;
857 r = ttm_bo_wait(&bo->tbo, true, no_wait);
858 ttm_bo_unreserve(&bo->tbo);
873 struct dma_resv *resv = bo->tbo.base.resv;