Lines Matching refs:robj
86 struct radeon_bo *robj = gem_to_radeon_bo(gobj);
88 if (robj) {
89 radeon_mn_unregister(robj);
90 radeon_bo_unref(&robj);
99 struct radeon_bo *robj;
121 flags, NULL, NULL, &robj);
133 *obj = &robj->tbo.base;
135 robj->pid = task_pid_nr(current);
138 list_add_tail(&robj->list, &rdev->gem.objects);
147 struct radeon_bo *robj;
152 robj = gem_to_radeon_bo(gobj);
165 r = dma_resv_wait_timeout(robj->tbo.base.resv,
176 if (domain == RADEON_GEM_DOMAIN_VRAM && robj->prime_shared_count) {
472 struct radeon_bo *robj;
478 robj = gem_to_radeon_bo(gobj);
479 if (radeon_ttm_tt_has_userptr(robj->rdev, robj->tbo.ttm)) {
483 *offset_p = radeon_bo_mmap_offset(robj);
501 struct radeon_bo *robj;
509 robj = gem_to_radeon_bo(gobj);
511 r = dma_resv_test_signaled(robj->tbo.base.resv, DMA_RESV_USAGE_READ);
517 cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
529 struct radeon_bo *robj;
538 robj = gem_to_radeon_bo(gobj);
540 ret = dma_resv_wait_timeout(robj->tbo.base.resv, DMA_RESV_USAGE_READ,
548 cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
551 robj->rdev->asic->mmio_hdp_flush(rdev);
562 struct radeon_bo *robj;
569 robj = gem_to_radeon_bo(gobj);
570 r = radeon_bo_set_tiling_flags(robj, args->tiling_flags, args->pitch);
771 struct radeon_bo *robj;
778 robj = gem_to_radeon_bo(gobj);
781 if (radeon_ttm_tt_has_userptr(robj->rdev, robj->tbo.ttm))
784 r = radeon_bo_reserve(robj, false);
790 args->value = robj->initial_domain;
793 robj->initial_domain = args->value & (RADEON_GEM_DOMAIN_VRAM |
801 radeon_bo_unreserve(robj);