Lines Matching defs:rdev
30 u32 si_gpu_check_soft_reset(struct radeon_device *rdev);
35 * @rdev: radeon_device pointer
41 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
43 u32 reset_mask = si_gpu_check_soft_reset(rdev);
52 radeon_ring_lockup_update(rdev, ring);
55 return radeon_ring_test_lockup(rdev, ring);
61 * @rdev: radeon_device pointer
69 void si_dma_vm_copy_pages(struct radeon_device *rdev,
95 * @rdev: radeon_device pointer
105 void si_dma_vm_write_pages(struct radeon_device *rdev,
125 value = radeon_vm_map_gart(rdev, addr);
142 * @rdev: radeon_device pointer
152 void si_dma_vm_set_pages(struct radeon_device *rdev,
187 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring,
221 * @rdev: radeon_device pointer
231 struct radeon_fence *si_copy_dma(struct radeon_device *rdev,
238 int ring_index = rdev->asic->copy.dma_ring_index;
239 struct radeon_ring *ring = &rdev->ring[ring_index];
248 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 11);
251 radeon_sync_free(rdev, &sync, NULL);
255 radeon_sync_resv(rdev, &sync, resv, false);
256 radeon_sync_rings(rdev, &sync, ring->idx);
272 r = radeon_fence_emit(rdev, &fence, ring->idx);
274 radeon_ring_unlock_undo(rdev, ring);
275 radeon_sync_free(rdev, &sync, NULL);
279 radeon_ring_unlock_commit(rdev, ring, false);
280 radeon_sync_free(rdev, &sync, fence);