Lines Matching defs:rdev
34 * @rdev: radeon_device pointer
40 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
42 u32 reset_mask = si_gpu_check_soft_reset(rdev);
51 radeon_ring_lockup_update(rdev, ring);
54 return radeon_ring_test_lockup(rdev, ring);
60 * @rdev: radeon_device pointer
68 void si_dma_vm_copy_pages(struct radeon_device *rdev,
94 * @rdev: radeon_device pointer
104 void si_dma_vm_write_pages(struct radeon_device *rdev,
124 value = radeon_vm_map_gart(rdev, addr);
141 * @rdev: radeon_device pointer
151 void si_dma_vm_set_pages(struct radeon_device *rdev,
186 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring,
220 * @rdev: radeon_device pointer
230 struct radeon_fence *si_copy_dma(struct radeon_device *rdev,
237 int ring_index = rdev->asic->copy.dma_ring_index;
238 struct radeon_ring *ring = &rdev->ring[ring_index];
247 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 11);
250 radeon_sync_free(rdev, &sync, NULL);
254 radeon_sync_resv(rdev, &sync, resv, false);
255 radeon_sync_rings(rdev, &sync, ring->idx);
271 r = radeon_fence_emit(rdev, &fence, ring->idx);
273 radeon_ring_unlock_undo(rdev, ring);
274 radeon_sync_free(rdev, &sync, NULL);
278 radeon_ring_unlock_commit(rdev, ring, false);
279 radeon_sync_free(rdev, &sync, fence);