Lines Matching defs:rdev
33 * @rdev: radeon_device pointer
40 void evergreen_dma_fence_ring_emit(struct radeon_device *rdev,
43 struct radeon_ring *ring = &rdev->ring[fence->ring];
44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr;
61 * @rdev: radeon_device pointer
66 void evergreen_dma_ring_ib_execute(struct radeon_device *rdev,
69 struct radeon_ring *ring = &rdev->ring[ib->ring];
71 if (rdev->wb.enabled) {
96 * @rdev: radeon_device pointer
106 struct radeon_fence *evergreen_copy_dma(struct radeon_device *rdev,
114 int ring_index = rdev->asic->copy.dma_ring_index;
115 struct radeon_ring *ring = &rdev->ring[ring_index];
124 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 11);
127 radeon_sync_free(rdev, &sync, NULL);
131 radeon_sync_resv(rdev, &sync, resv, false);
132 radeon_sync_rings(rdev, &sync, ring->idx);
148 r = radeon_fence_emit(rdev, &fence, ring->idx);
150 radeon_ring_unlock_undo(rdev, ring);
151 radeon_sync_free(rdev, &sync, NULL);
155 radeon_ring_unlock_commit(rdev, ring, false);
156 radeon_sync_free(rdev, &sync, fence);
164 * @rdev: radeon_device pointer
170 bool evergreen_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
172 u32 reset_mask = evergreen_gpu_check_soft_reset(rdev);
175 radeon_ring_lockup_update(rdev, ring);
178 return radeon_ring_test_lockup(rdev, ring);