/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_fence.c | 101 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_write() 117 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_read() 150 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit() 153 &ring->fence_drv.lock, in amdgpu_fence_emit() 156 amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr, in amdgpu_fence_emit() 159 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit() 204 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit_polling() 206 seq - ring->fence_drv.num_fences_mask, in amdgpu_fence_emit_polling() 211 amdgpu_ring_emit_fence(ring, ring->fence_drv in amdgpu_fence_emit_polling() [all...] |
H A D | mes_v10_1.c | 104 r = amdgpu_fence_wait_polling(ring, ring->fence_drv.sync_seq, in mes_v10_1_submit_pkt_and_poll_completion() 161 mes->ring.fence_drv.gpu_addr; in mes_v10_1_add_hw_queue() 163 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_add_hw_queue() 184 mes->ring.fence_drv.gpu_addr; in mes_v10_1_remove_hw_queue() 186 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_remove_hw_queue() 215 mes->ring.fence_drv.gpu_addr; in mes_v10_1_query_sched_status() 217 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_query_sched_status() 258 mes->ring.fence_drv.gpu_addr; in mes_v10_1_set_hw_resources() 260 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_set_hw_resources()
|
H A D | amdgpu_debugfs.c | 1360 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_fences_swap() 1363 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap() 1364 sync_seq = ring->fence_drv.sync_seq; in amdgpu_ib_preempt_fences_swap() 1421 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_mark_partial_job() 1474 length = ring->fence_drv.num_fences_mask + 1; in amdgpu_debugfs_ib_preempt() 1498 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt() 1499 ring->fence_drv.sync_seq) { in amdgpu_debugfs_ib_preempt()
|
H A D | amdgpu_job.c | 49 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 50 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
|
H A D | uvd_v6_0.c | 1062 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_ring_emit_pipeline_sync() 1063 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_ring_emit_pipeline_sync() 1091 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_enc_ring_emit_pipeline_sync() 1092 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_enc_ring_emit_pipeline_sync()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_fence.c | 103 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_write() 119 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_read() 163 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit() 172 &ring->fence_drv.lock, in amdgpu_fence_emit() 178 &ring->fence_drv.lock, in amdgpu_fence_emit() 183 amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr, in amdgpu_fence_emit() 186 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit() 234 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit_polling() 236 seq - ring->fence_drv in amdgpu_fence_emit_polling() [all...] |
H A D | amdgpu_ring_mux.c | 99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks() 109 le32_to_cpu(*(e->ring->fence_drv.cpu_addr + 2))) { in amdgpu_mux_resubmit_chunks() 475 last_seq = atomic_read(&ring->fence_drv.last_seq); in scan_and_remove_signaled_chunk() 538 chunk->sync_seq = READ_ONCE(ring->fence_drv.sync_seq); in amdgpu_ring_mux_end_ib() 571 mux->seqno_to_resubmit = ring->fence_drv.sync_seq; in amdgpu_mcbp_handle_trailing_fence_irq()
|
H A D | amdgpu_job.c | 63 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 64 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
|
H A D | amdgpu_debugfs.c | 1805 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_fences_swap() 1808 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap() 1809 sync_seq = ring->fence_drv.sync_seq; in amdgpu_ib_preempt_fences_swap() 1866 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_mark_partial_job() 1919 length = ring->fence_drv.num_fences_mask + 1; in amdgpu_debugfs_ib_preempt() 1941 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt() 1942 ring->fence_drv.sync_seq) { in amdgpu_debugfs_ib_preempt()
|
H A D | uvd_v6_0.c | 1088 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_ring_emit_pipeline_sync() 1089 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_ring_emit_pipeline_sync() 1117 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_enc_ring_emit_pipeline_sync() 1118 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_enc_ring_emit_pipeline_sync()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | radeon_fence.c | 69 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_write() 90 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_read() 120 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check() 146 (*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_emit() 177 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled() 224 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity() 226 last_emitted = rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_activity() 251 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity() 269 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local 273 fence_drv in radeon_fence_check_lockup() [all...] |
H A D | uvd_v2_2.c | 43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
|
H A D | evergreen_dma.c | 44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
|
H A D | uvd_v1_0.c | 85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
H A D | radeon_fence.c | 70 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_write() 91 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_read() 121 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check() 147 (*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_emit() 178 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled() 232 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity() 234 last_emitted = rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_activity() 259 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity() 277 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local 281 fence_drv in radeon_fence_check_lockup() [all...] |
H A D | uvd_v2_2.c | 43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
|
H A D | evergreen_dma.c | 45 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
|
H A D | uvd_v1_0.c | 85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/virtio/ |
H A D | virtgpu_fence.c | 78 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() 99 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() 117 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() 122 atomic64_set(&vgdev->fence_drv.last_seq, last_seq); in virtio_gpu_fence_event_process()
|
H A D | virtgpu_debugfs.c | 64 (u64)atomic64_read(&vgdev->fence_drv.last_seq), in virtio_gpu_debugfs_irq_info() 65 vgdev->fence_drv.sync_seq); in virtio_gpu_debugfs_irq_info()
|
H A D | virtgpu_kms.c | 130 vgdev->fence_drv.context = dma_fence_context_alloc(1); in virtio_gpu_init() 131 spin_lock_init(&vgdev->fence_drv.lock); in virtio_gpu_init() 132 INIT_LIST_HEAD(&vgdev->fence_drv.fences); in virtio_gpu_init()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/virtio/ |
H A D | virtgpu_fence.c | 79 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() 105 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() 130 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() 135 atomic64_set(&vgdev->fence_drv.last_fence_id, fence_id); in virtio_gpu_fence_event_process()
|
H A D | virtgpu_debugfs.c | 75 (u64)atomic64_read(&vgdev->fence_drv.last_fence_id), in virtio_gpu_debugfs_irq_info() 76 vgdev->fence_drv.current_fence_id); in virtio_gpu_debugfs_irq_info()
|
H A D | virtgpu_kms.c | 150 vgdev->fence_drv.context = dma_fence_context_alloc(1); in virtio_gpu_init() 151 spin_lock_init(&vgdev->fence_drv.lock); in virtio_gpu_init() 152 INIT_LIST_HEAD(&vgdev->fence_drv.fences); in virtio_gpu_init()
|
H A D | virtgpu_ioctl.c | 162 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_resource_create_ioctl() 249 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_transfer_from_host_ioctl() 309 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_transfer_to_host_ioctl()
|