/third_party/mesa3d/src/broadcom/vulkan/ |
H A D | v3dv_queue.c | 79 queue->last_job_syncs.syncs, 3, in queue_wait_idle() 111 &queue->last_job_syncs.syncs[V3DV_QUEUE_ANY], 1, in queue_wait_idle() 183 queue->last_job_syncs.syncs[queue_type], in export_perfmon_last_job_sync() 203 queue->last_job_syncs.syncs[V3DV_QUEUE_ANY], in export_perfmon_last_job_sync() 484 queue->last_job_syncs.syncs[V3DV_QUEUE_ANY], in process_waits() 514 queue->last_job_syncs.syncs[V3DV_QUEUE_ANY], in process_waits() 544 queue->last_job_syncs.syncs[V3DV_QUEUE_ANY], in process_signals() 615 struct drm_v3d_sem *syncs = in set_in_syncs() local 619 if (!syncs) in set_in_syncs() 623 syncs[ in set_in_syncs() 657 struct drm_v3d_sem *syncs = set_out_syncs() local [all...] |
H A D | v3dv_device.c | 1851 if (queue->last_job_syncs.syncs[i]) { in destroy_queue_syncs() 1853 queue->last_job_syncs.syncs[i]); in destroy_queue_syncs() 1879 &queue->last_job_syncs.syncs[i]); in queue_init()
|
H A D | v3dv_private.h | 264 uint32_t syncs[V3DV_QUEUE_COUNT]; member 999 /* List of syncs to wait before running a job */ 1003 /* List of syncs to signal when all jobs complete */
|
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_renderer_util.c | 25 .syncs = &sync, in vn_renderer_submit_simple_sync() 33 .syncs = &sync, in vn_renderer_submit_simple_sync()
|
H A D | vn_renderer.h | 104 /* syncs to update when the virtual sync queue is signaled */ 105 struct vn_renderer_sync *const *syncs; member 106 /* TODO allow NULL when syncs are all binary? */ 129 struct vn_renderer_sync *const *syncs; member 130 /* TODO allow NULL when syncs are all binary? */
|
H A D | vn_renderer_vtest.c | 451 struct vn_renderer_sync *const *syncs, in vtest_vcmd_sync_wait() 469 syncs[i]->sync_id, in vtest_vcmd_sync_wait() 564 /* write syncs */ in vtest_vcmd_submit_cmd2() 571 batch->syncs[j]->sync_id, in vtest_vcmd_submit_cmd2() 897 * is set to non-zero by GPU after the syncs signal. That would allow us in vtest_wait() 902 vtest_vcmd_sync_wait(vtest, flags, poll_timeout, wait->syncs, in vtest_wait() 448 vtest_vcmd_sync_wait(struct vtest *vtest, uint32_t flags, int poll_timeout, struct vn_renderer_sync *const *syncs, const uint64_t *vals, uint32_t count) vtest_vcmd_sync_wait() argument
|
H A D | vn_renderer_virtgpu.c | 418 struct virtgpu_sync *sync = (struct virtgpu_sync *)wait->syncs[i]; in sim_syncobj_wait() 491 struct vn_renderer_sync *const *syncs, in sim_submit_signal_syncs() 497 struct virtgpu_sync *sync = (struct virtgpu_sync *)syncs[i]; in sim_submit_signal_syncs() 563 ret = sim_submit_signal_syncs(gpu, args.fence_fd, batch->syncs, in sim_submit() 899 /* TODO replace wait->syncs by wait->sync_handles to avoid malloc/loop */ in virtgpu_ioctl_syncobj_timeline_wait() 905 struct virtgpu_sync *sync = (struct virtgpu_sync *)wait->syncs[i]; in virtgpu_ioctl_syncobj_timeline_wait() 1482 * TODO cache shmems as they are costly to set up and usually require syncs in virtgpu_init_shmem_blob_mem() 489 sim_submit_signal_syncs(struct virtgpu *gpu, int sync_fd, struct vn_renderer_sync *const *syncs, const uint64_t *sync_values, uint32_t sync_count, bool cpu) sim_submit_signal_syncs() argument
|
H A D | vn_queue.c | 861 .syncs = &sync, in vn_create_sync_file()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/synchronization/ |
H A D | vktGlobalPriorityQueueTests.cpp | 935 const syncs[] in createGlobalPriorityQueueTests() 951 for (const auto& sync : syncs) in createGlobalPriorityQueueTests()
|
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/synchronization/ |
H A D | vktGlobalPriorityQueueTests.cpp | 1037 const syncs[] in createGlobalPriorityQueueTests() 1053 for (const auto& sync : syncs) in createGlobalPriorityQueueTests()
|
/third_party/skia/third_party/externals/angle2/src/libANGLE/capture/ |
H A D | FrameCapture.cpp | 3135 const gl::SyncManager &syncs = apiState.getSyncManagerForCapture(); in CaptureShareGroupMidExecutionSetup() local 3136 for (const auto &syncIter : syncs) in CaptureShareGroupMidExecutionSetup()
|