/third_party/mesa3d/src/freedreno/drm/ |
H A D | freedreno_priv.h | 270 struct fd_ringbuffer *(*new_ringbuffer)(struct fd_submit *submit, 273 int (*flush)(struct fd_submit *submit, int in_fence_fd, 275 void (*destroy)(struct fd_submit *submit); 295 list_for_each_entry (struct fd_submit, submit, &dev->deferred_submits, node) { in fd_dev_count_deferred_cmds() 296 nr += fd_ringbuffer_cmd_count(submit->primary); in fd_dev_count_deferred_cmds() 357 /* Most recent index in submit's bo table, used to optimize the common 358 * case where a bo is used many times in the same submit.
|
H A D | freedreno_ringbuffer_sp.h | 41 /* A "softpin" implementation of submit/ringbuffer, which lowers CPU overhead 66 /* Flush args, potentially attached to the last submit in the list 74 struct list_head submit_list; /* includes this submit as last element */ 112 struct fd_submit *submit; member 122 uint32_t fd_submit_append_bo(struct fd_submit_sp *submit, struct fd_bo *bo);
|
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_renderer.h | 139 VkResult (*submit)(struct vn_renderer *renderer, member 140 const struct vn_renderer_submit *submit); 269 const struct vn_renderer_submit *submit) in vn_renderer_submit() 271 return renderer->ops.submit(renderer, submit); in vn_renderer_submit() 268 vn_renderer_submit(struct vn_renderer *renderer, const struct vn_renderer_submit *submit) vn_renderer_submit() argument
|
H A D | vn_renderer_util.c | 19 const struct vn_renderer_submit submit = { in vn_renderer_submit_simple_sync() local 38 result = vn_renderer_submit(renderer, &submit); in vn_renderer_submit_simple_sync()
|
/foundation/resourceschedule/ffrt/src/sync/ |
H A D | thread.cpp | 47 ffrt::submit([p, func, arg]() { in ffrt_thread_create() 86 ffrt::submit([thr]() { free(thr); }, {thr}); in ffrt_thread_detach()
|
/foundation/resourceschedule/ffrt/test/ut/testcase/ |
H A D | ut_execute_unit.cpp | 57 ffrt::submit([&]() { x.fetch_add(1); }); in HWTEST_F() 58 ffrt::submit([&]() { x.fetch_add(2); }, {}, {}, ffrt::task_attr().delay(1)); in HWTEST_F()
|
/third_party/skia/experimental/graphite/src/ |
H A D | Context.cpp | 46 void Context::submit(SyncToCpu syncToCpu) { in submit() function in skgpu::Context 47 // TODO: we want Gpu::submit to take an array of command buffers but, for now, it just takes in submit() 49 // command buffers and submit them as a block. in submit() 51 fGpu->submit(fRecordings[0]->fCommandBuffer); in submit()
|
/third_party/mesa3d/src/gallium/drivers/v3d/ |
H A D | v3dx_draw.c | 52 job->submit.bcl_start = job->bcl.bo->offset; in start_binning() 1012 * we need to submit the job before drawing and update the vertex count in v3d_draw_vbo() 1035 job->submit.in_sync_bcl = v3d->out_sync; in v3d_draw_vbo() 1042 job->submit.in_sync_bcl = v3d->out_sync; in v3d_draw_vbo() 1366 struct drm_v3d_submit_csd submit = { 0 }; in v3d_launch_grid() local 1399 submit.cfg[i] |= (v3d->compute_num_workgroups[i] << in v3d_launch_grid() 1421 submit.cfg[3] |= (wgs_per_sg & 0xf) << V3D_CSD_CFG3_WGS_PER_SG_SHIFT; in v3d_launch_grid() 1422 submit.cfg[3] |= in v3d_launch_grid() 1424 submit.cfg[3] |= (wg_size & 0xff) << V3D_CSD_CFG3_WG_SIZE_SHIFT; in v3d_launch_grid() 1428 submit in v3d_launch_grid() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gpu.h | 87 struct etnaviv_gem_submit *submit; member 205 void etnaviv_gpu_recover_hang(struct etnaviv_gem_submit *submit); 212 struct dma_fence *etnaviv_gpu_submit(struct etnaviv_gem_submit *submit);
|
/third_party/mesa3d/src/nouveau/drm-shim/ |
H A D | nouveau_noop.c | 111 struct drm_nouveau_gem_pushbuf *submit = arg; in nouveau_ioctl_gem_pushbuf() local 112 submit->vram_available = 3ULL << 30; in nouveau_ioctl_gem_pushbuf() 113 submit->gart_available = 1ULL << 40; in nouveau_ioctl_gem_pushbuf()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/adreno/ |
H A D | a6xx_gpu.c | 138 static void a6xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a6xx_submit() argument 140 unsigned int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT; in a6xx_submit() 144 struct msm_ringbuffer *ring = submit->ring; in a6xx_submit() 147 a6xx_set_pagetable(a6xx_gpu, ring, submit->queue->ctx); in a6xx_submit() 168 for (i = 0; i < submit->nr_cmds; i++) { in a6xx_submit() 169 switch (submit->cmd[i].type) { in a6xx_submit() 173 if (priv->lastctx == submit->queue->ctx) in a6xx_submit() 178 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a6xx_submit() 179 OUT_RING(ring, upper_32_bits(submit->cmd[i].iova)); in a6xx_submit() 180 OUT_RING(ring, submit in a6xx_submit() [all...] |
H A D | a4xx_gpu.c | 25 static void a4xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a4xx_submit() argument 28 struct msm_ringbuffer *ring = submit->ring; in a4xx_submit() 31 for (i = 0; i < submit->nr_cmds; i++) { in a4xx_submit() 32 switch (submit->cmd[i].type) { in a4xx_submit() 38 if (priv->lastctx == submit->queue->ctx) in a4xx_submit() 43 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a4xx_submit() 44 OUT_RING(ring, submit->cmd[i].size); in a4xx_submit() 51 OUT_RING(ring, submit->seqno); in a4xx_submit() 68 OUT_RING(ring, submit->seqno); in a4xx_submit() 629 .submit [all...] |
H A D | a2xx_gpu.c | 13 static void a2xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a2xx_submit() argument 16 struct msm_ringbuffer *ring = submit->ring; in a2xx_submit() 19 for (i = 0; i < submit->nr_cmds; i++) { in a2xx_submit() 20 switch (submit->cmd[i].type) { in a2xx_submit() 26 if (priv->lastctx == submit->queue->ctx) in a2xx_submit() 31 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a2xx_submit() 32 OUT_RING(ring, submit->cmd[i].size); in a2xx_submit() 39 OUT_RING(ring, submit->seqno); in a2xx_submit() 48 OUT_RING(ring, submit->seqno); in a2xx_submit() 479 .submit [all...] |
/third_party/mesa3d/src/virtio/venus-protocol/ |
H A D | vn_protocol_driver_pipeline.h | 2276 static inline void vn_submit_vkCreateGraphicsPipelines(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkGraphicsPipelineCreateInfo* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkPipeline* pPipelines, struct vn_instance_submit_command *submit) in vn_submit_vkCreateGraphicsPipelines() argument 2288 struct vn_cs_encoder *enc = vn_instance_submit_command_init(vn_instance, submit, cmd_data, cmd_size, reply_size); in vn_submit_vkCreateGraphicsPipelines() 2291 vn_instance_submit_command(vn_instance, submit); in vn_submit_vkCreateGraphicsPipelines() 2297 static inline void vn_submit_vkCreateComputePipelines(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkComputePipelineCreateInfo* pCreateInfos, const VkAllocationCallbacks* pAllocator, VkPipeline* pPipelines, struct vn_instance_submit_command *submit) in vn_submit_vkCreateComputePipelines() argument 2309 struct vn_cs_encoder *enc = vn_instance_submit_command_init(vn_instance, submit, cmd_data, cmd_size, reply_size); in vn_submit_vkCreateComputePipelines() 2312 vn_instance_submit_command(vn_instance, submit); in vn_submit_vkCreateComputePipelines() 2318 static inline void vn_submit_vkDestroyPipeline(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkPipeline pipeline, const VkAllocationCallbacks* pAllocator, struct vn_instance_submit_command *submit) in vn_submit_vkDestroyPipeline() argument 2330 struct vn_cs_encoder *enc = vn_instance_submit_command_init(vn_instance, submit, cmd_data, cmd_size, reply_size); in vn_submit_vkDestroyPipeline() 2333 vn_instance_submit_command(vn_instance, submit); in vn_submit_vkDestroyPipeline() 2343 struct vn_instance_submit_command submit; in vn_call_vkCreateGraphicsPipelines() local 2357 struct vn_instance_submit_command submit; vn_async_vkCreateGraphicsPipelines() local 2365 struct vn_instance_submit_command submit; vn_call_vkCreateComputePipelines() local 2379 struct vn_instance_submit_command submit; vn_async_vkCreateComputePipelines() local 2387 struct vn_instance_submit_command submit; vn_call_vkDestroyPipeline() local 2398 struct vn_instance_submit_command submit; vn_async_vkDestroyPipeline() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/msm/adreno/ |
H A D | a2xx_gpu.c | 13 static void a2xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a2xx_submit() argument 15 struct msm_ringbuffer *ring = submit->ring; in a2xx_submit() 18 for (i = 0; i < submit->nr_cmds; i++) { in a2xx_submit() 19 switch (submit->cmd[i].type) { in a2xx_submit() 25 if (gpu->cur_ctx_seqno == submit->queue->ctx->seqno) in a2xx_submit() 30 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a2xx_submit() 31 OUT_RING(ring, submit->cmd[i].size); in a2xx_submit() 38 OUT_RING(ring, submit->seqno); in a2xx_submit() 47 OUT_RING(ring, submit->seqno); in a2xx_submit() 498 .submit [all...] |
H A D | a4xx_gpu.c | 25 static void a4xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a4xx_submit() argument 27 struct msm_ringbuffer *ring = submit->ring; in a4xx_submit() 30 for (i = 0; i < submit->nr_cmds; i++) { in a4xx_submit() 31 switch (submit->cmd[i].type) { in a4xx_submit() 37 if (gpu->cur_ctx_seqno == submit->queue->ctx->seqno) in a4xx_submit() 42 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a4xx_submit() 43 OUT_RING(ring, submit->cmd[i].size); in a4xx_submit() 50 OUT_RING(ring, submit->seqno); in a4xx_submit() 67 OUT_RING(ring, submit->seqno); in a4xx_submit() 638 .submit [all...] |
/foundation/resourceschedule/soc_perf/services/core/src/ |
H A D | socperf_thread_wrap.cpp | 94 socperfQueue_.submit(initResourceNodeInfoFunc); in InitResourceNodeInfo() 116 socperfQueue_.submit(doFreqActionPackFunc); in DoFreqActionPack() 135 socperfQueue_.submit(updatePowerLimitBoostFreqFunc); in UpdatePowerLimitBoostFreq() 154 socperfQueue_.submit(updateThermalLimitBoostFreqFunc); in UpdateThermalLimitBoostFreq() 181 socperfQueue_.submit(updateLimitStatusFunc); in UpdateLimitStatus() 201 socperfQueue_.submit(updateLimitStatusFunc); in ClearAllAliveRequest() 322 socperfQueue_.submit(postDelayTaskFunc, taskAttr); in PostDelayTask()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/ |
H A D | vktBuildPrograms.cpp | 100 void submit (Task* task); 127 void TaskExecutor::submit (Task* task) in submit() function in vkt::__anon28945::TaskExecutor 171 submit(&syncTasks[ndx]); in waitForComplete() 463 executor.submit(&buildGlslTasks.back()); in buildPrograms() 478 executor.submit(&buildHlslTasks.back()); in buildPrograms() 493 executor.submit(&buildSpirvAsmTasks.back()); in buildPrograms() 514 executor.submit(&validationTasks.back()); in buildPrograms()
|
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/ |
H A D | vktBuildPrograms.cpp | 100 void submit (Task* task); 127 void TaskExecutor::submit (Task* task) in submit() function in vkt::__anon30112::TaskExecutor 171 submit(&syncTasks[ndx]); in waitForComplete() 458 executor.submit(&buildGlslTasks.back()); in buildPrograms() 473 executor.submit(&buildHlslTasks.back()); in buildPrograms() 488 executor.submit(&buildSpirvAsmTasks.back()); in buildPrograms() 511 executor.submit(&validationTasks.back()); in buildPrograms()
|
/foundation/communication/netmanager_ext/services/networksharemanager/src/ |
H A D | networkshare_upstreammonitor.cpp | 37 ffrtQueue.submit([weakMonitor = std::weak_ptr(this->NetworkMonitor_), netHandle]() mutable { in NetAvailable() 49 ffrtQueue.submit([weakMonitor = std::weak_ptr(this->NetworkMonitor_), netHandle, netAllCap]() mutable { in NetCapabilitiesChange() 61 ffrtQueue.submit([weakMonitor = std::weak_ptr(this->NetworkMonitor_), netHandle, info]() mutable { in NetConnectionPropertiesChange() 72 ffrtQueue.submit([weakMonitor = std::weak_ptr(this->NetworkMonitor_), netHandle]() mutable { in NetLost()
|
/third_party/mesa3d/src/freedreno/computerator/ |
H A D | a4xx.c | 227 cs_ibo_emit(struct fd_ringbuffer *ring, struct fd_submit *submit, in cs_ibo_emit() argument 278 struct fd_submit *submit) in a4xx_emit_grid() 281 submit, 0, FD_RINGBUFFER_PRIMARY | FD_RINGBUFFER_GROWABLE); in a4xx_emit_grid() 285 cs_ibo_emit(ring, submit, kernel); in a4xx_emit_grid() 277 a4xx_emit_grid(struct kernel *kernel, uint32_t grid[3], struct fd_submit *submit) a4xx_emit_grid() argument
|
H A D | main.c | 272 struct fd_submit *submit = fd_submit_new(pipe); in main() local 282 backend->emit_grid(kernel, grid, submit); in main() 287 fd_submit_flush(submit, -1, &fence); in main()
|
/third_party/python/Lib/concurrent/futures/ |
H A D | thread.py | 161 def submit(self, fn, /, *args, **kwargs): member in ThreadPoolExecutor 178 submit.__doc__ = _base.Executor.submit.__doc__
|
/third_party/mesa3d/src/amd/vulkan/winsys/amdgpu/ |
H A D | radv_amdgpu_cs.c | 147 /** Number of IBs to submit in the field ibs. */ 151 * IBs to submit. Those IBs will be submit together as single entity 689 * parent to submit multiple IBs. in radv_amdgpu_cs_execute_secondary() 994 /* Compute the number of IBs for this submit. */ in radv_amdgpu_winsys_cs_submit_fallback() 1087 /* Compute the number of IBs for this submit. */ in radv_amdgpu_winsys_cs_submit_sysmem() 1103 * IB per submit. in radv_amdgpu_winsys_cs_submit_sysmem() 1336 const struct radv_winsys_submit_info *submit, in radv_amdgpu_winsys_cs_submit_internal() 1342 if (!submit->cs_count) { in radv_amdgpu_winsys_cs_submit_internal() 1343 result = radv_amdgpu_cs_submit_zero(ctx, submit in radv_amdgpu_winsys_cs_submit_internal() 1335 radv_amdgpu_winsys_cs_submit_internal(struct radv_amdgpu_ctx *ctx, const struct radv_winsys_submit_info *submit, struct radv_winsys_sem_info *sem_info, bool can_patch) radv_amdgpu_winsys_cs_submit_internal() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/freedreno/ |
H A D | freedreno_batch.c | 57 return fd_submit_new_ringbuffer(batch->submit, sz, flags); in alloc_ring() 65 batch->submit = fd_submit_new(ctx->pipe); in batch_init() 82 /* Work around problems on earlier gens with submit merging, etc, in batch_init() 83 * by always creating a fence to request that the submit is flushed in batch_init() 150 if (!batch->submit) in cleanup_submit() 181 fd_submit_del(batch->submit); in cleanup_submit() 182 batch->submit = NULL; in cleanup_submit()
|