162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0+ 262306a36Sopenharmony_ci/* Copyright (C) 2018 Broadcom */ 362306a36Sopenharmony_ci 462306a36Sopenharmony_ci/** 562306a36Sopenharmony_ci * DOC: Broadcom V3D scheduling 662306a36Sopenharmony_ci * 762306a36Sopenharmony_ci * The shared DRM GPU scheduler is used to coordinate submitting jobs 862306a36Sopenharmony_ci * to the hardware. Each DRM fd (roughly a client process) gets its 962306a36Sopenharmony_ci * own scheduler entity, which will process jobs in order. The GPU 1062306a36Sopenharmony_ci * scheduler will round-robin between clients to submit the next job. 1162306a36Sopenharmony_ci * 1262306a36Sopenharmony_ci * For simplicity, and in order to keep latency low for interactive 1362306a36Sopenharmony_ci * jobs when bulk background jobs are queued up, we submit a new job 1462306a36Sopenharmony_ci * to the HW only when it has completed the last one, instead of 1562306a36Sopenharmony_ci * filling up the CT[01]Q FIFOs with jobs. Similarly, we use 1662306a36Sopenharmony_ci * drm_sched_job_add_dependency() to manage the dependency between bin and 1762306a36Sopenharmony_ci * render, instead of having the clients submit jobs using the HW's 1862306a36Sopenharmony_ci * semaphores to interlock between them. 1962306a36Sopenharmony_ci */ 2062306a36Sopenharmony_ci 2162306a36Sopenharmony_ci#include <linux/kthread.h> 2262306a36Sopenharmony_ci 2362306a36Sopenharmony_ci#include "v3d_drv.h" 2462306a36Sopenharmony_ci#include "v3d_regs.h" 2562306a36Sopenharmony_ci#include "v3d_trace.h" 2662306a36Sopenharmony_ci 2762306a36Sopenharmony_cistatic struct v3d_job * 2862306a36Sopenharmony_cito_v3d_job(struct drm_sched_job *sched_job) 2962306a36Sopenharmony_ci{ 3062306a36Sopenharmony_ci return container_of(sched_job, struct v3d_job, base); 3162306a36Sopenharmony_ci} 3262306a36Sopenharmony_ci 3362306a36Sopenharmony_cistatic struct v3d_bin_job * 3462306a36Sopenharmony_cito_bin_job(struct drm_sched_job *sched_job) 3562306a36Sopenharmony_ci{ 3662306a36Sopenharmony_ci return container_of(sched_job, struct v3d_bin_job, base.base); 3762306a36Sopenharmony_ci} 3862306a36Sopenharmony_ci 3962306a36Sopenharmony_cistatic struct v3d_render_job * 4062306a36Sopenharmony_cito_render_job(struct drm_sched_job *sched_job) 4162306a36Sopenharmony_ci{ 4262306a36Sopenharmony_ci return container_of(sched_job, struct v3d_render_job, base.base); 4362306a36Sopenharmony_ci} 4462306a36Sopenharmony_ci 4562306a36Sopenharmony_cistatic struct v3d_tfu_job * 4662306a36Sopenharmony_cito_tfu_job(struct drm_sched_job *sched_job) 4762306a36Sopenharmony_ci{ 4862306a36Sopenharmony_ci return container_of(sched_job, struct v3d_tfu_job, base.base); 4962306a36Sopenharmony_ci} 5062306a36Sopenharmony_ci 5162306a36Sopenharmony_cistatic struct v3d_csd_job * 5262306a36Sopenharmony_cito_csd_job(struct drm_sched_job *sched_job) 5362306a36Sopenharmony_ci{ 5462306a36Sopenharmony_ci return container_of(sched_job, struct v3d_csd_job, base.base); 5562306a36Sopenharmony_ci} 5662306a36Sopenharmony_ci 5762306a36Sopenharmony_cistatic void 5862306a36Sopenharmony_civ3d_sched_job_free(struct drm_sched_job *sched_job) 5962306a36Sopenharmony_ci{ 6062306a36Sopenharmony_ci struct v3d_job *job = to_v3d_job(sched_job); 6162306a36Sopenharmony_ci 6262306a36Sopenharmony_ci v3d_job_cleanup(job); 6362306a36Sopenharmony_ci} 6462306a36Sopenharmony_ci 6562306a36Sopenharmony_cistatic void 6662306a36Sopenharmony_civ3d_switch_perfmon(struct v3d_dev *v3d, struct v3d_job *job) 6762306a36Sopenharmony_ci{ 6862306a36Sopenharmony_ci if (job->perfmon != v3d->active_perfmon) 6962306a36Sopenharmony_ci v3d_perfmon_stop(v3d, v3d->active_perfmon, true); 7062306a36Sopenharmony_ci 7162306a36Sopenharmony_ci if (job->perfmon && v3d->active_perfmon != job->perfmon) 7262306a36Sopenharmony_ci v3d_perfmon_start(v3d, job->perfmon); 7362306a36Sopenharmony_ci} 7462306a36Sopenharmony_ci 7562306a36Sopenharmony_cistatic struct dma_fence *v3d_bin_job_run(struct drm_sched_job *sched_job) 7662306a36Sopenharmony_ci{ 7762306a36Sopenharmony_ci struct v3d_bin_job *job = to_bin_job(sched_job); 7862306a36Sopenharmony_ci struct v3d_dev *v3d = job->base.v3d; 7962306a36Sopenharmony_ci struct drm_device *dev = &v3d->drm; 8062306a36Sopenharmony_ci struct dma_fence *fence; 8162306a36Sopenharmony_ci unsigned long irqflags; 8262306a36Sopenharmony_ci 8362306a36Sopenharmony_ci if (unlikely(job->base.base.s_fence->finished.error)) 8462306a36Sopenharmony_ci return NULL; 8562306a36Sopenharmony_ci 8662306a36Sopenharmony_ci /* Lock required around bin_job update vs 8762306a36Sopenharmony_ci * v3d_overflow_mem_work(). 8862306a36Sopenharmony_ci */ 8962306a36Sopenharmony_ci spin_lock_irqsave(&v3d->job_lock, irqflags); 9062306a36Sopenharmony_ci v3d->bin_job = job; 9162306a36Sopenharmony_ci /* Clear out the overflow allocation, so we don't 9262306a36Sopenharmony_ci * reuse the overflow attached to a previous job. 9362306a36Sopenharmony_ci */ 9462306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_PTB_BPOS, 0); 9562306a36Sopenharmony_ci spin_unlock_irqrestore(&v3d->job_lock, irqflags); 9662306a36Sopenharmony_ci 9762306a36Sopenharmony_ci v3d_invalidate_caches(v3d); 9862306a36Sopenharmony_ci 9962306a36Sopenharmony_ci fence = v3d_fence_create(v3d, V3D_BIN); 10062306a36Sopenharmony_ci if (IS_ERR(fence)) 10162306a36Sopenharmony_ci return NULL; 10262306a36Sopenharmony_ci 10362306a36Sopenharmony_ci if (job->base.irq_fence) 10462306a36Sopenharmony_ci dma_fence_put(job->base.irq_fence); 10562306a36Sopenharmony_ci job->base.irq_fence = dma_fence_get(fence); 10662306a36Sopenharmony_ci 10762306a36Sopenharmony_ci trace_v3d_submit_cl(dev, false, to_v3d_fence(fence)->seqno, 10862306a36Sopenharmony_ci job->start, job->end); 10962306a36Sopenharmony_ci 11062306a36Sopenharmony_ci v3d_switch_perfmon(v3d, &job->base); 11162306a36Sopenharmony_ci 11262306a36Sopenharmony_ci /* Set the current and end address of the control list. 11362306a36Sopenharmony_ci * Writing the end register is what starts the job. 11462306a36Sopenharmony_ci */ 11562306a36Sopenharmony_ci if (job->qma) { 11662306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT0QMA, job->qma); 11762306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT0QMS, job->qms); 11862306a36Sopenharmony_ci } 11962306a36Sopenharmony_ci if (job->qts) { 12062306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT0QTS, 12162306a36Sopenharmony_ci V3D_CLE_CT0QTS_ENABLE | 12262306a36Sopenharmony_ci job->qts); 12362306a36Sopenharmony_ci } 12462306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT0QBA, job->start); 12562306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT0QEA, job->end); 12662306a36Sopenharmony_ci 12762306a36Sopenharmony_ci return fence; 12862306a36Sopenharmony_ci} 12962306a36Sopenharmony_ci 13062306a36Sopenharmony_cistatic struct dma_fence *v3d_render_job_run(struct drm_sched_job *sched_job) 13162306a36Sopenharmony_ci{ 13262306a36Sopenharmony_ci struct v3d_render_job *job = to_render_job(sched_job); 13362306a36Sopenharmony_ci struct v3d_dev *v3d = job->base.v3d; 13462306a36Sopenharmony_ci struct drm_device *dev = &v3d->drm; 13562306a36Sopenharmony_ci struct dma_fence *fence; 13662306a36Sopenharmony_ci 13762306a36Sopenharmony_ci if (unlikely(job->base.base.s_fence->finished.error)) 13862306a36Sopenharmony_ci return NULL; 13962306a36Sopenharmony_ci 14062306a36Sopenharmony_ci v3d->render_job = job; 14162306a36Sopenharmony_ci 14262306a36Sopenharmony_ci /* Can we avoid this flush? We need to be careful of 14362306a36Sopenharmony_ci * scheduling, though -- imagine job0 rendering to texture and 14462306a36Sopenharmony_ci * job1 reading, and them being executed as bin0, bin1, 14562306a36Sopenharmony_ci * render0, render1, so that render1's flush at bin time 14662306a36Sopenharmony_ci * wasn't enough. 14762306a36Sopenharmony_ci */ 14862306a36Sopenharmony_ci v3d_invalidate_caches(v3d); 14962306a36Sopenharmony_ci 15062306a36Sopenharmony_ci fence = v3d_fence_create(v3d, V3D_RENDER); 15162306a36Sopenharmony_ci if (IS_ERR(fence)) 15262306a36Sopenharmony_ci return NULL; 15362306a36Sopenharmony_ci 15462306a36Sopenharmony_ci if (job->base.irq_fence) 15562306a36Sopenharmony_ci dma_fence_put(job->base.irq_fence); 15662306a36Sopenharmony_ci job->base.irq_fence = dma_fence_get(fence); 15762306a36Sopenharmony_ci 15862306a36Sopenharmony_ci trace_v3d_submit_cl(dev, true, to_v3d_fence(fence)->seqno, 15962306a36Sopenharmony_ci job->start, job->end); 16062306a36Sopenharmony_ci 16162306a36Sopenharmony_ci v3d_switch_perfmon(v3d, &job->base); 16262306a36Sopenharmony_ci 16362306a36Sopenharmony_ci /* XXX: Set the QCFG */ 16462306a36Sopenharmony_ci 16562306a36Sopenharmony_ci /* Set the current and end address of the control list. 16662306a36Sopenharmony_ci * Writing the end register is what starts the job. 16762306a36Sopenharmony_ci */ 16862306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT1QBA, job->start); 16962306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CLE_CT1QEA, job->end); 17062306a36Sopenharmony_ci 17162306a36Sopenharmony_ci return fence; 17262306a36Sopenharmony_ci} 17362306a36Sopenharmony_ci 17462306a36Sopenharmony_cistatic struct dma_fence * 17562306a36Sopenharmony_civ3d_tfu_job_run(struct drm_sched_job *sched_job) 17662306a36Sopenharmony_ci{ 17762306a36Sopenharmony_ci struct v3d_tfu_job *job = to_tfu_job(sched_job); 17862306a36Sopenharmony_ci struct v3d_dev *v3d = job->base.v3d; 17962306a36Sopenharmony_ci struct drm_device *dev = &v3d->drm; 18062306a36Sopenharmony_ci struct dma_fence *fence; 18162306a36Sopenharmony_ci 18262306a36Sopenharmony_ci fence = v3d_fence_create(v3d, V3D_TFU); 18362306a36Sopenharmony_ci if (IS_ERR(fence)) 18462306a36Sopenharmony_ci return NULL; 18562306a36Sopenharmony_ci 18662306a36Sopenharmony_ci v3d->tfu_job = job; 18762306a36Sopenharmony_ci if (job->base.irq_fence) 18862306a36Sopenharmony_ci dma_fence_put(job->base.irq_fence); 18962306a36Sopenharmony_ci job->base.irq_fence = dma_fence_get(fence); 19062306a36Sopenharmony_ci 19162306a36Sopenharmony_ci trace_v3d_submit_tfu(dev, to_v3d_fence(fence)->seqno); 19262306a36Sopenharmony_ci 19362306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_IIA, job->args.iia); 19462306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_IIS, job->args.iis); 19562306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_ICA, job->args.ica); 19662306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_IUA, job->args.iua); 19762306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_IOA, job->args.ioa); 19862306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_IOS, job->args.ios); 19962306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_COEF0, job->args.coef[0]); 20062306a36Sopenharmony_ci if (job->args.coef[0] & V3D_TFU_COEF0_USECOEF) { 20162306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_COEF1, job->args.coef[1]); 20262306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_COEF2, job->args.coef[2]); 20362306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_COEF3, job->args.coef[3]); 20462306a36Sopenharmony_ci } 20562306a36Sopenharmony_ci /* ICFG kicks off the job. */ 20662306a36Sopenharmony_ci V3D_WRITE(V3D_TFU_ICFG, job->args.icfg | V3D_TFU_ICFG_IOC); 20762306a36Sopenharmony_ci 20862306a36Sopenharmony_ci return fence; 20962306a36Sopenharmony_ci} 21062306a36Sopenharmony_ci 21162306a36Sopenharmony_cistatic struct dma_fence * 21262306a36Sopenharmony_civ3d_csd_job_run(struct drm_sched_job *sched_job) 21362306a36Sopenharmony_ci{ 21462306a36Sopenharmony_ci struct v3d_csd_job *job = to_csd_job(sched_job); 21562306a36Sopenharmony_ci struct v3d_dev *v3d = job->base.v3d; 21662306a36Sopenharmony_ci struct drm_device *dev = &v3d->drm; 21762306a36Sopenharmony_ci struct dma_fence *fence; 21862306a36Sopenharmony_ci int i; 21962306a36Sopenharmony_ci 22062306a36Sopenharmony_ci v3d->csd_job = job; 22162306a36Sopenharmony_ci 22262306a36Sopenharmony_ci v3d_invalidate_caches(v3d); 22362306a36Sopenharmony_ci 22462306a36Sopenharmony_ci fence = v3d_fence_create(v3d, V3D_CSD); 22562306a36Sopenharmony_ci if (IS_ERR(fence)) 22662306a36Sopenharmony_ci return NULL; 22762306a36Sopenharmony_ci 22862306a36Sopenharmony_ci if (job->base.irq_fence) 22962306a36Sopenharmony_ci dma_fence_put(job->base.irq_fence); 23062306a36Sopenharmony_ci job->base.irq_fence = dma_fence_get(fence); 23162306a36Sopenharmony_ci 23262306a36Sopenharmony_ci trace_v3d_submit_csd(dev, to_v3d_fence(fence)->seqno); 23362306a36Sopenharmony_ci 23462306a36Sopenharmony_ci v3d_switch_perfmon(v3d, &job->base); 23562306a36Sopenharmony_ci 23662306a36Sopenharmony_ci for (i = 1; i <= 6; i++) 23762306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CSD_QUEUED_CFG0 + 4 * i, job->args.cfg[i]); 23862306a36Sopenharmony_ci /* CFG0 write kicks off the job. */ 23962306a36Sopenharmony_ci V3D_CORE_WRITE(0, V3D_CSD_QUEUED_CFG0, job->args.cfg[0]); 24062306a36Sopenharmony_ci 24162306a36Sopenharmony_ci return fence; 24262306a36Sopenharmony_ci} 24362306a36Sopenharmony_ci 24462306a36Sopenharmony_cistatic struct dma_fence * 24562306a36Sopenharmony_civ3d_cache_clean_job_run(struct drm_sched_job *sched_job) 24662306a36Sopenharmony_ci{ 24762306a36Sopenharmony_ci struct v3d_job *job = to_v3d_job(sched_job); 24862306a36Sopenharmony_ci struct v3d_dev *v3d = job->v3d; 24962306a36Sopenharmony_ci 25062306a36Sopenharmony_ci v3d_clean_caches(v3d); 25162306a36Sopenharmony_ci 25262306a36Sopenharmony_ci return NULL; 25362306a36Sopenharmony_ci} 25462306a36Sopenharmony_ci 25562306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 25662306a36Sopenharmony_civ3d_gpu_reset_for_timeout(struct v3d_dev *v3d, struct drm_sched_job *sched_job) 25762306a36Sopenharmony_ci{ 25862306a36Sopenharmony_ci enum v3d_queue q; 25962306a36Sopenharmony_ci 26062306a36Sopenharmony_ci mutex_lock(&v3d->reset_lock); 26162306a36Sopenharmony_ci 26262306a36Sopenharmony_ci /* block scheduler */ 26362306a36Sopenharmony_ci for (q = 0; q < V3D_MAX_QUEUES; q++) 26462306a36Sopenharmony_ci drm_sched_stop(&v3d->queue[q].sched, sched_job); 26562306a36Sopenharmony_ci 26662306a36Sopenharmony_ci if (sched_job) 26762306a36Sopenharmony_ci drm_sched_increase_karma(sched_job); 26862306a36Sopenharmony_ci 26962306a36Sopenharmony_ci /* get the GPU back into the init state */ 27062306a36Sopenharmony_ci v3d_reset(v3d); 27162306a36Sopenharmony_ci 27262306a36Sopenharmony_ci for (q = 0; q < V3D_MAX_QUEUES; q++) 27362306a36Sopenharmony_ci drm_sched_resubmit_jobs(&v3d->queue[q].sched); 27462306a36Sopenharmony_ci 27562306a36Sopenharmony_ci /* Unblock schedulers and restart their jobs. */ 27662306a36Sopenharmony_ci for (q = 0; q < V3D_MAX_QUEUES; q++) { 27762306a36Sopenharmony_ci drm_sched_start(&v3d->queue[q].sched, true); 27862306a36Sopenharmony_ci } 27962306a36Sopenharmony_ci 28062306a36Sopenharmony_ci mutex_unlock(&v3d->reset_lock); 28162306a36Sopenharmony_ci 28262306a36Sopenharmony_ci return DRM_GPU_SCHED_STAT_NOMINAL; 28362306a36Sopenharmony_ci} 28462306a36Sopenharmony_ci 28562306a36Sopenharmony_ci/* If the current address or return address have changed, then the GPU 28662306a36Sopenharmony_ci * has probably made progress and we should delay the reset. This 28762306a36Sopenharmony_ci * could fail if the GPU got in an infinite loop in the CL, but that 28862306a36Sopenharmony_ci * is pretty unlikely outside of an i-g-t testcase. 28962306a36Sopenharmony_ci */ 29062306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 29162306a36Sopenharmony_civ3d_cl_job_timedout(struct drm_sched_job *sched_job, enum v3d_queue q, 29262306a36Sopenharmony_ci u32 *timedout_ctca, u32 *timedout_ctra) 29362306a36Sopenharmony_ci{ 29462306a36Sopenharmony_ci struct v3d_job *job = to_v3d_job(sched_job); 29562306a36Sopenharmony_ci struct v3d_dev *v3d = job->v3d; 29662306a36Sopenharmony_ci u32 ctca = V3D_CORE_READ(0, V3D_CLE_CTNCA(q)); 29762306a36Sopenharmony_ci u32 ctra = V3D_CORE_READ(0, V3D_CLE_CTNRA(q)); 29862306a36Sopenharmony_ci 29962306a36Sopenharmony_ci if (*timedout_ctca != ctca || *timedout_ctra != ctra) { 30062306a36Sopenharmony_ci *timedout_ctca = ctca; 30162306a36Sopenharmony_ci *timedout_ctra = ctra; 30262306a36Sopenharmony_ci return DRM_GPU_SCHED_STAT_NOMINAL; 30362306a36Sopenharmony_ci } 30462306a36Sopenharmony_ci 30562306a36Sopenharmony_ci return v3d_gpu_reset_for_timeout(v3d, sched_job); 30662306a36Sopenharmony_ci} 30762306a36Sopenharmony_ci 30862306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 30962306a36Sopenharmony_civ3d_bin_job_timedout(struct drm_sched_job *sched_job) 31062306a36Sopenharmony_ci{ 31162306a36Sopenharmony_ci struct v3d_bin_job *job = to_bin_job(sched_job); 31262306a36Sopenharmony_ci 31362306a36Sopenharmony_ci return v3d_cl_job_timedout(sched_job, V3D_BIN, 31462306a36Sopenharmony_ci &job->timedout_ctca, &job->timedout_ctra); 31562306a36Sopenharmony_ci} 31662306a36Sopenharmony_ci 31762306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 31862306a36Sopenharmony_civ3d_render_job_timedout(struct drm_sched_job *sched_job) 31962306a36Sopenharmony_ci{ 32062306a36Sopenharmony_ci struct v3d_render_job *job = to_render_job(sched_job); 32162306a36Sopenharmony_ci 32262306a36Sopenharmony_ci return v3d_cl_job_timedout(sched_job, V3D_RENDER, 32362306a36Sopenharmony_ci &job->timedout_ctca, &job->timedout_ctra); 32462306a36Sopenharmony_ci} 32562306a36Sopenharmony_ci 32662306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 32762306a36Sopenharmony_civ3d_generic_job_timedout(struct drm_sched_job *sched_job) 32862306a36Sopenharmony_ci{ 32962306a36Sopenharmony_ci struct v3d_job *job = to_v3d_job(sched_job); 33062306a36Sopenharmony_ci 33162306a36Sopenharmony_ci return v3d_gpu_reset_for_timeout(job->v3d, sched_job); 33262306a36Sopenharmony_ci} 33362306a36Sopenharmony_ci 33462306a36Sopenharmony_cistatic enum drm_gpu_sched_stat 33562306a36Sopenharmony_civ3d_csd_job_timedout(struct drm_sched_job *sched_job) 33662306a36Sopenharmony_ci{ 33762306a36Sopenharmony_ci struct v3d_csd_job *job = to_csd_job(sched_job); 33862306a36Sopenharmony_ci struct v3d_dev *v3d = job->base.v3d; 33962306a36Sopenharmony_ci u32 batches = V3D_CORE_READ(0, V3D_CSD_CURRENT_CFG4); 34062306a36Sopenharmony_ci 34162306a36Sopenharmony_ci /* If we've made progress, skip reset and let the timer get 34262306a36Sopenharmony_ci * rearmed. 34362306a36Sopenharmony_ci */ 34462306a36Sopenharmony_ci if (job->timedout_batches != batches) { 34562306a36Sopenharmony_ci job->timedout_batches = batches; 34662306a36Sopenharmony_ci return DRM_GPU_SCHED_STAT_NOMINAL; 34762306a36Sopenharmony_ci } 34862306a36Sopenharmony_ci 34962306a36Sopenharmony_ci return v3d_gpu_reset_for_timeout(v3d, sched_job); 35062306a36Sopenharmony_ci} 35162306a36Sopenharmony_ci 35262306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_bin_sched_ops = { 35362306a36Sopenharmony_ci .run_job = v3d_bin_job_run, 35462306a36Sopenharmony_ci .timedout_job = v3d_bin_job_timedout, 35562306a36Sopenharmony_ci .free_job = v3d_sched_job_free, 35662306a36Sopenharmony_ci}; 35762306a36Sopenharmony_ci 35862306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_render_sched_ops = { 35962306a36Sopenharmony_ci .run_job = v3d_render_job_run, 36062306a36Sopenharmony_ci .timedout_job = v3d_render_job_timedout, 36162306a36Sopenharmony_ci .free_job = v3d_sched_job_free, 36262306a36Sopenharmony_ci}; 36362306a36Sopenharmony_ci 36462306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_tfu_sched_ops = { 36562306a36Sopenharmony_ci .run_job = v3d_tfu_job_run, 36662306a36Sopenharmony_ci .timedout_job = v3d_generic_job_timedout, 36762306a36Sopenharmony_ci .free_job = v3d_sched_job_free, 36862306a36Sopenharmony_ci}; 36962306a36Sopenharmony_ci 37062306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_csd_sched_ops = { 37162306a36Sopenharmony_ci .run_job = v3d_csd_job_run, 37262306a36Sopenharmony_ci .timedout_job = v3d_csd_job_timedout, 37362306a36Sopenharmony_ci .free_job = v3d_sched_job_free 37462306a36Sopenharmony_ci}; 37562306a36Sopenharmony_ci 37662306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_cache_clean_sched_ops = { 37762306a36Sopenharmony_ci .run_job = v3d_cache_clean_job_run, 37862306a36Sopenharmony_ci .timedout_job = v3d_generic_job_timedout, 37962306a36Sopenharmony_ci .free_job = v3d_sched_job_free 38062306a36Sopenharmony_ci}; 38162306a36Sopenharmony_ci 38262306a36Sopenharmony_ciint 38362306a36Sopenharmony_civ3d_sched_init(struct v3d_dev *v3d) 38462306a36Sopenharmony_ci{ 38562306a36Sopenharmony_ci int hw_jobs_limit = 1; 38662306a36Sopenharmony_ci int job_hang_limit = 0; 38762306a36Sopenharmony_ci int hang_limit_ms = 500; 38862306a36Sopenharmony_ci int ret; 38962306a36Sopenharmony_ci 39062306a36Sopenharmony_ci ret = drm_sched_init(&v3d->queue[V3D_BIN].sched, 39162306a36Sopenharmony_ci &v3d_bin_sched_ops, 39262306a36Sopenharmony_ci hw_jobs_limit, job_hang_limit, 39362306a36Sopenharmony_ci msecs_to_jiffies(hang_limit_ms), NULL, 39462306a36Sopenharmony_ci NULL, "v3d_bin", v3d->drm.dev); 39562306a36Sopenharmony_ci if (ret) 39662306a36Sopenharmony_ci return ret; 39762306a36Sopenharmony_ci 39862306a36Sopenharmony_ci ret = drm_sched_init(&v3d->queue[V3D_RENDER].sched, 39962306a36Sopenharmony_ci &v3d_render_sched_ops, 40062306a36Sopenharmony_ci hw_jobs_limit, job_hang_limit, 40162306a36Sopenharmony_ci msecs_to_jiffies(hang_limit_ms), NULL, 40262306a36Sopenharmony_ci NULL, "v3d_render", v3d->drm.dev); 40362306a36Sopenharmony_ci if (ret) 40462306a36Sopenharmony_ci goto fail; 40562306a36Sopenharmony_ci 40662306a36Sopenharmony_ci ret = drm_sched_init(&v3d->queue[V3D_TFU].sched, 40762306a36Sopenharmony_ci &v3d_tfu_sched_ops, 40862306a36Sopenharmony_ci hw_jobs_limit, job_hang_limit, 40962306a36Sopenharmony_ci msecs_to_jiffies(hang_limit_ms), NULL, 41062306a36Sopenharmony_ci NULL, "v3d_tfu", v3d->drm.dev); 41162306a36Sopenharmony_ci if (ret) 41262306a36Sopenharmony_ci goto fail; 41362306a36Sopenharmony_ci 41462306a36Sopenharmony_ci if (v3d_has_csd(v3d)) { 41562306a36Sopenharmony_ci ret = drm_sched_init(&v3d->queue[V3D_CSD].sched, 41662306a36Sopenharmony_ci &v3d_csd_sched_ops, 41762306a36Sopenharmony_ci hw_jobs_limit, job_hang_limit, 41862306a36Sopenharmony_ci msecs_to_jiffies(hang_limit_ms), NULL, 41962306a36Sopenharmony_ci NULL, "v3d_csd", v3d->drm.dev); 42062306a36Sopenharmony_ci if (ret) 42162306a36Sopenharmony_ci goto fail; 42262306a36Sopenharmony_ci 42362306a36Sopenharmony_ci ret = drm_sched_init(&v3d->queue[V3D_CACHE_CLEAN].sched, 42462306a36Sopenharmony_ci &v3d_cache_clean_sched_ops, 42562306a36Sopenharmony_ci hw_jobs_limit, job_hang_limit, 42662306a36Sopenharmony_ci msecs_to_jiffies(hang_limit_ms), NULL, 42762306a36Sopenharmony_ci NULL, "v3d_cache_clean", v3d->drm.dev); 42862306a36Sopenharmony_ci if (ret) 42962306a36Sopenharmony_ci goto fail; 43062306a36Sopenharmony_ci } 43162306a36Sopenharmony_ci 43262306a36Sopenharmony_ci return 0; 43362306a36Sopenharmony_ci 43462306a36Sopenharmony_cifail: 43562306a36Sopenharmony_ci v3d_sched_fini(v3d); 43662306a36Sopenharmony_ci return ret; 43762306a36Sopenharmony_ci} 43862306a36Sopenharmony_ci 43962306a36Sopenharmony_civoid 44062306a36Sopenharmony_civ3d_sched_fini(struct v3d_dev *v3d) 44162306a36Sopenharmony_ci{ 44262306a36Sopenharmony_ci enum v3d_queue q; 44362306a36Sopenharmony_ci 44462306a36Sopenharmony_ci for (q = 0; q < V3D_MAX_QUEUES; q++) { 44562306a36Sopenharmony_ci if (v3d->queue[q].sched.ready) 44662306a36Sopenharmony_ci drm_sched_fini(&v3d->queue[q].sched); 44762306a36Sopenharmony_ci } 44862306a36Sopenharmony_ci} 449