162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0+
262306a36Sopenharmony_ci/* Copyright (C) 2018 Broadcom */
362306a36Sopenharmony_ci
462306a36Sopenharmony_ci/**
562306a36Sopenharmony_ci * DOC: Broadcom V3D scheduling
662306a36Sopenharmony_ci *
762306a36Sopenharmony_ci * The shared DRM GPU scheduler is used to coordinate submitting jobs
862306a36Sopenharmony_ci * to the hardware.  Each DRM fd (roughly a client process) gets its
962306a36Sopenharmony_ci * own scheduler entity, which will process jobs in order.  The GPU
1062306a36Sopenharmony_ci * scheduler will round-robin between clients to submit the next job.
1162306a36Sopenharmony_ci *
1262306a36Sopenharmony_ci * For simplicity, and in order to keep latency low for interactive
1362306a36Sopenharmony_ci * jobs when bulk background jobs are queued up, we submit a new job
1462306a36Sopenharmony_ci * to the HW only when it has completed the last one, instead of
1562306a36Sopenharmony_ci * filling up the CT[01]Q FIFOs with jobs.  Similarly, we use
1662306a36Sopenharmony_ci * drm_sched_job_add_dependency() to manage the dependency between bin and
1762306a36Sopenharmony_ci * render, instead of having the clients submit jobs using the HW's
1862306a36Sopenharmony_ci * semaphores to interlock between them.
1962306a36Sopenharmony_ci */
2062306a36Sopenharmony_ci
2162306a36Sopenharmony_ci#include <linux/kthread.h>
2262306a36Sopenharmony_ci
2362306a36Sopenharmony_ci#include "v3d_drv.h"
2462306a36Sopenharmony_ci#include "v3d_regs.h"
2562306a36Sopenharmony_ci#include "v3d_trace.h"
2662306a36Sopenharmony_ci
2762306a36Sopenharmony_cistatic struct v3d_job *
2862306a36Sopenharmony_cito_v3d_job(struct drm_sched_job *sched_job)
2962306a36Sopenharmony_ci{
3062306a36Sopenharmony_ci	return container_of(sched_job, struct v3d_job, base);
3162306a36Sopenharmony_ci}
3262306a36Sopenharmony_ci
3362306a36Sopenharmony_cistatic struct v3d_bin_job *
3462306a36Sopenharmony_cito_bin_job(struct drm_sched_job *sched_job)
3562306a36Sopenharmony_ci{
3662306a36Sopenharmony_ci	return container_of(sched_job, struct v3d_bin_job, base.base);
3762306a36Sopenharmony_ci}
3862306a36Sopenharmony_ci
3962306a36Sopenharmony_cistatic struct v3d_render_job *
4062306a36Sopenharmony_cito_render_job(struct drm_sched_job *sched_job)
4162306a36Sopenharmony_ci{
4262306a36Sopenharmony_ci	return container_of(sched_job, struct v3d_render_job, base.base);
4362306a36Sopenharmony_ci}
4462306a36Sopenharmony_ci
4562306a36Sopenharmony_cistatic struct v3d_tfu_job *
4662306a36Sopenharmony_cito_tfu_job(struct drm_sched_job *sched_job)
4762306a36Sopenharmony_ci{
4862306a36Sopenharmony_ci	return container_of(sched_job, struct v3d_tfu_job, base.base);
4962306a36Sopenharmony_ci}
5062306a36Sopenharmony_ci
5162306a36Sopenharmony_cistatic struct v3d_csd_job *
5262306a36Sopenharmony_cito_csd_job(struct drm_sched_job *sched_job)
5362306a36Sopenharmony_ci{
5462306a36Sopenharmony_ci	return container_of(sched_job, struct v3d_csd_job, base.base);
5562306a36Sopenharmony_ci}
5662306a36Sopenharmony_ci
5762306a36Sopenharmony_cistatic void
5862306a36Sopenharmony_civ3d_sched_job_free(struct drm_sched_job *sched_job)
5962306a36Sopenharmony_ci{
6062306a36Sopenharmony_ci	struct v3d_job *job = to_v3d_job(sched_job);
6162306a36Sopenharmony_ci
6262306a36Sopenharmony_ci	v3d_job_cleanup(job);
6362306a36Sopenharmony_ci}
6462306a36Sopenharmony_ci
6562306a36Sopenharmony_cistatic void
6662306a36Sopenharmony_civ3d_switch_perfmon(struct v3d_dev *v3d, struct v3d_job *job)
6762306a36Sopenharmony_ci{
6862306a36Sopenharmony_ci	if (job->perfmon != v3d->active_perfmon)
6962306a36Sopenharmony_ci		v3d_perfmon_stop(v3d, v3d->active_perfmon, true);
7062306a36Sopenharmony_ci
7162306a36Sopenharmony_ci	if (job->perfmon && v3d->active_perfmon != job->perfmon)
7262306a36Sopenharmony_ci		v3d_perfmon_start(v3d, job->perfmon);
7362306a36Sopenharmony_ci}
7462306a36Sopenharmony_ci
7562306a36Sopenharmony_cistatic struct dma_fence *v3d_bin_job_run(struct drm_sched_job *sched_job)
7662306a36Sopenharmony_ci{
7762306a36Sopenharmony_ci	struct v3d_bin_job *job = to_bin_job(sched_job);
7862306a36Sopenharmony_ci	struct v3d_dev *v3d = job->base.v3d;
7962306a36Sopenharmony_ci	struct drm_device *dev = &v3d->drm;
8062306a36Sopenharmony_ci	struct dma_fence *fence;
8162306a36Sopenharmony_ci	unsigned long irqflags;
8262306a36Sopenharmony_ci
8362306a36Sopenharmony_ci	if (unlikely(job->base.base.s_fence->finished.error))
8462306a36Sopenharmony_ci		return NULL;
8562306a36Sopenharmony_ci
8662306a36Sopenharmony_ci	/* Lock required around bin_job update vs
8762306a36Sopenharmony_ci	 * v3d_overflow_mem_work().
8862306a36Sopenharmony_ci	 */
8962306a36Sopenharmony_ci	spin_lock_irqsave(&v3d->job_lock, irqflags);
9062306a36Sopenharmony_ci	v3d->bin_job = job;
9162306a36Sopenharmony_ci	/* Clear out the overflow allocation, so we don't
9262306a36Sopenharmony_ci	 * reuse the overflow attached to a previous job.
9362306a36Sopenharmony_ci	 */
9462306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_PTB_BPOS, 0);
9562306a36Sopenharmony_ci	spin_unlock_irqrestore(&v3d->job_lock, irqflags);
9662306a36Sopenharmony_ci
9762306a36Sopenharmony_ci	v3d_invalidate_caches(v3d);
9862306a36Sopenharmony_ci
9962306a36Sopenharmony_ci	fence = v3d_fence_create(v3d, V3D_BIN);
10062306a36Sopenharmony_ci	if (IS_ERR(fence))
10162306a36Sopenharmony_ci		return NULL;
10262306a36Sopenharmony_ci
10362306a36Sopenharmony_ci	if (job->base.irq_fence)
10462306a36Sopenharmony_ci		dma_fence_put(job->base.irq_fence);
10562306a36Sopenharmony_ci	job->base.irq_fence = dma_fence_get(fence);
10662306a36Sopenharmony_ci
10762306a36Sopenharmony_ci	trace_v3d_submit_cl(dev, false, to_v3d_fence(fence)->seqno,
10862306a36Sopenharmony_ci			    job->start, job->end);
10962306a36Sopenharmony_ci
11062306a36Sopenharmony_ci	v3d_switch_perfmon(v3d, &job->base);
11162306a36Sopenharmony_ci
11262306a36Sopenharmony_ci	/* Set the current and end address of the control list.
11362306a36Sopenharmony_ci	 * Writing the end register is what starts the job.
11462306a36Sopenharmony_ci	 */
11562306a36Sopenharmony_ci	if (job->qma) {
11662306a36Sopenharmony_ci		V3D_CORE_WRITE(0, V3D_CLE_CT0QMA, job->qma);
11762306a36Sopenharmony_ci		V3D_CORE_WRITE(0, V3D_CLE_CT0QMS, job->qms);
11862306a36Sopenharmony_ci	}
11962306a36Sopenharmony_ci	if (job->qts) {
12062306a36Sopenharmony_ci		V3D_CORE_WRITE(0, V3D_CLE_CT0QTS,
12162306a36Sopenharmony_ci			       V3D_CLE_CT0QTS_ENABLE |
12262306a36Sopenharmony_ci			       job->qts);
12362306a36Sopenharmony_ci	}
12462306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_CLE_CT0QBA, job->start);
12562306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_CLE_CT0QEA, job->end);
12662306a36Sopenharmony_ci
12762306a36Sopenharmony_ci	return fence;
12862306a36Sopenharmony_ci}
12962306a36Sopenharmony_ci
13062306a36Sopenharmony_cistatic struct dma_fence *v3d_render_job_run(struct drm_sched_job *sched_job)
13162306a36Sopenharmony_ci{
13262306a36Sopenharmony_ci	struct v3d_render_job *job = to_render_job(sched_job);
13362306a36Sopenharmony_ci	struct v3d_dev *v3d = job->base.v3d;
13462306a36Sopenharmony_ci	struct drm_device *dev = &v3d->drm;
13562306a36Sopenharmony_ci	struct dma_fence *fence;
13662306a36Sopenharmony_ci
13762306a36Sopenharmony_ci	if (unlikely(job->base.base.s_fence->finished.error))
13862306a36Sopenharmony_ci		return NULL;
13962306a36Sopenharmony_ci
14062306a36Sopenharmony_ci	v3d->render_job = job;
14162306a36Sopenharmony_ci
14262306a36Sopenharmony_ci	/* Can we avoid this flush?  We need to be careful of
14362306a36Sopenharmony_ci	 * scheduling, though -- imagine job0 rendering to texture and
14462306a36Sopenharmony_ci	 * job1 reading, and them being executed as bin0, bin1,
14562306a36Sopenharmony_ci	 * render0, render1, so that render1's flush at bin time
14662306a36Sopenharmony_ci	 * wasn't enough.
14762306a36Sopenharmony_ci	 */
14862306a36Sopenharmony_ci	v3d_invalidate_caches(v3d);
14962306a36Sopenharmony_ci
15062306a36Sopenharmony_ci	fence = v3d_fence_create(v3d, V3D_RENDER);
15162306a36Sopenharmony_ci	if (IS_ERR(fence))
15262306a36Sopenharmony_ci		return NULL;
15362306a36Sopenharmony_ci
15462306a36Sopenharmony_ci	if (job->base.irq_fence)
15562306a36Sopenharmony_ci		dma_fence_put(job->base.irq_fence);
15662306a36Sopenharmony_ci	job->base.irq_fence = dma_fence_get(fence);
15762306a36Sopenharmony_ci
15862306a36Sopenharmony_ci	trace_v3d_submit_cl(dev, true, to_v3d_fence(fence)->seqno,
15962306a36Sopenharmony_ci			    job->start, job->end);
16062306a36Sopenharmony_ci
16162306a36Sopenharmony_ci	v3d_switch_perfmon(v3d, &job->base);
16262306a36Sopenharmony_ci
16362306a36Sopenharmony_ci	/* XXX: Set the QCFG */
16462306a36Sopenharmony_ci
16562306a36Sopenharmony_ci	/* Set the current and end address of the control list.
16662306a36Sopenharmony_ci	 * Writing the end register is what starts the job.
16762306a36Sopenharmony_ci	 */
16862306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_CLE_CT1QBA, job->start);
16962306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_CLE_CT1QEA, job->end);
17062306a36Sopenharmony_ci
17162306a36Sopenharmony_ci	return fence;
17262306a36Sopenharmony_ci}
17362306a36Sopenharmony_ci
17462306a36Sopenharmony_cistatic struct dma_fence *
17562306a36Sopenharmony_civ3d_tfu_job_run(struct drm_sched_job *sched_job)
17662306a36Sopenharmony_ci{
17762306a36Sopenharmony_ci	struct v3d_tfu_job *job = to_tfu_job(sched_job);
17862306a36Sopenharmony_ci	struct v3d_dev *v3d = job->base.v3d;
17962306a36Sopenharmony_ci	struct drm_device *dev = &v3d->drm;
18062306a36Sopenharmony_ci	struct dma_fence *fence;
18162306a36Sopenharmony_ci
18262306a36Sopenharmony_ci	fence = v3d_fence_create(v3d, V3D_TFU);
18362306a36Sopenharmony_ci	if (IS_ERR(fence))
18462306a36Sopenharmony_ci		return NULL;
18562306a36Sopenharmony_ci
18662306a36Sopenharmony_ci	v3d->tfu_job = job;
18762306a36Sopenharmony_ci	if (job->base.irq_fence)
18862306a36Sopenharmony_ci		dma_fence_put(job->base.irq_fence);
18962306a36Sopenharmony_ci	job->base.irq_fence = dma_fence_get(fence);
19062306a36Sopenharmony_ci
19162306a36Sopenharmony_ci	trace_v3d_submit_tfu(dev, to_v3d_fence(fence)->seqno);
19262306a36Sopenharmony_ci
19362306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_IIA, job->args.iia);
19462306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_IIS, job->args.iis);
19562306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_ICA, job->args.ica);
19662306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_IUA, job->args.iua);
19762306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_IOA, job->args.ioa);
19862306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_IOS, job->args.ios);
19962306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_COEF0, job->args.coef[0]);
20062306a36Sopenharmony_ci	if (job->args.coef[0] & V3D_TFU_COEF0_USECOEF) {
20162306a36Sopenharmony_ci		V3D_WRITE(V3D_TFU_COEF1, job->args.coef[1]);
20262306a36Sopenharmony_ci		V3D_WRITE(V3D_TFU_COEF2, job->args.coef[2]);
20362306a36Sopenharmony_ci		V3D_WRITE(V3D_TFU_COEF3, job->args.coef[3]);
20462306a36Sopenharmony_ci	}
20562306a36Sopenharmony_ci	/* ICFG kicks off the job. */
20662306a36Sopenharmony_ci	V3D_WRITE(V3D_TFU_ICFG, job->args.icfg | V3D_TFU_ICFG_IOC);
20762306a36Sopenharmony_ci
20862306a36Sopenharmony_ci	return fence;
20962306a36Sopenharmony_ci}
21062306a36Sopenharmony_ci
21162306a36Sopenharmony_cistatic struct dma_fence *
21262306a36Sopenharmony_civ3d_csd_job_run(struct drm_sched_job *sched_job)
21362306a36Sopenharmony_ci{
21462306a36Sopenharmony_ci	struct v3d_csd_job *job = to_csd_job(sched_job);
21562306a36Sopenharmony_ci	struct v3d_dev *v3d = job->base.v3d;
21662306a36Sopenharmony_ci	struct drm_device *dev = &v3d->drm;
21762306a36Sopenharmony_ci	struct dma_fence *fence;
21862306a36Sopenharmony_ci	int i;
21962306a36Sopenharmony_ci
22062306a36Sopenharmony_ci	v3d->csd_job = job;
22162306a36Sopenharmony_ci
22262306a36Sopenharmony_ci	v3d_invalidate_caches(v3d);
22362306a36Sopenharmony_ci
22462306a36Sopenharmony_ci	fence = v3d_fence_create(v3d, V3D_CSD);
22562306a36Sopenharmony_ci	if (IS_ERR(fence))
22662306a36Sopenharmony_ci		return NULL;
22762306a36Sopenharmony_ci
22862306a36Sopenharmony_ci	if (job->base.irq_fence)
22962306a36Sopenharmony_ci		dma_fence_put(job->base.irq_fence);
23062306a36Sopenharmony_ci	job->base.irq_fence = dma_fence_get(fence);
23162306a36Sopenharmony_ci
23262306a36Sopenharmony_ci	trace_v3d_submit_csd(dev, to_v3d_fence(fence)->seqno);
23362306a36Sopenharmony_ci
23462306a36Sopenharmony_ci	v3d_switch_perfmon(v3d, &job->base);
23562306a36Sopenharmony_ci
23662306a36Sopenharmony_ci	for (i = 1; i <= 6; i++)
23762306a36Sopenharmony_ci		V3D_CORE_WRITE(0, V3D_CSD_QUEUED_CFG0 + 4 * i, job->args.cfg[i]);
23862306a36Sopenharmony_ci	/* CFG0 write kicks off the job. */
23962306a36Sopenharmony_ci	V3D_CORE_WRITE(0, V3D_CSD_QUEUED_CFG0, job->args.cfg[0]);
24062306a36Sopenharmony_ci
24162306a36Sopenharmony_ci	return fence;
24262306a36Sopenharmony_ci}
24362306a36Sopenharmony_ci
24462306a36Sopenharmony_cistatic struct dma_fence *
24562306a36Sopenharmony_civ3d_cache_clean_job_run(struct drm_sched_job *sched_job)
24662306a36Sopenharmony_ci{
24762306a36Sopenharmony_ci	struct v3d_job *job = to_v3d_job(sched_job);
24862306a36Sopenharmony_ci	struct v3d_dev *v3d = job->v3d;
24962306a36Sopenharmony_ci
25062306a36Sopenharmony_ci	v3d_clean_caches(v3d);
25162306a36Sopenharmony_ci
25262306a36Sopenharmony_ci	return NULL;
25362306a36Sopenharmony_ci}
25462306a36Sopenharmony_ci
25562306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
25662306a36Sopenharmony_civ3d_gpu_reset_for_timeout(struct v3d_dev *v3d, struct drm_sched_job *sched_job)
25762306a36Sopenharmony_ci{
25862306a36Sopenharmony_ci	enum v3d_queue q;
25962306a36Sopenharmony_ci
26062306a36Sopenharmony_ci	mutex_lock(&v3d->reset_lock);
26162306a36Sopenharmony_ci
26262306a36Sopenharmony_ci	/* block scheduler */
26362306a36Sopenharmony_ci	for (q = 0; q < V3D_MAX_QUEUES; q++)
26462306a36Sopenharmony_ci		drm_sched_stop(&v3d->queue[q].sched, sched_job);
26562306a36Sopenharmony_ci
26662306a36Sopenharmony_ci	if (sched_job)
26762306a36Sopenharmony_ci		drm_sched_increase_karma(sched_job);
26862306a36Sopenharmony_ci
26962306a36Sopenharmony_ci	/* get the GPU back into the init state */
27062306a36Sopenharmony_ci	v3d_reset(v3d);
27162306a36Sopenharmony_ci
27262306a36Sopenharmony_ci	for (q = 0; q < V3D_MAX_QUEUES; q++)
27362306a36Sopenharmony_ci		drm_sched_resubmit_jobs(&v3d->queue[q].sched);
27462306a36Sopenharmony_ci
27562306a36Sopenharmony_ci	/* Unblock schedulers and restart their jobs. */
27662306a36Sopenharmony_ci	for (q = 0; q < V3D_MAX_QUEUES; q++) {
27762306a36Sopenharmony_ci		drm_sched_start(&v3d->queue[q].sched, true);
27862306a36Sopenharmony_ci	}
27962306a36Sopenharmony_ci
28062306a36Sopenharmony_ci	mutex_unlock(&v3d->reset_lock);
28162306a36Sopenharmony_ci
28262306a36Sopenharmony_ci	return DRM_GPU_SCHED_STAT_NOMINAL;
28362306a36Sopenharmony_ci}
28462306a36Sopenharmony_ci
28562306a36Sopenharmony_ci/* If the current address or return address have changed, then the GPU
28662306a36Sopenharmony_ci * has probably made progress and we should delay the reset.  This
28762306a36Sopenharmony_ci * could fail if the GPU got in an infinite loop in the CL, but that
28862306a36Sopenharmony_ci * is pretty unlikely outside of an i-g-t testcase.
28962306a36Sopenharmony_ci */
29062306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
29162306a36Sopenharmony_civ3d_cl_job_timedout(struct drm_sched_job *sched_job, enum v3d_queue q,
29262306a36Sopenharmony_ci		    u32 *timedout_ctca, u32 *timedout_ctra)
29362306a36Sopenharmony_ci{
29462306a36Sopenharmony_ci	struct v3d_job *job = to_v3d_job(sched_job);
29562306a36Sopenharmony_ci	struct v3d_dev *v3d = job->v3d;
29662306a36Sopenharmony_ci	u32 ctca = V3D_CORE_READ(0, V3D_CLE_CTNCA(q));
29762306a36Sopenharmony_ci	u32 ctra = V3D_CORE_READ(0, V3D_CLE_CTNRA(q));
29862306a36Sopenharmony_ci
29962306a36Sopenharmony_ci	if (*timedout_ctca != ctca || *timedout_ctra != ctra) {
30062306a36Sopenharmony_ci		*timedout_ctca = ctca;
30162306a36Sopenharmony_ci		*timedout_ctra = ctra;
30262306a36Sopenharmony_ci		return DRM_GPU_SCHED_STAT_NOMINAL;
30362306a36Sopenharmony_ci	}
30462306a36Sopenharmony_ci
30562306a36Sopenharmony_ci	return v3d_gpu_reset_for_timeout(v3d, sched_job);
30662306a36Sopenharmony_ci}
30762306a36Sopenharmony_ci
30862306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
30962306a36Sopenharmony_civ3d_bin_job_timedout(struct drm_sched_job *sched_job)
31062306a36Sopenharmony_ci{
31162306a36Sopenharmony_ci	struct v3d_bin_job *job = to_bin_job(sched_job);
31262306a36Sopenharmony_ci
31362306a36Sopenharmony_ci	return v3d_cl_job_timedout(sched_job, V3D_BIN,
31462306a36Sopenharmony_ci				   &job->timedout_ctca, &job->timedout_ctra);
31562306a36Sopenharmony_ci}
31662306a36Sopenharmony_ci
31762306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
31862306a36Sopenharmony_civ3d_render_job_timedout(struct drm_sched_job *sched_job)
31962306a36Sopenharmony_ci{
32062306a36Sopenharmony_ci	struct v3d_render_job *job = to_render_job(sched_job);
32162306a36Sopenharmony_ci
32262306a36Sopenharmony_ci	return v3d_cl_job_timedout(sched_job, V3D_RENDER,
32362306a36Sopenharmony_ci				   &job->timedout_ctca, &job->timedout_ctra);
32462306a36Sopenharmony_ci}
32562306a36Sopenharmony_ci
32662306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
32762306a36Sopenharmony_civ3d_generic_job_timedout(struct drm_sched_job *sched_job)
32862306a36Sopenharmony_ci{
32962306a36Sopenharmony_ci	struct v3d_job *job = to_v3d_job(sched_job);
33062306a36Sopenharmony_ci
33162306a36Sopenharmony_ci	return v3d_gpu_reset_for_timeout(job->v3d, sched_job);
33262306a36Sopenharmony_ci}
33362306a36Sopenharmony_ci
33462306a36Sopenharmony_cistatic enum drm_gpu_sched_stat
33562306a36Sopenharmony_civ3d_csd_job_timedout(struct drm_sched_job *sched_job)
33662306a36Sopenharmony_ci{
33762306a36Sopenharmony_ci	struct v3d_csd_job *job = to_csd_job(sched_job);
33862306a36Sopenharmony_ci	struct v3d_dev *v3d = job->base.v3d;
33962306a36Sopenharmony_ci	u32 batches = V3D_CORE_READ(0, V3D_CSD_CURRENT_CFG4);
34062306a36Sopenharmony_ci
34162306a36Sopenharmony_ci	/* If we've made progress, skip reset and let the timer get
34262306a36Sopenharmony_ci	 * rearmed.
34362306a36Sopenharmony_ci	 */
34462306a36Sopenharmony_ci	if (job->timedout_batches != batches) {
34562306a36Sopenharmony_ci		job->timedout_batches = batches;
34662306a36Sopenharmony_ci		return DRM_GPU_SCHED_STAT_NOMINAL;
34762306a36Sopenharmony_ci	}
34862306a36Sopenharmony_ci
34962306a36Sopenharmony_ci	return v3d_gpu_reset_for_timeout(v3d, sched_job);
35062306a36Sopenharmony_ci}
35162306a36Sopenharmony_ci
35262306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_bin_sched_ops = {
35362306a36Sopenharmony_ci	.run_job = v3d_bin_job_run,
35462306a36Sopenharmony_ci	.timedout_job = v3d_bin_job_timedout,
35562306a36Sopenharmony_ci	.free_job = v3d_sched_job_free,
35662306a36Sopenharmony_ci};
35762306a36Sopenharmony_ci
35862306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_render_sched_ops = {
35962306a36Sopenharmony_ci	.run_job = v3d_render_job_run,
36062306a36Sopenharmony_ci	.timedout_job = v3d_render_job_timedout,
36162306a36Sopenharmony_ci	.free_job = v3d_sched_job_free,
36262306a36Sopenharmony_ci};
36362306a36Sopenharmony_ci
36462306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_tfu_sched_ops = {
36562306a36Sopenharmony_ci	.run_job = v3d_tfu_job_run,
36662306a36Sopenharmony_ci	.timedout_job = v3d_generic_job_timedout,
36762306a36Sopenharmony_ci	.free_job = v3d_sched_job_free,
36862306a36Sopenharmony_ci};
36962306a36Sopenharmony_ci
37062306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_csd_sched_ops = {
37162306a36Sopenharmony_ci	.run_job = v3d_csd_job_run,
37262306a36Sopenharmony_ci	.timedout_job = v3d_csd_job_timedout,
37362306a36Sopenharmony_ci	.free_job = v3d_sched_job_free
37462306a36Sopenharmony_ci};
37562306a36Sopenharmony_ci
37662306a36Sopenharmony_cistatic const struct drm_sched_backend_ops v3d_cache_clean_sched_ops = {
37762306a36Sopenharmony_ci	.run_job = v3d_cache_clean_job_run,
37862306a36Sopenharmony_ci	.timedout_job = v3d_generic_job_timedout,
37962306a36Sopenharmony_ci	.free_job = v3d_sched_job_free
38062306a36Sopenharmony_ci};
38162306a36Sopenharmony_ci
38262306a36Sopenharmony_ciint
38362306a36Sopenharmony_civ3d_sched_init(struct v3d_dev *v3d)
38462306a36Sopenharmony_ci{
38562306a36Sopenharmony_ci	int hw_jobs_limit = 1;
38662306a36Sopenharmony_ci	int job_hang_limit = 0;
38762306a36Sopenharmony_ci	int hang_limit_ms = 500;
38862306a36Sopenharmony_ci	int ret;
38962306a36Sopenharmony_ci
39062306a36Sopenharmony_ci	ret = drm_sched_init(&v3d->queue[V3D_BIN].sched,
39162306a36Sopenharmony_ci			     &v3d_bin_sched_ops,
39262306a36Sopenharmony_ci			     hw_jobs_limit, job_hang_limit,
39362306a36Sopenharmony_ci			     msecs_to_jiffies(hang_limit_ms), NULL,
39462306a36Sopenharmony_ci			     NULL, "v3d_bin", v3d->drm.dev);
39562306a36Sopenharmony_ci	if (ret)
39662306a36Sopenharmony_ci		return ret;
39762306a36Sopenharmony_ci
39862306a36Sopenharmony_ci	ret = drm_sched_init(&v3d->queue[V3D_RENDER].sched,
39962306a36Sopenharmony_ci			     &v3d_render_sched_ops,
40062306a36Sopenharmony_ci			     hw_jobs_limit, job_hang_limit,
40162306a36Sopenharmony_ci			     msecs_to_jiffies(hang_limit_ms), NULL,
40262306a36Sopenharmony_ci			     NULL, "v3d_render", v3d->drm.dev);
40362306a36Sopenharmony_ci	if (ret)
40462306a36Sopenharmony_ci		goto fail;
40562306a36Sopenharmony_ci
40662306a36Sopenharmony_ci	ret = drm_sched_init(&v3d->queue[V3D_TFU].sched,
40762306a36Sopenharmony_ci			     &v3d_tfu_sched_ops,
40862306a36Sopenharmony_ci			     hw_jobs_limit, job_hang_limit,
40962306a36Sopenharmony_ci			     msecs_to_jiffies(hang_limit_ms), NULL,
41062306a36Sopenharmony_ci			     NULL, "v3d_tfu", v3d->drm.dev);
41162306a36Sopenharmony_ci	if (ret)
41262306a36Sopenharmony_ci		goto fail;
41362306a36Sopenharmony_ci
41462306a36Sopenharmony_ci	if (v3d_has_csd(v3d)) {
41562306a36Sopenharmony_ci		ret = drm_sched_init(&v3d->queue[V3D_CSD].sched,
41662306a36Sopenharmony_ci				     &v3d_csd_sched_ops,
41762306a36Sopenharmony_ci				     hw_jobs_limit, job_hang_limit,
41862306a36Sopenharmony_ci				     msecs_to_jiffies(hang_limit_ms), NULL,
41962306a36Sopenharmony_ci				     NULL, "v3d_csd", v3d->drm.dev);
42062306a36Sopenharmony_ci		if (ret)
42162306a36Sopenharmony_ci			goto fail;
42262306a36Sopenharmony_ci
42362306a36Sopenharmony_ci		ret = drm_sched_init(&v3d->queue[V3D_CACHE_CLEAN].sched,
42462306a36Sopenharmony_ci				     &v3d_cache_clean_sched_ops,
42562306a36Sopenharmony_ci				     hw_jobs_limit, job_hang_limit,
42662306a36Sopenharmony_ci				     msecs_to_jiffies(hang_limit_ms), NULL,
42762306a36Sopenharmony_ci				     NULL, "v3d_cache_clean", v3d->drm.dev);
42862306a36Sopenharmony_ci		if (ret)
42962306a36Sopenharmony_ci			goto fail;
43062306a36Sopenharmony_ci	}
43162306a36Sopenharmony_ci
43262306a36Sopenharmony_ci	return 0;
43362306a36Sopenharmony_ci
43462306a36Sopenharmony_cifail:
43562306a36Sopenharmony_ci	v3d_sched_fini(v3d);
43662306a36Sopenharmony_ci	return ret;
43762306a36Sopenharmony_ci}
43862306a36Sopenharmony_ci
43962306a36Sopenharmony_civoid
44062306a36Sopenharmony_civ3d_sched_fini(struct v3d_dev *v3d)
44162306a36Sopenharmony_ci{
44262306a36Sopenharmony_ci	enum v3d_queue q;
44362306a36Sopenharmony_ci
44462306a36Sopenharmony_ci	for (q = 0; q < V3D_MAX_QUEUES; q++) {
44562306a36Sopenharmony_ci		if (v3d->queue[q].sched.ready)
44662306a36Sopenharmony_ci			drm_sched_fini(&v3d->queue[q].sched);
44762306a36Sopenharmony_ci	}
44862306a36Sopenharmony_ci}
449