Home
last modified time | relevance | path

Searched refs:queued (Results 1 - 10 of 10) sorted by relevance

/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_kinstr_jm.c867 struct kbase_jd_atom *const queued = kbase_gpu_inspect(kbdev, slot, 1); in kbasep_kinstr_jm_atom_hw_release() local
879 if (WARN_ON((submitted != katom) && (queued != katom))) { in kbasep_kinstr_jm_atom_hw_release()
883 if (queued == katom) { in kbasep_kinstr_jm_atom_hw_release()
890 if (queued && queued->gpu_rb_state == KBASE_ATOM_GPU_RB_SUBMITTED) { in kbasep_kinstr_jm_atom_hw_release()
891 kbase_kinstr_jm_atom_state_start(queued); in kbasep_kinstr_jm_atom_hw_release()
H A Dmali_kbase_jd.c157 * been queued are done before continuing. in kbase_jd_free_external_resources()
767 wake_up(&kctx->jctx.zero_jobs_wait); /* All events are safely queued now, and we can signal any waiter in jd_done_nolock()
851 int queued = 0; in jd_submit_atom() local
970 dev_dbg(kbdev->dev, "Atom %p status to queued\n", (void *)katom); in jd_submit_atom()
982 queued = 1; in jd_submit_atom()
987 if (!queued) { in jd_submit_atom()
1009 dev_dbg(kbdev->dev, "Atom %p status to queued\n", (void *)katom); in jd_submit_atom()
1112 if (queued && !IS_GPU_ATOM(katom)) { in jd_submit_atom()
1459 * has queued */ in kbase_jd_done_worker()
1639 * queued outsid in kbase_jd_zap_context()
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_kinstr_jm.c874 struct kbase_jd_atom *const queued = kbase_gpu_inspect(kbdev, slot, 1); in kbasep_kinstr_jm_atom_hw_release() local
884 if (WARN_ON((submitted != katom) && (queued != katom))) in kbasep_kinstr_jm_atom_hw_release()
887 if (queued == katom) in kbasep_kinstr_jm_atom_hw_release()
892 if (queued && queued->gpu_rb_state == KBASE_ATOM_GPU_RB_SUBMITTED) in kbasep_kinstr_jm_atom_hw_release()
893 kbase_kinstr_jm_atom_state_start(queued); in kbasep_kinstr_jm_atom_hw_release()
H A Dmali_kbase_jd.c165 * been queued are done before continuing. in kbase_jd_free_external_resources()
824 /* All events are safely queued now, and we can signal in jd_done_nolock()
913 int queued = 0; in jd_submit_atom() local
1037 dev_dbg(kbdev->dev, "Atom %pK status to queued\n", in jd_submit_atom()
1051 queued = 1; in jd_submit_atom()
1056 if (!queued) { in jd_submit_atom()
1079 dev_dbg(kbdev->dev, "Atom %pK status to queued\n", (void *)katom); in jd_submit_atom()
1195 if (queued && !IS_GPU_ATOM(katom)) in jd_submit_atom()
1564 * has queued in kbase_jd_done_worker()
1748 * queued outsid in kbase_jd_zap_context()
[all...]
/device/soc/rockchip/common/sdk_linux/kernel/sched/
H A Dcore.c540 * its already queued (either by us or someone else) and will get the in __wake_q_add()
593 * queued for wakeup.
1434 * Lock the task and the rq where the task is (or was) queued. in uclamp_update_active()
1895 * move_queued_task - move a queued task to new rq.
2015 bool queued, running; in do_set_cpus_allowed() local
2019 queued = task_on_rq_queued(p); in do_set_cpus_allowed()
2022 if (queued) { in do_set_cpus_allowed()
2036 if (queued) { in do_set_cpus_allowed()
2358 int running, queued; in wait_task_inactive() local
2398 queued in wait_task_inactive()
5302 int prio, oldprio, queued, running, rt_mutex_setprio() local
5434 bool queued, running; set_user_nice() local
5675 int retval, oldprio, newprio, queued, running; __sched_setscheduler() local
7254 bool queued, running; sched_setnuma() local
8550 int queued, running, sched_move_task() local
[all...]
H A Dfair.c314 /* runqueue on which this entity is (to be) queued */
4748 static void entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) in entity_tick() argument
4763 * queued ticks are scheduled to match the slice, so don't bother in entity_tick()
4766 if (queued) { in entity_tick()
11544 static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) in task_tick_fair() argument
11551 entity_tick(cfs_rq, se, queued); in task_tick_fair()
H A Dsched.h515 * are queued on this cfs_rq, so a weight of a entity should not be
703 int curr; /* highest queued rt task prio */
1911 void (*task_tick)(struct rq *rq, struct task_struct *p, int queued);
H A Drt.c107 /* We start is dequeued state, because no RT tasks are queued */ in init_rt_rq()
2254 * tasks queued on it (overloaded) needs to be notified that a CPU has opened
2255 * up that may be able to run one of its non-running queued RT tasks.
2266 * if its the only CPU with multiple RT tasks queued, and a large number
2701 static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued) in task_tick_rt() argument
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_jd.c225 * been queued are done before continuing. in kbase_jd_free_external_resources()
864 wake_up(&kctx->jctx.zero_jobs_wait); /* All events are safely queued now, and we can signal any waiter in jd_done_nolock()
932 int queued = 0; in jd_submit_atom() local
1055 queued = 1; in jd_submit_atom()
1060 if (!queued) { in jd_submit_atom()
1165 if (queued && !IS_GPU_ATOM(katom)) { in jd_submit_atom()
1526 * has queued */ in kbase_jd_done_worker()
1705 * queued outside the job scheduler. in kbase_jd_zap_context()
1742 * been queued are done before continuing. in kbase_jd_zap_context()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_jd.c230 * been queued are done before continuing. in kbase_jd_free_external_resources()
902 wake_up(&kctx->jctx.zero_jobs_wait); /* All events are safely queued now, and we can signal any waiter in jd_done_nolock()
969 int queued = 0; in jd_submit_atom() local
1103 queued = 1; in jd_submit_atom()
1108 if (!queued) { in jd_submit_atom()
1227 if (queued && !IS_GPU_ATOM(katom)) { in jd_submit_atom()
1600 * has queued */ in kbase_jd_done_worker()
1781 * queued outside the job scheduler. in kbase_jd_zap_context()
1817 * been queued are done before continuing. in kbase_jd_zap_context()

Completed in 38 milliseconds