Home
last modified time | relevance | path

Searched refs:kctx (Results 1 - 25 of 315) sorted by relevance

12345678910>>...13

/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_context.c37 struct kbase_context *kctx; in kbase_create_context() local
43 kctx = vzalloc(sizeof(*kctx)); in kbase_create_context()
44 if (!kctx) { in kbase_create_context()
51 kctx->kbdev = kbdev; in kbase_create_context()
52 kctx->as_nr = KBASEP_AS_NR_INVALID; in kbase_create_context()
53 atomic_set(&kctx->refcount, 0); in kbase_create_context()
55 kbase_ctx_flag_set(kctx, KCTX_COMPAT); in kbase_create_context()
58 kctx->timeline.owner_tgid = task_tgid_nr(current); in kbase_create_context()
60 atomic_set(&kctx in kbase_create_context()
214 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument
315 kbase_context_set_create_flags(struct kbase_context *kctx, u32 flags) kbase_context_set_create_flags() argument
[all...]
H A Dmali_kbase_js.c68 struct kbase_device *kbdev, struct kbase_context *kctx,
73 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback);
77 static int kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) in kbasep_js_trace_get_refcnt() argument
79 return atomic_read(&kctx->refcount); in kbasep_js_trace_get_refcnt()
82 static int kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) in kbasep_js_trace_get_refcnt() argument
85 CSTD_UNUSED(kctx); in kbasep_js_trace_get_refcnt()
137 bool kbasep_js_runpool_retain_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) in kbasep_js_runpool_retain_ctx_nolock() argument
144 KBASE_DEBUG_ASSERT(kctx != NULL); in kbasep_js_runpool_retain_ctx_nolock()
149 as_nr = kctx->as_nr; in kbasep_js_runpool_retain_ctx_nolock()
150 if (atomic_read(&kctx in kbasep_js_runpool_retain_ctx_nolock()
174 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_none_to_pull_prio() argument
194 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument
226 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument
259 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument
279 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument
307 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument
334 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument
350 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument
385 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument
526 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument
568 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument
623 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument
662 kbase_js_ctx_list_add_pullable_head_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument
703 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument
732 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument
770 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument
808 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local
835 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local
857 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument
894 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument
1008 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument
1114 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument
1133 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument
1163 kbasep_js_runpool_retain_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx() argument
1229 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument
1284 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument
1431 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument
1441 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument
1462 kbasep_js_runpool_release_ctx_and_katom_retained_state(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument
1497 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument
1508 kbasep_js_runpool_release_ctx_no_schedule(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument
1548 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_ctx() argument
1681 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_use_ctx() argument
1699 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_privileged_ctx() argument
1748 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_release_privileged_ctx() argument
1787 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local
1810 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local
1832 struct kbase_context *kctx, *n; kbasep_js_resume() local
1899 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_resolved_submit() argument
1978 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) kbase_js_evict_deps() argument
2005 kbase_js_pull(struct kbase_context *kctx, int js) kbase_js_pull() argument
2082 struct kbase_context *kctx = katom->kctx; js_return_worker() local
2185 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_unpull() argument
2203 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_complete_atom_wq() argument
2300 struct kbase_context *kctx = katom->kctx; kbase_js_complete_atom() local
2378 struct kbase_context *kctx; kbase_js_sched() local
2500 kbase_js_zap_context(struct kbase_context *kctx) kbase_js_zap_context() argument
2644 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) trace_get_refcnt() argument
2667 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) kbase_js_foreach_ctx_job() argument
[all...]
H A Dmali_kbase_ctx_sched.c67 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument
69 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx()
75 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && (kbdev->as_free & (1u << kctx->as_nr))) { in kbasep_ctx_sched_find_as_for_ctx()
76 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx()
90 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument
92 struct kbase_device *const kbdev = kctx->kbdev; in kbase_ctx_sched_retain_ctx()
99 if (atomic_inc_return(&kctx->refcount) == 1) { in kbase_ctx_sched_retain_ctx()
100 int const free_as = kbasep_ctx_sched_find_as_for_ctx(kctx); in kbase_ctx_sched_retain_ctx()
106 if (free_as != kctx in kbase_ctx_sched_retain_ctx()
132 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument
150 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument
161 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument
190 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local
[all...]
H A Dmali_kbase_mem_profile_debugfs.c32 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local
34 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
36 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show()
40 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
60 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) in kbasep_mem_profile_debugfs_insert() argument
64 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert()
66 dev_dbg(kctx->kbdev->dev, "initialised: %d", kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert()
68 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert()
92 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument
107 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_context.c42 struct kbase_context *kctx; in kbase_create_context() local
48 kctx = vzalloc(sizeof(*kctx)); in kbase_create_context()
50 if (!kctx) in kbase_create_context()
56 kctx->kbdev = kbdev; in kbase_create_context()
57 kctx->as_nr = KBASEP_AS_NR_INVALID; in kbase_create_context()
58 atomic_set(&kctx->refcount, 0); in kbase_create_context()
60 kbase_ctx_flag_set(kctx, KCTX_COMPAT); in kbase_create_context()
62 kctx->timeline.owner_tgid = task_tgid_nr(current); in kbase_create_context()
64 atomic_set(&kctx in kbase_create_context()
211 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument
311 kbase_context_set_create_flags(struct kbase_context *kctx, u32 flags) kbase_context_set_create_flags() argument
[all...]
H A Dmali_kbase_js.c71 struct kbase_device *kbdev, struct kbase_context *kctx,
77 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx,
83 struct kbase_context *kctx) in kbasep_js_trace_get_refcnt()
85 return atomic_read(&kctx->refcount); in kbasep_js_trace_get_refcnt()
89 struct kbase_context *kctx) in kbasep_js_trace_get_refcnt()
92 CSTD_UNUSED(kctx); in kbasep_js_trace_get_refcnt()
140 struct kbase_context *kctx) in kbasep_js_runpool_retain_ctx_nolock()
147 KBASE_DEBUG_ASSERT(kctx != NULL); in kbasep_js_runpool_retain_ctx_nolock()
152 as_nr = kctx->as_nr; in kbasep_js_runpool_retain_ctx_nolock()
153 if (atomic_read(&kctx in kbasep_js_runpool_retain_ctx_nolock()
82 kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_trace_get_refcnt() argument
88 kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_trace_get_refcnt() argument
139 kbasep_js_runpool_retain_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx_nolock() argument
179 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_none_to_pull_prio() argument
200 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument
232 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument
269 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument
290 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument
318 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument
345 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument
362 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument
398 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument
557 kbasep_js_kctx_init(struct kbase_context * const kctx) kbasep_js_kctx_init() argument
599 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument
653 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument
694 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument
736 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument
767 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument
808 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument
850 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local
878 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local
900 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument
934 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument
1064 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument
1172 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument
1193 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument
1225 kbasep_js_runpool_retain_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx() argument
1293 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument
1350 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument
1506 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument
1518 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument
1541 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument
1576 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument
1589 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument
1631 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_ctx() argument
1767 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_use_ctx() argument
1787 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_privileged_ctx() argument
1837 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_release_privileged_ctx() argument
1877 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local
1902 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local
1923 struct kbase_context *kctx, *n; kbasep_js_resume() local
1991 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_resolved_submit() argument
2075 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) kbase_js_evict_deps() argument
2105 kbase_js_pull(struct kbase_context *kctx, int js) kbase_js_pull() argument
2179 struct kbase_context *kctx = katom->kctx; js_return_worker() local
2285 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_unpull() argument
2303 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_complete_atom_wq() argument
2409 struct kbase_context *kctx = katom->kctx; kbase_js_complete_atom() local
2492 struct kbase_context *kctx; kbase_js_sched() local
2642 kbase_js_zap_context(struct kbase_context *kctx) kbase_js_zap_context() argument
2789 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) trace_get_refcnt() argument
2813 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) kbase_js_foreach_ctx_job() argument
[all...]
H A Dmali_kbase_ctx_sched.c68 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument
70 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx()
76 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && in kbasep_ctx_sched_find_as_for_ctx()
77 (kbdev->as_free & (1u << kctx->as_nr))) in kbasep_ctx_sched_find_as_for_ctx()
78 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx()
90 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument
92 struct kbase_device *const kbdev = kctx->kbdev; in kbase_ctx_sched_retain_ctx()
99 if (atomic_inc_return(&kctx->refcount) == 1) { in kbase_ctx_sched_retain_ctx()
100 int const free_as = kbasep_ctx_sched_find_as_for_ctx(kctx); in kbase_ctx_sched_retain_ctx()
107 if (free_as != kctx in kbase_ctx_sched_retain_ctx()
134 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument
151 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument
161 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument
189 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local
[all...]
H A Dmali_kbase_mem_profile_debugfs.c34 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local
36 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
38 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show()
42 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
62 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, in kbasep_mem_profile_debugfs_insert() argument
67 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert()
69 dev_dbg(kctx->kbdev->dev, "initialised: %d", in kbasep_mem_profile_debugfs_insert()
70 kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert()
72 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert()
99 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument
115 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/
H A Dmali_kbase_context.c52 /* Check if the kctx creation request is from a existing process. */ in find_process_node()
73 * @kctx: Pointer to kbase context.
83 static int kbase_insert_kctx_to_process(struct kbase_context *kctx) in kbase_insert_kctx_to_process() argument
85 struct rb_root *const prcs_root = &kctx->kbdev->process_root; in kbase_insert_kctx_to_process()
86 const pid_t tgid = kctx->tgid; in kbase_insert_kctx_to_process()
89 lockdep_assert_held(&kctx->kbdev->kctx_list_lock); in kbase_insert_kctx_to_process()
92 /* if the kctx is from new process then create a new kbase_process in kbase_insert_kctx_to_process()
122 kctx->kprcs = kprcs; in kbase_insert_kctx_to_process()
123 list_add(&kctx->kprcs_link, &kprcs->kctx_list); in kbase_insert_kctx_to_process()
128 int kbase_context_common_init(struct kbase_context *kctx) in kbase_context_common_init() argument
198 kbase_remove_kctx_from_process(struct kbase_context *kctx) kbase_remove_kctx_from_process() argument
219 kbase_context_common_term(struct kbase_context *kctx) kbase_context_common_term() argument
260 kbase_context_mem_pool_group_init(struct kbase_context *kctx) kbase_context_mem_pool_group_init() argument
266 kbase_context_mem_pool_group_term(struct kbase_context *kctx) kbase_context_mem_pool_group_term() argument
271 kbase_context_mmu_init(struct kbase_context *kctx) kbase_context_mmu_init() argument
278 kbase_context_mmu_term(struct kbase_context *kctx) kbase_context_mmu_term() argument
283 kbase_context_mem_alloc_page(struct kbase_context *kctx) kbase_context_mem_alloc_page() argument
297 kbase_context_mem_pool_free(struct kbase_context *kctx) kbase_context_mem_pool_free() argument
303 kbase_context_sticky_resource_term(struct kbase_context *kctx) kbase_context_sticky_resource_term() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/
H A Dmali_kbase_context.c48 /* Check if the kctx creation request is from a existing process.*/ in find_process_node()
69 * @kctx: Pointer to kbase context.
79 static int kbase_insert_kctx_to_process(struct kbase_context *kctx) in kbase_insert_kctx_to_process() argument
81 struct rb_root *const prcs_root = &kctx->kbdev->process_root; in kbase_insert_kctx_to_process()
82 const pid_t tgid = kctx->tgid; in kbase_insert_kctx_to_process()
85 lockdep_assert_held(&kctx->kbdev->kctx_list_lock); in kbase_insert_kctx_to_process()
89 /* if the kctx is from new process then create a new kbase_process in kbase_insert_kctx_to_process()
118 kctx->kprcs = kprcs; in kbase_insert_kctx_to_process()
119 list_add(&kctx->kprcs_link, &kprcs->kctx_list); in kbase_insert_kctx_to_process()
124 int kbase_context_common_init(struct kbase_context *kctx) in kbase_context_common_init() argument
178 kbase_context_add_to_dev_list(struct kbase_context *kctx) kbase_context_add_to_dev_list() argument
195 kbase_context_remove_from_dev_list(struct kbase_context *kctx) kbase_context_remove_from_dev_list() argument
221 kbase_remove_kctx_from_process(struct kbase_context *kctx) kbase_remove_kctx_from_process() argument
242 kbase_context_common_term(struct kbase_context *kctx) kbase_context_common_term() argument
267 kbase_context_mem_pool_group_init(struct kbase_context *kctx) kbase_context_mem_pool_group_init() argument
275 kbase_context_mem_pool_group_term(struct kbase_context *kctx) kbase_context_mem_pool_group_term() argument
280 kbase_context_mmu_init(struct kbase_context *kctx) kbase_context_mmu_init() argument
287 kbase_context_mmu_term(struct kbase_context *kctx) kbase_context_mmu_term() argument
292 kbase_context_mem_alloc_page(struct kbase_context *kctx) kbase_context_mem_alloc_page() argument
305 kbase_context_mem_pool_free(struct kbase_context *kctx) kbase_context_mem_pool_free() argument
314 kbase_context_sticky_resource_term(struct kbase_context *kctx) kbase_context_sticky_resource_term() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_ctx_sched.c32 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
34 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt()
37 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
39 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt()
80 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument
82 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx()
88 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && in kbasep_ctx_sched_find_as_for_ctx()
89 (kbdev->as_free & (1u << kctx->as_nr))) in kbasep_ctx_sched_find_as_for_ctx()
90 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx()
102 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument
150 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument
162 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument
187 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument
216 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local
313 kbase_ctx_sched_inc_refcount_nolock(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_nolock() argument
336 kbase_ctx_sched_inc_refcount(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount() argument
356 kbase_ctx_sched_release_ctx_lock(struct kbase_context *kctx) kbase_ctx_sched_release_ctx_lock() argument
373 kbase_ctx_sched_inc_refcount_if_as_valid(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_if_as_valid() argument
[all...]
H A Dmali_kbase_js.c77 struct kbase_device *kbdev, struct kbase_context *kctx,
83 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx,
88 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
90 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt()
93 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
95 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt()
143 * @kctx: Pointer to kbase context with ring buffer.
155 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument
158 struct jsctx_queue *rb = &kctx->jsctx_queue[prio][js]; in jsctx_rb_none_to_pull_prio()
160 lockdep_assert_held(&kctx in jsctx_rb_none_to_pull_prio()
183 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument
215 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach_prio() argument
275 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach() argument
297 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument
330 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument
358 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument
376 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument
418 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument
639 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument
684 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument
733 kbase_jsctx_slot_prio_blocked_clear(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_blocked_clear() argument
746 kbase_jsctx_slot_atoms_pulled(struct kbase_context *kctx, int js) kbase_jsctx_slot_atoms_pulled() argument
756 kbase_jsctx_slot_prio_is_blocked(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_is_blocked() argument
793 kbase_jsctx_slot_atom_pulled_inc(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_inc() argument
831 kbase_jsctx_slot_atom_pulled_dec(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_dec() argument
882 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument
925 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument
969 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument
1000 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument
1043 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument
1085 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local
1119 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local
1141 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument
1203 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument
1369 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument
1391 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument
1423 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local
1470 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local
1513 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument
1657 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument
1685 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument
1734 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument
1790 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument
1953 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument
1965 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument
1986 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument
2021 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument
2035 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument
2077 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
2221 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
2249 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
2322 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
2362 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2386 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2410 struct kbase_context *kctx, *n; global() local
2518 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
2585 struct kbase_context *const kctx = katom->kctx; global() local
2636 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument
2669 kbase_js_pull(struct kbase_context *kctx, int js) global() argument
2784 struct kbase_context *const kctx = start_katom->kctx; global() local
2893 struct kbase_context *const kctx = end_katom->kctx; global() local
2975 struct kbase_context *kctx = katom->kctx; global() local
3111 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
3146 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument
3242 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument
3282 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
3423 struct kbase_context *const kctx = end_katom->kctx; global() local
3467 struct kbase_context *kctx = katom->kctx; global() local
3551 struct kbase_context *const kctx = katom->kctx; global() local
3629 struct kbase_context *kctx; global() local
3816 kbase_js_zap_context(struct kbase_context *kctx) global() argument
3971 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
3995 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb *callback) global() argument
[all...]
H A Dmali_kbase_gwt.c26 struct kbase_context *kctx, in kbase_gpu_gwt_setup_page_permission()
39 err = kbase_mmu_update_pages(kctx, reg->start_pfn, in kbase_gpu_gwt_setup_page_permission()
45 dev_warn(kctx->kbdev->dev, "kbase_mmu_update_pages failure\n"); in kbase_gpu_gwt_setup_page_permission()
52 static void kbase_gpu_gwt_setup_pages(struct kbase_context *kctx, in kbase_gpu_gwt_setup_pages() argument
55 kbase_gpu_gwt_setup_page_permission(kctx, flag, in kbase_gpu_gwt_setup_pages()
56 rb_first(&(kctx->reg_rbtree_same))); in kbase_gpu_gwt_setup_pages()
57 kbase_gpu_gwt_setup_page_permission(kctx, flag, in kbase_gpu_gwt_setup_pages()
58 rb_first(&(kctx->reg_rbtree_custom))); in kbase_gpu_gwt_setup_pages()
62 int kbase_gpu_gwt_start(struct kbase_context *kctx) in kbase_gpu_gwt_start() argument
64 kbase_gpu_vm_lock(kctx); in kbase_gpu_gwt_start()
25 kbase_gpu_gwt_setup_page_permission( struct kbase_context *kctx, unsigned long flag, struct rb_node *node) kbase_gpu_gwt_setup_page_permission() argument
97 kbase_gpu_gwt_stop(struct kbase_context *kctx) kbase_gpu_gwt_stop() argument
144 kbase_gpu_gwt_collate(struct kbase_context *kctx, struct list_head *snapshot_list) kbase_gpu_gwt_collate() argument
174 kbase_gpu_gwt_dump(struct kbase_context *kctx, union kbase_ioctl_cinstr_gwt_dump *gwt_dump) kbase_gpu_gwt_dump() argument
[all...]
H A Dmali_kbase_mem.c71 static void free_partial_locked(struct kbase_context *kctx,
74 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument
93 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) in kbase_get_num_cpu_va_bits()
103 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument
108 kbase_ctx_reg_zone_get(kctx, KBASE_REG_ZONE_EXEC_VA); in kbase_gpu_va_to_rbtree()
114 rbtree = &kctx->reg_rbtree_exec; in kbase_gpu_va_to_rbtree()
119 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) { in kbase_gpu_va_to_rbtree()
125 kbase_ctx_reg_zone_get(kctx, in kbase_gpu_va_to_rbtree()
132 rbtree = &kctx->reg_rbtree_custom; in kbase_gpu_va_to_rbtree()
134 rbtree = &kctx in kbase_gpu_va_to_rbtree()
232 kbase_region_tracker_find_region_enclosing_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
272 kbase_region_tracker_find_region_base_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
555 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
702 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
748 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
765 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument
771 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
855 kbase_has_exec_va_zone_locked(struct kbase_context *kctx) kbase_has_exec_va_zone_locked() argument
865 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument
892 kbase_region_tracker_has_allocs(struct kbase_context *kctx) kbase_region_tracker_has_allocs() argument
947 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument
1016 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
1082 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument
1385 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local
1430 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local
1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument
1622 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
1698 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1742 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1762 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument
1792 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1845 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1944 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1977 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
2036 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
2100 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
2208 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local
2402 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local
2594 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
2624 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local
2718 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument
2746 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local
3090 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extension) kbase_check_alloc_sizes() argument
3210 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
3222 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
3234 struct kbase_context *kctx; global() member
3317 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
3341 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
3365 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
3390 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local
3417 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local
3452 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
3518 struct kbase_context *kctx; kbase_jit_destroy_worker() local
3542 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
3595 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument
3725 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument
3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument
3882 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument
3917 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument
3931 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument
3984 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument
4070 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument
4292 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
4353 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local
4377 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
4401 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
4446 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument
4491 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument
4554 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument
4628 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
4711 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
4781 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument
4822 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
4858 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
4917 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument
4934 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument
4949 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
4970 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument
4988 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
4995 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
H A Dmali_kbase_mem_profile_debugfs.c40 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local
42 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
44 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show()
48 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
69 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, in kbasep_mem_profile_debugfs_insert() argument
79 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert()
81 dev_dbg(kctx->kbdev->dev, "initialised: %d", in kbasep_mem_profile_debugfs_insert()
82 kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert()
84 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert()
113 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument
129 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_js.c71 (struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state);
75 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback);
79 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
81 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt()
84 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
86 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt()
139 * @kctx: Pointer to kbase context with ring buffer.
150 static inline bool jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument
153 struct jsctx_queue *rb = &kctx->jsctx_queue[prio][js]; in jsctx_rb_none_to_pull_prio()
155 lockdep_assert_held(&kctx in jsctx_rb_none_to_pull_prio()
176 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument
208 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument
257 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument
277 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument
307 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument
334 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument
352 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument
392 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument
526 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument
569 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument
627 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument
667 kbase_js_ctx_list_add_pullable_head_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument
709 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument
738 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument
778 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument
816 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local
847 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local
869 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument
918 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument
1046 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument
1069 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument
1099 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local
1148 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local
1195 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument
1331 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument
1357 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument
1403 kbasep_js_run_jobs_after_ctx_and_atom_release(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument
1455 kbasep_js_runpool_release_ctx_internal(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument
1604 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument
1614 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument
1632 kbasep_js_runpool_release_ctx_and_katom_retained_state(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument
1667 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument
1678 kbasep_js_runpool_release_ctx_no_schedule(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument
1718 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
1853 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
1874 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
1947 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
1985 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2006 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2029 struct kbase_context *kctx, *n; global() local
2131 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
2192 struct kbase_context *const kctx = katom->kctx; global() local
2234 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument
2264 kbase_js_pull(struct kbase_context *kctx, int js) global() argument
2370 struct kbase_context *const kctx = start_katom->kctx; global() local
2471 struct kbase_context *const kctx = end_katom->kctx; global() local
2547 struct kbase_context *kctx = katom->kctx; global() local
2674 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
2708 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument
2792 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument
2832 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
2960 struct kbase_context *const kctx = end_katom->kctx; global() local
3005 struct kbase_context *kctx = katom->kctx; global() local
3077 struct kbase_context *const kctx = katom->kctx; global() local
3148 struct kbase_context *kctx; global() local
3282 kbase_js_zap_context(struct kbase_context *kctx) global() argument
3428 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
3451 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) global() argument
[all...]
H A Dmali_kbase_ctx_sched.c32 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
34 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt()
37 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
39 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt()
79 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument
81 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx()
87 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && (kbdev->as_free & (1u << kctx->as_nr))) { in kbasep_ctx_sched_find_as_for_ctx()
88 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx()
102 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument
146 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument
158 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument
179 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument
209 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local
294 kbase_ctx_sched_inc_refcount_nolock(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_nolock() argument
317 kbase_ctx_sched_inc_refcount(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount() argument
339 kbase_ctx_sched_release_ctx_lock(struct kbase_context *kctx) kbase_ctx_sched_release_ctx_lock() argument
[all...]
H A Dmali_kbase_gwt.c26 static inline void kbase_gpu_gwt_setup_page_permission(struct kbase_context *kctx, unsigned long flag, in kbase_gpu_gwt_setup_page_permission() argument
37 err = kbase_mmu_update_pages(kctx, reg->start_pfn, kbase_get_gpu_phy_pages(reg), reg->gpu_alloc->nents, in kbase_gpu_gwt_setup_page_permission()
40 dev_warn(kctx->kbdev->dev, "kbase_mmu_update_pages failure\n"); in kbase_gpu_gwt_setup_page_permission()
48 static void kbase_gpu_gwt_setup_pages(struct kbase_context *kctx, unsigned long flag) in kbase_gpu_gwt_setup_pages() argument
50 kbase_gpu_gwt_setup_page_permission(kctx, flag, rb_first(&(kctx->reg_rbtree_same))); in kbase_gpu_gwt_setup_pages()
51 kbase_gpu_gwt_setup_page_permission(kctx, flag, rb_first(&(kctx->reg_rbtree_custom))); in kbase_gpu_gwt_setup_pages()
54 int kbase_gpu_gwt_start(struct kbase_context *kctx) in kbase_gpu_gwt_start() argument
56 kbase_gpu_vm_lock(kctx); in kbase_gpu_gwt_start()
88 kbase_gpu_gwt_stop(struct kbase_context *kctx) kbase_gpu_gwt_stop() argument
134 kbase_gpu_gwt_collate(struct kbase_context *kctx, struct list_head *snapshot_list) kbase_gpu_gwt_collate() argument
161 kbase_gpu_gwt_dump(struct kbase_context *kctx, union kbase_ioctl_cinstr_gwt_dump *gwt_dump) kbase_gpu_gwt_dump() argument
[all...]
H A Dmali_kbase_mem_profile_debugfs.c39 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local
41 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
43 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show()
47 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show()
68 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) in kbasep_mem_profile_debugfs_insert() argument
77 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert()
79 dev_dbg(kctx->kbdev->dev, "initialised: %d", kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert()
81 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert()
108 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument
123 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument
[all...]
H A Dmali_kbase_mem.c71 static void free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp);
73 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument
92 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) { in kbase_get_num_cpu_va_bits()
102 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn) in kbase_gpu_va_to_rbtree() argument
109 if (gpu_pfn >= kctx->exec_va_start) { in kbase_gpu_va_to_rbtree()
110 rbtree = &kctx->reg_rbtree_exec; in kbase_gpu_va_to_rbtree()
115 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) in kbase_gpu_va_to_rbtree()
120 same_va_end = kctx->same_va_end; in kbase_gpu_va_to_rbtree()
125 rbtree = &kctx->reg_rbtree_custom; in kbase_gpu_va_to_rbtree()
127 rbtree = &kctx in kbase_gpu_va_to_rbtree()
226 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
265 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
488 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
625 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
671 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
688 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument
693 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
764 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument
813 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
864 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument
1114 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local
1154 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local
1227 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument
1337 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
1393 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1440 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1459 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument
1488 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1537 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1629 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1661 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
1717 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
1779 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
1887 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local
2058 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local
2231 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
2254 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local
2330 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument
2355 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local
2691 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extent) kbase_check_alloc_sizes() argument
2792 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
2803 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
2815 struct kbase_context *kctx; global() member
2896 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
2922 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
2948 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
2975 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local
3001 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local
3034 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
3093 struct kbase_context *kctx; kbase_jit_destroy_worker() local
3116 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
3170 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument
3293 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument
3329 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas) kbase_jit_grow() argument
3447 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument
3481 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument
3496 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument
3546 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument
3619 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument
3829 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
3885 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local
3911 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
3934 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
3977 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument
4020 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument
4058 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument
4085 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument
4146 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
4220 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
4289 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument
4329 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
4362 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
4422 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument
4437 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument
4449 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
4472 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument
4491 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
4498 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_event.c28 * @kctx: Pointer to the Kbase context this event belongs to.
38 struct kbase_context *kctx; member
43 int kbase_csf_event_wait_add(struct kbase_context *kctx, in kbase_csf_event_wait_add() argument
53 event_cb->kctx = kctx; in kbase_csf_event_wait_add()
57 spin_lock_irqsave(&kctx->csf.event.lock, flags); in kbase_csf_event_wait_add()
58 list_add_tail(&event_cb->link, &kctx->csf.event.callback_list); in kbase_csf_event_wait_add()
59 dev_dbg(kctx->kbdev->dev, in kbase_csf_event_wait_add()
62 spin_unlock_irqrestore(&kctx->csf.event.lock, flags); in kbase_csf_event_wait_add()
70 void kbase_csf_event_wait_remove(struct kbase_context *kctx, in kbase_csf_event_wait_remove() argument
91 sync_update_notify_gpu(struct kbase_context *kctx) sync_update_notify_gpu() argument
111 kbase_csf_event_signal(struct kbase_context *kctx, bool notify_gpu) kbase_csf_event_signal() argument
156 kbase_csf_event_term(struct kbase_context *kctx) kbase_csf_event_term() argument
177 kbase_csf_event_init(struct kbase_context *const kctx) kbase_csf_event_init() argument
184 kbase_csf_event_remove_error(struct kbase_context *kctx, struct kbase_csf_notification *error) kbase_csf_event_remove_error() argument
194 kbase_csf_event_read_error(struct kbase_context *kctx, struct base_csf_notification *event_data) kbase_csf_event_read_error() argument
213 kbase_csf_event_add_error(struct kbase_context *const kctx, struct kbase_csf_notification *const error, struct base_csf_notification const *const data) kbase_csf_event_add_error() argument
239 kbase_csf_event_error_pending(struct kbase_context *kctx) kbase_csf_event_error_pending() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/backend/
H A Dmali_kbase_context_jm.c41 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument
43 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init()
44 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init()
45 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init()
46 kbasep_jd_debugfs_ctx_init(kctx); in kbase_context_debugfs_init()
50 void kbase_context_debugfs_term(struct kbase_context *const kctx) in kbase_context_debugfs_term() argument
52 debugfs_remove_recursive(kctx->kctx_dentry); in kbase_context_debugfs_term()
56 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument
58 CSTD_UNUSED(kctx); in kbase_context_debugfs_init()
62 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument
69 kbase_context_kbase_kinstr_jm_init(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_init() argument
74 kbase_context_kbase_kinstr_jm_term(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_term() argument
79 kbase_context_kbase_timer_setup(struct kbase_context *kctx) kbase_context_kbase_timer_setup() argument
87 kbase_context_submit_check(struct kbase_context *kctx) kbase_context_submit_check() argument
107 kbase_context_flush_jobs(struct kbase_context *kctx) kbase_context_flush_jobs() argument
113 kbase_context_free(struct kbase_context *kctx) kbase_context_free() argument
162 kbase_context_term_partial( struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument
178 struct kbase_context *kctx; kbase_create_context() local
229 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument
[all...]
H A Dmali_kbase_context_csf.c43 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument
45 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init()
46 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init()
47 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init()
48 kbase_csf_queue_group_debugfs_init(kctx); in kbase_context_debugfs_init()
49 kbase_csf_kcpu_debugfs_init(kctx); in kbase_context_debugfs_init()
50 kbase_csf_tiler_heap_debugfs_init(kctx); in kbase_context_debugfs_init()
51 kbase_csf_tiler_heap_total_debugfs_init(kctx); in kbase_context_debugfs_init()
52 kbase_csf_cpu_queue_debugfs_init(kctx); in kbase_context_debugfs_init()
56 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument
62 kbase_context_debugfs_init(struct kbase_context *const kctx) kbase_context_debugfs_init() argument
68 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument
75 kbase_context_free(struct kbase_context *kctx) kbase_context_free() argument
105 kbase_context_term_partial( struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument
121 struct kbase_context *kctx; kbase_create_context() local
172 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/backend/
H A Dmali_kbase_context_jm.c44 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument
46 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init()
47 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init()
48 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init()
49 kbasep_jd_debugfs_ctx_init(kctx); in kbase_context_debugfs_init()
50 kbase_debug_job_fault_context_init(kctx); in kbase_context_debugfs_init()
54 void kbase_context_debugfs_term(struct kbase_context *const kctx) in kbase_context_debugfs_term() argument
56 debugfs_remove_recursive(kctx->kctx_dentry); in kbase_context_debugfs_term()
57 kbase_debug_job_fault_context_term(kctx); in kbase_context_debugfs_term()
61 kbase_context_debugfs_init(struct kbase_context *const kctx) kbase_context_debugfs_init() argument
67 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument
74 kbase_context_kbase_kinstr_jm_init(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_init() argument
84 kbase_context_kbase_kinstr_jm_term(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_term() argument
89 kbase_context_kbase_timer_setup(struct kbase_context *kctx) kbase_context_kbase_timer_setup() argument
96 kbase_context_submit_check(struct kbase_context *kctx) kbase_context_submit_check() argument
136 kbase_context_term_partial(struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument
149 struct kbase_context *kctx; kbase_create_context() local
194 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/jm/
H A Dmali_kbase_jm_js.h80 * @kctx: The kbase_context to operate on
90 int kbasep_js_kctx_init(struct kbase_context *const kctx);
95 * @kctx: The kbase_context to operate on
109 void kbasep_js_kctx_term(struct kbase_context *kctx);
113 * @kctx: The kbase_context
129 static inline void kbase_jsctx_slot_prio_blocked_set(struct kbase_context *kctx, in kbase_jsctx_slot_prio_blocked_set() argument
133 &kctx->slot_tracking[js]; in kbase_jsctx_slot_prio_blocked_set()
135 lockdep_assert_held(&kctx->kbdev->hwaccess_lock); in kbase_jsctx_slot_prio_blocked_set()
137 "When marking slot %d as blocked for priority %d on a kctx, no atoms were pulled - the slot cannot become unblocked", in kbase_jsctx_slot_prio_blocked_set()
141 KBASE_KTRACE_ADD_JM_SLOT_INFO(kctx in kbase_jsctx_slot_prio_blocked_set()
152 kbase_jsctx_atoms_pulled(struct kbase_context *kctx) kbase_jsctx_atoms_pulled() argument
692 kbasep_js_is_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_is_submit_allowed() argument
721 kbasep_js_set_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_set_submit_allowed() argument
750 kbasep_js_clear_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_clear_submit_allowed() argument
859 kbase_js_runpool_inc_context_count( struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_runpool_inc_context_count() argument
896 kbase_js_runpool_dec_context_count( struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_runpool_dec_context_count() argument
[all...]

Completed in 18 milliseconds

12345678910>>...13