/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_context.c | 37 struct kbase_context *kctx;
in kbase_create_context() local 43 kctx = vzalloc(sizeof(*kctx));
in kbase_create_context() 44 if (!kctx) {
in kbase_create_context() 51 kctx->kbdev = kbdev;
in kbase_create_context() 52 kctx->as_nr = KBASEP_AS_NR_INVALID;
in kbase_create_context() 53 atomic_set(&kctx->refcount, 0);
in kbase_create_context() 55 kbase_ctx_flag_set(kctx, KCTX_COMPAT);
in kbase_create_context() 58 kctx->timeline.owner_tgid = task_tgid_nr(current);
in kbase_create_context() 60 atomic_set(&kctx in kbase_create_context() 214 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument 315 kbase_context_set_create_flags(struct kbase_context *kctx, u32 flags) kbase_context_set_create_flags() argument [all...] |
H A D | mali_kbase_js.c | 68 struct kbase_device *kbdev, struct kbase_context *kctx,
73 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback);
77 static int kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx)
in kbasep_js_trace_get_refcnt() argument 79 return atomic_read(&kctx->refcount);
in kbasep_js_trace_get_refcnt() 82 static int kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx)
in kbasep_js_trace_get_refcnt() argument 85 CSTD_UNUSED(kctx);
in kbasep_js_trace_get_refcnt() 137 bool kbasep_js_runpool_retain_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx)
in kbasep_js_runpool_retain_ctx_nolock() argument 144 KBASE_DEBUG_ASSERT(kctx != NULL);
in kbasep_js_runpool_retain_ctx_nolock() 149 as_nr = kctx->as_nr;
in kbasep_js_runpool_retain_ctx_nolock() 150 if (atomic_read(&kctx in kbasep_js_runpool_retain_ctx_nolock() 174 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_none_to_pull_prio() argument 194 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument 226 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument 259 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument 279 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument 307 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument 334 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument 350 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument 385 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument 526 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument 568 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument 623 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument 662 kbase_js_ctx_list_add_pullable_head_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument 703 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument 732 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument 770 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument 808 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local 835 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local 857 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument 894 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument 1008 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument 1114 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument 1133 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument 1163 kbasep_js_runpool_retain_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx() argument 1229 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument 1284 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument 1431 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument 1441 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument 1462 kbasep_js_runpool_release_ctx_and_katom_retained_state(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument 1497 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument 1508 kbasep_js_runpool_release_ctx_no_schedule(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument 1548 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_ctx() argument 1681 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_use_ctx() argument 1699 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_privileged_ctx() argument 1748 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_release_privileged_ctx() argument 1787 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local 1810 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local 1832 struct kbase_context *kctx, *n; kbasep_js_resume() local 1899 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_resolved_submit() argument 1978 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) kbase_js_evict_deps() argument 2005 kbase_js_pull(struct kbase_context *kctx, int js) kbase_js_pull() argument 2082 struct kbase_context *kctx = katom->kctx; js_return_worker() local 2185 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_unpull() argument 2203 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_complete_atom_wq() argument 2300 struct kbase_context *kctx = katom->kctx; kbase_js_complete_atom() local 2378 struct kbase_context *kctx; kbase_js_sched() local 2500 kbase_js_zap_context(struct kbase_context *kctx) kbase_js_zap_context() argument 2644 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) trace_get_refcnt() argument 2667 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) kbase_js_foreach_ctx_job() argument [all...] |
H A D | mali_kbase_ctx_sched.c | 67 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx)
in kbasep_ctx_sched_find_as_for_ctx() argument 69 struct kbase_device *const kbdev = kctx->kbdev;
in kbasep_ctx_sched_find_as_for_ctx() 75 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && (kbdev->as_free & (1u << kctx->as_nr))) {
in kbasep_ctx_sched_find_as_for_ctx() 76 return kctx->as_nr;
in kbasep_ctx_sched_find_as_for_ctx() 90 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx)
in kbase_ctx_sched_retain_ctx() argument 92 struct kbase_device *const kbdev = kctx->kbdev;
in kbase_ctx_sched_retain_ctx() 99 if (atomic_inc_return(&kctx->refcount) == 1) {
in kbase_ctx_sched_retain_ctx() 100 int const free_as = kbasep_ctx_sched_find_as_for_ctx(kctx);
in kbase_ctx_sched_retain_ctx() 106 if (free_as != kctx in kbase_ctx_sched_retain_ctx() 132 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument 150 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument 161 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument 190 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local [all...] |
H A D | mali_kbase_mem_profile_debugfs.c | 32 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local 34 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 36 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show() 40 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 60 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) in kbasep_mem_profile_debugfs_insert() argument 64 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert() 66 dev_dbg(kctx->kbdev->dev, "initialised: %d", kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert() 68 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert() 92 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument 107 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_context.c | 42 struct kbase_context *kctx; in kbase_create_context() local 48 kctx = vzalloc(sizeof(*kctx)); in kbase_create_context() 50 if (!kctx) in kbase_create_context() 56 kctx->kbdev = kbdev; in kbase_create_context() 57 kctx->as_nr = KBASEP_AS_NR_INVALID; in kbase_create_context() 58 atomic_set(&kctx->refcount, 0); in kbase_create_context() 60 kbase_ctx_flag_set(kctx, KCTX_COMPAT); in kbase_create_context() 62 kctx->timeline.owner_tgid = task_tgid_nr(current); in kbase_create_context() 64 atomic_set(&kctx in kbase_create_context() 211 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument 311 kbase_context_set_create_flags(struct kbase_context *kctx, u32 flags) kbase_context_set_create_flags() argument [all...] |
H A D | mali_kbase_js.c | 71 struct kbase_device *kbdev, struct kbase_context *kctx, 77 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, 83 struct kbase_context *kctx) in kbasep_js_trace_get_refcnt() 85 return atomic_read(&kctx->refcount); in kbasep_js_trace_get_refcnt() 89 struct kbase_context *kctx) in kbasep_js_trace_get_refcnt() 92 CSTD_UNUSED(kctx); in kbasep_js_trace_get_refcnt() 140 struct kbase_context *kctx) in kbasep_js_runpool_retain_ctx_nolock() 147 KBASE_DEBUG_ASSERT(kctx != NULL); in kbasep_js_runpool_retain_ctx_nolock() 152 as_nr = kctx->as_nr; in kbasep_js_runpool_retain_ctx_nolock() 153 if (atomic_read(&kctx in kbasep_js_runpool_retain_ctx_nolock() 82 kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_trace_get_refcnt() argument 88 kbasep_js_trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_trace_get_refcnt() argument 139 kbasep_js_runpool_retain_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx_nolock() argument 179 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_none_to_pull_prio() argument 200 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument 232 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument 269 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument 290 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument 318 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument 345 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument 362 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument 398 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument 557 kbasep_js_kctx_init(struct kbase_context * const kctx) kbasep_js_kctx_init() argument 599 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument 653 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument 694 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument 736 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument 767 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument 808 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument 850 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local 878 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local 900 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument 934 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument 1064 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument 1172 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument 1193 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument 1225 kbasep_js_runpool_retain_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_retain_ctx() argument 1293 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument 1350 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument 1506 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument 1518 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument 1541 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument 1576 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument 1589 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument 1631 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_ctx() argument 1767 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_use_ctx() argument 1787 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_schedule_privileged_ctx() argument 1837 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_release_privileged_ctx() argument 1877 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local 1902 struct kbase_context *kctx = kbdev->as_to_kctx[i]; kbasep_js_suspend() local 1923 struct kbase_context *kctx, *n; kbasep_js_resume() local 1991 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_resolved_submit() argument 2075 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) kbase_js_evict_deps() argument 2105 kbase_js_pull(struct kbase_context *kctx, int js) kbase_js_pull() argument 2179 struct kbase_context *kctx = katom->kctx; js_return_worker() local 2285 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_unpull() argument 2303 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_complete_atom_wq() argument 2409 struct kbase_context *kctx = katom->kctx; kbase_js_complete_atom() local 2492 struct kbase_context *kctx; kbase_js_sched() local 2642 kbase_js_zap_context(struct kbase_context *kctx) kbase_js_zap_context() argument 2789 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) trace_get_refcnt() argument 2813 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) kbase_js_foreach_ctx_job() argument [all...] |
H A D | mali_kbase_ctx_sched.c | 68 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument 70 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx() 76 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && in kbasep_ctx_sched_find_as_for_ctx() 77 (kbdev->as_free & (1u << kctx->as_nr))) in kbasep_ctx_sched_find_as_for_ctx() 78 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx() 90 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument 92 struct kbase_device *const kbdev = kctx->kbdev; in kbase_ctx_sched_retain_ctx() 99 if (atomic_inc_return(&kctx->refcount) == 1) { in kbase_ctx_sched_retain_ctx() 100 int const free_as = kbasep_ctx_sched_find_as_for_ctx(kctx); in kbase_ctx_sched_retain_ctx() 107 if (free_as != kctx in kbase_ctx_sched_retain_ctx() 134 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument 151 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument 161 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument 189 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local [all...] |
H A D | mali_kbase_mem_profile_debugfs.c | 34 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local 36 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 38 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show() 42 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 62 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, in kbasep_mem_profile_debugfs_insert() argument 67 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert() 69 dev_dbg(kctx->kbdev->dev, "initialised: %d", in kbasep_mem_profile_debugfs_insert() 70 kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert() 72 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert() 99 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument 115 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/ |
H A D | mali_kbase_context.c | 52 /* Check if the kctx creation request is from a existing process. */ in find_process_node() 73 * @kctx: Pointer to kbase context. 83 static int kbase_insert_kctx_to_process(struct kbase_context *kctx) in kbase_insert_kctx_to_process() argument 85 struct rb_root *const prcs_root = &kctx->kbdev->process_root; in kbase_insert_kctx_to_process() 86 const pid_t tgid = kctx->tgid; in kbase_insert_kctx_to_process() 89 lockdep_assert_held(&kctx->kbdev->kctx_list_lock); in kbase_insert_kctx_to_process() 92 /* if the kctx is from new process then create a new kbase_process in kbase_insert_kctx_to_process() 122 kctx->kprcs = kprcs; in kbase_insert_kctx_to_process() 123 list_add(&kctx->kprcs_link, &kprcs->kctx_list); in kbase_insert_kctx_to_process() 128 int kbase_context_common_init(struct kbase_context *kctx) in kbase_context_common_init() argument 198 kbase_remove_kctx_from_process(struct kbase_context *kctx) kbase_remove_kctx_from_process() argument 219 kbase_context_common_term(struct kbase_context *kctx) kbase_context_common_term() argument 260 kbase_context_mem_pool_group_init(struct kbase_context *kctx) kbase_context_mem_pool_group_init() argument 266 kbase_context_mem_pool_group_term(struct kbase_context *kctx) kbase_context_mem_pool_group_term() argument 271 kbase_context_mmu_init(struct kbase_context *kctx) kbase_context_mmu_init() argument 278 kbase_context_mmu_term(struct kbase_context *kctx) kbase_context_mmu_term() argument 283 kbase_context_mem_alloc_page(struct kbase_context *kctx) kbase_context_mem_alloc_page() argument 297 kbase_context_mem_pool_free(struct kbase_context *kctx) kbase_context_mem_pool_free() argument 303 kbase_context_sticky_resource_term(struct kbase_context *kctx) kbase_context_sticky_resource_term() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/ |
H A D | mali_kbase_context.c | 48 /* Check if the kctx creation request is from a existing process.*/ in find_process_node() 69 * @kctx: Pointer to kbase context. 79 static int kbase_insert_kctx_to_process(struct kbase_context *kctx) in kbase_insert_kctx_to_process() argument 81 struct rb_root *const prcs_root = &kctx->kbdev->process_root; in kbase_insert_kctx_to_process() 82 const pid_t tgid = kctx->tgid; in kbase_insert_kctx_to_process() 85 lockdep_assert_held(&kctx->kbdev->kctx_list_lock); in kbase_insert_kctx_to_process() 89 /* if the kctx is from new process then create a new kbase_process in kbase_insert_kctx_to_process() 118 kctx->kprcs = kprcs; in kbase_insert_kctx_to_process() 119 list_add(&kctx->kprcs_link, &kprcs->kctx_list); in kbase_insert_kctx_to_process() 124 int kbase_context_common_init(struct kbase_context *kctx) in kbase_context_common_init() argument 178 kbase_context_add_to_dev_list(struct kbase_context *kctx) kbase_context_add_to_dev_list() argument 195 kbase_context_remove_from_dev_list(struct kbase_context *kctx) kbase_context_remove_from_dev_list() argument 221 kbase_remove_kctx_from_process(struct kbase_context *kctx) kbase_remove_kctx_from_process() argument 242 kbase_context_common_term(struct kbase_context *kctx) kbase_context_common_term() argument 267 kbase_context_mem_pool_group_init(struct kbase_context *kctx) kbase_context_mem_pool_group_init() argument 275 kbase_context_mem_pool_group_term(struct kbase_context *kctx) kbase_context_mem_pool_group_term() argument 280 kbase_context_mmu_init(struct kbase_context *kctx) kbase_context_mmu_init() argument 287 kbase_context_mmu_term(struct kbase_context *kctx) kbase_context_mmu_term() argument 292 kbase_context_mem_alloc_page(struct kbase_context *kctx) kbase_context_mem_alloc_page() argument 305 kbase_context_mem_pool_free(struct kbase_context *kctx) kbase_context_mem_pool_free() argument 314 kbase_context_sticky_resource_term(struct kbase_context *kctx) kbase_context_sticky_resource_term() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_ctx_sched.c | 32 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 34 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt() 37 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 39 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt() 80 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument 82 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx() 88 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && in kbasep_ctx_sched_find_as_for_ctx() 89 (kbdev->as_free & (1u << kctx->as_nr))) in kbasep_ctx_sched_find_as_for_ctx() 90 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx() 102 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument 150 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument 162 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument 187 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument 216 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local 313 kbase_ctx_sched_inc_refcount_nolock(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_nolock() argument 336 kbase_ctx_sched_inc_refcount(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount() argument 356 kbase_ctx_sched_release_ctx_lock(struct kbase_context *kctx) kbase_ctx_sched_release_ctx_lock() argument 373 kbase_ctx_sched_inc_refcount_if_as_valid(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_if_as_valid() argument [all...] |
H A D | mali_kbase_js.c | 77 struct kbase_device *kbdev, struct kbase_context *kctx, 83 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, 88 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 90 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt() 93 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 95 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt() 143 * @kctx: Pointer to kbase context with ring buffer. 155 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument 158 struct jsctx_queue *rb = &kctx->jsctx_queue[prio][js]; in jsctx_rb_none_to_pull_prio() 160 lockdep_assert_held(&kctx in jsctx_rb_none_to_pull_prio() 183 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument 215 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach_prio() argument 275 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach() argument 297 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument 330 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument 358 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument 376 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument 418 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument 639 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument 684 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument 733 kbase_jsctx_slot_prio_blocked_clear(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_blocked_clear() argument 746 kbase_jsctx_slot_atoms_pulled(struct kbase_context *kctx, int js) kbase_jsctx_slot_atoms_pulled() argument 756 kbase_jsctx_slot_prio_is_blocked(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_is_blocked() argument 793 kbase_jsctx_slot_atom_pulled_inc(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_inc() argument 831 kbase_jsctx_slot_atom_pulled_dec(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_dec() argument 882 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument 925 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument 969 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument 1000 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument 1043 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument 1085 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local 1119 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local 1141 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument 1203 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument 1369 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument 1391 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument 1423 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local 1470 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local 1513 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument 1657 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument 1685 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument 1734 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument 1790 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument 1953 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument 1965 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument 1986 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument 2021 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument 2035 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument 2077 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 2221 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 2249 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 2322 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 2362 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2386 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2410 struct kbase_context *kctx, *n; global() local 2518 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 2585 struct kbase_context *const kctx = katom->kctx; global() local 2636 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument 2669 kbase_js_pull(struct kbase_context *kctx, int js) global() argument 2784 struct kbase_context *const kctx = start_katom->kctx; global() local 2893 struct kbase_context *const kctx = end_katom->kctx; global() local 2975 struct kbase_context *kctx = katom->kctx; global() local 3111 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 3146 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument 3242 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument 3282 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 3423 struct kbase_context *const kctx = end_katom->kctx; global() local 3467 struct kbase_context *kctx = katom->kctx; global() local 3551 struct kbase_context *const kctx = katom->kctx; global() local 3629 struct kbase_context *kctx; global() local 3816 kbase_js_zap_context(struct kbase_context *kctx) global() argument 3971 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 3995 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb *callback) global() argument [all...] |
H A D | mali_kbase_gwt.c | 26 struct kbase_context *kctx, in kbase_gpu_gwt_setup_page_permission() 39 err = kbase_mmu_update_pages(kctx, reg->start_pfn, in kbase_gpu_gwt_setup_page_permission() 45 dev_warn(kctx->kbdev->dev, "kbase_mmu_update_pages failure\n"); in kbase_gpu_gwt_setup_page_permission() 52 static void kbase_gpu_gwt_setup_pages(struct kbase_context *kctx, in kbase_gpu_gwt_setup_pages() argument 55 kbase_gpu_gwt_setup_page_permission(kctx, flag, in kbase_gpu_gwt_setup_pages() 56 rb_first(&(kctx->reg_rbtree_same))); in kbase_gpu_gwt_setup_pages() 57 kbase_gpu_gwt_setup_page_permission(kctx, flag, in kbase_gpu_gwt_setup_pages() 58 rb_first(&(kctx->reg_rbtree_custom))); in kbase_gpu_gwt_setup_pages() 62 int kbase_gpu_gwt_start(struct kbase_context *kctx) in kbase_gpu_gwt_start() argument 64 kbase_gpu_vm_lock(kctx); in kbase_gpu_gwt_start() 25 kbase_gpu_gwt_setup_page_permission( struct kbase_context *kctx, unsigned long flag, struct rb_node *node) kbase_gpu_gwt_setup_page_permission() argument 97 kbase_gpu_gwt_stop(struct kbase_context *kctx) kbase_gpu_gwt_stop() argument 144 kbase_gpu_gwt_collate(struct kbase_context *kctx, struct list_head *snapshot_list) kbase_gpu_gwt_collate() argument 174 kbase_gpu_gwt_dump(struct kbase_context *kctx, union kbase_ioctl_cinstr_gwt_dump *gwt_dump) kbase_gpu_gwt_dump() argument [all...] |
H A D | mali_kbase_mem.c | 71 static void free_partial_locked(struct kbase_context *kctx, 74 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument 93 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) in kbase_get_num_cpu_va_bits() 103 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument 108 kbase_ctx_reg_zone_get(kctx, KBASE_REG_ZONE_EXEC_VA); in kbase_gpu_va_to_rbtree() 114 rbtree = &kctx->reg_rbtree_exec; in kbase_gpu_va_to_rbtree() 119 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) { in kbase_gpu_va_to_rbtree() 125 kbase_ctx_reg_zone_get(kctx, in kbase_gpu_va_to_rbtree() 132 rbtree = &kctx->reg_rbtree_custom; in kbase_gpu_va_to_rbtree() 134 rbtree = &kctx in kbase_gpu_va_to_rbtree() 232 kbase_region_tracker_find_region_enclosing_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 272 kbase_region_tracker_find_region_base_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 555 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 702 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 748 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 765 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument 771 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 855 kbase_has_exec_va_zone_locked(struct kbase_context *kctx) kbase_has_exec_va_zone_locked() argument 865 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument 892 kbase_region_tracker_has_allocs(struct kbase_context *kctx) kbase_region_tracker_has_allocs() argument 947 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument 1016 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument 1082 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument 1385 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local 1430 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local 1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument 1622 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 1698 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1742 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1762 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument 1792 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1845 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1944 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1977 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 2036 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 2100 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 2208 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local 2402 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local 2594 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument 2624 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local 2718 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument 2746 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local 3090 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extension) kbase_check_alloc_sizes() argument 3210 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 3222 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 3234 struct kbase_context *kctx; global() member 3317 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 3341 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 3365 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 3390 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local 3417 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local 3452 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 3518 struct kbase_context *kctx; kbase_jit_destroy_worker() local 3542 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 3595 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument 3725 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument 3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument 3882 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument 3917 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument 3931 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument 3984 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument 4070 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument 4292 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 4353 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local 4377 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 4401 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 4446 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument 4491 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument 4554 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument 4628 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 4711 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 4781 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument 4822 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 4858 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 4917 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument 4934 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument 4949 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 4970 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument 4988 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 4995 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
H A D | mali_kbase_mem_profile_debugfs.c | 40 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local 42 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 44 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show() 48 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 69 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, in kbasep_mem_profile_debugfs_insert() argument 79 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert() 81 dev_dbg(kctx->kbdev->dev, "initialised: %d", in kbasep_mem_profile_debugfs_insert() 82 kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert() 84 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert() 113 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument 129 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_js.c | 71 (struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state); 75 static void kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback); 79 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 81 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt() 84 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 86 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt() 139 * @kctx: Pointer to kbase context with ring buffer. 150 static inline bool jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument 153 struct jsctx_queue *rb = &kctx->jsctx_queue[prio][js]; in jsctx_rb_none_to_pull_prio() 155 lockdep_assert_held(&kctx in jsctx_rb_none_to_pull_prio() 176 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument 208 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach_prio() argument 257 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb callback) jsctx_queue_foreach() argument 277 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument 307 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument 334 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument 352 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument 392 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument 526 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument 569 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument 627 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument 667 kbase_js_ctx_list_add_pullable_head_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument 709 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument 738 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument 778 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument 816 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local 847 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local 869 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument 918 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument 1046 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument 1069 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument 1099 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local 1148 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local 1195 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument 1331 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument 1357 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument 1403 kbasep_js_run_jobs_after_ctx_and_atom_release(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument 1455 kbasep_js_runpool_release_ctx_internal(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument 1604 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument 1614 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument 1632 kbasep_js_runpool_release_ctx_and_katom_retained_state(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument 1667 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument 1678 kbasep_js_runpool_release_ctx_no_schedule(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument 1718 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 1853 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 1874 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 1947 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 1985 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2006 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2029 struct kbase_context *kctx, *n; global() local 2131 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 2192 struct kbase_context *const kctx = katom->kctx; global() local 2234 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument 2264 kbase_js_pull(struct kbase_context *kctx, int js) global() argument 2370 struct kbase_context *const kctx = start_katom->kctx; global() local 2471 struct kbase_context *const kctx = end_katom->kctx; global() local 2547 struct kbase_context *kctx = katom->kctx; global() local 2674 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 2708 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument 2792 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument 2832 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 2960 struct kbase_context *const kctx = end_katom->kctx; global() local 3005 struct kbase_context *kctx = katom->kctx; global() local 3077 struct kbase_context *const kctx = katom->kctx; global() local 3148 struct kbase_context *kctx; global() local 3282 kbase_js_zap_context(struct kbase_context *kctx) global() argument 3428 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 3451 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb callback) global() argument [all...] |
H A D | mali_kbase_ctx_sched.c | 32 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 34 return atomic_read(&kctx->refcount); in kbase_ktrace_get_ctx_refcnt() 37 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 39 CSTD_UNUSED(kctx); in kbase_ktrace_get_ctx_refcnt() 79 static int kbasep_ctx_sched_find_as_for_ctx(struct kbase_context *kctx) in kbasep_ctx_sched_find_as_for_ctx() argument 81 struct kbase_device *const kbdev = kctx->kbdev; in kbasep_ctx_sched_find_as_for_ctx() 87 if ((kctx->as_nr != KBASEP_AS_NR_INVALID) && (kbdev->as_free & (1u << kctx->as_nr))) { in kbasep_ctx_sched_find_as_for_ctx() 88 return kctx->as_nr; in kbasep_ctx_sched_find_as_for_ctx() 102 int kbase_ctx_sched_retain_ctx(struct kbase_context *kctx) in kbase_ctx_sched_retain_ctx() argument 146 kbase_ctx_sched_retain_ctx_refcount(struct kbase_context *kctx) kbase_ctx_sched_retain_ctx_refcount() argument 158 kbase_ctx_sched_release_ctx(struct kbase_context *kctx) kbase_ctx_sched_release_ctx() argument 179 kbase_ctx_sched_remove_ctx(struct kbase_context *kctx) kbase_ctx_sched_remove_ctx() argument 209 struct kbase_context *kctx; kbase_ctx_sched_restore_all_as() local 294 kbase_ctx_sched_inc_refcount_nolock(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount_nolock() argument 317 kbase_ctx_sched_inc_refcount(struct kbase_context *kctx) kbase_ctx_sched_inc_refcount() argument 339 kbase_ctx_sched_release_ctx_lock(struct kbase_context *kctx) kbase_ctx_sched_release_ctx_lock() argument [all...] |
H A D | mali_kbase_gwt.c | 26 static inline void kbase_gpu_gwt_setup_page_permission(struct kbase_context *kctx, unsigned long flag, in kbase_gpu_gwt_setup_page_permission() argument 37 err = kbase_mmu_update_pages(kctx, reg->start_pfn, kbase_get_gpu_phy_pages(reg), reg->gpu_alloc->nents, in kbase_gpu_gwt_setup_page_permission() 40 dev_warn(kctx->kbdev->dev, "kbase_mmu_update_pages failure\n"); in kbase_gpu_gwt_setup_page_permission() 48 static void kbase_gpu_gwt_setup_pages(struct kbase_context *kctx, unsigned long flag) in kbase_gpu_gwt_setup_pages() argument 50 kbase_gpu_gwt_setup_page_permission(kctx, flag, rb_first(&(kctx->reg_rbtree_same))); in kbase_gpu_gwt_setup_pages() 51 kbase_gpu_gwt_setup_page_permission(kctx, flag, rb_first(&(kctx->reg_rbtree_custom))); in kbase_gpu_gwt_setup_pages() 54 int kbase_gpu_gwt_start(struct kbase_context *kctx) in kbase_gpu_gwt_start() argument 56 kbase_gpu_vm_lock(kctx); in kbase_gpu_gwt_start() 88 kbase_gpu_gwt_stop(struct kbase_context *kctx) kbase_gpu_gwt_stop() argument 134 kbase_gpu_gwt_collate(struct kbase_context *kctx, struct list_head *snapshot_list) kbase_gpu_gwt_collate() argument 161 kbase_gpu_gwt_dump(struct kbase_context *kctx, union kbase_ioctl_cinstr_gwt_dump *gwt_dump) kbase_gpu_gwt_dump() argument [all...] |
H A D | mali_kbase_mem_profile_debugfs.c | 39 struct kbase_context *kctx = sfile->private; in kbasep_mem_profile_seq_show() local 41 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 43 seq_write(sfile, kctx->mem_profile_data, kctx->mem_profile_size); in kbasep_mem_profile_seq_show() 47 mutex_unlock(&kctx->mem_profile_lock); in kbasep_mem_profile_seq_show() 68 int kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) in kbasep_mem_profile_debugfs_insert() argument 77 mutex_lock(&kctx->mem_profile_lock); in kbasep_mem_profile_debugfs_insert() 79 dev_dbg(kctx->kbdev->dev, "initialised: %d", kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZED)); in kbasep_mem_profile_debugfs_insert() 81 if (!kbase_ctx_flag(kctx, KCTX_MEM_PROFILE_INITIALIZE in kbasep_mem_profile_debugfs_insert() 108 kbasep_mem_profile_debugfs_remove(struct kbase_context *kctx) kbasep_mem_profile_debugfs_remove() argument 123 kbasep_mem_profile_debugfs_insert(struct kbase_context *kctx, char *data, size_t size) kbasep_mem_profile_debugfs_insert() argument [all...] |
H A D | mali_kbase_mem.c | 71 static void free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp); 73 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument 92 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) { in kbase_get_num_cpu_va_bits() 102 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn) in kbase_gpu_va_to_rbtree() argument 109 if (gpu_pfn >= kctx->exec_va_start) { in kbase_gpu_va_to_rbtree() 110 rbtree = &kctx->reg_rbtree_exec; in kbase_gpu_va_to_rbtree() 115 if (kbase_ctx_flag(kctx, KCTX_COMPAT)) in kbase_gpu_va_to_rbtree() 120 same_va_end = kctx->same_va_end; in kbase_gpu_va_to_rbtree() 125 rbtree = &kctx->reg_rbtree_custom; in kbase_gpu_va_to_rbtree() 127 rbtree = &kctx in kbase_gpu_va_to_rbtree() 226 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 265 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 488 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 625 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 671 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 688 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument 693 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 764 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument 813 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument 864 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument 1114 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local 1154 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local 1227 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument 1337 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 1393 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1440 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1459 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument 1488 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1537 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1629 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1661 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 1717 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 1779 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 1887 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local 2058 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local 2231 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument 2254 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local 2330 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument 2355 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local 2691 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extent) kbase_check_alloc_sizes() argument 2792 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 2803 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 2815 struct kbase_context *kctx; global() member 2896 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 2922 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 2948 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 2975 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local 3001 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local 3034 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 3093 struct kbase_context *kctx; kbase_jit_destroy_worker() local 3116 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 3170 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument 3293 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument 3329 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas) kbase_jit_grow() argument 3447 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument 3481 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument 3496 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument 3546 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument 3619 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument 3829 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 3885 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local 3911 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 3934 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 3977 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument 4020 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument 4058 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument 4085 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument 4146 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 4220 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 4289 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument 4329 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 4362 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 4422 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument 4437 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument 4449 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 4472 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument 4491 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 4498 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf_event.c | 28 * @kctx: Pointer to the Kbase context this event belongs to. 38 struct kbase_context *kctx; member 43 int kbase_csf_event_wait_add(struct kbase_context *kctx, in kbase_csf_event_wait_add() argument 53 event_cb->kctx = kctx; in kbase_csf_event_wait_add() 57 spin_lock_irqsave(&kctx->csf.event.lock, flags); in kbase_csf_event_wait_add() 58 list_add_tail(&event_cb->link, &kctx->csf.event.callback_list); in kbase_csf_event_wait_add() 59 dev_dbg(kctx->kbdev->dev, in kbase_csf_event_wait_add() 62 spin_unlock_irqrestore(&kctx->csf.event.lock, flags); in kbase_csf_event_wait_add() 70 void kbase_csf_event_wait_remove(struct kbase_context *kctx, in kbase_csf_event_wait_remove() argument 91 sync_update_notify_gpu(struct kbase_context *kctx) sync_update_notify_gpu() argument 111 kbase_csf_event_signal(struct kbase_context *kctx, bool notify_gpu) kbase_csf_event_signal() argument 156 kbase_csf_event_term(struct kbase_context *kctx) kbase_csf_event_term() argument 177 kbase_csf_event_init(struct kbase_context *const kctx) kbase_csf_event_init() argument 184 kbase_csf_event_remove_error(struct kbase_context *kctx, struct kbase_csf_notification *error) kbase_csf_event_remove_error() argument 194 kbase_csf_event_read_error(struct kbase_context *kctx, struct base_csf_notification *event_data) kbase_csf_event_read_error() argument 213 kbase_csf_event_add_error(struct kbase_context *const kctx, struct kbase_csf_notification *const error, struct base_csf_notification const *const data) kbase_csf_event_add_error() argument 239 kbase_csf_event_error_pending(struct kbase_context *kctx) kbase_csf_event_error_pending() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/backend/ |
H A D | mali_kbase_context_jm.c | 41 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument 43 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init() 44 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init() 45 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init() 46 kbasep_jd_debugfs_ctx_init(kctx); in kbase_context_debugfs_init() 50 void kbase_context_debugfs_term(struct kbase_context *const kctx) in kbase_context_debugfs_term() argument 52 debugfs_remove_recursive(kctx->kctx_dentry); in kbase_context_debugfs_term() 56 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument 58 CSTD_UNUSED(kctx); in kbase_context_debugfs_init() 62 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument 69 kbase_context_kbase_kinstr_jm_init(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_init() argument 74 kbase_context_kbase_kinstr_jm_term(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_term() argument 79 kbase_context_kbase_timer_setup(struct kbase_context *kctx) kbase_context_kbase_timer_setup() argument 87 kbase_context_submit_check(struct kbase_context *kctx) kbase_context_submit_check() argument 107 kbase_context_flush_jobs(struct kbase_context *kctx) kbase_context_flush_jobs() argument 113 kbase_context_free(struct kbase_context *kctx) kbase_context_free() argument 162 kbase_context_term_partial( struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument 178 struct kbase_context *kctx; kbase_create_context() local 229 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument [all...] |
H A D | mali_kbase_context_csf.c | 43 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument 45 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init() 46 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init() 47 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init() 48 kbase_csf_queue_group_debugfs_init(kctx); in kbase_context_debugfs_init() 49 kbase_csf_kcpu_debugfs_init(kctx); in kbase_context_debugfs_init() 50 kbase_csf_tiler_heap_debugfs_init(kctx); in kbase_context_debugfs_init() 51 kbase_csf_tiler_heap_total_debugfs_init(kctx); in kbase_context_debugfs_init() 52 kbase_csf_cpu_queue_debugfs_init(kctx); in kbase_context_debugfs_init() 56 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument 62 kbase_context_debugfs_init(struct kbase_context *const kctx) kbase_context_debugfs_init() argument 68 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument 75 kbase_context_free(struct kbase_context *kctx) kbase_context_free() argument 105 kbase_context_term_partial( struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument 121 struct kbase_context *kctx; kbase_create_context() local 172 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/backend/ |
H A D | mali_kbase_context_jm.c | 44 void kbase_context_debugfs_init(struct kbase_context *const kctx) in kbase_context_debugfs_init() argument 46 kbase_debug_mem_view_init(kctx); in kbase_context_debugfs_init() 47 kbase_mem_pool_debugfs_init(kctx->kctx_dentry, kctx); in kbase_context_debugfs_init() 48 kbase_jit_debugfs_init(kctx); in kbase_context_debugfs_init() 49 kbasep_jd_debugfs_ctx_init(kctx); in kbase_context_debugfs_init() 50 kbase_debug_job_fault_context_init(kctx); in kbase_context_debugfs_init() 54 void kbase_context_debugfs_term(struct kbase_context *const kctx) in kbase_context_debugfs_term() argument 56 debugfs_remove_recursive(kctx->kctx_dentry); in kbase_context_debugfs_term() 57 kbase_debug_job_fault_context_term(kctx); in kbase_context_debugfs_term() 61 kbase_context_debugfs_init(struct kbase_context *const kctx) kbase_context_debugfs_init() argument 67 kbase_context_debugfs_term(struct kbase_context *const kctx) kbase_context_debugfs_term() argument 74 kbase_context_kbase_kinstr_jm_init(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_init() argument 84 kbase_context_kbase_kinstr_jm_term(struct kbase_context *kctx) kbase_context_kbase_kinstr_jm_term() argument 89 kbase_context_kbase_timer_setup(struct kbase_context *kctx) kbase_context_kbase_timer_setup() argument 96 kbase_context_submit_check(struct kbase_context *kctx) kbase_context_submit_check() argument 136 kbase_context_term_partial(struct kbase_context *kctx, unsigned int i) kbase_context_term_partial() argument 149 struct kbase_context *kctx; kbase_create_context() local 194 kbase_destroy_context(struct kbase_context *kctx) kbase_destroy_context() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/jm/ |
H A D | mali_kbase_jm_js.h | 80 * @kctx: The kbase_context to operate on 90 int kbasep_js_kctx_init(struct kbase_context *const kctx); 95 * @kctx: The kbase_context to operate on 109 void kbasep_js_kctx_term(struct kbase_context *kctx); 113 * @kctx: The kbase_context 129 static inline void kbase_jsctx_slot_prio_blocked_set(struct kbase_context *kctx, in kbase_jsctx_slot_prio_blocked_set() argument 133 &kctx->slot_tracking[js]; in kbase_jsctx_slot_prio_blocked_set() 135 lockdep_assert_held(&kctx->kbdev->hwaccess_lock); in kbase_jsctx_slot_prio_blocked_set() 137 "When marking slot %d as blocked for priority %d on a kctx, no atoms were pulled - the slot cannot become unblocked", in kbase_jsctx_slot_prio_blocked_set() 141 KBASE_KTRACE_ADD_JM_SLOT_INFO(kctx in kbase_jsctx_slot_prio_blocked_set() 152 kbase_jsctx_atoms_pulled(struct kbase_context *kctx) kbase_jsctx_atoms_pulled() argument 692 kbasep_js_is_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_is_submit_allowed() argument 721 kbasep_js_set_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_set_submit_allowed() argument 750 kbasep_js_clear_submit_allowed( struct kbasep_js_device_data *js_devdata, struct kbase_context *kctx) kbasep_js_clear_submit_allowed() argument 859 kbase_js_runpool_inc_context_count( struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_runpool_inc_context_count() argument 896 kbase_js_runpool_dec_context_count( struct kbase_device *kbdev, struct kbase_context *kctx) kbase_js_runpool_dec_context_count() argument [all...] |