Home
last modified time | relevance | path

Searched defs:kctx (Results 201 - 225 of 239) sorted by relevance

12345678910

/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_kcpu.c46 struct kbase_context *const kctx = kcpu_queue->kctx; in kbase_kcpu_map_import_prepare() local
98 struct kbase_context *const kctx = kcpu_queue->kctx; in kbase_kcpu_unmap_import_prepare_internal() local
165 struct kbase_context *const kctx = queue->kctx; kbase_jit_add_to_pending_alloc_list() local
204 struct kbase_context *const kctx = queue->kctx; kbase_kcpu_jit_allocate_process() local
335 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_jit_allocate_prepare() local
425 kbase_kcpu_jit_retry_pending_allocs(struct kbase_context *kctx) kbase_kcpu_jit_retry_pending_allocs() argument
452 struct kbase_context *kctx = queue->kctx; kbase_kcpu_jit_free_process() local
507 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_jit_free_prepare() local
577 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_csf_queue_group_suspend_prepare() local
680 kbase_csf_queue_group_suspend_process(struct kbase_context *kctx, struct kbase_suspend_copy_buffer *sus_buf, u8 group_handle) kbase_csf_queue_group_suspend_process() argument
691 struct kbase_context *const kctx = kcpu_queue->kctx; event_cqs_callback() local
888 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_cqs_set_prepare() local
1151 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_cqs_set_operation_prepare() local
1192 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_csf_fence_wait_callback() local
1205 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_wait_cancel() local
1293 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_wait_prepare() local
1323 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_signal_process() local
1351 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_signal_prepare() local
1439 delete_queue(struct kbase_context *kctx, u32 id) delete_queue() argument
1942 kbase_csf_kcpu_queue_enqueue(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_enqueue *enq) kbase_csf_kcpu_queue_enqueue() argument
2113 kbase_csf_kcpu_queue_context_init(struct kbase_context *kctx) kbase_csf_kcpu_queue_context_init() argument
2134 kbase_csf_kcpu_queue_context_term(struct kbase_context *kctx) kbase_csf_kcpu_queue_context_term() argument
2151 kbase_csf_kcpu_queue_delete(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_delete *del) kbase_csf_kcpu_queue_delete() argument
2157 kbase_csf_kcpu_queue_new(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_new *newq) kbase_csf_kcpu_queue_new() argument
[all...]
H A Dmali_kbase_csf_firmware_no_mali.c473 struct kbase_context *kctx; in handle_internal_firmware_fatal() local
H A Dmali_kbase_csf.c55 static void put_user_pages_mmap_handle(struct kbase_context *kctx, in put_user_pages_mmap_handle() argument
81 static int get_user_pages_mmap_handle(struct kbase_context *kctx, in get_user_pages_mmap_handle() argument
115 gpu_munmap_user_io_pages(struct kbase_context *kctx, struct kbase_va_region *reg) gpu_munmap_user_io_pages() argument
214 kernel_unmap_user_io_pages(struct kbase_context *kctx, struct kbase_queue *queue) kernel_unmap_user_io_pages() argument
229 kernel_map_user_io_pages(struct kbase_context *kctx, struct kbase_queue *queue) kernel_map_user_io_pages() argument
300 kbase_csf_free_command_stream_user_pages(struct kbase_context *kctx, struct kbase_queue *queue) kbase_csf_free_command_stream_user_pages() argument
321 kbase_csf_alloc_command_stream_user_pages(struct kbase_context *kctx, struct kbase_queue *queue) kbase_csf_alloc_command_stream_user_pages() argument
388 find_queue_group(struct kbase_context *kctx, u8 group_handle) find_queue_group() argument
404 kbase_csf_queue_group_handle_is_valid(struct kbase_context *kctx, u8 group_handle) kbase_csf_queue_group_handle_is_valid() argument
416 find_queue(struct kbase_context *kctx, u64 base_addr) find_queue() argument
453 csf_queue_register_internal(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register *reg, struct kbase_ioctl_cs_queue_register_ex *reg_ex) csf_queue_register_internal() argument
602 kbase_csf_queue_register(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register *reg) kbase_csf_queue_register() argument
608 kbase_csf_queue_register_ex(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register_ex *reg) kbase_csf_queue_register_ex() argument
636 kbase_csf_queue_terminate(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_terminate *term) kbase_csf_queue_terminate() argument
694 kbase_csf_queue_bind(struct kbase_context *kctx, union kbase_ioctl_cs_queue_bind *bind) kbase_csf_queue_bind() argument
742 struct kbase_context *kctx = queue->kctx; get_bound_queue_group() local
778 struct kbase_context *kctx = pending_submission_worker() local
873 enqueue_gpu_submission_work(struct kbase_context *const kctx) enqueue_gpu_submission_work() argument
878 kbase_csf_queue_kick(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_kick *kick) kbase_csf_queue_kick() argument
915 unbind_stopped_queue(struct kbase_context *kctx, struct kbase_queue *queue) unbind_stopped_queue() argument
961 unbind_queue(struct kbase_context *kctx, struct kbase_queue *queue) unbind_queue() argument
976 struct kbase_context *kctx = queue->kctx; kbase_csf_queue_unbind() local
1006 struct kbase_context *kctx = queue->kctx; kbase_csf_queue_unbind_stopped() local
1026 find_free_group_handle(struct kbase_context *const kctx) find_free_group_handle() argument
1080 create_normal_suspend_buffer(struct kbase_context *const kctx, struct kbase_normal_suspend_buffer *s_buf) create_normal_suspend_buffer() argument
1256 create_suspend_buffers(struct kbase_context *const kctx, struct kbase_queue_group * const group) create_suspend_buffers() argument
1304 create_queue_group(struct kbase_context *const kctx, union kbase_ioctl_cs_queue_group_create *const create) create_queue_group() argument
1376 kbase_csf_queue_group_create(struct kbase_context *const kctx, union kbase_ioctl_cs_queue_group_create *const create) kbase_csf_queue_group_create() argument
1431 term_normal_suspend_buffer(struct kbase_context *const kctx, struct kbase_normal_suspend_buffer *s_buf) term_normal_suspend_buffer() argument
1490 struct kbase_context *kctx = group->kctx; kbase_csf_term_descheduled_queue_group() local
1531 struct kbase_context *kctx = group->kctx; term_queue_group() local
1555 struct kbase_context *kctx = group->kctx; remove_pending_group_fatal_error() local
1566 kbase_csf_queue_group_terminate(struct kbase_context *kctx, u8 group_handle) kbase_csf_queue_group_terminate() argument
1609 kbase_csf_queue_group_suspend(struct kbase_context *kctx, struct kbase_suspend_copy_buffer *sus_buf, u8 group_handle) kbase_csf_queue_group_suspend() argument
1665 kbase_csf_active_queue_groups_reset(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_csf_active_queue_groups_reset() argument
1698 kbase_csf_ctx_init(struct kbase_context *kctx) kbase_csf_ctx_init() argument
1757 kbase_csf_ctx_handle_fault(struct kbase_context *kctx, struct kbase_fault *fault) kbase_csf_ctx_handle_fault() argument
1811 kbase_csf_ctx_term(struct kbase_context *kctx) kbase_csf_ctx_term() argument
1936 handle_oom_event(struct kbase_context *const kctx, struct kbase_csf_cmd_stream_info const *const stream) handle_oom_event() argument
2026 struct kbase_context *const kctx = queue->kctx; kbase_queue_oom_event() local
2110 struct kbase_context *kctx = queue->kctx; oom_event_worker() local
2165 struct kbase_context *const kctx = group->kctx; timer_event_worker() local
2301 struct kbase_context *const kctx = queue->kctx; fatal_event_worker() local
2401 struct kbase_context *const kctx = queue->kctx; handle_queue_exception_event() local
[all...]
H A Dmali_kbase_csf_firmware.c1229 struct kbase_context *kctx; in handle_internal_firmware_fatal() local
H A Dmali_kbase_csf_defs.h338 struct kbase_context *kctx; member
458 struct kbase_context *kctx; member
561 struct kbase_context *kctx; member
H A Dmali_kbase_csf_scheduler.c1781 struct kbase_context *kctx = group->kctx; schedule_in_cycle() local
1807 struct kbase_context *const kctx = group->kctx; insert_group_to_runnable() local
1859 struct kbase_context *kctx = group->kctx; remove_group_from_runnable() local
1960 struct kbase_context *kctx = group->kctx; insert_group_to_idle_wait() local
1978 struct kbase_context *kctx = group->kctx; remove_group_from_idle_wait() local
2194 struct kbase_context *kctx = group->kctx; cleanup_csg_slot() local
2324 struct kbase_context *kctx = group->kctx; program_csg_slot() local
2495 struct kbase_context *kctx = group->kctx; sched_evict_group() local
2653 struct kbase_context *kctx = group->kctx; scheduler_group_schedule() local
2759 count_active_address_space(struct kbase_device *kbdev, struct kbase_context *kctx) count_active_address_space() argument
3351 kbase_csf_scheduler_evict_ctx_slots(struct kbase_device *kbdev, struct kbase_context *kctx, struct list_head *evicted_groups) kbase_csf_scheduler_evict_ctx_slots() argument
3633 scheduler_ctx_scan_groups(struct kbase_device *kbdev, struct kbase_context *kctx, int priority) scheduler_ctx_scan_groups() argument
3978 struct kbase_context *kctx; get_tock_top_group() local
4207 struct kbase_context *kctx; scheduler_prepare() local
4932 struct kbase_context *kctx; kbase_csf_scheduler_reset() local
5053 struct kbase_context *const kctx = group->kctx; kbase_csf_scheduler_group_copy_suspend_buf() local
5483 struct kbase_context *const kctx = container_of(work, check_group_sync_update_worker() local
5529 struct kbase_context *const kctx = param; check_group_sync_update_cb() local
5538 kbase_csf_scheduler_context_init(struct kbase_context *kctx) kbase_csf_scheduler_context_init() argument
5576 kbase_csf_scheduler_context_term(struct kbase_context *kctx) kbase_csf_scheduler_context_term() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_jd.c62 get_compat_pointer(struct kbase_context *kctx, const union kbase_pointer *p) in get_compat_pointer() argument
80 struct kbase_context *kctx = katom->kctx; in jd_run_atom() local
151 struct kbase_context *kctx; in kbase_jd_kds_waiters_add() local
658 struct kbase_context *kctx = katom->kctx; jd_check_force_failure() local
786 struct kbase_context *kctx = katom->kctx; jd_done_nolock() local
966 jd_submit_atom(struct kbase_context *kctx, const struct base_jd_atom_v2 *user_atom, struct kbase_jd_atom *katom) jd_submit_atom() argument
1278 kbase_jd_submit(struct kbase_context *kctx, void __user *user_addr, u32 nr_atoms, u32 stride, bool uk6_atom) kbase_jd_submit() argument
1459 struct kbase_context *kctx; kbase_jd_done_worker() local
1640 struct kbase_context *kctx; jd_cancel_worker() local
1704 struct kbase_context *kctx; kbase_jd_done() local
1741 struct kbase_context *kctx; kbase_jd_cancel() local
1763 kbase_jd_zap_context(struct kbase_context *kctx) kbase_jd_zap_context() argument
1827 kbase_jd_init(struct kbase_context *kctx) kbase_jd_init() argument
1892 kbase_jd_exit(struct kbase_context *kctx) kbase_jd_exit() argument
[all...]
H A Dmali_kbase_uku.h284 union kbase_pointer kctx; /**< base context created by process */ member
H A Dmali_kbase_softjobs.c51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job() local
61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job() local
71 struct kbase_context *kctx in kbasep_add_waiting_with_timeout() local
92 kbasep_read_soft_event_status( struct kbase_context *kctx, u64 evt, unsigned char *status) kbasep_read_soft_event_status() argument
109 kbasep_write_soft_event_status( struct kbase_context *kctx, u64 evt, unsigned char new_status) kbasep_write_soft_event_status() argument
139 struct kbase_context *kctx = katom->kctx; kbase_dump_cpu_gpu_time() local
200 struct kbase_context *kctx = katom->kctx; kbase_soft_event_wait_callback() local
215 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_complete_job() local
226 kbasep_complete_triggered_soft_events(struct kbase_context *kctx, u64 evt) kbasep_complete_triggered_soft_events() argument
273 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_check_atom() local
307 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_wait_timeout() local
347 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_wait_timeout_worker() local
359 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_timeout() local
376 struct kbase_context *kctx = from_timer(kctx, t, soft_job_timeout); kbasep_soft_job_timeout_worker() local
423 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_wait() local
444 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_update_locked() local
465 kbase_soft_event_update(struct kbase_context *kctx, u64 event, unsigned char new_status) kbase_soft_event_update() argument
720 kbase_mem_copy_from_extres_page(struct kbase_context *kctx, void *extres_page, struct page **pages, unsigned int nr_pages, unsigned int *target_page_nr, size_t offset, size_t *to_copy) kbase_mem_copy_from_extres_page() argument
792 kbase_mem_copy_from_extres(struct kbase_context *kctx, struct kbase_debug_copy_buffer *buf_data) kbase_mem_copy_from_extres() argument
910 struct kbase_context *kctx = katom->kctx; kbase_jit_allocate_prepare() local
986 struct kbase_context *kctx = katom->kctx; kbase_jit_allocate_process() local
1107 struct kbase_context *kctx = katom->kctx; kbase_jit_free_prepare() local
1117 struct kbase_context *kctx = katom->kctx; kbase_jit_free_process() local
1142 struct kbase_context *kctx = katom->kctx; kbasep_jit_free_finish_worker() local
1157 struct kbase_context *kctx = katom->kctx; kbase_jit_free_finish() local
1526 struct kbase_context *kctx = katom_iter->kctx; kbase_resume_suspended_soft_jobs() local
[all...]
H A Dmali_kbase_mem.c46 static struct rb_root *kbase_reg_flags_to_rbtree(struct kbase_context *kctx, in kbase_reg_flags_to_rbtree() argument
71 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument
98 kbase_region_tracker_insert(struct kbase_context *kctx, struct kbase_va_region *new_reg) kbase_region_tracker_insert() argument
132 kbase_region_tracker_find_region_enclosing_range_free( struct kbase_context *kctx, u64 start_pfn, size_t nr_pages) kbase_region_tracker_find_region_enclosing_range_free() argument
166 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
204 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
236 kbase_region_tracker_find_region_meeting_reqs(struct kbase_context *kctx, struct kbase_va_region *reg_reqs, size_t nr_pages, size_t align) kbase_region_tracker_find_region_meeting_reqs() argument
275 kbase_remove_va_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_remove_va_region() argument
350 kbase_insert_va_region_nolock(struct kbase_context *kctx, struct kbase_va_region *new_reg, struct kbase_va_region *at_reg, u64 start_pfn, size_t nr_pages) kbase_insert_va_region_nolock() argument
413 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
505 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *exec_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
540 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
550 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
648 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit() argument
790 kbase_alloc_free_region(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages, int zone) kbase_alloc_free_region() argument
884 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument
968 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
1009 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1053 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1073 kbase_sync_single(struct kbase_context *kctx, phys_addr_t cpu_pa, phys_addr_t gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1124 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1211 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1244 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
1294 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
1356 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
1455 struct kbase_context *kctx = alloc->imported.kctx; kbase_free_phy_pages_helper() local
1677 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
1688 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
1700 struct kbase_context *kctx; global() member
1783 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
1807 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
1831 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
1853 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
1884 struct kbase_context *kctx; kbase_jit_destroy_worker() local
1907 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
1920 kbase_jit_allocate(struct kbase_context *kctx, struct base_jit_alloc_info *info) kbase_jit_allocate() argument
2048 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
2062 struct kbase_context *kctx = reg->kctx; kbase_jit_backing_lost() local
2083 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
2104 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
2141 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
2250 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
2281 kbase_jd_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_umm_map() argument
2369 kbase_jd_umm_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_jd_umm_unmap() argument
2404 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm , u32 *kds_res_count, struct kds_resource **kds_resources, unsigned long *kds_access_bitmap, bool exclusive ) kbase_map_external_resource() argument
2480 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
2529 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
2590 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
2629 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
2636 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
H A Dmali_kbase_vinstr.c105 struct kbase_context *kctx; member
214 struct kbase_context *kctx = vinstr_ctx->kctx; in enable_hwcnt() local
247 struct kbase_context *kctx = vinstr_ctx->kctx; in disable_hwcnt() local
329 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_map_kernel_dump_buffer() local
357 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_unmap_kernel_dump_buffer() local
[all...]
H A Dmali_kbase_tlstream.c1391 static void kbase_create_timeline_objects(struct kbase_context *kctx) in kbase_create_timeline_objects() argument
1446 int kbase_tlstream_acquire(struct kbase_context *kctx, u32 flags) in kbase_tlstream_acquire() argument
[all...]
/third_party/openssl/test/
H A Devp_pkey_provided_test.c1356 EVP_PKEY_CTX *pctx = NULL, *ctx = NULL, *kctx = NULL; in test_ec_dup_no_operation() local
1389 EVP_PKEY_CTX *pctx = NULL, *ctx = NULL, *kctx = NULL; in test_ec_dup_keygen_operation() local
H A Devp_kdf_test.c26 EVP_KDF_CTX *kctx = EVP_KDF_CTX_new(kdf); in get_kdfbyname_libctx() local
62 EVP_KDF_CTX *kctx = NULL; in test_kdf_tls1_prf() local
85 EVP_KDF_CTX *kctx = NULL; in test_kdf_tls1_prf_invalid_digest() local
102 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_zero_output_size() local
122 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_empty_secret() local
140 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_1byte_secret() local
158 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_empty_seed() local
178 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_1byte_seed() local
223 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf() local
249 EVP_KDF_CTX *kctx = NULL; do_kdf_hkdf_gettables() local
305 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_invalid_digest() local
322 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_derive_set_params_fail() local
345 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_set_invalid_mode() local
369 EVP_KDF_CTX *kctx = NULL; do_kdf_hkdf_set_invalid_param() local
405 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_zero_output_size() local
425 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_empty_key() local
443 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_1byte_key() local
461 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_empty_salt() local
500 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf1() local
572 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2() local
604 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_output() local
631 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_large_output() local
662 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_salt() local
687 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_iterations() local
712 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_salt_pkcs5() local
749 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_iterations_pkcs5() local
786 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_invalid_digest() local
812 EVP_KDF_CTX *kctx; test_kdf_scrypt() local
855 EVP_KDF_CTX *kctx; test_kdf_ss_hash() local
893 EVP_KDF_CTX *kctx; test_kdf_x963() local
951 EVP_KDF_CTX *kctx; test_kdf_kbkdf_6803_128() local
1006 EVP_KDF_CTX *kctx; test_kdf_kbkdf_6803_256() local
1099 EVP_KDF_CTX *kctx; test_kdf_kbkdf_invalid_digest() local
1121 EVP_KDF_CTX *kctx; test_kdf_kbkdf_invalid_mac() local
1143 EVP_KDF_CTX *kctx; test_kdf_kbkdf_empty_key() local
1167 EVP_KDF_CTX *kctx; test_kdf_kbkdf_1byte_key() local
1189 EVP_KDF_CTX *kctx; test_kdf_kbkdf_zero_output_size() local
1215 EVP_KDF_CTX *kctx; test_kdf_kbkdf_8009_prf1() local
1255 EVP_KDF_CTX *kctx; test_kdf_kbkdf_8009_prf2() local
1306 EVP_KDF_CTX *kctx; test_kdf_kbkdf_fixedinfo() local
1360 EVP_KDF_CTX *kctx; test_kdf_ss_hmac() local
1401 EVP_KDF_CTX *kctx; test_kdf_ss_kmac() local
1453 EVP_KDF_CTX *kctx; test_kdf_sshkdf() local
1562 EVP_KDF_CTX *kctx = NULL; test_kdf_x942_asn1() local
1598 EVP_KDF_CTX *kctx; test_kdf_krb5kdf() local
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.c42 static struct rb_root *kbase_reg_flags_to_rbtree(struct kbase_context *kctx, struct kbase_va_region *reg) in kbase_reg_flags_to_rbtree() argument
66 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn) in kbase_gpu_va_to_rbtree() argument
94 kbase_region_tracker_insert(struct kbase_context *kctx, struct kbase_va_region *new_reg) kbase_region_tracker_insert() argument
128 kbase_region_tracker_find_region_enclosing_range_free(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages) kbase_region_tracker_find_region_enclosing_range_free() argument
162 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
200 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
232 kbase_region_tracker_find_region_meeting_reqs(struct kbase_context *kctx, struct kbase_va_region *reg_reqs, size_t nr_pages, size_t align) kbase_region_tracker_find_region_meeting_reqs() argument
272 kbase_remove_va_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_remove_va_region() argument
345 kbase_insert_va_region_nolock(struct kbase_context *kctx, struct kbase_va_region *new_reg, struct kbase_va_region *at_reg, u64 start_pfn, size_t nr_pages) kbase_insert_va_region_nolock() argument
397 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
492 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *exec_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
527 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
537 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
628 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit() argument
762 kbase_alloc_free_region(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages, int zone) kbase_alloc_free_region() argument
855 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument
938 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
979 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1026 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1045 kbase_sync_single(struct kbase_context *kctx, phys_addr_t cpu_pa, phys_addr_t gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1092 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1173 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1205 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
1255 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
1316 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
1410 struct kbase_context *kctx = alloc->imported.kctx; kbase_free_phy_pages_helper() local
1645 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
1656 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
1668 struct kbase_context *kctx; global() member
1749 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
1775 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
1801 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
1825 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
1853 struct kbase_context *kctx; kbase_jit_destroy_worker() local
1875 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
1888 kbase_jit_allocate(struct kbase_context *kctx, struct base_jit_alloc_info *info) kbase_jit_allocate() argument
2012 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
2026 struct kbase_context *kctx = reg->kctx; kbase_jit_backing_lost() local
2048 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
2069 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
2104 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
2194 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
2223 kbase_jd_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_umm_map() argument
2301 kbase_jd_umm_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_jd_umm_unmap() argument
2335 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm, u32 *kds_res_count, struct kds_resource **kds_resources, unsigned long *kds_access_bitmap, bool exclusive) kbase_map_external_resource() argument
2409 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
2454 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
2516 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
2553 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
2560 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
H A Dmali_kbase_vinstr.c93 struct kbase_context *kctx; member
189 struct kbase_context *kctx = vinstr_ctx->kctx; in enable_hwcnt() local
222 struct kbase_context *kctx = vinstr_ctx->kctx; in disable_hwcnt() local
297 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_map_kernel_dump_buffer() local
320 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_unmap_kernel_dump_buffer() local
[all...]
H A Dmali_kbase_tlstream.c1059 static void kbase_create_timeline_objects(struct kbase_context *kctx) in kbase_create_timeline_objects() argument
1107 int kbase_tlstream_acquire(struct kbase_context *kctx, u32 flags) in kbase_tlstream_acquire() argument
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_defs.h276 struct kbase_context *kctx; member
382 struct kbase_context *kctx; member
464 struct kbase_context *kctx; member
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c116 kbase_find_event_mem_region(struct kbase_context *kctx, u64 gpu_addr) in kbase_find_event_mem_region() argument
179 kbase_phy_alloc_mapping_init(struct kbase_context *kctx, struct kbase_va_region *reg, size_t vsize, size_t size) kbase_phy_alloc_mapping_init() argument
228 kbase_phy_alloc_mapping_term(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_phy_alloc_mapping_term() argument
245 kbase_phy_alloc_mapping_get(struct kbase_context *kctx, u64 gpu_addr, struct kbase_vmap_struct **out_kern_mapping) kbase_phy_alloc_mapping_get() argument
290 kbase_phy_alloc_mapping_put(struct kbase_context *kctx, struct kbase_vmap_struct *kern_mapping) kbase_phy_alloc_mapping_put() argument
307 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extent, u64 *flags, u64 *gpu_va) kbase_mem_alloc() argument
552 kbase_mem_query(struct kbase_context *kctx, u64 gpu_addr, u64 query, u64 *const out) kbase_mem_query() argument
686 struct kbase_context *kctx; kbase_mem_evictable_reclaim_count_objects() local
722 struct kbase_context *kctx; kbase_mem_evictable_reclaim_scan_objects() local
785 kbase_mem_evictable_init(struct kbase_context *kctx) kbase_mem_evictable_init() argument
807 kbase_mem_evictable_deinit(struct kbase_context *kctx) kbase_mem_evictable_deinit() argument
818 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_mark_reclaim() local
837 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_unmark_reclaim() local
855 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_make() local
879 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_unmake() local
924 kbase_mem_flags_change(struct kbase_context *kctx, u64 gpu_addr, unsigned int flags, unsigned int mask) kbase_mem_flags_change() argument
1075 kbase_mem_do_sync_imported(struct kbase_context *kctx, struct kbase_va_region *reg, enum kbase_sync_type sync_fn) kbase_mem_do_sync_imported() argument
1181 kbase_mem_umm_unmap_attachment(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap_attachment() argument
1209 kbase_mem_umm_map_attachment(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map_attachment() argument
1274 kbase_mem_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map() argument
1348 kbase_mem_umm_unmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap() argument
1376 get_umm_memory_group_id(struct kbase_context *kctx, struct dma_buf *dma_buf) get_umm_memory_group_id() argument
1409 kbase_mem_from_umm(struct kbase_context *kctx, int fd, u64 *va_pages, u64 *flags, u32 padding) kbase_mem_from_umm() argument
1565 kbase_mem_from_user_buffer(struct kbase_context *kctx, unsigned long address, unsigned long size, u64 *va_pages, u64 *flags) kbase_mem_from_user_buffer() argument
1780 kbase_mem_alias(struct kbase_context *kctx, u64 *flags, u64 stride, u64 nents, struct base_mem_aliasing_info *ai, u64 *num_pages) global() argument
1991 kbase_mem_import(struct kbase_context *kctx, enum base_mem_import_type type, void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages, u64 *flags) global() argument
2133 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument
2153 kbase_mem_shrink_cpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument
2182 kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages, u64 const old_pages) global() argument
2197 kbase_mem_commit(struct kbase_context *kctx, u64 gpu_addr, u64 new_pages) global() argument
2320 kbase_mem_shrink(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages) global() argument
2516 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
2612 kbase_free_unused_jit_allocations(struct kbase_context *kctx) global() argument
2630 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
2695 kbase_os_mem_map_lock(struct kbase_context *kctx) global() argument
2701 kbase_os_mem_map_unlock(struct kbase_context *kctx) global() argument
2707 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2775 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument
2949 kbase_sync_mem_regions(struct kbase_context *kctx, struct kbase_vmap_struct *map, enum kbase_sync_type dest) global() argument
2980 kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, u64 offset_bytes, size_t size, struct kbase_vmap_struct *map) global() argument
3066 kbase_vmap_prot(struct kbase_context *kctx, u64 gpu_addr, size_t size, unsigned long prot_request, struct kbase_vmap_struct *map) global() argument
3113 kbase_vmap(struct kbase_context *kctx, u64 gpu_addr, size_t size, struct kbase_vmap_struct *map) global() argument
3125 kbase_vunmap_phy_pages(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument
3143 kbase_vunmap(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument
3165 kbasep_os_process_page_usage_update(struct kbase_context *kctx, int pages) global() argument
3184 kbasep_os_process_page_usage_drain(struct kbase_context *kctx) global() argument
3213 struct kbase_context *kctx; global() local
3223 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3280 struct kbase_context *kctx; global() local
3389 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3464 struct kbase_context *kctx = vma->vm_private_data; global() local
3480 struct kbase_context *kctx = vma->vm_private_data; global() local
3501 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
H A Dmali_kbase_mem.c73 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument
102 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn) in kbase_gpu_va_to_rbtree() argument
226 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
265 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
488 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
625 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
671 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
688 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument
693 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
764 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument
813 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
864 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument
1114 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local
1154 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local
1227 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument
1337 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
1393 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1440 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1459 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument
1488 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1537 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1629 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1661 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
1717 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
1779 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
1887 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local
2058 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local
2231 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
2254 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local
2330 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument
2355 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local
2691 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extent) kbase_check_alloc_sizes() argument
2792 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
2803 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
2815 struct kbase_context *kctx; global() member
2896 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
2922 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
2948 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
2975 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local
3001 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local
3034 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
3093 struct kbase_context *kctx; kbase_jit_destroy_worker() local
3116 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
3170 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument
3293 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument
3329 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas) kbase_jit_grow() argument
3447 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument
3481 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument
3496 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument
3546 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument
3619 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument
3829 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
3885 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local
3911 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
3934 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
3977 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument
4020 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument
4058 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument
4085 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument
4146 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
4220 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
4289 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument
4329 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
4362 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
4422 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument
4437 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument
4449 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
4472 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument
4491 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
4498 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c111 kbase_find_event_mem_region( struct kbase_context *kctx, u64 gpu_addr) kbase_find_event_mem_region() argument
172 kbase_phy_alloc_mapping_init(struct kbase_context *kctx, struct kbase_va_region *reg, size_t vsize, size_t size) kbase_phy_alloc_mapping_init() argument
216 kbase_phy_alloc_mapping_term(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_phy_alloc_mapping_term() argument
233 kbase_phy_alloc_mapping_get(struct kbase_context *kctx, u64 gpu_addr, struct kbase_vmap_struct **out_kern_mapping) kbase_phy_alloc_mapping_get() argument
277 kbase_phy_alloc_mapping_put(struct kbase_context *kctx, struct kbase_vmap_struct *kern_mapping) kbase_phy_alloc_mapping_put() argument
295 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extension, u64 *flags, u64 *gpu_va, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mem_alloc() argument
520 kbase_mem_query(struct kbase_context *kctx, u64 gpu_addr, u64 query, u64 * const out) kbase_mem_query() argument
640 struct kbase_context *kctx; kbase_mem_evictable_reclaim_count_objects() local
677 struct kbase_context *kctx; kbase_mem_evictable_reclaim_scan_objects() local
729 kbase_mem_evictable_init(struct kbase_context *kctx) kbase_mem_evictable_init() argument
747 kbase_mem_evictable_deinit(struct kbase_context *kctx) kbase_mem_evictable_deinit() argument
758 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_mark_reclaim() local
781 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_unmark_reclaim() local
803 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_make() local
829 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_unmake() local
881 kbase_mem_flags_change(struct kbase_context *kctx, u64 gpu_addr, unsigned int flags, unsigned int mask) kbase_mem_flags_change() argument
1028 kbase_mem_do_sync_imported(struct kbase_context *kctx, struct kbase_va_region *reg, enum kbase_sync_type sync_fn) kbase_mem_do_sync_imported() argument
1130 kbase_mem_umm_unmap_attachment(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap_attachment() argument
1158 kbase_mem_umm_map_attachment(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map_attachment() argument
1221 kbase_mem_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map() argument
1301 kbase_mem_umm_unmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap() argument
1331 get_umm_memory_group_id(struct kbase_context *kctx, struct dma_buf *dma_buf) get_umm_memory_group_id() argument
1365 kbase_mem_from_umm(struct kbase_context *kctx, int fd, u64 *va_pages, u64 *flags, u32 padding) kbase_mem_from_umm() argument
1513 kbase_mem_from_user_buffer( struct kbase_context *kctx, unsigned long address, unsigned long size, u64 *va_pages, u64 *flags) kbase_mem_from_user_buffer() argument
1725 kbase_mem_alias(struct kbase_context *kctx, u64 *flags, u64 stride, u64 nents, struct base_mem_aliasing_info *ai, u64 *num_pages) global() argument
1952 kbase_mem_import(struct kbase_context *kctx, enum base_mem_import_type type, void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages, u64 *flags) global() argument
2096 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages, enum kbase_caller_mmu_sync_info mmu_sync_info) global() argument
2118 kbase_mem_shrink_cpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument
2146 kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages, u64 const old_pages) global() argument
2159 kbase_mem_commit(struct kbase_context *kctx, u64 gpu_addr, u64 new_pages) global() argument
2291 kbase_mem_shrink(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages) global() argument
2484 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
2577 kbase_free_unused_jit_allocations(struct kbase_context *kctx) global() argument
2594 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
2661 kbase_os_mem_map_lock(struct kbase_context *kctx) global() argument
2667 kbase_os_mem_map_unlock(struct kbase_context *kctx) global() argument
2673 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2752 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument
2926 kbase_sync_mem_regions(struct kbase_context *kctx, struct kbase_vmap_struct *map, enum kbase_sync_type dest) global() argument
2956 kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, u64 offset_bytes, size_t size, struct kbase_vmap_struct *map) global() argument
3033 kbase_vmap_prot(struct kbase_context *kctx, u64 gpu_addr, size_t size, unsigned long prot_request, struct kbase_vmap_struct *map) global() argument
3078 kbase_vmap(struct kbase_context *kctx, u64 gpu_addr, size_t size, struct kbase_vmap_struct *map) global() argument
3091 kbase_vunmap_phy_pages(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument
3109 kbase_vunmap(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument
3131 kbasep_os_process_page_usage_update(struct kbase_context *kctx, int pages) global() argument
3150 kbasep_os_process_page_usage_drain(struct kbase_context *kctx) global() argument
3178 struct kbase_context *kctx; global() local
3188 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3239 struct kbase_context *kctx; global() local
3362 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3436 struct kbase_context *kctx = vma->vm_private_data; global() local
3452 struct kbase_context *kctx = vma->vm_private_data; global() local
3493 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
H A Dmali_kbase_mem.h69 struct kbase_context *kctx; member
150 struct kbase_context *kctx; member
163 struct kbase_context *kctx; member
565 kbase_va_region_alloc_get( struct kbase_context *kctx, struct kbase_va_region *region) kbase_va_region_alloc_get() argument
580 kbase_va_region_alloc_put( struct kbase_context *kctx, struct kbase_va_region *region) kbase_va_region_alloc_put() argument
637 kbase_alloc_create( struct kbase_context *kctx, size_t nr_pages, enum kbase_memory_type type, int group_id) kbase_alloc_create() argument
696 kbase_reg_prepare_native(struct kbase_va_region *reg, struct kbase_context *kctx, int group_id) kbase_reg_prepare_native() argument
1315 kbase_process_page_usage_inc(struct kbase_context *kctx, int pages) kbase_process_page_usage_inc() argument
1331 kbase_process_page_usage_dec(struct kbase_context *kctx, int pages) kbase_process_page_usage_dec() argument
1704 kbase_jit_request_phys_increase_locked(struct kbase_context *kctx, size_t needed_pages) kbase_jit_request_phys_increase_locked() argument
1744 kbase_jit_request_phys_increase(struct kbase_context *kctx, size_t needed_pages) kbase_jit_request_phys_increase() argument
1780 kbase_jit_done_phys_increase(struct kbase_context *kctx, size_t needed_pages) kbase_jit_done_phys_increase() argument
1935 kbase_link_event_mem_page(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_link_event_mem_page() argument
1955 kbase_unlink_event_mem_page(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_unlink_event_mem_page() argument
2075 kbase_ctx_reg_zone_init(struct kbase_context *kctx, unsigned long zone_bits, u64 base_pfn, u64 va_size_pages) kbase_ctx_reg_zone_init() argument
2102 kbase_ctx_reg_zone_get_nolock(struct kbase_context *kctx, unsigned long zone_bits) kbase_ctx_reg_zone_get_nolock() argument
2120 kbase_ctx_reg_zone_get(struct kbase_context *kctx, unsigned long zone_bits) kbase_ctx_reg_zone_get() argument
2137 kbase_mem_allow_alloc(struct kbase_context *kctx) kbase_mem_allow_alloc() argument
[all...]
H A Dmali_kbase_js.c88 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
93 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument
155 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument
183 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument
215 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach_prio() argument
275 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach() argument
297 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument
330 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument
358 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument
376 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument
418 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument
639 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument
684 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument
733 kbase_jsctx_slot_prio_blocked_clear(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_blocked_clear() argument
746 kbase_jsctx_slot_atoms_pulled(struct kbase_context *kctx, int js) kbase_jsctx_slot_atoms_pulled() argument
756 kbase_jsctx_slot_prio_is_blocked(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_is_blocked() argument
793 kbase_jsctx_slot_atom_pulled_inc(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_inc() argument
831 kbase_jsctx_slot_atom_pulled_dec(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_dec() argument
882 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument
925 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument
969 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument
1000 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument
1043 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument
1085 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local
1119 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local
1141 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument
1203 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument
1369 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument
1391 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument
1423 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local
1470 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local
1513 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument
1657 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument
1685 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument
1734 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument
1790 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument
1953 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument
1965 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument
1986 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument
2021 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument
2035 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument
2077 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
2221 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument
2249 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
2322 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
2362 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2386 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local
2410 struct kbase_context *kctx, *n; global() local
2518 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
2585 struct kbase_context *const kctx = katom->kctx; global() local
2636 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument
2669 kbase_js_pull(struct kbase_context *kctx, int js) global() argument
2784 struct kbase_context *const kctx = start_katom->kctx; global() local
2893 struct kbase_context *const kctx = end_katom->kctx; global() local
2975 struct kbase_context *kctx = katom->kctx; global() local
3111 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
3146 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument
3242 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument
3282 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument
3423 struct kbase_context *const kctx = end_katom->kctx; global() local
3467 struct kbase_context *kctx = katom->kctx; global() local
3551 struct kbase_context *const kctx = katom->kctx; global() local
3629 struct kbase_context *kctx; global() local
3816 kbase_js_zap_context(struct kbase_context *kctx) global() argument
3971 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument
3995 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb *callback) global() argument
[all...]
H A Dmali_kbase_mem.c74 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument
103 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument
232 kbase_region_tracker_find_region_enclosing_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument
272 kbase_region_tracker_find_region_base_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument
555 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument
702 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument
748 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument
765 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument
771 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument
855 kbase_has_exec_va_zone_locked(struct kbase_context *kctx) kbase_has_exec_va_zone_locked() argument
865 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument
892 kbase_region_tracker_has_allocs(struct kbase_context *kctx) kbase_region_tracker_has_allocs() argument
947 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument
1016 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
1082 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument
1385 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local
1430 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local
1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument
1622 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument
1698 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument
1742 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument
1762 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument
1792 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument
1845 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument
1944 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument
1977 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument
2036 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument
2100 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument
2208 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local
2402 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local
2594 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
2624 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local
2718 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument
2746 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local
3090 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extension) kbase_check_alloc_sizes() argument
3210 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument
3222 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument
3234 struct kbase_context *kctx; global() member
3317 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local
3341 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local
3365 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local
3390 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local
3417 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local
3452 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument
3518 struct kbase_context *kctx; kbase_jit_destroy_worker() local
3542 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument
3595 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument
3725 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument
3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument
3882 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument
3917 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument
3931 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument
3984 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument
4070 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument
4292 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument
4353 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local
4377 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument
4401 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument
4446 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument
4491 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument
4554 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument
4628 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument
4711 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument
4781 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument
4822 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument
4858 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument
4917 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument
4934 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument
4949 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument
4970 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument
4988 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument
4995 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument
[all...]
/third_party/wpa_supplicant/wpa_supplicant-2.9/src/crypto/
H A Dcrypto_openssl.c1974 EVP_PKEY_CTX *kctx = NULL; in crypto_ecdh_init() local

Completed in 47 milliseconds

12345678910