| /device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_kcpu.c | 46 struct kbase_context *const kctx = kcpu_queue->kctx; in kbase_kcpu_map_import_prepare() local 98 struct kbase_context *const kctx = kcpu_queue->kctx; in kbase_kcpu_unmap_import_prepare_internal() local 165 struct kbase_context *const kctx = queue->kctx; kbase_jit_add_to_pending_alloc_list() local 204 struct kbase_context *const kctx = queue->kctx; kbase_kcpu_jit_allocate_process() local 335 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_jit_allocate_prepare() local 425 kbase_kcpu_jit_retry_pending_allocs(struct kbase_context *kctx) kbase_kcpu_jit_retry_pending_allocs() argument 452 struct kbase_context *kctx = queue->kctx; kbase_kcpu_jit_free_process() local 507 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_jit_free_prepare() local 577 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_csf_queue_group_suspend_prepare() local 680 kbase_csf_queue_group_suspend_process(struct kbase_context *kctx, struct kbase_suspend_copy_buffer *sus_buf, u8 group_handle) kbase_csf_queue_group_suspend_process() argument 691 struct kbase_context *const kctx = kcpu_queue->kctx; event_cqs_callback() local 888 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_cqs_set_prepare() local 1151 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_cqs_set_operation_prepare() local 1192 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_csf_fence_wait_callback() local 1205 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_wait_cancel() local 1293 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_wait_prepare() local 1323 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_signal_process() local 1351 struct kbase_context *const kctx = kcpu_queue->kctx; kbase_kcpu_fence_signal_prepare() local 1439 delete_queue(struct kbase_context *kctx, u32 id) delete_queue() argument 1942 kbase_csf_kcpu_queue_enqueue(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_enqueue *enq) kbase_csf_kcpu_queue_enqueue() argument 2113 kbase_csf_kcpu_queue_context_init(struct kbase_context *kctx) kbase_csf_kcpu_queue_context_init() argument 2134 kbase_csf_kcpu_queue_context_term(struct kbase_context *kctx) kbase_csf_kcpu_queue_context_term() argument 2151 kbase_csf_kcpu_queue_delete(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_delete *del) kbase_csf_kcpu_queue_delete() argument 2157 kbase_csf_kcpu_queue_new(struct kbase_context *kctx, struct kbase_ioctl_kcpu_queue_new *newq) kbase_csf_kcpu_queue_new() argument [all...] |
| H A D | mali_kbase_csf_firmware_no_mali.c | 473 struct kbase_context *kctx; in handle_internal_firmware_fatal() local
|
| H A D | mali_kbase_csf.c | 55 static void put_user_pages_mmap_handle(struct kbase_context *kctx, in put_user_pages_mmap_handle() argument 81 static int get_user_pages_mmap_handle(struct kbase_context *kctx, in get_user_pages_mmap_handle() argument 115 gpu_munmap_user_io_pages(struct kbase_context *kctx, struct kbase_va_region *reg) gpu_munmap_user_io_pages() argument 214 kernel_unmap_user_io_pages(struct kbase_context *kctx, struct kbase_queue *queue) kernel_unmap_user_io_pages() argument 229 kernel_map_user_io_pages(struct kbase_context *kctx, struct kbase_queue *queue) kernel_map_user_io_pages() argument 300 kbase_csf_free_command_stream_user_pages(struct kbase_context *kctx, struct kbase_queue *queue) kbase_csf_free_command_stream_user_pages() argument 321 kbase_csf_alloc_command_stream_user_pages(struct kbase_context *kctx, struct kbase_queue *queue) kbase_csf_alloc_command_stream_user_pages() argument 388 find_queue_group(struct kbase_context *kctx, u8 group_handle) find_queue_group() argument 404 kbase_csf_queue_group_handle_is_valid(struct kbase_context *kctx, u8 group_handle) kbase_csf_queue_group_handle_is_valid() argument 416 find_queue(struct kbase_context *kctx, u64 base_addr) find_queue() argument 453 csf_queue_register_internal(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register *reg, struct kbase_ioctl_cs_queue_register_ex *reg_ex) csf_queue_register_internal() argument 602 kbase_csf_queue_register(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register *reg) kbase_csf_queue_register() argument 608 kbase_csf_queue_register_ex(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_register_ex *reg) kbase_csf_queue_register_ex() argument 636 kbase_csf_queue_terminate(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_terminate *term) kbase_csf_queue_terminate() argument 694 kbase_csf_queue_bind(struct kbase_context *kctx, union kbase_ioctl_cs_queue_bind *bind) kbase_csf_queue_bind() argument 742 struct kbase_context *kctx = queue->kctx; get_bound_queue_group() local 778 struct kbase_context *kctx = pending_submission_worker() local 873 enqueue_gpu_submission_work(struct kbase_context *const kctx) enqueue_gpu_submission_work() argument 878 kbase_csf_queue_kick(struct kbase_context *kctx, struct kbase_ioctl_cs_queue_kick *kick) kbase_csf_queue_kick() argument 915 unbind_stopped_queue(struct kbase_context *kctx, struct kbase_queue *queue) unbind_stopped_queue() argument 961 unbind_queue(struct kbase_context *kctx, struct kbase_queue *queue) unbind_queue() argument 976 struct kbase_context *kctx = queue->kctx; kbase_csf_queue_unbind() local 1006 struct kbase_context *kctx = queue->kctx; kbase_csf_queue_unbind_stopped() local 1026 find_free_group_handle(struct kbase_context *const kctx) find_free_group_handle() argument 1080 create_normal_suspend_buffer(struct kbase_context *const kctx, struct kbase_normal_suspend_buffer *s_buf) create_normal_suspend_buffer() argument 1256 create_suspend_buffers(struct kbase_context *const kctx, struct kbase_queue_group * const group) create_suspend_buffers() argument 1304 create_queue_group(struct kbase_context *const kctx, union kbase_ioctl_cs_queue_group_create *const create) create_queue_group() argument 1376 kbase_csf_queue_group_create(struct kbase_context *const kctx, union kbase_ioctl_cs_queue_group_create *const create) kbase_csf_queue_group_create() argument 1431 term_normal_suspend_buffer(struct kbase_context *const kctx, struct kbase_normal_suspend_buffer *s_buf) term_normal_suspend_buffer() argument 1490 struct kbase_context *kctx = group->kctx; kbase_csf_term_descheduled_queue_group() local 1531 struct kbase_context *kctx = group->kctx; term_queue_group() local 1555 struct kbase_context *kctx = group->kctx; remove_pending_group_fatal_error() local 1566 kbase_csf_queue_group_terminate(struct kbase_context *kctx, u8 group_handle) kbase_csf_queue_group_terminate() argument 1609 kbase_csf_queue_group_suspend(struct kbase_context *kctx, struct kbase_suspend_copy_buffer *sus_buf, u8 group_handle) kbase_csf_queue_group_suspend() argument 1665 kbase_csf_active_queue_groups_reset(struct kbase_device *kbdev, struct kbase_context *kctx) kbase_csf_active_queue_groups_reset() argument 1698 kbase_csf_ctx_init(struct kbase_context *kctx) kbase_csf_ctx_init() argument 1757 kbase_csf_ctx_handle_fault(struct kbase_context *kctx, struct kbase_fault *fault) kbase_csf_ctx_handle_fault() argument 1811 kbase_csf_ctx_term(struct kbase_context *kctx) kbase_csf_ctx_term() argument 1936 handle_oom_event(struct kbase_context *const kctx, struct kbase_csf_cmd_stream_info const *const stream) handle_oom_event() argument 2026 struct kbase_context *const kctx = queue->kctx; kbase_queue_oom_event() local 2110 struct kbase_context *kctx = queue->kctx; oom_event_worker() local 2165 struct kbase_context *const kctx = group->kctx; timer_event_worker() local 2301 struct kbase_context *const kctx = queue->kctx; fatal_event_worker() local 2401 struct kbase_context *const kctx = queue->kctx; handle_queue_exception_event() local [all...] |
| H A D | mali_kbase_csf_firmware.c | 1229 struct kbase_context *kctx; in handle_internal_firmware_fatal() local
|
| H A D | mali_kbase_csf_defs.h | 338 struct kbase_context *kctx; member 458 struct kbase_context *kctx; member 561 struct kbase_context *kctx; member
|
| H A D | mali_kbase_csf_scheduler.c | 1781 struct kbase_context *kctx = group->kctx; schedule_in_cycle() local 1807 struct kbase_context *const kctx = group->kctx; insert_group_to_runnable() local 1859 struct kbase_context *kctx = group->kctx; remove_group_from_runnable() local 1960 struct kbase_context *kctx = group->kctx; insert_group_to_idle_wait() local 1978 struct kbase_context *kctx = group->kctx; remove_group_from_idle_wait() local 2194 struct kbase_context *kctx = group->kctx; cleanup_csg_slot() local 2324 struct kbase_context *kctx = group->kctx; program_csg_slot() local 2495 struct kbase_context *kctx = group->kctx; sched_evict_group() local 2653 struct kbase_context *kctx = group->kctx; scheduler_group_schedule() local 2759 count_active_address_space(struct kbase_device *kbdev, struct kbase_context *kctx) count_active_address_space() argument 3351 kbase_csf_scheduler_evict_ctx_slots(struct kbase_device *kbdev, struct kbase_context *kctx, struct list_head *evicted_groups) kbase_csf_scheduler_evict_ctx_slots() argument 3633 scheduler_ctx_scan_groups(struct kbase_device *kbdev, struct kbase_context *kctx, int priority) scheduler_ctx_scan_groups() argument 3978 struct kbase_context *kctx; get_tock_top_group() local 4207 struct kbase_context *kctx; scheduler_prepare() local 4932 struct kbase_context *kctx; kbase_csf_scheduler_reset() local 5053 struct kbase_context *const kctx = group->kctx; kbase_csf_scheduler_group_copy_suspend_buf() local 5483 struct kbase_context *const kctx = container_of(work, check_group_sync_update_worker() local 5529 struct kbase_context *const kctx = param; check_group_sync_update_cb() local 5538 kbase_csf_scheduler_context_init(struct kbase_context *kctx) kbase_csf_scheduler_context_init() argument 5576 kbase_csf_scheduler_context_term(struct kbase_context *kctx) kbase_csf_scheduler_context_term() argument [all...] |
| /device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_jd.c | 62 get_compat_pointer(struct kbase_context *kctx, const union kbase_pointer *p) in get_compat_pointer() argument 80 struct kbase_context *kctx = katom->kctx; in jd_run_atom() local 151 struct kbase_context *kctx; in kbase_jd_kds_waiters_add() local 658 struct kbase_context *kctx = katom->kctx; jd_check_force_failure() local 786 struct kbase_context *kctx = katom->kctx; jd_done_nolock() local 966 jd_submit_atom(struct kbase_context *kctx, const struct base_jd_atom_v2 *user_atom, struct kbase_jd_atom *katom) jd_submit_atom() argument 1278 kbase_jd_submit(struct kbase_context *kctx, void __user *user_addr, u32 nr_atoms, u32 stride, bool uk6_atom) kbase_jd_submit() argument 1459 struct kbase_context *kctx; kbase_jd_done_worker() local 1640 struct kbase_context *kctx; jd_cancel_worker() local 1704 struct kbase_context *kctx; kbase_jd_done() local 1741 struct kbase_context *kctx; kbase_jd_cancel() local 1763 kbase_jd_zap_context(struct kbase_context *kctx) kbase_jd_zap_context() argument 1827 kbase_jd_init(struct kbase_context *kctx) kbase_jd_init() argument 1892 kbase_jd_exit(struct kbase_context *kctx) kbase_jd_exit() argument [all...] |
| H A D | mali_kbase_uku.h | 284 union kbase_pointer kctx; /**< base context created by process */ member
|
| H A D | mali_kbase_softjobs.c | 51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job() local 61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job() local 71 struct kbase_context *kctx in kbasep_add_waiting_with_timeout() local 92 kbasep_read_soft_event_status( struct kbase_context *kctx, u64 evt, unsigned char *status) kbasep_read_soft_event_status() argument 109 kbasep_write_soft_event_status( struct kbase_context *kctx, u64 evt, unsigned char new_status) kbasep_write_soft_event_status() argument 139 struct kbase_context *kctx = katom->kctx; kbase_dump_cpu_gpu_time() local 200 struct kbase_context *kctx = katom->kctx; kbase_soft_event_wait_callback() local 215 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_complete_job() local 226 kbasep_complete_triggered_soft_events(struct kbase_context *kctx, u64 evt) kbasep_complete_triggered_soft_events() argument 273 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_check_atom() local 307 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_wait_timeout() local 347 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_wait_timeout_worker() local 359 struct kbase_context *kctx = katom->kctx; kbase_fence_debug_timeout() local 376 struct kbase_context *kctx = from_timer(kctx, t, soft_job_timeout); kbasep_soft_job_timeout_worker() local 423 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_wait() local 444 struct kbase_context *kctx = katom->kctx; kbasep_soft_event_update_locked() local 465 kbase_soft_event_update(struct kbase_context *kctx, u64 event, unsigned char new_status) kbase_soft_event_update() argument 720 kbase_mem_copy_from_extres_page(struct kbase_context *kctx, void *extres_page, struct page **pages, unsigned int nr_pages, unsigned int *target_page_nr, size_t offset, size_t *to_copy) kbase_mem_copy_from_extres_page() argument 792 kbase_mem_copy_from_extres(struct kbase_context *kctx, struct kbase_debug_copy_buffer *buf_data) kbase_mem_copy_from_extres() argument 910 struct kbase_context *kctx = katom->kctx; kbase_jit_allocate_prepare() local 986 struct kbase_context *kctx = katom->kctx; kbase_jit_allocate_process() local 1107 struct kbase_context *kctx = katom->kctx; kbase_jit_free_prepare() local 1117 struct kbase_context *kctx = katom->kctx; kbase_jit_free_process() local 1142 struct kbase_context *kctx = katom->kctx; kbasep_jit_free_finish_worker() local 1157 struct kbase_context *kctx = katom->kctx; kbase_jit_free_finish() local 1526 struct kbase_context *kctx = katom_iter->kctx; kbase_resume_suspended_soft_jobs() local [all...] |
| H A D | mali_kbase_mem.c | 46 static struct rb_root *kbase_reg_flags_to_rbtree(struct kbase_context *kctx, in kbase_reg_flags_to_rbtree() argument 71 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument 98 kbase_region_tracker_insert(struct kbase_context *kctx, struct kbase_va_region *new_reg) kbase_region_tracker_insert() argument 132 kbase_region_tracker_find_region_enclosing_range_free( struct kbase_context *kctx, u64 start_pfn, size_t nr_pages) kbase_region_tracker_find_region_enclosing_range_free() argument 166 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 204 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 236 kbase_region_tracker_find_region_meeting_reqs(struct kbase_context *kctx, struct kbase_va_region *reg_reqs, size_t nr_pages, size_t align) kbase_region_tracker_find_region_meeting_reqs() argument 275 kbase_remove_va_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_remove_va_region() argument 350 kbase_insert_va_region_nolock(struct kbase_context *kctx, struct kbase_va_region *new_reg, struct kbase_va_region *at_reg, u64 start_pfn, size_t nr_pages) kbase_insert_va_region_nolock() argument 413 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 505 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *exec_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 540 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 550 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 648 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit() argument 790 kbase_alloc_free_region(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages, int zone) kbase_alloc_free_region() argument 884 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument 968 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 1009 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1053 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1073 kbase_sync_single(struct kbase_context *kctx, phys_addr_t cpu_pa, phys_addr_t gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1124 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1211 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1244 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 1294 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 1356 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 1455 struct kbase_context *kctx = alloc->imported.kctx; kbase_free_phy_pages_helper() local 1677 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 1688 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 1700 struct kbase_context *kctx; global() member 1783 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 1807 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 1831 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 1853 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 1884 struct kbase_context *kctx; kbase_jit_destroy_worker() local 1907 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 1920 kbase_jit_allocate(struct kbase_context *kctx, struct base_jit_alloc_info *info) kbase_jit_allocate() argument 2048 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 2062 struct kbase_context *kctx = reg->kctx; kbase_jit_backing_lost() local 2083 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 2104 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 2141 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 2250 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 2281 kbase_jd_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_umm_map() argument 2369 kbase_jd_umm_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_jd_umm_unmap() argument 2404 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm , u32 *kds_res_count, struct kds_resource **kds_resources, unsigned long *kds_access_bitmap, bool exclusive ) kbase_map_external_resource() argument 2480 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 2529 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 2590 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 2629 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 2636 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
| H A D | mali_kbase_vinstr.c | 105 struct kbase_context *kctx; member 214 struct kbase_context *kctx = vinstr_ctx->kctx; in enable_hwcnt() local 247 struct kbase_context *kctx = vinstr_ctx->kctx; in disable_hwcnt() local 329 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_map_kernel_dump_buffer() local 357 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_unmap_kernel_dump_buffer() local [all...] |
| H A D | mali_kbase_tlstream.c | 1391 static void kbase_create_timeline_objects(struct kbase_context *kctx) in kbase_create_timeline_objects() argument 1446 int kbase_tlstream_acquire(struct kbase_context *kctx, u32 flags) in kbase_tlstream_acquire() argument [all...] |
| /third_party/openssl/test/ |
| H A D | evp_pkey_provided_test.c | 1356 EVP_PKEY_CTX *pctx = NULL, *ctx = NULL, *kctx = NULL; in test_ec_dup_no_operation() local 1389 EVP_PKEY_CTX *pctx = NULL, *ctx = NULL, *kctx = NULL; in test_ec_dup_keygen_operation() local
|
| H A D | evp_kdf_test.c | 26 EVP_KDF_CTX *kctx = EVP_KDF_CTX_new(kdf); in get_kdfbyname_libctx() local 62 EVP_KDF_CTX *kctx = NULL; in test_kdf_tls1_prf() local 85 EVP_KDF_CTX *kctx = NULL; in test_kdf_tls1_prf_invalid_digest() local 102 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_zero_output_size() local 122 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_empty_secret() local 140 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_1byte_secret() local 158 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_empty_seed() local 178 EVP_KDF_CTX *kctx = NULL; test_kdf_tls1_prf_1byte_seed() local 223 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf() local 249 EVP_KDF_CTX *kctx = NULL; do_kdf_hkdf_gettables() local 305 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_invalid_digest() local 322 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_derive_set_params_fail() local 345 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_set_invalid_mode() local 369 EVP_KDF_CTX *kctx = NULL; do_kdf_hkdf_set_invalid_param() local 405 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_zero_output_size() local 425 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_empty_key() local 443 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_1byte_key() local 461 EVP_KDF_CTX *kctx = NULL; test_kdf_hkdf_empty_salt() local 500 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf1() local 572 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2() local 604 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_output() local 631 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_large_output() local 662 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_salt() local 687 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_iterations() local 712 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_salt_pkcs5() local 749 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_small_iterations_pkcs5() local 786 EVP_KDF_CTX *kctx = NULL; test_kdf_pbkdf2_invalid_digest() local 812 EVP_KDF_CTX *kctx; test_kdf_scrypt() local 855 EVP_KDF_CTX *kctx; test_kdf_ss_hash() local 893 EVP_KDF_CTX *kctx; test_kdf_x963() local 951 EVP_KDF_CTX *kctx; test_kdf_kbkdf_6803_128() local 1006 EVP_KDF_CTX *kctx; test_kdf_kbkdf_6803_256() local 1099 EVP_KDF_CTX *kctx; test_kdf_kbkdf_invalid_digest() local 1121 EVP_KDF_CTX *kctx; test_kdf_kbkdf_invalid_mac() local 1143 EVP_KDF_CTX *kctx; test_kdf_kbkdf_empty_key() local 1167 EVP_KDF_CTX *kctx; test_kdf_kbkdf_1byte_key() local 1189 EVP_KDF_CTX *kctx; test_kdf_kbkdf_zero_output_size() local 1215 EVP_KDF_CTX *kctx; test_kdf_kbkdf_8009_prf1() local 1255 EVP_KDF_CTX *kctx; test_kdf_kbkdf_8009_prf2() local 1306 EVP_KDF_CTX *kctx; test_kdf_kbkdf_fixedinfo() local 1360 EVP_KDF_CTX *kctx; test_kdf_ss_hmac() local 1401 EVP_KDF_CTX *kctx; test_kdf_ss_kmac() local 1453 EVP_KDF_CTX *kctx; test_kdf_sshkdf() local 1562 EVP_KDF_CTX *kctx = NULL; test_kdf_x942_asn1() local 1598 EVP_KDF_CTX *kctx; test_kdf_krb5kdf() local [all...] |
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_mem.c | 42 static struct rb_root *kbase_reg_flags_to_rbtree(struct kbase_context *kctx, struct kbase_va_region *reg)
in kbase_reg_flags_to_rbtree() argument 66 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn)
in kbase_gpu_va_to_rbtree() argument 94 kbase_region_tracker_insert(struct kbase_context *kctx, struct kbase_va_region *new_reg) kbase_region_tracker_insert() argument 128 kbase_region_tracker_find_region_enclosing_range_free(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages) kbase_region_tracker_find_region_enclosing_range_free() argument 162 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 200 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 232 kbase_region_tracker_find_region_meeting_reqs(struct kbase_context *kctx, struct kbase_va_region *reg_reqs, size_t nr_pages, size_t align) kbase_region_tracker_find_region_meeting_reqs() argument 272 kbase_remove_va_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_remove_va_region() argument 345 kbase_insert_va_region_nolock(struct kbase_context *kctx, struct kbase_va_region *new_reg, struct kbase_va_region *at_reg, u64 start_pfn, size_t nr_pages) kbase_insert_va_region_nolock() argument 397 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 492 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *exec_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 527 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 537 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 628 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit() argument 762 kbase_alloc_free_region(struct kbase_context *kctx, u64 start_pfn, size_t nr_pages, int zone) kbase_alloc_free_region() argument 855 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument 938 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 979 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1026 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1045 kbase_sync_single(struct kbase_context *kctx, phys_addr_t cpu_pa, phys_addr_t gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1092 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1173 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1205 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 1255 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 1316 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 1410 struct kbase_context *kctx = alloc->imported.kctx; kbase_free_phy_pages_helper() local 1645 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 1656 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 1668 struct kbase_context *kctx; global() member 1749 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 1775 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 1801 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 1825 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 1853 struct kbase_context *kctx; kbase_jit_destroy_worker() local 1875 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 1888 kbase_jit_allocate(struct kbase_context *kctx, struct base_jit_alloc_info *info) kbase_jit_allocate() argument 2012 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 2026 struct kbase_context *kctx = reg->kctx; kbase_jit_backing_lost() local 2048 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 2069 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 2104 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 2194 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 2223 kbase_jd_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_umm_map() argument 2301 kbase_jd_umm_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_jd_umm_unmap() argument 2335 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm, u32 *kds_res_count, struct kds_resource **kds_resources, unsigned long *kds_access_bitmap, bool exclusive) kbase_map_external_resource() argument 2409 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 2454 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 2516 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 2553 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 2560 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
| H A D | mali_kbase_vinstr.c | 93 struct kbase_context *kctx;
member 189 struct kbase_context *kctx = vinstr_ctx->kctx;
in enable_hwcnt() local 222 struct kbase_context *kctx = vinstr_ctx->kctx;
in disable_hwcnt() local 297 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_map_kernel_dump_buffer() local 320 struct kbase_context *kctx = vinstr_ctx->kctx; kbasep_vinstr_unmap_kernel_dump_buffer() local [all...] |
| H A D | mali_kbase_tlstream.c | 1059 static void kbase_create_timeline_objects(struct kbase_context *kctx)
in kbase_create_timeline_objects() argument 1107 int kbase_tlstream_acquire(struct kbase_context *kctx, u32 flags)
in kbase_tlstream_acquire() argument [all...] |
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_defs.h | 276 struct kbase_context *kctx; member 382 struct kbase_context *kctx; member 464 struct kbase_context *kctx; member
|
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_linux.c | 116 kbase_find_event_mem_region(struct kbase_context *kctx, u64 gpu_addr) in kbase_find_event_mem_region() argument 179 kbase_phy_alloc_mapping_init(struct kbase_context *kctx, struct kbase_va_region *reg, size_t vsize, size_t size) kbase_phy_alloc_mapping_init() argument 228 kbase_phy_alloc_mapping_term(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_phy_alloc_mapping_term() argument 245 kbase_phy_alloc_mapping_get(struct kbase_context *kctx, u64 gpu_addr, struct kbase_vmap_struct **out_kern_mapping) kbase_phy_alloc_mapping_get() argument 290 kbase_phy_alloc_mapping_put(struct kbase_context *kctx, struct kbase_vmap_struct *kern_mapping) kbase_phy_alloc_mapping_put() argument 307 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extent, u64 *flags, u64 *gpu_va) kbase_mem_alloc() argument 552 kbase_mem_query(struct kbase_context *kctx, u64 gpu_addr, u64 query, u64 *const out) kbase_mem_query() argument 686 struct kbase_context *kctx; kbase_mem_evictable_reclaim_count_objects() local 722 struct kbase_context *kctx; kbase_mem_evictable_reclaim_scan_objects() local 785 kbase_mem_evictable_init(struct kbase_context *kctx) kbase_mem_evictable_init() argument 807 kbase_mem_evictable_deinit(struct kbase_context *kctx) kbase_mem_evictable_deinit() argument 818 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_mark_reclaim() local 837 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_unmark_reclaim() local 855 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_make() local 879 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_unmake() local 924 kbase_mem_flags_change(struct kbase_context *kctx, u64 gpu_addr, unsigned int flags, unsigned int mask) kbase_mem_flags_change() argument 1075 kbase_mem_do_sync_imported(struct kbase_context *kctx, struct kbase_va_region *reg, enum kbase_sync_type sync_fn) kbase_mem_do_sync_imported() argument 1181 kbase_mem_umm_unmap_attachment(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap_attachment() argument 1209 kbase_mem_umm_map_attachment(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map_attachment() argument 1274 kbase_mem_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map() argument 1348 kbase_mem_umm_unmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap() argument 1376 get_umm_memory_group_id(struct kbase_context *kctx, struct dma_buf *dma_buf) get_umm_memory_group_id() argument 1409 kbase_mem_from_umm(struct kbase_context *kctx, int fd, u64 *va_pages, u64 *flags, u32 padding) kbase_mem_from_umm() argument 1565 kbase_mem_from_user_buffer(struct kbase_context *kctx, unsigned long address, unsigned long size, u64 *va_pages, u64 *flags) kbase_mem_from_user_buffer() argument 1780 kbase_mem_alias(struct kbase_context *kctx, u64 *flags, u64 stride, u64 nents, struct base_mem_aliasing_info *ai, u64 *num_pages) global() argument 1991 kbase_mem_import(struct kbase_context *kctx, enum base_mem_import_type type, void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages, u64 *flags) global() argument 2133 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument 2153 kbase_mem_shrink_cpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument 2182 kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages, u64 const old_pages) global() argument 2197 kbase_mem_commit(struct kbase_context *kctx, u64 gpu_addr, u64 new_pages) global() argument 2320 kbase_mem_shrink(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages) global() argument 2516 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 2612 kbase_free_unused_jit_allocations(struct kbase_context *kctx) global() argument 2630 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 2695 kbase_os_mem_map_lock(struct kbase_context *kctx) global() argument 2701 kbase_os_mem_map_unlock(struct kbase_context *kctx) global() argument 2707 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2775 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument 2949 kbase_sync_mem_regions(struct kbase_context *kctx, struct kbase_vmap_struct *map, enum kbase_sync_type dest) global() argument 2980 kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, u64 offset_bytes, size_t size, struct kbase_vmap_struct *map) global() argument 3066 kbase_vmap_prot(struct kbase_context *kctx, u64 gpu_addr, size_t size, unsigned long prot_request, struct kbase_vmap_struct *map) global() argument 3113 kbase_vmap(struct kbase_context *kctx, u64 gpu_addr, size_t size, struct kbase_vmap_struct *map) global() argument 3125 kbase_vunmap_phy_pages(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument 3143 kbase_vunmap(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument 3165 kbasep_os_process_page_usage_update(struct kbase_context *kctx, int pages) global() argument 3184 kbasep_os_process_page_usage_drain(struct kbase_context *kctx) global() argument 3213 struct kbase_context *kctx; global() local 3223 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3280 struct kbase_context *kctx; global() local 3389 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3464 struct kbase_context *kctx = vma->vm_private_data; global() local 3480 struct kbase_context *kctx = vma->vm_private_data; global() local 3501 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
| H A D | mali_kbase_mem.c | 73 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument 102 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, u64 gpu_pfn) in kbase_gpu_va_to_rbtree() argument 226 kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 265 kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 488 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 625 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 671 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 688 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument 693 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 764 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument 813 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument 864 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument 1114 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local 1154 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local 1227 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_gpu_mmap() argument 1337 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 1393 kbasep_find_enclosing_cpu_mapping(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1440 kbasep_find_enclosing_cpu_mapping_offset(struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1459 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument 1488 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1537 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1629 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1661 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 1717 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 1779 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 1887 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local 2058 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local 2231 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument 2254 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local 2330 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument 2355 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local 2691 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extent) kbase_check_alloc_sizes() argument 2792 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 2803 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 2815 struct kbase_context *kctx; global() member 2896 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 2922 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 2948 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 2975 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local 3001 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local 3034 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 3093 struct kbase_context *kctx; kbase_jit_destroy_worker() local 3116 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 3170 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument 3293 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument 3329 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas) kbase_jit_grow() argument 3447 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument 3481 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument 3496 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument 3546 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument 3619 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument 3829 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 3885 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local 3911 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 3934 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 3977 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument 4020 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument 4058 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument 4085 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument 4146 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 4220 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 4289 kbase_map_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument 4329 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 4362 kbase_sticky_resource_acquire(struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 4422 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument 4437 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument 4449 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 4472 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument 4491 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 4498 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
| /device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_linux.c | 111 kbase_find_event_mem_region( struct kbase_context *kctx, u64 gpu_addr) kbase_find_event_mem_region() argument 172 kbase_phy_alloc_mapping_init(struct kbase_context *kctx, struct kbase_va_region *reg, size_t vsize, size_t size) kbase_phy_alloc_mapping_init() argument 216 kbase_phy_alloc_mapping_term(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_phy_alloc_mapping_term() argument 233 kbase_phy_alloc_mapping_get(struct kbase_context *kctx, u64 gpu_addr, struct kbase_vmap_struct **out_kern_mapping) kbase_phy_alloc_mapping_get() argument 277 kbase_phy_alloc_mapping_put(struct kbase_context *kctx, struct kbase_vmap_struct *kern_mapping) kbase_phy_alloc_mapping_put() argument 295 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extension, u64 *flags, u64 *gpu_va, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mem_alloc() argument 520 kbase_mem_query(struct kbase_context *kctx, u64 gpu_addr, u64 query, u64 * const out) kbase_mem_query() argument 640 struct kbase_context *kctx; kbase_mem_evictable_reclaim_count_objects() local 677 struct kbase_context *kctx; kbase_mem_evictable_reclaim_scan_objects() local 729 kbase_mem_evictable_init(struct kbase_context *kctx) kbase_mem_evictable_init() argument 747 kbase_mem_evictable_deinit(struct kbase_context *kctx) kbase_mem_evictable_deinit() argument 758 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_mark_reclaim() local 781 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_mem_evictable_unmark_reclaim() local 803 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_make() local 829 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; kbase_mem_evictable_unmake() local 881 kbase_mem_flags_change(struct kbase_context *kctx, u64 gpu_addr, unsigned int flags, unsigned int mask) kbase_mem_flags_change() argument 1028 kbase_mem_do_sync_imported(struct kbase_context *kctx, struct kbase_va_region *reg, enum kbase_sync_type sync_fn) kbase_mem_do_sync_imported() argument 1130 kbase_mem_umm_unmap_attachment(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap_attachment() argument 1158 kbase_mem_umm_map_attachment(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map_attachment() argument 1221 kbase_mem_umm_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_umm_map() argument 1301 kbase_mem_umm_unmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_mem_umm_unmap() argument 1331 get_umm_memory_group_id(struct kbase_context *kctx, struct dma_buf *dma_buf) get_umm_memory_group_id() argument 1365 kbase_mem_from_umm(struct kbase_context *kctx, int fd, u64 *va_pages, u64 *flags, u32 padding) kbase_mem_from_umm() argument 1513 kbase_mem_from_user_buffer( struct kbase_context *kctx, unsigned long address, unsigned long size, u64 *va_pages, u64 *flags) kbase_mem_from_user_buffer() argument 1725 kbase_mem_alias(struct kbase_context *kctx, u64 *flags, u64 stride, u64 nents, struct base_mem_aliasing_info *ai, u64 *num_pages) global() argument 1952 kbase_mem_import(struct kbase_context *kctx, enum base_mem_import_type type, void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages, u64 *flags) global() argument 2096 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages, enum kbase_caller_mmu_sync_info mmu_sync_info) global() argument 2118 kbase_mem_shrink_cpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages) global() argument 2146 kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages, u64 const old_pages) global() argument 2159 kbase_mem_commit(struct kbase_context *kctx, u64 gpu_addr, u64 new_pages) global() argument 2291 kbase_mem_shrink(struct kbase_context *const kctx, struct kbase_va_region *const reg, u64 const new_pages) global() argument 2484 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 2577 kbase_free_unused_jit_allocations(struct kbase_context *kctx) global() argument 2594 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 2661 kbase_os_mem_map_lock(struct kbase_context *kctx) global() argument 2667 kbase_os_mem_map_unlock(struct kbase_context *kctx) global() argument 2673 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2752 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument 2926 kbase_sync_mem_regions(struct kbase_context *kctx, struct kbase_vmap_struct *map, enum kbase_sync_type dest) global() argument 2956 kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, u64 offset_bytes, size_t size, struct kbase_vmap_struct *map) global() argument 3033 kbase_vmap_prot(struct kbase_context *kctx, u64 gpu_addr, size_t size, unsigned long prot_request, struct kbase_vmap_struct *map) global() argument 3078 kbase_vmap(struct kbase_context *kctx, u64 gpu_addr, size_t size, struct kbase_vmap_struct *map) global() argument 3091 kbase_vunmap_phy_pages(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument 3109 kbase_vunmap(struct kbase_context *kctx, struct kbase_vmap_struct *map) global() argument 3131 kbasep_os_process_page_usage_update(struct kbase_context *kctx, int pages) global() argument 3150 kbasep_os_process_page_usage_drain(struct kbase_context *kctx) global() argument 3178 struct kbase_context *kctx; global() local 3188 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3239 struct kbase_context *kctx; global() local 3362 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3436 struct kbase_context *kctx = vma->vm_private_data; global() local 3452 struct kbase_context *kctx = vma->vm_private_data; global() local 3493 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
| H A D | mali_kbase_mem.h | 69 struct kbase_context *kctx; member 150 struct kbase_context *kctx; member 163 struct kbase_context *kctx; member 565 kbase_va_region_alloc_get( struct kbase_context *kctx, struct kbase_va_region *region) kbase_va_region_alloc_get() argument 580 kbase_va_region_alloc_put( struct kbase_context *kctx, struct kbase_va_region *region) kbase_va_region_alloc_put() argument 637 kbase_alloc_create( struct kbase_context *kctx, size_t nr_pages, enum kbase_memory_type type, int group_id) kbase_alloc_create() argument 696 kbase_reg_prepare_native(struct kbase_va_region *reg, struct kbase_context *kctx, int group_id) kbase_reg_prepare_native() argument 1315 kbase_process_page_usage_inc(struct kbase_context *kctx, int pages) kbase_process_page_usage_inc() argument 1331 kbase_process_page_usage_dec(struct kbase_context *kctx, int pages) kbase_process_page_usage_dec() argument 1704 kbase_jit_request_phys_increase_locked(struct kbase_context *kctx, size_t needed_pages) kbase_jit_request_phys_increase_locked() argument 1744 kbase_jit_request_phys_increase(struct kbase_context *kctx, size_t needed_pages) kbase_jit_request_phys_increase() argument 1780 kbase_jit_done_phys_increase(struct kbase_context *kctx, size_t needed_pages) kbase_jit_done_phys_increase() argument 1935 kbase_link_event_mem_page(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_link_event_mem_page() argument 1955 kbase_unlink_event_mem_page(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_unlink_event_mem_page() argument 2075 kbase_ctx_reg_zone_init(struct kbase_context *kctx, unsigned long zone_bits, u64 base_pfn, u64 va_size_pages) kbase_ctx_reg_zone_init() argument 2102 kbase_ctx_reg_zone_get_nolock(struct kbase_context *kctx, unsigned long zone_bits) kbase_ctx_reg_zone_get_nolock() argument 2120 kbase_ctx_reg_zone_get(struct kbase_context *kctx, unsigned long zone_bits) kbase_ctx_reg_zone_get() argument 2137 kbase_mem_allow_alloc(struct kbase_context *kctx) kbase_mem_allow_alloc() argument [all...] |
| H A D | mali_kbase_js.c | 88 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 93 static int kbase_ktrace_get_ctx_refcnt(struct kbase_context *kctx) in kbase_ktrace_get_ctx_refcnt() argument 155 jsctx_rb_none_to_pull_prio(struct kbase_context *kctx, int js, int prio) in jsctx_rb_none_to_pull_prio() argument 183 jsctx_rb_none_to_pull(struct kbase_context *kctx, int js) jsctx_rb_none_to_pull() argument 215 jsctx_queue_foreach_prio(struct kbase_context *kctx, int js, int prio, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach_prio() argument 275 jsctx_queue_foreach(struct kbase_context *kctx, int js, kbasep_js_ctx_job_cb *callback) jsctx_queue_foreach() argument 297 jsctx_rb_peek_prio(struct kbase_context *kctx, int js, int prio) jsctx_rb_peek_prio() argument 330 jsctx_rb_peek(struct kbase_context *kctx, int js) jsctx_rb_peek() argument 358 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_pull() argument 376 jsctx_tree_add(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_tree_add() argument 418 jsctx_rb_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) jsctx_rb_unpull() argument 639 kbasep_js_kctx_init(struct kbase_context *const kctx) kbasep_js_kctx_init() argument 684 kbasep_js_kctx_term(struct kbase_context *kctx) kbasep_js_kctx_term() argument 733 kbase_jsctx_slot_prio_blocked_clear(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_blocked_clear() argument 746 kbase_jsctx_slot_atoms_pulled(struct kbase_context *kctx, int js) kbase_jsctx_slot_atoms_pulled() argument 756 kbase_jsctx_slot_prio_is_blocked(struct kbase_context *kctx, int js, int sched_prio) kbase_jsctx_slot_prio_is_blocked() argument 793 kbase_jsctx_slot_atom_pulled_inc(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_inc() argument 831 kbase_jsctx_slot_atom_pulled_dec(struct kbase_context *kctx, const struct kbase_jd_atom *katom) kbase_jsctx_slot_atom_pulled_dec() argument 882 kbase_js_ctx_list_add_pullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_nolock() argument 925 kbase_js_ctx_list_add_pullable_head_nolock( struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head_nolock() argument 969 kbase_js_ctx_list_add_pullable_head(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_pullable_head() argument 1000 kbase_js_ctx_list_add_unpullable_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_add_unpullable_nolock() argument 1043 kbase_js_ctx_list_remove_nolock(struct kbase_device *kbdev, struct kbase_context *kctx, int js) kbase_js_ctx_list_remove_nolock() argument 1085 struct kbase_context *kctx; kbase_js_ctx_list_pop_head_nolock() local 1119 struct kbase_context *kctx; kbase_js_ctx_list_pop_head() local 1141 kbase_js_ctx_pullable(struct kbase_context *kctx, int js, bool is_scheduled) kbase_js_ctx_pullable() argument 1203 kbase_js_dep_validate(struct kbase_context *kctx, struct kbase_jd_atom *katom) kbase_js_dep_validate() argument 1369 kbase_js_set_ctx_priority(struct kbase_context *kctx, int new_priority) kbase_js_set_ctx_priority() argument 1391 kbase_js_update_ctx_priority(struct kbase_context *kctx) kbase_js_update_ctx_priority() argument 1423 struct kbase_context *const kctx = start_katom->kctx; js_add_start_rp() local 1470 struct kbase_context *const kctx = end_katom->kctx; js_add_end_rp() local 1513 kbasep_js_add_job(struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_add_job() argument 1657 kbasep_js_remove_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *atom) kbasep_js_remove_job() argument 1685 kbasep_js_remove_cancelled_job(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom) kbasep_js_remove_cancelled_job() argument 1734 kbasep_js_run_jobs_after_ctx_and_atom_release( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state, bool runpool_ctx_attr_change) kbasep_js_run_jobs_after_ctx_and_atom_release() argument 1790 kbasep_js_runpool_release_ctx_internal( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_internal() argument 1953 kbasep_js_runpool_release_ctx_nolock(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_nolock() argument 1965 kbasep_js_runpool_requeue_or_kill_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, bool has_pm_ref) kbasep_js_runpool_requeue_or_kill_ctx() argument 1986 kbasep_js_runpool_release_ctx_and_katom_retained_state( struct kbase_device *kbdev, struct kbase_context *kctx, struct kbasep_js_atom_retained_state *katom_retained_state) kbasep_js_runpool_release_ctx_and_katom_retained_state() argument 2021 kbasep_js_runpool_release_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx() argument 2035 kbasep_js_runpool_release_ctx_no_schedule( struct kbase_device *kbdev, struct kbase_context *kctx) kbasep_js_runpool_release_ctx_no_schedule() argument 2077 kbasep_js_schedule_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 2221 kbase_js_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int js) global() argument 2249 kbasep_js_schedule_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 2322 kbasep_js_release_privileged_ctx(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 2362 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2386 struct kbase_context *kctx = kbdev->as_to_kctx[i]; global() local 2410 struct kbase_context *kctx, *n; global() local 2518 kbase_js_dep_resolved_submit(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 2585 struct kbase_context *const kctx = katom->kctx; global() local 2636 kbase_js_evict_deps(struct kbase_context *kctx, struct kbase_jd_atom *katom, int js, int prio) global() argument 2669 kbase_js_pull(struct kbase_context *kctx, int js) global() argument 2784 struct kbase_context *const kctx = start_katom->kctx; global() local 2893 struct kbase_context *const kctx = end_katom->kctx; global() local 2975 struct kbase_context *kctx = katom->kctx; global() local 3111 kbase_js_unpull(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 3146 js_complete_start_rp(struct kbase_context *kctx, struct kbase_jd_atom *const start_katom) global() argument 3242 js_complete_end_rp(struct kbase_context *kctx, struct kbase_jd_atom *const end_katom) global() argument 3282 kbase_js_complete_atom_wq(struct kbase_context *kctx, struct kbase_jd_atom *katom) global() argument 3423 struct kbase_context *const kctx = end_katom->kctx; global() local 3467 struct kbase_context *kctx = katom->kctx; global() local 3551 struct kbase_context *const kctx = katom->kctx; global() local 3629 struct kbase_context *kctx; global() local 3816 kbase_js_zap_context(struct kbase_context *kctx) global() argument 3971 trace_get_refcnt(struct kbase_device *kbdev, struct kbase_context *kctx) global() argument 3995 kbase_js_foreach_ctx_job(struct kbase_context *kctx, kbasep_js_ctx_job_cb *callback) global() argument [all...] |
| H A D | mali_kbase_mem.c | 74 static size_t kbase_get_num_cpu_va_bits(struct kbase_context *kctx) in kbase_get_num_cpu_va_bits() argument 103 static struct rb_root *kbase_gpu_va_to_rbtree(struct kbase_context *kctx, in kbase_gpu_va_to_rbtree() argument 232 kbase_region_tracker_find_region_enclosing_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_enclosing_address() argument 272 kbase_region_tracker_find_region_base_address( struct kbase_context *kctx, u64 gpu_addr) kbase_region_tracker_find_region_base_address() argument 555 kbase_add_va_region(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align) kbase_add_va_region() argument 702 kbase_region_tracker_ds_init(struct kbase_context *kctx, struct kbase_va_region *same_va_reg, struct kbase_va_region *custom_va_reg) kbase_region_tracker_ds_init() argument 748 kbase_region_tracker_term(struct kbase_context *kctx) kbase_region_tracker_term() argument 765 kbase_get_same_va_bits(struct kbase_context *kctx) kbase_get_same_va_bits() argument 771 kbase_region_tracker_init(struct kbase_context *kctx) kbase_region_tracker_init() argument 855 kbase_has_exec_va_zone_locked(struct kbase_context *kctx) kbase_has_exec_va_zone_locked() argument 865 kbase_has_exec_va_zone(struct kbase_context *kctx) kbase_has_exec_va_zone() argument 892 kbase_region_tracker_has_allocs(struct kbase_context *kctx) kbase_region_tracker_has_allocs() argument 947 kbase_region_tracker_init_jit_64(struct kbase_context *kctx, u64 jit_va_pages) kbase_region_tracker_init_jit_64() argument 1016 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument 1082 kbase_region_tracker_init_exec(struct kbase_context *kctx, u64 exec_va_pages) kbase_region_tracker_init_exec() argument 1385 struct kbase_context *kctx = NULL; kbase_reg_flags_to_kctx() local 1430 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_free_alloced_region() local 1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument 1622 kbase_gpu_munmap(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_gpu_munmap() argument 1698 kbasep_find_enclosing_cpu_mapping( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping() argument 1742 kbasep_find_enclosing_cpu_mapping_offset( struct kbase_context *kctx, unsigned long uaddr, size_t size, u64 *offset) kbasep_find_enclosing_cpu_mapping_offset() argument 1762 kbasep_find_enclosing_gpu_mapping_start_and_offset(struct kbase_context *kctx, u64 gpu_addr, size_t size, u64 *start, u64 *offset) kbasep_find_enclosing_gpu_mapping_start_and_offset() argument 1792 kbase_sync_single(struct kbase_context *kctx, struct tagged_addr t_cpu_pa, struct tagged_addr t_gpu_pa, off_t offset, size_t size, enum kbase_sync_type sync_fn) kbase_sync_single() argument 1845 kbase_do_syncset(struct kbase_context *kctx, struct basep_syncset *sset, enum kbase_sync_type sync_fn) kbase_do_syncset() argument 1944 kbase_sync_now(struct kbase_context *kctx, struct basep_syncset *sset) kbase_sync_now() argument 1977 kbase_mem_free_region(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_mem_free_region() argument 2036 kbase_mem_free(struct kbase_context *kctx, u64 gpu_addr) kbase_mem_free() argument 2100 kbase_update_region_flags(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned long flags) kbase_update_region_flags() argument 2208 struct kbase_context *kctx; kbase_alloc_phy_pages_helper() local 2402 struct kbase_context *kctx; kbase_alloc_phy_pages_helper_locked() local 2594 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument 2624 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper() local 2718 free_partial_locked(struct kbase_context *kctx, struct kbase_mem_pool *pool, struct tagged_addr tp) free_partial_locked() argument 2746 struct kbase_context *kctx = alloc->imported.native.kctx; kbase_free_phy_pages_helper_locked() local 3090 kbase_check_alloc_sizes(struct kbase_context *kctx, unsigned long flags, u64 va_pages, u64 commit_pages, u64 large_extension) kbase_check_alloc_sizes() argument 3210 kbase_gpu_vm_lock(struct kbase_context *kctx) kbase_gpu_vm_lock() argument 3222 kbase_gpu_vm_unlock(struct kbase_context *kctx) kbase_gpu_vm_unlock() argument 3234 struct kbase_context *kctx; global() member 3317 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_count_get() local 3341 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_vm_get() local 3365 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_phys_get() local 3390 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_used_get() local 3417 struct kbase_context *kctx = data->kctx; kbase_jit_debugfs_trim_get() local 3452 kbase_jit_debugfs_init(struct kbase_context *kctx) kbase_jit_debugfs_init() argument 3518 struct kbase_context *kctx; kbase_jit_destroy_worker() local 3542 kbase_jit_init(struct kbase_context *kctx) kbase_jit_init() argument 3595 kbase_mem_jit_trim_pages_from_region(struct kbase_context *kctx, struct kbase_va_region *reg, size_t pages_needed, size_t *freed, bool shrink) kbase_mem_jit_trim_pages_from_region() argument 3725 kbase_mem_jit_trim_pages(struct kbase_context *kctx, size_t pages_needed) kbase_mem_jit_trim_pages() argument 3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument 3882 trace_jit_stats(struct kbase_context *kctx, u32 bin_id, u32 max_allocations) trace_jit_stats() argument 3917 get_jit_phys_backing(struct kbase_context *kctx) get_jit_phys_backing() argument 3931 kbase_jit_trim_necessary_pages(struct kbase_context *kctx, size_t needed_pages) kbase_jit_trim_necessary_pages() argument 3984 jit_allow_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) jit_allow_allocate() argument 4070 kbase_jit_allocate(struct kbase_context *kctx, const struct base_jit_alloc_info *info, bool ignore_pressure_limit) kbase_jit_allocate() argument 4292 kbase_jit_free(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jit_free() argument 4353 struct kbase_context *kctx = kbase_reg_flags_to_kctx(reg); kbase_jit_backing_lost() local 4377 kbase_jit_evict(struct kbase_context *kctx) kbase_jit_evict() argument 4401 kbase_jit_term(struct kbase_context *kctx) kbase_jit_term() argument 4446 kbase_trace_jit_report_gpu_mem_trace_enabled(struct kbase_context *kctx, struct kbase_va_region *reg, unsigned int flags) kbase_trace_jit_report_gpu_mem_trace_enabled() argument 4491 kbase_jit_report_update_pressure(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_used_pages, unsigned int flags) kbase_jit_report_update_pressure() argument 4554 kbase_jd_user_buf_pin_pages(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_pin_pages() argument 4628 kbase_jd_user_buf_map(struct kbase_context *kctx, struct kbase_va_region *reg) kbase_jd_user_buf_map() argument 4711 kbase_jd_user_buf_unmap(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc, bool writeable) kbase_jd_user_buf_unmap() argument 4781 kbase_map_external_resource( struct kbase_context *kctx, struct kbase_va_region *reg, struct mm_struct *locked_mm) kbase_map_external_resource() argument 4822 kbase_unmap_external_resource(struct kbase_context *kctx, struct kbase_va_region *reg, struct kbase_mem_phy_alloc *alloc) kbase_unmap_external_resource() argument 4858 kbase_sticky_resource_acquire( struct kbase_context *kctx, u64 gpu_addr) kbase_sticky_resource_acquire() argument 4917 find_sticky_resource_meta(struct kbase_context *kctx, u64 gpu_addr) find_sticky_resource_meta() argument 4934 release_sticky_resource_meta(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta) release_sticky_resource_meta() argument 4949 kbase_sticky_resource_release(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release() argument 4970 kbase_sticky_resource_release_force(struct kbase_context *kctx, struct kbase_ctx_ext_res_meta *meta, u64 gpu_addr) kbase_sticky_resource_release_force() argument 4988 kbase_sticky_resource_init(struct kbase_context *kctx) kbase_sticky_resource_init() argument 4995 kbase_sticky_resource_term(struct kbase_context *kctx) kbase_sticky_resource_term() argument [all...] |
| /third_party/wpa_supplicant/wpa_supplicant-2.9/src/crypto/ |
| H A D | crypto_openssl.c | 1974 EVP_PKEY_CTX *kctx = NULL; in crypto_ecdh_init() local
|