/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu.c | 145 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 150 enum kbase_caller_mmu_sync_info mmu_sync_info); 159 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 165 enum kbase_caller_mmu_sync_info mmu_sync_info); 312 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_gpu_mmu_handle_write_faulting_as() local 326 .mmu_sync_info = mmu_sync_info, in kbase_gpu_mmu_handle_write_faulting_as() 684 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_mmu_page_fault_worker() local 877 .mmu_sync_info = mmu_sync_info, in kbase_mmu_page_fault_worker() 1368 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_single_page() argument 1697 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_pages() argument 1740 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_flush_invalidate_noretain() local 1777 kbase_mmu_flush_invalidate_as(struct kbase_device *kbdev, struct kbase_as *as, u64 vpfn, size_t nr, bool sync, u32 kctx_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_as() argument 1850 kbase_mmu_flush_invalidate_no_ctx(struct kbase_device *kbdev, u64 vpfn, size_t nr, bool sync, int as_nr, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_no_ctx() argument 1863 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate() argument 2013 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_teardown_pages() local 2284 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_update_pages() local [all...] |
H A D | mali_kbase_mmu.h | 139 enum kbase_caller_mmu_sync_info mmu_sync_info); 143 enum kbase_caller_mmu_sync_info mmu_sync_info);
|
H A D | mali_kbase_mmu_hw.h | 83 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 90 enum kbase_caller_mmu_sync_info mmu_sync_info; member
|
H A D | mali_kbase_mmu_hw_direct.c | 287 if (op_param->mmu_sync_info == CALLER_MMU_SYNC) in kbase_mmu_hw_do_operation()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_linux.h | 48 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 55 enum kbase_caller_mmu_sync_info mmu_sync_info); 174 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 186 enum kbase_caller_mmu_sync_info mmu_sync_info);
|
H A D | mali_kbase_mem_linux.c | 297 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_mem_alloc() 474 mmu_sync_info) != 0) { in kbase_mem_alloc() 835 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_mem_evictable_unmake() local 868 gpu_alloc->evicted, 0, mmu_sync_info); in kbase_mem_evictable_unmake() 1231 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_mem_umm_map() local 1263 alloc->group_id, mmu_sync_info); in kbase_mem_umm_map() 1280 KBASE_MEM_GROUP_SINK, mmu_sync_info); in kbase_mem_umm_map() 1737 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; local 1913 mmu_sync_info) != 0) { 1961 const enum kbase_caller_mmu_sync_info mmu_sync_info local 295 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extension, u64 *flags, u64 *gpu_va, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mem_alloc() argument 2096 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages, enum kbase_caller_mmu_sync_info mmu_sync_info) global() argument 2170 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; global() local 2686 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; global() local [all...] |
H A D | mali_kbase_dummy_job_wa.c | 287 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_dummy_job_wa_load() local 384 &gpu_va, mmu_sync_info); in kbase_dummy_job_wa_load()
|
H A D | mali_kbase_mem.c | 1504 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_gpu_mmap() 1550 group_id, mmu_sync_info); in kbase_gpu_mmap() 1563 group_id, mmu_sync_info); in kbase_gpu_mmap() 1575 group_id, mmu_sync_info); in kbase_gpu_mmap() 1600 KBASE_MEM_GROUP_SINK, mmu_sync_info); in kbase_gpu_mmap() 3766 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_jit_grow() 3863 old_size, mmu_sync_info); in kbase_jit_grow() 4081 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_jit_allocate() local 4176 mmu_sync_info); in kbase_jit_allocate() 4241 mmu_sync_info); in kbase_jit_allocate() 1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument 3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument 4646 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_jd_user_buf_map() local [all...] |
H A D | mali_kbase_hwcnt_backend_jm.c | 651 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbasep_hwcnt_backend_jm_dump_alloc() local 666 mmu_sync_info); in kbasep_hwcnt_backend_jm_dump_alloc()
|
H A D | mali_kbase_hwcnt_backend_csf_if_fw.c | 314 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc() local 365 mmu_sync_info); in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc()
|
H A D | mali_kbase_mem.h | 1191 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops. 1197 enum kbase_caller_mmu_sync_info mmu_sync_info);
|
H A D | mali_kbase_core_linux.c | 830 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_api_mem_alloc() local 866 mmu_sync_info); in kbase_api_mem_alloc()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf_heap_context_alloc.c | 165 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_csf_heap_context_allocator_alloc() local 179 &ctx_alloc->gpu_va, mmu_sync_info); in kbase_csf_heap_context_allocator_alloc()
|
H A D | mali_kbase_csf_tiler_heap.c | 190 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in create_chunk() local 208 &chunk->gpu_va, mmu_sync_info); in create_chunk()
|
H A D | mali_kbase_csf.c | 160 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in gpu_mmap_user_io_pages() local 188 KBASE_MEM_GROUP_CSF_IO, mmu_sync_info); in gpu_mmap_user_io_pages() 197 mmu_sync_info); in gpu_mmap_user_io_pages() 1092 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in create_normal_suspend_buffer() local 1133 KBASE_MEM_GROUP_CSF_FW, mmu_sync_info); in create_normal_suspend_buffer() 1182 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in create_protected_suspend_buffer() local 1218 KBASE_MEM_GROUP_CSF_FW, mmu_sync_info); in create_protected_suspend_buffer()
|