Home
last modified time | relevance | path

Searched refs:mmu_sync_info (Results 1 - 15 of 15) sorted by relevance

/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.c145 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
150 enum kbase_caller_mmu_sync_info mmu_sync_info);
159 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
165 enum kbase_caller_mmu_sync_info mmu_sync_info);
312 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_gpu_mmu_handle_write_faulting_as() local
326 .mmu_sync_info = mmu_sync_info, in kbase_gpu_mmu_handle_write_faulting_as()
684 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_mmu_page_fault_worker() local
877 .mmu_sync_info = mmu_sync_info, in kbase_mmu_page_fault_worker()
1368 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_single_page() argument
1697 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_pages() argument
1740 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_flush_invalidate_noretain() local
1777 kbase_mmu_flush_invalidate_as(struct kbase_device *kbdev, struct kbase_as *as, u64 vpfn, size_t nr, bool sync, u32 kctx_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_as() argument
1850 kbase_mmu_flush_invalidate_no_ctx(struct kbase_device *kbdev, u64 vpfn, size_t nr, bool sync, int as_nr, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_no_ctx() argument
1863 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate() argument
2013 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_teardown_pages() local
2284 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_mmu_update_pages() local
[all...]
H A Dmali_kbase_mmu.h139 enum kbase_caller_mmu_sync_info mmu_sync_info);
143 enum kbase_caller_mmu_sync_info mmu_sync_info);
H A Dmali_kbase_mmu_hw.h83 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
90 enum kbase_caller_mmu_sync_info mmu_sync_info; member
H A Dmali_kbase_mmu_hw_direct.c287 if (op_param->mmu_sync_info == CALLER_MMU_SYNC) in kbase_mmu_hw_do_operation()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.h48 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
55 enum kbase_caller_mmu_sync_info mmu_sync_info);
174 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
186 enum kbase_caller_mmu_sync_info mmu_sync_info);
H A Dmali_kbase_mem_linux.c297 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_mem_alloc()
474 mmu_sync_info) != 0) { in kbase_mem_alloc()
835 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_mem_evictable_unmake() local
868 gpu_alloc->evicted, 0, mmu_sync_info); in kbase_mem_evictable_unmake()
1231 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_mem_umm_map() local
1263 alloc->group_id, mmu_sync_info); in kbase_mem_umm_map()
1280 KBASE_MEM_GROUP_SINK, mmu_sync_info); in kbase_mem_umm_map()
1737 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; local
1913 mmu_sync_info) != 0) {
1961 const enum kbase_caller_mmu_sync_info mmu_sync_info local
295 kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extension, u64 *flags, u64 *gpu_va, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mem_alloc() argument
2096 kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages, enum kbase_caller_mmu_sync_info mmu_sync_info) global() argument
2170 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; global() local
2686 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; global() local
[all...]
H A Dmali_kbase_dummy_job_wa.c287 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_dummy_job_wa_load() local
384 &gpu_va, mmu_sync_info); in kbase_dummy_job_wa_load()
H A Dmali_kbase_mem.c1504 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_gpu_mmap()
1550 group_id, mmu_sync_info); in kbase_gpu_mmap()
1563 group_id, mmu_sync_info); in kbase_gpu_mmap()
1575 group_id, mmu_sync_info); in kbase_gpu_mmap()
1600 KBASE_MEM_GROUP_SINK, mmu_sync_info); in kbase_gpu_mmap()
3766 enum kbase_caller_mmu_sync_info mmu_sync_info) in kbase_jit_grow()
3863 old_size, mmu_sync_info); in kbase_jit_grow()
4081 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in kbase_jit_allocate() local
4176 mmu_sync_info); in kbase_jit_allocate()
4241 mmu_sync_info); in kbase_jit_allocate()
1502 kbase_gpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, u64 addr, size_t nr_pages, size_t align, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_gpu_mmap() argument
3762 kbase_jit_grow(struct kbase_context *kctx, const struct base_jit_alloc_info *info, struct kbase_va_region *reg, struct kbase_sub_alloc **prealloc_sas, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_jit_grow() argument
4646 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; kbase_jd_user_buf_map() local
[all...]
H A Dmali_kbase_hwcnt_backend_jm.c651 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbasep_hwcnt_backend_jm_dump_alloc() local
666 mmu_sync_info); in kbasep_hwcnt_backend_jm_dump_alloc()
H A Dmali_kbase_hwcnt_backend_csf_if_fw.c314 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc() local
365 mmu_sync_info); in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc()
H A Dmali_kbase_mem.h1191 * @mmu_sync_info: Indicates whether this call is synchronous wrt MMU ops.
1197 enum kbase_caller_mmu_sync_info mmu_sync_info);
H A Dmali_kbase_core_linux.c830 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_api_mem_alloc() local
866 mmu_sync_info); in kbase_api_mem_alloc()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_heap_context_alloc.c165 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in kbase_csf_heap_context_allocator_alloc() local
179 &ctx_alloc->gpu_va, mmu_sync_info); in kbase_csf_heap_context_allocator_alloc()
H A Dmali_kbase_csf_tiler_heap.c190 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_SYNC; in create_chunk() local
208 &chunk->gpu_va, mmu_sync_info); in create_chunk()
H A Dmali_kbase_csf.c160 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in gpu_mmap_user_io_pages() local
188 KBASE_MEM_GROUP_CSF_IO, mmu_sync_info); in gpu_mmap_user_io_pages()
197 mmu_sync_info); in gpu_mmap_user_io_pages()
1092 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in create_normal_suspend_buffer() local
1133 KBASE_MEM_GROUP_CSF_FW, mmu_sync_info); in create_normal_suspend_buffer()
1182 const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; in create_protected_suspend_buffer() local
1218 KBASE_MEM_GROUP_CSF_FW, mmu_sync_info); in create_protected_suspend_buffer()

Completed in 29 milliseconds