/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu_hw.h | 86 u64 vpfn; member
|
H A D | mali_kbase_mmu_mode_aarch64.c | 192 static void entry_set_pte(u64 *pgd, u64 vpfn, phys_addr_t phy) in entry_set_pte() argument
|
H A D | mali_kbase_mmu.c | 1185 mmu_get_next_pgd(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, phys_addr_t *pgd, u64 vpfn, int level) mmu_get_next_pgd() argument 1237 mmu_get_pgd_at_level(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, int level, phys_addr_t *out_pgd) mmu_get_pgd_at_level() argument 1265 mmu_get_bottom_pgd(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, phys_addr_t *out_pgd) mmu_get_bottom_pgd() argument 1279 u64 vpfn = from_vpfn; mmu_insert_pages_failure_recovery() local 1368 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_single_page() argument 1697 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_pages() argument 1730 kbase_mmu_flush_invalidate_noretain(struct kbase_context *kctx, u64 vpfn, size_t nr) kbase_mmu_flush_invalidate_noretain() argument 1777 kbase_mmu_flush_invalidate_as(struct kbase_device *kbdev, struct kbase_as *as, u64 vpfn, size_t nr, bool sync, u32 kctx_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_as() argument 1850 kbase_mmu_flush_invalidate_no_ctx(struct kbase_device *kbdev, u64 vpfn, size_t nr, bool sync, int as_nr, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate_no_ctx() argument 1863 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_flush_invalidate() argument 1948 kbase_mmu_update_and_free_parent_pgds(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, phys_addr_t *pgds, u64 vpfn, int level) kbase_mmu_update_and_free_parent_pgds() argument 2001 kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, size_t nr, int as_nr) kbase_mmu_teardown_pages() argument 2174 kbase_mmu_update_pages_no_flush(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages_no_flush() argument 2275 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu_hw_direct.c | 173 int kbase_mmu_hw_do_operation(struct kbase_device *kbdev, struct kbase_as *as, u64 vpfn, u32 nr, u32 op, in kbase_mmu_hw_do_operation() argument
|
H A D | mali_kbase_mmu.c | 953 static int mmu_get_next_pgd(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, phys_addr_t *pgd, u64 vpfn, in mmu_get_next_pgd() argument 1002 mmu_get_pgd_at_level(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, int level, phys_addr_t *out_pgd) mmu_get_pgd_at_level() argument 1025 mmu_get_bottom_pgd(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, phys_addr_t *out_pgd) mmu_get_bottom_pgd() argument 1034 u64 vpfn = from_vpfn; mmu_insert_pages_failure_recovery() local 1101 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_insert_single_page() argument 1379 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id) kbase_mmu_insert_pages() argument 1408 kbase_mmu_flush_invalidate_noretain(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate_noretain() argument 1440 kbase_mmu_flush_invalidate_as(struct kbase_device *kbdev, struct kbase_as *as, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate_as() argument 1478 kbase_mmu_flush_invalidate_no_ctx(struct kbase_device *kbdev, u64 vpfn, size_t nr, bool sync, int as_nr) kbase_mmu_flush_invalidate_no_ctx() argument 1486 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate() argument 1568 kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, size_t nr, int as_nr) kbase_mmu_teardown_pages() argument 1707 kbase_mmu_update_pages_no_flush(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages_no_flush() argument 1786 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/backend/gpu/ |
H A D | mali_kbase_mmu_hw_direct.c | 273 int kbase_mmu_hw_do_operation(struct kbase_device *kbdev, struct kbase_as *as, struct kbase_context *kctx, u64 vpfn, in kbase_mmu_hw_do_operation() argument
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/backend/gpu/ |
H A D | mali_kbase_mmu_hw_direct.c | 302 kbase_mmu_hw_do_operation(struct kbase_device *kbdev, struct kbase_as *as, struct kbase_context *kctx, u64 vpfn, u32 nr, u32 op, unsigned int handling_irq) kbase_mmu_hw_do_operation() argument
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mmu.c | 516 static phys_addr_t mmu_insert_pages_recover_get_next_pgd(struct kbase_context *kctx, phys_addr_t pgd, u64 vpfn, in argument 445 mmu_get_next_pgd(struct kbase_context *kctx, phys_addr_t *pgd, u64 vpfn, int level) mmu_get_next_pgd() argument 493 mmu_get_bottom_pgd(struct kbase_context *kctx, u64 vpfn, phys_addr_t *out_pgd) mmu_get_bottom_pgd() argument 545 mmu_insert_pages_recover_get_bottom_pgd(struct kbase_context *kctx, u64 vpfn) mmu_insert_pages_recover_get_bottom_pgd() argument 563 mmu_insert_pages_failure_recovery(struct kbase_context *kctx, u64 vpfn, size_t nr) mmu_insert_pages_failure_recovery() argument 613 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, phys_addr_t phys, size_t nr, unsigned long flags) kbase_mmu_insert_single_page() argument 725 kbase_mmu_insert_pages_no_flush(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_insert_pages_no_flush() argument 840 kbase_mmu_insert_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_insert_pages() argument 864 kbase_mmu_flush_invalidate_noretain(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate_noretain() argument 907 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate() argument 1042 kbase_mmu_teardown_pages(struct kbase_context *kctx, u64 vpfn, size_t nr) kbase_mmu_teardown_pages() argument 1125 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_update_pages() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mmu.c | 410 static int mmu_get_next_pgd(struct kbase_context *kctx, phys_addr_t *pgd, u64 vpfn, int level)
in mmu_get_next_pgd() argument 456 static int mmu_get_bottom_pgd(struct kbase_context *kctx, u64 vpfn, phys_addr_t *out_pgd)
in mmu_get_bottom_pgd() argument 478 static phys_addr_t mmu_insert_pages_recover_get_next_pgd(struct kbase_context *kctx, phys_addr_t pgd, u64 vpfn,
in mmu_insert_pages_recover_get_next_pgd() argument 508 mmu_insert_pages_recover_get_bottom_pgd(struct kbase_context *kctx, u64 vpfn) mmu_insert_pages_recover_get_bottom_pgd() argument 526 mmu_insert_pages_failure_recovery(struct kbase_context *kctx, u64 vpfn, size_t nr) mmu_insert_pages_failure_recovery() argument 577 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, phys_addr_t phys, size_t nr, unsigned long flags) kbase_mmu_insert_single_page() argument 682 kbase_mmu_insert_pages_no_flush(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_insert_pages_no_flush() argument 791 kbase_mmu_insert_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_insert_pages() argument 813 kbase_mmu_flush_invalidate_noretain(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate_noretain() argument 856 kbase_mmu_flush_invalidate(struct kbase_context *kctx, u64 vpfn, size_t nr, bool sync) kbase_mmu_flush_invalidate() argument 989 kbase_mmu_teardown_pages(struct kbase_context *kctx, u64 vpfn, size_t nr) kbase_mmu_teardown_pages() argument 1072 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags) kbase_mmu_update_pages() argument [all...] |