/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_pool_group.c | 47 struct kbase_mem_pool_group *const mem_pools, in kbase_mem_pool_group_init() 55 err = kbase_mem_pool_init(&mem_pools->small[gid], in kbase_mem_pool_group_init() 63 err = kbase_mem_pool_init(&mem_pools->large[gid], in kbase_mem_pool_group_init() 70 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init() 86 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init() 87 kbase_mem_pool_term(&mem_pools->large[gid]); in kbase_mem_pool_group_init() 95 struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_mark_dying() 100 kbase_mem_pool_mark_dying(&mem_pools->small[gid]); in kbase_mem_pool_group_mark_dying() 101 kbase_mem_pool_mark_dying(&mem_pools->large[gid]); in kbase_mem_pool_group_mark_dying() 106 struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_term() 46 kbase_mem_pool_group_init( struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev, const struct kbase_mem_pool_group_config *const configs, struct kbase_mem_pool_group *next_pools) kbase_mem_pool_group_init() argument 94 kbase_mem_pool_group_mark_dying( struct kbase_mem_pool_group *const mem_pools) kbase_mem_pool_group_mark_dying() argument 105 kbase_mem_pool_group_term( struct kbase_mem_pool_group *const mem_pools) kbase_mem_pool_group_term() argument [all...] |
H A D | mali_kbase_mem_pool_debugfs.c | 31 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_trim() local 33 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_trim() 37 kbase_mem_pool_trim(&mem_pools[index], value); in kbase_mem_pool_debugfs_trim() 43 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_set_max_size() local 45 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_set_max_size() 49 kbase_mem_pool_set_max_size(&mem_pools[index], value); in kbase_mem_pool_debugfs_set_max_size() 54 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_size() local 56 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_size() 60 return kbase_mem_pool_size(&mem_pools[index]); in kbase_mem_pool_debugfs_size() 65 struct kbase_mem_pool *const mem_pools in kbase_mem_pool_debugfs_max_size() local [all...] |
H A D | mali_kbase_mem_pool_group.h | 51 * empty pool in @mem_pools will attempt to allocate from the equivalent pool 54 * is full in @mem_pools. Pages are zeroed before they spill over to another 57 * @mem_pools: Set of memory pools to initialize 61 * is no free memory in one of the @mem_pools 65 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *mem_pools, 77 * @mem_pools: Set of memory pools to mark 79 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *mem_pools); 87 * @mem_pools: Set of memory pools to terminate 89 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *mem_pools);
|
H A D | mali_kbase_hwcnt_backend_csf_if_fw.c | 345 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], num_pages, in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc() 394 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], num_pages, in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc() 511 &fw_ctx->kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbasep_hwcnt_backend_csf_if_fw_ring_buf_free()
|
H A D | mali_kbase_core_linux.c | 3443 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_size() 3457 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_size() 3475 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_max_size() 3489 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_max_size() 3517 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_size() 3543 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_size() 3571 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_max_size() 3596 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_max_size() 3627 kbdev->mem_pools.small, 1, kbase_mem_pool_debugfs_max_size); in show_simplified_mem_pool_max_size() 3660 kbdev->mem_pools in set_simplified_mem_pool_max_size() [all...] |
H A D | mali_kbase_mem.c | 1295 err = kbase_mem_pool_group_init(&kbdev->mem_pools, kbdev, in kbase_mem_init() 1320 kbase_mem_pool_group_term(&kbdev->mem_pools); in kbase_mem_term() 2249 &kctx->mem_pools.large[alloc->group_id], in kbase_alloc_phy_pages_helper() 2299 &kctx->mem_pools.large[ in kbase_alloc_phy_pages_helper() 2305 &kctx->mem_pools.large[alloc->group_id], in kbase_alloc_phy_pages_helper() 2319 &kctx->mem_pools.large[ in kbase_alloc_phy_pages_helper() 2352 &kctx->mem_pools.small[alloc->group_id], in kbase_alloc_phy_pages_helper() 2608 &kctx->mem_pools.large[group_id], in free_partial() 2660 &kctx->mem_pools.large[alloc->group_id], in kbase_free_phy_pages_helper() 2684 &kctx->mem_pools in kbase_free_phy_pages_helper() [all...] |
H A D | mali_kbase_defs.h | 711 * @mem_pools: Global pools of free physical memory pages which can 999 struct kbase_mem_pool_group mem_pools; member 1537 * @mem_pools: Context-specific pools of free physical memory pages. 1777 struct kbase_mem_pool_group mem_pools; member
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_pool_group.c | 42 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev, in kbase_mem_pool_group_init() argument 49 err = kbase_mem_pool_init(&mem_pools->small[gid], &configs->small[gid], KBASE_MEM_POOL_4KB_PAGE_TABLE_ORDER, in kbase_mem_pool_group_init() 52 err = kbase_mem_pool_init(&mem_pools->large[gid], &configs->large[gid], KBASE_MEM_POOL_2MB_PAGE_TABLE_ORDER, in kbase_mem_pool_group_init() 55 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init() 73 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init() 74 kbase_mem_pool_term(&mem_pools->large[gid]); in kbase_mem_pool_group_init() 81 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_mark_dying() argument 86 kbase_mem_pool_mark_dying(&mem_pools->small[gid]); in kbase_mem_pool_group_mark_dying() 87 kbase_mem_pool_mark_dying(&mem_pools->large[gid]); in kbase_mem_pool_group_mark_dying() 91 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_term() argument [all...] |
H A D | mali_kbase_mem_pool_debugfs.c | 31 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_trim() local 33 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_trim() 37 kbase_mem_pool_trim(&mem_pools[index], value); in kbase_mem_pool_debugfs_trim() 42 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_set_max_size() local 44 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_set_max_size() 48 kbase_mem_pool_set_max_size(&mem_pools[index], value); in kbase_mem_pool_debugfs_set_max_size() 53 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_size() local 55 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_size() 59 return kbase_mem_pool_size(&mem_pools[index]); in kbase_mem_pool_debugfs_size() 64 struct kbase_mem_pool *const mem_pools in kbase_mem_pool_debugfs_max_size() local [all...] |
H A D | mali_kbase_mem_pool_group.h | 51 * empty pool in @mem_pools will attempt to allocate from the equivalent pool 54 * is full in @mem_pools. Pages are zeroed before they spill over to another 57 * @mem_pools: Set of memory pools to initialize 61 * is no free memory in one of the @mem_pools 65 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev, 76 * @mem_pools: Set of memory pools to mark 78 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *const mem_pools); 86 * @mem_pools: Set of memory pools to terminate 88 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *const mem_pools);
|
H A D | mali_kbase_defs.h | 674 * @mem_pools: Global pools of free physical memory pages which can 914 struct kbase_mem_pool_group mem_pools; member 1360 * @mem_pools: Context-specific pools of free physical memory pages. 1600 struct kbase_mem_pool_group mem_pools; member
|
H A D | mali_kbase_mem.c | 1029 err = kbase_mem_pool_group_init(&kbdev->mem_pools, kbdev, &mem_pool_defaults, NULL); in kbase_mem_init() 1054 kbase_mem_pool_group_term(&kbdev->mem_pools); in kbase_mem_term() 1925 res = kbase_mem_pool_alloc_pages(&kctx->mem_pools.large[alloc->group_id], nr_lp * (SZ_2M / SZ_4K), tp, true); in kbase_alloc_phy_pages_helper() 1966 np = kbase_mem_pool_alloc(&kctx->mem_pools.large[alloc->group_id]); in kbase_alloc_phy_pages_helper() 1971 err = kbase_mem_pool_grow(&kctx->mem_pools.large[alloc->group_id], 1); in kbase_alloc_phy_pages_helper() 1984 kbase_mem_pool_free(&kctx->mem_pools.large[alloc->group_id], np, false); in kbase_alloc_phy_pages_helper() 2015 res = kbase_mem_pool_alloc_pages(&kctx->mem_pools.small[alloc->group_id], nr_left, tp, false); in kbase_alloc_phy_pages_helper() 2243 kbase_mem_pool_free(&kctx->mem_pools.large[group_id], head_page, true); in free_partial() 2287 kbase_mem_pool_free_pages(&kctx->mem_pools.large[alloc->group_id], 0x200, start_free, syncback, reclaimed); in kbase_free_phy_pages_helper() 2304 kbase_mem_pool_free_pages(&kctx->mem_pools in kbase_free_phy_pages_helper() [all...] |
H A D | mali_kbase_core_linux.c | 3316 buf, PAGE_SIZE, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_size() 3332 buf, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_size() 3352 buf, PAGE_SIZE, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_max_size() 3368 buf, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_max_size() 3399 buf, PAGE_SIZE, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_size() 3428 buf, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_size() 3459 buf, PAGE_SIZE, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_max_size() 3487 buf, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_max_size()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/ |
H A D | mali_kbase_context.c | 262 return kbase_mem_pool_group_init(&kctx->mem_pools, kctx->kbdev, &kctx->kbdev->mem_pool_defaults, in kbase_context_mem_pool_group_init() 263 &kctx->kbdev->mem_pools); in kbase_context_mem_pool_group_init() 268 kbase_mem_pool_group_term(&kctx->mem_pools); in kbase_context_mem_pool_group_term() 287 p = kbase_mem_alloc_page(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK]); in kbase_context_mem_alloc_page() 300 kbase_mem_pool_free(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK], as_page(kctx->aliasing_sink_page), false); in kbase_context_mem_pool_free()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/ |
H A D | mali_kbase_context.c | 269 return kbase_mem_pool_group_init(&kctx->mem_pools, in kbase_context_mem_pool_group_init() 272 &kctx->kbdev->mem_pools); in kbase_context_mem_pool_group_init() 277 kbase_mem_pool_group_term(&kctx->mem_pools); in kbase_context_mem_pool_group_term() 296 p = kbase_mem_alloc_page(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK]); in kbase_context_mem_alloc_page() 309 &kctx->mem_pools.small[KBASE_MEM_GROUP_SINK], in kbase_context_mem_pool_free()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu.c | 492 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc() 496 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc() 1077 &kctx->mem_pools.large[ in kbase_mmu_page_fault_worker() 1089 &kctx->mem_pools.small[ in kbase_mmu_page_fault_worker() 1140 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd() 1177 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], p, false); in kbase_mmu_alloc_pgd() 1426 &kbdev->mem_pools.small[ in kbase_mmu_insert_single_page() 1514 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], in kbase_mmu_free_pgd() 1600 &kbdev->mem_pools.small[mmut->group_id], in kbase_mmu_insert_pages_no_flush() 2217 &kbdev->mem_pools in kbase_mmu_update_pages_no_flush() [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu.c | 349 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc() 353 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc() 852 struct kbase_mem_pool *const lp_mem_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker() 859 struct kbase_mem_pool *const mem_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker() 909 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd() 945 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], p, false); in kbase_mmu_alloc_pgd() 1158 err = kbase_mem_pool_grow(&kbdev->mem_pools.small[kctx->mmu.group_id], MIDGARD_MMU_BOTTOMLEVEL); in kbase_mmu_insert_single_page() 1225 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], tmp_p, false); in cleanup_empty_pte() 1304 err = kbase_mem_pool_grow(&kbdev->mem_pools.small[mmut->group_id], cur_level); in kbase_mmu_insert_pages_no_flush() 1749 err = kbase_mem_pool_grow(&kbdev->mem_pools in kbase_mmu_update_pages_no_flush() [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/backend/ |
H A D | mali_kbase_context_csf.c | 196 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()
|
H A D | mali_kbase_context_jm.c | 259 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf.c | 309 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_free_command_stream_user_pages() 337 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_alloc_command_stream_user_pages() 379 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_alloc_command_stream_user_pages() 1112 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in create_normal_suspend_buffer() 1148 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], nr_pages, in create_normal_suspend_buffer() 1450 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in term_normal_suspend_buffer() 2903 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_doorbell_mapping_term() 2921 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_doorbell_mapping_init() 2942 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], page, in kbase_csf_free_dummy_user_reg_page() 2957 &kbdev->mem_pools in kbase_csf_setup_dummy_user_reg_page() [all...] |
H A D | mali_kbase_csf_firmware.c | 558 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in parse_memory_setup_entry() 669 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in parse_memory_setup_entry() 2000 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_term() 2351 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init() 2402 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init() 2433 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_term()
|
H A D | mali_kbase_csf_firmware_no_mali.c | 1348 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init() 1399 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init() 1430 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_term()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/backend/ |
H A D | mali_kbase_context_jm.c | 213 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()
|