Home
last modified time | relevance | path

Searched refs:mem_pools (Results 1 - 23 of 23) sorted by relevance

/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_pool_group.c47 struct kbase_mem_pool_group *const mem_pools, in kbase_mem_pool_group_init()
55 err = kbase_mem_pool_init(&mem_pools->small[gid], in kbase_mem_pool_group_init()
63 err = kbase_mem_pool_init(&mem_pools->large[gid], in kbase_mem_pool_group_init()
70 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init()
86 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init()
87 kbase_mem_pool_term(&mem_pools->large[gid]); in kbase_mem_pool_group_init()
95 struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_mark_dying()
100 kbase_mem_pool_mark_dying(&mem_pools->small[gid]); in kbase_mem_pool_group_mark_dying()
101 kbase_mem_pool_mark_dying(&mem_pools->large[gid]); in kbase_mem_pool_group_mark_dying()
106 struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_term()
46 kbase_mem_pool_group_init( struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev, const struct kbase_mem_pool_group_config *const configs, struct kbase_mem_pool_group *next_pools) kbase_mem_pool_group_init() argument
94 kbase_mem_pool_group_mark_dying( struct kbase_mem_pool_group *const mem_pools) kbase_mem_pool_group_mark_dying() argument
105 kbase_mem_pool_group_term( struct kbase_mem_pool_group *const mem_pools) kbase_mem_pool_group_term() argument
[all...]
H A Dmali_kbase_mem_pool_debugfs.c31 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_trim() local
33 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_trim()
37 kbase_mem_pool_trim(&mem_pools[index], value); in kbase_mem_pool_debugfs_trim()
43 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_set_max_size() local
45 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_set_max_size()
49 kbase_mem_pool_set_max_size(&mem_pools[index], value); in kbase_mem_pool_debugfs_set_max_size()
54 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_size() local
56 if (WARN_ON(!mem_pools) || in kbase_mem_pool_debugfs_size()
60 return kbase_mem_pool_size(&mem_pools[index]); in kbase_mem_pool_debugfs_size()
65 struct kbase_mem_pool *const mem_pools in kbase_mem_pool_debugfs_max_size() local
[all...]
H A Dmali_kbase_mem_pool_group.h51 * empty pool in @mem_pools will attempt to allocate from the equivalent pool
54 * is full in @mem_pools. Pages are zeroed before they spill over to another
57 * @mem_pools: Set of memory pools to initialize
61 * is no free memory in one of the @mem_pools
65 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *mem_pools,
77 * @mem_pools: Set of memory pools to mark
79 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *mem_pools);
87 * @mem_pools: Set of memory pools to terminate
89 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *mem_pools);
H A Dmali_kbase_hwcnt_backend_csf_if_fw.c345 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], num_pages, in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc()
394 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], num_pages, in kbasep_hwcnt_backend_csf_if_fw_ring_buf_alloc()
511 &fw_ctx->kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbasep_hwcnt_backend_csf_if_fw_ring_buf_free()
H A Dmali_kbase_core_linux.c3443 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_size()
3457 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_size()
3475 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_max_size()
3489 kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_max_size()
3517 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_size()
3543 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_size()
3571 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_max_size()
3596 kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_max_size()
3627 kbdev->mem_pools.small, 1, kbase_mem_pool_debugfs_max_size); in show_simplified_mem_pool_max_size()
3660 kbdev->mem_pools in set_simplified_mem_pool_max_size()
[all...]
H A Dmali_kbase_mem.c1295 err = kbase_mem_pool_group_init(&kbdev->mem_pools, kbdev, in kbase_mem_init()
1320 kbase_mem_pool_group_term(&kbdev->mem_pools); in kbase_mem_term()
2249 &kctx->mem_pools.large[alloc->group_id], in kbase_alloc_phy_pages_helper()
2299 &kctx->mem_pools.large[ in kbase_alloc_phy_pages_helper()
2305 &kctx->mem_pools.large[alloc->group_id], in kbase_alloc_phy_pages_helper()
2319 &kctx->mem_pools.large[ in kbase_alloc_phy_pages_helper()
2352 &kctx->mem_pools.small[alloc->group_id], in kbase_alloc_phy_pages_helper()
2608 &kctx->mem_pools.large[group_id], in free_partial()
2660 &kctx->mem_pools.large[alloc->group_id], in kbase_free_phy_pages_helper()
2684 &kctx->mem_pools in kbase_free_phy_pages_helper()
[all...]
H A Dmali_kbase_defs.h711 * @mem_pools: Global pools of free physical memory pages which can
999 struct kbase_mem_pool_group mem_pools; member
1537 * @mem_pools: Context-specific pools of free physical memory pages.
1777 struct kbase_mem_pool_group mem_pools; member
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_pool_group.c42 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev, in kbase_mem_pool_group_init() argument
49 err = kbase_mem_pool_init(&mem_pools->small[gid], &configs->small[gid], KBASE_MEM_POOL_4KB_PAGE_TABLE_ORDER, in kbase_mem_pool_group_init()
52 err = kbase_mem_pool_init(&mem_pools->large[gid], &configs->large[gid], KBASE_MEM_POOL_2MB_PAGE_TABLE_ORDER, in kbase_mem_pool_group_init()
55 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init()
73 kbase_mem_pool_term(&mem_pools->small[gid]); in kbase_mem_pool_group_init()
74 kbase_mem_pool_term(&mem_pools->large[gid]); in kbase_mem_pool_group_init()
81 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_mark_dying() argument
86 kbase_mem_pool_mark_dying(&mem_pools->small[gid]); in kbase_mem_pool_group_mark_dying()
87 kbase_mem_pool_mark_dying(&mem_pools->large[gid]); in kbase_mem_pool_group_mark_dying()
91 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *const mem_pools) in kbase_mem_pool_group_term() argument
[all...]
H A Dmali_kbase_mem_pool_debugfs.c31 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_trim() local
33 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_trim()
37 kbase_mem_pool_trim(&mem_pools[index], value); in kbase_mem_pool_debugfs_trim()
42 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_set_max_size() local
44 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_set_max_size()
48 kbase_mem_pool_set_max_size(&mem_pools[index], value); in kbase_mem_pool_debugfs_set_max_size()
53 struct kbase_mem_pool *const mem_pools = array; in kbase_mem_pool_debugfs_size() local
55 if (WARN_ON(!mem_pools) || WARN_ON(index >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_debugfs_size()
59 return kbase_mem_pool_size(&mem_pools[index]); in kbase_mem_pool_debugfs_size()
64 struct kbase_mem_pool *const mem_pools in kbase_mem_pool_debugfs_max_size() local
[all...]
H A Dmali_kbase_mem_pool_group.h51 * empty pool in @mem_pools will attempt to allocate from the equivalent pool
54 * is full in @mem_pools. Pages are zeroed before they spill over to another
57 * @mem_pools: Set of memory pools to initialize
61 * is no free memory in one of the @mem_pools
65 int kbase_mem_pool_group_init(struct kbase_mem_pool_group *const mem_pools, struct kbase_device *const kbdev,
76 * @mem_pools: Set of memory pools to mark
78 void kbase_mem_pool_group_mark_dying(struct kbase_mem_pool_group *const mem_pools);
86 * @mem_pools: Set of memory pools to terminate
88 void kbase_mem_pool_group_term(struct kbase_mem_pool_group *const mem_pools);
H A Dmali_kbase_defs.h674 * @mem_pools: Global pools of free physical memory pages which can
914 struct kbase_mem_pool_group mem_pools; member
1360 * @mem_pools: Context-specific pools of free physical memory pages.
1600 struct kbase_mem_pool_group mem_pools; member
H A Dmali_kbase_mem.c1029 err = kbase_mem_pool_group_init(&kbdev->mem_pools, kbdev, &mem_pool_defaults, NULL); in kbase_mem_init()
1054 kbase_mem_pool_group_term(&kbdev->mem_pools); in kbase_mem_term()
1925 res = kbase_mem_pool_alloc_pages(&kctx->mem_pools.large[alloc->group_id], nr_lp * (SZ_2M / SZ_4K), tp, true); in kbase_alloc_phy_pages_helper()
1966 np = kbase_mem_pool_alloc(&kctx->mem_pools.large[alloc->group_id]); in kbase_alloc_phy_pages_helper()
1971 err = kbase_mem_pool_grow(&kctx->mem_pools.large[alloc->group_id], 1); in kbase_alloc_phy_pages_helper()
1984 kbase_mem_pool_free(&kctx->mem_pools.large[alloc->group_id], np, false); in kbase_alloc_phy_pages_helper()
2015 res = kbase_mem_pool_alloc_pages(&kctx->mem_pools.small[alloc->group_id], nr_left, tp, false); in kbase_alloc_phy_pages_helper()
2243 kbase_mem_pool_free(&kctx->mem_pools.large[group_id], head_page, true); in free_partial()
2287 kbase_mem_pool_free_pages(&kctx->mem_pools.large[alloc->group_id], 0x200, start_free, syncback, reclaimed); in kbase_free_phy_pages_helper()
2304 kbase_mem_pool_free_pages(&kctx->mem_pools in kbase_free_phy_pages_helper()
[all...]
H A Dmali_kbase_core_linux.c3316 buf, PAGE_SIZE, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_size()
3332 buf, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_size()
3352 buf, PAGE_SIZE, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_mem_pool_max_size()
3368 buf, kbdev->mem_pools.small, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_mem_pool_max_size()
3399 buf, PAGE_SIZE, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_size()
3428 buf, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_size()
3459 buf, PAGE_SIZE, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in show_lp_mem_pool_max_size()
3487 buf, kbdev->mem_pools.large, MEMORY_GROUP_MANAGER_NR_GROUPS, in set_lp_mem_pool_max_size()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/
H A Dmali_kbase_context.c262 return kbase_mem_pool_group_init(&kctx->mem_pools, kctx->kbdev, &kctx->kbdev->mem_pool_defaults, in kbase_context_mem_pool_group_init()
263 &kctx->kbdev->mem_pools); in kbase_context_mem_pool_group_init()
268 kbase_mem_pool_group_term(&kctx->mem_pools); in kbase_context_mem_pool_group_term()
287 p = kbase_mem_alloc_page(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK]); in kbase_context_mem_alloc_page()
300 kbase_mem_pool_free(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK], as_page(kctx->aliasing_sink_page), false); in kbase_context_mem_pool_free()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/
H A Dmali_kbase_context.c269 return kbase_mem_pool_group_init(&kctx->mem_pools, in kbase_context_mem_pool_group_init()
272 &kctx->kbdev->mem_pools); in kbase_context_mem_pool_group_init()
277 kbase_mem_pool_group_term(&kctx->mem_pools); in kbase_context_mem_pool_group_term()
296 p = kbase_mem_alloc_page(&kctx->mem_pools.small[KBASE_MEM_GROUP_SINK]); in kbase_context_mem_alloc_page()
309 &kctx->mem_pools.small[KBASE_MEM_GROUP_SINK], in kbase_context_mem_pool_free()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.c492 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc()
496 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc()
1077 &kctx->mem_pools.large[ in kbase_mmu_page_fault_worker()
1089 &kctx->mem_pools.small[ in kbase_mmu_page_fault_worker()
1140 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd()
1177 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], p, false); in kbase_mmu_alloc_pgd()
1426 &kbdev->mem_pools.small[ in kbase_mmu_insert_single_page()
1514 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], in kbase_mmu_free_pgd()
1600 &kbdev->mem_pools.small[mmut->group_id], in kbase_mmu_insert_pages_no_flush()
2217 &kbdev->mem_pools in kbase_mmu_update_pages_no_flush()
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.c349 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc()
353 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc()
852 struct kbase_mem_pool *const lp_mem_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
859 struct kbase_mem_pool *const mem_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
909 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd()
945 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], p, false); in kbase_mmu_alloc_pgd()
1158 err = kbase_mem_pool_grow(&kbdev->mem_pools.small[kctx->mmu.group_id], MIDGARD_MMU_BOTTOMLEVEL); in kbase_mmu_insert_single_page()
1225 kbase_mem_pool_free(&kbdev->mem_pools.small[mmut->group_id], tmp_p, false); in cleanup_empty_pte()
1304 err = kbase_mem_pool_grow(&kbdev->mem_pools.small[mmut->group_id], cur_level); in kbase_mmu_insert_pages_no_flush()
1749 err = kbase_mem_pool_grow(&kbdev->mem_pools in kbase_mmu_update_pages_no_flush()
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/context/backend/
H A Dmali_kbase_context_csf.c196 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()
H A Dmali_kbase_context_jm.c259 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf.c309 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_free_command_stream_user_pages()
337 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_alloc_command_stream_user_pages()
379 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_IO], in kbase_csf_alloc_command_stream_user_pages()
1112 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in create_normal_suspend_buffer()
1148 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], nr_pages, in create_normal_suspend_buffer()
1450 &kctx->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in term_normal_suspend_buffer()
2903 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_doorbell_mapping_term()
2921 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_doorbell_mapping_init()
2942 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], page, in kbase_csf_free_dummy_user_reg_page()
2957 &kbdev->mem_pools in kbase_csf_setup_dummy_user_reg_page()
[all...]
H A Dmali_kbase_csf_firmware.c558 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in parse_memory_setup_entry()
669 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in parse_memory_setup_entry()
2000 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_term()
2351 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init()
2402 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init()
2433 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_term()
H A Dmali_kbase_csf_firmware_no_mali.c1348 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init()
1399 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_init()
1430 &kbdev->mem_pools.small[KBASE_MEM_GROUP_CSF_FW], in kbase_csf_firmware_mcu_shared_mapping_term()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/context/backend/
H A Dmali_kbase_context_jm.c213 kbase_mem_pool_group_mark_dying(&kctx->mem_pools); in kbase_destroy_context()

Completed in 45 milliseconds