Home
last modified time | relevance | path

Searched refs:group_id (Results 1 - 25 of 32) sorted by relevance

12

/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_native_mgm.c33 * @group_id: A physical memory group ID, which must be valid but is not used.
45 struct memory_group_manager_device *mgm_dev, int group_id, in kbase_native_mgm_alloc()
62 CSTD_UNUSED(group_id); in kbase_native_mgm_alloc()
71 * @group_id: A physical memory group ID, which must be valid but is not used.
82 int group_id, struct page *page, unsigned int order) in kbase_native_mgm_free()
85 CSTD_UNUSED(group_id); in kbase_native_mgm_free()
94 * @group_id: A physical memory group ID, which must be valid but is not used.
108 struct memory_group_manager_device *mgm_dev, int group_id, in kbase_native_mgm_vmf_insert_pfn_prot()
113 CSTD_UNUSED(group_id); in kbase_native_mgm_vmf_insert_pfn_prot()
123 * @group_id
44 kbase_native_mgm_alloc( struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, unsigned int order) kbase_native_mgm_alloc() argument
81 kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, unsigned int order) kbase_native_mgm_free() argument
107 kbase_native_mgm_vmf_insert_pfn_prot( struct memory_group_manager_device *mgm_dev, int group_id, struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, pgprot_t pgprot) kbase_native_mgm_vmf_insert_pfn_prot() argument
133 kbase_native_mgm_update_gpu_pte(struct memory_group_manager_device *mgm_dev, int group_id, int mmu_level, u64 pte) kbase_native_mgm_update_gpu_pte() argument
[all...]
H A Dmali_kbase_mem_pool.c167 pool->group_id, gfp, pool->order); in kbase_mem_alloc_page()
176 pool->group_id, p, pool->order); in kbase_mem_alloc_page()
201 pool->group_id, p, pool->order); in kbase_mem_pool_free_page()
361 int group_id, in kbase_mem_pool_init()
365 if (WARN_ON(group_id < 0) || in kbase_mem_pool_init()
366 WARN_ON(group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_init()
373 pool->group_id = group_id; in kbase_mem_pool_init()
358 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument
H A Dmali_kbase_mem.h128 * @group_id: A memory group ID to be passed to a platform-specific
146 u8 group_id; member
639 enum kbase_memory_type type, int group_id) in kbase_alloc_create()
687 alloc->group_id = group_id; in kbase_alloc_create()
697 struct kbase_context *kctx, int group_id) in kbase_reg_prepare_native()
705 KBASE_MEM_TYPE_NATIVE, group_id); in kbase_reg_prepare_native()
715 KBASE_MEM_TYPE_NATIVE, group_id); in kbase_reg_prepare_native()
789 * @group_id: A memory group ID to be passed to a platform-specific
814 int group_id,
637 kbase_alloc_create( struct kbase_context *kctx, size_t nr_pages, enum kbase_memory_type type, int group_id) kbase_alloc_create() argument
696 kbase_reg_prepare_native(struct kbase_va_region *reg, struct kbase_context *kctx, int group_id) kbase_reg_prepare_native() argument
[all...]
H A Dmali_kbase_mem.c1017 int max_allocations, int trim_level, int group_id, in kbase_region_tracker_init_jit()
1025 if (group_id < 0 || group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS) in kbase_region_tracker_init_jit()
1068 kctx->jit_group_id = group_id; in kbase_region_tracker_init_jit()
1511 int group_id; in kbase_gpu_mmap() local
1533 group_id = alloc->group_id; in kbase_gpu_mmap()
1550 group_id, mmu_sync_info); in kbase_gpu_mmap()
1563 group_id, mmu_sync_info); in kbase_gpu_mmap()
1575 group_id, mmu_sync_inf in kbase_gpu_mmap()
1016 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
2594 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
[all...]
H A Dmali_kbase_mem_linux.c609 *out |= kbase_mem_group_id_set(reg->cpu_alloc->group_id); in kbase_mem_query()
1007 reg->gpu_alloc->group_id); in kbase_mem_flags_change()
1263 alloc->group_id, mmu_sync_info); in kbase_mem_umm_map()
1334 int group_id = BASE_MEM_GROUP_DEFAULT; in get_umm_memory_group_id() local
1343 group_id = kctx->kbdev->mgm_dev->ops.mgm_get_import_memory_id( in get_umm_memory_group_id()
1347 return group_id; in get_umm_memory_group_id()
1373 int group_id; in kbase_mem_from_umm() local
1441 group_id = get_umm_memory_group_id(kctx, dma_buf); in kbase_mem_from_umm()
1444 KBASE_MEM_TYPE_IMPORTED_UMM, group_id); in kbase_mem_from_umm()
2112 kctx->as_nr, reg->gpu_alloc->group_id,
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_native_mgm.c34 * @group_id: A physical memory group ID, which must be valid but is not used.
45 static struct page *kbase_native_mgm_alloc(struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, in kbase_native_mgm_alloc() argument
60 CSTD_UNUSED(group_id); in kbase_native_mgm_alloc()
69 * @group_id: A physical memory group ID, which must be valid but is not used.
79 static void kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, in kbase_native_mgm_free() argument
83 CSTD_UNUSED(group_id); in kbase_native_mgm_free()
92 * @group_id: A physical memory group ID, which must be valid but is not used.
105 static vm_fault_t kbase_native_mgm_vmf_insert_pfn_prot(struct memory_group_manager_device *mgm_dev, int group_id, in kbase_native_mgm_vmf_insert_pfn_prot() argument
110 CSTD_UNUSED(group_id); in kbase_native_mgm_vmf_insert_pfn_prot()
120 * @group_id
129 kbase_native_mgm_update_gpu_pte(struct memory_group_manager_device *mgm_dev, int group_id, int mmu_level, u64 pte) kbase_native_mgm_update_gpu_pte() argument
[all...]
H A Dmali_base_kernel.h737 * @group_id: Physical memory group ID. Range is 0..(BASE_MEM_GROUP_COUNT-1).
741 static inline base_context_create_flags base_context_mmu_group_id_set(int const group_id) in base_context_mmu_group_id_set() argument
743 LOCAL_ASSERT(group_id >= 0); in base_context_mmu_group_id_set()
744 LOCAL_ASSERT(group_id < BASE_MEM_GROUP_COUNT); in base_context_mmu_group_id_set()
745 return BASEP_CONTEXT_MMU_GROUP_ID_MASK & ((base_context_create_flags)group_id << BASEP_CONTEXT_MMU_GROUP_ID_SHIFT); in base_context_mmu_group_id_set()
H A Dmali_kbase_mem.h122 * @group_id: A memory group ID to be passed to a platform-specific
139 u8 group_id; member
559 enum kbase_memory_type type, int group_id) in kbase_alloc_create()
606 alloc->group_id = group_id; in kbase_alloc_create()
615 static inline int kbase_reg_prepare_native(struct kbase_va_region *reg, struct kbase_context *kctx, int group_id) in kbase_reg_prepare_native() argument
622 reg->cpu_alloc = kbase_alloc_create(kctx, reg->nr_pages, KBASE_MEM_TYPE_NATIVE, group_id); in kbase_reg_prepare_native()
631 reg->gpu_alloc = kbase_alloc_create(kctx, reg->nr_pages, KBASE_MEM_TYPE_NATIVE, group_id); in kbase_reg_prepare_native()
703 * @group_id: A memory group ID to be passed to a platform-specific
726 int group_id, struc
558 kbase_alloc_create(struct kbase_context *kctx, size_t nr_pages, enum kbase_memory_type type, int group_id) kbase_alloc_create() argument
[all...]
H A Dmali_kbase_mem_pool.c167 p = kbdev->mgm_dev->ops.mgm_alloc_page(kbdev->mgm_dev, pool->group_id, gfp, pool->order); in kbase_mem_alloc_page()
173 kbdev->mgm_dev->ops.mgm_free_page(kbdev->mgm_dev, pool->group_id, p, pool->order); in kbase_mem_alloc_page()
197 kbdev->mgm_dev->ops.mgm_free_page(kbdev->mgm_dev, pool->group_id, p, pool->order); in kbase_mem_pool_free_page()
362 int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) in kbase_mem_pool_init()
364 if (WARN_ON(group_id < 0) || WARN_ON(group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in kbase_mem_pool_init()
371 pool->group_id = group_id; in kbase_mem_pool_init()
361 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument
H A Dmali_kbase_mem.c814 int group_id, u64 phys_pages_limit) in kbase_region_tracker_init_jit()
822 if (group_id < 0 || group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS) { in kbase_region_tracker_init_jit()
852 kctx->jit_group_id = group_id; in kbase_region_tracker_init_jit()
1234 int group_id; in kbase_gpu_mmap() local
1258 group_id = alloc->group_id; in kbase_gpu_mmap()
1269 alloc->imported.alias.aliased[i].length, reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap()
1278 (reg->flags & mask & gwt_mask) | attr, group_id); in kbase_gpu_mmap()
1286 kbase_reg_current_backed_size(reg), reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap()
813 kbase_region_tracker_init_jit(struct kbase_context *kctx, u64 jit_va_pages, int max_allocations, int trim_level, int group_id, u64 phys_pages_limit) kbase_region_tracker_init_jit() argument
2231 free_partial(struct kbase_context *kctx, int group_id, struct tagged_addr tp) free_partial() argument
[all...]
H A Dmali_kbase_ioctl.h263 * @group_id: Group ID to be used for physical allocations
276 __u8 group_id; member
289 * @group_id: Group ID to be used for physical allocations
300 __u8 group_id; member
H A Dmali_kbase_mem_linux.c657 *out |= base_mem_group_id_set(reg->cpu_alloc->group_id); in kbase_mem_query()
1051 new_flags, reg->gpu_alloc->group_id); in kbase_mem_flags_change()
1311 alloc->group_id); in kbase_mem_umm_map()
1379 int group_id = BASE_MEM_GROUP_DEFAULT; in get_umm_memory_group_id() local
1387 group_id = kctx->kbdev->mgm_dev->ops.mgm_get_import_memory_id( in get_umm_memory_group_id()
1391 return group_id; in get_umm_memory_group_id()
1418 int group_id; in kbase_mem_from_umm() local
1490 group_id = get_umm_memory_group_id(kctx, dma_buf); in kbase_mem_from_umm()
1493 KBASE_MEM_TYPE_IMPORTED_UMM, group_id); in kbase_mem_from_umm()
2148 kctx->as_nr, reg->gpu_alloc->group_id);
[all...]
H A Dmali_kbase_defs.h252 * @group_id: A memory group ID to be passed to a platform-specific
263 u8 group_id; member
428 * @group_id: A memory group ID to be passed to a platform-specific
448 u8 group_id; member
H A Dmali_kbase_gwt.c38 reg->flags & flag, reg->gpu_alloc->group_id); in kbase_gpu_gwt_setup_page_permission()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.h59 * @group_id: The physical group ID from which to allocate GPU page tables.
64 int kbase_mmu_init(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, struct kbase_context *kctx, int group_id);
93 * @group_id: The physical memory group in which the page was allocated.
104 int group_id);
107 struct tagged_addr *phys, size_t nr, unsigned long flags, int group_id);
109 size_t nr, unsigned long flags, int as_nr, int group_id);
111 unsigned long flags, int group_id);
115 unsigned long flags, int const group_id);
H A Dmali_kbase_mmu.c113 unsigned long flags, int group_id);
270 region->flags, region->gpu_alloc->group_id); in kbase_gpu_mmu_handle_write_fault()
341 if (WARN_ON(region->gpu_alloc->group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS)) { in page_fault_try_alloc()
349 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc()
353 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc()
657 if (region->gpu_alloc->group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS) { in kbase_mmu_page_fault_worker()
761 region->flags, region->gpu_alloc->group_id); in kbase_mmu_page_fault_worker()
852 struct kbase_mem_pool *const lp_mem_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
859 struct kbase_mem_pool *const mem_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
909 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd()
1101 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_insert_single_page() argument
1239 kbase_mmu_create_ate(struct kbase_device *const kbdev, struct tagged_addr const phy, unsigned long const flags, int const level, int const group_id) kbase_mmu_create_ate() argument
1248 kbase_mmu_insert_pages_no_flush(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, const u64 start_vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_insert_pages_no_flush() argument
1379 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id) kbase_mmu_insert_pages() argument
1707 kbase_mmu_update_pages_no_flush(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages_no_flush() argument
1786 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages() argument
1847 kbase_mmu_init(struct kbase_device *const kbdev, struct kbase_mmu_table *const mmut, struct kbase_context *const kctx, int const group_id) kbase_mmu_init() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.h82 * @group_id: The physical group ID from which to allocate GPU page tables.
88 struct kbase_context *kctx, int group_id);
117 * @group_id: The physical memory group in which the page was allocated.
128 struct tagged_addr phy, unsigned long flags, int level, int group_id);
134 unsigned long flags, int group_id);
138 unsigned long flags, int as_nr, int group_id,
142 unsigned long flags, int group_id,
150 unsigned long flags, int const group_id);
H A Dmali_kbase_mmu.c197 unsigned long flags, int group_id);
408 1, region->flags, region->gpu_alloc->group_id); in kbase_gpu_mmu_handle_write_fault()
483 if (WARN_ON(region->gpu_alloc->group_id >= in page_fault_try_alloc()
492 root_pool = &kctx->mem_pools.large[region->gpu_alloc->group_id]; in page_fault_try_alloc()
496 root_pool = &kctx->mem_pools.small[region->gpu_alloc->group_id]; in page_fault_try_alloc()
816 if (region->gpu_alloc->group_id >= MEMORY_GROUP_MANAGER_NR_GROUPS) { in kbase_mmu_page_fault_worker()
959 new_pages, region->flags, region->gpu_alloc->group_id); in kbase_mmu_page_fault_worker()
1078 region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
1090 region->gpu_alloc->group_id]; in kbase_mmu_page_fault_worker()
1140 p = kbase_mem_pool_alloc(&kbdev->mem_pools.small[mmut->group_id]); in kbase_mmu_alloc_pgd()
1368 kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, size_t nr, unsigned long flags, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_single_page() argument
1530 kbase_mmu_create_ate(struct kbase_device *const kbdev, struct tagged_addr const phy, unsigned long const flags, int const level, int const group_id) kbase_mmu_create_ate() argument
1541 kbase_mmu_insert_pages_no_flush(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, const u64 start_vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_insert_pages_no_flush() argument
1697 kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr, int const group_id, enum kbase_caller_mmu_sync_info mmu_sync_info) kbase_mmu_insert_pages() argument
2174 kbase_mmu_update_pages_no_flush(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages_no_flush() argument
2275 kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys, size_t nr, unsigned long flags, int const group_id) kbase_mmu_update_pages() argument
2339 kbase_mmu_init(struct kbase_device *const kbdev, struct kbase_mmu_table *const mmut, struct kbase_context *const kctx, int const group_id) kbase_mmu_init() argument
[all...]
/device/soc/rockchip/common/sdk_linux/include/linux/
H A Dmemory_group_manager.h54 * @group_id: A physical memory group ID. The meaning of this is defined
64 struct page *(*mgm_alloc_page)(struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask,
72 * @group_id: A physical memory group ID. The meaning of this is
78 * @group_id and @order.
82 void (*mgm_free_page)(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page,
108 * @group_id: A physical memory group ID. The meaning of this is
124 u64 (*mgm_update_gpu_pte)(struct memory_group_manager_device *mgm_dev, int group_id, int mmu_level, u64 pte);
131 * @group_id: A physical memory group ID. The meaning of this is
147 vm_fault_t (*mgm_vmf_insert_pfn_prot)(struct memory_group_manager_device *mgm_dev, int group_id,
/device/soc/rockchip/rk3588/kernel/include/linux/
H A Dmemory_group_manager.h54 * @group_id: A physical memory group ID. The meaning of this is defined
65 struct memory_group_manager_device *mgm_dev, int group_id,
73 * @group_id: A physical memory group ID. The meaning of this is
79 * @group_id and @order.
84 struct memory_group_manager_device *mgm_dev, int group_id,
111 * @group_id: A physical memory group ID. The meaning of this is
128 int group_id, int mmu_level, u64 pte);
135 * @group_id: A physical memory group ID. The meaning of this is
152 struct memory_group_manager_device *mgm_dev, int group_id,
/device/soc/rockchip/rk3568/hardware/mpp/mpp/legacy/inc/
H A Dmpp_buffer_impl.h60 RK_U32 group_id; member
90 RK_U32 group_id; member
122 RK_U32 group_id; member
/device/soc/rockchip/rk3588/kernel/include/uapi/gpu/arm/bifrost/
H A Dmali_base_kernel.h717 #define BASE_CONTEXT_MMU_GROUP_ID_SET(group_id) \
719 ((base_context_create_flags)(group_id) \
H A Dmali_kbase_ioctl.h247 * @group_id: Group ID to be used for physical allocations
260 __u8 group_id; member
274 * @group_id: Group ID to be used for physical allocations
285 __u8 group_id; member
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_ioctl.h234 * @group_id: Group ID to be used for physical allocations.
251 __u8 group_id; member
/device/soc/rockchip/rk3588/kernel/include/uapi/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_ioctl.h346 * @in.group_id: Group ID to be used for physical allocations.
361 __u8 group_id; member

Completed in 35 milliseconds

12