Home
last modified time | relevance | path

Searched refs:page_count (Results 1 - 25 of 35) sorted by relevance

12

/device/soc/hisilicon/hi3751v350/sdk_linux/source/common/drv/osal/
H A Dosal_addr.c224 unsigned int page_count; in osal_blockmem_vmap() local
233 page_count = (size + PAGE_SIZE - 1) / PAGE_SIZE; in osal_blockmem_vmap()
234 pages = vmalloc(page_count * sizeof(struct page *)); in osal_blockmem_vmap()
240 ret = memset_s(pages, page_count * sizeof(struct page *), 0, page_count * sizeof(struct page *)); in osal_blockmem_vmap()
245 for (i = 0; i < page_count; i++) { in osal_blockmem_vmap()
249 vaddr = vmap(pages, page_count, VM_MAP, PAGE_KERNEL); in osal_blockmem_vmap()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory_os_alloc.c123 if (page_count(page) == 1) { in mali_mem_os_put_page()
133 u32 page_count) in mali_mem_os_resize_pages()
141 if (mem_from->count < start_page + page_count) { in mali_mem_os_resize_pages()
147 if (i >= start_page && i < start_page + page_count) { in mali_mem_os_resize_pages()
162 size_t page_count = PAGE_ALIGN(size) / MALI_OSK_MALI_PAGE_SIZE; in mali_mem_os_alloc_pages() local
163 size_t remaining = page_count; in mali_mem_os_alloc_pages()
179 os_mem->count = page_count; in mali_mem_os_alloc_pages()
230 os_mem->count = (page_count - remaining) + i; in mali_mem_os_alloc_pages()
245 os_mem->count = (page_count - remaining) + i; in mali_mem_os_alloc_pages()
262 os_mem->count = (page_count in mali_mem_os_alloc_pages()
132 mali_mem_os_resize_pages(mali_mem_os_mem *mem_from, mali_mem_os_mem *mem_to, u32 start_page, u32 page_count) mali_mem_os_resize_pages() argument
[all...]
H A Dmali_memory_block_alloc.c146 size_t page_count = PAGE_ALIGN(size) / MALI_OSK_MALI_PAGE_SIZE; in mali_mem_block_alloc() local
156 if (atomic_read(&info->free_num) > page_count) { in mali_mem_block_alloc()
159 if (page_count > 0) { in mali_mem_block_alloc()
169 page_count--; in mali_mem_block_alloc()
H A Dmali_memory_swap_alloc.c523 size_t page_count = PAGE_ALIGN(size) / PAGE_SIZE; in mali_mem_swap_alloc_pages() local
531 MALI_DEBUG_ASSERT(page_count <= MALI_SWAP_GLOBAL_SWAP_FILE_INDEX_RESERVE); in mali_mem_swap_alloc_pages()
543 swap_mem->count = page_count; in mali_mem_swap_alloc_pages()
544 index = mali_mem_swap_idx_range_alloc(page_count); in mali_mem_swap_alloc_pages()
550 for (i = 0; i < page_count; i++) { in mali_mem_swap_alloc_pages()
557 mali_mem_swap_idx_range_free(index + i, page_count - i); in mali_mem_swap_alloc_pages()
567 mali_mem_swap_idx_range_free(index + i + 1, page_count - i - 1); in mali_mem_swap_alloc_pages()
H A Dmali_memory_os_alloc.h45 u32 page_count);
H A Dmali_memory_manager.c286 u32 dec_size, page_count; in mali_mem_resize() local
294 page_count = dec_size / MALI_MMU_PAGE_SIZE; in mali_mem_resize()
299 mali_mem_os_resize_pages(&mem_backend->os_mem, &tmp_os_mem, physical_size / MALI_MMU_PAGE_SIZE, page_count); in mali_mem_resize()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory_os_alloc.c118 if (1 == page_count(page)) { in mali_mem_os_put_page()
128 _mali_osk_errcode_t mali_mem_os_resize_pages(mali_mem_os_mem *mem_from, mali_mem_os_mem *mem_to, u32 start_page, u32 page_count) in mali_mem_os_resize_pages() argument
136 if (mem_from->count < start_page + page_count) { in mali_mem_os_resize_pages()
141 if (i >= start_page && i < start_page + page_count) { in mali_mem_os_resize_pages()
157 size_t page_count = PAGE_ALIGN(size) / _MALI_OSK_MALI_PAGE_SIZE; in mali_mem_os_alloc_pages() local
158 size_t remaining = page_count; in mali_mem_os_alloc_pages()
173 os_mem->count = page_count; in mali_mem_os_alloc_pages()
224 os_mem->count = (page_count - remaining) + i; in mali_mem_os_alloc_pages()
243 os_mem->count = (page_count - remaining) + i; in mali_mem_os_alloc_pages()
260 os_mem->count = (page_count in mali_mem_os_alloc_pages()
[all...]
H A Dmali_memory_block_alloc.c148 size_t page_count = PAGE_ALIGN(size) / _MALI_OSK_MALI_PAGE_SIZE; in mali_mem_block_alloc() local
158 if (atomic_read(&info->free_num) > page_count) { in mali_mem_block_alloc()
160 if (page_count > 0) { in mali_mem_block_alloc()
170 page_count--; in mali_mem_block_alloc()
H A Dmali_memory_swap_alloc.c524 size_t page_count = PAGE_ALIGN(size) / PAGE_SIZE; in mali_mem_swap_alloc_pages() local
532 MALI_DEBUG_ASSERT(page_count <= MALI_SWAP_GLOBAL_SWAP_FILE_INDEX_RESERVE); in mali_mem_swap_alloc_pages()
543 swap_mem->count = page_count; in mali_mem_swap_alloc_pages()
544 index = mali_mem_swap_idx_range_alloc(page_count); in mali_mem_swap_alloc_pages()
551 for (i = 0; i < page_count; i++) { in mali_mem_swap_alloc_pages()
559 mali_mem_swap_idx_range_free(index + i, page_count - i); in mali_mem_swap_alloc_pages()
570 mali_mem_swap_idx_range_free(index + i + 1, page_count - i - 1); in mali_mem_swap_alloc_pages()
H A Dmali_memory_os_alloc.h44 _mali_osk_errcode_t mali_mem_os_resize_pages(mali_mem_os_mem *mem_from, mali_mem_os_mem *mem_to, u32 start_page, u32 page_count);
/device/soc/rockchip/common/vendor/drivers/firmware/
H A Drockchip_sip.c131 unsigned int page_count; in sip_map() local
141 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); in sip_map()
145 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL); in sip_map()
147 pr_err("%s: Failed to allocate array for %u pages\n", __func__, page_count); in sip_map()
151 for (i = 0; i < page_count; i++) { in sip_map()
155 vaddr = vmap(pages, page_count, VM_MAP, prot); in sip_map()
/device/soc/rockchip/rk3588/kernel/drivers/firmware/
H A Drockchip_sip.c133 unsigned int page_count; in sip_map() local
142 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); in sip_map()
146 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL); in sip_map()
149 __func__, page_count); in sip_map()
153 for (i = 0; i < page_count; i++) in sip_map()
156 vaddr = vmap(pages, page_count, VM_MAP, prot); in sip_map()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/common/
H A Dmali_mmu_page_directory.c123 int i, page_count; in mali_mmu_pagedir_map() local
153 page_count = (start_address + MALI_MMU_VIRTUAL_PAGE_SIZE - mali_address) / MALI_MMU_PAGE_SIZE; in mali_mmu_pagedir_map()
154 pagedir->page_entries_usage_count[i] += page_count; in mali_mmu_pagedir_map()
157 page_count = (mali_address + size - start_address) / MALI_MMU_PAGE_SIZE; in mali_mmu_pagedir_map()
158 pagedir->page_entries_usage_count[i] += page_count; in mali_mmu_pagedir_map()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/common/
H A Dmali_mmu_page_directory.c122 int i, page_count; in mali_mmu_pagedir_map() local
152 page_count = (start_address + MALI_MMU_VIRTUAL_PAGE_SIZE - mali_address) / MALI_MMU_PAGE_SIZE; in mali_mmu_pagedir_map()
153 pagedir->page_entries_usage_count[i] += page_count; in mali_mmu_pagedir_map()
156 page_count = (mali_address + size - start_address) / MALI_MMU_PAGE_SIZE; in mali_mmu_pagedir_map()
157 pagedir->page_entries_usage_count[i] += page_count; in mali_mmu_pagedir_map()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem_linux.c2187 size_t page_count = PFN_UP(offset + size); local
2198 /* check if page_count calculation will wrap */
2210 /* check if page_index + page_count will wrap */
2211 if (-1UL - page_count < page_index)
2214 if (page_index + page_count > kbase_reg_current_backed_size(reg))
2229 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL);
2233 for (i = 0; i < page_count; i++)
2245 cpu_addr = vmap(pages, page_count, VM_MAP, prot);
2278 for (i = 1; page_count > 2 && i < page_count
2328 size_t page_count = PFN_UP(offset + size); global() local
[all...]
H A Dmali_kbase_tlstream.h148 void __kbase_tlstream_aux_pagesalloc(u32 ctx_nr, u64 page_count);
565 * @page_count: number of pages used by the context
567 #define KBASE_TLSTREAM_AUX_PAGESALLOC(ctx_nr, page_count) \
568 __TRACE_IF_ENABLED(aux_pagesalloc, ctx_nr, page_count)
H A Dmali_kbase_mem.c1134 u64 page_off, page_count; in kbase_do_syncset() local
1167 page_count = (size + offset + (PAGE_SIZE - 1)) >> PAGE_SHIFT; in kbase_do_syncset()
1172 page_off + page_count > reg->nr_pages) { in kbase_do_syncset()
1187 for (i = 1; page_count > 2 && i < page_count - 1; i++) { in kbase_do_syncset()
1197 if (page_count > 1 && cpu_pa[page_off + page_count - 1]) { in kbase_do_syncset()
1200 kbase_sync_single(kctx, cpu_pa[page_off + page_count - 1], in kbase_do_syncset()
1201 gpu_pa[page_off + page_count - 1], 0, sz, in kbase_do_syncset()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_tlstream.h136 void __kbase_tlstream_aux_pagesalloc(u32 ctx_nr, u64 page_count);
525 * @page_count: number of pages used by the context
527 #define KBASE_TLSTREAM_AUX_PAGESALLOC(ctx_nr, page_count) __TRACE_IF_ENABLED(aux_pagesalloc, ctx_nr, page_count)
H A Dmali_kbase_mem_linux.c2179 size_t page_count = PFN_UP(offset + size); local
2191 /* check if page_count calculation will wrap */
2205 /* check if page_index + page_count will wrap */
2206 if (-1UL - page_count < page_index) {
2210 if (page_index + page_count > kbase_reg_current_backed_size(reg)) {
2229 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL);
2234 for (i = 0; i < page_count; i++) {
2247 cpu_addr = vmap(pages, page_count, VM_MAP, prot);
2281 for (i = 1; page_count > KBASE_CPU_PAGE && i < page_count
2329 size_t page_count = PFN_UP(offset + size); global() local
[all...]
H A Dmali_kbase_mem.c1101 u64 page_off, page_count; in kbase_do_syncset() local
1133 page_count = (size + offset + (PAGE_SIZE - 1)) >> PAGE_SHIFT; in kbase_do_syncset()
1137 if (page_off > reg->nr_pages || page_off + page_count > reg->nr_pages) { in kbase_do_syncset()
1151 for (i = 1; page_count > 0x2 && i < page_count - 1; i++) { in kbase_do_syncset()
1161 if (page_count > 1 && cpu_pa[page_off + page_count - 1]) { in kbase_do_syncset()
1164 kbase_sync_single(kctx, cpu_pa[page_off + page_count - 1], gpu_pa[page_off + page_count - 1], 0, sz, sync_fn); in kbase_do_syncset()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c2955 size_t const page_count = PFN_UP(offset + map->size); local
2965 for (i = 1; page_count > 0x2 && i < page_count - 1; i++) {
2972 if (page_count > 1) {
2973 cpu_pa = map->cpu_pages[page_count - 1];
2974 gpu_pa = map->gpu_pages[page_count - 1];
2986 size_t page_count = PFN_UP(offset_in_page + size); local
2997 /* check if page_count calculation will wrap */
3004 /* check if page_index + page_count will wrap */
3005 if (-1UL - page_count < page_inde
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c2931 size_t const page_count = PFN_UP(offset + map->size); local
2941 for (i = 1; page_count > 2 && i < page_count - 1; i++) {
2948 if (page_count > 1) {
2949 cpu_pa = map->cpu_pages[page_count - 1];
2950 gpu_pa = map->gpu_pages[page_count - 1];
2962 size_t page_count = PFN_UP(offset_in_page + size); local
2972 /* check if page_count calculation will wrap */
2978 /* check if page_index + page_count will wrap */
2979 if (-1UL - page_count < page_inde
[all...]
H A Dmali_kbase_hwcnt_backend_jm.c735 size_t page_count; in kbasep_hwcnt_backend_jm_create() local
778 page_count = PFN_UP(info->dump_bytes); in kbasep_hwcnt_backend_jm_create()
781 page_count); in kbasep_hwcnt_backend_jm_create()
/device/soc/rockchip/rk3588/kernel/drivers/video/rockchip/rga3/
H A Drga2_mmu_info.c526 unsigned long page_count, in rga2_mmu_set_channel_internal()
540 rga2_sgt_to_page_table(sgt, mmu_base, page_count, false); in rga2_mmu_set_channel_internal()
542 page_count = (page_count + 15) & (~15); in rga2_mmu_set_channel_internal()
543 rga2_sgt_to_page_table(sgt, mmu_base, page_count, true); in rga2_mmu_set_channel_internal()
546 return page_count; in rga2_mmu_set_channel_internal()
523 rga2_mmu_set_channel_internal(struct rga_scheduler_t *scheduler, struct rga_internal_buffer *internal_buffer, uint32_t *mmu_base, unsigned long page_count, uint32_t **virt_flush_base, uint32_t *virt_flush_count, int map_flag) rga2_mmu_set_channel_internal() argument
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/backend/gpu/
H A Dmali_kbase_model_dummy.h160 size_t page_count);

Completed in 35 milliseconds

12