/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/ump/linux/ |
H A D | ump_osk_low_level_mem.c | 38 static void ump_vma_open(struct vm_area_struct *vma); 39 static void ump_vma_close(struct vm_area_struct *vma); 41 static int ump_cpu_page_fault_handler(struct vm_area_struct *vma, struct vm_fault *vmf); 43 static unsigned long ump_cpu_page_fault_handler(struct vm_area_struct *vma, unsigned long address); 61 static int ump_cpu_page_fault_handler(struct vm_area_struct *vma, struct vm_fault *vmf) in ump_cpu_page_fault_handler() argument 63 static unsigned long ump_cpu_page_fault_handler(struct vm_area_struct *vma, unsigned long address) in ump_cpu_page_fault_handler() 71 MSG_ERR(("VMA: 0x%08lx, virtual address: 0x%08lx\n", (unsigned long)vma, address)); in ump_cpu_page_fault_handler() 80 static void ump_vma_open(struct vm_area_struct *vma) in ump_vma_open() argument 85 vma_usage_tracker = (ump_vma_usage_tracker *)vma->vm_private_data; in ump_vma_open() 90 DBG_MSG(4, ("VMA open, VMA reference count incremented. VMA: 0x%08lx, reference count: %d\n", (unsigned long)vma, new_va in ump_vma_open() 93 ump_vma_close(struct vm_area_struct *vma) ump_vma_close() argument 128 struct vm_area_struct *vma; _ump_osk_mem_mapregion_init() local 174 struct vm_area_struct *vma; _ump_osk_mem_mapregion_term() local 192 struct vm_area_struct *vma; _ump_osk_mem_mapregion_map() local [all...] |
H A D | ump_kernel_linux.c | 86 static int ump_file_mmap(struct file *filp, struct vm_area_struct *vma); 394 static int ump_file_mmap(struct file *filp, struct vm_area_struct *vma) in ump_file_mmap() argument 410 args.size = vma->vm_end - vma->vm_start; in ump_file_mmap() 411 args._ukk_private = vma; in ump_file_mmap() 412 args.secure_id = vma->vm_pgoff; in ump_file_mmap() 415 vma->vm_flags |= VM_DONTCOPY; in ump_file_mmap() 417 DBG_MSG(4, ("UMP vma->flags: %x\n", vma->vm_flags)); in ump_file_mmap()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory.c | 40 static void mali_mem_vma_open(struct vm_area_struct *vma) in mali_mem_vma_open() argument 42 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_open() 43 MALI_DEBUG_PRINT(MALI_KERNEL_LEVEL_CODE, ("Open called on vma %p\n", vma)); in mali_mem_vma_open() 49 static void mali_mem_vma_close(struct vm_area_struct *vma) in mali_mem_vma_close() argument 52 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_close() 55 vma->vm_private_data = NULL; in mali_mem_vma_close() 60 struct vm_area_struct *vma = vmf->vma; in mali_mem_vma_fault() local 61 mali_mem_allocation *alloc = (mali_mem_allocation *)vma in mali_mem_vma_fault() 150 mali_mmap(struct file *filp, struct vm_area_struct *vma) mali_mmap() argument [all...] |
H A D | mali_memory_cow.c | 365 if (alloc->cpu_mapping.vma != NULL) {
in mali_memory_cow_modify_range() 367 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma != NULL);
in mali_memory_cow_modify_range() 368 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size);
in mali_memory_cow_modify_range() 371 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size);
in mali_memory_cow_modify_range() 373 ret = mali_mem_cow_cpu_map_pages_locked(backend, alloc->cpu_mapping.vma,
in mali_memory_cow_modify_range() 374 alloc->cpu_mapping.vma->vm_start + range_start,
in mali_memory_cow_modify_range() 383 alloc->cpu_mapping.vma->vm_flags |= VM_PFNMAP;
in mali_memory_cow_modify_range() 384 alloc->cpu_mapping.vma in mali_memory_cow_modify_range() 515 mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma) mali_mem_cow_cpu_map() argument 546 mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long vaddr, int num) mali_mem_cow_cpu_map_pages_locked() argument [all...] |
H A D | mali_memory_cow.h | 22 int mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma);
23 mali_osk_errcode_t mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma,
|
H A D | mali_memory_os_alloc.h | 52 int mali_mem_os_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma); 54 mali_osk_errcode_t mali_mem_os_resize_cpu_map_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma,
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory.c | 42 static void mali_mem_vma_open(struct vm_area_struct *vma) in mali_mem_vma_open() argument 44 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_open() 45 MALI_DEBUG_PRINT(4, ("Open called on vma %p\n", vma)); in mali_mem_vma_open() 51 static void mali_mem_vma_close(struct vm_area_struct *vma) in mali_mem_vma_close() argument 54 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_close() 57 vma->vm_private_data = NULL; in mali_mem_vma_close() 62 struct vm_area_struct *vma = vmf->vma; in mali_mem_vma_fault() local 63 mali_mem_allocation *alloc = (mali_mem_allocation *)vma in mali_mem_vma_fault() 152 mali_mmap(struct file *filp, struct vm_area_struct *vma) mali_mmap() argument [all...] |
H A D | mali_memory_cow.c | 378 if (NULL != alloc->cpu_mapping.vma) { in mali_memory_cow_modify_range() 380 MALI_DEBUG_ASSERT(NULL != alloc->cpu_mapping.vma); in mali_memory_cow_modify_range() 381 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range() 384 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range() 386 ret = mali_mem_cow_cpu_map_pages_locked(backend, alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size / _MALI_OSK_MALI_PAGE_SIZE); in mali_memory_cow_modify_range() 394 alloc->cpu_mapping.vma->vm_flags |= VM_PFNMAP; in mali_memory_cow_modify_range() 395 alloc->cpu_mapping.vma in mali_memory_cow_modify_range() 522 mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma) mali_mem_cow_cpu_map() argument 553 mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long vaddr, int num) mali_mem_cow_cpu_map_pages_locked() argument [all...] |
H A D | mali_memory_cow.h | 18 int mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma); 20 struct vm_area_struct *vma,
|
H A D | mali_memory_os_alloc.h | 50 int mali_mem_os_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma); 52 _mali_osk_errcode_t mali_mem_os_resize_cpu_map_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long start_vaddr, u32 mappig_size);
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/thirdparty/ |
H A D | mali_kbase_mmap.c | 152 struct vm_area_struct *vma; in kbase_unmapped_area_topdown() local 188 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 189 if (vma->rb_subtree_gap < length) { in kbase_unmapped_area_topdown() 195 gap_start = vma->vm_prev ? vma->vm_prev->vm_end : 0; in kbase_unmapped_area_topdown() 196 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown() 197 struct vm_area_struct *right = rb_entry(vma->vm_rb.rb_right, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 199 vma = right; in kbase_unmapped_area_topdown() 206 gap_end = vma->vm_start; in kbase_unmapped_area_topdown() 223 if (vma in kbase_unmapped_area_topdown() [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/thirdparty/ |
H A D | mali_kbase_mmap.c | 136 struct vm_area_struct *vma; in kbase_unmapped_area_topdown() local 168 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 169 if (vma->rb_subtree_gap < length) in kbase_unmapped_area_topdown() 174 gap_start = vma->vm_prev ? vma->vm_prev->vm_end : 0; in kbase_unmapped_area_topdown() 175 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown() 177 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown() 180 vma = right; in kbase_unmapped_area_topdown() 187 gap_end = vma->vm_start; in kbase_unmapped_area_topdown() 203 if (vma in kbase_unmapped_area_topdown() [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_linux.c | 92 struct vm_area_struct *vma); 94 struct vm_area_struct *vma); 104 struct vm_area_struct *vma); 2358 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument 2360 struct kbase_cpu_mapping *map = vma->vm_private_data; 2368 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument 2370 struct kbase_cpu_mapping *map = vma->vm_private_data; 2405 static struct kbase_aliased *get_aliased_alloc(struct vm_area_struct *vma, argument 2436 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, argument 2442 struct vm_area_struct *vma 2516 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 2630 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 2707 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2775 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument 3211 kbase_special_vm_close(struct vm_area_struct *vma) global() argument 3223 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3271 kbase_csf_user_io_pages_vm_open(struct vm_area_struct *vma) global() argument 3277 kbase_csf_user_io_pages_vm_close(struct vm_area_struct *vma) global() argument 3297 kbase_csf_user_io_pages_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument 3389 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3462 kbase_csf_user_reg_vm_close(struct vm_area_struct *vma) global() argument 3472 kbase_csf_user_reg_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument 3501 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
H A D | mali_kbase_mem_linux.h | 148 * @vma: Pointer to the struct containing the info where the GPU allocation 153 int kbase_context_mmap(struct kbase_context *kctx, struct vm_area_struct *vma); 441 static inline vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, in vmf_insert_pfn_prot() argument 450 if (pgprot_val(pgprot) != pgprot_val(vma->vm_page_prot)) { in vmf_insert_pfn_prot() 454 err = vm_insert_pfn(vma, addr, pfn); in vmf_insert_pfn_prot() 456 err = vm_insert_pfn_prot(vma, addr, pfn, pgprot); in vmf_insert_pfn_prot()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_linux.c | 90 struct vm_area_struct *vma); 92 struct vm_area_struct *vma); 101 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma); 2328 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument 2330 struct kbase_cpu_mapping *map = vma->vm_private_data; 2338 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument 2340 struct kbase_cpu_mapping *map = vma->vm_private_data; 2373 static struct kbase_aliased *get_aliased_alloc(struct vm_area_struct *vma, argument 2405 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, argument 2411 struct vm_area_struct *vma 2484 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 2594 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 2673 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2752 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument 3176 kbase_special_vm_close(struct vm_area_struct *vma) global() argument 3188 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3230 kbase_csf_user_io_pages_vm_open(struct vm_area_struct *vma) global() argument 3236 kbase_csf_user_io_pages_vm_close(struct vm_area_struct *vma) global() argument 3271 kbase_csf_user_io_pages_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument 3362 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument 3434 kbase_csf_user_reg_vm_close(struct vm_area_struct *vma) global() argument 3444 kbase_csf_user_reg_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument 3493 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
H A D | mali_kbase_mem_linux.h | 145 * @vma: Pointer to the struct containing the info where the GPU allocation 150 int kbase_context_mmap(struct kbase_context *kctx, struct vm_area_struct *vma); 439 static inline vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, in vmf_insert_pfn_prot() argument 447 if (pgprot_val(pgprot) != pgprot_val(vma->vm_page_prot)) in vmf_insert_pfn_prot() 450 err = vm_insert_pfn(vma, addr, pfn); in vmf_insert_pfn_prot() 452 err = vm_insert_pfn_prot(vma, addr, pfn, pgprot); in vmf_insert_pfn_prot()
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | mm.h | 217 * per a vma. In ELF, the number of sections is represented in unsigned short.
303 #define VM_DONTCOPY 0x00020000 /* Do not copy this vma on fork */
321 #define VM_HUGEPAGE 0x20000000 /* MADV_HUGEPAGE marked this vma */
322 #define VM_NOHUGEPAGE 0x40000000 /* MADV_NOHUGEPAGE marked this vma */
532 * vm_fault is filled by the pagefault handler and passed to the vma's
533 * ->fault function. The vma's ->fault is responsible for returning a bitmask
542 struct vm_area_struct *vma; /* Target VMA */
member 545 pgoff_t pgoff; /* Logical page offset based on vma */
610 int (*access)(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write);
612 /* Called by the /proc/PID/maps code to ask the vma whethe 647 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) vma_init() argument 657 vma_set_anonymous(struct vm_area_struct *vma) vma_set_anonymous() argument 662 vma_is_anonymous(struct vm_area_struct *vma) vma_is_anonymous() argument 667 vma_is_temporary_stack(struct vm_area_struct *vma) vma_is_temporary_stack() argument 682 vma_is_foreign(struct vm_area_struct *vma) vma_is_foreign() argument 695 vma_is_accessible(struct vm_area_struct *vma) vma_is_accessible() argument 707 vma_is_shmem(struct vm_area_struct *vma) vma_is_shmem() argument 1008 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) maybe_mkwrite() argument 1738 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) handle_mm_fault() argument 2561 vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert) vma_adjust() argument 2688 struct vm_area_struct *vma = find_vma(mm, start_addr); find_vma_intersection() local 2696 vm_start_gap(struct vm_area_struct *vma) vm_start_gap() argument 2709 vm_end_gap(struct vm_area_struct *vma) vm_end_gap() argument 2722 vma_pages(struct vm_area_struct *vma) vma_pages() argument 2730 struct vm_area_struct *vma = find_vma(mm, vm_start); find_exact_vma() local 2739 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) range_in_vma() argument 2752 vma_set_page_prot(struct vm_area_struct *vma) vma_set_page_prot() argument 2775 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vmf_insert_page() argument 2788 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument 3133 vma_is_special_huge(const struct vm_area_struct *vma) vma_is_special_huge() argument 3210 seal_check_future_write(int seals, struct vm_area_struct *vma) seal_check_future_write() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem_linux.c | 49 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma); 1606 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument 1608 struct kbase_cpu_mapping *map = vma->vm_private_data; 1616 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument 1618 struct kbase_cpu_mapping *map = vma->vm_private_data; 1654 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) argument 1659 struct vm_area_struct *vma = vmf->vma; 1661 struct kbase_cpu_mapping *map = vma->vm_private_data; 1689 while (i < map->alloc->nents && (addr < vma 1713 kbase_cpu_mmap(struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 1813 kbase_trace_buffer_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kaddr) global() argument 1898 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 1970 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2038 kbase_mmap(struct file *file, struct vm_area_struct *vma) global() argument 2412 kbase_special_vm_close(struct vm_area_struct *vma) global() argument 2424 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
/device/soc/rockchip/common/sdk_linux/ipc/ |
H A D | shm.c | 91 static void shm_open(struct vm_area_struct *vma);
92 static void shm_close(struct vm_area_struct *vma);
271 static int __shm_open(struct vm_area_struct *vma)
in __shm_open() argument 273 struct file *file = vma->vm_file;
in __shm_open() 296 static void shm_open(struct vm_area_struct *vma)
in shm_open() argument 298 int err = __shm_open(vma);
in shm_open() 353 * remove the attach descriptor vma.
358 static void shm_close(struct vm_area_struct *vma)
in shm_close() argument 360 struct file *file = vma->vm_file;
in shm_close() 519 struct file *file = vmf->vma in shm_fault() 525 shm_split(struct vm_area_struct *vma, unsigned long addr) shm_split() argument 537 shm_pagesize(struct vm_area_struct *vma) shm_pagesize() argument 550 shm_set_policy(struct vm_area_struct *vma, struct mempolicy *new) shm_set_policy() argument 562 shm_get_policy(struct vm_area_struct *vma, unsigned long addr) shm_get_policy() argument 578 shm_mmap(struct file *file, struct vm_area_struct *vma) shm_mmap() argument 1750 struct vm_area_struct *vma; ksys_shmdt() local [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem_linux.c | 44 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma);
1610 static void kbase_cpu_vm_open(struct vm_area_struct *vma)
argument 1612 struct kbase_cpu_mapping *map = vma->vm_private_data;
1620 static void kbase_cpu_vm_close(struct vm_area_struct *vma)
argument 1622 struct kbase_cpu_mapping *map = vma->vm_private_data;
1658 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
argument 1663 struct vm_area_struct *vma = vmf->vma;
1665 struct kbase_cpu_mapping *map = vma->vm_private_data;
1695 while (i < map->alloc->nents && (addr < vma 1717 kbase_cpu_mmap(struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument 1816 kbase_trace_buffer_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kaddr) global() argument 1901 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument 1972 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument 2036 kbase_mmap(struct file *file, struct vm_area_struct *vma) global() argument 2410 kbase_special_vm_close(struct vm_area_struct *vma) global() argument 2422 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument [all...] |
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/osal/linux/mmz/ |
H A D | mmz_userdev.c | 464 error_mmz("mmb<%s> vma is invalid.\n", p->mmb->name); 469 error_mmz("mmb<%s> vma is invalid.\n", p->mmb->name); 886 int mmz_userdev_mmap(struct file *file, struct vm_area_struct *vma) argument 890 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 891 unsigned long size = vma->vm_end - vma->vm_start; 926 vma->vm_page_prot = __pgprot(pgprot_val(vma->vm_page_prot) 929 vma->vm_page_prot = pgprot_writecombine(vma [all...] |
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/ |
H A D | drm_gem.c | 1017 * drm_gem_vm_open - vma->ops->open implementation for GEM 1018 * @vma: VM area structure 1023 void drm_gem_vm_open(struct vm_area_struct *vma) in drm_gem_vm_open() argument 1025 struct drm_gem_object *obj = vma->vm_private_data; in drm_gem_vm_open() 1032 * drm_gem_vm_close - vma->ops->close implementation for GEM 1033 * @vma: VM area structure 1038 void drm_gem_vm_close(struct vm_area_struct *vma) in drm_gem_vm_close() argument 1040 struct drm_gem_object *obj = vma->vm_private_data; in drm_gem_vm_close() 1050 * @vma: VMA for the area to be mapped 1071 struct vm_area_struct *vma) in drm_gem_mmap_obj() 1070 drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size, struct vm_area_struct *vma) drm_gem_mmap_obj() argument 1131 drm_gem_mmap(struct file *filp, struct vm_area_struct *vma) drm_gem_mmap() argument [all...] |
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc.c | 229 u32 offset = intel_guc_ggtt_offset(guc, guc->log.vma) >> PAGE_SHIFT; in guc_ctl_log_params_flags() 503 * @rsa_offset: rsa offset w.r.t ggtt base of huc vma 655 struct i915_vma *vma; in intel_guc_allocate_vma() local 664 vma = i915_vma_instance(obj, >->ggtt->vm, NULL); in intel_guc_allocate_vma() 665 if (IS_ERR(vma)) { in intel_guc_allocate_vma() 669 flags = PIN_OFFSET_BIAS | i915_ggtt_pin_bias(vma); in intel_guc_allocate_vma() 670 ret = i915_ggtt_pin(vma, NULL, 0, flags); in intel_guc_allocate_vma() 672 vma = ERR_PTR(ret); in intel_guc_allocate_vma() 676 return i915_vma_make_unshrinkable(vma); in intel_guc_allocate_vma() 680 return vma; in intel_guc_allocate_vma() 697 struct i915_vma *vma; intel_guc_allocate_and_map_vma() local [all...] |
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/rockchip/ |
H A D | rockchip_drm_gem.c | 504 static int rockchip_drm_gem_object_mmap_iommu(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap_iommu() argument 508 unsigned long user_count = vma_pages(vma); in rockchip_drm_gem_object_mmap_iommu() 513 return vm_map_pages(vma, rk_obj->pages, count); in rockchip_drm_gem_object_mmap_iommu() 516 static int rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap_dma() argument 521 return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, obj->size, rk_obj->dma_attrs); in rockchip_drm_gem_object_mmap_dma() 524 static int rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap() argument 531 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in rockchip_drm_gem_object_mmap() 538 vma->vm_flags &= ~VM_PFNMAP; in rockchip_drm_gem_object_mmap() 544 ret = rockchip_drm_gem_object_mmap_iommu(obj, vma); in rockchip_drm_gem_object_mmap() 556 rockchip_gem_mmap_buf(struct drm_gem_object *obj, struct vm_area_struct *vma) rockchip_gem_mmap_buf() argument 569 rockchip_gem_mmap(struct file *filp, struct vm_area_struct *vma) rockchip_gem_mmap() argument [all...] |
/device/soc/rockchip/common/sdk_linux/drivers/dma-buf/ |
H A D | dma-buf.c | 160 static void dma_buf_vma_open(struct vm_area_struct *vma) in dma_buf_vma_open() argument 162 struct dma_buf *dmabuf = vma->vm_file->private_data; in dma_buf_vma_open() 165 /* call the heap provided vma open() op */ in dma_buf_vma_open() 167 dmabuf->exp_vm_ops->open(vma); in dma_buf_vma_open() 171 static void dma_buf_vma_close(struct vm_area_struct *vma) in dma_buf_vma_close() argument 173 struct dma_buf *dmabuf = vma->vm_file->private_data; in dma_buf_vma_close() 178 /* call the heap provided vma close() op */ in dma_buf_vma_close() 180 dmabuf->exp_vm_ops->close(vma); in dma_buf_vma_close() 184 static int dma_buf_do_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) in dma_buf_do_mmap() argument 186 /* call this first because the exporter might override vma in dma_buf_do_mmap() 204 dma_buf_do_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) dma_buf_do_mmap() argument 210 dma_buf_mmap_internal(struct file *file, struct vm_area_struct *vma) dma_buf_mmap_internal() argument 1331 dma_buf_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma, unsigned long pgoff) dma_buf_mmap() argument [all...] |