Home
last modified time | relevance | path

Searched refs:vma (Results 1 - 25 of 83) sorted by relevance

1234

/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/ump/linux/
H A Dump_osk_low_level_mem.c38 static void ump_vma_open(struct vm_area_struct *vma);
39 static void ump_vma_close(struct vm_area_struct *vma);
41 static int ump_cpu_page_fault_handler(struct vm_area_struct *vma, struct vm_fault *vmf);
43 static unsigned long ump_cpu_page_fault_handler(struct vm_area_struct *vma, unsigned long address);
61 static int ump_cpu_page_fault_handler(struct vm_area_struct *vma, struct vm_fault *vmf) in ump_cpu_page_fault_handler() argument
63 static unsigned long ump_cpu_page_fault_handler(struct vm_area_struct *vma, unsigned long address) in ump_cpu_page_fault_handler()
71 MSG_ERR(("VMA: 0x%08lx, virtual address: 0x%08lx\n", (unsigned long)vma, address)); in ump_cpu_page_fault_handler()
80 static void ump_vma_open(struct vm_area_struct *vma) in ump_vma_open() argument
85 vma_usage_tracker = (ump_vma_usage_tracker *)vma->vm_private_data; in ump_vma_open()
90 DBG_MSG(4, ("VMA open, VMA reference count incremented. VMA: 0x%08lx, reference count: %d\n", (unsigned long)vma, new_va in ump_vma_open()
93 ump_vma_close(struct vm_area_struct *vma) ump_vma_close() argument
128 struct vm_area_struct *vma; _ump_osk_mem_mapregion_init() local
174 struct vm_area_struct *vma; _ump_osk_mem_mapregion_term() local
192 struct vm_area_struct *vma; _ump_osk_mem_mapregion_map() local
[all...]
H A Dump_kernel_linux.c86 static int ump_file_mmap(struct file *filp, struct vm_area_struct *vma);
394 static int ump_file_mmap(struct file *filp, struct vm_area_struct *vma) in ump_file_mmap() argument
410 args.size = vma->vm_end - vma->vm_start; in ump_file_mmap()
411 args._ukk_private = vma; in ump_file_mmap()
412 args.secure_id = vma->vm_pgoff; in ump_file_mmap()
415 vma->vm_flags |= VM_DONTCOPY; in ump_file_mmap()
417 DBG_MSG(4, ("UMP vma->flags: %x\n", vma->vm_flags)); in ump_file_mmap()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory.c40 static void mali_mem_vma_open(struct vm_area_struct *vma) in mali_mem_vma_open() argument
42 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_open()
43 MALI_DEBUG_PRINT(MALI_KERNEL_LEVEL_CODE, ("Open called on vma %p\n", vma)); in mali_mem_vma_open()
49 static void mali_mem_vma_close(struct vm_area_struct *vma) in mali_mem_vma_close() argument
52 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_close()
55 vma->vm_private_data = NULL; in mali_mem_vma_close()
60 struct vm_area_struct *vma = vmf->vma; in mali_mem_vma_fault() local
61 mali_mem_allocation *alloc = (mali_mem_allocation *)vma in mali_mem_vma_fault()
150 mali_mmap(struct file *filp, struct vm_area_struct *vma) mali_mmap() argument
[all...]
H A Dmali_memory_cow.c365 if (alloc->cpu_mapping.vma != NULL) { in mali_memory_cow_modify_range()
367 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma != NULL); in mali_memory_cow_modify_range()
368 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range()
371 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range()
373 ret = mali_mem_cow_cpu_map_pages_locked(backend, alloc->cpu_mapping.vma, in mali_memory_cow_modify_range()
374 alloc->cpu_mapping.vma->vm_start + range_start, in mali_memory_cow_modify_range()
383 alloc->cpu_mapping.vma->vm_flags |= VM_PFNMAP; in mali_memory_cow_modify_range()
384 alloc->cpu_mapping.vma in mali_memory_cow_modify_range()
515 mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma) mali_mem_cow_cpu_map() argument
546 mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long vaddr, int num) mali_mem_cow_cpu_map_pages_locked() argument
[all...]
H A Dmali_memory_cow.h22 int mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma);
23 mali_osk_errcode_t mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma,
H A Dmali_memory_os_alloc.h52 int mali_mem_os_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma);
54 mali_osk_errcode_t mali_mem_os_resize_cpu_map_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma,
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory.c42 static void mali_mem_vma_open(struct vm_area_struct *vma) in mali_mem_vma_open() argument
44 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_open()
45 MALI_DEBUG_PRINT(4, ("Open called on vma %p\n", vma)); in mali_mem_vma_open()
51 static void mali_mem_vma_close(struct vm_area_struct *vma) in mali_mem_vma_close() argument
54 mali_mem_allocation *alloc = (mali_mem_allocation *)vma->vm_private_data; in mali_mem_vma_close()
57 vma->vm_private_data = NULL; in mali_mem_vma_close()
62 struct vm_area_struct *vma = vmf->vma; in mali_mem_vma_fault() local
63 mali_mem_allocation *alloc = (mali_mem_allocation *)vma in mali_mem_vma_fault()
152 mali_mmap(struct file *filp, struct vm_area_struct *vma) mali_mmap() argument
[all...]
H A Dmali_memory_cow.c378 if (NULL != alloc->cpu_mapping.vma) { in mali_memory_cow_modify_range()
380 MALI_DEBUG_ASSERT(NULL != alloc->cpu_mapping.vma); in mali_memory_cow_modify_range()
381 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range()
384 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range()
386 ret = mali_mem_cow_cpu_map_pages_locked(backend, alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size / _MALI_OSK_MALI_PAGE_SIZE); in mali_memory_cow_modify_range()
394 alloc->cpu_mapping.vma->vm_flags |= VM_PFNMAP; in mali_memory_cow_modify_range()
395 alloc->cpu_mapping.vma in mali_memory_cow_modify_range()
522 mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma) mali_mem_cow_cpu_map() argument
553 mali_mem_cow_cpu_map_pages_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long vaddr, int num) mali_mem_cow_cpu_map_pages_locked() argument
[all...]
H A Dmali_memory_cow.h18 int mali_mem_cow_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma);
20 struct vm_area_struct *vma,
H A Dmali_memory_os_alloc.h50 int mali_mem_os_cpu_map(mali_mem_backend *mem_bkend, struct vm_area_struct *vma);
52 _mali_osk_errcode_t mali_mem_os_resize_cpu_map_locked(mali_mem_backend *mem_bkend, struct vm_area_struct *vma, unsigned long start_vaddr, u32 mappig_size);
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/thirdparty/
H A Dmali_kbase_mmap.c152 struct vm_area_struct *vma; in kbase_unmapped_area_topdown() local
188 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
189 if (vma->rb_subtree_gap < length) { in kbase_unmapped_area_topdown()
195 gap_start = vma->vm_prev ? vma->vm_prev->vm_end : 0; in kbase_unmapped_area_topdown()
196 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
197 struct vm_area_struct *right = rb_entry(vma->vm_rb.rb_right, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
199 vma = right; in kbase_unmapped_area_topdown()
206 gap_end = vma->vm_start; in kbase_unmapped_area_topdown()
223 if (vma in kbase_unmapped_area_topdown()
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/thirdparty/
H A Dmali_kbase_mmap.c136 struct vm_area_struct *vma; in kbase_unmapped_area_topdown() local
168 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
169 if (vma->rb_subtree_gap < length) in kbase_unmapped_area_topdown()
174 gap_start = vma->vm_prev ? vma->vm_prev->vm_end : 0; in kbase_unmapped_area_topdown()
175 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
177 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown()
180 vma = right; in kbase_unmapped_area_topdown()
187 gap_end = vma->vm_start; in kbase_unmapped_area_topdown()
203 if (vma in kbase_unmapped_area_topdown()
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c92 struct vm_area_struct *vma);
94 struct vm_area_struct *vma);
104 struct vm_area_struct *vma);
2358 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument
2360 struct kbase_cpu_mapping *map = vma->vm_private_data;
2368 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument
2370 struct kbase_cpu_mapping *map = vma->vm_private_data;
2405 static struct kbase_aliased *get_aliased_alloc(struct vm_area_struct *vma, argument
2436 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, argument
2442 struct vm_area_struct *vma
2516 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
2630 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
2707 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2775 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument
3211 kbase_special_vm_close(struct vm_area_struct *vma) global() argument
3223 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3271 kbase_csf_user_io_pages_vm_open(struct vm_area_struct *vma) global() argument
3277 kbase_csf_user_io_pages_vm_close(struct vm_area_struct *vma) global() argument
3297 kbase_csf_user_io_pages_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument
3389 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3462 kbase_csf_user_reg_vm_close(struct vm_area_struct *vma) global() argument
3472 kbase_csf_user_reg_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument
3501 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
H A Dmali_kbase_mem_linux.h148 * @vma: Pointer to the struct containing the info where the GPU allocation
153 int kbase_context_mmap(struct kbase_context *kctx, struct vm_area_struct *vma);
441 static inline vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, in vmf_insert_pfn_prot() argument
450 if (pgprot_val(pgprot) != pgprot_val(vma->vm_page_prot)) { in vmf_insert_pfn_prot()
454 err = vm_insert_pfn(vma, addr, pfn); in vmf_insert_pfn_prot()
456 err = vm_insert_pfn_prot(vma, addr, pfn, pgprot); in vmf_insert_pfn_prot()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c90 struct vm_area_struct *vma);
92 struct vm_area_struct *vma);
101 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma);
2328 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument
2330 struct kbase_cpu_mapping *map = vma->vm_private_data;
2338 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument
2340 struct kbase_cpu_mapping *map = vma->vm_private_data;
2373 static struct kbase_aliased *get_aliased_alloc(struct vm_area_struct *vma, argument
2405 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, argument
2411 struct vm_area_struct *vma
2484 kbase_cpu_mmap(struct kbase_context *kctx, struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
2594 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
2673 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2752 kbase_context_mmap(struct kbase_context *const kctx, struct vm_area_struct *const vma) global() argument
3176 kbase_special_vm_close(struct vm_area_struct *vma) global() argument
3188 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3230 kbase_csf_user_io_pages_vm_open(struct vm_area_struct *vma) global() argument
3236 kbase_csf_user_io_pages_vm_close(struct vm_area_struct *vma) global() argument
3271 kbase_csf_user_io_pages_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument
3362 kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
3434 kbase_csf_user_reg_vm_close(struct vm_area_struct *vma) global() argument
3444 kbase_csf_user_reg_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) global() argument
3493 kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
H A Dmali_kbase_mem_linux.h145 * @vma: Pointer to the struct containing the info where the GPU allocation
150 int kbase_context_mmap(struct kbase_context *kctx, struct vm_area_struct *vma);
439 static inline vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, in vmf_insert_pfn_prot() argument
447 if (pgprot_val(pgprot) != pgprot_val(vma->vm_page_prot)) in vmf_insert_pfn_prot()
450 err = vm_insert_pfn(vma, addr, pfn); in vmf_insert_pfn_prot()
452 err = vm_insert_pfn_prot(vma, addr, pfn, pgprot); in vmf_insert_pfn_prot()
/device/soc/rockchip/common/sdk_linux/include/linux/
H A Dmm.h217 * per a vma. In ELF, the number of sections is represented in unsigned short.
303 #define VM_DONTCOPY 0x00020000 /* Do not copy this vma on fork */
321 #define VM_HUGEPAGE 0x20000000 /* MADV_HUGEPAGE marked this vma */
322 #define VM_NOHUGEPAGE 0x40000000 /* MADV_NOHUGEPAGE marked this vma */
532 * vm_fault is filled by the pagefault handler and passed to the vma's
533 * ->fault function. The vma's ->fault is responsible for returning a bitmask
542 struct vm_area_struct *vma; /* Target VMA */ member
545 pgoff_t pgoff; /* Logical page offset based on vma */
610 int (*access)(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write);
612 /* Called by the /proc/PID/maps code to ask the vma whethe
647 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) vma_init() argument
657 vma_set_anonymous(struct vm_area_struct *vma) vma_set_anonymous() argument
662 vma_is_anonymous(struct vm_area_struct *vma) vma_is_anonymous() argument
667 vma_is_temporary_stack(struct vm_area_struct *vma) vma_is_temporary_stack() argument
682 vma_is_foreign(struct vm_area_struct *vma) vma_is_foreign() argument
695 vma_is_accessible(struct vm_area_struct *vma) vma_is_accessible() argument
707 vma_is_shmem(struct vm_area_struct *vma) vma_is_shmem() argument
1008 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) maybe_mkwrite() argument
1738 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) handle_mm_fault() argument
2561 vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert) vma_adjust() argument
2688 struct vm_area_struct *vma = find_vma(mm, start_addr); find_vma_intersection() local
2696 vm_start_gap(struct vm_area_struct *vma) vm_start_gap() argument
2709 vm_end_gap(struct vm_area_struct *vma) vm_end_gap() argument
2722 vma_pages(struct vm_area_struct *vma) vma_pages() argument
2730 struct vm_area_struct *vma = find_vma(mm, vm_start); find_exact_vma() local
2739 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) range_in_vma() argument
2752 vma_set_page_prot(struct vm_area_struct *vma) vma_set_page_prot() argument
2775 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vmf_insert_page() argument
2788 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
3133 vma_is_special_huge(const struct vm_area_struct *vma) vma_is_special_huge() argument
3210 seal_check_future_write(int seals, struct vm_area_struct *vma) seal_check_future_write() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem_linux.c49 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma);
1606 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument
1608 struct kbase_cpu_mapping *map = vma->vm_private_data;
1616 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument
1618 struct kbase_cpu_mapping *map = vma->vm_private_data;
1654 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) argument
1659 struct vm_area_struct *vma = vmf->vma;
1661 struct kbase_cpu_mapping *map = vma->vm_private_data;
1689 while (i < map->alloc->nents && (addr < vma
1713 kbase_cpu_mmap(struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
1813 kbase_trace_buffer_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kaddr) global() argument
1898 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
1970 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2038 kbase_mmap(struct file *file, struct vm_area_struct *vma) global() argument
2412 kbase_special_vm_close(struct vm_area_struct *vma) global() argument
2424 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
/device/soc/rockchip/common/sdk_linux/ipc/
H A Dshm.c91 static void shm_open(struct vm_area_struct *vma);
92 static void shm_close(struct vm_area_struct *vma);
271 static int __shm_open(struct vm_area_struct *vma) in __shm_open() argument
273 struct file *file = vma->vm_file; in __shm_open()
296 static void shm_open(struct vm_area_struct *vma) in shm_open() argument
298 int err = __shm_open(vma); in shm_open()
353 * remove the attach descriptor vma.
358 static void shm_close(struct vm_area_struct *vma) in shm_close() argument
360 struct file *file = vma->vm_file; in shm_close()
519 struct file *file = vmf->vma in shm_fault()
525 shm_split(struct vm_area_struct *vma, unsigned long addr) shm_split() argument
537 shm_pagesize(struct vm_area_struct *vma) shm_pagesize() argument
550 shm_set_policy(struct vm_area_struct *vma, struct mempolicy *new) shm_set_policy() argument
562 shm_get_policy(struct vm_area_struct *vma, unsigned long addr) shm_get_policy() argument
578 shm_mmap(struct file *file, struct vm_area_struct *vma) shm_mmap() argument
1750 struct vm_area_struct *vma; ksys_shmdt() local
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem_linux.c44 static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma);
1610 static void kbase_cpu_vm_open(struct vm_area_struct *vma) argument
1612 struct kbase_cpu_mapping *map = vma->vm_private_data;
1620 static void kbase_cpu_vm_close(struct vm_area_struct *vma) argument
1622 struct kbase_cpu_mapping *map = vma->vm_private_data;
1658 static vm_fault_t kbase_cpu_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) argument
1663 struct vm_area_struct *vma = vmf->vma;
1665 struct kbase_cpu_mapping *map = vma->vm_private_data;
1695 while (i < map->alloc->nents && (addr < vma
1717 kbase_cpu_mmap(struct kbase_va_region *reg, struct vm_area_struct *vma, void *kaddr, size_t nr_pages, unsigned long aligned_offset, int free_on_close) global() argument
1816 kbase_trace_buffer_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kaddr) global() argument
1901 kbase_mmu_dump_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **const reg, void **const kmap_addr) global() argument
1972 kbasep_reg_mmap(struct kbase_context *kctx, struct vm_area_struct *vma, struct kbase_va_region **regm, size_t *nr_pages, size_t *aligned_offset) global() argument
2036 kbase_mmap(struct file *file, struct vm_area_struct *vma) global() argument
2410 kbase_special_vm_close(struct vm_area_struct *vma) global() argument
2422 kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) global() argument
[all...]
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/osal/linux/mmz/
H A Dmmz_userdev.c464 error_mmz("mmb<%s> vma is invalid.\n", p->mmb->name);
469 error_mmz("mmb<%s> vma is invalid.\n", p->mmb->name);
886 int mmz_userdev_mmap(struct file *file, struct vm_area_struct *vma) argument
890 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
891 unsigned long size = vma->vm_end - vma->vm_start;
926 vma->vm_page_prot = __pgprot(pgprot_val(vma->vm_page_prot)
929 vma->vm_page_prot = pgprot_writecombine(vma
[all...]
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/
H A Ddrm_gem.c1017 * drm_gem_vm_open - vma->ops->open implementation for GEM
1018 * @vma: VM area structure
1023 void drm_gem_vm_open(struct vm_area_struct *vma) in drm_gem_vm_open() argument
1025 struct drm_gem_object *obj = vma->vm_private_data; in drm_gem_vm_open()
1032 * drm_gem_vm_close - vma->ops->close implementation for GEM
1033 * @vma: VM area structure
1038 void drm_gem_vm_close(struct vm_area_struct *vma) in drm_gem_vm_close() argument
1040 struct drm_gem_object *obj = vma->vm_private_data; in drm_gem_vm_close()
1050 * @vma: VMA for the area to be mapped
1071 struct vm_area_struct *vma) in drm_gem_mmap_obj()
1070 drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size, struct vm_area_struct *vma) drm_gem_mmap_obj() argument
1131 drm_gem_mmap(struct file *filp, struct vm_area_struct *vma) drm_gem_mmap() argument
[all...]
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc.c229 u32 offset = intel_guc_ggtt_offset(guc, guc->log.vma) >> PAGE_SHIFT; in guc_ctl_log_params_flags()
503 * @rsa_offset: rsa offset w.r.t ggtt base of huc vma
655 struct i915_vma *vma; in intel_guc_allocate_vma() local
664 vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL); in intel_guc_allocate_vma()
665 if (IS_ERR(vma)) { in intel_guc_allocate_vma()
669 flags = PIN_OFFSET_BIAS | i915_ggtt_pin_bias(vma); in intel_guc_allocate_vma()
670 ret = i915_ggtt_pin(vma, NULL, 0, flags); in intel_guc_allocate_vma()
672 vma = ERR_PTR(ret); in intel_guc_allocate_vma()
676 return i915_vma_make_unshrinkable(vma); in intel_guc_allocate_vma()
680 return vma; in intel_guc_allocate_vma()
697 struct i915_vma *vma; intel_guc_allocate_and_map_vma() local
[all...]
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/rockchip/
H A Drockchip_drm_gem.c504 static int rockchip_drm_gem_object_mmap_iommu(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap_iommu() argument
508 unsigned long user_count = vma_pages(vma); in rockchip_drm_gem_object_mmap_iommu()
513 return vm_map_pages(vma, rk_obj->pages, count); in rockchip_drm_gem_object_mmap_iommu()
516 static int rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap_dma() argument
521 return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, obj->size, rk_obj->dma_attrs); in rockchip_drm_gem_object_mmap_dma()
524 static int rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) in rockchip_drm_gem_object_mmap() argument
531 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in rockchip_drm_gem_object_mmap()
538 vma->vm_flags &= ~VM_PFNMAP; in rockchip_drm_gem_object_mmap()
544 ret = rockchip_drm_gem_object_mmap_iommu(obj, vma); in rockchip_drm_gem_object_mmap()
556 rockchip_gem_mmap_buf(struct drm_gem_object *obj, struct vm_area_struct *vma) rockchip_gem_mmap_buf() argument
569 rockchip_gem_mmap(struct file *filp, struct vm_area_struct *vma) rockchip_gem_mmap() argument
[all...]
/device/soc/rockchip/common/sdk_linux/drivers/dma-buf/
H A Ddma-buf.c160 static void dma_buf_vma_open(struct vm_area_struct *vma) in dma_buf_vma_open() argument
162 struct dma_buf *dmabuf = vma->vm_file->private_data; in dma_buf_vma_open()
165 /* call the heap provided vma open() op */ in dma_buf_vma_open()
167 dmabuf->exp_vm_ops->open(vma); in dma_buf_vma_open()
171 static void dma_buf_vma_close(struct vm_area_struct *vma) in dma_buf_vma_close() argument
173 struct dma_buf *dmabuf = vma->vm_file->private_data; in dma_buf_vma_close()
178 /* call the heap provided vma close() op */ in dma_buf_vma_close()
180 dmabuf->exp_vm_ops->close(vma); in dma_buf_vma_close()
184 static int dma_buf_do_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) in dma_buf_do_mmap() argument
186 /* call this first because the exporter might override vma in dma_buf_do_mmap()
204 dma_buf_do_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) dma_buf_do_mmap() argument
210 dma_buf_mmap_internal(struct file *file, struct vm_area_struct *vma) dma_buf_mmap_internal() argument
1331 dma_buf_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma, unsigned long pgoff) dma_buf_mmap() argument
[all...]

Completed in 27 milliseconds

1234