/kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
H A D | i915_vma.h | 51 static inline bool i915_vma_is_active(const struct i915_vma *vma) in i915_vma_is_active() argument 53 return !i915_active_is_idle(&vma->active); in i915_vma_is_active() 60 int __must_check _i915_vma_move_to_active(struct i915_vma *vma, 65 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, in i915_vma_move_to_active() argument 68 return _i915_vma_move_to_active(vma, rq, &rq->fence, flags); in i915_vma_move_to_active() 73 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) in i915_vma_is_ggtt() argument 75 return test_bit(I915_VMA_GGTT_BIT, __i915_vma_flags(vma)); in i915_vma_is_ggtt() 78 static inline bool i915_vma_is_dpt(const struct i915_vma *vma) in i915_vma_is_dpt() argument 80 return i915_is_dpt(vma->vm); in i915_vma_is_dpt() 83 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) in i915_vma_has_ggtt_write() argument 88 i915_vma_set_ggtt_write(struct i915_vma *vma) i915_vma_set_ggtt_write() argument 94 i915_vma_unset_ggtt_write(struct i915_vma *vma) i915_vma_unset_ggtt_write() argument 102 i915_vma_is_map_and_fenceable(const struct i915_vma *vma) i915_vma_is_map_and_fenceable() argument 107 i915_vma_set_userfault(struct i915_vma *vma) i915_vma_set_userfault() argument 113 i915_vma_unset_userfault(struct i915_vma *vma) i915_vma_unset_userfault() argument 118 i915_vma_has_userfault(const struct i915_vma *vma) i915_vma_has_userfault() argument 123 i915_vma_is_closed(const struct i915_vma *vma) i915_vma_is_closed() argument 129 __i915_vma_size(const struct i915_vma *vma) __i915_vma_size() argument 144 i915_vma_size(const struct i915_vma *vma) i915_vma_size() argument 151 __i915_vma_offset(const struct i915_vma *vma) __i915_vma_offset() argument 167 i915_vma_offset(const struct i915_vma *vma) i915_vma_offset() argument 173 i915_ggtt_offset(const struct i915_vma *vma) i915_ggtt_offset() argument 183 i915_ggtt_pin_bias(struct i915_vma *vma) i915_ggtt_pin_bias() argument 188 i915_vma_get(struct i915_vma *vma) i915_vma_get() argument 194 i915_vma_tryget(struct i915_vma *vma) i915_vma_tryget() argument 202 i915_vma_put(struct i915_vma *vma) i915_vma_put() argument 208 i915_vma_compare(struct i915_vma *vma, struct i915_address_space *vm, const struct i915_gtt_view *view) i915_vma_compare() argument 278 i915_vma_lock(struct i915_vma *vma) i915_vma_lock() argument 283 i915_vma_unlock(struct i915_vma *vma) i915_vma_unlock() argument 293 i915_vma_pin(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) i915_vma_pin() argument 316 i915_vma_pin_count(const struct i915_vma *vma) i915_vma_pin_count() argument 321 i915_vma_is_pinned(const struct i915_vma *vma) i915_vma_is_pinned() argument 326 __i915_vma_pin(struct i915_vma *vma) __i915_vma_pin() argument 332 __i915_vma_unpin(struct i915_vma *vma) __i915_vma_unpin() argument 338 i915_vma_unpin(struct i915_vma *vma) i915_vma_unpin() argument 344 i915_vma_is_bound(const struct i915_vma *vma, unsigned int where) i915_vma_is_bound() argument 400 __i915_vma_unpin_fence(struct i915_vma *vma) __i915_vma_unpin_fence() argument 415 i915_vma_unpin_fence(struct i915_vma *vma) i915_vma_unpin_fence() argument 421 i915_vma_fence_id(const struct i915_vma *vma) i915_vma_fence_id() argument 428 i915_vma_is_scanout(const struct i915_vma *vma) i915_vma_is_scanout() argument 433 i915_vma_mark_scanout(struct i915_vma *vma) i915_vma_mark_scanout() argument 438 i915_vma_clear_scanout(struct i915_vma *vma) i915_vma_clear_scanout() argument 466 i915_vma_sync(struct i915_vma *vma) i915_vma_sync() argument 482 i915_vma_get_current_resource(struct i915_vma *vma) i915_vma_get_current_resource() argument [all...] |
H A D | i915_vma.c | 46 static inline void assert_vma_held_evict(const struct i915_vma *vma) in assert_vma_held_evict() argument 53 if (kref_read(&vma->vm->ref)) in assert_vma_held_evict() 54 assert_object_held_shared(vma->obj); in assert_vma_held_evict() 64 static void i915_vma_free(struct i915_vma *vma) in i915_vma_free() argument 66 return kmem_cache_free(slab_vmas, vma); in i915_vma_free() 73 static void vma_print_allocator(struct i915_vma *vma, const char *reason) in vma_print_allocator() argument 77 if (!vma->node.stack) { in vma_print_allocator() 78 drm_dbg(vma->obj->base.dev, in vma_print_allocator() 79 "vma.node [%08llx + %08llx] %s: unknown owner\n", in vma_print_allocator() 80 vma in vma_print_allocator() 92 vma_print_allocator(struct i915_vma *vma, const char *reason) vma_print_allocator() argument 119 struct i915_vma *vma; vma_create() local 261 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); i915_vma_lookup() local 295 struct i915_vma *vma; i915_vma_instance() local 373 i915_vma_wait_for_bind(struct i915_vma *vma) i915_vma_wait_for_bind() argument 393 i915_vma_verify_bind_complete(struct i915_vma *vma) i915_vma_verify_bind_complete() argument 415 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, struct i915_vma *vma) i915_vma_resource_init_from_vma() argument 440 i915_vma_bind(struct i915_vma *vma, unsigned int pat_index, u32 flags, struct i915_vma_work *work, struct i915_vma_resource *vma_res) i915_vma_bind() argument 549 i915_vma_pin_iomap(struct i915_vma *vma) i915_vma_pin_iomap() argument 617 i915_vma_flush_writes(struct i915_vma *vma) i915_vma_flush_writes() argument 623 i915_vma_unpin_iomap(struct i915_vma *vma) i915_vma_unpin_iomap() argument 637 struct i915_vma *vma; i915_vma_unpin_and_release() local 655 i915_vma_misplaced(const struct i915_vma *vma, u64 size, u64 alignment, u64 flags) i915_vma_misplaced() argument 689 __i915_vma_set_map_and_fenceable(struct i915_vma *vma) __i915_vma_set_map_and_fenceable() argument 708 i915_gem_valid_gtt_space(struct i915_vma *vma, unsigned long color) i915_gem_valid_gtt_space() argument 756 i915_vma_insert(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) i915_vma_insert() argument 895 i915_vma_detach(struct i915_vma *vma) i915_vma_detach() argument 908 try_qad_pin(struct i915_vma *vma, unsigned int flags) try_qad_pin() argument 1268 __i915_vma_get_pages(struct i915_vma *vma) __i915_vma_get_pages() argument 1315 i915_vma_get_pages(struct i915_vma *vma) i915_vma_get_pages() argument 1362 __vma_put_pages(struct i915_vma *vma, unsigned int count) __vma_put_pages() argument 1378 i915_vma_put_pages(struct i915_vma *vma) i915_vma_put_pages() argument 1386 vma_unbind_pages(struct i915_vma *vma) vma_unbind_pages() argument 1400 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) i915_vma_pin_ww() argument 1584 __i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) __i915_ggtt_pin() argument 1619 i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) i915_ggtt_pin() argument 1651 struct i915_vma *vma; i915_ggtt_clear_scanout() local 1661 __vma_close(struct i915_vma *vma, struct intel_gt *gt) __vma_close() argument 1679 i915_vma_close(struct i915_vma *vma) i915_vma_close() argument 1696 __i915_vma_remove_closed(struct i915_vma *vma) __i915_vma_remove_closed() argument 1701 i915_vma_reopen(struct i915_vma *vma) i915_vma_reopen() argument 1711 force_unbind(struct i915_vma *vma) force_unbind() argument 1721 release_references(struct i915_vma *vma, struct intel_gt *gt, bool vm_ddestroy) release_references() argument 1775 i915_vma_destroy_locked(struct i915_vma *vma) i915_vma_destroy_locked() argument 1784 i915_vma_destroy(struct i915_vma *vma) i915_vma_destroy() argument 1803 struct i915_vma *vma, *next; i915_vma_parked() local 1846 __i915_vma_iounmap(struct i915_vma *vma) __i915_vma_iounmap() argument 1860 i915_vma_revoke_mmap(struct i915_vma *vma) i915_vma_revoke_mmap() argument 1884 __i915_request_await_bind(struct i915_request *rq, struct i915_vma *vma) __i915_request_await_bind() argument 1889 __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq) __i915_vma_move_to_active() argument 1901 _i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, struct dma_fence *fence, unsigned int flags) _i915_vma_move_to_active() argument 1976 __i915_vma_evict(struct i915_vma *vma, bool async) __i915_vma_evict() argument 2056 __i915_vma_unbind(struct i915_vma *vma) __i915_vma_unbind() argument 2087 __i915_vma_unbind_async(struct i915_vma *vma) __i915_vma_unbind_async() argument 2122 i915_vma_unbind(struct i915_vma *vma) i915_vma_unbind() argument 2160 i915_vma_unbind_async(struct i915_vma *vma, bool trylock_vm) i915_vma_unbind_async() argument 2222 i915_vma_unbind_unlocked(struct i915_vma *vma) i915_vma_unbind_unlocked() argument 2233 i915_vma_make_unshrinkable(struct i915_vma *vma) i915_vma_make_unshrinkable() argument 2239 i915_vma_make_shrinkable(struct i915_vma *vma) i915_vma_make_shrinkable() argument 2244 i915_vma_make_purgeable(struct i915_vma *vma) i915_vma_make_purgeable() argument [all...] |
H A D | i915_gem_evict.c | 41 static bool dying_vma(struct i915_vma *vma) in dying_vma() argument 43 return !kref_read(&vma->obj->base.refcount); in dying_vma() 54 * Not everything in the GGTT is tracked via vma (otherwise we in ggtt_flush() 67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) in grab_vma() argument 73 if (i915_gem_object_get_rcu(vma->obj)) { in grab_vma() 74 if (!i915_gem_object_trylock(vma->obj, ww)) { in grab_vma() 75 i915_gem_object_put(vma->obj); in grab_vma() 80 atomic_and(~I915_VMA_PIN_MASK, &vma->flags); in grab_vma() 86 static void ungrab_vma(struct i915_vma *vma) in ungrab_vma() argument 88 if (dying_vma(vma)) in ungrab_vma() 96 mark_free(struct drm_mm_scan *scan, struct i915_gem_ww_ctx *ww, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) mark_free() argument 112 defer_evict(struct i915_vma *vma) defer_evict() argument 157 struct i915_vma *vma, *next; i915_gem_evict_something() local 337 struct i915_vma *vma, *next; i915_gem_evict_for_node() local 478 struct i915_vma *vma, *vn; i915_gem_evict_vm() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | i915_vma.c | 51 void i915_vma_free(struct i915_vma *vma) in i915_vma_free() argument 53 return kmem_cache_free(global.slab_vmas, vma); in i915_vma_free() 60 static void vma_print_allocator(struct i915_vma *vma, const char *reason) in vma_print_allocator() argument 66 if (!vma->node.stack) { in vma_print_allocator() 67 DRM_DEBUG_DRIVER("vma.node [%08llx + %08llx] %s: unknown owner\n", in vma_print_allocator() 68 vma->node.start, vma->node.size, reason); in vma_print_allocator() 72 nr_entries = stack_depot_fetch(vma->node.stack, &entries); in vma_print_allocator() 74 DRM_DEBUG_DRIVER("vma.node [%08llx + %08llx] %s: inserted at %s\n", in vma_print_allocator() 75 vma in vma_print_allocator() 80 vma_print_allocator(struct i915_vma *vma, const char *reason) vma_print_allocator() argument 108 struct i915_vma *vma; vma_create() local 241 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); vma_lookup() local 275 struct i915_vma *vma; i915_vma_instance() local 296 struct i915_vma *vma; global() member 306 struct i915_vma *vma = vw->vma; __vma_bind() local 346 i915_vma_wait_for_bind(struct i915_vma *vma) i915_vma_wait_for_bind() argument 376 i915_vma_bind(struct i915_vma *vma, enum i915_cache_level cache_level, u32 flags, struct i915_vma_work *work) i915_vma_bind() argument 449 i915_vma_pin_iomap(struct i915_vma *vma) i915_vma_pin_iomap() argument 495 i915_vma_flush_writes(struct i915_vma *vma) i915_vma_flush_writes() argument 501 i915_vma_unpin_iomap(struct i915_vma *vma) i915_vma_unpin_iomap() argument 513 struct i915_vma *vma; i915_vma_unpin_and_release() local 531 i915_vma_misplaced(const struct i915_vma *vma, u64 size, u64 alignment, u64 flags) i915_vma_misplaced() argument 561 __i915_vma_set_map_and_fenceable(struct i915_vma *vma) __i915_vma_set_map_and_fenceable() argument 579 i915_gem_valid_gtt_space(struct i915_vma *vma, unsigned long color) i915_gem_valid_gtt_space() argument 626 i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) i915_vma_insert() argument 735 i915_vma_detach(struct i915_vma *vma) i915_vma_detach() argument 748 try_qad_pin(struct i915_vma *vma, unsigned int flags) try_qad_pin() argument 791 vma_get_pages(struct i915_vma *vma) vma_get_pages() argument 824 __vma_put_pages(struct i915_vma *vma, unsigned int count) __vma_put_pages() argument 838 vma_put_pages(struct i915_vma *vma) vma_put_pages() argument 846 vma_unbind_pages(struct i915_vma *vma) vma_unbind_pages() argument 860 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) i915_vma_pin_ww() argument 1018 i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) i915_ggtt_pin() argument 1046 __vma_close(struct i915_vma *vma, struct intel_gt *gt) __vma_close() argument 1064 i915_vma_close(struct i915_vma *vma) i915_vma_close() argument 1081 __i915_vma_remove_closed(struct i915_vma *vma) __i915_vma_remove_closed() argument 1090 i915_vma_reopen(struct i915_vma *vma) i915_vma_reopen() argument 1098 struct i915_vma *vma = container_of(ref, typeof(*vma), ref); i915_vma_release() local 1128 struct i915_vma *vma, *next; i915_vma_parked() local 1163 __i915_vma_iounmap(struct i915_vma *vma) __i915_vma_iounmap() argument 1174 i915_vma_revoke_mmap(struct i915_vma *vma) i915_vma_revoke_mmap() argument 1198 __i915_request_await_bind(struct i915_request *rq, struct i915_vma *vma) __i915_request_await_bind() argument 1203 __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq) __i915_vma_move_to_active() argument 1217 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, unsigned int flags) i915_vma_move_to_active() argument 1262 __i915_vma_evict(struct i915_vma *vma) __i915_vma_evict() argument 1305 __i915_vma_unbind(struct i915_vma *vma) __i915_vma_unbind() argument 1335 i915_vma_unbind(struct i915_vma *vma) i915_vma_unbind() argument 1371 i915_vma_make_unshrinkable(struct i915_vma *vma) i915_vma_make_unshrinkable() argument 1377 i915_vma_make_shrinkable(struct i915_vma *vma) i915_vma_make_shrinkable() argument 1382 i915_vma_make_purgeable(struct i915_vma *vma) i915_vma_make_purgeable() argument [all...] |
H A D | i915_vma.h | 50 static inline bool i915_vma_is_active(const struct i915_vma *vma) in i915_vma_is_active() argument 52 return !i915_active_is_idle(&vma->active); in i915_vma_is_active() 55 int __must_check __i915_vma_move_to_active(struct i915_vma *vma, 57 int __must_check i915_vma_move_to_active(struct i915_vma *vma, 63 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) in i915_vma_is_ggtt() argument 65 return test_bit(I915_VMA_GGTT_BIT, __i915_vma_flags(vma)); in i915_vma_is_ggtt() 68 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) in i915_vma_has_ggtt_write() argument 70 return test_bit(I915_VMA_GGTT_WRITE_BIT, __i915_vma_flags(vma)); in i915_vma_has_ggtt_write() 73 static inline void i915_vma_set_ggtt_write(struct i915_vma *vma) in i915_vma_set_ggtt_write() argument 75 GEM_BUG_ON(!i915_vma_is_ggtt(vma)); in i915_vma_set_ggtt_write() 79 i915_vma_unset_ggtt_write(struct i915_vma *vma) i915_vma_unset_ggtt_write() argument 87 i915_vma_is_map_and_fenceable(const struct i915_vma *vma) i915_vma_is_map_and_fenceable() argument 92 i915_vma_set_userfault(struct i915_vma *vma) i915_vma_set_userfault() argument 98 i915_vma_unset_userfault(struct i915_vma *vma) i915_vma_unset_userfault() argument 103 i915_vma_has_userfault(const struct i915_vma *vma) i915_vma_has_userfault() argument 108 i915_vma_is_closed(const struct i915_vma *vma) i915_vma_is_closed() argument 113 i915_ggtt_offset(const struct i915_vma *vma) i915_ggtt_offset() argument 122 i915_ggtt_pin_bias(struct i915_vma *vma) i915_ggtt_pin_bias() argument 127 i915_vma_get(struct i915_vma *vma) i915_vma_get() argument 133 i915_vma_tryget(struct i915_vma *vma) i915_vma_tryget() argument 141 i915_vma_put(struct i915_vma *vma) i915_vma_put() argument 152 i915_vma_compare(struct i915_vma *vma, struct i915_address_space *vm, const struct i915_ggtt_view *view) i915_vma_compare() argument 213 __i915_vma_get(struct i915_vma *vma) __i915_vma_get() argument 222 __i915_vma_put(struct i915_vma *vma) __i915_vma_put() argument 229 i915_vma_lock(struct i915_vma *vma) i915_vma_lock() argument 234 i915_vma_unlock(struct i915_vma *vma) i915_vma_unlock() argument 244 i915_vma_pin(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) i915_vma_pin() argument 252 i915_vma_pin_count(const struct i915_vma *vma) i915_vma_pin_count() argument 257 i915_vma_is_pinned(const struct i915_vma *vma) i915_vma_is_pinned() argument 262 __i915_vma_pin(struct i915_vma *vma) __i915_vma_pin() argument 268 __i915_vma_unpin(struct i915_vma *vma) __i915_vma_unpin() argument 274 i915_vma_unpin(struct i915_vma *vma) i915_vma_unpin() argument 280 i915_vma_is_bound(const struct i915_vma *vma, unsigned int where) i915_vma_is_bound() argument 317 i915_vma_first_page(struct i915_vma *vma) i915_vma_first_page() argument 343 __i915_vma_unpin_fence(struct i915_vma *vma) __i915_vma_unpin_fence() argument 358 i915_vma_unpin_fence(struct i915_vma *vma) i915_vma_unpin_fence() argument 390 i915_vma_sync(struct i915_vma *vma) i915_vma_sync() argument [all...] |
H A D | i915_gem_evict.c | 42 * Not everything in the GGTT is tracked via vma (otherwise we in ggtt_flush() 53 struct i915_vma *vma, in mark_free() 57 if (i915_vma_is_pinned(vma)) in mark_free() 60 list_add(&vma->evict_link, unwind); in mark_free() 61 return drm_mm_scan_add_block(scan, &vma->node); in mark_free() 78 * This function is used by the object/vma binding code. 96 struct i915_vma *vma, *next; in i915_gem_evict_something() local 130 list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) { in i915_gem_evict_something() 131 if (vma == active) { /* now seen this vma twic in i915_gem_evict_something() 52 mark_free(struct drm_mm_scan *scan, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) mark_free() argument 259 struct i915_vma *vma, *next; i915_gem_evict_for_node() local 377 struct i915_vma *vma, *vn; i915_gem_evict_vm() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_vm.c | 61 struct vm_area_struct *vma; member 65 static void drm_vm_open(struct vm_area_struct *vma); 66 static void drm_vm_close(struct vm_area_struct *vma); 69 struct vm_area_struct *vma) in drm_io_prot() 71 pgprot_t tmp = vm_get_page_prot(vma->vm_flags); in drm_io_prot() 83 if (efi_range_is_wc(vma->vm_start, vma->vm_end - in drm_io_prot() 84 vma->vm_start)) in drm_io_prot() 94 static pgprot_t drm_dma_prot(uint32_t map_type, struct vm_area_struct *vma) in drm_dma_prot() argument 96 pgprot_t tmp = vm_get_page_prot(vma in drm_dma_prot() 68 drm_io_prot(struct drm_local_map *map, struct vm_area_struct *vma) drm_io_prot() argument 117 struct vm_area_struct *vma = vmf->vma; drm_vm_fault() local 206 struct vm_area_struct *vma = vmf->vma; drm_vm_shm_fault() local 235 drm_vm_shm_close(struct vm_area_struct *vma) drm_vm_shm_close() argument 306 struct vm_area_struct *vma = vmf->vma; drm_vm_dma_fault() local 341 struct vm_area_struct *vma = vmf->vma; drm_vm_sg_fault() local 394 drm_vm_open_locked(struct drm_device *dev, struct vm_area_struct *vma) drm_vm_open_locked() argument 410 drm_vm_open(struct vm_area_struct *vma) drm_vm_open() argument 420 drm_vm_close_locked(struct drm_device *dev, struct vm_area_struct *vma) drm_vm_close_locked() argument 445 drm_vm_close(struct vm_area_struct *vma) drm_vm_close() argument 465 drm_mmap_dma(struct file *filp, struct vm_area_struct *vma) drm_mmap_dma() argument 528 drm_mmap_locked(struct file *filp, struct vm_area_struct *vma) drm_mmap_locked() argument 642 drm_legacy_mmap(struct file *filp, struct vm_area_struct *vma) drm_legacy_mmap() argument 662 struct drm_vma_entry *vma, *vma_temp; drm_legacy_vma_flush() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_vm.c | 59 struct vm_area_struct *vma; member 63 static void drm_vm_open(struct vm_area_struct *vma); 64 static void drm_vm_close(struct vm_area_struct *vma); 67 struct vm_area_struct *vma) in drm_io_prot() 69 pgprot_t tmp = vm_get_page_prot(vma->vm_flags); in drm_io_prot() 78 if (efi_range_is_wc(vma->vm_start, vma->vm_end - in drm_io_prot() 79 vma->vm_start)) in drm_io_prot() 89 static pgprot_t drm_dma_prot(uint32_t map_type, struct vm_area_struct *vma) in drm_dma_prot() argument 91 pgprot_t tmp = vm_get_page_prot(vma in drm_dma_prot() 66 drm_io_prot(struct drm_local_map *map, struct vm_area_struct *vma) drm_io_prot() argument 112 struct vm_area_struct *vma = vmf->vma; drm_vm_fault() local 201 struct vm_area_struct *vma = vmf->vma; drm_vm_shm_fault() local 230 drm_vm_shm_close(struct vm_area_struct *vma) drm_vm_shm_close() argument 301 struct vm_area_struct *vma = vmf->vma; drm_vm_dma_fault() local 336 struct vm_area_struct *vma = vmf->vma; drm_vm_sg_fault() local 389 drm_vm_open_locked(struct drm_device *dev, struct vm_area_struct *vma) drm_vm_open_locked() argument 405 drm_vm_open(struct vm_area_struct *vma) drm_vm_open() argument 415 drm_vm_close_locked(struct drm_device *dev, struct vm_area_struct *vma) drm_vm_close_locked() argument 440 drm_vm_close(struct vm_area_struct *vma) drm_vm_close() argument 460 drm_mmap_dma(struct file *filp, struct vm_area_struct *vma) drm_mmap_dma() argument 523 drm_mmap_locked(struct file *filp, struct vm_area_struct *vma) drm_mmap_locked() argument 637 drm_legacy_mmap(struct file *filp, struct vm_area_struct *vma) drm_legacy_mmap() argument 657 struct drm_vma_entry *vma, *vma_temp; drm_legacy_vma_flush() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | mremap.c | 72 static pud_t *alloc_new_pud(struct mm_struct *mm, struct vm_area_struct *vma, in alloc_new_pud() argument 86 static pmd_t *alloc_new_pmd(struct mm_struct *mm, struct vm_area_struct *vma, in alloc_new_pmd() argument 92 pud = alloc_new_pud(mm, vma, addr); in alloc_new_pmd() 105 static void take_rmap_locks(struct vm_area_struct *vma) in take_rmap_locks() argument 107 if (vma->vm_file) in take_rmap_locks() 108 i_mmap_lock_write(vma->vm_file->f_mapping); in take_rmap_locks() 109 if (vma->anon_vma) in take_rmap_locks() 110 anon_vma_lock_write(vma->anon_vma); in take_rmap_locks() 113 static void drop_rmap_locks(struct vm_area_struct *vma) in drop_rmap_locks() argument 115 if (vma in drop_rmap_locks() 136 move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, unsigned long old_addr, unsigned long old_end, struct vm_area_struct *new_vma, pmd_t *new_pmd, unsigned long new_addr, bool need_rmap_locks) move_ptes() argument 236 move_normal_pmd(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pmd_t *old_pmd, pmd_t *new_pmd) move_normal_pmd() argument 295 move_normal_pmd(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pmd_t *old_pmd, pmd_t *new_pmd) move_normal_pmd() argument 304 move_normal_pud(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pud_t *old_pud, pud_t *new_pud) move_normal_pud() argument 344 move_normal_pud(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pud_t *old_pud, pud_t *new_pud) move_normal_pud() argument 353 move_huge_pud(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pud_t *old_pud, pud_t *new_pud) move_huge_pud() argument 393 move_huge_pud(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pud_t *old_pud, pud_t *new_pud) move_huge_pud() argument 451 move_pgt_entry(enum pgt_entry entry, struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, void *old_entry, void *new_entry, bool need_rmap_locks) move_pgt_entry() argument 492 move_page_tables(struct vm_area_struct *vma, unsigned long old_addr, struct vm_area_struct *new_vma, unsigned long new_addr, unsigned long len, bool need_rmap_locks) move_page_tables() argument 583 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr, bool *locked, unsigned long flags, struct vm_userfaultfd_ctx *uf, struct list_head *uf_unmap) move_vma() argument 750 struct vm_area_struct *vma; vma_to_resize() local 807 struct vm_area_struct *vma; mremap_to() local 887 vma_expandable(struct vm_area_struct *vma, unsigned long delta) vma_expandable() argument 913 struct vm_area_struct *vma; SYSCALL_DEFINE5() local [all...] |
H A D | mmap.c | 86 struct vm_area_struct *vma, struct vm_area_struct *prev, 95 /* Update vma->vm_page_prot to reflect vma->vm_flags. */ 96 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot() argument 98 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() 101 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot() 102 if (vma_wants_writenotify(vma, vm_page_prot)) { in vma_set_page_prot() 106 /* remove_protection_ptes reads vma->vm_page_prot without mmap_lock */ in vma_set_page_prot() 107 WRITE_ONCE(vma->vm_page_prot, vm_page_prot); in vma_set_page_prot() 113 static void __remove_shared_vm_struct(struct vm_area_struct *vma, in __remove_shared_vm_struct() argument 128 unlink_file_vma(struct vm_area_struct *vma) unlink_file_vma() argument 143 remove_vma(struct vm_area_struct *vma, bool unreachable) remove_vma() argument 291 struct vm_area_struct *vma; validate_mm() local 356 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct *vma) anon_vma_interval_tree_pre_update_vma() argument 365 anon_vma_interval_tree_post_update_vma(struct vm_area_struct *vma) anon_vma_interval_tree_post_update_vma() argument 377 struct vm_area_struct *vma; count_vma_pages_range() local 390 __vma_link_file(struct vm_area_struct *vma, struct address_space *mapping) __vma_link_file() argument 401 vma_link(struct mm_struct *mm, struct vm_area_struct *vma) vma_link() argument 434 init_multi_vma_prep(struct vma_prepare *vp, struct vm_area_struct *vma, struct vm_area_struct *next, struct vm_area_struct *remove, struct vm_area_struct *remove2) init_multi_vma_prep() argument 458 init_vma_prep(struct vma_prepare *vp, struct vm_area_struct *vma) init_vma_prep() argument 636 vma_expand(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *next) vma_expand() argument 692 vma_shrink(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff) vma_shrink() argument 726 is_mergeable_vma(struct vm_area_struct *vma, struct file *file, unsigned long vm_flags, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name, bool may_remove_vma) is_mergeable_vma() argument 752 is_mergeable_anon_vma(struct anon_vma *anon_vma1, struct anon_vma *anon_vma2, struct vm_area_struct *vma) is_mergeable_anon_vma() argument 779 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name) can_vma_merge_before() argument 802 can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name) can_vma_merge_after() argument 878 struct vm_area_struct *vma, *adjust, *remove, *remove2; vma_merge() local 1110 find_mergeable_anon_vma(struct vm_area_struct *vma) find_mergeable_anon_vma() argument 1483 vma_is_shared_writable(struct vm_area_struct *vma) vma_is_shared_writable() argument 1489 vma_fs_can_writeback(struct vm_area_struct *vma) vma_fs_can_writeback() argument 1503 vma_needs_dirty_tracking(struct vm_area_struct *vma) vma_needs_dirty_tracking() argument 1526 vma_wants_writenotify(struct vm_area_struct *vma, pgprot_t vm_page_prot) vma_wants_writenotify() argument 1716 struct vm_area_struct *vma, *prev; generic_get_unmapped_area() local 1763 struct vm_area_struct *vma, *prev; generic_get_unmapped_area_topdown() local 1918 struct vm_area_struct *vma; find_vma_prev() local 1933 acct_stack_growth(struct vm_area_struct *vma, unsigned long size, unsigned long grow) acct_stack_growth() argument 1972 expand_upwards(struct vm_area_struct *vma, unsigned long address) expand_upwards() argument 2072 expand_downwards(struct vm_area_struct *vma, unsigned long address) expand_downwards() argument 2178 expand_stack_locked(struct vm_area_struct *vma, unsigned long address) expand_stack_locked() argument 2185 struct vm_area_struct *vma, *prev; find_extend_vma_locked() local 2200 expand_stack_locked(struct vm_area_struct *vma, unsigned long address) expand_stack_locked() argument 2209 struct vm_area_struct *vma; find_extend_vma_locked() local 2235 vma_expand_ok(struct vm_area_struct *vma, unsigned long addr) vma_expand_ok() argument 2246 vma_expand_up(struct vm_area_struct *vma, unsigned long addr) vma_expand_up() argument 2255 vma_expand_down(struct vm_area_struct *vma, unsigned long addr) vma_expand_down() argument 2287 struct vm_area_struct *vma, *prev; expand_stack() local 2322 struct vm_area_struct *vma; remove_mt() local 2342 unmap_region(struct mm_struct *mm, struct ma_state *mas, struct vm_area_struct *vma, struct vm_area_struct *prev, struct vm_area_struct *next, unsigned long start, unsigned long end, unsigned long tree_end, bool mm_wr_locked) unmap_region() argument 2366 __split_vma(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long addr, int new_below) __split_vma() argument 2448 split_vma(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long addr, int new_below) split_vma() argument 2472 do_vmi_align_munmap(struct vma_iterator *vmi, struct vm_area_struct *vma, struct mm_struct *mm, unsigned long start, unsigned long end, struct list_head *uf, bool unlock) do_vmi_align_munmap() argument 2643 struct vm_area_struct *vma; do_vmi_munmap() local 2692 struct vm_area_struct *vma = NULL; mmap_region() local 2982 struct vm_area_struct *vma; SYSCALL_DEFINE5() local 3072 do_vma_munmap(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *uf, bool unlock) do_vma_munmap() argument 3094 do_brk_flags(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long addr, unsigned long len, unsigned long flags) do_brk_flags() argument 3178 struct vm_area_struct *vma = NULL; vm_brk_flags() local 3232 struct vm_area_struct *vma; exit_mmap() local 3296 insert_vm_struct(struct mm_struct *mm, struct vm_area_struct *vma) insert_vm_struct() argument 3341 struct vm_area_struct *vma = *vmap; copy_vma() local 3469 special_mapping_close(struct vm_area_struct *vma) special_mapping_close() argument 3473 special_mapping_name(struct vm_area_struct *vma) special_mapping_name() argument 3491 special_mapping_split(struct vm_area_struct *vma, unsigned long addr) special_mapping_split() argument 3519 struct vm_area_struct *vma = vmf->vma; special_mapping_fault() local 3554 struct vm_area_struct *vma; __install_special_mapping() local 3585 vma_is_special_mapping(const struct vm_area_struct *vma, const struct vm_special_mapping *sm) vma_is_special_mapping() argument 3615 struct vm_area_struct *vma = __install_special_mapping( install_special_mapping() local 3705 struct vm_area_struct *vma; mm_take_all_locks() local 3801 struct vm_area_struct *vma; mm_drop_all_locks() local [all...] |
H A D | nommu.c | 99 struct vm_area_struct *vma; in kobjsize() local 101 vma = find_vma(current->mm, (unsigned long)objp); in kobjsize() 102 if (vma) in kobjsize() 103 return vma->vm_end - vma->vm_start; in kobjsize() 115 * @vma: memory mapping 123 int follow_pfn(struct vm_area_struct *vma, unsigned long address, in follow_pfn() argument 126 if (!(vma->vm_flags & (VM_IO | VM_PFNMAP))) in follow_pfn() 172 struct vm_area_struct *vma; in __vmalloc_user_flags() local 175 vma in __vmalloc_user_flags() 353 vm_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vm_insert_page() argument 360 vm_map_pages(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages() argument 367 vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages_zero() argument 547 setup_vma_to_mm(struct vm_area_struct *vma, struct mm_struct *mm) setup_vma_to_mm() argument 563 cleanup_vma_from_mm(struct vm_area_struct *vma) cleanup_vma_from_mm() argument 582 delete_vma_from_mm(struct vm_area_struct *vma) delete_vma_from_mm() argument 601 delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) delete_vma() argument 641 struct vm_area_struct *vma; lock_mm_and_find_vma() local 654 expand_stack_locked(struct vm_area_struct *vma, unsigned long addr) expand_stack_locked() argument 673 struct vm_area_struct *vma; find_vma_exact() local 895 do_mmap_shared_file(struct vm_area_struct *vma) do_mmap_shared_file() argument 916 do_mmap_private(struct vm_area_struct *vma, struct vm_region *region, unsigned long len, unsigned long capabilities) do_mmap_private() argument 1024 struct vm_area_struct *vma; do_mmap() local 1308 split_vma(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long addr, int new_below) split_vma() argument 1386 vmi_shrink_vma(struct vma_iterator *vmi, struct vm_area_struct *vma, unsigned long from, unsigned long to) vmi_shrink_vma() argument 1431 struct vm_area_struct *vma; do_munmap() local 1513 struct vm_area_struct *vma; exit_mmap() local 1553 struct vm_area_struct *vma; do_mremap() local 1597 follow_page(struct vm_area_struct *vma, unsigned long address, unsigned int foll_flags) follow_page() argument 1603 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() argument 1614 vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len) vm_iomap_memory() argument 1624 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) remap_vmalloc_range() argument 1657 struct vm_area_struct *vma; __access_remote_vm() local 1741 struct vm_area_struct *vma; nommu_shrink_inode_mappings() local [all...] |
H A D | mprotect.c | 43 bool can_change_pte_writable(struct vm_area_struct *vma, unsigned long addr, in can_change_pte_writable() argument 48 if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE))) in can_change_pte_writable() 56 if (vma_soft_dirty_enabled(vma) && !pte_soft_dirty(pte)) in can_change_pte_writable() 60 if (userfaultfd_pte_wp(vma, pte)) in can_change_pte_writable() 63 if (!(vma->vm_flags & VM_SHARED)) { in can_change_pte_writable() 70 page = vm_normal_page(vma, addr, pte); in can_change_pte_writable() 85 struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, in change_pte_range() 97 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in change_pte_range() 102 if (prot_numa && !(vma->vm_flags & VM_SHARED) && in change_pte_range() 103 atomic_read(&vma in change_pte_range() 84 change_pte_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t newprot, unsigned long cp_flags) change_pte_range() argument 297 pgtable_split_needed(struct vm_area_struct *vma, unsigned long cp_flags) pgtable_split_needed() argument 312 pgtable_populate_needed(struct vm_area_struct *vma, unsigned long cp_flags) pgtable_populate_needed() argument 354 change_pmd_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, pgprot_t newprot, unsigned long cp_flags) change_pmd_range() argument 438 change_pud_range(struct mmu_gather *tlb, struct vm_area_struct *vma, p4d_t *p4d, unsigned long addr, unsigned long end, pgprot_t newprot, unsigned long cp_flags) change_pud_range() argument 461 change_p4d_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t newprot, unsigned long cp_flags) change_p4d_range() argument 484 change_protection_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end, pgprot_t newprot, unsigned long cp_flags) change_protection_range() argument 514 change_protection(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long cp_flags) change_protection() argument 576 mprotect_fixup(struct vma_iterator *vmi, struct mmu_gather *tlb, struct vm_area_struct *vma, struct vm_area_struct **pprev, unsigned long start, unsigned long end, unsigned long newflags) mprotect_fixup() argument 695 struct vm_area_struct *vma, *prev; do_mprotect_pkey() local [all...] |
H A D | madvise.c | 46 * Any behaviour which results in changes to the vma->vm_flags needs to 94 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument 96 mmap_assert_locked(vma->vm_mm); in anon_vma_name() 98 return vma->anon_name; in anon_vma_name() 102 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument 105 struct anon_vma_name *orig_name = anon_vma_name(vma); in replace_anon_vma_name() 108 vma->anon_name = NULL; in replace_anon_vma_name() 116 vma->anon_name = anon_vma_name_reuse(anon_name); in replace_anon_vma_name() 122 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument 132 * Update the vm_flags on region of a vma, splittin 137 madvise_update_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *anon_name) madvise_update_vma() argument 192 struct vm_area_struct *vma = walk->private; swapin_walk_pmd_entry() local 238 shmem_swapin_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) shmem_swapin_range() argument 279 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument 327 can_do_file_pageout(struct vm_area_struct *vma) can_do_file_pageout() argument 350 struct vm_area_struct *vma = walk->vma; madvise_cold_or_pageout_pte_range() local 544 madvise_cold_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end) madvise_cold_page_range() argument 558 can_madv_lru_vma(struct vm_area_struct *vma) can_madv_lru_vma() argument 563 madvise_cold(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start_addr, unsigned long end_addr) madvise_cold() argument 582 madvise_pageout_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end) madvise_pageout_page_range() argument 596 madvise_pageout(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start_addr, unsigned long end_addr) madvise_pageout() argument 631 struct vm_area_struct *vma = walk->vma; madvise_free_pte_range() local 768 madvise_free_single_vma(struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr) madvise_free_single_vma() argument 822 madvise_dontneed_single_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) madvise_dontneed_single_vma() argument 829 madvise_dontneed_free_valid_vma(struct vm_area_struct *vma, unsigned long start, unsigned long *end, int behavior) madvise_dontneed_free_valid_vma() argument 859 madvise_dontneed_free(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_dontneed_free() argument 913 madvise_populate(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_populate() argument 974 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument 1024 madvise_vma_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long behavior) madvise_vma_behavior() argument 1229 madvise_walk_vmas(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned long arg, int (*visit)(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long arg)) madvise_walk_vmas() argument 1235 struct vm_area_struct *vma; madvise_walk_vmas() local 1288 madvise_vma_anon_name(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long anon_name) madvise_vma_anon_name() argument [all...] |
H A D | memory.c | 366 struct vm_area_struct *vma, unsigned long floor, in free_pgtables() 370 unsigned long addr = vma->vm_start; in free_pgtables() 380 * Hide vma from rmap and truncate_pagecache before freeing in free_pgtables() 384 vma_start_write(vma); in free_pgtables() 385 unlink_anon_vmas(vma); in free_pgtables() 386 unlink_file_vma(vma); in free_pgtables() 388 if (is_vm_hugetlb_page(vma)) { in free_pgtables() 389 hugetlb_free_pgd_range(tlb, addr, vma->vm_end, in free_pgtables() 395 while (next && next->vm_start <= vma->vm_end + PMD_SIZE in free_pgtables() 397 vma in free_pgtables() 365 free_pgtables(struct mmu_gather *tlb, struct ma_state *mas, struct vm_area_struct *vma, unsigned long floor, unsigned long ceiling, bool mm_wr_locked) free_pgtables() argument 490 print_bad_pte(struct vm_area_struct *vma, unsigned long addr, pte_t pte, struct page *page) print_bad_pte() argument 583 vm_normal_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte) vm_normal_page() argument 646 vm_normal_folio(struct vm_area_struct *vma, unsigned long addr, pte_t pte) vm_normal_folio() argument 657 vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd) vm_normal_page_pmd() argument 698 restore_exclusive_pte(struct vm_area_struct *vma, struct page *page, unsigned long address, pte_t *ptep) restore_exclusive_pte() argument 746 try_restore_exclusive_pte(pte_t *src_pte, struct vm_area_struct *vma, unsigned long addr) try_restore_exclusive_pte() argument 984 page_copy_prealloc(struct mm_struct *src_mm, struct vm_area_struct *vma, unsigned long addr) page_copy_prealloc() argument 1382 zap_install_uffd_wp_if_needed(struct vm_area_struct *vma, unsigned long addr, pte_t *pte, struct zap_details *details, pte_t pteval) zap_install_uffd_wp_if_needed() argument 1396 zap_pte_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, struct zap_details *details) zap_pte_range() argument 1548 zap_pmd_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, struct zap_details *details) zap_pmd_range() argument 1590 zap_pud_range(struct mmu_gather *tlb, struct vm_area_struct *vma, p4d_t *p4d, unsigned long addr, unsigned long end, struct zap_details *details) zap_pud_range() argument 1619 zap_p4d_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, struct zap_details *details) zap_p4d_range() argument 1638 unmap_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end, struct zap_details *details) unmap_page_range() argument 1659 unmap_single_vma(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, struct zap_details *details, bool mm_wr_locked) unmap_single_vma() argument 1724 unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas, struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, unsigned long tree_end, bool mm_wr_locked) unmap_vmas() argument 1759 zap_page_range_single(struct vm_area_struct *vma, unsigned long address, unsigned long size, struct zap_details *details) zap_page_range_single() argument 1794 zap_vma_ptes(struct vm_area_struct *vma, unsigned long address, unsigned long size) zap_vma_ptes() argument 1845 insert_page_into_pte_locked(struct vm_area_struct *vma, pte_t *pte, unsigned long addr, struct page *page, pgprot_t prot) insert_page_into_pte_locked() argument 1865 insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page, pgprot_t prot) insert_page() argument 1885 insert_page_in_batch_locked(struct vm_area_struct *vma, pte_t *pte, unsigned long addr, struct page *page, pgprot_t prot) insert_page_in_batch_locked() argument 1901 insert_pages(struct vm_area_struct *vma, unsigned long addr, struct page **pages, unsigned long *num, pgprot_t prot) insert_pages() argument 1974 vm_insert_pages(struct vm_area_struct *vma, unsigned long addr, struct page **pages, unsigned long *num) vm_insert_pages() argument 2020 vm_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vm_insert_page() argument 2047 __vm_map_pages(struct vm_area_struct *vma, struct page **pages, unsigned long num, unsigned long offset) __vm_map_pages() argument 2090 vm_map_pages(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages() argument 2110 vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages_zero() argument 2117 insert_pfn(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, pgprot_t prot, bool mkwrite) insert_pfn() argument 2204 vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, pgprot_t pgprot) vmf_insert_pfn_prot() argument 2252 vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) vmf_insert_pfn() argument 2259 vm_mixed_ok(struct vm_area_struct *vma, pfn_t pfn) vm_mixed_ok() argument 2273 __vm_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, bool mkwrite) __vm_insert_mixed() argument 2319 vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) vmf_insert_mixed() argument 2331 vmf_insert_mixed_mkwrite(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) vmf_insert_mixed_mkwrite() argument 2440 remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range_notrack() argument 2505 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() argument 2536 vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len) vm_iomap_memory() argument 2820 struct vm_area_struct *vma = vmf->vma; __wp_page_copy_user() local 2912 __get_fault_gfp_mask(struct vm_area_struct *vma) __get_fault_gfp_mask() argument 2967 struct vm_area_struct *vma = vmf->vma; fault_dirty_shared_page() local 3021 struct vm_area_struct *vma = vmf->vma; global() variable 3065 struct vm_area_struct *vma = vmf->vma; wp_page_copy() local 3271 struct vm_area_struct *vma = vmf->vma; wp_pfn_shared() local 3295 struct vm_area_struct *vma = vmf->vma; global() variable 3358 struct vm_area_struct *vma = vmf->vma; global() variable 3471 unmap_mapping_range_vma(struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, struct zap_details *details) unmap_mapping_range_vma() argument 3483 struct vm_area_struct *vma; unmap_mapping_range_tree() local 3605 struct vm_area_struct *vma = vmf->vma; remove_device_exclusive_entry() local 3644 should_try_to_free_swap(struct folio *folio, struct vm_area_struct *vma, unsigned int fault_flags) should_try_to_free_swap() argument 3740 struct vm_area_struct *vma = vmf->vma; do_swap_page() local 4106 struct vm_area_struct *vma = vmf->vma; do_anonymous_page() local 4239 struct vm_area_struct *vma = vmf->vma; __do_fault() local 4296 struct vm_area_struct *vma = vmf->vma; deposit_prealloc_pte() local 4309 struct vm_area_struct *vma = vmf->vma; do_set_pmd() local 4389 struct vm_area_struct *vma = vmf->vma; set_pte_range() local 4448 struct vm_area_struct *vma = vmf->vma; finish_fault() local 4641 struct vm_area_struct *vma = vmf->vma; do_cow_fault() local 4685 struct vm_area_struct *vma = vmf->vma; do_shared_fault() local 4736 struct vm_area_struct *vma = vmf->vma; do_fault() local 4778 numa_migrate_prep(struct page *page, struct vm_area_struct *vma, unsigned long addr, int page_nid, int *flags) numa_migrate_prep() argument 4797 struct vm_area_struct *vma = vmf->vma; do_numa_page() local 4914 struct vm_area_struct *vma = vmf->vma; create_huge_pmd() local 4925 struct vm_area_struct *vma = vmf->vma; wp_huge_pmd() local 4954 struct vm_area_struct *vma = vmf->vma; create_huge_pud() local 4968 struct vm_area_struct *vma = vmf->vma; wp_huge_pud() local 5087 __handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags) __handle_mm_fault() argument 5250 lru_gen_enter_fault(struct vm_area_struct *vma) lru_gen_enter_fault() argument 5261 lru_gen_enter_fault(struct vm_area_struct *vma) lru_gen_enter_fault() argument 5270 sanitize_fault_flags(struct vm_area_struct *vma, unsigned int *flags) sanitize_fault_flags() argument 5311 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) handle_mm_fault() argument 5428 struct vm_area_struct *vma; lock_mm_and_find_vma() local 5491 struct vm_area_struct *vma; lock_vma_under_rcu() local 5678 follow_pfn(struct vm_area_struct *vma, unsigned long address, unsigned long *pfn) follow_pfn() argument 5698 follow_phys(struct vm_area_struct *vma, unsigned long address, unsigned int flags, unsigned long *prot, resource_size_t *phys) follow_phys() argument 5738 generic_access_phys(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write) generic_access_phys() argument 5815 struct vm_area_struct *vma = NULL; __access_remote_vm() local 5920 struct vm_area_struct *vma; print_vma_addr() local 6059 copy_user_gigantic_page(struct folio *dst, struct folio *src, unsigned long addr, struct vm_area_struct *vma, unsigned int pages_per_huge_page) copy_user_gigantic_page() argument 6085 struct vm_area_struct *vma; global() member 6100 copy_user_large_folio(struct folio *dst, struct folio *src, unsigned long addr_hint, struct vm_area_struct *vma) copy_user_large_folio() argument [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | mmap.c | 87 struct vm_area_struct *vma, struct vm_area_struct *prev, 131 /* Update vma->vm_page_prot to reflect vma->vm_flags. */ 132 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot() argument 134 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() 137 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot() 138 if (vma_wants_writenotify(vma, vm_page_prot)) { in vma_set_page_prot() 142 /* remove_protection_ptes reads vma->vm_page_prot without mmap_lock */ in vma_set_page_prot() 143 WRITE_ONCE(vma->vm_page_prot, vm_page_prot); in vma_set_page_prot() 149 static void __remove_shared_vm_struct(struct vm_area_struct *vma, in __remove_shared_vm_struct() argument 166 unlink_file_vma(struct vm_area_struct *vma) unlink_file_vma() argument 181 remove_vma(struct vm_area_struct *vma) remove_vma() argument 296 vma_compute_gap(struct vm_area_struct *vma) vma_compute_gap() argument 318 vma_compute_subtree_gap(struct vm_area_struct *vma) vma_compute_subtree_gap() argument 344 struct vm_area_struct *vma; browse_rb() local 389 struct vm_area_struct *vma; validate_mm_rb() local 402 struct vm_area_struct *vma = mm->mmap; validate_mm() local 450 vma_gap_update(struct vm_area_struct *vma) vma_gap_update() argument 459 vma_rb_insert(struct vm_area_struct *vma, struct rb_root *root) vma_rb_insert() argument 468 __vma_rb_erase(struct vm_area_struct *vma, struct rb_root *root) __vma_rb_erase() argument 478 vma_rb_erase_ignore(struct vm_area_struct *vma, struct rb_root *root, struct vm_area_struct *ignore) vma_rb_erase_ignore() argument 496 vma_rb_erase(struct vm_area_struct *vma, struct rb_root *root) vma_rb_erase() argument 517 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct *vma) anon_vma_interval_tree_pre_update_vma() argument 526 anon_vma_interval_tree_post_update_vma(struct vm_area_struct *vma) anon_vma_interval_tree_post_update_vma() argument 577 vma_next(struct mm_struct *mm, struct vm_area_struct *vma) vma_next() argument 616 struct vm_area_struct *vma; count_vma_pages_range() local 640 __vma_link_rb(struct mm_struct *mm, struct vm_area_struct *vma, struct rb_node **rb_link, struct rb_node *rb_parent) __vma_link_rb() argument 664 __vma_link_file(struct vm_area_struct *vma) __vma_link_file() argument 684 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) __vma_link() argument 692 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) vma_link() argument 717 __insert_vm_struct(struct mm_struct *mm, struct vm_area_struct *vma) __insert_vm_struct() argument 729 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *ignore) __vma_unlink() argument 746 __vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert, struct vm_area_struct *expand) __vma_adjust() argument 1036 is_mergeable_vma(struct vm_area_struct *vma, struct file *file, unsigned long vm_flags, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name) is_mergeable_vma() argument 1062 is_mergeable_anon_vma(struct anon_vma *anon_vma1, struct anon_vma *anon_vma2, struct vm_area_struct *vma) is_mergeable_anon_vma() argument 1088 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name) can_vma_merge_before() argument 1110 can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff, struct vm_userfaultfd_ctx vm_userfaultfd_ctx, struct anon_vma_name *anon_name) can_vma_merge_after() argument 1323 find_mergeable_anon_vma(struct vm_area_struct *vma) find_mergeable_anon_vma() argument 1470 struct vm_area_struct *vma = find_vma(mm, addr); do_mmap() local 1712 vma_wants_writenotify(struct vm_area_struct *vma, pgprot_t vm_page_prot) vma_wants_writenotify() argument 1769 struct vm_area_struct *vma, *prev, *merge; mmap_region() local 1971 struct vm_area_struct *vma; unmapped_area() local 2067 struct vm_area_struct *vma; unmapped_area_topdown() local 2204 struct vm_area_struct *vma, *prev; arch_get_unmapped_area() local 2249 struct vm_area_struct *vma, *prev; arch_get_unmapped_area_topdown() local 2351 struct vm_area_struct *vma; find_vma() local 2388 struct vm_area_struct *vma; find_vma_prev() local 2406 acct_stack_growth(struct vm_area_struct *vma, unsigned long size, unsigned long grow) acct_stack_growth() argument 2452 expand_upwards(struct vm_area_struct *vma, unsigned long address) expand_upwards() argument 2542 expand_downwards(struct vm_area_struct *vma, unsigned long address) expand_downwards() argument 2633 expand_stack(struct vm_area_struct *vma, unsigned long address) expand_stack() argument 2641 struct vm_area_struct *vma, *prev; find_extend_vma() local 2655 expand_stack(struct vm_area_struct *vma, unsigned long address) expand_stack() argument 2663 struct vm_area_struct *vma; find_extend_vma() local 2691 remove_vma_list(struct mm_struct *mm, struct vm_area_struct *vma) remove_vma_list() argument 2714 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) unmap_region() argument 2752 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() argument 2793 __split_vma(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, int new_below) __split_vma() argument 2857 split_vma(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, int new_below) split_vma() argument 2875 struct vm_area_struct *vma, *prev, *last; __do_munmap() local 3039 struct vm_area_struct *vma; SYSCALL_DEFINE5() local 3138 struct vm_area_struct *vma, *prev; do_brk_flags() local 3240 struct vm_area_struct *vma; exit_mmap() local 3311 insert_vm_struct(struct mm_struct *mm, struct vm_area_struct *vma) insert_vm_struct() argument 3352 struct vm_area_struct *vma = *vmap; copy_vma() local 3469 special_mapping_close(struct vm_area_struct *vma) special_mapping_close() argument 3473 special_mapping_name(struct vm_area_struct *vma) special_mapping_name() argument 3507 struct vm_area_struct *vma = vmf->vma; special_mapping_fault() local 3542 struct vm_area_struct *vma; __install_special_mapping() local 3572 vma_is_special_mapping(const struct vm_area_struct *vma, const struct vm_special_mapping *sm) vma_is_special_mapping() argument 3602 struct vm_area_struct *vma = __install_special_mapping( install_special_mapping() local 3691 struct vm_area_struct *vma; mm_take_all_locks() local 3771 struct vm_area_struct *vma; mm_drop_all_locks() local [all...] |
H A D | nommu.c | 100 struct vm_area_struct *vma; in kobjsize() local 102 vma = find_vma(current->mm, (unsigned long)objp); in kobjsize() 103 if (vma) in kobjsize() 104 return vma->vm_end - vma->vm_start; in kobjsize() 116 * @vma: memory mapping 124 int follow_pfn(struct vm_area_struct *vma, unsigned long address, in follow_pfn() argument 127 if (!(vma->vm_flags & (VM_IO | VM_PFNMAP))) in follow_pfn() 173 struct vm_area_struct *vma; in __vmalloc_user_flags() local 176 vma in __vmalloc_user_flags() 363 vm_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vm_insert_page() argument 370 vm_map_pages(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages() argument 377 vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages, unsigned long num) vm_map_pages_zero() argument 563 add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma) add_vma_to_mm() argument 627 delete_vma_from_mm(struct vm_area_struct *vma) delete_vma_from_mm() argument 663 delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) delete_vma() argument 679 struct vm_area_struct *vma; find_vma() local 714 expand_stack(struct vm_area_struct *vma, unsigned long address) expand_stack() argument 727 struct vm_area_struct *vma; find_vma_exact() local 954 do_mmap_shared_file(struct vm_area_struct *vma) do_mmap_shared_file() argument 975 do_mmap_private(struct vm_area_struct *vma, struct vm_region *region, unsigned long len, unsigned long capabilities) do_mmap_private() argument 1078 struct vm_area_struct *vma; do_mmap() local 1354 split_vma(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, int new_below) split_vma() argument 1417 shrink_vma(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long from, unsigned long to) shrink_vma() argument 1458 struct vm_area_struct *vma; do_munmap() local 1538 struct vm_area_struct *vma; exit_mmap() local 1572 struct vm_area_struct *vma; do_mremap() local 1616 follow_page(struct vm_area_struct *vma, unsigned long address, unsigned int foll_flags) follow_page() argument 1622 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() argument 1633 vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len) vm_iomap_memory() argument 1643 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) remap_vmalloc_range() argument 1681 struct vm_area_struct *vma; __access_remote_vm() local 1765 struct vm_area_struct *vma; nommu_shrink_inode_mappings() local [all...] |
H A D | mremap.c | 59 static pmd_t *alloc_new_pmd(struct mm_struct *mm, struct vm_area_struct *vma, in alloc_new_pmd() argument 84 static void take_rmap_locks(struct vm_area_struct *vma) in take_rmap_locks() argument 86 if (vma->vm_file) in take_rmap_locks() 87 i_mmap_lock_write(vma->vm_file->f_mapping); in take_rmap_locks() 88 if (vma->anon_vma) in take_rmap_locks() 89 anon_vma_lock_write(vma->anon_vma); in take_rmap_locks() 92 static void drop_rmap_locks(struct vm_area_struct *vma) in drop_rmap_locks() argument 94 if (vma->anon_vma) in drop_rmap_locks() 95 anon_vma_unlock_write(vma->anon_vma); in drop_rmap_locks() 96 if (vma in drop_rmap_locks() 115 move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, unsigned long old_addr, unsigned long old_end, struct vm_area_struct *new_vma, pmd_t *new_pmd, unsigned long new_addr, bool need_rmap_locks) move_ptes() argument 195 move_normal_pmd(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pmd_t *old_pmd, pmd_t *new_pmd) move_normal_pmd() argument 254 move_page_tables(struct vm_area_struct *vma, unsigned long old_addr, struct vm_area_struct *new_vma, unsigned long new_addr, unsigned long len, bool need_rmap_locks) move_page_tables() argument 333 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr, bool *locked, unsigned long flags, struct vm_userfaultfd_ctx *uf, struct list_head *uf_unmap) move_vma() argument 480 struct vm_area_struct *vma = find_vma(mm, addr); vma_to_resize() local 552 struct vm_area_struct *vma; mremap_to() local 639 vma_expandable(struct vm_area_struct *vma, unsigned long delta) vma_expandable() argument 664 struct vm_area_struct *vma; SYSCALL_DEFINE5() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/display/ |
H A D | intel_fb_pin.c | 30 struct i915_vma *vma; in intel_pin_fb_obj_dpt() local 73 vma = i915_vma_instance(obj, vm, view); in intel_pin_fb_obj_dpt() 74 if (IS_ERR(vma)) { in intel_pin_fb_obj_dpt() 75 ret = PTR_ERR(vma); in intel_pin_fb_obj_dpt() 79 if (i915_vma_misplaced(vma, 0, alignment, 0)) { in intel_pin_fb_obj_dpt() 80 ret = i915_vma_unbind(vma); in intel_pin_fb_obj_dpt() 85 ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL); in intel_pin_fb_obj_dpt() 90 vma = ERR_PTR(ret); in intel_pin_fb_obj_dpt() 94 vma->display_alignment = max(vma in intel_pin_fb_obj_dpt() 117 struct i915_vma *vma; intel_pin_and_fence_fb_obj() local 230 intel_unpin_fb_vma(struct i915_vma *vma, unsigned long flags) intel_unpin_fb_vma() argument 243 struct i915_vma *vma; intel_plane_pin_fb() local 285 struct i915_vma *vma; intel_plane_unpin_fb() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/ |
H A D | nouveau_vmm.c | 29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap() argument 31 if (vma->mem) { in nouveau_vma_unmap() 32 nvif_vmm_unmap(&vma->vmm->vmm, vma->addr); in nouveau_vma_unmap() 33 vma->mem = NULL; in nouveau_vma_unmap() 38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument 40 struct nvif_vma tmp = { .addr = vma->addr }; in nouveau_vma_map() 41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp); in nouveau_vma_map() 44 vma->mem = mem; in nouveau_vma_map() 51 struct nouveau_vma *vma; in nouveau_vma_find() local 64 struct nouveau_vma *vma = *pvma; nouveau_vma_del() local 81 struct nouveau_vma *vma; nouveau_vma_new() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/ |
H A D | nouveau_vmm.c | 29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap() argument 31 if (vma->mem) { in nouveau_vma_unmap() 32 nvif_vmm_unmap(&vma->vmm->vmm, vma->addr); in nouveau_vma_unmap() 33 vma->mem = NULL; in nouveau_vma_unmap() 38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument 40 struct nvif_vma tmp = { .addr = vma->addr }; in nouveau_vma_map() 41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp); in nouveau_vma_map() 44 vma->mem = mem; in nouveau_vma_map() 51 struct nouveau_vma *vma; in nouveau_vma_find() local 64 struct nouveau_vma *vma = *pvma; nouveau_vma_del() local 81 struct nouveau_vma *vma; nouveau_vma_new() local [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | userfaultfd_k.h | 93 extern long uffd_wp_range(struct vm_area_struct *vma, 97 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, in is_mergeable_vm_userfaultfd_ctx() argument 100 return vma->vm_userfaultfd_ctx.ctx == vm_ctx.ctx; in is_mergeable_vm_userfaultfd_ctx() 114 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) in uffd_disable_huge_pmd_share() argument 116 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_huge_pmd_share() 126 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) in uffd_disable_fault_around() argument 128 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_fault_around() 131 static inline bool userfaultfd_missing(struct vm_area_struct *vma) in userfaultfd_missing() argument 133 return vma->vm_flags & VM_UFFD_MISSING; in userfaultfd_missing() 136 static inline bool userfaultfd_wp(struct vm_area_struct *vma) in userfaultfd_wp() argument 141 userfaultfd_minor(struct vm_area_struct *vma) userfaultfd_minor() argument 146 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) userfaultfd_pte_wp() argument 152 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) userfaultfd_huge_pmd_wp() argument 158 userfaultfd_armed(struct vm_area_struct *vma) userfaultfd_armed() argument 163 vma_can_userfault(struct vm_area_struct *vma, unsigned long vm_flags) vma_can_userfault() argument 210 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, struct vm_userfaultfd_ctx vm_ctx) is_mergeable_vm_userfaultfd_ctx() argument 216 userfaultfd_missing(struct vm_area_struct *vma) userfaultfd_missing() argument 221 userfaultfd_wp(struct vm_area_struct *vma) userfaultfd_wp() argument 226 userfaultfd_minor(struct vm_area_struct *vma) userfaultfd_minor() argument 231 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) userfaultfd_pte_wp() argument 237 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) userfaultfd_huge_pmd_wp() argument 244 userfaultfd_armed(struct vm_area_struct *vma) userfaultfd_armed() argument 249 dup_userfaultfd(struct vm_area_struct *vma, struct list_head *l) dup_userfaultfd() argument 259 mremap_userfaultfd_prep(struct vm_area_struct *vma, struct vm_userfaultfd_ctx *ctx) mremap_userfaultfd_prep() argument 271 userfaultfd_remove(struct vm_area_struct *vma, unsigned long start, unsigned long end) userfaultfd_remove() argument 278 userfaultfd_unmap_prep(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *uf) userfaultfd_unmap_prep() argument 290 uffd_disable_fault_around(struct vm_area_struct *vma) uffd_disable_fault_around() argument 295 userfaultfd_wp_unpopulated(struct vm_area_struct *vma) userfaultfd_wp_unpopulated() argument 302 userfaultfd_wp_use_markers(struct vm_area_struct *vma) userfaultfd_wp_use_markers() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 750 struct nvkm_vma *vma = kzalloc(sizeof(*vma), GFP_KERNEL); in nvkm_vma_new() local 751 if (vma) { in nvkm_vma_new() 752 vma->addr = addr; in nvkm_vma_new() 753 vma->size = size; in nvkm_vma_new() 754 vma->page = NVKM_VMA_PAGE_NONE; in nvkm_vma_new() 755 vma->refd = NVKM_VMA_PAGE_NONE; in nvkm_vma_new() 757 return vma; in nvkm_vma_new() 761 nvkm_vma_tail(struct nvkm_vma *vma, u64 tail) in nvkm_vma_tail() argument 765 BUG_ON(vma in nvkm_vma_tail() 785 nvkm_vmm_free_remove(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_free_remove() argument 791 nvkm_vmm_free_delete(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_free_delete() argument 799 nvkm_vmm_free_insert(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_free_insert() argument 827 nvkm_vmm_node_remove(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_node_remove() argument 833 nvkm_vmm_node_delete(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_node_delete() argument 841 nvkm_vmm_node_insert(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_node_insert() argument 867 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_node_search() local 883 nvkm_vmm_node_merge(struct nvkm_vmm *vmm, struct nvkm_vma *prev, struct nvkm_vma *vma, struct nvkm_vma *next, u64 size) nvkm_vmm_node_merge() argument 925 nvkm_vmm_node_split(struct nvkm_vmm *vmm, struct nvkm_vma *vma, u64 addr, u64 size) nvkm_vmm_node_split() argument 952 nvkm_vma_dump(struct nvkm_vma *vma) nvkm_vma_dump() argument 971 struct nvkm_vma *vma; nvkm_vmm_dump() local 980 struct nvkm_vma *vma; nvkm_vmm_dtor() local 987 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_dtor() local 1021 struct nvkm_vma *vma; nvkm_vmm_ctor_managed() local 1042 struct nvkm_vma *vma; nvkm_vmm_ctor() local 1151 nvkm_vmm_pfn_split_merge(struct nvkm_vmm *vmm, struct nvkm_vma *vma, u64 addr, u64 size, u8 page, bool map) nvkm_vmm_pfn_split_merge() argument 1176 struct nvkm_vma *vma = nvkm_vmm_node_search(vmm, addr); nvkm_vmm_pfn_unmap() local 1213 struct nvkm_vma *vma, *tmp; nvkm_vmm_pfn_map() local 1333 nvkm_vmm_unmap_region(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_unmap_region() argument 1350 nvkm_vmm_unmap_locked(struct nvkm_vmm *vmm, struct nvkm_vma *vma, bool pfn) nvkm_vmm_unmap_locked() argument 1365 nvkm_vmm_unmap(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_unmap() argument 1375 nvkm_vmm_map_valid(struct nvkm_vmm *vmm, struct nvkm_vma *vma, void *argv, u32 argc, struct nvkm_vmm_map *map) nvkm_vmm_map_valid() argument 1411 nvkm_vmm_map_choose(struct nvkm_vmm *vmm, struct nvkm_vma *vma, void *argv, u32 argc, struct nvkm_vmm_map *map) nvkm_vmm_map_choose() argument 1423 nvkm_vmm_map_locked(struct nvkm_vmm *vmm, struct nvkm_vma *vma, void *argv, u32 argc, struct nvkm_vmm_map *map) nvkm_vmm_map_locked() argument 1509 nvkm_vmm_map(struct nvkm_vmm *vmm, struct nvkm_vma *vma, void *argv, u32 argc, struct nvkm_vmm_map *map) nvkm_vmm_map() argument 1521 nvkm_vmm_put_region(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_put_region() argument 1540 nvkm_vmm_put_locked(struct nvkm_vmm *vmm, struct nvkm_vma *vma) nvkm_vmm_put_locked() argument 1625 struct nvkm_vma *vma = *pvma; nvkm_vmm_put() local 1640 struct nvkm_vma *vma = NULL, *tmp; nvkm_vmm_get_locked() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/msm/ |
H A D | msm_gem_vma.c | 41 /* Actually unmap memory for the vma */ 42 void msm_gem_vma_purge(struct msm_gem_vma *vma) in msm_gem_vma_purge() argument 44 struct msm_gem_address_space *aspace = vma->aspace; in msm_gem_vma_purge() 45 unsigned size = vma->node.size; in msm_gem_vma_purge() 48 if (!vma->mapped) in msm_gem_vma_purge() 51 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); in msm_gem_vma_purge() 53 vma->mapped = false; in msm_gem_vma_purge() 56 /* Map and pin vma: */ 58 msm_gem_vma_map(struct msm_gem_vma *vma, int prot, in msm_gem_vma_map() argument 61 struct msm_gem_address_space *aspace = vma in msm_gem_vma_map() 94 msm_gem_vma_close(struct msm_gem_vma *vma) msm_gem_vma_close() argument 112 struct msm_gem_vma *vma; msm_gem_vma_new() local 124 msm_gem_vma_init(struct msm_gem_vma *vma, int size, u64 range_start, u64 range_end) msm_gem_vma_init() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/ |
H A D | msm_gem_vma.c | 40 /* Actually unmap memory for the vma */ 42 struct msm_gem_vma *vma) in msm_gem_purge_vma() 44 unsigned size = vma->node.size << PAGE_SHIFT; in msm_gem_purge_vma() 46 /* Print a message if we try to purge a vma in use */ in msm_gem_purge_vma() 47 if (WARN_ON(vma->inuse > 0)) in msm_gem_purge_vma() 51 if (!vma->mapped) in msm_gem_purge_vma() 55 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); in msm_gem_purge_vma() 57 vma->mapped = false; in msm_gem_purge_vma() 62 struct msm_gem_vma *vma) in msm_gem_unmap_vma() 64 if (!WARN_ON(!vma in msm_gem_unmap_vma() 41 msm_gem_purge_vma(struct msm_gem_address_space *aspace, struct msm_gem_vma *vma) msm_gem_purge_vma() argument 61 msm_gem_unmap_vma(struct msm_gem_address_space *aspace, struct msm_gem_vma *vma) msm_gem_unmap_vma() argument 69 msm_gem_map_vma(struct msm_gem_address_space *aspace, struct msm_gem_vma *vma, int prot, struct sg_table *sgt, int npages) msm_gem_map_vma() argument 100 msm_gem_close_vma(struct msm_gem_address_space *aspace, struct msm_gem_vma *vma) msm_gem_close_vma() argument 117 msm_gem_init_vma(struct msm_gem_address_space *aspace, struct msm_gem_vma *vma, int npages, u64 range_start, u64 range_end) msm_gem_init_vma() argument [all...] |
/kernel/linux/linux-5.10/drivers/pci/ |
H A D | mmap.c | 23 struct vm_area_struct *vma, in pci_mmap_page_range() 31 vma->vm_pgoff -= start >> PAGE_SHIFT; in pci_mmap_page_range() 32 return pci_mmap_resource_range(pdev, bar, vma, mmap_state, in pci_mmap_page_range() 44 struct vm_area_struct *vma, in pci_mmap_resource_range() 51 if (vma->vm_pgoff + vma_pages(vma) > size) in pci_mmap_resource_range() 55 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); in pci_mmap_resource_range() 57 vma->vm_page_prot = pgprot_device(vma in pci_mmap_resource_range() 22 pci_mmap_page_range(struct pci_dev *pdev, int bar, struct vm_area_struct *vma, enum pci_mmap_state mmap_state, int write_combine) pci_mmap_page_range() argument 43 pci_mmap_resource_range(struct pci_dev *pdev, int bar, struct vm_area_struct *vma, enum pci_mmap_state mmap_state, int write_combine) pci_mmap_resource_range() argument 81 pci_mmap_resource_range(struct pci_dev *pdev, int bar, struct vm_area_struct *vma, enum pci_mmap_state mmap_state, int write_combine) pci_mmap_resource_range() argument [all...] |