Lines Matching defs:addr
113 #define untagged_addr(addr) (addr)
246 #define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE)
249 #define PAGE_ALIGNED(addr) IS_ALIGNED((unsigned long)(addr), PAGE_SIZE)
593 int (*split)(struct vm_area_struct *area, unsigned long addr);
610 int (*access)(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write);
629 * (vma,addr) marked as MPOL_SHARED. The shared policy infrastructure
631 * get_policy() must NOT add a ref if the policy at (vma,addr) is not
633 * If no [shared/vma] mempolicy exists at the addr, get_policy() op
637 struct mempolicy *(*get_policy)(struct vm_area_struct *vma, unsigned long addr);
641 * page for @addr. This is useful if the default behavior
644 struct page *(*find_special_page)(struct vm_area_struct *vma, unsigned long addr);
770 struct page *vmalloc_to_page(const void *addr);
771 unsigned long vmalloc_to_pfn(const void *addr);
829 extern void kvfree(const void *addr);
830 extern void kvfree_sensitive(const void *addr, size_t len);
1702 struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte);
1703 struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd);
1711 void free_pgd_range(struct mmu_gather *tlb, unsigned long addr, unsigned long end, unsigned long floor,
1720 int generic_access_phys(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write);
1770 extern int access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, unsigned int gup_flags);
1771 extern int access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, unsigned int gup_flags);
1772 extern int __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len,
1842 struct page *get_dump_page(unsigned long addr);
1897 static inline bool get_user_page_fast_only(unsigned long addr, unsigned int gup_flags, struct page **pagep)
1899 return get_user_pages_fast_only(addr, 1, gup_flags, pagep) == 1;
2043 extern pte_t *__get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl);
2044 static inline pte_t *get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl)
2047 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl));
2566 extern struct vm_area_struct *vma_merge(struct mm_struct *, struct vm_area_struct *prev, unsigned long addr,
2570 extern int __split_vma(struct mm_struct *, struct vm_area_struct *, unsigned long addr, int new_below);
2571 extern int split_vma(struct mm_struct *, struct vm_area_struct *, unsigned long addr, int new_below);
2575 extern struct vm_area_struct *copy_vma(struct vm_area_struct **, unsigned long addr, unsigned long len, pgoff_t pgoff,
2602 extern struct vm_area_struct *_install_special_mapping(struct mm_struct *mm, unsigned long addr, unsigned long len,
2605 extern int install_special_mapping(struct mm_struct *mm, unsigned long addr, unsigned long len, unsigned long flags,
2613 extern unsigned long mmap_region(struct file *file, unsigned long addr, unsigned long len, vm_flags_t vm_flags,
2615 extern unsigned long do_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot,
2622 extern int __mm_populate(unsigned long addr, unsigned long len, int ignore_errors);
2623 static inline void mm_populate(unsigned long addr, unsigned long len)
2626 (void)__mm_populate(addr, len, 1);
2629 static inline void mm_populate(unsigned long addr, unsigned long len)
2679 /* Look up the first VMA which satisfies addr < vm_end, NULL if none. */
2680 extern struct vm_area_struct *find_vma(struct mm_struct *mm, unsigned long addr);
2681 extern struct vm_area_struct *find_vma_prev(struct mm_struct *mm, unsigned long addr, struct vm_area_struct **pprev);
2762 struct vm_area_struct *find_extend_vma(struct mm_struct *, unsigned long addr);
2763 int remap_pfn_range(struct vm_area_struct *, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t);
2764 int vm_insert_page(struct vm_area_struct *, unsigned long addr, struct page *);
2765 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr, struct page **pages, unsigned long *num);
2768 vm_fault_t vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn);
2769 vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, pgprot_t pgprot);
2770 vm_fault_t vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn);
2771 vm_fault_t vmf_insert_mixed_prot(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, pgprot_t pgprot);
2772 vm_fault_t vmf_insert_mixed_mkwrite(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn);
2775 static inline vm_fault_t vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page)
2777 int err = vm_insert_page(vma, addr, page);
2788 static inline int io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn,
2791 return remap_pfn_range(vma, addr, pfn, size, pgprot_decrypted(prot));
2898 typedef int (*pte_fn_t)(pte_t *pte, unsigned long addr, void *data);
3003 extern int in_gate_area_no_mm(unsigned long addr);
3004 extern int in_gate_area(struct mm_struct *mm, unsigned long addr);
3010 static inline int in_gate_area_no_mm(unsigned long addr)
3014 static inline int in_gate_area(struct mm_struct *mm, unsigned long addr)
3047 pgd_t *vmemmap_pgd_populate(unsigned long addr, int node);
3048 p4d_t *vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node);
3049 pud_t *vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node);
3050 pmd_t *vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node);
3051 pte_t *vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node, struct vmem_altmap *altmap);