Lines Matching refs:start
1707 void unmap_vmas(struct mmu_gather *tlb, struct vm_area_struct *start_vma, unsigned long start, unsigned long end);
1735 void unmap_mapping_pages(struct address_space *mapping, pgoff_t start, pgoff_t nr, bool even_cows);
1755 static inline void unmap_mapping_pages(struct address_space *mapping, pgoff_t start, pgoff_t nr, bool even_cows)
1775 long get_user_pages_remote(struct mm_struct *mm, unsigned long start, unsigned long nr_pages, unsigned int gup_flags,
1777 long pin_user_pages_remote(struct mm_struct *mm, unsigned long start, unsigned long nr_pages, unsigned int gup_flags,
1779 long get_user_pages(unsigned long start, unsigned long nr_pages, unsigned int gup_flags, struct page **pages,
1781 long pin_user_pages(unsigned long start, unsigned long nr_pages, unsigned int gup_flags, struct page **pages,
1783 long get_user_pages_locked(unsigned long start, unsigned long nr_pages, unsigned int gup_flags, struct page **pages,
1785 long pin_user_pages_locked(unsigned long start, unsigned long nr_pages, unsigned int gup_flags, struct page **pages,
1787 long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages, struct page **pages, unsigned int gup_flags);
1788 long pin_user_pages_unlocked(unsigned long start, unsigned long nr_pages, struct page **pages, unsigned int gup_flags);
1790 int get_user_pages_fast(unsigned long start, int nr_pages, unsigned int gup_flags, struct page **pages);
1791 int pin_user_pages_fast(unsigned long start, int nr_pages, unsigned int gup_flags, struct page **pages);
1810 int get_vaddr_frames(unsigned long start, unsigned int nr_pfns, unsigned int gup_flags, struct frame_vector *vec);
1841 int get_kernel_page(unsigned long start, int write, struct page **pages);
1886 extern unsigned long change_protection(struct vm_area_struct *vma, unsigned long start, unsigned long end,
1888 extern int mprotect_fixup(struct vm_area_struct *vma, struct vm_area_struct **pprev, unsigned long start,
1894 int get_user_pages_fast_only(unsigned long start, int nr_pages, unsigned int gup_flags, struct page **pages);
1895 int pin_user_pages_fast_only(unsigned long start, int nr_pages, unsigned int gup_flags, struct page **pages);
2402 * Free reserved pages within range [PAGE_ALIGN(start), end & PAGE_MASK)
2407 extern unsigned long free_reserved_area(void *start, void *end, int poison, const char *s);
2420 extern void reserve_bootmem_region(phys_addr_t start, phys_addr_t end);
2534 struct vm_area_struct *vma_interval_tree_iter_first(struct rb_root_cached *root, unsigned long start,
2536 struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node, unsigned long start,
2539 #define vma_interval_tree_foreach(vma, root, start, last) \
2540 for (vma = vma_interval_tree_iter_first(root, start, last); vma; \
2541 vma = vma_interval_tree_iter_next(vma, start, last))
2545 struct anon_vma_chain *anon_vma_interval_tree_iter_first(struct rb_root_cached *root, unsigned long start,
2547 struct anon_vma_chain *anon_vma_interval_tree_iter_next(struct anon_vma_chain *node, unsigned long start,
2553 #define anon_vma_interval_tree_foreach(avc, root, start, last) \
2554 for (avc = anon_vma_interval_tree_iter_first(root, start, last); avc; \
2555 avc = anon_vma_interval_tree_iter_next(avc, start, last))
2559 extern int __vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff,
2561 static inline int vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff,
2564 return __vma_adjust(vma, start, end, pgoff, insert, NULL);
2579 static inline int check_data_rlimit(unsigned long rlim, unsigned long new, unsigned long start, unsigned long end_data,
2583 if (((new - start) + (end_data - start_data)) > rlim) {
2609 unsigned long randomize_page(unsigned long start, unsigned long range);
2619 extern int do_madvise(struct mm_struct *mm, unsigned long start, size_t len_in, int behavior);
2739 static inline bool range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end)
2741 return (vma && vma->vm_start <= start && end <= vma->vm_end);
2759 unsigned long change_prot_numa(struct vm_area_struct *vma, unsigned long start, unsigned long end);
2773 int vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len);
2811 0x20 /* if a disk transfer is needed, start the IO \
3056 int vmemmap_populate_basepages(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap);
3057 int vmemmap_populate(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap);
3060 void vmemmap_free(unsigned long start, unsigned long end, struct vmem_altmap *altmap);
3194 pgoff_t bitmap_pgoff, unsigned long *bitmap, pgoff_t *start,
3237 int madvise_set_anon_name(struct mm_struct *mm, unsigned long start, unsigned long len_in,
3240 static inline int madvise_set_anon_name(struct mm_struct *mm, unsigned long start, unsigned long len_in,