Lines Matching refs:start
347 * Decide whether we really walk over the current vma on [@start, @end)
352 static int walk_page_test(unsigned long start, unsigned long end,
359 return ops->test_walk(start, end, walk);
372 err = ops->pte_hole(start, end, -1, walk);
378 static int __walk_page_range(unsigned long start, unsigned long end,
386 err = ops->pre_vma(start, end, walk);
393 err = walk_hugetlb_range(start, end, walk);
395 err = walk_pgd_range(start, end, walk);
433 * @start: start address of the virtual address range
439 * within the virtual address range [@start, @end). During walking, we can do
470 int walk_page_range(struct mm_struct *mm, unsigned long start,
483 if (start >= end)
491 vma = find_vma(walk.mm, start);
497 err = ops->pte_hole(start, next, -1, &walk);
498 } else if (start < vma->vm_start) { /* outside vma */
502 err = ops->pte_hole(start, next, -1, &walk);
509 err = walk_page_test(start, next, &walk);
521 err = __walk_page_range(start, next, &walk);
525 } while (start = next, start < end);
532 * @start: start address of the virtual address range
543 int walk_page_range_novma(struct mm_struct *mm, unsigned long start,
556 if (start >= end || !walk.mm)
561 return walk_pgd_range(start, end, &walk);
564 int walk_page_range_vma(struct vm_area_struct *vma, unsigned long start,
575 if (start >= end || !walk.mm)
577 if (start < vma->vm_start || end > vma->vm_end)
582 return __walk_page_range(start, end, &walk);