Lines Matching refs:end

254 				   unsigned long start, unsigned long end,
260 seq_put_hex_ll(m, "-", end, 8);
282 unsigned long start, end;
294 end = vma->vm_end;
295 CALL_HCK_LITE_HOOK(hideaddr_header_prefix_lhck, &start, &end, &flags, m, vma);
296 show_vma_header_prefix(m, start, end, flags, pgoff, dev, ino);
510 static int smaps_pte_hole(unsigned long addr, unsigned long end,
518 linear_page_index(vma, end));
623 static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
642 for (; addr != end; pte++, addr += PAGE_SIZE)
735 unsigned long addr, unsigned long end,
1172 unsigned long end, struct mm_walk *walk)
1206 for (; addr != end; pte++, addr += PAGE_SIZE) {
1231 static int clear_refs_test_walk(unsigned long start, unsigned long end,
1379 static int pagemap_pte_hole(unsigned long start, unsigned long end,
1386 while (addr < end) {
1393 hole_end = min(end, vma->vm_start);
1395 hole_end = end;
1409 for (; addr < min(end, vma->vm_end); addr += PAGE_SIZE) {
1473 static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end,
1534 for (; addr != end; addr += PAGE_SIZE) {
1553 * We can assume that @vma always points to a valid one and @end never
1561 for (; addr < end; pte++, addr += PAGE_SIZE) {
1579 unsigned long addr, unsigned long end,
1612 for (; addr != end; addr += PAGE_SIZE) {
1703 unsigned long end;
1711 end = start_vaddr + ((count / PM_ENTRY_BYTES) << PAGE_SHIFT);
1712 if (end >= start_vaddr && end < mm->task_size)
1713 end_vaddr = end;
1723 unsigned long end;
1726 end = (start_vaddr + PAGEMAP_WALK_SIZE) & PAGEMAP_WALK_MASK;
1728 if (end < start_vaddr || end > end_vaddr)
1729 end = end_vaddr;
1733 ret = walk_page_range(mm, start_vaddr, end, &pagemap_ops, &pm);
1735 start_vaddr = end;
1881 unsigned long end, struct mm_walk *walk)
1914 } while (pte++, addr += PAGE_SIZE, addr != end);
1921 unsigned long addr, unsigned long end, struct mm_walk *walk)
1939 unsigned long addr, unsigned long end, struct mm_walk *walk)