Lines Matching defs:nr_pages
65 unsigned long nr_pages);
450 unsigned long nr_pages;
478 nr_pages = DIV_ROUND_UP(m.num, XEN_PFN_PER_PAGE);
479 if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT)))
521 m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) {
526 ret = alloc_empty_pages(vma, nr_pages);
533 m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) {
537 if (privcmd_vma_range_is_mapped(vma, m.addr, nr_pages)) {
582 struct page *pages[], unsigned int nr_pages, unsigned int *pinned)
593 if (requested > nr_pages)
603 nr_pages -= page_count;
613 static void unlock_pages(struct page *pages[], unsigned int nr_pages)
615 unpin_user_pages_dirty_lock(pages, nr_pages, true);
623 unsigned int nr_pages = 0;
665 nr_pages += DIV_ROUND_UP(
670 pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL);
682 rc = lock_pages(kbufs, kdata.num, pages, nr_pages, &pinned);
959 unsigned long nr_pages)
961 return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT,