Lines Matching refs:page
30 * This is called when relaxing access to a hugepage. It's also called in the page
88 * lookup in page tables with local interrupts disabled. For huge pages
90 * pmd_t we want to prevent transit from pmd pointing to page table
91 * to pmd pointing to huge page (and back) while interrupts are disabled.
92 * We clear pmd to possibly replace it with page table pointer in
126 * this PMD pte entry to a regular level 0 PTE by a parallel page fault.
147 pmd_t mk_pmd(struct page *page, pgprot_t pgprot)
149 return pfn_pmd(page_to_pfn(page), pgprot);
279 * If we have taken up all the fragments mark PTE page NULL
292 struct page *page;
297 page = alloc_page(gfp);
298 if (!page)
300 if (!pgtable_pmd_page_ctor(page)) {
301 __free_pages(page, 0);
305 atomic_set(&page->pt_frag_refcount, 1);
307 ret = page_address(page);
310 * allocated page.
318 * the allocated page with single fragement
322 atomic_set(&page->pt_frag_refcount, PMD_FRAG_NR);
343 struct page *page = virt_to_page(pmd);
345 if (PageReserved(page))
346 return free_reserved_page(page);
348 BUG_ON(atomic_read(&page->pt_frag_refcount) <= 0);
349 if (atomic_dec_and_test(&page->pt_frag_refcount)) {
350 pgtable_pmd_page_dtor(page);
351 __free_page(page);
456 * pmd page. Hence if we have different pmd page we need to withdraw during pmd