Lines Matching defs:page
33 #include <asm/page.h>
60 * bits determine what page size we use for kernel linear
73 * if these larger page sizes are not supported by the cpu.
132 /* Sanitize what we got from the firmware, by page aligning
180 struct page *mem_map_zero __read_mostly;
288 struct page *page;
290 page = pfn_to_page(pfn);
291 if (page) {
292 struct folio *folio = page_folio(page);
473 * is merely the zero page. The 'bigcore' testcase in GDB
896 * physical page address to get the true physical address.
1056 * It is, however, slow to check every single page if it matches a
1909 printk("Allocated %ld bytes for kernel page tables.\n",
1921 void __kernel_map_pages(struct page *page, int numpages, int enable)
1923 unsigned long phys_start = page_to_pfn(page) << PAGE_SHIFT;
1965 * address, so we can use all that our page tables
2009 * as 4-level page table cannot support more than
2265 /* paging_init() sets up the page tables */
2299 * When a page gets marked as dcache-dirty, we store the
2403 /* Set kernel pgd to upper alias so physical page computations
2528 * Set up the zero page, mark it reserved, so that page count
2529 * is not manipulated when freeing the page from user ptes.
2533 prom_printf("paging_init: Cannot alloc zero page.\n");
2551 * and if so the freeing below will free invalid page structs.
2562 unsigned long page;
2564 page = (addr +
2570 free_reserved_page(virt_to_page(page));
2812 pte_t mk_pte_io(unsigned long page, pgprot_t prot, int space, unsigned long page_size)
2816 pte_val(pte) = page | pgprot_val(pgprot_noncached(prot));
2855 * cheetah+ page size encodings.
2873 * cheetah+ page size encodings.
2899 struct page *page = alloc_page(GFP_KERNEL | __GFP_ZERO);
2902 if (page)
2903 pte = (pte_t *) page_address(page);
2950 struct page *page;
2952 page = container_of(head, struct page, rcu_head);
2953 __pte_free((pgtable_t)page_address(page));
2958 struct page *page;
2960 page = virt_to_page(pgtable);
2961 call_rcu(&page->rcu_head, pte_free_now);
3043 /* When changing the page size fields, we
3152 void copy_user_highpage(struct page *to, struct page *from,
3163 /* If this page has ADI enabled, copy over any ADI tags
3187 void copy_highpage(struct page *to, struct page *from)