18c2ecf20Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0
28c2ecf20Sopenharmony_ci#include <linux/mm.h>
38c2ecf20Sopenharmony_ci#include <linux/gfp.h>
48c2ecf20Sopenharmony_ci#include <linux/hugetlb.h>
58c2ecf20Sopenharmony_ci#include <asm/pgalloc.h>
68c2ecf20Sopenharmony_ci#include <asm/tlb.h>
78c2ecf20Sopenharmony_ci#include <asm/fixmap.h>
88c2ecf20Sopenharmony_ci#include <asm/mtrr.h>
98c2ecf20Sopenharmony_ci
108c2ecf20Sopenharmony_ci#ifdef CONFIG_DYNAMIC_PHYSICAL_MASK
118c2ecf20Sopenharmony_ciphys_addr_t physical_mask __ro_after_init = (1ULL << __PHYSICAL_MASK_SHIFT) - 1;
128c2ecf20Sopenharmony_ciEXPORT_SYMBOL(physical_mask);
138c2ecf20Sopenharmony_ci#endif
148c2ecf20Sopenharmony_ci
158c2ecf20Sopenharmony_ci#ifdef CONFIG_HIGHPTE
168c2ecf20Sopenharmony_ci#define PGTABLE_HIGHMEM __GFP_HIGHMEM
178c2ecf20Sopenharmony_ci#else
188c2ecf20Sopenharmony_ci#define PGTABLE_HIGHMEM 0
198c2ecf20Sopenharmony_ci#endif
208c2ecf20Sopenharmony_ci
218c2ecf20Sopenharmony_ci#ifndef CONFIG_PARAVIRT
228c2ecf20Sopenharmony_cistatic inline
238c2ecf20Sopenharmony_civoid paravirt_tlb_remove_table(struct mmu_gather *tlb, void *table)
248c2ecf20Sopenharmony_ci{
258c2ecf20Sopenharmony_ci	tlb_remove_page(tlb, table);
268c2ecf20Sopenharmony_ci}
278c2ecf20Sopenharmony_ci#endif
288c2ecf20Sopenharmony_ci
298c2ecf20Sopenharmony_cigfp_t __userpte_alloc_gfp = GFP_PGTABLE_USER | PGTABLE_HIGHMEM;
308c2ecf20Sopenharmony_ci
318c2ecf20Sopenharmony_cipgtable_t pte_alloc_one(struct mm_struct *mm)
328c2ecf20Sopenharmony_ci{
338c2ecf20Sopenharmony_ci	return __pte_alloc_one(mm, __userpte_alloc_gfp);
348c2ecf20Sopenharmony_ci}
358c2ecf20Sopenharmony_ci
368c2ecf20Sopenharmony_cistatic int __init setup_userpte(char *arg)
378c2ecf20Sopenharmony_ci{
388c2ecf20Sopenharmony_ci	if (!arg)
398c2ecf20Sopenharmony_ci		return -EINVAL;
408c2ecf20Sopenharmony_ci
418c2ecf20Sopenharmony_ci	/*
428c2ecf20Sopenharmony_ci	 * "userpte=nohigh" disables allocation of user pagetables in
438c2ecf20Sopenharmony_ci	 * high memory.
448c2ecf20Sopenharmony_ci	 */
458c2ecf20Sopenharmony_ci	if (strcmp(arg, "nohigh") == 0)
468c2ecf20Sopenharmony_ci		__userpte_alloc_gfp &= ~__GFP_HIGHMEM;
478c2ecf20Sopenharmony_ci	else
488c2ecf20Sopenharmony_ci		return -EINVAL;
498c2ecf20Sopenharmony_ci	return 0;
508c2ecf20Sopenharmony_ci}
518c2ecf20Sopenharmony_ciearly_param("userpte", setup_userpte);
528c2ecf20Sopenharmony_ci
538c2ecf20Sopenharmony_civoid ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte)
548c2ecf20Sopenharmony_ci{
558c2ecf20Sopenharmony_ci	pgtable_pte_page_dtor(pte);
568c2ecf20Sopenharmony_ci	paravirt_release_pte(page_to_pfn(pte));
578c2ecf20Sopenharmony_ci	paravirt_tlb_remove_table(tlb, pte);
588c2ecf20Sopenharmony_ci}
598c2ecf20Sopenharmony_ci
608c2ecf20Sopenharmony_ci#if CONFIG_PGTABLE_LEVELS > 2
618c2ecf20Sopenharmony_civoid ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd)
628c2ecf20Sopenharmony_ci{
638c2ecf20Sopenharmony_ci	struct page *page = virt_to_page(pmd);
648c2ecf20Sopenharmony_ci	paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT);
658c2ecf20Sopenharmony_ci	/*
668c2ecf20Sopenharmony_ci	 * NOTE! For PAE, any changes to the top page-directory-pointer-table
678c2ecf20Sopenharmony_ci	 * entries need a full cr3 reload to flush.
688c2ecf20Sopenharmony_ci	 */
698c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_PAE
708c2ecf20Sopenharmony_ci	tlb->need_flush_all = 1;
718c2ecf20Sopenharmony_ci#endif
728c2ecf20Sopenharmony_ci	pgtable_pmd_page_dtor(page);
738c2ecf20Sopenharmony_ci	paravirt_tlb_remove_table(tlb, page);
748c2ecf20Sopenharmony_ci}
758c2ecf20Sopenharmony_ci
768c2ecf20Sopenharmony_ci#if CONFIG_PGTABLE_LEVELS > 3
778c2ecf20Sopenharmony_civoid ___pud_free_tlb(struct mmu_gather *tlb, pud_t *pud)
788c2ecf20Sopenharmony_ci{
798c2ecf20Sopenharmony_ci	paravirt_release_pud(__pa(pud) >> PAGE_SHIFT);
808c2ecf20Sopenharmony_ci	paravirt_tlb_remove_table(tlb, virt_to_page(pud));
818c2ecf20Sopenharmony_ci}
828c2ecf20Sopenharmony_ci
838c2ecf20Sopenharmony_ci#if CONFIG_PGTABLE_LEVELS > 4
848c2ecf20Sopenharmony_civoid ___p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d)
858c2ecf20Sopenharmony_ci{
868c2ecf20Sopenharmony_ci	paravirt_release_p4d(__pa(p4d) >> PAGE_SHIFT);
878c2ecf20Sopenharmony_ci	paravirt_tlb_remove_table(tlb, virt_to_page(p4d));
888c2ecf20Sopenharmony_ci}
898c2ecf20Sopenharmony_ci#endif	/* CONFIG_PGTABLE_LEVELS > 4 */
908c2ecf20Sopenharmony_ci#endif	/* CONFIG_PGTABLE_LEVELS > 3 */
918c2ecf20Sopenharmony_ci#endif	/* CONFIG_PGTABLE_LEVELS > 2 */
928c2ecf20Sopenharmony_ci
938c2ecf20Sopenharmony_cistatic inline void pgd_list_add(pgd_t *pgd)
948c2ecf20Sopenharmony_ci{
958c2ecf20Sopenharmony_ci	struct page *page = virt_to_page(pgd);
968c2ecf20Sopenharmony_ci
978c2ecf20Sopenharmony_ci	list_add(&page->lru, &pgd_list);
988c2ecf20Sopenharmony_ci}
998c2ecf20Sopenharmony_ci
1008c2ecf20Sopenharmony_cistatic inline void pgd_list_del(pgd_t *pgd)
1018c2ecf20Sopenharmony_ci{
1028c2ecf20Sopenharmony_ci	struct page *page = virt_to_page(pgd);
1038c2ecf20Sopenharmony_ci
1048c2ecf20Sopenharmony_ci	list_del(&page->lru);
1058c2ecf20Sopenharmony_ci}
1068c2ecf20Sopenharmony_ci
1078c2ecf20Sopenharmony_ci#define UNSHARED_PTRS_PER_PGD				\
1088c2ecf20Sopenharmony_ci	(SHARED_KERNEL_PMD ? KERNEL_PGD_BOUNDARY : PTRS_PER_PGD)
1098c2ecf20Sopenharmony_ci#define MAX_UNSHARED_PTRS_PER_PGD			\
1108c2ecf20Sopenharmony_ci	max_t(size_t, KERNEL_PGD_BOUNDARY, PTRS_PER_PGD)
1118c2ecf20Sopenharmony_ci
1128c2ecf20Sopenharmony_ci
1138c2ecf20Sopenharmony_cistatic void pgd_set_mm(pgd_t *pgd, struct mm_struct *mm)
1148c2ecf20Sopenharmony_ci{
1158c2ecf20Sopenharmony_ci	virt_to_page(pgd)->pt_mm = mm;
1168c2ecf20Sopenharmony_ci}
1178c2ecf20Sopenharmony_ci
1188c2ecf20Sopenharmony_cistruct mm_struct *pgd_page_get_mm(struct page *page)
1198c2ecf20Sopenharmony_ci{
1208c2ecf20Sopenharmony_ci	return page->pt_mm;
1218c2ecf20Sopenharmony_ci}
1228c2ecf20Sopenharmony_ci
1238c2ecf20Sopenharmony_cistatic void pgd_ctor(struct mm_struct *mm, pgd_t *pgd)
1248c2ecf20Sopenharmony_ci{
1258c2ecf20Sopenharmony_ci	/* If the pgd points to a shared pagetable level (either the
1268c2ecf20Sopenharmony_ci	   ptes in non-PAE, or shared PMD in PAE), then just copy the
1278c2ecf20Sopenharmony_ci	   references from swapper_pg_dir. */
1288c2ecf20Sopenharmony_ci	if (CONFIG_PGTABLE_LEVELS == 2 ||
1298c2ecf20Sopenharmony_ci	    (CONFIG_PGTABLE_LEVELS == 3 && SHARED_KERNEL_PMD) ||
1308c2ecf20Sopenharmony_ci	    CONFIG_PGTABLE_LEVELS >= 4) {
1318c2ecf20Sopenharmony_ci		clone_pgd_range(pgd + KERNEL_PGD_BOUNDARY,
1328c2ecf20Sopenharmony_ci				swapper_pg_dir + KERNEL_PGD_BOUNDARY,
1338c2ecf20Sopenharmony_ci				KERNEL_PGD_PTRS);
1348c2ecf20Sopenharmony_ci	}
1358c2ecf20Sopenharmony_ci
1368c2ecf20Sopenharmony_ci	/* list required to sync kernel mapping updates */
1378c2ecf20Sopenharmony_ci	if (!SHARED_KERNEL_PMD) {
1388c2ecf20Sopenharmony_ci		pgd_set_mm(pgd, mm);
1398c2ecf20Sopenharmony_ci		pgd_list_add(pgd);
1408c2ecf20Sopenharmony_ci	}
1418c2ecf20Sopenharmony_ci}
1428c2ecf20Sopenharmony_ci
1438c2ecf20Sopenharmony_cistatic void pgd_dtor(pgd_t *pgd)
1448c2ecf20Sopenharmony_ci{
1458c2ecf20Sopenharmony_ci	if (SHARED_KERNEL_PMD)
1468c2ecf20Sopenharmony_ci		return;
1478c2ecf20Sopenharmony_ci
1488c2ecf20Sopenharmony_ci	spin_lock(&pgd_lock);
1498c2ecf20Sopenharmony_ci	pgd_list_del(pgd);
1508c2ecf20Sopenharmony_ci	spin_unlock(&pgd_lock);
1518c2ecf20Sopenharmony_ci}
1528c2ecf20Sopenharmony_ci
1538c2ecf20Sopenharmony_ci/*
1548c2ecf20Sopenharmony_ci * List of all pgd's needed for non-PAE so it can invalidate entries
1558c2ecf20Sopenharmony_ci * in both cached and uncached pgd's; not needed for PAE since the
1568c2ecf20Sopenharmony_ci * kernel pmd is shared. If PAE were not to share the pmd a similar
1578c2ecf20Sopenharmony_ci * tactic would be needed. This is essentially codepath-based locking
1588c2ecf20Sopenharmony_ci * against pageattr.c; it is the unique case in which a valid change
1598c2ecf20Sopenharmony_ci * of kernel pagetables can't be lazily synchronized by vmalloc faults.
1608c2ecf20Sopenharmony_ci * vmalloc faults work because attached pagetables are never freed.
1618c2ecf20Sopenharmony_ci * -- nyc
1628c2ecf20Sopenharmony_ci */
1638c2ecf20Sopenharmony_ci
1648c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_PAE
1658c2ecf20Sopenharmony_ci/*
1668c2ecf20Sopenharmony_ci * In PAE mode, we need to do a cr3 reload (=tlb flush) when
1678c2ecf20Sopenharmony_ci * updating the top-level pagetable entries to guarantee the
1688c2ecf20Sopenharmony_ci * processor notices the update.  Since this is expensive, and
1698c2ecf20Sopenharmony_ci * all 4 top-level entries are used almost immediately in a
1708c2ecf20Sopenharmony_ci * new process's life, we just pre-populate them here.
1718c2ecf20Sopenharmony_ci *
1728c2ecf20Sopenharmony_ci * Also, if we're in a paravirt environment where the kernel pmd is
1738c2ecf20Sopenharmony_ci * not shared between pagetables (!SHARED_KERNEL_PMDS), we allocate
1748c2ecf20Sopenharmony_ci * and initialize the kernel pmds here.
1758c2ecf20Sopenharmony_ci */
1768c2ecf20Sopenharmony_ci#define PREALLOCATED_PMDS	UNSHARED_PTRS_PER_PGD
1778c2ecf20Sopenharmony_ci#define MAX_PREALLOCATED_PMDS	MAX_UNSHARED_PTRS_PER_PGD
1788c2ecf20Sopenharmony_ci
1798c2ecf20Sopenharmony_ci/*
1808c2ecf20Sopenharmony_ci * We allocate separate PMDs for the kernel part of the user page-table
1818c2ecf20Sopenharmony_ci * when PTI is enabled. We need them to map the per-process LDT into the
1828c2ecf20Sopenharmony_ci * user-space page-table.
1838c2ecf20Sopenharmony_ci */
1848c2ecf20Sopenharmony_ci#define PREALLOCATED_USER_PMDS	 (boot_cpu_has(X86_FEATURE_PTI) ? \
1858c2ecf20Sopenharmony_ci					KERNEL_PGD_PTRS : 0)
1868c2ecf20Sopenharmony_ci#define MAX_PREALLOCATED_USER_PMDS KERNEL_PGD_PTRS
1878c2ecf20Sopenharmony_ci
1888c2ecf20Sopenharmony_civoid pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd)
1898c2ecf20Sopenharmony_ci{
1908c2ecf20Sopenharmony_ci	paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
1918c2ecf20Sopenharmony_ci
1928c2ecf20Sopenharmony_ci	/* Note: almost everything apart from _PAGE_PRESENT is
1938c2ecf20Sopenharmony_ci	   reserved at the pmd (PDPT) level. */
1948c2ecf20Sopenharmony_ci	set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT));
1958c2ecf20Sopenharmony_ci
1968c2ecf20Sopenharmony_ci	/*
1978c2ecf20Sopenharmony_ci	 * According to Intel App note "TLBs, Paging-Structure Caches,
1988c2ecf20Sopenharmony_ci	 * and Their Invalidation", April 2007, document 317080-001,
1998c2ecf20Sopenharmony_ci	 * section 8.1: in PAE mode we explicitly have to flush the
2008c2ecf20Sopenharmony_ci	 * TLB via cr3 if the top-level pgd is changed...
2018c2ecf20Sopenharmony_ci	 */
2028c2ecf20Sopenharmony_ci	flush_tlb_mm(mm);
2038c2ecf20Sopenharmony_ci}
2048c2ecf20Sopenharmony_ci#else  /* !CONFIG_X86_PAE */
2058c2ecf20Sopenharmony_ci
2068c2ecf20Sopenharmony_ci/* No need to prepopulate any pagetable entries in non-PAE modes. */
2078c2ecf20Sopenharmony_ci#define PREALLOCATED_PMDS	0
2088c2ecf20Sopenharmony_ci#define MAX_PREALLOCATED_PMDS	0
2098c2ecf20Sopenharmony_ci#define PREALLOCATED_USER_PMDS	 0
2108c2ecf20Sopenharmony_ci#define MAX_PREALLOCATED_USER_PMDS 0
2118c2ecf20Sopenharmony_ci#endif	/* CONFIG_X86_PAE */
2128c2ecf20Sopenharmony_ci
2138c2ecf20Sopenharmony_cistatic void free_pmds(struct mm_struct *mm, pmd_t *pmds[], int count)
2148c2ecf20Sopenharmony_ci{
2158c2ecf20Sopenharmony_ci	int i;
2168c2ecf20Sopenharmony_ci
2178c2ecf20Sopenharmony_ci	for (i = 0; i < count; i++)
2188c2ecf20Sopenharmony_ci		if (pmds[i]) {
2198c2ecf20Sopenharmony_ci			pgtable_pmd_page_dtor(virt_to_page(pmds[i]));
2208c2ecf20Sopenharmony_ci			free_page((unsigned long)pmds[i]);
2218c2ecf20Sopenharmony_ci			mm_dec_nr_pmds(mm);
2228c2ecf20Sopenharmony_ci		}
2238c2ecf20Sopenharmony_ci}
2248c2ecf20Sopenharmony_ci
2258c2ecf20Sopenharmony_cistatic int preallocate_pmds(struct mm_struct *mm, pmd_t *pmds[], int count)
2268c2ecf20Sopenharmony_ci{
2278c2ecf20Sopenharmony_ci	int i;
2288c2ecf20Sopenharmony_ci	bool failed = false;
2298c2ecf20Sopenharmony_ci	gfp_t gfp = GFP_PGTABLE_USER;
2308c2ecf20Sopenharmony_ci
2318c2ecf20Sopenharmony_ci	if (mm == &init_mm)
2328c2ecf20Sopenharmony_ci		gfp &= ~__GFP_ACCOUNT;
2338c2ecf20Sopenharmony_ci
2348c2ecf20Sopenharmony_ci	for (i = 0; i < count; i++) {
2358c2ecf20Sopenharmony_ci		pmd_t *pmd = (pmd_t *)__get_free_page(gfp);
2368c2ecf20Sopenharmony_ci		if (!pmd)
2378c2ecf20Sopenharmony_ci			failed = true;
2388c2ecf20Sopenharmony_ci		if (pmd && !pgtable_pmd_page_ctor(virt_to_page(pmd))) {
2398c2ecf20Sopenharmony_ci			free_page((unsigned long)pmd);
2408c2ecf20Sopenharmony_ci			pmd = NULL;
2418c2ecf20Sopenharmony_ci			failed = true;
2428c2ecf20Sopenharmony_ci		}
2438c2ecf20Sopenharmony_ci		if (pmd)
2448c2ecf20Sopenharmony_ci			mm_inc_nr_pmds(mm);
2458c2ecf20Sopenharmony_ci		pmds[i] = pmd;
2468c2ecf20Sopenharmony_ci	}
2478c2ecf20Sopenharmony_ci
2488c2ecf20Sopenharmony_ci	if (failed) {
2498c2ecf20Sopenharmony_ci		free_pmds(mm, pmds, count);
2508c2ecf20Sopenharmony_ci		return -ENOMEM;
2518c2ecf20Sopenharmony_ci	}
2528c2ecf20Sopenharmony_ci
2538c2ecf20Sopenharmony_ci	return 0;
2548c2ecf20Sopenharmony_ci}
2558c2ecf20Sopenharmony_ci
2568c2ecf20Sopenharmony_ci/*
2578c2ecf20Sopenharmony_ci * Mop up any pmd pages which may still be attached to the pgd.
2588c2ecf20Sopenharmony_ci * Normally they will be freed by munmap/exit_mmap, but any pmd we
2598c2ecf20Sopenharmony_ci * preallocate which never got a corresponding vma will need to be
2608c2ecf20Sopenharmony_ci * freed manually.
2618c2ecf20Sopenharmony_ci */
2628c2ecf20Sopenharmony_cistatic void mop_up_one_pmd(struct mm_struct *mm, pgd_t *pgdp)
2638c2ecf20Sopenharmony_ci{
2648c2ecf20Sopenharmony_ci	pgd_t pgd = *pgdp;
2658c2ecf20Sopenharmony_ci
2668c2ecf20Sopenharmony_ci	if (pgd_val(pgd) != 0) {
2678c2ecf20Sopenharmony_ci		pmd_t *pmd = (pmd_t *)pgd_page_vaddr(pgd);
2688c2ecf20Sopenharmony_ci
2698c2ecf20Sopenharmony_ci		pgd_clear(pgdp);
2708c2ecf20Sopenharmony_ci
2718c2ecf20Sopenharmony_ci		paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT);
2728c2ecf20Sopenharmony_ci		pmd_free(mm, pmd);
2738c2ecf20Sopenharmony_ci		mm_dec_nr_pmds(mm);
2748c2ecf20Sopenharmony_ci	}
2758c2ecf20Sopenharmony_ci}
2768c2ecf20Sopenharmony_ci
2778c2ecf20Sopenharmony_cistatic void pgd_mop_up_pmds(struct mm_struct *mm, pgd_t *pgdp)
2788c2ecf20Sopenharmony_ci{
2798c2ecf20Sopenharmony_ci	int i;
2808c2ecf20Sopenharmony_ci
2818c2ecf20Sopenharmony_ci	for (i = 0; i < PREALLOCATED_PMDS; i++)
2828c2ecf20Sopenharmony_ci		mop_up_one_pmd(mm, &pgdp[i]);
2838c2ecf20Sopenharmony_ci
2848c2ecf20Sopenharmony_ci#ifdef CONFIG_PAGE_TABLE_ISOLATION
2858c2ecf20Sopenharmony_ci
2868c2ecf20Sopenharmony_ci	if (!boot_cpu_has(X86_FEATURE_PTI))
2878c2ecf20Sopenharmony_ci		return;
2888c2ecf20Sopenharmony_ci
2898c2ecf20Sopenharmony_ci	pgdp = kernel_to_user_pgdp(pgdp);
2908c2ecf20Sopenharmony_ci
2918c2ecf20Sopenharmony_ci	for (i = 0; i < PREALLOCATED_USER_PMDS; i++)
2928c2ecf20Sopenharmony_ci		mop_up_one_pmd(mm, &pgdp[i + KERNEL_PGD_BOUNDARY]);
2938c2ecf20Sopenharmony_ci#endif
2948c2ecf20Sopenharmony_ci}
2958c2ecf20Sopenharmony_ci
2968c2ecf20Sopenharmony_cistatic void pgd_prepopulate_pmd(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmds[])
2978c2ecf20Sopenharmony_ci{
2988c2ecf20Sopenharmony_ci	p4d_t *p4d;
2998c2ecf20Sopenharmony_ci	pud_t *pud;
3008c2ecf20Sopenharmony_ci	int i;
3018c2ecf20Sopenharmony_ci
3028c2ecf20Sopenharmony_ci	if (PREALLOCATED_PMDS == 0) /* Work around gcc-3.4.x bug */
3038c2ecf20Sopenharmony_ci		return;
3048c2ecf20Sopenharmony_ci
3058c2ecf20Sopenharmony_ci	p4d = p4d_offset(pgd, 0);
3068c2ecf20Sopenharmony_ci	pud = pud_offset(p4d, 0);
3078c2ecf20Sopenharmony_ci
3088c2ecf20Sopenharmony_ci	for (i = 0; i < PREALLOCATED_PMDS; i++, pud++) {
3098c2ecf20Sopenharmony_ci		pmd_t *pmd = pmds[i];
3108c2ecf20Sopenharmony_ci
3118c2ecf20Sopenharmony_ci		if (i >= KERNEL_PGD_BOUNDARY)
3128c2ecf20Sopenharmony_ci			memcpy(pmd, (pmd_t *)pgd_page_vaddr(swapper_pg_dir[i]),
3138c2ecf20Sopenharmony_ci			       sizeof(pmd_t) * PTRS_PER_PMD);
3148c2ecf20Sopenharmony_ci
3158c2ecf20Sopenharmony_ci		pud_populate(mm, pud, pmd);
3168c2ecf20Sopenharmony_ci	}
3178c2ecf20Sopenharmony_ci}
3188c2ecf20Sopenharmony_ci
3198c2ecf20Sopenharmony_ci#ifdef CONFIG_PAGE_TABLE_ISOLATION
3208c2ecf20Sopenharmony_cistatic void pgd_prepopulate_user_pmd(struct mm_struct *mm,
3218c2ecf20Sopenharmony_ci				     pgd_t *k_pgd, pmd_t *pmds[])
3228c2ecf20Sopenharmony_ci{
3238c2ecf20Sopenharmony_ci	pgd_t *s_pgd = kernel_to_user_pgdp(swapper_pg_dir);
3248c2ecf20Sopenharmony_ci	pgd_t *u_pgd = kernel_to_user_pgdp(k_pgd);
3258c2ecf20Sopenharmony_ci	p4d_t *u_p4d;
3268c2ecf20Sopenharmony_ci	pud_t *u_pud;
3278c2ecf20Sopenharmony_ci	int i;
3288c2ecf20Sopenharmony_ci
3298c2ecf20Sopenharmony_ci	u_p4d = p4d_offset(u_pgd, 0);
3308c2ecf20Sopenharmony_ci	u_pud = pud_offset(u_p4d, 0);
3318c2ecf20Sopenharmony_ci
3328c2ecf20Sopenharmony_ci	s_pgd += KERNEL_PGD_BOUNDARY;
3338c2ecf20Sopenharmony_ci	u_pud += KERNEL_PGD_BOUNDARY;
3348c2ecf20Sopenharmony_ci
3358c2ecf20Sopenharmony_ci	for (i = 0; i < PREALLOCATED_USER_PMDS; i++, u_pud++, s_pgd++) {
3368c2ecf20Sopenharmony_ci		pmd_t *pmd = pmds[i];
3378c2ecf20Sopenharmony_ci
3388c2ecf20Sopenharmony_ci		memcpy(pmd, (pmd_t *)pgd_page_vaddr(*s_pgd),
3398c2ecf20Sopenharmony_ci		       sizeof(pmd_t) * PTRS_PER_PMD);
3408c2ecf20Sopenharmony_ci
3418c2ecf20Sopenharmony_ci		pud_populate(mm, u_pud, pmd);
3428c2ecf20Sopenharmony_ci	}
3438c2ecf20Sopenharmony_ci
3448c2ecf20Sopenharmony_ci}
3458c2ecf20Sopenharmony_ci#else
3468c2ecf20Sopenharmony_cistatic void pgd_prepopulate_user_pmd(struct mm_struct *mm,
3478c2ecf20Sopenharmony_ci				     pgd_t *k_pgd, pmd_t *pmds[])
3488c2ecf20Sopenharmony_ci{
3498c2ecf20Sopenharmony_ci}
3508c2ecf20Sopenharmony_ci#endif
3518c2ecf20Sopenharmony_ci/*
3528c2ecf20Sopenharmony_ci * Xen paravirt assumes pgd table should be in one page. 64 bit kernel also
3538c2ecf20Sopenharmony_ci * assumes that pgd should be in one page.
3548c2ecf20Sopenharmony_ci *
3558c2ecf20Sopenharmony_ci * But kernel with PAE paging that is not running as a Xen domain
3568c2ecf20Sopenharmony_ci * only needs to allocate 32 bytes for pgd instead of one page.
3578c2ecf20Sopenharmony_ci */
3588c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_PAE
3598c2ecf20Sopenharmony_ci
3608c2ecf20Sopenharmony_ci#include <linux/slab.h>
3618c2ecf20Sopenharmony_ci
3628c2ecf20Sopenharmony_ci#define PGD_SIZE	(PTRS_PER_PGD * sizeof(pgd_t))
3638c2ecf20Sopenharmony_ci#define PGD_ALIGN	32
3648c2ecf20Sopenharmony_ci
3658c2ecf20Sopenharmony_cistatic struct kmem_cache *pgd_cache;
3668c2ecf20Sopenharmony_ci
3678c2ecf20Sopenharmony_civoid __init pgtable_cache_init(void)
3688c2ecf20Sopenharmony_ci{
3698c2ecf20Sopenharmony_ci	/*
3708c2ecf20Sopenharmony_ci	 * When PAE kernel is running as a Xen domain, it does not use
3718c2ecf20Sopenharmony_ci	 * shared kernel pmd. And this requires a whole page for pgd.
3728c2ecf20Sopenharmony_ci	 */
3738c2ecf20Sopenharmony_ci	if (!SHARED_KERNEL_PMD)
3748c2ecf20Sopenharmony_ci		return;
3758c2ecf20Sopenharmony_ci
3768c2ecf20Sopenharmony_ci	/*
3778c2ecf20Sopenharmony_ci	 * when PAE kernel is not running as a Xen domain, it uses
3788c2ecf20Sopenharmony_ci	 * shared kernel pmd. Shared kernel pmd does not require a whole
3798c2ecf20Sopenharmony_ci	 * page for pgd. We are able to just allocate a 32-byte for pgd.
3808c2ecf20Sopenharmony_ci	 * During boot time, we create a 32-byte slab for pgd table allocation.
3818c2ecf20Sopenharmony_ci	 */
3828c2ecf20Sopenharmony_ci	pgd_cache = kmem_cache_create("pgd_cache", PGD_SIZE, PGD_ALIGN,
3838c2ecf20Sopenharmony_ci				      SLAB_PANIC, NULL);
3848c2ecf20Sopenharmony_ci}
3858c2ecf20Sopenharmony_ci
3868c2ecf20Sopenharmony_cistatic inline pgd_t *_pgd_alloc(void)
3878c2ecf20Sopenharmony_ci{
3888c2ecf20Sopenharmony_ci	/*
3898c2ecf20Sopenharmony_ci	 * If no SHARED_KERNEL_PMD, PAE kernel is running as a Xen domain.
3908c2ecf20Sopenharmony_ci	 * We allocate one page for pgd.
3918c2ecf20Sopenharmony_ci	 */
3928c2ecf20Sopenharmony_ci	if (!SHARED_KERNEL_PMD)
3938c2ecf20Sopenharmony_ci		return (pgd_t *)__get_free_pages(GFP_PGTABLE_USER,
3948c2ecf20Sopenharmony_ci						 PGD_ALLOCATION_ORDER);
3958c2ecf20Sopenharmony_ci
3968c2ecf20Sopenharmony_ci	/*
3978c2ecf20Sopenharmony_ci	 * Now PAE kernel is not running as a Xen domain. We can allocate
3988c2ecf20Sopenharmony_ci	 * a 32-byte slab for pgd to save memory space.
3998c2ecf20Sopenharmony_ci	 */
4008c2ecf20Sopenharmony_ci	return kmem_cache_alloc(pgd_cache, GFP_PGTABLE_USER);
4018c2ecf20Sopenharmony_ci}
4028c2ecf20Sopenharmony_ci
4038c2ecf20Sopenharmony_cistatic inline void _pgd_free(pgd_t *pgd)
4048c2ecf20Sopenharmony_ci{
4058c2ecf20Sopenharmony_ci	if (!SHARED_KERNEL_PMD)
4068c2ecf20Sopenharmony_ci		free_pages((unsigned long)pgd, PGD_ALLOCATION_ORDER);
4078c2ecf20Sopenharmony_ci	else
4088c2ecf20Sopenharmony_ci		kmem_cache_free(pgd_cache, pgd);
4098c2ecf20Sopenharmony_ci}
4108c2ecf20Sopenharmony_ci#else
4118c2ecf20Sopenharmony_ci
4128c2ecf20Sopenharmony_cistatic inline pgd_t *_pgd_alloc(void)
4138c2ecf20Sopenharmony_ci{
4148c2ecf20Sopenharmony_ci	return (pgd_t *)__get_free_pages(GFP_PGTABLE_USER,
4158c2ecf20Sopenharmony_ci					 PGD_ALLOCATION_ORDER);
4168c2ecf20Sopenharmony_ci}
4178c2ecf20Sopenharmony_ci
4188c2ecf20Sopenharmony_cistatic inline void _pgd_free(pgd_t *pgd)
4198c2ecf20Sopenharmony_ci{
4208c2ecf20Sopenharmony_ci	free_pages((unsigned long)pgd, PGD_ALLOCATION_ORDER);
4218c2ecf20Sopenharmony_ci}
4228c2ecf20Sopenharmony_ci#endif /* CONFIG_X86_PAE */
4238c2ecf20Sopenharmony_ci
4248c2ecf20Sopenharmony_cipgd_t *pgd_alloc(struct mm_struct *mm)
4258c2ecf20Sopenharmony_ci{
4268c2ecf20Sopenharmony_ci	pgd_t *pgd;
4278c2ecf20Sopenharmony_ci	pmd_t *u_pmds[MAX_PREALLOCATED_USER_PMDS];
4288c2ecf20Sopenharmony_ci	pmd_t *pmds[MAX_PREALLOCATED_PMDS];
4298c2ecf20Sopenharmony_ci
4308c2ecf20Sopenharmony_ci	pgd = _pgd_alloc();
4318c2ecf20Sopenharmony_ci
4328c2ecf20Sopenharmony_ci	if (pgd == NULL)
4338c2ecf20Sopenharmony_ci		goto out;
4348c2ecf20Sopenharmony_ci
4358c2ecf20Sopenharmony_ci	mm->pgd = pgd;
4368c2ecf20Sopenharmony_ci
4378c2ecf20Sopenharmony_ci	if (preallocate_pmds(mm, pmds, PREALLOCATED_PMDS) != 0)
4388c2ecf20Sopenharmony_ci		goto out_free_pgd;
4398c2ecf20Sopenharmony_ci
4408c2ecf20Sopenharmony_ci	if (preallocate_pmds(mm, u_pmds, PREALLOCATED_USER_PMDS) != 0)
4418c2ecf20Sopenharmony_ci		goto out_free_pmds;
4428c2ecf20Sopenharmony_ci
4438c2ecf20Sopenharmony_ci	if (paravirt_pgd_alloc(mm) != 0)
4448c2ecf20Sopenharmony_ci		goto out_free_user_pmds;
4458c2ecf20Sopenharmony_ci
4468c2ecf20Sopenharmony_ci	/*
4478c2ecf20Sopenharmony_ci	 * Make sure that pre-populating the pmds is atomic with
4488c2ecf20Sopenharmony_ci	 * respect to anything walking the pgd_list, so that they
4498c2ecf20Sopenharmony_ci	 * never see a partially populated pgd.
4508c2ecf20Sopenharmony_ci	 */
4518c2ecf20Sopenharmony_ci	spin_lock(&pgd_lock);
4528c2ecf20Sopenharmony_ci
4538c2ecf20Sopenharmony_ci	pgd_ctor(mm, pgd);
4548c2ecf20Sopenharmony_ci	pgd_prepopulate_pmd(mm, pgd, pmds);
4558c2ecf20Sopenharmony_ci	pgd_prepopulate_user_pmd(mm, pgd, u_pmds);
4568c2ecf20Sopenharmony_ci
4578c2ecf20Sopenharmony_ci	spin_unlock(&pgd_lock);
4588c2ecf20Sopenharmony_ci
4598c2ecf20Sopenharmony_ci	return pgd;
4608c2ecf20Sopenharmony_ci
4618c2ecf20Sopenharmony_ciout_free_user_pmds:
4628c2ecf20Sopenharmony_ci	free_pmds(mm, u_pmds, PREALLOCATED_USER_PMDS);
4638c2ecf20Sopenharmony_ciout_free_pmds:
4648c2ecf20Sopenharmony_ci	free_pmds(mm, pmds, PREALLOCATED_PMDS);
4658c2ecf20Sopenharmony_ciout_free_pgd:
4668c2ecf20Sopenharmony_ci	_pgd_free(pgd);
4678c2ecf20Sopenharmony_ciout:
4688c2ecf20Sopenharmony_ci	return NULL;
4698c2ecf20Sopenharmony_ci}
4708c2ecf20Sopenharmony_ci
4718c2ecf20Sopenharmony_civoid pgd_free(struct mm_struct *mm, pgd_t *pgd)
4728c2ecf20Sopenharmony_ci{
4738c2ecf20Sopenharmony_ci	pgd_mop_up_pmds(mm, pgd);
4748c2ecf20Sopenharmony_ci	pgd_dtor(pgd);
4758c2ecf20Sopenharmony_ci	paravirt_pgd_free(mm, pgd);
4768c2ecf20Sopenharmony_ci	_pgd_free(pgd);
4778c2ecf20Sopenharmony_ci}
4788c2ecf20Sopenharmony_ci
4798c2ecf20Sopenharmony_ci/*
4808c2ecf20Sopenharmony_ci * Used to set accessed or dirty bits in the page table entries
4818c2ecf20Sopenharmony_ci * on other architectures. On x86, the accessed and dirty bits
4828c2ecf20Sopenharmony_ci * are tracked by hardware. However, do_wp_page calls this function
4838c2ecf20Sopenharmony_ci * to also make the pte writeable at the same time the dirty bit is
4848c2ecf20Sopenharmony_ci * set. In that case we do actually need to write the PTE.
4858c2ecf20Sopenharmony_ci */
4868c2ecf20Sopenharmony_ciint ptep_set_access_flags(struct vm_area_struct *vma,
4878c2ecf20Sopenharmony_ci			  unsigned long address, pte_t *ptep,
4888c2ecf20Sopenharmony_ci			  pte_t entry, int dirty)
4898c2ecf20Sopenharmony_ci{
4908c2ecf20Sopenharmony_ci	int changed = !pte_same(*ptep, entry);
4918c2ecf20Sopenharmony_ci
4928c2ecf20Sopenharmony_ci	if (changed && dirty)
4938c2ecf20Sopenharmony_ci		set_pte(ptep, entry);
4948c2ecf20Sopenharmony_ci
4958c2ecf20Sopenharmony_ci	return changed;
4968c2ecf20Sopenharmony_ci}
4978c2ecf20Sopenharmony_ci
4988c2ecf20Sopenharmony_ci#ifdef CONFIG_TRANSPARENT_HUGEPAGE
4998c2ecf20Sopenharmony_ciint pmdp_set_access_flags(struct vm_area_struct *vma,
5008c2ecf20Sopenharmony_ci			  unsigned long address, pmd_t *pmdp,
5018c2ecf20Sopenharmony_ci			  pmd_t entry, int dirty)
5028c2ecf20Sopenharmony_ci{
5038c2ecf20Sopenharmony_ci	int changed = !pmd_same(*pmdp, entry);
5048c2ecf20Sopenharmony_ci
5058c2ecf20Sopenharmony_ci	VM_BUG_ON(address & ~HPAGE_PMD_MASK);
5068c2ecf20Sopenharmony_ci
5078c2ecf20Sopenharmony_ci	if (changed && dirty) {
5088c2ecf20Sopenharmony_ci		set_pmd(pmdp, entry);
5098c2ecf20Sopenharmony_ci		/*
5108c2ecf20Sopenharmony_ci		 * We had a write-protection fault here and changed the pmd
5118c2ecf20Sopenharmony_ci		 * to to more permissive. No need to flush the TLB for that,
5128c2ecf20Sopenharmony_ci		 * #PF is architecturally guaranteed to do that and in the
5138c2ecf20Sopenharmony_ci		 * worst-case we'll generate a spurious fault.
5148c2ecf20Sopenharmony_ci		 */
5158c2ecf20Sopenharmony_ci	}
5168c2ecf20Sopenharmony_ci
5178c2ecf20Sopenharmony_ci	return changed;
5188c2ecf20Sopenharmony_ci}
5198c2ecf20Sopenharmony_ci
5208c2ecf20Sopenharmony_ciint pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address,
5218c2ecf20Sopenharmony_ci			  pud_t *pudp, pud_t entry, int dirty)
5228c2ecf20Sopenharmony_ci{
5238c2ecf20Sopenharmony_ci	int changed = !pud_same(*pudp, entry);
5248c2ecf20Sopenharmony_ci
5258c2ecf20Sopenharmony_ci	VM_BUG_ON(address & ~HPAGE_PUD_MASK);
5268c2ecf20Sopenharmony_ci
5278c2ecf20Sopenharmony_ci	if (changed && dirty) {
5288c2ecf20Sopenharmony_ci		set_pud(pudp, entry);
5298c2ecf20Sopenharmony_ci		/*
5308c2ecf20Sopenharmony_ci		 * We had a write-protection fault here and changed the pud
5318c2ecf20Sopenharmony_ci		 * to to more permissive. No need to flush the TLB for that,
5328c2ecf20Sopenharmony_ci		 * #PF is architecturally guaranteed to do that and in the
5338c2ecf20Sopenharmony_ci		 * worst-case we'll generate a spurious fault.
5348c2ecf20Sopenharmony_ci		 */
5358c2ecf20Sopenharmony_ci	}
5368c2ecf20Sopenharmony_ci
5378c2ecf20Sopenharmony_ci	return changed;
5388c2ecf20Sopenharmony_ci}
5398c2ecf20Sopenharmony_ci#endif
5408c2ecf20Sopenharmony_ci
5418c2ecf20Sopenharmony_ciint ptep_test_and_clear_young(struct vm_area_struct *vma,
5428c2ecf20Sopenharmony_ci			      unsigned long addr, pte_t *ptep)
5438c2ecf20Sopenharmony_ci{
5448c2ecf20Sopenharmony_ci	int ret = 0;
5458c2ecf20Sopenharmony_ci
5468c2ecf20Sopenharmony_ci	if (pte_young(*ptep))
5478c2ecf20Sopenharmony_ci		ret = test_and_clear_bit(_PAGE_BIT_ACCESSED,
5488c2ecf20Sopenharmony_ci					 (unsigned long *) &ptep->pte);
5498c2ecf20Sopenharmony_ci
5508c2ecf20Sopenharmony_ci	return ret;
5518c2ecf20Sopenharmony_ci}
5528c2ecf20Sopenharmony_ci
5538c2ecf20Sopenharmony_ci#ifdef CONFIG_TRANSPARENT_HUGEPAGE
5548c2ecf20Sopenharmony_ciint pmdp_test_and_clear_young(struct vm_area_struct *vma,
5558c2ecf20Sopenharmony_ci			      unsigned long addr, pmd_t *pmdp)
5568c2ecf20Sopenharmony_ci{
5578c2ecf20Sopenharmony_ci	int ret = 0;
5588c2ecf20Sopenharmony_ci
5598c2ecf20Sopenharmony_ci	if (pmd_young(*pmdp))
5608c2ecf20Sopenharmony_ci		ret = test_and_clear_bit(_PAGE_BIT_ACCESSED,
5618c2ecf20Sopenharmony_ci					 (unsigned long *)pmdp);
5628c2ecf20Sopenharmony_ci
5638c2ecf20Sopenharmony_ci	return ret;
5648c2ecf20Sopenharmony_ci}
5658c2ecf20Sopenharmony_ciint pudp_test_and_clear_young(struct vm_area_struct *vma,
5668c2ecf20Sopenharmony_ci			      unsigned long addr, pud_t *pudp)
5678c2ecf20Sopenharmony_ci{
5688c2ecf20Sopenharmony_ci	int ret = 0;
5698c2ecf20Sopenharmony_ci
5708c2ecf20Sopenharmony_ci	if (pud_young(*pudp))
5718c2ecf20Sopenharmony_ci		ret = test_and_clear_bit(_PAGE_BIT_ACCESSED,
5728c2ecf20Sopenharmony_ci					 (unsigned long *)pudp);
5738c2ecf20Sopenharmony_ci
5748c2ecf20Sopenharmony_ci	return ret;
5758c2ecf20Sopenharmony_ci}
5768c2ecf20Sopenharmony_ci#endif
5778c2ecf20Sopenharmony_ci
5788c2ecf20Sopenharmony_ciint ptep_clear_flush_young(struct vm_area_struct *vma,
5798c2ecf20Sopenharmony_ci			   unsigned long address, pte_t *ptep)
5808c2ecf20Sopenharmony_ci{
5818c2ecf20Sopenharmony_ci	/*
5828c2ecf20Sopenharmony_ci	 * On x86 CPUs, clearing the accessed bit without a TLB flush
5838c2ecf20Sopenharmony_ci	 * doesn't cause data corruption. [ It could cause incorrect
5848c2ecf20Sopenharmony_ci	 * page aging and the (mistaken) reclaim of hot pages, but the
5858c2ecf20Sopenharmony_ci	 * chance of that should be relatively low. ]
5868c2ecf20Sopenharmony_ci	 *
5878c2ecf20Sopenharmony_ci	 * So as a performance optimization don't flush the TLB when
5888c2ecf20Sopenharmony_ci	 * clearing the accessed bit, it will eventually be flushed by
5898c2ecf20Sopenharmony_ci	 * a context switch or a VM operation anyway. [ In the rare
5908c2ecf20Sopenharmony_ci	 * event of it not getting flushed for a long time the delay
5918c2ecf20Sopenharmony_ci	 * shouldn't really matter because there's no real memory
5928c2ecf20Sopenharmony_ci	 * pressure for swapout to react to. ]
5938c2ecf20Sopenharmony_ci	 */
5948c2ecf20Sopenharmony_ci	return ptep_test_and_clear_young(vma, address, ptep);
5958c2ecf20Sopenharmony_ci}
5968c2ecf20Sopenharmony_ci
5978c2ecf20Sopenharmony_ci#ifdef CONFIG_TRANSPARENT_HUGEPAGE
5988c2ecf20Sopenharmony_ciint pmdp_clear_flush_young(struct vm_area_struct *vma,
5998c2ecf20Sopenharmony_ci			   unsigned long address, pmd_t *pmdp)
6008c2ecf20Sopenharmony_ci{
6018c2ecf20Sopenharmony_ci	int young;
6028c2ecf20Sopenharmony_ci
6038c2ecf20Sopenharmony_ci	VM_BUG_ON(address & ~HPAGE_PMD_MASK);
6048c2ecf20Sopenharmony_ci
6058c2ecf20Sopenharmony_ci	young = pmdp_test_and_clear_young(vma, address, pmdp);
6068c2ecf20Sopenharmony_ci	if (young)
6078c2ecf20Sopenharmony_ci		flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
6088c2ecf20Sopenharmony_ci
6098c2ecf20Sopenharmony_ci	return young;
6108c2ecf20Sopenharmony_ci}
6118c2ecf20Sopenharmony_ci#endif
6128c2ecf20Sopenharmony_ci
6138c2ecf20Sopenharmony_ci/**
6148c2ecf20Sopenharmony_ci * reserve_top_address - reserves a hole in the top of kernel address space
6158c2ecf20Sopenharmony_ci * @reserve - size of hole to reserve
6168c2ecf20Sopenharmony_ci *
6178c2ecf20Sopenharmony_ci * Can be used to relocate the fixmap area and poke a hole in the top
6188c2ecf20Sopenharmony_ci * of kernel address space to make room for a hypervisor.
6198c2ecf20Sopenharmony_ci */
6208c2ecf20Sopenharmony_civoid __init reserve_top_address(unsigned long reserve)
6218c2ecf20Sopenharmony_ci{
6228c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_32
6238c2ecf20Sopenharmony_ci	BUG_ON(fixmaps_set > 0);
6248c2ecf20Sopenharmony_ci	__FIXADDR_TOP = round_down(-reserve, 1 << PMD_SHIFT) - PAGE_SIZE;
6258c2ecf20Sopenharmony_ci	printk(KERN_INFO "Reserving virtual address space above 0x%08lx (rounded to 0x%08lx)\n",
6268c2ecf20Sopenharmony_ci	       -reserve, __FIXADDR_TOP + PAGE_SIZE);
6278c2ecf20Sopenharmony_ci#endif
6288c2ecf20Sopenharmony_ci}
6298c2ecf20Sopenharmony_ci
6308c2ecf20Sopenharmony_ciint fixmaps_set;
6318c2ecf20Sopenharmony_ci
6328c2ecf20Sopenharmony_civoid __native_set_fixmap(enum fixed_addresses idx, pte_t pte)
6338c2ecf20Sopenharmony_ci{
6348c2ecf20Sopenharmony_ci	unsigned long address = __fix_to_virt(idx);
6358c2ecf20Sopenharmony_ci
6368c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_64
6378c2ecf20Sopenharmony_ci       /*
6388c2ecf20Sopenharmony_ci	* Ensure that the static initial page tables are covering the
6398c2ecf20Sopenharmony_ci	* fixmap completely.
6408c2ecf20Sopenharmony_ci	*/
6418c2ecf20Sopenharmony_ci	BUILD_BUG_ON(__end_of_permanent_fixed_addresses >
6428c2ecf20Sopenharmony_ci		     (FIXMAP_PMD_NUM * PTRS_PER_PTE));
6438c2ecf20Sopenharmony_ci#endif
6448c2ecf20Sopenharmony_ci
6458c2ecf20Sopenharmony_ci	if (idx >= __end_of_fixed_addresses) {
6468c2ecf20Sopenharmony_ci		BUG();
6478c2ecf20Sopenharmony_ci		return;
6488c2ecf20Sopenharmony_ci	}
6498c2ecf20Sopenharmony_ci	set_pte_vaddr(address, pte);
6508c2ecf20Sopenharmony_ci	fixmaps_set++;
6518c2ecf20Sopenharmony_ci}
6528c2ecf20Sopenharmony_ci
6538c2ecf20Sopenharmony_civoid native_set_fixmap(unsigned /* enum fixed_addresses */ idx,
6548c2ecf20Sopenharmony_ci		       phys_addr_t phys, pgprot_t flags)
6558c2ecf20Sopenharmony_ci{
6568c2ecf20Sopenharmony_ci	/* Sanitize 'prot' against any unsupported bits: */
6578c2ecf20Sopenharmony_ci	pgprot_val(flags) &= __default_kernel_pte_mask;
6588c2ecf20Sopenharmony_ci
6598c2ecf20Sopenharmony_ci	__native_set_fixmap(idx, pfn_pte(phys >> PAGE_SHIFT, flags));
6608c2ecf20Sopenharmony_ci}
6618c2ecf20Sopenharmony_ci
6628c2ecf20Sopenharmony_ci#ifdef CONFIG_HAVE_ARCH_HUGE_VMAP
6638c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_5LEVEL
6648c2ecf20Sopenharmony_ci/**
6658c2ecf20Sopenharmony_ci * p4d_set_huge - setup kernel P4D mapping
6668c2ecf20Sopenharmony_ci *
6678c2ecf20Sopenharmony_ci * No 512GB pages yet -- always return 0
6688c2ecf20Sopenharmony_ci */
6698c2ecf20Sopenharmony_ciint p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot)
6708c2ecf20Sopenharmony_ci{
6718c2ecf20Sopenharmony_ci	return 0;
6728c2ecf20Sopenharmony_ci}
6738c2ecf20Sopenharmony_ci
6748c2ecf20Sopenharmony_ci/**
6758c2ecf20Sopenharmony_ci * p4d_clear_huge - clear kernel P4D mapping when it is set
6768c2ecf20Sopenharmony_ci *
6778c2ecf20Sopenharmony_ci * No 512GB pages yet -- always return 0
6788c2ecf20Sopenharmony_ci */
6798c2ecf20Sopenharmony_ciint p4d_clear_huge(p4d_t *p4d)
6808c2ecf20Sopenharmony_ci{
6818c2ecf20Sopenharmony_ci	return 0;
6828c2ecf20Sopenharmony_ci}
6838c2ecf20Sopenharmony_ci#endif
6848c2ecf20Sopenharmony_ci
6858c2ecf20Sopenharmony_ci/**
6868c2ecf20Sopenharmony_ci * pud_set_huge - setup kernel PUD mapping
6878c2ecf20Sopenharmony_ci *
6888c2ecf20Sopenharmony_ci * MTRRs can override PAT memory types with 4KiB granularity. Therefore, this
6898c2ecf20Sopenharmony_ci * function sets up a huge page only if any of the following conditions are met:
6908c2ecf20Sopenharmony_ci *
6918c2ecf20Sopenharmony_ci * - MTRRs are disabled, or
6928c2ecf20Sopenharmony_ci *
6938c2ecf20Sopenharmony_ci * - MTRRs are enabled and the range is completely covered by a single MTRR, or
6948c2ecf20Sopenharmony_ci *
6958c2ecf20Sopenharmony_ci * - MTRRs are enabled and the corresponding MTRR memory type is WB, which
6968c2ecf20Sopenharmony_ci *   has no effect on the requested PAT memory type.
6978c2ecf20Sopenharmony_ci *
6988c2ecf20Sopenharmony_ci * Callers should try to decrease page size (1GB -> 2MB -> 4K) if the bigger
6998c2ecf20Sopenharmony_ci * page mapping attempt fails.
7008c2ecf20Sopenharmony_ci *
7018c2ecf20Sopenharmony_ci * Returns 1 on success and 0 on failure.
7028c2ecf20Sopenharmony_ci */
7038c2ecf20Sopenharmony_ciint pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
7048c2ecf20Sopenharmony_ci{
7058c2ecf20Sopenharmony_ci	u8 mtrr, uniform;
7068c2ecf20Sopenharmony_ci
7078c2ecf20Sopenharmony_ci	mtrr = mtrr_type_lookup(addr, addr + PUD_SIZE, &uniform);
7088c2ecf20Sopenharmony_ci	if ((mtrr != MTRR_TYPE_INVALID) && (!uniform) &&
7098c2ecf20Sopenharmony_ci	    (mtrr != MTRR_TYPE_WRBACK))
7108c2ecf20Sopenharmony_ci		return 0;
7118c2ecf20Sopenharmony_ci
7128c2ecf20Sopenharmony_ci	/* Bail out if we are we on a populated non-leaf entry: */
7138c2ecf20Sopenharmony_ci	if (pud_present(*pud) && !pud_huge(*pud))
7148c2ecf20Sopenharmony_ci		return 0;
7158c2ecf20Sopenharmony_ci
7168c2ecf20Sopenharmony_ci	set_pte((pte_t *)pud, pfn_pte(
7178c2ecf20Sopenharmony_ci		(u64)addr >> PAGE_SHIFT,
7188c2ecf20Sopenharmony_ci		__pgprot(protval_4k_2_large(pgprot_val(prot)) | _PAGE_PSE)));
7198c2ecf20Sopenharmony_ci
7208c2ecf20Sopenharmony_ci	return 1;
7218c2ecf20Sopenharmony_ci}
7228c2ecf20Sopenharmony_ci
7238c2ecf20Sopenharmony_ci/**
7248c2ecf20Sopenharmony_ci * pmd_set_huge - setup kernel PMD mapping
7258c2ecf20Sopenharmony_ci *
7268c2ecf20Sopenharmony_ci * See text over pud_set_huge() above.
7278c2ecf20Sopenharmony_ci *
7288c2ecf20Sopenharmony_ci * Returns 1 on success and 0 on failure.
7298c2ecf20Sopenharmony_ci */
7308c2ecf20Sopenharmony_ciint pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
7318c2ecf20Sopenharmony_ci{
7328c2ecf20Sopenharmony_ci	u8 mtrr, uniform;
7338c2ecf20Sopenharmony_ci
7348c2ecf20Sopenharmony_ci	mtrr = mtrr_type_lookup(addr, addr + PMD_SIZE, &uniform);
7358c2ecf20Sopenharmony_ci	if ((mtrr != MTRR_TYPE_INVALID) && (!uniform) &&
7368c2ecf20Sopenharmony_ci	    (mtrr != MTRR_TYPE_WRBACK)) {
7378c2ecf20Sopenharmony_ci		pr_warn_once("%s: Cannot satisfy [mem %#010llx-%#010llx] with a huge-page mapping due to MTRR override.\n",
7388c2ecf20Sopenharmony_ci			     __func__, addr, addr + PMD_SIZE);
7398c2ecf20Sopenharmony_ci		return 0;
7408c2ecf20Sopenharmony_ci	}
7418c2ecf20Sopenharmony_ci
7428c2ecf20Sopenharmony_ci	/* Bail out if we are we on a populated non-leaf entry: */
7438c2ecf20Sopenharmony_ci	if (pmd_present(*pmd) && !pmd_huge(*pmd))
7448c2ecf20Sopenharmony_ci		return 0;
7458c2ecf20Sopenharmony_ci
7468c2ecf20Sopenharmony_ci	set_pte((pte_t *)pmd, pfn_pte(
7478c2ecf20Sopenharmony_ci		(u64)addr >> PAGE_SHIFT,
7488c2ecf20Sopenharmony_ci		__pgprot(protval_4k_2_large(pgprot_val(prot)) | _PAGE_PSE)));
7498c2ecf20Sopenharmony_ci
7508c2ecf20Sopenharmony_ci	return 1;
7518c2ecf20Sopenharmony_ci}
7528c2ecf20Sopenharmony_ci
7538c2ecf20Sopenharmony_ci/**
7548c2ecf20Sopenharmony_ci * pud_clear_huge - clear kernel PUD mapping when it is set
7558c2ecf20Sopenharmony_ci *
7568c2ecf20Sopenharmony_ci * Returns 1 on success and 0 on failure (no PUD map is found).
7578c2ecf20Sopenharmony_ci */
7588c2ecf20Sopenharmony_ciint pud_clear_huge(pud_t *pud)
7598c2ecf20Sopenharmony_ci{
7608c2ecf20Sopenharmony_ci	if (pud_large(*pud)) {
7618c2ecf20Sopenharmony_ci		pud_clear(pud);
7628c2ecf20Sopenharmony_ci		return 1;
7638c2ecf20Sopenharmony_ci	}
7648c2ecf20Sopenharmony_ci
7658c2ecf20Sopenharmony_ci	return 0;
7668c2ecf20Sopenharmony_ci}
7678c2ecf20Sopenharmony_ci
7688c2ecf20Sopenharmony_ci/**
7698c2ecf20Sopenharmony_ci * pmd_clear_huge - clear kernel PMD mapping when it is set
7708c2ecf20Sopenharmony_ci *
7718c2ecf20Sopenharmony_ci * Returns 1 on success and 0 on failure (no PMD map is found).
7728c2ecf20Sopenharmony_ci */
7738c2ecf20Sopenharmony_ciint pmd_clear_huge(pmd_t *pmd)
7748c2ecf20Sopenharmony_ci{
7758c2ecf20Sopenharmony_ci	if (pmd_large(*pmd)) {
7768c2ecf20Sopenharmony_ci		pmd_clear(pmd);
7778c2ecf20Sopenharmony_ci		return 1;
7788c2ecf20Sopenharmony_ci	}
7798c2ecf20Sopenharmony_ci
7808c2ecf20Sopenharmony_ci	return 0;
7818c2ecf20Sopenharmony_ci}
7828c2ecf20Sopenharmony_ci
7838c2ecf20Sopenharmony_ci/*
7848c2ecf20Sopenharmony_ci * Until we support 512GB pages, skip them in the vmap area.
7858c2ecf20Sopenharmony_ci */
7868c2ecf20Sopenharmony_ciint p4d_free_pud_page(p4d_t *p4d, unsigned long addr)
7878c2ecf20Sopenharmony_ci{
7888c2ecf20Sopenharmony_ci	return 0;
7898c2ecf20Sopenharmony_ci}
7908c2ecf20Sopenharmony_ci
7918c2ecf20Sopenharmony_ci#ifdef CONFIG_X86_64
7928c2ecf20Sopenharmony_ci/**
7938c2ecf20Sopenharmony_ci * pud_free_pmd_page - Clear pud entry and free pmd page.
7948c2ecf20Sopenharmony_ci * @pud: Pointer to a PUD.
7958c2ecf20Sopenharmony_ci * @addr: Virtual address associated with pud.
7968c2ecf20Sopenharmony_ci *
7978c2ecf20Sopenharmony_ci * Context: The pud range has been unmapped and TLB purged.
7988c2ecf20Sopenharmony_ci * Return: 1 if clearing the entry succeeded. 0 otherwise.
7998c2ecf20Sopenharmony_ci *
8008c2ecf20Sopenharmony_ci * NOTE: Callers must allow a single page allocation.
8018c2ecf20Sopenharmony_ci */
8028c2ecf20Sopenharmony_ciint pud_free_pmd_page(pud_t *pud, unsigned long addr)
8038c2ecf20Sopenharmony_ci{
8048c2ecf20Sopenharmony_ci	pmd_t *pmd, *pmd_sv;
8058c2ecf20Sopenharmony_ci	pte_t *pte;
8068c2ecf20Sopenharmony_ci	int i;
8078c2ecf20Sopenharmony_ci
8088c2ecf20Sopenharmony_ci	pmd = pud_pgtable(*pud);
8098c2ecf20Sopenharmony_ci	pmd_sv = (pmd_t *)__get_free_page(GFP_KERNEL);
8108c2ecf20Sopenharmony_ci	if (!pmd_sv)
8118c2ecf20Sopenharmony_ci		return 0;
8128c2ecf20Sopenharmony_ci
8138c2ecf20Sopenharmony_ci	for (i = 0; i < PTRS_PER_PMD; i++) {
8148c2ecf20Sopenharmony_ci		pmd_sv[i] = pmd[i];
8158c2ecf20Sopenharmony_ci		if (!pmd_none(pmd[i]))
8168c2ecf20Sopenharmony_ci			pmd_clear(&pmd[i]);
8178c2ecf20Sopenharmony_ci	}
8188c2ecf20Sopenharmony_ci
8198c2ecf20Sopenharmony_ci	pud_clear(pud);
8208c2ecf20Sopenharmony_ci
8218c2ecf20Sopenharmony_ci	/* INVLPG to clear all paging-structure caches */
8228c2ecf20Sopenharmony_ci	flush_tlb_kernel_range(addr, addr + PAGE_SIZE-1);
8238c2ecf20Sopenharmony_ci
8248c2ecf20Sopenharmony_ci	for (i = 0; i < PTRS_PER_PMD; i++) {
8258c2ecf20Sopenharmony_ci		if (!pmd_none(pmd_sv[i])) {
8268c2ecf20Sopenharmony_ci			pte = (pte_t *)pmd_page_vaddr(pmd_sv[i]);
8278c2ecf20Sopenharmony_ci			free_page((unsigned long)pte);
8288c2ecf20Sopenharmony_ci		}
8298c2ecf20Sopenharmony_ci	}
8308c2ecf20Sopenharmony_ci
8318c2ecf20Sopenharmony_ci	free_page((unsigned long)pmd_sv);
8328c2ecf20Sopenharmony_ci
8338c2ecf20Sopenharmony_ci	pgtable_pmd_page_dtor(virt_to_page(pmd));
8348c2ecf20Sopenharmony_ci	free_page((unsigned long)pmd);
8358c2ecf20Sopenharmony_ci
8368c2ecf20Sopenharmony_ci	return 1;
8378c2ecf20Sopenharmony_ci}
8388c2ecf20Sopenharmony_ci
8398c2ecf20Sopenharmony_ci/**
8408c2ecf20Sopenharmony_ci * pmd_free_pte_page - Clear pmd entry and free pte page.
8418c2ecf20Sopenharmony_ci * @pmd: Pointer to a PMD.
8428c2ecf20Sopenharmony_ci * @addr: Virtual address associated with pmd.
8438c2ecf20Sopenharmony_ci *
8448c2ecf20Sopenharmony_ci * Context: The pmd range has been unmapped and TLB purged.
8458c2ecf20Sopenharmony_ci * Return: 1 if clearing the entry succeeded. 0 otherwise.
8468c2ecf20Sopenharmony_ci */
8478c2ecf20Sopenharmony_ciint pmd_free_pte_page(pmd_t *pmd, unsigned long addr)
8488c2ecf20Sopenharmony_ci{
8498c2ecf20Sopenharmony_ci	pte_t *pte;
8508c2ecf20Sopenharmony_ci
8518c2ecf20Sopenharmony_ci	pte = (pte_t *)pmd_page_vaddr(*pmd);
8528c2ecf20Sopenharmony_ci	pmd_clear(pmd);
8538c2ecf20Sopenharmony_ci
8548c2ecf20Sopenharmony_ci	/* INVLPG to clear all paging-structure caches */
8558c2ecf20Sopenharmony_ci	flush_tlb_kernel_range(addr, addr + PAGE_SIZE-1);
8568c2ecf20Sopenharmony_ci
8578c2ecf20Sopenharmony_ci	free_page((unsigned long)pte);
8588c2ecf20Sopenharmony_ci
8598c2ecf20Sopenharmony_ci	return 1;
8608c2ecf20Sopenharmony_ci}
8618c2ecf20Sopenharmony_ci
8628c2ecf20Sopenharmony_ci#else /* !CONFIG_X86_64 */
8638c2ecf20Sopenharmony_ci
8648c2ecf20Sopenharmony_ciint pud_free_pmd_page(pud_t *pud, unsigned long addr)
8658c2ecf20Sopenharmony_ci{
8668c2ecf20Sopenharmony_ci	return pud_none(*pud);
8678c2ecf20Sopenharmony_ci}
8688c2ecf20Sopenharmony_ci
8698c2ecf20Sopenharmony_ci/*
8708c2ecf20Sopenharmony_ci * Disable free page handling on x86-PAE. This assures that ioremap()
8718c2ecf20Sopenharmony_ci * does not update sync'd pmd entries. See vmalloc_sync_one().
8728c2ecf20Sopenharmony_ci */
8738c2ecf20Sopenharmony_ciint pmd_free_pte_page(pmd_t *pmd, unsigned long addr)
8748c2ecf20Sopenharmony_ci{
8758c2ecf20Sopenharmony_ci	return pmd_none(*pmd);
8768c2ecf20Sopenharmony_ci}
8778c2ecf20Sopenharmony_ci
8788c2ecf20Sopenharmony_ci#endif /* CONFIG_X86_64 */
8798c2ecf20Sopenharmony_ci#endif	/* CONFIG_HAVE_ARCH_HUGE_VMAP */
880