Lines Matching defs:pmd

53  * The pmd table for the upper-most set of pages.
76 pmdval_t pmd;
84 .pmd = PMD_SECT_UNCACHED,
89 .pmd = PMD_SECT_BUFFERED,
94 .pmd = PMD_SECT_WT,
99 .pmd = PMD_SECT_WB,
104 .pmd = PMD_SECT_WBWA,
114 * via the "pmd" value. This is used to ensure that on ARMv6 and later,
119 void __init init_default_cache_policy(unsigned long pmd)
123 initial_pmd_value = pmd;
125 pmd &= PMD_SECT_CACHE_MASK;
128 if (cache_policies[i].pmd == pmd) {
378 pmd_t *pmd;
387 pmd = fixmap_pmd(FIXADDR_TOP);
388 pmd_populate_kernel(&init_mm, pmd, bm_pte);
663 mem_types[MT_MEMORY_RWX].prot_sect |= ecc_mask | cp->pmd;
665 mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd;
667 mem_types[MT_MEMORY_RO].prot_sect |= ecc_mask | cp->pmd;
671 mem_types[MT_ROM].prot_sect |= cp->pmd;
673 switch (cp->pmd) {
729 static pte_t * __init arm_pte_alloc(pmd_t *pmd, unsigned long addr,
733 if (pmd_none(*pmd)) {
735 __pmd_populate(pmd, __pa(pte), prot);
737 BUG_ON(pmd_bad(*pmd));
738 return pte_offset_kernel(pmd, addr);
741 static pte_t * __init early_pte_alloc(pmd_t *pmd, unsigned long addr,
744 return arm_pte_alloc(pmd, addr, prot, early_alloc);
747 static void __init alloc_init_pte(pmd_t *pmd, unsigned long addr,
753 pte_t *pte = arm_pte_alloc(pmd, addr, type->prot_l1, alloc);
761 static void __init __map_init_section(pmd_t *pmd, unsigned long addr,
765 pmd_t *p = pmd;
778 pmd++;
781 *pmd = __pmd(phys | type->prot_sect | (ng ? PMD_SECT_nG : 0));
783 } while (pmd++, addr += SECTION_SIZE, addr != end);
793 pmd_t *pmd = pmd_offset(pud, addr);
809 __map_init_section(pmd, addr, next, phys, type, ng);
811 alloc_init_pte(pmd, addr, next,
817 } while (pmd++, addr = next, addr != end);
899 pmd_t *pmd = pmd_offset(pud, addr);
903 *pmd++ = __pmd(phys | type->prot_sect | PMD_SECT_SUPER |
1074 pmd_t *pmd;
1088 pmd = pmd_off_k(addr);
1089 if (pmd_none(*pmd))
1100 pmd = pmd_off_k(addr) + 1;
1101 if (pmd_none(*pmd))
1218 * Find the first non-pmd-aligned page, and point
1220 * limit down to be pmd-aligned, which happens at the
1224 * bank can be non-pmd-aligned. The only exception is
1248 * Round the memblock limit down to a pmd size. This
1250 * last full pmd, which should be mapped.
1286 * equal to MODULES_VADDR and then we exit the pmd clearing. If we
1369 * Clear page table except top pmd used by early fixmaps