Lines Matching refs:pmd
547 pmd_t *pmd; /* Pointer to pmd entry matching
567 * is not NULL, otherwise pmd.
1703 struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd);
2163 int __pte_alloc(struct mm_struct *mm, pmd_t *pmd);
2164 int __pte_alloc_kernel(pmd_t *pmd);
2214 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd)
2216 return ptlock_ptr(pmd_page(*pmd));
2240 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd)
2279 #define pte_offset_map_lock(mm, pmd, address, ptlp) \
2281 spinlock_t *__ptl = pte_lockptr(mm, pmd); \
2282 pte_t *__pte = pte_offset_map(pmd, address); \
2294 #define pte_alloc(mm, pmd) (unlikely(pmd_none(*(pmd))) && __pte_alloc(mm, pmd))
2296 #define pte_alloc_map(mm, pmd, address) (pte_alloc(mm, pmd) ? NULL : pte_offset_map(pmd, address))
2298 #define pte_alloc_map_lock(mm, pmd, address, ptlp) \
2299 (pte_alloc(mm, pmd) ? NULL : pte_offset_map_lock(mm, pmd, address, ptlp))
2301 #define pte_alloc_kernel(pmd, address) \
2302 ((unlikely(pmd_none(*(pmd))) && __pte_alloc_kernel(pmd)) ? NULL : pte_offset_kernel(pmd, address))
2306 static struct page *pmd_to_page(pmd_t *pmd)
2309 return virt_to_page((void *)((unsigned long)pmd & mask));
2312 static inline spinlock_t *pmd_lockptr(struct mm_struct *mm, pmd_t *pmd)
2314 return ptlock_ptr(pmd_to_page(pmd));
2333 #define pmd_huge_pte(mm, pmd) (pmd_to_page(pmd)->pmd_huge_pte)
2337 static inline spinlock_t *pmd_lockptr(struct mm_struct *mm, pmd_t *pmd)
2350 #define pmd_huge_pte(mm, pmd) ((mm)->pmd_huge_pte)
2354 static inline spinlock_t *pmd_lock(struct mm_struct *mm, pmd_t *pmd)
2356 spinlock_t *ptl = pmd_lockptr(mm, pmd);
2824 #define FOLL_SPLIT_PMD 0x20000 /* split huge pmd before returning */
3051 pte_t *vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node, struct vmem_altmap *altmap);