Lines Matching refs:haddr
652 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
701 folio_add_new_anon_rmap(folio, vma, haddr);
704 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry);
761 struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd,
770 set_pmd_at(mm, haddr, pmd, entry);
779 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
781 if (!transhuge_vma_suitable(vma, haddr))
816 haddr, vmf->pmd, zero_page);
827 folio = vma_alloc_folio(gfp, HPAGE_PMD_ORDER, vma, haddr, true);
1295 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
1361 if (pmdp_set_access_flags(vma, haddr, vmf->pmd, entry, 1))
1494 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
1517 page = vm_normal_page_pmd(vma, haddr, pmd);
1532 target_nid = numa_migrate_prep(page, vma, haddr, page_nid,
1570 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, pmd);
1992 unsigned long haddr)
1994 VM_BUG_ON(haddr & ~HPAGE_PUD_MASK);
1995 VM_BUG_ON_VMA(vma->vm_start > haddr, vma);
1996 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PUD_SIZE, vma);
2001 pudp_huge_clear_flush(vma, haddr, pud);
2026 unsigned long haddr, pmd_t *pmd)
2043 old_pmd = pmdp_huge_clear_flush(vma, haddr, pmd);
2048 pte = pte_offset_map(&_pmd, haddr);
2050 for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) {
2067 unsigned long haddr, bool freeze)
2079 VM_BUG_ON(haddr & ~HPAGE_PMD_MASK);
2080 VM_BUG_ON_VMA(vma->vm_start > haddr, vma);
2081 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PMD_SIZE, vma);
2088 old_pmd = pmdp_huge_clear_flush(vma, haddr, pmd);
2125 return __split_huge_zero_page_pmd(vma, haddr, pmd);
2148 old_pmd = pmdp_invalidate(vma, haddr, pmd);
2205 pte = pte_offset_map(&_pmd, haddr);
2207 for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) {
3258 unsigned long haddr = address & HPAGE_PMD_MASK;
3286 page_add_anon_rmap(new, vma, haddr, rmap_flags);
3291 set_pmd_at(mm, haddr, pvmw->pmd, pmde);