Lines Matching refs:pvmw
779 struct page_vma_mapped_walk pvmw = {
786 while (page_vma_mapped_walk(&pvmw)) {
787 address = pvmw.address;
794 page_vma_mapped_walk_done(&pvmw);
799 if (pvmw.pte) {
801 pvmw.pte)) {
815 pvmw.pmd))
913 struct page_vma_mapped_walk pvmw = {
931 while (page_vma_mapped_walk(&pvmw)) {
934 address = pvmw.address;
935 if (pvmw.pte) {
937 pte_t *pte = pvmw.pte;
950 pmd_t *pmd = pvmw.pmd;
1390 struct page_vma_mapped_walk pvmw = {
1408 pvmw.flags = PVMW_SYNC;
1445 while (page_vma_mapped_walk(&pvmw)) {
1448 if (!pvmw.pte && (flags & TTU_MIGRATION)) {
1451 set_pmd_migration_entry(&pvmw, page);
1458 page_vma_mapped_walk_done(&pvmw);
1479 page_vma_mapped_walk_done(&pvmw);
1487 VM_BUG_ON_PAGE(!pvmw.pte, page);
1489 subpage = page - page_to_pfn(page) + pte_pfn(*pvmw.pte);
1490 address = pvmw.address;
1499 if (huge_pmd_unshare(mm, vma, &address, pvmw.pte)) {
1521 page_vma_mapped_walk_done(&pvmw);
1532 pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte);
1550 set_pte_at(mm, pvmw.address, pvmw.pte, swp_pte);
1567 flush_cache_page(vma, address, pte_pfn(*pvmw.pte));
1577 pteval = ptep_get_and_clear(mm, address, pvmw.pte);
1581 pteval = ptep_clear_flush(vma, address, pvmw.pte);
1596 pvmw.pte, pteval,
1600 set_pte_at(mm, address, pvmw.pte, pteval);
1627 set_pte_at(mm, address, pvmw.pte, pteval);
1629 page_vma_mapped_walk_done(&pvmw);
1645 set_pte_at(mm, address, pvmw.pte, swp_pte);
1663 page_vma_mapped_walk_done(&pvmw);
1704 set_pte_at(mm, address, pvmw.pte, pteval);
1707 page_vma_mapped_walk_done(&pvmw);
1712 set_pte_at(mm, address, pvmw.pte, pteval);
1714 page_vma_mapped_walk_done(&pvmw);
1718 set_pte_at(mm, address, pvmw.pte, pteval);
1720 page_vma_mapped_walk_done(&pvmw);
1736 set_pte_at(mm, address, pvmw.pte, swp_pte);