/kernel/linux/linux-5.10/mm/ |
H A D | migrate.c | 404 struct page *newpage, struct page *page, int extra_count) in migrate_page_move_mapping() 418 newpage->index = page->index; in migrate_page_move_mapping() 419 newpage->mapping = page->mapping; in migrate_page_move_mapping() 421 __SetPageSwapBacked(newpage); in migrate_page_move_mapping() 427 newzone = page_zone(newpage); in migrate_page_move_mapping() 444 newpage->index = page->index; in migrate_page_move_mapping() 445 newpage->mapping = page->mapping; in migrate_page_move_mapping() 446 page_ref_add(newpage, nr); /* add cache reference */ in migrate_page_move_mapping() 448 __SetPageSwapBacked(newpage); in migrate_page_move_mapping() 450 SetPageSwapCache(newpage); in migrate_page_move_mapping() 403 migrate_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page, int extra_count) migrate_page_move_mapping() argument 525 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument 608 migrate_page_states(struct page *newpage, struct page *page) migrate_page_states() argument 685 migrate_page_copy(struct page *newpage, struct page *page) migrate_page_copy() argument 700 migrate_page_extra(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, int extra_count) migrate_page_extra() argument 726 migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) migrate_page() argument 773 __buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, bool check_refs) __buffer_migrate_page() argument 857 buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page() argument 870 buffer_migrate_page_norefs(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page_norefs() argument 921 fallback_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) fallback_migrate_page() argument 958 move_to_new_page(struct page *newpage, struct page *page, enum migrate_mode mode) move_to_new_page() argument 1038 __unmap_and_move(struct page *page, struct page *newpage, int force, enum migrate_mode mode) __unmap_and_move() argument 1193 struct page *newpage = NULL; unmap_and_move() local 2031 struct page *newpage; alloc_misplaced_dst_page() local 3018 struct page *newpage = migrate_pfn_to_page(migrate->dst[i]); migrate_vma_pages() local 3106 struct page *newpage = migrate_pfn_to_page(migrate->dst[i]); migrate_vma_finalize() local [all...] |
H A D | balloon_compaction.c | 234 struct page *newpage, struct page *page, in balloon_page_migrate() 248 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); in balloon_page_migrate() 250 return balloon->migratepage(balloon, newpage, page, mode); in balloon_page_migrate() 233 balloon_page_migrate(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) balloon_page_migrate() argument
|
H A D | zsmalloc.c | 1880 struct page *newpage, struct page *oldpage) in replace_sub_page() 1889 pages[idx] = newpage; in replace_sub_page() 1896 set_first_obj_offset(newpage, get_first_obj_offset(oldpage)); in replace_sub_page() 1898 newpage->index = oldpage->index; in replace_sub_page() 1899 __SetPageMovable(newpage, page_mapping(oldpage)); in replace_sub_page() 1958 static int zs_page_migrate(struct address_space *mapping, struct page *newpage, in zs_page_migrate() argument 2018 d_addr = kmap_atomic(newpage); in zs_page_migrate() 2032 new_obj = (unsigned long)location_to_obj(newpage, in zs_page_migrate() 2039 replace_sub_page(class, zspage, newpage, page); in zs_page_migrate() 2040 get_page(newpage); in zs_page_migrate() 1879 replace_sub_page(struct size_class *class, struct zspage *zspage, struct page *newpage, struct page *oldpage) replace_sub_page() argument [all...] |
H A D | internal.h | 500 static inline void mlock_migrate_page(struct page *newpage, struct page *page) in mlock_migrate_page() argument 507 SetPageMlocked(newpage); in mlock_migrate_page() 508 __mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages); in mlock_migrate_page()
|
H A D | shmem.c | 1628 struct page *oldpage, *newpage; in shmem_replace_page() local 1644 newpage = shmem_alloc_page(gfp, info, index); in shmem_replace_page() 1645 if (!newpage) in shmem_replace_page() 1648 get_page(newpage); in shmem_replace_page() 1649 copy_highpage(newpage, oldpage); in shmem_replace_page() 1650 flush_dcache_page(newpage); in shmem_replace_page() 1652 __SetPageLocked(newpage); in shmem_replace_page() 1653 __SetPageSwapBacked(newpage); in shmem_replace_page() 1654 SetPageUptodate(newpage); in shmem_replace_page() 1655 set_page_private(newpage, entr in shmem_replace_page() [all...] |
H A D | z3fold.c | 1616 static int z3fold_page_migrate(struct address_space *mapping, struct page *newpage, in z3fold_page_migrate() argument 1626 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); in z3fold_page_migrate() 1642 new_zhdr = page_address(newpage); in z3fold_page_migrate() 1644 newpage->private = page->private; in z3fold_page_migrate() 1658 get_page(newpage); in z3fold_page_migrate() 1666 set_bit(NEEDS_COMPACTING, &newpage->private); in z3fold_page_migrate() 1669 list_add(&newpage->lru, &pool->lru); in z3fold_page_migrate() 1671 __SetPageMovable(newpage, new_mapping); in z3fold_page_migrate()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | migrate.h | 40 struct page *newpage, struct page *page, 43 struct page *newpage, struct page *page, 53 extern void migrate_page_states(struct page *newpage, struct page *page); 54 extern void migrate_page_copy(struct page *newpage, struct page *page); 56 struct page *newpage, struct page *page); 58 struct page *newpage, struct page *page, int extra_count); 75 static inline void migrate_page_states(struct page *newpage, struct page *page) in migrate_page_states() argument 79 static inline void migrate_page_copy(struct page *newpage, in migrate_page_copy() argument 83 struct page *newpage, struct page *page) in migrate_huge_page_move_mapping() 82 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument
|
H A D | page_owner.h | 15 extern void __copy_page_owner(struct page *oldpage, struct page *newpage); 39 static inline void copy_page_owner(struct page *oldpage, struct page *newpage) in copy_page_owner() argument 42 __copy_page_owner(oldpage, newpage); in copy_page_owner() 66 static inline void copy_page_owner(struct page *oldpage, struct page *newpage) in copy_page_owner() argument
|
H A D | balloon_compaction.h | 58 int (*migratepage)(struct balloon_dev_info *, struct page *newpage, 87 struct page *newpage, 168 static inline int balloon_page_migrate(struct page *newpage, in balloon_page_migrate() argument
|
H A D | ksm.h | 55 void ksm_migrate_page(struct page *newpage, struct page *oldpage); 86 static inline void ksm_migrate_page(struct page *newpage, struct page *oldpage) in ksm_migrate_page() argument
|
/kernel/linux/linux-6.6/mm/ |
H A D | migrate_device.c | 690 struct page *newpage = migrate_pfn_to_page(dst_pfns[i]); in __migrate_device_pages() local 695 if (!newpage) { in __migrate_device_pages() 722 migrate_vma_insert_page(migrate, addr, newpage, in __migrate_device_pages() 729 if (is_device_private_page(newpage) || in __migrate_device_pages() 730 is_device_coherent_page(newpage)) { in __migrate_device_pages() 748 } else if (is_zone_device_page(newpage)) { in __migrate_device_pages() 757 r = migrate_folio_extra(mapping, page_folio(newpage), in __migrate_device_pages() 761 r = migrate_folio(mapping, page_folio(newpage), in __migrate_device_pages() 818 struct page *newpage = migrate_pfn_to_page(dst_pfns[i]); in migrate_device_finalize() local 822 if (newpage) { in migrate_device_finalize() [all...] |
H A D | balloon_compaction.c | 232 static int balloon_page_migrate(struct page *newpage, struct page *page, in balloon_page_migrate() argument 246 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); in balloon_page_migrate() 248 return balloon->migratepage(balloon, newpage, page, mode); in balloon_page_migrate()
|
H A D | z3fold.c | 1285 static int z3fold_page_migrate(struct page *newpage, struct page *page, in z3fold_page_migrate() argument 1293 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); in z3fold_page_migrate() 1309 new_zhdr = page_address(newpage); in z3fold_page_migrate() 1311 newpage->private = page->private; in z3fold_page_migrate() 1323 get_page(newpage); in z3fold_page_migrate() 1331 set_bit(NEEDS_COMPACTING, &newpage->private); in z3fold_page_migrate() 1333 __SetPageMovable(newpage, &z3fold_mops); in z3fold_page_migrate()
|
H A D | zsmalloc.c | 1756 struct page *newpage, struct page *oldpage) in replace_sub_page() 1765 pages[idx] = newpage; in replace_sub_page() 1772 set_first_obj_offset(newpage, get_first_obj_offset(oldpage)); in replace_sub_page() 1774 newpage->index = oldpage->index; in replace_sub_page() 1775 __SetPageMovable(newpage, &zsmalloc_mops); in replace_sub_page() 1798 static int zs_page_migrate(struct page *newpage, struct page *page, in zs_page_migrate() argument 1841 d_addr = kmap_atomic(newpage); in zs_page_migrate() 1851 new_obj = (unsigned long)location_to_obj(newpage, in zs_page_migrate() 1858 replace_sub_page(class, zspage, newpage, page); in zs_page_migrate() 1867 get_page(newpage); in zs_page_migrate() 1755 replace_sub_page(struct size_class *class, struct zspage *zspage, struct page *newpage, struct page *oldpage) replace_sub_page() argument [all...] |
/kernel/linux/linux-5.10/fs/cachefiles/ |
H A D | rdwr.c | 235 struct page *newpage, *backpage; in cachefiles_read_backing_file_one() local 254 newpage = NULL; in cachefiles_read_backing_file_one() 261 if (!newpage) { in cachefiles_read_backing_file_one() 262 newpage = __page_cache_alloc(cachefiles_gfp); in cachefiles_read_backing_file_one() 263 if (!newpage) in cachefiles_read_backing_file_one() 267 ret = add_to_page_cache_lru(newpage, bmapping, in cachefiles_read_backing_file_one() 278 _debug("- new %p", newpage); in cachefiles_read_backing_file_one() 280 backpage = newpage; in cachefiles_read_backing_file_one() 281 newpage = NULL; in cachefiles_read_backing_file_one() 314 if (newpage) { in cachefiles_read_backing_file_one() 471 struct page *newpage = NULL, *netpage, *_n, *backpage = NULL; cachefiles_read_backing_file() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/platforms/pseries/ |
H A D | cmm.c | 519 struct page *newpage, struct page *page, in cmm_migratepage() 525 * loan/"inflate" the newpage first. in cmm_migratepage() 532 if (plpar_page_set_loaned(newpage)) { in cmm_migratepage() 539 get_page(newpage); in cmm_migratepage() 546 if (page_zone(page) != page_zone(newpage)) { in cmm_migratepage() 548 adjust_managed_page_count(newpage, -1); in cmm_migratepage() 552 balloon_page_insert(b_dev_info, newpage); in cmm_migratepage() 518 cmm_migratepage(struct balloon_dev_info *b_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) cmm_migratepage() argument
|
/kernel/linux/linux-6.6/arch/powerpc/platforms/pseries/ |
H A D | cmm.c | 501 struct page *newpage, struct page *page, in cmm_migratepage() 507 * loan/"inflate" the newpage first. in cmm_migratepage() 514 if (plpar_page_set_loaned(newpage)) { in cmm_migratepage() 521 get_page(newpage); in cmm_migratepage() 528 if (page_zone(page) != page_zone(newpage)) { in cmm_migratepage() 530 adjust_managed_page_count(newpage, -1); in cmm_migratepage() 534 balloon_page_insert(b_dev_info, newpage); in cmm_migratepage() 500 cmm_migratepage(struct balloon_dev_info *b_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) cmm_migratepage() argument
|
/kernel/linux/linux-5.10/drivers/virtio/ |
H A D | virtio_balloon.c | 743 * @newpage: page that will replace the isolated page after migration finishes. 744 * @page : the isolated (old) page that is about to be migrated to newpage. 751 * 1) insert newpage into vb->pages list and update the host about it; 758 struct page *newpage, struct page *page, enum migrate_mode mode) in virtballoon_migratepage() 775 get_page(newpage); /* balloon reference */ in virtballoon_migratepage() 783 page_zone(page) != page_zone(newpage)) { in virtballoon_migratepage() 785 adjust_managed_page_count(newpage, -1); in virtballoon_migratepage() 788 /* balloon's page migration 1st step -- inflate "newpage" */ in virtballoon_migratepage() 790 balloon_page_insert(vb_dev_info, newpage); in virtballoon_migratepage() 795 set_page_pfns(vb, vb->pfns, newpage); in virtballoon_migratepage() 757 virtballoon_migratepage(struct balloon_dev_info *vb_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) virtballoon_migratepage() argument [all...] |
/kernel/linux/linux-6.6/drivers/virtio/ |
H A D | virtio_balloon.c | 736 * @newpage: page that will replace the isolated page after migration finishes. 737 * @page : the isolated (old) page that is about to be migrated to newpage. 744 * 1) insert newpage into vb->pages list and update the host about it; 751 struct page *newpage, struct page *page, enum migrate_mode mode) in virtballoon_migratepage() 768 get_page(newpage); /* balloon reference */ in virtballoon_migratepage() 776 page_zone(page) != page_zone(newpage)) { in virtballoon_migratepage() 778 adjust_managed_page_count(newpage, -1); in virtballoon_migratepage() 781 /* balloon's page migration 1st step -- inflate "newpage" */ in virtballoon_migratepage() 783 balloon_page_insert(vb_dev_info, newpage); in virtballoon_migratepage() 788 set_page_pfns(vb, vb->pfns, newpage); in virtballoon_migratepage() 750 virtballoon_migratepage(struct balloon_dev_info *vb_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) virtballoon_migratepage() argument [all...] |
/kernel/linux/linux-5.10/fs/fuse/ |
H A D | dev.c | 799 struct page *newpage; in fuse_try_move_page() local 825 newpage = buf->page; in fuse_try_move_page() 827 if (!PageUptodate(newpage)) in fuse_try_move_page() 828 SetPageUptodate(newpage); in fuse_try_move_page() 830 ClearPageMappedToDisk(newpage); in fuse_try_move_page() 832 if (fuse_check_page(newpage) != 0) in fuse_try_move_page() 848 err = replace_page_cache_page(oldpage, newpage, GFP_KERNEL); in fuse_try_move_page() 850 unlock_page(newpage); in fuse_try_move_page() 854 get_page(newpage); in fuse_try_move_page() 857 lru_cache_add(newpage); in fuse_try_move_page() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | balloon_compaction.h | 58 int (*migratepage)(struct balloon_dev_info *, struct page *newpage,
|
/kernel/linux/linux-5.10/drivers/misc/ |
H A D | vmw_balloon.c | 1755 * @newpage: the page to which @page should be migrated. 1767 struct page *newpage, struct page *page, in vmballoon_migratepage() 1818 vmballoon_add_page(b, 0, newpage); in vmballoon_migratepage() 1823 status = vmballoon_status_page(b, 0, &newpage); in vmballoon_migratepage() 1841 get_page(newpage); in vmballoon_migratepage() 1849 * On inflation success, we already took a reference for the @newpage. in vmballoon_migratepage() 1854 balloon_page_insert(&b->b_dev_info, newpage); in vmballoon_migratepage() 1766 vmballoon_migratepage(struct balloon_dev_info *b_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) vmballoon_migratepage() argument
|
/kernel/linux/linux-5.10/fs/ubifs/ |
H A D | file.c | 1465 struct page *newpage, struct page *page, enum migrate_mode mode) in ubifs_migrate_page() 1469 rc = migrate_page_move_mapping(mapping, newpage, page, 0); in ubifs_migrate_page() 1475 attach_page_private(newpage, (void *)1); in ubifs_migrate_page() 1479 migrate_page_copy(newpage, page); in ubifs_migrate_page() 1481 migrate_page_states(newpage, page); in ubifs_migrate_page() 1464 ubifs_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) ubifs_migrate_page() argument
|
/kernel/linux/linux-5.10/fs/hugetlbfs/ |
H A D | inode.c | 966 struct page *newpage, struct page *page, in hugetlbfs_migrate_page() 971 rc = migrate_huge_page_move_mapping(mapping, newpage, page); in hugetlbfs_migrate_page() 982 set_page_private(newpage, page_private(page)); in hugetlbfs_migrate_page() 987 migrate_page_copy(newpage, page); in hugetlbfs_migrate_page() 989 migrate_page_states(newpage, page); in hugetlbfs_migrate_page() 965 hugetlbfs_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) hugetlbfs_migrate_page() argument
|
/kernel/linux/linux-6.6/drivers/misc/ |
H A D | vmw_balloon.c | 1734 * @newpage: the page to which @page should be migrated. 1746 struct page *newpage, struct page *page, in vmballoon_migratepage() 1797 vmballoon_add_page(b, 0, newpage); in vmballoon_migratepage() 1802 status = vmballoon_status_page(b, 0, &newpage); in vmballoon_migratepage() 1820 get_page(newpage); in vmballoon_migratepage() 1828 * On inflation success, we already took a reference for the @newpage. in vmballoon_migratepage() 1833 balloon_page_insert(&b->b_dev_info, newpage); in vmballoon_migratepage() 1745 vmballoon_migratepage(struct balloon_dev_info *b_dev_info, struct page *newpage, struct page *page, enum migrate_mode mode) vmballoon_migratepage() argument
|