Lines Matching refs:new_page
1070 struct page *new_page;
1089 new_page = khugepaged_alloc_page(hpage, gfp, node);
1090 if (!new_page) {
1095 if (unlikely(mem_cgroup_charge(new_page, mm, gfp))) {
1099 count_memcg_page_event(new_page, THP_COLLAPSE_ALLOC);
1190 __collapse_huge_page_copy(pte, new_page, vma, address, pte_ptl,
1193 __SetPageUptodate(new_page);
1196 _pmd = mk_huge_pmd(new_page, vma->vm_page_prot);
1208 page_add_new_anon_rmap(new_page, vma, address, true);
1209 lru_cache_add_inactive_or_unevictable(new_page, vma);
1681 struct page *new_page;
1694 new_page = khugepaged_alloc_page(hpage, gfp, node);
1695 if (!new_page) {
1700 if (unlikely(mem_cgroup_charge(new_page, mm, gfp))) {
1704 count_memcg_page_event(new_page, THP_COLLAPSE_ALLOC);
1719 __SetPageLocked(new_page);
1721 __SetPageSwapBacked(new_page);
1722 new_page->index = start;
1723 new_page->mapping = mapping;
1726 * At this point the new_page is locked and not up-to-date.
1754 xas_store(&xas, new_page);
1896 xas_store(&xas, new_page);
1905 __inc_node_page_state(new_page, NR_SHMEM_THPS);
1907 __inc_node_page_state(new_page, NR_FILE_THPS);
1912 __mod_lruvec_page_state(new_page, NR_FILE_PAGES, nr_none);
1914 __mod_lruvec_page_state(new_page, NR_SHMEM, nr_none);
1931 clear_highpage(new_page + (index % HPAGE_PMD_NR));
1934 copy_highpage(new_page + (page->index % HPAGE_PMD_NR),
1946 clear_highpage(new_page + (index % HPAGE_PMD_NR));
1950 SetPageUptodate(new_page);
1951 page_ref_add(new_page, HPAGE_PMD_NR - 1);
1953 set_page_dirty(new_page);
1954 lru_cache_add(new_page);
2001 new_page->mapping = NULL;
2004 unlock_page(new_page);