/kernel/linux/linux-5.10/include/asm-generic/ |
H A D | memory_model.h | 46 struct pglist_data *__pgdat = NODE_DATA(page_to_nid(__pg)); \
|
/kernel/linux/linux-5.10/mm/ |
H A D | page_ext.c | 128 base = NODE_DATA(page_to_nid(page))->node_page_ext; in lookup_page_ext() 137 index = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext()
|
H A D | list_lru.c | 117 int nid = page_to_nid(virt_to_page(item)); in list_lru_add() 141 int nid = page_to_nid(virt_to_page(item)); in list_lru_del()
|
H A D | sparse.c | 45 int page_to_nid(const struct page *page) in page_to_nid() function 49 EXPORT_SYMBOL(page_to_nid); variable
|
H A D | slub.c | 1211 struct kmem_cache_node *n = get_node(s, page_to_nid(page)); in free_debug_processing() 1817 inc_slabs_node(s, page_to_nid(page), page->objects); in allocate_slab() 1872 dec_slabs_node(s, page_to_nid(page), page->objects); in discard_slab() 2161 struct kmem_cache_node *n = get_node(s, page_to_nid(page)); in deactivate_slab() 2326 n2 = get_node(s, page_to_nid(page)); in unfreeze_partials() 2500 if (node != NUMA_NO_NODE && page_to_nid(page) != node) in node_match() 3012 n = get_node(s, page_to_nid(page)); in __slab_free() 3535 if (page_to_nid(page) != node) { in early_kmem_cache_node_alloc() 4683 node_set(page_to_nid(virt_to_page(track)), l->nodes); in add_location() 4714 node_set(page_to_nid(virt_to_pag in add_location() [all...] |
H A D | hugetlb.c | 1070 int nid = page_to_nid(page); in enqueue_huge_page() 1274 if (cma_release(hugetlb_cma[page_to_nid(page)], page, 1 << order)) in free_gigantic_page() 1346 h->nr_huge_pages_node[page_to_nid(page)]--; in update_and_free_page() 1454 int nid = page_to_nid(page); in __free_huge_page() 1732 prep_new_huge_page(h, page, page_to_nid(page)); in alloc_fresh_huge_page() 1829 int nid = page_to_nid(head); in dissolve_free_huge_page() 1934 h->surplus_huge_pages_node[page_to_nid(page)]++; in alloc_surplus_huge_page() 2514 prep_new_huge_page(h, page, page_to_nid(page)); in gather_bootmem_prealloc() 2624 h->free_huge_pages_node[page_to_nid(page)]--; in try_to_free_low() 5722 int old_nid = page_to_nid(oldpag in move_hugetlb_state() [all...] |
H A D | mempolicy.c | 453 int nid = page_to_nid(page); in queue_pages_required() 926 err = page_to_nid(p); in lookup_node() 2137 if (page && page_to_nid(page) == nid) { in alloc_page_interleave() 2463 int curnid = page_to_nid(page); in mpol_misplaced()
|
H A D | huge_memory.c | 490 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() 500 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() 1444 page_nid = page_to_nid(page); in do_huge_pmd_numa_page() 2652 struct pglist_data *pgdata = NODE_DATA(page_to_nid(head)); in split_huge_page_to_list() 2818 memcg_set_shrinker_bit(memcg, page_to_nid(page), in deferred_split_huge_page()
|
H A D | memremap.c | 149 nid = page_to_nid(first_page); in pageunmap_range()
|
H A D | migrate.c | 1570 nid = page_to_nid(page); in alloc_migration_target() 1665 if (page_to_nid(page) == node) in add_page_for_migration() 1844 err = page ? page_to_nid(page) : -ENOENT; in do_pages_stat_array()
|
/kernel/linux/linux-6.6/kernel/dma/ |
H A D | contiguous.c | 420 if (cma_release(dma_contiguous_pernuma_area[page_to_nid(page)], in dma_free_contiguous() 423 if (cma_release(dma_contiguous_numa_area[page_to_nid(page)], in dma_free_contiguous()
|
/kernel/linux/linux-6.6/mm/ |
H A D | page_ext.c | 162 base = NODE_DATA(page_to_nid(page))->node_page_ext; in lookup_page_ext() 171 index = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext()
|
H A D | hugetlb_vmemmap.c | 321 int nid = page_to_nid((struct page *)start); in vmemmap_remap_free() 385 int nid = page_to_nid((struct page *)start); in alloc_vmemmap_page_list()
|
H A D | list_lru.c | 121 int nid = page_to_nid(virt_to_page(item)); in list_lru_add() 145 int nid = page_to_nid(virt_to_page(item)); in list_lru_del()
|
H A D | sparse.c | 46 int page_to_nid(const struct page *page) in page_to_nid() function 50 EXPORT_SYMBOL(page_to_nid); variable
|
H A D | migrate.c | 605 bool f_toptier = node_is_toptier(page_to_nid(&folio->page)); in folio_migrate_flags() 606 bool t_toptier = node_is_toptier(page_to_nid(&newfolio->page)); in folio_migrate_flags() 2099 if (page_to_nid(page) == node) in add_page_for_migration() 2304 err = page_to_nid(page); in do_pages_stat_array() 2601 if (!node_is_toptier(page_to_nid(page)) && node_is_toptier(node)) in migrate_misplaced_page()
|
/kernel/linux/linux-5.10/arch/m68k/include/asm/ |
H A D | page_mm.h | 165 pgdat = &pg_data_map[page_to_nid(__p)]; \
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | mm.h | 1650 extern int page_to_nid(const struct page *page); 1652 static inline int page_to_nid(const struct page *page) in page_to_nid() function 1662 return page_to_nid(&folio->page); in folio_nid() 1762 return page_to_nid(page); /* XXX */ in page_cpupid_xchg_last() 1772 return page_to_nid(page); /* XXX */ in page_cpupid_last() 1870 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone() 1875 return NODE_DATA(page_to_nid(page)); in page_pgdat()
|
/kernel/linux/linux-5.10/kernel/dma/ |
H A D | contiguous.c | 360 if (cma_release(dma_contiguous_pernuma_area[page_to_nid(page)], in dma_free_contiguous()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | mm.h | 1328 extern int page_to_nid(const struct page *page); 1330 static inline int page_to_nid(const struct page *page) in page_to_nid() function 1405 return page_to_nid(page); /* XXX */ in page_cpupid_xchg_last() 1410 return page_to_nid(page); /* XXX */ in page_cpupid_last() 1489 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone() 1494 return NODE_DATA(page_to_nid(page)); in page_pgdat()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/liquidio/ |
H A D | octeon_network.h | 342 unlikely(page_to_nid(pg_info->page) != numa_node_id())) { in recv_buffer_recycle()
|
/kernel/linux/linux-5.10/net/core/ |
H A D | page_pool.c | 121 pref_nid = numa_mem_id(); /* will be zero like page_to_nid() */ in page_pool_refill_alloc_cache() 133 if (likely(page_to_nid(page) == pref_nid)) { in page_pool_refill_alloc_cache()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/liquidio/ |
H A D | octeon_network.h | 342 unlikely(page_to_nid(pg_info->page) != numa_node_id())) { in recv_buffer_recycle()
|
/kernel/linux/linux-5.10/fs/proc/ |
H A D | task_mmu.c | 1780 md->node[page_to_nid(page)] += nr_pages; in gather_stats() 1799 nid = page_to_nid(page); in can_gather_numa_stats() 1824 nid = page_to_nid(page); in can_gather_numa_stats_pmd()
|
/kernel/linux/linux-6.6/fs/proc/ |
H A D | task_mmu.c | 1828 md->node[page_to_nid(page)] += nr_pages; in gather_stats() 1847 nid = page_to_nid(page); in can_gather_numa_stats() 1872 nid = page_to_nid(page); in can_gather_numa_stats_pmd()
|