/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | init.c | 333 unsigned long start_pfn, unsigned long end_pfn, in save_mr() 336 if (start_pfn < end_pfn) { in save_mr() 340 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 411 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 427 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 429 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 431 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 433 if (end_pfn > limit_pfn) in split_mem_range() 434 end_pfn = limit_pfn; in split_mem_range() 435 if (start_pfn < end_pfn) { in split_mem_range() 332 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) save_mr() argument 510 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) add_pfn_range_mapped() argument 523 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) pfn_range_is_mapped() argument 580 unsigned long start_pfn, end_pfn; init_range_memory_mapping() local 994 unsigned long start_pfn, end_pfn; memblock_find_dma_reserve() local [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | init.c | 328 unsigned long start_pfn, unsigned long end_pfn, in save_mr() 331 if (start_pfn < end_pfn) { in save_mr() 335 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 406 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 422 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 424 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 426 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 428 if (end_pfn > limit_pfn) in split_mem_range() 429 end_pfn = limit_pfn; in split_mem_range() 430 if (start_pfn < end_pfn) { in split_mem_range() 327 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) save_mr() argument 505 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) add_pfn_range_mapped() argument 518 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) pfn_range_is_mapped() argument 575 unsigned long start_pfn, end_pfn; init_range_memory_mapping() local 1005 unsigned long start_pfn, end_pfn; memblock_find_dma_reserve() local [all...] |
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
H A D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanne [all...] |
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
H A D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanne [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | page_isolation.c | 147 * @end_pfn: The upper PFN of the range to be isolated. 148 * start_pfn/end_pfn must be aligned to pageblock_order. 183 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 191 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in start_isolate_page_range() 194 pfn < end_pfn; in start_isolate_page_range() 221 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 228 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in undo_isolate_page_range() 231 pfn < end_pfn; in undo_isolate_page_range() 241 * all pages in [start_pfn...end_pfn) must be in the same zone. 247 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument 284 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) test_pages_isolated() argument [all...] |
H A D | page_idle.c | 124 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 134 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 135 if (end_pfn > max_pfn) in page_idle_bitmap_read() 136 end_pfn = max_pfn; in page_idle_bitmap_read() 138 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 169 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 179 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 180 if (end_pfn > max_pfn) in page_idle_bitmap_write() 181 end_pfn = max_pfn; in page_idle_bitmap_write() 183 for (; pfn < end_pfn; pf in page_idle_bitmap_write() [all...] |
H A D | memory_hotplug.c | 234 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 245 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 248 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 313 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local 341 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages() 343 cur_nr_pages = min(end_pfn - pfn, in __add_pages() 354 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 357 unsigned long end_pfn) in find_smallest_section_pfn() 359 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 375 /* find the biggest valid pfn in the range [start_pfn, end_pfn) 355 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_smallest_section_pfn() argument 376 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_biggest_section_pfn() argument 400 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) shrink_zone_span() argument 475 const unsigned long end_pfn = start_pfn + nr_pages; remove_pfn_range_from_zone() local 536 const unsigned long end_pfn = pfn + nr_pages; __remove_pages() local 613 const unsigned long end_pfn = start_pfn + nr_pages; online_pages_range() local 1193 test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn) test_pages_in_a_zone() argument 1280 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) do_migrate_range() argument 1464 const unsigned long end_pfn = start_pfn + nr_pages; offline_pages() local [all...] |
H A D | page_alloc.c | 441 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 449 if (prev_end_pfn != end_pfn) { in defer_init() 450 prev_end_pfn = end_pfn; in defer_init() 455 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) in defer_init() 480 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 1500 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local 1502 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 1580 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local 1586 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid() 1589 state->last_end = end_pfn; in __early_pfn_to_nid() 1637 __pageblock_pfn_to_page(unsigned long start_pfn, unsigned long end_pfn, struct zone *zone) __pageblock_pfn_to_page() argument 1752 deferred_free_pages(unsigned long pfn, unsigned long end_pfn) deferred_free_pages() argument 1778 deferred_init_pages(struct zone *zone, unsigned long pfn, unsigned long end_pfn) deferred_init_pages() argument 1844 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) deferred_init_maxorder() argument 1888 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) deferred_init_memmap_chunk() argument 2372 move_freepages(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, int migratetype, int *num_movable) move_freepages() argument 2417 unsigned long start_pfn, end_pfn, pfn; move_freepages_block() local 6216 unsigned long pfn, end_pfn = start_pfn + size; memmap_init_zone() local 6276 unsigned long pfn, end_pfn = start_pfn + nr_pages; memmap_init_zone_device() local 6401 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) memmap_init_zone_range() argument 6427 unsigned long start_pfn, end_pfn; memmap_init() local 6690 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) get_pfn_range_for_nid() argument 6811 unsigned long start_pfn, end_pfn; __absent_pages_in_range() local 6829 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) absent_pages_in_range() argument 6864 unsigned long start_pfn, end_pfn; zone_absent_pages_in_node() local 7238 unsigned long end_pfn = 0; free_area_init_node() local 7347 unsigned long start_pfn, end_pfn; early_calculate_totalpages() local 7478 unsigned long start_pfn, end_pfn; find_zone_movable_pfns_for_nodes() local 7565 unsigned long start_pfn, end_pfn; find_zone_movable_pfns_for_nodes() local 7621 unsigned long start_pfn, end_pfn; free_area_init() local 8794 unsigned long end_pfn = start_pfn + nr_pages; __alloc_contig_pages() local 8803 unsigned long i, end_pfn = start_pfn + nr_pages; pfn_range_valid_contig() local 8942 __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) __offline_isolated_pages() argument [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | mm_init.c | 300 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local 303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages() 304 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages() 462 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local 480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes() 484 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes() 490 kernel_pages = min(end_pfn, usable_startpfn) in find_zone_movable_pfns_for_nodes() 499 if (end_pfn <= usable_startpfn) { in find_zone_movable_pfns_for_nodes() 507 zone_movable_pfn[nid] = end_pfn; in find_zone_movable_pfns_for_nodes() 515 * start_pfn->end_pfn in find_zone_movable_pfns_for_nodes() 549 unsigned long start_pfn, end_pfn; find_zone_movable_pfns_for_nodes() local 602 unsigned long start_pfn, end_pfn; __early_pfn_to_nid() local 672 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) defer_init() argument 732 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) defer_init() argument 752 unsigned long end_pfn = PFN_UP(end); reserve_bootmem_region() local 853 unsigned long pfn, end_pfn = start_pfn + size; memmap_init_range() local 909 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) memmap_init_zone_range() argument 935 unsigned long start_pfn, end_pfn; memmap_init() local 1042 unsigned long pfn, end_pfn = head_pfn + nr_pages; memmap_init_compound() local 1068 unsigned long pfn, end_pfn = start_pfn + nr_pages; memmap_init_zone_device() local 1151 unsigned long start_pfn, end_pfn; __absent_pages_in_range() local 1169 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) absent_pages_in_range() argument 1195 unsigned long start_pfn, end_pfn; zone_absent_pages_in_node() local 1700 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) get_pfn_range_for_nid() argument 1722 unsigned long end_pfn = 0; free_area_init_node() local 1806 unsigned long start_pfn, end_pfn; free_area_init() local 2031 deferred_free_pages(unsigned long pfn, unsigned long end_pfn) deferred_free_pages() argument 2056 deferred_init_pages(struct zone *zone, unsigned long pfn, unsigned long end_pfn) deferred_init_pages() argument 2121 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) deferred_init_maxorder() argument 2165 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) deferred_init_memmap_chunk() argument [all...] |
H A D | page_idle.c | 120 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 130 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 131 if (end_pfn > max_pfn) in page_idle_bitmap_read() 132 end_pfn = max_pfn; in page_idle_bitmap_read() 134 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 165 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 175 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 176 if (end_pfn > max_pfn) in page_idle_bitmap_write() 177 end_pfn = max_pfn; in page_idle_bitmap_write() 179 for (; pfn < end_pfn; pf in page_idle_bitmap_write() [all...] |
H A D | page_isolation.c | 19 * This function checks whether the range [start_pfn, end_pfn) includes 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, in has_unmovable_pages() argument 41 pageblock_start_pfn(end_pfn - 1)); in has_unmovable_pages() 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in has_unmovable_pages() 144 * present in [start_pfn, end_pfn). The pageblock must intersect with 145 * [start_pfn, end_pfn). 148 unsigned long start_pfn, unsigned long end_pfn) in set_migratetype_isolate() 171 * Pass the intersection of [start_pfn, end_pfn) and the page's pageblock in set_migratetype_isolate() 176 end_pfn); in set_migratetype_isolate() 486 * @end_pfn 147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) set_migratetype_isolate() argument 528 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype, int flags, gfp_t gfp_flags) start_isolate_page_range() argument 582 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype) undo_isolate_page_range() argument 607 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, int flags) __test_page_isolated_in_pageblock() argument 654 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) test_pages_isolated() argument [all...] |
H A D | memory_hotplug.c | 373 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local 400 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages() 402 cur_nr_pages = min(end_pfn - pfn, in __add_pages() 414 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 417 unsigned long end_pfn) in find_smallest_section_pfn() 419 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 435 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */ 438 unsigned long end_pfn) in find_biggest_section_pfn() 443 pfn = end_pfn - 1; in find_biggest_section_pfn() 461 unsigned long end_pfn) in shrink_zone_span() 415 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_smallest_section_pfn() argument 436 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_biggest_section_pfn() argument 460 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) shrink_zone_span() argument 507 unsigned long end_pfn = zone_end_pfn(zone); update_pgdat_span() local 532 const unsigned long end_pfn = start_pfn + nr_pages; remove_pfn_range_from_zone() local 577 const unsigned long end_pfn = pfn + nr_pages; __remove_pages() local 647 const unsigned long end_pfn = start_pfn + nr_pages; online_pages_range() local 968 unsigned long online_pages = 0, max_pages, end_pfn; auto_movable_zone_for_pfn() local 1092 unsigned long end_pfn = pfn + nr_pages; mhp_init_memmap_on_memory() local 1117 unsigned long end_pfn = pfn + nr_pages; mhp_deinit_memmap_on_memory() local 1707 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) do_migrate_range() argument 1878 const unsigned long end_pfn = start_pfn + nr_pages; offline_pages() local [all...] |
/kernel/linux/linux-5.10/arch/mips/loongson64/ |
H A D | numa.c | 105 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local 121 end_pfn = start_pfn + node_psize; in szmem() 125 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", in szmem() 126 start_pfn, end_pfn, num_physpages); in szmem() 133 end_pfn = start_pfn + node_psize; in szmem() 137 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", in szmem() 138 start_pfn, end_pfn, num_physpages); in szmem() 155 unsigned long start_pfn, end_pfn; in node_mem_init() local 161 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 162 pr_info("Node%d: start_pfn=0x%lx, end_pfn in node_mem_init() [all...] |
/kernel/linux/linux-5.10/arch/sparc/mm/ |
H A D | init_32.c | 66 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 68 if (end_pfn <= max_low_pfn) in calc_highpages() 74 nr += end_pfn - start_pfn; in calc_highpages() 236 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 241 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 244 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 285 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 287 if (end_pfn <= highstart_pfn) in mem_init() 293 map_high_region(start_pfn, end_pfn); in mem_init()
|
/kernel/linux/linux-5.10/arch/sh/mm/ |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 34 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
/kernel/linux/linux-6.6/arch/sh/mm/ |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 34 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
/kernel/linux/linux-6.6/arch/sparc/mm/ |
H A D | init_32.c | 65 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 67 if (end_pfn <= max_low_pfn) in calc_highpages() 73 nr += end_pfn - start_pfn; in calc_highpages() 235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 243 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 284 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 286 if (end_pfn <= highstart_pfn) in mem_init() 292 map_high_region(start_pfn, end_pfn); in mem_init()
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/ |
H A D | memtrace.c | 115 u64 start_pfn, end_pfn, nr_pages, pfn; in memtrace_alloc_node() local 123 end_pfn = node_end_pfn(nid); in memtrace_alloc_node() 127 end_pfn = round_down(end_pfn - nr_pages, nr_pages); in memtrace_alloc_node() 130 for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { in memtrace_alloc_node() 143 end_pfn = base_pfn + nr_pages; in memtrace_alloc_node() 144 for (pfn = base_pfn; pfn < end_pfn; pfn += bytes>> PAGE_SHIFT) { in memtrace_alloc_node()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | node.h | 97 unsigned long end_pfn, 101 unsigned long end_pfn, in register_memory_blocks_under_node() 121 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() local 126 register_memory_blocks_under_node(nid, start_pfn, end_pfn, in register_one_node() 100 register_memory_blocks_under_node(int nid, unsigned long start_pfn, unsigned long end_pfn, enum meminit_context context) register_memory_blocks_under_node() argument
|
H A D | page-isolation.h | 40 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 43 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 46 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | page-isolation.h | 43 * Changes migrate type in [start_pfn, end_pfn) to be MIGRATE_ISOLATE. 46 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 51 * target range is [start_pfn, end_pfn) 54 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 58 * Test all pages in [start_pfn, end_pfn) are isolated or not. 60 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
H A D | node.h | 103 unsigned long end_pfn, 107 unsigned long end_pfn, in link_mem_sections() 126 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() local 132 link_mem_sections(nid, start_pfn, end_pfn, MEMINIT_EARLY); in register_one_node() 106 link_mem_sections(int nid, unsigned long start_pfn, unsigned long end_pfn, enum meminit_context context) link_mem_sections() argument
|
/kernel/linux/linux-6.6/arch/mips/loongson64/ |
H A D | numa.c | 88 unsigned long start_pfn, end_pfn; in node_mem_init() local 97 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 98 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n", in node_mem_init() 99 node, start_pfn, end_pfn); in node_mem_init() 112 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 122 max_low_pfn = end_pfn; in node_mem_init()
|