/kernel/linux/linux-5.10/mm/ |
H A D | memory_hotplug.c | 173 static void register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 180 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section() 208 static void register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 215 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section() 354 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 356 unsigned long start_pfn, in find_smallest_section_pfn() 359 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 360 if (unlikely(!pfn_to_online_page(start_pfn))) in find_smallest_section_pfn() 363 if (unlikely(pfn_to_nid(start_pfn) ! in find_smallest_section_pfn() 355 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_smallest_section_pfn() argument 376 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_biggest_section_pfn() argument 400 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) shrink_zone_span() argument 471 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) remove_pfn_range_from_zone() argument 611 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) online_pages_range() argument 661 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) resize_zone_range() argument 672 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) resize_pgdat_range() argument 692 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) move_pfn_range_to_zone() argument 729 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_kernel_zone_for_pfn() argument 745 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_zone_for_pfn() argument 769 zone_for_pfn_range(int online_type, int nid, unsigned long start_pfn, unsigned long nr_pages) zone_for_pfn_range() argument 1193 test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn) test_pages_in_a_zone() argument 1280 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) do_migrate_range() argument 1453 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) count_system_ram_pages_cb() argument 1462 offline_pages(unsigned long start_pfn, unsigned long nr_pages) offline_pages() argument [all...] |
H A D | page_isolation.c | 146 * @start_pfn: The lower PFN of the range to be isolated. 148 * start_pfn/end_pfn must be aligned to pageblock_order. 183 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 190 BUG_ON(!IS_ALIGNED(start_pfn, pageblock_nr_pages)); in start_isolate_page_range() 193 for (pfn = start_pfn; in start_isolate_page_range() 206 for (pfn = start_pfn; in start_isolate_page_range() 221 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 227 BUG_ON(!IS_ALIGNED(start_pfn, pageblock_nr_pages)); in undo_isolate_page_range() 230 for (pfn = start_pfn; in undo_isolate_page_range() 241 * all pages in [start_pfn 284 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) test_pages_isolated() argument [all...] |
H A D | page_ext.c | 292 static int __meminit online_page_ext(unsigned long start_pfn, in online_page_ext() argument 299 start = SECTION_ALIGN_DOWN(start_pfn); in online_page_ext() 300 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 305 * "start_pfn" passed to us is a pfn which is an arg for in online_page_ext() 306 * online__pages(), and start_pfn should exist. in online_page_ext() 308 nid = pfn_to_nid(start_pfn); in online_page_ext() 324 static int __meminit offline_page_ext(unsigned long start_pfn, in offline_page_ext() argument 329 start = SECTION_ALIGN_DOWN(start_pfn); in offline_page_ext() 330 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 346 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 378 unsigned long start_pfn, end_pfn; page_ext_init() local [all...] |
/kernel/linux/linux-5.10/arch/mips/loongson64/ |
H A D | numa.c | 105 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local 119 start_pfn = ((node_id << 44) + mem_start) >> PAGE_SHIFT; in szmem() 121 end_pfn = start_pfn + node_psize; in szmem() 125 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", in szmem() 126 start_pfn, end_pfn, num_physpages); in szmem() 127 memblock_add_node(PFN_PHYS(start_pfn), in szmem() 131 start_pfn = ((node_id << 44) + mem_start) >> PAGE_SHIFT; in szmem() 133 end_pfn = start_pfn + node_psize; in szmem() 137 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", in szmem() 138 start_pfn, end_pf in szmem() 155 unsigned long start_pfn, end_pfn; node_mem_init() local [all...] |
/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | init.c | 333 unsigned long start_pfn, unsigned long end_pfn, in save_mr() 336 if (start_pfn < end_pfn) { in save_mr() 339 mr[nr_range].start = start_pfn<<PAGE_SHIFT; in save_mr() 411 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 418 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 435 if (start_pfn < end_pfn) { in split_mem_range() 436 nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0); in split_mem_range() 441 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 450 if (start_pfn < end_pfn) { in split_mem_range() 451 nr_range = save_mr(mr, nr_range, start_pfn, end_pf in split_mem_range() 332 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) save_mr() argument 510 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) add_pfn_range_mapped() argument 523 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) pfn_range_is_mapped() argument 580 unsigned long start_pfn, end_pfn; init_range_memory_mapping() local 994 unsigned long start_pfn, end_pfn; memblock_find_dma_reserve() local [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | init.c | 328 unsigned long start_pfn, unsigned long end_pfn, in save_mr() 331 if (start_pfn < end_pfn) { in save_mr() 334 mr[nr_range].start = start_pfn<<PAGE_SHIFT; in save_mr() 406 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 413 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 430 if (start_pfn < end_pfn) { in split_mem_range() 431 nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0); in split_mem_range() 436 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 445 if (start_pfn < end_pfn) { in split_mem_range() 446 nr_range = save_mr(mr, nr_range, start_pfn, end_pf in split_mem_range() 327 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) save_mr() argument 505 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) add_pfn_range_mapped() argument 518 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) pfn_range_is_mapped() argument 575 unsigned long start_pfn, end_pfn; init_range_memory_mapping() local 1005 unsigned long start_pfn, end_pfn; memblock_find_dma_reserve() local [all...] |
/kernel/linux/linux-5.10/arch/sparc/mm/ |
H A D | init_32.c | 65 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; in calc_highpages() local 71 if (start_pfn < max_low_pfn) in calc_highpages() 72 start_pfn = max_low_pfn; in calc_highpages() 74 nr += end_pfn - start_pfn; in calc_highpages() 134 unsigned long start_pfn, bytes_avail, size; in bootmem_init() local 172 start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end)); in bootmem_init() 175 start_pfn >>= PAGE_SHIFT; in bootmem_init() 193 size = (start_pfn << PAGE_SHIFT) - phys_base; in bootmem_init() 236 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 241 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pf in map_high_region() 284 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; mem_init() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | mm_init.c | 300 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local 303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages() 304 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages() 462 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local 480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes() 483 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes() 484 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes() 488 if (start_pfn < usable_startpfn) { in find_zone_movable_pfns_for_nodes() 491 - start_pfn; in find_zone_movable_pfns_for_nodes() 549 unsigned long start_pfn, end_pfn; find_zone_movable_pfns_for_nodes() local 602 unsigned long start_pfn, end_pfn; __early_pfn_to_nid() local 751 unsigned long start_pfn = PFN_DOWN(start); reserve_bootmem_region() local 848 memmap_init_range(unsigned long size, int nid, unsigned long zone, unsigned long start_pfn, unsigned long zone_end_pfn, enum meminit_context context, struct vmem_altmap *altmap, int migratetype) memmap_init_range() argument 909 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) memmap_init_zone_range() argument 935 unsigned long start_pfn, end_pfn; memmap_init() local 1063 memmap_init_zone_device(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct dev_pagemap *pgmap) memmap_init_zone_device() argument 1151 unsigned long start_pfn, end_pfn; __absent_pages_in_range() local 1169 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) absent_pages_in_range() argument 1195 unsigned long start_pfn, end_pfn; zone_absent_pages_in_node() local 1700 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) get_pfn_range_for_nid() argument 1721 unsigned long start_pfn = 0; free_area_init_node() local 1806 unsigned long start_pfn, end_pfn; free_area_init() local 2121 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) deferred_init_maxorder() argument 2165 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) deferred_init_memmap_chunk() argument [all...] |
H A D | memory_hotplug.c | 414 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 416 unsigned long start_pfn, in find_smallest_section_pfn() 419 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 420 if (unlikely(!pfn_to_online_page(start_pfn))) in find_smallest_section_pfn() 423 if (unlikely(pfn_to_nid(start_pfn) != nid)) in find_smallest_section_pfn() 426 if (zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 429 return start_pfn; in find_smallest_section_pfn() 435 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */ 437 unsigned long start_pfn, in find_biggest_section_pfn() 415 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_smallest_section_pfn() argument 436 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) find_biggest_section_pfn() argument 460 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) shrink_zone_span() argument 528 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) remove_pfn_range_from_zone() argument 645 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) online_pages_range() argument 705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) resize_zone_range() argument 716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) resize_pgdat_range() argument 750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) move_pfn_range_to_zone() argument 897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_kernel_zone_for_pfn() argument 1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_zone_for_pfn() argument 1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) zone_for_pfn_range() argument 1707 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) do_migrate_range() argument 1863 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) count_system_ram_pages_cb() argument 1875 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) offline_pages() argument [all...] |
H A D | page_isolation.c | 19 * This function checks whether the range [start_pfn, end_pfn) includes 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, in has_unmovable_pages() argument 36 struct page *page = pfn_to_page(start_pfn); in has_unmovable_pages() 40 VM_BUG_ON(pageblock_start_pfn(start_pfn) != in has_unmovable_pages() 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in has_unmovable_pages() 144 * present in [start_pfn, end_pfn). The pageblock must intersect with 145 * [start_pfn, end_pfn). 148 unsigned long start_pfn, unsigned long end_pfn) in set_migratetype_isolate() 171 * Pass the intersection of [start_pfn, end_pfn) and the page's pageblock in set_migratetype_isolate() 174 check_unmovable_start = max(page_to_pfn(page), start_pfn); in set_migratetype_isolate() 147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) set_migratetype_isolate() argument 309 unsigned long start_pfn; isolate_single_pageblock() local 528 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype, int flags, gfp_t gfp_flags) start_isolate_page_range() argument 582 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype) undo_isolate_page_range() argument 654 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) test_pages_isolated() argument [all...] |
H A D | page_ext.c | 354 static int __meminit online_page_ext(unsigned long start_pfn, in online_page_ext() argument 361 start = SECTION_ALIGN_DOWN(start_pfn); in online_page_ext() 362 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 367 * "start_pfn" passed to us is a pfn which is an arg for in online_page_ext() 368 * online__pages(), and start_pfn should exist. in online_page_ext() 370 nid = pfn_to_nid(start_pfn); in online_page_ext() 387 static void __meminit offline_page_ext(unsigned long start_pfn, in offline_page_ext() argument 392 start = SECTION_ALIGN_DOWN(start_pfn); in offline_page_ext() 393 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 421 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 451 unsigned long start_pfn, end_pfn; page_ext_init() local [all...] |
/kernel/linux/linux-5.10/arch/x86/xen/ |
H A D | setup.c | 94 static void __init xen_add_extra_mem(unsigned long start_pfn, in xen_add_extra_mem() argument 106 xen_extra_mem[i].start_pfn = start_pfn; in xen_add_extra_mem() 111 if (xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns == in xen_add_extra_mem() 112 start_pfn) { in xen_add_extra_mem() 120 memblock_reserve(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns)); in xen_add_extra_mem() 123 static void __init xen_del_extra_mem(unsigned long start_pfn, in xen_del_extra_mem() argument 130 start_r = xen_extra_mem[i].start_pfn; in xen_del_extra_mem() 134 if (start_r == start_pfn) { in xen_del_extra_mem() 136 xen_extra_mem[i].start_pfn in xen_del_extra_mem() 254 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages) xen_set_identity_and_release_chunk() argument 333 xen_do_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) xen_do_set_identity_and_remap_chunk() argument 387 xen_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn) xen_set_identity_and_remap_chunk() argument 443 xen_count_remap_pages( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pages) xen_count_remap_pages() argument 453 xen_foreach_remap_area(unsigned long nr_pages, unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long last_val)) xen_foreach_remap_area() argument 476 unsigned long start_pfn = PFN_DOWN(start); xen_foreach_remap_area() local [all...] |
/kernel/linux/linux-6.6/arch/x86/xen/ |
H A D | setup.c | 99 static void __init xen_add_extra_mem(unsigned long start_pfn, in xen_add_extra_mem() argument 111 xen_extra_mem[i].start_pfn = start_pfn; in xen_add_extra_mem() 116 if (xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns == in xen_add_extra_mem() 117 start_pfn) { in xen_add_extra_mem() 125 memblock_reserve(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns)); in xen_add_extra_mem() 128 static void __init xen_del_extra_mem(unsigned long start_pfn, in xen_del_extra_mem() argument 135 start_r = xen_extra_mem[i].start_pfn; in xen_del_extra_mem() 139 if (start_r == start_pfn) { in xen_del_extra_mem() 141 xen_extra_mem[i].start_pfn in xen_del_extra_mem() 259 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages) xen_set_identity_and_release_chunk() argument 334 xen_do_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) xen_do_set_identity_and_remap_chunk() argument 388 xen_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn) xen_set_identity_and_remap_chunk() argument 444 xen_count_remap_pages( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pages) xen_count_remap_pages() argument 454 xen_foreach_remap_area(unsigned long nr_pages, unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long last_val)) xen_foreach_remap_area() argument 477 unsigned long start_pfn = PFN_DOWN(start); xen_foreach_remap_area() local [all...] |
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | page_isolation.h | 13 unsigned long start_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 20 __field(unsigned long, start_pfn) 26 __entry->start_pfn = start_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | page_isolation.h | 13 unsigned long start_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 20 __field(unsigned long, start_pfn) 26 __entry->start_pfn = start_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
|
H A D | compaction.h | 17 unsigned long start_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 25 __field(unsigned long, start_pfn) 32 __entry->start_pfn = start_pfn; 39 __entry->start_pfn, 48 unsigned long start_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 59 unsigned long start_pfn, 64 TP_ARGS(start_pfn, end_pf [all...] |
/kernel/linux/linux-6.6/arch/powerpc/platforms/powernv/ |
H A D | memtrace.c | 91 static void memtrace_clear_range(unsigned long start_pfn, in memtrace_clear_range() argument 97 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) { in memtrace_clear_range() 106 flush_dcache_range_chunked((unsigned long)pfn_to_kaddr(start_pfn), in memtrace_clear_range() 107 (unsigned long)pfn_to_kaddr(start_pfn + nr_pages), in memtrace_clear_range() 114 unsigned long pfn, start_pfn; in memtrace_alloc_node() local 125 start_pfn = page_to_pfn(page); in memtrace_alloc_node() 132 memtrace_clear_range(start_pfn, nr_pages); in memtrace_alloc_node() 138 for (pfn = start_pfn; pfn < start_pfn in memtrace_alloc_node() 217 const unsigned long start_pfn = PHYS_PFN(start); memtrace_free() local [all...] |
/kernel/linux/linux-6.6/arch/sparc/mm/ |
H A D | init_32.c | 64 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; in calc_highpages() local 70 if (start_pfn < max_low_pfn) in calc_highpages() 71 start_pfn = max_low_pfn; in calc_highpages() 73 nr += end_pfn - start_pfn; in calc_highpages() 133 unsigned long start_pfn, bytes_avail, size; in bootmem_init() local 171 start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end)); in bootmem_init() 174 start_pfn >>= PAGE_SHIFT; in bootmem_init() 192 size = (start_pfn << PAGE_SHIFT) - phys_base; in bootmem_init() 235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pf in map_high_region() 283 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; mem_init() local [all...] |
/kernel/linux/linux-5.10/arch/sh/mm/ |
H A D | init.c | 210 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 223 NODE_DATA(nid)->node_start_pfn = start_pfn; in allocate_pgdat() 224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 229 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem() 234 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem() 247 unsigned long start_pfn; in early_reserve_mem() local 255 start_pfn = PFN_UP(__pa(_end)); in early_reserve_mem() 263 memblock_reserve(start, (PFN_PHYS(start_pfn) in early_reserve_mem() 411 unsigned long start_pfn = PFN_DOWN(start); arch_add_memory() local 429 unsigned long start_pfn = PFN_DOWN(start); arch_remove_memory() local [all...] |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 33 start_pfn = PFN_DOWN(start); in setup_bootmem_node() 41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 51 NODE_DATA(nid)->node_start_pfn = start_pfn; in setup_bootmem_node() 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
/kernel/linux/linux-6.6/arch/sh/mm/ |
H A D | init.c | 210 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 223 NODE_DATA(nid)->node_start_pfn = start_pfn; in allocate_pgdat() 224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 229 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem() 234 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem() 247 unsigned long start_pfn; in early_reserve_mem() local 255 start_pfn = PFN_UP(__pa(_end)); in early_reserve_mem() 263 memblock_reserve(start, (PFN_PHYS(start_pfn) in early_reserve_mem() 403 unsigned long start_pfn = PFN_DOWN(start); arch_add_memory() local 420 unsigned long start_pfn = PFN_DOWN(start); arch_remove_memory() local [all...] |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 33 start_pfn = PFN_DOWN(start); in setup_bootmem_node() 41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 51 NODE_DATA(nid)->node_start_pfn = start_pfn; in setup_bootmem_node() 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
/kernel/linux/linux-5.10/drivers/hv/ |
H A D | hv_balloon.c | 421 * The range start_pfn : end_pfn specifies the range 423 * start_pfn : ha_end_pfn specifies the range that we have 433 unsigned long start_pfn; member 446 unsigned long start_pfn; member 584 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) in has_pfn_is_backed() 591 static unsigned long hv_page_offline_check(unsigned long start_pfn, in hv_page_offline_check() argument 594 unsigned long pfn = start_pfn, count = 0; in hv_page_offline_check() 598 while (pfn < start_pfn + nr_pages) { in hv_page_offline_check() 605 while ((pfn >= has->start_pfn) && in hv_page_offline_check() 607 (pfn < start_pfn in hv_page_offline_check() 687 hv_bring_pgs_online(struct hv_hotadd_state *has, unsigned long start_pfn, unsigned long size) hv_bring_pgs_online() argument 703 unsigned long start_pfn; hv_mem_hot_add() local 782 pfn_covered(unsigned long start_pfn, unsigned long pfn_cnt) pfn_covered() argument 845 unsigned long start_pfn = pg_start; handle_pg_range() local [all...] |
/kernel/linux/linux-6.6/drivers/hv/ |
H A D | hv_balloon.c | 425 * The range start_pfn : end_pfn specifies the range 427 * start_pfn : ha_end_pfn specifies the range that we have 437 unsigned long start_pfn; member 450 unsigned long start_pfn; member 603 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) in has_pfn_is_backed() 610 static unsigned long hv_page_offline_check(unsigned long start_pfn, in hv_page_offline_check() argument 613 unsigned long pfn = start_pfn, count = 0; in hv_page_offline_check() 617 while (pfn < start_pfn + nr_pages) { in hv_page_offline_check() 624 while ((pfn >= has->start_pfn) && in hv_page_offline_check() 626 (pfn < start_pfn in hv_page_offline_check() 707 hv_bring_pgs_online(struct hv_hotadd_state *has, unsigned long start_pfn, unsigned long size) hv_bring_pgs_online() argument 723 unsigned long start_pfn; hv_mem_hot_add() local 799 pfn_covered(unsigned long start_pfn, unsigned long pfn_cnt) pfn_covered() argument 860 unsigned long start_pfn = pg_start; handle_pg_range() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | memory_hotplug.h | 117 extern struct zone *test_pages_in_a_zone(unsigned long start_pfn, 119 extern void __offline_isolated_pages(unsigned long start_pfn, 147 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, 151 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 155 static inline int add_pages(int nid, unsigned long start_pfn, in add_pages() argument 158 return __add_pages(nid, start_pfn, nr_pages, params); in add_pages() 161 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 315 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages); 323 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages) in offline_pages() argument 348 extern void move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, [all...] |