Lines Matching refs:start_pfn

300 	unsigned long start_pfn, end_pfn;
303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
304 unsigned long pages = end_pfn - start_pfn;
462 unsigned long start_pfn, end_pfn;
480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
483 start_pfn = max(start_pfn, zone_movable_pfn[nid]);
484 if (start_pfn >= end_pfn)
488 if (start_pfn < usable_startpfn) {
491 - start_pfn;
510 start_pfn = usable_startpfn;
515 * start_pfn->end_pfn. Calculate size_pages as the
518 size_pages = end_pfn - start_pfn;
521 zone_movable_pfn[nid] = start_pfn + size_pages;
549 unsigned long start_pfn, end_pfn;
554 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
602 unsigned long start_pfn, end_pfn;
608 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn);
610 state->last_start = start_pfn;
751 unsigned long start_pfn = PFN_DOWN(start);
754 for (; start_pfn < end_pfn; start_pfn++) {
755 if (pfn_valid(start_pfn)) {
756 struct page *page = pfn_to_page(start_pfn);
758 init_reserved_page(start_pfn, nid);
849 unsigned long start_pfn, unsigned long zone_end_pfn,
853 unsigned long pfn, end_pfn = start_pfn + size;
871 if (start_pfn == altmap->base_pfn)
872 start_pfn += altmap->reserve;
877 for (pfn = start_pfn; pfn < end_pfn; ) {
910 unsigned long start_pfn,
918 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn);
921 if (start_pfn >= end_pfn)
924 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn,
927 if (*hole_pfn < start_pfn)
928 init_unavailable_range(*hole_pfn, start_pfn, zone_id, nid);
935 unsigned long start_pfn, end_pfn;
939 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
948 memmap_init_zone_range(zone, start_pfn, end_pfn,
1064 unsigned long start_pfn,
1068 unsigned long pfn, end_pfn = start_pfn + nr_pages;
1085 start_pfn = altmap->base_pfn + vmem_altmap_offset(altmap);
1086 nr_pages = end_pfn - start_pfn;
1089 for (pfn = start_pfn; pfn < end_pfn; pfn += pfns_per_compound) {
1151 unsigned long start_pfn, end_pfn;
1154 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
1155 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn);
1157 nr_absent -= end_pfn - start_pfn;
1164 * @start_pfn: The start PFN to start searching for holes
1169 unsigned long __init absent_pages_in_range(unsigned long start_pfn,
1172 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn);
1195 unsigned long start_pfn, end_pfn;
1199 start_pfn = clamp(memblock_region_memory_base_pfn(r),
1206 nr_absent += end_pfn - start_pfn;
1210 nr_absent += end_pfn - start_pfn;
1693 * @start_pfn: Passed by reference. On return, it will have the node start_pfn.
1701 unsigned long *start_pfn, unsigned long *end_pfn)
1706 *start_pfn = -1UL;
1710 *start_pfn = min(*start_pfn, this_start_pfn);
1714 if (*start_pfn == -1UL)
1715 *start_pfn = 0;
1721 unsigned long start_pfn = 0;
1727 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
1730 pgdat->node_start_pfn = start_pfn;
1733 if (start_pfn != end_pfn) {
1735 (u64)start_pfn << PAGE_SHIFT,
1738 calculate_node_totalpages(pgdat, start_pfn, end_pfn);
1806 unsigned long start_pfn, end_pfn;
1816 start_pfn = PHYS_PFN(memblock_start_of_DRAM());
1828 end_pfn = max(max_zone_pfn[zone], start_pfn);
1829 arch_zone_lowest_possible_pfn[zone] = start_pfn;
1832 start_pfn = end_pfn;
1870 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
1872 (u64)start_pfn << PAGE_SHIFT,
1874 subsection_map_init(start_pfn, end_pfn - start_pfn);
2121 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn,
2124 unsigned long mo_pfn = ALIGN(*start_pfn + 1, MAX_ORDER_NR_PAGES);
2125 unsigned long spfn = *start_pfn, epfn = *end_pfn;
2130 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) {
2133 if (mo_pfn <= *start_pfn)
2137 nr_pages += deferred_init_pages(zone, *start_pfn, t);
2140 *start_pfn = mo_pfn;
2165 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn,
2172 deferred_init_mem_pfn_range_in_zone(&i, zone, &spfn, &epfn, start_pfn);