Lines Matching refs:pfn

184 static void subsection_mask_set(unsigned long *map, unsigned long pfn,
187 int idx = subsection_map_index(pfn);
188 int end = subsection_map_index(pfn + nr_pages - 1);
193 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages)
195 int end_sec = pfn_to_section_nr(pfn + nr_pages - 1);
196 unsigned long nr, start_sec = pfn_to_section_nr(pfn);
206 - (pfn & ~PAGE_SECTION_MASK));
208 subsection_mask_set(ms->usage->subsection_map, pfn, pfns);
211 pfns, subsection_map_index(pfn),
212 subsection_map_index(pfn + pfns - 1));
214 pfn += pfns;
219 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages)
227 unsigned long pfn;
244 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) {
245 unsigned long section = pfn_to_section_nr(pfn);
275 * Subtle, we encode the real pfn into the mem_map such that
276 * the identity pfn - section_mem_map will return the actual
428 struct page __init *__populate_section_memmap(unsigned long pfn,
521 unsigned long pfn = section_nr_to_pfn(pnum);
526 map = __populate_section_memmap(pfn, PAGES_PER_SECTION,
591 /* Mark all memory sections within the pfn range as online */
594 unsigned long pfn;
596 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
597 unsigned long section_nr = pfn_to_section_nr(pfn);
609 /* Mark all memory sections within the pfn range as offline */
612 unsigned long pfn;
614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
615 unsigned long section_nr = pfn_to_section_nr(pfn);
631 static struct page * __meminit populate_section_memmap(unsigned long pfn,
635 return __populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);
638 static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,
641 unsigned long start = (unsigned long) pfn_to_page(pfn);
654 static int clear_subsection_map(unsigned long pfn, unsigned long nr_pages)
658 struct mem_section *ms = __pfn_to_section(pfn);
662 subsection_mask_set(map, pfn, nr_pages);
668 pfn, nr_pages))
681 static int fill_subsection_map(unsigned long pfn, unsigned long nr_pages)
683 struct mem_section *ms = __pfn_to_section(pfn);
688 subsection_mask_set(map, pfn, nr_pages);
703 static struct page * __meminit populate_section_memmap(unsigned long pfn,
711 static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,
714 kvfree(pfn_to_page(pfn));
747 static int clear_subsection_map(unsigned long pfn, unsigned long nr_pages)
757 static int fill_subsection_map(unsigned long pfn, unsigned long nr_pages)
779 static void section_deactivate(unsigned long pfn, unsigned long nr_pages,
782 struct mem_section *ms = __pfn_to_section(pfn);
787 if (clear_subsection_map(pfn, nr_pages))
792 unsigned long section_nr = pfn_to_section_nr(pfn);
820 depopulate_section_memmap(pfn, nr_pages, altmap);
828 static struct page * __meminit section_activate(int nid, unsigned long pfn,
832 struct mem_section *ms = __pfn_to_section(pfn);
844 rc = fill_subsection_map(pfn, nr_pages);
860 return pfn_to_page(pfn);
862 memmap = populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);
864 section_deactivate(pfn, nr_pages, altmap);
874 * @start_pfn: start pfn of the memory range
925 void sparse_remove_section(unsigned long pfn, unsigned long nr_pages,
928 struct mem_section *ms = __pfn_to_section(pfn);
933 section_deactivate(pfn, nr_pages, altmap);