Lines Matching refs:usage

208 		subsection_mask_set(ms->usage->subsection_map, pfn, pfns);
302 struct mem_section_usage *usage, unsigned long flags)
307 ms->usage = usage;
335 struct mem_section_usage *usage;
352 usage = memblock_alloc_try_nid(size, SMP_CACHE_BYTES, goal, limit, nid);
353 if (!usage && limit) {
357 return usage;
361 struct mem_section_usage *usage)
375 usemap_snr = pfn_to_section_nr(__pa(usage) >> PAGE_SHIFT);
411 struct mem_section_usage *usage)
509 struct mem_section_usage *usage;
513 usage = sparse_early_usemaps_alloc_pgdat_section(NODE_DATA(nid),
515 if (!usage) {
535 check_usemap_section_nr(nid, usage);
536 sparse_init_one_section(__nr_to_section(pnum), pnum, map, usage,
538 usage = (void *) usage + mem_section_usage_size();
659 unsigned long *subsection_map = ms->usage
660 ? &ms->usage->subsection_map[0] : NULL;
677 return bitmap_empty(&ms->usage->subsection_map[0],
690 subsection_map = &ms->usage->subsection_map[0];
775 * usage map, but still need to free the vmemmap range.
797 * ms->usage array.
802 * When removing an early section, the usage map is kept (as the
803 * usage maps of other sections fall into the same page). It
805 * longer an early section. If the usage map is PageReserved, it
808 if (!PageReserved(virt_to_page(ms->usage))) {
809 kfree_rcu(ms->usage, rcu);
810 WRITE_ONCE(ms->usage, NULL);
833 struct mem_section_usage *usage = NULL;
837 if (!ms->usage) {
838 usage = kzalloc(mem_section_usage_size(), GFP_KERNEL);
839 if (!usage)
841 ms->usage = usage;
846 if (usage)
847 ms->usage = NULL;
848 kfree(usage);
920 sparse_init_one_section(ms, section_nr, memmap, ms->usage, 0);