/kernel/linux/linux-6.6/mm/ |
H A D | page_counter.c | 47 * @nr_pages: number of pages to cancel 49 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) in page_counter_cancel() argument 53 new = atomic_long_sub_return(nr_pages, &counter->usage); in page_counter_cancel() 55 if (WARN_ONCE(new < 0, "page_counter underflow: %ld nr_pages=%lu\n", in page_counter_cancel() 56 new, nr_pages)) { in page_counter_cancel() 66 * @nr_pages: number of pages to charge 70 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) in page_counter_charge() argument 77 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_charge() 91 * @nr_pages: number of pages to charge 98 unsigned long nr_pages, in page_counter_try_charge() 97 page_counter_try_charge(struct page_counter *counter, unsigned long nr_pages, struct page_counter **fail) page_counter_try_charge() argument 153 page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) page_counter_uncharge() argument 171 page_counter_set_max(struct page_counter *counter, unsigned long nr_pages) page_counter_set_max() argument 210 page_counter_set_min(struct page_counter *counter, unsigned long nr_pages) page_counter_set_min() argument 227 page_counter_set_low(struct page_counter *counter, unsigned long nr_pages) page_counter_set_low() argument 246 page_counter_memparse(const char *buf, const char *max, unsigned long *nr_pages) page_counter_memparse() argument [all...] |
H A D | percpu-km.c | 55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local 65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 71 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk() 78 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk() 89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local 98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk() 109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info() 118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() [all...] |
H A D | gup_test.c | 11 unsigned long nr_pages, unsigned int gup_test_flags) in put_back_pages() 18 for (i = 0; i < nr_pages; i++) in put_back_pages() 25 unpin_user_pages(pages, nr_pages); in put_back_pages() 29 unpin_user_pages(pages, nr_pages); in put_back_pages() 31 for (i = 0; i < nr_pages; i++) in put_back_pages() 40 unsigned long nr_pages) in verify_dma_pinned() 49 for (i = 0; i < nr_pages; i++) { in verify_dma_pinned() 70 unsigned long nr_pages) in dump_pages_test() 80 if (gup->which_pages[i] > nr_pages) { in dump_pages_test() 104 unsigned long i, nr_pages, add in __gup_test_ioctl() local 10 put_back_pages(unsigned int cmd, struct page **pages, unsigned long nr_pages, unsigned int gup_test_flags) put_back_pages() argument 39 verify_dma_pinned(unsigned int cmd, struct page **pages, unsigned long nr_pages) verify_dma_pinned() argument 69 dump_pages_test(struct gup_test *gup, struct page **pages, unsigned long nr_pages) dump_pages_test() argument 224 long nr_pages, cur_pages, addr, remaining_pages; pin_longterm_test_start() local [all...] |
H A D | sparse.c | 185 unsigned long nr_pages) in subsection_mask_set() 188 int end = subsection_map_index(pfn + nr_pages - 1); in subsection_mask_set() 193 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) in subsection_map_init() argument 195 int end_sec = pfn_to_section_nr(pfn + nr_pages - 1); in subsection_map_init() 198 if (!nr_pages) in subsection_map_init() 205 pfns = min(nr_pages, PAGES_PER_SECTION in subsection_map_init() 215 nr_pages -= pfns; in subsection_map_init() 219 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) in subsection_map_init() argument 429 unsigned long nr_pages, int nid, struct vmem_altmap *altmap, in __populate_section_memmap() 632 unsigned long nr_pages, in in populate_section_memmap() 184 subsection_mask_set(unsigned long *map, unsigned long pfn, unsigned long nr_pages) subsection_mask_set() argument 428 __populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) __populate_section_memmap() argument 631 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) populate_section_memmap() argument 638 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) depopulate_section_memmap() argument 654 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) clear_subsection_map() argument 681 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) fill_subsection_map() argument 703 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) populate_section_memmap() argument 711 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) depopulate_section_memmap() argument 720 unsigned long magic, nr_pages; free_map_bootmem() local 747 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) clear_subsection_map() argument 757 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) fill_subsection_map() argument 779 section_deactivate(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) section_deactivate() argument 828 section_activate(int nid, unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) section_activate() argument 890 sparse_add_section(int nid, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) sparse_add_section() argument 925 sparse_remove_section(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) sparse_remove_section() argument [all...] |
H A D | memory_hotplug.c | 60 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); in memory_block_memmap_on_memory_pages() local 69 return pageblock_align(nr_pages); in memory_block_memmap_on_memory_pages() 70 return nr_pages; in memory_block_memmap_on_memory_pages() 300 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) in check_pfn_span() argument 317 if (!IS_ALIGNED(pfn | nr_pages, min_align)) in check_pfn_span() 370 int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, in __add_pages() argument 373 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() 381 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false)); in __add_pages() 388 || vmem_altmap_offset(altmap) > nr_pages) { in __add_pages() 395 if (check_pfn_span(pfn, nr_pages)) { in __add_pages() 528 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) remove_pfn_range_from_zone() argument 574 __remove_pages(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) __remove_pages() argument 645 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) online_pages_range() argument 682 node_states_check_changes_online(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) node_states_check_changes_online() argument 705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) resize_zone_range() argument 716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) resize_pgdat_range() argument 750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) move_pfn_range_to_zone() argument 843 auto_movable_can_online_movable(int nid, struct memory_group *group, unsigned long nr_pages) auto_movable_can_online_movable() argument 897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_kernel_zone_for_pfn() argument 963 auto_movable_zone_for_pfn(int nid, struct memory_group *group, unsigned long pfn, unsigned long nr_pages) auto_movable_zone_for_pfn() argument 1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_zone_for_pfn() argument 1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) zone_for_pfn_range() argument 1068 adjust_present_page_count(struct page *page, struct memory_group *group, long nr_pages) adjust_present_page_count() argument 1089 mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, struct zone *zone) mhp_init_memmap_on_memory() argument 1115 mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages) mhp_deinit_memmap_on_memory() argument 1138 online_pages(unsigned long pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) online_pages() argument 1190 adjust_present_page_count(pfn_to_page(pfn), group, nr_pages); online_pages() local 1816 node_states_check_changes_offline(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) node_states_check_changes_offline() argument 1863 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) count_system_ram_pages_cb() argument 1875 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) offline_pages() argument [all...] |
H A D | hugetlb_cgroup.c | 190 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local 205 nr_pages = compound_nr(page); in hugetlb_cgroup_move_parent() 209 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 213 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent() 255 static int __hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in __hugetlb_cgroup_charge_cgroup() argument 282 nr_pages, &counter)) { in __hugetlb_cgroup_charge_cgroup() 298 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument 301 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, false); in hugetlb_cgroup_charge_cgroup() 304 int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup_rsvd() argument 307 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, pt in hugetlb_cgroup_charge_cgroup_rsvd() 311 __hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio, bool rsvd) __hugetlb_cgroup_commit_charge() argument 332 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) hugetlb_cgroup_commit_charge() argument 339 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) hugetlb_cgroup_commit_charge_rsvd() argument 349 __hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio, bool rsvd) __hugetlb_cgroup_uncharge_folio() argument 364 nr_pages); __hugetlb_cgroup_uncharge_folio() local 381 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) hugetlb_cgroup_uncharge_folio() argument 387 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) hugetlb_cgroup_uncharge_folio_rsvd() argument 393 __hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, bool rsvd) __hugetlb_cgroup_uncharge_cgroup() argument 411 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup() argument 417 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup_rsvd() argument 435 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) hugetlb_cgroup_uncharge_file_region() argument 599 unsigned long nr_pages; hugetlb_cgroup_write() local [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | page_counter.c | 48 * @nr_pages: number of pages to cancel 50 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) in page_counter_cancel() argument 54 new = atomic_long_sub_return(nr_pages, &counter->usage); in page_counter_cancel() 63 * @nr_pages: number of pages to charge 67 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) in page_counter_charge() argument 74 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_charge() 88 * @nr_pages: number of pages to charge 95 unsigned long nr_pages, in page_counter_try_charge() 116 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_try_charge() 118 atomic_long_sub(nr_pages, in page_counter_try_charge() 94 page_counter_try_charge(struct page_counter *counter, unsigned long nr_pages, struct page_counter **fail) page_counter_try_charge() argument 151 page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) page_counter_uncharge() argument 169 page_counter_set_max(struct page_counter *counter, unsigned long nr_pages) page_counter_set_max() argument 208 page_counter_set_min(struct page_counter *counter, unsigned long nr_pages) page_counter_set_min() argument 225 page_counter_set_low(struct page_counter *counter, unsigned long nr_pages) page_counter_set_low() argument 244 page_counter_memparse(const char *buf, const char *max, unsigned long *nr_pages) page_counter_memparse() argument [all...] |
H A D | memory_hotplug.c | 234 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 238 nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT; in register_page_bootmem_info_node() 241 for (i = 0; i < nr_pages; i++, page++) in register_page_bootmem_info_node() 261 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages, in check_pfn_span() argument 280 || !IS_ALIGNED(nr_pages, min_align)) { in check_pfn_span() 282 reason, pfn, pfn + nr_pages - 1); in check_pfn_span() 289 unsigned long nr_pages) in check_hotplug_memory_addressable() 291 const u64 max_addr = PFN_PHYS(pfn + nr_pages) - 1; in check_hotplug_memory_addressable() 310 int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, in __add_pages() argument 313 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() 288 check_hotplug_memory_addressable(unsigned long pfn, unsigned long nr_pages) check_hotplug_memory_addressable() argument 471 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) remove_pfn_range_from_zone() argument 510 __remove_section(unsigned long pfn, unsigned long nr_pages, unsigned long map_offset, struct vmem_altmap *altmap) __remove_section() argument 533 __remove_pages(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) __remove_pages() argument 611 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) online_pages_range() argument 630 node_states_check_changes_online(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) node_states_check_changes_online() argument 661 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) resize_zone_range() argument 672 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) resize_pgdat_range() argument 692 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) move_pfn_range_to_zone() argument 729 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_kernel_zone_for_pfn() argument 745 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) default_zone_for_pfn() argument 769 zone_for_pfn_range(int online_type, int nid, unsigned long start_pfn, unsigned long nr_pages) zone_for_pfn_range() argument 781 online_pages(unsigned long pfn, unsigned long nr_pages, int online_type, int nid) online_pages() argument 1387 node_states_check_changes_offline(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) node_states_check_changes_offline() argument 1453 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) count_system_ram_pages_cb() argument 1462 offline_pages(unsigned long start_pfn, unsigned long nr_pages) offline_pages() argument [all...] |
H A D | sparse.c | 212 unsigned long nr_pages) in subsection_mask_set() 215 int end = subsection_map_index(pfn + nr_pages - 1); in subsection_mask_set() 220 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) in subsection_map_init() argument 222 int end_sec = pfn_to_section_nr(pfn + nr_pages - 1); in subsection_map_init() 225 if (!nr_pages) in subsection_map_init() 232 pfns = min(nr_pages, PAGES_PER_SECTION in subsection_map_init() 242 nr_pages -= pfns; in subsection_map_init() 246 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) in subsection_map_init() argument 446 unsigned long nr_pages, int nid, struct vmem_altmap *altmap) in __populate_section_memmap() 652 unsigned long nr_pages, in in populate_section_memmap() 211 subsection_mask_set(unsigned long *map, unsigned long pfn, unsigned long nr_pages) subsection_mask_set() argument 445 __populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap) __populate_section_memmap() argument 651 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap) populate_section_memmap() argument 657 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) depopulate_section_memmap() argument 673 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) clear_subsection_map() argument 700 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) fill_subsection_map() argument 722 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap) populate_section_memmap() argument 729 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) depopulate_section_memmap() argument 738 unsigned long magic, nr_pages; free_map_bootmem() local 765 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) clear_subsection_map() argument 775 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) fill_subsection_map() argument 797 section_deactivate(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) section_deactivate() argument 846 section_activate(int nid, unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) section_activate() argument 906 sparse_add_section(int nid, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap) sparse_add_section() argument 941 clear_hwpoisoned_pages(struct page *memmap, int nr_pages) clear_hwpoisoned_pages() argument 962 clear_hwpoisoned_pages(struct page *memmap, int nr_pages) clear_hwpoisoned_pages() argument 967 sparse_remove_section(struct mem_section *ms, unsigned long pfn, unsigned long nr_pages, unsigned long map_offset, struct vmem_altmap *altmap) sparse_remove_section() argument [all...] |
H A D | percpu-km.c | 50 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local 60 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 66 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk() 73 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk() 84 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local 93 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk() 104 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 112 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info() 113 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 115 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() [all...] |
H A D | hugetlb_cgroup.c | 164 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local 178 nr_pages = compound_nr(page); in hugetlb_cgroup_move_parent() 182 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 186 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent() 231 static int __hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in __hugetlb_cgroup_charge_cgroup() argument 258 nr_pages, &counter)) { in __hugetlb_cgroup_charge_cgroup() 274 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument 277 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, false); in hugetlb_cgroup_charge_cgroup() 280 int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup_rsvd() argument 283 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, pt in hugetlb_cgroup_charge_cgroup_rsvd() 287 __hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page, bool rsvd) __hugetlb_cgroup_commit_charge() argument 298 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge() argument 305 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge_rsvd() argument 315 __hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page, bool rsvd) __hugetlb_cgroup_uncharge_page() argument 330 nr_pages); __hugetlb_cgroup_uncharge_page() local 338 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page() argument 344 hugetlb_cgroup_uncharge_page_rsvd(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page_rsvd() argument 350 __hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, bool rsvd) __hugetlb_cgroup_uncharge_cgroup() argument 368 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup() argument 374 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup_rsvd() argument 392 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) hugetlb_cgroup_uncharge_file_region() argument 503 unsigned long nr_pages; hugetlb_cgroup_write() local [all...] |
H A D | gup_benchmark.c | 25 unsigned long nr_pages) in put_back_pages() 32 for (i = 0; i < nr_pages; i++) in put_back_pages() 39 unpin_user_pages(pages, nr_pages); in put_back_pages() 45 unsigned long nr_pages) in verify_dma_pinned() 54 for (i = 0; i < nr_pages; i++) { in verify_dma_pinned() 71 unsigned long i, nr_pages, addr, next; in __gup_benchmark_ioctl() local 81 nr_pages = gup->size / PAGE_SIZE; in __gup_benchmark_ioctl() 82 pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL); in __gup_benchmark_ioctl() 140 /* Shifting the meaning of nr_pages: now it is actual number pinned: */ in __gup_benchmark_ioctl() 141 nr_pages in __gup_benchmark_ioctl() 24 put_back_pages(unsigned int cmd, struct page **pages, unsigned long nr_pages) put_back_pages() argument 44 verify_dma_pinned(unsigned int cmd, struct page **pages, unsigned long nr_pages) verify_dma_pinned() argument [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | hugetlb_cgroup.h | 128 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, 130 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, 132 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, 135 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, 138 extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, 140 extern void hugetlb_cgroup_uncharge_page_rsvd(int idx, unsigned long nr_pages, 143 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, 145 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, 153 unsigned long nr_pages, 163 unsigned long nr_pages, in hugetlb_cgroup_uncharge_file_region() 161 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) hugetlb_cgroup_uncharge_file_region() argument 211 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) hugetlb_cgroup_charge_cgroup() argument 217 hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) hugetlb_cgroup_charge_cgroup_rsvd() argument 224 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge() argument 231 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge_rsvd() argument 237 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page() argument 242 hugetlb_cgroup_uncharge_page_rsvd(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page_rsvd() argument 247 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup() argument 254 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup_rsvd() argument [all...] |
H A D | page_counter.h | 51 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); 52 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); 54 unsigned long nr_pages, 56 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); 57 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); 58 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); 61 unsigned long nr_pages) in page_counter_set_high() 63 WRITE_ONCE(counter->high, nr_pages); in page_counter_set_high() 66 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); 68 unsigned long *nr_pages); 60 page_counter_set_high(struct page_counter *counter, unsigned long nr_pages) page_counter_set_high() argument [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | hugetlb_cgroup.h | 138 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, 140 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, 142 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, 145 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, 148 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, 150 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, 153 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, 155 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, 163 unsigned long nr_pages, 173 unsigned long nr_pages, in hugetlb_cgroup_uncharge_file_region() 171 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) hugetlb_cgroup_uncharge_file_region() argument 218 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) hugetlb_cgroup_charge_cgroup() argument 224 hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) hugetlb_cgroup_charge_cgroup_rsvd() argument 231 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) hugetlb_cgroup_commit_charge() argument 238 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) hugetlb_cgroup_commit_charge_rsvd() argument 244 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) hugetlb_cgroup_uncharge_folio() argument 249 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) hugetlb_cgroup_uncharge_folio_rsvd() argument 254 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup() argument 261 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) hugetlb_cgroup_uncharge_cgroup_rsvd() argument [all...] |
H A D | page_counter.h | 60 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); 61 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); 63 unsigned long nr_pages, 65 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); 66 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); 67 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); 70 unsigned long nr_pages) in page_counter_set_high() 72 WRITE_ONCE(counter->high, nr_pages); in page_counter_set_high() 75 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); 77 unsigned long *nr_pages); 69 page_counter_set_high(struct page_counter *counter, unsigned long nr_pages) page_counter_set_high() argument [all...] |
H A D | memory_hotplug.h | 154 long nr_pages); 156 extern int mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, 158 extern void mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages); 159 extern int online_pages(unsigned long pfn, unsigned long nr_pages, 188 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, 192 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 197 unsigned long nr_pages, struct mhp_params *params) in add_pages() 199 return __add_pages(nid, start_pfn, nr_pages, params); in add_pages() 202 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 308 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages, 196 add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, struct mhp_params *params) add_pages() argument 317 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) offline_pages() argument [all...] |
/kernel/linux/linux-5.10/fs/iomap/ |
H A D | swapfile.c | 19 unsigned long nr_pages; /* number of pages collected */ member 32 unsigned long nr_pages; in iomap_swapfile_add_extent() local 39 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent() 41 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent() 54 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent() 55 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent() 71 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent() 75 isi->nr_pages in iomap_swapfile_add_extent() [all...] |
/kernel/linux/linux-6.6/fs/iomap/ |
H A D | swapfile.c | 19 unsigned long nr_pages; /* number of pages collected */ member 33 unsigned long nr_pages; in iomap_swapfile_add_extent() local 40 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent() 42 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent() 55 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent() 56 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent() 72 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent() 76 isi->nr_pages in iomap_swapfile_add_extent() [all...] |
/kernel/linux/linux-5.10/drivers/media/v4l2-core/ |
H A D | videobuf-dma-sg.c | 63 int nr_pages) in videobuf_vmalloc_to_sg() 69 sglist = vzalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_vmalloc_to_sg() 72 sg_init_table(sglist, nr_pages); in videobuf_vmalloc_to_sg() 73 for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { in videobuf_vmalloc_to_sg() 93 int nr_pages, int offset, size_t size) in videobuf_pages_to_sg() 100 sglist = vmalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_pages_to_sg() 103 sg_init_table(sglist, nr_pages); in videobuf_pages_to_sg() 111 for (i = 1; i < nr_pages; i++) { in videobuf_pages_to_sg() 174 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked() 175 dma->pages = kmalloc_array(dma->nr_pages, sizeo in videobuf_dma_init_user_locked() 62 videobuf_vmalloc_to_sg(unsigned char *virt, int nr_pages) videobuf_vmalloc_to_sg() argument 92 videobuf_pages_to_sg(struct page **pages, int nr_pages, int offset, size_t size) videobuf_pages_to_sg() argument 210 videobuf_dma_init_kernel(struct videobuf_dmabuf *dma, int direction, unsigned long nr_pages) videobuf_dma_init_kernel() argument 269 videobuf_dma_init_overlay(struct videobuf_dmabuf *dma, int direction, dma_addr_t addr, unsigned long nr_pages) videobuf_dma_init_overlay() argument [all...] |
/kernel/linux/linux-6.6/drivers/media/v4l2-core/ |
H A D | videobuf-dma-sg.c | 63 int nr_pages) in videobuf_vmalloc_to_sg() 69 sglist = vzalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_vmalloc_to_sg() 72 sg_init_table(sglist, nr_pages); in videobuf_vmalloc_to_sg() 73 for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { in videobuf_vmalloc_to_sg() 93 int nr_pages, int offset, size_t size) in videobuf_pages_to_sg() 100 sglist = vmalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_pages_to_sg() 103 sg_init_table(sglist, nr_pages); in videobuf_pages_to_sg() 111 for (i = 1; i < nr_pages; i++) { in videobuf_pages_to_sg() 173 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked() 174 dma->pages = kmalloc_array(dma->nr_pages, sizeo in videobuf_dma_init_user_locked() 62 videobuf_vmalloc_to_sg(unsigned char *virt, int nr_pages) videobuf_vmalloc_to_sg() argument 92 videobuf_pages_to_sg(struct page **pages, int nr_pages, int offset, size_t size) videobuf_pages_to_sg() argument 206 videobuf_dma_init_kernel(struct videobuf_dmabuf *dma, int direction, unsigned long nr_pages) videobuf_dma_init_kernel() argument 265 videobuf_dma_init_overlay(struct videobuf_dmabuf *dma, int direction, dma_addr_t addr, unsigned long nr_pages) videobuf_dma_init_overlay() argument [all...] |
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | balloon.c | 402 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument 408 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation() 409 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation() 412 for (i = 0; i < nr_pages; i++) { in increase_reservation() 414 nr_pages = i; in increase_reservation() 422 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation() 441 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument 449 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation() 450 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation() 452 for (i = 0; i < nr_pages; in decrease_reservation() 583 add_ballooned_pages(int nr_pages) add_ballooned_pages() argument 616 alloc_xenballooned_pages(int nr_pages, struct page **pages) alloc_xenballooned_pages() argument 669 free_xenballooned_pages(int nr_pages, struct page **pages) free_xenballooned_pages() argument [all...] |
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | balloon.c | 386 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument 392 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation() 393 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation() 396 for (i = 0; i < nr_pages; i++) { in increase_reservation() 398 nr_pages = i; in increase_reservation() 406 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation() 425 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument 433 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation() 434 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation() 436 for (i = 0; i < nr_pages; in decrease_reservation() 567 add_ballooned_pages(unsigned int nr_pages) add_ballooned_pages() argument 600 xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages) xen_alloc_ballooned_pages() argument 653 xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages) xen_free_ballooned_pages() argument [all...] |
H A D | unpopulated-alloc.c | 34 static int fill_list(unsigned int nr_pages) in fill_list() argument 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() 154 * @nr_pages: Number of pages 158 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_unpopulated_pages() argument 169 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_alloc_unpopulated_pages() 172 if (list_count < nr_pages) { in xen_alloc_unpopulated_pages() 173 ret = fill_list(nr_pages - list_count); in xen_alloc_unpopulated_pages() 178 for (i = 0; i < nr_pages; i++) { in xen_alloc_unpopulated_pages() 211 * @nr_pages: Number of pages 214 void xen_free_unpopulated_pages(unsigned int nr_pages, struc argument [all...] |
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/ |
H A D | memtrace.c | 72 unsigned long nr_pages) in memtrace_clear_range() 81 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) { in memtrace_clear_range() 89 static bool memtrace_offline_pages(u32 nid, u64 start_pfn, u64 nr_pages) in memtrace_offline_pages() argument 92 const unsigned long size = PFN_PHYS(nr_pages); in memtrace_offline_pages() 100 if (offline_pages(start_pfn, nr_pages)) { in memtrace_offline_pages() 115 u64 start_pfn, end_pfn, nr_pages, pfn; in memtrace_alloc_node() local 124 nr_pages = size >> PAGE_SHIFT; in memtrace_alloc_node() 127 end_pfn = round_down(end_pfn - nr_pages, nr_pages); in memtrace_alloc_node() 130 for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { in memtrace_alloc_node() 71 memtrace_clear_range(unsigned long start_pfn, unsigned long nr_pages) memtrace_clear_range() argument [all...] |