/kernel/linux/linux-5.10/include/linux/ |
H A D | gfp.h | 516 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, 520 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument 522 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages()
|
H A D | hugetlb.h | 511 struct page *alloc_huge_page_nodemask(struct hstate *h, int preferred_nid, 784 alloc_huge_page_nodemask(struct hstate *h, int preferred_nid, in alloc_huge_page_nodemask() argument
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | gfp.h | 183 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, 185 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, 188 unsigned long __alloc_pages_bulk(gfp_t gfp, int preferred_nid,
|
H A D | hugetlb.h | 749 struct folio *alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid, 1057 alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid,
|
/kernel/linux/linux-6.6/mm/ |
H A D | page_alloc.c | 4210 int preferred_nid, nodemask_t *nodemask, in prepare_alloc_pages() 4215 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() 4260 * @preferred_nid: The preferred NUMA node ID to allocate from 4277 unsigned long __alloc_pages_bulk(gfp_t gfp, int preferred_nid, in __alloc_pages_bulk() argument 4331 if (!prepare_alloc_pages(gfp, 0, preferred_nid, nodemask, &ac, &alloc_gfp, &alloc_flags)) in __alloc_pages_bulk() 4413 page = __alloc_pages(gfp, 0, preferred_nid, nodemask); in __alloc_pages_bulk() 4429 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, in __alloc_pages() argument 4454 if (!prepare_alloc_pages(gfp, order, preferred_nid, nodemask, &ac, in __alloc_pages() 4494 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, in __folio_alloc() argument 4498 preferred_nid, nodemas in __folio_alloc() 4209 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_gfp, unsigned int *alloc_flags) prepare_alloc_pages() argument [all...] |
H A D | mempolicy.c | 2190 int preferred_nid; in vma_alloc_folio() local 2263 preferred_nid = policy_node(gfp, pol, node); in vma_alloc_folio() 2264 folio = __folio_alloc(gfp, order, preferred_nid, nmask); in vma_alloc_folio()
|
H A D | hugetlb.c | 2502 struct folio *alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid, in alloc_hugetlb_folio_nodemask() argument 2510 preferred_nid, nmask); in alloc_hugetlb_folio_nodemask() 2518 return alloc_migrate_hugetlb_folio(h, gfp_mask, preferred_nid, nmask); in alloc_hugetlb_folio_nodemask()
|
/kernel/linux/linux-5.10/mm/ |
H A D | mempolicy.c | 2174 int preferred_nid; in alloc_pages_vma() local 2229 preferred_nid = policy_node(gfp, pol, node); in alloc_pages_vma() 2230 page = __alloc_pages_nodemask(gfp, order, preferred_nid, nmask); in alloc_pages_vma()
|
H A D | page_alloc.c | 4960 int preferred_nid, nodemask_t *nodemask, in prepare_alloc_pages() 4965 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() 5014 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, in __alloc_pages_nodemask() argument 5033 if (!prepare_alloc_pages(gfp_mask, order, preferred_nid, nodemask, &ac, &alloc_mask, &alloc_flags)) in __alloc_pages_nodemask() 4959 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_mask, unsigned int *alloc_flags) prepare_alloc_pages() argument
|
H A D | hugetlb.c | 1985 struct page *alloc_huge_page_nodemask(struct hstate *h, int preferred_nid, in alloc_huge_page_nodemask() argument 1992 page = dequeue_huge_page_nodemask(h, gfp_mask, preferred_nid, nmask); in alloc_huge_page_nodemask() 2000 return alloc_migrate_huge_page(h, gfp_mask, preferred_nid, nmask); in alloc_huge_page_nodemask()
|