/kernel/linux/linux-5.10/include/trace/events/ |
H A D | kmem.h | 281 int alloc_order, int fallback_order, 285 alloc_order, fallback_order, 290 __field( int, alloc_order ) 299 __entry->alloc_order = alloc_order; 307 TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d", 310 __entry->alloc_order,
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | kmem.h | 269 int alloc_order, int fallback_order, 273 alloc_order, fallback_order, 278 __field( int, alloc_order ) 287 __entry->alloc_order = alloc_order; 295 TP_printk("page=%p pfn=0x%lx alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d", 298 __entry->alloc_order,
|
/kernel/linux/linux-5.10/arch/powerpc/sysdev/xive/ |
H A D | spapr.c | 534 unsigned int alloc_order; in xive_spapr_cleanup_queue() local 543 alloc_order = xive_alloc_order(xive_queue_shift); in xive_spapr_cleanup_queue() 545 uv_unshare_page(PHYS_PFN(__pa(q->qpage)), 1 << alloc_order); in xive_spapr_cleanup_queue() 546 free_pages((unsigned long)q->qpage, alloc_order); in xive_spapr_cleanup_queue()
|
H A D | native.c | 233 unsigned int alloc_order; in xive_native_cleanup_queue() local 240 alloc_order = xive_alloc_order(xive_queue_shift); in xive_native_cleanup_queue() 241 free_pages((unsigned long)q->qpage, alloc_order); in xive_native_cleanup_queue()
|
H A D | common.c | 1554 unsigned int alloc_order; in xive_queue_page_alloc() local 1558 alloc_order = xive_alloc_order(queue_shift); in xive_queue_page_alloc() 1559 pages = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, alloc_order); in xive_queue_page_alloc()
|
/kernel/linux/linux-6.6/arch/powerpc/sysdev/xive/ |
H A D | spapr.c | 548 unsigned int alloc_order; in xive_spapr_cleanup_queue() local 557 alloc_order = xive_alloc_order(xive_queue_shift); in xive_spapr_cleanup_queue() 559 uv_unshare_page(PHYS_PFN(__pa(q->qpage)), 1 << alloc_order); in xive_spapr_cleanup_queue() 560 free_pages((unsigned long)q->qpage, alloc_order); in xive_spapr_cleanup_queue()
|
H A D | native.c | 232 unsigned int alloc_order; in xive_native_cleanup_queue() local 239 alloc_order = xive_alloc_order(xive_queue_shift); in xive_native_cleanup_queue() 240 free_pages((unsigned long)q->qpage, alloc_order); in xive_native_cleanup_queue()
|
H A D | common.c | 1694 unsigned int alloc_order; in xive_queue_page_alloc() local 1698 alloc_order = xive_alloc_order(queue_shift); in xive_queue_page_alloc() 1699 pages = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, alloc_order); in xive_queue_page_alloc()
|
/kernel/linux/linux-5.10/drivers/lightnvm/ |
H A D | pblk-rb.c | 80 unsigned int alloc_order, order, iter; in pblk_rb_init() local 104 alloc_order = power_size; in pblk_rb_init() 105 if (alloc_order >= max_order) { in pblk_rb_init() 107 iter = (1 << (alloc_order - max_order)); in pblk_rb_init() 109 order = alloc_order; in pblk_rb_init()
|
/kernel/linux/linux-5.10/drivers/dma/ioat/ |
H A D | dma.h | 120 * @alloc_order: log2 of the number of allocated descriptors 130 u16 alloc_order; member 310 return 1 << ioat_chan->alloc_order; in ioat_ring_size()
|
H A D | init.c | 612 const int total_descs = 1 << ioat_chan->alloc_order; in ioat_free_chan_resources() 664 ioat_chan->alloc_order = 0; in ioat_free_chan_resources() 689 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources() 718 ioat_chan->alloc_order = order; in ioat_alloc_chan_resources() 754 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources()
|
H A D | sysfs.c | 115 return sprintf(page, "%d\n", (1 << ioat_chan->alloc_order) & ~1); in ring_size_show()
|
/kernel/linux/linux-5.10/Documentation/trace/postprocess/ |
H A D | trace-pagealloc-postprocess.pl | 80 my $regex_fragdetails_default = 'page=([0-9a-f]*) pfn=([0-9]*) alloc_order=([-0-9]*) fallback_order=([-0-9]*) pageblock_order=([-0-9]*) alloc_migratetype=([-0-9]*) fallback_migratetype=([-0-9]*) fragmenting=([-0-9]) change_ownership=([-0-9])'; 132 "alloc_order", "fallback_order", "pageblock_order", 242 my ($alloc_order, $fallback_order, $pageblock_order); 254 $alloc_order = $3;
|
/kernel/linux/linux-6.6/Documentation/trace/postprocess/ |
H A D | trace-pagealloc-postprocess.pl | 80 my $regex_fragdetails_default = 'page=([0-9a-f]*) pfn=([0-9]*) alloc_order=([-0-9]*) fallback_order=([-0-9]*) pageblock_order=([-0-9]*) alloc_migratetype=([-0-9]*) fallback_migratetype=([-0-9]*) fragmenting=([-0-9]) change_ownership=([-0-9])'; 132 "alloc_order", "fallback_order", "pageblock_order", 242 my ($alloc_order, $fallback_order, $pageblock_order); 254 $alloc_order = $3;
|
/kernel/linux/linux-6.6/drivers/dma/ioat/ |
H A D | dma.h | 121 * @alloc_order: log2 of the number of allocated descriptors 131 u16 alloc_order; member 309 return 1 << ioat_chan->alloc_order; in ioat_ring_size()
|
H A D | init.c | 612 const int total_descs = 1 << ioat_chan->alloc_order; in ioat_free_chan_resources() 664 ioat_chan->alloc_order = 0; in ioat_free_chan_resources() 689 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources() 718 ioat_chan->alloc_order = order; in ioat_alloc_chan_resources() 754 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources()
|
H A D | sysfs.c | 115 return sprintf(page, "%d\n", (1 << ioat_chan->alloc_order) & ~1); in ring_size_show()
|
/kernel/linux/linux-5.10/mm/ |
H A D | vmscan.c | 3783 static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_order, in kswapd_try_to_sleep() argument 3814 wakeup_kcompactd(pgdat, alloc_order, highest_zoneidx); in kswapd_try_to_sleep() 3882 unsigned int alloc_order, reclaim_order; in kswapd() local 3914 alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order); in kswapd() 3919 kswapd_try_to_sleep(pgdat, alloc_order, reclaim_order, in kswapd() 3923 alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order); in kswapd() 3946 * request (alloc_order). in kswapd() 3949 alloc_order); in kswapd() 3956 reclaim_order = balance_pgdat(pgdat, alloc_order, in kswapd() 3961 if (reclaim_order < alloc_order) in kswapd() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 286 struct xgbe_page_alloc *pa, int alloc_order, in xgbe_alloc_pages() 295 order = alloc_order; in xgbe_alloc_pages() 285 xgbe_alloc_pages(struct xgbe_prv_data *pdata, struct xgbe_page_alloc *pa, int alloc_order, int node) xgbe_alloc_pages() argument
|
/kernel/linux/linux-6.6/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 286 struct xgbe_page_alloc *pa, int alloc_order, in xgbe_alloc_pages() 295 order = alloc_order; in xgbe_alloc_pages() 285 xgbe_alloc_pages(struct xgbe_prv_data *pdata, struct xgbe_page_alloc *pa, int alloc_order, int node) xgbe_alloc_pages() argument
|
/kernel/linux/linux-5.10/drivers/net/ethernet/brocade/bna/ |
H A D | bnad.h | 251 int alloc_order; member
|
H A D | bnad.c | 252 unmap_q->alloc_order = -1; in bnad_rxq_alloc_uninit() 271 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 275 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 279 unmap_q->alloc_order = order; in bnad_rxq_alloc_init() 352 alloc_size = PAGE_SIZE << unmap_q->alloc_order; in bnad_rxq_refill_page() 360 unmap_q->alloc_order); in bnad_rxq_refill_page()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/brocade/bna/ |
H A D | bnad.h | 251 int alloc_order; member
|
H A D | bnad.c | 252 unmap_q->alloc_order = -1; in bnad_rxq_alloc_uninit() 271 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 275 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 279 unmap_q->alloc_order = order; in bnad_rxq_alloc_init() 352 alloc_size = PAGE_SIZE << unmap_q->alloc_order; in bnad_rxq_refill_page() 360 unmap_q->alloc_order); in bnad_rxq_refill_page()
|
/kernel/linux/linux-6.6/mm/ |
H A D | vmscan.c | 7646 static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_order, in kswapd_try_to_sleep() argument 7677 wakeup_kcompactd(pgdat, alloc_order, highest_zoneidx); in kswapd_try_to_sleep() 7745 unsigned int alloc_order, reclaim_order; in kswapd() local 7775 alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order); in kswapd() 7780 kswapd_try_to_sleep(pgdat, alloc_order, reclaim_order, in kswapd() 7784 alloc_order = READ_ONCE(pgdat->kswapd_order); in kswapd() 7807 * request (alloc_order). in kswapd() 7810 alloc_order); in kswapd() 7814 reclaim_order = balance_pgdat(pgdat, alloc_order, in kswapd() 7816 if (reclaim_order < alloc_order) in kswapd() [all...] |