/kernel/linux/linux-6.6/mm/ |
H A D | fail_page_alloc.c | 10 u32 min_order; member 15 .min_order = 1, 28 if (order < fail_page_alloc.min_order) in __should_fail_alloc_page() 59 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order); in fail_page_alloc_debugfs()
|
H A D | slub.c | 4116 unsigned int min_order = slub_min_order; in calc_slab_order() local 4119 if (order_objects(min_order, size) > MAX_OBJS_PER_PAGE) in calc_slab_order() 4122 for (order = max(min_order, (unsigned int)get_order(min_objects * size)); in calc_slab_order()
|
H A D | page_alloc.c | 2016 int min_order = order; in __rmqueue_fallback() local 2027 min_order = pageblock_order; in __rmqueue_fallback() 2034 for (current_order = MAX_ORDER; current_order >= min_order; in __rmqueue_fallback()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | intel_memory_region.c | 74 unsigned int min_order = 0; in __intel_memory_region_get_pages_buddy() local 81 min_order = ilog2(mem->min_page_size) - in __intel_memory_region_get_pages_buddy() 87 min_order = ilog2(size) - ilog2(mem->mm.chunk_size); in __intel_memory_region_get_pages_buddy() 103 GEM_BUG_ON(order < min_order); in __intel_memory_region_get_pages_buddy() 110 if (order-- == min_order) in __intel_memory_region_get_pages_buddy()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_buddy.c | 678 unsigned int min_order, order; in drm_buddy_alloc_blocks() local 710 min_order = ilog2(min_page_size) - ilog2(mm->chunk_size); in drm_buddy_alloc_blocks() 715 BUG_ON(order < min_order); in drm_buddy_alloc_blocks() 728 if (order-- == min_order) { in drm_buddy_alloc_blocks()
|
/kernel/linux/linux-5.10/drivers/video/fbdev/vermilion/ |
H A D | vermilion.c | 74 unsigned min_order) in vmlfb_alloc_vram_area() 91 } while (va->logical == 0 && max_order > min_order); in vmlfb_alloc_vram_area() 73 vmlfb_alloc_vram_area(struct vram_area *va, unsigned max_order, unsigned min_order) vmlfb_alloc_vram_area() argument
|
/kernel/linux/linux-6.6/drivers/video/fbdev/vermilion/ |
H A D | vermilion.c | 75 unsigned min_order) in vmlfb_alloc_vram_area() 92 } while (va->logical == 0 && max_order > min_order); in vmlfb_alloc_vram_area() 74 vmlfb_alloc_vram_area(struct vram_area *va, unsigned max_order, unsigned min_order) vmlfb_alloc_vram_area() argument
|
/kernel/linux/linux-5.10/mm/ |
H A D | page_alloc.c | 2772 int min_order = order; in __rmqueue_fallback() local 2783 min_order = pageblock_order; in __rmqueue_fallback() 2790 for (current_order = MAX_ORDER - 1; current_order >= min_order; in __rmqueue_fallback() 3534 u32 min_order; member 3539 .min_order = 1, 3550 if (order < fail_page_alloc.min_order) in __should_fail_alloc_page() 3577 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order); in fail_page_alloc_debugfs()
|
H A D | slub.c | 3409 unsigned int min_order = slub_min_order; in slab_order() local 3412 if (order_objects(min_order, size) > MAX_OBJS_PER_PAGE) in slab_order() 3415 for (order = max(min_order, (unsigned int)get_order(min_objects * size)); in slab_order()
|
/kernel/linux/linux-6.6/fs/ext4/ |
H A D | mballoc.c | 996 int i, order, min_order; in ext4_mb_choose_next_group_best_avail() local 1011 min_order = order - sbi->s_mb_best_avail_max_trim_order; in ext4_mb_choose_next_group_best_avail() 1012 if (min_order < 0) in ext4_mb_choose_next_group_best_avail() 1013 min_order = 0; in ext4_mb_choose_next_group_best_avail() 1021 if (1 << min_order < num_stripe_clusters) in ext4_mb_choose_next_group_best_avail() 1026 min_order = fls(num_stripe_clusters) - 1; in ext4_mb_choose_next_group_best_avail() 1029 if (1 << min_order < ac->ac_o_ex.fe_len) in ext4_mb_choose_next_group_best_avail() 1030 min_order = fls(ac->ac_o_ex.fe_len); in ext4_mb_choose_next_group_best_avail() 1032 for (i = order; i >= min_order; i--) { in ext4_mb_choose_next_group_best_avail()
|
/kernel/linux/linux-5.10/drivers/md/bcache/ |
H A D | btree.c | 587 static int mca_reap(struct btree *b, unsigned int min_order, bool flush) in mca_reap() argument 599 if (b->keys.page_order < min_order) in mca_reap()
|
/kernel/linux/linux-6.6/drivers/md/bcache/ |
H A D | btree.c | 608 static int mca_reap(struct btree *b, unsigned int min_order, bool flush) in mca_reap() argument 620 if (b->keys.page_order < min_order) in mca_reap()
|