Home
last modified time | relevance | path

Searched refs:order (Results 1 - 25 of 965) sorted by relevance

12345678910>>...39

/kernel/linux/linux-6.6/scripts/atomic/
H A Dgen-atomic-fallback.sh8 #gen_template_fallback(template, meta, pfx, name, sfx, order, atomic, int, args...)
16 local order="$1"; shift
28 #gen_order_fallback(meta, pfx, name, sfx, order, atomic, int, args...)
35 local order="$1"; shift
37 local tmpl_order=${order#_}
39 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
42 #gen_proto_fallback(meta, pfx, name, sfx, order, atomic, int, args...)
49 local order="$1"; shift
51 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
52 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "
[all...]
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/nvhe/
H A Dpage_alloc.c28 * __find_buddy_nocheck(pool, page 0, order 0) => page 1
29 * __find_buddy_nocheck(pool, page 0, order 1) => page 2
30 * __find_buddy_nocheck(pool, page 1, order 0) => page 0
31 * __find_buddy_nocheck(pool, page 2, order 0) => page 3
35 unsigned short order) in __find_buddy_nocheck()
39 addr ^= (PAGE_SIZE << order); in __find_buddy_nocheck()
54 unsigned short order) in __find_buddy_avail()
56 struct hyp_page *buddy = __find_buddy_nocheck(pool, p, order); in __find_buddy_avail()
58 if (!buddy || buddy->order != order || budd in __find_buddy_avail()
33 __find_buddy_nocheck(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __find_buddy_nocheck() argument
52 __find_buddy_avail(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __find_buddy_avail() argument
97 unsigned short order = p->order; __hyp_attach_page() local
130 __hyp_extract_page(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __hyp_extract_page() argument
186 unsigned short order = p->order; hyp_split_page() local
198 hyp_alloc_pages(struct hyp_pool *pool, unsigned short order) hyp_alloc_pages() argument
[all...]
/kernel/linux/linux-5.10/include/trace/events/
H A Dcompaction.h173 int order,
177 TP_ARGS(order, gfp_mask, prio),
180 __field(int, order)
186 __entry->order = order;
191 TP_printk("order=%d gfp_mask=%s priority=%d",
192 __entry->order,
200 int order,
203 TP_ARGS(zone, order, ret),
208 __field(int, order)
[all...]
H A Dkmem.h154 TP_PROTO(struct page *page, unsigned int order),
156 TP_ARGS(page, order),
160 __field( unsigned int, order )
165 __entry->order = order;
168 TP_printk("page=%p pfn=%lu order=%d",
171 __entry->order)
188 TP_printk("page=%p pfn=%lu order=0",
195 TP_PROTO(struct page *page, unsigned int order,
198 TP_ARGS(page, order, gfp_flag
[all...]
/kernel/linux/linux-6.6/include/trace/events/
H A Dcompaction.h168 int order,
172 TP_ARGS(order, gfp_mask, prio),
175 __field(int, order)
181 __entry->order = order;
186 TP_printk("order=%d gfp_mask=%s priority=%d",
187 __entry->order,
195 int order,
198 TP_ARGS(zone, order, ret),
203 __field(int, order)
[all...]
H A Dkmem.h138 TP_PROTO(struct page *page, unsigned int order),
140 TP_ARGS(page, order),
144 __field( unsigned int, order )
149 __entry->order = order;
152 TP_printk("page=%p pfn=0x%lx order=%d",
155 __entry->order)
172 TP_printk("page=%p pfn=0x%lx order=0",
179 TP_PROTO(struct page *page, unsigned int order,
182 TP_ARGS(page, order, gfp_flag
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/
H A Di915_buddy.c15 pr_err("block info: header=%llx, state=%u, order=%d, offset=%llx size=%llx root=%s buddy=%s\n", in __igt_dump_block()
202 unsigned int order; in igt_check_mm() local
218 order = i915_buddy_block_order(root); in igt_check_mm()
221 if (order != mm->max_order) { in igt_check_mm()
222 pr_err("max order root missing\n"); in igt_check_mm()
242 block = list_first_entry_or_null(&mm->free_list[order], in igt_check_mm()
246 pr_err("root mismatch at order=%u\n", order); in igt_check_mm()
307 int *order; in igt_buddy_alloc_smoke() local
320 order in igt_buddy_alloc_smoke()
329 int order; igt_buddy_alloc_smoke() local
412 unsigned int order; igt_buddy_alloc_pessimistic() local
505 int order; igt_buddy_alloc_optimistic() local
556 int order, top; igt_buddy_alloc_pathological() local
[all...]
/kernel/linux/linux-6.6/mm/
H A Dpage_alloc.c233 static void __free_pages_ok(struct page *page, unsigned int order,
310 static bool page_contains_unaccepted(struct page *page, unsigned int order);
311 static void accept_page(struct page *page, unsigned int order);
312 static bool try_to_accept_memory(struct zone *zone, unsigned int order);
338 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
340 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
527 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument
530 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex()
531 VM_BUG_ON(order != pageblock_order); in order_to_pindex()
535 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDE in order_to_pindex()
543 int order = pindex / MIGRATE_PCPTYPES; pindex_to_order() local
555 pcp_allowed_order(unsigned int order) pcp_allowed_order() argument
566 free_the_page(struct page *page, unsigned int order) free_the_page() argument
586 prep_compound_page(struct page *page, unsigned int order) prep_compound_page() argument
612 set_buddy_order(struct page *page, unsigned int order) set_buddy_order() argument
630 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument
661 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument
669 add_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list() argument
679 add_to_free_list_tail(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list_tail() argument
693 move_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) move_to_free_list() argument
701 del_page_from_free_list(struct page *page, struct zone *zone, unsigned int order) del_page_from_free_list() argument
730 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) buddy_merge_likely() argument
770 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) __free_one_page() argument
864 split_free_page(struct page *free_page, unsigned int order, unsigned long split_pfn_offset) split_free_page() argument
1082 free_pages_prepare(struct page *page, unsigned int order, fpi_t fpi_flags) free_pages_prepare() argument
1148 PAGE_SIZE << order); free_pages_prepare() local
1150 PAGE_SIZE << order); free_pages_prepare() local
1195 unsigned int order; free_pcppages_bulk() local
1249 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, int migratetype, fpi_t fpi_flags) free_one_page() argument
1265 __free_pages_ok(struct page *page, unsigned int order, fpi_t fpi_flags) __free_pages_ok() argument
1294 __free_pages_core(struct page *page, unsigned int order) __free_pages_core() argument
1445 check_new_pages(struct page *page, unsigned int order) check_new_pages() argument
1487 post_alloc_hook(struct page *page, unsigned int order, gfp_t gfp_flags) post_alloc_hook() argument
1547 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) prep_new_page() argument
1572 __rmqueue_smallest(struct zone *zone, unsigned int order, int migratetype) __rmqueue_smallest() argument
1611 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument
1617 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument
1632 unsigned int order; move_freepages() local
1708 can_steal_fallback(unsigned int order, int start_mt) can_steal_fallback() argument
1855 find_suitable_fallback(struct free_area *area, unsigned int order, int migratetype, bool only_stealable, bool *can_steal) find_suitable_fallback() argument
1936 int order; unreserve_highatomic_pageblock() local
2011 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) __rmqueue_fallback() argument
2089 __rmqueue_with_cma_reuse(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue_with_cma_reuse() argument
2114 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue() argument
2157 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) rmqueue_bulk() argument
2341 free_unref_page_prepare(struct page *page, unsigned long pfn, unsigned int order) free_unref_page_prepare() argument
2401 free_unref_page_commit(struct zone *zone, struct per_cpu_pages *pcp, struct page *page, int migratetype, unsigned int order) free_unref_page_commit() argument
2431 free_unref_page(struct page *page, unsigned int order) free_unref_page() argument
2452 free_one_page(page_zone(page), page, pfn, order, migratetype, FPI_NONE); free_unref_page() local
2563 split_page(struct page *page, unsigned int order) split_page() argument
2577 __isolate_free_page(struct page *page, unsigned int order) __isolate_free_page() argument
2629 __putback_isolated_page(struct page *page, unsigned int order, int mt) __putback_isolated_page() argument
2668 rmqueue_buddy(struct zone *preferred_zone, struct zone *zone, unsigned int order, unsigned int alloc_flags, int migratetype) rmqueue_buddy() argument
2710 __rmqueue_pcplist(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags, struct per_cpu_pages *pcp, struct list_head *list) __rmqueue_pcplist() argument
2750 rmqueue_pcplist(struct zone *preferred_zone, struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) rmqueue_pcplist() argument
2797 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) rmqueue() argument
2832 should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) should_fail_alloc_page() argument
2838 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) __zone_watermark_unusable_free() argument
2869 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) __zone_watermark_ok() argument
2947 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) zone_watermark_ok() argument
2954 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) zone_watermark_fast() argument
2999 zone_watermark_ok_safe(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx) zone_watermark_ok_safe() argument
3082 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) get_page_from_freelist() argument
3287 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) __alloc_pages_cpuset_fallback() argument
3307 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_may_oom() argument
3402 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument
3461 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument
3527 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument
3536 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument
3648 __perform_reclaim(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac) __perform_reclaim() argument
3674 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_direct_reclaim() argument
3707 wake_all_kswapds(unsigned int order, gfp_t gfp_mask, const struct alloc_context *ac) wake_all_kswapds() argument
3727 gfp_to_alloc_flags(gfp_t gfp_mask, unsigned int order) gfp_to_alloc_flags() argument
3828 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) should_reclaim_retry() argument
3933 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct alloc_context *ac) __alloc_pages_slowpath() argument
4209 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_gfp, unsigned int *alloc_flags) prepare_alloc_pages() argument
4429 __alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, nodemask_t *nodemask) __alloc_pages() argument
4494 __folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, nodemask_t *nodemask) __folio_alloc() argument
4512 __get_free_pages(gfp_t gfp_mask, unsigned int order) __get_free_pages() argument
4549 __free_pages(struct page *page, unsigned int order) __free_pages() argument
4562 free_pages(unsigned long addr, unsigned int order) free_pages() argument
4566 __free_pages(virt_to_page((void *)addr), order); free_pages() local
4698 make_alloc_exact(unsigned long addr, unsigned int order, size_t size) make_alloc_exact() argument
4735 unsigned int order = get_order(size); alloc_pages_exact() local
4760 unsigned int order = get_order(size); alloc_pages_exact_nid() local
6152 int order; alloc_contig_range() local
6437 unsigned int order; __offline_isolated_pages() local
6480 unsigned int order; is_free_buddy_page() local
6537 unsigned int order; take_page_off_buddy() local
6627 page_contains_unaccepted(struct page *page, unsigned int order) page_contains_unaccepted() argument
6635 accept_page(struct page *page, unsigned int order) accept_page() argument
6676 try_to_accept_memory(struct zone *zone, unsigned int order) try_to_accept_memory() argument
6726 page_contains_unaccepted(struct page *page, unsigned int order) page_contains_unaccepted() argument
6731 accept_page(struct page *page, unsigned int order) accept_page() argument
6735 try_to_accept_memory(struct zone *zone, unsigned int order) try_to_accept_memory() argument
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/
H A Dttm_pool.c54 * @vaddr: original vaddr return for the mapping and order in the lower bits
78 /* Allocate pages of size 1 << order with the given gfp_flags */
80 unsigned int order) in ttm_pool_alloc_page()
87 /* Don't set the __GFP_COMP flag for higher order allocations. in ttm_pool_alloc_page()
91 if (order) in ttm_pool_alloc_page()
96 p = alloc_pages_node(pool->nid, gfp_flags, order); in ttm_pool_alloc_page()
98 p->private = order; in ttm_pool_alloc_page()
106 if (order) in ttm_pool_alloc_page()
109 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page()
122 dma->vaddr = (unsigned long)vaddr | order; in ttm_pool_alloc_page()
79 ttm_pool_alloc_page(struct ttm_pool *pool, gfp_t gfp_flags, unsigned int order) ttm_pool_alloc_page() argument
132 ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order, struct page *p) ttm_pool_free_page() argument
185 ttm_pool_map(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr) ttm_pool_map() argument
258 ttm_pool_type_init(struct ttm_pool_type *pt, struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_type_init() argument
286 ttm_pool_select_type(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_select_type() argument
349 ttm_pool_page_allocated(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr, unsigned long *num_pages, struct page ***pages) ttm_pool_page_allocated() argument
388 unsigned int order; ttm_pool_free_range() local
429 unsigned int order; ttm_pool_alloc() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/lib/
H A Ddrm_random.c15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument
23 swap(order[i], order[j]); in drm_random_reorder()
30 unsigned int *order, i; in drm_random_order() local
32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order()
33 if (!order) in drm_random_order()
34 return order; in drm_random_order()
37 order[i] = i; in drm_random_order()
39 drm_random_reorder(order, coun in drm_random_order()
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/lib/
H A Ddrm_random.c16 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument
24 swap(order[i], order[j]); in drm_random_reorder()
31 unsigned int *order, i; in drm_random_order() local
33 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order()
34 if (!order) in drm_random_order()
35 return order; in drm_random_order()
38 order[i] = i; in drm_random_order()
40 drm_random_reorder(order, coun in drm_random_order()
[all...]
/kernel/linux/linux-6.6/mm/kmsan/
H A Dinit.c105 * by their order: when kmsan_memblock_free_pages() is called for the first
106 * time with a certain order, it is reserved as a shadow block, for the second
109 * after which held_back[order] can be used again.
114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument
118 if (!held_back[order].shadow) { in kmsan_memblock_free_pages()
119 held_back[order].shadow = page; in kmsan_memblock_free_pages()
122 if (!held_back[order].origin) { in kmsan_memblock_free_pages()
123 held_back[order].origin = page; in kmsan_memblock_free_pages()
126 shadow = held_back[order].shadow; in kmsan_memblock_free_pages()
127 origin = held_back[order] in kmsan_memblock_free_pages()
139 int order; global() member
[all...]
/kernel/linux/linux-6.6/arch/riscv/kvm/
H A Dtlb.c22 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa()
26 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa()
33 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
38 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
50 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa()
54 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa()
61 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
66 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
81 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva()
85 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva()
20 kvm_riscv_local_hfence_gvma_vmid_gpa(unsigned long vmid, gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_local_hfence_gvma_vmid_gpa() argument
49 kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_local_hfence_gvma_gpa() argument
77 kvm_riscv_local_hfence_vvma_asid_gva(unsigned long vmid, unsigned long asid, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_local_hfence_vvma_asid_gva() argument
119 kvm_riscv_local_hfence_vvma_gva(unsigned long vmid, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_local_hfence_vvma_gva() argument
332 kvm_riscv_hfence_gvma_vmid_gpa(struct kvm *kvm, unsigned long hbase, unsigned long hmask, gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_hfence_gvma_vmid_gpa() argument
355 kvm_riscv_hfence_vvma_asid_gva(struct kvm *kvm, unsigned long hbase, unsigned long hmask, unsigned long gva, unsigned long gvsz, unsigned long order, unsigned long asid) kvm_riscv_hfence_vvma_asid_gva() argument
384 kvm_riscv_hfence_vvma_gva(struct kvm *kvm, unsigned long hbase, unsigned long hmask, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_hfence_vvma_gva() argument
[all...]
/kernel/linux/linux-5.10/mm/
H A Dpage_alloc.c243 * change gfp_allowed_mask in order to avoid using I/O during memory allocations
282 static void __free_pages_ok(struct page *page, unsigned int order,
362 * many cases very high-order allocations like THP are likely to be
419 static inline void kasan_free_nondeferred_pages(struct page *page, int order) in kasan_free_nondeferred_pages() argument
422 kasan_free_pages(page, order); in kasan_free_nondeferred_pages()
683 * Higher-order pages are called "compound pages". They are structured thusly:
693 * The first tail page's ->compound_order holds the order of allocation.
694 * This usage means that zero-order pages may not be compound.
703 void prep_compound_page(struct page *page, unsigned int order) in prep_compound_page() argument
706 int nr_pages = 1 << order; in prep_compound_page()
767 set_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) set_page_guard() argument
785 clear_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) clear_page_guard() argument
798 set_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) set_page_guard() argument
800 clear_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) clear_page_guard() argument
804 set_buddy_order(struct page *page, unsigned int order) set_buddy_order() argument
823 page_is_buddy(struct page *page, struct page *buddy, unsigned int order) page_is_buddy() argument
856 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument
887 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument
895 add_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list() argument
905 add_to_free_list_tail(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list_tail() argument
919 move_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) move_to_free_list() argument
927 del_page_from_free_list(struct page *page, struct zone *zone, unsigned int order) del_page_from_free_list() argument
949 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) buddy_merge_likely() argument
994 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) __free_one_page() argument
1212 free_pages_prepare(struct page *page, unsigned int order, bool check_free) free_pages_prepare() argument
1269 PAGE_SIZE << order); free_pages_prepare() local
1271 PAGE_SIZE << order); free_pages_prepare() local
1433 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, int migratetype, fpi_t fpi_flags) free_one_page() argument
1521 __free_pages_ok(struct page *page, unsigned int order, fpi_t fpi_flags) __free_pages_ok() argument
1534 free_one_page(page_zone(page), page, pfn, order, migratetype, __free_pages_ok() local
1539 __free_pages_core(struct page *page, unsigned int order) __free_pages_core() argument
1612 memblock_free_pages(struct page *page, unsigned long pfn, unsigned int order) memblock_free_pages() argument
2007 deferred_grow_zone(struct zone *zone, unsigned int order) deferred_grow_zone() argument
2074 _deferred_grow_zone(struct zone *zone, unsigned int order) _deferred_grow_zone() argument
2262 check_new_pages(struct page *page, unsigned int order) check_new_pages() argument
2275 post_alloc_hook(struct page *page, unsigned int order, gfp_t gfp_flags) post_alloc_hook() argument
2289 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) prep_new_page() argument
2317 __rmqueue_smallest(struct zone *zone, unsigned int order, int migratetype) __rmqueue_smallest() argument
2357 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument
2363 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument
2378 unsigned int order; move_freepages() local
2459 can_steal_fallback(unsigned int order, int start_mt) can_steal_fallback() argument
2607 find_suitable_fallback(struct free_area *area, unsigned int order, int migratetype, bool only_stealable, bool *can_steal) find_suitable_fallback() argument
2692 int order; unreserve_highatomic_pageblock() local
2767 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) __rmqueue_fallback() argument
2845 __rmqueue_with_cma_reuse(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue_with_cma_reuse() argument
2870 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue() argument
2915 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) rmqueue_bulk() argument
3144 unsigned int order, t; mark_free_pages() local
3296 split_page(struct page *page, unsigned int order) split_page() argument
3310 __isolate_free_page(struct page *page, unsigned int order) __isolate_free_page() argument
3367 __putback_isolated_page(struct page *page, unsigned int order, int mt) __putback_isolated_page() argument
3457 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) rmqueue() argument
3548 __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) __should_fail_alloc_page() argument
3588 __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) __should_fail_alloc_page() argument
3595 should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) should_fail_alloc_page() argument
3601 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) __zone_watermark_unusable_free() argument
3630 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) __zone_watermark_ok() argument
3694 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) zone_watermark_ok() argument
3701 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) zone_watermark_fast() argument
3745 zone_watermark_ok_safe(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx) zone_watermark_ok_safe() argument
3829 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) get_page_from_freelist() argument
4023 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) __alloc_pages_cpuset_fallback() argument
4043 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_may_oom() argument
4137 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument
4192 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument
4270 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument
4279 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument
4383 __perform_reclaim(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac) __perform_reclaim() argument
4411 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_direct_reclaim() argument
4446 wake_all_kswapds(unsigned int order, gfp_t gfp_mask, const struct alloc_context *ac) wake_all_kswapds() argument
4557 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) should_reclaim_retry() argument
4681 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct alloc_context *ac) __alloc_pages_slowpath() argument
4959 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_mask, unsigned int *alloc_flags) prepare_alloc_pages() argument
5014 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask) __alloc_pages_nodemask() argument
5082 __get_free_pages(gfp_t gfp_mask, unsigned int order) __get_free_pages() argument
5099 free_the_page(struct page *page, unsigned int order) free_the_page() argument
5107 __free_pages(struct page *page, unsigned int order) __free_pages() argument
5120 free_pages(unsigned long addr, unsigned int order) free_pages() argument
5124 __free_pages(virt_to_page((void *)addr), order); free_pages() local
5161 int order = get_order(nc->size); __page_frag_cache_refill() local
5287 make_alloc_exact(unsigned long addr, unsigned int order, size_t size) make_alloc_exact() argument
5294 split_page(virt_to_page((void *)addr), order); make_alloc_exact() local
5320 unsigned int order = get_order(size); alloc_pages_exact() local
5345 unsigned int order = get_order(size); alloc_pages_exact_nid() local
5734 unsigned int order; show_free_areas() local
6341 unsigned int order, t; zone_init_free_lists() local
6975 unsigned int order; set_pageblock_order() local
8662 unsigned int order; alloc_contig_range() local
8947 unsigned int order; __offline_isolated_pages() local
8989 unsigned int order; is_free_buddy_page() local
9046 unsigned int order; take_page_off_buddy() local
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dcompaction.h61 * Number of free order-0 pages that should be available above given watermark
65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
69 * free scanner may have up to 1 << order pages on its list and then in compact_gap()
70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap()
71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap()
80 return 2UL << order; in compact_gap()
91 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
92 extern int fragmentation_index(struct zone *zone, unsigned int order);
94 unsigned int order, unsigned int alloc_flags,
98 extern enum compact_result compaction_suitable(struct zone *zone, int order,
191 compaction_suitable(struct zone *zone, int order, int alloc_flags, int highest_zoneidx) compaction_suitable() argument
197 defer_compaction(struct zone *zone, int order) defer_compaction() argument
201 compaction_deferred(struct zone *zone, int order) compaction_deferred() argument
234 wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx) wakeup_kcompactd() argument
[all...]
H A Dgfp.h157 * of so called costly allocations (with order > %PAGE_ALLOC_COSTLY_ORDER).
395 * The zone fallback order is MOVABLE=>HIGHMEM=>NORMAL=>DMA32=>DMA.
503 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
506 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
516 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid,
520 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument
522 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages()
530 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument
535 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node()
544 unsigned int order) in alloc_pages_node()
543 alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) alloc_pages_node() argument
556 alloc_pages(gfp_t gfp_mask, unsigned int order) alloc_pages() argument
566 alloc_pages(gfp_t gfp_mask, unsigned int order) alloc_pages() argument
[all...]
/kernel/linux/linux-5.10/tools/testing/radix-tree/
H A Dmultiorder.c3 * multiorder.c: Multi-order radix tree entry testing
16 unsigned order) in item_insert_order()
18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order()
19 struct item *item = item_create(index, order); in item_insert_order()
42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local
47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration()
53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration()
58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration()
60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration()
66 assert(item->order in multiorder_iteration()
15 item_insert_order(struct xarray *xa, unsigned long index, unsigned order) item_insert_order() argument
82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; multiorder_tagged_iteration() local
167 unsigned int order = RADIX_TREE_MAP_SHIFT - 1; creator_func() local
[all...]
/kernel/linux/linux-6.6/tools/testing/radix-tree/
H A Dmultiorder.c3 * multiorder.c: Multi-order radix tree entry testing
16 unsigned order) in item_insert_order()
18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order()
19 struct item *item = item_create(index, order); in item_insert_order()
42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local
47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration()
53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration()
58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration()
60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration()
66 assert(item->order in multiorder_iteration()
15 item_insert_order(struct xarray *xa, unsigned long index, unsigned order) item_insert_order() argument
82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; multiorder_tagged_iteration() local
167 unsigned int order = RADIX_TREE_MAP_SHIFT - 1; creator_func() local
218 unsigned int order; load_creator() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/tests/
H A Ddrm_buddy_test.c21 static inline u64 get_size(int order, u64 chunk_size) in get_size() argument
23 return (1 << order) * chunk_size; in get_size()
49 kunit_err(test, "block info: header=%llx, state=%u, order=%d, offset=%llx size=%llx root=%d buddy=%d\n", in __dump_block()
228 unsigned int order; in check_mm() local
244 order = drm_buddy_block_order(root); in check_mm()
247 if (order != mm->max_order) { in check_mm()
248 kunit_err(test, "max order root missing\n"); in check_mm()
268 block = list_first_entry_or_null(&mm->free_list[order], in check_mm()
271 kunit_err(test, "root mismatch at order=%u\n", order); in check_mm()
331 int order, top; drm_test_buddy_alloc_pathological() local
411 int *order; drm_test_buddy_alloc_smoke() local
432 int order, err; drm_test_buddy_alloc_smoke() local
503 unsigned int order; drm_test_buddy_alloc_pessimistic() local
600 int order; drm_test_buddy_alloc_optimistic() local
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dcompaction.h61 * Number of free order-0 pages that should be available above given watermark
65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
69 * free scanner may have up to 1 << order pages on its list and then in compact_gap()
70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap()
71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap()
80 return 2UL << order; in compact_gap()
85 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
86 extern int fragmentation_index(struct zone *zone, unsigned int order);
88 unsigned int order, unsigned int alloc_flags,
92 extern bool compaction_suitable(struct zone *zone, int order,
110 compaction_suitable(struct zone *zone, int order, int highest_zoneidx) compaction_suitable() argument
123 wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx) wakeup_kcompactd() argument
[all...]
/kernel/linux/linux-6.6/kernel/bpf/
H A Dcgroup_iter.c13 * 1. Walk the descendants of a cgroup in pre-order.
14 * 2. Walk the descendants of a cgroup in post-order.
18 * For walking descendants, cgroup_iter can walk in either pre-order or
19 * post-order. For walking ancestors, the iter walks up from a cgroup to
40 * EOPNOTSUPP. In order to work around, the user may have to update their
54 int order; member
77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start()
79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start()
110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next()
112 else if (p->order in cgroup_iter_seq_next()
200 int order = linfo->cgroup.order; bpf_iter_attach_cgroup() local
[all...]
/kernel/linux/linux-5.10/drivers/media/pci/cx18/
H A Dcx18-mailbox.c230 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
239 mb = &order->mb; in epu_dma_done()
246 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
252 mdl_ack = order->mdl_ack; in epu_dma_done()
276 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
323 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
326 char *str = order->str; in epu_debug()
328 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
334 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
336 switch (order in epu_cmd()
363 free_in_work_order(struct cx18 *cx, struct cx18_in_work_order *order) free_in_work_order() argument
370 struct cx18_in_work_order *order = cx18_in_work_handler() local
382 mb_ack_irq(struct cx18 *cx, struct cx18_in_work_order *order) mb_ack_irq() argument
416 epu_dma_done_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_dma_done_irq() argument
444 epu_debug_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_debug_irq() argument
465 epu_cmd_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_cmd_irq() argument
500 struct cx18_in_work_order *order = NULL; alloc_in_work_order_irq() local
524 struct cx18_in_work_order *order; cx18_api_epu_cmd_irq() local
[all...]
/kernel/linux/linux-6.6/drivers/media/pci/cx18/
H A Dcx18-mailbox.c231 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
240 mb = &order->mb; in epu_dma_done()
247 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
253 mdl_ack = order->mdl_ack; in epu_dma_done()
277 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
324 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
327 char *str = order->str; in epu_debug()
329 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
335 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
337 switch (order in epu_cmd()
364 free_in_work_order(struct cx18 *cx, struct cx18_in_work_order *order) free_in_work_order() argument
371 struct cx18_in_work_order *order = cx18_in_work_handler() local
383 mb_ack_irq(struct cx18 *cx, struct cx18_in_work_order *order) mb_ack_irq() argument
417 epu_dma_done_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_dma_done_irq() argument
445 epu_debug_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_debug_irq() argument
466 epu_cmd_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_cmd_irq() argument
501 struct cx18_in_work_order *order = NULL; alloc_in_work_order_irq() local
525 struct cx18_in_work_order *order; cx18_api_epu_cmd_irq() local
[all...]
/kernel/linux/linux-5.10/Documentation/trace/postprocess/
H A Dtrace-vmscan-postprocess.pl31 # Per-order events
110 my $regex_direct_begin_default = 'order=([0-9]*) may_writepage=([0-9]*) gfp_flags=([A-Z_|]*)';
112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)';
114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) zid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)';
115 my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned=([0-9]*) nr_skipped=([0-9]*) nr_taken=([0-9]*) lru=([a-z_]*)';
166 # Verify fields are in the right order
187 "order", "may_writepage",
196 "nid", "order");
204 "nid", "zid", "order", "gfp_flags");
208 "isolate_mode", "classzone_idx", "order",
[all...]
/kernel/linux/linux-6.6/Documentation/trace/postprocess/
H A Dtrace-vmscan-postprocess.pl31 # Per-order events
110 my $regex_direct_begin_default = 'order=([0-9]*) may_writepage=([0-9]*) gfp_flags=([A-Z_|]*)';
112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)';
114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) zid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)';
115 my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned=([0-9]*) nr_skipped=([0-9]*) nr_taken=([0-9]*) lru=([a-z_]*)';
166 # Verify fields are in the right order
187 "order", "may_writepage",
196 "nid", "order");
204 "nid", "zid", "order", "gfp_flags");
208 "isolate_mode", "classzone_idx", "order",
[all...]

Completed in 20 milliseconds

12345678910>>...39