/kernel/linux/linux-6.6/scripts/atomic/ |
H A D | gen-atomic-fallback.sh | 8 #gen_template_fallback(template, meta, pfx, name, sfx, order, atomic, int, args...) 16 local order="$1"; shift 28 #gen_order_fallback(meta, pfx, name, sfx, order, atomic, int, args...) 35 local order="$1"; shift 37 local tmpl_order=${order#_} 39 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@" 42 #gen_proto_fallback(meta, pfx, name, sfx, order, atomic, int, args...) 49 local order="$1"; shift 51 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 52 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" " [all...] |
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/nvhe/ |
H A D | page_alloc.c | 28 * __find_buddy_nocheck(pool, page 0, order 0) => page 1 29 * __find_buddy_nocheck(pool, page 0, order 1) => page 2 30 * __find_buddy_nocheck(pool, page 1, order 0) => page 0 31 * __find_buddy_nocheck(pool, page 2, order 0) => page 3 35 unsigned short order) in __find_buddy_nocheck() 39 addr ^= (PAGE_SIZE << order); in __find_buddy_nocheck() 54 unsigned short order) in __find_buddy_avail() 56 struct hyp_page *buddy = __find_buddy_nocheck(pool, p, order); in __find_buddy_avail() 58 if (!buddy || buddy->order != order || budd in __find_buddy_avail() 33 __find_buddy_nocheck(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __find_buddy_nocheck() argument 52 __find_buddy_avail(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __find_buddy_avail() argument 97 unsigned short order = p->order; __hyp_attach_page() local 130 __hyp_extract_page(struct hyp_pool *pool, struct hyp_page *p, unsigned short order) __hyp_extract_page() argument 186 unsigned short order = p->order; hyp_split_page() local 198 hyp_alloc_pages(struct hyp_pool *pool, unsigned short order) hyp_alloc_pages() argument [all...] |
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | compaction.h | 173 int order, 177 TP_ARGS(order, gfp_mask, prio), 180 __field(int, order) 186 __entry->order = order; 191 TP_printk("order=%d gfp_mask=%s priority=%d", 192 __entry->order, 200 int order, 203 TP_ARGS(zone, order, ret), 208 __field(int, order) [all...] |
H A D | kmem.h | 154 TP_PROTO(struct page *page, unsigned int order), 156 TP_ARGS(page, order), 160 __field( unsigned int, order ) 165 __entry->order = order; 168 TP_printk("page=%p pfn=%lu order=%d", 171 __entry->order) 188 TP_printk("page=%p pfn=%lu order=0", 195 TP_PROTO(struct page *page, unsigned int order, 198 TP_ARGS(page, order, gfp_flag [all...] |
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | compaction.h | 168 int order, 172 TP_ARGS(order, gfp_mask, prio), 175 __field(int, order) 181 __entry->order = order; 186 TP_printk("order=%d gfp_mask=%s priority=%d", 187 __entry->order, 195 int order, 198 TP_ARGS(zone, order, ret), 203 __field(int, order) [all...] |
H A D | kmem.h | 138 TP_PROTO(struct page *page, unsigned int order), 140 TP_ARGS(page, order), 144 __field( unsigned int, order ) 149 __entry->order = order; 152 TP_printk("page=%p pfn=0x%lx order=%d", 155 __entry->order) 172 TP_printk("page=%p pfn=0x%lx order=0", 179 TP_PROTO(struct page *page, unsigned int order, 182 TP_ARGS(page, order, gfp_flag [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/ |
H A D | i915_buddy.c | 15 pr_err("block info: header=%llx, state=%u, order=%d, offset=%llx size=%llx root=%s buddy=%s\n", in __igt_dump_block() 202 unsigned int order; in igt_check_mm() local 218 order = i915_buddy_block_order(root); in igt_check_mm() 221 if (order != mm->max_order) { in igt_check_mm() 222 pr_err("max order root missing\n"); in igt_check_mm() 242 block = list_first_entry_or_null(&mm->free_list[order], in igt_check_mm() 246 pr_err("root mismatch at order=%u\n", order); in igt_check_mm() 307 int *order; in igt_buddy_alloc_smoke() local 320 order in igt_buddy_alloc_smoke() 329 int order; igt_buddy_alloc_smoke() local 412 unsigned int order; igt_buddy_alloc_pessimistic() local 505 int order; igt_buddy_alloc_optimistic() local 556 int order, top; igt_buddy_alloc_pathological() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | page_alloc.c | 233 static void __free_pages_ok(struct page *page, unsigned int order, 310 static bool page_contains_unaccepted(struct page *page, unsigned int order); 311 static void accept_page(struct page *page, unsigned int order); 312 static bool try_to_accept_memory(struct zone *zone, unsigned int order); 338 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument 340 return deferred_grow_zone(zone, order); in _deferred_grow_zone() 527 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument 530 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex() 531 VM_BUG_ON(order != pageblock_order); in order_to_pindex() 535 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDE in order_to_pindex() 543 int order = pindex / MIGRATE_PCPTYPES; pindex_to_order() local 555 pcp_allowed_order(unsigned int order) pcp_allowed_order() argument 566 free_the_page(struct page *page, unsigned int order) free_the_page() argument 586 prep_compound_page(struct page *page, unsigned int order) prep_compound_page() argument 612 set_buddy_order(struct page *page, unsigned int order) set_buddy_order() argument 630 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument 661 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument 669 add_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list() argument 679 add_to_free_list_tail(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list_tail() argument 693 move_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) move_to_free_list() argument 701 del_page_from_free_list(struct page *page, struct zone *zone, unsigned int order) del_page_from_free_list() argument 730 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) buddy_merge_likely() argument 770 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) __free_one_page() argument 864 split_free_page(struct page *free_page, unsigned int order, unsigned long split_pfn_offset) split_free_page() argument 1082 free_pages_prepare(struct page *page, unsigned int order, fpi_t fpi_flags) free_pages_prepare() argument 1148 PAGE_SIZE << order); free_pages_prepare() local 1150 PAGE_SIZE << order); free_pages_prepare() local 1195 unsigned int order; free_pcppages_bulk() local 1249 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, int migratetype, fpi_t fpi_flags) free_one_page() argument 1265 __free_pages_ok(struct page *page, unsigned int order, fpi_t fpi_flags) __free_pages_ok() argument 1294 __free_pages_core(struct page *page, unsigned int order) __free_pages_core() argument 1445 check_new_pages(struct page *page, unsigned int order) check_new_pages() argument 1487 post_alloc_hook(struct page *page, unsigned int order, gfp_t gfp_flags) post_alloc_hook() argument 1547 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) prep_new_page() argument 1572 __rmqueue_smallest(struct zone *zone, unsigned int order, int migratetype) __rmqueue_smallest() argument 1611 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument 1617 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument 1632 unsigned int order; move_freepages() local 1708 can_steal_fallback(unsigned int order, int start_mt) can_steal_fallback() argument 1855 find_suitable_fallback(struct free_area *area, unsigned int order, int migratetype, bool only_stealable, bool *can_steal) find_suitable_fallback() argument 1936 int order; unreserve_highatomic_pageblock() local 2011 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) __rmqueue_fallback() argument 2089 __rmqueue_with_cma_reuse(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue_with_cma_reuse() argument 2114 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue() argument 2157 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) rmqueue_bulk() argument 2341 free_unref_page_prepare(struct page *page, unsigned long pfn, unsigned int order) free_unref_page_prepare() argument 2401 free_unref_page_commit(struct zone *zone, struct per_cpu_pages *pcp, struct page *page, int migratetype, unsigned int order) free_unref_page_commit() argument 2431 free_unref_page(struct page *page, unsigned int order) free_unref_page() argument 2452 free_one_page(page_zone(page), page, pfn, order, migratetype, FPI_NONE); free_unref_page() local 2563 split_page(struct page *page, unsigned int order) split_page() argument 2577 __isolate_free_page(struct page *page, unsigned int order) __isolate_free_page() argument 2629 __putback_isolated_page(struct page *page, unsigned int order, int mt) __putback_isolated_page() argument 2668 rmqueue_buddy(struct zone *preferred_zone, struct zone *zone, unsigned int order, unsigned int alloc_flags, int migratetype) rmqueue_buddy() argument 2710 __rmqueue_pcplist(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags, struct per_cpu_pages *pcp, struct list_head *list) __rmqueue_pcplist() argument 2750 rmqueue_pcplist(struct zone *preferred_zone, struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) rmqueue_pcplist() argument 2797 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) rmqueue() argument 2832 should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) should_fail_alloc_page() argument 2838 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) __zone_watermark_unusable_free() argument 2869 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) __zone_watermark_ok() argument 2947 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) zone_watermark_ok() argument 2954 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) zone_watermark_fast() argument 2999 zone_watermark_ok_safe(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx) zone_watermark_ok_safe() argument 3082 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) get_page_from_freelist() argument 3287 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) __alloc_pages_cpuset_fallback() argument 3307 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_may_oom() argument 3402 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument 3461 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument 3527 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument 3536 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument 3648 __perform_reclaim(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac) __perform_reclaim() argument 3674 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_direct_reclaim() argument 3707 wake_all_kswapds(unsigned int order, gfp_t gfp_mask, const struct alloc_context *ac) wake_all_kswapds() argument 3727 gfp_to_alloc_flags(gfp_t gfp_mask, unsigned int order) gfp_to_alloc_flags() argument 3828 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) should_reclaim_retry() argument 3933 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct alloc_context *ac) __alloc_pages_slowpath() argument 4209 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_gfp, unsigned int *alloc_flags) prepare_alloc_pages() argument 4429 __alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, nodemask_t *nodemask) __alloc_pages() argument 4494 __folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, nodemask_t *nodemask) __folio_alloc() argument 4512 __get_free_pages(gfp_t gfp_mask, unsigned int order) __get_free_pages() argument 4549 __free_pages(struct page *page, unsigned int order) __free_pages() argument 4562 free_pages(unsigned long addr, unsigned int order) free_pages() argument 4566 __free_pages(virt_to_page((void *)addr), order); free_pages() local 4698 make_alloc_exact(unsigned long addr, unsigned int order, size_t size) make_alloc_exact() argument 4735 unsigned int order = get_order(size); alloc_pages_exact() local 4760 unsigned int order = get_order(size); alloc_pages_exact_nid() local 6152 int order; alloc_contig_range() local 6437 unsigned int order; __offline_isolated_pages() local 6480 unsigned int order; is_free_buddy_page() local 6537 unsigned int order; take_page_off_buddy() local 6627 page_contains_unaccepted(struct page *page, unsigned int order) page_contains_unaccepted() argument 6635 accept_page(struct page *page, unsigned int order) accept_page() argument 6676 try_to_accept_memory(struct zone *zone, unsigned int order) try_to_accept_memory() argument 6726 page_contains_unaccepted(struct page *page, unsigned int order) page_contains_unaccepted() argument 6731 accept_page(struct page *page, unsigned int order) accept_page() argument 6735 try_to_accept_memory(struct zone *zone, unsigned int order) try_to_accept_memory() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 54 * @vaddr: original vaddr return for the mapping and order in the lower bits 78 /* Allocate pages of size 1 << order with the given gfp_flags */ 80 unsigned int order) in ttm_pool_alloc_page() 87 /* Don't set the __GFP_COMP flag for higher order allocations. in ttm_pool_alloc_page() 91 if (order) in ttm_pool_alloc_page() 96 p = alloc_pages_node(pool->nid, gfp_flags, order); in ttm_pool_alloc_page() 98 p->private = order; in ttm_pool_alloc_page() 106 if (order) in ttm_pool_alloc_page() 109 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page() 122 dma->vaddr = (unsigned long)vaddr | order; in ttm_pool_alloc_page() 79 ttm_pool_alloc_page(struct ttm_pool *pool, gfp_t gfp_flags, unsigned int order) ttm_pool_alloc_page() argument 132 ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order, struct page *p) ttm_pool_free_page() argument 185 ttm_pool_map(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr) ttm_pool_map() argument 258 ttm_pool_type_init(struct ttm_pool_type *pt, struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_type_init() argument 286 ttm_pool_select_type(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_select_type() argument 349 ttm_pool_page_allocated(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr, unsigned long *num_pages, struct page ***pages) ttm_pool_page_allocated() argument 388 unsigned int order; ttm_pool_free_range() local 429 unsigned int order; ttm_pool_alloc() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/lib/ |
H A D | drm_random.c | 15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 23 swap(order[i], order[j]); in drm_random_reorder() 30 unsigned int *order, i; in drm_random_order() local 32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 33 if (!order) in drm_random_order() 34 return order; in drm_random_order() 37 order[i] = i; in drm_random_order() 39 drm_random_reorder(order, coun in drm_random_order() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/lib/ |
H A D | drm_random.c | 16 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 24 swap(order[i], order[j]); in drm_random_reorder() 31 unsigned int *order, i; in drm_random_order() local 33 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 34 if (!order) in drm_random_order() 35 return order; in drm_random_order() 38 order[i] = i; in drm_random_order() 40 drm_random_reorder(order, coun in drm_random_order() [all...] |
/kernel/linux/linux-6.6/mm/kmsan/ |
H A D | init.c | 105 * by their order: when kmsan_memblock_free_pages() is called for the first 106 * time with a certain order, it is reserved as a shadow block, for the second 109 * after which held_back[order] can be used again. 114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument 118 if (!held_back[order].shadow) { in kmsan_memblock_free_pages() 119 held_back[order].shadow = page; in kmsan_memblock_free_pages() 122 if (!held_back[order].origin) { in kmsan_memblock_free_pages() 123 held_back[order].origin = page; in kmsan_memblock_free_pages() 126 shadow = held_back[order].shadow; in kmsan_memblock_free_pages() 127 origin = held_back[order] in kmsan_memblock_free_pages() 139 int order; global() member [all...] |
/kernel/linux/linux-6.6/arch/riscv/kvm/ |
H A D | tlb.c | 22 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() 26 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa() 33 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 38 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 50 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() 54 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa() 61 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 66 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 81 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() 85 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva() 20 kvm_riscv_local_hfence_gvma_vmid_gpa(unsigned long vmid, gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_local_hfence_gvma_vmid_gpa() argument 49 kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_local_hfence_gvma_gpa() argument 77 kvm_riscv_local_hfence_vvma_asid_gva(unsigned long vmid, unsigned long asid, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_local_hfence_vvma_asid_gva() argument 119 kvm_riscv_local_hfence_vvma_gva(unsigned long vmid, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_local_hfence_vvma_gva() argument 332 kvm_riscv_hfence_gvma_vmid_gpa(struct kvm *kvm, unsigned long hbase, unsigned long hmask, gpa_t gpa, gpa_t gpsz, unsigned long order) kvm_riscv_hfence_gvma_vmid_gpa() argument 355 kvm_riscv_hfence_vvma_asid_gva(struct kvm *kvm, unsigned long hbase, unsigned long hmask, unsigned long gva, unsigned long gvsz, unsigned long order, unsigned long asid) kvm_riscv_hfence_vvma_asid_gva() argument 384 kvm_riscv_hfence_vvma_gva(struct kvm *kvm, unsigned long hbase, unsigned long hmask, unsigned long gva, unsigned long gvsz, unsigned long order) kvm_riscv_hfence_vvma_gva() argument [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | page_alloc.c | 243 * change gfp_allowed_mask in order to avoid using I/O during memory allocations 282 static void __free_pages_ok(struct page *page, unsigned int order, 362 * many cases very high-order allocations like THP are likely to be 419 static inline void kasan_free_nondeferred_pages(struct page *page, int order) in kasan_free_nondeferred_pages() argument 422 kasan_free_pages(page, order); in kasan_free_nondeferred_pages() 683 * Higher-order pages are called "compound pages". They are structured thusly: 693 * The first tail page's ->compound_order holds the order of allocation. 694 * This usage means that zero-order pages may not be compound. 703 void prep_compound_page(struct page *page, unsigned int order) in prep_compound_page() argument 706 int nr_pages = 1 << order; in prep_compound_page() 767 set_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) set_page_guard() argument 785 clear_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) clear_page_guard() argument 798 set_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) set_page_guard() argument 800 clear_page_guard(struct zone *zone, struct page *page, unsigned int order, int migratetype) clear_page_guard() argument 804 set_buddy_order(struct page *page, unsigned int order) set_buddy_order() argument 823 page_is_buddy(struct page *page, struct page *buddy, unsigned int order) page_is_buddy() argument 856 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument 887 compaction_capture(struct capture_control *capc, struct page *page, int order, int migratetype) compaction_capture() argument 895 add_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list() argument 905 add_to_free_list_tail(struct page *page, struct zone *zone, unsigned int order, int migratetype) add_to_free_list_tail() argument 919 move_to_free_list(struct page *page, struct zone *zone, unsigned int order, int migratetype) move_to_free_list() argument 927 del_page_from_free_list(struct page *page, struct zone *zone, unsigned int order) del_page_from_free_list() argument 949 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) buddy_merge_likely() argument 994 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) __free_one_page() argument 1212 free_pages_prepare(struct page *page, unsigned int order, bool check_free) free_pages_prepare() argument 1269 PAGE_SIZE << order); free_pages_prepare() local 1271 PAGE_SIZE << order); free_pages_prepare() local 1433 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, int migratetype, fpi_t fpi_flags) free_one_page() argument 1521 __free_pages_ok(struct page *page, unsigned int order, fpi_t fpi_flags) __free_pages_ok() argument 1534 free_one_page(page_zone(page), page, pfn, order, migratetype, __free_pages_ok() local 1539 __free_pages_core(struct page *page, unsigned int order) __free_pages_core() argument 1612 memblock_free_pages(struct page *page, unsigned long pfn, unsigned int order) memblock_free_pages() argument 2007 deferred_grow_zone(struct zone *zone, unsigned int order) deferred_grow_zone() argument 2074 _deferred_grow_zone(struct zone *zone, unsigned int order) _deferred_grow_zone() argument 2262 check_new_pages(struct page *page, unsigned int order) check_new_pages() argument 2275 post_alloc_hook(struct page *page, unsigned int order, gfp_t gfp_flags) post_alloc_hook() argument 2289 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) prep_new_page() argument 2317 __rmqueue_smallest(struct zone *zone, unsigned int order, int migratetype) __rmqueue_smallest() argument 2357 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument 2363 __rmqueue_cma_fallback(struct zone *zone, unsigned int order) __rmqueue_cma_fallback() argument 2378 unsigned int order; move_freepages() local 2459 can_steal_fallback(unsigned int order, int start_mt) can_steal_fallback() argument 2607 find_suitable_fallback(struct free_area *area, unsigned int order, int migratetype, bool only_stealable, bool *can_steal) find_suitable_fallback() argument 2692 int order; unreserve_highatomic_pageblock() local 2767 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) __rmqueue_fallback() argument 2845 __rmqueue_with_cma_reuse(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue_with_cma_reuse() argument 2870 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) __rmqueue() argument 2915 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) rmqueue_bulk() argument 3144 unsigned int order, t; mark_free_pages() local 3296 split_page(struct page *page, unsigned int order) split_page() argument 3310 __isolate_free_page(struct page *page, unsigned int order) __isolate_free_page() argument 3367 __putback_isolated_page(struct page *page, unsigned int order, int mt) __putback_isolated_page() argument 3457 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) rmqueue() argument 3548 __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) __should_fail_alloc_page() argument 3588 __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) __should_fail_alloc_page() argument 3595 should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) should_fail_alloc_page() argument 3601 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) __zone_watermark_unusable_free() argument 3630 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) __zone_watermark_ok() argument 3694 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) zone_watermark_ok() argument 3701 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) zone_watermark_fast() argument 3745 zone_watermark_ok_safe(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx) zone_watermark_ok_safe() argument 3829 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) get_page_from_freelist() argument 4023 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) __alloc_pages_cpuset_fallback() argument 4043 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_may_oom() argument 4137 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument 4192 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument 4270 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) __alloc_pages_direct_compact() argument 4279 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) should_compact_retry() argument 4383 __perform_reclaim(gfp_t gfp_mask, unsigned int order, const struct alloc_context *ac) __perform_reclaim() argument 4411 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) __alloc_pages_direct_reclaim() argument 4446 wake_all_kswapds(unsigned int order, gfp_t gfp_mask, const struct alloc_context *ac) wake_all_kswapds() argument 4557 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) should_reclaim_retry() argument 4681 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct alloc_context *ac) __alloc_pages_slowpath() argument 4959 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_mask, unsigned int *alloc_flags) prepare_alloc_pages() argument 5014 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask) __alloc_pages_nodemask() argument 5082 __get_free_pages(gfp_t gfp_mask, unsigned int order) __get_free_pages() argument 5099 free_the_page(struct page *page, unsigned int order) free_the_page() argument 5107 __free_pages(struct page *page, unsigned int order) __free_pages() argument 5120 free_pages(unsigned long addr, unsigned int order) free_pages() argument 5124 __free_pages(virt_to_page((void *)addr), order); free_pages() local 5161 int order = get_order(nc->size); __page_frag_cache_refill() local 5287 make_alloc_exact(unsigned long addr, unsigned int order, size_t size) make_alloc_exact() argument 5294 split_page(virt_to_page((void *)addr), order); make_alloc_exact() local 5320 unsigned int order = get_order(size); alloc_pages_exact() local 5345 unsigned int order = get_order(size); alloc_pages_exact_nid() local 5734 unsigned int order; show_free_areas() local 6341 unsigned int order, t; zone_init_free_lists() local 6975 unsigned int order; set_pageblock_order() local 8662 unsigned int order; alloc_contig_range() local 8947 unsigned int order; __offline_isolated_pages() local 8989 unsigned int order; is_free_buddy_page() local 9046 unsigned int order; take_page_off_buddy() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | compaction.h | 61 * Number of free order-0 pages that should be available above given watermark 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 69 * free scanner may have up to 1 << order pages on its list and then in compact_gap() 70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap() 71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap() 80 return 2UL << order; in compact_gap() 91 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 92 extern int fragmentation_index(struct zone *zone, unsigned int order); 94 unsigned int order, unsigned int alloc_flags, 98 extern enum compact_result compaction_suitable(struct zone *zone, int order, 191 compaction_suitable(struct zone *zone, int order, int alloc_flags, int highest_zoneidx) compaction_suitable() argument 197 defer_compaction(struct zone *zone, int order) defer_compaction() argument 201 compaction_deferred(struct zone *zone, int order) compaction_deferred() argument 234 wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx) wakeup_kcompactd() argument [all...] |
H A D | gfp.h | 157 * of so called costly allocations (with order > %PAGE_ALLOC_COSTLY_ORDER). 395 * The zone fallback order is MOVABLE=>HIGHMEM=>NORMAL=>DMA32=>DMA. 503 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 506 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 516 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, 520 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument 522 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages() 530 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 535 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node() 544 unsigned int order) in alloc_pages_node() 543 alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) alloc_pages_node() argument 556 alloc_pages(gfp_t gfp_mask, unsigned int order) alloc_pages() argument 566 alloc_pages(gfp_t gfp_mask, unsigned int order) alloc_pages() argument [all...] |
/kernel/linux/linux-5.10/tools/testing/radix-tree/ |
H A D | multiorder.c | 3 * multiorder.c: Multi-order radix tree entry testing 16 unsigned order) in item_insert_order() 18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order() 19 struct item *item = item_create(index, order); in item_insert_order() 42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local 47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration() 53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration() 58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration() 60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration() 66 assert(item->order in multiorder_iteration() 15 item_insert_order(struct xarray *xa, unsigned long index, unsigned order) item_insert_order() argument 82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; multiorder_tagged_iteration() local 167 unsigned int order = RADIX_TREE_MAP_SHIFT - 1; creator_func() local [all...] |
/kernel/linux/linux-6.6/tools/testing/radix-tree/ |
H A D | multiorder.c | 3 * multiorder.c: Multi-order radix tree entry testing 16 unsigned order) in item_insert_order() 18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order() 19 struct item *item = item_create(index, order); in item_insert_order() 42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local 47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration() 53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration() 58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration() 60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration() 66 assert(item->order in multiorder_iteration() 15 item_insert_order(struct xarray *xa, unsigned long index, unsigned order) item_insert_order() argument 82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; multiorder_tagged_iteration() local 167 unsigned int order = RADIX_TREE_MAP_SHIFT - 1; creator_func() local 218 unsigned int order; load_creator() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/tests/ |
H A D | drm_buddy_test.c | 21 static inline u64 get_size(int order, u64 chunk_size) in get_size() argument 23 return (1 << order) * chunk_size; in get_size() 49 kunit_err(test, "block info: header=%llx, state=%u, order=%d, offset=%llx size=%llx root=%d buddy=%d\n", in __dump_block() 228 unsigned int order; in check_mm() local 244 order = drm_buddy_block_order(root); in check_mm() 247 if (order != mm->max_order) { in check_mm() 248 kunit_err(test, "max order root missing\n"); in check_mm() 268 block = list_first_entry_or_null(&mm->free_list[order], in check_mm() 271 kunit_err(test, "root mismatch at order=%u\n", order); in check_mm() 331 int order, top; drm_test_buddy_alloc_pathological() local 411 int *order; drm_test_buddy_alloc_smoke() local 432 int order, err; drm_test_buddy_alloc_smoke() local 503 unsigned int order; drm_test_buddy_alloc_pessimistic() local 600 int order; drm_test_buddy_alloc_optimistic() local [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | compaction.h | 61 * Number of free order-0 pages that should be available above given watermark 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 69 * free scanner may have up to 1 << order pages on its list and then in compact_gap() 70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap() 71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap() 80 return 2UL << order; in compact_gap() 85 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 86 extern int fragmentation_index(struct zone *zone, unsigned int order); 88 unsigned int order, unsigned int alloc_flags, 92 extern bool compaction_suitable(struct zone *zone, int order, 110 compaction_suitable(struct zone *zone, int order, int highest_zoneidx) compaction_suitable() argument 123 wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx) wakeup_kcompactd() argument [all...] |
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | cgroup_iter.c | 13 * 1. Walk the descendants of a cgroup in pre-order. 14 * 2. Walk the descendants of a cgroup in post-order. 18 * For walking descendants, cgroup_iter can walk in either pre-order or 19 * post-order. For walking ancestors, the iter walks up from a cgroup to 40 * EOPNOTSUPP. In order to work around, the user may have to update their 54 int order; member 77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start() 79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start() 110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next() 112 else if (p->order in cgroup_iter_seq_next() 200 int order = linfo->cgroup.order; bpf_iter_attach_cgroup() local [all...] |
/kernel/linux/linux-5.10/drivers/media/pci/cx18/ |
H A D | cx18-mailbox.c | 230 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 239 mb = &order->mb; in epu_dma_done() 246 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 252 mdl_ack = order->mdl_ack; in epu_dma_done() 276 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 323 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 326 char *str = order->str; in epu_debug() 328 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 334 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 336 switch (order in epu_cmd() 363 free_in_work_order(struct cx18 *cx, struct cx18_in_work_order *order) free_in_work_order() argument 370 struct cx18_in_work_order *order = cx18_in_work_handler() local 382 mb_ack_irq(struct cx18 *cx, struct cx18_in_work_order *order) mb_ack_irq() argument 416 epu_dma_done_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_dma_done_irq() argument 444 epu_debug_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_debug_irq() argument 465 epu_cmd_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_cmd_irq() argument 500 struct cx18_in_work_order *order = NULL; alloc_in_work_order_irq() local 524 struct cx18_in_work_order *order; cx18_api_epu_cmd_irq() local [all...] |
/kernel/linux/linux-6.6/drivers/media/pci/cx18/ |
H A D | cx18-mailbox.c | 231 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 240 mb = &order->mb; in epu_dma_done() 247 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 253 mdl_ack = order->mdl_ack; in epu_dma_done() 277 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 324 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 327 char *str = order->str; in epu_debug() 329 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 335 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 337 switch (order in epu_cmd() 364 free_in_work_order(struct cx18 *cx, struct cx18_in_work_order *order) free_in_work_order() argument 371 struct cx18_in_work_order *order = cx18_in_work_handler() local 383 mb_ack_irq(struct cx18 *cx, struct cx18_in_work_order *order) mb_ack_irq() argument 417 epu_dma_done_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_dma_done_irq() argument 445 epu_debug_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_debug_irq() argument 466 epu_cmd_irq(struct cx18 *cx, struct cx18_in_work_order *order) epu_cmd_irq() argument 501 struct cx18_in_work_order *order = NULL; alloc_in_work_order_irq() local 525 struct cx18_in_work_order *order; cx18_api_epu_cmd_irq() local [all...] |
/kernel/linux/linux-5.10/Documentation/trace/postprocess/ |
H A D | trace-vmscan-postprocess.pl | 31 # Per-order events 110 my $regex_direct_begin_default = 'order=([0-9]*) may_writepage=([0-9]*) gfp_flags=([A-Z_|]*)'; 112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)'; 114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) zid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)'; 115 my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned=([0-9]*) nr_skipped=([0-9]*) nr_taken=([0-9]*) lru=([a-z_]*)'; 166 # Verify fields are in the right order 187 "order", "may_writepage", 196 "nid", "order"); 204 "nid", "zid", "order", "gfp_flags"); 208 "isolate_mode", "classzone_idx", "order", [all...] |
/kernel/linux/linux-6.6/Documentation/trace/postprocess/ |
H A D | trace-vmscan-postprocess.pl | 31 # Per-order events 110 my $regex_direct_begin_default = 'order=([0-9]*) may_writepage=([0-9]*) gfp_flags=([A-Z_|]*)'; 112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)'; 114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) zid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)'; 115 my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned=([0-9]*) nr_skipped=([0-9]*) nr_taken=([0-9]*) lru=([a-z_]*)'; 166 # Verify fields are in the right order 187 "order", "may_writepage", 196 "nid", "order"); 204 "nid", "zid", "order", "gfp_flags"); 208 "isolate_mode", "classzone_idx", "order", [all...] |