Home
last modified time | relevance | path

Searched refs:page (Results 1 - 25 of 6349) sorted by relevance

12345678910>>...254

/kernel/linux/linux-5.10/include/linux/
H A Dpage_ref.h7 #include <linux/page-flags.h>
29 extern void __page_ref_set(struct page *page, int v);
30 extern void __page_ref_mod(struct page *page, int v);
31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret);
32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret);
33 extern void __page_ref_mod_unless(struct page *pag
41 __page_ref_set(struct page *page, int v) __page_ref_set() argument
44 __page_ref_mod(struct page *page, int v) __page_ref_mod() argument
47 __page_ref_mod_and_test(struct page *page, int v, int ret) __page_ref_mod_and_test() argument
50 __page_ref_mod_and_return(struct page *page, int v, int ret) __page_ref_mod_and_return() argument
53 __page_ref_mod_unless(struct page *page, int v, int u) __page_ref_mod_unless() argument
56 __page_ref_freeze(struct page *page, int v, int ret) __page_ref_freeze() argument
59 __page_ref_unfreeze(struct page *page, int v) __page_ref_unfreeze() argument
65 page_ref_count(struct page *page) page_ref_count() argument
70 page_count(struct page *page) page_count() argument
75 set_page_count(struct page *page, int v) set_page_count() argument
86 init_page_count(struct page *page) init_page_count() argument
91 page_ref_add(struct page *page, int nr) page_ref_add() argument
98 page_ref_sub(struct page *page, int nr) page_ref_sub() argument
105 page_ref_sub_return(struct page *page, int nr) page_ref_sub_return() argument
114 page_ref_inc(struct page *page) page_ref_inc() argument
121 page_ref_dec(struct page *page) page_ref_dec() argument
128 page_ref_sub_and_test(struct page *page, int nr) page_ref_sub_and_test() argument
137 page_ref_inc_return(struct page *page) page_ref_inc_return() argument
146 page_ref_dec_and_test(struct page *page) page_ref_dec_and_test() argument
155 page_ref_dec_return(struct page *page) page_ref_dec_return() argument
164 page_ref_add_unless(struct page *page, int nr, int u) page_ref_add_unless() argument
173 page_ref_freeze(struct page *page, int count) page_ref_freeze() argument
182 page_ref_unfreeze(struct page *page, int count) page_ref_unfreeze() argument
184 VM_BUG_ON_PAGE(page_count(page) != 0, page); page_ref_unfreeze() local
[all...]
H A Dpage-flags.h3 * Macros for manipulating and testing page->flags
18 * Various page->flags bits:
20 * PG_reserved is set for special pages. The "struct page" of such a page
25 * - Pages reserved or allocated early during boot (before the page allocator
27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much
29 * be given to the page allocator.
32 * - The zero page(s)
33 * - Pages not added to the page allocator when onlining a section because
49 * Consequently, PG_reserved for a page mappe
194 compound_head(struct page *page) compound_head() argument
203 PageTail(struct page *page) PageTail() argument
208 PageCompound(struct page *page) PageCompound() argument
214 PagePoisoned(const struct page *page) PagePoisoned() argument
222 page_init_poison(struct page *page, size_t size) page_init_poison() argument
416 PageSwapCache(struct page *page) PageSwapCache() argument
506 PageMappingFlags(struct page *page) PageMappingFlags() argument
511 PageAnon(struct page *page) PageAnon() argument
517 __PageMovable(struct page *page) __PageMovable() argument
530 PageKsm(struct page *page) PageKsm() argument
542 PageUptodate(struct page *page) PageUptodate() argument
561 __SetPageUptodate(struct page *page) __SetPageUptodate() argument
563 VM_BUG_ON_PAGE(PageTail(page), page); __SetPageUptodate() local
568 SetPageUptodate(struct page *page) SetPageUptodate() argument
570 VM_BUG_ON_PAGE(PageTail(page), page); SetPageUptodate() local
590 set_page_writeback(struct page *page) set_page_writeback() argument
595 set_page_writeback_keepwrite(struct page *page) set_page_writeback_keepwrite() argument
602 set_compound_head(struct page *page, struct page *head) set_compound_head() argument
607 clear_compound_head(struct page *page) clear_compound_head() argument
613 ClearPageCompound(struct page *page) ClearPageCompound() argument
630 page_huge_active(struct page *page) page_huge_active() argument
646 PageTransHuge(struct page *page) PageTransHuge() argument
648 VM_BUG_ON_PAGE(PageTail(page), page); PageTransHuge() local
657 PageTransCompound(struct page *page) PageTransCompound() argument
683 PageTransCompoundMap(struct page *page) PageTransCompoundMap() argument
704 PageTransTail(struct page *page) PageTransTail() argument
754 page_has_type(struct page *page) page_has_type() argument
824 PageSlabPfmemalloc(struct page *page) PageSlabPfmemalloc() argument
830 SetPageSlabPfmemalloc(struct page *page) SetPageSlabPfmemalloc() argument
836 __ClearPageSlabPfmemalloc(struct page *page) __ClearPageSlabPfmemalloc() argument
842 ClearPageSlabPfmemalloc(struct page *page) ClearPageSlabPfmemalloc() argument
892 page_has_private(struct page *page) page_has_private() argument
[all...]
H A Dpage_idle.h6 #include <linux/page-flags.h>
12 static inline bool page_is_young(struct page *page) in page_is_young() argument
14 return PageYoung(page); in page_is_young()
17 static inline void set_page_young(struct page *page) in set_page_young() argument
19 SetPageYoung(page); in set_page_young()
22 static inline bool test_and_clear_page_young(struct page *page) in test_and_clear_page_young() argument
24 return TestClearPageYoung(page); in test_and_clear_page_young()
27 page_is_idle(struct page *page) page_is_idle() argument
32 set_page_idle(struct page *page) set_page_idle() argument
37 clear_page_idle(struct page *page) clear_page_idle() argument
48 page_is_young(struct page *page) page_is_young() argument
58 set_page_young(struct page *page) set_page_young() argument
68 test_and_clear_page_young(struct page *page) test_and_clear_page_young() argument
78 page_is_idle(struct page *page) page_is_idle() argument
88 set_page_idle(struct page *page) set_page_idle() argument
98 clear_page_idle(struct page *page) clear_page_idle() argument
111 page_is_young(struct page *page) page_is_young() argument
116 set_page_young(struct page *page) set_page_young() argument
120 test_and_clear_page_young(struct page *page) test_and_clear_page_young() argument
125 page_is_idle(struct page *page) page_is_idle() argument
130 set_page_idle(struct page *page) set_page_idle() argument
134 clear_page_idle(struct page *page) clear_page_idle() argument
[all...]
H A Dpagemap.h158 void release_pages(struct page **pages, int nr);
161 * speculatively take a reference to a page.
162 * If the page is free (_refcount == 0), then _refcount is untouched, and 0
166 * been used to lookup the page in the pagecache radix-tree (or page table):
172 * page has been finished with, no matter what it is subsequently allocated
179 * 1. find page in radix tree
181 * 3. check the page is still in pagecache (if no, goto 1)
186 * B. remove page from pagecache
187 * C. free the page
204 __page_cache_add_speculative(struct page *page, int count) __page_cache_add_speculative() argument
219 VM_BUG_ON_PAGE(page_count(page) == 0, page); __page_cache_add_speculative() local
232 VM_BUG_ON_PAGE(PageTail(page), page); __page_cache_add_speculative() local
237 page_cache_get_speculative(struct page *page) page_cache_get_speculative() argument
242 page_cache_add_speculative(struct page *page, int count) page_cache_add_speculative() argument
255 attach_page_private(struct page *page, void *data) attach_page_private() argument
271 detach_page_private(struct page *page) detach_page_private() argument
507 page_to_index(struct page *page) page_to_index() argument
529 page_to_pgoff(struct page *page) page_to_pgoff() argument
539 page_offset(struct page *page) page_offset() argument
544 page_file_offset(struct page *page) page_file_offset() argument
564 struct page *page; global() member
570 struct page *page; global() member
598 trylock_page(struct page *page) trylock_page() argument
607 lock_page(struct page *page) lock_page() argument
619 lock_page_killable(struct page *page) lock_page_killable() argument
635 lock_page_async(struct page *page, struct wait_page_queue *wait) lock_page_async() argument
650 lock_page_or_retry(struct page *page, struct mm_struct *mm, unsigned int flags) lock_page_or_retry() argument
671 wait_on_page_locked(struct page *page) wait_on_page_locked() argument
677 wait_on_page_locked_killable(struct page *page) wait_on_page_locked_killable() argument
768 add_to_page_cache(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache() argument
857 page_cache_async_readahead(struct address_space *mapping, struct file_ra_state *ra, struct file *file, struct page *page, pgoff_t index, unsigned long req_count) page_cache_async_readahead() argument
876 struct page *page; readahead_page() local
899 struct page *page; __readahead_batch() local
912 VM_BUG_ON_PAGE(PageTail(page), page); __readahead_batch() local
997 page_mkwrite_check_truncate(struct page *page, struct inode *inode) page_mkwrite_check_truncate() argument
1029 i_blocks_per_page(struct inode *inode, struct page *page) i_blocks_per_page() argument
[all...]
H A Dmigrate.h10 typedef struct page *new_page_t(struct page *page, unsigned long private);
11 typedef void free_page_t(struct page *page, unsigned long private);
17 * - negative errno on page migration failure;
18 * - zero on page migration success;
40 struct page *newpage, struct page *page,
66 alloc_migration_target(struct page *page, unsigned long private) alloc_migration_target() argument
69 isolate_movable_page(struct page *page, isolate_mode_t mode) isolate_movable_page() argument
75 migrate_page_states(struct page *newpage, struct page *page) migrate_page_states() argument
79 migrate_page_copy(struct page *newpage, struct page *page) migrate_page_copy() argument
82 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument
95 PageMovable(struct page *page) PageMovable() argument
96 __SetPageMovable(struct page *page, struct address_space *mapping) __SetPageMovable() argument
100 __ClearPageMovable(struct page *page) __ClearPageMovable() argument
114 migrate_misplaced_page(struct page *page, struct vm_area_struct *vma, int node) migrate_misplaced_page() argument
128 migrate_misplaced_transhuge_page(struct mm_struct *mm, struct vm_area_struct *vma, pmd_t *pmd, pmd_t entry, unsigned long address, struct page *page, int node) migrate_misplaced_transhuge_page() argument
[all...]
H A Dballoon_compaction.h7 * Balloon page migration makes use of the general non-lru movable page
10 * page->private is used to reference the responsible balloon device.
11 * page->mapping is used in context of non-lru page migration to reference
12 * the address space operations for page isolation/migration/compaction.
14 * As the page isolation scanning step a compaction thread does is a lockless
15 * procedure (from a page standpoint), it might bring some racy situations while
16 * performing balloon page compaction. In order to sort out these racy scenarios
17 * and safely perform balloon's page compactio
99 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) balloon_page_insert() argument
116 balloon_page_delete(struct page *page) balloon_page_delete() argument
133 balloon_page_device(struct page *page) balloon_page_device() argument
145 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) balloon_page_insert() argument
152 balloon_page_delete(struct page *page) balloon_page_delete() argument
158 balloon_page_isolate(struct page *page) balloon_page_isolate() argument
163 balloon_page_putback(struct page *page) balloon_page_putback() argument
168 balloon_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) balloon_page_migrate() argument
188 balloon_page_push(struct list_head *pages, struct page *page) balloon_page_push() argument
202 struct page *page = list_first_entry_or_null(pages, struct page, lru); balloon_page_pop() local
[all...]
H A Dhighmem.h15 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) in flush_anon_page() argument
21 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument
35 extern void *kmap_atomic_high_prot(struct page *page, pgprot_t prot);
47 void *kmap_high(struct page *page);
48 static inline void *kmap(struct page *page) in kmap() argument
63 kunmap(struct page *page) kunmap() argument
84 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
137 kmap(struct page *page) kmap() argument
143 kunmap_high(struct page *page) kunmap_high() argument
147 kunmap(struct page *page) kunmap() argument
154 kmap_atomic(struct page *page) kmap_atomic() argument
307 zero_user_segment(struct page *page, unsigned start, unsigned end) zero_user_segment() argument
313 zero_user(struct page *page, unsigned start, unsigned size) zero_user() argument
350 memcpy_from_page(char *to, struct page *page, size_t offset, size_t len) memcpy_from_page() argument
359 memcpy_to_page(struct page *page, size_t offset, const char *from, size_t len) memcpy_to_page() argument
[all...]
H A Dpage_owner.h11 extern void __reset_page_owner(struct page *page, unsigned int order);
12 extern void __set_page_owner(struct page *page,
14 extern void __split_page_owner(struct page *page, unsigned int nr);
15 extern void __copy_page_owner(struct page *oldpage, struct page *newpage);
16 extern void __set_page_owner_migrate_reason(struct page *page, in
21 reset_page_owner(struct page *page, unsigned int order) reset_page_owner() argument
27 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument
34 split_page_owner(struct page *page, unsigned int nr) split_page_owner() argument
44 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument
49 dump_page_owner(struct page *page) dump_page_owner() argument
55 reset_page_owner(struct page *page, unsigned int order) reset_page_owner() argument
58 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument
62 split_page_owner(struct page *page, unsigned int order) split_page_owner() argument
69 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument
72 dump_page_owner(struct page *page) dump_page_owner() argument
[all...]
H A Dhugetlb_cgroup.h25 * Minimum page order trackable by hugetlb cgroup.
27 * The second tail page (hpage[2]) is the fault usage cgroup.
28 * The third tail page (hpage[3]) is the reservation usage cgroup.
62 __hugetlb_cgroup_from_page(struct page *page, bool rsvd) in __hugetlb_cgroup_from_page() argument
64 VM_BUG_ON_PAGE(!PageHuge(page), page); in __hugetlb_cgroup_from_page()
66 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in __hugetlb_cgroup_from_page()
69 return (struct hugetlb_cgroup *)page[3].private; in __hugetlb_cgroup_from_page()
71 return (struct hugetlb_cgroup *)page[ in __hugetlb_cgroup_from_page()
74 hugetlb_cgroup_from_page(struct page *page) hugetlb_cgroup_from_page() argument
80 hugetlb_cgroup_from_page_rsvd(struct page *page) hugetlb_cgroup_from_page_rsvd() argument
85 __set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg, bool rsvd) __set_hugetlb_cgroup() argument
99 set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup() argument
105 set_hugetlb_cgroup_rsvd(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup_rsvd() argument
168 hugetlb_cgroup_from_page(struct page *page) hugetlb_cgroup_from_page() argument
174 hugetlb_cgroup_from_page_resv(struct page *page) hugetlb_cgroup_from_page_resv() argument
180 hugetlb_cgroup_from_page_rsvd(struct page *page) hugetlb_cgroup_from_page_rsvd() argument
185 set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup() argument
191 set_hugetlb_cgroup_rsvd(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup_rsvd() argument
224 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge() argument
231 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge_rsvd() argument
237 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page() argument
242 hugetlb_cgroup_uncharge_page_rsvd(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page_rsvd() argument
[all...]
H A Dmm.h27 #include <linux/page-flags.h>
100 #include <asm/page.h>
138 * a zero page mapping on a read fault.
141 * related to the physical page in case of virtualization.
154 /* This function must be updated when the size of struct page grows above 80
161 static inline void __mm_zero_struct_page(struct page *page) in __mm_zero_struct_page() argument
163 unsigned long *_pp = (void *)page; in __mm_zero_struct_page()
165 /* Check that struct page is either 56, 64, 72, or 80 bytes */ in __mm_zero_struct_page()
166 BUILD_BUG_ON(sizeof(struct page) in __mm_zero_struct_page()
552 struct page *page; /* ->fault handlers should return a global() member
735 put_page_testzero(struct page *page) put_page_testzero() argument
737 VM_BUG_ON_PAGE(page_ref_count(page) == 0, page); put_page_testzero() local
747 get_page_unless_zero(struct page *page) get_page_unless_zero() argument
836 compound_mapcount(struct page *page) compound_mapcount() argument
848 page_mapcount_reset(struct page *page) page_mapcount_reset() argument
863 page_mapcount(struct page *page) page_mapcount() argument
874 total_mapcount(struct page *page) total_mapcount() argument
878 page_trans_huge_mapcount(struct page *page, int *total_mapcount) page_trans_huge_mapcount() argument
890 struct page *page = virt_to_page(x); virt_to_head_page() local
922 set_compound_page_dtor(struct page *page, enum compound_dtor_id compound_dtor) set_compound_page_dtor() argument
929 destroy_compound_page(struct page *page) destroy_compound_page() argument
935 compound_order(struct page *page) compound_order() argument
942 hpage_pincount_available(struct page *page) hpage_pincount_available() argument
958 compound_pincount(struct page *page) compound_pincount() argument
965 set_compound_order(struct page *page, unsigned int order) set_compound_order() argument
972 compound_nr(struct page *page) compound_nr() argument
980 page_size(struct page *page) page_size() argument
986 page_shift(struct page *page) page_shift() argument
1115 page_zonenum(const struct page *page) page_zonenum() argument
1122 is_zone_device_page(const struct page *page) is_zone_device_page() argument
1129 is_zone_device_page(const struct page *page) is_zone_device_page() argument
1139 page_is_devmap_managed(struct page *page) page_is_devmap_managed() argument
1158 page_is_devmap_managed(struct page *page) page_is_devmap_managed() argument
1163 put_devmap_managed_page(struct page *page) put_devmap_managed_page() argument
1168 is_device_private_page(const struct page *page) is_device_private_page() argument
1176 is_pci_p2pdma_page(const struct page *page) is_pci_p2pdma_page() argument
1188 get_page(struct page *page) get_page() argument
1195 VM_BUG_ON_PAGE(page_ref_zero_or_close_to_overflow(page), page); get_page() local
1201 try_get_page(struct page *page) try_get_page() argument
1210 put_page(struct page *page) put_page() argument
1293 page_maybe_dma_pinned(struct page *page) page_maybe_dma_pinned() argument
1322 page_zone_id(struct page *page) page_zone_id() argument
1330 page_to_nid(const struct page *page) page_to_nid() argument
1376 page_cpupid_xchg_last(struct page *page, int cpupid) page_cpupid_xchg_last() argument
1381 page_cpupid_last(struct page *page) page_cpupid_last() argument
1385 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument
1390 page_cpupid_last(struct page *page) page_cpupid_last() argument
1397 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument
1403 page_cpupid_xchg_last(struct page *page, int cpupid) page_cpupid_xchg_last() argument
1408 page_cpupid_last(struct page *page) page_cpupid_last() argument
1438 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument
1456 page_kasan_tag(const struct page *page) page_kasan_tag() argument
1466 page_kasan_tag_set(struct page *page, u8 tag) page_kasan_tag_set() argument
1473 page_kasan_tag_reset(struct page *page) page_kasan_tag_reset() argument
1478 page_kasan_tag(const struct page *page) page_kasan_tag() argument
1483 page_kasan_tag_set(struct page *page, u8 tag) page_kasan_tag_set() argument
1484 page_kasan_tag_reset(struct page *page) page_kasan_tag_reset() argument
1487 page_zone(const struct page *page) page_zone() argument
1492 page_pgdat(const struct page *page) page_pgdat() argument
1498 set_page_section(struct page *page, unsigned long section) set_page_section() argument
1504 page_to_section(const struct page *page) page_to_section() argument
1510 set_page_zone(struct page *page, enum zone_type zone) set_page_zone() argument
1516 set_page_node(struct page *page, unsigned long node) set_page_node() argument
1522 set_page_links(struct page *page, enum zone_type zone, unsigned long node, unsigned long pfn) set_page_links() argument
1533 page_memcg(struct page *page) page_memcg() argument
1537 page_memcg_rcu(struct page *page) page_memcg_rcu() argument
1543 page_memcg(struct page *page) page_memcg() argument
1547 page_memcg_rcu(struct page *page) page_memcg_rcu() argument
1559 lowmem_page_address(const struct page *page) lowmem_page_address() argument
1569 page_address(const struct page *page) page_address() argument
1573 set_page_address(struct page *page, void *address) set_page_address() argument
1599 page_file_mapping(struct page *page) page_file_mapping() argument
1613 page_index(struct page *page) page_index() argument
1629 page_is_pfmemalloc(struct page *page) page_is_pfmemalloc() argument
1642 set_page_pfmemalloc(struct page *page) set_page_pfmemalloc() argument
1647 clear_page_pfmemalloc(struct page *page) clear_page_pfmemalloc() argument
1752 unmap_mapping_page(struct page *page) unmap_mapping_page() argument
1867 cancel_dirty_page(struct page *page) cancel_dirty_page() argument
1972 mm_counter_file(struct page *page) mm_counter_file() argument
1979 mm_counter(struct page *page) mm_counter() argument
2199 ptlock_ptr(struct page *page) ptlock_ptr() argument
2208 ptlock_alloc(struct page *page) ptlock_alloc() argument
2213 ptlock_free(struct page *page) ptlock_free() argument
2217 ptlock_ptr(struct page *page) ptlock_ptr() argument
2228 ptlock_init(struct page *page) ptlock_init() argument
2253 ptlock_init(struct page *page) ptlock_init() argument
2254 ptlock_free(struct page *page) ptlock_free() argument
2263 pgtable_pte_page_ctor(struct page *page) pgtable_pte_page_ctor() argument
2272 pgtable_pte_page_dtor(struct page *page) pgtable_pte_page_dtor() argument
2319 pmd_ptlock_init(struct page *page) pmd_ptlock_init() argument
2327 pmd_ptlock_free(struct page *page) pmd_ptlock_free() argument
2344 pmd_ptlock_init(struct page *page) pmd_ptlock_init() argument
2345 pmd_ptlock_free(struct page *page) pmd_ptlock_free() argument
2358 pgtable_pmd_page_ctor(struct page *page) pgtable_pmd_page_ctor() argument
2367 pgtable_pmd_page_dtor(struct page *page) pgtable_pmd_page_dtor() argument
2420 __free_reserved_page(struct page *page) __free_reserved_page() argument
2427 free_reserved_page(struct page *page) free_reserved_page() argument
2433 mark_page_reserved(struct page *page) mark_page_reserved() argument
2814 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vmf_insert_page() argument
2947 kernel_poison_pages(struct page *page, int numpages, int enable) kernel_poison_pages() argument
3009 kernel_map_pages(struct page *page, int numpages, int enable) kernel_map_pages() argument
3018 kernel_map_pages(struct page *page, int numpages, int enable) kernel_map_pages() argument
3020 kernel_page_present(struct page *page) kernel_page_present() argument
3190 page_is_guard(struct page *page) page_is_guard() argument
3200 page_is_guard(struct page *page) page_is_guard() argument
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dpage_ref.h7 #include <linux/page-flags.h>
29 extern void __page_ref_set(struct page *page, int v);
30 extern void __page_ref_mod(struct page *page, int v);
31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret);
32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret);
33 extern void __page_ref_mod_unless(struct page *pag
41 __page_ref_set(struct page *page, int v) __page_ref_set() argument
44 __page_ref_mod(struct page *page, int v) __page_ref_mod() argument
47 __page_ref_mod_and_test(struct page *page, int v, int ret) __page_ref_mod_and_test() argument
50 __page_ref_mod_and_return(struct page *page, int v, int ret) __page_ref_mod_and_return() argument
53 __page_ref_mod_unless(struct page *page, int v, int u) __page_ref_mod_unless() argument
56 __page_ref_freeze(struct page *page, int v, int ret) __page_ref_freeze() argument
59 __page_ref_unfreeze(struct page *page, int v) __page_ref_unfreeze() argument
65 page_ref_count(const struct page *page) page_ref_count() argument
92 page_count(const struct page *page) page_count() argument
97 set_page_count(struct page *page, int v) set_page_count() argument
113 init_page_count(struct page *page) init_page_count() argument
118 page_ref_add(struct page *page, int nr) page_ref_add() argument
130 page_ref_sub(struct page *page, int nr) page_ref_sub() argument
142 page_ref_sub_return(struct page *page, int nr) page_ref_sub_return() argument
156 page_ref_inc(struct page *page) page_ref_inc() argument
168 page_ref_dec(struct page *page) page_ref_dec() argument
180 page_ref_sub_and_test(struct page *page, int nr) page_ref_sub_and_test() argument
194 page_ref_inc_return(struct page *page) page_ref_inc_return() argument
208 page_ref_dec_and_test(struct page *page) page_ref_dec_and_test() argument
222 page_ref_dec_return(struct page *page) page_ref_dec_return() argument
236 page_ref_add_unless(struct page *page, int nr, int u) page_ref_add_unless() argument
316 page_ref_freeze(struct page *page, int count) page_ref_freeze() argument
330 page_ref_unfreeze(struct page *page, int count) page_ref_unfreeze() argument
332 VM_BUG_ON_PAGE(page_count(page) != 0, page); page_ref_unfreeze() local
[all...]
H A Dpage-flags.h3 * Macros for manipulating and testing page->flags
18 * Various page->flags bits:
20 * PG_reserved is set for special pages. The "struct page" of such a page
25 * - Pages reserved or allocated early during boot (before the page allocator
27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much
29 * be given to the page allocator.
32 * - The zero page(s)
33 * - Pages not added to the page allocator when onlining a section because
49 * Consequently, PG_reserved for a page mappe
215 page_fixed_fake_head(const struct page *page) page_fixed_fake_head() argument
241 page_fixed_fake_head(const struct page *page) page_fixed_fake_head() argument
247 page_is_fake_head(struct page *page) page_is_fake_head() argument
252 _compound_head(const struct page *page) _compound_head() argument
291 PageTail(struct page *page) PageTail() argument
296 PageCompound(struct page *page) PageCompound() argument
303 PagePoisoned(const struct page *page) PagePoisoned() argument
311 page_init_poison(struct page *page, size_t size) page_init_poison() argument
318 struct page *page = &folio->page; folio_flags() local
320 VM_BUG_ON_PGFLAGS(PageTail(page), page); folio_flags() local
556 PageSwapCache(struct page *page) PageSwapCache() argument
666 PageMappingFlags(struct page *page) PageMappingFlags() argument
676 PageAnon(struct page *page) PageAnon() argument
687 __PageMovable(struct page *page) __PageMovable() argument
706 PageKsm(struct page *page) PageKsm() argument
743 PageUptodate(struct page *page) PageUptodate() argument
765 __SetPageUptodate(struct page *page) __SetPageUptodate() argument
770 SetPageUptodate(struct page *page) SetPageUptodate() argument
785 test_set_page_writeback(struct page *page) test_set_page_writeback() argument
795 PageHead(struct page *page) PageHead() argument
816 set_compound_head(struct page *page, struct page *head) set_compound_head() argument
821 clear_compound_head(struct page *page) clear_compound_head() argument
827 ClearPageCompound(struct page *page) ClearPageCompound() argument
871 PageTransHuge(struct page *page) PageTransHuge() argument
873 VM_BUG_ON_PAGE(PageTail(page), page); PageTransHuge() local
882 PageTransCompound(struct page *page) PageTransCompound() argument
892 PageTransTail(struct page *page) PageTransTail() argument
922 is_page_hwpoison(struct page *page) is_page_hwpoison() argument
956 page_has_type(struct page *page) page_has_type() argument
1040 PageAnonExclusive(struct page *page) PageAnonExclusive() argument
1043 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); PageAnonExclusive() local
1047 SetPageAnonExclusive(struct page *page) SetPageAnonExclusive() argument
1050 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); SetPageAnonExclusive() local
1054 ClearPageAnonExclusive(struct page *page) ClearPageAnonExclusive() argument
1057 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); ClearPageAnonExclusive() local
1061 __ClearPageAnonExclusive(struct page *page) __ClearPageAnonExclusive() argument
1064 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); __ClearPageAnonExclusive() local
1120 page_has_private(struct page *page) page_has_private() argument
[all...]
H A Dhighmem-internal.h10 void *__kmap_local_page_prot(struct page *page, pgprot_t prot);
35 void *kmap_high(struct page *page);
36 void kunmap_high(struct page *page);
38 struct page *__kmap_to_page(void *addr);
40 static inline void *kmap(struct page *page) in kmap() argument
45 if (!PageHighMem(page)) in kmap()
53 kunmap(struct page *page) kunmap() argument
71 kmap_local_page(struct page *page) kmap_local_page() argument
78 struct page *page = folio_page(folio, offset / PAGE_SIZE); kmap_local_folio() local
82 kmap_local_page_prot(struct page *page, pgprot_t prot) kmap_local_page_prot() argument
97 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
108 kmap_atomic(struct page *page) kmap_atomic() argument
167 kmap(struct page *page) kmap() argument
173 kunmap_high(struct page *page) kunmap_high() argument
176 kunmap(struct page *page) kunmap() argument
183 kmap_local_page(struct page *page) kmap_local_page() argument
193 kmap_local_page_prot(struct page *page, pgprot_t prot) kmap_local_page_prot() argument
210 kmap_atomic(struct page *page) kmap_atomic() argument
220 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
[all...]
H A Dpage_owner.h11 extern void __reset_page_owner(struct page *page, unsigned short order);
12 extern void __set_page_owner(struct page *page,
14 extern void __split_page_owner(struct page *page, unsigned int nr);
16 extern void __set_page_owner_migrate_reason(struct page *page, int reason);
17 extern void __dump_page_owner(const struct page *page);
21 reset_page_owner(struct page *page, unsigned short order) reset_page_owner() argument
27 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) set_page_owner() argument
34 split_page_owner(struct page *page, unsigned int nr) split_page_owner() argument
44 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument
49 dump_page_owner(const struct page *page) dump_page_owner() argument
55 reset_page_owner(struct page *page, unsigned short order) reset_page_owner() argument
58 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument
62 split_page_owner(struct page *page, unsigned short order) split_page_owner() argument
69 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument
72 dump_page_owner(const struct page *page) dump_page_owner() argument
[all...]
/kernel/linux/linux-5.10/mm/
H A Dswap.c45 /* How many pages do we try to swap or page in/out together? */
79 static void __page_cache_release(struct page *page) in __page_cache_release() argument
81 if (PageLRU(page)) { in __page_cache_release()
82 pg_data_t *pgdat = page_pgdat(page); in __page_cache_release()
87 lruvec = mem_cgroup_page_lruvec(page, pgdat); in __page_cache_release()
88 VM_BUG_ON_PAGE(!PageLRU(page), page); in __page_cache_release()
89 __ClearPageLRU(page); in __page_cache_release()
90 del_page_from_lru_list(page, lruve in __page_cache_release()
96 __put_single_page(struct page *page) __put_single_page() argument
103 __put_compound_page(struct page *page) __put_compound_page() argument
116 __put_page(struct page *page) __put_page() argument
206 pagevec_lru_move_fn(struct pagevec *pvec, void (*move_fn)(struct page *page, struct lruvec *lruvec, void *arg), void *arg) pagevec_lru_move_fn() argument
216 struct page *page = pvec->pages[i]; pagevec_lru_move_fn() local
235 pagevec_move_tail_fn(struct page *page, struct lruvec *lruvec, void *arg) pagevec_move_tail_fn() argument
265 rotate_reclaimable_page(struct page *page) rotate_reclaimable_page() argument
312 lru_note_cost_page(struct page *page) lru_note_cost_page() argument
324 __activate_page(struct page *page, struct lruvec *lruvec, void *arg) __activate_page() argument
357 activate_page(struct page *page) activate_page() argument
377 activate_page(struct page *page) activate_page() argument
388 __lru_cache_activate_page(struct page *page) __lru_cache_activate_page() argument
428 mark_page_accessed(struct page *page) mark_page_accessed() argument
468 lru_cache_add(struct page *page) lru_cache_add() argument
473 VM_BUG_ON_PAGE(PageLRU(page), page); lru_cache_add() local
492 lru_cache_add_inactive_or_unevictable(struct page *page, struct vm_area_struct *vma) lru_cache_add_inactive_or_unevictable() argument
497 VM_BUG_ON_PAGE(PageLRU(page), page); lru_cache_add_inactive_or_unevictable() local
534 lru_deactivate_file_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_deactivate_file_fn() argument
582 lru_deactivate_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_deactivate_fn() argument
600 lru_lazyfree_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_lazyfree_fn() argument
672 deactivate_file_page(struct page *page) deactivate_file_page() argument
701 deactivate_page(struct page *page) deactivate_page() argument
722 mark_page_lazyfree(struct page *page) mark_page_lazyfree() argument
886 struct page *page = pages[i]; release_pages() local
985 lru_add_page_tail(struct page *page, struct page *page_tail, struct lruvec *lruvec, struct list_head *list) lru_add_page_tail() argument
989 VM_BUG_ON_PAGE(PageCompound(page_tail), page); lru_add_page_tail() local
990 VM_BUG_ON_PAGE(PageLRU(page_tail), page); lru_add_page_tail() local
1016 __pagevec_lru_add_fn(struct page *page, struct lruvec *lruvec, void *arg) __pagevec_lru_add_fn() argument
1023 VM_BUG_ON_PAGE(PageLRU(page), page); __pagevec_lru_add_fn() local
1127 struct page *page = pvec->pages[i]; pagevec_remove_exceptionals() local
1201 put_devmap_managed_page(struct page *page) put_devmap_managed_page() argument
[all...]
H A Dfilemap.c63 * finished 'unifying' the page and buffer cache and SMP-threaded the
64 * page-cache, 21.05.1999, Ingo Molnar <mingo@redhat.com>
121 struct page *page, void *shadow) in page_cache_delete()
123 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete()
129 if (!PageHuge(page)) { in page_cache_delete()
130 xas_set_order(&xas, page->index, compound_order(page)); in page_cache_delete()
131 nr = compound_nr(page); in page_cache_delete()
134 VM_BUG_ON_PAGE(!PageLocked(page), pag in page_cache_delete()
120 page_cache_delete(struct address_space *mapping, struct page *page, void *shadow) page_cache_delete() argument
135 VM_BUG_ON_PAGE(PageTail(page), page); page_cache_delete() local
157 unaccount_page_cache_page(struct address_space *mapping, struct page *page) unaccount_page_cache_page() argument
172 VM_BUG_ON_PAGE(PageTail(page), page); unaccount_page_cache_page() local
173 VM_BUG_ON_PAGE(page_mapped(page), page); unaccount_page_cache_page() local
232 __delete_from_page_cache(struct page *page, void *shadow) __delete_from_page_cache() argument
242 page_cache_free_page(struct address_space *mapping, struct page *page) page_cache_free_page() argument
253 VM_BUG_ON_PAGE(page_count(page) <= 0, page); page_cache_free_page() local
267 delete_from_page_cache(struct page *page) delete_from_page_cache() argument
301 struct page *page; page_cache_delete_batch() local
476 struct page *page; filemap_range_has_page() local
525 struct page *page = pvec.pages[i]; __filemap_fdatawait_range() local
830 __add_to_page_cache_locked(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp, void **shadowp) __add_to_page_cache_locked() argument
841 VM_BUG_ON_PAGE(PageSwapBacked(page), page); __add_to_page_cache_locked() local
928 add_to_page_cache_locked(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache_locked() argument
936 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache_lru() argument
969 struct page *page; __page_cache_alloc() local
1000 page_waitqueue(struct page *page) page_waitqueue() argument
1100 wake_up_page_bit(struct page *page, int bit_nr) wake_up_page_bit() argument
1154 wake_up_page(struct page *page, int bit) wake_up_page() argument
1180 trylock_page_bit_common(struct page *page, int bit_nr, struct wait_queue_entry *wait) trylock_page_bit_common() argument
1196 wait_on_page_bit_common(wait_queue_head_t *q, struct page *page, int bit_nr, int state, enum behavior behavior) wait_on_page_bit_common() argument
1335 wait_on_page_bit(struct page *page, int bit_nr) wait_on_page_bit() argument
1342 wait_on_page_bit_killable(struct page *page, int bit_nr) wait_on_page_bit_killable() argument
1349 __wait_on_page_locked_async(struct page *page, struct wait_page_queue *wait, bool set) __wait_on_page_locked_async() argument
1379 wait_on_page_locked_async(struct page *page, struct wait_page_queue *wait) wait_on_page_locked_async() argument
1397 put_and_wait_on_page_locked(struct page *page) put_and_wait_on_page_locked() argument
1413 add_page_wait_queue(struct page *page, wait_queue_entry_t *waiter) add_page_wait_queue() argument
1463 unlock_page(struct page *page) unlock_page() argument
1477 end_page_writeback(struct page *page) end_page_writeback() argument
1511 page_endio(struct page *page, bool is_write, int err) page_endio() argument
1541 struct page *page = compound_head(__page); __lock_page() local
1550 struct page *page = compound_head(__page); __lock_page_killable() local
1557 __lock_page_async(struct page *page, struct wait_page_queue *wait) __lock_page_async() argument
1573 __lock_page_or_retry(struct page *page, struct mm_struct *mm, unsigned int flags) __lock_page_or_retry() argument
1693 struct page *page; find_get_entry() local
1742 struct page *page; find_lock_entry() local
1794 struct page *page; pagecache_get_page() local
1903 struct page *page; find_get_entries() local
1977 struct page *page; find_get_pages_range() local
2042 struct page *page; find_get_pages_contig() local
2099 struct page *page; find_get_pages_range_tag() local
2226 struct page *page; generic_file_buffered_read() local
2579 lock_page_maybe_drop_mmap(struct vm_fault *vmf, struct page *page, struct file **fpin) lock_page_maybe_drop_mmap() argument
2669 do_async_mmap_readahead(struct vm_fault *vmf, struct page *page) do_async_mmap_readahead() argument
2726 struct page *page; filemap_fault() local
2769 VM_BUG_ON_PAGE(page_to_pgoff(page) != offset, page); filemap_fault() local
2849 struct page *head, *page; filemap_map_pages() local
2914 struct page *page = vmf->page; filemap_page_mkwrite() local
2985 wait_on_page_read(struct page *page) wait_on_page_read() argument
3003 struct page *page; do_read_cache_page() local
3160 pagecache_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) pagecache_write_end() argument
3280 struct page *page; grab_cache_page_write_begin() local
3305 struct page *page; generic_perform_write() local
3523 try_to_release_page(struct page *page, gfp_t gfp_mask) try_to_release_page() argument
[all...]
H A Dmigrate.c86 int isolate_movable_page(struct page *page, isolate_mode_t mode) in isolate_movable_page() argument
94 * In case we 'win' a race for a movable page being freed under us and in isolate_movable_page()
97 * release this page, thus avoiding a nasty leakage. in isolate_movable_page()
99 if (unlikely(!get_page_unless_zero(page))) in isolate_movable_page()
103 * Check PageMovable before holding a PG_lock because page's owner in isolate_movable_page()
104 * assumes anybody doesn't touch PG_lock of newly allocated page in isolate_movable_page()
105 * so unconditionally grabbing the lock ruins page's owner side. in isolate_movable_page()
107 if (unlikely(!__PageMovable(page))) in isolate_movable_page()
111 * compaction threads can race against page migratio in isolate_movable_page()
148 putback_movable_page(struct page *page) putback_movable_page() argument
171 struct page *page; putback_movable_pages() local
205 remove_migration_pte(struct page *page, struct vm_area_struct *vma, unsigned long addr, void *old) remove_migration_pte() argument
218 VM_BUG_ON_PAGE(PageTail(page), page); remove_migration_pte() local
317 struct page *page; __migration_entry_wait() local
364 struct page *page; pmd_migration_entry_wait() local
380 expected_page_refs(struct address_space *mapping, struct page *page) expected_page_refs() argument
403 migrate_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page, int extra_count) migrate_page_move_mapping() argument
454 VM_BUG_ON_PAGE(PageSwapCache(page), page); migrate_page_move_mapping() local
525 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument
608 migrate_page_states(struct page *newpage, struct page *page) migrate_page_states() argument
619 VM_BUG_ON_PAGE(PageUnevictable(page), page); migrate_page_states() local
685 migrate_page_copy(struct page *newpage, struct page *page) migrate_page_copy() argument
700 migrate_page_extra(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, int extra_count) migrate_page_extra() argument
726 migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) migrate_page() argument
773 __buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, bool check_refs) __buffer_migrate_page() argument
857 buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page() argument
870 buffer_migrate_page_norefs(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page_norefs() argument
880 writeout(struct address_space *mapping, struct page *page) writeout() argument
921 fallback_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) fallback_migrate_page() argument
958 move_to_new_page(struct page *newpage, struct page *page, enum migrate_mode mode) move_to_new_page() argument
1038 __unmap_and_move(struct page *page, struct page *newpage, int force, enum migrate_mode mode) __unmap_and_move() argument
1135 VM_BUG_ON_PAGE(PageAnon(page), page); __unmap_and_move() local
1186 unmap_and_move(new_page_t get_new_page, free_page_t put_new_page, unsigned long private, struct page *page, int force, enum migrate_mode mode, enum migrate_reason reason) unmap_and_move() argument
1446 struct page *page; migrate_pages() local
1557 alloc_migration_target(struct page *page, unsigned long private) alloc_migration_target() argument
1642 struct page *page; add_page_for_migration() local
1830 struct page *page; do_pages_stat_array() local
2027 alloc_misplaced_dst_page(struct page *page, unsigned long data) alloc_misplaced_dst_page() argument
2042 numamigrate_isolate_page(pg_data_t *pgdat, struct page *page) numamigrate_isolate_page() argument
2082 struct page *page = pmd_page(pmd); pmd_trans_migrating() local
2091 migrate_misplaced_page(struct page *page, struct vm_area_struct *vma, int node) migrate_misplaced_page() argument
2146 migrate_misplaced_transhuge_page(struct mm_struct *mm, struct vm_area_struct *vma, pmd_t *pmd, pmd_t entry, unsigned long address, struct page *page, int node) migrate_misplaced_transhuge_page() argument
2338 struct page *page; migrate_vma_collect_pmd() local
2381 struct page *page; migrate_vma_collect_pmd() local
2544 migrate_vma_check_page(struct page *page, struct page *fault_page) migrate_vma_check_page() argument
2608 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_prepare() local
2683 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_prepare() local
2716 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_unmap() local
2737 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_unmap() local
2869 migrate_vma_insert_page(struct migrate_vma *migrate, unsigned long addr, struct page *page, unsigned long *src, unsigned long *dst) migrate_vma_insert_page() argument
3019 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_pages() local
3107 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_finalize() local
[all...]
H A Drmap.c10 * Provides methods for unmapping each kind of mapped page:
25 * page->flags PG_locked (lock_page) * (see huegtlbfs below)
28 * hugetlb_fault_mutex (hugetlbfs specific page fault mutex)
50 * hugetlb_fault_mutex (hugetlbfs specific page fault mutex)
51 * page->flags PG_locked (lock_page)
274 * searches where page is mapped.
466 * Getting a lock on a stable anon_vma from a page off the LRU is tricky!
470 * have been relevant to this page.
472 * The page might have been remapped to a different anon_vma or the anon_vma
477 * ensure that any anon_vma obtained from the page wil
489 page_get_anon_vma(struct page *page) page_get_anon_vma() argument
532 page_lock_anon_vma_read(struct page *page) page_lock_anon_vma_read() argument
712 page_address_in_vma(struct page *page, struct vm_area_struct *vma) page_address_in_vma() argument
775 page_referenced_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) page_referenced_one() argument
862 page_referenced(struct page *page, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) page_referenced() argument
910 page_mkclean_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) page_mkclean_one() argument
992 page_mkclean(struct page *page) page_mkclean() argument
1027 page_move_anon_rmap(struct page *page, struct vm_area_struct *vma) page_move_anon_rmap() argument
1052 __page_set_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, int exclusive) __page_set_anon_rmap() argument
1081 __page_check_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) __page_check_anon_rmap() argument
1096 VM_BUG_ON_PAGE(page_anon_vma(page)->root != vma->anon_vma->root, page); __page_check_anon_rmap() local
1098 page); __page_check_anon_rmap() local
1113 page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, bool compound) page_add_anon_rmap() argument
1124 do_page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, int flags) do_page_add_anon_rmap() argument
1182 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, bool compound) page_add_new_anon_rmap() argument
1199 VM_BUG_ON_PAGE(PageTransCompound(page), page); page_add_new_anon_rmap() local
1214 page_add_file_rmap(struct page *page, bool compound) page_add_file_rmap() argument
1247 page_remove_file_rmap(struct page *page, bool compound) page_remove_file_rmap() argument
1288 page_remove_anon_compound_rmap(struct page *page) page_remove_anon_compound_rmap() argument
1339 page_remove_rmap(struct page *page, bool compound) page_remove_rmap() argument
1386 try_to_unmap_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) try_to_unmap_one() argument
1775 page_not_mapped(struct page *page) page_not_mapped() argument
1790 try_to_unmap(struct page *page, enum ttu_flags flags) try_to_unmap() argument
1834 try_to_munlock(struct page *page) try_to_munlock() argument
1859 rmap_walk_anon_lock(struct page *page, struct rmap_walk_control *rwc) rmap_walk_anon_lock() argument
1895 rmap_walk_anon(struct page *page, struct rmap_walk_control *rwc, bool locked) rmap_walk_anon() argument
1948 rmap_walk_file(struct page *page, struct rmap_walk_control *rwc, bool locked) rmap_walk_file() argument
1991 rmap_walk(struct page *page, struct rmap_walk_control *rwc) rmap_walk() argument
2002 rmap_walk_locked(struct page *page, struct rmap_walk_control *rwc) rmap_walk_locked() argument
2005 VM_BUG_ON_PAGE(PageKsm(page), page); rmap_walk_locked() local
2018 hugepage_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) hugepage_add_anon_rmap() argument
2032 hugepage_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) hugepage_add_new_anon_rmap() argument
[all...]
H A Dpage_io.c10 * Add swap of shared pages through the page cache. 20.2.1998. Stephen Tweedie
30 struct page *page, bio_end_io_t end_io) in get_swap_bio()
38 bio->bi_iter.bi_sector = map_swap_page(page, &bdev); in get_swap_bio()
43 bio_add_page(bio, page, thp_size(page), 0); in get_swap_bio()
50 struct page *page = bio_first_page_all(bio); in end_swap_bio_write() local
53 SetPageError(page); in end_swap_bio_write()
55 * We failed to write the page ou in end_swap_bio_write()
29 get_swap_bio(gfp_t gfp_flags, struct page *page, bio_end_io_t end_io) get_swap_bio() argument
74 struct page *page = bio_first_page_all(bio); end_swap_bio_read() local
198 swap_writepage(struct page *page, struct writeback_control *wbc) swap_writepage() argument
227 count_swpout_vm_event(struct page *page) count_swpout_vm_event() argument
237 bio_associate_blkg_from_page(struct bio *bio, struct page *page) bio_associate_blkg_from_page() argument
253 __swap_writepage(struct page *page, struct writeback_control *wbc, bio_end_io_t end_write_func) __swap_writepage() argument
324 swap_readpage(struct page *page, bool synchronous) swap_readpage() argument
335 VM_BUG_ON_PAGE(PageUptodate(page), page); swap_readpage() local
405 swap_set_page_dirty(struct page *page) swap_set_page_dirty() argument
[all...]
H A Dtruncate.c30 * Regular page slots are stabilized by the page lock even without the tree
83 struct page *page = pvec->pages[i]; in truncate_exceptional_pvec_entries() local
86 if (!xa_is_value(page)) { in truncate_exceptional_pvec_entries()
87 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries()
99 __clear_shadow_entry(mapping, index, page); in truncate_exceptional_pvec_entries()
138 * do_invalidatepage - invalidate part or all of a page
139 * @page: the page whic
152 do_invalidatepage(struct page *page, unsigned int offset, unsigned int length) do_invalidatepage() argument
176 truncate_cleanup_page(struct page *page) truncate_cleanup_page() argument
202 invalidate_complete_page(struct address_space *mapping, struct page *page) invalidate_complete_page() argument
217 truncate_inode_page(struct address_space *mapping, struct page *page) truncate_inode_page() argument
219 VM_BUG_ON_PAGE(PageTail(page), page); truncate_inode_page() local
232 generic_error_remove_page(struct address_space *mapping, struct page *page) generic_error_remove_page() argument
252 invalidate_inode_page(struct page *page) invalidate_inode_page() argument
338 struct page *page = pvec.pages[i]; truncate_inode_pages_range() local
372 struct page *page = find_lock_page(mapping, start - 1); truncate_inode_pages_range() local
391 struct page *page = find_lock_page(mapping, end); truncate_inode_pages_range() local
430 struct page *page = pvec.pages[i]; truncate_inode_pages_range() local
543 struct page *page = pvec.pages[i]; __invalidate_mapping_pages() local
655 invalidate_complete_page2(struct address_space *mapping, struct page *page) invalidate_complete_page2() argument
683 do_launder_page(struct address_space *mapping, struct page *page) do_launder_page() argument
723 struct page *page = pvec.pages[i]; invalidate_inode_pages2_range() local
886 struct page *page; pagecache_isize_extended() local
[all...]
H A Dpage_isolation.c7 #include <linux/page-isolation.h>
18 static int set_migratetype_isolate(struct page *page, int migratetype, int isol_flags) in set_migratetype_isolate() argument
20 struct zone *zone = page_zone(page); in set_migratetype_isolate()
21 struct page *unmovable; in set_migratetype_isolate()
31 if (is_migrate_isolate_page(page)) { in set_migratetype_isolate()
40 unmovable = has_unmovable_pages(zone, page, migratetype, isol_flags); in set_migratetype_isolate()
43 int mt = get_pageblock_migratetype(page); in set_migratetype_isolate()
45 set_pageblock_migratetype(page, MIGRATE_ISOLATE); in set_migratetype_isolate()
47 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLAT in set_migratetype_isolate()
68 unset_migratetype_isolate(struct page *page, unsigned migratetype) unset_migratetype_isolate() argument
133 struct page *page; __first_valid_page() local
188 struct page *page; start_isolate_page_range() local
209 struct page *page = pfn_to_online_page(pfn); start_isolate_page_range() local
225 struct page *page; undo_isolate_page_range() local
250 struct page *page; __test_page_isolated_in_pageblock() local
288 struct page *page; test_pages_isolated() local
[all...]
/kernel/linux/linux-6.6/mm/
H A Dfolio-compat.c13 struct address_space *page_mapping(struct page *page) in page_mapping() argument
15 return folio_mapping(page_folio(page)); in page_mapping()
19 void unlock_page(struct page *page) in unlock_page() argument
21 return folio_unlock(page_folio(page)); in unlock_page()
25 void end_page_writeback(struct page *page) in end_page_writeback() argument
27 return folio_end_writeback(page_folio(page)); in end_page_writeback()
31 void wait_on_page_writeback(struct page *pag argument
37 wait_for_stable_page(struct page *page) wait_for_stable_page() argument
43 mark_page_accessed(struct page *page) mark_page_accessed() argument
49 set_page_writeback(struct page *page) set_page_writeback() argument
55 set_page_dirty(struct page *page) set_page_dirty() argument
61 __set_page_dirty_nobuffers(struct page *page) __set_page_dirty_nobuffers() argument
67 clear_page_dirty_for_io(struct page *page) clear_page_dirty_for_io() argument
73 redirty_page_for_writepage(struct writeback_control *wbc, struct page *page) redirty_page_for_writepage() argument
80 lru_cache_add_inactive_or_unevictable(struct page *page, struct vm_area_struct *vma) lru_cache_add_inactive_or_unevictable() argument
86 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t index, gfp_t gfp) add_to_page_cache_lru() argument
114 isolate_lru_page(struct page *page) isolate_lru_page() argument
121 putback_lru_page(struct page *page) putback_lru_page() argument
127 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) page_add_new_anon_rmap() argument
130 VM_BUG_ON_PAGE(PageTail(page), page); page_add_new_anon_rmap() local
[all...]
/kernel/linux/linux-6.6/fs/btrfs/
H A Dsubpage.c14 * - Only support 64K page size for now
15 * This is to make metadata handling easier, as 64K page would ensure
16 * all nodesize would fit inside one page, thus we don't need to handle
22 * - Metadata can't cross 64K page boundary
32 * needed range, other unrelated range in the same page will not be touched.
35 * The writeback is still for the full page, but we will only submit
36 * the dirty extent buffers in the page.
38 * This means, if we have a metadata page like this:
54 * record the status of each sector inside a page. This provides the extra
58 * Since we have multiple tree blocks inside one page, w
67 btrfs_is_subpage(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_is_subpage() argument
118 btrfs_attach_subpage(const struct btrfs_fs_info *fs_info, struct page *page, enum btrfs_subpage_type type) btrfs_attach_subpage() argument
142 btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_detach_subpage() argument
194 btrfs_page_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_inc_eb_refs() argument
209 btrfs_page_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_dec_eb_refs() argument
225 btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_assert() argument
241 btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_start_reader() argument
252 btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_end_reader() argument
276 btrfs_subpage_clamp_range(struct page *page, u64 *start, u32 *len) btrfs_subpage_clamp_range() argument
294 btrfs_subpage_start_writer(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_start_writer() argument
308 btrfs_subpage_end_and_test_writer(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_end_and_test_writer() argument
340 btrfs_page_start_writer_lock(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_start_writer_lock() argument
357 btrfs_page_end_writer_lock(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_end_writer_lock() argument
387 btrfs_subpage_set_uptodate(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_uptodate() argument
402 btrfs_subpage_clear_uptodate(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_uptodate() argument
416 btrfs_subpage_set_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_dirty() argument
440 btrfs_subpage_clear_and_test_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_and_test_dirty() argument
457 btrfs_subpage_clear_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_dirty() argument
467 btrfs_subpage_set_writeback(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_writeback() argument
481 btrfs_subpage_clear_writeback(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_writeback() argument
498 btrfs_subpage_set_ordered(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_ordered() argument
512 btrfs_subpage_clear_ordered(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_ordered() argument
527 btrfs_subpage_set_checked(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_checked() argument
542 btrfs_subpage_clear_checked(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_checked() argument
656 btrfs_page_assert_not_dirty(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_assert_not_dirty() argument
687 btrfs_page_unlock_writer(struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_unlock_writer() argument
719 btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_dump_bitmap() argument
[all...]
H A Dsubpage.h42 * Structure to trace status of each sector inside a page, attached to
43 * page::private for both data and metadata inodes.
50 * page.
51 * Data relies on @readers to unlock the page when last reader finished.
52 * While metadata doesn't need page unlock, it needs to prevent
53 * page::private get cleared before the last end_page_read().
76 bool btrfs_is_subpage(const struct btrfs_fs_info *fs_info, struct page *page);
80 struct page *page, enu
[all...]
/kernel/linux/linux-6.6/arch/arm/include/asm/
H A Dpage.h3 * arch/arm/include/asm/page.h
10 /* PAGE_SHIFT determines the page size */
19 #include <asm/page-nommu.h>
30 * page-based copying and clearing for user space for the particular
107 struct page;
111 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr);
112 void (*cpu_copy_user_highpage)(struct page *to, struct page *from,
116 void fa_copy_user_highpage(struct page *t
[all...]

Completed in 19 milliseconds

12345678910>>...254