/kernel/linux/linux-5.10/include/linux/ |
H A D | page_ref.h | 7 #include <linux/page-flags.h> 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *pag 41 __page_ref_set(struct page *page, int v) __page_ref_set() argument 44 __page_ref_mod(struct page *page, int v) __page_ref_mod() argument 47 __page_ref_mod_and_test(struct page *page, int v, int ret) __page_ref_mod_and_test() argument 50 __page_ref_mod_and_return(struct page *page, int v, int ret) __page_ref_mod_and_return() argument 53 __page_ref_mod_unless(struct page *page, int v, int u) __page_ref_mod_unless() argument 56 __page_ref_freeze(struct page *page, int v, int ret) __page_ref_freeze() argument 59 __page_ref_unfreeze(struct page *page, int v) __page_ref_unfreeze() argument 65 page_ref_count(struct page *page) page_ref_count() argument 70 page_count(struct page *page) page_count() argument 75 set_page_count(struct page *page, int v) set_page_count() argument 86 init_page_count(struct page *page) init_page_count() argument 91 page_ref_add(struct page *page, int nr) page_ref_add() argument 98 page_ref_sub(struct page *page, int nr) page_ref_sub() argument 105 page_ref_sub_return(struct page *page, int nr) page_ref_sub_return() argument 114 page_ref_inc(struct page *page) page_ref_inc() argument 121 page_ref_dec(struct page *page) page_ref_dec() argument 128 page_ref_sub_and_test(struct page *page, int nr) page_ref_sub_and_test() argument 137 page_ref_inc_return(struct page *page) page_ref_inc_return() argument 146 page_ref_dec_and_test(struct page *page) page_ref_dec_and_test() argument 155 page_ref_dec_return(struct page *page) page_ref_dec_return() argument 164 page_ref_add_unless(struct page *page, int nr, int u) page_ref_add_unless() argument 173 page_ref_freeze(struct page *page, int count) page_ref_freeze() argument 182 page_ref_unfreeze(struct page *page, int count) page_ref_unfreeze() argument 184 VM_BUG_ON_PAGE(page_count(page) != 0, page); page_ref_unfreeze() local [all...] |
H A D | page-flags.h | 3 * Macros for manipulating and testing page->flags 18 * Various page->flags bits: 20 * PG_reserved is set for special pages. The "struct page" of such a page 25 * - Pages reserved or allocated early during boot (before the page allocator 27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much 29 * be given to the page allocator. 32 * - The zero page(s) 33 * - Pages not added to the page allocator when onlining a section because 49 * Consequently, PG_reserved for a page mappe 194 compound_head(struct page *page) compound_head() argument 203 PageTail(struct page *page) PageTail() argument 208 PageCompound(struct page *page) PageCompound() argument 214 PagePoisoned(const struct page *page) PagePoisoned() argument 222 page_init_poison(struct page *page, size_t size) page_init_poison() argument 416 PageSwapCache(struct page *page) PageSwapCache() argument 506 PageMappingFlags(struct page *page) PageMappingFlags() argument 511 PageAnon(struct page *page) PageAnon() argument 517 __PageMovable(struct page *page) __PageMovable() argument 530 PageKsm(struct page *page) PageKsm() argument 542 PageUptodate(struct page *page) PageUptodate() argument 561 __SetPageUptodate(struct page *page) __SetPageUptodate() argument 563 VM_BUG_ON_PAGE(PageTail(page), page); __SetPageUptodate() local 568 SetPageUptodate(struct page *page) SetPageUptodate() argument 570 VM_BUG_ON_PAGE(PageTail(page), page); SetPageUptodate() local 590 set_page_writeback(struct page *page) set_page_writeback() argument 595 set_page_writeback_keepwrite(struct page *page) set_page_writeback_keepwrite() argument 602 set_compound_head(struct page *page, struct page *head) set_compound_head() argument 607 clear_compound_head(struct page *page) clear_compound_head() argument 613 ClearPageCompound(struct page *page) ClearPageCompound() argument 630 page_huge_active(struct page *page) page_huge_active() argument 646 PageTransHuge(struct page *page) PageTransHuge() argument 648 VM_BUG_ON_PAGE(PageTail(page), page); PageTransHuge() local 657 PageTransCompound(struct page *page) PageTransCompound() argument 683 PageTransCompoundMap(struct page *page) PageTransCompoundMap() argument 704 PageTransTail(struct page *page) PageTransTail() argument 754 page_has_type(struct page *page) page_has_type() argument 824 PageSlabPfmemalloc(struct page *page) PageSlabPfmemalloc() argument 830 SetPageSlabPfmemalloc(struct page *page) SetPageSlabPfmemalloc() argument 836 __ClearPageSlabPfmemalloc(struct page *page) __ClearPageSlabPfmemalloc() argument 842 ClearPageSlabPfmemalloc(struct page *page) ClearPageSlabPfmemalloc() argument 892 page_has_private(struct page *page) page_has_private() argument [all...] |
H A D | page_idle.h | 6 #include <linux/page-flags.h> 12 static inline bool page_is_young(struct page *page) in page_is_young() argument 14 return PageYoung(page); in page_is_young() 17 static inline void set_page_young(struct page *page) in set_page_young() argument 19 SetPageYoung(page); in set_page_young() 22 static inline bool test_and_clear_page_young(struct page *page) in test_and_clear_page_young() argument 24 return TestClearPageYoung(page); in test_and_clear_page_young() 27 page_is_idle(struct page *page) page_is_idle() argument 32 set_page_idle(struct page *page) set_page_idle() argument 37 clear_page_idle(struct page *page) clear_page_idle() argument 48 page_is_young(struct page *page) page_is_young() argument 58 set_page_young(struct page *page) set_page_young() argument 68 test_and_clear_page_young(struct page *page) test_and_clear_page_young() argument 78 page_is_idle(struct page *page) page_is_idle() argument 88 set_page_idle(struct page *page) set_page_idle() argument 98 clear_page_idle(struct page *page) clear_page_idle() argument 111 page_is_young(struct page *page) page_is_young() argument 116 set_page_young(struct page *page) set_page_young() argument 120 test_and_clear_page_young(struct page *page) test_and_clear_page_young() argument 125 page_is_idle(struct page *page) page_is_idle() argument 130 set_page_idle(struct page *page) set_page_idle() argument 134 clear_page_idle(struct page *page) clear_page_idle() argument [all...] |
H A D | pagemap.h | 158 void release_pages(struct page **pages, int nr); 161 * speculatively take a reference to a page. 162 * If the page is free (_refcount == 0), then _refcount is untouched, and 0 166 * been used to lookup the page in the pagecache radix-tree (or page table): 172 * page has been finished with, no matter what it is subsequently allocated 179 * 1. find page in radix tree 181 * 3. check the page is still in pagecache (if no, goto 1) 186 * B. remove page from pagecache 187 * C. free the page 204 __page_cache_add_speculative(struct page *page, int count) __page_cache_add_speculative() argument 219 VM_BUG_ON_PAGE(page_count(page) == 0, page); __page_cache_add_speculative() local 232 VM_BUG_ON_PAGE(PageTail(page), page); __page_cache_add_speculative() local 237 page_cache_get_speculative(struct page *page) page_cache_get_speculative() argument 242 page_cache_add_speculative(struct page *page, int count) page_cache_add_speculative() argument 255 attach_page_private(struct page *page, void *data) attach_page_private() argument 271 detach_page_private(struct page *page) detach_page_private() argument 507 page_to_index(struct page *page) page_to_index() argument 529 page_to_pgoff(struct page *page) page_to_pgoff() argument 539 page_offset(struct page *page) page_offset() argument 544 page_file_offset(struct page *page) page_file_offset() argument 564 struct page *page; global() member 570 struct page *page; global() member 598 trylock_page(struct page *page) trylock_page() argument 607 lock_page(struct page *page) lock_page() argument 619 lock_page_killable(struct page *page) lock_page_killable() argument 635 lock_page_async(struct page *page, struct wait_page_queue *wait) lock_page_async() argument 650 lock_page_or_retry(struct page *page, struct mm_struct *mm, unsigned int flags) lock_page_or_retry() argument 671 wait_on_page_locked(struct page *page) wait_on_page_locked() argument 677 wait_on_page_locked_killable(struct page *page) wait_on_page_locked_killable() argument 768 add_to_page_cache(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache() argument 857 page_cache_async_readahead(struct address_space *mapping, struct file_ra_state *ra, struct file *file, struct page *page, pgoff_t index, unsigned long req_count) page_cache_async_readahead() argument 876 struct page *page; readahead_page() local 899 struct page *page; __readahead_batch() local 912 VM_BUG_ON_PAGE(PageTail(page), page); __readahead_batch() local 997 page_mkwrite_check_truncate(struct page *page, struct inode *inode) page_mkwrite_check_truncate() argument 1029 i_blocks_per_page(struct inode *inode, struct page *page) i_blocks_per_page() argument [all...] |
H A D | migrate.h | 10 typedef struct page *new_page_t(struct page *page, unsigned long private); 11 typedef void free_page_t(struct page *page, unsigned long private); 17 * - negative errno on page migration failure; 18 * - zero on page migration success; 40 struct page *newpage, struct page *page, 66 alloc_migration_target(struct page *page, unsigned long private) alloc_migration_target() argument 69 isolate_movable_page(struct page *page, isolate_mode_t mode) isolate_movable_page() argument 75 migrate_page_states(struct page *newpage, struct page *page) migrate_page_states() argument 79 migrate_page_copy(struct page *newpage, struct page *page) migrate_page_copy() argument 82 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument 95 PageMovable(struct page *page) PageMovable() argument 96 __SetPageMovable(struct page *page, struct address_space *mapping) __SetPageMovable() argument 100 __ClearPageMovable(struct page *page) __ClearPageMovable() argument 114 migrate_misplaced_page(struct page *page, struct vm_area_struct *vma, int node) migrate_misplaced_page() argument 128 migrate_misplaced_transhuge_page(struct mm_struct *mm, struct vm_area_struct *vma, pmd_t *pmd, pmd_t entry, unsigned long address, struct page *page, int node) migrate_misplaced_transhuge_page() argument [all...] |
H A D | balloon_compaction.h | 7 * Balloon page migration makes use of the general non-lru movable page 10 * page->private is used to reference the responsible balloon device. 11 * page->mapping is used in context of non-lru page migration to reference 12 * the address space operations for page isolation/migration/compaction. 14 * As the page isolation scanning step a compaction thread does is a lockless 15 * procedure (from a page standpoint), it might bring some racy situations while 16 * performing balloon page compaction. In order to sort out these racy scenarios 17 * and safely perform balloon's page compactio 99 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) balloon_page_insert() argument 116 balloon_page_delete(struct page *page) balloon_page_delete() argument 133 balloon_page_device(struct page *page) balloon_page_device() argument 145 balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) balloon_page_insert() argument 152 balloon_page_delete(struct page *page) balloon_page_delete() argument 158 balloon_page_isolate(struct page *page) balloon_page_isolate() argument 163 balloon_page_putback(struct page *page) balloon_page_putback() argument 168 balloon_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) balloon_page_migrate() argument 188 balloon_page_push(struct list_head *pages, struct page *page) balloon_page_push() argument 202 struct page *page = list_first_entry_or_null(pages, struct page, lru); balloon_page_pop() local [all...] |
H A D | highmem.h | 15 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) in flush_anon_page() argument 21 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument 35 extern void *kmap_atomic_high_prot(struct page *page, pgprot_t prot); 47 void *kmap_high(struct page *page); 48 static inline void *kmap(struct page *page) in kmap() argument 63 kunmap(struct page *page) kunmap() argument 84 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 137 kmap(struct page *page) kmap() argument 143 kunmap_high(struct page *page) kunmap_high() argument 147 kunmap(struct page *page) kunmap() argument 154 kmap_atomic(struct page *page) kmap_atomic() argument 307 zero_user_segment(struct page *page, unsigned start, unsigned end) zero_user_segment() argument 313 zero_user(struct page *page, unsigned start, unsigned size) zero_user() argument 350 memcpy_from_page(char *to, struct page *page, size_t offset, size_t len) memcpy_from_page() argument 359 memcpy_to_page(struct page *page, size_t offset, const char *from, size_t len) memcpy_to_page() argument [all...] |
H A D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned int order); 12 extern void __set_page_owner(struct page *page, 14 extern void __split_page_owner(struct page *page, unsigned int nr); 15 extern void __copy_page_owner(struct page *oldpage, struct page *newpage); 16 extern void __set_page_owner_migrate_reason(struct page *page, in 21 reset_page_owner(struct page *page, unsigned int order) reset_page_owner() argument 27 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument 34 split_page_owner(struct page *page, unsigned int nr) split_page_owner() argument 44 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument 49 dump_page_owner(struct page *page) dump_page_owner() argument 55 reset_page_owner(struct page *page, unsigned int order) reset_page_owner() argument 58 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument 62 split_page_owner(struct page *page, unsigned int order) split_page_owner() argument 69 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument 72 dump_page_owner(struct page *page) dump_page_owner() argument [all...] |
H A D | hugetlb_cgroup.h | 25 * Minimum page order trackable by hugetlb cgroup. 27 * The second tail page (hpage[2]) is the fault usage cgroup. 28 * The third tail page (hpage[3]) is the reservation usage cgroup. 62 __hugetlb_cgroup_from_page(struct page *page, bool rsvd) in __hugetlb_cgroup_from_page() argument 64 VM_BUG_ON_PAGE(!PageHuge(page), page); in __hugetlb_cgroup_from_page() 66 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in __hugetlb_cgroup_from_page() 69 return (struct hugetlb_cgroup *)page[3].private; in __hugetlb_cgroup_from_page() 71 return (struct hugetlb_cgroup *)page[ in __hugetlb_cgroup_from_page() 74 hugetlb_cgroup_from_page(struct page *page) hugetlb_cgroup_from_page() argument 80 hugetlb_cgroup_from_page_rsvd(struct page *page) hugetlb_cgroup_from_page_rsvd() argument 85 __set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg, bool rsvd) __set_hugetlb_cgroup() argument 99 set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup() argument 105 set_hugetlb_cgroup_rsvd(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup_rsvd() argument 168 hugetlb_cgroup_from_page(struct page *page) hugetlb_cgroup_from_page() argument 174 hugetlb_cgroup_from_page_resv(struct page *page) hugetlb_cgroup_from_page_resv() argument 180 hugetlb_cgroup_from_page_rsvd(struct page *page) hugetlb_cgroup_from_page_rsvd() argument 185 set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup() argument 191 set_hugetlb_cgroup_rsvd(struct page *page, struct hugetlb_cgroup *h_cg) set_hugetlb_cgroup_rsvd() argument 224 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge() argument 231 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct page *page) hugetlb_cgroup_commit_charge_rsvd() argument 237 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page() argument 242 hugetlb_cgroup_uncharge_page_rsvd(int idx, unsigned long nr_pages, struct page *page) hugetlb_cgroup_uncharge_page_rsvd() argument [all...] |
H A D | mm.h | 27 #include <linux/page-flags.h> 100 #include <asm/page.h> 138 * a zero page mapping on a read fault. 141 * related to the physical page in case of virtualization. 154 /* This function must be updated when the size of struct page grows above 80 161 static inline void __mm_zero_struct_page(struct page *page) in __mm_zero_struct_page() argument 163 unsigned long *_pp = (void *)page; in __mm_zero_struct_page() 165 /* Check that struct page is either 56, 64, 72, or 80 bytes */ in __mm_zero_struct_page() 166 BUILD_BUG_ON(sizeof(struct page) in __mm_zero_struct_page() 552 struct page *page; /* ->fault handlers should return a global() member 735 put_page_testzero(struct page *page) put_page_testzero() argument 737 VM_BUG_ON_PAGE(page_ref_count(page) == 0, page); put_page_testzero() local 747 get_page_unless_zero(struct page *page) get_page_unless_zero() argument 836 compound_mapcount(struct page *page) compound_mapcount() argument 848 page_mapcount_reset(struct page *page) page_mapcount_reset() argument 863 page_mapcount(struct page *page) page_mapcount() argument 874 total_mapcount(struct page *page) total_mapcount() argument 878 page_trans_huge_mapcount(struct page *page, int *total_mapcount) page_trans_huge_mapcount() argument 890 struct page *page = virt_to_page(x); virt_to_head_page() local 922 set_compound_page_dtor(struct page *page, enum compound_dtor_id compound_dtor) set_compound_page_dtor() argument 929 destroy_compound_page(struct page *page) destroy_compound_page() argument 935 compound_order(struct page *page) compound_order() argument 942 hpage_pincount_available(struct page *page) hpage_pincount_available() argument 958 compound_pincount(struct page *page) compound_pincount() argument 965 set_compound_order(struct page *page, unsigned int order) set_compound_order() argument 972 compound_nr(struct page *page) compound_nr() argument 980 page_size(struct page *page) page_size() argument 986 page_shift(struct page *page) page_shift() argument 1115 page_zonenum(const struct page *page) page_zonenum() argument 1122 is_zone_device_page(const struct page *page) is_zone_device_page() argument 1129 is_zone_device_page(const struct page *page) is_zone_device_page() argument 1139 page_is_devmap_managed(struct page *page) page_is_devmap_managed() argument 1158 page_is_devmap_managed(struct page *page) page_is_devmap_managed() argument 1163 put_devmap_managed_page(struct page *page) put_devmap_managed_page() argument 1168 is_device_private_page(const struct page *page) is_device_private_page() argument 1176 is_pci_p2pdma_page(const struct page *page) is_pci_p2pdma_page() argument 1188 get_page(struct page *page) get_page() argument 1195 VM_BUG_ON_PAGE(page_ref_zero_or_close_to_overflow(page), page); get_page() local 1201 try_get_page(struct page *page) try_get_page() argument 1210 put_page(struct page *page) put_page() argument 1293 page_maybe_dma_pinned(struct page *page) page_maybe_dma_pinned() argument 1322 page_zone_id(struct page *page) page_zone_id() argument 1330 page_to_nid(const struct page *page) page_to_nid() argument 1376 page_cpupid_xchg_last(struct page *page, int cpupid) page_cpupid_xchg_last() argument 1381 page_cpupid_last(struct page *page) page_cpupid_last() argument 1385 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument 1390 page_cpupid_last(struct page *page) page_cpupid_last() argument 1397 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument 1403 page_cpupid_xchg_last(struct page *page, int cpupid) page_cpupid_xchg_last() argument 1408 page_cpupid_last(struct page *page) page_cpupid_last() argument 1438 page_cpupid_reset_last(struct page *page) page_cpupid_reset_last() argument 1456 page_kasan_tag(const struct page *page) page_kasan_tag() argument 1466 page_kasan_tag_set(struct page *page, u8 tag) page_kasan_tag_set() argument 1473 page_kasan_tag_reset(struct page *page) page_kasan_tag_reset() argument 1478 page_kasan_tag(const struct page *page) page_kasan_tag() argument 1483 page_kasan_tag_set(struct page *page, u8 tag) page_kasan_tag_set() argument 1484 page_kasan_tag_reset(struct page *page) page_kasan_tag_reset() argument 1487 page_zone(const struct page *page) page_zone() argument 1492 page_pgdat(const struct page *page) page_pgdat() argument 1498 set_page_section(struct page *page, unsigned long section) set_page_section() argument 1504 page_to_section(const struct page *page) page_to_section() argument 1510 set_page_zone(struct page *page, enum zone_type zone) set_page_zone() argument 1516 set_page_node(struct page *page, unsigned long node) set_page_node() argument 1522 set_page_links(struct page *page, enum zone_type zone, unsigned long node, unsigned long pfn) set_page_links() argument 1533 page_memcg(struct page *page) page_memcg() argument 1537 page_memcg_rcu(struct page *page) page_memcg_rcu() argument 1543 page_memcg(struct page *page) page_memcg() argument 1547 page_memcg_rcu(struct page *page) page_memcg_rcu() argument 1559 lowmem_page_address(const struct page *page) lowmem_page_address() argument 1569 page_address(const struct page *page) page_address() argument 1573 set_page_address(struct page *page, void *address) set_page_address() argument 1599 page_file_mapping(struct page *page) page_file_mapping() argument 1613 page_index(struct page *page) page_index() argument 1629 page_is_pfmemalloc(struct page *page) page_is_pfmemalloc() argument 1642 set_page_pfmemalloc(struct page *page) set_page_pfmemalloc() argument 1647 clear_page_pfmemalloc(struct page *page) clear_page_pfmemalloc() argument 1752 unmap_mapping_page(struct page *page) unmap_mapping_page() argument 1867 cancel_dirty_page(struct page *page) cancel_dirty_page() argument 1972 mm_counter_file(struct page *page) mm_counter_file() argument 1979 mm_counter(struct page *page) mm_counter() argument 2199 ptlock_ptr(struct page *page) ptlock_ptr() argument 2208 ptlock_alloc(struct page *page) ptlock_alloc() argument 2213 ptlock_free(struct page *page) ptlock_free() argument 2217 ptlock_ptr(struct page *page) ptlock_ptr() argument 2228 ptlock_init(struct page *page) ptlock_init() argument 2253 ptlock_init(struct page *page) ptlock_init() argument 2254 ptlock_free(struct page *page) ptlock_free() argument 2263 pgtable_pte_page_ctor(struct page *page) pgtable_pte_page_ctor() argument 2272 pgtable_pte_page_dtor(struct page *page) pgtable_pte_page_dtor() argument 2319 pmd_ptlock_init(struct page *page) pmd_ptlock_init() argument 2327 pmd_ptlock_free(struct page *page) pmd_ptlock_free() argument 2344 pmd_ptlock_init(struct page *page) pmd_ptlock_init() argument 2345 pmd_ptlock_free(struct page *page) pmd_ptlock_free() argument 2358 pgtable_pmd_page_ctor(struct page *page) pgtable_pmd_page_ctor() argument 2367 pgtable_pmd_page_dtor(struct page *page) pgtable_pmd_page_dtor() argument 2420 __free_reserved_page(struct page *page) __free_reserved_page() argument 2427 free_reserved_page(struct page *page) free_reserved_page() argument 2433 mark_page_reserved(struct page *page) mark_page_reserved() argument 2814 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) vmf_insert_page() argument 2947 kernel_poison_pages(struct page *page, int numpages, int enable) kernel_poison_pages() argument 3009 kernel_map_pages(struct page *page, int numpages, int enable) kernel_map_pages() argument 3018 kernel_map_pages(struct page *page, int numpages, int enable) kernel_map_pages() argument 3020 kernel_page_present(struct page *page) kernel_page_present() argument 3190 page_is_guard(struct page *page) page_is_guard() argument 3200 page_is_guard(struct page *page) page_is_guard() argument [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | page_ref.h | 7 #include <linux/page-flags.h> 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *pag 41 __page_ref_set(struct page *page, int v) __page_ref_set() argument 44 __page_ref_mod(struct page *page, int v) __page_ref_mod() argument 47 __page_ref_mod_and_test(struct page *page, int v, int ret) __page_ref_mod_and_test() argument 50 __page_ref_mod_and_return(struct page *page, int v, int ret) __page_ref_mod_and_return() argument 53 __page_ref_mod_unless(struct page *page, int v, int u) __page_ref_mod_unless() argument 56 __page_ref_freeze(struct page *page, int v, int ret) __page_ref_freeze() argument 59 __page_ref_unfreeze(struct page *page, int v) __page_ref_unfreeze() argument 65 page_ref_count(const struct page *page) page_ref_count() argument 92 page_count(const struct page *page) page_count() argument 97 set_page_count(struct page *page, int v) set_page_count() argument 113 init_page_count(struct page *page) init_page_count() argument 118 page_ref_add(struct page *page, int nr) page_ref_add() argument 130 page_ref_sub(struct page *page, int nr) page_ref_sub() argument 142 page_ref_sub_return(struct page *page, int nr) page_ref_sub_return() argument 156 page_ref_inc(struct page *page) page_ref_inc() argument 168 page_ref_dec(struct page *page) page_ref_dec() argument 180 page_ref_sub_and_test(struct page *page, int nr) page_ref_sub_and_test() argument 194 page_ref_inc_return(struct page *page) page_ref_inc_return() argument 208 page_ref_dec_and_test(struct page *page) page_ref_dec_and_test() argument 222 page_ref_dec_return(struct page *page) page_ref_dec_return() argument 236 page_ref_add_unless(struct page *page, int nr, int u) page_ref_add_unless() argument 316 page_ref_freeze(struct page *page, int count) page_ref_freeze() argument 330 page_ref_unfreeze(struct page *page, int count) page_ref_unfreeze() argument 332 VM_BUG_ON_PAGE(page_count(page) != 0, page); page_ref_unfreeze() local [all...] |
H A D | page-flags.h | 3 * Macros for manipulating and testing page->flags 18 * Various page->flags bits: 20 * PG_reserved is set for special pages. The "struct page" of such a page 25 * - Pages reserved or allocated early during boot (before the page allocator 27 * initial vmemmap, initial page tables, crashkernel, elfcorehdr, and much 29 * be given to the page allocator. 32 * - The zero page(s) 33 * - Pages not added to the page allocator when onlining a section because 49 * Consequently, PG_reserved for a page mappe 215 page_fixed_fake_head(const struct page *page) page_fixed_fake_head() argument 241 page_fixed_fake_head(const struct page *page) page_fixed_fake_head() argument 247 page_is_fake_head(struct page *page) page_is_fake_head() argument 252 _compound_head(const struct page *page) _compound_head() argument 291 PageTail(struct page *page) PageTail() argument 296 PageCompound(struct page *page) PageCompound() argument 303 PagePoisoned(const struct page *page) PagePoisoned() argument 311 page_init_poison(struct page *page, size_t size) page_init_poison() argument 318 struct page *page = &folio->page; folio_flags() local 320 VM_BUG_ON_PGFLAGS(PageTail(page), page); folio_flags() local 556 PageSwapCache(struct page *page) PageSwapCache() argument 666 PageMappingFlags(struct page *page) PageMappingFlags() argument 676 PageAnon(struct page *page) PageAnon() argument 687 __PageMovable(struct page *page) __PageMovable() argument 706 PageKsm(struct page *page) PageKsm() argument 743 PageUptodate(struct page *page) PageUptodate() argument 765 __SetPageUptodate(struct page *page) __SetPageUptodate() argument 770 SetPageUptodate(struct page *page) SetPageUptodate() argument 785 test_set_page_writeback(struct page *page) test_set_page_writeback() argument 795 PageHead(struct page *page) PageHead() argument 816 set_compound_head(struct page *page, struct page *head) set_compound_head() argument 821 clear_compound_head(struct page *page) clear_compound_head() argument 827 ClearPageCompound(struct page *page) ClearPageCompound() argument 871 PageTransHuge(struct page *page) PageTransHuge() argument 873 VM_BUG_ON_PAGE(PageTail(page), page); PageTransHuge() local 882 PageTransCompound(struct page *page) PageTransCompound() argument 892 PageTransTail(struct page *page) PageTransTail() argument 922 is_page_hwpoison(struct page *page) is_page_hwpoison() argument 956 page_has_type(struct page *page) page_has_type() argument 1040 PageAnonExclusive(struct page *page) PageAnonExclusive() argument 1043 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); PageAnonExclusive() local 1047 SetPageAnonExclusive(struct page *page) SetPageAnonExclusive() argument 1050 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); SetPageAnonExclusive() local 1054 ClearPageAnonExclusive(struct page *page) ClearPageAnonExclusive() argument 1057 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); ClearPageAnonExclusive() local 1061 __ClearPageAnonExclusive(struct page *page) __ClearPageAnonExclusive() argument 1064 VM_BUG_ON_PGFLAGS(PageHuge(page) && !PageHead(page), page); __ClearPageAnonExclusive() local 1120 page_has_private(struct page *page) page_has_private() argument [all...] |
H A D | highmem-internal.h | 10 void *__kmap_local_page_prot(struct page *page, pgprot_t prot); 35 void *kmap_high(struct page *page); 36 void kunmap_high(struct page *page); 38 struct page *__kmap_to_page(void *addr); 40 static inline void *kmap(struct page *page) in kmap() argument 45 if (!PageHighMem(page)) in kmap() 53 kunmap(struct page *page) kunmap() argument 71 kmap_local_page(struct page *page) kmap_local_page() argument 78 struct page *page = folio_page(folio, offset / PAGE_SIZE); kmap_local_folio() local 82 kmap_local_page_prot(struct page *page, pgprot_t prot) kmap_local_page_prot() argument 97 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 108 kmap_atomic(struct page *page) kmap_atomic() argument 167 kmap(struct page *page) kmap() argument 173 kunmap_high(struct page *page) kunmap_high() argument 176 kunmap(struct page *page) kunmap() argument 183 kmap_local_page(struct page *page) kmap_local_page() argument 193 kmap_local_page_prot(struct page *page, pgprot_t prot) kmap_local_page_prot() argument 210 kmap_atomic(struct page *page) kmap_atomic() argument 220 kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument [all...] |
H A D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned short order); 12 extern void __set_page_owner(struct page *page, 14 extern void __split_page_owner(struct page *page, unsigned int nr); 16 extern void __set_page_owner_migrate_reason(struct page *page, int reason); 17 extern void __dump_page_owner(const struct page *page); 21 reset_page_owner(struct page *page, unsigned short order) reset_page_owner() argument 27 set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) set_page_owner() argument 34 split_page_owner(struct page *page, unsigned int nr) split_page_owner() argument 44 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument 49 dump_page_owner(const struct page *page) dump_page_owner() argument 55 reset_page_owner(struct page *page, unsigned short order) reset_page_owner() argument 58 set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) set_page_owner() argument 62 split_page_owner(struct page *page, unsigned short order) split_page_owner() argument 69 set_page_owner_migrate_reason(struct page *page, int reason) set_page_owner_migrate_reason() argument 72 dump_page_owner(const struct page *page) dump_page_owner() argument [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | swap.c | 45 /* How many pages do we try to swap or page in/out together? */ 79 static void __page_cache_release(struct page *page) in __page_cache_release() argument 81 if (PageLRU(page)) { in __page_cache_release() 82 pg_data_t *pgdat = page_pgdat(page); in __page_cache_release() 87 lruvec = mem_cgroup_page_lruvec(page, pgdat); in __page_cache_release() 88 VM_BUG_ON_PAGE(!PageLRU(page), page); in __page_cache_release() 89 __ClearPageLRU(page); in __page_cache_release() 90 del_page_from_lru_list(page, lruve in __page_cache_release() 96 __put_single_page(struct page *page) __put_single_page() argument 103 __put_compound_page(struct page *page) __put_compound_page() argument 116 __put_page(struct page *page) __put_page() argument 206 pagevec_lru_move_fn(struct pagevec *pvec, void (*move_fn)(struct page *page, struct lruvec *lruvec, void *arg), void *arg) pagevec_lru_move_fn() argument 216 struct page *page = pvec->pages[i]; pagevec_lru_move_fn() local 235 pagevec_move_tail_fn(struct page *page, struct lruvec *lruvec, void *arg) pagevec_move_tail_fn() argument 265 rotate_reclaimable_page(struct page *page) rotate_reclaimable_page() argument 312 lru_note_cost_page(struct page *page) lru_note_cost_page() argument 324 __activate_page(struct page *page, struct lruvec *lruvec, void *arg) __activate_page() argument 357 activate_page(struct page *page) activate_page() argument 377 activate_page(struct page *page) activate_page() argument 388 __lru_cache_activate_page(struct page *page) __lru_cache_activate_page() argument 428 mark_page_accessed(struct page *page) mark_page_accessed() argument 468 lru_cache_add(struct page *page) lru_cache_add() argument 473 VM_BUG_ON_PAGE(PageLRU(page), page); lru_cache_add() local 492 lru_cache_add_inactive_or_unevictable(struct page *page, struct vm_area_struct *vma) lru_cache_add_inactive_or_unevictable() argument 497 VM_BUG_ON_PAGE(PageLRU(page), page); lru_cache_add_inactive_or_unevictable() local 534 lru_deactivate_file_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_deactivate_file_fn() argument 582 lru_deactivate_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_deactivate_fn() argument 600 lru_lazyfree_fn(struct page *page, struct lruvec *lruvec, void *arg) lru_lazyfree_fn() argument 672 deactivate_file_page(struct page *page) deactivate_file_page() argument 701 deactivate_page(struct page *page) deactivate_page() argument 722 mark_page_lazyfree(struct page *page) mark_page_lazyfree() argument 886 struct page *page = pages[i]; release_pages() local 985 lru_add_page_tail(struct page *page, struct page *page_tail, struct lruvec *lruvec, struct list_head *list) lru_add_page_tail() argument 989 VM_BUG_ON_PAGE(PageCompound(page_tail), page); lru_add_page_tail() local 990 VM_BUG_ON_PAGE(PageLRU(page_tail), page); lru_add_page_tail() local 1016 __pagevec_lru_add_fn(struct page *page, struct lruvec *lruvec, void *arg) __pagevec_lru_add_fn() argument 1023 VM_BUG_ON_PAGE(PageLRU(page), page); __pagevec_lru_add_fn() local 1127 struct page *page = pvec->pages[i]; pagevec_remove_exceptionals() local 1201 put_devmap_managed_page(struct page *page) put_devmap_managed_page() argument [all...] |
H A D | filemap.c | 63 * finished 'unifying' the page and buffer cache and SMP-threaded the 64 * page-cache, 21.05.1999, Ingo Molnar <mingo@redhat.com> 121 struct page *page, void *shadow) in page_cache_delete() 123 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete() 129 if (!PageHuge(page)) { in page_cache_delete() 130 xas_set_order(&xas, page->index, compound_order(page)); in page_cache_delete() 131 nr = compound_nr(page); in page_cache_delete() 134 VM_BUG_ON_PAGE(!PageLocked(page), pag in page_cache_delete() 120 page_cache_delete(struct address_space *mapping, struct page *page, void *shadow) page_cache_delete() argument 135 VM_BUG_ON_PAGE(PageTail(page), page); page_cache_delete() local 157 unaccount_page_cache_page(struct address_space *mapping, struct page *page) unaccount_page_cache_page() argument 172 VM_BUG_ON_PAGE(PageTail(page), page); unaccount_page_cache_page() local 173 VM_BUG_ON_PAGE(page_mapped(page), page); unaccount_page_cache_page() local 232 __delete_from_page_cache(struct page *page, void *shadow) __delete_from_page_cache() argument 242 page_cache_free_page(struct address_space *mapping, struct page *page) page_cache_free_page() argument 253 VM_BUG_ON_PAGE(page_count(page) <= 0, page); page_cache_free_page() local 267 delete_from_page_cache(struct page *page) delete_from_page_cache() argument 301 struct page *page; page_cache_delete_batch() local 476 struct page *page; filemap_range_has_page() local 525 struct page *page = pvec.pages[i]; __filemap_fdatawait_range() local 830 __add_to_page_cache_locked(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp, void **shadowp) __add_to_page_cache_locked() argument 841 VM_BUG_ON_PAGE(PageSwapBacked(page), page); __add_to_page_cache_locked() local 928 add_to_page_cache_locked(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache_locked() argument 936 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t offset, gfp_t gfp_mask) add_to_page_cache_lru() argument 969 struct page *page; __page_cache_alloc() local 1000 page_waitqueue(struct page *page) page_waitqueue() argument 1100 wake_up_page_bit(struct page *page, int bit_nr) wake_up_page_bit() argument 1154 wake_up_page(struct page *page, int bit) wake_up_page() argument 1180 trylock_page_bit_common(struct page *page, int bit_nr, struct wait_queue_entry *wait) trylock_page_bit_common() argument 1196 wait_on_page_bit_common(wait_queue_head_t *q, struct page *page, int bit_nr, int state, enum behavior behavior) wait_on_page_bit_common() argument 1335 wait_on_page_bit(struct page *page, int bit_nr) wait_on_page_bit() argument 1342 wait_on_page_bit_killable(struct page *page, int bit_nr) wait_on_page_bit_killable() argument 1349 __wait_on_page_locked_async(struct page *page, struct wait_page_queue *wait, bool set) __wait_on_page_locked_async() argument 1379 wait_on_page_locked_async(struct page *page, struct wait_page_queue *wait) wait_on_page_locked_async() argument 1397 put_and_wait_on_page_locked(struct page *page) put_and_wait_on_page_locked() argument 1413 add_page_wait_queue(struct page *page, wait_queue_entry_t *waiter) add_page_wait_queue() argument 1463 unlock_page(struct page *page) unlock_page() argument 1477 end_page_writeback(struct page *page) end_page_writeback() argument 1511 page_endio(struct page *page, bool is_write, int err) page_endio() argument 1541 struct page *page = compound_head(__page); __lock_page() local 1550 struct page *page = compound_head(__page); __lock_page_killable() local 1557 __lock_page_async(struct page *page, struct wait_page_queue *wait) __lock_page_async() argument 1573 __lock_page_or_retry(struct page *page, struct mm_struct *mm, unsigned int flags) __lock_page_or_retry() argument 1693 struct page *page; find_get_entry() local 1742 struct page *page; find_lock_entry() local 1794 struct page *page; pagecache_get_page() local 1903 struct page *page; find_get_entries() local 1977 struct page *page; find_get_pages_range() local 2042 struct page *page; find_get_pages_contig() local 2099 struct page *page; find_get_pages_range_tag() local 2226 struct page *page; generic_file_buffered_read() local 2579 lock_page_maybe_drop_mmap(struct vm_fault *vmf, struct page *page, struct file **fpin) lock_page_maybe_drop_mmap() argument 2669 do_async_mmap_readahead(struct vm_fault *vmf, struct page *page) do_async_mmap_readahead() argument 2726 struct page *page; filemap_fault() local 2769 VM_BUG_ON_PAGE(page_to_pgoff(page) != offset, page); filemap_fault() local 2849 struct page *head, *page; filemap_map_pages() local 2914 struct page *page = vmf->page; filemap_page_mkwrite() local 2985 wait_on_page_read(struct page *page) wait_on_page_read() argument 3003 struct page *page; do_read_cache_page() local 3160 pagecache_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) pagecache_write_end() argument 3280 struct page *page; grab_cache_page_write_begin() local 3305 struct page *page; generic_perform_write() local 3523 try_to_release_page(struct page *page, gfp_t gfp_mask) try_to_release_page() argument [all...] |
H A D | migrate.c | 86 int isolate_movable_page(struct page *page, isolate_mode_t mode) in isolate_movable_page() argument 94 * In case we 'win' a race for a movable page being freed under us and in isolate_movable_page() 97 * release this page, thus avoiding a nasty leakage. in isolate_movable_page() 99 if (unlikely(!get_page_unless_zero(page))) in isolate_movable_page() 103 * Check PageMovable before holding a PG_lock because page's owner in isolate_movable_page() 104 * assumes anybody doesn't touch PG_lock of newly allocated page in isolate_movable_page() 105 * so unconditionally grabbing the lock ruins page's owner side. in isolate_movable_page() 107 if (unlikely(!__PageMovable(page))) in isolate_movable_page() 111 * compaction threads can race against page migratio in isolate_movable_page() 148 putback_movable_page(struct page *page) putback_movable_page() argument 171 struct page *page; putback_movable_pages() local 205 remove_migration_pte(struct page *page, struct vm_area_struct *vma, unsigned long addr, void *old) remove_migration_pte() argument 218 VM_BUG_ON_PAGE(PageTail(page), page); remove_migration_pte() local 317 struct page *page; __migration_entry_wait() local 364 struct page *page; pmd_migration_entry_wait() local 380 expected_page_refs(struct address_space *mapping, struct page *page) expected_page_refs() argument 403 migrate_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page, int extra_count) migrate_page_move_mapping() argument 454 VM_BUG_ON_PAGE(PageSwapCache(page), page); migrate_page_move_mapping() local 525 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) migrate_huge_page_move_mapping() argument 608 migrate_page_states(struct page *newpage, struct page *page) migrate_page_states() argument 619 VM_BUG_ON_PAGE(PageUnevictable(page), page); migrate_page_states() local 685 migrate_page_copy(struct page *newpage, struct page *page) migrate_page_copy() argument 700 migrate_page_extra(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, int extra_count) migrate_page_extra() argument 726 migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) migrate_page() argument 773 __buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode, bool check_refs) __buffer_migrate_page() argument 857 buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page() argument 870 buffer_migrate_page_norefs(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) buffer_migrate_page_norefs() argument 880 writeout(struct address_space *mapping, struct page *page) writeout() argument 921 fallback_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) fallback_migrate_page() argument 958 move_to_new_page(struct page *newpage, struct page *page, enum migrate_mode mode) move_to_new_page() argument 1038 __unmap_and_move(struct page *page, struct page *newpage, int force, enum migrate_mode mode) __unmap_and_move() argument 1135 VM_BUG_ON_PAGE(PageAnon(page), page); __unmap_and_move() local 1186 unmap_and_move(new_page_t get_new_page, free_page_t put_new_page, unsigned long private, struct page *page, int force, enum migrate_mode mode, enum migrate_reason reason) unmap_and_move() argument 1446 struct page *page; migrate_pages() local 1557 alloc_migration_target(struct page *page, unsigned long private) alloc_migration_target() argument 1642 struct page *page; add_page_for_migration() local 1830 struct page *page; do_pages_stat_array() local 2027 alloc_misplaced_dst_page(struct page *page, unsigned long data) alloc_misplaced_dst_page() argument 2042 numamigrate_isolate_page(pg_data_t *pgdat, struct page *page) numamigrate_isolate_page() argument 2082 struct page *page = pmd_page(pmd); pmd_trans_migrating() local 2091 migrate_misplaced_page(struct page *page, struct vm_area_struct *vma, int node) migrate_misplaced_page() argument 2146 migrate_misplaced_transhuge_page(struct mm_struct *mm, struct vm_area_struct *vma, pmd_t *pmd, pmd_t entry, unsigned long address, struct page *page, int node) migrate_misplaced_transhuge_page() argument 2338 struct page *page; migrate_vma_collect_pmd() local 2381 struct page *page; migrate_vma_collect_pmd() local 2544 migrate_vma_check_page(struct page *page, struct page *fault_page) migrate_vma_check_page() argument 2608 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_prepare() local 2683 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_prepare() local 2716 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_unmap() local 2737 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_unmap() local 2869 migrate_vma_insert_page(struct migrate_vma *migrate, unsigned long addr, struct page *page, unsigned long *src, unsigned long *dst) migrate_vma_insert_page() argument 3019 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_pages() local 3107 struct page *page = migrate_pfn_to_page(migrate->src[i]); migrate_vma_finalize() local [all...] |
H A D | rmap.c | 10 * Provides methods for unmapping each kind of mapped page: 25 * page->flags PG_locked (lock_page) * (see huegtlbfs below) 28 * hugetlb_fault_mutex (hugetlbfs specific page fault mutex) 50 * hugetlb_fault_mutex (hugetlbfs specific page fault mutex) 51 * page->flags PG_locked (lock_page) 274 * searches where page is mapped. 466 * Getting a lock on a stable anon_vma from a page off the LRU is tricky! 470 * have been relevant to this page. 472 * The page might have been remapped to a different anon_vma or the anon_vma 477 * ensure that any anon_vma obtained from the page wil 489 page_get_anon_vma(struct page *page) page_get_anon_vma() argument 532 page_lock_anon_vma_read(struct page *page) page_lock_anon_vma_read() argument 712 page_address_in_vma(struct page *page, struct vm_area_struct *vma) page_address_in_vma() argument 775 page_referenced_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) page_referenced_one() argument 862 page_referenced(struct page *page, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) page_referenced() argument 910 page_mkclean_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) page_mkclean_one() argument 992 page_mkclean(struct page *page) page_mkclean() argument 1027 page_move_anon_rmap(struct page *page, struct vm_area_struct *vma) page_move_anon_rmap() argument 1052 __page_set_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, int exclusive) __page_set_anon_rmap() argument 1081 __page_check_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) __page_check_anon_rmap() argument 1096 VM_BUG_ON_PAGE(page_anon_vma(page)->root != vma->anon_vma->root, page); __page_check_anon_rmap() local 1098 page); __page_check_anon_rmap() local 1113 page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, bool compound) page_add_anon_rmap() argument 1124 do_page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, int flags) do_page_add_anon_rmap() argument 1182 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address, bool compound) page_add_new_anon_rmap() argument 1199 VM_BUG_ON_PAGE(PageTransCompound(page), page); page_add_new_anon_rmap() local 1214 page_add_file_rmap(struct page *page, bool compound) page_add_file_rmap() argument 1247 page_remove_file_rmap(struct page *page, bool compound) page_remove_file_rmap() argument 1288 page_remove_anon_compound_rmap(struct page *page) page_remove_anon_compound_rmap() argument 1339 page_remove_rmap(struct page *page, bool compound) page_remove_rmap() argument 1386 try_to_unmap_one(struct page *page, struct vm_area_struct *vma, unsigned long address, void *arg) try_to_unmap_one() argument 1775 page_not_mapped(struct page *page) page_not_mapped() argument 1790 try_to_unmap(struct page *page, enum ttu_flags flags) try_to_unmap() argument 1834 try_to_munlock(struct page *page) try_to_munlock() argument 1859 rmap_walk_anon_lock(struct page *page, struct rmap_walk_control *rwc) rmap_walk_anon_lock() argument 1895 rmap_walk_anon(struct page *page, struct rmap_walk_control *rwc, bool locked) rmap_walk_anon() argument 1948 rmap_walk_file(struct page *page, struct rmap_walk_control *rwc, bool locked) rmap_walk_file() argument 1991 rmap_walk(struct page *page, struct rmap_walk_control *rwc) rmap_walk() argument 2002 rmap_walk_locked(struct page *page, struct rmap_walk_control *rwc) rmap_walk_locked() argument 2005 VM_BUG_ON_PAGE(PageKsm(page), page); rmap_walk_locked() local 2018 hugepage_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) hugepage_add_anon_rmap() argument 2032 hugepage_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) hugepage_add_new_anon_rmap() argument [all...] |
H A D | page_io.c | 10 * Add swap of shared pages through the page cache. 20.2.1998. Stephen Tweedie 30 struct page *page, bio_end_io_t end_io) in get_swap_bio() 38 bio->bi_iter.bi_sector = map_swap_page(page, &bdev); in get_swap_bio() 43 bio_add_page(bio, page, thp_size(page), 0); in get_swap_bio() 50 struct page *page = bio_first_page_all(bio); in end_swap_bio_write() local 53 SetPageError(page); in end_swap_bio_write() 55 * We failed to write the page ou in end_swap_bio_write() 29 get_swap_bio(gfp_t gfp_flags, struct page *page, bio_end_io_t end_io) get_swap_bio() argument 74 struct page *page = bio_first_page_all(bio); end_swap_bio_read() local 198 swap_writepage(struct page *page, struct writeback_control *wbc) swap_writepage() argument 227 count_swpout_vm_event(struct page *page) count_swpout_vm_event() argument 237 bio_associate_blkg_from_page(struct bio *bio, struct page *page) bio_associate_blkg_from_page() argument 253 __swap_writepage(struct page *page, struct writeback_control *wbc, bio_end_io_t end_write_func) __swap_writepage() argument 324 swap_readpage(struct page *page, bool synchronous) swap_readpage() argument 335 VM_BUG_ON_PAGE(PageUptodate(page), page); swap_readpage() local 405 swap_set_page_dirty(struct page *page) swap_set_page_dirty() argument [all...] |
H A D | truncate.c | 30 * Regular page slots are stabilized by the page lock even without the tree 83 struct page *page = pvec->pages[i]; in truncate_exceptional_pvec_entries() local 86 if (!xa_is_value(page)) { in truncate_exceptional_pvec_entries() 87 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries() 99 __clear_shadow_entry(mapping, index, page); in truncate_exceptional_pvec_entries() 138 * do_invalidatepage - invalidate part or all of a page 139 * @page: the page whic 152 do_invalidatepage(struct page *page, unsigned int offset, unsigned int length) do_invalidatepage() argument 176 truncate_cleanup_page(struct page *page) truncate_cleanup_page() argument 202 invalidate_complete_page(struct address_space *mapping, struct page *page) invalidate_complete_page() argument 217 truncate_inode_page(struct address_space *mapping, struct page *page) truncate_inode_page() argument 219 VM_BUG_ON_PAGE(PageTail(page), page); truncate_inode_page() local 232 generic_error_remove_page(struct address_space *mapping, struct page *page) generic_error_remove_page() argument 252 invalidate_inode_page(struct page *page) invalidate_inode_page() argument 338 struct page *page = pvec.pages[i]; truncate_inode_pages_range() local 372 struct page *page = find_lock_page(mapping, start - 1); truncate_inode_pages_range() local 391 struct page *page = find_lock_page(mapping, end); truncate_inode_pages_range() local 430 struct page *page = pvec.pages[i]; truncate_inode_pages_range() local 543 struct page *page = pvec.pages[i]; __invalidate_mapping_pages() local 655 invalidate_complete_page2(struct address_space *mapping, struct page *page) invalidate_complete_page2() argument 683 do_launder_page(struct address_space *mapping, struct page *page) do_launder_page() argument 723 struct page *page = pvec.pages[i]; invalidate_inode_pages2_range() local 886 struct page *page; pagecache_isize_extended() local [all...] |
H A D | page_isolation.c | 7 #include <linux/page-isolation.h> 18 static int set_migratetype_isolate(struct page *page, int migratetype, int isol_flags) in set_migratetype_isolate() argument 20 struct zone *zone = page_zone(page); in set_migratetype_isolate() 21 struct page *unmovable; in set_migratetype_isolate() 31 if (is_migrate_isolate_page(page)) { in set_migratetype_isolate() 40 unmovable = has_unmovable_pages(zone, page, migratetype, isol_flags); in set_migratetype_isolate() 43 int mt = get_pageblock_migratetype(page); in set_migratetype_isolate() 45 set_pageblock_migratetype(page, MIGRATE_ISOLATE); in set_migratetype_isolate() 47 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLAT in set_migratetype_isolate() 68 unset_migratetype_isolate(struct page *page, unsigned migratetype) unset_migratetype_isolate() argument 133 struct page *page; __first_valid_page() local 188 struct page *page; start_isolate_page_range() local 209 struct page *page = pfn_to_online_page(pfn); start_isolate_page_range() local 225 struct page *page; undo_isolate_page_range() local 250 struct page *page; __test_page_isolated_in_pageblock() local 288 struct page *page; test_pages_isolated() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | folio-compat.c | 13 struct address_space *page_mapping(struct page *page) in page_mapping() argument 15 return folio_mapping(page_folio(page)); in page_mapping() 19 void unlock_page(struct page *page) in unlock_page() argument 21 return folio_unlock(page_folio(page)); in unlock_page() 25 void end_page_writeback(struct page *page) in end_page_writeback() argument 27 return folio_end_writeback(page_folio(page)); in end_page_writeback() 31 void wait_on_page_writeback(struct page *pag argument 37 wait_for_stable_page(struct page *page) wait_for_stable_page() argument 43 mark_page_accessed(struct page *page) mark_page_accessed() argument 49 set_page_writeback(struct page *page) set_page_writeback() argument 55 set_page_dirty(struct page *page) set_page_dirty() argument 61 __set_page_dirty_nobuffers(struct page *page) __set_page_dirty_nobuffers() argument 67 clear_page_dirty_for_io(struct page *page) clear_page_dirty_for_io() argument 73 redirty_page_for_writepage(struct writeback_control *wbc, struct page *page) redirty_page_for_writepage() argument 80 lru_cache_add_inactive_or_unevictable(struct page *page, struct vm_area_struct *vma) lru_cache_add_inactive_or_unevictable() argument 86 add_to_page_cache_lru(struct page *page, struct address_space *mapping, pgoff_t index, gfp_t gfp) add_to_page_cache_lru() argument 114 isolate_lru_page(struct page *page) isolate_lru_page() argument 121 putback_lru_page(struct page *page) putback_lru_page() argument 127 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) page_add_new_anon_rmap() argument 130 VM_BUG_ON_PAGE(PageTail(page), page); page_add_new_anon_rmap() local [all...] |
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | subpage.c | 14 * - Only support 64K page size for now 15 * This is to make metadata handling easier, as 64K page would ensure 16 * all nodesize would fit inside one page, thus we don't need to handle 22 * - Metadata can't cross 64K page boundary 32 * needed range, other unrelated range in the same page will not be touched. 35 * The writeback is still for the full page, but we will only submit 36 * the dirty extent buffers in the page. 38 * This means, if we have a metadata page like this: 54 * record the status of each sector inside a page. This provides the extra 58 * Since we have multiple tree blocks inside one page, w 67 btrfs_is_subpage(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_is_subpage() argument 118 btrfs_attach_subpage(const struct btrfs_fs_info *fs_info, struct page *page, enum btrfs_subpage_type type) btrfs_attach_subpage() argument 142 btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_detach_subpage() argument 194 btrfs_page_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_inc_eb_refs() argument 209 btrfs_page_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_dec_eb_refs() argument 225 btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_assert() argument 241 btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_start_reader() argument 252 btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_end_reader() argument 276 btrfs_subpage_clamp_range(struct page *page, u64 *start, u32 *len) btrfs_subpage_clamp_range() argument 294 btrfs_subpage_start_writer(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_start_writer() argument 308 btrfs_subpage_end_and_test_writer(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_end_and_test_writer() argument 340 btrfs_page_start_writer_lock(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_start_writer_lock() argument 357 btrfs_page_end_writer_lock(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_end_writer_lock() argument 387 btrfs_subpage_set_uptodate(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_uptodate() argument 402 btrfs_subpage_clear_uptodate(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_uptodate() argument 416 btrfs_subpage_set_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_dirty() argument 440 btrfs_subpage_clear_and_test_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_and_test_dirty() argument 457 btrfs_subpage_clear_dirty(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_dirty() argument 467 btrfs_subpage_set_writeback(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_writeback() argument 481 btrfs_subpage_clear_writeback(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_writeback() argument 498 btrfs_subpage_set_ordered(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_ordered() argument 512 btrfs_subpage_clear_ordered(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_ordered() argument 527 btrfs_subpage_set_checked(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_set_checked() argument 542 btrfs_subpage_clear_checked(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_clear_checked() argument 656 btrfs_page_assert_not_dirty(const struct btrfs_fs_info *fs_info, struct page *page) btrfs_page_assert_not_dirty() argument 687 btrfs_page_unlock_writer(struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_page_unlock_writer() argument 719 btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info, struct page *page, u64 start, u32 len) btrfs_subpage_dump_bitmap() argument [all...] |
H A D | subpage.h | 42 * Structure to trace status of each sector inside a page, attached to 43 * page::private for both data and metadata inodes. 50 * page. 51 * Data relies on @readers to unlock the page when last reader finished. 52 * While metadata doesn't need page unlock, it needs to prevent 53 * page::private get cleared before the last end_page_read(). 76 bool btrfs_is_subpage(const struct btrfs_fs_info *fs_info, struct page *page); 80 struct page *page, enu [all...] |
/kernel/linux/linux-6.6/arch/arm/include/asm/ |
H A D | page.h | 3 * arch/arm/include/asm/page.h 10 /* PAGE_SHIFT determines the page size */ 19 #include <asm/page-nommu.h> 30 * page-based copying and clearing for user space for the particular 107 struct page; 111 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr); 112 void (*cpu_copy_user_highpage)(struct page *to, struct page *from, 116 void fa_copy_user_highpage(struct page *t [all...] |