/kernel/linux/linux-5.10/mm/ |
H A D | sparse.c | 35 * If we did not store the node number in the page then we have to 37 * node the page belongs to. 45 int page_to_nid(const struct page *page) in page_to_nid() argument 47 return section_to_node_table[page_to_section(page)]; in page_to_nid() 304 * physical page frame number. 306 static unsigned long sparse_encode_mem_map(struct page *mem_map, unsigned long pnum) in sparse_encode_mem_map() 319 struct page *sparse_decode_mem_map(unsigned long coded_mem_map, unsigned long pnum) in sparse_decode_mem_map() 323 return ((struct page *)coded_mem_map) + section_nr_to_pfn(pnum); in sparse_decode_mem_map() 328 unsigned long pnum, struct page *mem_ma in sparse_init_one_section() 739 struct page *page = virt_to_page(memmap); free_map_bootmem() local [all...] |
/kernel/linux/linux-6.6/drivers/vdpa/vdpa_user/ |
H A D | iova_domain.c | 141 struct page *page; in do_bounce() local 147 page = pfn_to_page(pfn); in do_bounce() 149 memcpy_from_page(addr, page, offset, sz); in do_bounce() 151 memcpy_to_page(page, offset, addr, sz); in do_bounce() 189 static struct page * 195 struct page *page = NULL; in vduse_domain_get_coherent_page() local 202 page = pfn_to_page((map->addr + iova - map->start) >> PAGE_SHIFT); in vduse_domain_get_coherent_page() 203 get_page(page); in vduse_domain_get_coherent_page() 214 struct page *page = NULL; vduse_domain_get_bounce_page() local 297 struct page *page = NULL; vduse_domain_remove_user_bounce_pages() local 376 vduse_domain_map_page(struct vduse_iova_domain *domain, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, unsigned long attrs) vduse_domain_map_page() argument 489 struct page *page; vduse_domain_mmap_fault() local [all...] |
/kernel/linux/linux-5.10/arch/x86/lib/ |
H A D | mmx_32.c | 136 static void fast_clear_page(void *page) in fast_clear_page() argument 156 : : "r" (page) : "memory"); in fast_clear_page() 157 page += 64; in fast_clear_page() 256 static void fast_clear_page(void *page) in fast_clear_page() argument 284 : : "r" (page) : "memory"); in fast_clear_page() 285 page += 128; in fast_clear_page() 345 * Favour MMX for page clear and copy: 347 static void slow_zero_page(void *page) in slow_zero_page() argument 356 :"a" (0), "1" (page), "0" (1024) in slow_zero_page() 360 void mmx_clear_page(void *page) in mmx_clear_page() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
H A D | shmem_utils.c | 54 struct page **pages; in shmem_pin_map() 97 struct page *page; in __shmem_rw() local 100 page = shmem_read_mapping_page_gfp(file->f_mapping, pfn, in __shmem_rw() 102 if (IS_ERR(page)) in __shmem_rw() 103 return PTR_ERR(page); in __shmem_rw() 105 vaddr = kmap(page); in __shmem_rw() 108 set_page_dirty(page); in __shmem_rw() 112 mark_page_accessed(page); in __shmem_rw() 113 kunmap(page); in __shmem_rw() [all...] |
/kernel/linux/linux-5.10/fs/jfs/ |
H A D | jfs_metapage.h | 23 sector_t index; /* block address of page */ 27 struct page *page; member 65 * hold_metapage and put_metapage are used in conjunction. The page lock 93 struct page *page = mp->page; in metapage_nohomeok() local 94 lock_page(page); in metapage_nohomeok() 97 get_page(page); in metapage_nohomeok() 98 wait_on_page_writeback(page); in metapage_nohomeok() [all...] |
/kernel/linux/linux-6.6/fs/jfs/ |
H A D | jfs_metapage.h | 23 sector_t index; /* block address of page */ 27 struct page *page; member 65 * hold_metapage and put_metapage are used in conjunction. The page lock 93 struct page *page = mp->page; in metapage_nohomeok() local 94 lock_page(page); in metapage_nohomeok() 97 get_page(page); in metapage_nohomeok() 98 wait_on_page_writeback(page); in metapage_nohomeok() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | ksm.h | 69 * When do_swap_page() first faults in from swap what used to be a KSM page, 73 * needed to reconstitute a cross-anon_vma KSM page: for now it has to make 77 * but what if the vma was unmerged while the page was swapped out? 79 struct page *ksm_might_need_to_copy(struct page *page, 86 void collect_procs_ksm(struct page *page, struct list_head *to_kill, 119 static inline void collect_procs_ksm(struct page *page, in collect_procs_ksm() argument 132 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) ksm_might_need_to_copy() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_phys.c | 55 sg_assign_page(sg, (struct page *)vaddr); in i915_gem_object_get_pages_phys() 61 struct page *page; in i915_gem_object_get_pages_phys() local 64 page = shmem_read_mapping_page(mapping, i); in i915_gem_object_get_pages_phys() 65 if (IS_ERR(page)) in i915_gem_object_get_pages_phys() 68 src = kmap_atomic(page); in i915_gem_object_get_pages_phys() 73 put_page(page); in i915_gem_object_get_pages_phys() 107 struct page *page; in i915_gem_object_put_pages_phys() local 110 page in i915_gem_object_put_pages_phys() [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | writeback.h | 53 this for each page written */ 69 unsigned for_reclaim:1; /* Invoked from the page allocator */ 137 * on page writeback completions [end_page_writeback()]. Those 141 * We use page writeout completions because we are interested in 147 * of this period itself is measured in page writeback completions. 214 void __inode_attach_wb(struct inode *inode, struct page *page); 219 void wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page, 228 * @page 234 inode_attach_wb(struct inode *inode, struct page *page) inode_attach_wb() argument 296 inode_attach_wb(struct inode *inode, struct page *page) inode_attach_wb() argument 324 wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page, size_t bytes) wbc_account_cgroup_owner() argument [all...] |
/kernel/linux/linux-5.10/fs/verity/ |
H A D | verify.c | 42 /* Extract a hash from a hash page */ 43 static void extract_hash(struct page *hpage, unsigned int hoffset, in extract_hash() 70 * Verify a single data page against the file's Merkle tree. 74 * only ascend the tree until an already-verified page is seen, as indicated by 75 * the PageChecked bit being set; then verify the path to that page. 81 * Note that multiple processes may race to verify a hash page and mark it 84 * Return: true if the page is valid, else false. 87 struct ahash_request *req, struct page *data_page, in verify_page() 97 struct page *hpages[FS_VERITY_MAX_LEVELS]; in verify_page() 104 pr_debug_ratelimited("Verifying data page in verify_page() 196 fsverity_verify_page(struct page *page) fsverity_verify_page() argument 259 struct page *page = bv->bv_page; fsverity_verify_bio() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_phys.c | 60 sg_assign_page(sg, (struct page *)vaddr); in i915_gem_object_get_pages_phys() 66 struct page *page; in i915_gem_object_get_pages_phys() local 69 page = shmem_read_mapping_page(mapping, i); in i915_gem_object_get_pages_phys() 70 if (IS_ERR(page)) in i915_gem_object_get_pages_phys() 73 src = kmap_atomic(page); in i915_gem_object_get_pages_phys() 78 put_page(page); in i915_gem_object_get_pages_phys() 84 /* We're no longer struct page backed */ in i915_gem_object_get_pages_phys() 114 struct page *page; in i915_gem_object_put_pages_phys() local [all...] |
/kernel/linux/linux-6.6/drivers/hwmon/pmbus/ |
H A D | mp2888.c | 39 static int mp2888_read_byte_data(struct i2c_client *client, int page, int reg) in mp2888_read_byte_data() argument 94 mp2888_read_phase(struct i2c_client *client, struct mp2888_data *data, int page, int phase, u8 reg) in mp2888_read_phase() argument 98 ret = pmbus_read_word_data(client, page, phase, reg); in mp2888_read_phase() 128 mp2888_read_phases(struct i2c_client *client, struct mp2888_data *data, int page, int phase) in mp2888_read_phases() argument 134 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS1_2); in mp2888_read_phases() 137 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS3_4); in mp2888_read_phases() 140 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS5_6); in mp2888_read_phases() 143 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS7_8); in mp2888_read_phases() 146 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS9_10); in mp2888_read_phases() 154 static int mp2888_read_word_data(struct i2c_client *client, int page, in argument 263 mp2888_write_word_data(struct i2c_client *client, int page, int reg, u16 word) mp2888_write_word_data() argument [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | memfd.c | 34 struct page *page; in memfd_tag_pins() local 41 xas_for_each(xas, page, ULONG_MAX) { in memfd_tag_pins() 43 if (!xa_is_value(page) && in memfd_tag_pins() 44 PageTransHuge(page) && !PageHuge(page)) in memfd_tag_pins() 47 if (!xa_is_value(page) && in memfd_tag_pins() 48 page_count(page) - total_mapcount(page) != cache_count) in memfd_tag_pins() 51 xas_set(xas, page in memfd_tag_pins() 78 struct page *page; memfd_wait_for_pins() local [all...] |
/kernel/linux/linux-5.10/fs/reiserfs/ |
H A D | inode.c | 24 int reiserfs_commit_write(struct file *f, struct page *page, 169 * about what happens for reading a file. For each page, 185 * If this page has a file tail in it, and 186 * it was read in by get_block_create_0, the page data is valid, 188 * it. So, look through this page, and check all the mapped buffers 193 static inline void fix_tail_page_for_writing(struct page *page) in fix_tail_page_for_writing() argument 197 if (page && page_has_buffers(page)) { in fix_tail_page_for_writing() 2197 struct page *page; grab_tail_page() local 2269 struct page *page = NULL; reiserfs_truncate_file() local 2522 reiserfs_write_full_page(struct page *page, struct writeback_control *wbc) reiserfs_write_full_page() argument 2738 reiserfs_readpage(struct file *f, struct page *page) reiserfs_readpage() argument 2743 reiserfs_writepage(struct page *page, struct writeback_control *wbc) reiserfs_writepage() argument 2762 struct page *page; reiserfs_write_begin() local 2830 __reiserfs_write_begin(struct page *page, unsigned from, unsigned len) __reiserfs_write_begin() argument 2890 reiserfs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) reiserfs_write_end() argument 2993 reiserfs_commit_write(struct file *f, struct page *page, unsigned from, unsigned to) reiserfs_commit_write() argument 3153 reiserfs_invalidatepage(struct page *page, unsigned int offset, unsigned int length) reiserfs_invalidatepage() argument 3206 reiserfs_set_page_dirty(struct page *page) reiserfs_set_page_dirty() argument 3225 reiserfs_releasepage(struct page *page, gfp_t unused_gfp_flags) reiserfs_releasepage() argument [all...] |
/kernel/linux/linux-5.10/arch/csky/mm/ |
H A D | dma-mapping.c | 18 struct page *page = phys_to_page(paddr); in cache_op() local 19 void *start = __va(page_to_phys(page)); in cache_op() 29 if (PageHighMem(page)) { in cache_op() 30 start = kmap_atomic(page); in cache_op() 42 page++; in cache_op() 54 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 56 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-5.10/arch/nds32/kernel/ |
H A D | dma.c | 16 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); in cache_op() local 24 if (PageHighMem(page)) { in cache_op() 29 page += offset >> PAGE_SHIFT; in cache_op() 35 addr = kmap_atomic(page); in cache_op() 44 page++; in cache_op() 79 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 81 cache_op(page_to_phys(page), size, cpu_dma_wbinval_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | memory.h | 6 #include <asm/page.h> 36 #define hyp_page_to_pfn(page) ((struct hyp_page *)(page) - hyp_vmemmap) 37 #define hyp_page_to_phys(page) hyp_pfn_to_phys((hyp_page_to_pfn(page))) 38 #define hyp_page_to_virt(page) __hyp_va(hyp_page_to_phys(page)) 39 #define hyp_page_to_pool(page) (((struct hyp_page *)page)->pool)
|
/kernel/linux/linux-6.6/arch/csky/mm/ |
H A D | dma-mapping.c | 17 struct page *page = phys_to_page(paddr); in cache_op() local 18 void *start = __va(page_to_phys(page)); in cache_op() 28 if (PageHighMem(page)) { in cache_op() 29 start = kmap_atomic(page); in cache_op() 41 page++; in cache_op() 53 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 55 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-5.10/drivers/tee/optee/ |
H A D | shm_pool.c | 19 struct page *page; in pool_op_alloc() local 22 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc() 23 if (!page) in pool_op_alloc() 26 shm->kaddr = page_address(page); in pool_op_alloc() 27 shm->paddr = page_to_phys(page); in pool_op_alloc() 36 struct page **pages; in pool_op_alloc() 45 pages[i] = page + i; in pool_op_alloc() 58 __free_pages(page, order); in pool_op_alloc() 84 * optee_shm_pool_alloc_pages() - create page [all...] |
/kernel/linux/linux-5.10/fs/ocfs2/ |
H A D | aops.h | 14 struct page *page, 18 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno, 22 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages); 43 struct page **pagep, void **fsdata, 44 struct buffer_head *di_bh, struct page *mmap_page); 46 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
|
/kernel/linux/linux-5.10/include/asm-generic/ |
H A D | pgalloc.h | 11 * __pte_alloc_one_kernel - allocate a page for PTE-level kernel page table 15 * anything beyond simple page allocation. 26 * pte_alloc_one_kernel - allocate a page for PTE-level kernel page table 38 * pte_free_kernel - free PTE-level kernel page table page 40 * @pte: pointer to the memory containing the page table 48 * __pte_alloc_one - allocate a page for PTE-level user page tabl 121 struct page *page; pmd_alloc_one() local [all...] |
/kernel/linux/linux-6.6/fs/ocfs2/ |
H A D | aops.h | 12 struct page *page, 16 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno, 20 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages); 41 struct page **pagep, void **fsdata, 42 struct buffer_head *di_bh, struct page *mmap_page); 44 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
|
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | mem-reservation.c | 23 * Use one extent per PAGE_SIZE to avoid to break down the page into 30 struct page **pages, in __xenmem_reservation_va_mapping_update() 36 struct page *page = pages[i]; in __xenmem_reservation_va_mapping_update() local 37 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_update() 40 BUG_ON(!page); in __xenmem_reservation_va_mapping_update() 44 * different page granularity. in __xenmem_reservation_va_mapping_update() 59 struct page **pages) in __xenmem_reservation_va_mapping_reset() 64 struct page *page in __xenmem_reservation_va_mapping_reset() local [all...] |
/third_party/node/deps/v8/src/heap/ |
H A D | paged-spaces-inl.h | 60 void PagedSpace::UnlinkFreeListCategories(Page* page) { in UnlinkFreeListCategories() argument 61 DCHECK_EQ(this, page->owner()); in UnlinkFreeListCategories() 62 page->ForAllFreeListCategories([this](FreeListCategory* category) { in UnlinkFreeListCategories() 67 size_t PagedSpace::RelinkFreeListCategories(Page* page) { in RelinkFreeListCategories() argument 68 DCHECK_EQ(this, page->owner()); in RelinkFreeListCategories() 70 page->ForAllFreeListCategories([this, &added](FreeListCategory* category) { in RelinkFreeListCategories() 75 DCHECK_IMPLIES(!page->IsFlagSet(Page::NEVER_ALLOCATE_ON_PAGE), in RelinkFreeListCategories() 76 page->AvailableInFreeList() == in RelinkFreeListCategories() 77 page->AvailableInFreeListFromAllocatedBytes()); in RelinkFreeListCategories()
|
/kernel/linux/linux-5.10/arch/xtensa/include/asm/ |
H A D | pgalloc.h | 26 #define pmd_populate(mm, pmdp, page) \ 27 (pmd_val(*(pmdp)) = ((unsigned long)page_to_virt(page))) 57 struct page *page; in pte_alloc_one() local 59 page = __pte_alloc_one(mm, GFP_PGTABLE_USER); in pte_alloc_one() 60 if (!page) in pte_alloc_one() 62 ptes_clear(page_address(page)); in pte_alloc_one() 63 return page; in pte_alloc_one()
|