/kernel/linux/linux-6.6/arch/nios2/include/asm/ |
H A D | cacheflush.h | 16 * This flag is used to indicate that the page pointed to by a pte is clean 31 void flush_dcache_page(struct page *page); 36 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, 44 extern void copy_to_user_page(struct vm_area_struct *vma, struct page *page, 47 extern void copy_from_user_page(struct vm_area_struct *vma, struct page *page,
|
/kernel/linux/linux-5.10/include/xen/interface/io/ |
H A D | fbif.h | 90 /* shared page */ 95 #define XENFB_IN_RING(page) \ 96 ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS)) 97 #define XENFB_IN_RING_REF(page, idx) \ 98 (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN]) 103 #define XENFB_OUT_RING(page) \ 104 ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS)) 105 #define XENFB_OUT_RING_REF(page, idx) \ 106 (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN]) 119 * Framebuffer page director [all...] |
/kernel/linux/linux-5.10/fs/efs/ |
H A D | symlink.c | 15 static int efs_symlink_readpage(struct file *file, struct page *page) in efs_symlink_readpage() argument 17 char *link = page_address(page); in efs_symlink_readpage() 19 struct inode * inode = page->mapping->host; in efs_symlink_readpage() 42 SetPageUptodate(page); in efs_symlink_readpage() 43 unlock_page(page); in efs_symlink_readpage() 46 SetPageError(page); in efs_symlink_readpage() 47 unlock_page(page); in efs_symlink_readpage()
|
/kernel/linux/linux-6.6/include/xen/interface/io/ |
H A D | fbif.h | 73 /* shared page */ 78 #define XENFB_IN_RING(page) \ 79 ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS)) 80 #define XENFB_IN_RING_REF(page, idx) \ 81 (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN]) 86 #define XENFB_OUT_RING(page) \ 87 ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS)) 88 #define XENFB_OUT_RING_REF(page, idx) \ 89 (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN]) 102 * Framebuffer page director [all...] |
/kernel/linux/linux-5.10/arch/csky/mm/ |
H A D | dma-mapping.c | 18 struct page *page = phys_to_page(paddr); in cache_op() local 19 void *start = __va(page_to_phys(page)); in cache_op() 29 if (PageHighMem(page)) { in cache_op() 30 start = kmap_atomic(page); in cache_op() 42 page++; in cache_op() 54 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 56 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-5.10/arch/nds32/kernel/ |
H A D | dma.c | 16 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); in cache_op() local 24 if (PageHighMem(page)) { in cache_op() 29 page += offset >> PAGE_SHIFT; in cache_op() 35 addr = kmap_atomic(page); in cache_op() 44 page++; in cache_op() 79 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 81 cache_op(page_to_phys(page), size, cpu_dma_wbinval_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | memory.h | 6 #include <asm/page.h> 36 #define hyp_page_to_pfn(page) ((struct hyp_page *)(page) - hyp_vmemmap) 37 #define hyp_page_to_phys(page) hyp_pfn_to_phys((hyp_page_to_pfn(page))) 38 #define hyp_page_to_virt(page) __hyp_va(hyp_page_to_phys(page)) 39 #define hyp_page_to_pool(page) (((struct hyp_page *)page)->pool)
|
/kernel/linux/linux-6.6/arch/csky/mm/ |
H A D | dma-mapping.c | 17 struct page *page = phys_to_page(paddr); in cache_op() local 18 void *start = __va(page_to_phys(page)); in cache_op() 28 if (PageHighMem(page)) { in cache_op() 29 start = kmap_atomic(page); in cache_op() 41 page++; in cache_op() 53 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 55 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
|
/kernel/linux/linux-5.10/drivers/tee/optee/ |
H A D | shm_pool.c | 19 struct page *page; in pool_op_alloc() local 22 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc() 23 if (!page) in pool_op_alloc() 26 shm->kaddr = page_address(page); in pool_op_alloc() 27 shm->paddr = page_to_phys(page); in pool_op_alloc() 36 struct page **pages; in pool_op_alloc() 45 pages[i] = page + i; in pool_op_alloc() 58 __free_pages(page, order); in pool_op_alloc() 84 * optee_shm_pool_alloc_pages() - create page [all...] |
/kernel/linux/linux-5.10/fs/ocfs2/ |
H A D | aops.h | 14 struct page *page, 18 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno, 22 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages); 43 struct page **pagep, void **fsdata, 44 struct buffer_head *di_bh, struct page *mmap_page); 46 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
|
/kernel/linux/linux-5.10/include/asm-generic/ |
H A D | pgalloc.h | 11 * __pte_alloc_one_kernel - allocate a page for PTE-level kernel page table 15 * anything beyond simple page allocation. 26 * pte_alloc_one_kernel - allocate a page for PTE-level kernel page table 38 * pte_free_kernel - free PTE-level kernel page table page 40 * @pte: pointer to the memory containing the page table 48 * __pte_alloc_one - allocate a page for PTE-level user page tabl 121 struct page *page; pmd_alloc_one() local [all...] |
/kernel/linux/linux-6.6/fs/ocfs2/ |
H A D | aops.h | 12 struct page *page, 16 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno, 20 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages); 41 struct page **pagep, void **fsdata, 42 struct buffer_head *di_bh, struct page *mmap_page); 44 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
|
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | mem-reservation.c | 23 * Use one extent per PAGE_SIZE to avoid to break down the page into 30 struct page **pages, in __xenmem_reservation_va_mapping_update() 36 struct page *page = pages[i]; in __xenmem_reservation_va_mapping_update() local 37 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_update() 40 BUG_ON(!page); in __xenmem_reservation_va_mapping_update() 44 * different page granularity. in __xenmem_reservation_va_mapping_update() 59 struct page **pages) in __xenmem_reservation_va_mapping_reset() 64 struct page *page in __xenmem_reservation_va_mapping_reset() local [all...] |
/third_party/node/deps/v8/src/heap/ |
H A D | paged-spaces-inl.h | 60 void PagedSpace::UnlinkFreeListCategories(Page* page) { in UnlinkFreeListCategories() argument 61 DCHECK_EQ(this, page->owner()); in UnlinkFreeListCategories() 62 page->ForAllFreeListCategories([this](FreeListCategory* category) { in UnlinkFreeListCategories() 67 size_t PagedSpace::RelinkFreeListCategories(Page* page) { in RelinkFreeListCategories() argument 68 DCHECK_EQ(this, page->owner()); in RelinkFreeListCategories() 70 page->ForAllFreeListCategories([this, &added](FreeListCategory* category) { in RelinkFreeListCategories() 75 DCHECK_IMPLIES(!page->IsFlagSet(Page::NEVER_ALLOCATE_ON_PAGE), in RelinkFreeListCategories() 76 page->AvailableInFreeList() == in RelinkFreeListCategories() 77 page->AvailableInFreeListFromAllocatedBytes()); in RelinkFreeListCategories()
|
/kernel/linux/linux-5.10/arch/c6x/include/asm/ |
H A D | cacheflush.h | 16 #include <asm/page.h> 28 #define flush_icache_page(vma, page) \ 31 L1D_cache_block_writeback_invalidate(page_address(page), \ 32 (unsigned long) page_address(page) + PAGE_SIZE)); \ 33 L1P_cache_block_invalidate(page_address(page), \ 34 (unsigned long) page_address(page) + PAGE_SIZE)); \ 37 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
|
/kernel/linux/linux-6.6/arch/ia64/include/asm/ |
H A D | cacheflush.h | 10 #include <linux/page-flags.h> 13 #include <asm/page.h> 22 static inline void flush_dcache_page(struct page *page) in flush_dcache_page() argument 24 flush_dcache_folio(page_folio(page)); in flush_dcache_page() 31 #define flush_icache_user_page(vma, page, user_addr, len) \ 33 unsigned long _addr = (unsigned long) page_address(page) + ((user_addr) & ~PAGE_MASK); \
|
/kernel/linux/linux-5.10/include/xen/ |
H A D | page.h | 5 #include <asm/page.h> 7 /* The hypercall interface supports only 4KB page */ 20 #define page_to_xen_pfn(page) \ 21 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT)) 28 #include <asm/xen/page.h> 30 /* Return the GFN associated to the first 4KB of the page */ 31 static inline unsigned long xen_page_to_gfn(struct page *page) in xen_page_to_gfn() argument 33 return pfn_to_gfn(page_to_xen_pfn(page)); in xen_page_to_gfn()
|
/kernel/linux/linux-6.6/include/xen/ |
H A D | page.h | 5 #include <asm/page.h> 7 /* The hypercall interface supports only 4KB page */ 20 #define page_to_xen_pfn(page) \ 21 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT)) 28 #include <asm/xen/page.h> 30 /* Return the GFN associated to the first 4KB of the page */ 31 static inline unsigned long xen_page_to_gfn(struct page *page) in xen_page_to_gfn() argument 33 return pfn_to_gfn(page_to_xen_pfn(page)); in xen_page_to_gfn()
|
/kernel/linux/linux-5.10/fs/hugetlbfs/ |
H A D | inode.c | 122 * Mask used when checking the page offset value passed in via system 155 * page based offset in vm_pgoff could be sufficiently large to in hugetlbfs_file_mmap() 165 /* must be huge page aligned */ in hugetlbfs_file_mmap() 289 hugetlbfs_read_actor(struct page *page, unsigned long offset, in hugetlbfs_read_actor() argument 306 n = copy_page_to_iter(&page[i], offset, chunksize, to); in hugetlbfs_read_actor() 318 * Support for read() - Find the page attached to f_mapping and copy out the 335 struct page *page; in hugetlbfs_read_iter() local 338 /* nr is the maximum number of bytes to copy from this page */ in hugetlbfs_read_iter() 392 hugetlbfs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) hugetlbfs_write_end() argument 400 remove_huge_page(struct page *page) remove_huge_page() argument 490 struct page *page = pvec.pages[i]; remove_inode_hugepages() local 687 struct page *page; hugetlbfs_fallocate() local 957 hugetlbfs_set_page_dirty(struct page *page) hugetlbfs_set_page_dirty() argument 965 hugetlbfs_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) hugetlbfs_migrate_page() argument 994 hugetlbfs_error_remove_page(struct address_space *mapping, struct page *page) hugetlbfs_error_remove_page() argument [all...] |
/third_party/skia/third_party/externals/harfbuzz/src/ |
H A D | hb-bit-set.hh | 32 #include "hb-bit-page.hh" 131 page_t *page = page_for (g, true); if (unlikely (!page)) return; in add() local 132 page->add (g); in add() 143 page_t *page = page_for (a, true); if (unlikely (!page)) return false; in add_range() local 144 page->add_range (a, b); in add_range() 148 page_t *page = page_for (a, true); if (unlikely (!page)) return false; in add_range() local 149 page in add_range() 173 page_t *page = page_for (g, v); if (unlikely (v && !page)) return; set_array() local 213 page_t *page = page_for (g, v); if (unlikely (v && !page)) return false; set_sorted_array() local 291 page_t *page = page_for (a); del_range() local 302 page_t *page = page_for (b); del_range() local 311 const page_t *page = page_for (g); get() local 723 const auto& page = pages[map.index]; get_min() local 736 const auto& page = pages[map.index]; get_max() local [all...] |
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | dma-map-ops.h | 20 struct page *(*alloc_pages)(struct device *dev, size_t size, 23 void (*free_pages)(struct device *dev, size_t size, struct page *vaddr, 37 dma_addr_t (*map_page)(struct device *dev, struct page *page, 120 struct page *dma_alloc_from_contiguous(struct device *dev, size_t count, 122 bool dma_release_from_contiguous(struct device *dev, struct page *pages, 124 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp); 125 void dma_free_contiguous(struct device *dev, struct page *page, size_t size); 143 static inline struct page *dma_alloc_from_contiguou in dma_contiguous_reserve_area() 161 dma_free_contiguous(struct device *dev, struct page *page, size_t size) dma_free_contiguous() argument 326 arch_dma_prep_coherent(struct page *page, size_t size) arch_dma_prep_coherent() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_vm.c | 109 * \return pointer to the page structure. 111 * Find the right map and if it's AGP memory find the real physical page to 112 * map, get the page, increment the use count and return it. 147 struct page *page; in drm_vm_fault() local 157 * It's AGP memory - find the real physical page to map in drm_vm_fault() 169 * Get the page, inc the use count, and return it in drm_vm_fault() 172 page = agpmem->memory->pages[offset]; in drm_vm_fault() 173 get_page(page); in drm_vm_fault() 174 vmf->page in drm_vm_fault() 210 struct page *page; drm_vm_shm_fault() local 312 struct page *page; drm_vm_dma_fault() local 349 struct page *page; drm_vm_sg_fault() local [all...] |
/kernel/linux/linux-5.10/fs/hfsplus/ |
H A D | btree.c | 139 struct page *page; in hfs_btree_open() local 161 page = read_mapping_page(mapping, 0, NULL); in hfs_btree_open() 162 if (IS_ERR(page)) in hfs_btree_open() 166 head = (struct hfs_btree_header_rec *)(kmap(page) + in hfs_btree_open() 243 kunmap(page); in hfs_btree_open() 244 put_page(page); in hfs_btree_open() 248 put_page(page); in hfs_btree_open() 286 struct page *page; in hfs_btree_write() local 451 struct page *page; hfs_bmap_free() local [all...] |
/kernel/linux/linux-5.10/fs/ext4/ |
H A D | move_extent.c | 119 * @index1: page index 120 * @index2: page index 121 * @page: result page vector 127 pgoff_t index1, pgoff_t index2, struct page *page[2]) in mext_page_double_lock() 142 page[0] = grab_cache_page_write_begin(mapping[0], index1, fl); in mext_page_double_lock() 143 if (!page[0]) in mext_page_double_lock() 146 page[1] = grab_cache_page_write_begin(mapping[1], index2, fl); in mext_page_double_lock() 147 if (!page[ in mext_page_double_lock() 126 mext_page_double_lock(struct inode *inode1, struct inode *inode2, pgoff_t index1, pgoff_t index2, struct page *page[2]) mext_page_double_lock() argument 167 mext_page_mkuptodate(struct page *page, unsigned from, unsigned to) mext_page_mkuptodate() argument [all...] |
/kernel/linux/linux-5.10/fs/erofs/ |
H A D | inode.c | 12 * if inode is successfully read, return its inode page (or sometimes 13 * the inode payload page if it's an extended inode) in order to fill 16 static struct page *erofs_read_inode(struct inode *inode, in erofs_read_inode() 25 struct page *page; in erofs_read_inode() local 37 page = erofs_get_meta_page(sb, blkaddr); in erofs_read_inode() 38 if (IS_ERR(page)) { in erofs_read_inode() 39 erofs_err(sb, "failed to get inode (nid: %llu) page, err %ld", in erofs_read_inode() 40 vi->nid, PTR_ERR(page)); in erofs_read_inode() 41 return page; in erofs_read_inode() 236 struct page *page; erofs_fill_inode() local [all...] |