Home
last modified time | relevance | path

Searched refs:page (Results 576 - 600 of 6878) sorted by relevance

1...<<21222324252627282930>>...276

/kernel/linux/linux-6.6/arch/nios2/include/asm/
H A Dcacheflush.h16 * This flag is used to indicate that the page pointed to by a pte is clean
31 void flush_dcache_page(struct page *page);
36 void flush_icache_pages(struct vm_area_struct *vma, struct page *page,
44 extern void copy_to_user_page(struct vm_area_struct *vma, struct page *page,
47 extern void copy_from_user_page(struct vm_area_struct *vma, struct page *page,
/kernel/linux/linux-5.10/include/xen/interface/io/
H A Dfbif.h90 /* shared page */
95 #define XENFB_IN_RING(page) \
96 ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS))
97 #define XENFB_IN_RING_REF(page, idx) \
98 (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN])
103 #define XENFB_OUT_RING(page) \
104 ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS))
105 #define XENFB_OUT_RING_REF(page, idx) \
106 (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN])
119 * Framebuffer page director
[all...]
/kernel/linux/linux-5.10/fs/efs/
H A Dsymlink.c15 static int efs_symlink_readpage(struct file *file, struct page *page) in efs_symlink_readpage() argument
17 char *link = page_address(page); in efs_symlink_readpage()
19 struct inode * inode = page->mapping->host; in efs_symlink_readpage()
42 SetPageUptodate(page); in efs_symlink_readpage()
43 unlock_page(page); in efs_symlink_readpage()
46 SetPageError(page); in efs_symlink_readpage()
47 unlock_page(page); in efs_symlink_readpage()
/kernel/linux/linux-6.6/include/xen/interface/io/
H A Dfbif.h73 /* shared page */
78 #define XENFB_IN_RING(page) \
79 ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS))
80 #define XENFB_IN_RING_REF(page, idx) \
81 (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN])
86 #define XENFB_OUT_RING(page) \
87 ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS))
88 #define XENFB_OUT_RING_REF(page, idx) \
89 (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN])
102 * Framebuffer page director
[all...]
/kernel/linux/linux-5.10/arch/csky/mm/
H A Ddma-mapping.c18 struct page *page = phys_to_page(paddr); in cache_op() local
19 void *start = __va(page_to_phys(page)); in cache_op()
29 if (PageHighMem(page)) { in cache_op()
30 start = kmap_atomic(page); in cache_op()
42 page++; in cache_op()
54 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
56 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
/kernel/linux/linux-5.10/arch/nds32/kernel/
H A Ddma.c16 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); in cache_op() local
24 if (PageHighMem(page)) { in cache_op()
29 page += offset >> PAGE_SHIFT; in cache_op()
35 addr = kmap_atomic(page); in cache_op()
44 page++; in cache_op()
79 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
81 cache_op(page_to_phys(page), size, cpu_dma_wbinval_range); in arch_dma_prep_coherent()
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/include/nvhe/
H A Dmemory.h6 #include <asm/page.h>
36 #define hyp_page_to_pfn(page) ((struct hyp_page *)(page) - hyp_vmemmap)
37 #define hyp_page_to_phys(page) hyp_pfn_to_phys((hyp_page_to_pfn(page)))
38 #define hyp_page_to_virt(page) __hyp_va(hyp_page_to_phys(page))
39 #define hyp_page_to_pool(page) (((struct hyp_page *)page)->pool)
/kernel/linux/linux-6.6/arch/csky/mm/
H A Ddma-mapping.c17 struct page *page = phys_to_page(paddr); in cache_op() local
18 void *start = __va(page_to_phys(page)); in cache_op()
28 if (PageHighMem(page)) { in cache_op()
29 start = kmap_atomic(page); in cache_op()
41 page++; in cache_op()
53 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
55 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
/kernel/linux/linux-5.10/drivers/tee/optee/
H A Dshm_pool.c19 struct page *page; in pool_op_alloc() local
22 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc()
23 if (!page) in pool_op_alloc()
26 shm->kaddr = page_address(page); in pool_op_alloc()
27 shm->paddr = page_to_phys(page); in pool_op_alloc()
36 struct page **pages; in pool_op_alloc()
45 pages[i] = page + i; in pool_op_alloc()
58 __free_pages(page, order); in pool_op_alloc()
84 * optee_shm_pool_alloc_pages() - create page
[all...]
/kernel/linux/linux-5.10/fs/ocfs2/
H A Daops.h14 struct page *page,
18 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno,
22 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages);
43 struct page **pagep, void **fsdata,
44 struct buffer_head *di_bh, struct page *mmap_page);
46 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
/kernel/linux/linux-5.10/include/asm-generic/
H A Dpgalloc.h11 * __pte_alloc_one_kernel - allocate a page for PTE-level kernel page table
15 * anything beyond simple page allocation.
26 * pte_alloc_one_kernel - allocate a page for PTE-level kernel page table
38 * pte_free_kernel - free PTE-level kernel page table page
40 * @pte: pointer to the memory containing the page table
48 * __pte_alloc_one - allocate a page for PTE-level user page tabl
121 struct page *page; pmd_alloc_one() local
[all...]
/kernel/linux/linux-6.6/fs/ocfs2/
H A Daops.h12 struct page *page,
16 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno,
20 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages);
41 struct page **pagep, void **fsdata,
42 struct buffer_head *di_bh, struct page *mmap_page);
44 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
/kernel/linux/linux-6.6/drivers/xen/
H A Dmem-reservation.c23 * Use one extent per PAGE_SIZE to avoid to break down the page into
30 struct page **pages, in __xenmem_reservation_va_mapping_update()
36 struct page *page = pages[i]; in __xenmem_reservation_va_mapping_update() local
37 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_update()
40 BUG_ON(!page); in __xenmem_reservation_va_mapping_update()
44 * different page granularity. in __xenmem_reservation_va_mapping_update()
59 struct page **pages) in __xenmem_reservation_va_mapping_reset()
64 struct page *page in __xenmem_reservation_va_mapping_reset() local
[all...]
/third_party/node/deps/v8/src/heap/
H A Dpaged-spaces-inl.h60 void PagedSpace::UnlinkFreeListCategories(Page* page) { in UnlinkFreeListCategories() argument
61 DCHECK_EQ(this, page->owner()); in UnlinkFreeListCategories()
62 page->ForAllFreeListCategories([this](FreeListCategory* category) { in UnlinkFreeListCategories()
67 size_t PagedSpace::RelinkFreeListCategories(Page* page) { in RelinkFreeListCategories() argument
68 DCHECK_EQ(this, page->owner()); in RelinkFreeListCategories()
70 page->ForAllFreeListCategories([this, &added](FreeListCategory* category) { in RelinkFreeListCategories()
75 DCHECK_IMPLIES(!page->IsFlagSet(Page::NEVER_ALLOCATE_ON_PAGE), in RelinkFreeListCategories()
76 page->AvailableInFreeList() == in RelinkFreeListCategories()
77 page->AvailableInFreeListFromAllocatedBytes()); in RelinkFreeListCategories()
/kernel/linux/linux-5.10/arch/c6x/include/asm/
H A Dcacheflush.h16 #include <asm/page.h>
28 #define flush_icache_page(vma, page) \
31 L1D_cache_block_writeback_invalidate(page_address(page), \
32 (unsigned long) page_address(page) + PAGE_SIZE)); \
33 L1P_cache_block_invalidate(page_address(page), \
34 (unsigned long) page_address(page) + PAGE_SIZE)); \
37 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
/kernel/linux/linux-6.6/arch/ia64/include/asm/
H A Dcacheflush.h10 #include <linux/page-flags.h>
13 #include <asm/page.h>
22 static inline void flush_dcache_page(struct page *page) in flush_dcache_page() argument
24 flush_dcache_folio(page_folio(page)); in flush_dcache_page()
31 #define flush_icache_user_page(vma, page, user_addr, len) \
33 unsigned long _addr = (unsigned long) page_address(page) + ((user_addr) & ~PAGE_MASK); \
/kernel/linux/linux-5.10/include/xen/
H A Dpage.h5 #include <asm/page.h>
7 /* The hypercall interface supports only 4KB page */
20 #define page_to_xen_pfn(page) \
21 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
28 #include <asm/xen/page.h>
30 /* Return the GFN associated to the first 4KB of the page */
31 static inline unsigned long xen_page_to_gfn(struct page *page) in xen_page_to_gfn() argument
33 return pfn_to_gfn(page_to_xen_pfn(page)); in xen_page_to_gfn()
/kernel/linux/linux-6.6/include/xen/
H A Dpage.h5 #include <asm/page.h>
7 /* The hypercall interface supports only 4KB page */
20 #define page_to_xen_pfn(page) \
21 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
28 #include <asm/xen/page.h>
30 /* Return the GFN associated to the first 4KB of the page */
31 static inline unsigned long xen_page_to_gfn(struct page *page) in xen_page_to_gfn() argument
33 return pfn_to_gfn(page_to_xen_pfn(page)); in xen_page_to_gfn()
/kernel/linux/linux-5.10/fs/hugetlbfs/
H A Dinode.c122 * Mask used when checking the page offset value passed in via system
155 * page based offset in vm_pgoff could be sufficiently large to in hugetlbfs_file_mmap()
165 /* must be huge page aligned */ in hugetlbfs_file_mmap()
289 hugetlbfs_read_actor(struct page *page, unsigned long offset, in hugetlbfs_read_actor() argument
306 n = copy_page_to_iter(&page[i], offset, chunksize, to); in hugetlbfs_read_actor()
318 * Support for read() - Find the page attached to f_mapping and copy out the
335 struct page *page; in hugetlbfs_read_iter() local
338 /* nr is the maximum number of bytes to copy from this page */ in hugetlbfs_read_iter()
392 hugetlbfs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) hugetlbfs_write_end() argument
400 remove_huge_page(struct page *page) remove_huge_page() argument
490 struct page *page = pvec.pages[i]; remove_inode_hugepages() local
687 struct page *page; hugetlbfs_fallocate() local
957 hugetlbfs_set_page_dirty(struct page *page) hugetlbfs_set_page_dirty() argument
965 hugetlbfs_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) hugetlbfs_migrate_page() argument
994 hugetlbfs_error_remove_page(struct address_space *mapping, struct page *page) hugetlbfs_error_remove_page() argument
[all...]
/third_party/skia/third_party/externals/harfbuzz/src/
H A Dhb-bit-set.hh32 #include "hb-bit-page.hh"
131 page_t *page = page_for (g, true); if (unlikely (!page)) return; in add() local
132 page->add (g); in add()
143 page_t *page = page_for (a, true); if (unlikely (!page)) return false; in add_range() local
144 page->add_range (a, b); in add_range()
148 page_t *page = page_for (a, true); if (unlikely (!page)) return false; in add_range() local
149 page in add_range()
173 page_t *page = page_for (g, v); if (unlikely (v && !page)) return; set_array() local
213 page_t *page = page_for (g, v); if (unlikely (v && !page)) return false; set_sorted_array() local
291 page_t *page = page_for (a); del_range() local
302 page_t *page = page_for (b); del_range() local
311 const page_t *page = page_for (g); get() local
723 const auto& page = pages[map.index]; get_min() local
736 const auto& page = pages[map.index]; get_max() local
[all...]
/device/soc/rockchip/common/sdk_linux/include/linux/
H A Ddma-map-ops.h20 struct page *(*alloc_pages)(struct device *dev, size_t size,
23 void (*free_pages)(struct device *dev, size_t size, struct page *vaddr,
37 dma_addr_t (*map_page)(struct device *dev, struct page *page,
120 struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
122 bool dma_release_from_contiguous(struct device *dev, struct page *pages,
124 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
125 void dma_free_contiguous(struct device *dev, struct page *page, size_t size);
143 static inline struct page *dma_alloc_from_contiguou in dma_contiguous_reserve_area()
161 dma_free_contiguous(struct device *dev, struct page *page, size_t size) dma_free_contiguous() argument
326 arch_dma_prep_coherent(struct page *page, size_t size) arch_dma_prep_coherent() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/
H A Ddrm_vm.c109 * \return pointer to the page structure.
111 * Find the right map and if it's AGP memory find the real physical page to
112 * map, get the page, increment the use count and return it.
147 struct page *page; in drm_vm_fault() local
157 * It's AGP memory - find the real physical page to map in drm_vm_fault()
169 * Get the page, inc the use count, and return it in drm_vm_fault()
172 page = agpmem->memory->pages[offset]; in drm_vm_fault()
173 get_page(page); in drm_vm_fault()
174 vmf->page in drm_vm_fault()
210 struct page *page; drm_vm_shm_fault() local
312 struct page *page; drm_vm_dma_fault() local
349 struct page *page; drm_vm_sg_fault() local
[all...]
/kernel/linux/linux-5.10/fs/hfsplus/
H A Dbtree.c139 struct page *page; in hfs_btree_open() local
161 page = read_mapping_page(mapping, 0, NULL); in hfs_btree_open()
162 if (IS_ERR(page)) in hfs_btree_open()
166 head = (struct hfs_btree_header_rec *)(kmap(page) + in hfs_btree_open()
243 kunmap(page); in hfs_btree_open()
244 put_page(page); in hfs_btree_open()
248 put_page(page); in hfs_btree_open()
286 struct page *page; in hfs_btree_write() local
451 struct page *page; hfs_bmap_free() local
[all...]
/kernel/linux/linux-5.10/fs/ext4/
H A Dmove_extent.c119 * @index1: page index
120 * @index2: page index
121 * @page: result page vector
127 pgoff_t index1, pgoff_t index2, struct page *page[2]) in mext_page_double_lock()
142 page[0] = grab_cache_page_write_begin(mapping[0], index1, fl); in mext_page_double_lock()
143 if (!page[0]) in mext_page_double_lock()
146 page[1] = grab_cache_page_write_begin(mapping[1], index2, fl); in mext_page_double_lock()
147 if (!page[ in mext_page_double_lock()
126 mext_page_double_lock(struct inode *inode1, struct inode *inode2, pgoff_t index1, pgoff_t index2, struct page *page[2]) mext_page_double_lock() argument
167 mext_page_mkuptodate(struct page *page, unsigned from, unsigned to) mext_page_mkuptodate() argument
[all...]
/kernel/linux/linux-5.10/fs/erofs/
H A Dinode.c12 * if inode is successfully read, return its inode page (or sometimes
13 * the inode payload page if it's an extended inode) in order to fill
16 static struct page *erofs_read_inode(struct inode *inode, in erofs_read_inode()
25 struct page *page; in erofs_read_inode() local
37 page = erofs_get_meta_page(sb, blkaddr); in erofs_read_inode()
38 if (IS_ERR(page)) { in erofs_read_inode()
39 erofs_err(sb, "failed to get inode (nid: %llu) page, err %ld", in erofs_read_inode()
40 vi->nid, PTR_ERR(page)); in erofs_read_inode()
41 return page; in erofs_read_inode()
236 struct page *page; erofs_fill_inode() local
[all...]

Completed in 10 milliseconds

1...<<21222324252627282930>>...276