Home
last modified time | relevance | path

Searched refs:page (Results 551 - 575 of 6707) sorted by relevance

1...<<21222324252627282930>>...269

/kernel/linux/linux-5.10/mm/
H A Dsparse.c35 * If we did not store the node number in the page then we have to
37 * node the page belongs to.
45 int page_to_nid(const struct page *page) in page_to_nid() argument
47 return section_to_node_table[page_to_section(page)]; in page_to_nid()
304 * physical page frame number.
306 static unsigned long sparse_encode_mem_map(struct page *mem_map, unsigned long pnum) in sparse_encode_mem_map()
319 struct page *sparse_decode_mem_map(unsigned long coded_mem_map, unsigned long pnum) in sparse_decode_mem_map()
323 return ((struct page *)coded_mem_map) + section_nr_to_pfn(pnum); in sparse_decode_mem_map()
328 unsigned long pnum, struct page *mem_ma in sparse_init_one_section()
739 struct page *page = virt_to_page(memmap); free_map_bootmem() local
[all...]
/kernel/linux/linux-6.6/drivers/vdpa/vdpa_user/
H A Diova_domain.c141 struct page *page; in do_bounce() local
147 page = pfn_to_page(pfn); in do_bounce()
149 memcpy_from_page(addr, page, offset, sz); in do_bounce()
151 memcpy_to_page(page, offset, addr, sz); in do_bounce()
189 static struct page *
195 struct page *page = NULL; in vduse_domain_get_coherent_page() local
202 page = pfn_to_page((map->addr + iova - map->start) >> PAGE_SHIFT); in vduse_domain_get_coherent_page()
203 get_page(page); in vduse_domain_get_coherent_page()
214 struct page *page = NULL; vduse_domain_get_bounce_page() local
297 struct page *page = NULL; vduse_domain_remove_user_bounce_pages() local
376 vduse_domain_map_page(struct vduse_iova_domain *domain, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, unsigned long attrs) vduse_domain_map_page() argument
489 struct page *page; vduse_domain_mmap_fault() local
[all...]
/kernel/linux/linux-5.10/arch/x86/lib/
H A Dmmx_32.c136 static void fast_clear_page(void *page) in fast_clear_page() argument
156 : : "r" (page) : "memory"); in fast_clear_page()
157 page += 64; in fast_clear_page()
256 static void fast_clear_page(void *page) in fast_clear_page() argument
284 : : "r" (page) : "memory"); in fast_clear_page()
285 page += 128; in fast_clear_page()
345 * Favour MMX for page clear and copy:
347 static void slow_zero_page(void *page) in slow_zero_page() argument
356 :"a" (0), "1" (page), "0" (1024) in slow_zero_page()
360 void mmx_clear_page(void *page) in mmx_clear_page() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
H A Dshmem_utils.c54 struct page **pages; in shmem_pin_map()
97 struct page *page; in __shmem_rw() local
100 page = shmem_read_mapping_page_gfp(file->f_mapping, pfn, in __shmem_rw()
102 if (IS_ERR(page)) in __shmem_rw()
103 return PTR_ERR(page); in __shmem_rw()
105 vaddr = kmap(page); in __shmem_rw()
108 set_page_dirty(page); in __shmem_rw()
112 mark_page_accessed(page); in __shmem_rw()
113 kunmap(page); in __shmem_rw()
[all...]
/kernel/linux/linux-5.10/fs/jfs/
H A Djfs_metapage.h23 sector_t index; /* block address of page */
27 struct page *page; member
65 * hold_metapage and put_metapage are used in conjunction. The page lock
93 struct page *page = mp->page; in metapage_nohomeok() local
94 lock_page(page); in metapage_nohomeok()
97 get_page(page); in metapage_nohomeok()
98 wait_on_page_writeback(page); in metapage_nohomeok()
[all...]
/kernel/linux/linux-6.6/fs/jfs/
H A Djfs_metapage.h23 sector_t index; /* block address of page */
27 struct page *page; member
65 * hold_metapage and put_metapage are used in conjunction. The page lock
93 struct page *page = mp->page; in metapage_nohomeok() local
94 lock_page(page); in metapage_nohomeok()
97 get_page(page); in metapage_nohomeok()
98 wait_on_page_writeback(page); in metapage_nohomeok()
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dksm.h69 * When do_swap_page() first faults in from swap what used to be a KSM page,
73 * needed to reconstitute a cross-anon_vma KSM page: for now it has to make
77 * but what if the vma was unmerged while the page was swapped out?
79 struct page *ksm_might_need_to_copy(struct page *page,
86 void collect_procs_ksm(struct page *page, struct list_head *to_kill,
119 static inline void collect_procs_ksm(struct page *page, in collect_procs_ksm() argument
132 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) ksm_might_need_to_copy() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/
H A Di915_gem_phys.c55 sg_assign_page(sg, (struct page *)vaddr); in i915_gem_object_get_pages_phys()
61 struct page *page; in i915_gem_object_get_pages_phys() local
64 page = shmem_read_mapping_page(mapping, i); in i915_gem_object_get_pages_phys()
65 if (IS_ERR(page)) in i915_gem_object_get_pages_phys()
68 src = kmap_atomic(page); in i915_gem_object_get_pages_phys()
73 put_page(page); in i915_gem_object_get_pages_phys()
107 struct page *page; in i915_gem_object_put_pages_phys() local
110 page in i915_gem_object_put_pages_phys()
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dwriteback.h53 this for each page written */
69 unsigned for_reclaim:1; /* Invoked from the page allocator */
137 * on page writeback completions [end_page_writeback()]. Those
141 * We use page writeout completions because we are interested in
147 * of this period itself is measured in page writeback completions.
214 void __inode_attach_wb(struct inode *inode, struct page *page);
219 void wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page,
228 * @page
234 inode_attach_wb(struct inode *inode, struct page *page) inode_attach_wb() argument
296 inode_attach_wb(struct inode *inode, struct page *page) inode_attach_wb() argument
324 wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page, size_t bytes) wbc_account_cgroup_owner() argument
[all...]
/kernel/linux/linux-5.10/fs/verity/
H A Dverify.c42 /* Extract a hash from a hash page */
43 static void extract_hash(struct page *hpage, unsigned int hoffset, in extract_hash()
70 * Verify a single data page against the file's Merkle tree.
74 * only ascend the tree until an already-verified page is seen, as indicated by
75 * the PageChecked bit being set; then verify the path to that page.
81 * Note that multiple processes may race to verify a hash page and mark it
84 * Return: true if the page is valid, else false.
87 struct ahash_request *req, struct page *data_page, in verify_page()
97 struct page *hpages[FS_VERITY_MAX_LEVELS]; in verify_page()
104 pr_debug_ratelimited("Verifying data page in verify_page()
196 fsverity_verify_page(struct page *page) fsverity_verify_page() argument
259 struct page *page = bv->bv_page; fsverity_verify_bio() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/
H A Di915_gem_phys.c60 sg_assign_page(sg, (struct page *)vaddr); in i915_gem_object_get_pages_phys()
66 struct page *page; in i915_gem_object_get_pages_phys() local
69 page = shmem_read_mapping_page(mapping, i); in i915_gem_object_get_pages_phys()
70 if (IS_ERR(page)) in i915_gem_object_get_pages_phys()
73 src = kmap_atomic(page); in i915_gem_object_get_pages_phys()
78 put_page(page); in i915_gem_object_get_pages_phys()
84 /* We're no longer struct page backed */ in i915_gem_object_get_pages_phys()
114 struct page *page; in i915_gem_object_put_pages_phys() local
[all...]
/kernel/linux/linux-6.6/drivers/hwmon/pmbus/
H A Dmp2888.c39 static int mp2888_read_byte_data(struct i2c_client *client, int page, int reg) in mp2888_read_byte_data() argument
94 mp2888_read_phase(struct i2c_client *client, struct mp2888_data *data, int page, int phase, u8 reg) in mp2888_read_phase() argument
98 ret = pmbus_read_word_data(client, page, phase, reg); in mp2888_read_phase()
128 mp2888_read_phases(struct i2c_client *client, struct mp2888_data *data, int page, int phase) in mp2888_read_phases() argument
134 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS1_2); in mp2888_read_phases()
137 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS3_4); in mp2888_read_phases()
140 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS5_6); in mp2888_read_phases()
143 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS7_8); in mp2888_read_phases()
146 ret = mp2888_read_phase(client, data, page, phase, MP2888_MFR_READ_CS9_10); in mp2888_read_phases()
154 static int mp2888_read_word_data(struct i2c_client *client, int page, in argument
263 mp2888_write_word_data(struct i2c_client *client, int page, int reg, u16 word) mp2888_write_word_data() argument
[all...]
/kernel/linux/linux-6.6/mm/
H A Dmemfd.c34 struct page *page; in memfd_tag_pins() local
41 xas_for_each(xas, page, ULONG_MAX) { in memfd_tag_pins()
43 if (!xa_is_value(page) && in memfd_tag_pins()
44 PageTransHuge(page) && !PageHuge(page)) in memfd_tag_pins()
47 if (!xa_is_value(page) && in memfd_tag_pins()
48 page_count(page) - total_mapcount(page) != cache_count) in memfd_tag_pins()
51 xas_set(xas, page in memfd_tag_pins()
78 struct page *page; memfd_wait_for_pins() local
[all...]
/kernel/linux/linux-5.10/fs/reiserfs/
H A Dinode.c24 int reiserfs_commit_write(struct file *f, struct page *page,
169 * about what happens for reading a file. For each page,
185 * If this page has a file tail in it, and
186 * it was read in by get_block_create_0, the page data is valid,
188 * it. So, look through this page, and check all the mapped buffers
193 static inline void fix_tail_page_for_writing(struct page *page) in fix_tail_page_for_writing() argument
197 if (page && page_has_buffers(page)) { in fix_tail_page_for_writing()
2197 struct page *page; grab_tail_page() local
2269 struct page *page = NULL; reiserfs_truncate_file() local
2522 reiserfs_write_full_page(struct page *page, struct writeback_control *wbc) reiserfs_write_full_page() argument
2738 reiserfs_readpage(struct file *f, struct page *page) reiserfs_readpage() argument
2743 reiserfs_writepage(struct page *page, struct writeback_control *wbc) reiserfs_writepage() argument
2762 struct page *page; reiserfs_write_begin() local
2830 __reiserfs_write_begin(struct page *page, unsigned from, unsigned len) __reiserfs_write_begin() argument
2890 reiserfs_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) reiserfs_write_end() argument
2993 reiserfs_commit_write(struct file *f, struct page *page, unsigned from, unsigned to) reiserfs_commit_write() argument
3153 reiserfs_invalidatepage(struct page *page, unsigned int offset, unsigned int length) reiserfs_invalidatepage() argument
3206 reiserfs_set_page_dirty(struct page *page) reiserfs_set_page_dirty() argument
3225 reiserfs_releasepage(struct page *page, gfp_t unused_gfp_flags) reiserfs_releasepage() argument
[all...]
/kernel/linux/linux-5.10/arch/csky/mm/
H A Ddma-mapping.c18 struct page *page = phys_to_page(paddr); in cache_op() local
19 void *start = __va(page_to_phys(page)); in cache_op()
29 if (PageHighMem(page)) { in cache_op()
30 start = kmap_atomic(page); in cache_op()
42 page++; in cache_op()
54 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
56 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
/kernel/linux/linux-5.10/arch/nds32/kernel/
H A Ddma.c16 struct page *page = pfn_to_page(paddr >> PAGE_SHIFT); in cache_op() local
24 if (PageHighMem(page)) { in cache_op()
29 page += offset >> PAGE_SHIFT; in cache_op()
35 addr = kmap_atomic(page); in cache_op()
44 page++; in cache_op()
79 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
81 cache_op(page_to_phys(page), size, cpu_dma_wbinval_range); in arch_dma_prep_coherent()
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/include/nvhe/
H A Dmemory.h6 #include <asm/page.h>
36 #define hyp_page_to_pfn(page) ((struct hyp_page *)(page) - hyp_vmemmap)
37 #define hyp_page_to_phys(page) hyp_pfn_to_phys((hyp_page_to_pfn(page)))
38 #define hyp_page_to_virt(page) __hyp_va(hyp_page_to_phys(page))
39 #define hyp_page_to_pool(page) (((struct hyp_page *)page)->pool)
/kernel/linux/linux-6.6/arch/csky/mm/
H A Ddma-mapping.c17 struct page *page = phys_to_page(paddr); in cache_op() local
18 void *start = __va(page_to_phys(page)); in cache_op()
28 if (PageHighMem(page)) { in cache_op()
29 start = kmap_atomic(page); in cache_op()
41 page++; in cache_op()
53 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument
55 cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range); in arch_dma_prep_coherent()
/kernel/linux/linux-5.10/drivers/tee/optee/
H A Dshm_pool.c19 struct page *page; in pool_op_alloc() local
22 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc()
23 if (!page) in pool_op_alloc()
26 shm->kaddr = page_address(page); in pool_op_alloc()
27 shm->paddr = page_to_phys(page); in pool_op_alloc()
36 struct page **pages; in pool_op_alloc()
45 pages[i] = page + i; in pool_op_alloc()
58 __free_pages(page, order); in pool_op_alloc()
84 * optee_shm_pool_alloc_pages() - create page
[all...]
/kernel/linux/linux-5.10/fs/ocfs2/
H A Daops.h14 struct page *page,
18 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno,
22 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages);
43 struct page **pagep, void **fsdata,
44 struct buffer_head *di_bh, struct page *mmap_page);
46 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
/kernel/linux/linux-5.10/include/asm-generic/
H A Dpgalloc.h11 * __pte_alloc_one_kernel - allocate a page for PTE-level kernel page table
15 * anything beyond simple page allocation.
26 * pte_alloc_one_kernel - allocate a page for PTE-level kernel page table
38 * pte_free_kernel - free PTE-level kernel page table page
40 * @pte: pointer to the memory containing the page table
48 * __pte_alloc_one - allocate a page for PTE-level user page tabl
121 struct page *page; pmd_alloc_one() local
[all...]
/kernel/linux/linux-6.6/fs/ocfs2/
H A Daops.h12 struct page *page,
16 int ocfs2_map_page_blocks(struct page *page, u64 *p_blkno,
20 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages);
41 struct page **pagep, void **fsdata,
42 struct buffer_head *di_bh, struct page *mmap_page);
44 int ocfs2_read_inline_data(struct inode *inode, struct page *page,
/kernel/linux/linux-6.6/drivers/xen/
H A Dmem-reservation.c23 * Use one extent per PAGE_SIZE to avoid to break down the page into
30 struct page **pages, in __xenmem_reservation_va_mapping_update()
36 struct page *page = pages[i]; in __xenmem_reservation_va_mapping_update() local
37 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_update()
40 BUG_ON(!page); in __xenmem_reservation_va_mapping_update()
44 * different page granularity. in __xenmem_reservation_va_mapping_update()
59 struct page **pages) in __xenmem_reservation_va_mapping_reset()
64 struct page *page in __xenmem_reservation_va_mapping_reset() local
[all...]
/third_party/node/deps/v8/src/heap/
H A Dpaged-spaces-inl.h60 void PagedSpace::UnlinkFreeListCategories(Page* page) { in UnlinkFreeListCategories() argument
61 DCHECK_EQ(this, page->owner()); in UnlinkFreeListCategories()
62 page->ForAllFreeListCategories([this](FreeListCategory* category) { in UnlinkFreeListCategories()
67 size_t PagedSpace::RelinkFreeListCategories(Page* page) { in RelinkFreeListCategories() argument
68 DCHECK_EQ(this, page->owner()); in RelinkFreeListCategories()
70 page->ForAllFreeListCategories([this, &added](FreeListCategory* category) { in RelinkFreeListCategories()
75 DCHECK_IMPLIES(!page->IsFlagSet(Page::NEVER_ALLOCATE_ON_PAGE), in RelinkFreeListCategories()
76 page->AvailableInFreeList() == in RelinkFreeListCategories()
77 page->AvailableInFreeListFromAllocatedBytes()); in RelinkFreeListCategories()
/kernel/linux/linux-5.10/arch/xtensa/include/asm/
H A Dpgalloc.h26 #define pmd_populate(mm, pmdp, page) \
27 (pmd_val(*(pmdp)) = ((unsigned long)page_to_virt(page)))
57 struct page *page; in pte_alloc_one() local
59 page = __pte_alloc_one(mm, GFP_PGTABLE_USER); in pte_alloc_one()
60 if (!page) in pte_alloc_one()
62 ptes_clear(page_address(page)); in pte_alloc_one()
63 return page; in pte_alloc_one()

Completed in 11 milliseconds

1...<<21222324252627282930>>...269