/kernel/linux/linux-6.6/fs/ecryptfs/ |
H A D | read_write.c | 233 * @page_index: Page index in @page_for_ecryptfs from which to start 247 pgoff_t page_index, in ecryptfs_read_lower_page_segment() 255 offset = ((((loff_t)page_index) << PAGE_SHIFT) + offset_in_page); in ecryptfs_read_lower_page_segment() 246 ecryptfs_read_lower_page_segment(struct page *page_for_ecryptfs, pgoff_t page_index, size_t offset_in_page, size_t size, struct inode *ecryptfs_inode) ecryptfs_read_lower_page_segment() argument
|
/kernel/linux/linux-6.6/tools/testing/selftests/mm/ |
H A D | gup_test.c | 193 long page_index = strtol(argv[optind], 0, 0) + 1; in main() local 195 gup.which_pages[extra_arg_count] = page_index; in main()
|
/kernel/linux/common_modules/xpm/core/ |
H A D | xpm_report.c | 146 JSTR_PAIR(page_type, %s)", "JVAL_PAIR(page_index, %lu)", " in set_integrity_content() 149 info->page_type, info->page_index, info->vm_pgprot, in set_integrity_content() 189 info->page_index = param->page->index; in xpm_set_report_info()
|
H A D | xpm_report.h | 66 pgoff_t page_index; member
|
/kernel/linux/linux-5.10/include/linux/qed/ |
H A D | qed_chain.h | 285 u32 page_index = 0; in qed_chain_advance_page() local 304 page_index = *(u16 *)page_to_inc; in qed_chain_advance_page() 308 page_index = *(u32 *)page_to_inc; in qed_chain_advance_page() 310 *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr; in qed_chain_advance_page()
|
/kernel/linux/linux-6.6/include/linux/qed/ |
H A D | qed_chain.h | 286 u32 page_index = 0; in qed_chain_advance_page() local 305 page_index = *(u16 *)page_to_inc; in qed_chain_advance_page() 309 page_index = *(u32 *)page_to_inc; in qed_chain_advance_page() 311 *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr; in qed_chain_advance_page()
|
/kernel/linux/linux-5.10/fs/jfs/ |
H A D | jfs_metapage.c | 584 unsigned long page_index; in __get_metapage() local 592 page_index = lblock >> l2BlocksPerPage; in __get_metapage() 593 page_offset = (lblock - (page_index << l2BlocksPerPage)) << l2bsize; in __get_metapage() 614 page = grab_cache_page(mapping, page_index); in __get_metapage() 621 page = read_mapping_page(mapping, page_index, NULL); in __get_metapage()
|
/kernel/linux/linux-6.6/fs/jfs/ |
H A D | jfs_metapage.c | 583 unsigned long page_index; in __get_metapage() local 591 page_index = lblock >> l2BlocksPerPage; in __get_metapage() 592 page_offset = (lblock - (page_index << l2BlocksPerPage)) << l2bsize; in __get_metapage() 613 page = grab_cache_page(mapping, page_index); in __get_metapage() 620 page = read_mapping_page(mapping, page_index, NULL); in __get_metapage()
|
/kernel/linux/linux-6.6/arch/x86/kernel/cpu/sgx/ |
H A D | encl.h | 110 int sgx_encl_alloc_backing(struct sgx_encl *encl, unsigned long page_index,
|
H A D | main.c | 302 pgoff_t page_index; in sgx_reclaim_pages() local 334 page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base); in sgx_reclaim_pages() 337 ret = sgx_encl_alloc_backing(encl_page->encl, page_index, &backing[i]); in sgx_reclaim_pages()
|
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | scrub.c | 1201 int page_index; in scrub_handle_errored_block() local 1203 for (page_index = 0; page_index < sblock->page_count; in scrub_handle_errored_block() 1204 page_index++) { in scrub_handle_errored_block() 1205 sblock->pagev[page_index]->sblock = NULL; in scrub_handle_errored_block() 1206 recover = sblock->pagev[page_index]->recover; in scrub_handle_errored_block() 1209 sblock->pagev[page_index]->recover = in scrub_handle_errored_block() 1212 scrub_page_put(sblock->pagev[page_index]); in scrub_handle_errored_block() 1281 int page_index = 0; in scrub_setup_recheck_block() local 1321 BUG_ON(page_index > in scrub_setup_recheck_block() [all...] |
/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_buf.c | 1482 int page_index; in xfs_buf_ioapply_map() local 1491 page_index = 0; in xfs_buf_ioapply_map() 1494 page_index++; in xfs_buf_ioapply_map() 1517 for (; size && nr_pages; nr_pages--, page_index++) { in xfs_buf_ioapply_map() 1523 rbytes = bio_add_page(bio, bp->b_pages[page_index], nbytes, in xfs_buf_ioapply_map() 1741 int page_index, page_offset, csize; in xfs_buf_zero() local 1743 page_index = (boff + bp->b_offset) >> PAGE_SHIFT; in xfs_buf_zero() 1745 page = bp->b_pages[page_index]; in xfs_buf_zero()
|
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_buf.c | 1461 int page_index; in xfs_buf_ioapply_map() local 1470 page_index = 0; in xfs_buf_ioapply_map() 1473 page_index++; in xfs_buf_ioapply_map() 1494 for (; size && nr_pages; nr_pages--, page_index++) { in xfs_buf_ioapply_map() 1500 rbytes = bio_add_page(bio, bp->b_pages[page_index], nbytes, in xfs_buf_ioapply_map() 1733 int page_index, page_offset, csize; in xfs_buf_zero() local 1735 page_index = (boff + bp->b_offset) >> PAGE_SHIFT; in xfs_buf_zero() 1737 page = bp->b_pages[page_index]; in xfs_buf_zero()
|
/kernel/linux/linux-5.10/drivers/misc/vmw_vmci/ |
H A D | vmci_queue_pair.c | 338 const u64 page_index = in qp_memcpy_to_queue_iter() local 346 va = kmap(kernel_if->u.h.page[page_index]); in qp_memcpy_to_queue_iter() 348 va = kernel_if->u.g.vas[page_index + 1]; in qp_memcpy_to_queue_iter() 360 kunmap(kernel_if->u.h.page[page_index]); in qp_memcpy_to_queue_iter() 365 kunmap(kernel_if->u.h.page[page_index]); in qp_memcpy_to_queue_iter() 385 const u64 page_index = in qp_memcpy_from_queue_iter() local 394 va = kmap(kernel_if->u.h.page[page_index]); in qp_memcpy_from_queue_iter() 396 va = kernel_if->u.g.vas[page_index + 1]; in qp_memcpy_from_queue_iter() 408 kunmap(kernel_if->u.h.page[page_index]); in qp_memcpy_from_queue_iter() 413 kunmap(kernel_if->u.h.page[page_index]); in qp_memcpy_from_queue_iter() [all...] |
/kernel/linux/linux-5.10/fs/hmdfs/ |
H A D | hmdfs_client.c | 330 unsigned long page_index = ctx->page->index; in hmdfs_writepage_cb() local 332 trace_hmdfs_writepage_cb_enter(peer, info->remote_ino, page_index, ret); in hmdfs_writepage_cb() 355 trace_hmdfs_writepage_cb_exit(peer, info->remote_ino, page_index, ret); in hmdfs_writepage_cb() 401 unsigned long page_index = page->index; in hmdfs_client_recv_readpage() local 413 info->remote_ino, page_index, ret); in hmdfs_client_recv_readpage()
|
/kernel/linux/linux-6.6/fs/hmdfs/ |
H A D | hmdfs_client.c | 329 unsigned long page_index = ctx->page->index; in hmdfs_writepage_cb() local 331 trace_hmdfs_writepage_cb_enter(peer, info->remote_ino, page_index, ret); in hmdfs_writepage_cb() 354 trace_hmdfs_writepage_cb_exit(peer, info->remote_ino, page_index, ret); in hmdfs_writepage_cb() 400 unsigned long page_index = page->index; in hmdfs_client_recv_readpage() local 412 info->remote_ino, page_index, ret); in hmdfs_client_recv_readpage()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem_linux.c | 2185 unsigned long page_index; local 2208 page_index = (gpu_addr >> PAGE_SHIFT) - reg->start_pfn; 2210 /* check if page_index + page_count will wrap */ 2211 if (-1UL - page_count < page_index) 2214 if (page_index + page_count > kbase_reg_current_backed_size(reg)) 2234 pages[i] = pfn_to_page(PFN_DOWN(page_array[page_index + i])); 2254 map->cpu_pages = &kbase_get_cpu_phy_pages(reg)[page_index]; 2256 map->gpu_pages = &kbase_get_gpu_phy_pages(reg)[page_index];
|
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | raid56.c | 248 int page_index = offset >> PAGE_SHIFT; in index_stripe_sectors() local 250 ASSERT(page_index < rbio->nr_pages); in index_stripe_sectors() 251 rbio->stripe_sectors[i].page = rbio->stripe_pages[page_index]; in index_stripe_sectors() 2751 const int page_index = offset_in_full_stripe >> PAGE_SHIFT; in raid56_parity_cache_data_pages() local 2773 struct page *dst = rbio->stripe_pages[page_nr + page_index]; in raid56_parity_cache_data_pages() 2777 for (int sector_nr = sectors_per_page * page_index; in raid56_parity_cache_data_pages() 2778 sector_nr < sectors_per_page * (page_index + 1); in raid56_parity_cache_data_pages()
|
/kernel/linux/linux-5.10/fs/nfs/ |
H A D | dir.c | 151 unsigned long page_index; member 360 desc->page_index++; in nfs_readdir_search_array() 478 nfsi->page_index + 1, -1); in nfs_force_use_readdirplus() 742 return read_cache_page(desc->file->f_mapping, desc->page_index, in get_cache_page() 747 * Returns 0 if desc->dir_cookie was found on page desc->page_index 767 nfsi->page_index = desc->page_index; in find_and_lock_cache_page() 783 if (desc->page_index == 0) { in readdir_search_pagecache() 865 desc->page_index = 0; in uncached_readdir() 935 desc->page_index in nfs_readdir() [all...] |
/kernel/linux/linux-5.10/fs/nilfs2/ |
H A D | btnode.c | 149 pgoff_t index = page_index(page); in nilfs_btnode_delete()
|
/kernel/linux/linux-6.6/fs/nilfs2/ |
H A D | btnode.c | 149 pgoff_t index = page_index(page); in nilfs_btnode_delete()
|
/third_party/skia/third_party/externals/piex/src/binary_parse/ |
H A D | range_checked_byte_ptr.h | 83 // void getPage(size_t page_index, const unsigned char** begin, 121 // with index "page_index". 138 virtual void getPage(size_t page_index, const unsigned char **begin,
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | user_sdma.c | 1520 size_t page_index; in add_mapping_to_sdma_packet() local 1530 page_index = PFN_DOWN(start - cache_entry->rb.addr); in add_mapping_to_sdma_packet() 1532 if (page_index >= cache_entry->npages) { in add_mapping_to_sdma_packet() 1534 "Request for page_index %zu >= cache_entry->npages %u", in add_mapping_to_sdma_packet() 1535 page_index, cache_entry->npages); in add_mapping_to_sdma_packet() 1555 cache_entry->pages[page_index], in add_mapping_to_sdma_packet() 1566 "sdma_txadd_page failed %d page_index %lu page_offset %u from_this_page %u", in add_mapping_to_sdma_packet() 1567 ret, page_index, page_offset, from_this_page); in add_mapping_to_sdma_packet()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | pagemap.h | 436 return page_index(head) == (index & ~(thp_nr_pages(head) - 1UL)); in thp_contains() 546 return ((loff_t)page_index(page)) << PAGE_SHIFT; in page_file_offset()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem_linux.c | 2177 unsigned long page_index;
local 2203 page_index = (gpu_addr >> PAGE_SHIFT) - reg->start_pfn;
2205 /* check if page_index + page_count will wrap */
2206 if (-1UL - page_count < page_index) {
2210 if (page_index + page_count > kbase_reg_current_backed_size(reg)) {
2235 pages[i] = pfn_to_page(PFN_DOWN(page_array[page_index + i]));
2257 map->cpu_pages = &kbase_get_cpu_phy_pages(reg)[page_index];
2259 map->gpu_pages = &kbase_get_gpu_phy_pages(reg)[page_index];
|