/kernel/linux/linux-6.6/drivers/s390/cio/ |
H A D | vfio_ccw_cp.c | 94 int unpinned = 0, npage = 1; in page_array_unpin() local 98 dma_addr_t *last = &first[npage]; in page_array_unpin() 100 if (unpinned + npage < pa_nr && in page_array_unpin() 101 *first + npage * PAGE_SIZE == *last && in page_array_unpin() 103 npage++; in page_array_unpin() 107 vfio_unpin_pages(vdev, *first, npage); in page_array_unpin() 108 unpinned += npage; in page_array_unpin() 109 npage = 1; in page_array_unpin() 133 int pinned = 0, npage = 1; in page_array_pin() local 138 dma_addr_t *last = &first[npage]; in page_array_pin() [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | vfio.h | 98 int npage, int prot, 101 unsigned long *user_pfn, int npage); 132 int npage, int prot, unsigned long *phys_pfn); 134 int npage); 137 unsigned long *user_iova_pfn, int npage, 140 unsigned long *user_iova_pfn, int npage);
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 577 int count, npage; in mtr_map_region() local 585 npage = 0; in mtr_map_region() 586 while (offset < end && npage < max_count) { in mtr_map_region() 593 for (i = 0; i < count && npage < max_count; i++) { in mtr_map_region() 594 addr = pages[npage]; in mtr_map_region() 597 npage++; in mtr_map_region() 602 return npage; in mtr_map_region() 703 int npage; in mtr_map_bufs() local 712 npage = hns_roce_get_umem_bufs(hr_dev, pages, page_count, in mtr_map_bufs() 715 npage in mtr_map_bufs() 808 u32 npage; hns_roce_mtr_find() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/gma500/ |
H A D | gtt.c | 95 set_pages_array_wc(pages, r->npage); in psb_gtt_insert() 99 for (i = r->roll; i < r->npage; i++) { in psb_gtt_insert() 137 for (i = 0; i < r->npage; i++) in psb_gtt_remove() 140 set_pages_array_wb(r->pages, r->npage); in psb_gtt_remove() 159 if (roll >= r->npage) { in psb_gtt_roll() 173 for (i = r->roll; i < r->npage; i++) { in psb_gtt_roll() 204 gt->npage = gt->gem.size / PAGE_SIZE; in psb_gtt_attach_pages() 255 gt->npage, 0, 0, PSB_MMU_CACHED_MEMORY); in psb_gtt_pin() 296 (gpu_base + gt->offset), gt->npage, 0, 0); in psb_gtt_unpin()
|
H A D | gtt.h | 39 int npage; /* Number of backing pages */ member
|
H A D | gma_display.c | 395 if (gt->npage > 4) in gma_crtc_cursor_set() 398 cursor_pages = gt->npage; in gma_crtc_cursor_set()
|
/kernel/linux/linux-6.6/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 413 bool lock_cap, long npage) in mm_lock_acct() 420 ret = __account_locked_vm(mm, abs(npage), npage > 0, task, lock_cap); in mm_lock_acct() 425 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) in vfio_lock_acct() argument 430 if (!npage) in vfio_lock_acct() 437 ret = mm_lock_acct(dma->task, mm, dma->lock_cap, npage); in vfio_lock_acct() 439 dma->locked_vm += npage; in vfio_lock_acct() 614 long npage, unsigned long *pfn_base, in vfio_pin_pages_remote() 636 while (npage) { in vfio_pin_pages_remote() 639 long req_pages = min_t(long, npage, batc in vfio_pin_pages_remote() 412 mm_lock_acct(struct task_struct *task, struct mm_struct *mm, bool lock_cap, long npage) mm_lock_acct() argument 613 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) vfio_pin_pages_remote() argument 724 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) vfio_unpin_pages_remote() argument 796 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, dma_addr_t user_iova, int npage, int prot, struct page **pages) vfio_iommu_type1_pin_pages() argument 916 vfio_iommu_type1_unpin_pages(void *iommu_data, dma_addr_t user_iova, int npage) vfio_iommu_type1_unpin_pages() argument 1430 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument 1464 long npage; vfio_pin_map_dma() local 1529 long npage = dma->locked_vm; vfio_change_dma_owner() local 1736 long npage; vfio_iommu_replay() local [all...] |
H A D | vfio.h | 217 int npage, int prot, 220 dma_addr_t user_iova, int npage); 248 dma_addr_t iova, int npage, 251 dma_addr_t iova, int npage); 293 dma_addr_t iova, int npage, in vfio_device_container_pin_pages() 300 dma_addr_t iova, int npage) in vfio_device_container_unpin_pages() 292 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_device_container_pin_pages() argument 299 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) vfio_device_container_unpin_pages() argument
|
H A D | container.c | 532 dma_addr_t iova, int npage, in vfio_device_container_pin_pages() 539 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_device_container_pin_pages() 545 npage, prot, pages); in vfio_device_container_pin_pages() 549 dma_addr_t iova, int npage) in vfio_device_container_unpin_pages() 553 if (WARN_ON(npage <= 0 || npage > VFIO_PIN_PAGES_MAX_ENTRIES)) in vfio_device_container_unpin_pages() 557 npage); in vfio_device_container_unpin_pages() 531 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_device_container_pin_pages() argument 548 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) vfio_device_container_unpin_pages() argument
|
H A D | vfio_main.c | 1524 * @npage [in] : count of pages to be pinned. This count should not 1534 int npage, int prot, struct page **pages) in vfio_pin_pages() 1537 if (!pages || !npage || WARN_ON(!vfio_assert_device_open(device))) in vfio_pin_pages() 1543 npage, prot, pages); in vfio_pin_pages() 1557 npage * PAGE_SIZE, pages, in vfio_pin_pages() 1561 return npage; in vfio_pin_pages() 1571 * @npage [in] : count of pages to be unpinned. This count should not 1574 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) in vfio_unpin_pages() argument 1582 vfio_device_container_unpin_pages(device, iova, npage); in vfio_unpin_pages() 1590 npage * PAGE_SIZ in vfio_unpin_pages() 1533 vfio_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_pin_pages() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 637 int npage; in mtr_map_region() local 648 npage = 0; in mtr_map_region() 657 addr = to_hr_hw_page_addr(pages[npage]); in mtr_map_region() 659 addr = pages[npage]; in mtr_map_region() 662 npage++; in mtr_map_region() 820 int npage; in mtr_get_pages() local 824 npage = hns_roce_get_umem_bufs(hr_dev, pages, count, 0, in mtr_get_pages() 827 npage = hns_roce_get_kmem_bufs(hr_dev, pages, count, 0, in mtr_get_pages() 830 if (mtr->hem_cfg.is_direct && npage > 1) { in mtr_get_pages() 831 err = mtr_check_direct_pages(pages, npage, page_shif in mtr_get_pages() 889 int npage; hns_roce_mtr_find() local [all...] |
/kernel/linux/linux-5.10/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 387 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) in vfio_lock_acct() argument 392 if (!npage) in vfio_lock_acct() 401 ret = __account_locked_vm(mm, abs(npage), npage > 0, dma->task, in vfio_lock_acct() 562 long npage, unsigned long *pfn_base, in vfio_pin_pages_remote() 600 for (vaddr += PAGE_SIZE, iova += PAGE_SIZE; pinned < npage; in vfio_pin_pages_remote() 643 unsigned long pfn, long npage, in vfio_unpin_pages_remote() 649 for (i = 0; i < npage; i++, iova += PAGE_SIZE) { in vfio_unpin_pages_remote() 717 int npage, int prot, in vfio_iommu_type1_pin_pages() 749 for (i = 0; i < npage; in vfio_iommu_type1_pin_pages() 561 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) vfio_pin_pages_remote() argument 642 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) vfio_unpin_pages_remote() argument 714 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, unsigned long *user_pfn, int npage, int prot, unsigned long *phys_pfn) vfio_iommu_type1_pin_pages() argument 820 vfio_iommu_type1_unpin_pages(void *iommu_data, unsigned long *user_pfn, int npage) vfio_iommu_type1_unpin_pages() argument 1296 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument 1329 long npage; vfio_pin_map_dma() local 1560 long npage; vfio_iommu_replay() local [all...] |
H A D | vfio.c | 1889 * @npage [in] : count of elements in user_pfn array. This count should not 1895 int vfio_pin_pages(struct device *dev, unsigned long *user_pfn, int npage, in vfio_pin_pages() argument 1903 if (!dev || !user_pfn || !phys_pfn || !npage) in vfio_pin_pages() 1906 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_pin_pages() 1927 npage, prot, phys_pfn); in vfio_pin_pages() 1944 * @npage [in] : count of elements in user_pfn array. This count should not 1948 int vfio_unpin_pages(struct device *dev, unsigned long *user_pfn, int npage) in vfio_unpin_pages() argument 1955 if (!dev || !user_pfn || !npage) in vfio_unpin_pages() 1958 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_unpin_pages() 1973 npage); in vfio_unpin_pages() 2006 vfio_group_pin_pages(struct vfio_group *group, unsigned long *user_iova_pfn, int npage, int prot, unsigned long *phys_pfn) vfio_group_pin_pages() argument 2054 vfio_group_unpin_pages(struct vfio_group *group, unsigned long *user_iova_pfn, int npage) vfio_group_unpin_pages() argument [all...] |
/kernel/linux/linux-5.10/fs/f2fs/ |
H A D | node.c | 725 struct page *npage[4]; in f2fs_get_dnode_of_data() local 738 npage[0] = dn->inode_page; in f2fs_get_dnode_of_data() 740 if (!npage[0]) { in f2fs_get_dnode_of_data() 741 npage[0] = f2fs_get_node_page(sbi, nids[0]); in f2fs_get_dnode_of_data() 742 if (IS_ERR(npage[0])) in f2fs_get_dnode_of_data() 743 return PTR_ERR(npage[0]); in f2fs_get_dnode_of_data() 749 f2fs_put_page(npage[0], 1); in f2fs_get_dnode_of_data() 753 parent = npage[0]; in f2fs_get_dnode_of_data() 756 dn->inode_page = npage[0]; in f2fs_get_dnode_of_data() 771 npage[ in f2fs_get_dnode_of_data() 1140 struct page *npage; f2fs_truncate_xattr_node() local [all...] |
H A D | inline.c | 274 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) in f2fs_recover_inline_data() argument 289 if (IS_INODE(npage)) in f2fs_recover_inline_data() 290 ri = F2FS_INODE(npage); in f2fs_recover_inline_data() 301 src_addr = inline_data_addr(inode, npage); in f2fs_recover_inline_data()
|
/kernel/linux/linux-6.6/fs/f2fs/ |
H A D | node.c | 764 struct page *npage[4]; in f2fs_get_dnode_of_data() local 777 npage[0] = dn->inode_page; in f2fs_get_dnode_of_data() 779 if (!npage[0]) { in f2fs_get_dnode_of_data() 780 npage[0] = f2fs_get_node_page(sbi, nids[0]); in f2fs_get_dnode_of_data() 781 if (IS_ERR(npage[0])) in f2fs_get_dnode_of_data() 782 return PTR_ERR(npage[0]); in f2fs_get_dnode_of_data() 788 f2fs_put_page(npage[0], 1); in f2fs_get_dnode_of_data() 792 parent = npage[0]; in f2fs_get_dnode_of_data() 795 dn->inode_page = npage[0]; in f2fs_get_dnode_of_data() 810 npage[ in f2fs_get_dnode_of_data() 1217 struct page *npage; f2fs_truncate_xattr_node() local [all...] |
H A D | inline.c | 275 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) in f2fs_recover_inline_data() argument 290 if (IS_INODE(npage)) in f2fs_recover_inline_data() 291 ri = F2FS_INODE(npage); in f2fs_recover_inline_data() 302 src_addr = inline_data_addr(inode, npage); in f2fs_recover_inline_data()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gvt/ |
H A D | kvmgt.c | 157 int npage; in gvt_unpin_guest_page() local 162 for (npage = 0; npage < total_pages; npage++) { in gvt_unpin_guest_page() 163 unsigned long cur_gfn = gfn + npage; in gvt_unpin_guest_page() 177 int npage; in gvt_pin_guest_page() local 185 for (npage = 0; npage < total_pages; npage++) { in gvt_pin_guest_page() 186 unsigned long cur_gfn = gfn + npage; in gvt_pin_guest_page() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() local 165 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), in mthca_array_init() 170 for (i = 0; i < npage; ++i) { in mthca_array_init()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() local 160 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), in mthca_array_init() 165 for (i = 0; i < npage; ++i) { in mthca_array_init()
|
/kernel/linux/linux-5.10/lib/ |
H A D | kfifo.c | 315 struct page *npage; in setup_sgl_buf() local 319 npage = virt_to_page(buf); in setup_sgl_buf() 320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf() 325 page = npage; in setup_sgl_buf()
|
/kernel/linux/linux-6.6/lib/ |
H A D | kfifo.c | 315 struct page *npage; in setup_sgl_buf() local 319 npage = virt_to_page(buf); in setup_sgl_buf() 320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf() 325 page = npage; in setup_sgl_buf()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gvt/ |
H A D | kvmgt.c | 141 int npage; in gvt_pin_guest_page() local 148 for (npage = 0; npage < total_pages; npage++) { in gvt_pin_guest_page() 149 dma_addr_t cur_iova = (gfn + npage) << PAGE_SHIFT; in gvt_pin_guest_page() 160 if (npage == 0) in gvt_pin_guest_page() 162 else if (page_to_pfn(base_page) + npage != page_to_pfn(cur_page)) { in gvt_pin_guest_page() 164 npage++; in gvt_pin_guest_page() 172 if (npage) in gvt_pin_guest_page() 173 gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZ in gvt_pin_guest_page() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | vfio.h | 315 int npage, int prot, struct page **pages); 316 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage);
|
/kernel/linux/linux-6.6/drivers/iommu/iommufd/ |
H A D | pages.c | 671 unsigned int npage = 0; in batch_rw() local 677 copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data, in batch_rw() 683 npage++; in batch_rw() 684 if (npage == batch->npfns[cur]) { in batch_rw() 685 npage = 0; in batch_rw()
|