Home
last modified time | relevance | path

Searched refs:npage (Results 1 - 25 of 28) sorted by relevance

12

/kernel/linux/linux-6.6/drivers/s390/cio/
H A Dvfio_ccw_cp.c94 int unpinned = 0, npage = 1; in page_array_unpin() local
98 dma_addr_t *last = &first[npage]; in page_array_unpin()
100 if (unpinned + npage < pa_nr && in page_array_unpin()
101 *first + npage * PAGE_SIZE == *last && in page_array_unpin()
103 npage++; in page_array_unpin()
107 vfio_unpin_pages(vdev, *first, npage); in page_array_unpin()
108 unpinned += npage; in page_array_unpin()
109 npage = 1; in page_array_unpin()
133 int pinned = 0, npage = 1; in page_array_pin() local
138 dma_addr_t *last = &first[npage]; in page_array_pin()
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dvfio.h98 int npage, int prot,
101 unsigned long *user_pfn, int npage);
132 int npage, int prot, unsigned long *phys_pfn);
134 int npage);
137 unsigned long *user_iova_pfn, int npage,
140 unsigned long *user_iova_pfn, int npage);
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/
H A Dhns_roce_mr.c577 int count, npage; in mtr_map_region() local
585 npage = 0; in mtr_map_region()
586 while (offset < end && npage < max_count) { in mtr_map_region()
593 for (i = 0; i < count && npage < max_count; i++) { in mtr_map_region()
594 addr = pages[npage]; in mtr_map_region()
597 npage++; in mtr_map_region()
602 return npage; in mtr_map_region()
703 int npage; in mtr_map_bufs() local
712 npage = hns_roce_get_umem_bufs(hr_dev, pages, page_count, in mtr_map_bufs()
715 npage in mtr_map_bufs()
808 u32 npage; hns_roce_mtr_find() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/gma500/
H A Dgtt.c95 set_pages_array_wc(pages, r->npage); in psb_gtt_insert()
99 for (i = r->roll; i < r->npage; i++) { in psb_gtt_insert()
137 for (i = 0; i < r->npage; i++) in psb_gtt_remove()
140 set_pages_array_wb(r->pages, r->npage); in psb_gtt_remove()
159 if (roll >= r->npage) { in psb_gtt_roll()
173 for (i = r->roll; i < r->npage; i++) { in psb_gtt_roll()
204 gt->npage = gt->gem.size / PAGE_SIZE; in psb_gtt_attach_pages()
255 gt->npage, 0, 0, PSB_MMU_CACHED_MEMORY); in psb_gtt_pin()
296 (gpu_base + gt->offset), gt->npage, 0, 0); in psb_gtt_unpin()
H A Dgtt.h39 int npage; /* Number of backing pages */ member
H A Dgma_display.c395 if (gt->npage > 4) in gma_crtc_cursor_set()
398 cursor_pages = gt->npage; in gma_crtc_cursor_set()
/kernel/linux/linux-6.6/drivers/vfio/
H A Dvfio_iommu_type1.c413 bool lock_cap, long npage) in mm_lock_acct()
420 ret = __account_locked_vm(mm, abs(npage), npage > 0, task, lock_cap); in mm_lock_acct()
425 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) in vfio_lock_acct() argument
430 if (!npage) in vfio_lock_acct()
437 ret = mm_lock_acct(dma->task, mm, dma->lock_cap, npage); in vfio_lock_acct()
439 dma->locked_vm += npage; in vfio_lock_acct()
614 long npage, unsigned long *pfn_base, in vfio_pin_pages_remote()
636 while (npage) { in vfio_pin_pages_remote()
639 long req_pages = min_t(long, npage, batc in vfio_pin_pages_remote()
412 mm_lock_acct(struct task_struct *task, struct mm_struct *mm, bool lock_cap, long npage) mm_lock_acct() argument
613 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) vfio_pin_pages_remote() argument
724 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) vfio_unpin_pages_remote() argument
796 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, dma_addr_t user_iova, int npage, int prot, struct page **pages) vfio_iommu_type1_pin_pages() argument
916 vfio_iommu_type1_unpin_pages(void *iommu_data, dma_addr_t user_iova, int npage) vfio_iommu_type1_unpin_pages() argument
1430 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument
1464 long npage; vfio_pin_map_dma() local
1529 long npage = dma->locked_vm; vfio_change_dma_owner() local
1736 long npage; vfio_iommu_replay() local
[all...]
H A Dvfio.h217 int npage, int prot,
220 dma_addr_t user_iova, int npage);
248 dma_addr_t iova, int npage,
251 dma_addr_t iova, int npage);
293 dma_addr_t iova, int npage, in vfio_device_container_pin_pages()
300 dma_addr_t iova, int npage) in vfio_device_container_unpin_pages()
292 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_device_container_pin_pages() argument
299 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) vfio_device_container_unpin_pages() argument
H A Dcontainer.c532 dma_addr_t iova, int npage, in vfio_device_container_pin_pages()
539 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_device_container_pin_pages()
545 npage, prot, pages); in vfio_device_container_pin_pages()
549 dma_addr_t iova, int npage) in vfio_device_container_unpin_pages()
553 if (WARN_ON(npage <= 0 || npage > VFIO_PIN_PAGES_MAX_ENTRIES)) in vfio_device_container_unpin_pages()
557 npage); in vfio_device_container_unpin_pages()
531 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_device_container_pin_pages() argument
548 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) vfio_device_container_unpin_pages() argument
H A Dvfio_main.c1524 * @npage [in] : count of pages to be pinned. This count should not
1534 int npage, int prot, struct page **pages) in vfio_pin_pages()
1537 if (!pages || !npage || WARN_ON(!vfio_assert_device_open(device))) in vfio_pin_pages()
1543 npage, prot, pages); in vfio_pin_pages()
1557 npage * PAGE_SIZE, pages, in vfio_pin_pages()
1561 return npage; in vfio_pin_pages()
1571 * @npage [in] : count of pages to be unpinned. This count should not
1574 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) in vfio_unpin_pages() argument
1582 vfio_device_container_unpin_pages(device, iova, npage); in vfio_unpin_pages()
1590 npage * PAGE_SIZ in vfio_unpin_pages()
1533 vfio_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) vfio_pin_pages() argument
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
H A Dhns_roce_mr.c637 int npage; in mtr_map_region() local
648 npage = 0; in mtr_map_region()
657 addr = to_hr_hw_page_addr(pages[npage]); in mtr_map_region()
659 addr = pages[npage]; in mtr_map_region()
662 npage++; in mtr_map_region()
820 int npage; in mtr_get_pages() local
824 npage = hns_roce_get_umem_bufs(hr_dev, pages, count, 0, in mtr_get_pages()
827 npage = hns_roce_get_kmem_bufs(hr_dev, pages, count, 0, in mtr_get_pages()
830 if (mtr->hem_cfg.is_direct && npage > 1) { in mtr_get_pages()
831 err = mtr_check_direct_pages(pages, npage, page_shif in mtr_get_pages()
889 int npage; hns_roce_mtr_find() local
[all...]
/kernel/linux/linux-5.10/drivers/vfio/
H A Dvfio_iommu_type1.c387 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) in vfio_lock_acct() argument
392 if (!npage) in vfio_lock_acct()
401 ret = __account_locked_vm(mm, abs(npage), npage > 0, dma->task, in vfio_lock_acct()
562 long npage, unsigned long *pfn_base, in vfio_pin_pages_remote()
600 for (vaddr += PAGE_SIZE, iova += PAGE_SIZE; pinned < npage; in vfio_pin_pages_remote()
643 unsigned long pfn, long npage, in vfio_unpin_pages_remote()
649 for (i = 0; i < npage; i++, iova += PAGE_SIZE) { in vfio_unpin_pages_remote()
717 int npage, int prot, in vfio_iommu_type1_pin_pages()
749 for (i = 0; i < npage; in vfio_iommu_type1_pin_pages()
561 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) vfio_pin_pages_remote() argument
642 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) vfio_unpin_pages_remote() argument
714 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, unsigned long *user_pfn, int npage, int prot, unsigned long *phys_pfn) vfio_iommu_type1_pin_pages() argument
820 vfio_iommu_type1_unpin_pages(void *iommu_data, unsigned long *user_pfn, int npage) vfio_iommu_type1_unpin_pages() argument
1296 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument
1329 long npage; vfio_pin_map_dma() local
1560 long npage; vfio_iommu_replay() local
[all...]
H A Dvfio.c1889 * @npage [in] : count of elements in user_pfn array. This count should not
1895 int vfio_pin_pages(struct device *dev, unsigned long *user_pfn, int npage, in vfio_pin_pages() argument
1903 if (!dev || !user_pfn || !phys_pfn || !npage) in vfio_pin_pages()
1906 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_pin_pages()
1927 npage, prot, phys_pfn); in vfio_pin_pages()
1944 * @npage [in] : count of elements in user_pfn array. This count should not
1948 int vfio_unpin_pages(struct device *dev, unsigned long *user_pfn, int npage) in vfio_unpin_pages() argument
1955 if (!dev || !user_pfn || !npage) in vfio_unpin_pages()
1958 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) in vfio_unpin_pages()
1973 npage); in vfio_unpin_pages()
2006 vfio_group_pin_pages(struct vfio_group *group, unsigned long *user_iova_pfn, int npage, int prot, unsigned long *phys_pfn) vfio_group_pin_pages() argument
2054 vfio_group_unpin_pages(struct vfio_group *group, unsigned long *user_iova_pfn, int npage) vfio_group_unpin_pages() argument
[all...]
/kernel/linux/linux-5.10/fs/f2fs/
H A Dnode.c725 struct page *npage[4]; in f2fs_get_dnode_of_data() local
738 npage[0] = dn->inode_page; in f2fs_get_dnode_of_data()
740 if (!npage[0]) { in f2fs_get_dnode_of_data()
741 npage[0] = f2fs_get_node_page(sbi, nids[0]); in f2fs_get_dnode_of_data()
742 if (IS_ERR(npage[0])) in f2fs_get_dnode_of_data()
743 return PTR_ERR(npage[0]); in f2fs_get_dnode_of_data()
749 f2fs_put_page(npage[0], 1); in f2fs_get_dnode_of_data()
753 parent = npage[0]; in f2fs_get_dnode_of_data()
756 dn->inode_page = npage[0]; in f2fs_get_dnode_of_data()
771 npage[ in f2fs_get_dnode_of_data()
1140 struct page *npage; f2fs_truncate_xattr_node() local
[all...]
H A Dinline.c274 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) in f2fs_recover_inline_data() argument
289 if (IS_INODE(npage)) in f2fs_recover_inline_data()
290 ri = F2FS_INODE(npage); in f2fs_recover_inline_data()
301 src_addr = inline_data_addr(inode, npage); in f2fs_recover_inline_data()
/kernel/linux/linux-6.6/fs/f2fs/
H A Dnode.c764 struct page *npage[4]; in f2fs_get_dnode_of_data() local
777 npage[0] = dn->inode_page; in f2fs_get_dnode_of_data()
779 if (!npage[0]) { in f2fs_get_dnode_of_data()
780 npage[0] = f2fs_get_node_page(sbi, nids[0]); in f2fs_get_dnode_of_data()
781 if (IS_ERR(npage[0])) in f2fs_get_dnode_of_data()
782 return PTR_ERR(npage[0]); in f2fs_get_dnode_of_data()
788 f2fs_put_page(npage[0], 1); in f2fs_get_dnode_of_data()
792 parent = npage[0]; in f2fs_get_dnode_of_data()
795 dn->inode_page = npage[0]; in f2fs_get_dnode_of_data()
810 npage[ in f2fs_get_dnode_of_data()
1217 struct page *npage; f2fs_truncate_xattr_node() local
[all...]
H A Dinline.c275 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) in f2fs_recover_inline_data() argument
290 if (IS_INODE(npage)) in f2fs_recover_inline_data()
291 ri = F2FS_INODE(npage); in f2fs_recover_inline_data()
302 src_addr = inline_data_addr(inode, npage); in f2fs_recover_inline_data()
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gvt/
H A Dkvmgt.c157 int npage; in gvt_unpin_guest_page() local
162 for (npage = 0; npage < total_pages; npage++) { in gvt_unpin_guest_page()
163 unsigned long cur_gfn = gfn + npage; in gvt_unpin_guest_page()
177 int npage; in gvt_pin_guest_page() local
185 for (npage = 0; npage < total_pages; npage++) { in gvt_pin_guest_page()
186 unsigned long cur_gfn = gfn + npage; in gvt_pin_guest_page()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() local
165 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), in mthca_array_init()
170 for (i = 0; i < npage; ++i) { in mthca_array_init()
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() local
160 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), in mthca_array_init()
165 for (i = 0; i < npage; ++i) { in mthca_array_init()
/kernel/linux/linux-5.10/lib/
H A Dkfifo.c315 struct page *npage; in setup_sgl_buf() local
319 npage = virt_to_page(buf); in setup_sgl_buf()
320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf()
325 page = npage; in setup_sgl_buf()
/kernel/linux/linux-6.6/lib/
H A Dkfifo.c315 struct page *npage; in setup_sgl_buf() local
319 npage = virt_to_page(buf); in setup_sgl_buf()
320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf()
325 page = npage; in setup_sgl_buf()
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gvt/
H A Dkvmgt.c141 int npage; in gvt_pin_guest_page() local
148 for (npage = 0; npage < total_pages; npage++) { in gvt_pin_guest_page()
149 dma_addr_t cur_iova = (gfn + npage) << PAGE_SHIFT; in gvt_pin_guest_page()
160 if (npage == 0) in gvt_pin_guest_page()
162 else if (page_to_pfn(base_page) + npage != page_to_pfn(cur_page)) { in gvt_pin_guest_page()
164 npage++; in gvt_pin_guest_page()
172 if (npage) in gvt_pin_guest_page()
173 gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZ in gvt_pin_guest_page()
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dvfio.h315 int npage, int prot, struct page **pages);
316 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage);
/kernel/linux/linux-6.6/drivers/iommu/iommufd/
H A Dpages.c671 unsigned int npage = 0; in batch_rw() local
677 copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data, in batch_rw()
683 npage++; in batch_rw()
684 if (npage == batch->npfns[cur]) { in batch_rw()
685 npage = 0; in batch_rw()

Completed in 34 milliseconds

12