/kernel/linux/linux-6.6/tools/testing/selftests/mm/ |
H A D | mremap_dontunmap.c | 45 unsigned long num_pages = 1; in kernel_support_for_mremap_dontunmap() local 46 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, in kernel_support_for_mremap_dontunmap() 53 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap() 58 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 62 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 78 unsigned long num_pages = size / page_size; in check_region_contains_byte() local 82 for (i = 0; i < num_pages; ++i) { in check_region_contains_byte() 97 unsigned long num_pages = 5; in mremap_dontunmap_simple() local 100 mmap(NULL, num_pages * page_siz in mremap_dontunmap_simple() 130 unsigned long num_pages = 5; mremap_dontunmap_simple_shmem() local 183 unsigned long num_pages = 5; mremap_dontunmap_simple_fixed() local 242 unsigned long num_pages = 10; mremap_dontunmap_partial_mapping() local [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/vm/ |
H A D | mremap_dontunmap.c | 48 unsigned long num_pages = 1; in kernel_support_for_mremap_dontunmap() local 49 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, in kernel_support_for_mremap_dontunmap() 56 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap() 61 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 65 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 81 unsigned long num_pages = size / page_size; in check_region_contains_byte() local 85 for (i = 0; i < num_pages; ++i) { in check_region_contains_byte() 100 unsigned long num_pages = 5; in mremap_dontunmap_simple() local 103 mmap(NULL, num_pages * page_siz in mremap_dontunmap_simple() 135 unsigned long num_pages = 5; mremap_dontunmap_simple_fixed() local 194 unsigned long num_pages = 10; mremap_dontunmap_partial_mapping() local [all...] |
/kernel/linux/linux-5.10/drivers/staging/gasket/ |
H A D | gasket_page_table.c | 468 uint num_pages, int is_simple_mapping) in gasket_perform_mapping() 477 for (i = 0; i < num_pages; i++) { in gasket_perform_mapping() 576 ulong dev_addr, uint num_pages) in gasket_alloc_simple_entries() 580 num_pages)) in gasket_alloc_simple_entries() 592 u64 __iomem *slots, uint num_pages, in gasket_perform_unmapping() 600 for (i = 0; i < num_pages; i++) { in gasket_perform_unmapping() 630 ulong dev_addr, uint num_pages) in gasket_unmap_simple_pages() 635 pg_tbl->base_slot + slot, num_pages, 1); in gasket_unmap_simple_pages() 643 ulong dev_addr, uint num_pages) in gasket_unmap_extended_pages() 649 remain = num_pages; in gasket_unmap_extended_pages() 465 gasket_perform_mapping(struct gasket_page_table *pg_tbl, struct gasket_page_table_entry *ptes, u64 __iomem *slots, ulong host_addr, uint num_pages, int is_simple_mapping) gasket_perform_mapping() argument 575 gasket_alloc_simple_entries(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_alloc_simple_entries() argument 590 gasket_perform_unmapping(struct gasket_page_table *pg_tbl, struct gasket_page_table_entry *ptes, u64 __iomem *slots, uint num_pages, int is_simple_mapping) gasket_perform_unmapping() argument 629 gasket_unmap_simple_pages(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_unmap_simple_pages() argument 642 gasket_unmap_extended_pages(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_unmap_extended_pages() argument 707 gasket_is_simple_dev_addr_bad(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_is_simple_dev_addr_bad() argument 745 gasket_is_extended_dev_addr_bad(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_is_extended_dev_addr_bad() argument 801 gasket_page_table_unmap_nolock(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_page_table_unmap_nolock() argument 817 gasket_map_simple_pages(struct gasket_page_table *pg_tbl, ulong host_addr, ulong dev_addr, uint num_pages) gasket_map_simple_pages() argument 953 gasket_map_extended_pages(struct gasket_page_table *pg_tbl, ulong host_addr, ulong dev_addr, uint num_pages) gasket_map_extended_pages() argument 1012 gasket_page_table_map(struct gasket_page_table *pg_tbl, ulong host_addr, ulong dev_addr, uint num_pages) gasket_page_table_map() argument 1044 gasket_page_table_unmap(struct gasket_page_table *pg_tbl, ulong dev_addr, uint num_pages) gasket_page_table_unmap() argument 1153 uint num_pages = bytes / PAGE_SIZE; gasket_page_table_is_dev_addr_bad() local 1232 unsigned int num_pages = size / PAGE_SIZE; gasket_set_user_virt() local 1258 unsigned int num_pages = DIV_ROUND_UP(size, PAGE_SIZE); gasket_alloc_coherent_memory() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, in __qib_release_user_pages() argument 43 unpin_user_pages_dirty_lock(p, num_pages, dirty); in __qib_release_user_pages() 85 * @num_pages: the number of pages 90 * now, num_pages is always 1, but that will probably change at some point 94 int qib_get_user_pages(unsigned long start_page, size_t num_pages, in qib_get_user_pages() argument 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 110 for (got = 0; got < num_pages; got += ret) { in qib_get_user_pages() 112 num_pages - got, in qib_get_user_pages() 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 130 void qib_release_user_pages(struct page **p, size_t num_pages) in qib_release_user_pages() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, in __qib_release_user_pages() argument 43 unpin_user_pages_dirty_lock(p, num_pages, dirty); in __qib_release_user_pages() 85 * @num_pages: the number of pages 90 * now, num_pages is always 1, but that will probably change at some point 94 int qib_get_user_pages(unsigned long start_page, size_t num_pages, in qib_get_user_pages() argument 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 110 for (got = 0; got < num_pages; got += ret) { in qib_get_user_pages() 112 num_pages - got, in qib_get_user_pages() 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 130 void qib_release_user_pages(struct page **p, size_t num_pages) in qib_release_user_pages() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 167 unsigned int num_pages = last - first; in ttm_pool_apply_caching() local 169 if (!num_pages) in ttm_pool_apply_caching() 176 return set_pages_array_wc(first, num_pages); in ttm_pool_apply_caching() 178 return set_pages_array_uc(first, num_pages); in ttm_pool_apply_caching() 213 unsigned int num_pages) in ttm_pool_unmap() 219 dma_unmap_page(pool->dev, dma_addr, (long)num_pages << PAGE_SHIFT, in ttm_pool_unmap() 226 unsigned int i, num_pages = 1 << pt->order; in ttm_pool_type_give() local 228 for (i = 0; i < num_pages; ++i) { in ttm_pool_type_give() 317 unsigned int num_pages; in ttm_pool_shrink() local 328 num_pages in ttm_pool_shrink() 212 ttm_pool_unmap(struct ttm_pool *pool, dma_addr_t dma_addr, unsigned int num_pages) ttm_pool_unmap() argument 349 ttm_pool_page_allocated(struct ttm_pool *pool, unsigned int order, struct page *p, dma_addr_t **dma_addr, unsigned long *num_pages, struct page ***pages) ttm_pool_page_allocated() argument 422 pgoff_t num_pages = tt->num_pages; ttm_pool_alloc() local 617 unsigned long num_pages = atomic_long_read(&allocated_pages); ttm_pool_shrinker_count() local 754 ttm_pool_mgr_init(unsigned long num_pages) ttm_pool_mgr_init() argument [all...] |
H A D | ttm_tt.c | 100 ttm->pages = kvcalloc(ttm->num_pages, sizeof(void*), GFP_KERNEL); in ttm_tt_alloc_page_directory() 109 ttm->pages = kvcalloc(ttm->num_pages, sizeof(*ttm->pages) + in ttm_dma_tt_alloc_page_directory() 114 ttm->dma_address = (void *)(ttm->pages + ttm->num_pages); in ttm_dma_tt_alloc_page_directory() 120 ttm->dma_address = kvcalloc(ttm->num_pages, sizeof(*ttm->dma_address), in ttm_sg_tt_alloc_page_directory() 139 ttm->num_pages = (PAGE_ALIGN(bo->base.size) >> PAGE_SHIFT) + extra_pages; in ttm_tt_init_fields() 212 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin() 252 loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT; in ttm_tt_swapout() 268 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout() 288 return ttm->num_pages; in ttm_tt_swapout() 308 atomic_long_add(ttm->num_pages, in ttm_tt_populate() 394 ttm_tt_mgr_init(unsigned long num_pages, unsigned long num_dma32_pages) ttm_tt_mgr_init() argument [all...] |
/kernel/linux/linux-5.10/drivers/staging/vc04_services/interface/vchiq_arm/ |
H A D | vchiq_2835_arm.c | 42 unsigned int num_pages; member 288 pagelistinfo->num_pages, pagelistinfo->dma_dir); in cleanup_pagelistinfo() 292 unpin_user_pages(pagelistinfo->pages, pagelistinfo->num_pages); in cleanup_pagelistinfo() 314 unsigned int num_pages, offset, i, k; in create_pagelist() local 328 num_pages = DIV_ROUND_UP(count + offset, PAGE_SIZE); in create_pagelist() 330 if (num_pages > (SIZE_MAX - sizeof(struct pagelist) - in create_pagelist() 337 (num_pages * sizeof(u32)) + in create_pagelist() 338 (num_pages * sizeof(pages[0]) + in create_pagelist() 339 (num_pages * sizeof(struct scatterlist))) + in create_pagelist() 354 pages = (struct page **)(addrs + num_pages); in create_pagelist() 503 unsigned int num_pages = pagelistinfo->num_pages; free_pagelist() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_gmr.c | 40 unsigned long num_pages, in vmw_gmr2_bind() 48 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind() 49 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind() 59 define_cmd.numPages = num_pages; in vmw_gmr2_bind() 74 while (num_pages > 0) { in vmw_gmr2_bind() 75 unsigned long nr = min(num_pages, (unsigned long)VMW_PPN_PER_REMAP); in vmw_gmr2_bind() 95 num_pages -= nr; in vmw_gmr2_bind() 129 unsigned long num_pages, in vmw_gmr_bind() 142 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_i in vmw_gmr_bind() 38 vmw_gmr2_bind(struct vmw_private *dev_priv, struct vmw_piter *iter, unsigned long num_pages, int gmr_id) vmw_gmr2_bind() argument 127 vmw_gmr_bind(struct vmw_private *dev_priv, const struct vmw_sg_table *vsgt, unsigned long num_pages, int gmr_id) vmw_gmr_bind() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_gmr.c | 39 unsigned long num_pages, in vmw_gmr2_bind() 47 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind() 48 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind() 58 define_cmd.numPages = num_pages; in vmw_gmr2_bind() 73 while (num_pages > 0) { in vmw_gmr2_bind() 74 unsigned long nr = min_t(unsigned long, num_pages, VMW_PPN_PER_REMAP); in vmw_gmr2_bind() 94 num_pages -= nr; in vmw_gmr2_bind() 128 unsigned long num_pages, in vmw_gmr_bind() 141 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_i in vmw_gmr_bind() 37 vmw_gmr2_bind(struct vmw_private *dev_priv, struct vmw_piter *iter, unsigned long num_pages, int gmr_id) vmw_gmr2_bind() argument 126 vmw_gmr_bind(struct vmw_private *dev_priv, const struct vmw_sg_table *vsgt, unsigned long num_pages, int gmr_id) vmw_gmr_bind() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/xen/ |
H A D | xen_drm_front_gem.c | 29 size_t num_pages; member 48 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); in gem_alloc_pages_array() 49 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array() 103 ret = xen_alloc_unpopulated_pages(xen_obj->num_pages, in gem_create() 107 xen_obj->num_pages, ret); in gem_create() 119 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); in gem_create() 156 xen_free_unpopulated_pages(xen_obj->num_pages, in xen_drm_front_gem_free_object_unlocked() 184 xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_get_sg_table() 209 NULL, xen_obj->num_pages); in xen_drm_front_gem_import_sg_table() 256 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); in gem_mmap_obj() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/xen/ |
H A D | xen_drm_front_gem.c | 28 size_t num_pages; member 47 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); in gem_alloc_pages_array() 48 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array() 92 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_object_mmap() 158 ret = xen_alloc_unpopulated_pages(xen_obj->num_pages, in gem_create() 162 xen_obj->num_pages, ret); in gem_create() 174 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); in gem_create() 211 xen_free_unpopulated_pages(xen_obj->num_pages, in xen_drm_front_gem_free_object_unlocked() 239 xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_get_sg_table() 264 xen_obj->num_pages); in xen_drm_front_gem_import_sg_table() [all...] |
/kernel/linux/linux-6.6/tools/testing/scatterlist/ |
H A D | main.c | 10 unsigned num_pages; member 40 printf("%u input PFNs:", test->num_pages); in fail() 41 for (i = 0; i < test->num_pages; i++) in fail() 87 int left_pages = test->pfn_app ? test->num_pages : 0; in main() 92 set_pages(pages, test->pfn, test->num_pages); in main() 96 &append, pages, test->num_pages, 0, test->size, in main() 100 &append.sgt, pages, test->num_pages, 0, in main() 109 set_pages(pages, test->pfn_app, test->num_pages); in main() 111 &append, pages, test->num_pages, 0, test->size, in main()
|
/kernel/linux/linux-6.6/drivers/infiniband/sw/siw/ |
H A D | siw_mem.c | 63 static void siw_free_plist(struct siw_page_chunk *chunk, int num_pages, in siw_free_plist() argument 66 unpin_user_pages_dirty_lock(chunk->plist, num_pages, dirty); in siw_free_plist() 72 int i, num_pages = umem->num_pages; in siw_umem_release() local 74 for (i = 0; num_pages; i++) { in siw_umem_release() 75 int to_free = min_t(int, PAGES_PER_CHUNK, num_pages); in siw_umem_release() 80 num_pages -= to_free; in siw_umem_release() 82 atomic64_sub(umem->num_pages, &mm_s->pinned_vm); in siw_umem_release() 372 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 381 num_pages in siw_umem_get() [all...] |
/kernel/linux/linux-6.6/arch/x86/platform/efi/ |
H A D | memmap.c | 118 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_split_count() 183 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_insert() 192 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert() 199 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert() 205 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 213 md->num_pages = (m_end - m_start + 1) >> in efi_memmap_insert() 220 md->num_pages = (end - m_end) >> in efi_memmap_insert() 227 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 234 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert()
|
/kernel/linux/linux-5.10/drivers/media/common/videobuf2/ |
H A D | videobuf2-dma-sg.c | 49 unsigned int num_pages; member 106 int num_pages; in vb2_dma_sg_alloc() local 120 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc() 128 buf->pages = kvmalloc_array(buf->num_pages, sizeof(struct page *), in vb2_dma_sg_alloc() 138 buf->num_pages, 0, size, GFP_KERNEL); in vb2_dma_sg_alloc() 161 __func__, buf->num_pages); in vb2_dma_sg_alloc() 168 num_pages = buf->num_pages; in vb2_dma_sg_alloc() 169 while (num_pages--) in vb2_dma_sg_alloc() 170 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc() [all...] |
/kernel/linux/linux-6.6/drivers/media/common/videobuf2/ |
H A D | videobuf2-dma-sg.c | 49 unsigned int num_pages; member 107 int num_pages; in vb2_dma_sg_alloc() local 121 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc() 129 buf->pages = kvcalloc(buf->num_pages, sizeof(struct page *), GFP_KERNEL); in vb2_dma_sg_alloc() 138 buf->num_pages, 0, size, GFP_KERNEL); in vb2_dma_sg_alloc() 162 __func__, buf->num_pages); in vb2_dma_sg_alloc() 169 num_pages = buf->num_pages; in vb2_dma_sg_alloc() 170 while (num_pages--) in vb2_dma_sg_alloc() 171 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
H A D | siw_mem.c | 63 static void siw_free_plist(struct siw_page_chunk *chunk, int num_pages, in siw_free_plist() argument 66 unpin_user_pages_dirty_lock(chunk->plist, num_pages, dirty); in siw_free_plist() 72 int i, num_pages = umem->num_pages; in siw_umem_release() local 74 for (i = 0; num_pages; i++) { in siw_umem_release() 75 int to_free = min_t(int, PAGES_PER_CHUNK, num_pages); in siw_umem_release() 80 num_pages -= to_free; in siw_umem_release() 82 atomic64_sub(umem->num_pages, &mm_s->pinned_vm); in siw_umem_release() 372 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 381 num_pages in siw_umem_get() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_cache.c | 61 unsigned long num_pages) in drm_cache_flush_clflush() 66 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush() 75 * @num_pages: Number of pages in the array. 81 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument 86 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages() 96 for (i = 0; i < num_pages; i++) { in drm_clflush_pages() 60 drm_cache_flush_clflush(struct page *pages[], unsigned long num_pages) drm_cache_flush_clflush() argument
|
/kernel/linux/linux-5.10/net/ceph/ |
H A D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument 17 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector() 26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument 30 for (i = 0; i < num_pages; i++) in ceph_release_page_vector() 39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 47 for (i = 0; i < num_pages; i++) { in ceph_alloc_page_vector()
|
/kernel/linux/linux-6.6/net/ceph/ |
H A D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument 17 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector() 26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument 30 for (i = 0; i < num_pages; i++) in ceph_release_page_vector() 39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 47 for (i = 0; i < num_pages; i++) { in ceph_alloc_page_vector()
|
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | xen-front-pgdir-shbuf.c | 169 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); in get_num_pages_dir() 196 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; in guest_calc_num_grefs() 217 unmap_ops = kcalloc(buf->num_pages, sizeof(*unmap_ops), in backend_unmap() 222 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 231 buf->num_pages); in backend_unmap() 233 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 262 map_ops = kcalloc(buf->num_pages, sizeof(*map_ops), GFP_KERNEL); in backend_map() 266 buf->backend_map_handles = kcalloc(buf->num_pages, in backend_map() 280 grefs_left = buf->num_pages; in backend_map() 304 ret = gnttab_map_refs(map_ops, NULL, buf->pages, buf->num_pages); in backend_map() [all...] |
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | xen-front-pgdir-shbuf.c | 160 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); in get_num_pages_dir() 187 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; in guest_calc_num_grefs() 208 unmap_ops = kcalloc(buf->num_pages, sizeof(*unmap_ops), in backend_unmap() 213 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 222 buf->num_pages); in backend_unmap() 224 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 253 map_ops = kcalloc(buf->num_pages, sizeof(*map_ops), GFP_KERNEL); in backend_map() 257 buf->backend_map_handles = kcalloc(buf->num_pages, in backend_map() 271 grefs_left = buf->num_pages; in backend_map() 295 ret = gnttab_map_refs(map_ops, NULL, buf->pages, buf->num_pages); in backend_map() [all...] |
/kernel/linux/linux-6.6/arch/riscv/include/asm/ |
H A D | set_memory.h | 21 int num_pages)) in set_kernel_memory() 25 int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT; in set_kernel_memory() local 27 return set_memory(start, num_pages); in set_kernel_memory() 37 int num_pages)) in set_kernel_memory() 19 set_kernel_memory(char *startp, char *endp, int (*set_memory)(unsigned long start, int num_pages)) set_kernel_memory() argument 35 set_kernel_memory(char *startp, char *endp, int (*set_memory)(unsigned long start, int num_pages)) set_kernel_memory() argument
|
/kernel/linux/linux-6.6/drivers/gpu/drm/gma500/ |
H A D | mmu.c | 478 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() 496 rows = num_pages / desired_tile_stride; in psb_mmu_flush_ptes() 498 desired_tile_stride = num_pages; in psb_mmu_flush_ptes() 526 unsigned long address, uint32_t num_pages) in psb_mmu_remove_pfn_sequence() 537 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_remove_pfn_sequence() 554 psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); in psb_mmu_remove_pfn_sequence() 565 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_remove_pages() 579 rows = num_pages / desired_tile_stride; in psb_mmu_remove_pages() 581 desired_tile_stride = num_pages; in psb_mmu_remove_pages() 611 psb_mmu_flush_ptes(pd, f_address, num_pages, in psb_mmu_remove_pages() 477 psb_mmu_flush_ptes(struct psb_mmu_pd *pd, unsigned long address, uint32_t num_pages, uint32_t desired_tile_stride, uint32_t hw_tile_stride) psb_mmu_flush_ptes() argument 525 psb_mmu_remove_pfn_sequence(struct psb_mmu_pd *pd, unsigned long address, uint32_t num_pages) psb_mmu_remove_pfn_sequence() argument 564 psb_mmu_remove_pages(struct psb_mmu_pd *pd, unsigned long address, uint32_t num_pages, uint32_t desired_tile_stride, uint32_t hw_tile_stride) psb_mmu_remove_pages() argument 620 psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd, uint32_t start_pfn, unsigned long address, uint32_t num_pages, int type) psb_mmu_insert_pfn_sequence() argument 666 psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages, unsigned long address, uint32_t num_pages, uint32_t desired_tile_stride, uint32_t hw_tile_stride, int type) psb_mmu_insert_pages() argument [all...] |