Home
last modified time | relevance | path

Searched refs:pages_to_map (Results 1 - 8 of 8) sorted by relevance

/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/
H A Dumr.c700 size_t pages_to_map = 0; in mlx5r_umr_update_xlt() local
722 pages_to_map = ALIGN(npages, page_align); in mlx5r_umr_update_xlt()
735 pages_to_map = min_t(size_t, pages_to_map, max_pages); in mlx5r_umr_update_xlt()
743 pages_mapped < pages_to_map && !err; in mlx5r_umr_update_xlt()
745 npages = min_t(int, pages_iter, pages_to_map - pages_mapped); in mlx5r_umr_update_xlt()
754 if (pages_mapped + pages_iter >= pages_to_map) in mlx5r_umr_update_xlt()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
H A Dmr.c1037 size_t pages_to_map = 0; in mlx5_ib_update_xlt() local
1058 pages_to_map = ALIGN(npages, page_align); in mlx5_ib_update_xlt()
1059 size = desc_size * pages_to_map; in mlx5_ib_update_xlt()
1091 pages_to_map = min_t(size_t, pages_to_map, max_pages); in mlx5_ib_update_xlt()
1114 pages_mapped < pages_to_map && !err; in mlx5_ib_update_xlt()
1116 npages = min_t(int, pages_iter, pages_to_map - pages_mapped); in mlx5_ib_update_xlt()
1134 if (pages_mapped + pages_iter >= pages_to_map) { in mlx5_ib_update_xlt()
/kernel/linux/linux-5.10/drivers/net/xen-netback/
H A Dcommon.h173 struct page *pages_to_map[MAX_PENDING_REQS]; member
H A Dnetback.c345 queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx]; in xenvif_tx_create_map_op()
1411 queue->pages_to_map, in xenvif_tx_action()
/kernel/linux/linux-6.6/drivers/net/xen-netback/
H A Dcommon.h173 struct page *pages_to_map[MAX_PENDING_REQS]; member
H A Dnetback.c345 queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx]; in xenvif_tx_create_map_op()
1400 queue->pages_to_map, in xenvif_tx_action()
/kernel/linux/linux-6.6/net/ipv4/
H A Dtcp.c1996 unsigned int pages_to_map, in tcp_zerocopy_vm_insert_batch()
2003 unsigned long pages_remaining = pages_to_map; in tcp_zerocopy_vm_insert_batch()
2009 pages_mapped = pages_to_map - (unsigned int)pages_remaining; in tcp_zerocopy_vm_insert_batch()
2089 unsigned int pages_to_map = 0; in tcp_zerocopy_receive() local
2172 pages[pages_to_map++] = page; in tcp_zerocopy_receive()
2176 if (pages_to_map == TCP_ZEROCOPY_PAGE_BATCH_SIZE || in tcp_zerocopy_receive()
2182 pages_to_map, in tcp_zerocopy_receive()
2188 pages_to_map = 0; in tcp_zerocopy_receive()
2191 if (pages_to_map) { in tcp_zerocopy_receive()
2192 ret = tcp_zerocopy_vm_insert_batch(vma, pages, pages_to_map, in tcp_zerocopy_receive()
1994 tcp_zerocopy_vm_insert_batch(struct vm_area_struct *vma, struct page **pages, unsigned int pages_to_map, unsigned long *address, u32 *length, u32 *seq, struct tcp_zerocopy_receive *zc, u32 total_bytes_to_map) tcp_zerocopy_vm_insert_batch() argument
[all...]
/kernel/linux/linux-5.10/net/ipv4/
H A Dtcp.c1876 unsigned long pages_to_map, in tcp_zerocopy_vm_insert_batch()
1882 unsigned long pages_remaining = pages_to_map; in tcp_zerocopy_vm_insert_batch()
1887 bytes_mapped = PAGE_SIZE * (pages_to_map - pages_remaining); in tcp_zerocopy_vm_insert_batch()
1874 tcp_zerocopy_vm_insert_batch(struct vm_area_struct *vma, struct page **pages, unsigned long pages_to_map, unsigned long *insert_addr, u32 *length_with_pending, u32 *seq, struct tcp_zerocopy_receive *zc) tcp_zerocopy_vm_insert_batch() argument

Completed in 19 milliseconds