Home
last modified time | relevance | path

Searched refs:XEN_PAGE_SIZE (Results 1 - 25 of 65) sorted by relevance

123

/kernel/linux/linux-5.10/include/xen/
H A Dpage.h9 #define XEN_PAGE_SIZE (_AC(1, UL) << XEN_PAGE_SHIFT) macro
10 #define XEN_PAGE_MASK (~(XEN_PAGE_SIZE-1))
14 * We assume that PAGE_SIZE is a multiple of XEN_PAGE_SIZE
23 #define XEN_PFN_PER_PAGE (PAGE_SIZE / XEN_PAGE_SIZE)
26 #define XEN_PFN_UP(x) (((x) + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT)
/kernel/linux/linux-6.6/include/xen/
H A Dpage.h9 #define XEN_PAGE_SIZE (_AC(1, UL) << XEN_PAGE_SHIFT) macro
10 #define XEN_PAGE_MASK (~(XEN_PAGE_SIZE-1))
14 * We assume that PAGE_SIZE is a multiple of XEN_PAGE_SIZE
23 #define XEN_PFN_PER_PAGE (PAGE_SIZE / XEN_PAGE_SIZE)
26 #define XEN_PFN_UP(x) (((x) + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT)
/kernel/linux/linux-5.10/drivers/xen/
H A Dmem-reservation.c45 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_update()
76 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_reset()
H A Darm-device.c24 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_unmap_device_mmio()
58 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_map_device_mmio()
H A Dbiomerge.c11 #if XEN_PAGE_SIZE == PAGE_SIZE in xen_biovec_phys_mergeable()
H A Dgrant-table.c746 vaddr = xen_remap(addr, XEN_PAGE_SIZE * max_nr_gframes); in gnttab_setup_auto_xlat_frames()
1131 glen = min_t(unsigned int, XEN_PAGE_SIZE - goffset, len); in gnttab_foreach_grant_in_range()
1156 fn(pfn_to_gfn(xen_pfn), goffset, XEN_PAGE_SIZE, data); in gnttab_foreach_grant() local
1158 goffset += XEN_PAGE_SIZE; in gnttab_foreach_grant()
1423 .grefs_per_grant_frame = XEN_PAGE_SIZE /
1435 .grefs_per_grant_frame = XEN_PAGE_SIZE /
/kernel/linux/linux-6.6/drivers/xen/
H A Dmem-reservation.c46 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_update()
72 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_reset()
H A Darm-device.c24 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_unmap_device_mmio()
58 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_map_device_mmio()
H A Dbiomerge.c11 #if XEN_PAGE_SIZE == PAGE_SIZE in xen_biovec_phys_mergeable()
H A Dgrant-dma-ops.c95 ret = alloc_pages_exact(n_pages * XEN_PAGE_SIZE, gfp); in xen_grant_dma_alloc()
102 free_pages_exact(ret, n_pages * XEN_PAGE_SIZE); in xen_grant_dma_alloc()
142 free_pages_exact(vaddr, n_pages * XEN_PAGE_SIZE); in xen_grant_dma_free()
H A Dgrant-table.c830 vaddr = memremap(addr, XEN_PAGE_SIZE * max_nr_gframes, MEMREMAP_WB); in gnttab_setup_auto_xlat_frames()
1215 glen = min_t(unsigned int, XEN_PAGE_SIZE - goffset, len); in gnttab_foreach_grant_in_range()
1240 fn(pfn_to_gfn(xen_pfn), goffset, XEN_PAGE_SIZE, data); in gnttab_foreach_grant() local
1242 goffset += XEN_PAGE_SIZE; in gnttab_foreach_grant()
1507 .grefs_per_grant_frame = XEN_PAGE_SIZE /
1518 .grefs_per_grant_frame = XEN_PAGE_SIZE /
/kernel/linux/linux-5.10/arch/arm/xen/
H A Dmm.c59 if (size + cflush.offset > XEN_PAGE_SIZE) in dma_cache_maint()
60 cflush.length = XEN_PAGE_SIZE - cflush.offset; in dma_cache_maint()
/kernel/linux/linux-6.6/arch/arm/xen/
H A Dmm.c57 if (size + cflush.offset > XEN_PAGE_SIZE) in dma_cache_maint()
58 cflush.length = XEN_PAGE_SIZE - cflush.offset; in dma_cache_maint()
/kernel/linux/linux-6.6/drivers/xen/xenbus/
H A Dxenbus_client.c382 unsigned long ring_size = nr_pages * XEN_PAGE_SIZE; in xenbus_setup_ring()
413 addr += XEN_PAGE_SIZE; in xenbus_setup_ring()
451 free_pages_exact(*vaddr, nr_pages * XEN_PAGE_SIZE); in xenbus_teardown_ring()
752 area = get_vm_area(XEN_PAGE_SIZE * nr_grefs, VM_IOREMAP); in xenbus_map_ring_pv()
756 XEN_PAGE_SIZE * nr_grefs, map_ring_apply, info)) in xenbus_map_ring_pv()
815 addr = (unsigned long)vaddr + (XEN_PAGE_SIZE * i); in xenbus_unmap_ring_pv()
/kernel/linux/linux-5.10/drivers/net/xen-netback/
H A Dcommon.h68 #define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE)
69 #define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE)
89 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1)
H A Drx.c48 needed = DIV_ROUND_UP(skb->len, XEN_PAGE_SIZE); in xenvif_update_needed_slots()
379 chunk_len = min_t(size_t, frag_len, XEN_PAGE_SIZE - offset); in xenvif_rx_next_chunk()
380 chunk_len = min_t(size_t, chunk_len, XEN_PAGE_SIZE - in xenvif_rx_next_chunk()
411 } while (offset < XEN_PAGE_SIZE && pkt->remaining_len > 0); in xenvif_rx_data_slot()
H A Dhash.c349 len > XEN_PAGE_SIZE / sizeof(*mapping)) in xenvif_set_hash_mapping()
354 if (copy_op[0].dest.offset + copy_op[0].len > XEN_PAGE_SIZE) { in xenvif_set_hash_mapping()
356 copy_op[1].source.offset = XEN_PAGE_SIZE - copy_op[0].dest.offset; in xenvif_set_hash_mapping()
/kernel/linux/linux-6.6/drivers/net/xen-netback/
H A Dcommon.h68 #define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE)
69 #define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE)
89 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1)
H A Drx.c48 needed = DIV_ROUND_UP(skb->len, XEN_PAGE_SIZE); in xenvif_update_needed_slots()
379 chunk_len = min_t(size_t, frag_len, XEN_PAGE_SIZE - offset); in xenvif_rx_next_chunk()
380 chunk_len = min_t(size_t, chunk_len, XEN_PAGE_SIZE - in xenvif_rx_next_chunk()
411 } while (offset < XEN_PAGE_SIZE && pkt->remaining_len > 0); in xenvif_rx_data_slot()
H A Dhash.c349 len > XEN_PAGE_SIZE / sizeof(*mapping)) in xenvif_set_hash_mapping()
354 if (copy_op[0].dest.offset + copy_op[0].len > XEN_PAGE_SIZE) { in xenvif_set_hash_mapping()
356 copy_op[1].source.offset = XEN_PAGE_SIZE - copy_op[0].dest.offset; in xenvif_set_hash_mapping()
/kernel/linux/linux-5.10/drivers/xen/xenbus/
H A Dxenbus_client.c403 vaddr = vaddr + XEN_PAGE_SIZE; in xenbus_grant_ring()
709 area = get_vm_area(XEN_PAGE_SIZE * nr_grefs, VM_IOREMAP); in xenbus_map_ring_pv()
713 XEN_PAGE_SIZE * nr_grefs, map_ring_apply, info)) in xenbus_map_ring_pv()
772 addr = (unsigned long)vaddr + (XEN_PAGE_SIZE * i); in xenbus_unmap_ring_pv()
/kernel/linux/linux-5.10/sound/xen/
H A Dxen_snd_front.c371 /* At the moment we only support case with XEN_PAGE_SIZE == PAGE_SIZE */ in xen_drv_init()
372 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init()
373 pr_err(XENSND_DRIVER_NAME ": different kernel and Xen page sizes are not supported: XEN_PAGE_SIZE (%lu) != PAGE_SIZE (%lu)\n", in xen_drv_init()
374 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
/kernel/linux/linux-6.6/drivers/block/xen-blkback/
H A Dcommon.h59 #define XEN_PAGES_PER_SEGMENT (PAGE_SIZE / XEN_PAGE_SIZE)
62 (XEN_PAGE_SIZE/sizeof(struct blkif_request_segment))
/kernel/linux/linux-6.6/sound/xen/
H A Dxen_snd_front.c371 /* At the moment we only support case with XEN_PAGE_SIZE == PAGE_SIZE */ in xen_drv_init()
372 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init()
373 pr_err(XENSND_DRIVER_NAME ": different kernel and Xen page sizes are not supported: XEN_PAGE_SIZE (%lu) != PAGE_SIZE (%lu)\n", in xen_drv_init()
374 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
/kernel/linux/linux-5.10/drivers/block/xen-blkback/
H A Dxenbus.c221 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
222 size = __RING_SIZE(sring_native, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
231 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
232 size = __RING_SIZE(sring_x86_32, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
241 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
242 size = __RING_SIZE(sring_x86_64, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()

Completed in 28 milliseconds

123