Lines Matching defs:page
54 #include <xen/page.h>
139 struct page *grant_tx_page[NET_TX_RING_SIZE];
273 struct page *page;
281 page = page_pool_alloc_pages(queue->page_pool,
283 if (unlikely(!page)) {
287 skb_add_rx_frag(skb, 0, page, 0, 0, PAGE_SIZE);
312 struct page *page;
330 page = skb_frag_page(&skb_shinfo(skb)->frags[0]);
335 page,
466 struct page *page;
467 struct xen_netif_tx_request *tx; /* Last request on ring page */
480 struct page *page = info->page;
493 queue->grant_tx_page[id] = page;
520 gnttab_for_one_grant(info->page, offset, len, xennet_tx_setup_grant, info);
537 struct page *page,
540 /* Skip unused frames from start of page */
541 page += offset >> PAGE_SHIFT;
545 info->page = page;
548 gnttab_foreach_grant_in_range(page, offset, len,
552 page++;
560 * might be a compound page.
575 /* Skip unused frames from start of page */
620 .page = virt_to_page(xdpf->data),
714 struct page *page;
751 page = virt_to_page(skb->data);
758 * avoid exposing contiguous data on the granted page not belonging to
767 page = virt_to_page(skb->data);
785 info.page = page;
789 page++;
823 xennet_make_txreqs(&info, page, offset, len);
967 static u32 xennet_run_xdp(struct netfront_queue *queue, struct page *pdata,
1083 /* currently only a single page contains data */
1446 struct page *page;
1456 page = skb_frag_page(&skb_shinfo(skb)->frags[0]);
1458 /* gnttab_end_foreign_access() needs a page ref until
1461 get_page(page);
1462 gnttab_end_foreign_access(ref, page);
1801 static void xennet_end_access(int ref, void *page)
1803 /* This frees the page as a side-effect */
1805 gnttab_end_foreign_access(ref, virt_to_page(page));
2227 /* use page pool recycling instead of buddy allocator */
2230 dev_err(&info->xbdev->dev, "can't allocate page pool\n");