Lines Matching defs:page
932 /* Initialize the BD for every fragment in the page. */
954 /* Initialize the BD for every fragment in the page. */
975 struct page *page = txq->tx_buf[i].buf_p;
977 if (page)
978 page_pool_put_page(page->pp, page, 0, false);
1390 struct page *page;
1424 /* Tx processing cannot call any XDP (or page pool) APIs if
1427 * however, we can't use the page pool from IRQ context.
1440 page = txq->tx_buf[index].buf_p;
1502 page_pool_put_page(page->pp, page, 0, true);
1547 struct page *new_page;
1552 rxq->rx_skb_info[index].page = new_page;
1565 struct page *page;
1614 page = virt_to_head_page(xdp->data);
1615 page_pool_put_page(rxq->page_pool, page, sync, true);
1650 struct page *page;
1714 page = rxq->rx_skb_info[index].page;
1719 prefetch(page_address(page));
1725 xdp_prepare_buff(&xdp, page_address(page),
1737 skb = build_skb(page_address(page), PAGE_SIZE);
1739 page_pool_recycle_direct(rxq->page_pool, page);
1751 data = page_address(page) + FEC_ENET_XDP_HEADROOM;
3246 page_pool_put_full_page(rxq->page_pool, rxq->rx_skb_info[i].page, false);
3273 struct page *page = txq->tx_buf[i].buf_p;
3275 page_pool_put_page(page->pp, page, 0, false);
3361 struct page *page;
3374 page = page_pool_dev_alloc_pages(rxq->page_pool);
3375 if (!page)
3378 phys_addr = page_pool_get_dma_addr(page) + FEC_ENET_XDP_HEADROOM;
3381 rxq->rx_skb_info[i].page = page;
3832 struct page *page;
3834 page = virt_to_page(xdpb->data);
3835 dma_addr = page_pool_get_dma_addr(page) +
3840 txq->tx_buf[index].buf_p = page;