Lines Matching defs:page
124 member of struct page.
126 struct page *drbd_pp_pool;
643 * which are page aligned on the sender will end up page aligned on the
1502 * to the page into the skb, and to hand it over to the NIC. In
1505 * As soon as the page was really sent over the network put_page()
1509 * reaches 0 the page will be freed. ]
1522 static int _drbd_no_send_page(struct drbd_peer_device *peer_device, struct page *page,
1530 addr = kmap(page) + offset;
1532 kunmap(page);
1538 static int _drbd_send_page(struct drbd_peer_device *peer_device, struct page *page,
1551 * __page_cache_release a page that would actually still be referenced
1553 if (!drbd_disable_sendpage && sendpage_ok(page))
1560 bvec_set_page(&bvec, page, len, offset);
1593 /* hint all but last page with MSG_MORE */
1612 /* hint all but last page with MSG_MORE */
1628 struct page *page = peer_req->pages;
1632 /* hint all but last page with MSG_MORE */
1633 page_chain_for_each(page) {
1636 err = _drbd_send_page(peer_device, page, 0, l,
1637 page_chain_next(page) ? MSG_MORE : 0);
2072 struct page *page;
2075 page = drbd_pp_pool;
2076 drbd_pp_pool = (struct page *)page_private(page);
2077 __free_page(page);
2103 struct page *page;
2152 page = alloc_page(GFP_HIGHUSER);
2153 if (!page)
2155 set_page_private(page, (unsigned long)drbd_pp_pool);
2156 drbd_pp_pool = page;
2221 __free_page(device->md_io.page);
2734 device->md_io.page = alloc_page(GFP_KERNEL);
2735 if (!device->md_io.page)
2821 __free_page(device->md_io.page);