Lines Matching refs:buffer

125 #undef  RX_COUNT_BUFFERS    /* define to calculate RX buffer stats */
443 /* cp->lock held. note: the last put_page will free the buffer */
448 __free_pages(page->buffer, cp->page_order);
474 page->buffer = alloc_pages(flags, cp->page_order);
475 if (!page->buffer)
477 page->dma_addr = dma_map_page(&cp->pdev->dev, page->buffer, 0,
563 if (page_count(page->buffer) > 1)
1352 if (page_count(page->buffer) == 1)
1371 /* swap if buffer is in use */
1372 if (page_count(page0[index]->buffer) > 1) {
1401 rxd[i].buffer = cpu_to_le64(page->dma_addr);
1863 daddr = le64_to_cpu(txd->buffer);
1870 /* tiny buffer may follow */
1962 addr = cas_page_map(page->buffer);
1998 addr = cas_page_map(page->buffer);
2017 __skb_frag_set_page(frag, page->buffer);
2043 __skb_frag_set_page(frag, page->buffer);
2051 addr = cas_page_map(page->buffer);
2075 addr = cas_page_map(page->buffer);
2095 addr = cas_page_map(page->buffer);
2151 * do any additional locking here. stick the buffer
2162 /* put rx descriptor back on ring. if a buffer is in use by a higher
2173 cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr);
2209 /* make a new buffer if it's still in use */
2210 if (page_count(page[entry]->buffer) > 1) {
2227 cp->init_rxds[ring][entry].buffer =
2255 * small packets: should be copied header + data in single buffer.
2256 * large packets: header and data in a single buffer.
2257 * split packets: header in a separate buffer from data.
2723 txd->buffer = cpu_to_le64(mapping);
3885 /* first buffer is never a tiny buffer and so
3888 daddr = le64_to_cpu(txd[ent].buffer);
3897 /* next buffer might by a tiny buffer.