Lines Matching refs:buf_info

336 	BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp);
339 skb = tq->buf_info[eop_idx].skb;
341 tq->buf_info[eop_idx].skb = NULL;
346 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp,
408 tbi = tq->buf_info + tq->tx_ring.next2comp;
420 BUG_ON(tq->buf_info[i].skb != NULL ||
421 tq->buf_info[i].map_type != VMXNET3_MAP_NONE);
454 if (tq->buf_info) {
456 tq->tx_ring.size * sizeof(tq->buf_info[0]),
457 tq->buf_info, tq->buf_info_pa);
458 tq->buf_info = NULL;
496 memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size);
498 tq->buf_info[i].map_type = VMXNET3_MAP_NONE;
511 tq->comp_ring.base || tq->buf_info);
537 sz = tq->tx_ring.size * sizeof(tq->buf_info[0]);
538 tq->buf_info = dma_alloc_coherent(&adapter->pdev->dev, sz,
540 if (!tq->buf_info)
570 struct vmxnet3_rx_buf_info *rbi_base = rq->buf_info[ring_idx];
702 tbi = tq->buf_info + tq->tx_ring.next2fill;
730 tbi = tq->buf_info + tq->tx_ring.next2fill;
765 tbi = tq->buf_info + tq->tx_ring.next2fill;
803 /* set the last buf_info for the pkt */
1405 rbi = rq->buf_info[ring_idx] + idx;
1687 rq->buf_info[ring_idx][i].skb) {
1690 dev_kfree_skb(rq->buf_info[ring_idx][i].skb);
1691 rq->buf_info[ring_idx][i].skb = NULL;
1693 rq->buf_info[ring_idx][i].page) {
1696 put_page(rq->buf_info[ring_idx][i].page);
1697 rq->buf_info[ring_idx][i].page = NULL;
1729 if (rq->buf_info[i]) {
1731 BUG_ON(rq->buf_info[i][j].page != NULL);
1761 if (rq->buf_info[0]) {
1764 dma_free_coherent(&adapter->pdev->dev, sz, rq->buf_info[0],
1766 rq->buf_info[0] = rq->buf_info[1] = NULL;
1796 /* initialize buf_info */
1801 rq->buf_info[0][i].buf_type = VMXNET3_RX_BUF_SKB;
1802 rq->buf_info[0][i].len = adapter->skb_buf_size;
1804 rq->buf_info[0][i].buf_type = VMXNET3_RX_BUF_PAGE;
1805 rq->buf_info[0][i].len = PAGE_SIZE;
1809 rq->buf_info[1][i].buf_type = VMXNET3_RX_BUF_PAGE;
1810 rq->buf_info[1][i].len = PAGE_SIZE;
1914 rq->buf_info[0] = bi;
1915 rq->buf_info[1] = bi + rq->rx_ring[0].size;