Lines Matching refs:tx_buff
636 memset(tx_pool->tx_buff, 0,
684 kfree(tx_pool->tx_buff);
715 tx_pool->tx_buff = kcalloc(num_entries,
718 if (!tx_pool->tx_buff)
1269 struct ibmvnic_tx_buff *tx_buff;
1273 if (!tx_pool || !tx_pool->tx_buff)
1279 tx_buff = &tx_pool->tx_buff[i];
1280 if (tx_buff && tx_buff->skb) {
1281 dev_kfree_skb_any(tx_buff->skb);
1282 tx_buff->skb = NULL;
1540 struct ibmvnic_tx_buff *tx_buff = NULL;
1627 tx_buff = &tx_pool->tx_buff[index];
1628 tx_buff->skb = skb;
1629 tx_buff->data_dma[0] = data_dma_addr;
1630 tx_buff->data_len[0] = skb->len;
1631 tx_buff->index = index;
1632 tx_buff->pool_index = queue_num;
1633 tx_buff->last_frag = true;
1680 build_hdr_descs_arr(tx_buff, &num_entries, *hdrs);
1682 tx_buff->num_entries = num_entries;
1683 tx_buff->indir_arr[0] = tx_crq;
1684 tx_buff->indir_dma = dma_map_single(dev, tx_buff->indir_arr,
1685 sizeof(tx_buff->indir_arr),
1687 if (dma_mapping_error(dev, tx_buff->indir_dma)) {
1689 tx_buff->skb = NULL;
1698 (u64)tx_buff->indir_dma,
1700 dma_unmap_single(dev, tx_buff->indir_dma,
1701 sizeof(tx_buff->indir_arr), DMA_TO_DEVICE);
1703 tx_buff->num_entries = num_entries;
1711 tx_buff->skb = NULL;
3122 txbuff = &tx_pool->tx_buff[index];