Lines Matching refs:txr
247 static inline u32 bnx2_tx_avail(struct bnx2 *bp, struct bnx2_tx_ring_info *txr)
254 diff = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons);
697 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
699 if (txr->tx_desc_ring) {
701 txr->tx_desc_ring,
702 txr->tx_desc_mapping);
703 txr->tx_desc_ring = NULL;
705 kfree(txr->tx_buf_ring);
706 txr->tx_buf_ring = NULL;
749 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
751 txr->tx_buf_ring = kzalloc(SW_TXBD_RING_SIZE, GFP_KERNEL);
752 if (!txr->tx_buf_ring)
755 txr->tx_desc_ring =
757 &txr->tx_desc_mapping, GFP_KERNEL);
758 if (!txr->tx_desc_ring)
2844 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
2854 sw_cons = txr->tx_cons;
2863 tx_buf = &txr->tx_buf_ring[sw_ring_cons];
2894 tx_buf = &txr->tx_buf_ring[BNX2_TX_RING_IDX(sw_cons)];
2914 txr->hw_tx_cons = hw_cons;
2915 txr->tx_cons = sw_cons;
2925 (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh)) {
2928 (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh))
3388 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
3392 (bnx2_get_hw_tx_cons(bnapi) != txr->hw_tx_cons))
3482 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
3485 if (bnx2_get_hw_tx_cons(bnapi) != txr->hw_tx_cons)
5128 struct bnx2_tx_ring_info *txr;
5134 txr = &bnapi->tx_ring;
5137 txr->tx_cons = 0;
5138 txr->hw_tx_cons = 0;
5148 bnx2_init_tx_context(struct bnx2 *bp, u32 cid, struct bnx2_tx_ring_info *txr)
5170 val = (u64) txr->tx_desc_mapping >> 32;
5173 val = (u64) txr->tx_desc_mapping & 0xffffffff;
5183 struct bnx2_tx_ring_info *txr;
5186 txr = &bnapi->tx_ring;
5195 txbd = &txr->tx_desc_ring[BNX2_MAX_TX_DESC_CNT];
5197 txbd->tx_bd_haddr_hi = (u64) txr->tx_desc_mapping >> 32;
5198 txbd->tx_bd_haddr_lo = (u64) txr->tx_desc_mapping & 0xffffffff;
5200 txr->tx_prod = 0;
5201 txr->tx_prod_bseq = 0;
5203 txr->tx_bidx_addr = MB_GET_CID_ADDR(cid) + BNX2_L2CTX_TX_HOST_BIDX;
5204 txr->tx_bseq_addr = MB_GET_CID_ADDR(cid) + BNX2_L2CTX_TX_HOST_BSEQ;
5206 bnx2_init_tx_context(bp, cid, txr);
5433 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring;
5436 if (!txr->tx_buf_ring)
5440 struct bnx2_sw_tx_bd *tx_buf = &txr->tx_buf_ring[j];
5459 tx_buf = &txr->tx_buf_ring[BNX2_TX_RING_IDX(j)];
5814 struct bnx2_tx_ring_info *txr;
5819 txr = &tx_napi->tx_ring;
5862 txbd = &txr->tx_desc_ring[BNX2_TX_RING_IDX(txr->tx_prod)];
5870 txr->tx_prod = BNX2_NEXT_TX_BD(txr->tx_prod);
5871 txr->tx_prod_bseq += pkt_size;
5873 BNX2_WR16(bp, txr->tx_bidx_addr, txr->tx_prod);
5874 BNX2_WR(bp, txr->tx_bseq_addr, txr->tx_prod_bseq);
5888 if (bnx2_get_hw_tx_cons(tx_napi) != txr->tx_prod)
6599 struct bnx2_tx_ring_info *txr;
6605 txr = &bnapi->tx_ring;
6608 if (unlikely(bnx2_tx_avail(bp, txr) <
6616 prod = txr->tx_prod;
6669 tx_buf = &txr->tx_buf_ring[ring_prod];
6673 txbd = &txr->tx_desc_ring[ring_prod];
6689 txbd = &txr->tx_desc_ring[ring_prod];
6696 dma_unmap_addr_set(&txr->tx_buf_ring[ring_prod], mapping,
6713 txr->tx_prod_bseq += skb->len;
6715 BNX2_WR16(bp, txr->tx_bidx_addr, prod);
6716 BNX2_WR(bp, txr->tx_bseq_addr, txr->tx_prod_bseq);
6718 txr->tx_prod = prod;
6720 if (unlikely(bnx2_tx_avail(bp, txr) <= MAX_SKB_FRAGS)) {
6729 if (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh)
6739 prod = txr->tx_prod;
6741 tx_buf = &txr->tx_buf_ring[ring_prod];
6750 tx_buf = &txr->tx_buf_ring[ring_prod];