/kernel/linux/linux-5.10/drivers/net/ethernet/stmicro/stmmac/ |
H A D | ring_mode.c | 46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 48 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 66 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 67 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 68 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 80 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 81 tx_q->tx_skbuff_dma[entry].len = nopaged_len; in jumbo_frm() 82 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 127 if (unlikely(tx_q->tx_skbuff_dma[entr in clean_desc3() [all...] |
H A D | chain_mode.c | 41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 60 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 72 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 73 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 150 if (tx_q->tx_skbuff_dma[entry].last_segment && !priv->extend_desc && in clean_desc3()
|
H A D | stmmac_main.c | 1404 if (tx_q->tx_skbuff_dma[i].buf) { in stmmac_free_tx_buffer() 1405 if (tx_q->tx_skbuff_dma[i].map_as_page) in stmmac_free_tx_buffer() 1407 tx_q->tx_skbuff_dma[i].buf, in stmmac_free_tx_buffer() 1408 tx_q->tx_skbuff_dma[i].len, in stmmac_free_tx_buffer() 1412 tx_q->tx_skbuff_dma[i].buf, in stmmac_free_tx_buffer() 1413 tx_q->tx_skbuff_dma[i].len, in stmmac_free_tx_buffer() 1420 tx_q->tx_skbuff_dma[i].buf = 0; in stmmac_free_tx_buffer() 1421 tx_q->tx_skbuff_dma[i].map_as_page = false; in stmmac_free_tx_buffer() 1545 tx_q->tx_skbuff_dma[i].buf = 0; in init_dma_tx_desc_rings() 1546 tx_q->tx_skbuff_dma[ in init_dma_tx_desc_rings() [all...] |
H A D | stmmac.h | 56 struct stmmac_tx_info *tx_skbuff_dma; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/stmicro/stmmac/ |
H A D | ring_mode.c | 46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 48 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 66 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 67 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 68 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 80 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 81 tx_q->tx_skbuff_dma[entry].len = nopaged_len; in jumbo_frm() 82 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm() 125 if (unlikely(tx_q->tx_skbuff_dma[entr in clean_desc3() [all...] |
H A D | chain_mode.c | 41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 60 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 72 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 73 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 148 if (tx_q->tx_skbuff_dma[entry].last_segment && !priv->extend_desc && in clean_desc3()
|
H A D | stmmac_main.c | 1519 if (tx_q->tx_skbuff_dma[i].buf && in stmmac_free_tx_buffer() 1520 tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) { in stmmac_free_tx_buffer() 1521 if (tx_q->tx_skbuff_dma[i].map_as_page) in stmmac_free_tx_buffer() 1523 tx_q->tx_skbuff_dma[i].buf, in stmmac_free_tx_buffer() 1524 tx_q->tx_skbuff_dma[i].len, in stmmac_free_tx_buffer() 1528 tx_q->tx_skbuff_dma[i].buf, in stmmac_free_tx_buffer() 1529 tx_q->tx_skbuff_dma[i].len, in stmmac_free_tx_buffer() 1534 (tx_q->tx_skbuff_dma[i].buf_type == STMMAC_TXBUF_T_XDP_TX || in stmmac_free_tx_buffer() 1535 tx_q->tx_skbuff_dma[i].buf_type == STMMAC_TXBUF_T_XDP_NDO)) { in stmmac_free_tx_buffer() 1540 if (tx_q->tx_skbuff_dma[ in stmmac_free_tx_buffer() [all...] |
H A D | stmmac.h | 73 struct stmmac_tx_info *tx_skbuff_dma; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_main.c | 411 tx_ring->tx_skbuff_dma = devm_kcalloc(dev, tx_rsize, in init_tx_ring() 413 if (!tx_ring->tx_skbuff_dma) in init_tx_ring() 622 if (txqueue->tx_skbuff_dma[dma_desc]) in tx_free_ring_skbufs() 624 txqueue->tx_skbuff_dma[dma_desc], in tx_free_ring_skbufs() 630 txqueue->tx_skbuff_dma[dma_desc] = 0; in tx_free_ring_skbufs() 762 if (likely(tqueue->tx_skbuff_dma[entry])) { in sxgbe_tx_queue_clean() 764 tqueue->tx_skbuff_dma[entry], in sxgbe_tx_queue_clean() 767 tqueue->tx_skbuff_dma[entry] = 0; in sxgbe_tx_queue_clean() 1364 tqueue->tx_skbuff_dma[entry] = tx_desc->tdes01; in sxgbe_xmit()
|
H A D | sxgbe_common.h | 381 dma_addr_t *tx_skbuff_dma; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_main.c | 411 tx_ring->tx_skbuff_dma = devm_kcalloc(dev, tx_rsize, in init_tx_ring() 413 if (!tx_ring->tx_skbuff_dma) in init_tx_ring() 622 if (txqueue->tx_skbuff_dma[dma_desc]) in tx_free_ring_skbufs() 624 txqueue->tx_skbuff_dma[dma_desc], in tx_free_ring_skbufs() 630 txqueue->tx_skbuff_dma[dma_desc] = 0; in tx_free_ring_skbufs() 762 if (likely(tqueue->tx_skbuff_dma[entry])) { in sxgbe_tx_queue_clean() 764 tqueue->tx_skbuff_dma[entry], in sxgbe_tx_queue_clean() 767 tqueue->tx_skbuff_dma[entry] = 0; in sxgbe_tx_queue_clean() 1367 tqueue->tx_skbuff_dma[entry] = tx_desc->tdes01; in sxgbe_xmit()
|
H A D | sxgbe_common.h | 381 dma_addr_t *tx_skbuff_dma; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/via/ |
H A D | via-rhine.c | 447 dma_addr_t tx_skbuff_dma[TX_RING_SIZE]; member 1337 if (rp->tx_skbuff_dma[i]) { in free_tbufs() 1339 rp->tx_skbuff_dma[i], in free_tbufs() 1807 rp->tx_skbuff_dma[entry] = 0; in rhine_start_tx() 1812 rp->tx_skbuff_dma[entry] = in rhine_start_tx() 1815 if (dma_mapping_error(hwdev, rp->tx_skbuff_dma[entry])) { in rhine_start_tx() 1817 rp->tx_skbuff_dma[entry] = 0; in rhine_start_tx() 1821 rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_skbuff_dma[entry]); in rhine_start_tx() 1975 if (rp->tx_skbuff_dma[entry]) { in rhine_tx() 1977 rp->tx_skbuff_dma[entr in rhine_tx() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/via/ |
H A D | via-rhine.c | 447 dma_addr_t tx_skbuff_dma[TX_RING_SIZE]; member 1334 if (rp->tx_skbuff_dma[i]) { in free_tbufs() 1336 rp->tx_skbuff_dma[i], in free_tbufs() 1804 rp->tx_skbuff_dma[entry] = 0; in rhine_start_tx() 1809 rp->tx_skbuff_dma[entry] = in rhine_start_tx() 1812 if (dma_mapping_error(hwdev, rp->tx_skbuff_dma[entry])) { in rhine_start_tx() 1814 rp->tx_skbuff_dma[entry] = 0; in rhine_start_tx() 1818 rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_skbuff_dma[entry]); in rhine_start_tx() 1972 if (rp->tx_skbuff_dma[entry]) { in rhine_tx() 1974 rp->tx_skbuff_dma[entr in rhine_tx() [all...] |