Home
last modified time | relevance | path

Searched refs:tx_next (Results 1 - 19 of 19) sorted by relevance

/kernel/linux/linux-5.10/drivers/net/ethernet/aeroflex/
H A Dgreth.c324 greth->tx_next = 0; in greth_init_rings()
425 bdp = greth->tx_bd_base + greth->tx_next; in greth_start_xmit()
433 greth->tx_bufs_length[greth->tx_next] = skb->len & GRETH_BD_LEN; in greth_start_xmit()
436 if (greth->tx_next == GRETH_TXBD_NUM_MASK) { in greth_start_xmit()
440 greth->tx_next = NEXT_TX(greth->tx_next); in greth_start_xmit()
454 static inline u16 greth_num_free_bds(u16 tx_last, u16 tx_next) in greth_num_free_bds() argument
456 if (tx_next < tx_last) in greth_num_free_bds()
457 return (tx_last - tx_next) - 1; in greth_num_free_bds()
459 return GRETH_TXBD_NUM - (tx_next in greth_num_free_bds()
[all...]
H A Dgreth.h109 u16 tx_next; member
/kernel/linux/linux-6.6/drivers/net/ethernet/aeroflex/
H A Dgreth.c324 greth->tx_next = 0; in greth_init_rings()
425 bdp = greth->tx_bd_base + greth->tx_next; in greth_start_xmit()
433 greth->tx_bufs_length[greth->tx_next] = skb->len & GRETH_BD_LEN; in greth_start_xmit()
436 if (greth->tx_next == GRETH_TXBD_NUM_MASK) { in greth_start_xmit()
440 greth->tx_next = NEXT_TX(greth->tx_next); in greth_start_xmit()
454 static inline u16 greth_num_free_bds(u16 tx_last, u16 tx_next) in greth_num_free_bds() argument
456 if (tx_next < tx_last) in greth_num_free_bds()
457 return (tx_last - tx_next) - 1; in greth_num_free_bds()
459 return GRETH_TXBD_NUM - (tx_next in greth_num_free_bds()
[all...]
H A Dgreth.h109 u16 tx_next; member
/kernel/linux/linux-5.10/drivers/net/can/
H A Dat91_can.c132 unsigned int tx_next; member
265 return (priv->tx_next & get_next_mb_mask(priv)) + get_mb_tx_first(priv); in get_tx_next_mb()
270 return (priv->tx_next >> get_next_prio_shift(priv)) & 0xf; in get_tx_next_prio()
348 priv->tx_next = priv->tx_echo = 0; in at91_setup_mailboxes()
441 * We use the priv->tx_next as counter for the next transmission
445 * priv->tx_next = (prio << get_next_prio_shift(priv)) |
492 * tx_next buffer prio and mailbox equals 0. in at91_start_xmit()
497 priv->tx_next++; in at91_start_xmit()
500 (priv->tx_next & get_next_mask(priv)) == 0) in at91_start_xmit()
826 * We iterate from priv->tx_echo to priv->tx_next an
[all...]
/kernel/linux/linux-6.6/drivers/net/can/
H A Dat91_can.c132 unsigned int tx_next; member
265 return (priv->tx_next & get_next_mb_mask(priv)) + get_mb_tx_first(priv); in get_tx_next_mb()
270 return (priv->tx_next >> get_next_prio_shift(priv)) & 0xf; in get_tx_next_prio()
348 priv->tx_next = priv->tx_echo = 0; in at91_setup_mailboxes()
440 * We use the priv->tx_next as counter for the next transmission
444 * priv->tx_next = (prio << get_next_prio_shift(priv)) |
487 * tx_next buffer prio and mailbox equals 0. in at91_start_xmit()
492 priv->tx_next++; in at91_start_xmit()
495 (priv->tx_next & get_next_mask(priv)) == 0) in at91_start_xmit()
818 * We iterate from priv->tx_echo to priv->tx_next an
[all...]
/kernel/linux/linux-6.6/drivers/net/wwan/t7xx/
H A Dt7xx_hif_cldma.h79 struct cldma_request *tx_next; member
H A Dt7xx_hif_cldma.c70 queue->tx_next = NULL; in md_cd_queue_struct_reset()
311 req = list_prev_entry_circular(queue->tx_next, &queue->tr_ring->gpd_ring, entry); in t7xx_cldma_txq_empty_hndl()
516 queue->tx_next = req; in t7xx_cldma_q_reset()
949 tx_req = queue->tx_next; in t7xx_cldma_send_skb()
955 queue->tx_next = list_next_entry_circular(tx_req, gpd_ring, entry); in t7xx_cldma_send_skb()
1123 t7xx_cldma_hw_set_start_addr(hw_info, qno_t, md_ctrl->txq[qno_t].tx_next->gpd_addr, in t7xx_cldma_resume_early()
/kernel/linux/linux-5.10/drivers/net/ethernet/aurora/
H A Dnb8800.c423 next = priv->tx_next; in nb8800_xmit()
461 priv->tx_next = next; in nb8800_xmit()
490 unsigned int limit = priv->tx_next; in nb8800_tx_done()
815 priv->tx_next = 0; in nb8800_dma_reset()
885 priv->tx_done == priv->tx_next, in nb8800_dma_stop()
H A Dnb8800.h272 u32 tx_next; member
/kernel/linux/linux-6.6/drivers/net/ethernet/actions/
H A Dowl-emac.c699 unsigned int tx_next; in owl_emac_tx_complete() local
722 tx_next = ring->tail; in owl_emac_tx_complete()
724 while ((tx_next = owl_emac_ring_get_next(ring, tx_next)) != ring->head) { in owl_emac_tx_complete()
725 status = READ_ONCE(ring->descs[tx_next].status); in owl_emac_tx_complete()
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/octeon/
H A Docteon_mgmt.c127 unsigned int tx_next; member
991 p->tx_next = 0; in octeon_mgmt_open()
1312 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit()
1313 p->tx_next = (p->tx_next + 1) % OCTEON_MGMT_TX_RING_SIZE; in octeon_mgmt_xmit()
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/octeon/
H A Docteon_mgmt.c127 unsigned int tx_next; member
988 p->tx_next = 0; in octeon_mgmt_open()
1309 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit()
1310 p->tx_next = (p->tx_next + 1) % OCTEON_MGMT_TX_RING_SIZE; in octeon_mgmt_xmit()
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
H A Dsdma.c2477 struct sdma_txreq *tx, *tx_next; in sdma_send_txlist() local
2485 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist()
2518 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist()
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/
H A Dsdma.c2435 struct sdma_txreq *tx, *tx_next; in sdma_send_txlist() local
2443 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist()
2476 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist()
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/
H A Dsky2.h2228 u16 tx_next; /* debug only */ member
H A Dsky2.c2024 sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); in sky2_tx_complete()
4562 for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; in sky2_debug_show()
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/
H A Dsky2.h2228 u16 tx_next; /* debug only */ member
H A Dsky2.c2024 sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); in sky2_tx_complete()
4428 for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; in sky2_debug_show()

Completed in 89 milliseconds