/kernel/linux/linux-5.10/drivers/net/ethernet/google/gve/ |
H A D | gve_tx.c | 155 netdev_tx_reset_queue(tx->netdev_txq); in gve_tx_free_ring() 227 tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); in gve_tx_alloc_ring() 334 netif_tx_stop_queue(tx->netdev_txq); in gve_maybe_stop_tx() 352 netif_tx_start_queue(tx->netdev_txq); in gve_maybe_stop_tx() 499 netdev_tx_sent_queue(tx->netdev_txq, skb->len); in gve_tx() 505 if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) in gve_tx() 554 netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes); in gve_clean_tx_done() 561 if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) && in gve_clean_tx_done() 564 netif_tx_wake_queue(tx->netdev_txq); in gve_clean_tx_done()
|
H A D | gve.h | 142 struct netdev_queue *netdev_txq; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 493 struct netdev_queue *netdev_txq; in hinic_lb_xmit_frame() local 535 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame() 536 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_lb_xmit_frame() 554 struct netdev_queue *netdev_txq; in hinic_xmit_frame() local 624 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame() 625 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_xmit_frame() 700 struct netdev_queue *netdev_txq; in free_tx_poll() local 744 netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id); in free_tx_poll() 746 __netif_tx_lock(netdev_txq, smp_processor_id()); in free_tx_poll() 750 __netif_tx_unlock(netdev_txq); in free_tx_poll() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 495 struct netdev_queue *netdev_txq; in hinic_lb_xmit_frame() local 537 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame() 538 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_lb_xmit_frame() 556 struct netdev_queue *netdev_txq; in hinic_xmit_frame() local 626 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame() 627 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_xmit_frame() 702 struct netdev_queue *netdev_txq; in free_tx_poll() local 746 netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id); in free_tx_poll() 748 __netif_tx_lock(netdev_txq, smp_processor_id()); in free_tx_poll() 752 __netif_tx_unlock(netdev_txq); in free_tx_poll() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 441 struct netdev_queue *netdev_txq; in qede_tx_int() local 445 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_tx_int() 467 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_tx_int() 480 if (unlikely(netif_tx_queue_stopped(netdev_txq))) { in qede_tx_int() 491 __netif_tx_lock(netdev_txq, smp_processor_id()); in qede_tx_int() 493 if ((netif_tx_queue_stopped(netdev_txq)) && in qede_tx_int() 497 netif_tx_wake_queue(netdev_txq); in qede_tx_int() 502 __netif_tx_unlock(netdev_txq); in qede_tx_int() 1499 struct netdev_queue *netdev_txq; in qede_start_xmit() local 1518 netdev_txq in qede_start_xmit() [all...] |
H A D | qede_main.c | 1723 struct netdev_queue *netdev_txq; in qede_empty_tx_queue() local 1726 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_empty_tx_queue() 1750 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_empty_tx_queue() 2711 struct netdev_queue *netdev_txq; in qede_is_txq_full() local 2713 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_is_txq_full() 2714 if (netif_xmit_stopped(netdev_txq)) in qede_is_txq_full()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 435 struct netdev_queue *netdev_txq; in qede_tx_int() local 439 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_tx_int() 461 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_tx_int() 474 if (unlikely(netif_tx_queue_stopped(netdev_txq))) { in qede_tx_int() 485 __netif_tx_lock(netdev_txq, smp_processor_id()); in qede_tx_int() 487 if ((netif_tx_queue_stopped(netdev_txq)) && in qede_tx_int() 491 netif_tx_wake_queue(netdev_txq); in qede_tx_int() 496 __netif_tx_unlock(netdev_txq); in qede_tx_int() 1481 struct netdev_queue *netdev_txq; in qede_start_xmit() local 1500 netdev_txq in qede_start_xmit() [all...] |
H A D | qede_main.c | 1791 struct netdev_queue *netdev_txq; in qede_empty_tx_queue() local 1794 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_empty_tx_queue() 1818 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_empty_tx_queue() 2809 struct netdev_queue *netdev_txq; in qede_is_txq_full() local 2811 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_is_txq_full() 2812 if (netif_xmit_stopped(netdev_txq)) in qede_is_txq_full()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/google/gve/ |
H A D | gve_tx_dqo.c | 267 tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); in gve_tx_alloc_ring_dqo() 380 netdev_tx_reset_queue(tx->netdev_txq); in gve_tx_free_rings_dqo() 420 netif_tx_stop_queue(tx->netdev_txq); in gve_maybe_stop_tx_dqo() 433 netif_tx_start_queue(tx->netdev_txq); in gve_maybe_stop_tx_dqo() 909 netdev_tx_sent_queue(tx->netdev_txq, skb->len); in gve_try_tx_skb() 935 if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) in gve_tx_dqo() 1237 netdev_tx_completed_queue(tx->netdev_txq, in gve_clean_tx_done_dqo() 1264 if (netif_tx_queue_stopped(tx->netdev_txq) && in gve_tx_poll_dqo() 1267 netif_tx_wake_queue(tx->netdev_txq); in gve_tx_poll_dqo()
|
H A D | gve_tx.c | 210 netdev_tx_reset_queue(tx->netdev_txq); in gve_tx_free_ring() 283 tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); in gve_tx_alloc_ring() 436 netif_tx_stop_queue(tx->netdev_txq); in gve_maybe_stop_tx() 708 netdev_tx_sent_queue(tx->netdev_txq, skb->len); in gve_tx() 715 if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) in gve_tx() 864 netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes); in gve_clean_tx_done() 871 if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) && in gve_clean_tx_done() 874 netif_tx_wake_queue(tx->netdev_txq); in gve_clean_tx_done()
|
H A D | gve.h | 553 struct netdev_queue *netdev_txq; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_txrx.c | 28 struct netdev_queue *netdev_txq; in ionic_txq_poke_doorbell() local 32 netdev_txq = netdev_get_tx_queue(netdev, q->index); in ionic_txq_poke_doorbell() 34 HARD_TX_LOCK(netdev, netdev_txq, smp_processor_id()); in ionic_txq_poke_doorbell() 37 HARD_TX_UNLOCK(netdev, netdev_txq); in ionic_txq_poke_doorbell() 52 HARD_TX_UNLOCK(netdev, netdev_txq); in ionic_txq_poke_doorbell()
|