Lines Matching defs:txq
281 static int txq2vq(int txq)
283 return txq * 2 + 1;
511 /* when vi->curr_queue_pairs > nr_cpu_ids, the txq/sq is only used for xdp tx on
516 * lock/unlock of txq 3. make sparse happy. It is difficult for two inline
520 struct netdev_queue *txq; \
527 txq = netdev_get_tx_queue(v->dev, qp); \
528 __netif_tx_acquire(txq); \
531 txq = netdev_get_tx_queue(v->dev, qp); \
532 __netif_tx_lock(txq, raw_smp_processor_id()); \
538 struct netdev_queue *txq; \
541 txq = netdev_get_tx_queue(v->dev, (q) - v->sq); \
543 __netif_tx_release(txq); \
545 __netif_tx_unlock(txq); \
1505 struct netdev_queue *txq = netdev_get_tx_queue(vi->dev, index);
1510 if (__netif_tx_trylock(txq)) {
1512 __netif_tx_unlock(txq);
1516 netif_tx_wake_queue(txq);
1588 struct netdev_queue *txq;
1598 txq = netdev_get_tx_queue(vi->dev, index);
1599 __netif_tx_lock(txq, raw_smp_processor_id());
1610 __netif_tx_unlock(txq);
1615 __netif_tx_lock(txq, raw_smp_processor_id());
1617 __netif_tx_unlock(txq);
1624 netif_tx_wake_queue(txq);
1682 struct netdev_queue *txq = netdev_get_tx_queue(dev, qnum);
1739 if (kick || netif_xmit_stopped(txq)) {