Home
last modified time | relevance | path

Searched refs:txq (Results 1 - 25 of 630) sorted by relevance

12345678910>>...26

/kernel/linux/linux-6.6/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_tx.c54 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; in t7xx_dpmaif_update_drb_rd_idx() local
58 if (!txq->que_started) in t7xx_dpmaif_update_drb_rd_idx()
61 old_sw_rd_idx = txq->drb_rd_idx; in t7xx_dpmaif_update_drb_rd_idx()
71 drb_cnt = txq->drb_size_cnt - old_sw_rd_idx + new_hw_rd_idx; in t7xx_dpmaif_update_drb_rd_idx()
73 spin_lock_irqsave(&txq->tx_lock, flags); in t7xx_dpmaif_update_drb_rd_idx()
74 txq->drb_rd_idx = new_hw_rd_idx; in t7xx_dpmaif_update_drb_rd_idx()
75 spin_unlock_irqrestore(&txq->tx_lock, flags); in t7xx_dpmaif_update_drb_rd_idx()
83 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_nu in t7xx_dpmaif_release_tx_buffer() local
137 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; t7xx_dpmaif_tx_release() local
153 t7xx_dpmaif_drb_ring_not_empty(struct dpmaif_tx_queue *txq) t7xx_dpmaif_drb_ring_not_empty() argument
160 struct dpmaif_tx_queue *txq = container_of(work, struct dpmaif_tx_queue, dpmaif_tx_work); t7xx_dpmaif_tx_done() local
248 struct dpmaif_tx_queue *txq; t7xx_dpmaif_add_skb_to_ring() local
349 struct dpmaif_tx_queue *txq; t7xx_select_tx_queue() local
358 t7xx_txq_drb_wr_available(struct dpmaif_tx_queue *txq) t7xx_txq_drb_wr_available() argument
370 t7xx_txq_burst_send_skb(struct dpmaif_tx_queue *txq) t7xx_txq_burst_send_skb() argument
416 struct dpmaif_tx_queue *txq; t7xx_do_tx_hw_push() local
510 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[txq_number]; t7xx_dpmaif_tx_send_skb() local
537 t7xx_dpmaif_tx_drb_buf_init(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_drb_buf_init() argument
563 t7xx_dpmaif_tx_free_drb_skb(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_free_drb_skb() argument
587 t7xx_dpmaif_tx_drb_buf_rel(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_drb_buf_rel() argument
607 t7xx_dpmaif_txq_init(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_init() argument
633 t7xx_dpmaif_txq_free(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_free() argument
647 struct dpmaif_tx_queue *txq; t7xx_dpmaif_tx_stop() local
665 t7xx_dpmaif_txq_flush_rel(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_flush_rel() argument
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c29 struct hfi1_ipoib_txq *txq; member
47 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_used() argument
49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs, in hfi1_ipoib_used()
50 txq->tx_ring.complete_txreqs); in hfi1_ipoib_used()
53 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_stop_txq() argument
55 trace_hfi1_txq_stop(txq); in hfi1_ipoib_stop_txq()
56 if (atomic_inc_return(&txq->tx_ring.stops) == 1) in hfi1_ipoib_stop_txq()
57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_stop_txq()
60 static void hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_wake_txq() argument
67 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_hwat() argument
73 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_lwat() argument
79 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_depth() argument
89 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_stopped() argument
134 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_ring() argument
156 struct hfi1_ipoib_txq *txq = hfi1_ipoib_poll_tx_ring() local
353 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_common() local
404 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_submit_tx_list() argument
422 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_flush_tx_list() argument
439 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) hfi1_ipoib_submit_tx() argument
460 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_single() local
511 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_list() local
619 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_sleep() local
660 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_wakeup() local
673 struct hfi1_ipoib_txq *txq = hfi1_ipoib_flush_txq() local
705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local
756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local
770 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_list() argument
799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_deinit() local
822 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_enable() local
834 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_disable() local
844 struct hfi1_ipoib_txq *txq = &priv->txqs[q]; hfi1_ipoib_tx_timeout() local
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c31 * @txq: txq on which skb was output
39 struct hfi1_ipoib_txq *txq; member
47 struct hfi1_ipoib_txq *txq; member
59 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_used() argument
61 return hfi1_ipoib_txreqs(txq->sent_txreqs, in hfi1_ipoib_used()
62 atomic64_read(&txq->complete_txreqs)); in hfi1_ipoib_used()
65 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_stop_txq() argument
67 trace_hfi1_txq_stop(txq); in hfi1_ipoib_stop_txq()
68 if (atomic_inc_return(&txq in hfi1_ipoib_stop_txq()
72 hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) hfi1_ipoib_wake_txq() argument
79 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_hwat() argument
85 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_lwat() argument
91 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_depth() argument
101 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_stopped() argument
146 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq, int budget) hfi1_ipoib_drain_tx_ring() argument
184 struct hfi1_ipoib_txq *txq = &priv->txqs[napi - priv->tx_napis]; hfi1_ipoib_process_tx_ring() local
214 struct hfi1_ipoib_txq *txq = tx->txq; hfi1_ipoib_add_tx() local
428 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_submit_tx_list() argument
446 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_flush_tx_list() argument
463 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) hfi1_ipoib_submit_tx() argument
485 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_single() local
530 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_list() local
633 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_sleep() local
671 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_wakeup() local
684 struct hfi1_ipoib_txq *txq = hfi1_ipoib_flush_txq() local
731 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local
777 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local
796 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_list() argument
826 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_deinit() local
852 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_enable() local
864 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_disable() local
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/queue/
H A Dtx.c79 for (txq_id = 0; txq_id < ARRAY_SIZE(trans->txqs.txq); txq_id++) { in iwl_txq_gen2_tx_stop()
80 if (!trans->txqs.txq[txq_id]) in iwl_txq_gen2_tx_stop()
90 struct iwl_txq *txq, u16 byte_cnt, in iwl_pcie_gen2_update_byte_tbl()
93 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_update_byte_tbl()
98 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) in iwl_pcie_gen2_update_byte_tbl()
114 struct iwl_gen3_bc_tbl *scd_bc_tbl_gen3 = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl()
122 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl()
136 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) in iwl_txq_inc_wr_ptr() argument
89 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_pcie_gen2_update_byte_tbl() argument
185 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_tfd() argument
529 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) iwl_txq_gen2_build_tx_amsdu() argument
616 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) iwl_txq_gen2_build_tx() argument
711 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) iwl_txq_gen2_build_tfd() argument
784 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_tx() local
879 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_unmap() local
909 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_memory() argument
941 struct iwl_txq *txq; iwl_txq_gen2_free() local
996 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_init() argument
1048 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_log_scd_error() argument
1080 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); iwl_txq_stuck_timer() local
1096 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_alloc() argument
1167 struct iwl_txq *txq; iwl_txq_dyn_alloc_dma() local
1211 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) iwl_txq_alloc_response() argument
1262 struct iwl_txq *txq = NULL; iwl_txq_dyn_alloc() local
1407 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) iwl_txq_gen1_tfd_unmap() argument
1461 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_txq_gen1_update_byte_cnt_tbl() argument
1505 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen1_inval_byte_cnt_tbl() argument
[all...]
H A Dtx.h69 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument
71 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma()
83 struct iwl_txq *txq) in iwl_wake_queue()
85 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue()
86 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); in iwl_wake_queue()
87 iwl_op_mode_queue_not_full(trans->op_mode, txq->id); in iwl_wake_queue()
92 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd()
95 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd()
97 return txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd()
100 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in
82 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) iwl_wake_queue() argument
91 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) iwl_txq_get_tfd() argument
114 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_stop() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/queue/
H A Dtx.c23 struct iwl_txq *txq, u16 byte_cnt, in iwl_pcie_gen2_update_byte_tbl()
26 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_update_byte_tbl()
31 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) in iwl_pcie_gen2_update_byte_tbl()
47 struct iwl_gen3_bc_tbl_entry *scd_bc_tbl_gen3 = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl()
55 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl()
69 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) in iwl_txq_inc_wr_ptr() argument
71 lockdep_assert_held(&txq->lock); in iwl_txq_inc_wr_ptr()
73 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq in iwl_txq_inc_wr_ptr()
22 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_pcie_gen2_update_byte_tbl() argument
162 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_tfd() argument
464 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) iwl_txq_gen2_build_tx_amsdu() argument
551 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) iwl_txq_gen2_build_tx() argument
646 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) iwl_txq_gen2_build_tfd() argument
726 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_tx() local
821 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_unmap() local
851 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_memory() argument
883 struct iwl_txq *txq; iwl_txq_gen2_free() local
938 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_init() argument
990 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_log_scd_error() argument
1022 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); iwl_txq_stuck_timer() local
1047 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_alloc() argument
1129 struct iwl_txq *txq; iwl_txq_dyn_alloc_dma() local
1172 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) iwl_txq_alloc_response() argument
1229 struct iwl_txq *txq; iwl_txq_dyn_alloc() local
1391 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) iwl_txq_gen1_tfd_unmap() argument
1437 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_txq_gen1_update_byte_cnt_tbl() argument
1481 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen1_inval_byte_cnt_tbl() argument
1515 iwl_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_free_tfd() argument
1552 iwl_txq_progress(struct iwl_txq *txq) iwl_txq_progress() argument
1580 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_reclaim() local
1705 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_set_q_ptrs() local
1721 struct iwl_txq *txq = trans->txqs.txq[queue]; iwl_trans_txq_freeze_timer() local
1773 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_trans_txq_send_hcmd_sync() local
[all...]
H A Dtx.h16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument
18 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma()
30 struct iwl_txq *txq) in iwl_wake_queue()
32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue()
33 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); in iwl_wake_queue()
34 iwl_op_mode_queue_not_full(trans->op_mode, txq->id); in iwl_wake_queue()
39 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd()
42 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd()
44 return (u8 *)txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd()
47 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in
29 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) iwl_wake_queue() argument
38 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) iwl_txq_get_tfd() argument
61 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_stop() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/pcie/
H A Dtx.c130 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr()
133 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr()
135 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr()
158 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr()
167 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr()
168 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr()
170 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr()
178 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local
183 spin_lock_bh(&txq in iwl_pcie_txq_check_wrptrs()
129 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_inc_wr_ptr() argument
217 iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_free_tfd() argument
249 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) iwl_pcie_txq_build_tfd() argument
300 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_unmap() local
353 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_free() local
464 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_tx_reset() local
727 iwl_pcie_txq_progress(struct iwl_txq *txq) iwl_pcie_txq_progress() argument
755 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_reclaim() local
871 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_set_q_ptrs() local
931 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_cmdq_reclaim() local
1005 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_enable() local
1115 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_set_shared_mode() local
1171 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_enqueue_hcmd() local
1455 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_hcmd_complete() local
1534 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_send_hcmd_sync() local
1645 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) iwl_fill_data_tbs() argument
1716 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument
1896 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument
1916 struct iwl_txq *txq; iwl_trans_pcie_tx() local
[all...]
H A Dtx-gen2.c78 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local
159 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd()
161 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd()
162 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd()
165 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { in iwl_pcie_gen2_enqueue_hcmd()
166 spin_unlock_irqrestore(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd()
174 out_cmd = txq in iwl_pcie_gen2_enqueue_hcmd()
315 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_gen2_send_hcmd_sync() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c74 * @txq: Logical Tx Queue
76 void hinic_txq_clean_stats(struct hinic_txq *txq) in hinic_txq_clean_stats() argument
78 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_clean_stats()
92 * @txq: Logical Tx Queue
95 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) in hinic_txq_get_stats() argument
97 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_get_stats()
113 * @txq: Logical Tx Queue
115 static void txq_stats_init(struct hinic_txq *txq) in txq_stats_init() argument
117 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in txq_stats_init()
120 hinic_txq_clean_stats(txq); in txq_stats_init()
497 struct hinic_txq *txq; hinic_lb_xmit_frame() local
558 struct hinic_txq *txq; hinic_xmit_frame() local
663 free_all_tx_skbs(struct hinic_txq *txq) free_all_tx_skbs() argument
697 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); free_tx_poll() local
777 struct hinic_txq *txq = data; tx_irq() local
794 tx_request_irq(struct hinic_txq *txq) tx_request_irq() argument
839 tx_free_irq(struct hinic_txq *txq) tx_free_irq() argument
855 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) hinic_init_txq() argument
920 hinic_clean_txq(struct hinic_txq *txq) hinic_clean_txq() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c75 * @txq: Logical Tx Queue
77 static void hinic_txq_clean_stats(struct hinic_txq *txq) in hinic_txq_clean_stats() argument
79 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_clean_stats()
93 * @txq: Logical Tx Queue
96 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) in hinic_txq_get_stats() argument
98 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_get_stats()
114 * @txq: Logical Tx Queue
116 static void txq_stats_init(struct hinic_txq *txq) in txq_stats_init() argument
118 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in txq_stats_init()
121 hinic_txq_clean_stats(txq); in txq_stats_init()
499 struct hinic_txq *txq; hinic_lb_xmit_frame() local
560 struct hinic_txq *txq; hinic_xmit_frame() local
665 free_all_tx_skbs(struct hinic_txq *txq) free_all_tx_skbs() argument
699 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); free_tx_poll() local
779 struct hinic_txq *txq = data; tx_irq() local
796 tx_request_irq(struct hinic_txq *txq) tx_request_irq() argument
842 tx_free_irq(struct hinic_txq *txq) tx_free_irq() argument
858 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) hinic_init_txq() argument
923 hinic_clean_txq(struct hinic_txq *txq) hinic_clean_txq() argument
[all...]
/kernel/linux/linux-6.6/include/net/
H A Dnetdev_queues.h34 #define netif_txq_try_stop(txq, get_desc, start_thrs) \
38 netif_tx_stop_queue(txq); \
50 netif_tx_start_queue(txq); \
58 * @txq: struct netdev_queue to stop/start
75 #define netif_txq_maybe_stop(txq, get_desc, stop_thrs, start_thrs) \
81 _res = netif_txq_try_stop(txq, get_desc, start_thrs); \
100 * @txq: struct netdev_queue to stop/start
118 #define __netif_txq_completed_wake(txq, pkts, bytes, \
128 netdev_txq_completed_mb(txq, pkts, bytes); \
133 if (unlikely(netif_tx_queue_stopped(txq))
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/pcie/
H A Dtx.c73 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr()
76 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr()
78 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr()
101 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr()
110 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr()
111 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr()
113 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr()
121 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local
126 spin_lock_bh(&txq in iwl_pcie_txq_check_wrptrs()
72 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_inc_wr_ptr() argument
135 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) iwl_pcie_txq_build_tfd() argument
189 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_unmap() local
239 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_free() local
350 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_tx_reset() local
653 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_cmdq_reclaim() local
723 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_enable() local
833 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_set_shared_mode() local
888 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_enqueue_hcmd() local
1169 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_hcmd_complete() local
1227 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) iwl_fill_data_tbs() argument
1279 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument
1410 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument
1429 struct iwl_txq *txq; iwl_trans_pcie_tx() local
[all...]
H A Dtx-gen2.c31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local
112 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd()
114 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd()
115 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd()
118 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { in iwl_pcie_gen2_enqueue_hcmd()
119 spin_unlock_irqrestore(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd()
127 out_cmd = txq in iwl_pcie_gen2_enqueue_hcmd()
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/ath/ath9k/
H A Dxmit.c50 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
53 int tx_flags, struct ath_txq *txq,
56 struct ath_txq *txq, struct list_head *bf_q,
59 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
67 struct ath_txq *txq,
101 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq)
102 __releases(&txq->axq_lock)
109 skb_queue_splice_init(&txq->complete_q, &q);
110 spin_unlock_bh(&txq->axq_lock);
129 struct ath_txq *txq in ath9k_wake_tx_queue() local
164 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) ath_txq_skb_done() argument
189 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); ath_tid_pull() local
234 struct ath_txq *txq = tid->txq; ath_tx_flush_tid() local
316 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tid_drain() argument
344 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) ath_tx_set_retry() argument
439 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) ath_tx_complete_aggr() argument
681 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) ath_tx_process_buffer() argument
894 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) ath_tx_get_tid_subframe() argument
983 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_aggr() argument
1347 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) ath_tx_fill_desc() argument
1440 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_burst() argument
1473 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tx_sched_aggr() argument
1520 struct ath_txq *txq; ath_tx_aggr_start() local
1560 struct ath_txq *txq = txtid->txq; ath_tx_aggr_stop() local
1592 struct ath_txq *txq; ath_tx_aggr_wakeup() local
1636 struct ath_txq *txq = sc->tx.uapsdq; ath9k_release_buffered_frames() local
1750 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; ath_txq_setup() local
1812 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) ath_drain_txq_list() argument
1845 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) ath_draintxq() argument
1873 struct ath_txq *txq; ath_drain_all_txq() local
1911 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_cleanupq() argument
1920 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) ath_txq_schedule() argument
1961 struct ath_txq *txq; ath_txq_schedule_all() local
1981 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) ath_tx_txqaddbuf() argument
2048 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_send_normal() argument
2153 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_setup_buffer() argument
2282 struct ath_txq *txq = txctl->txq; ath_tx_start() local
2406 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) ath_tx_complete() argument
2458 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) ath_tx_complete_buf() argument
2571 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_processq() argument
2665 struct ath_txq *txq; ath_tx_edma_tasklet() local
2832 struct ath_txq *txq; ath_tx_node_cleanup() local
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/ath/ath9k/
H A Dxmit.c56 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
59 int tx_flags, struct ath_txq *txq,
62 struct ath_txq *txq, struct list_head *bf_q,
65 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
73 struct ath_txq *txq,
107 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq)
108 __releases(&txq->axq_lock)
115 skb_queue_splice_init(&txq->complete_q, &q);
116 spin_unlock_bh(&txq->axq_lock);
135 struct ath_txq *txq in ath9k_wake_tx_queue() local
211 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) ath_txq_skb_done() argument
236 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); ath_tid_pull() local
281 struct ath_txq *txq = tid->txq; ath_tx_flush_tid() local
363 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tid_drain() argument
391 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) ath_tx_set_retry() argument
488 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) ath_tx_complete_aggr() argument
730 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) ath_tx_process_buffer() argument
943 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) ath_tx_get_tid_subframe() argument
1032 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_aggr() argument
1408 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) ath_tx_fill_desc() argument
1501 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_burst() argument
1534 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tx_sched_aggr() argument
1581 struct ath_txq *txq; ath_tx_aggr_start() local
1621 struct ath_txq *txq = txtid->txq; ath_tx_aggr_stop() local
1653 struct ath_txq *txq; ath_tx_aggr_wakeup() local
1697 struct ath_txq *txq = sc->tx.uapsdq; ath9k_release_buffered_frames() local
1809 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; ath_txq_setup() local
1871 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) ath_drain_txq_list() argument
1904 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) ath_draintxq() argument
1932 struct ath_txq *txq; ath_drain_all_txq() local
1970 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_cleanupq() argument
1979 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) ath_txq_schedule() argument
2020 struct ath_txq *txq; ath_txq_schedule_all() local
2040 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) ath_tx_txqaddbuf() argument
2107 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_send_normal() argument
2212 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_setup_buffer() argument
2341 struct ath_txq *txq = txctl->txq; ath_tx_start() local
2465 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) ath_tx_complete() argument
2517 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) ath_tx_complete_buf() argument
2630 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_processq() argument
2724 struct ath_txq *txq; ath_tx_edma_tasklet() local
2891 struct ath_txq *txq; ath_tx_node_cleanup() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dselq.h30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix() argument
32 while (unlikely(txq >= num_channels)) in mlx5e_txq_to_ch_ix()
33 txq -= num_channels; in mlx5e_txq_to_ch_ix()
34 return txq; in mlx5e_txq_to_ch_ix()
37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix_htb() argument
39 if (unlikely(txq >= num_channels)) { in mlx5e_txq_to_ch_ix_htb()
40 if (unlikely(txq >= num_channels << 3)) in mlx5e_txq_to_ch_ix_htb()
41 txq %= num_channels; in mlx5e_txq_to_ch_ix_htb()
44 txq -= num_channels; in mlx5e_txq_to_ch_ix_htb()
45 while (txq > in mlx5e_txq_to_ch_ix_htb()
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/atheros/alx/
H A Dmain.c54 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) in alx_free_txbuf() argument
56 struct alx_buffer *txb = &txq->bufs[entry]; in alx_free_txbuf()
59 dma_unmap_single(txq->dev, in alx_free_txbuf()
150 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping()
153 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) in alx_get_tx_queue() argument
155 return netdev_get_tx_queue(txq->netdev, txq->queue_idx); in alx_get_tx_queue()
158 static inline int alx_tpd_avail(struct alx_tx_queue *txq) in alx_tpd_avail() argument
160 if (txq->write_idx >= txq in alx_tpd_avail()
165 alx_clean_tx_irq(struct alx_tx_queue *txq) alx_clean_tx_irq() argument
489 alx_free_txring_buf(struct alx_tx_queue *txq) alx_free_txring_buf() argument
617 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) alx_alloc_tx_ring() argument
743 struct alx_tx_queue *txq; alx_alloc_napis() local
1437 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) alx_map_tx_skb() argument
1508 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) alx_start_xmit_ring() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/alx/
H A Dmain.c53 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) in alx_free_txbuf() argument
55 struct alx_buffer *txb = &txq->bufs[entry]; in alx_free_txbuf()
58 dma_unmap_single(txq->dev, in alx_free_txbuf()
149 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping()
152 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) in alx_get_tx_queue() argument
154 return netdev_get_tx_queue(txq->netdev, txq->queue_idx); in alx_get_tx_queue()
157 static inline int alx_tpd_avail(struct alx_tx_queue *txq) in alx_tpd_avail() argument
159 if (txq->write_idx >= txq in alx_tpd_avail()
164 alx_clean_tx_irq(struct alx_tx_queue *txq) alx_clean_tx_irq() argument
488 alx_free_txring_buf(struct alx_tx_queue *txq) alx_free_txring_buf() argument
616 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) alx_alloc_tx_ring() argument
742 struct alx_tx_queue *txq; alx_alloc_napis() local
1460 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) alx_map_tx_skb() argument
1531 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) alx_start_xmit_ring() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/
H A Dmv643xx_eth.c178 #define IS_TSO_HEADER(txq, addr) \
179 ((addr >= txq->tso_hdrs_dma) && \
180 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
407 struct tx_queue txq[8]; member
439 /* rxq/txq helper functions *************************************************/
445 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument
447 return container_of(txq, struct mv643xx_eth_private, txq[txq in txq_to_mp()
466 txq_reset_hw_ptr(struct tx_queue *txq) txq_reset_hw_ptr() argument
476 txq_enable(struct tx_queue *txq) txq_enable() argument
482 txq_disable(struct tx_queue *txq) txq_disable() argument
492 txq_maybe_wake(struct tx_queue *txq) txq_maybe_wake() argument
725 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) txq_put_data_tso() argument
773 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) txq_put_hdr_tso() argument
815 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_tso() argument
888 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) txq_submit_frag_skb() argument
926 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_skb() argument
994 struct tx_queue *txq; mv643xx_eth_xmit() local
1029 txq_kick(struct tx_queue *txq) txq_kick() argument
1054 txq_reclaim(struct tx_queue *txq, int budget, int force) txq_reclaim() argument
1162 txq_set_rate(struct tx_queue *txq, int rate, int burst) txq_set_rate() argument
1180 txq_set_fixed_prio_mode(struct tx_queue *txq) txq_set_fixed_prio_mode() argument
1259 struct tx_queue *txq = mp->txq + i; mv643xx_eth_get_stats() local
2025 struct tx_queue *txq = mp->txq + index; txq_init() local
2111 txq_deinit(struct tx_queue *txq) txq_deinit() argument
2201 struct tx_queue *txq = mp->txq + i; handle_link_event() local
2344 struct tx_queue *txq = mp->txq + i; port_start() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/
H A Dmv643xx_eth.c179 #define IS_TSO_HEADER(txq, addr) \
180 ((addr >= txq->tso_hdrs_dma) && \
181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
408 struct tx_queue txq[8]; member
440 /* rxq/txq helper functions *************************************************/
446 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument
448 return container_of(txq, struct mv643xx_eth_private, txq[txq in txq_to_mp()
467 txq_reset_hw_ptr(struct tx_queue *txq) txq_reset_hw_ptr() argument
477 txq_enable(struct tx_queue *txq) txq_enable() argument
483 txq_disable(struct tx_queue *txq) txq_disable() argument
493 txq_maybe_wake(struct tx_queue *txq) txq_maybe_wake() argument
727 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) txq_put_data_tso() argument
775 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) txq_put_hdr_tso() argument
818 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_tso() argument
891 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) txq_submit_frag_skb() argument
929 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_skb() argument
997 struct tx_queue *txq; mv643xx_eth_xmit() local
1032 txq_kick(struct tx_queue *txq) txq_kick() argument
1057 txq_reclaim(struct tx_queue *txq, int budget, int force) txq_reclaim() argument
1165 txq_set_rate(struct tx_queue *txq, int rate, int burst) txq_set_rate() argument
1183 txq_set_fixed_prio_mode(struct tx_queue *txq) txq_set_fixed_prio_mode() argument
1262 struct tx_queue *txq = mp->txq + i; mv643xx_eth_get_stats() local
2036 struct tx_queue *txq = mp->txq + index; txq_init() local
2122 txq_deinit(struct tx_queue *txq) txq_deinit() argument
2212 struct tx_queue *txq = mp->txq + i; handle_link_event() local
2355 struct tx_queue *txq = mp->txq + i; port_start() local
[all...]
H A Dmvneta.c135 #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8)
773 /* Increment txq get counter */
774 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) in mvneta_txq_inc_get() argument
776 txq->txq_get_index++; in mvneta_txq_inc_get()
777 if (txq->txq_get_index == txq->size) in mvneta_txq_inc_get()
778 txq->txq_get_index = 0; in mvneta_txq_inc_get()
781 /* Increment txq put counter */
782 static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq) in mvneta_txq_inc_put() argument
963 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) mvneta_txq_pend_desc_add() argument
982 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) mvneta_txq_next_desc_get() argument
993 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) mvneta_txq_desc_put() argument
1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_port_up() local
1498 int rxq, txq; mvneta_defaults_set() local
1708 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) mvneta_tx_done_pkts_coal_set() argument
1734 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) mvneta_txq_sent_desc_dec() argument
1752 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sent_desc_num_get() argument
1768 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sent_desc_proc() argument
1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) mvneta_txq_bufs_free() argument
1913 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_done() argument
2091 mvneta_xdp_submit_frame(struct mvneta_port *pp, struct mvneta_tx_queue *txq, struct xdp_frame *xdpf, int *nxmit_byte, bool dma_map) mvneta_xdp_submit_frame() argument
2177 struct mvneta_tx_queue *txq; mvneta_xdp_xmit_back() local
2219 struct mvneta_tx_queue *txq; mvneta_xdp_xmit() local
2670 mvneta_free_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_free_tso_hdrs() argument
2686 mvneta_alloc_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_alloc_tso_hdrs() argument
2706 mvneta_get_tso_hdr(struct mvneta_tx_queue *txq, dma_addr_t *dma) mvneta_get_tso_hdr() argument
2718 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_tx_queue *txq, struct tso_t *tso, int size, bool is_last) mvneta_tso_put_hdr() argument
2742 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) mvneta_tso_put_data() argument
2775 mvneta_release_descs(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int first, int num) mvneta_release_descs() argument
2803 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) mvneta_tx_tso() argument
2862 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) mvneta_tx_frag_process() argument
2915 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; mvneta_tx() local
3005 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_done_force() argument
3025 struct mvneta_tx_queue *txq; mvneta_tx_done_gbe() local
3524 mvneta_txq_sw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sw_init() argument
3569 mvneta_txq_hw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_hw_init() argument
3584 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_init() argument
3599 mvneta_txq_sw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sw_deinit() argument
3621 mvneta_txq_hw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_hw_deinit() argument
3633 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_deinit() argument
4691 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_ethtool_set_coalesce() local
5221 int rxq, txq, tc, ret; mvneta_setup_mqprio() local
5356 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_init() local
5795 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_suspend() local
5848 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_resume() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c75 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) in qede_free_tx_pkt() argument
77 u16 idx = txq->sw_tx_cons; in qede_free_tx_pkt()
78 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt()
83 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt()
88 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n", in qede_free_tx_pkt()
89 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt()
95 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq in qede_free_tx_pkt()
130 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) qede_free_failed_tx_pkt() argument
242 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) map_frag_to_bd() argument
289 qede_update_tx_producer(struct qede_tx_queue *txq) qede_update_tx_producer() argument
305 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) qede_xdp_xmit() argument
393 qede_txq_has_work(struct qede_tx_queue *txq) qede_txq_has_work() argument
406 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_xdp_tx_int() argument
438 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_tx_int() argument
1500 struct qede_tx_queue *txq; qede_start_xmit() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c76 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) in qede_free_tx_pkt() argument
78 u16 idx = txq->sw_tx_cons; in qede_free_tx_pkt()
79 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt()
84 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt()
89 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n", in qede_free_tx_pkt()
90 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt()
96 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq in qede_free_tx_pkt()
131 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) qede_free_failed_tx_pkt() argument
243 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) map_frag_to_bd() argument
288 qede_update_tx_producer(struct qede_tx_queue *txq) qede_update_tx_producer() argument
304 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) qede_xdp_xmit() argument
387 qede_txq_has_work(struct qede_tx_queue *txq) qede_txq_has_work() argument
400 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_xdp_tx_int() argument
432 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_tx_int() argument
1482 struct qede_tx_queue *txq; qede_start_xmit() local
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/
H A Dtx.c9 mt76_txq_get_qid(struct ieee80211_txq *txq) in mt76_txq_get_qid() argument
11 if (!txq->sta) in mt76_txq_get_qid()
14 return txq->ac; in mt76_txq_get_qid()
21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local
30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn()
31 mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_tx_check_agg_ssn()
306 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); in mt76_txq_dequeue() local
311 skb = ieee80211_tx_dequeue(phy->hw, txq); in mt76_txq_dequeue()
352 struct ieee80211_txq *txq in mt76_release_buffered_frames() local
388 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); mt76_txq_send_burst() local
453 struct ieee80211_txq *txq; mt76_txq_schedule_list() local
479 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); mt76_txq_schedule_list() local
547 struct ieee80211_txq *txq = sta->txq[i]; mt76_stop_tx_queues() local
564 mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq) mt76_wake_tx_queue() argument
[all...]

Completed in 31 milliseconds

12345678910>>...26