/kernel/linux/linux-6.6/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_tx.c | 54 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; in t7xx_dpmaif_update_drb_rd_idx() local 58 if (!txq->que_started) in t7xx_dpmaif_update_drb_rd_idx() 61 old_sw_rd_idx = txq->drb_rd_idx; in t7xx_dpmaif_update_drb_rd_idx() 71 drb_cnt = txq->drb_size_cnt - old_sw_rd_idx + new_hw_rd_idx; in t7xx_dpmaif_update_drb_rd_idx() 73 spin_lock_irqsave(&txq->tx_lock, flags); in t7xx_dpmaif_update_drb_rd_idx() 74 txq->drb_rd_idx = new_hw_rd_idx; in t7xx_dpmaif_update_drb_rd_idx() 75 spin_unlock_irqrestore(&txq->tx_lock, flags); in t7xx_dpmaif_update_drb_rd_idx() 83 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_nu in t7xx_dpmaif_release_tx_buffer() local 137 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; t7xx_dpmaif_tx_release() local 153 t7xx_dpmaif_drb_ring_not_empty(struct dpmaif_tx_queue *txq) t7xx_dpmaif_drb_ring_not_empty() argument 160 struct dpmaif_tx_queue *txq = container_of(work, struct dpmaif_tx_queue, dpmaif_tx_work); t7xx_dpmaif_tx_done() local 248 struct dpmaif_tx_queue *txq; t7xx_dpmaif_add_skb_to_ring() local 349 struct dpmaif_tx_queue *txq; t7xx_select_tx_queue() local 358 t7xx_txq_drb_wr_available(struct dpmaif_tx_queue *txq) t7xx_txq_drb_wr_available() argument 370 t7xx_txq_burst_send_skb(struct dpmaif_tx_queue *txq) t7xx_txq_burst_send_skb() argument 416 struct dpmaif_tx_queue *txq; t7xx_do_tx_hw_push() local 510 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[txq_number]; t7xx_dpmaif_tx_send_skb() local 537 t7xx_dpmaif_tx_drb_buf_init(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_drb_buf_init() argument 563 t7xx_dpmaif_tx_free_drb_skb(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_free_drb_skb() argument 587 t7xx_dpmaif_tx_drb_buf_rel(struct dpmaif_tx_queue *txq) t7xx_dpmaif_tx_drb_buf_rel() argument 607 t7xx_dpmaif_txq_init(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_init() argument 633 t7xx_dpmaif_txq_free(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_free() argument 647 struct dpmaif_tx_queue *txq; t7xx_dpmaif_tx_stop() local 665 t7xx_dpmaif_txq_flush_rel(struct dpmaif_tx_queue *txq) t7xx_dpmaif_txq_flush_rel() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 29 struct hfi1_ipoib_txq *txq; member 47 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_used() argument 49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs, in hfi1_ipoib_used() 50 txq->tx_ring.complete_txreqs); in hfi1_ipoib_used() 53 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_stop_txq() argument 55 trace_hfi1_txq_stop(txq); in hfi1_ipoib_stop_txq() 56 if (atomic_inc_return(&txq->tx_ring.stops) == 1) in hfi1_ipoib_stop_txq() 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_stop_txq() 60 static void hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_wake_txq() argument 67 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_hwat() argument 73 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_lwat() argument 79 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_depth() argument 89 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_stopped() argument 134 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_ring() argument 156 struct hfi1_ipoib_txq *txq = hfi1_ipoib_poll_tx_ring() local 353 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_common() local 404 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_submit_tx_list() argument 422 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_flush_tx_list() argument 439 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) hfi1_ipoib_submit_tx() argument 460 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_single() local 511 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_list() local 619 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_sleep() local 660 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_wakeup() local 673 struct hfi1_ipoib_txq *txq = hfi1_ipoib_flush_txq() local 705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local 756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local 770 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_list() argument 799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_deinit() local 822 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_enable() local 834 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_disable() local 844 struct hfi1_ipoib_txq *txq = &priv->txqs[q]; hfi1_ipoib_tx_timeout() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 31 * @txq: txq on which skb was output 39 struct hfi1_ipoib_txq *txq; member 47 struct hfi1_ipoib_txq *txq; member 59 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_used() argument 61 return hfi1_ipoib_txreqs(txq->sent_txreqs, in hfi1_ipoib_used() 62 atomic64_read(&txq->complete_txreqs)); in hfi1_ipoib_used() 65 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_stop_txq() argument 67 trace_hfi1_txq_stop(txq); in hfi1_ipoib_stop_txq() 68 if (atomic_inc_return(&txq in hfi1_ipoib_stop_txq() 72 hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) hfi1_ipoib_wake_txq() argument 79 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_hwat() argument 85 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) hfi1_ipoib_ring_lwat() argument 91 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_depth() argument 101 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) hfi1_ipoib_check_queue_stopped() argument 146 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq, int budget) hfi1_ipoib_drain_tx_ring() argument 184 struct hfi1_ipoib_txq *txq = &priv->txqs[napi - priv->tx_napis]; hfi1_ipoib_process_tx_ring() local 214 struct hfi1_ipoib_txq *txq = tx->txq; hfi1_ipoib_add_tx() local 428 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_submit_tx_list() argument 446 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) hfi1_ipoib_flush_tx_list() argument 463 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) hfi1_ipoib_submit_tx() argument 485 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_single() local 530 struct hfi1_ipoib_txq *txq = txp->txq; hfi1_ipoib_send_dma_list() local 633 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_sleep() local 671 struct hfi1_ipoib_txq *txq = hfi1_ipoib_sdma_wakeup() local 684 struct hfi1_ipoib_txq *txq = hfi1_ipoib_flush_txq() local 731 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local 777 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_init() local 796 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) hfi1_ipoib_drain_tx_list() argument 826 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_txreq_deinit() local 852 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_enable() local 864 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; hfi1_ipoib_napi_tx_disable() local [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/queue/ |
H A D | tx.c | 79 for (txq_id = 0; txq_id < ARRAY_SIZE(trans->txqs.txq); txq_id++) { in iwl_txq_gen2_tx_stop() 80 if (!trans->txqs.txq[txq_id]) in iwl_txq_gen2_tx_stop() 90 struct iwl_txq *txq, u16 byte_cnt, in iwl_pcie_gen2_update_byte_tbl() 93 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_update_byte_tbl() 98 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) in iwl_pcie_gen2_update_byte_tbl() 114 struct iwl_gen3_bc_tbl *scd_bc_tbl_gen3 = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl() 122 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl() 136 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) in iwl_txq_inc_wr_ptr() argument 89 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_pcie_gen2_update_byte_tbl() argument 185 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_tfd() argument 529 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) iwl_txq_gen2_build_tx_amsdu() argument 616 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) iwl_txq_gen2_build_tx() argument 711 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) iwl_txq_gen2_build_tfd() argument 784 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_tx() local 879 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_unmap() local 909 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_memory() argument 941 struct iwl_txq *txq; iwl_txq_gen2_free() local 996 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_init() argument 1048 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_log_scd_error() argument 1080 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); iwl_txq_stuck_timer() local 1096 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_alloc() argument 1167 struct iwl_txq *txq; iwl_txq_dyn_alloc_dma() local 1211 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) iwl_txq_alloc_response() argument 1262 struct iwl_txq *txq = NULL; iwl_txq_dyn_alloc() local 1407 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) iwl_txq_gen1_tfd_unmap() argument 1461 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_txq_gen1_update_byte_cnt_tbl() argument 1505 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen1_inval_byte_cnt_tbl() argument [all...] |
H A D | tx.h | 69 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument 71 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma() 83 struct iwl_txq *txq) in iwl_wake_queue() 85 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue() 86 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); in iwl_wake_queue() 87 iwl_op_mode_queue_not_full(trans->op_mode, txq->id); in iwl_wake_queue() 92 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd() 95 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd() 97 return txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd() 100 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in 82 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) iwl_wake_queue() argument 91 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) iwl_txq_get_tfd() argument 114 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_stop() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/queue/ |
H A D | tx.c | 23 struct iwl_txq *txq, u16 byte_cnt, in iwl_pcie_gen2_update_byte_tbl() 26 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_update_byte_tbl() 31 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) in iwl_pcie_gen2_update_byte_tbl() 47 struct iwl_gen3_bc_tbl_entry *scd_bc_tbl_gen3 = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl() 55 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; in iwl_pcie_gen2_update_byte_tbl() 69 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) in iwl_txq_inc_wr_ptr() argument 71 lockdep_assert_held(&txq->lock); in iwl_txq_inc_wr_ptr() 73 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq in iwl_txq_inc_wr_ptr() 22 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_pcie_gen2_update_byte_tbl() argument 162 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_tfd() argument 464 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) iwl_txq_gen2_build_tx_amsdu() argument 551 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) iwl_txq_gen2_build_tx() argument 646 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) iwl_txq_gen2_build_tfd() argument 726 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_tx() local 821 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_gen2_unmap() local 851 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen2_free_memory() argument 883 struct iwl_txq *txq; iwl_txq_gen2_free() local 938 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_init() argument 990 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_log_scd_error() argument 1022 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); iwl_txq_stuck_timer() local 1047 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) iwl_txq_alloc() argument 1129 struct iwl_txq *txq; iwl_txq_dyn_alloc_dma() local 1172 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) iwl_txq_alloc_response() argument 1229 struct iwl_txq *txq; iwl_txq_dyn_alloc() local 1391 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) iwl_txq_gen1_tfd_unmap() argument 1437 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) iwl_txq_gen1_update_byte_cnt_tbl() argument 1481 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_gen1_inval_byte_cnt_tbl() argument 1515 iwl_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_free_tfd() argument 1552 iwl_txq_progress(struct iwl_txq *txq) iwl_txq_progress() argument 1580 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_reclaim() local 1705 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_txq_set_q_ptrs() local 1721 struct iwl_txq *txq = trans->txqs.txq[queue]; iwl_trans_txq_freeze_timer() local 1773 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_trans_txq_send_hcmd_sync() local [all...] |
H A D | tx.h | 16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument 18 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma() 30 struct iwl_txq *txq) in iwl_wake_queue() 32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue() 33 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); in iwl_wake_queue() 34 iwl_op_mode_queue_not_full(trans->op_mode, txq->id); in iwl_wake_queue() 39 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd() 42 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd() 44 return (u8 *)txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd() 47 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in 29 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) iwl_wake_queue() argument 38 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) iwl_txq_get_tfd() argument 61 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) iwl_txq_stop() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | tx.c | 130 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr() 133 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr() 135 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr() 158 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr() 167 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr() 168 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr() 170 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr() 178 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 183 spin_lock_bh(&txq in iwl_pcie_txq_check_wrptrs() 129 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_inc_wr_ptr() argument 217 iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_free_tfd() argument 249 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) iwl_pcie_txq_build_tfd() argument 300 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_unmap() local 353 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_free() local 464 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_tx_reset() local 727 iwl_pcie_txq_progress(struct iwl_txq *txq) iwl_pcie_txq_progress() argument 755 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_reclaim() local 871 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_set_q_ptrs() local 931 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_cmdq_reclaim() local 1005 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_enable() local 1115 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_set_shared_mode() local 1171 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_enqueue_hcmd() local 1455 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_hcmd_complete() local 1534 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_send_hcmd_sync() local 1645 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) iwl_fill_data_tbs() argument 1716 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument 1896 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument 1916 struct iwl_txq *txq; iwl_trans_pcie_tx() local [all...] |
H A D | tx-gen2.c | 78 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local 159 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 161 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 162 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 165 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { in iwl_pcie_gen2_enqueue_hcmd() 166 spin_unlock_irqrestore(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 174 out_cmd = txq in iwl_pcie_gen2_enqueue_hcmd() 315 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_gen2_send_hcmd_sync() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 74 * @txq: Logical Tx Queue 76 void hinic_txq_clean_stats(struct hinic_txq *txq) in hinic_txq_clean_stats() argument 78 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_clean_stats() 92 * @txq: Logical Tx Queue 95 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) in hinic_txq_get_stats() argument 97 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_get_stats() 113 * @txq: Logical Tx Queue 115 static void txq_stats_init(struct hinic_txq *txq) in txq_stats_init() argument 117 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in txq_stats_init() 120 hinic_txq_clean_stats(txq); in txq_stats_init() 497 struct hinic_txq *txq; hinic_lb_xmit_frame() local 558 struct hinic_txq *txq; hinic_xmit_frame() local 663 free_all_tx_skbs(struct hinic_txq *txq) free_all_tx_skbs() argument 697 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); free_tx_poll() local 777 struct hinic_txq *txq = data; tx_irq() local 794 tx_request_irq(struct hinic_txq *txq) tx_request_irq() argument 839 tx_free_irq(struct hinic_txq *txq) tx_free_irq() argument 855 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) hinic_init_txq() argument 920 hinic_clean_txq(struct hinic_txq *txq) hinic_clean_txq() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 75 * @txq: Logical Tx Queue 77 static void hinic_txq_clean_stats(struct hinic_txq *txq) in hinic_txq_clean_stats() argument 79 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_clean_stats() 93 * @txq: Logical Tx Queue 96 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) in hinic_txq_get_stats() argument 98 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_get_stats() 114 * @txq: Logical Tx Queue 116 static void txq_stats_init(struct hinic_txq *txq) in txq_stats_init() argument 118 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in txq_stats_init() 121 hinic_txq_clean_stats(txq); in txq_stats_init() 499 struct hinic_txq *txq; hinic_lb_xmit_frame() local 560 struct hinic_txq *txq; hinic_xmit_frame() local 665 free_all_tx_skbs(struct hinic_txq *txq) free_all_tx_skbs() argument 699 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); free_tx_poll() local 779 struct hinic_txq *txq = data; tx_irq() local 796 tx_request_irq(struct hinic_txq *txq) tx_request_irq() argument 842 tx_free_irq(struct hinic_txq *txq) tx_free_irq() argument 858 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) hinic_init_txq() argument 923 hinic_clean_txq(struct hinic_txq *txq) hinic_clean_txq() argument [all...] |
/kernel/linux/linux-6.6/include/net/ |
H A D | netdev_queues.h | 34 #define netif_txq_try_stop(txq, get_desc, start_thrs) \ 38 netif_tx_stop_queue(txq); \ 50 netif_tx_start_queue(txq); \ 58 * @txq: struct netdev_queue to stop/start 75 #define netif_txq_maybe_stop(txq, get_desc, stop_thrs, start_thrs) \ 81 _res = netif_txq_try_stop(txq, get_desc, start_thrs); \ 100 * @txq: struct netdev_queue to stop/start 118 #define __netif_txq_completed_wake(txq, pkts, bytes, \ 128 netdev_txq_completed_mb(txq, pkts, bytes); \ 133 if (unlikely(netif_tx_queue_stopped(txq)) [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | tx.c | 73 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr() 76 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr() 78 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr() 101 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr() 110 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr() 111 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr() 113 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr() 121 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 126 spin_lock_bh(&txq in iwl_pcie_txq_check_wrptrs() 72 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) iwl_pcie_txq_inc_wr_ptr() argument 135 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) iwl_pcie_txq_build_tfd() argument 189 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_unmap() local 239 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_txq_free() local 350 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_tx_reset() local 653 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_pcie_cmdq_reclaim() local 723 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_enable() local 833 struct iwl_txq *txq = trans->txqs.txq[txq_id]; iwl_trans_pcie_txq_set_shared_mode() local 888 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_enqueue_hcmd() local 1169 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; iwl_pcie_hcmd_complete() local 1227 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) iwl_fill_data_tbs() argument 1279 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument 1410 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) iwl_fill_data_tbs_amsdu() argument 1429 struct iwl_txq *txq; iwl_trans_pcie_tx() local [all...] |
H A D | tx-gen2.c | 31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local 112 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 114 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 115 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 118 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { in iwl_pcie_gen2_enqueue_hcmd() 119 spin_unlock_irqrestore(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 127 out_cmd = txq in iwl_pcie_gen2_enqueue_hcmd() [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/ath/ath9k/ |
H A D | xmit.c | 50 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 53 int tx_flags, struct ath_txq *txq, 56 struct ath_txq *txq, struct list_head *bf_q, 59 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 67 struct ath_txq *txq, 101 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq) 102 __releases(&txq->axq_lock) 109 skb_queue_splice_init(&txq->complete_q, &q); 110 spin_unlock_bh(&txq->axq_lock); 129 struct ath_txq *txq in ath9k_wake_tx_queue() local 164 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) ath_txq_skb_done() argument 189 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); ath_tid_pull() local 234 struct ath_txq *txq = tid->txq; ath_tx_flush_tid() local 316 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tid_drain() argument 344 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) ath_tx_set_retry() argument 439 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) ath_tx_complete_aggr() argument 681 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) ath_tx_process_buffer() argument 894 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) ath_tx_get_tid_subframe() argument 983 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_aggr() argument 1347 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) ath_tx_fill_desc() argument 1440 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_burst() argument 1473 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tx_sched_aggr() argument 1520 struct ath_txq *txq; ath_tx_aggr_start() local 1560 struct ath_txq *txq = txtid->txq; ath_tx_aggr_stop() local 1592 struct ath_txq *txq; ath_tx_aggr_wakeup() local 1636 struct ath_txq *txq = sc->tx.uapsdq; ath9k_release_buffered_frames() local 1750 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; ath_txq_setup() local 1812 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) ath_drain_txq_list() argument 1845 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) ath_draintxq() argument 1873 struct ath_txq *txq; ath_drain_all_txq() local 1911 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_cleanupq() argument 1920 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) ath_txq_schedule() argument 1961 struct ath_txq *txq; ath_txq_schedule_all() local 1981 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) ath_tx_txqaddbuf() argument 2048 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_send_normal() argument 2153 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_setup_buffer() argument 2282 struct ath_txq *txq = txctl->txq; ath_tx_start() local 2406 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) ath_tx_complete() argument 2458 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) ath_tx_complete_buf() argument 2571 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_processq() argument 2665 struct ath_txq *txq; ath_tx_edma_tasklet() local 2832 struct ath_txq *txq; ath_tx_node_cleanup() local [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/ath/ath9k/ |
H A D | xmit.c | 56 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 59 int tx_flags, struct ath_txq *txq, 62 struct ath_txq *txq, struct list_head *bf_q, 65 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 73 struct ath_txq *txq, 107 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq) 108 __releases(&txq->axq_lock) 115 skb_queue_splice_init(&txq->complete_q, &q); 116 spin_unlock_bh(&txq->axq_lock); 135 struct ath_txq *txq in ath9k_wake_tx_queue() local 211 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) ath_txq_skb_done() argument 236 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); ath_tid_pull() local 281 struct ath_txq *txq = tid->txq; ath_tx_flush_tid() local 363 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tid_drain() argument 391 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) ath_tx_set_retry() argument 488 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) ath_tx_complete_aggr() argument 730 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) ath_tx_process_buffer() argument 943 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) ath_tx_get_tid_subframe() argument 1032 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_aggr() argument 1408 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) ath_tx_fill_desc() argument 1501 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) ath_tx_form_burst() argument 1534 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) ath_tx_sched_aggr() argument 1581 struct ath_txq *txq; ath_tx_aggr_start() local 1621 struct ath_txq *txq = txtid->txq; ath_tx_aggr_stop() local 1653 struct ath_txq *txq; ath_tx_aggr_wakeup() local 1697 struct ath_txq *txq = sc->tx.uapsdq; ath9k_release_buffered_frames() local 1809 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; ath_txq_setup() local 1871 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) ath_drain_txq_list() argument 1904 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) ath_draintxq() argument 1932 struct ath_txq *txq; ath_drain_all_txq() local 1970 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_cleanupq() argument 1979 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) ath_txq_schedule() argument 2020 struct ath_txq *txq; ath_txq_schedule_all() local 2040 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) ath_tx_txqaddbuf() argument 2107 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_send_normal() argument 2212 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) ath_tx_setup_buffer() argument 2341 struct ath_txq *txq = txctl->txq; ath_tx_start() local 2465 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) ath_tx_complete() argument 2517 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) ath_tx_complete_buf() argument 2630 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) ath_tx_processq() argument 2724 struct ath_txq *txq; ath_tx_edma_tasklet() local 2891 struct ath_txq *txq; ath_tx_node_cleanup() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | selq.h | 30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix() argument 32 while (unlikely(txq >= num_channels)) in mlx5e_txq_to_ch_ix() 33 txq -= num_channels; in mlx5e_txq_to_ch_ix() 34 return txq; in mlx5e_txq_to_ch_ix() 37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix_htb() argument 39 if (unlikely(txq >= num_channels)) { in mlx5e_txq_to_ch_ix_htb() 40 if (unlikely(txq >= num_channels << 3)) in mlx5e_txq_to_ch_ix_htb() 41 txq %= num_channels; in mlx5e_txq_to_ch_ix_htb() 44 txq -= num_channels; in mlx5e_txq_to_ch_ix_htb() 45 while (txq > in mlx5e_txq_to_ch_ix_htb() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 54 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) in alx_free_txbuf() argument 56 struct alx_buffer *txb = &txq->bufs[entry]; in alx_free_txbuf() 59 dma_unmap_single(txq->dev, in alx_free_txbuf() 150 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping() 153 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) in alx_get_tx_queue() argument 155 return netdev_get_tx_queue(txq->netdev, txq->queue_idx); in alx_get_tx_queue() 158 static inline int alx_tpd_avail(struct alx_tx_queue *txq) in alx_tpd_avail() argument 160 if (txq->write_idx >= txq in alx_tpd_avail() 165 alx_clean_tx_irq(struct alx_tx_queue *txq) alx_clean_tx_irq() argument 489 alx_free_txring_buf(struct alx_tx_queue *txq) alx_free_txring_buf() argument 617 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) alx_alloc_tx_ring() argument 743 struct alx_tx_queue *txq; alx_alloc_napis() local 1437 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) alx_map_tx_skb() argument 1508 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) alx_start_xmit_ring() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 53 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) in alx_free_txbuf() argument 55 struct alx_buffer *txb = &txq->bufs[entry]; in alx_free_txbuf() 58 dma_unmap_single(txq->dev, in alx_free_txbuf() 149 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping() 152 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) in alx_get_tx_queue() argument 154 return netdev_get_tx_queue(txq->netdev, txq->queue_idx); in alx_get_tx_queue() 157 static inline int alx_tpd_avail(struct alx_tx_queue *txq) in alx_tpd_avail() argument 159 if (txq->write_idx >= txq in alx_tpd_avail() 164 alx_clean_tx_irq(struct alx_tx_queue *txq) alx_clean_tx_irq() argument 488 alx_free_txring_buf(struct alx_tx_queue *txq) alx_free_txring_buf() argument 616 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) alx_alloc_tx_ring() argument 742 struct alx_tx_queue *txq; alx_alloc_napis() local 1460 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) alx_map_tx_skb() argument 1531 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) alx_start_xmit_ring() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 178 #define IS_TSO_HEADER(txq, addr) \ 179 ((addr >= txq->tso_hdrs_dma) && \ 180 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 407 struct tx_queue txq[8]; member 439 /* rxq/txq helper functions *************************************************/ 445 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument 447 return container_of(txq, struct mv643xx_eth_private, txq[txq in txq_to_mp() 466 txq_reset_hw_ptr(struct tx_queue *txq) txq_reset_hw_ptr() argument 476 txq_enable(struct tx_queue *txq) txq_enable() argument 482 txq_disable(struct tx_queue *txq) txq_disable() argument 492 txq_maybe_wake(struct tx_queue *txq) txq_maybe_wake() argument 725 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) txq_put_data_tso() argument 773 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) txq_put_hdr_tso() argument 815 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_tso() argument 888 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) txq_submit_frag_skb() argument 926 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_skb() argument 994 struct tx_queue *txq; mv643xx_eth_xmit() local 1029 txq_kick(struct tx_queue *txq) txq_kick() argument 1054 txq_reclaim(struct tx_queue *txq, int budget, int force) txq_reclaim() argument 1162 txq_set_rate(struct tx_queue *txq, int rate, int burst) txq_set_rate() argument 1180 txq_set_fixed_prio_mode(struct tx_queue *txq) txq_set_fixed_prio_mode() argument 1259 struct tx_queue *txq = mp->txq + i; mv643xx_eth_get_stats() local 2025 struct tx_queue *txq = mp->txq + index; txq_init() local 2111 txq_deinit(struct tx_queue *txq) txq_deinit() argument 2201 struct tx_queue *txq = mp->txq + i; handle_link_event() local 2344 struct tx_queue *txq = mp->txq + i; port_start() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 179 #define IS_TSO_HEADER(txq, addr) \ 180 ((addr >= txq->tso_hdrs_dma) && \ 181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 408 struct tx_queue txq[8]; member 440 /* rxq/txq helper functions *************************************************/ 446 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument 448 return container_of(txq, struct mv643xx_eth_private, txq[txq in txq_to_mp() 467 txq_reset_hw_ptr(struct tx_queue *txq) txq_reset_hw_ptr() argument 477 txq_enable(struct tx_queue *txq) txq_enable() argument 483 txq_disable(struct tx_queue *txq) txq_disable() argument 493 txq_maybe_wake(struct tx_queue *txq) txq_maybe_wake() argument 727 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) txq_put_data_tso() argument 775 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) txq_put_hdr_tso() argument 818 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_tso() argument 891 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) txq_submit_frag_skb() argument 929 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) txq_submit_skb() argument 997 struct tx_queue *txq; mv643xx_eth_xmit() local 1032 txq_kick(struct tx_queue *txq) txq_kick() argument 1057 txq_reclaim(struct tx_queue *txq, int budget, int force) txq_reclaim() argument 1165 txq_set_rate(struct tx_queue *txq, int rate, int burst) txq_set_rate() argument 1183 txq_set_fixed_prio_mode(struct tx_queue *txq) txq_set_fixed_prio_mode() argument 1262 struct tx_queue *txq = mp->txq + i; mv643xx_eth_get_stats() local 2036 struct tx_queue *txq = mp->txq + index; txq_init() local 2122 txq_deinit(struct tx_queue *txq) txq_deinit() argument 2212 struct tx_queue *txq = mp->txq + i; handle_link_event() local 2355 struct tx_queue *txq = mp->txq + i; port_start() local [all...] |
H A D | mvneta.c | 135 #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8) 773 /* Increment txq get counter */ 774 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) in mvneta_txq_inc_get() argument 776 txq->txq_get_index++; in mvneta_txq_inc_get() 777 if (txq->txq_get_index == txq->size) in mvneta_txq_inc_get() 778 txq->txq_get_index = 0; in mvneta_txq_inc_get() 781 /* Increment txq put counter */ 782 static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq) in mvneta_txq_inc_put() argument 963 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) mvneta_txq_pend_desc_add() argument 982 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) mvneta_txq_next_desc_get() argument 993 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) mvneta_txq_desc_put() argument 1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_port_up() local 1498 int rxq, txq; mvneta_defaults_set() local 1708 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) mvneta_tx_done_pkts_coal_set() argument 1734 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) mvneta_txq_sent_desc_dec() argument 1752 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sent_desc_num_get() argument 1768 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sent_desc_proc() argument 1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) mvneta_txq_bufs_free() argument 1913 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_done() argument 2091 mvneta_xdp_submit_frame(struct mvneta_port *pp, struct mvneta_tx_queue *txq, struct xdp_frame *xdpf, int *nxmit_byte, bool dma_map) mvneta_xdp_submit_frame() argument 2177 struct mvneta_tx_queue *txq; mvneta_xdp_xmit_back() local 2219 struct mvneta_tx_queue *txq; mvneta_xdp_xmit() local 2670 mvneta_free_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_free_tso_hdrs() argument 2686 mvneta_alloc_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_alloc_tso_hdrs() argument 2706 mvneta_get_tso_hdr(struct mvneta_tx_queue *txq, dma_addr_t *dma) mvneta_get_tso_hdr() argument 2718 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_tx_queue *txq, struct tso_t *tso, int size, bool is_last) mvneta_tso_put_hdr() argument 2742 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) mvneta_tso_put_data() argument 2775 mvneta_release_descs(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int first, int num) mvneta_release_descs() argument 2803 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) mvneta_tx_tso() argument 2862 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) mvneta_tx_frag_process() argument 2915 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; mvneta_tx() local 3005 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_done_force() argument 3025 struct mvneta_tx_queue *txq; mvneta_tx_done_gbe() local 3524 mvneta_txq_sw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sw_init() argument 3569 mvneta_txq_hw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_hw_init() argument 3584 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_init() argument 3599 mvneta_txq_sw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_sw_deinit() argument 3621 mvneta_txq_hw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_hw_deinit() argument 3633 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) mvneta_txq_deinit() argument 4691 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_ethtool_set_coalesce() local 5221 int rxq, txq, tc, ret; mvneta_setup_mqprio() local 5356 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_init() local 5795 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_suspend() local 5848 struct mvneta_tx_queue *txq = &pp->txqs[queue]; mvneta_resume() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 75 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) in qede_free_tx_pkt() argument 77 u16 idx = txq->sw_tx_cons; in qede_free_tx_pkt() 78 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 83 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 88 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n", in qede_free_tx_pkt() 89 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt() 95 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq in qede_free_tx_pkt() 130 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) qede_free_failed_tx_pkt() argument 242 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) map_frag_to_bd() argument 289 qede_update_tx_producer(struct qede_tx_queue *txq) qede_update_tx_producer() argument 305 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) qede_xdp_xmit() argument 393 qede_txq_has_work(struct qede_tx_queue *txq) qede_txq_has_work() argument 406 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_xdp_tx_int() argument 438 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_tx_int() argument 1500 struct qede_tx_queue *txq; qede_start_xmit() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 76 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) in qede_free_tx_pkt() argument 78 u16 idx = txq->sw_tx_cons; in qede_free_tx_pkt() 79 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 84 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 89 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n", in qede_free_tx_pkt() 90 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt() 96 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq in qede_free_tx_pkt() 131 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) qede_free_failed_tx_pkt() argument 243 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) map_frag_to_bd() argument 288 qede_update_tx_producer(struct qede_tx_queue *txq) qede_update_tx_producer() argument 304 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) qede_xdp_xmit() argument 387 qede_txq_has_work(struct qede_tx_queue *txq) qede_txq_has_work() argument 400 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_xdp_tx_int() argument 432 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) qede_tx_int() argument 1482 struct qede_tx_queue *txq; qede_start_xmit() local [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/ |
H A D | tx.c | 9 mt76_txq_get_qid(struct ieee80211_txq *txq) in mt76_txq_get_qid() argument 11 if (!txq->sta) in mt76_txq_get_qid() 14 return txq->ac; in mt76_txq_get_qid() 21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local 30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn() 31 mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_tx_check_agg_ssn() 306 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); in mt76_txq_dequeue() local 311 skb = ieee80211_tx_dequeue(phy->hw, txq); in mt76_txq_dequeue() 352 struct ieee80211_txq *txq in mt76_release_buffered_frames() local 388 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); mt76_txq_send_burst() local 453 struct ieee80211_txq *txq; mt76_txq_schedule_list() local 479 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); mt76_txq_schedule_list() local 547 struct ieee80211_txq *txq = sta->txq[i]; mt76_stop_tx_queues() local 564 mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq) mt76_wake_tx_queue() argument [all...] |