Home
last modified time | relevance | path

Searched refs:rxq (Results 1 - 25 of 324) sorted by relevance

12345678910>>...13

/kernel/linux/linux-6.6/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_rx.c83 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local
84 struct dpmaif_bat_request *bat_req = rxq->bat_req; in t7xx_dpmaif_update_bat_wr_idx()
87 if (!rxq->que_started) { in t7xx_dpmaif_update_bat_wr_idx()
88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); in t7xx_dpmaif_update_bat_wr_idx()
235 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, in t7xx_dpmaifq_release_pit_entry() argument
238 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info; in t7xx_dpmaifq_release_pit_entry()
242 if (!rxq->que_started) in t7xx_dpmaifq_release_pit_entry()
245 if (rel_entry_num >= rxq->pit_size_cnt) { in t7xx_dpmaifq_release_pit_entry()
246 dev_err(rxq in t7xx_dpmaifq_release_pit_entry()
275 t7xx_frag_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) t7xx_frag_bat_cur_bid_check() argument
392 t7xx_dpmaif_set_frag_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct sk_buff *skb) t7xx_dpmaif_set_frag_to_skb() argument
424 t7xx_dpmaif_get_frag(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, const struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_get_frag() argument
445 t7xx_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) t7xx_bat_cur_bid_check() argument
461 t7xx_dpmaif_check_pit_seq(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pit) t7xx_dpmaif_check_pit_seq() argument
499 t7xx_dpmaif_release_bat_entry(const struct dpmaif_rx_queue *rxq, const unsigned int rel_entry_num, const enum bat_type buf_type) t7xx_dpmaif_release_bat_entry() argument
555 t7xx_dpmaif_pit_release_and_add(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_pit_release_and_add() argument
570 t7xx_dpmaif_bat_release_and_add(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_bat_release_and_add() argument
592 t7xx_dpmaif_frag_bat_release_and_add(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_frag_bat_release_and_add() argument
610 t7xx_dpmaif_parse_msg_pit(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *msg_pit, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_parse_msg_pit() argument
622 t7xx_dpmaif_set_data_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_set_data_to_skb() argument
658 t7xx_dpmaif_get_rx_pkt(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_get_rx_pkt() argument
679 t7xx_dpmaifq_rx_notify_hw(struct dpmaif_rx_queue *rxq) t7xx_dpmaifq_rx_notify_hw() argument
693 t7xx_dpmaif_rx_skb(struct dpmaif_rx_queue *rxq, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_rx_skb() argument
717 t7xx_dpmaif_rx_start(struct dpmaif_rx_queue *rxq, const unsigned int pit_cnt, const unsigned int budget, int *once_more) t7xx_dpmaif_rx_start() argument
798 t7xx_dpmaifq_poll_pit(struct dpmaif_rx_queue *rxq) t7xx_dpmaifq_poll_pit() argument
816 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; t7xx_dpmaif_napi_rx_data_collect() local
833 struct dpmaif_rx_queue *rxq = container_of(napi, struct dpmaif_rx_queue, napi); t7xx_dpmaif_napi_rx_poll() local
891 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_irq_rx_done() local
1006 t7xx_dpmaif_rx_alloc(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_rx_alloc() argument
1030 t7xx_dpmaif_rx_buf_free(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_rx_buf_free() argument
1063 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_bat_release_work() local
1116 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[i]; t7xx_dpmaif_rx_stop() local
1130 t7xx_dpmaif_stop_rxq(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_stop_rxq() argument
[all...]
H A Dt7xx_hif_dpmaif.c227 rx_q = &dpmaif_ctrl->rxq[rx_idx]; in t7xx_dpmaif_rxtx_sw_allocs()
267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
294 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_sw_release()
303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local
314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start()
315 rxq->que_started = true; in t7xx_dpmaif_start()
316 rxq->index = i; in t7xx_dpmaif_start()
317 rxq->budget = rxq in t7xx_dpmaif_start()
432 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_start_txrx_qs() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/
H A Dhinic_rx.c51 * @rxq: Logical Rx Queue
53 void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument
55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats()
68 * @rxq: Logical Rx Queue
71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument
73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats()
89 * @rxq: Logical Rx Queue
91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument
93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init()
96 hinic_rxq_clean_stats(rxq); in rxq_stats_init()
99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) rx_csum() argument
126 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) rx_alloc_skb() argument
164 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) rx_unmap_skb() argument
181 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) rx_free_skb() argument
194 rx_alloc_pkts(struct hinic_rxq *rxq) rx_alloc_pkts() argument
246 free_all_rx_skbs(struct hinic_rxq *rxq) free_all_rx_skbs() argument
274 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) rx_recv_jumbo_pkt() argument
357 rxq_recv(struct hinic_rxq *rxq, int budget) rxq_recv() argument
462 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); rx_poll() local
481 rx_add_napi(struct hinic_rxq *rxq) rx_add_napi() argument
489 rx_del_napi(struct hinic_rxq *rxq) rx_del_napi() argument
497 struct hinic_rxq *rxq = (struct hinic_rxq *)data; rx_irq() local
515 rx_request_irq(struct hinic_rxq *rxq) rx_request_irq() argument
565 rx_free_irq(struct hinic_rxq *rxq) rx_free_irq() argument
582 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) hinic_init_rxq() argument
625 hinic_clean_rxq(struct hinic_rxq *rxq) hinic_clean_rxq() argument
[all...]
H A Dhinic_rx.h44 void hinic_rxq_clean_stats(struct hinic_rxq *rxq);
46 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats);
48 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq,
51 void hinic_clean_rxq(struct hinic_rxq *rxq);
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/
H A Dhinic_rx.c51 * @rxq: Logical Rx Queue
53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument
55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats()
68 * @rxq: Logical Rx Queue
71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument
73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats()
89 * @rxq: Logical Rx Queue
91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument
93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init()
96 hinic_rxq_clean_stats(rxq); in rxq_stats_init()
99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) rx_csum() argument
127 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) rx_alloc_skb() argument
163 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) rx_unmap_skb() argument
180 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) rx_free_skb() argument
193 rx_alloc_pkts(struct hinic_rxq *rxq) rx_alloc_pkts() argument
243 free_all_rx_skbs(struct hinic_rxq *rxq) free_all_rx_skbs() argument
271 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) rx_recv_jumbo_pkt() argument
354 rxq_recv(struct hinic_rxq *rxq, int budget) rxq_recv() argument
459 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); rx_poll() local
478 rx_add_napi(struct hinic_rxq *rxq) rx_add_napi() argument
487 rx_del_napi(struct hinic_rxq *rxq) rx_del_napi() argument
495 struct hinic_rxq *rxq = (struct hinic_rxq *)data; rx_irq() local
513 rx_request_irq(struct hinic_rxq *rxq) rx_request_irq() argument
563 rx_free_irq(struct hinic_rxq *rxq) rx_free_irq() argument
580 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) hinic_init_rxq() argument
623 hinic_clean_rxq(struct hinic_rxq *rxq) hinic_clean_rxq() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/pcie/
H A Drx.c55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free.
58 * or reused - added immediately to the iwl->rxq->rx_free list.
66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them
76 * detached from the iwl->rxq. The driver 'processed' index is updated.
77 * + If there are no allocated buffers in iwl->rxq->rx_free,
104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue
108 * rxq
119 iwl_rxq_space(const struct iwl_rxq *rxq) iwl_rxq_space() argument
166 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_inc_wr_ptr() argument
209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local
220 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) iwl_pcie_restock_bd() argument
244 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxmq_restock() argument
291 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxsq_restock() argument
349 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_restock() argument
429 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) iwl_pcie_rxq_alloc_rbs() argument
613 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_allocator_get() argument
678 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_free_rxq_dma() argument
713 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_alloc_rxq_dma() argument
757 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_alloc_rxq_dma() local
807 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_alloc() local
835 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_hw_init() argument
991 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) iwl_pcie_rx_init_rxb_lists() argument
1005 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); iwl_pcie_napi_poll() local
1032 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); iwl_pcie_napi_poll_msix() local
1071 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_napi_sync() local
1108 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; _iwl_pcie_rx_init() local
1229 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_free() local
1246 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) iwl_pcie_rx_move_to_allocator() argument
1260 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) iwl_pcie_rx_reuse_rbd() argument
1292 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) iwl_pcie_rx_handle_rb() argument
1442 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) iwl_pcie_get_rxb() argument
1500 struct iwl_rxq *rxq; iwl_pcie_rx_handle() local
1643 struct iwl_rxq *rxq; iwl_pcie_irq_rx_msix_handler() local
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/pcie/
H A Drx.c112 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free.
115 * or reused - added immediately to the iwl->rxq->rx_free list.
123 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them
133 * detached from the iwl->rxq. The driver 'processed' index is updated.
134 * + If there are no allocated buffers in iwl->rxq->rx_free,
161 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue
165 * rxq
176 iwl_rxq_space(const struct iwl_rxq *rxq) iwl_rxq_space() argument
223 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_inc_wr_ptr() argument
263 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local
274 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) iwl_pcie_restock_bd() argument
298 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxmq_restock() argument
345 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxsq_restock() argument
403 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_restock() argument
483 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) iwl_pcie_rxq_alloc_rbs() argument
667 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_allocator_get() argument
723 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_free_rxq_dma() argument
765 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_alloc_rxq_dma() argument
826 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_alloc_rxq_dma() local
878 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_alloc() local
903 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_hw_init() argument
1061 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) iwl_pcie_rx_init_rxb_lists() argument
1107 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; _iwl_pcie_rx_init() local
1222 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_free() local
1237 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) iwl_pcie_rx_move_to_allocator() argument
1251 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) iwl_pcie_rx_reuse_rbd() argument
1283 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) iwl_pcie_rx_handle_rb() argument
1432 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) iwl_pcie_get_rxb() argument
1481 struct iwl_rxq *rxq; iwl_pcie_rx_handle() local
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/
H A Dnetdev_rx.c194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init()
197 if (!rx->rxq) { in hfi1_netdev_rxq_init()
203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local
205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); in hfi1_netdev_rxq_init()
209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init()
210 rxq->rx = rx; in hfi1_netdev_rxq_init()
211 rxq->rcd->napi = &rxq in hfi1_netdev_rxq_init()
230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; hfi1_netdev_rxq_init() local
250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; hfi1_netdev_rxq_deinit() local
268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; enable_queues() local
286 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; disable_queues() local
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
H A Dnetdev_rx.c194 priv->rxq = kcalloc_node(priv->num_rx_q, sizeof(struct hfi1_netdev_rxq), in hfi1_netdev_rxq_init()
197 if (!priv->rxq) { in hfi1_netdev_rxq_init()
203 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; in hfi1_netdev_rxq_init() local
205 rc = hfi1_netdev_allot_ctxt(priv, &rxq->rcd); in hfi1_netdev_rxq_init()
209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init()
210 rxq->priv = priv; in hfi1_netdev_rxq_init()
211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init()
213 i, rxq in hfi1_netdev_rxq_init()
230 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; hfi1_netdev_rxq_init() local
251 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; hfi1_netdev_rxq_deinit() local
269 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; enable_queues() local
287 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; disable_queues() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c26 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument
37 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer()
38 rxq->filled_buffers--; in qede_alloc_rx_buffer()
49 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer()
50 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer()
51 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer()
56 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer()
62 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer()
66 rxq in qede_alloc_rx_buffer()
508 qede_has_rx_work(struct qede_rx_queue *rxq) qede_has_rx_work() argument
521 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) qede_rx_bd_ring_consume() argument
530 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_reuse_page() argument
553 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) qede_recycle_rx_bd_ring() argument
564 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_realloc_rx_buffer() argument
594 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_update_rx_prod() argument
643 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) qede_skb_receive() argument
670 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) qede_fill_frag_skb() argument
746 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_build_skb() argument
765 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) qede_tpa_rx_build_skb() argument
796 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_rx_build_skb() argument
837 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) qede_tpa_start() argument
963 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) qede_tpa_cont() argument
982 struct qede_rx_queue *rxq = fp->rxq; qede_tpa_end() local
1085 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) qede_rx_xdp() argument
1184 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) qede_rx_build_jumbo() argument
1241 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) qede_rx_process_tpa_cqe() argument
1261 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) qede_rx_process_cqe() argument
1359 struct qede_rx_queue *rxq = fp->rxq; qede_rx_int() local
[all...]
H A Dqede_main.c902 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array()
903 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); in qede_free_fp_array()
904 kfree(fp->rxq); in qede_free_fp_array()
963 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); in qede_alloc_fp_array()
964 if (!fp->rxq) in qede_alloc_fp_array()
1460 struct qede_rx_queue *rxq) in qede_free_rx_buffers()
1464 for (i = rxq->sw_rx_cons; i != rxq in qede_free_rx_buffers()
1459 qede_free_rx_buffers(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_free_rx_buffers() argument
1479 qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_free_mem_rxq() argument
1492 qede_set_tpa_param(struct qede_rx_queue *rxq) qede_set_tpa_param() argument
1504 qede_alloc_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_alloc_mem_rxq() argument
2215 struct qede_rx_queue *rxq = fp->rxq; qede_start_queues() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c27 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument
38 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer()
39 rxq->filled_buffers--; in qede_alloc_rx_buffer()
50 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer()
51 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer()
52 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer()
57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer()
63 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer()
67 rxq in qede_alloc_rx_buffer()
502 qede_has_rx_work(struct qede_rx_queue *rxq) qede_has_rx_work() argument
515 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) qede_rx_bd_ring_consume() argument
524 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_reuse_page() argument
547 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) qede_recycle_rx_bd_ring() argument
558 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_realloc_rx_buffer() argument
588 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_update_rx_prod() argument
637 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) qede_skb_receive() argument
664 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) qede_fill_frag_skb() argument
740 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_build_skb() argument
759 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) qede_tpa_rx_build_skb() argument
790 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_rx_build_skb() argument
831 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) qede_tpa_start() argument
957 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) qede_tpa_cont() argument
976 struct qede_rx_queue *rxq = fp->rxq; qede_tpa_end() local
1079 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) qede_rx_xdp() argument
1169 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) qede_rx_build_jumbo() argument
1223 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) qede_rx_process_tpa_cqe() argument
1243 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) qede_rx_process_cqe() argument
1341 struct qede_rx_queue *rxq = fp->rxq; qede_rx_int() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c548 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) in mana_get_rxbuf_pre() argument
550 struct net_device *ndev = rxq->ndev; in mana_get_rxbuf_pre()
562 if (mpc->rxbpre_datasize != rxq->datasize) { in mana_get_rxbuf_pre()
564 mpc->rxbpre_datasize, rxq->datasize); in mana_get_rxbuf_pre()
568 if (mpc->rxbpre_alloc_size != rxq->alloc_size) { in mana_get_rxbuf_pre()
570 mpc->rxbpre_alloc_size, rxq->alloc_size); in mana_get_rxbuf_pre()
574 if (mpc->rxbpre_headroom != rxq->headroom) { in mana_get_rxbuf_pre()
576 mpc->rxbpre_headroom, rxq->headroom); in mana_get_rxbuf_pre()
1258 static int mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) in mana_fence_rq() argument
1264 init_completion(&rxq in mana_fence_rq()
1300 struct mana_rxq *rxq; mana_fence_rqs() local
1461 mana_post_pkt_rxq(struct mana_rxq *rxq) mana_post_pkt_rxq() argument
1481 mana_build_skb(struct mana_rxq *rxq, void *buf_va, uint pkt_len, struct xdp_buff *xdp) mana_build_skb() argument
1501 mana_rx_skb(void *buf_va, bool from_pool, struct mana_rxcomp_oob *cqe, struct mana_rxq *rxq) mana_rx_skb() argument
1602 mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev, dma_addr_t *da, bool *from_pool, bool is_napi) mana_get_rxfrag() argument
1653 mana_refill_rx_oob(struct device *dev, struct mana_rxq *rxq, struct mana_recv_buf_oob *rxoob, void **old_buf, bool *old_fp) mana_refill_rx_oob() argument
1675 mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq, struct gdma_comp *cqe) mana_process_rx_cqe() argument
1745 struct mana_rxq *rxq = cq->rxq; mana_poll_rx_cq() local
1986 mana_destroy_rxq(struct mana_port_context *apc, struct mana_rxq *rxq, bool validate_state) mana_destroy_rxq() argument
2045 mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key, struct mana_rxq *rxq, struct device *dev) mana_fill_rx_oob() argument
2074 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) mana_alloc_rx_wqe() argument
2114 mana_push_wqe(struct mana_rxq *rxq) mana_push_wqe() argument
2132 mana_create_page_pool(struct mana_rxq *rxq, struct gdma_context *gc) mana_create_page_pool() argument
2163 struct mana_rxq *rxq; mana_create_rxq() local
2283 struct mana_rxq *rxq; mana_add_rx_queues() local
2307 struct mana_rxq *rxq; mana_destroy_vport() local
[all...]
/kernel/linux/linux-5.10/drivers/atm/
H A Dambassador.c673 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local
678 spin_lock_irqsave (&rxq->lock, flags); in rx_give()
680 if (rxq->pending < rxq->maximum) { in rx_give()
681 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give()
683 *rxq->in.ptr = *rx; in rx_give()
684 rxq->pending++; in rx_give()
685 rxq->in.ptr = NEXTQ (rxq in rx_give()
698 amb_rxq * rxq = &dev->rxq[pool]; rx_take() local
731 amb_rxq * rxq = &dev->rxq[pool]; drain_rx_pool() local
769 amb_rxq * rxq; fill_rx_pool() local
1575 amb_rxq * rxq = &dev->rxq[pool]; create_queues() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/
H A Dmvneta.c109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3))
134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq)
857 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add()
864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
876 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get()
880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
888 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update()
856 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) mvneta_rxq_non_occup_desc_add() argument
875 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_busy_desc_num_get() argument
887 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) mvneta_rxq_desc_num_update() argument
922 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) mvneta_rxq_next_desc_get() argument
945 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) mvneta_rxq_offset_set() argument
1002 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) mvneta_rxq_buf_size_set() argument
1017 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_disable() argument
1028 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_enable() argument
1039 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_long_pool_set() argument
1052 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_short_pool_set() argument
1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_port_up() local
1498 int rxq, txq; mvneta_defaults_set() local
1685 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_pkts_coal_set() argument
1695 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_time_coal_set() argument
1722 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) mvneta_rx_desc_fill() argument
1935 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) mvneta_rx_refill() argument
1986 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_drop_pkts() argument
2041 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rx_refill_queue() argument
2072 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len) mvneta_xdp_put_buff() argument
2257 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) mvneta_run_xdp() argument
2314 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_rx_frame() argument
2349 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_add_rx_fragment() argument
2425 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) mvneta_rx_swbm() argument
2538 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) mvneta_rx_hwbm() argument
3346 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) mvneta_create_page_pool() argument
3390 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) mvneta_rxq_fill() argument
3439 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_sw_init() argument
3456 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_hw_init() argument
3489 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_init() argument
3505 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_deinit() argument
4343 int rxq; mvneta_percpu_elect() local
4683 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_ethtool_set_coalesce() local
5144 mvneta_map_vlan_prio_to_rxq(struct mvneta_port *pp, u8 pri, u8 rxq) mvneta_map_vlan_prio_to_rxq() argument
5221 int rxq, txq, tc, ret; mvneta_setup_mqprio() local
5368 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_init() local
5789 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_suspend() local
5841 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_resume() local
[all...]
H A Dmv643xx_eth.c399 struct rx_queue rxq[8]; member
440 /* rxq/txq helper functions *************************************************/
441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
451 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
457 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
506 rxq_process(struct rx_queue *rxq, int budget) rxq_process() argument
598 rxq_refill(struct rx_queue *rxq, int budget) rxq_refill() argument
1940 struct rx_queue *rxq = mp->rxq + index; rxq_init() local
2005 rxq_deinit(struct rx_queue *rxq) rxq_deinit() argument
2383 struct rx_queue *rxq = mp->rxq + i; port_start() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/
H A Dmvneta.c129 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq)
817 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add()
824 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
830 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
836 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get()
840 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
848 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update()
856 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
876 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq in mvneta_rxq_desc_num_update()
816 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) mvneta_rxq_non_occup_desc_add() argument
835 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_busy_desc_num_get() argument
847 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) mvneta_rxq_desc_num_update() argument
882 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) mvneta_rxq_next_desc_get() argument
905 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) mvneta_rxq_offset_set() argument
962 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) mvneta_rxq_buf_size_set() argument
977 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_disable() argument
988 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_enable() argument
999 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_long_pool_set() argument
1012 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_short_pool_set() argument
1229 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_port_up() local
1458 int rxq, txq; mvneta_defaults_set() local
1645 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_pkts_coal_set() argument
1655 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_time_coal_set() argument
1682 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) mvneta_rx_desc_fill() argument
1890 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) mvneta_rx_refill() argument
1941 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_drop_pkts() argument
1996 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rx_refill_queue() argument
2027 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len, bool napi) mvneta_xdp_put_buff() argument
2170 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) mvneta_run_xdp() argument
2227 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_rx_frame() argument
2268 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_add_rx_fragment() argument
2306 mvneta_swbm_build_skb(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, u32 desc_status) mvneta_swbm_build_skb() argument
2336 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) mvneta_rx_swbm() argument
2451 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) mvneta_rx_hwbm() argument
3200 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) mvneta_create_page_pool() argument
3243 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) mvneta_rxq_fill() argument
3292 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_sw_init() argument
3309 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_hw_init() argument
3342 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_init() argument
3358 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_deinit() argument
4150 int rxq; mvneta_percpu_elect() local
4490 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_ethtool_set_coalesce() local
4976 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_init() local
5357 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_suspend() local
5409 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_resume() local
[all...]
H A Dmv643xx_eth.c398 struct rx_queue rxq[8]; member
439 /* rxq/txq helper functions *************************************************/
440 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
442 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
450 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
452 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
453 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
456 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
505 rxq_process(struct rx_queue *rxq, int budget) rxq_process() argument
597 rxq_refill(struct rx_queue *rxq, int budget) rxq_refill() argument
1929 struct rx_queue *rxq = mp->rxq + index; rxq_init() local
1994 rxq_deinit(struct rx_queue *rxq) rxq_deinit() argument
2372 struct rx_queue *rxq = mp->rxq + i; port_start() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/atheros/alx/
H A Dmain.c74 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local
80 next = cur = rxq->write_idx; in alx_refill_rx_ring()
83 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
85 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
86 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
128 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
135 rxq->write_idx = cur; in alx_refill_rx_ring()
217 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument
226 alx = netdev_priv(rxq in alx_clean_rx_irq()
507 alx_free_rxring_buf(struct alx_rx_queue *rxq) alx_free_rxring_buf() argument
631 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) alx_alloc_rx_ring() argument
742 struct alx_rx_queue *rxq; alx_alloc_napis() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/alx/
H A Dmain.c73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local
79 next = cur = rxq->write_idx; in alx_refill_rx_ring()
82 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
84 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
85 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
127 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
134 rxq->write_idx = cur; in alx_refill_rx_ring()
216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument
225 alx = netdev_priv(rxq in alx_clean_rx_irq()
506 alx_free_rxring_buf(struct alx_rx_queue *rxq) alx_free_rxring_buf() argument
630 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) alx_alloc_rx_ring() argument
741 struct alx_rx_queue *rxq; alx_alloc_napis() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/
H A Dhisi_femac.c121 struct hisi_femac_queue rxq; member
212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local
218 pos = rxq->head; in hisi_femac_rx_refill()
220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill()
222 if (unlikely(rxq->skb[pos])) { in hisi_femac_rx_refill()
224 pos, rxq->skb[pos]); in hisi_femac_rx_refill()
237 rxq->dma_phys[pos] = addr; in hisi_femac_rx_refill()
238 rxq in hisi_femac_rx_refill()
248 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_rx() local
388 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_free_skb_rings() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/
H A Dhisi_femac.c121 struct hisi_femac_queue rxq; member
212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local
218 pos = rxq->head; in hisi_femac_rx_refill()
220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill()
222 if (unlikely(rxq->skb[pos])) { in hisi_femac_rx_refill()
224 pos, rxq->skb[pos]); in hisi_femac_rx_refill()
237 rxq->dma_phys[pos] = addr; in hisi_femac_rx_refill()
238 rxq in hisi_femac_rx_refill()
248 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_rx() local
388 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_free_skb_rings() local
[all...]
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/
H A Dxdp_hw_metadata.c53 int rxq; variable
231 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) in verify_metadata() argument
234 struct pollfd fds[rxq + 1]; in verify_metadata()
241 for (i = 0; i < rxq; i++) { in verify_metadata()
247 fds[rxq].fd = server_fd; in verify_metadata()
248 fds[rxq].events = POLLIN; in verify_metadata()
249 fds[rxq].revents = 0; in verify_metadata()
253 ret = poll(fds, rxq + 1, 1000); in verify_metadata()
262 if (fds[rxq].revents) in verify_metadata()
265 for (i = 0; i < rxq; in verify_metadata()
[all...]
/kernel/linux/linux-5.10/samples/bpf/
H A Dxdp_rxq_info_user.c194 struct record *rxq; member
234 rec->rxq = alloc_record_per_rxq(); in alloc_stats_record()
236 rec->rxq[i].cpu = alloc_record_per_cpu(); in alloc_stats_record()
248 free(r->rxq[i].cpu); in free_stats_record()
250 free(r->rxq); in free_stats_record()
294 map_collect_percpu(fd, i, &rec->rxq[i]); in stats_collect()
343 int rxq; in stats_print() local
383 for (rxq = 0; rxq < nr_rxqs; rxq in stats_print()
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_trace.h75 TP_PROTO(const struct funeth_rxq *rxq,
81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec),
90 __string(devname, rxq->netdev->name)
94 __entry->qidx = rxq->qidx;
95 __entry->cq_head = rxq->cq_head;
100 __assign_str(devname, rxq->netdev->name);

Completed in 37 milliseconds

12345678910>>...13