/kernel/linux/linux-6.6/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_rx.c | 83 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local 84 struct dpmaif_bat_request *bat_req = rxq->bat_req; in t7xx_dpmaif_update_bat_wr_idx() 87 if (!rxq->que_started) { in t7xx_dpmaif_update_bat_wr_idx() 88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); in t7xx_dpmaif_update_bat_wr_idx() 235 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, in t7xx_dpmaifq_release_pit_entry() argument 238 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info; in t7xx_dpmaifq_release_pit_entry() 242 if (!rxq->que_started) in t7xx_dpmaifq_release_pit_entry() 245 if (rel_entry_num >= rxq->pit_size_cnt) { in t7xx_dpmaifq_release_pit_entry() 246 dev_err(rxq in t7xx_dpmaifq_release_pit_entry() 275 t7xx_frag_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) t7xx_frag_bat_cur_bid_check() argument 392 t7xx_dpmaif_set_frag_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct sk_buff *skb) t7xx_dpmaif_set_frag_to_skb() argument 424 t7xx_dpmaif_get_frag(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, const struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_get_frag() argument 445 t7xx_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) t7xx_bat_cur_bid_check() argument 461 t7xx_dpmaif_check_pit_seq(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pit) t7xx_dpmaif_check_pit_seq() argument 499 t7xx_dpmaif_release_bat_entry(const struct dpmaif_rx_queue *rxq, const unsigned int rel_entry_num, const enum bat_type buf_type) t7xx_dpmaif_release_bat_entry() argument 555 t7xx_dpmaif_pit_release_and_add(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_pit_release_and_add() argument 570 t7xx_dpmaif_bat_release_and_add(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_bat_release_and_add() argument 592 t7xx_dpmaif_frag_bat_release_and_add(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_frag_bat_release_and_add() argument 610 t7xx_dpmaif_parse_msg_pit(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *msg_pit, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_parse_msg_pit() argument 622 t7xx_dpmaif_set_data_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_set_data_to_skb() argument 658 t7xx_dpmaif_get_rx_pkt(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_get_rx_pkt() argument 679 t7xx_dpmaifq_rx_notify_hw(struct dpmaif_rx_queue *rxq) t7xx_dpmaifq_rx_notify_hw() argument 693 t7xx_dpmaif_rx_skb(struct dpmaif_rx_queue *rxq, struct dpmaif_cur_rx_skb_info *skb_info) t7xx_dpmaif_rx_skb() argument 717 t7xx_dpmaif_rx_start(struct dpmaif_rx_queue *rxq, const unsigned int pit_cnt, const unsigned int budget, int *once_more) t7xx_dpmaif_rx_start() argument 798 t7xx_dpmaifq_poll_pit(struct dpmaif_rx_queue *rxq) t7xx_dpmaifq_poll_pit() argument 816 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; t7xx_dpmaif_napi_rx_data_collect() local 833 struct dpmaif_rx_queue *rxq = container_of(napi, struct dpmaif_rx_queue, napi); t7xx_dpmaif_napi_rx_poll() local 891 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_irq_rx_done() local 1006 t7xx_dpmaif_rx_alloc(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_rx_alloc() argument 1030 t7xx_dpmaif_rx_buf_free(const struct dpmaif_rx_queue *rxq) t7xx_dpmaif_rx_buf_free() argument 1063 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_bat_release_work() local 1116 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[i]; t7xx_dpmaif_rx_stop() local 1130 t7xx_dpmaif_stop_rxq(struct dpmaif_rx_queue *rxq) t7xx_dpmaif_stop_rxq() argument [all...] |
H A D | t7xx_hif_dpmaif.c | 227 rx_q = &dpmaif_ctrl->rxq[rx_idx]; in t7xx_dpmaif_rxtx_sw_allocs() 267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs() 294 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_sw_release() 303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local 314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start() 315 rxq->que_started = true; in t7xx_dpmaif_start() 316 rxq->index = i; in t7xx_dpmaif_start() 317 rxq->budget = rxq in t7xx_dpmaif_start() 432 struct dpmaif_rx_queue *rxq; t7xx_dpmaif_start_txrx_qs() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_rx.c | 51 * @rxq: Logical Rx Queue 53 void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument 55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats() 68 * @rxq: Logical Rx Queue 71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument 73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats() 89 * @rxq: Logical Rx Queue 91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument 93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init() 96 hinic_rxq_clean_stats(rxq); in rxq_stats_init() 99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) rx_csum() argument 126 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) rx_alloc_skb() argument 164 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) rx_unmap_skb() argument 181 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) rx_free_skb() argument 194 rx_alloc_pkts(struct hinic_rxq *rxq) rx_alloc_pkts() argument 246 free_all_rx_skbs(struct hinic_rxq *rxq) free_all_rx_skbs() argument 274 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) rx_recv_jumbo_pkt() argument 357 rxq_recv(struct hinic_rxq *rxq, int budget) rxq_recv() argument 462 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); rx_poll() local 481 rx_add_napi(struct hinic_rxq *rxq) rx_add_napi() argument 489 rx_del_napi(struct hinic_rxq *rxq) rx_del_napi() argument 497 struct hinic_rxq *rxq = (struct hinic_rxq *)data; rx_irq() local 515 rx_request_irq(struct hinic_rxq *rxq) rx_request_irq() argument 565 rx_free_irq(struct hinic_rxq *rxq) rx_free_irq() argument 582 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) hinic_init_rxq() argument 625 hinic_clean_rxq(struct hinic_rxq *rxq) hinic_clean_rxq() argument [all...] |
H A D | hinic_rx.h | 44 void hinic_rxq_clean_stats(struct hinic_rxq *rxq); 46 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats); 48 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, 51 void hinic_clean_rxq(struct hinic_rxq *rxq);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_rx.c | 51 * @rxq: Logical Rx Queue 53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument 55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats() 68 * @rxq: Logical Rx Queue 71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument 73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats() 89 * @rxq: Logical Rx Queue 91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument 93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init() 96 hinic_rxq_clean_stats(rxq); in rxq_stats_init() 99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) rx_csum() argument 127 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) rx_alloc_skb() argument 163 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) rx_unmap_skb() argument 180 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) rx_free_skb() argument 193 rx_alloc_pkts(struct hinic_rxq *rxq) rx_alloc_pkts() argument 243 free_all_rx_skbs(struct hinic_rxq *rxq) free_all_rx_skbs() argument 271 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) rx_recv_jumbo_pkt() argument 354 rxq_recv(struct hinic_rxq *rxq, int budget) rxq_recv() argument 459 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); rx_poll() local 478 rx_add_napi(struct hinic_rxq *rxq) rx_add_napi() argument 487 rx_del_napi(struct hinic_rxq *rxq) rx_del_napi() argument 495 struct hinic_rxq *rxq = (struct hinic_rxq *)data; rx_irq() local 513 rx_request_irq(struct hinic_rxq *rxq) rx_request_irq() argument 563 rx_free_irq(struct hinic_rxq *rxq) rx_free_irq() argument 580 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) hinic_init_rxq() argument 623 hinic_clean_rxq(struct hinic_rxq *rxq) hinic_clean_rxq() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | rx.c | 55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free. 58 * or reused - added immediately to the iwl->rxq->rx_free list. 66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them 76 * detached from the iwl->rxq. The driver 'processed' index is updated. 77 * + If there are no allocated buffers in iwl->rxq->rx_free, 104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue 108 * rxq 119 iwl_rxq_space(const struct iwl_rxq *rxq) iwl_rxq_space() argument 166 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_inc_wr_ptr() argument 209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local 220 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) iwl_pcie_restock_bd() argument 244 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxmq_restock() argument 291 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxsq_restock() argument 349 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_restock() argument 429 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) iwl_pcie_rxq_alloc_rbs() argument 613 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_allocator_get() argument 678 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_free_rxq_dma() argument 713 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_alloc_rxq_dma() argument 757 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_alloc_rxq_dma() local 807 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_alloc() local 835 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_hw_init() argument 991 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) iwl_pcie_rx_init_rxb_lists() argument 1005 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); iwl_pcie_napi_poll() local 1032 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); iwl_pcie_napi_poll_msix() local 1071 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_napi_sync() local 1108 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; _iwl_pcie_rx_init() local 1229 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_free() local 1246 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) iwl_pcie_rx_move_to_allocator() argument 1260 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) iwl_pcie_rx_reuse_rbd() argument 1292 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) iwl_pcie_rx_handle_rb() argument 1442 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) iwl_pcie_get_rxb() argument 1500 struct iwl_rxq *rxq; iwl_pcie_rx_handle() local 1643 struct iwl_rxq *rxq; iwl_pcie_irq_rx_msix_handler() local [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | rx.c | 112 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free. 115 * or reused - added immediately to the iwl->rxq->rx_free list. 123 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them 133 * detached from the iwl->rxq. The driver 'processed' index is updated. 134 * + If there are no allocated buffers in iwl->rxq->rx_free, 161 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue 165 * rxq 176 iwl_rxq_space(const struct iwl_rxq *rxq) iwl_rxq_space() argument 223 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_inc_wr_ptr() argument 263 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local 274 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) iwl_pcie_restock_bd() argument 298 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxmq_restock() argument 345 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxsq_restock() argument 403 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rxq_restock() argument 483 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) iwl_pcie_rxq_alloc_rbs() argument 667 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_allocator_get() argument 723 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_free_rxq_dma() argument 765 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_alloc_rxq_dma() argument 826 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_alloc_rxq_dma() local 878 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_alloc() local 903 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) iwl_pcie_rx_hw_init() argument 1061 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) iwl_pcie_rx_init_rxb_lists() argument 1107 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; _iwl_pcie_rx_init() local 1222 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rx_free() local 1237 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) iwl_pcie_rx_move_to_allocator() argument 1251 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) iwl_pcie_rx_reuse_rbd() argument 1283 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) iwl_pcie_rx_handle_rb() argument 1432 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) iwl_pcie_get_rxb() argument 1481 struct iwl_rxq *rxq; iwl_pcie_rx_handle() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | netdev_rx.c | 194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init() 197 if (!rx->rxq) { in hfi1_netdev_rxq_init() 203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local 205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); in hfi1_netdev_rxq_init() 209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init() 210 rxq->rx = rx; in hfi1_netdev_rxq_init() 211 rxq->rcd->napi = &rxq in hfi1_netdev_rxq_init() 230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; hfi1_netdev_rxq_init() local 250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; hfi1_netdev_rxq_deinit() local 268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; enable_queues() local 286 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; disable_queues() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | netdev_rx.c | 194 priv->rxq = kcalloc_node(priv->num_rx_q, sizeof(struct hfi1_netdev_rxq), in hfi1_netdev_rxq_init() 197 if (!priv->rxq) { in hfi1_netdev_rxq_init() 203 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; in hfi1_netdev_rxq_init() local 205 rc = hfi1_netdev_allot_ctxt(priv, &rxq->rcd); in hfi1_netdev_rxq_init() 209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init() 210 rxq->priv = priv; in hfi1_netdev_rxq_init() 211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init() 213 i, rxq in hfi1_netdev_rxq_init() 230 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; hfi1_netdev_rxq_init() local 251 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; hfi1_netdev_rxq_deinit() local 269 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; enable_queues() local 287 struct hfi1_netdev_rxq *rxq = &priv->rxq[i]; disable_queues() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 26 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument 37 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer() 38 rxq->filled_buffers--; in qede_alloc_rx_buffer() 49 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer() 50 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer() 51 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer() 56 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer() 62 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer() 66 rxq in qede_alloc_rx_buffer() 508 qede_has_rx_work(struct qede_rx_queue *rxq) qede_has_rx_work() argument 521 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) qede_rx_bd_ring_consume() argument 530 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_reuse_page() argument 553 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) qede_recycle_rx_bd_ring() argument 564 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_realloc_rx_buffer() argument 594 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_update_rx_prod() argument 643 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) qede_skb_receive() argument 670 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) qede_fill_frag_skb() argument 746 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_build_skb() argument 765 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) qede_tpa_rx_build_skb() argument 796 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_rx_build_skb() argument 837 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) qede_tpa_start() argument 963 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) qede_tpa_cont() argument 982 struct qede_rx_queue *rxq = fp->rxq; qede_tpa_end() local 1085 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) qede_rx_xdp() argument 1184 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) qede_rx_build_jumbo() argument 1241 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) qede_rx_process_tpa_cqe() argument 1261 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) qede_rx_process_cqe() argument 1359 struct qede_rx_queue *rxq = fp->rxq; qede_rx_int() local [all...] |
H A D | qede_main.c | 902 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array() 903 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); in qede_free_fp_array() 904 kfree(fp->rxq); in qede_free_fp_array() 963 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); in qede_alloc_fp_array() 964 if (!fp->rxq) in qede_alloc_fp_array() 1460 struct qede_rx_queue *rxq) in qede_free_rx_buffers() 1464 for (i = rxq->sw_rx_cons; i != rxq in qede_free_rx_buffers() 1459 qede_free_rx_buffers(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_free_rx_buffers() argument 1479 qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_free_mem_rxq() argument 1492 qede_set_tpa_param(struct qede_rx_queue *rxq) qede_set_tpa_param() argument 1504 qede_alloc_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_alloc_mem_rxq() argument 2215 struct qede_rx_queue *rxq = fp->rxq; qede_start_queues() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 27 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument 38 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer() 39 rxq->filled_buffers--; in qede_alloc_rx_buffer() 50 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer() 51 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer() 52 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer() 57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer() 63 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer() 67 rxq in qede_alloc_rx_buffer() 502 qede_has_rx_work(struct qede_rx_queue *rxq) qede_has_rx_work() argument 515 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) qede_rx_bd_ring_consume() argument 524 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_reuse_page() argument 547 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) qede_recycle_rx_bd_ring() argument 558 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) qede_realloc_rx_buffer() argument 588 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) qede_update_rx_prod() argument 637 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) qede_skb_receive() argument 664 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) qede_fill_frag_skb() argument 740 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_build_skb() argument 759 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) qede_tpa_rx_build_skb() argument 790 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) qede_rx_build_skb() argument 831 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) qede_tpa_start() argument 957 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) qede_tpa_cont() argument 976 struct qede_rx_queue *rxq = fp->rxq; qede_tpa_end() local 1079 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) qede_rx_xdp() argument 1169 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) qede_rx_build_jumbo() argument 1223 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) qede_rx_process_tpa_cqe() argument 1243 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) qede_rx_process_cqe() argument 1341 struct qede_rx_queue *rxq = fp->rxq; qede_rx_int() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_en.c | 548 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) in mana_get_rxbuf_pre() argument 550 struct net_device *ndev = rxq->ndev; in mana_get_rxbuf_pre() 562 if (mpc->rxbpre_datasize != rxq->datasize) { in mana_get_rxbuf_pre() 564 mpc->rxbpre_datasize, rxq->datasize); in mana_get_rxbuf_pre() 568 if (mpc->rxbpre_alloc_size != rxq->alloc_size) { in mana_get_rxbuf_pre() 570 mpc->rxbpre_alloc_size, rxq->alloc_size); in mana_get_rxbuf_pre() 574 if (mpc->rxbpre_headroom != rxq->headroom) { in mana_get_rxbuf_pre() 576 mpc->rxbpre_headroom, rxq->headroom); in mana_get_rxbuf_pre() 1258 static int mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) in mana_fence_rq() argument 1264 init_completion(&rxq in mana_fence_rq() 1300 struct mana_rxq *rxq; mana_fence_rqs() local 1461 mana_post_pkt_rxq(struct mana_rxq *rxq) mana_post_pkt_rxq() argument 1481 mana_build_skb(struct mana_rxq *rxq, void *buf_va, uint pkt_len, struct xdp_buff *xdp) mana_build_skb() argument 1501 mana_rx_skb(void *buf_va, bool from_pool, struct mana_rxcomp_oob *cqe, struct mana_rxq *rxq) mana_rx_skb() argument 1602 mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev, dma_addr_t *da, bool *from_pool, bool is_napi) mana_get_rxfrag() argument 1653 mana_refill_rx_oob(struct device *dev, struct mana_rxq *rxq, struct mana_recv_buf_oob *rxoob, void **old_buf, bool *old_fp) mana_refill_rx_oob() argument 1675 mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq, struct gdma_comp *cqe) mana_process_rx_cqe() argument 1745 struct mana_rxq *rxq = cq->rxq; mana_poll_rx_cq() local 1986 mana_destroy_rxq(struct mana_port_context *apc, struct mana_rxq *rxq, bool validate_state) mana_destroy_rxq() argument 2045 mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key, struct mana_rxq *rxq, struct device *dev) mana_fill_rx_oob() argument 2074 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) mana_alloc_rx_wqe() argument 2114 mana_push_wqe(struct mana_rxq *rxq) mana_push_wqe() argument 2132 mana_create_page_pool(struct mana_rxq *rxq, struct gdma_context *gc) mana_create_page_pool() argument 2163 struct mana_rxq *rxq; mana_create_rxq() local 2283 struct mana_rxq *rxq; mana_add_rx_queues() local 2307 struct mana_rxq *rxq; mana_destroy_vport() local [all...] |
/kernel/linux/linux-5.10/drivers/atm/ |
H A D | ambassador.c | 673 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local 678 spin_lock_irqsave (&rxq->lock, flags); in rx_give() 680 if (rxq->pending < rxq->maximum) { in rx_give() 681 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give() 683 *rxq->in.ptr = *rx; in rx_give() 684 rxq->pending++; in rx_give() 685 rxq->in.ptr = NEXTQ (rxq in rx_give() 698 amb_rxq * rxq = &dev->rxq[pool]; rx_take() local 731 amb_rxq * rxq = &dev->rxq[pool]; drain_rx_pool() local 769 amb_rxq * rxq; fill_rx_pool() local 1575 amb_rxq * rxq = &dev->rxq[pool]; create_queues() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3)) 134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) 857 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() 864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 876 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() 880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get() 888 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() 856 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) mvneta_rxq_non_occup_desc_add() argument 875 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_busy_desc_num_get() argument 887 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) mvneta_rxq_desc_num_update() argument 922 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) mvneta_rxq_next_desc_get() argument 945 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) mvneta_rxq_offset_set() argument 1002 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) mvneta_rxq_buf_size_set() argument 1017 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_disable() argument 1028 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_enable() argument 1039 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_long_pool_set() argument 1052 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_short_pool_set() argument 1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_port_up() local 1498 int rxq, txq; mvneta_defaults_set() local 1685 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_pkts_coal_set() argument 1695 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_time_coal_set() argument 1722 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) mvneta_rx_desc_fill() argument 1935 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) mvneta_rx_refill() argument 1986 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_drop_pkts() argument 2041 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rx_refill_queue() argument 2072 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len) mvneta_xdp_put_buff() argument 2257 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) mvneta_run_xdp() argument 2314 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_rx_frame() argument 2349 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_add_rx_fragment() argument 2425 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) mvneta_rx_swbm() argument 2538 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) mvneta_rx_hwbm() argument 3346 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) mvneta_create_page_pool() argument 3390 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) mvneta_rxq_fill() argument 3439 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_sw_init() argument 3456 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_hw_init() argument 3489 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_init() argument 3505 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_deinit() argument 4343 int rxq; mvneta_percpu_elect() local 4683 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_ethtool_set_coalesce() local 5144 mvneta_map_vlan_prio_to_rxq(struct mvneta_port *pp, u8 pri, u8 rxq) mvneta_map_vlan_prio_to_rxq() argument 5221 int rxq, txq, tc, ret; mvneta_setup_mqprio() local 5368 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_init() local 5789 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_suspend() local 5841 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_resume() local [all...] |
H A D | mv643xx_eth.c | 399 struct rx_queue rxq[8]; member 440 /* rxq/txq helper functions *************************************************/ 441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument 443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 451 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument 453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() 454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable() 457 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument 506 rxq_process(struct rx_queue *rxq, int budget) rxq_process() argument 598 rxq_refill(struct rx_queue *rxq, int budget) rxq_refill() argument 1940 struct rx_queue *rxq = mp->rxq + index; rxq_init() local 2005 rxq_deinit(struct rx_queue *rxq) rxq_deinit() argument 2383 struct rx_queue *rxq = mp->rxq + i; port_start() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 129 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) 817 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() 824 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 830 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 836 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() 840 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get() 848 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() 856 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update() 876 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq in mvneta_rxq_desc_num_update() 816 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) mvneta_rxq_non_occup_desc_add() argument 835 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_busy_desc_num_get() argument 847 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) mvneta_rxq_desc_num_update() argument 882 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) mvneta_rxq_next_desc_get() argument 905 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) mvneta_rxq_offset_set() argument 962 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) mvneta_rxq_buf_size_set() argument 977 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_disable() argument 988 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_bm_enable() argument 999 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_long_pool_set() argument 1012 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_short_pool_set() argument 1229 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_port_up() local 1458 int rxq, txq; mvneta_defaults_set() local 1645 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_pkts_coal_set() argument 1655 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) mvneta_rx_time_coal_set() argument 1682 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) mvneta_rx_desc_fill() argument 1890 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) mvneta_rx_refill() argument 1941 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_drop_pkts() argument 1996 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rx_refill_queue() argument 2027 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len, bool napi) mvneta_xdp_put_buff() argument 2170 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) mvneta_run_xdp() argument 2227 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_rx_frame() argument 2268 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) mvneta_swbm_add_rx_fragment() argument 2306 mvneta_swbm_build_skb(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, u32 desc_status) mvneta_swbm_build_skb() argument 2336 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) mvneta_rx_swbm() argument 2451 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) mvneta_rx_hwbm() argument 3200 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) mvneta_create_page_pool() argument 3243 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) mvneta_rxq_fill() argument 3292 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_sw_init() argument 3309 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_hw_init() argument 3342 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_init() argument 3358 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) mvneta_rxq_deinit() argument 4150 int rxq; mvneta_percpu_elect() local 4490 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_ethtool_set_coalesce() local 4976 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_init() local 5357 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_suspend() local 5409 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; mvneta_resume() local [all...] |
H A D | mv643xx_eth.c | 398 struct rx_queue rxq[8]; member 439 /* rxq/txq helper functions *************************************************/ 440 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument 442 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 450 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument 452 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() 453 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable() 456 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument 505 rxq_process(struct rx_queue *rxq, int budget) rxq_process() argument 597 rxq_refill(struct rx_queue *rxq, int budget) rxq_refill() argument 1929 struct rx_queue *rxq = mp->rxq + index; rxq_init() local 1994 rxq_deinit(struct rx_queue *rxq) rxq_deinit() argument 2372 struct rx_queue *rxq = mp->rxq + i; port_start() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 74 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local 80 next = cur = rxq->write_idx; in alx_refill_rx_ring() 83 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 85 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring() 86 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring() 128 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 135 rxq->write_idx = cur; in alx_refill_rx_ring() 217 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument 226 alx = netdev_priv(rxq in alx_clean_rx_irq() 507 alx_free_rxring_buf(struct alx_rx_queue *rxq) alx_free_rxring_buf() argument 631 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) alx_alloc_rx_ring() argument 742 struct alx_rx_queue *rxq; alx_alloc_napis() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local 79 next = cur = rxq->write_idx; in alx_refill_rx_ring() 82 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 84 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring() 85 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring() 127 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 134 rxq->write_idx = cur; in alx_refill_rx_ring() 216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument 225 alx = netdev_priv(rxq in alx_clean_rx_irq() 506 alx_free_rxring_buf(struct alx_rx_queue *rxq) alx_free_rxring_buf() argument 630 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) alx_alloc_rx_ring() argument 741 struct alx_rx_queue *rxq; alx_alloc_napis() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/ |
H A D | hisi_femac.c | 121 struct hisi_femac_queue rxq; member 212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local 218 pos = rxq->head; in hisi_femac_rx_refill() 220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill() 222 if (unlikely(rxq->skb[pos])) { in hisi_femac_rx_refill() 224 pos, rxq->skb[pos]); in hisi_femac_rx_refill() 237 rxq->dma_phys[pos] = addr; in hisi_femac_rx_refill() 238 rxq in hisi_femac_rx_refill() 248 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_rx() local 388 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_free_skb_rings() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/ |
H A D | hisi_femac.c | 121 struct hisi_femac_queue rxq; member 212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local 218 pos = rxq->head; in hisi_femac_rx_refill() 220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill() 222 if (unlikely(rxq->skb[pos])) { in hisi_femac_rx_refill() 224 pos, rxq->skb[pos]); in hisi_femac_rx_refill() 237 rxq->dma_phys[pos] = addr; in hisi_femac_rx_refill() 238 rxq in hisi_femac_rx_refill() 248 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_rx() local 388 struct hisi_femac_queue *rxq = &priv->rxq; hisi_femac_free_skb_rings() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/ |
H A D | xdp_hw_metadata.c | 53 int rxq; variable 231 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) in verify_metadata() argument 234 struct pollfd fds[rxq + 1]; in verify_metadata() 241 for (i = 0; i < rxq; i++) { in verify_metadata() 247 fds[rxq].fd = server_fd; in verify_metadata() 248 fds[rxq].events = POLLIN; in verify_metadata() 249 fds[rxq].revents = 0; in verify_metadata() 253 ret = poll(fds, rxq + 1, 1000); in verify_metadata() 262 if (fds[rxq].revents) in verify_metadata() 265 for (i = 0; i < rxq; in verify_metadata() [all...] |
/kernel/linux/linux-5.10/samples/bpf/ |
H A D | xdp_rxq_info_user.c | 194 struct record *rxq; member 234 rec->rxq = alloc_record_per_rxq(); in alloc_stats_record() 236 rec->rxq[i].cpu = alloc_record_per_cpu(); in alloc_stats_record() 248 free(r->rxq[i].cpu); in free_stats_record() 250 free(r->rxq); in free_stats_record() 294 map_collect_percpu(fd, i, &rec->rxq[i]); in stats_collect() 343 int rxq; in stats_print() local 383 for (rxq = 0; rxq < nr_rxqs; rxq in stats_print() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_trace.h | 75 TP_PROTO(const struct funeth_rxq *rxq, 81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec), 90 __string(devname, rxq->netdev->name) 94 __entry->qidx = rxq->qidx; 95 __entry->cq_head = rxq->cq_head; 100 __assign_str(devname, rxq->netdev->name);
|