/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 59 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, in mlx5e_xmit_xdp_buff() argument 89 dma_addr = dma_map_single(sq->pdev, xdptxd.data, xdptxd.len, in mlx5e_xmit_xdp_buff() 91 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff() 109 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd.len, in mlx5e_xmit_xdp_buff() 117 return INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe, in mlx5e_xmit_xdp_buff() 118 mlx5e_xmit_xdp_frame, sq, &xdptxd, &xdpi, 0); in mlx5e_xmit_xdp_buff() 166 static u16 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) in mlx5e_xdpsq_get_next_pi() argument 168 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdpsq_get_next_pi() 171 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi() 176 wi = &sq in mlx5e_xdpsq_get_next_pi() 195 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_session_start() argument 217 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_complete() argument 245 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check_mpwqe() argument 265 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdp_info *xdpi, int check_result) mlx5e_xmit_xdp_frame_mpwqe() argument 299 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check() argument 312 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdp_info *xdpi, int check_result) mlx5e_xmit_xdp_frame() argument 366 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, bool recycle) mlx5e_free_xdpsq_desc() argument 400 struct mlx5e_xdpsq *sq; mlx5e_poll_xdpsq_cq() local 464 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) mlx5e_free_xdpsq_descs() argument 488 struct mlx5e_xdpsq *sq; mlx5e_xdp_xmit() local 562 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) mlx5e_set_xmit_fp() argument [all...] |
H A D | reporter_tx.c | 6 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) in mlx5e_wait_for_sq_flush() argument 12 if (sq->cc == sq->pc) in mlx5e_wait_for_sq_flush() 18 netdev_err(sq->channel->netdev, in mlx5e_wait_for_sq_flush() 19 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n", in mlx5e_wait_for_sq_flush() 20 sq->sqn, sq->cc, sq->pc); in mlx5e_wait_for_sq_flush() 25 static void mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) in mlx5e_reset_txqsq_cc_pc() argument 39 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_err_cqe_recover() local 87 struct mlx5e_txqsq *sq; global() member 96 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_timeout_recover() local 144 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output() argument 261 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_diagnose() local 281 struct mlx5e_txqsq *sq = ctx; mlx5e_tx_reporter_dump_sq() local 377 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_dump_all_sqs() local 405 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_err_cqe() argument 419 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_timeout() argument [all...] |
H A D | xdp.h | 52 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq); 54 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq); 55 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw); 60 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, 64 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, 68 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq)); 69 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq)); 99 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) in mlx5e_xmit_xdp_doorbell() argument 101 if (sq->doorbell_cseg) { in mlx5e_xmit_xdp_doorbell() 102 mlx5e_notify_hw(&sq in mlx5e_xmit_xdp_doorbell() 110 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) mlx5e_xdp_get_inline_state() argument 140 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) mlx5e_xdp_mpwqe_add_dseg() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_tx.c | 46 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) in mlx5e_dma_unmap_wqe_err() argument 52 mlx5e_dma_get(sq, --sq->dma_fifo_pc); in mlx5e_dma_unmap_wqe_err() 54 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma); in mlx5e_dma_unmap_wqe_err() 119 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_eseg_csum() argument 123 if (unlikely(mlx5e_ipsec_txwqe_build_eseg_csum(sq, skb, eseg))) in mlx5e_txwqe_build_eseg_csum() 131 sq->stats->csum_partial_inner++; in mlx5e_txwqe_build_eseg_csum() 134 sq->stats->csum_partial++; in mlx5e_txwqe_build_eseg_csum() 139 sq->stats->csum_partial++; in mlx5e_txwqe_build_eseg_csum() 142 sq in mlx5e_txwqe_build_eseg_csum() 149 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb, int *hopbyhop) mlx5e_tx_get_gso_ihs() argument 177 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) mlx5e_txwqe_build_dsegs() argument 243 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) mlx5e_tx_wqe_inline_mode() argument 262 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) mlx5e_sq_xmit_prepare() argument 342 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) mlx5e_tx_check_stop() argument 350 mlx5e_tx_flush(struct mlx5e_txqsq *sq) mlx5e_tx_flush() argument 371 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_txwqe_complete() argument 421 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) mlx5e_sq_xmit_wqe() argument 512 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_same_eseg() argument 520 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_session_start() argument 544 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_is_active() argument 549 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) mlx5e_tx_mpwqe_add_dseg() argument 567 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_complete() argument 599 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_sq_xmit_mpwqe() argument 648 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_ensure_complete() argument 663 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) mlx5e_txwqe_build_eseg() argument 680 struct mlx5e_txqsq *sq; mlx5e_xmit() local 733 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_tx_wi_dma_unmap() argument 745 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_consume_skb() argument 763 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_tx_wi_consume_fifo_skbs() argument 775 mlx5e_txqsq_wake(struct mlx5e_txqsq *sq) mlx5e_txqsq_wake() argument 789 struct mlx5e_txqsq *sq; mlx5e_poll_tx_cq() local 890 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) mlx5e_tx_wi_kfree_fifo_skbs() argument 898 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) mlx5e_free_txqsq_descs() argument 973 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) mlx5i_sq_xmit() argument [all...] |
H A D | en_txrx.c | 49 static void mlx5e_handle_tx_dim(struct mlx5e_txqsq *sq) in mlx5e_handle_tx_dim() argument 51 struct mlx5e_sq_stats *stats = sq->stats; in mlx5e_handle_tx_dim() 54 if (unlikely(!test_bit(MLX5E_SQ_STATE_DIM, &sq->state))) in mlx5e_handle_tx_dim() 57 dim_update_sample(sq->cq.event_ctr, stats->packets, stats->bytes, &dim_sample); in mlx5e_handle_tx_dim() 58 net_dim(&sq->dim, dim_sample); in mlx5e_handle_tx_dim() 73 void mlx5e_trigger_irq(struct mlx5e_icosq *sq) in mlx5e_trigger_irq() argument 75 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_trigger_irq() 77 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_trigger_irq() 79 sq->db.wqe_info[pi] = (struct mlx5e_icosq_wqe_info) { in mlx5e_trigger_irq() 84 nopwqe = mlx5e_post_nop(wq, sq in mlx5e_trigger_irq() 157 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); mlx5e_napi_poll() local 225 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); mlx5e_napi_poll() local [all...] |
/kernel/linux/linux-6.6/drivers/nvme/target/ |
H A D | fabrics-cmd-auth.c | 17 struct nvmet_sq *sq = container_of(to_delayed_work(work), in nvmet_auth_expired_work() local 21 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); in nvmet_auth_expired_work() 22 sq->dhchap_step = NVME_AUTH_DHCHAP_MESSAGE_NEGOTIATE; in nvmet_auth_expired_work() 23 sq->dhchap_tid = -1; in nvmet_auth_expired_work() 26 void nvmet_auth_sq_init(struct nvmet_sq *sq) in nvmet_auth_sq_init() argument 29 INIT_DELAYED_WORK(&sq->auth_expired_work, nvmet_auth_expired_work); in nvmet_auth_sq_init() 30 sq->authenticated = false; in nvmet_auth_sq_init() 31 sq in nvmet_auth_sq_init() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 61 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, in mlx5e_xmit_xdp_buff() argument 96 dma_addr = dma_map_single(sq->pdev, xdptxd->data, xdptxd->len, in mlx5e_xmit_xdp_buff() 98 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff() 105 if (unlikely(!INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe, in mlx5e_xmit_xdp_buff() 106 mlx5e_xmit_xdp_frame, sq, xdptxd, 0))) in mlx5e_xmit_xdp_buff() 110 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff() 112 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff() 114 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff() 126 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd->len, DMA_BIDIRECTIONAL); in mlx5e_xmit_xdp_buff() 140 dma_sync_single_for_device(sq in mlx5e_xmit_xdp_buff() 303 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) mlx5e_xdpsq_get_next_pi() argument 332 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_session_start() argument 354 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_complete() argument 382 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check_mpwqe() argument 404 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result) mlx5e_xmit_xdp_frame_mpwqe() argument 464 mlx5e_xmit_xdp_frame_check_stop_room(struct mlx5e_xdpsq *sq, int stop_room) mlx5e_xmit_xdp_frame_check_stop_room() argument 476 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check() argument 482 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result) mlx5e_xmit_xdp_frame() argument 610 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq) mlx5e_free_xdpsq_desc() argument 686 struct mlx5e_xdpsq *sq; mlx5e_poll_xdpsq_cq() local 754 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) mlx5e_free_xdpsq_descs() argument 786 struct mlx5e_xdpsq *sq; mlx5e_xdp_xmit() local 903 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) mlx5e_set_xmit_fp() argument [all...] |
H A D | reporter_tx.c | 22 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) in mlx5e_wait_for_sq_flush() argument 24 struct mlx5_core_dev *dev = sq->mdev; in mlx5e_wait_for_sq_flush() 30 if (sq->cc == sq->pc) in mlx5e_wait_for_sq_flush() 36 netdev_err(sq->netdev, in mlx5e_wait_for_sq_flush() 37 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n", in mlx5e_wait_for_sq_flush() 38 sq->sqn, sq->cc, sq in mlx5e_wait_for_sq_flush() 43 mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) mlx5e_reset_txqsq_cc_pc() argument 53 mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq) mlx5e_health_sq_put_sw_state() argument 78 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_err_cqe_recover() local 130 struct mlx5e_txqsq *sq; global() member 139 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_timeout_recover() local 224 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output_sq_common() argument 276 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output() argument 458 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_diagnose() local 491 struct mlx5e_txqsq *sq = ctx; mlx5e_tx_reporter_dump_sq() local 597 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_dump_all_sqs() local 607 struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq; mlx5e_tx_reporter_dump_all_sqs() local 636 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_err_cqe() argument 650 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_timeout() argument [all...] |
H A D | xdp.h | 106 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq); 108 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq); 109 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw); 116 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, 119 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, 122 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq)); 123 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq)); 153 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) in mlx5e_xmit_xdp_doorbell() argument 155 if (sq->doorbell_cseg) { in mlx5e_xmit_xdp_doorbell() 156 mlx5e_notify_hw(&sq in mlx5e_xmit_xdp_doorbell() 164 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) mlx5e_xdp_get_inline_state() argument 195 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) mlx5e_xdp_mpwqe_add_dseg() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_tx.c | 43 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) in mlx5e_dma_unmap_wqe_err() argument 49 mlx5e_dma_get(sq, --sq->dma_fifo_pc); in mlx5e_dma_unmap_wqe_err() 51 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma); in mlx5e_dma_unmap_wqe_err() 151 ipsec_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in ipsec_txwqe_build_eseg_csum() argument 157 sq->stats->csum_partial_inner++; in ipsec_txwqe_build_eseg_csum() 159 sq->stats->csum_partial++; in ipsec_txwqe_build_eseg_csum() 164 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_eseg_csum() argument 173 sq->stats->csum_partial_inner++; in mlx5e_txwqe_build_eseg_csum() 176 sq in mlx5e_txwqe_build_eseg_csum() 191 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb) mlx5e_tx_get_gso_ihs() argument 213 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) mlx5e_txwqe_build_dsegs() argument 278 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) mlx5e_tx_wqe_inline_mode() argument 297 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) mlx5e_sq_xmit_prepare() argument 370 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) mlx5e_tx_check_stop() argument 379 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, bool xmit_more) mlx5e_txwqe_complete() argument 411 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) mlx5e_sq_xmit_wqe() argument 472 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_same_eseg() argument 480 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_session_start() argument 504 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_is_active() argument 509 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) mlx5e_tx_mpwqe_add_dseg() argument 527 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_complete() argument 559 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_sq_xmit_mpwqe() argument 609 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_ensure_complete() argument 616 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) mlx5e_txwqe_build_eseg() argument 635 struct mlx5e_txqsq *sq; mlx5e_xmit() local 676 mlx5e_sq_xmit_simple(struct mlx5e_txqsq *sq, struct sk_buff *skb, bool xmit_more) mlx5e_sq_xmit_simple() argument 691 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_tx_wi_dma_unmap() argument 703 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_consume_skb() argument 717 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_tx_wi_consume_fifo_skbs() argument 732 struct mlx5e_txqsq *sq; mlx5e_poll_tx_cq() local 839 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) mlx5e_tx_wi_kfree_fifo_skbs() argument 847 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) mlx5e_free_txqsq_descs() argument 922 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) mlx5i_sq_xmit() argument [all...] |
/kernel/linux/linux-5.10/tools/io_uring/ |
H A D | queue.c | 70 struct io_uring_sq *sq = &ring->sq; in io_uring_submit() local 71 const unsigned mask = *sq->kring_mask; in io_uring_submit() 81 if (*sq->khead != *sq->ktail) { in io_uring_submit() 82 submitted = *sq->kring_entries; in io_uring_submit() 86 if (sq->sqe_head == sq->sqe_tail) in io_uring_submit() 93 ktail = ktail_next = *sq->ktail; in io_uring_submit() 94 to_submit = sq in io_uring_submit() 143 struct io_uring_sq *sq = &ring->sq; io_uring_get_sqe() local [all...] |
H A D | setup.c | 11 struct io_uring_sq *sq, struct io_uring_cq *cq) in io_uring_mmap() 17 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap() 18 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE, in io_uring_mmap() 22 sq->khead = ptr + p->sq_off.head; in io_uring_mmap() 23 sq->ktail = ptr + p->sq_off.tail; in io_uring_mmap() 24 sq->kring_mask = ptr + p->sq_off.ring_mask; in io_uring_mmap() 25 sq->kring_entries = ptr + p->sq_off.ring_entries; in io_uring_mmap() 26 sq->kflags = ptr + p->sq_off.flags; in io_uring_mmap() 27 sq->kdropped = ptr + p->sq_off.dropped; in io_uring_mmap() 28 sq in io_uring_mmap() 10 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) io_uring_mmap() argument 100 struct io_uring_sq *sq = &ring->sq; io_uring_queue_exit() local [all...] |
/kernel/linux/linux-5.10/sound/oss/dmasound/ |
H A D | dmasound_core.c | 411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) in sq_allocate_buffers() argument 415 if (sq->buffers) in sq_allocate_buffers() 417 sq->numBufs = num; in sq_allocate_buffers() 418 sq->bufSize = size; in sq_allocate_buffers() 419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL); in sq_allocate_buffers() 420 if (!sq->buffers) in sq_allocate_buffers() 423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); in sq_allocate_buffers() 424 if (!sq->buffers[i]) { in sq_allocate_buffers() 426 dmasound.mach.dma_free(sq->buffers[i], size); in sq_allocate_buffers() 427 kfree(sq in sq_allocate_buffers() 435 sq_release_buffers(struct sound_queue *sq) sq_release_buffers() argument 448 sq_setup(struct sound_queue *sq) sq_setup() argument 688 sq_init_waitqueue(struct sound_queue *sq) sq_init_waitqueue() argument 707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) sq_open2() argument 953 set_queue_frags(struct sound_queue *sq, int bufs, int size) set_queue_frags() argument [all...] |
/kernel/linux/linux-6.6/sound/oss/dmasound/ |
H A D | dmasound_core.c | 411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) in sq_allocate_buffers() argument 415 if (sq->buffers) in sq_allocate_buffers() 417 sq->numBufs = num; in sq_allocate_buffers() 418 sq->bufSize = size; in sq_allocate_buffers() 419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL); in sq_allocate_buffers() 420 if (!sq->buffers) in sq_allocate_buffers() 423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); in sq_allocate_buffers() 424 if (!sq->buffers[i]) { in sq_allocate_buffers() 426 dmasound.mach.dma_free(sq->buffers[i], size); in sq_allocate_buffers() 427 kfree(sq in sq_allocate_buffers() 435 sq_release_buffers(struct sound_queue *sq) sq_release_buffers() argument 448 sq_setup(struct sound_queue *sq) sq_setup() argument 688 sq_init_waitqueue(struct sound_queue *sq) sq_init_waitqueue() argument 707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) sq_open2() argument 953 set_queue_frags(struct sound_queue *sq, int bufs, int size) set_queue_frags() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_tx.c | 16 mlx5e_ktls_dumps_num_wqes(struct mlx5e_txqsq *sq, unsigned int nfrags, in mlx5e_ktls_dumps_num_wqes() argument 22 return nfrags + DIV_ROUND_UP(sync_len, sq->hw_mtu); in mlx5e_ktls_dumps_num_wqes() 25 u16 mlx5e_ktls_get_stop_room(struct mlx5e_txqsq *sq) in mlx5e_ktls_get_stop_room() argument 29 num_dumps = mlx5e_ktls_dumps_num_wqes(sq, MAX_SKB_FRAGS, TLS_MAX_PAYLOAD_SIZE); in mlx5e_ktls_get_stop_room() 138 static void tx_fill_wi(struct mlx5e_txqsq *sq, in tx_fill_wi() argument 142 struct mlx5e_tx_wqe_info *wi = &sq->db.wqe_info[pi]; in tx_fill_wi() 162 post_static_params(struct mlx5e_txqsq *sq, in post_static_params() argument 170 pi = mlx5e_txqsq_get_next_pi(sq, num_wqebbs); in post_static_params() 171 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi); in post_static_params() 172 mlx5e_ktls_build_static_params(wqe, sq in post_static_params() 180 post_progress_params(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, bool fence) post_progress_params() argument 197 mlx5e_ktls_tx_post_param_wqes(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, bool skip_static_post, bool fence_first_post) mlx5e_ktls_tx_post_param_wqes() argument 278 tx_post_resync_params(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, u64 rcd_sn) tx_post_resync_params() argument 299 tx_post_resync_dump(struct mlx5e_txqsq *sq, skb_frag_t *frag, u32 tisn, bool first) tx_post_resync_dump() argument 340 mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_ktls_tx_handle_resync_dump_comp() argument 356 tx_post_fence_nop(struct mlx5e_txqsq *sq) tx_post_fence_nop() argument 367 mlx5e_ktls_tx_handle_ooo(struct mlx5e_ktls_offload_context_tx *priv_tx, struct mlx5e_txqsq *sq, int datalen, u32 seq) mlx5e_ktls_tx_handle_ooo() argument 442 mlx5e_ktls_handle_tx_skb(struct tls_context *tls_ctx, struct mlx5e_txqsq *sq, struct sk_buff *skb, int datalen, struct mlx5e_accel_tx_tls_state *state) mlx5e_ktls_handle_tx_skb() argument [all...] |
H A D | ktls_rx.c | 128 static void icosq_fill_wi(struct mlx5e_icosq *sq, u16 pi, in icosq_fill_wi() argument 131 sq->db.wqe_info[pi] = *wi; in icosq_fill_wi() 135 post_static_params(struct mlx5e_icosq *sq, in post_static_params() argument 144 if (unlikely(!mlx5e_wqc_has_room_for(&sq->wq, sq->cc, sq->pc, room))) in post_static_params() 147 pi = mlx5e_icosq_get_next_pi(sq, num_wqebbs); in post_static_params() 148 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi); in post_static_params() 149 mlx5e_ktls_build_static_params(wqe, sq->pc, sq in post_static_params() 165 post_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx, u32 next_record_tcp_sn) post_progress_params() argument 200 struct mlx5e_icosq *sq; post_rx_param_wqes() local 251 resync_post_get_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx) resync_post_get_progress_params() argument 336 struct mlx5e_icosq *sq; resync_handle_work() local 369 struct mlx5e_icosq *sq; resync_handle_seq_match() local 398 mlx5e_ktls_handle_get_psv_completion(struct mlx5e_icosq_wqe_info *wi, struct mlx5e_icosq *sq) mlx5e_ktls_handle_get_psv_completion() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | tx.c | 46 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, in mlx5e_xsk_tx_post_err() argument 49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); in mlx5e_xsk_tx_post_err() 50 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi]; in mlx5e_xsk_tx_post_err() 56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); in mlx5e_xsk_tx_post_err() 57 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, *xdpi); in mlx5e_xsk_tx_post_err() 58 sq->doorbell_cseg = &nopwqe->ctrl; in mlx5e_xsk_tx_post_err() 61 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigne argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | tx.c | 52 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, in mlx5e_xsk_tx_post_err() argument 55 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); in mlx5e_xsk_tx_post_err() 56 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi]; in mlx5e_xsk_tx_post_err() 62 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); in mlx5e_xsk_tx_post_err() 63 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, xdpi); in mlx5e_xsk_tx_post_err() 64 sq->doorbell_cseg = &nopwqe->ctrl; in mlx5e_xsk_tx_post_err() 67 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigne argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 38 struct otx2_snd_queue *sq; in otx2_qos_sq_aura_pool_init() local 76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init() 77 sq->sqb_count = 0; in otx2_qos_sq_aura_pool_init() 78 sq->sqb_ptrs = kcalloc(num_sqbs, sizeof(*sq->sqb_ptrs), GFP_KERNEL); in otx2_qos_sq_aura_pool_init() 79 if (!sq->sqb_ptrs) { in otx2_qos_sq_aura_pool_init() 89 sq->sqb_ptrs[sq->sqb_count++] = (u64)bufptr; in otx2_qos_sq_aura_pool_init() 96 if (!sq in otx2_qos_sq_aura_pool_init() 120 struct otx2_snd_queue *sq; otx2_qos_sq_free_sqbs() local 263 struct otx2_snd_queue *sq; otx2_qos_disable_sq() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 80 struct otx2_snd_queue *sq, in otx2_snd_pkt_handler() 96 sg = &sq->sg[snd_comp->sqe_id]; in otx2_snd_pkt_handler() 102 timestamp = ((u64 *)sq->timestamps->base)[snd_comp->sqe_id]; in otx2_snd_pkt_handler() 372 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[cq->cint_idx], in otx2_tx_napi_handler() 442 static void otx2_sqe_flush(struct otx2_snd_queue *sq, int size) in otx2_sqe_flush() argument 450 memcpy(sq->lmt_addr, sq->sqe_base, size); in otx2_sqe_flush() 451 status = otx2_lmt_flush(sq->io_addr); in otx2_sqe_flush() 454 sq->head++; in otx2_sqe_flush() 455 sq in otx2_sqe_flush() 78 otx2_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe, int budget, int *tx_pkts, int *tx_bytes) otx2_snd_pkt_handler() argument 460 otx2_sqe_add_sg(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int num_segs, int *offset) otx2_sqe_add_sg() argument 505 otx2_sqe_add_ext(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int *offset) otx2_sqe_add_ext() argument 560 otx2_sqe_add_mem(struct otx2_snd_queue *sq, int *offset, int alg, u64 iova) otx2_sqe_add_mem() argument 575 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) otx2_sqe_add_hdr() argument 622 otx2_dma_map_tso_skb(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int sqe, int hdr_len) otx2_dma_map_tso_skb() argument 655 otx2_tso_frag_dma_addr(struct otx2_snd_queue *sq, struct sk_buff *skb, int seg, u64 seg_addr, int hdr_len, int sqe) otx2_tso_frag_dma_addr() argument 673 otx2_sqe_tso_add_sg(struct otx2_snd_queue *sq, struct sg_list *list, int *offset) otx2_sqe_tso_add_sg() argument 704 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_tso() argument 825 otx2_set_txtstamp(struct otx2_nic *pfvf, struct sk_buff *skb, struct otx2_snd_queue *sq, int *offset) otx2_set_txtstamp() argument 840 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_skb() argument 934 struct otx2_snd_queue *sq; otx2_cleanup_tx_cqes() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 19 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, 504 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() 508 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE, in nicvf_init_snd_queue() 513 sq->desc = sq->dmem.base; in nicvf_init_snd_queue() 514 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); in nicvf_init_snd_queue() 515 if (!sq->skbuff) in nicvf_init_snd_queue() 518 sq->head = 0; in nicvf_init_snd_queue() 519 sq->tail = 0; in nicvf_init_snd_queue() 520 sq in nicvf_init_snd_queue() 503 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument 553 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) nicvf_unmap_sndq_buffers() argument 571 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) nicvf_free_snd_queue() argument 862 struct snd_queue *sq; nicvf_snd_queue_config() local 1118 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_get_sq_desc() argument 1134 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) nicvf_rollback_sq_desc() argument 1142 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_put_sq_desc() argument 1152 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) nicvf_get_nxt_sqentry() argument 1159 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument 1179 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument 1205 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) nicvf_xdp_sq_doorbell() argument 1221 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) nicvf_xdp_sq_add_hdr_subdesc() argument 1235 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) nicvf_xdp_sq_append_pkt() argument 1329 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) nicvf_sq_add_hdr_subdesc() argument 1418 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) nicvf_sq_add_gather_subdesc() argument 1437 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) nicvf_sq_add_cqe_subdesc() argument 1485 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) nicvf_sq_append_tso() argument 1549 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) nicvf_sq_append_skb() argument 1826 struct snd_queue *sq; nicvf_update_sq_stats() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 20 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, 505 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() 509 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE, in nicvf_init_snd_queue() 514 sq->desc = sq->dmem.base; in nicvf_init_snd_queue() 515 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); in nicvf_init_snd_queue() 516 if (!sq->skbuff) in nicvf_init_snd_queue() 519 sq->head = 0; in nicvf_init_snd_queue() 520 sq->tail = 0; in nicvf_init_snd_queue() 521 sq in nicvf_init_snd_queue() 504 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument 554 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) nicvf_unmap_sndq_buffers() argument 572 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) nicvf_free_snd_queue() argument 863 struct snd_queue *sq; nicvf_snd_queue_config() local 1119 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_get_sq_desc() argument 1135 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) nicvf_rollback_sq_desc() argument 1143 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_put_sq_desc() argument 1153 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) nicvf_get_nxt_sqentry() argument 1160 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument 1180 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument 1206 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) nicvf_xdp_sq_doorbell() argument 1222 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) nicvf_xdp_sq_add_hdr_subdesc() argument 1236 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) nicvf_xdp_sq_append_pkt() argument 1330 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) nicvf_sq_add_hdr_subdesc() argument 1419 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) nicvf_sq_add_gather_subdesc() argument 1438 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) nicvf_sq_add_cqe_subdesc() argument 1486 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) nicvf_sq_append_tso() argument 1550 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) nicvf_sq_append_skb() argument 1827 struct snd_queue *sq; nicvf_update_sq_stats() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/net/ |
H A D | io_uring_zerocopy_tx.c | 108 struct io_uring_sq sq; member 169 struct io_uring_sq *sq, struct io_uring_cq *cq) in io_uring_mmap() 175 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap() 176 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE, in io_uring_mmap() 180 sq->khead = ptr + p->sq_off.head; in io_uring_mmap() 181 sq->ktail = ptr + p->sq_off.tail; in io_uring_mmap() 182 sq->kring_mask = ptr + p->sq_off.ring_mask; in io_uring_mmap() 183 sq->kring_entries = ptr + p->sq_off.ring_entries; in io_uring_mmap() 184 sq->kflags = ptr + p->sq_off.flags; in io_uring_mmap() 185 sq in io_uring_mmap() 168 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) io_uring_mmap() argument 238 struct io_uring_sq *sq = &ring->sq; io_uring_submit() local 293 struct io_uring_sq *sq = &ring->sq; io_uring_get_sqe() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 8 (qinfo)->sq.head = prefix##_ATQH; \ 9 (qinfo)->sq.tail = prefix##_ATQT; \ 10 (qinfo)->sq.len = prefix##_ATQLEN; \ 11 (qinfo)->sq.bah = prefix##_ATQBAH; \ 12 (qinfo)->sq.bal = prefix##_ATQBAL; \ 13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \ 14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \ 15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \ 16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \ 64 if (cq->sq in ice_check_sq_alive() 807 struct ice_ctl_q_ring *sq = &cq->sq; ice_clean_sq() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 8 (qinfo)->sq.head = prefix##_ATQH; \ 9 (qinfo)->sq.tail = prefix##_ATQT; \ 10 (qinfo)->sq.len = prefix##_ATQLEN; \ 11 (qinfo)->sq.bah = prefix##_ATQBAH; \ 12 (qinfo)->sq.bal = prefix##_ATQBAL; \ 13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \ 14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \ 15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \ 16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \ 77 if (cq->sq in ice_check_sq_alive() 860 struct ice_ctl_q_ring *sq = &cq->sq; ice_clean_sq() local [all...] |