Home
last modified time | relevance | path

Searched refs:sq (Results 1 - 25 of 457) sorted by relevance

12345678910>>...19

/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dxdp.c59 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, in mlx5e_xmit_xdp_buff() argument
89 dma_addr = dma_map_single(sq->pdev, xdptxd.data, xdptxd.len, in mlx5e_xmit_xdp_buff()
91 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff()
109 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd.len, in mlx5e_xmit_xdp_buff()
117 return INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe, in mlx5e_xmit_xdp_buff()
118 mlx5e_xmit_xdp_frame, sq, &xdptxd, &xdpi, 0); in mlx5e_xmit_xdp_buff()
166 static u16 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) in mlx5e_xdpsq_get_next_pi() argument
168 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdpsq_get_next_pi()
171 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi()
176 wi = &sq in mlx5e_xdpsq_get_next_pi()
195 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_session_start() argument
217 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_complete() argument
245 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check_mpwqe() argument
265 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdp_info *xdpi, int check_result) mlx5e_xmit_xdp_frame_mpwqe() argument
299 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check() argument
312 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdp_info *xdpi, int check_result) mlx5e_xmit_xdp_frame() argument
366 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, bool recycle) mlx5e_free_xdpsq_desc() argument
400 struct mlx5e_xdpsq *sq; mlx5e_poll_xdpsq_cq() local
464 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) mlx5e_free_xdpsq_descs() argument
488 struct mlx5e_xdpsq *sq; mlx5e_xdp_xmit() local
562 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) mlx5e_set_xmit_fp() argument
[all...]
H A Dreporter_tx.c6 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) in mlx5e_wait_for_sq_flush() argument
12 if (sq->cc == sq->pc) in mlx5e_wait_for_sq_flush()
18 netdev_err(sq->channel->netdev, in mlx5e_wait_for_sq_flush()
19 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n", in mlx5e_wait_for_sq_flush()
20 sq->sqn, sq->cc, sq->pc); in mlx5e_wait_for_sq_flush()
25 static void mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) in mlx5e_reset_txqsq_cc_pc() argument
39 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_err_cqe_recover() local
87 struct mlx5e_txqsq *sq; global() member
96 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_timeout_recover() local
144 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output() argument
261 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_diagnose() local
281 struct mlx5e_txqsq *sq = ctx; mlx5e_tx_reporter_dump_sq() local
377 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_dump_all_sqs() local
405 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_err_cqe() argument
419 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_timeout() argument
[all...]
H A Dxdp.h52 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq);
54 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq);
55 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw);
60 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq,
64 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq,
68 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq));
69 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq));
99 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) in mlx5e_xmit_xdp_doorbell() argument
101 if (sq->doorbell_cseg) { in mlx5e_xmit_xdp_doorbell()
102 mlx5e_notify_hw(&sq in mlx5e_xmit_xdp_doorbell()
110 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) mlx5e_xdp_get_inline_state() argument
140 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) mlx5e_xdp_mpwqe_add_dseg() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_tx.c46 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) in mlx5e_dma_unmap_wqe_err() argument
52 mlx5e_dma_get(sq, --sq->dma_fifo_pc); in mlx5e_dma_unmap_wqe_err()
54 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma); in mlx5e_dma_unmap_wqe_err()
119 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_eseg_csum() argument
123 if (unlikely(mlx5e_ipsec_txwqe_build_eseg_csum(sq, skb, eseg))) in mlx5e_txwqe_build_eseg_csum()
131 sq->stats->csum_partial_inner++; in mlx5e_txwqe_build_eseg_csum()
134 sq->stats->csum_partial++; in mlx5e_txwqe_build_eseg_csum()
139 sq->stats->csum_partial++; in mlx5e_txwqe_build_eseg_csum()
142 sq in mlx5e_txwqe_build_eseg_csum()
149 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb, int *hopbyhop) mlx5e_tx_get_gso_ihs() argument
177 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) mlx5e_txwqe_build_dsegs() argument
243 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) mlx5e_tx_wqe_inline_mode() argument
262 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) mlx5e_sq_xmit_prepare() argument
342 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) mlx5e_tx_check_stop() argument
350 mlx5e_tx_flush(struct mlx5e_txqsq *sq) mlx5e_tx_flush() argument
371 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_txwqe_complete() argument
421 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) mlx5e_sq_xmit_wqe() argument
512 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_same_eseg() argument
520 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_session_start() argument
544 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_is_active() argument
549 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) mlx5e_tx_mpwqe_add_dseg() argument
567 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_complete() argument
599 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_sq_xmit_mpwqe() argument
648 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_ensure_complete() argument
663 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) mlx5e_txwqe_build_eseg() argument
680 struct mlx5e_txqsq *sq; mlx5e_xmit() local
733 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_tx_wi_dma_unmap() argument
745 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_consume_skb() argument
763 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_tx_wi_consume_fifo_skbs() argument
775 mlx5e_txqsq_wake(struct mlx5e_txqsq *sq) mlx5e_txqsq_wake() argument
789 struct mlx5e_txqsq *sq; mlx5e_poll_tx_cq() local
890 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) mlx5e_tx_wi_kfree_fifo_skbs() argument
898 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) mlx5e_free_txqsq_descs() argument
973 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) mlx5i_sq_xmit() argument
[all...]
H A Den_txrx.c49 static void mlx5e_handle_tx_dim(struct mlx5e_txqsq *sq) in mlx5e_handle_tx_dim() argument
51 struct mlx5e_sq_stats *stats = sq->stats; in mlx5e_handle_tx_dim()
54 if (unlikely(!test_bit(MLX5E_SQ_STATE_DIM, &sq->state))) in mlx5e_handle_tx_dim()
57 dim_update_sample(sq->cq.event_ctr, stats->packets, stats->bytes, &dim_sample); in mlx5e_handle_tx_dim()
58 net_dim(&sq->dim, dim_sample); in mlx5e_handle_tx_dim()
73 void mlx5e_trigger_irq(struct mlx5e_icosq *sq) in mlx5e_trigger_irq() argument
75 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_trigger_irq()
77 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_trigger_irq()
79 sq->db.wqe_info[pi] = (struct mlx5e_icosq_wqe_info) { in mlx5e_trigger_irq()
84 nopwqe = mlx5e_post_nop(wq, sq in mlx5e_trigger_irq()
157 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); mlx5e_napi_poll() local
225 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); mlx5e_napi_poll() local
[all...]
/kernel/linux/linux-6.6/drivers/nvme/target/
H A Dfabrics-cmd-auth.c17 struct nvmet_sq *sq = container_of(to_delayed_work(work), in nvmet_auth_expired_work() local
21 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); in nvmet_auth_expired_work()
22 sq->dhchap_step = NVME_AUTH_DHCHAP_MESSAGE_NEGOTIATE; in nvmet_auth_expired_work()
23 sq->dhchap_tid = -1; in nvmet_auth_expired_work()
26 void nvmet_auth_sq_init(struct nvmet_sq *sq) in nvmet_auth_sq_init() argument
29 INIT_DELAYED_WORK(&sq->auth_expired_work, nvmet_auth_expired_work); in nvmet_auth_sq_init()
30 sq->authenticated = false; in nvmet_auth_sq_init()
31 sq in nvmet_auth_sq_init()
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dxdp.c61 mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq, in mlx5e_xmit_xdp_buff() argument
96 dma_addr = dma_map_single(sq->pdev, xdptxd->data, xdptxd->len, in mlx5e_xmit_xdp_buff()
98 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff()
105 if (unlikely(!INDIRECT_CALL_2(sq->xmit_xdp_frame, mlx5e_xmit_xdp_frame_mpwqe, in mlx5e_xmit_xdp_buff()
106 mlx5e_xmit_xdp_frame, sq, xdptxd, 0))) in mlx5e_xmit_xdp_buff()
110 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff()
112 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff()
114 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, in mlx5e_xmit_xdp_buff()
126 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd->len, DMA_BIDIRECTIONAL); in mlx5e_xmit_xdp_buff()
140 dma_sync_single_for_device(sq in mlx5e_xmit_xdp_buff()
303 mlx5e_xdpsq_get_next_pi(struct mlx5e_xdpsq *sq, u16 size) mlx5e_xdpsq_get_next_pi() argument
332 mlx5e_xdp_mpwqe_session_start(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_session_start() argument
354 mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq) mlx5e_xdp_mpwqe_complete() argument
382 mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check_mpwqe() argument
404 mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result) mlx5e_xmit_xdp_frame_mpwqe() argument
464 mlx5e_xmit_xdp_frame_check_stop_room(struct mlx5e_xdpsq *sq, int stop_room) mlx5e_xmit_xdp_frame_check_stop_room() argument
476 mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq) mlx5e_xmit_xdp_frame_check() argument
482 mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, int check_result) mlx5e_xmit_xdp_frame() argument
610 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq) mlx5e_free_xdpsq_desc() argument
686 struct mlx5e_xdpsq *sq; mlx5e_poll_xdpsq_cq() local
754 mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq) mlx5e_free_xdpsq_descs() argument
786 struct mlx5e_xdpsq *sq; mlx5e_xdp_xmit() local
903 mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw) mlx5e_set_xmit_fp() argument
[all...]
H A Dreporter_tx.c22 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) in mlx5e_wait_for_sq_flush() argument
24 struct mlx5_core_dev *dev = sq->mdev; in mlx5e_wait_for_sq_flush()
30 if (sq->cc == sq->pc) in mlx5e_wait_for_sq_flush()
36 netdev_err(sq->netdev, in mlx5e_wait_for_sq_flush()
37 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n", in mlx5e_wait_for_sq_flush()
38 sq->sqn, sq->cc, sq in mlx5e_wait_for_sq_flush()
43 mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) mlx5e_reset_txqsq_cc_pc() argument
53 mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq) mlx5e_health_sq_put_sw_state() argument
78 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_err_cqe_recover() local
130 struct mlx5e_txqsq *sq; global() member
139 struct mlx5e_txqsq *sq; mlx5e_tx_reporter_timeout_recover() local
224 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output_sq_common() argument
276 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq, int tc) mlx5e_tx_reporter_build_diagnose_output() argument
458 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_diagnose() local
491 struct mlx5e_txqsq *sq = ctx; mlx5e_tx_reporter_dump_sq() local
597 struct mlx5e_txqsq *sq = &c->sq[tc]; mlx5e_tx_reporter_dump_all_sqs() local
607 struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq; mlx5e_tx_reporter_dump_all_sqs() local
636 mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_err_cqe() argument
650 mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq) mlx5e_reporter_tx_timeout() argument
[all...]
H A Dxdp.h106 void mlx5e_xdp_mpwqe_complete(struct mlx5e_xdpsq *sq);
108 void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq);
109 void mlx5e_set_xmit_fp(struct mlx5e_xdpsq *sq, bool is_mpw);
116 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame_mpwqe(struct mlx5e_xdpsq *sq,
119 INDIRECT_CALLABLE_DECLARE(bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq,
122 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check_mpwqe(struct mlx5e_xdpsq *sq));
123 INDIRECT_CALLABLE_DECLARE(int mlx5e_xmit_xdp_frame_check(struct mlx5e_xdpsq *sq));
153 static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) in mlx5e_xmit_xdp_doorbell() argument
155 if (sq->doorbell_cseg) { in mlx5e_xmit_xdp_doorbell()
156 mlx5e_notify_hw(&sq in mlx5e_xmit_xdp_doorbell()
164 mlx5e_xdp_get_inline_state(struct mlx5e_xdpsq *sq, bool cur) mlx5e_xdp_get_inline_state() argument
195 mlx5e_xdp_mpwqe_add_dseg(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd, struct mlx5e_xdpsq_stats *stats) mlx5e_xdp_mpwqe_add_dseg() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_tx.c43 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) in mlx5e_dma_unmap_wqe_err() argument
49 mlx5e_dma_get(sq, --sq->dma_fifo_pc); in mlx5e_dma_unmap_wqe_err()
51 mlx5e_tx_dma_unmap(sq->pdev, last_pushed_dma); in mlx5e_dma_unmap_wqe_err()
151 ipsec_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in ipsec_txwqe_build_eseg_csum() argument
157 sq->stats->csum_partial_inner++; in ipsec_txwqe_build_eseg_csum()
159 sq->stats->csum_partial++; in ipsec_txwqe_build_eseg_csum()
164 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_eseg_csum() argument
173 sq->stats->csum_partial_inner++; in mlx5e_txwqe_build_eseg_csum()
176 sq in mlx5e_txwqe_build_eseg_csum()
191 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb) mlx5e_tx_get_gso_ihs() argument
213 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, unsigned char *skb_data, u16 headlen, struct mlx5_wqe_data_seg *dseg) mlx5e_txwqe_build_dsegs() argument
278 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel) mlx5e_tx_wqe_inline_mode() argument
297 mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5e_tx_attr *attr) mlx5e_sq_xmit_prepare() argument
370 mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) mlx5e_tx_check_stop() argument
379 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, u8 num_dma, struct mlx5e_tx_wqe_info *wi, struct mlx5_wqe_ctrl_seg *cseg, bool xmit_more) mlx5e_txwqe_complete() argument
411 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, const struct mlx5e_tx_attr *attr, const struct mlx5e_tx_wqe_attr *wqe_attr, struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) mlx5e_sq_xmit_wqe() argument
472 mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_same_eseg() argument
480 mlx5e_tx_mpwqe_session_start(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg) mlx5e_tx_mpwqe_session_start() argument
504 mlx5e_tx_mpwqe_session_is_active(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_is_active() argument
509 mlx5e_tx_mpwqe_add_dseg(struct mlx5e_txqsq *sq, struct mlx5e_xmit_data *txd) mlx5e_tx_mpwqe_add_dseg() argument
527 mlx5e_tx_mpwqe_session_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_session_complete() argument
559 mlx5e_sq_xmit_mpwqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, bool xmit_more) mlx5e_sq_xmit_mpwqe() argument
609 mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq) mlx5e_tx_mpwqe_ensure_complete() argument
616 mlx5e_txwqe_build_eseg(struct mlx5e_priv *priv, struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5e_accel_tx_state *accel, struct mlx5_wqe_eth_seg *eseg, u16 ihs) mlx5e_txwqe_build_eseg() argument
635 struct mlx5e_txqsq *sq; mlx5e_xmit() local
676 mlx5e_sq_xmit_simple(struct mlx5e_txqsq *sq, struct sk_buff *skb, bool xmit_more) mlx5e_sq_xmit_simple() argument
691 mlx5e_tx_wi_dma_unmap(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_tx_wi_dma_unmap() argument
703 mlx5e_consume_skb(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_consume_skb() argument
717 mlx5e_tx_wi_consume_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, struct mlx5_cqe64 *cqe, int napi_budget) mlx5e_tx_wi_consume_fifo_skbs() argument
732 struct mlx5e_txqsq *sq; mlx5e_poll_tx_cq() local
839 mlx5e_tx_wi_kfree_fifo_skbs(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi) mlx5e_tx_wi_kfree_fifo_skbs() argument
847 mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq) mlx5e_free_txqsq_descs() argument
922 mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, struct mlx5_av *av, u32 dqpn, u32 dqkey, bool xmit_more) mlx5i_sq_xmit() argument
[all...]
/kernel/linux/linux-5.10/tools/io_uring/
H A Dqueue.c70 struct io_uring_sq *sq = &ring->sq; in io_uring_submit() local
71 const unsigned mask = *sq->kring_mask; in io_uring_submit()
81 if (*sq->khead != *sq->ktail) { in io_uring_submit()
82 submitted = *sq->kring_entries; in io_uring_submit()
86 if (sq->sqe_head == sq->sqe_tail) in io_uring_submit()
93 ktail = ktail_next = *sq->ktail; in io_uring_submit()
94 to_submit = sq in io_uring_submit()
143 struct io_uring_sq *sq = &ring->sq; io_uring_get_sqe() local
[all...]
H A Dsetup.c11 struct io_uring_sq *sq, struct io_uring_cq *cq) in io_uring_mmap()
17 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap()
18 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE, in io_uring_mmap()
22 sq->khead = ptr + p->sq_off.head; in io_uring_mmap()
23 sq->ktail = ptr + p->sq_off.tail; in io_uring_mmap()
24 sq->kring_mask = ptr + p->sq_off.ring_mask; in io_uring_mmap()
25 sq->kring_entries = ptr + p->sq_off.ring_entries; in io_uring_mmap()
26 sq->kflags = ptr + p->sq_off.flags; in io_uring_mmap()
27 sq->kdropped = ptr + p->sq_off.dropped; in io_uring_mmap()
28 sq in io_uring_mmap()
10 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) io_uring_mmap() argument
100 struct io_uring_sq *sq = &ring->sq; io_uring_queue_exit() local
[all...]
/kernel/linux/linux-5.10/sound/oss/dmasound/
H A Ddmasound_core.c411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) in sq_allocate_buffers() argument
415 if (sq->buffers) in sq_allocate_buffers()
417 sq->numBufs = num; in sq_allocate_buffers()
418 sq->bufSize = size; in sq_allocate_buffers()
419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL); in sq_allocate_buffers()
420 if (!sq->buffers) in sq_allocate_buffers()
423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); in sq_allocate_buffers()
424 if (!sq->buffers[i]) { in sq_allocate_buffers()
426 dmasound.mach.dma_free(sq->buffers[i], size); in sq_allocate_buffers()
427 kfree(sq in sq_allocate_buffers()
435 sq_release_buffers(struct sound_queue *sq) sq_release_buffers() argument
448 sq_setup(struct sound_queue *sq) sq_setup() argument
688 sq_init_waitqueue(struct sound_queue *sq) sq_init_waitqueue() argument
707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) sq_open2() argument
953 set_queue_frags(struct sound_queue *sq, int bufs, int size) set_queue_frags() argument
[all...]
/kernel/linux/linux-6.6/sound/oss/dmasound/
H A Ddmasound_core.c411 static int sq_allocate_buffers(struct sound_queue *sq, int num, int size) in sq_allocate_buffers() argument
415 if (sq->buffers) in sq_allocate_buffers()
417 sq->numBufs = num; in sq_allocate_buffers()
418 sq->bufSize = size; in sq_allocate_buffers()
419 sq->buffers = kmalloc_array (num, sizeof(char *), GFP_KERNEL); in sq_allocate_buffers()
420 if (!sq->buffers) in sq_allocate_buffers()
423 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); in sq_allocate_buffers()
424 if (!sq->buffers[i]) { in sq_allocate_buffers()
426 dmasound.mach.dma_free(sq->buffers[i], size); in sq_allocate_buffers()
427 kfree(sq in sq_allocate_buffers()
435 sq_release_buffers(struct sound_queue *sq) sq_release_buffers() argument
448 sq_setup(struct sound_queue *sq) sq_setup() argument
688 sq_init_waitqueue(struct sound_queue *sq) sq_init_waitqueue() argument
707 sq_open2(struct sound_queue *sq, struct file *file, fmode_t mode, int numbufs, int bufsize) sq_open2() argument
953 set_queue_frags(struct sound_queue *sq, int bufs, int size) set_queue_frags() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dktls_tx.c16 mlx5e_ktls_dumps_num_wqes(struct mlx5e_txqsq *sq, unsigned int nfrags, in mlx5e_ktls_dumps_num_wqes() argument
22 return nfrags + DIV_ROUND_UP(sync_len, sq->hw_mtu); in mlx5e_ktls_dumps_num_wqes()
25 u16 mlx5e_ktls_get_stop_room(struct mlx5e_txqsq *sq) in mlx5e_ktls_get_stop_room() argument
29 num_dumps = mlx5e_ktls_dumps_num_wqes(sq, MAX_SKB_FRAGS, TLS_MAX_PAYLOAD_SIZE); in mlx5e_ktls_get_stop_room()
138 static void tx_fill_wi(struct mlx5e_txqsq *sq, in tx_fill_wi() argument
142 struct mlx5e_tx_wqe_info *wi = &sq->db.wqe_info[pi]; in tx_fill_wi()
162 post_static_params(struct mlx5e_txqsq *sq, in post_static_params() argument
170 pi = mlx5e_txqsq_get_next_pi(sq, num_wqebbs); in post_static_params()
171 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi); in post_static_params()
172 mlx5e_ktls_build_static_params(wqe, sq in post_static_params()
180 post_progress_params(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, bool fence) post_progress_params() argument
197 mlx5e_ktls_tx_post_param_wqes(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, bool skip_static_post, bool fence_first_post) mlx5e_ktls_tx_post_param_wqes() argument
278 tx_post_resync_params(struct mlx5e_txqsq *sq, struct mlx5e_ktls_offload_context_tx *priv_tx, u64 rcd_sn) tx_post_resync_params() argument
299 tx_post_resync_dump(struct mlx5e_txqsq *sq, skb_frag_t *frag, u32 tisn, bool first) tx_post_resync_dump() argument
340 mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq, struct mlx5e_tx_wqe_info *wi, u32 *dma_fifo_cc) mlx5e_ktls_tx_handle_resync_dump_comp() argument
356 tx_post_fence_nop(struct mlx5e_txqsq *sq) tx_post_fence_nop() argument
367 mlx5e_ktls_tx_handle_ooo(struct mlx5e_ktls_offload_context_tx *priv_tx, struct mlx5e_txqsq *sq, int datalen, u32 seq) mlx5e_ktls_tx_handle_ooo() argument
442 mlx5e_ktls_handle_tx_skb(struct tls_context *tls_ctx, struct mlx5e_txqsq *sq, struct sk_buff *skb, int datalen, struct mlx5e_accel_tx_tls_state *state) mlx5e_ktls_handle_tx_skb() argument
[all...]
H A Dktls_rx.c128 static void icosq_fill_wi(struct mlx5e_icosq *sq, u16 pi, in icosq_fill_wi() argument
131 sq->db.wqe_info[pi] = *wi; in icosq_fill_wi()
135 post_static_params(struct mlx5e_icosq *sq, in post_static_params() argument
144 if (unlikely(!mlx5e_wqc_has_room_for(&sq->wq, sq->cc, sq->pc, room))) in post_static_params()
147 pi = mlx5e_icosq_get_next_pi(sq, num_wqebbs); in post_static_params()
148 wqe = MLX5E_TLS_FETCH_SET_STATIC_PARAMS_WQE(sq, pi); in post_static_params()
149 mlx5e_ktls_build_static_params(wqe, sq->pc, sq in post_static_params()
165 post_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx, u32 next_record_tcp_sn) post_progress_params() argument
200 struct mlx5e_icosq *sq; post_rx_param_wqes() local
251 resync_post_get_progress_params(struct mlx5e_icosq *sq, struct mlx5e_ktls_offload_context_rx *priv_rx) resync_post_get_progress_params() argument
336 struct mlx5e_icosq *sq; resync_handle_work() local
369 struct mlx5e_icosq *sq; resync_handle_seq_match() local
398 mlx5e_ktls_handle_get_psv_completion(struct mlx5e_icosq_wqe_info *wi, struct mlx5e_icosq *sq) mlx5e_ktls_handle_get_psv_completion() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dtx.c46 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, in mlx5e_xsk_tx_post_err() argument
49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); in mlx5e_xsk_tx_post_err()
50 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi]; in mlx5e_xsk_tx_post_err()
56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); in mlx5e_xsk_tx_post_err()
57 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, *xdpi); in mlx5e_xsk_tx_post_err()
58 sq->doorbell_cseg = &nopwqe->ctrl; in mlx5e_xsk_tx_post_err()
61 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigne argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dtx.c52 static void mlx5e_xsk_tx_post_err(struct mlx5e_xdpsq *sq, in mlx5e_xsk_tx_post_err() argument
55 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); in mlx5e_xsk_tx_post_err()
56 struct mlx5e_xdp_wqe_info *wi = &sq->db.wqe_info[pi]; in mlx5e_xsk_tx_post_err()
62 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); in mlx5e_xsk_tx_post_err()
63 mlx5e_xdpi_fifo_push(&sq->db.xdpi_fifo, xdpi); in mlx5e_xsk_tx_post_err()
64 sq->doorbell_cseg = &nopwqe->ctrl; in mlx5e_xsk_tx_post_err()
67 bool mlx5e_xsk_tx(struct mlx5e_xdpsq *sq, unsigne argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dqos_sq.c38 struct otx2_snd_queue *sq; in otx2_qos_sq_aura_pool_init() local
76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init()
77 sq->sqb_count = 0; in otx2_qos_sq_aura_pool_init()
78 sq->sqb_ptrs = kcalloc(num_sqbs, sizeof(*sq->sqb_ptrs), GFP_KERNEL); in otx2_qos_sq_aura_pool_init()
79 if (!sq->sqb_ptrs) { in otx2_qos_sq_aura_pool_init()
89 sq->sqb_ptrs[sq->sqb_count++] = (u64)bufptr; in otx2_qos_sq_aura_pool_init()
96 if (!sq in otx2_qos_sq_aura_pool_init()
120 struct otx2_snd_queue *sq; otx2_qos_sq_free_sqbs() local
263 struct otx2_snd_queue *sq; otx2_qos_disable_sq() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c80 struct otx2_snd_queue *sq, in otx2_snd_pkt_handler()
96 sg = &sq->sg[snd_comp->sqe_id]; in otx2_snd_pkt_handler()
102 timestamp = ((u64 *)sq->timestamps->base)[snd_comp->sqe_id]; in otx2_snd_pkt_handler()
372 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[cq->cint_idx], in otx2_tx_napi_handler()
442 static void otx2_sqe_flush(struct otx2_snd_queue *sq, int size) in otx2_sqe_flush() argument
450 memcpy(sq->lmt_addr, sq->sqe_base, size); in otx2_sqe_flush()
451 status = otx2_lmt_flush(sq->io_addr); in otx2_sqe_flush()
454 sq->head++; in otx2_sqe_flush()
455 sq in otx2_sqe_flush()
78 otx2_snd_pkt_handler(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, struct otx2_snd_queue *sq, struct nix_cqe_tx_s *cqe, int budget, int *tx_pkts, int *tx_bytes) otx2_snd_pkt_handler() argument
460 otx2_sqe_add_sg(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int num_segs, int *offset) otx2_sqe_add_sg() argument
505 otx2_sqe_add_ext(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int *offset) otx2_sqe_add_ext() argument
560 otx2_sqe_add_mem(struct otx2_snd_queue *sq, int *offset, int alg, u64 iova) otx2_sqe_add_mem() argument
575 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) otx2_sqe_add_hdr() argument
622 otx2_dma_map_tso_skb(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, int sqe, int hdr_len) otx2_dma_map_tso_skb() argument
655 otx2_tso_frag_dma_addr(struct otx2_snd_queue *sq, struct sk_buff *skb, int seg, u64 seg_addr, int hdr_len, int sqe) otx2_tso_frag_dma_addr() argument
673 otx2_sqe_tso_add_sg(struct otx2_snd_queue *sq, struct sg_list *list, int *offset) otx2_sqe_tso_add_sg() argument
704 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_tso() argument
825 otx2_set_txtstamp(struct otx2_nic *pfvf, struct sk_buff *skb, struct otx2_snd_queue *sq, int *offset) otx2_set_txtstamp() argument
840 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_skb() argument
934 struct otx2_snd_queue *sq; otx2_cleanup_tx_cqes() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/
H A Dnicvf_queues.c19 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry,
504 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue()
508 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE, in nicvf_init_snd_queue()
513 sq->desc = sq->dmem.base; in nicvf_init_snd_queue()
514 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); in nicvf_init_snd_queue()
515 if (!sq->skbuff) in nicvf_init_snd_queue()
518 sq->head = 0; in nicvf_init_snd_queue()
519 sq->tail = 0; in nicvf_init_snd_queue()
520 sq in nicvf_init_snd_queue()
503 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument
553 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) nicvf_unmap_sndq_buffers() argument
571 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) nicvf_free_snd_queue() argument
862 struct snd_queue *sq; nicvf_snd_queue_config() local
1118 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_get_sq_desc() argument
1134 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) nicvf_rollback_sq_desc() argument
1142 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_put_sq_desc() argument
1152 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) nicvf_get_nxt_sqentry() argument
1159 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument
1179 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument
1205 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) nicvf_xdp_sq_doorbell() argument
1221 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) nicvf_xdp_sq_add_hdr_subdesc() argument
1235 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) nicvf_xdp_sq_append_pkt() argument
1329 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) nicvf_sq_add_hdr_subdesc() argument
1418 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) nicvf_sq_add_gather_subdesc() argument
1437 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) nicvf_sq_add_cqe_subdesc() argument
1485 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) nicvf_sq_append_tso() argument
1549 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) nicvf_sq_append_skb() argument
1826 struct snd_queue *sq; nicvf_update_sq_stats() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/thunder/
H A Dnicvf_queues.c20 static inline void nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry,
505 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue()
509 err = nicvf_alloc_q_desc_mem(nic, &sq->dmem, q_len, SND_QUEUE_DESC_SIZE, in nicvf_init_snd_queue()
514 sq->desc = sq->dmem.base; in nicvf_init_snd_queue()
515 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); in nicvf_init_snd_queue()
516 if (!sq->skbuff) in nicvf_init_snd_queue()
519 sq->head = 0; in nicvf_init_snd_queue()
520 sq->tail = 0; in nicvf_init_snd_queue()
521 sq in nicvf_init_snd_queue()
504 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument
554 nicvf_unmap_sndq_buffers(struct nicvf *nic, struct snd_queue *sq, int hdr_sqe, u8 subdesc_cnt) nicvf_unmap_sndq_buffers() argument
572 nicvf_free_snd_queue(struct nicvf *nic, struct snd_queue *sq) nicvf_free_snd_queue() argument
863 struct snd_queue *sq; nicvf_snd_queue_config() local
1119 nicvf_get_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_get_sq_desc() argument
1135 nicvf_rollback_sq_desc(struct snd_queue *sq, int qentry, int desc_cnt) nicvf_rollback_sq_desc() argument
1143 nicvf_put_sq_desc(struct snd_queue *sq, int desc_cnt) nicvf_put_sq_desc() argument
1153 nicvf_get_nxt_sqentry(struct snd_queue *sq, int qentry) nicvf_get_nxt_sqentry() argument
1160 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument
1180 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument
1206 nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num) nicvf_xdp_sq_doorbell() argument
1222 nicvf_xdp_sq_add_hdr_subdesc(struct snd_queue *sq, int qentry, int subdesc_cnt, u64 data, int len) nicvf_xdp_sq_add_hdr_subdesc() argument
1236 nicvf_xdp_sq_append_pkt(struct nicvf *nic, struct snd_queue *sq, u64 bufaddr, u64 dma_addr, u16 len) nicvf_xdp_sq_append_pkt() argument
1330 nicvf_sq_add_hdr_subdesc(struct nicvf *nic, struct snd_queue *sq, int qentry, int subdesc_cnt, struct sk_buff *skb, int len) nicvf_sq_add_hdr_subdesc() argument
1419 nicvf_sq_add_gather_subdesc(struct snd_queue *sq, int qentry, int size, u64 data) nicvf_sq_add_gather_subdesc() argument
1438 nicvf_sq_add_cqe_subdesc(struct snd_queue *sq, int qentry, int tso_sqe, struct sk_buff *skb) nicvf_sq_add_cqe_subdesc() argument
1486 nicvf_sq_append_tso(struct nicvf *nic, struct snd_queue *sq, int sq_num, int qentry, struct sk_buff *skb) nicvf_sq_append_tso() argument
1550 nicvf_sq_append_skb(struct nicvf *nic, struct snd_queue *sq, struct sk_buff *skb, u8 sq_num) nicvf_sq_append_skb() argument
1827 struct snd_queue *sq; nicvf_update_sq_stats() local
[all...]
/kernel/linux/linux-6.6/tools/testing/selftests/net/
H A Dio_uring_zerocopy_tx.c108 struct io_uring_sq sq; member
169 struct io_uring_sq *sq, struct io_uring_cq *cq) in io_uring_mmap()
175 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap()
176 ptr = mmap(0, sq->ring_sz, PROT_READ | PROT_WRITE, in io_uring_mmap()
180 sq->khead = ptr + p->sq_off.head; in io_uring_mmap()
181 sq->ktail = ptr + p->sq_off.tail; in io_uring_mmap()
182 sq->kring_mask = ptr + p->sq_off.ring_mask; in io_uring_mmap()
183 sq->kring_entries = ptr + p->sq_off.ring_entries; in io_uring_mmap()
184 sq->kflags = ptr + p->sq_off.flags; in io_uring_mmap()
185 sq in io_uring_mmap()
168 io_uring_mmap(int fd, struct io_uring_params *p, struct io_uring_sq *sq, struct io_uring_cq *cq) io_uring_mmap() argument
238 struct io_uring_sq *sq = &ring->sq; io_uring_submit() local
293 struct io_uring_sq *sq = &ring->sq; io_uring_get_sqe() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/
H A Dice_controlq.c8 (qinfo)->sq.head = prefix##_ATQH; \
9 (qinfo)->sq.tail = prefix##_ATQT; \
10 (qinfo)->sq.len = prefix##_ATQLEN; \
11 (qinfo)->sq.bah = prefix##_ATQBAH; \
12 (qinfo)->sq.bal = prefix##_ATQBAL; \
13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \
14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \
15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \
16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \
64 if (cq->sq in ice_check_sq_alive()
807 struct ice_ctl_q_ring *sq = &cq->sq; ice_clean_sq() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/
H A Dice_controlq.c8 (qinfo)->sq.head = prefix##_ATQH; \
9 (qinfo)->sq.tail = prefix##_ATQT; \
10 (qinfo)->sq.len = prefix##_ATQLEN; \
11 (qinfo)->sq.bah = prefix##_ATQBAH; \
12 (qinfo)->sq.bal = prefix##_ATQBAL; \
13 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \
14 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \
15 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \
16 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \
77 if (cq->sq in ice_check_sq_alive()
860 struct ice_ctl_q_ring *sq = &cq->sq; ice_clean_sq() local
[all...]

Completed in 19 milliseconds

12345678910>>...19