Home
last modified time | relevance | path

Searched refs:cur_edge (Results 1 - 8 of 8) sorted by relevance

/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
H A Dwr.c33 * @cur_edge: Updated current edge.
36 u32 wqe_sz, void **cur_edge) in handle_post_send_edge()
40 if (likely(*seg != *cur_edge)) in handle_post_send_edge()
44 *cur_edge = get_sq_edge(sq, idx); in handle_post_send_edge()
52 * @cur_edge: Updated current edge.
58 static inline void memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, in memcpy_send_wqe() argument
63 size_t leftlen = *cur_edge - *seg; in memcpy_send_wqe()
74 handle_post_send_edge(sq, seg, *wqe_sz, cur_edge); in memcpy_send_wqe()
105 void **seg, int *size, void **cur_edge) in set_eth_seg()
129 copysz = min_t(u64, *cur_edge in set_eth_seg()
35 handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, u32 wqe_sz, void **cur_edge) handle_post_send_edge() argument
104 set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_eth_seg() argument
481 set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **wqe, int *wqe_sz, void **cur_edge) set_data_inl_seg() argument
637 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_sig_data_segment() argument
778 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_pi_umr_wr() argument
856 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) set_reg_wr() argument
910 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_linv_wr() argument
942 __begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq, bool send_signaled, bool solicited) __begin_wqe() argument
967 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) begin_wqe() argument
977 finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) finish_wqe() argument
1015 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_local_inv() argument
1024 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_reg_mr() argument
1033 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) handle_psv() argument
1064 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) handle_reg_mr_integrity() argument
1152 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) handle_qpt_rc() argument
1215 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_hw_gsi() argument
1225 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_ud() argument
1246 handle_qpt_reg_umr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_qpt_reg_umr() argument
1284 void *cur_edge; mlx5_ib_post_send() local
[all...]
H A Dmlx5_ib.h285 void *cur_edge; member
H A Dqp.c1092 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in _create_kernel_qp()
4089 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in __mlx5_ib_modify_qp()
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/
H A Dwr.h49 * @cur_edge: Updated current edge.
52 u32 wqe_sz, void **cur_edge) in handle_post_send_edge()
56 if (likely(*seg != *cur_edge)) in handle_post_send_edge()
60 *cur_edge = get_sq_edge(sq, idx); in handle_post_send_edge()
68 * @cur_edge: Updated current edge.
74 static inline void mlx5r_memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, in mlx5r_memcpy_send_wqe() argument
79 size_t leftlen = *cur_edge - *seg; in mlx5r_memcpy_send_wqe()
90 handle_post_send_edge(sq, seg, *wqe_sz, cur_edge); in mlx5r_memcpy_send_wqe()
97 int *size, void **cur_edge, int nreq, __be32 general_id,
100 void *seg, u8 size, void *cur_edge, unsigne
51 handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, u32 wqe_sz, void **cur_edge) handle_post_send_edge() argument
[all...]
H A Dwr.c55 void **seg, int *size, void **cur_edge) in set_eth_seg()
79 copysz = min_t(u64, *cur_edge - (void *)eseg->inline_hdr.start, in set_eth_seg()
88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_eth_seg()
91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, in set_eth_seg()
261 void **wqe, int *wqe_sz, void **cur_edge) in set_data_inl_seg()
287 cur_edge); in set_data_inl_seg()
289 leftlen = *cur_edge - *wqe; in set_data_inl_seg()
420 void **cur_edge) in set_sig_data_segment()
514 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_sig_data_segment()
523 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_sig_data_segment()
54 set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_eth_seg() argument
260 set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **wqe, int *wqe_sz, void **cur_edge) set_data_inl_seg() argument
416 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_sig_data_segment() argument
557 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_pi_umr_wr() argument
635 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) set_reg_wr() argument
692 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_linv_wr() argument
724 mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, unsigned int *idx, int *size, void **cur_edge, int nreq, __be32 general_id, bool send_signaled, bool solicited) mlx5r_begin_wqe() argument
748 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) begin_wqe() argument
758 mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) mlx5r_finish_wqe() argument
794 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_local_inv() argument
803 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_reg_mr() argument
812 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) handle_psv() argument
843 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) handle_reg_mr_integrity() argument
931 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) handle_qpt_rc() argument
994 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_hw_gsi() argument
1004 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_ud() argument
1059 void *cur_edge; mlx5_ib_post_send() local
[all...]
H A Dumr.c242 void *cur_edge, *seg; in mlx5r_umr_post_send() local
252 err = mlx5r_begin_wqe(qp, &seg, &ctrl, &idx, &size, &cur_edge, 0, in mlx5r_umr_post_send()
259 mlx5r_memcpy_send_wqe(&qp->sq, &cur_edge, &seg, &size, wqe, wqe_size); in mlx5r_umr_post_send()
262 mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, id.wr_id, 0, in mlx5r_umr_post_send()
H A Dmlx5_ib.h376 void *cur_edge; member
H A Dqp.c1151 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in _create_kernel_qp()
4421 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in __mlx5_ib_modify_qp()

Completed in 27 milliseconds