/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | wr.c | 33 * @cur_edge: Updated current edge. 36 u32 wqe_sz, void **cur_edge) in handle_post_send_edge() 40 if (likely(*seg != *cur_edge)) in handle_post_send_edge() 44 *cur_edge = get_sq_edge(sq, idx); in handle_post_send_edge() 52 * @cur_edge: Updated current edge. 58 static inline void memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, in memcpy_send_wqe() argument 63 size_t leftlen = *cur_edge - *seg; in memcpy_send_wqe() 74 handle_post_send_edge(sq, seg, *wqe_sz, cur_edge); in memcpy_send_wqe() 105 void **seg, int *size, void **cur_edge) in set_eth_seg() 129 copysz = min_t(u64, *cur_edge in set_eth_seg() 35 handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, u32 wqe_sz, void **cur_edge) handle_post_send_edge() argument 104 set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_eth_seg() argument 481 set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **wqe, int *wqe_sz, void **cur_edge) set_data_inl_seg() argument 637 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_sig_data_segment() argument 778 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_pi_umr_wr() argument 856 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) set_reg_wr() argument 910 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_linv_wr() argument 942 __begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq, bool send_signaled, bool solicited) __begin_wqe() argument 967 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) begin_wqe() argument 977 finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) finish_wqe() argument 1015 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_local_inv() argument 1024 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_reg_mr() argument 1033 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) handle_psv() argument 1064 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) handle_reg_mr_integrity() argument 1152 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) handle_qpt_rc() argument 1215 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_hw_gsi() argument 1225 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_ud() argument 1246 handle_qpt_reg_umr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_qpt_reg_umr() argument 1284 void *cur_edge; mlx5_ib_post_send() local [all...] |
H A D | mlx5_ib.h | 285 void *cur_edge; member
|
H A D | qp.c | 1092 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in _create_kernel_qp() 4089 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in __mlx5_ib_modify_qp()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | wr.h | 49 * @cur_edge: Updated current edge. 52 u32 wqe_sz, void **cur_edge) in handle_post_send_edge() 56 if (likely(*seg != *cur_edge)) in handle_post_send_edge() 60 *cur_edge = get_sq_edge(sq, idx); in handle_post_send_edge() 68 * @cur_edge: Updated current edge. 74 static inline void mlx5r_memcpy_send_wqe(struct mlx5_ib_wq *sq, void **cur_edge, in mlx5r_memcpy_send_wqe() argument 79 size_t leftlen = *cur_edge - *seg; in mlx5r_memcpy_send_wqe() 90 handle_post_send_edge(sq, seg, *wqe_sz, cur_edge); in mlx5r_memcpy_send_wqe() 97 int *size, void **cur_edge, int nreq, __be32 general_id, 100 void *seg, u8 size, void *cur_edge, unsigne 51 handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, u32 wqe_sz, void **cur_edge) handle_post_send_edge() argument [all...] |
H A D | wr.c | 55 void **seg, int *size, void **cur_edge) in set_eth_seg() 79 copysz = min_t(u64, *cur_edge - (void *)eseg->inline_hdr.start, in set_eth_seg() 88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_eth_seg() 91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, in set_eth_seg() 261 void **wqe, int *wqe_sz, void **cur_edge) in set_data_inl_seg() 287 cur_edge); in set_data_inl_seg() 289 leftlen = *cur_edge - *wqe; in set_data_inl_seg() 420 void **cur_edge) in set_sig_data_segment() 514 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_sig_data_segment() 523 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_sig_data_segment() 54 set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_eth_seg() argument 260 set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **wqe, int *wqe_sz, void **cur_edge) set_data_inl_seg() argument 416 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_sig_data_segment() argument 557 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_pi_umr_wr() argument 635 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) set_reg_wr() argument 692 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) set_linv_wr() argument 724 mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, unsigned int *idx, int *size, void **cur_edge, int nreq, __be32 general_id, bool send_signaled, bool solicited) mlx5r_begin_wqe() argument 748 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) begin_wqe() argument 758 mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) mlx5r_finish_wqe() argument 794 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_local_inv() argument 803 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) handle_reg_mr() argument 812 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) handle_psv() argument 843 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) handle_reg_mr_integrity() argument 931 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) handle_qpt_rc() argument 994 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_hw_gsi() argument 1004 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) handle_qpt_ud() argument 1059 void *cur_edge; mlx5_ib_post_send() local [all...] |
H A D | umr.c | 242 void *cur_edge, *seg; in mlx5r_umr_post_send() local 252 err = mlx5r_begin_wqe(qp, &seg, &ctrl, &idx, &size, &cur_edge, 0, in mlx5r_umr_post_send() 259 mlx5r_memcpy_send_wqe(&qp->sq, &cur_edge, &seg, &size, wqe, wqe_size); in mlx5r_umr_post_send() 262 mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, id.wr_id, 0, in mlx5r_umr_post_send()
|
H A D | mlx5_ib.h | 376 void *cur_edge; member
|
H A D | qp.c | 1151 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in _create_kernel_qp() 4421 qp->sq.cur_edge = get_sq_edge(&qp->sq, 0); in __mlx5_ib_modify_qp()
|