Home
last modified time | relevance | path

Searched refs:wqe_idx (Results 1 - 25 of 25) sorted by relevance

/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/
H A Di40iw_uk.c51 u32 wqe_idx, peek_head; in i40iw_nop_1() local
57 wqe_idx = I40IW_RING_GETCURRENT_HEAD(qp->sq_ring); in i40iw_nop_1()
58 wqe = qp->sq_base[wqe_idx].elem; in i40iw_nop_1()
60 qp->sq_wrtrk_array[wqe_idx].wqe_size = I40IW_QP_WQE_MIN_SIZE; in i40iw_nop_1()
120 * @wqe_idx: wqe index
122 static void i40iw_qp_ring_push_db(struct i40iw_qp_uk *qp, u32 wqe_idx) in i40iw_qp_ring_push_db() argument
124 set_32bit_val(qp->push_db, 0, LS_32((wqe_idx >> 2), I40E_PFPE_WQEALLOC_WQE_DESC_INDEX) | qp->qp_id); in i40iw_qp_ring_push_db()
131 * @wqe_idx: return wqe index
135 u32 *wqe_idx, in i40iw_qp_get_next_send_wqe()
149 *wqe_idx in i40iw_qp_get_next_send_wqe()
134 i40iw_qp_get_next_send_wqe(struct i40iw_qp_uk *qp, u32 *wqe_idx, u8 wqe_size, u32 total_size, u64 wr_id ) i40iw_qp_get_next_send_wqe() argument
220 i40iw_qp_get_next_recv_wqe(struct i40iw_qp_uk *qp, u32 *wqe_idx) i40iw_qp_get_next_recv_wqe() argument
252 u32 i, wqe_idx; i40iw_rdma_write() local
322 u32 wqe_idx; i40iw_rdma_read() local
370 u32 i, wqe_idx, total_size = 0, byte_off; i40iw_send() local
431 u32 wqe_idx; i40iw_inline_rdma_write() local
506 u32 wqe_idx; i40iw_inline_send() local
576 u32 wqe_idx; i40iw_stag_local_invalidate() local
618 u32 wqe_idx; i40iw_mw_bind() local
663 u32 total_size = 0, wqe_idx, i, byte_off; i40iw_post_receive() local
756 u32 wqe_idx, q_type, array_idx = 0; i40iw_cq_poll_completion() local
1133 u32 wqe_idx; i40iw_nop() local
[all...]
H A Di40iw_puda.c48 static void i40iw_ilq_putback_rcvbuf(struct i40iw_sc_qp *qp, u32 wqe_idx);
105 * @wqe_idx: wqe index to use
109 static void i40iw_puda_post_recvbuf(struct i40iw_puda_rsrc *rsrc, u32 wqe_idx, in i40iw_puda_post_recvbuf() argument
116 qp->qp_uk.rq_wrid_array[wqe_idx] = (uintptr_t)buf; in i40iw_puda_post_recvbuf()
117 wqe = qp->qp_uk.rq_base[wqe_idx].elem; in i40iw_puda_post_recvbuf()
119 "%s: wqe_idx= %d buf = %p wqe = %p\n", __func__, in i40iw_puda_post_recvbuf()
120 wqe_idx, buf, wqe); in i40iw_puda_post_recvbuf()
204 * @wqe_idx: wqe index for caller
206 static u64 *i40iw_puda_get_next_send_wqe(struct i40iw_qp_uk *qp, u32 *wqe_idx) in i40iw_puda_get_next_send_wqe() argument
211 *wqe_idx in i40iw_puda_get_next_send_wqe()
377 u32 wqe_idx; i40iw_puda_send() local
989 i40iw_ilq_putback_rcvbuf(struct i40iw_sc_qp *qp, u32 wqe_idx) i40iw_ilq_putback_rcvbuf() argument
1445 u32 wqe_idx = ieq->compl_rxwqe_idx; i40iw_ieq_receive() local
[all...]
H A Di40iw_user.h407 u64 *i40iw_qp_get_next_send_wqe(struct i40iw_qp_uk *qp, u32 *wqe_idx,
412 u64 *i40iw_qp_get_next_recv_wqe(struct i40iw_qp_uk *qp, u32 *wqe_idx);
413 u64 *i40iw_qp_get_next_srq_wqe(struct i40iw_srq_uk *srq, u32 *wqe_idx);
H A Di40iw_ctrl.c602 * @wqe_idx: WQE index for next WQE on CQP SQ
605 u64 scratch, u32 *wqe_idx) in i40iw_sc_cqp_get_next_send_wqe_idx()
620 I40IW_ATOMIC_RING_MOVE_HEAD(cqp->sq_ring, *wqe_idx, ret_code); in i40iw_sc_cqp_get_next_send_wqe_idx()
624 if (!*wqe_idx) in i40iw_sc_cqp_get_next_send_wqe_idx()
627 wqe = cqp->sq_base[*wqe_idx].elem; in i40iw_sc_cqp_get_next_send_wqe_idx()
628 cqp->scratch_array[*wqe_idx] = scratch; in i40iw_sc_cqp_get_next_send_wqe_idx()
641 u32 wqe_idx; in i40iw_sc_cqp_get_next_send_wqe() local
643 return i40iw_sc_cqp_get_next_send_wqe_idx(cqp, scratch, &wqe_idx); in i40iw_sc_cqp_get_next_send_wqe()
726 u32 wqe_idx; in i40iw_sc_ccq_get_cqe_info() local
748 wqe_idx in i40iw_sc_ccq_get_cqe_info()
604 i40iw_sc_cqp_get_next_send_wqe_idx(struct i40iw_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) i40iw_sc_cqp_get_next_send_wqe_idx() argument
1864 u16 wqe_idx; i40iw_sc_get_next_aeqe() local
3357 u32 wqe_idx; i40iw_sc_mr_fast_register() local
3737 u32 wqe_idx; cqp_sds_wqe_fill() local
[all...]
H A Di40iw_utils.c461 u32 i, pending_work, wqe_idx; in i40iw_cleanup_pending_cqp_op() local
464 wqe_idx = I40IW_RING_GETCURRENT_TAIL(cqp->sc_cqp.sq_ring); in i40iw_cleanup_pending_cqp_op()
466 cqp_request = (struct i40iw_cqp_request *)(unsigned long)cqp->scratch_array[wqe_idx]; in i40iw_cleanup_pending_cqp_op()
469 wqe_idx = (wqe_idx + 1) % I40IW_RING_GETSIZE(cqp->sc_cqp.sq_ring); in i40iw_cleanup_pending_cqp_op()
H A Di40iw_puda.h66 u32 wqe_idx; member
H A Di40iw_hw.c418 ctx_info->err_rq_idx = info->wqe_idx; in i40iw_process_aeq()
H A Di40iw_type.h766 u16 wqe_idx; member
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/
H A Duk.c62 u32 wqe_idx; in irdma_nop_1() local
68 wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_nop_1()
69 wqe = qp->sq_base[wqe_idx].elem; in irdma_nop_1()
71 qp->sq_wrtrk_array[wqe_idx].quanta = IRDMA_QP_WQE_MIN_QUANTA; in irdma_nop_1()
92 * @qp_wqe_idx: wqe_idx
97 u32 wqe_idx; in irdma_clr_wqes() local
100 wqe_idx = (qp_wqe_idx + 128) % qp->sq_ring.size; in irdma_clr_wqes()
101 sq = qp->sq_base + wqe_idx; in irdma_clr_wqes()
102 if (wqe_idx) in irdma_clr_wqes()
149 * @wqe_idx
154 irdma_qp_get_next_send_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx, u16 quanta, u32 total_size, struct irdma_post_sq_info *info) irdma_qp_get_next_send_wqe() argument
206 irdma_qp_get_next_recv_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx) irdma_qp_get_next_recv_wqe() argument
238 u32 i, wqe_idx; irdma_uk_rdma_write() local
336 u32 wqe_idx; irdma_uk_rdma_read() local
413 u32 i, wqe_idx, total_size = 0, byte_off; irdma_uk_send() local
660 u32 wqe_idx; irdma_uk_inline_rdma_write() local
728 u32 wqe_idx; irdma_uk_inline_send() local
801 u32 wqe_idx; irdma_uk_stag_local_invalidate() local
844 u32 wqe_idx, i, byte_off; irdma_uk_post_receive() local
977 u32 wqe_idx; irdma_uk_cq_poll_cmpl() local
1535 u32 wqe_idx; irdma_nop() local
[all...]
H A Dpuda.c15 struct irdma_puda_buf *buf, u32 wqe_idx);
75 * @wqe_idx: wqe index to use
79 static void irdma_puda_post_recvbuf(struct irdma_puda_rsrc *rsrc, u32 wqe_idx, in irdma_puda_post_recvbuf() argument
89 qp->qp_uk.rq_wrid_array[wqe_idx] = (uintptr_t)buf; in irdma_puda_post_recvbuf()
90 wqe = qp->qp_uk.rq_base[wqe_idx].elem; in irdma_puda_post_recvbuf()
189 * @wqe_idx: wqe index for caller
192 u32 *wqe_idx) in irdma_puda_get_next_send_wqe()
196 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_puda_get_next_send_wqe()
197 if (!*wqe_idx) in irdma_puda_get_next_send_wqe()
203 return qp->sq_base[*wqe_idx] in irdma_puda_get_next_send_wqe()
191 irdma_puda_get_next_send_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx) irdma_puda_get_next_send_wqe() argument
430 u32 wqe_idx; irdma_puda_send() local
1120 irdma_ilq_putback_rcvbuf(struct irdma_sc_qp *qp, struct irdma_puda_buf *buf, u32 wqe_idx) irdma_ilq_putback_rcvbuf() argument
1684 u32 wqe_idx = ieq->compl_rxwqe_idx; irdma_ieq_receive() local
[all...]
H A Dutils.c537 u32 i, pending_work, wqe_idx; in irdma_cleanup_pending_cqp_op() local
540 wqe_idx = IRDMA_RING_CURRENT_TAIL(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op()
543 cqp->scratch_array[wqe_idx]; in irdma_cleanup_pending_cqp_op()
546 wqe_idx = (wqe_idx + 1) % IRDMA_RING_SIZE(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op()
2459 u32 wqe_idx; in irdma_generate_flush_completions() local
2476 wqe_idx = sq_ring->tail; in irdma_generate_flush_completions()
2479 cmpl->cpi.wr_id = qp->sq_wrtrk_array[wqe_idx].wrid; in irdma_generate_flush_completions()
2480 sw_wqe = qp->sq_base[wqe_idx].elem; in irdma_generate_flush_completions()
2520 wqe_idx in irdma_generate_flush_completions()
[all...]
H A Duser.h398 __le64 *irdma_qp_get_next_send_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx,
401 __le64 *irdma_qp_get_next_recv_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx);
H A Dpuda.h38 u32 wqe_idx; member
H A Dtype.h949 u16 wqe_idx; member
1481 u32 *wqe_idx);
1490 u32 wqe_idx; in irdma_sc_cqp_get_next_send_wqe() local
1492 return irdma_sc_cqp_get_next_send_wqe_idx(cqp, scratch, &wqe_idx); in irdma_sc_cqp_get_next_send_wqe()
H A Dctrl.c1297 u32 wqe_idx; in irdma_sc_mr_fast_register() local
1311 wqe = irdma_qp_get_next_send_wqe(&qp->qp_uk, &wqe_idx, in irdma_sc_mr_fast_register()
1316 irdma_clr_wqes(&qp->qp_uk, wqe_idx); in irdma_sc_mr_fast_register()
1319 "MR: wr_id[%llxh] wqe_idx[%04d] location[%p]\n", in irdma_sc_mr_fast_register()
1320 info->wr_id, wqe_idx, in irdma_sc_mr_fast_register() local
1321 &qp->qp_uk.sq_wrtrk_array[wqe_idx].wrid); in irdma_sc_mr_fast_register()
3258 * @wqe_idx: WQE index of CQP SQ
3261 u32 *wqe_idx) in irdma_sc_cqp_get_next_send_wqe_idx()
3273 IRDMA_ATOMIC_RING_MOVE_HEAD(cqp->sq_ring, *wqe_idx, ret_code); in irdma_sc_cqp_get_next_send_wqe_idx()
3278 if (!*wqe_idx) in irdma_sc_cqp_get_next_send_wqe_idx()
3260 irdma_sc_cqp_get_next_send_wqe_idx(struct irdma_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) irdma_sc_cqp_get_next_send_wqe_idx() argument
3351 u32 wqe_idx; irdma_sc_ccq_get_cqe_info() local
4422 u32 wqe_idx; cqp_sds_wqe_fill() local
[all...]
H A Dhw.c399 ctx_info->roce_info->err_rq_idx = info->wqe_idx; in irdma_process_aeq()
409 ctx_info->iwarp_info->err_rq_idx = info->wqe_idx; in irdma_process_aeq()
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/
H A Derdma_cq.c117 u16 wqe_idx, depth; in erdma_poll_one_cqe() local
132 wqe_idx = be32_to_cpu(cqe->qe_idx); in erdma_poll_one_cqe()
148 wqe_hdr = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, in erdma_poll_one_cqe()
152 wqe_idx + 1; in erdma_poll_one_cqe()
157 wc->wr_id = id_table[wqe_idx & (depth - 1)]; in erdma_poll_one_cqe()
H A Derdma_qp.c208 const struct ib_send_wr *send_wr, u16 wqe_idx, in fill_inline_data()
215 wqe_idx += (sgl_offset >> SQEBB_SHIFT); in fill_inline_data()
217 data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, in fill_inline_data()
238 wqe_idx += (sgl_offset >> SQEBB_SHIFT); in fill_inline_data()
241 data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, in fill_inline_data()
255 u16 wqe_idx, u32 sgl_offset, __le32 *length_field) in fill_sgl()
268 wqe_idx += (sgl_offset >> SQEBB_SHIFT); in fill_sgl()
270 sgl = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, in fill_sgl()
207 fill_inline_data(struct erdma_qp *qp, const struct ib_send_wr *send_wr, u16 wqe_idx, u32 sgl_offset, __le32 *length_field) fill_inline_data() argument
254 fill_sgl(struct erdma_qp *qp, const struct ib_send_wr *send_wr, u16 wqe_idx, u32 sgl_offset, __le32 *length_field) fill_sgl() argument
/kernel/linux/linux-5.10/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c1593 u32 qpn = 0, wqe_idx = 0; in ocrdma_discard_cqes() local
1626 wqe_idx = (le32_to_cpu(cqe->rq.buftag_qpn) >> in ocrdma_discard_cqes()
1629 BUG_ON(wqe_idx < 1); in ocrdma_discard_cqes()
1632 ocrdma_srq_toggle_bit(qp->srq, wqe_idx - 1); in ocrdma_discard_cqes()
2402 u32 wqe_idx) in ocrdma_update_wc()
2408 hdr = ocrdma_hwq_head_from_idx(&qp->sq, wqe_idx); in ocrdma_update_wc()
2410 ibwc->wr_id = qp->wqe_wr_id_tbl[wqe_idx].wrid; in ocrdma_update_wc()
2563 u32 wqe_idx; in ocrdma_poll_success_scqe() local
2574 wqe_idx = (le32_to_cpu(cqe->wq.wqeidx) & in ocrdma_poll_success_scqe()
2576 if (tail != wqe_idx) in ocrdma_poll_success_scqe()
2401 ocrdma_update_wc(struct ocrdma_qp *qp, struct ib_wc *ibwc, u32 wqe_idx) ocrdma_update_wc() argument
2632 u32 wqe_idx; ocrdma_update_free_srq_cqe() local
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c1594 u32 qpn = 0, wqe_idx = 0; in ocrdma_discard_cqes() local
1627 wqe_idx = (le32_to_cpu(cqe->rq.buftag_qpn) >> in ocrdma_discard_cqes()
1630 BUG_ON(wqe_idx < 1); in ocrdma_discard_cqes()
1633 ocrdma_srq_toggle_bit(qp->srq, wqe_idx - 1); in ocrdma_discard_cqes()
2400 u32 wqe_idx) in ocrdma_update_wc()
2406 hdr = ocrdma_hwq_head_from_idx(&qp->sq, wqe_idx); in ocrdma_update_wc()
2408 ibwc->wr_id = qp->wqe_wr_id_tbl[wqe_idx].wrid; in ocrdma_update_wc()
2561 u32 wqe_idx; in ocrdma_poll_success_scqe() local
2572 wqe_idx = (le32_to_cpu(cqe->wq.wqeidx) & in ocrdma_poll_success_scqe()
2574 if (tail != wqe_idx) in ocrdma_poll_success_scqe()
2399 ocrdma_update_wc(struct ocrdma_qp *qp, struct ib_wc *ibwc, u32 wqe_idx) ocrdma_update_wc() argument
2630 u32 wqe_idx; ocrdma_update_free_srq_cqe() local
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/
H A Dhns_roce_hw_v2.c698 unsigned int wqe_idx; in hns_roce_v2_post_send() local
721 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v2_post_send()
731 wqe = hns_roce_get_send_wqe(qp, wqe_idx); in hns_roce_v2_post_send()
732 qp->sq.wrid[wqe_idx] = wr->wr_id; in hns_roce_v2_post_send()
805 u32 wqe_idx, u32 max_sge) in fill_rq_wqe()
809 wqe = hns_roce_get_recv_wqe(hr_qp, wqe_idx); in fill_rq_wqe()
820 u32 wqe_idx, nreq, max_sge; in hns_roce_v2_post_recv() local
850 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v2_post_recv()
851 fill_rq_wqe(hr_qp, wr, wqe_idx, max_sge); in hns_roce_v2_post_recv()
852 hr_qp->rq.wrid[wqe_idx] in hns_roce_v2_post_recv()
804 fill_rq_wqe(struct hns_roce_qp *hr_qp, const struct ib_recv_wr *wr, u32 wqe_idx, u32 max_sge) fill_rq_wqe() argument
916 get_srq_wqe_idx(struct hns_roce_srq *srq, u32 *wqe_idx) get_srq_wqe_idx() argument
930 fill_wqe_idx(struct hns_roce_srq *srq, unsigned int wqe_idx) fill_wqe_idx() argument
961 u32 wqe_idx; hns_roce_v2_post_srq_recv() local
3879 u16 wqe_idx; hns_roce_v2_poll_one() local
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
H A Dhns_roce_hw_v2.c651 unsigned int wqe_idx; in hns_roce_v2_post_send() local
674 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v2_post_send()
684 wqe = hns_roce_get_send_wqe(qp, wqe_idx); in hns_roce_v2_post_send()
685 qp->sq.wrid[wqe_idx] = wr->wr_id; in hns_roce_v2_post_send()
737 u32 wqe_idx; in hns_roce_v2_post_recv() local
759 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v2_post_recv()
769 wqe = hns_roce_get_recv_wqe(hr_qp, wqe_idx); in hns_roce_v2_post_recv()
785 sge_list = hr_qp->rq_inl_buf.wqe_list[wqe_idx].sg_list; in hns_roce_v2_post_recv()
786 hr_qp->rq_inl_buf.wqe_list[wqe_idx].sge_cnt = in hns_roce_v2_post_recv()
795 hr_qp->rq.wrid[wqe_idx] in hns_roce_v2_post_recv()
851 int wqe_idx; find_empty_entry() local
874 int wqe_idx; hns_roce_v2_post_srq_recv() local
[all...]
H A Dhns_roce_hw_v1.c79 u32 wqe_idx; in hns_roce_v1_post_send() local
99 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v1_post_send()
109 wqe = hns_roce_get_send_wqe(qp, wqe_idx); in hns_roce_v1_post_send()
110 qp->sq.wrid[wqe_idx] = wr->wr_id; in hns_roce_v1_post_send()
354 unsigned int wqe_idx; in hns_roce_v1_post_recv() local
370 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v1_post_recv()
380 ctrl = hns_roce_get_recv_wqe(hr_qp, wqe_idx); in hns_roce_v1_post_recv()
392 hr_qp->rq.wrid[wqe_idx] = wr->wr_id; in hns_roce_v1_post_recv()
/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c1722 u32 wqe_idx; in bnxt_qplib_post_send() local
1745 swq = bnxt_qplib_get_swqe(sq, &wqe_idx); in bnxt_qplib_post_send()
1934 bnxt_qplib_swq_mod_start(sq, wqe_idx); in bnxt_qplib_post_send()
1972 u32 wqe_idx; in bnxt_qplib_post_recv() local
1991 swq = bnxt_qplib_get_swqe(rq, &wqe_idx); in bnxt_qplib_post_recv()
2020 base_hdr->wr_id[0] = cpu_to_le32(wqe_idx); in bnxt_qplib_post_recv()
2022 bnxt_qplib_swq_mod_start(rq, wqe_idx); in bnxt_qplib_post_recv()
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c1752 u32 wqe_idx; in bnxt_qplib_post_send() local
1775 swq = bnxt_qplib_get_swqe(sq, &wqe_idx); in bnxt_qplib_post_send()
1964 bnxt_qplib_swq_mod_start(sq, wqe_idx); in bnxt_qplib_post_send()
2002 u32 wqe_idx; in bnxt_qplib_post_recv() local
2021 swq = bnxt_qplib_get_swqe(rq, &wqe_idx); in bnxt_qplib_post_recv()
2050 base_hdr->wr_id[0] = cpu_to_le32(wqe_idx); in bnxt_qplib_post_recv()
2052 bnxt_qplib_swq_mod_start(rq, wqe_idx); in bnxt_qplib_post_recv()

Completed in 59 milliseconds