/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | cq.c | 173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local 180 cq_context = mailbox->buf; in mlx4_cq_modify() 181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify() 182 cq_context->cq_period = cpu_to_be16(period); in mlx4_cq_modify() 195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local 203 cq_context = mailbox->buf; in mlx4_cq_resize() 204 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); in mlx4_cq_resize() 205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize() 207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize() 208 cq_context in mlx4_cq_resize() 350 struct mlx4_cq_context *cq_context; mlx4_cq_alloc() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
H A D | cq.c | 173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local 180 cq_context = mailbox->buf; in mlx4_cq_modify() 181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify() 182 cq_context->cq_period = cpu_to_be16(period); in mlx4_cq_modify() 195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local 203 cq_context = mailbox->buf; in mlx4_cq_resize() 204 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); in mlx4_cq_resize() 205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize() 207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize() 208 cq_context in mlx4_cq_resize() 351 struct mlx4_cq_context *cq_context; mlx4_cq_alloc() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2_dfx.c | 12 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_query_cqc_info() local 20 cq_context = mailbox->buf; in hns_roce_v2_query_cqc_info() 29 memcpy(buffer, cq_context, sizeof(*cq_context)); in hns_roce_v2_query_cqc_info()
|
H A D | hns_roce_hw_v1.c | 799 free_mr->mr_free_cq->ib_cq.cq_context = NULL; in hns_roce_v1_rsv_lp_qp() 2071 struct hns_roce_cq_context *cq_context = mb_buf; in hns_roce_v1_write_cqc() local 2075 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v1_write_cqc() 2082 /* Register cq_context members */ in hns_roce_v1_write_cqc() 2083 roce_set_field(cq_context->cqc_byte_4, in hns_roce_v1_write_cqc() 2086 roce_set_field(cq_context->cqc_byte_4, CQ_CONTEXT_CQC_BYTE_4_CQN_M, in hns_roce_v1_write_cqc() 2089 cq_context->cq_bt_l = cpu_to_le32((u32)dma_handle); in hns_roce_v1_write_cqc() 2091 roce_set_field(cq_context->cqc_byte_12, in hns_roce_v1_write_cqc() 2095 roce_set_field(cq_context in hns_roce_v1_write_cqc() [all...] |
H A D | hns_roce_hw_v2.c | 3053 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local 3055 cq_context = mb_buf; in hns_roce_v2_write_cqc() 3056 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc() 3058 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_CQ_ST_M, in hns_roce_v2_write_cqc() 3060 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_ARM_ST_M, in hns_roce_v2_write_cqc() 3062 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_SHIFT_M, in hns_roce_v2_write_cqc() 3064 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_CEQN_M, in hns_roce_v2_write_cqc() 3067 roce_set_field(cq_context->byte_8_cqn, V2_CQC_BYTE_8_CQN_M, in hns_roce_v2_write_cqc() 3070 roce_set_field(cq_context in hns_roce_v2_write_cqc() 5329 struct hns_roce_v2_cq_context *cq_context; hns_roce_v2_modify_cq() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 773 struct mthca_cq_context *cq_context; in mthca_init_cq() local 811 cq_context = mailbox->buf; in mthca_init_cq() 824 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq() 825 cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK | in mthca_init_cq() 828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq() 830 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq() 832 cq_context in mthca_init_cq() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 773 struct mthca_cq_context *cq_context; in mthca_init_cq() local 811 cq_context = mailbox->buf; in mthca_init_cq() 824 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq() 825 cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK | in mthca_init_cq() 828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq() 830 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq() 832 cq_context in mthca_init_cq() [all...] |
/kernel/linux/linux-5.10/net/smc/ |
H A D | smc_wr.h | 119 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context); 128 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
H A D | smc_wr.c | 147 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_tx_cq_handler() argument 149 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_tx_cq_handler() 436 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_rx_cq_handler() argument 438 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_rx_cq_handler()
|
/kernel/linux/linux-6.6/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 105 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post() 119 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_cq_post()
|
/kernel/linux/linux-6.6/net/smc/ |
H A D | smc_wr.h | 131 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context); 136 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 54 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete() 118 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
|
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | uverbs_std_types_cq.c | 43 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq() 125 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_METHOD_CQ_CREATE()
|
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 370 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 455 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv() 477 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_wake() 584 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_done()
|
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 366 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 462 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv() 482 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_wake() 585 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_done()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | cq.c | 92 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), c_eqn); in mlx5_core_create_cq() 208 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | uverbs_std_types_cq.c | 43 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq() 125 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_METHOD_CQ_CREATE()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/ |
H A D | erdma_eq.c | 64 cq->ibcq.cq_context); in erdma_aeq_event_handler() 153 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in erdma_ceq_completion_handler()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | cq.c | 92 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), in mlx5_create_cq() 219 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 2592 free_mr->rsv_cq->ib_cq.cq_context = NULL; in free_mr_init_cq() 3541 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local 3543 cq_context = mb_buf; in hns_roce_v2_write_cqc() 3544 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc() 3546 hr_reg_write(cq_context, CQC_CQ_ST, V2_CQ_STATE_VALID); in hns_roce_v2_write_cqc() 3547 hr_reg_write(cq_context, CQC_ARM_ST, NO_ARMED); in hns_roce_v2_write_cqc() 3548 hr_reg_write(cq_context, CQC_SHIFT, ilog2(hr_cq->cq_depth)); in hns_roce_v2_write_cqc() 3549 hr_reg_write(cq_context, CQC_CEQN, hr_cq->vector); in hns_roce_v2_write_cqc() 3550 hr_reg_write(cq_context, CQC_CQ in hns_roce_v2_write_cqc() 5711 struct hns_roce_v2_cq_context *cq_context; hns_roce_v2_modify_cq() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 765 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user() 888 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel() 916 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler() 977 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq() 1315 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 786 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user() 913 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel() 941 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler() 1002 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq() 1349 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 116 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 182 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
|