/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 87 /* qp_context flags */ 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 568 struct mthca_qp_context *qp_context; in __mthca_modify_qp() local 578 qp_context = &qp_param->context; in __mthca_modify_qp() 581 qp_context->flags = cpu_to_be32((to_mthca_state(new_state) << 28) | in __mthca_modify_qp() 583 qp_context->flags |= cpu_to_be32(MTHCA_QP_BIT_DE); in __mthca_modify_qp() 585 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 590 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 593 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_REARM << 11); in __mthca_modify_qp() 596 qp_context in __mthca_modify_qp() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 87 /* qp_context flags */ 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 568 struct mthca_qp_context *qp_context; in __mthca_modify_qp() local 578 qp_context = &qp_param->context; in __mthca_modify_qp() 581 qp_context->flags = cpu_to_be32((to_mthca_state(new_state) << 28) | in __mthca_modify_qp() 583 qp_context->flags |= cpu_to_be32(MTHCA_QP_BIT_DE); in __mthca_modify_qp() 585 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 590 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 593 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_REARM << 11); in __mthca_modify_qp() 596 qp_context in __mthca_modify_qp() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | security.c | 174 if (sec->qp->event_handler && sec->qp->qp_context) { in qp_to_error() 177 sec->qp->qp_context); in qp_to_error() 185 if (qp->event_handler && qp->qp_context) { in qp_to_error() 189 qp->qp_context); in qp_to_error()
|
H A D | verbs.c | 1104 event->element.qp->event_handler(event, event->element.qp->qp_context); in __ib_shared_qp_event_handler() 1110 void *qp_context) in __ib_open_qp() 1131 qp->qp_context = qp_context; in __ib_open_qp() 1157 qp_open_attr->qp_context); in ib_open_qp() 1170 qp->qp_context = qp; in create_xrc_qp_user() 1179 qp_init_attr->qp_context); in create_xrc_qp_user() 1248 qp->qp_context = qp_init_attr->qp_context; in ib_create_qp() 1108 __ib_open_qp(struct ib_qp *real_qp, void (*event_handler)(struct ib_event *, void *), void *qp_context) __ib_open_qp() argument
|
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | security.c | 174 if (sec->qp->event_handler && sec->qp->qp_context) { in qp_to_error() 177 sec->qp->qp_context); in qp_to_error() 185 if (qp->event_handler && qp->qp_context) { in qp_to_error() 189 qp->qp_context); in qp_to_error()
|
H A D | verbs.c | 1109 event->element.qp->event_handler(event, event->element.qp->qp_context); in __ib_shared_qp_event_handler() 1115 void *qp_context) in __ib_open_qp() 1136 qp->qp_context = qp_context; in __ib_open_qp() 1162 qp_open_attr->qp_context); in ib_open_qp() 1175 qp->qp_context = qp; in create_xrc_qp_user() 1184 qp_init_attr->qp_context); in create_xrc_qp_user() 1223 qp->qp_context = attr->qp_context; in create_qp() 1113 __ib_open_qp(struct ib_qp *real_qp, void (*event_handler)(struct ib_event *, void *), void *qp_context) __ib_open_qp() argument
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 459 cmd->qp_context.pri_path.grh_mylmc = params->smac_index; in mlx4_update_qp() 474 cmd->qp_context.pri_path.fl |= in mlx4_update_qp() 482 cmd->qp_context.param3 |= cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_update_qp() 487 cmd->qp_context.rate_limit_params = cpu_to_be16((params->rate_unit << 14) | params->rate_val); in mlx4_update_qp() 498 cmd->qp_context.qos_vport = params->qos_vport; in mlx4_update_qp()
|
H A D | resource_tracker.c | 4324 smac_index = cmd->qp_context.pri_path.grh_mylmc; in mlx4_UPDATE_QP_wrapper() 5270 ctx->qp_context.qos_vport = work->qos_vport; in update_qos_vpp() 5361 upd_context->qp_context.param3 = qp->param3; in mlx4_vf_immed_vlan_work_handler() 5362 upd_context->qp_context.pri_path.vlan_control = qp->vlan_control; in mlx4_vf_immed_vlan_work_handler() 5363 upd_context->qp_context.pri_path.fvl_rx = qp->fvl_rx; in mlx4_vf_immed_vlan_work_handler() 5364 upd_context->qp_context.pri_path.vlan_index = qp->vlan_index; in mlx4_vf_immed_vlan_work_handler() 5365 upd_context->qp_context.pri_path.fl = qp->pri_path_fl; in mlx4_vf_immed_vlan_work_handler() 5366 upd_context->qp_context.pri_path.feup = qp->feup; in mlx4_vf_immed_vlan_work_handler() 5367 upd_context->qp_context.pri_path.sched_queue = in mlx4_vf_immed_vlan_work_handler() 5370 upd_context->qp_context in mlx4_vf_immed_vlan_work_handler() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 464 cmd->qp_context.pri_path.grh_mylmc = params->smac_index; in mlx4_update_qp() 479 cmd->qp_context.pri_path.fl |= in mlx4_update_qp() 487 cmd->qp_context.param3 |= cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_update_qp() 492 cmd->qp_context.rate_limit_params = cpu_to_be16((params->rate_unit << 14) | params->rate_val); in mlx4_update_qp() 503 cmd->qp_context.qos_vport = params->qos_vport; in mlx4_update_qp()
|
H A D | resource_tracker.c | 4325 smac_index = cmd->qp_context.pri_path.grh_mylmc; in mlx4_UPDATE_QP_wrapper() 5271 ctx->qp_context.qos_vport = work->qos_vport; in update_qos_vpp() 5362 upd_context->qp_context.param3 = qp->param3; in mlx4_vf_immed_vlan_work_handler() 5363 upd_context->qp_context.pri_path.vlan_control = qp->vlan_control; in mlx4_vf_immed_vlan_work_handler() 5364 upd_context->qp_context.pri_path.fvl_rx = qp->fvl_rx; in mlx4_vf_immed_vlan_work_handler() 5365 upd_context->qp_context.pri_path.vlan_index = qp->vlan_index; in mlx4_vf_immed_vlan_work_handler() 5366 upd_context->qp_context.pri_path.fl = qp->pri_path_fl; in mlx4_vf_immed_vlan_work_handler() 5367 upd_context->qp_context.pri_path.feup = qp->feup; in mlx4_vf_immed_vlan_work_handler() 5368 upd_context->qp_context.pri_path.sched_queue = in mlx4_vf_immed_vlan_work_handler() 5371 upd_context->qp_context in mlx4_vf_immed_vlan_work_handler() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 123 attr.qp_context = isert_conn; in isert_create_qp() 517 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connected_handler() 631 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_disconnected_handler() 658 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connect_error() 701 isert_conn = cma_id->qp->qp_context; in isert_cma_handler() 1321 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_recv_done() 1380 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_recv_done() 1578 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_write_done() 1620 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_read_done() 1696 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_send_done() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 122 attr.qp_context = isert_conn; in isert_create_qp() 518 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connected_handler() 632 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_disconnected_handler() 659 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connect_error() 702 isert_conn = cma_id->qp->qp_context; in isert_cma_handler() 1321 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_recv_done() 1380 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_recv_done() 1578 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_write_done() 1620 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_read_done() 1696 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_send_done() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.c | 493 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 498 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 505 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.c | 492 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 497 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 504 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | gsi.c | 226 .qp_context = gsi->rx_qp->qp_context, in create_gsi_ud_qp()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | gsi.c | 204 .qp_context = gsi->rx_qp->qp_context, in create_gsi_ud_qp()
|
/kernel/linux/linux-5.10/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 564 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 659 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_task_rsp() 734 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_dataout_comp()
|
/kernel/linux/linux-5.10/net/smc/ |
H A D | smc_wr.c | 87 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 367 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_wr_rx_demultiplex() 390 link = wc[i].qp->qp_context; in smc_wr_rx_process_cqes()
|
/kernel/linux/linux-6.6/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 535 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 636 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_task_rsp() 699 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_dataout_comp()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
|
/kernel/linux/linux-5.10/drivers/nvme/target/ |
H A D | rdma.c | 703 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_send_done() 756 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_read_data_done() 789 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_write_data_done() 1011 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_recv_done() 1279 qp_attr.qp_context = queue; in nvmet_rdma_create_queue_ib() 1708 * @queue: nvmet rdma queue (cm id qp_context) 1759 queue = cm_id->qp->qp_context; in nvmet_rdma_cm_handler()
|
/kernel/linux/linux-6.6/drivers/nvme/target/ |
H A D | rdma.c | 704 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_send_done() 757 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_read_data_done() 790 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_write_data_done() 1012 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_recv_done() 1279 qp_attr.qp_context = queue; in nvmet_rdma_create_queue_ib() 1708 * @queue: nvmet rdma queue (cm id qp_context) 1759 queue = cm_id->qp->qp_context; in nvmet_rdma_cm_handler()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
H A D | qib_ruc.c | 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp()
|
/kernel/linux/linux-5.10/include/linux/mlx4/ |
H A D | qp.h | 240 struct mlx4_qp_context qp_context; member
|