Home
last modified time | relevance | path

Searched refs:mqp (Results 1 - 25 of 42) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
H A Dgsi.c47 static void generate_completions(struct mlx5_ib_qp *mqp) in generate_completions() argument
49 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; in generate_completions()
50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions()
73 struct mlx5_ib_qp *mqp = container_of(gsi, struct mlx5_ib_qp, gsi); in handle_single_completion() local
82 wr->wc.qp = &mqp->ibqp; in handle_single_completion()
84 generate_completions(mqp); in handle_single_completion()
88 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp, in mlx5_ib_create_gsi() argument
106 gsi = &mqp->gsi; in mlx5_ib_create_gsi()
185 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp) in mlx5_ib_destroy_gsi() argument
187 struct mlx5_ib_dev *dev = to_mdev(mqp in mlx5_ib_destroy_gsi()
286 struct mlx5_ib_qp *mqp; setup_qp() local
357 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_modify_qp() local
378 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_query_qp() local
388 mlx5_ib_add_outstanding_wr(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr, struct ib_wc *wc) mlx5_ib_add_outstanding_wr() argument
420 mlx5_ib_gsi_silent_drop(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr) mlx5_ib_gsi_silent_drop() argument
461 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_post_send() local
506 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_post_recv() local
[all...]
H A Dqp.c1257 err = mlx5_core_create_sq_tracked(dev, in, inlen, &sq->base.mqp); in create_raw_packet_qp_sq()
1277 mlx5_core_destroy_sq_tracked(dev, &sq->base.mqp); in destroy_raw_packet_qp_sq()
1300 struct mlx5_ib_qp *mqp = rq->base.container_mibqp; in create_raw_packet_qp_rq() local
1329 if (mqp->flags & IB_QP_CREATE_SCATTER_FCS) in create_raw_packet_qp_rq()
1347 err = mlx5_core_create_rq_tracked(dev, in, inlen, &rq->base.mqp); in create_raw_packet_qp_rq()
1357 mlx5_core_destroy_rq_tracked(dev, &rq->base.mqp); in destroy_raw_packet_qp_rq()
1390 MLX5_SET(tirc, tirc, inline_rqn, rq->base.mqp.qpn); in create_raw_packet_qp_tir()
1451 resp->sqn = sq->base.mqp.qpn; in create_raw_packet_qp()
1456 sq->base.mqp.event = mlx5_ib_qp_event; in create_raw_packet_qp()
1476 resp->rqn = rq->base.mqp in create_raw_packet_qp()
2850 mlx5_ib_destroy_dct(struct mlx5_ib_qp *mqp) mlx5_ib_destroy_dct() argument
3020 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_destroy_qp() local
3689 struct mlx5_ib_qp *mqp = to_mqp(qp); get_tx_affinity() local
3746 struct mlx5_ib_qp *mqp = to_mqp(qp); __mlx5_ib_qp_set_counter() local
4670 mlx5_ib_dct_query_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *mqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) mlx5_ib_dct_query_qp() argument
5477 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_qp_set_counter() local
[all...]
H A Dqpc.c193 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct()
206 destroy_resource_common(dev, &dct->mqp); in _mlx5_core_destroy_dct()
217 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_create_dct()
275 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_drain_dct()
513 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_dct_query()
H A Dmlx5_ib.h335 struct mlx5_core_qp mqp; member
1058 static inline struct mlx5_ib_qp *to_mibqp(struct mlx5_core_qp *mqp) in to_mibqp() argument
1060 return container_of(mqp, struct mlx5_ib_qp_base, mqp)->container_mibqp; in to_mibqp()
1342 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp,
1344 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp);
H A Dfs.c1151 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_create_flow() local
1238 if (mqp->is_rss) in mlx5_ib_create_flow()
1239 dst->tir_num = mqp->rss_qp.tirn; in mlx5_ib_create_flow()
1241 dst->tir_num = mqp->raw_packet_qp.rq.tirn; in mlx5_ib_create_flow()
1246 underlay_qpn = (mqp->flags & IB_QP_CREATE_SOURCE_QPN) ? in mlx5_ib_create_flow()
1247 mqp->underlay_qpn : in mlx5_ib_create_flow()
1872 struct mlx5_ib_qp *mqp; in get_dests() local
1882 mqp = to_mqp(*qp); in get_dests()
1883 if (mqp->is_rss) in get_dests()
1884 *dest_id = mqp in get_dests()
[all...]
H A Dib_rep.c155 sq->base.mqp.qpn); in create_flow_rule_vport_sq()
H A Dmain.c2592 struct mlx5_ib_qp *mqp = to_mqp(ibqp); in mlx5_ib_mcg_attach() local
2599 if (mqp->flags & IB_QP_CREATE_SOURCE_QPN) { in mlx5_ib_mcg_attach()
2716 struct mlx5_ib_qp *mqp; in mlx5_ib_handle_internal_error() local
2728 list_for_each_entry(mqp, &ibdev->qp_list, qps_list) { in mlx5_ib_handle_internal_error()
2729 spin_lock_irqsave(&mqp->sq.lock, flags_qp); in mlx5_ib_handle_internal_error()
2730 if (mqp->sq.tail != mqp->sq.head) { in mlx5_ib_handle_internal_error()
2731 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx5_ib_handle_internal_error()
2734 mqp->ibqp.send_cq->comp_handler) { in mlx5_ib_handle_internal_error()
2743 spin_unlock_irqrestore(&mqp in mlx5_ib_handle_internal_error()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/
H A Dgsi.c47 static void generate_completions(struct mlx5_ib_qp *mqp) in generate_completions() argument
49 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; in generate_completions()
50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions()
73 struct mlx5_ib_qp *mqp = container_of(gsi, struct mlx5_ib_qp, gsi); in handle_single_completion() local
82 wr->wc.qp = &mqp->ibqp; in handle_single_completion()
84 generate_completions(mqp); in handle_single_completion()
88 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp, in mlx5_ib_create_gsi() argument
106 gsi = &mqp->gsi; in mlx5_ib_create_gsi()
168 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp) in mlx5_ib_destroy_gsi() argument
170 struct mlx5_ib_dev *dev = to_mdev(mqp in mlx5_ib_destroy_gsi()
264 struct mlx5_ib_qp *mqp; setup_qp() local
324 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_modify_qp() local
349 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_query_qp() local
359 mlx5_ib_add_outstanding_wr(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr, struct ib_wc *wc) mlx5_ib_add_outstanding_wr() argument
391 mlx5_ib_gsi_silent_drop(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr) mlx5_ib_gsi_silent_drop() argument
432 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_post_send() local
477 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_gsi_post_recv() local
[all...]
H A Dqp.c334 err = mlx5_core_qp_query(dev, &qp->trans_qp.base.mqp, outb, outlen, in mlx5_ib_qp_err_syndrome()
1408 err = mlx5_core_create_sq_tracked(dev, in, inlen, &sq->base.mqp); in create_raw_packet_qp_sq()
1428 mlx5_core_destroy_sq_tracked(dev, &sq->base.mqp); in destroy_raw_packet_qp_sq()
1436 struct mlx5_ib_qp *mqp = rq->base.container_mibqp; in create_raw_packet_qp_rq() local
1477 if (mqp->flags & IB_QP_CREATE_SCATTER_FCS) in create_raw_packet_qp_rq()
1495 err = mlx5_core_create_rq_tracked(dev, in, inlen, &rq->base.mqp); in create_raw_packet_qp_rq()
1505 mlx5_core_destroy_rq_tracked(dev, &rq->base.mqp); in destroy_raw_packet_qp_rq()
1538 MLX5_SET(tirc, tirc, inline_rqn, rq->base.mqp.qpn); in create_raw_packet_qp_tir()
1600 resp->sqn = sq->base.mqp.qpn; in create_raw_packet_qp()
1605 sq->base.mqp in create_raw_packet_qp()
3188 mlx5_ib_destroy_dct(struct mlx5_ib_qp *mqp) mlx5_ib_destroy_dct() argument
3339 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_destroy_qp() local
4020 struct mlx5_ib_qp *mqp = to_mqp(qp); get_tx_affinity() local
4077 struct mlx5_ib_qp *mqp = to_mqp(qp); __mlx5_ib_qp_set_counter() local
4997 mlx5_ib_dct_query_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *mqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) mlx5_ib_dct_query_qp() argument
5809 struct mlx5_ib_qp *mqp = to_mqp(qp); mlx5_ib_qp_set_counter() local
[all...]
H A Dqpc.c199 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct()
210 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_create_dct()
268 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_drain_dct()
293 tmp = xa_cmpxchg_irq(&table->dct_xa, dct->mqp.qpn, dct, XA_ZERO_ENTRY, GFP_KERNEL); in mlx5_core_destroy_dct()
299 xa_cmpxchg_irq(&table->dct_xa, dct->mqp.qpn, XA_ZERO_ENTRY, dct, 0); in mlx5_core_destroy_dct()
302 xa_erase_irq(&table->dct_xa, dct->mqp.qpn); in mlx5_core_destroy_dct()
539 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_dct_query()
H A Dmlx5_ib.h426 struct mlx5_core_qp mqp; member
1197 static inline struct mlx5_ib_qp *to_mibqp(struct mlx5_core_qp *mqp) in to_mibqp() argument
1199 return container_of(mqp, struct mlx5_ib_qp_base, mqp)->container_mibqp; in to_mibqp()
1471 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp,
1473 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp);
H A Dfs.c1280 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_create_flow() local
1366 if (mqp->is_rss) in mlx5_ib_create_flow()
1367 dst->tir_num = mqp->rss_qp.tirn; in mlx5_ib_create_flow()
1369 dst->tir_num = mqp->raw_packet_qp.rq.tirn; in mlx5_ib_create_flow()
1374 underlay_qpn = (mqp->flags & IB_QP_CREATE_SOURCE_QPN) ? in mlx5_ib_create_flow()
1375 mqp->underlay_qpn : in mlx5_ib_create_flow()
1851 struct mlx5_ib_qp *mqp; in get_dests() local
1861 mqp = to_mqp(*qp); in get_dests()
1862 if (mqp->is_rss) in get_dests()
1863 *dest_id = mqp in get_dests()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/
H A Dmcast.c77 struct rvt_mcast_qp *mqp; in rvt_mcast_qp_alloc() local
79 mqp = kmalloc(sizeof(*mqp), GFP_KERNEL); in rvt_mcast_qp_alloc()
80 if (!mqp) in rvt_mcast_qp_alloc()
83 mqp->qp = qp; in rvt_mcast_qp_alloc()
87 return mqp; in rvt_mcast_qp_alloc()
90 static void rvt_mcast_qp_free(struct rvt_mcast_qp *mqp) in rvt_mcast_qp_free() argument
92 struct rvt_qp *qp = mqp->qp; in rvt_mcast_qp_free()
97 kfree(mqp); in rvt_mcast_qp_free()
186 * @mqp
193 rvt_mcast_add(struct rvt_dev_info *rdi, struct rvt_ibport *ibp, struct rvt_mcast *mcast, struct rvt_mcast_qp *mqp) rvt_mcast_add() argument
286 struct rvt_mcast_qp *mqp; rvt_attach_mcast() local
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/
H A Dmcast.c35 struct rvt_mcast_qp *mqp; in rvt_mcast_qp_alloc() local
37 mqp = kmalloc(sizeof(*mqp), GFP_KERNEL); in rvt_mcast_qp_alloc()
38 if (!mqp) in rvt_mcast_qp_alloc()
41 mqp->qp = qp; in rvt_mcast_qp_alloc()
45 return mqp; in rvt_mcast_qp_alloc()
48 static void rvt_mcast_qp_free(struct rvt_mcast_qp *mqp) in rvt_mcast_qp_free() argument
50 struct rvt_qp *qp = mqp->qp; in rvt_mcast_qp_free()
55 kfree(mqp); in rvt_mcast_qp_free()
144 * @mqp
151 rvt_mcast_add(struct rvt_dev_info *rdi, struct rvt_ibport *ibp, struct rvt_mcast *mcast, struct rvt_mcast_qp *mqp) rvt_mcast_add() argument
244 struct rvt_mcast_qp *mqp; rvt_attach_mcast() local
[all...]
/kernel/linux/linux-5.10/drivers/scsi/arm/
H A Dmsgqueue.c121 struct msgqueue_entry **mqp; in msgqueue_addmsg() local
133 mqp = &msgq->qe; in msgqueue_addmsg()
134 while (*mqp) in msgqueue_addmsg()
135 mqp = &(*mqp)->next; in msgqueue_addmsg()
137 *mqp = mq; in msgqueue_addmsg()
/kernel/linux/linux-6.6/drivers/scsi/arm/
H A Dmsgqueue.c121 struct msgqueue_entry **mqp; in msgqueue_addmsg() local
133 mqp = &msgq->qe; in msgqueue_addmsg()
134 while (*mqp) in msgqueue_addmsg()
135 mqp = &(*mqp)->next; in msgqueue_addmsg()
137 *mqp = mq; in msgqueue_addmsg()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/
H A Dmain.c1267 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in add_gid_entry() local
1276 if (mlx4_ib_add_mc(mdev, mqp, gid)) { in add_gid_entry()
1277 ge->port = mqp->port; in add_gid_entry()
1281 mutex_lock(&mqp->mutex); in add_gid_entry()
1282 list_add_tail(&ge->list, &mqp->gid_list); in add_gid_entry()
1283 mutex_unlock(&mqp->mutex); in add_gid_entry()
1304 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, in mlx4_ib_add_mc() argument
1310 if (!mqp->port) in mlx4_ib_add_mc()
1314 ndev = mdev->iboe.netdevs[mqp->port - 1]; in mlx4_ib_add_mc()
1845 struct mlx4_ib_qp *mqp in mlx4_ib_mcg_attach() local
1923 struct mlx4_ib_qp *mqp = to_mqp(ibqp); mlx4_ib_mcg_detach() local
2949 mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, int is_attach) mlx4_ib_steer_qp_reg() argument
3073 struct mlx4_ib_qp *mqp; mlx4_ib_handle_catas_error() local
[all...]
H A Dqp.c110 return qp->mqp.qpn >= dev->dev->phys_caps.base_tunnel_sqpn && in is_tunnel_qp()
111 qp->mqp.qpn < dev->dev->phys_caps.base_tunnel_sqpn + in is_tunnel_qp()
122 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_sqp()
123 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 3); in is_sqp()
129 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy || in is_sqp()
130 qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp1_proxy) { in is_sqp()
150 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_qp0()
151 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 1); in is_qp0()
157 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy) { in is_qp0()
627 qp->mqp in create_qp_rss()
1628 struct mlx4_ib_qp *mqp = to_mqp(qp); _mlx4_ib_destroy_qp() local
1657 struct mlx4_ib_qp *mqp = to_mqp(qp); mlx4_ib_destroy_qp() local
1857 mlx4_set_path(struct mlx4_ib_dev *dev, const struct ib_qp_attr *qp, enum ib_qp_attr_mask qp_attr_mask, struct mlx4_ib_qp *mqp, struct mlx4_qp_path *path, u8 port, u16 vlan_id, u8 *smac) mlx4_set_path() argument
1869 mlx4_set_alt_path(struct mlx4_ib_dev *dev, const struct ib_qp_attr *qp, enum ib_qp_attr_mask qp_attr_mask, struct mlx4_ib_qp *mqp, struct mlx4_qp_path *path, u8 port) mlx4_set_alt_path() argument
2793 struct mlx4_ib_qp *mqp = to_mqp(ibqp); mlx4_ib_modify_qp() local
[all...]
H A Dmlx4_ib.h327 struct mlx4_qp mqp; member
718 static inline struct mlx4_ib_qp *to_mibqp(struct mlx4_qp *mqp) in to_mibqp() argument
720 return container_of(mqp, struct mlx4_ib_qp, mqp); in to_mibqp()
851 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
909 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/
H A Dmain.c1267 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in add_gid_entry() local
1276 if (mlx4_ib_add_mc(mdev, mqp, gid)) { in add_gid_entry()
1277 ge->port = mqp->port; in add_gid_entry()
1281 mutex_lock(&mqp->mutex); in add_gid_entry()
1282 list_add_tail(&ge->list, &mqp->gid_list); in add_gid_entry()
1283 mutex_unlock(&mqp->mutex); in add_gid_entry()
1304 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, in mlx4_ib_add_mc() argument
1310 if (!mqp->port) in mlx4_ib_add_mc()
1314 ndev = mdev->iboe.netdevs[mqp->port - 1]; in mlx4_ib_add_mc()
1841 struct mlx4_ib_qp *mqp in mlx4_ib_mcg_attach() local
1919 struct mlx4_ib_qp *mqp = to_mqp(ibqp); mlx4_ib_mcg_detach() local
2943 mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, int is_attach) mlx4_ib_steer_qp_reg() argument
3071 struct mlx4_ib_qp *mqp; mlx4_ib_handle_catas_error() local
[all...]
H A Dqp.c118 return qp->mqp.qpn >= dev->dev->phys_caps.base_tunnel_sqpn && in is_tunnel_qp()
119 qp->mqp.qpn < dev->dev->phys_caps.base_tunnel_sqpn + in is_tunnel_qp()
130 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_sqp()
131 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 3); in is_sqp()
137 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy || in is_sqp()
138 qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp1_proxy) { in is_sqp()
158 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_qp0()
159 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 1); in is_qp0()
165 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy) { in is_qp0()
662 qp->mqp in create_qp_rss()
1663 struct mlx4_ib_qp *mqp = to_mqp(qp); _mlx4_ib_destroy_qp() local
1690 struct mlx4_ib_qp *mqp = to_mqp(qp); mlx4_ib_destroy_qp() local
1890 mlx4_set_path(struct mlx4_ib_dev *dev, const struct ib_qp_attr *qp, enum ib_qp_attr_mask qp_attr_mask, struct mlx4_ib_qp *mqp, struct mlx4_qp_path *path, u8 port, u16 vlan_id, u8 *smac) mlx4_set_path() argument
1902 mlx4_set_alt_path(struct mlx4_ib_dev *dev, const struct ib_qp_attr *qp, enum ib_qp_attr_mask qp_attr_mask, struct mlx4_ib_qp *mqp, struct mlx4_qp_path *path, u8 port) mlx4_set_alt_path() argument
2826 struct mlx4_ib_qp *mqp = to_mqp(ibqp); mlx4_ib_modify_qp() local
[all...]
H A Dmlx4_ib.h328 struct mlx4_qp mqp; member
720 static inline struct mlx4_ib_qp *to_mibqp(struct mlx4_qp *mqp) in to_mibqp() argument
722 return container_of(mqp, struct mlx4_ib_qp, mqp); in to_mibqp()
852 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
910 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
/kernel/linux/linux-5.10/arch/sparc/kernel/
H A Dpci_msi.c335 } *mqp; in sparc64_pbm_msi_init() local
355 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
357 if (!mqp) in sparc64_pbm_msi_init()
358 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
360 if (!mqp || len != sizeof(struct msiq_prop)) in sparc64_pbm_msi_init()
363 pbm->msiq_first = mqp->first_msiq; in sparc64_pbm_msi_init()
364 pbm->msiq_first_devino = mqp->first_devino; in sparc64_pbm_msi_init()
/kernel/linux/linux-6.6/arch/sparc/kernel/
H A Dpci_msi.c337 } *mqp; in sparc64_pbm_msi_init() local
357 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
359 if (!mqp) in sparc64_pbm_msi_init()
360 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
362 if (!mqp || len != sizeof(struct msiq_prop)) in sparc64_pbm_msi_init()
365 pbm->msiq_first = mqp->first_msiq; in sparc64_pbm_msi_init()
366 pbm->msiq_first_devino = mqp->first_devino; in sparc64_pbm_msi_init()
/kernel/linux/linux-5.10/drivers/vdpa/mlx5/net/
H A Dmlx5_vnet.c68 struct mlx5_core_qp mqp; member
417 vqp->mqp.uid = ndev->mvdev.res.uid; in qp_create()
418 vqp->mqp.qpn = MLX5_GET(create_qp_out, out, qpn); in qp_create()
440 MLX5_SET(destroy_qp_in, in, qpn, vqp->mqp.qpn); in qp_destroy()
443 mlx5_vdpa_warn(&ndev->mvdev, "destroy qp 0x%x\n", vqp->mqp.qpn); in qp_destroy()
800 MLX5_SET(virtio_q, vq_ctx, event_qpn_or_msix, mvq->fwqp.mqp.qpn); in create_virtqueue()
852 return fw ? mvq->vqqp.mqp.qpn : mvq->fwqp.mqp.qpn; in get_rqpn()
857 return fw ? mvq->fwqp.mqp.qpn : mvq->vqqp.mqp in get_qpn()
[all...]

Completed in 47 milliseconds

12