/kernel/linux/linux-6.6/drivers/pmdomain/imx/ |
H A D | scu-pd.c | 404 struct imx_sc_pm_domain *sc_pd; in imx_scu_pd_xlate() local 406 sc_pd = to_imx_sc_pd(pd_data->domains[i]); in imx_scu_pd_xlate() 407 if (sc_pd->rsrc == spec->args[0]) { in imx_scu_pd_xlate() 408 domain = &sc_pd->pd; in imx_scu_pd_xlate() 420 struct imx_sc_pm_domain *sc_pd; in imx_scu_add_pm_domain() local 427 sc_pd = devm_kzalloc(dev, sizeof(*sc_pd), GFP_KERNEL); in imx_scu_add_pm_domain() 428 if (!sc_pd) in imx_scu_add_pm_domain() 431 sc_pd->rsrc = pd_ranges->rsrc + idx; in imx_scu_add_pm_domain() 432 sc_pd in imx_scu_add_pm_domain() 479 struct imx_sc_pm_domain *sc_pd; imx_scu_init_pm_domains() local [all...] |
/kernel/linux/linux-5.10/drivers/firmware/imx/ |
H A D | scu-pd.c | 241 struct imx_sc_pm_domain *sc_pd; in imx_scu_pd_xlate() local 243 sc_pd = to_imx_sc_pd(pd_data->domains[i]); in imx_scu_pd_xlate() 244 if (sc_pd->rsrc == spec->args[0]) { in imx_scu_pd_xlate() 245 domain = &sc_pd->pd; in imx_scu_pd_xlate() 257 struct imx_sc_pm_domain *sc_pd; in imx_scu_add_pm_domain() local 263 sc_pd = devm_kzalloc(dev, sizeof(*sc_pd), GFP_KERNEL); in imx_scu_add_pm_domain() 264 if (!sc_pd) in imx_scu_add_pm_domain() 267 sc_pd->rsrc = pd_ranges->rsrc + idx; in imx_scu_add_pm_domain() 268 sc_pd in imx_scu_add_pm_domain() 305 struct imx_sc_pm_domain *sc_pd; imx_scu_init_pm_domains() local [all...] |
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 426 newxprt->sc_pd = ib_alloc_pd(dev, 0); in svc_rdma_accept() 427 if (IS_ERR(newxprt->sc_pd)) { in svc_rdma_accept() 428 trace_svcrdma_pd_err(newxprt, PTR_ERR(newxprt->sc_pd)); in svc_rdma_accept() 453 dprintk("svcrdma: newxprt->sc_cm_id=%p, newxprt->sc_pd=%p\n", in svc_rdma_accept() 454 newxprt->sc_cm_id, newxprt->sc_pd); in svc_rdma_accept() 460 ret = rdma_create_qp(newxprt->sc_cm_id, newxprt->sc_pd, &qp_attr); in svc_rdma_accept() 573 if (rdma->sc_pd && !IS_ERR(rdma->sc_pd)) in __svc_rdma_free() 574 ib_dealloc_pd(rdma->sc_pd); in __svc_rdma_free()
|
H A D | svc_rdma_sendto.c | 149 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_send_ctxt_alloc() 151 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_send_ctxt_alloc() 167 ctxt->sc_sges[i].lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_send_ctxt_alloc() 189 ib_dma_unmap_single(rdma->sc_pd->device, in svc_rdma_send_ctxts_destroy() 312 ib_dma_sync_single_for_device(rdma->sc_pd->device, in svc_rdma_send()
|
H A D | svc_rdma_recvfrom.c | 140 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_recv_ctxt_alloc() 142 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_recv_ctxt_alloc() 154 ctxt->rc_recv_sge.lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_recv_ctxt_alloc() 170 ib_dma_unmap_single(rdma->sc_pd->device, ctxt->rc_recv_sge.addr, in svc_rdma_recv_ctxt_destroy() 334 ib_dma_sync_single_for_cpu(rdma->sc_pd->device, in svc_rdma_wc_receive()
|
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 427 newxprt->sc_pd = ib_alloc_pd(dev, 0); in svc_rdma_accept() 428 if (IS_ERR(newxprt->sc_pd)) { in svc_rdma_accept() 429 trace_svcrdma_pd_err(newxprt, PTR_ERR(newxprt->sc_pd)); in svc_rdma_accept() 454 dprintk("svcrdma: newxprt->sc_cm_id=%p, newxprt->sc_pd=%p\n", in svc_rdma_accept() 455 newxprt->sc_cm_id, newxprt->sc_pd); in svc_rdma_accept() 461 ret = rdma_create_qp(newxprt->sc_cm_id, newxprt->sc_pd, &qp_attr); in svc_rdma_accept() 567 if (rdma->sc_pd && !IS_ERR(rdma->sc_pd)) in __svc_rdma_free() 568 ib_dealloc_pd(rdma->sc_pd); in __svc_rdma_free()
|
H A D | svc_rdma_recvfrom.c | 139 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_recv_ctxt_alloc() 141 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_recv_ctxt_alloc() 157 ctxt->rc_recv_sge.lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_recv_ctxt_alloc() 172 ib_dma_unmap_single(rdma->sc_pd->device, ctxt->rc_recv_sge.addr, in svc_rdma_recv_ctxt_destroy() 817 ib_dma_sync_single_for_cpu(rdma_xprt->sc_pd->device, in svc_rdma_recvfrom()
|
H A D | svc_rdma_sendto.c | 139 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_send_ctxt_alloc() 141 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_send_ctxt_alloc() 157 ctxt->sc_sges[i].lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_send_ctxt_alloc() 180 ib_dma_unmap_single(rdma->sc_pd->device, in svc_rdma_send_ctxts_destroy() 321 ib_dma_sync_single_for_device(rdma->sc_pd->device, in svc_rdma_send()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_verbs.h | 50 struct i40iw_sc_pd sc_pd; member
|
H A D | i40iw_puda.h | 117 struct i40iw_sc_pd sc_pd; member
|
H A D | i40iw_verbs.c | 265 struct i40iw_sc_pd *sc_pd; in i40iw_alloc_pd() local 279 sc_pd = &iwpd->sc_pd; in i40iw_alloc_pd() 284 dev->iw_pd_ops->pd_init(dev, sc_pd, pd_id, ucontext->abi_ver); in i40iw_alloc_pd() 292 dev->iw_pd_ops->pd_init(dev, sc_pd, pd_id, -1); in i40iw_alloc_pd() 601 init_info.pd = &iwpd->sc_pd; in i40iw_create_qp() 1513 info->pd_id = iwpd->sc_pd.pd_id; in i40iw_hw_alloc_stag() 1686 stag_info->pd_id = iwpd->sc_pd.pd_id; in i40iw_hwreg_mr() 2010 info->pd_id = cpu_to_le32(iwpd->sc_pd.pd_id & 0x00007fff); in i40iw_dereg_mr()
|
H A D | i40iw_puda.c | 577 qp->pd = &rsrc->sc_pd; in i40iw_puda_qp_create() 934 dev->iw_pd_ops->pd_init(dev, &rsrc->sc_pd, info->pd_id, -1); in i40iw_puda_create_rsrc()
|
H A D | i40iw_utils.c | 596 i40iw_free_resource(iwdev, iwdev->allocated_pds, iwpd->sc_pd.pd_id); in i40iw_rem_pdusecount()
|
H A D | i40iw_cm.c | 3404 iwarp_info->pd_id = iwqp->iwpd->sc_pd.pd_id; in i40iw_cm_init_tsa_conn()
|
/kernel/linux/linux-5.10/include/linux/sunrpc/ |
H A D | svc_rdma.h | 92 struct ib_pd *sc_pd; member
|
/kernel/linux/linux-6.6/include/linux/sunrpc/ |
H A D | svc_rdma.h | 90 struct ib_pd *sc_pd; member
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | verbs.h | 27 struct irdma_sc_pd sc_pd; member
|
H A D | puda.h | 102 struct irdma_sc_pd sc_pd; member
|
H A D | verbs.c | 380 struct irdma_sc_pd *sc_pd; in irdma_alloc_pd() local 392 sc_pd = &iwpd->sc_pd; in irdma_alloc_pd() 397 irdma_sc_pd_init(dev, sc_pd, pd_id, ucontext->abi_ver); in irdma_alloc_pd() 405 irdma_sc_pd_init(dev, sc_pd, pd_id, IRDMA_ABI_VER); in irdma_alloc_pd() 425 irdma_free_rsrc(iwdev->rf, iwdev->rf->allocated_pds, iwpd->sc_pd.pd_id); in irdma_dealloc_pd() 945 init_info.pd = &iwpd->sc_pd; in irdma_create_qp() 954 init_info.qp_uk_init_info.abi_ver = iwpd->sc_pd.abi_ver; in irdma_create_qp() 1248 ctx_info->roce_info->pd_id = iwpd->sc_pd.pd_id; in irdma_modify_qp_roce() 2557 info->pd_id = iwpd->sc_pd in irdma_hw_alloc_mw() [all...] |
H A D | puda.c | 666 qp->pd = &rsrc->sc_pd; in irdma_puda_qp_create() 1057 irdma_sc_pd_init(dev, &rsrc->sc_pd, info->pd_id, info->abi_ver); in irdma_puda_create_rsrc()
|
H A D | cm.c | 3406 iwarp_info->pd_id = iwqp->iwpd->sc_pd.pd_id; in irdma_cm_init_tsa_conn()
|