/kernel/linux/linux-5.10/net/rds/ |
H A D | ib_frmr.c | 36 rds_transition_frwr_state(struct rds_ib_mr *ibmr, in rds_transition_frwr_state() argument 40 if (cmpxchg(&ibmr->u.frmr.fr_state, in rds_transition_frwr_state() 43 /* enforce order of ibmr->u.frmr.fr_state update in rds_transition_frwr_state() 47 atomic_dec(&ibmr->ic->i_fastreg_inuse_count); in rds_transition_frwr_state() 57 struct rds_ib_mr *ibmr = NULL; in rds_ib_alloc_frmr() local 66 ibmr = rds_ib_try_reuse_ibmr(pool); in rds_ib_alloc_frmr() 67 if (ibmr) in rds_ib_alloc_frmr() 68 return ibmr; in rds_ib_alloc_frmr() 70 ibmr = kzalloc_node(sizeof(*ibmr), GFP_KERNE in rds_ib_alloc_frmr() 106 rds_ib_free_frmr(struct rds_ib_mr *ibmr, bool drop) rds_ib_free_frmr() argument 123 rds_ib_post_reg_frmr(struct rds_ib_mr *ibmr) rds_ib_post_reg_frmr() argument 187 rds_ib_map_frmr(struct rds_ib_device *rds_ibdev, struct rds_ib_mr_pool *pool, struct rds_ib_mr *ibmr, struct scatterlist *sg, unsigned int sg_len) rds_ib_map_frmr() argument 266 rds_ib_post_inv(struct rds_ib_mr *ibmr) rds_ib_post_inv() argument 324 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; rds_ib_mr_cqe_handler() local 360 struct rds_ib_mr *ibmr, *next; rds_ib_unreg_frmr() local 406 struct rds_ib_mr *ibmr = NULL; rds_ib_reg_frmr() local 437 rds_ib_free_frmr_list(struct rds_ib_mr *ibmr) rds_ib_free_frmr_list() argument [all...] |
H A D | ib_rdma.c | 200 struct rds_ib_mr *ibmr = NULL; in rds_ib_reuse_mr() local 208 ibmr = llist_entry(ret, struct rds_ib_mr, llnode); in rds_ib_reuse_mr() 215 return ibmr; in rds_ib_reuse_mr() 220 struct rds_ib_mr *ibmr = trans_private; in rds_ib_sync_mr() local 221 struct rds_ib_device *rds_ibdev = ibmr->device; in rds_ib_sync_mr() 223 if (ibmr->odp) in rds_ib_sync_mr() 228 ib_dma_sync_sg_for_cpu(rds_ibdev->dev, ibmr->sg, in rds_ib_sync_mr() 229 ibmr->sg_dma_len, DMA_BIDIRECTIONAL); in rds_ib_sync_mr() 232 ib_dma_sync_sg_for_device(rds_ibdev->dev, ibmr->sg, in rds_ib_sync_mr() 233 ibmr in rds_ib_sync_mr() 238 __rds_ib_teardown_mr(struct rds_ib_mr *ibmr) __rds_ib_teardown_mr() argument 269 rds_ib_teardown_mr(struct rds_ib_mr *ibmr) rds_ib_teardown_mr() argument 298 struct rds_ib_mr *ibmr; llist_append_to_list() local 323 struct rds_ib_mr *ibmr; list_to_llist_nodes() local 345 struct rds_ib_mr *ibmr; rds_ib_flush_mr_pool() local 442 struct rds_ib_mr *ibmr = NULL; rds_ib_try_reuse_ibmr() local 486 struct rds_ib_mr *ibmr = trans_private; rds_ib_free_mr() local 545 struct rds_ib_mr *ibmr = trans_private; rds_ib_get_lkey() local 556 struct rds_ib_mr *ibmr = NULL; rds_ib_get_mr() local 696 struct rds_ib_mr *ibmr; rds_ib_odp_mr_worker() local [all...] |
/kernel/linux/linux-6.6/net/rds/ |
H A D | ib_frmr.c | 36 rds_transition_frwr_state(struct rds_ib_mr *ibmr, in rds_transition_frwr_state() argument 40 if (cmpxchg(&ibmr->u.frmr.fr_state, in rds_transition_frwr_state() 43 /* enforce order of ibmr->u.frmr.fr_state update in rds_transition_frwr_state() 47 atomic_dec(&ibmr->ic->i_fastreg_inuse_count); in rds_transition_frwr_state() 57 struct rds_ib_mr *ibmr = NULL; in rds_ib_alloc_frmr() local 66 ibmr = rds_ib_try_reuse_ibmr(pool); in rds_ib_alloc_frmr() 67 if (ibmr) in rds_ib_alloc_frmr() 68 return ibmr; in rds_ib_alloc_frmr() 70 ibmr = kzalloc_node(sizeof(*ibmr), GFP_KERNE in rds_ib_alloc_frmr() 106 rds_ib_free_frmr(struct rds_ib_mr *ibmr, bool drop) rds_ib_free_frmr() argument 123 rds_ib_post_reg_frmr(struct rds_ib_mr *ibmr) rds_ib_post_reg_frmr() argument 187 rds_ib_map_frmr(struct rds_ib_device *rds_ibdev, struct rds_ib_mr_pool *pool, struct rds_ib_mr *ibmr, struct scatterlist *sg, unsigned int sg_len) rds_ib_map_frmr() argument 266 rds_ib_post_inv(struct rds_ib_mr *ibmr) rds_ib_post_inv() argument 324 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; rds_ib_mr_cqe_handler() local 360 struct rds_ib_mr *ibmr, *next; rds_ib_unreg_frmr() local 406 struct rds_ib_mr *ibmr = NULL; rds_ib_reg_frmr() local 437 rds_ib_free_frmr_list(struct rds_ib_mr *ibmr) rds_ib_free_frmr_list() argument [all...] |
H A D | ib_rdma.c | 200 struct rds_ib_mr *ibmr = NULL; in rds_ib_reuse_mr() local 208 ibmr = llist_entry(ret, struct rds_ib_mr, llnode); in rds_ib_reuse_mr() 215 return ibmr; in rds_ib_reuse_mr() 220 struct rds_ib_mr *ibmr = trans_private; in rds_ib_sync_mr() local 221 struct rds_ib_device *rds_ibdev = ibmr->device; in rds_ib_sync_mr() 223 if (ibmr->odp) in rds_ib_sync_mr() 228 ib_dma_sync_sg_for_cpu(rds_ibdev->dev, ibmr->sg, in rds_ib_sync_mr() 229 ibmr->sg_dma_len, DMA_BIDIRECTIONAL); in rds_ib_sync_mr() 232 ib_dma_sync_sg_for_device(rds_ibdev->dev, ibmr->sg, in rds_ib_sync_mr() 233 ibmr in rds_ib_sync_mr() 238 __rds_ib_teardown_mr(struct rds_ib_mr *ibmr) __rds_ib_teardown_mr() argument 269 rds_ib_teardown_mr(struct rds_ib_mr *ibmr) rds_ib_teardown_mr() argument 298 struct rds_ib_mr *ibmr; llist_append_to_list() local 323 struct rds_ib_mr *ibmr; list_to_llist_nodes() local 345 struct rds_ib_mr *ibmr; rds_ib_flush_mr_pool() local 442 struct rds_ib_mr *ibmr = NULL; rds_ib_try_reuse_ibmr() local 486 struct rds_ib_mr *ibmr = trans_private; rds_ib_free_mr() local 545 struct rds_ib_mr *ibmr = trans_private; rds_ib_get_lkey() local 556 struct rds_ib_mr *ibmr = NULL; rds_ib_get_mr() local 696 struct rds_ib_mr *ibmr; rds_ib_odp_mr_worker() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_mr.c | 94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr() 95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr() 97 return &mr->ibmr; in pvrdma_get_dma_mr() 182 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr() 183 mr->ibmr.rkey = resp->rkey; in pvrdma_reg_user_mr() 185 return &mr->ibmr; in pvrdma_reg_user_mr() 254 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr() 255 mr->ibmr.rkey = resp->rkey; in pvrdma_alloc_mr() 259 return &mr->ibmr; in pvrdma_alloc_mr() 272 * @ibmr 277 pvrdma_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) pvrdma_dereg_mr() argument 302 pvrdma_set_page(struct ib_mr *ibmr, u64 addr) pvrdma_set_page() argument 313 pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) pvrdma_map_mr_sg() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_mr.c | 94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr() 95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr() 97 return &mr->ibmr; in pvrdma_get_dma_mr() 182 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr() 183 mr->ibmr.rkey = resp->rkey; in pvrdma_reg_user_mr() 185 return &mr->ibmr; in pvrdma_reg_user_mr() 254 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr() 255 mr->ibmr.rkey = resp->rkey; in pvrdma_alloc_mr() 259 return &mr->ibmr; in pvrdma_alloc_mr() 272 * @ibmr 277 pvrdma_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) pvrdma_dereg_mr() argument 302 pvrdma_set_page(struct ib_mr *ibmr, u64 addr) pvrdma_set_page() argument 313 pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) pvrdma_map_mr_sg() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
H A D | mr.h | 54 struct ib_mr ibmr; member 59 static inline struct rvt_mr *to_imr(struct ib_mr *ibmr) in to_imr() argument 61 return container_of(ibmr, struct rvt_mr, ibmr); in to_imr() 72 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata); 75 int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
|
H A D | trace_mr.h | 186 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 187 TP_ARGS(ibmr, sg_nents, sg_offset), 189 RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) 198 RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) 199 __entry->ibmr_iova = ibmr->iova; 200 __entry->iova = to_imr(ibmr)->mr.iova; 201 __entry->user_base = to_imr(ibmr)->mr.user_base; 202 __entry->ibmr_length = to_imr(ibmr)->mr.length;
|
H A D | mr.c | 295 * ib_reg_phys_mr() will initialize mr->ibmr except for in __rvt_alloc_mr() 301 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr() 302 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr() 356 ret = &mr->ibmr; in rvt_get_dma_mr() 428 return &mr->ibmr; in rvt_reg_user_mr() 542 * @ibmr: the memory region to free 550 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in rvt_dereg_mr() argument 552 struct rvt_mr *mr = to_imr(ibmr); in rvt_dereg_mr() 588 return &mr->ibmr; in rvt_alloc_mr() 593 * @ibmr 598 rvt_set_page(struct ib_mr *ibmr, u64 addr) rvt_set_page() argument 629 rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) rvt_map_mr_sg() argument 655 rvt_fast_reg_mr(struct rvt_qp *qp, struct ib_mr *ibmr, u32 key, int access) rvt_fast_reg_mr() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/ |
H A D | mr.h | 12 struct ib_mr ibmr; member 17 static inline struct rvt_mr *to_imr(struct ib_mr *ibmr) in to_imr() argument 19 return container_of(ibmr, struct rvt_mr, ibmr); in to_imr() 30 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata); 33 int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
|
H A D | trace_mr.h | 144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 145 TP_ARGS(ibmr, sg_nents, sg_offset), 147 RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) 156 RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device)); 157 __entry->ibmr_iova = ibmr->iova; 158 __entry->iova = to_imr(ibmr)->mr.iova; 159 __entry->user_base = to_imr(ibmr)->mr.user_base; 160 __entry->ibmr_length = to_imr(ibmr)->mr.length;
|
H A D | mr.c | 253 * ib_reg_phys_mr() will initialize mr->ibmr except for in __rvt_alloc_mr() 259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr() 260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr() 314 ret = &mr->ibmr; in rvt_get_dma_mr() 387 return &mr->ibmr; in rvt_reg_user_mr() 501 * @ibmr: the memory region to free 509 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in rvt_dereg_mr() argument 511 struct rvt_mr *mr = to_imr(ibmr); in rvt_dereg_mr() 547 return &mr->ibmr; in rvt_alloc_mr() 552 * @ibmr 557 rvt_set_page(struct ib_mr *ibmr, u64 addr) rvt_set_page() argument 588 rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) rvt_map_mr_sg() argument 614 rvt_fast_reg_mr(struct rvt_qp *qp, struct ib_mr *ibmr, u32 key, int access) rvt_fast_reg_mr() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | mr.c | 1096 mr->ibmr.lkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 1097 mr->ibmr.rkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 1100 return &mr->ibmr; in mlx5_ib_get_dma_mr() 1132 mr->ibmr.lkey = mr->mmkey.key; in set_mr_fields() 1133 mr->ibmr.rkey = mr->mmkey.key; in set_mr_fields() 1134 mr->ibmr.length = length; in set_mr_fields() 1135 mr->ibmr.device = &dev->ib_dev; in set_mr_fields() 1136 mr->ibmr.iova = iova; in set_mr_fields() 1193 mr->ibmr.pd = pd; in alloc_cacheable_mr() 1202 * If ibmr i 1863 mlx5_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) mlx5_ib_dereg_mr() argument 2282 mlx5_ib_check_mr_status(struct ib_mr *ibmr, u32 check_mask, struct ib_mr_status *mr_status) mlx5_ib_check_mr_status() argument 2324 mlx5_ib_map_pa_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_pa_mr_sg_pi() argument 2417 mlx5_set_page(struct ib_mr *ibmr, u64 addr) mlx5_set_page() argument 2431 mlx5_set_page_pi(struct ib_mr *ibmr, u64 addr) mlx5_set_page_pi() argument 2447 mlx5_ib_map_mtt_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_mtt_mr_sg_pi() argument 2512 mlx5_ib_map_klm_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_klm_mr_sg_pi() argument 2545 mlx5_ib_map_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_mr_sg_pi() argument 2606 mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) mlx5_ib_map_mr_sg() argument [all...] |
H A D | restrack.c | 81 static int fill_stat_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_stat_mr_entry() argument 83 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_stat_mr_entry() 115 static int fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry_raw() argument 117 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry_raw() 123 static int fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry() argument 125 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry()
|
/kernel/linux/linux-6.6/drivers/infiniband/sw/rxe/ |
H A D | rxe_mr.c | 29 switch (mr->ibmr.type) { in mr_check_range() 35 if (iova < mr->ibmr.iova || in mr_check_range() 36 iova + length > mr->ibmr.iova + mr->ibmr.length) { in mr_check_range() 52 /* set ibmr->l/rkey and also copy into private l/rkey in rxe_mr_init() 57 mr->lkey = mr->ibmr.lkey = key; in rxe_mr_init() 58 mr->rkey = mr->ibmr.rkey = key; in rxe_mr_init() 61 mr->ibmr.page_size = PAGE_SIZE; in rxe_mr_init() 72 mr->ibmr.type = IB_MR_TYPE_DMA; in rxe_mr_init_dma() 77 return (iova >> mr->page_shift) - (mr->ibmr in rxe_mr_iova_to_index() 208 rxe_set_page(struct ib_mr *ibmr, u64 dma_addr) rxe_set_page() argument 231 rxe_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sgl, int sg_nents, unsigned int *sg_offset) rxe_map_mr_sg() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | mr.c | 842 mr->ibmr.lkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 843 mr->ibmr.rkey = mr->mmkey.key; in mlx5_ib_get_dma_mr() 846 return &mr->ibmr; in mlx5_ib_get_dma_mr() 1005 mr->ibmr.pd = pd; in alloc_mr_from_cache() 1106 wr.pd = mr->ibmr.pd; in mlx5_ib_update_xlt() 1166 * If ibmr is NULL it will be allocated by reg_create. 1167 * Else, the given ibmr will be used. 1169 static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd, in reg_create() argument 1184 mr = ibmr ? to_mmr(ibmr) in reg_create() 1720 mlx5_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) mlx5_ib_dereg_mr() argument 2089 mlx5_ib_check_mr_status(struct ib_mr *ibmr, u32 check_mask, struct ib_mr_status *mr_status) mlx5_ib_check_mr_status() argument 2131 mlx5_ib_map_pa_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_pa_mr_sg_pi() argument 2224 mlx5_set_page(struct ib_mr *ibmr, u64 addr) mlx5_set_page() argument 2238 mlx5_set_page_pi(struct ib_mr *ibmr, u64 addr) mlx5_set_page_pi() argument 2254 mlx5_ib_map_mtt_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_mtt_mr_sg_pi() argument 2319 mlx5_ib_map_klm_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_klm_mr_sg_pi() argument 2352 mlx5_ib_map_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, int data_sg_nents, unsigned int *data_sg_offset, struct scatterlist *meta_sg, int meta_sg_nents, unsigned int *meta_sg_offset) mlx5_ib_map_mr_sg_pi() argument 2413 mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) mlx5_ib_map_mr_sg() argument [all...] |
H A D | restrack.c | 81 static int fill_stat_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_stat_mr_entry() argument 83 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_stat_mr_entry() 115 static int fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry_raw() argument 117 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry_raw() 123 static int fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry() argument 125 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr() 79 return &mr->ibmr; in mlx4_ib_get_dma_mr() 441 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr() 442 mr->ibmr.page_size = 1U << shift; in mlx4_ib_reg_user_mr() 444 return &mr->ibmr; in mlx4_ib_reg_user_mr() 587 struct ib_device *device = mr->ibmr.device; in mlx4_free_priv_pages() 596 int mlx4_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in mlx4_ib_dereg_mr() argument 598 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_ib_dereg_mr() 688 mlx4_set_page(struct ib_mr *ibmr, u64 addr) mlx4_set_page() argument 700 mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) mlx4_ib_map_mr_sg() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr() 79 return &mr->ibmr; in mlx4_ib_get_dma_mr() 441 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr() 442 mr->ibmr.page_size = 1U << shift; in mlx4_ib_reg_user_mr() 444 return &mr->ibmr; in mlx4_ib_reg_user_mr() 587 struct ib_device *device = mr->ibmr.device; in mlx4_free_priv_pages() 596 int mlx4_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in mlx4_ib_dereg_mr() argument 598 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_ib_dereg_mr() 688 mlx4_set_page(struct ib_mr *ibmr, u64 addr) mlx4_set_page() argument 700 mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) mlx4_ib_map_mr_sg() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mana/ |
H A D | mr.c | 69 mr->ibmr.lkey = resp.lkey; in mana_ib_gd_create_mr() 70 mr->ibmr.rkey = resp.rkey; in mana_ib_gd_create_mr() 164 return &mr->ibmr; in mana_ib_reg_user_mr() 178 int mana_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in mana_ib_dereg_mr() argument 180 struct mana_ib_mr *mr = container_of(ibmr, struct mana_ib_mr, ibmr); in mana_ib_dereg_mr() 181 struct ib_device *ibdev = ibmr->device; in mana_ib_dereg_mr()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib.h | 69 struct ib_mr ibmr; member 125 struct usnic_ib_mr *to_umr(struct ib_mr *ibmr) in to_umr() argument 127 return container_of(ibmr, struct usnic_ib_mr, ibmr); in to_umr()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib.h | 69 struct ib_mr ibmr; member 125 struct usnic_ib_mr *to_umr(struct ib_mr *ibmr) in to_umr() argument 127 return container_of(ibmr, struct usnic_ib_mr, ibmr); in to_umr()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 251 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_get_dma_mr() 253 return &mr->ibmr; in hns_roce_get_dma_mr() 288 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_reg_user_mr() 290 return &mr->ibmr; in hns_roce_reg_user_mr() 301 static int rereg_mr_trans(struct ib_mr *ibmr, int flags, in rereg_mr_trans() argument 307 struct hns_roce_dev *hr_dev = to_hr_dev(ibmr->device); in rereg_mr_trans() 309 struct hns_roce_mr *mr = to_hr_mr(ibmr); in rereg_mr_trans() 330 int hns_roce_rereg_user_mr(struct ib_mr *ibmr, in argument 400 hns_roce_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) hns_roce_dereg_mr() argument 468 hns_roce_set_page(struct ib_mr *ibmr, u64 addr) hns_roce_set_page() argument 480 hns_roce_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) hns_roce_map_mr_sg() argument [all...] |
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 289 struct ib_mr *ibmr; in frwr_map() local 320 ibmr = mr->frwr.fr_mr; in frwr_map() 321 n = ib_map_mr_sg(ibmr, mr->mr_sg, dma_nents, NULL, PAGE_SIZE); in frwr_map() 325 ibmr->iova &= 0x00000000ffffffff; in frwr_map() 326 ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32; in frwr_map() 327 key = (u8)(ibmr->rkey & 0x000000FF); in frwr_map() 328 ib_update_fast_reg_key(ibmr, ++key); in frwr_map() 331 reg_wr->mr = ibmr; in frwr_map() 332 reg_wr->key = ibmr->rkey; in frwr_map() 337 mr->mr_handle = ibmr in frwr_map() [all...] |
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 292 struct ib_mr *ibmr; in frwr_map() local 318 ibmr = mr->mr_ibmr; in frwr_map() 319 n = ib_map_mr_sg(ibmr, mr->mr_sg, dma_nents, NULL, PAGE_SIZE); in frwr_map() 323 ibmr->iova &= 0x00000000ffffffff; in frwr_map() 324 ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32; in frwr_map() 325 key = (u8)(ibmr->rkey & 0x000000FF); in frwr_map() 326 ib_update_fast_reg_key(ibmr, ++key); in frwr_map() 329 reg_wr->mr = ibmr; in frwr_map() 330 reg_wr->key = ibmr->rkey; in frwr_map() 335 mr->mr_handle = ibmr in frwr_map() [all...] |