Home
last modified time | relevance | path

Searched refs:mtt (Results 1 - 25 of 69) sorted by relevance

123

/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
H A Dmthca_mr.c211 struct mthca_mtt *mtt; in __mthca_alloc_mtt() local
217 mtt = kmalloc(sizeof *mtt, GFP_KERNEL); in __mthca_alloc_mtt()
218 if (!mtt) in __mthca_alloc_mtt()
221 mtt->buddy = buddy; in __mthca_alloc_mtt()
222 mtt->order = 0; in __mthca_alloc_mtt()
224 ++mtt->order; in __mthca_alloc_mtt()
226 mtt->first_seg = mthca_alloc_mtt_range(dev, mtt->order, buddy); in __mthca_alloc_mtt()
227 if (mtt in __mthca_alloc_mtt()
240 mthca_free_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt) mthca_free_mtt() argument
254 __mthca_write_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) __mthca_write_mtt() argument
315 mthca_tavor_write_mtt_seg(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_tavor_write_mtt_seg() argument
329 mthca_arbel_write_mtt_seg(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_arbel_write_mtt_seg() argument
358 mthca_write_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_write_mtt() argument
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/
H A Dmthca_mr.c209 struct mthca_mtt *mtt; in __mthca_alloc_mtt() local
215 mtt = kmalloc(sizeof *mtt, GFP_KERNEL); in __mthca_alloc_mtt()
216 if (!mtt) in __mthca_alloc_mtt()
219 mtt->buddy = buddy; in __mthca_alloc_mtt()
220 mtt->order = 0; in __mthca_alloc_mtt()
222 ++mtt->order; in __mthca_alloc_mtt()
224 mtt->first_seg = mthca_alloc_mtt_range(dev, mtt->order, buddy); in __mthca_alloc_mtt()
225 if (mtt in __mthca_alloc_mtt()
238 mthca_free_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt) mthca_free_mtt() argument
252 __mthca_write_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) __mthca_write_mtt() argument
313 mthca_tavor_write_mtt_seg(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_tavor_write_mtt_seg() argument
327 mthca_arbel_write_mtt_seg(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_arbel_write_mtt_seg() argument
356 mthca_write_mtt(struct mthca_dev *dev, struct mthca_mtt *mtt, int start_index, u64 *buffer_list, int list_len) mthca_write_mtt() argument
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/
H A Derdma_verbs.c25 struct erdma_mtt *mtt = mem->mtt; in assemble_qbuf_mtt_for_cmd() local
28 *addr0 = mtt->buf_dma; in assemble_qbuf_mtt_for_cmd()
32 *addr0 = mtt->buf[0]; in assemble_qbuf_mtt_for_cmd()
33 memcpy(addr1, mtt->buf + 1, MTT_SIZE(mem->mtt_nents - 1)); in assemble_qbuf_mtt_for_cmd()
143 if (mr->mem.mtt->continuous) { in regmr_cmd()
144 req.phy_addr[0] = mr->mem.mtt->buf_dma; in regmr_cmd()
147 req.phy_addr[0] = sg_dma_address(mr->mem.mtt->sglist); in regmr_cmd()
148 mtt_level = mr->mem.mtt->level; in regmr_cmd()
151 memcpy(req.phy_addr, mr->mem.mtt in regmr_cmd()
543 struct erdma_mtt *mtt = mem->mtt; erdma_fill_bottom_mtt() local
557 struct erdma_mtt *mtt; erdma_create_cont_mtt() local
585 erdma_destroy_mtt_buf_sg(struct erdma_dev *dev, struct erdma_mtt *mtt) erdma_destroy_mtt_buf_sg() argument
592 erdma_destroy_scatter_mtt(struct erdma_dev *dev, struct erdma_mtt *mtt) erdma_destroy_scatter_mtt() argument
600 erdma_init_middle_mtt(struct erdma_mtt *mtt, struct erdma_mtt *low_mtt) erdma_init_middle_mtt() argument
610 erdma_create_mtt_buf_sg(struct erdma_dev *dev, struct erdma_mtt *mtt) erdma_create_mtt_buf_sg() argument
652 struct erdma_mtt *mtt; erdma_create_scatter_mtt() local
686 struct erdma_mtt *mtt, *tmp_mtt; erdma_create_mtt() local
736 erdma_destroy_mtt(struct erdma_dev *dev, struct erdma_mtt *mtt) erdma_destroy_mtt() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/
H A Dmr.c195 struct mlx4_mtt *mtt) in mlx4_mtt_init()
200 mtt->order = -1; in mlx4_mtt_init()
201 mtt->page_shift = MLX4_ICM_PAGE_SHIFT; in mlx4_mtt_init()
204 mtt->page_shift = page_shift; in mlx4_mtt_init()
206 for (mtt->order = 0, i = 1; i < npages; i <<= 1) in mlx4_mtt_init()
207 ++mtt->order; in mlx4_mtt_init()
209 mtt->offset = mlx4_alloc_mtt_range(dev, mtt->order); in mlx4_mtt_init()
210 if (mtt->offset == -1) in mlx4_mtt_init()
244 mlx4_warn(dev, "Failed to free mtt rang in mlx4_free_mtt_range()
194 mlx4_mtt_init(struct mlx4_dev *dev, int npages, int page_shift, struct mlx4_mtt *mtt) mlx4_mtt_init() argument
251 mlx4_mtt_cleanup(struct mlx4_dev *dev, struct mlx4_mtt *mtt) mlx4_mtt_cleanup() argument
260 mlx4_mtt_addr(struct mlx4_dev *dev, struct mlx4_mtt *mtt) mlx4_mtt_addr() argument
692 mlx4_write_mtt_chunk(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) mlx4_write_mtt_chunk() argument
718 __mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) __mlx4_write_mtt() argument
746 mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) mlx4_write_mtt() argument
790 mlx4_buf_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_buf *buf) mlx4_buf_write_mtt() argument
[all...]
H A Dresource_tracker.c110 struct res_mtt *mtt; member
158 struct res_mtt *mtt; member
170 struct res_mtt *mtt; member
181 struct res_mtt *mtt; member
193 struct res_mtt *mtt; member
456 dev->quotas.mtt = dev->caps.num_mtts - dev->caps.reserved_mtts; in mlx4_init_quotas()
468 dev->quotas.mtt = in mlx4_init_quotas()
2744 int size, struct res_mtt *mtt) in check_mtt_range()
2746 int res_start = mtt->com.res_id; in check_mtt_range()
2747 int res_size = (1 << mtt in check_mtt_range()
2743 check_mtt_range(struct mlx4_dev *dev, int slave, int start, int size, struct res_mtt *mtt) check_mtt_range() argument
2762 struct res_mtt *mtt; mlx4_SW2HW_MPT_wrapper() local
2959 struct res_mtt *mtt; mlx4_RST2INIT_QP_wrapper() local
3105 struct res_mtt *mtt; mlx4_SW2HW_EQ_wrapper() local
3163 struct res_mtt *mtt; get_containing_mtt() local
3268 struct mlx4_mtt mtt; mlx4_WRITE_MTT_wrapper() local
3444 struct res_mtt *mtt; mlx4_SW2HW_CQ_wrapper() local
3529 struct res_mtt *mtt; handle_resize() local
3617 struct res_mtt *mtt; mlx4_SW2HW_SRQ_wrapper() local
4933 struct res_mtt *mtt; rem_slave_mtts() local
[all...]
H A Dcq.c192 int entries, struct mlx4_mtt *mtt) in mlx4_cq_resize()
205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize()
206 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_cq_resize()
342 struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, in mlx4_cq_alloc()
384 cq_context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in mlx4_cq_alloc()
386 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_cq_alloc()
191 mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq, int entries, struct mlx4_mtt *mtt) mlx4_cq_resize() argument
341 mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, unsigned vector, int collapsed, int timestamp_en, void *buf_addr, bool user_cq) mlx4_cq_alloc() argument
H A Dqp.c87 static int __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in __mlx4_qp_modify() argument
164 u64 mtt_addr = mlx4_mtt_addr(dev, mtt); in __mlx4_qp_modify()
167 context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in __mlx4_qp_modify()
208 int mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in mlx4_qp_modify() argument
214 return __mlx4_qp_modify(dev, mtt, cur_state, new_state, context, in mlx4_qp_modify()
915 int mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in mlx4_qp_to_ready() argument
933 err = mlx4_qp_modify(dev, mtt, states[i], states[i + 1], in mlx4_qp_to_ready()
H A Dsrq.c163 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq) in mlx4_srq_alloc()
193 srq_context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in mlx4_srq_alloc()
195 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_srq_alloc()
162 mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd, struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq) mlx4_srq_alloc() argument
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/
H A Dmr.c195 struct mlx4_mtt *mtt) in mlx4_mtt_init()
200 mtt->order = -1; in mlx4_mtt_init()
201 mtt->page_shift = MLX4_ICM_PAGE_SHIFT; in mlx4_mtt_init()
204 mtt->page_shift = page_shift; in mlx4_mtt_init()
206 for (mtt->order = 0, i = 1; i < npages; i <<= 1) in mlx4_mtt_init()
207 ++mtt->order; in mlx4_mtt_init()
209 mtt->offset = mlx4_alloc_mtt_range(dev, mtt->order); in mlx4_mtt_init()
210 if (mtt->offset == -1) in mlx4_mtt_init()
244 mlx4_warn(dev, "Failed to free mtt rang in mlx4_free_mtt_range()
194 mlx4_mtt_init(struct mlx4_dev *dev, int npages, int page_shift, struct mlx4_mtt *mtt) mlx4_mtt_init() argument
251 mlx4_mtt_cleanup(struct mlx4_dev *dev, struct mlx4_mtt *mtt) mlx4_mtt_cleanup() argument
260 mlx4_mtt_addr(struct mlx4_dev *dev, struct mlx4_mtt *mtt) mlx4_mtt_addr() argument
692 mlx4_write_mtt_chunk(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) mlx4_write_mtt_chunk() argument
718 __mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) __mlx4_write_mtt() argument
746 mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) mlx4_write_mtt() argument
790 mlx4_buf_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_buf *buf) mlx4_buf_write_mtt() argument
[all...]
H A Dresource_tracker.c110 struct res_mtt *mtt; member
158 struct res_mtt *mtt; member
170 struct res_mtt *mtt; member
181 struct res_mtt *mtt; member
193 struct res_mtt *mtt; member
456 dev->quotas.mtt = dev->caps.num_mtts - dev->caps.reserved_mtts; in mlx4_init_quotas()
468 dev->quotas.mtt = in mlx4_init_quotas()
2745 int size, struct res_mtt *mtt) in check_mtt_range()
2747 int res_start = mtt->com.res_id; in check_mtt_range()
2748 int res_size = (1 << mtt in check_mtt_range()
2744 check_mtt_range(struct mlx4_dev *dev, int slave, int start, int size, struct res_mtt *mtt) check_mtt_range() argument
2763 struct res_mtt *mtt; mlx4_SW2HW_MPT_wrapper() local
2960 struct res_mtt *mtt; mlx4_RST2INIT_QP_wrapper() local
3106 struct res_mtt *mtt; mlx4_SW2HW_EQ_wrapper() local
3164 struct res_mtt *mtt; get_containing_mtt() local
3269 struct mlx4_mtt mtt; mlx4_WRITE_MTT_wrapper() local
3445 struct res_mtt *mtt; mlx4_SW2HW_CQ_wrapper() local
3530 struct res_mtt *mtt; handle_resize() local
3618 struct res_mtt *mtt; mlx4_SW2HW_SRQ_wrapper() local
4934 struct res_mtt *mtt; rem_slave_mtts() local
[all...]
H A Dcq.c192 int entries, struct mlx4_mtt *mtt) in mlx4_cq_resize()
205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize()
206 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_cq_resize()
343 struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, in mlx4_cq_alloc()
385 cq_context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in mlx4_cq_alloc()
387 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_cq_alloc()
191 mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq, int entries, struct mlx4_mtt *mtt) mlx4_cq_resize() argument
342 mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, unsigned vector, int collapsed, int timestamp_en, void *buf_addr, bool user_cq) mlx4_cq_alloc() argument
H A Dqp.c92 static int __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in __mlx4_qp_modify() argument
169 u64 mtt_addr = mlx4_mtt_addr(dev, mtt); in __mlx4_qp_modify()
172 context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in __mlx4_qp_modify()
213 int mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in mlx4_qp_modify() argument
219 return __mlx4_qp_modify(dev, mtt, cur_state, new_state, context, in mlx4_qp_modify()
919 int mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, in mlx4_qp_to_ready() argument
937 err = mlx4_qp_modify(dev, mtt, states[i], states[i + 1], in mlx4_qp_to_ready()
H A Dsrq.c163 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq) in mlx4_srq_alloc()
193 srq_context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; in mlx4_srq_alloc()
195 mtt_addr = mlx4_mtt_addr(dev, mtt); in mlx4_srq_alloc()
162 mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd, struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq) mlx4_srq_alloc() argument
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/
H A Dmr.c95 struct mlx4_mtt *mtt, in mlx4_ib_umem_write_mtt_block()
134 err = mlx4_write_mtt(dev->dev, mtt, *start_index, in mlx4_ib_umem_write_mtt_block()
163 * mtt size accordingly. Here we take the last right bit which in mlx4_ib_umem_calc_block_mtt()
175 * reduce the mtt size accordingly. in mlx4_ib_umem_calc_block_mtt()
182 int mlx4_ib_umem_write_mtt(struct mlx4_ib_dev *dev, struct mlx4_mtt *mtt, in mlx4_ib_umem_write_mtt() argument
200 mtt_shift = mtt->page_shift; in mlx4_ib_umem_write_mtt()
211 * If len is malaligned, write an extra mtt entry to cover the in mlx4_ib_umem_write_mtt()
214 err = mlx4_ib_umem_write_mtt_block(dev, mtt, mtt_size, in mlx4_ib_umem_write_mtt()
229 * If len is malaligned, write an extra mtt entry to cover in mlx4_ib_umem_write_mtt()
232 err = mlx4_ib_umem_write_mtt_block(dev, mtt, mtt_siz in mlx4_ib_umem_write_mtt()
94 mlx4_ib_umem_write_mtt_block(struct mlx4_ib_dev *dev, struct mlx4_mtt *mtt, u64 mtt_size, u64 mtt_shift, u64 len, u64 cur_start_addr, u64 *pages, int *start_index, int *npages) mlx4_ib_umem_write_mtt_block() argument
[all...]
H A Dsrq.c120 PAGE_SHIFT, &srq->mtt); in mlx4_ib_create_srq()
124 err = mlx4_ib_umem_write_mtt(dev, &srq->mtt, srq->umem); in mlx4_ib_create_srq()
160 &srq->mtt); in mlx4_ib_create_srq()
164 err = mlx4_buf_write_mtt(dev->dev, &srq->mtt, &srq->buf); in mlx4_ib_create_srq()
182 &srq->mtt, srq->db.dma, &srq->msrq); in mlx4_ib_create_srq()
206 mlx4_mtt_cleanup(dev->dev, &srq->mtt); in mlx4_ib_create_srq()
270 mlx4_mtt_cleanup(dev->dev, &msrq->mtt); in mlx4_ib_destroy_srq()
H A Dcq.c113 &buf->mtt); in mlx4_ib_alloc_cq_buf()
117 err = mlx4_buf_write_mtt(dev->dev, &buf->mtt, &buf->buf); in mlx4_ib_alloc_cq_buf()
124 mlx4_mtt_cleanup(dev->dev, &buf->mtt); in mlx4_ib_alloc_cq_buf()
153 err = mlx4_mtt_init(dev->dev, n, shift, &buf->mtt); in mlx4_ib_get_cq_umem()
158 err = mlx4_ib_umem_write_mtt(dev, &buf->mtt, *umem); in mlx4_ib_get_cq_umem()
165 mlx4_mtt_cleanup(dev->dev, &buf->mtt); in mlx4_ib_get_cq_umem()
247 err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, cq->db.dma, in mlx4_ib_create_cq()
277 mlx4_mtt_cleanup(dev->dev, &cq->buf.mtt); in mlx4_ib_create_cq()
383 struct mlx4_mtt mtt; in mlx4_ib_resize_cq() local
421 mtt in mlx4_ib_resize_cq()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/
H A Dmr.c95 struct mlx4_mtt *mtt, in mlx4_ib_umem_write_mtt_block()
134 err = mlx4_write_mtt(dev->dev, mtt, *start_index, in mlx4_ib_umem_write_mtt_block()
163 * mtt size accordingly. Here we take the last right bit which in mlx4_ib_umem_calc_block_mtt()
175 * reduce the mtt size accordingly. in mlx4_ib_umem_calc_block_mtt()
182 int mlx4_ib_umem_write_mtt(struct mlx4_ib_dev *dev, struct mlx4_mtt *mtt, in mlx4_ib_umem_write_mtt() argument
200 mtt_shift = mtt->page_shift; in mlx4_ib_umem_write_mtt()
211 * If len is malaligned, write an extra mtt entry to cover the in mlx4_ib_umem_write_mtt()
214 err = mlx4_ib_umem_write_mtt_block(dev, mtt, mtt_size, in mlx4_ib_umem_write_mtt()
229 * If len is malaligned, write an extra mtt entry to cover in mlx4_ib_umem_write_mtt()
232 err = mlx4_ib_umem_write_mtt_block(dev, mtt, mtt_siz in mlx4_ib_umem_write_mtt()
94 mlx4_ib_umem_write_mtt_block(struct mlx4_ib_dev *dev, struct mlx4_mtt *mtt, u64 mtt_size, u64 mtt_shift, u64 len, u64 cur_start_addr, u64 *pages, int *start_index, int *npages) mlx4_ib_umem_write_mtt_block() argument
[all...]
H A Dsrq.c124 PAGE_SHIFT, &srq->mtt); in mlx4_ib_create_srq()
128 err = mlx4_ib_umem_write_mtt(dev, &srq->mtt, srq->umem); in mlx4_ib_create_srq()
164 &srq->mtt); in mlx4_ib_create_srq()
168 err = mlx4_buf_write_mtt(dev->dev, &srq->mtt, &srq->buf); in mlx4_ib_create_srq()
186 &srq->mtt, srq->db.dma, &srq->msrq); in mlx4_ib_create_srq()
210 mlx4_mtt_cleanup(dev->dev, &srq->mtt); in mlx4_ib_create_srq()
274 mlx4_mtt_cleanup(dev->dev, &msrq->mtt); in mlx4_ib_destroy_srq()
H A Dcq.c113 &buf->mtt); in mlx4_ib_alloc_cq_buf()
117 err = mlx4_buf_write_mtt(dev->dev, &buf->mtt, &buf->buf); in mlx4_ib_alloc_cq_buf()
124 mlx4_mtt_cleanup(dev->dev, &buf->mtt); in mlx4_ib_alloc_cq_buf()
153 err = mlx4_mtt_init(dev->dev, n, shift, &buf->mtt); in mlx4_ib_get_cq_umem()
158 err = mlx4_ib_umem_write_mtt(dev, &buf->mtt, *umem); in mlx4_ib_get_cq_umem()
165 mlx4_mtt_cleanup(dev->dev, &buf->mtt); in mlx4_ib_get_cq_umem()
247 err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, cq->db.dma, in mlx4_ib_create_cq()
277 mlx4_mtt_cleanup(dev->dev, &cq->buf.mtt); in mlx4_ib_create_cq()
383 struct mlx4_mtt mtt; in mlx4_ib_resize_cq() local
421 mtt in mlx4_ib_resize_cq()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/
H A Dumr.c619 struct mlx5_mtt *mtt; in mlx5r_umr_update_mr_pas() local
628 mtt = mlx5r_umr_create_xlt( in mlx5r_umr_update_mr_pas()
630 sizeof(*mtt), flags); in mlx5r_umr_update_mr_pas()
631 if (!mtt) in mlx5r_umr_update_mr_pas()
641 cur_mtt = mtt; in mlx5r_umr_update_mr_pas()
643 if (cur_mtt == (void *)mtt + sg.length) { in mlx5r_umr_update_mr_pas()
656 cur_mtt = mtt; in mlx5r_umr_update_mr_pas()
669 final_size = (void *)cur_mtt - (void *)mtt; in mlx5r_umr_update_mr_pas()
679 mlx5r_umr_unmap_free_xlt(dev, mtt, &sg); in mlx5r_umr_update_mr_pas()
H A Dodp.c133 struct mlx5_ib_mr *mtt = xa_load(&imr->implicit_children, idx); in populate_klm() local
136 if (mtt) { in populate_klm()
137 pklm->key = cpu_to_be32(mtt->ibmr.lkey); in populate_klm()
536 struct mlx5_ib_mr *mtt; in mlx5_ib_free_odp_mr() local
543 xa_for_each(&mr->implicit_children, idx, mtt) { in mlx5_ib_free_odp_mr()
545 mlx5_ib_dereg_mr(&mtt->ibmr, NULL); in mlx5_ib_free_odp_mr()
623 struct mlx5_ib_mr *mtt; in pagefault_implicit_mr() local
627 mtt = xa_load(&imr->implicit_children, idx); in pagefault_implicit_mr()
628 if (unlikely(!mtt)) { in pagefault_implicit_mr()
630 mtt in pagefault_implicit_mr()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
H A Dodp.c135 struct mlx5_ib_mr *mtt = xa_load(&imr->implicit_children, idx); in populate_klm() local
138 if (mtt) { in populate_klm()
139 pklm->key = cpu_to_be32(mtt->ibmr.lkey); in populate_klm()
589 struct mlx5_ib_mr *mtt; in mlx5_ib_free_implicit_mr() local
621 xa_for_each (&imr->implicit_children, idx, mtt) { in mlx5_ib_free_implicit_mr()
623 list_add(&mtt->odp_destroy.elm, &destroy_list); in mlx5_ib_free_implicit_mr()
640 list_for_each_entry_safe (mtt, tmp, &destroy_list, odp_destroy.elm) in mlx5_ib_free_implicit_mr()
641 free_implicit_child_mr(mtt, false); in mlx5_ib_free_implicit_mr()
741 struct mlx5_ib_mr *mtt; in pagefault_implicit_mr() local
744 mtt in pagefault_implicit_mr()
[all...]
/kernel/linux/linux-5.10/include/linux/mlx4/
H A Ddevice.h684 struct mlx4_mtt mtt; member
689 struct mlx4_mtt mtt; member
835 int mtt; member
1102 struct mlx4_mtt *mtt);
1103 void mlx4_mtt_cleanup(struct mlx4_dev *dev, struct mlx4_mtt *mtt);
1104 u64 mlx4_mtt_addr(struct mlx4_dev *dev, struct mlx4_mtt *mtt);
1114 int mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
1116 int mlx4_buf_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
1127 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
1140 struct mlx4_mtt *mtt, u6
[all...]
H A Dqp.h477 int mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
485 int mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
/kernel/linux/linux-6.6/include/linux/mlx4/
H A Ddevice.h684 struct mlx4_mtt mtt; member
689 struct mlx4_mtt mtt; member
835 int mtt; member
1121 struct mlx4_mtt *mtt);
1122 void mlx4_mtt_cleanup(struct mlx4_dev *dev, struct mlx4_mtt *mtt);
1123 u64 mlx4_mtt_addr(struct mlx4_dev *dev, struct mlx4_mtt *mtt);
1133 int mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
1135 int mlx4_buf_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
1146 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
1159 struct mlx4_mtt *mtt, u6
[all...]

Completed in 41 milliseconds

123