Lines Matching refs:in

12 static int get_pas_size(struct mlx5_srq_attr *in)
14 u32 log_page_size = in->log_page_size + 12;
15 u32 log_srq_size = in->log_size;
16 u32 log_rq_stride = in->wqe_shift;
17 u32 page_offset = in->page_offset;
27 static void set_wq(void *wq, struct mlx5_srq_attr *in)
29 MLX5_SET(wq, wq, wq_signature, !!(in->flags
31 MLX5_SET(wq, wq, log_wq_pg_sz, in->log_page_size);
32 MLX5_SET(wq, wq, log_wq_stride, in->wqe_shift + 4);
33 MLX5_SET(wq, wq, log_wq_sz, in->log_size);
34 MLX5_SET(wq, wq, page_offset, in->page_offset);
35 MLX5_SET(wq, wq, lwm, in->lwm);
36 MLX5_SET(wq, wq, pd, in->pd);
37 MLX5_SET64(wq, wq, dbr_addr, in->db_record);
40 static void set_srqc(void *srqc, struct mlx5_srq_attr *in)
42 MLX5_SET(srqc, srqc, wq_signature, !!(in->flags
44 MLX5_SET(srqc, srqc, log_page_size, in->log_page_size);
45 MLX5_SET(srqc, srqc, log_rq_stride, in->wqe_shift);
46 MLX5_SET(srqc, srqc, log_srq_size, in->log_size);
47 MLX5_SET(srqc, srqc, page_offset, in->page_offset);
48 MLX5_SET(srqc, srqc, lwm, in->lwm);
49 MLX5_SET(srqc, srqc, pd, in->pd);
50 MLX5_SET64(srqc, srqc, dbr_addr, in->db_record);
51 MLX5_SET(srqc, srqc, xrcd, in->xrcd);
52 MLX5_SET(srqc, srqc, cqn, in->cqn);
55 static void get_wq(void *wq, struct mlx5_srq_attr *in)
58 in->flags &= MLX5_SRQ_FLAG_WQ_SIG;
59 in->log_page_size = MLX5_GET(wq, wq, log_wq_pg_sz);
60 in->wqe_shift = MLX5_GET(wq, wq, log_wq_stride) - 4;
61 in->log_size = MLX5_GET(wq, wq, log_wq_sz);
62 in->page_offset = MLX5_GET(wq, wq, page_offset);
63 in->lwm = MLX5_GET(wq, wq, lwm);
64 in->pd = MLX5_GET(wq, wq, pd);
65 in->db_record = MLX5_GET64(wq, wq, dbr_addr);
68 static void get_srqc(void *srqc, struct mlx5_srq_attr *in)
71 in->flags &= MLX5_SRQ_FLAG_WQ_SIG;
72 in->log_page_size = MLX5_GET(srqc, srqc, log_page_size);
73 in->wqe_shift = MLX5_GET(srqc, srqc, log_rq_stride);
74 in->log_size = MLX5_GET(srqc, srqc, log_srq_size);
75 in->page_offset = MLX5_GET(srqc, srqc, page_offset);
76 in->lwm = MLX5_GET(srqc, srqc, lwm);
77 in->pd = MLX5_GET(srqc, srqc, pd);
78 in->db_record = MLX5_GET64(srqc, srqc, dbr_addr);
95 static int __set_srq_page_size(struct mlx5_srq_attr *in,
100 in->log_page_size = order_base_2(page_size) - MLX5_ADAPTER_PAGE_SHIFT;
102 if (WARN_ON(get_pas_size(in) !=
103 ib_umem_num_dma_blocks(in->umem, page_size) * sizeof(u64)))
108 #define set_srq_page_size(in, typ, log_pgsz_fld) \
109 __set_srq_page_size(in, mlx5_umem_find_best_quantized_pgoff( \
110 (in)->umem, typ, log_pgsz_fld, \
112 64, &(in)->page_offset))
115 struct mlx5_srq_attr *in)
125 if (in->umem) {
126 err = set_srq_page_size(in, srqc, log_page_size);
131 pas_size = get_pas_size(in);
137 MLX5_SET(create_srq_in, create_in, uid, in->uid);
141 set_srqc(srqc, in);
142 if (in->umem)
144 in->umem,
145 1UL << (in->log_page_size + MLX5_ADAPTER_PAGE_SHIFT),
148 memcpy(pas, in->pas, pas_size);
158 srq->uid = in->uid;
166 u32 in[MLX5_ST_SZ_DW(destroy_srq_in)] = {};
168 MLX5_SET(destroy_srq_in, in, opcode, MLX5_CMD_OP_DESTROY_SRQ);
169 MLX5_SET(destroy_srq_in, in, srqn, srq->srqn);
170 MLX5_SET(destroy_srq_in, in, uid, srq->uid);
172 return mlx5_cmd_exec_in(dev->mdev, destroy_srq, in);
178 u32 in[MLX5_ST_SZ_DW(arm_rq_in)] = {};
180 MLX5_SET(arm_rq_in, in, opcode, MLX5_CMD_OP_ARM_RQ);
181 MLX5_SET(arm_rq_in, in, op_mod, MLX5_ARM_RQ_IN_OP_MOD_SRQ);
182 MLX5_SET(arm_rq_in, in, srq_number, srq->srqn);
183 MLX5_SET(arm_rq_in, in, lwm, lwm);
184 MLX5_SET(arm_rq_in, in, uid, srq->uid);
186 return mlx5_cmd_exec_in(dev->mdev, arm_rq, in);
192 u32 in[MLX5_ST_SZ_DW(query_srq_in)] = {};
201 MLX5_SET(query_srq_in, in, opcode, MLX5_CMD_OP_QUERY_SRQ);
202 MLX5_SET(query_srq_in, in, srqn, srq->srqn);
203 err = mlx5_cmd_exec_inout(dev->mdev, query_srq, in, srq_out);
218 struct mlx5_srq_attr *in)
228 if (in->umem) {
229 err = set_srq_page_size(in, xrc_srqc, log_page_size);
234 pas_size = get_pas_size(in);
240 MLX5_SET(create_xrc_srq_in, create_in, uid, in->uid);
245 set_srqc(xrc_srqc, in);
246 MLX5_SET(xrc_srqc, xrc_srqc, user_index, in->user_index);
247 if (in->umem)
249 in->umem,
250 1UL << (in->log_page_size + MLX5_ADAPTER_PAGE_SHIFT),
253 memcpy(pas, in->pas, pas_size);
264 srq->uid = in->uid;
273 u32 in[MLX5_ST_SZ_DW(destroy_xrc_srq_in)] = {};
275 MLX5_SET(destroy_xrc_srq_in, in, opcode, MLX5_CMD_OP_DESTROY_XRC_SRQ);
276 MLX5_SET(destroy_xrc_srq_in, in, xrc_srqn, srq->srqn);
277 MLX5_SET(destroy_xrc_srq_in, in, uid, srq->uid);
279 return mlx5_cmd_exec_in(dev->mdev, destroy_xrc_srq, in);
285 u32 in[MLX5_ST_SZ_DW(arm_xrc_srq_in)] = {};
287 MLX5_SET(arm_xrc_srq_in, in, opcode, MLX5_CMD_OP_ARM_XRC_SRQ);
288 MLX5_SET(arm_xrc_srq_in, in, op_mod,
290 MLX5_SET(arm_xrc_srq_in, in, xrc_srqn, srq->srqn);
291 MLX5_SET(arm_xrc_srq_in, in, lwm, lwm);
292 MLX5_SET(arm_xrc_srq_in, in, uid, srq->uid);
294 return mlx5_cmd_exec_in(dev->mdev, arm_xrc_srq, in);
301 u32 in[MLX5_ST_SZ_DW(query_xrc_srq_in)] = {};
310 MLX5_SET(query_xrc_srq_in, in, opcode, MLX5_CMD_OP_QUERY_XRC_SRQ);
311 MLX5_SET(query_xrc_srq_in, in, xrc_srqn, srq->srqn);
313 err = mlx5_cmd_exec_inout(dev->mdev, query_xrc_srq, in, xrcsrq_out);
329 struct mlx5_srq_attr *in)
341 if (in->umem) {
342 err = set_srq_page_size(in, wq, log_wq_pg_sz);
347 pas_size = get_pas_size(in);
361 MLX5_SET(create_rmp_in, create_in, uid, in->uid);
364 set_wq(wq, in);
365 if (in->umem)
367 in->umem,
368 1UL << (in->log_page_size + MLX5_ADAPTER_PAGE_SHIFT),
371 memcpy(pas, in->pas, pas_size);
377 srq->uid = in->uid;
388 u32 in[MLX5_ST_SZ_DW(destroy_rmp_in)] = {};
390 MLX5_SET(destroy_rmp_in, in, opcode, MLX5_CMD_OP_DESTROY_RMP);
391 MLX5_SET(destroy_rmp_in, in, rmpn, srq->srqn);
392 MLX5_SET(destroy_rmp_in, in, uid, srq->uid);
393 return mlx5_cmd_exec_in(dev->mdev, destroy_rmp, in);
400 void *in = NULL;
411 in = kvzalloc(inlen, GFP_KERNEL);
413 if (!in || !out) {
418 rmpc = MLX5_ADDR_OF(modify_rmp_in, in, ctx);
419 bitmask = MLX5_ADDR_OF(modify_rmp_in, in, bitmask);
422 MLX5_SET(modify_rmp_in, in, rmp_state, MLX5_RMPC_STATE_RDY);
423 MLX5_SET(modify_rmp_in, in, rmpn, srq->srqn);
424 MLX5_SET(modify_rmp_in, in, uid, srq->uid);
428 MLX5_SET(modify_rmp_in, in, opcode, MLX5_CMD_OP_MODIFY_RMP);
430 err = mlx5_cmd_exec_inout(dev->mdev, modify_rmp, in, out);
433 kvfree(in);
476 struct mlx5_srq_attr *in)
487 if (in->umem) {
488 err = set_srq_page_size(in, wq, log_wq_pg_sz);
493 pas_size = get_pas_size(in);
503 set_wq(wq, in);
504 if (in->umem)
506 in->umem,
507 1UL << (in->log_page_size + MLX5_ADAPTER_PAGE_SHIFT),
510 memcpy(pas, in->pas, pas_size);
512 if (in->type == IB_SRQT_TM) {
514 if (in->flags & MLX5_SRQ_FLAG_RNDV)
518 in->tm_log_list_size);
520 MLX5_SET(xrqc, xrqc, user_index, in->user_index);
521 MLX5_SET(xrqc, xrqc, cqn, in->cqn);
523 MLX5_SET(create_xrq_in, create_in, uid, in->uid);
529 srq->uid = in->uid;
537 u32 in[MLX5_ST_SZ_DW(destroy_xrq_in)] = {};
539 MLX5_SET(destroy_xrq_in, in, opcode, MLX5_CMD_OP_DESTROY_XRQ);
540 MLX5_SET(destroy_xrq_in, in, xrqn, srq->srqn);
541 MLX5_SET(destroy_xrq_in, in, uid, srq->uid);
543 return mlx5_cmd_exec_in(dev->mdev, destroy_xrq, in);
550 u32 in[MLX5_ST_SZ_DW(arm_rq_in)] = {};
552 MLX5_SET(arm_rq_in, in, opcode, MLX5_CMD_OP_ARM_RQ);
553 MLX5_SET(arm_rq_in, in, op_mod, MLX5_ARM_RQ_IN_OP_MOD_XRQ);
554 MLX5_SET(arm_rq_in, in, srq_number, srq->srqn);
555 MLX5_SET(arm_rq_in, in, lwm, lwm);
556 MLX5_SET(arm_rq_in, in, uid, srq->uid);
558 return mlx5_cmd_exec_in(dev->mdev, arm_rq, in);
564 u32 in[MLX5_ST_SZ_DW(query_xrq_in)] = {};
574 MLX5_SET(query_xrq_in, in, opcode, MLX5_CMD_OP_QUERY_XRQ);
575 MLX5_SET(query_xrq_in, in, xrqn, srq->srqn);
577 err = mlx5_cmd_exec_inout(dev->mdev, query_xrq, in, xrq_out);
601 struct mlx5_srq_attr *in)
604 return create_srq_cmd(dev, srq, in);
607 return create_xrc_srq_cmd(dev, srq, in);
609 return create_xrq_cmd(dev, srq, in);
611 return create_rmp_cmd(dev, srq, in);
630 struct mlx5_srq_attr *in)
635 switch (in->type) {
646 err = create_srq_split(dev, srq, in);
680 * because we are storing in pre-allocated space xarray