Lines Matching refs:mcq

48 static void mlx5_ib_cq_event(struct mlx5_core_cq *mcq, enum mlx5_event type)
50 struct mlx5_ib_cq *cq = container_of(mcq, struct mlx5_ib_cq, mcq);
57 type, mcq->cqn);
84 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
96 return get_sw_cqe(cq, cq->mcq.cons_index);
461 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
463 ++cq->mcq.cons_index;
520 "Requestor" : "Responder", cq->mcq.cqn);
557 cq->mcq.cqn, sig->err_item.key,
583 cq->mcq.cqn);
627 mlx5_cq_set_ci(&cq->mcq);
650 mlx5_cq_arm(&cq->mcq,
653 uar_page, to_mcq(ibcq)->mcq.cons_index);
866 cq->mcq.set_ci_db = cq->db.db;
867 cq->mcq.arm_db = cq->db.db + 1;
868 cq->mcq.cqe_sz = cqe_size;
989 err = mlx5_core_create_cq(dev->mdev, &cq->mcq, cqb, inlen, out, sizeof(out));
993 mlx5_ib_dbg(dev, "cqn 0x%x\n", cq->mcq.cqn);
995 cq->mcq.tasklet_ctx.comp = mlx5_ib_cq_comp;
997 cq->mcq.comp = mlx5_ib_cq_comp;
998 cq->mcq.event = mlx5_ib_cq_event;
1003 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof(__u32))) {
1013 mlx5_core_destroy_cq(dev->mdev, &cq->mcq);
1027 struct mlx5_ib_cq *mcq = to_mcq(cq);
1030 ret = mlx5_core_destroy_cq(dev->mdev, &mcq->mcq);
1035 destroy_cq_user(mcq, udata);
1037 destroy_cq_kernel(dev, mcq);
1063 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); prod_index++)
1064 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe)
1070 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) {
1072 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
1079 dest64 = (cq->mcq.cqe_sz == 64) ? dest : dest + 64;
1081 memcpy(dest, cqe, cq->mcq.cqe_sz);
1088 cq->mcq.cons_index += nfreed;
1093 mlx5_cq_set_ci(&cq->mcq);
1110 struct mlx5_ib_cq *mcq = to_mcq(cq);
1119 err = mlx5_core_modify_cq_moderation(dev->mdev, &mcq->mcq,
1122 mlx5_ib_warn(dev, "modify cq 0x%x failed\n", mcq->mcq.cqn);
1206 i = cq->mcq.cons_index;
1233 cq->mcq.cqn);
1237 ++cq->mcq.cons_index;
1326 MLX5_SET(modify_cq_in, in, cqn, cq->mcq.cqn);
1328 err = mlx5_core_modify_cq(dev->mdev, &cq->mcq, in, inlen);