Lines Matching refs:err_cqe
62 static void dump_err_cqe(struct c4iw_dev *dev, struct t4_cqe *err_cqe)
64 __be64 *p = (void *)err_cqe;
69 CQE_QPID(err_cqe), CQE_OPCODE(err_cqe),
70 CQE_STATUS(err_cqe), CQE_TYPE(err_cqe), ntohl(err_cqe->len),
71 CQE_WRID_HI(err_cqe), CQE_WRID_LOW(err_cqe));
82 if (RQ_TYPE(err_cqe) && (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE ||
83 CQE_OPCODE(err_cqe) == FW_RI_READ_RESP))
84 print_tpte(dev, CQE_WRID_STAG(err_cqe));
89 struct t4_cqe *err_cqe,
96 dump_err_cqe(dev, err_cqe);
120 void c4iw_ev_dispatch(struct c4iw_dev *dev, struct t4_cqe *err_cqe)
127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe));
130 CQE_QPID(err_cqe),
131 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe),
132 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe),
133 CQE_WRID_LOW(err_cqe));
138 if (SQ_TYPE(err_cqe))
145 cqid, CQE_QPID(err_cqe),
146 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe),
147 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe),
148 CQE_WRID_LOW(err_cqe));
158 if (RQ_TYPE(err_cqe) &&
159 (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE)) {
160 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_REQ_ERR);
164 switch (CQE_STATUS(err_cqe)) {
179 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_ACCESS_ERR);
186 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_DEVICE_FATAL);
206 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL);
211 CQE_STATUS(err_cqe), qhp->wq.sq.qid);
212 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL);