Lines Matching refs:chp
87 static void post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp,
105 event.device = chp->ibcq.device;
107 event.element.cq = &chp->ibcq;
113 if (t4_clear_cq_armed(&chp->cq)) {
114 spin_lock_irqsave(&chp->comp_handler_lock, flag);
115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context);
116 spin_unlock_irqrestore(&chp->comp_handler_lock, flag);
122 struct c4iw_cq *chp;
142 chp = get_chp(dev, cqid);
143 if (!chp) {
154 atomic_inc(&chp->refcnt);
160 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_REQ_ERR);
179 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_ACCESS_ERR);
186 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_DEVICE_FATAL);
206 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL);
212 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL);
216 if (atomic_dec_and_test(&chp->refcnt))
217 wake_up(&chp->wait);
225 struct c4iw_cq *chp;
229 chp = xa_load(&dev->cqs, qid);
230 if (chp) {
231 atomic_inc(&chp->refcnt);
233 t4_clear_cq_armed(&chp->cq);
234 spin_lock_irqsave(&chp->comp_handler_lock, flag);
235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context);
236 spin_unlock_irqrestore(&chp->comp_handler_lock, flag);
237 if (atomic_dec_and_test(&chp->refcnt))
238 wake_up(&chp->wait);