Lines Matching defs:iwqp

239  * @iwqp: iwarp qp to modify
245 void i40iw_next_iw_state(struct i40iw_qp *iwqp,
266 if (iwqp->sc_qp.term_flags && (state == I40IW_QP_STATE_ERROR))
268 iwqp->hw_iwarp_state = state;
269 i40iw_hw_modify_qp(iwqp->iwdev, iwqp, &info, 0);
284 struct i40iw_qp *iwqp = NULL;
308 iwqp = iwdev->qp_table[info->qp_cq_id];
309 if (!iwqp) {
316 i40iw_qp_add_ref(&iwqp->ibqp);
318 qp = &iwqp->sc_qp;
319 spin_lock_irqsave(&iwqp->lock, flags);
320 iwqp->hw_tcp_state = info->tcp_state;
321 iwqp->hw_iwarp_state = info->iwarp_state;
322 iwqp->last_aeq = info->ae_id;
323 spin_unlock_irqrestore(&iwqp->lock, flags);
324 ctx_info = &iwqp->ctx_info;
335 if (atomic_inc_return(&iwqp->close_timer_started) == 1) {
336 iwqp->hw_tcp_state = I40IW_TCP_STATE_CLOSE_WAIT;
337 if ((iwqp->hw_tcp_state == I40IW_TCP_STATE_CLOSE_WAIT) &&
338 (iwqp->ibqp_state == IB_QPS_RTS)) {
339 i40iw_next_iw_state(iwqp,
341 i40iw_cm_disconn(iwqp);
343 iwqp->cm_id->add_ref(iwqp->cm_id);
344 i40iw_schedule_cm_timer(iwqp->cm_node,
345 (struct i40iw_puda_buf *)iwqp,
353 i40iw_cm_disconn(iwqp);
357 i40iw_next_iw_state(iwqp, I40IW_QP_STATE_ERROR, 1, 0, 0);
358 i40iw_cm_disconn(iwqp);
361 if (atomic_read(&iwqp->close_timer_started))
363 i40iw_cm_disconn(iwqp);
366 i40iw_qp_suspend_resume(dev, &iwqp->sc_qp, false);
421 ret = dev->iw_priv_qp_ops->qp_setctx(&iwqp->sc_qp,
422 iwqp->host_ctx.va,
429 i40iw_qp_rem_ref(&iwqp->ibqp);
677 struct i40iw_qp *iwqp = (struct i40iw_qp *)qp->back_qp;
694 complete(&iwqp->sq_drained);
695 complete(&iwqp->rq_drained);
701 complete(&iwqp->sq_drained);
704 complete(&iwqp->rq_drained);
709 complete(&iwqp->sq_drained);
710 complete(&iwqp->rq_drained);
832 * @iwqp: qp to flush wqes
834 void i40iw_flush_wqes(struct i40iw_device *iwdev, struct i40iw_qp *iwqp)
839 struct i40iw_sc_qp *qp = &iwqp->sc_qp;
850 (void)i40iw_hw_flush_wqes(iwdev, &iwqp->sc_qp, &info, true);