Lines Matching refs:psn

130 			     struct rvt_qp *qp, u32 psn, int diff, bool fecn);
348 qpriv->flow_state.psn = 0;
801 fs->psn = 0;
828 fs->psn = 0;
1515 flow->flow_state.spsn = qpriv->flow_state.psn;
1519 qpriv->flow_state.psn += flow->npkts;
1675 u32 psn, u16 *fidx)
1685 if (cmp_psn(psn, flow->flow_state.ib_spsn) >= 0 &&
1686 cmp_psn(psn, flow->flow_state.ib_lpsn) <= 0) {
1710 /* This is the IB psn used to send the request */
1793 trace_hfi1_tid_req_build_read_req(qp, 0, wqe->wr.opcode, wqe->psn,
1832 if ((qpriv->flow_state.psn + npkts) > MAX_TID_FLOW_PSN - 1) {
1868 /* Calculate the next segment start psn.*/
1886 u32 bth0, u32 psn, u64 vaddr, u32 len)
1944 flow->flow_state.ib_spsn = psn;
1958 e->psn = psn;
1959 e->lpsn = psn + flow->npkts - 1;
1971 req->r_flow_psn = e->psn;
1973 trace_hfi1_tid_req_rcv_read_req(qp, 0, e->opcode, e->psn, e->lpsn,
1980 struct rvt_qp *qp, u32 psn, int diff)
1992 trace_hfi1_rsp_tid_rcv_error(qp, psn);
1993 trace_hfi1_tid_rdma_rcv_err(qp, 0, psn, diff);
2008 e = find_prev_entry(qp, psn, &prev, NULL, &old_req);
2014 req->r_flow_psn = psn;
2015 trace_hfi1_tid_req_rcv_err(qp, 0, e->opcode, e->psn, e->lpsn, req);
2030 if (psn != e->psn || len != req->total_len)
2055 if (tid_rdma_rcv_read_request(qp, e, packet, ohdr, bth0, psn,
2126 delta_psn(psn, fstate->resp_ib_psn),
2129 delta_psn(psn, fstate->resp_ib_psn);
2139 req->cur_seg = delta_psn(psn, e->psn);
2155 trace_hfi1_tid_req_rcv_err(qp, 0, e->opcode, e->psn,
2188 * It's possible to receive a retry psn that is earlier than an RNRNAK
2189 * psn. In this case, the rnrnak state should be cleared.
2231 u32 bth0, psn, len, rkey;
2243 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
2244 trace_hfi1_rsp_rcv_tid_read_req(qp, psn);
2259 diff = delta_psn(psn, qp->r_psn);
2261 tid_rdma_rcv_err(packet, ohdr, qp, psn, diff, fecn);
2288 if (tid_rdma_rcv_read_request(qp, e, packet, ohdr, bth0, psn, vaddr,
2300 qp->r_psn += e->lpsn - e->psn + 1;
2408 find_tid_request(struct rvt_qp *qp, u32 psn, enum ib_wr_opcode opcode)
2420 if (cmp_psn(psn, wqe->psn) >= 0 &&
2421 cmp_psn(psn, wqe->lpsn) <= 0) {
2522 req->e.swqe->psn, req->e.swqe->lpsn,
2650 u8 rte, u32 psn, u32 ibpsn)
2671 /* If the psn is out of valid range, drop the packet */
2734 trace_hfi1_tid_req_read_kdeth_eflags(qp, 0, wqe->wr.opcode, wqe->psn,
2754 diff = cmp_psn(psn,
2780 if (cmp_psn(fpsn, psn) == 0) {
2787 mask_psn(psn + 1);
2851 u32 qp_num, psn, ibpsn;
2876 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
2916 ret = handle_read_kdeth_eflags(rcd, packet, rcv_type, rte, psn,
2938 trace_hfi1_eflags_err_write(qp, rcv_type, rte, psn);
2939 trace_hfi1_rsp_handle_kdeth_eflags(qp, psn);
2941 trace_hfi1_tid_req_handle_kdeth_eflags(qp, 0, e->opcode, e->psn,
2966 diff = cmp_psn(psn,
2979 if (psn == full_flow_psn(flow,
2983 mask_psn(psn + 1);
3055 wqe->psn, wqe->lpsn,
3123 trace_hfi1_tid_req_restart_req(qp, 0, wqe->wr.opcode, wqe->psn,
3323 wqe->lpsn = wqe->psn;
3345 wqe->psn, wqe->lpsn,
3498 trace_hfi1_tid_req_write_alloc_res(qp, 0, e->opcode, e->psn,
3536 if (qpriv->flow_state.psn + npkts > MAX_TID_FLOW_PSN - 1) {
3590 qp->r_psn = e->psn + req->alloc_seg;
3656 u32 bth0, psn, len, rkey, num_segs;
3667 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
3668 trace_hfi1_rsp_rcv_tid_write_req(qp, psn);
3681 diff = delta_psn(psn, qp->r_psn);
3683 tid_rdma_rcv_err(packet, ohdr, qp, psn, diff, fecn);
3741 e->psn = psn;
3755 req->r_flow_psn = e->psn;
3773 trace_hfi1_tid_req_rcv_write_req(qp, 0, e->opcode, e->psn, e->lpsn,
3836 trace_hfi1_tid_req_build_write_resp(qp, 0, e->opcode, e->psn, e->lpsn,
4036 u32 opcode, aeth, psn, flow_psn, i, tidlen = 0, pktlen;
4041 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
4048 if (cmp_psn(psn, qp->s_next_psn) >= 0)
4052 if (unlikely(cmp_psn(psn, qp->s_last_psn) <= 0))
4064 if (cmp_psn(psn, qp->s_last_psn + 1) != 0)
4089 if (!do_rc_ack(qp, aeth, psn, opcode, 0, rcd))
4092 trace_hfi1_ack(qp, psn);
4106 flow->flow_state.resp_ib_psn = psn;
4143 trace_hfi1_tid_req_rcv_write_resp(qp, 0, wqe->wr.opcode, wqe->psn,
4149 if (!cmp_psn(psn, wqe->psn)) {
4150 req->r_last_acked = mask_psn(wqe->psn - 1);
4268 u32 psn, next;
4273 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
4284 if (cmp_psn(psn, full_flow_psn(flow, flow->flow_state.lpsn))) {
4287 if (cmp_psn(psn, flow->flow_state.r_next_psn))
4290 flow->flow_state.r_next_psn = mask_psn(psn + 1);
4311 len += delta_psn(psn,
4329 priv->r_next_psn_kdeth = mask_psn(psn + 1);
4334 flow->flow_state.r_next_psn = mask_psn(psn + 1);
4337 rcd->flows[flow->idx].psn = psn & HFI1_KDETH_BTH_SEQ_MASK;
4348 trace_hfi1_rsp_rcv_tid_write_data(qp, psn);
4349 trace_hfi1_tid_req_rcv_write_data(qp, 0, e->opcode, e->psn, e->lpsn,
4408 static bool hfi1_tid_rdma_is_resync_psn(u32 psn)
4410 return (bool)((psn & HFI1_KDETH_BTH_SEQ_MASK) ==
4493 u32 aeth, psn, req_psn, ack_psn, flpsn, resync_psn, ack_kpsn;
4499 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
4505 trace_hfi1_rcv_tid_ack(qp, aeth, psn, req_psn, resync_psn);
4509 cmp_psn(psn, qpriv->s_resync_psn))
4513 if (hfi1_tid_rdma_is_resync_psn(psn))
4516 ack_kpsn = psn;
4531 trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
4537 if (cmp_psn(psn, full_flow_psn(flow, flow->flow_state.spsn)) < 0 ||
4548 trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
4567 trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
4573 if (!hfi1_tid_rdma_is_resync_psn(psn)) {
4614 psn = mask_psn(psn + 1);
4615 generation = psn >> HFI1_KDETH_BTH_SEQ_SHIFT;
4636 req->r_ack_psn = psn;
4713 if (cmp_psn(psn, flpsn) > 0)
4812 qp, 0, wqe->wr.opcode, wqe->psn, wqe->lpsn, req);
4868 u32 psn, generation, idx, gen_next;
4873 psn = mask_psn(be32_to_cpu(ohdr->bth[2]));
4875 generation = mask_psn(psn + 1) >> HFI1_KDETH_BTH_SEQ_SHIFT;
4903 fs->psn = 0;
4924 trace_hfi1_tid_req_rcv_resync(qp, 0, e->opcode, e->psn,
4941 flow->flow_state.spsn = fs->psn;
4947 fs->psn += flow->npkts;
5058 trace_hfi1_tid_req_make_tid_pkt(qp, 0, wqe->wr.opcode, wqe->psn,
5097 wqe->psn, wqe->lpsn, req);
5219 trace_hfi1_rsp_make_tid_ack(qp, e->psn);
5220 trace_hfi1_tid_req_make_tid_ack(qp, 0, e->opcode, e->psn, e->lpsn,
5294 trace_hfi1_tid_req_make_tid_ack(qp, 0, e->opcode, e->psn, e->lpsn,
5491 struct rvt_qp *qp, u32 psn, int diff, bool fecn)
5495 tid_rdma_rcv_error(packet, ohdr, qp, psn, diff);