Lines Matching refs:req
575 struct cpl_abort_req *req = cplhdr(skb);
578 req->cmd = CPL_ABORT_NO_RST;
664 struct cpl_get_tcb *req;
665 int wrlen = roundup(sizeof(*req), 16);
667 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
672 req = (struct cpl_get_tcb *) skb_put(skb, wrlen);
673 memset(req, 0, wrlen);
674 INIT_TP_WR(req, ep->hwtid);
675 OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_GET_TCB, ep->hwtid));
676 req->reply_ctrl = htons(REPLY_CHAN_V(0) | QUEUENO_V(ep->rss_qid));
715 struct cpl_act_open_req *req = NULL;
828 req = skb_put(skb, wrlen);
829 INIT_TP_WR(req, 0);
834 req = (struct cpl_act_open_req *)t5req;
839 req = (struct cpl_act_open_req *)t6req;
849 OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ,
851 req->local_port = la->sin_port;
852 req->peer_port = ra->sin_port;
853 req->local_ip = la->sin_addr.s_addr;
854 req->peer_ip = ra->sin_addr.s_addr;
855 req->opt0 = cpu_to_be64(opt0);
858 req->params = cpu_to_be32(params);
859 req->opt2 = cpu_to_be32(opt2);
944 struct fw_ofld_tx_data_wr *req;
954 wrlen = roundup(mpalen + sizeof(*req), 16);
962 req = skb_put_zero(skb, wrlen);
963 req->op_to_immdlen = cpu_to_be32(
967 req->flowid_len16 = cpu_to_be32(
970 req->plen = cpu_to_be32(mpalen);
971 req->tunnel_to_proxy = cpu_to_be32(
975 mpa = (struct mpa_message *)(req + 1);
1048 struct fw_ofld_tx_data_wr *req;
1059 wrlen = roundup(mpalen + sizeof(*req), 16);
1068 req = skb_put_zero(skb, wrlen);
1069 req->op_to_immdlen = cpu_to_be32(
1073 req->flowid_len16 = cpu_to_be32(
1076 req->plen = cpu_to_be32(mpalen);
1077 req->tunnel_to_proxy = cpu_to_be32(
1081 mpa = (struct mpa_message *)(req + 1);
1128 struct fw_ofld_tx_data_wr *req;
1139 wrlen = roundup(mpalen + sizeof(*req), 16);
1148 req = skb_put_zero(skb, wrlen);
1149 req->op_to_immdlen = cpu_to_be32(
1153 req->flowid_len16 = cpu_to_be32(
1156 req->plen = cpu_to_be32(mpalen);
1157 req->tunnel_to_proxy = cpu_to_be32(
1161 mpa = (struct mpa_message *)(req + 1);
1217 struct cpl_act_establish *req = cplhdr(skb);
1218 unsigned short tcp_opt = ntohs(req->tcp_opt);
1219 unsigned int tid = GET_TID(req);
1220 unsigned int atid = TID_TID_G(ntohl(req->tos_atid));
1227 be32_to_cpu(req->snd_isn), be32_to_cpu(req->rcv_isn));
1237 ep->snd_seq = be32_to_cpu(req->snd_isn);
1238 ep->rcv_seq = be32_to_cpu(req->rcv_isn);
1961 struct fw_ofld_connection_wr *req;
1967 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
1971 req = __skb_put_zero(skb, sizeof(*req));
1972 req->op_compl = htonl(WR_OP_V(FW_OFLD_CONNECTION_WR));
1973 req->len16_pkd = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)));
1974 req->le.filter = cpu_to_be32(cxgb4_select_ntuple(
1978 req->le.lport = sin->sin_port;
1979 req->le.u.ipv4.lip = sin->sin_addr.s_addr;
1981 req->le.pport = sin->sin_port;
1982 req->le.u.ipv4.pip = sin->sin_addr.s_addr;
1983 req->tcb.t_state_to_astid =
1986 req->tcb.cplrxdataack_cplpassacceptrpl =
1988 req->tcb.tx_max = (__force __be32) jiffies;
1989 req->tcb.rcv_adv = htons(1);
2003 req->tcb.opt0 = (__force __be64) (TCAM_BYPASS_F |
2015 req->tcb.opt2 = (__force __be32) (PACE_V(1) |
2021 req->tcb.opt2 |= (__force __be32)TSTAMPS_EN_F;
2023 req->tcb.opt2 |= (__force __be32)SACK_EN_F;
2025 req->tcb.opt2 |= (__force __be32)WND_SCALE_EN_F;
2026 req->tcb.opt0 = cpu_to_be64((__force u64)req->tcb.opt0);
2027 req->tcb.opt2 = cpu_to_be32((__force u32)req->tcb.opt2);
2416 struct cpl_pass_accept_req *req)
2429 enable_tcp_timestamps && req->tcpopt.tstamp,
2454 if (enable_tcp_timestamps && req->tcpopt.tstamp)
2456 if (enable_tcp_sack && req->tcpopt.sack)
2462 u32 hlen = ntohl(req->hdr_len);
2465 tcph = (const void *)(req + 1) + ETH_HDR_LEN_G(hlen) +
2468 tcph = (const void *)(req + 1) +
2519 struct cpl_pass_accept_req *req = cplhdr(skb);
2520 unsigned int stid = PASS_OPEN_TID_G(ntohl(req->tos_stid));
2522 unsigned int hwtid = GET_TID(req);
2528 u16 peer_mss = ntohs(req->tcpopt.mss);
2548 tos = PASS_OPEN_TOS_G(ntohl(req->tos_stid));
2550 cxgb_get_4tuple(req, parent_ep->com.dev->rdev.lldi.adapter_type,
2596 ((enable_tcp_timestamps && req->tcpopt.tstamp) ? 12 : 0);
2657 if (accept_cr(child_ep, skb, req)) {
2682 struct cpl_pass_establish *req = cplhdr(skb);
2683 unsigned int tid = GET_TID(req);
2685 u16 tcp_opt = ntohs(req->tcp_opt);
2689 ep->snd_seq = be32_to_cpu(req->snd_isn);
2690 ep->rcv_seq = be32_to_cpu(req->rcv_isn);
2817 struct cpl_abort_req_rss6 *req = cplhdr(skb);
2823 unsigned int tid = GET_TID(req);
2833 status = ABORT_RSS_STATUS_G(be32_to_cpu(req->srqidx_status));
2895 be32_to_cpu(req->srqidx_status));
3749 struct cpl_fw6_msg_ofld_connection_wr_rpl *req)
3752 int atid = be32_to_cpu(req->tid);
3755 (__force u32) req->tid);
3759 switch (req->retval) {
3776 __func__, req->retval);
3780 req->retval, atid);
3784 connect_reply_upcall(ep, status2errno(req->retval));
3800 struct cpl_fw6_msg_ofld_connection_wr_rpl *req)
3806 rpl_skb = (struct sk_buff *)(unsigned long)req->cookie;
3807 if (req->retval) {
3808 pr_err("%s passive open failure %d\n", __func__, req->retval);
3817 (__force u32) req->tid)));
3902 struct cpl_fw6_msg_ofld_connection_wr_rpl *req;
3909 req = (struct cpl_fw6_msg_ofld_connection_wr_rpl *)rpl->data;
3910 switch (req->t_state) {
3912 active_ofld_conn_reply(dev, skb, req);
3915 passive_ofld_conn_reply(dev, skb, req);
3919 __func__, req->t_state);
3935 struct cpl_pass_accept_req *req;
3948 __skb_pull(skb, sizeof(*req) + sizeof(struct rss_header));
3958 req = __skb_push(skb, sizeof(*req));
3959 memset(req, 0, sizeof(*req));
3960 req->l2info = cpu_to_be16(SYN_INTF_V(intf) |
3967 req->hdr_len =
3973 req->hdr_len |= cpu_to_be32(TCP_HDR_LEN_V(tcp_hdr_len) |
3978 req->hdr_len |= cpu_to_be32(T6_TCP_HDR_LEN_V(tcp_hdr_len) |
3982 req->vlan = vlantag;
3983 req->len = len;
3984 req->tos_stid = cpu_to_be32(PASS_OPEN_TID_V(stid) |
3986 req->tcpopt.mss = htons(tmp_opt.mss_clamp);
3988 req->tcpopt.wsf = tmp_opt.snd_wscale;
3989 req->tcpopt.tstamp = tmp_opt.saw_tstamp;
3991 req->tcpopt.sack = 1;
3992 OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_PASS_ACCEPT_REQ, 0));
4003 struct fw_ofld_connection_wr *req;
4010 req = __skb_put_zero(req_skb, sizeof(*req));
4011 req->op_compl = htonl(WR_OP_V(FW_OFLD_CONNECTION_WR) | FW_WR_COMPL_F);
4012 req->len16_pkd = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)));
4013 req->le.version_cpl = htonl(FW_OFLD_CONNECTION_WR_CPL_F);
4014 req->le.filter = (__force __be32) filter;
4015 req->le.lport = lport;
4016 req->le.pport = rport;
4017 req->le.u.ipv4.lip = laddr;
4018 req->le.u.ipv4.pip = raddr;
4019 req->tcb.rcv_nxt = htonl(rcv_isn + 1);
4020 req->tcb.rcv_adv = htons(window);
4021 req->tcb.t_state_to_astid =
4031 req->tcb.opt2 = htonl(RSS_QUEUE_V(rss_qid));
4039 req->tcb.opt0 = cpu_to_be64(MSS_IDX_V(0xF));
4040 req->cookie = (uintptr_t)skb;
4069 struct cpl_pass_accept_req *req = (void *)(rss + 1);
4120 eh = (struct ethhdr *)(req + 1);
4123 vlan_eh = (struct vlan_ethhdr *)(req + 1);
4406 struct cpl_abort_req_rss *req = cplhdr(skb);
4408 unsigned int tid = GET_TID(req);
4417 if (cxgb_is_neg_adv(req->status)) {
4419 ep->hwtid, req->status,
4420 neg_adv_str(req->status));