Lines Matching refs:req

575 	struct cpl_abort_req *req = cplhdr(skb);
578 req->cmd = CPL_ABORT_NO_RST;
664 struct cpl_get_tcb *req;
665 int wrlen = roundup(sizeof(*req), 16);
667 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
672 req = (struct cpl_get_tcb *) skb_put(skb, wrlen);
673 memset(req, 0, wrlen);
674 INIT_TP_WR(req, ep->hwtid);
675 OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_GET_TCB, ep->hwtid));
676 req->reply_ctrl = htons(REPLY_CHAN_V(0) | QUEUENO_V(ep->rss_qid));
715 struct cpl_act_open_req *req = NULL;
828 req = skb_put(skb, wrlen);
829 INIT_TP_WR(req, 0);
834 req = (struct cpl_act_open_req *)t5req;
839 req = (struct cpl_act_open_req *)t6req;
849 OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ,
851 req->local_port = la->sin_port;
852 req->peer_port = ra->sin_port;
853 req->local_ip = la->sin_addr.s_addr;
854 req->peer_ip = ra->sin_addr.s_addr;
855 req->opt0 = cpu_to_be64(opt0);
858 req->params = cpu_to_be32(params);
859 req->opt2 = cpu_to_be32(opt2);
944 struct fw_ofld_tx_data_wr *req;
954 wrlen = roundup(mpalen + sizeof(*req), 16);
962 req = skb_put_zero(skb, wrlen);
963 req->op_to_immdlen = cpu_to_be32(
967 req->flowid_len16 = cpu_to_be32(
970 req->plen = cpu_to_be32(mpalen);
971 req->tunnel_to_proxy = cpu_to_be32(
975 mpa = (struct mpa_message *)(req + 1);
1048 struct fw_ofld_tx_data_wr *req;
1059 wrlen = roundup(mpalen + sizeof(*req), 16);
1068 req = skb_put_zero(skb, wrlen);
1069 req->op_to_immdlen = cpu_to_be32(
1073 req->flowid_len16 = cpu_to_be32(
1076 req->plen = cpu_to_be32(mpalen);
1077 req->tunnel_to_proxy = cpu_to_be32(
1081 mpa = (struct mpa_message *)(req + 1);
1128 struct fw_ofld_tx_data_wr *req;
1139 wrlen = roundup(mpalen + sizeof(*req), 16);
1148 req = skb_put_zero(skb, wrlen);
1149 req->op_to_immdlen = cpu_to_be32(
1153 req->flowid_len16 = cpu_to_be32(
1156 req->plen = cpu_to_be32(mpalen);
1157 req->tunnel_to_proxy = cpu_to_be32(
1161 mpa = (struct mpa_message *)(req + 1);
1217 struct cpl_act_establish *req = cplhdr(skb);
1218 unsigned short tcp_opt = ntohs(req->tcp_opt);
1219 unsigned int tid = GET_TID(req);
1220 unsigned int atid = TID_TID_G(ntohl(req->tos_atid));
1227 be32_to_cpu(req->snd_isn), be32_to_cpu(req->rcv_isn));
1237 ep->snd_seq = be32_to_cpu(req->snd_isn);
1238 ep->rcv_seq = be32_to_cpu(req->rcv_isn);
1961 struct fw_ofld_connection_wr *req;
1967 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
1971 req = __skb_put_zero(skb, sizeof(*req));
1972 req->op_compl = htonl(WR_OP_V(FW_OFLD_CONNECTION_WR));
1973 req->len16_pkd = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)));
1974 req->le.filter = cpu_to_be32(cxgb4_select_ntuple(
1978 req->le.lport = sin->sin_port;
1979 req->le.u.ipv4.lip = sin->sin_addr.s_addr;
1981 req->le.pport = sin->sin_port;
1982 req->le.u.ipv4.pip = sin->sin_addr.s_addr;
1983 req->tcb.t_state_to_astid =
1986 req->tcb.cplrxdataack_cplpassacceptrpl =
1988 req->tcb.tx_max = (__force __be32) jiffies;
1989 req->tcb.rcv_adv = htons(1);
2003 req->tcb.opt0 = (__force __be64) (TCAM_BYPASS_F |
2015 req->tcb.opt2 = (__force __be32) (PACE_V(1) |
2021 req->tcb.opt2 |= (__force __be32)TSTAMPS_EN_F;
2023 req->tcb.opt2 |= (__force __be32)SACK_EN_F;
2025 req->tcb.opt2 |= (__force __be32)WND_SCALE_EN_F;
2026 req->tcb.opt0 = cpu_to_be64((__force u64)req->tcb.opt0);
2027 req->tcb.opt2 = cpu_to_be32((__force u32)req->tcb.opt2);
2416 struct cpl_pass_accept_req *req)
2429 enable_tcp_timestamps && req->tcpopt.tstamp,
2454 if (enable_tcp_timestamps && req->tcpopt.tstamp)
2456 if (enable_tcp_sack && req->tcpopt.sack)
2462 u32 hlen = ntohl(req->hdr_len);
2465 tcph = (const void *)(req + 1) + ETH_HDR_LEN_G(hlen) +
2468 tcph = (const void *)(req + 1) +
2513 struct cpl_pass_accept_req *req = cplhdr(skb);
2514 unsigned int stid = PASS_OPEN_TID_G(ntohl(req->tos_stid));
2516 unsigned int hwtid = GET_TID(req);
2522 u16 peer_mss = ntohs(req->tcpopt.mss);
2542 tos = PASS_OPEN_TOS_G(ntohl(req->tos_stid));
2544 cxgb_get_4tuple(req, parent_ep->com.dev->rdev.lldi.adapter_type,
2590 ((enable_tcp_timestamps && req->tcpopt.tstamp) ? 12 : 0);
2651 if (accept_cr(child_ep, skb, req)) {
2676 struct cpl_pass_establish *req = cplhdr(skb);
2677 unsigned int tid = GET_TID(req);
2679 u16 tcp_opt = ntohs(req->tcp_opt);
2686 ep->snd_seq = be32_to_cpu(req->snd_isn);
2687 ep->rcv_seq = be32_to_cpu(req->rcv_isn);
2814 struct cpl_abort_req_rss6 *req = cplhdr(skb);
2820 unsigned int tid = GET_TID(req);
2830 status = ABORT_RSS_STATUS_G(be32_to_cpu(req->srqidx_status));
2892 be32_to_cpu(req->srqidx_status));
3746 struct cpl_fw6_msg_ofld_connection_wr_rpl *req)
3749 int atid = be32_to_cpu(req->tid);
3752 (__force u32) req->tid);
3756 switch (req->retval) {
3773 __func__, req->retval);
3777 req->retval, atid);
3781 connect_reply_upcall(ep, status2errno(req->retval));
3797 struct cpl_fw6_msg_ofld_connection_wr_rpl *req)
3803 rpl_skb = (struct sk_buff *)(unsigned long)req->cookie;
3804 if (req->retval) {
3805 pr_err("%s passive open failure %d\n", __func__, req->retval);
3814 (__force u32) req->tid)));
3899 struct cpl_fw6_msg_ofld_connection_wr_rpl *req;
3906 req = (struct cpl_fw6_msg_ofld_connection_wr_rpl *)rpl->data;
3907 switch (req->t_state) {
3909 active_ofld_conn_reply(dev, skb, req);
3912 passive_ofld_conn_reply(dev, skb, req);
3916 __func__, req->t_state);
3932 struct cpl_pass_accept_req *req;
3945 __skb_pull(skb, sizeof(*req) + sizeof(struct rss_header));
3955 req = __skb_push(skb, sizeof(*req));
3956 memset(req, 0, sizeof(*req));
3957 req->l2info = cpu_to_be16(SYN_INTF_V(intf) |
3964 req->hdr_len =
3970 req->hdr_len |= cpu_to_be32(TCP_HDR_LEN_V(tcp_hdr_len) |
3975 req->hdr_len |= cpu_to_be32(T6_TCP_HDR_LEN_V(tcp_hdr_len) |
3979 req->vlan = vlantag;
3980 req->len = len;
3981 req->tos_stid = cpu_to_be32(PASS_OPEN_TID_V(stid) |
3983 req->tcpopt.mss = htons(tmp_opt.mss_clamp);
3985 req->tcpopt.wsf = tmp_opt.snd_wscale;
3986 req->tcpopt.tstamp = tmp_opt.saw_tstamp;
3988 req->tcpopt.sack = 1;
3989 OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_PASS_ACCEPT_REQ, 0));
4000 struct fw_ofld_connection_wr *req;
4007 req = __skb_put_zero(req_skb, sizeof(*req));
4008 req->op_compl = htonl(WR_OP_V(FW_OFLD_CONNECTION_WR) | FW_WR_COMPL_F);
4009 req->len16_pkd = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)));
4010 req->le.version_cpl = htonl(FW_OFLD_CONNECTION_WR_CPL_F);
4011 req->le.filter = (__force __be32) filter;
4012 req->le.lport = lport;
4013 req->le.pport = rport;
4014 req->le.u.ipv4.lip = laddr;
4015 req->le.u.ipv4.pip = raddr;
4016 req->tcb.rcv_nxt = htonl(rcv_isn + 1);
4017 req->tcb.rcv_adv = htons(window);
4018 req->tcb.t_state_to_astid =
4028 req->tcb.opt2 = htonl(RSS_QUEUE_V(rss_qid));
4036 req->tcb.opt0 = cpu_to_be64(MSS_IDX_V(0xF));
4037 req->cookie = (uintptr_t)skb;
4066 struct cpl_pass_accept_req *req = (void *)(rss + 1);
4117 eh = (struct ethhdr *)(req + 1);
4120 vlan_eh = (struct vlan_ethhdr *)(req + 1);
4407 struct cpl_abort_req_rss *req = cplhdr(skb);
4409 unsigned int tid = GET_TID(req);
4418 if (cxgb_is_neg_adv(req->status)) {
4420 ep->hwtid, req->status,
4421 neg_adv_str(req->status));