Lines Matching defs:gl

212 cxgbit_copy_frags(struct sk_buff *skb, const struct pkt_gl *gl,
219 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page,
220 gl->frags[0].offset + offset,
221 gl->frags[0].size - offset);
222 for (i = 1; i < gl->nfrags; i++)
224 gl->frags[i].page,
225 gl->frags[i].offset,
226 gl->frags[i].size);
228 skb_shinfo(skb)->nr_frags += gl->nfrags;
231 get_page(gl->frags[gl->nfrags - 1].page);
235 cxgbit_lro_add_packet_gl(struct sk_buff *skb, u8 op, const struct pkt_gl *gl)
243 struct cpl_iscsi_hdr *cpl = (struct cpl_iscsi_hdr *)gl->va;
249 pdu_cb->hdr = gl->va + offset;
253 if (unlikely(gl->nfrags > 1))
258 struct cpl_iscsi_data *cpl = (struct cpl_iscsi_data *)gl->va;
265 pdu_cb->nr_dfrags = gl->nfrags;
271 cpl = (struct cpl_rx_iscsi_cmp *)gl->va;
275 pdu_cb->hdr = gl->va + offset;
281 if (unlikely(gl->nfrags > 1))
299 cxgbit_copy_frags(skb, gl, offset);
301 pdu_cb->frags += gl->nfrags;
309 cxgbit_lro_init_skb(struct cxgbit_sock *csk, u8 op, const struct pkt_gl *gl,
373 const struct pkt_gl *gl, struct t4_lro_mgr *lro_mgr,
393 skb = cxgbit_lro_init_skb(csk, op, gl, rsp, napi);
406 if ((gl && (((skb_shinfo(skb)->nr_frags + gl->nfrags) >
413 if (gl)
414 cxgbit_lro_add_packet_gl(skb, op, gl);
428 const struct pkt_gl *gl, struct t4_lro_mgr *lro_mgr,
455 rpl = gl ? (struct cpl_tx_data *)gl->va :
467 if (!gl) {
483 if (unlikely(op != *(u8 *)gl->va)) {
485 gl->va, be64_to_cpu(*rsp),
486 get_unaligned_be64(gl->va),
487 gl->tot_len);
493 if (!cxgbit_lro_receive(csk, op, rsp, gl, lro_mgr,
499 skb = cxgb4_pktgl_to_skb(gl, RX_PULL_LEN, RX_PULL_LEN);