Lines Matching defs:pfrag
262 struct page_frag *pfrag,
268 if (skb_frag_page(frag) == pfrag->page &&
269 skb_frag_off(frag) + skb_frag_size(frag) == pfrag->offset) {
273 skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
276 get_page(pfrag->page);
279 pfrag->offset += size;
320 struct page_frag *pfrag,
332 if (unlikely(pfrag->size - pfrag->offset < prot->tag_size) &&
333 !skb_page_frag_refill(prot->tag_size, pfrag, sk->sk_allocation)) {
336 pfrag = &dummy_tag_frag;
338 tls_append_frag(record, pfrag, prot->tag_size);
347 struct page_frag *pfrag,
358 skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
361 get_page(pfrag->page);
362 pfrag->offset += prepend_size;
372 struct page_frag *pfrag,
378 if (unlikely(!skb_page_frag_refill(prepend_size, pfrag,
385 ret = tls_create_new_record(offload_ctx, pfrag, prepend_size);
389 if (pfrag->size > pfrag->offset)
393 if (!sk_page_frag_refill(sk, pfrag))
434 struct page_frag *pfrag;
463 pfrag = sk_page_frag(sk);
471 rc = tls_do_allocation(sk, ctx, pfrag, prot->prepend_size);
523 copy = min_t(size_t, copy, pfrag->size - pfrag->offset);
525 rc = tls_device_copy_data(page_address(pfrag->page) +
526 pfrag->offset, copy,
530 tls_append_frag(record, pfrag, copy);
548 pfrag, record_type);