Lines Matching refs:size
122 * a sensible code size.
882 struct sk_buff *sk_stream_alloc_skb(struct sock *sk, int size, gfp_t gfp,
887 if (likely(!size)) {
900 size = ALIGN(size, 4);
905 skb = alloc_skb_fclone(size + sk->sk_prot->max_header, gfp);
918 * Make sure that we have exactly size bytes
921 skb->reserved_tailroom = skb->end - skb->tail - size;
985 size_t size, int flags)
1018 while (size > 0) {
1041 if (copy > size)
1042 copy = size;
1078 size -= copy;
1079 if (!size)
1127 size_t size, int flags)
1130 return sock_no_sendpage_locked(sk, page, offset, size, flags);
1134 return do_tcp_sendpages(sk, page, offset, size, flags);
1139 size_t size, int flags)
1144 ret = tcp_sendpage_locked(sk, page, offset, size, flags);
1160 int *copied, size_t size,
1181 tp->fastopen_req->size = size;
1207 int tcp_sendmsg_locked(struct sock *sk, struct msghdr *msg, size_t size)
1221 if (flags & MSG_ZEROCOPY && size && sock_flag(sk, SOCK_ZEROCOPY)) {
1223 uarg = sock_zerocopy_realloc(sk, size, skb_zcopy(skb));
1236 err = tcp_sendmsg_fastopen(sk, msg, &copied_syn, size, uarg);
1260 copied = tcp_send_rcvq(sk, msg, size);
1360 copy = min_t(int, copy, pfrag->size - pfrag->offset);
1460 int tcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t size)
1465 ret = tcp_sendmsg_locked(sk, msg, size);