/kernel/linux/linux-5.10/net/core/ |
H A D | stream.c | 211 WARN_ON(sk->sk_wmem_queued); in sk_stream_kill_queues()
|
H A D | sock.c | 1923 newsk->sk_wmem_queued = 0; in sk_clone_lock() 2648 if (sk->sk_wmem_queued < wmem0) in __sk_mem_raise_allocated() 2662 sk_mem_pages(sk->sk_wmem_queued + in __sk_mem_raise_allocated() 2676 if (sk->sk_wmem_queued + size >= sk->sk_sndbuf) in __sk_mem_raise_allocated() 2700 * memory_pressure use sk_wmem_queued as write buffer accounting. 3332 mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued); in sk_get_meminfo()
|
/kernel/linux/linux-6.6/net/core/ |
H A D | stream.c | 213 WARN_ON_ONCE(sk->sk_wmem_queued); in sk_stream_kill_queues()
|
H A D | sock.c | 2339 newsk->sk_wmem_queued = 0; in sk_clone_lock() 3088 if (sk->sk_wmem_queued < wmem0) in __sk_mem_raise_allocated() 3102 sk_mem_pages(sk->sk_wmem_queued + in __sk_mem_raise_allocated() 3116 if (sk->sk_wmem_queued + size >= sk->sk_sndbuf) { in __sk_mem_raise_allocated() 3145 * memory_pressure use sk_wmem_queued as write buffer accounting. 3756 mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued); in sk_get_meminfo()
|
/kernel/linux/linux-5.10/include/net/ |
H A D | sock.h | 275 * @sk_wmem_queued: persistent queue size 440 int sk_wmem_queued; member 978 return READ_ONCE(sk->sk_wmem_queued) >> 1; in sk_stream_min_wspace() 983 return READ_ONCE(sk->sk_sndbuf) - READ_ONCE(sk->sk_wmem_queued); in sk_stream_wspace() 988 WRITE_ONCE(sk->sk_wmem_queued, sk->sk_wmem_queued + val); in sk_wmem_queued_add() 1319 if (READ_ONCE(sk->sk_wmem_queued) >= READ_ONCE(sk->sk_sndbuf)) in __sk_stream_memory_free() 2401 val = min(sk->sk_sndbuf, sk->sk_wmem_queued >> 1); in sk_stream_moderate_sndbuf()
|
H A D | tcp.h | 292 if (sk->sk_wmem_queued > SOCK_MIN_SNDBUF && in tcp_out_of_memory()
|
/kernel/linux/linux-6.6/include/net/ |
H A D | sock.h | 270 * @sk_wmem_queued: persistent queue size 450 int sk_wmem_queued; member 1043 return READ_ONCE(sk->sk_wmem_queued) >> 1; in sk_stream_min_wspace() 1048 return READ_ONCE(sk->sk_sndbuf) - READ_ONCE(sk->sk_wmem_queued); in sk_stream_wspace() 1053 WRITE_ONCE(sk->sk_wmem_queued, sk->sk_wmem_queued + val); in sk_wmem_queued_add() 1389 if (READ_ONCE(sk->sk_wmem_queued) >= READ_ONCE(sk->sk_sndbuf)) in __sk_stream_memory_free() 1651 unused_mem = sk->sk_reserved_mem - sk->sk_wmem_queued - in sk_unused_reserved_mem() 2574 val = min(sk->sk_sndbuf, sk->sk_wmem_queued >> 1); in sk_stream_moderate_sndbuf()
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | sock.h | 121 __entry->wmem_queued = READ_ONCE(sk->sk_wmem_queued);
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | mptcp.h | 54 __entry->ratio = div_u64((u64)ssk->sk_wmem_queued << 32, __entry->pace);
|
H A D | sock.h | 121 __entry->wmem_queued = READ_ONCE(sk->sk_wmem_queued);
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_io.c | 709 sk->sk_wmem_queued -= total_size; in chtls_push_frames() 827 sk->sk_wmem_queued += skb->truesize; in skb_entail() 901 sk->sk_wmem_queued += copy; in chtls_skb_copy_to_page_nocache() 907 return (cdev->max_host_sndbuf - sk->sk_wmem_queued > 0); in csk_mem_free() 1217 sk->sk_wmem_queued -= skb->truesize; in chtls_sendmsg() 1297 sk->sk_wmem_queued += copy; in chtls_sendpage()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_io.c | 709 sk->sk_wmem_queued -= total_size; in chtls_push_frames() 827 sk->sk_wmem_queued += skb->truesize; in skb_entail() 901 sk->sk_wmem_queued += copy; in chtls_skb_copy_to_page_nocache() 907 return (cdev->max_host_sndbuf - sk->sk_wmem_queued > 0); in csk_mem_free() 1230 sk->sk_wmem_queued -= skb->truesize; in chtls_sendmsg()
|
/kernel/linux/linux-5.10/net/sctp/ |
H A D | proc.c | 287 READ_ONCE(sk->sk_wmem_queued), in sctp_assocs_seq_show()
|
H A D | diag.c | 173 mem[SK_MEMINFO_WMEM_QUEUED] = sk->sk_wmem_queued; in inet_sctp_diag_fill()
|
/kernel/linux/linux-6.6/net/sctp/ |
H A D | proc.c | 285 READ_ONCE(sk->sk_wmem_queued), in sctp_assocs_seq_show()
|
H A D | diag.c | 173 mem[SK_MEMINFO_WMEM_QUEUED] = sk->sk_wmem_queued; in inet_sctp_diag_fill()
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
H A D | tcp_nip_output.c | 372 sk->sk_wmem_queued += skb->truesize; in tcp_nip_connect_queue_skb() 616 sk->sk_wmem_queued += skb->truesize; in tcp_nip_queue_skb()
|
H A D | tcp_nip.c | 477 WARN_ON(sk->sk_wmem_queued); in sk_nip_stream_kill_queues() 1425 sk->sk_wmem_queued += skb->truesize; in skb_nip_entail()
|
/kernel/linux/linux-5.10/net/sched/ |
H A D | em_meta.c | 449 dst->value = READ_ONCE(sk->sk_wmem_queued); in META_COLLECTOR()
|
/kernel/linux/linux-6.6/net/sched/ |
H A D | em_meta.c | 452 dst->value = READ_ONCE(sk->sk_wmem_queued); in META_COLLECTOR()
|
/kernel/linux/linux-5.10/net/kcm/ |
H A D | kcmsock.c | 707 sk->sk_wmem_queued -= sent; in kcm_write_msgs() 848 sk->sk_wmem_queued += size; in kcm_sendpage()
|
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_output.c | 1549 /* tcp_sendmsg() can overshoot sk_wmem_queued by one full size skb. in tcp_fragment() 1555 if (unlikely((sk->sk_wmem_queued >> 1) > limit && in tcp_fragment() 3169 min_t(u32, sk->sk_wmem_queued + (sk->sk_wmem_queued >> 2), in __tcp_retransmit_skb()
|
H A D | inet_diag.c | 274 .idiag_wmem = READ_ONCE(sk->sk_wmem_queued), in inet_sk_diag_fill()
|
/kernel/linux/linux-6.6/net/mptcp/ |
H A D | protocol.c | 1451 linger_time = div_u64((u64)READ_ONCE(ssk->sk_wmem_queued) << 32, pace); in mptcp_subflow_get_send() 1479 wmem = READ_ONCE(ssk->sk_wmem_queued); in mptcp_subflow_get_send()
|
/kernel/linux/linux-5.10/drivers/block/drbd/ |
H A D | drbd_worker.c | 636 int queued = sk->sk_wmem_queued; in make_resync_request()
|