/kernel/linux/linux-5.10/include/net/ |
H A D | inet_connection_sock.h | 160 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 167 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca() 191 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 196 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 201 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 209 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() 223 pr_debug("inet_csk BUG: unknown timer value\n"); in inet_csk_clear_xmit_timer() 275 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() 298 pr_debug("inet_csk BU in inet_csk_reset_xmit_timer() [all...] |
/kernel/linux/linux-6.6/include/net/ |
H A D | inet_connection_sock.h | 148 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 155 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca() 179 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 184 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 189 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 197 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() 211 pr_debug("inet_csk BUG: unknown timer value\n"); in inet_csk_clear_xmit_timer() 222 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() 240 pr_debug("inet_csk BU in inet_csk_reset_xmit_timer() [all...] |
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
H A D | tcp_nip_timer.c | 54 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_delack_timer_handler() 114 inet_csk(sk)->icsk_retransmits, boundary); in retransmits_nip_timed_out() 115 return inet_csk(sk)->icsk_retransmits > boundary; in retransmits_nip_timed_out() 120 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_write_timeout() 151 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_retransmit_timer() 218 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_probe_timer() 267 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_write_timer_handler() 314 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_keepalive_is_timeout()
|
H A D | tcp_nip_output.c | 85 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_nip_mtu_to_mss() 131 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_event_new_data_sent() 148 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) { in tcp_nip_check_probe_timer() 153 } else if (inet_csk(sk)->icsk_pending != ICSK_TIME_PROBE0) { in tcp_nip_check_probe_timer() 155 tcp_sk(sk)->packets_out, inet_csk(sk)->icsk_pending); in tcp_nip_check_probe_timer() 171 struct inet_connection_sock *icsk = inet_csk(sk); in __nip_tcp_select_window() 340 inet_csk(sk)->icsk_rto = get_nip_rto() == 0 ? TCP_TIMEOUT_INIT : (HZ / get_nip_rto()); in tcp_nip_connect_init() 341 inet_csk(sk)->icsk_retransmits = 0; in tcp_nip_connect_init() 488 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_nip_transmit_skb() 655 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, inet_csk(s in __tcp_nip_connect() [all...] |
H A D | tcp_nip_input.c | 151 inet_csk(sk)->icsk_ack.pingpong = 1; in tcp_nip_fin() 538 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_data_queue() 660 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_send_delayed_ack() 692 inet_csk(sk)->icsk_ack.rcv_mss = tcp_nip_current_mss(sk); // TCP_BASE_MSS in __tcp_nip_ack_snd_check() 695 if (((tp->rcv_nxt - tp->rcv_wup) > get_ack_num() * inet_csk(sk)->icsk_ack.rcv_mss && in __tcp_nip_ack_snd_check() 741 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_rearm_rto() 751 u32 rto = inet_csk(sk)->icsk_rto; in tcp_nip_rearm_rto() 793 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_nip_clean_rtx_queue() 1000 struct inet_connection_sock *newicsk = inet_csk(newsk); in tcp_nip_create_openreq_child() 1289 struct inet_connection_sock *icsk = inet_csk(s in tcp_nip_ack_probe() [all...] |
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_recovery.c | 18 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) in tcp_rack_reo_wnd() 114 timeout, inet_csk(sk)->icsk_rto); in tcp_rack_mark_lost() 162 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { in tcp_rack_reo_timeout() 164 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rack_reo_timeout() 169 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) in tcp_rack_reo_timeout() 222 const u8 state = inet_csk(sk)->icsk_ca_state; in tcp_newreno_mark_lost()
|
H A D | tcp_timer.c | 28 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() 45 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout() 215 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out() 233 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() 292 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() 358 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() 409 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() 443 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rtx_probe0_timed_out() 489 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() 637 struct inet_connection_sock *icsk = inet_csk(s in tcp_write_timer_handler() [all...] |
H A D | tcp_dctcp.h | 29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
H A D | inet_connection_sock.c | 405 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port() 407 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port() 422 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() 470 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() 563 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() 574 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() 586 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers_sync() 736 reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); in inet_csk_reqsk_queue_drop() 755 struct inet_connection_sock *icsk = inet_csk(sk_listener); in reqsk_timer_handler() 836 struct inet_connection_sock *icsk = inet_csk(news in inet_clone_ulp() [all...] |
H A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp()
|
H A D | tcp_output.c | 69 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 155 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart() 166 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() 1113 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb() 1241 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() 1697 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() 1739 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu() 1762 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init() 1784 inet_csk(sk)->icsk_pmtu_cookie is last pmtu, seen by this function. 1794 NOTE2. inet_csk(s [all...] |
H A D | inet_hashtables.c | 109 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash() 124 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port() 126 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port() 151 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port() 216 hlist_add_tail_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, in inet_hash2() 219 hlist_add_head_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, in inet_hash2() 230 WARN_ON_ONCE(hlist_unhashed(&inet_csk(sk)->icsk_listen_portaddr_node))) in inet_unhash2() 236 hlist_del_init_rcu(&inet_csk(sk)->icsk_listen_portaddr_node); in inet_unhash2() 617 struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash; in inet_reuseport_add_sock() 627 inet_csk(sk in inet_reuseport_add_sock() [all...] |
H A D | tcp_fastopen.c | 49 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher() 94 q = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_reset_cipher() 260 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child() 264 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child() 327 fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check() 598 u32 timeouts = inet_csk(sk)->icsk_retransmits; in tcp_fastopen_active_detect_blackhole()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_recovery.c | 13 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) in tcp_rack_reo_wnd() 109 timeout, inet_csk(sk)->icsk_rto); in tcp_rack_mark_lost() 158 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { in tcp_rack_reo_timeout() 160 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rack_reo_timeout() 165 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) in tcp_rack_reo_timeout() 219 const u8 state = inet_csk(sk)->icsk_ca_state; in tcp_newreno_mark_lost()
|
H A D | tcp_timer.c | 28 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() 46 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout() 217 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out() 235 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() 300 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() 370 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() 424 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() 464 rcv_delta = inet_csk(sk)->icsk_timeout - tp->rcv_tstamp; in tcp_rtx_probe0_timed_out() 489 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() 644 struct inet_connection_sock *icsk = inet_csk(s in tcp_write_timer_handler() [all...] |
H A D | inet_connection_sock.c | 584 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port() 586 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port() 587 WARN_ON(inet_csk(sk)->icsk_bind2_hash != tb2); in inet_csk_get_port() 611 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() 659 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() 756 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() 767 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() 976 reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); in inet_csk_reqsk_queue_drop() 1018 reqsk_queue_migrated(&inet_csk(nsk)->icsk_accept_queue, req); in reqsk_timer_handler() 1024 icsk = inet_csk(sk_listene in reqsk_timer_handler() [all...] |
H A D | tcp_dctcp.h | 29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
H A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp()
|
H A D | tcp_output.c | 67 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 153 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart() 164 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() 267 if (unlikely(inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOMEM)) in tcp_select_window() 1123 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb() 1246 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() 1698 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() 1740 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu() 1763 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init() 1785 inet_csk(s [all...] |
H A D | inet_hashtables.c | 173 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash() 175 inet_csk(sk)->icsk_bind2_hash = tb2; in inet_bind_hash() 194 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port() 196 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port() 201 if (inet_csk(sk)->icsk_bind2_hash) { in __inet_put_port() 202 struct inet_bind2_bucket *tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_put_port() 205 inet_csk(sk)->icsk_bind2_hash = NULL; in __inet_put_port() 239 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port() 240 tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_inherit_port() 713 struct inet_bind_bucket *tb = inet_csk(s in inet_reuseport_add_sock() [all...] |
H A D | tcp_cong.c | 39 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() 222 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() 241 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() 256 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control() 275 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_congestion_control() 417 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_congestion_control()
|
H A D | tcp_fastopen.c | 43 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher() 82 q = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_reset_cipher() 243 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child() 247 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child() 311 fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check() 582 u32 timeouts = inet_csk(sk)->icsk_retransmits; in tcp_fastopen_active_detect_blackhole()
|
/kernel/linux/linux-5.10/net/dccp/ |
H A D | output.c | 46 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_transmit_skb() 161 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_sync_mss() 268 inet_csk(sk)->icsk_rto, in dccp_xmit_packet() 381 if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0) in dccp_retransmit_skb() 385 inet_csk(sk)->icsk_retransmits++; in dccp_retransmit_skb() 512 int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk); in dccp_send_reset() 537 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_connect() 581 inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; in dccp_send_ack() 601 struct inet_connection_sock *icsk = inet_csk(sk);
|
/kernel/linux/linux-6.6/net/dccp/ |
H A D | output.c | 46 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_transmit_skb() 163 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_sync_mss() 272 inet_csk(sk)->icsk_rto, in dccp_xmit_packet() 390 if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0) in dccp_retransmit_skb() 394 inet_csk(sk)->icsk_retransmits++; in dccp_retransmit_skb() 521 int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk); in dccp_send_reset() 546 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_connect() 590 inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; in dccp_send_ack() 610 struct inet_connection_sock *icsk = inet_csk(sk);
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/progs/ |
H A D | bpf_dctcp.c | 126 new_state != BPF_CORE_READ_BITFIELD(inet_csk(sk), icsk_ca_state)) in BPF_PROG() 159 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 163 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|