Lines Matching refs:sk
26 static u32 tcp_clamp_rto_to_user_timeout(const struct sock *sk)
28 struct inet_connection_sock *icsk = inet_csk(sk);
32 start_ts = tcp_sk(sk)->retrans_stamp;
35 elapsed = tcp_time_stamp(tcp_sk(sk)) - start_ts;
43 u32 tcp_clamp_probe0_to_user_timeout(const struct sock *sk, u32 when)
45 struct inet_connection_sock *icsk = inet_csk(sk);
63 * @sk: The socket the error has appeared on.
68 static void tcp_write_err(struct sock *sk)
70 sk->sk_err = sk->sk_err_soft ? : ETIMEDOUT;
71 sk->sk_error_report(sk);
73 tcp_write_queue_purge(sk);
74 tcp_done(sk);
75 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT);
80 * @sk: pointer to current socket
102 static int tcp_out_of_resources(struct sock *sk, bool do_reset)
104 struct tcp_sock *tp = tcp_sk(sk);
113 if (sk->sk_err_soft)
116 if (tcp_check_oom(sk, shift)) {
124 tcp_send_active_reset(sk, GFP_ATOMIC);
125 tcp_done(sk);
126 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY);
130 if (!check_net(sock_net(sk))) {
132 tcp_done(sk);
141 * @sk: Pointer to the current socket.
144 static int tcp_orphan_retries(struct sock *sk, bool alive)
146 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */
149 if (sk->sk_err_soft && !alive)
160 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk)
162 const struct net *net = sock_net(sk);
173 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1;
177 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss);
179 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
182 static unsigned int tcp_model_timeout(struct sock *sk,
198 * @sk: The current socket
209 static bool retransmits_timed_out(struct sock *sk,
215 if (!inet_csk(sk)->icsk_retransmits)
218 start_ts = tcp_sk(sk)->retrans_stamp;
222 if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))
223 rto_base = tcp_timeout_init(sk);
224 timeout = tcp_model_timeout(sk, boundary, rto_base);
227 return (s32)(tcp_time_stamp(tcp_sk(sk)) - start_ts - timeout) >= 0;
231 static int tcp_write_timeout(struct sock *sk)
233 struct inet_connection_sock *icsk = inet_csk(sk);
234 struct tcp_sock *tp = tcp_sk(sk);
235 struct net *net = sock_net(sk);
239 if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) {
241 __dst_negative_advice(sk);
246 if (retransmits_timed_out(sk, READ_ONCE(net->ipv4.sysctl_tcp_retries1), 0)) {
248 tcp_mtu_probing(icsk, sk);
250 __dst_negative_advice(sk);
254 if (sock_flag(sk, SOCK_DEAD)) {
257 retry_until = tcp_orphan_retries(sk, alive);
259 !retransmits_timed_out(sk, retry_until, 0);
261 if (tcp_out_of_resources(sk, do_reset))
266 expired = retransmits_timed_out(sk, retry_until,
268 tcp_fastopen_active_detect_blackhole(sk, expired);
271 tcp_call_bpf_3arg(sk, BPF_SOCK_OPS_RTO_CB,
277 tcp_write_err(sk);
281 if (sk_rethink_txhash(sk)) {
283 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH);
290 void tcp_delack_timer_handler(struct sock *sk)
292 struct inet_connection_sock *icsk = inet_csk(sk);
294 sk_mem_reclaim_partial(sk);
296 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) ||
301 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout);
306 if (inet_csk_ack_scheduled(sk)) {
307 if (!inet_csk_in_pingpong_mode(sk)) {
314 inet_csk_exit_pingpong_mode(sk);
317 tcp_mstamp_refresh(tcp_sk(sk));
318 tcp_send_ack(sk);
319 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS);
323 if (tcp_under_memory_pressure(sk))
324 sk_mem_reclaim(sk);
341 struct sock *sk = &icsk->icsk_inet.sk;
343 bh_lock_sock(sk);
344 if (!sock_owned_by_user(sk)) {
345 tcp_delack_timer_handler(sk);
347 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED);
349 if (!test_and_set_bit(TCP_DELACK_TIMER_DEFERRED, &sk->sk_tsq_flags))
350 sock_hold(sk);
352 bh_unlock_sock(sk);
353 sock_put(sk);
356 static void tcp_probe_timer(struct sock *sk)
358 struct inet_connection_sock *icsk = inet_csk(sk);
359 struct sk_buff *skb = tcp_send_head(sk);
360 struct tcp_sock *tp = tcp_sk(sk);
384 max_probes = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retries2);
385 if (sock_flag(sk, SOCK_DEAD)) {
388 max_probes = tcp_orphan_retries(sk, alive);
391 if (tcp_out_of_resources(sk, true))
396 abort: tcp_write_err(sk);
399 tcp_send_probe0(sk);
405 * sk here is the child socket, not the parent (listener) socket.
407 static void tcp_fastopen_synack_timer(struct sock *sk, struct request_sock *req)
409 struct inet_connection_sock *icsk = inet_csk(sk);
410 struct tcp_sock *tp = tcp_sk(sk);
417 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_synack_retries) + 1;
420 tcp_write_err(sk);
425 tcp_enter_loss(sk);
431 inet_rtx_syn_ack(sk, req);
436 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
440 static bool tcp_rtx_probe0_timed_out(const struct sock *sk,
443 const struct inet_connection_sock *icsk = inet_csk(sk);
445 const struct tcp_sock *tp = tcp_sk(sk);
476 * @sk: Pointer to the current socket.
485 void tcp_retransmit_timer(struct sock *sk)
487 struct tcp_sock *tp = tcp_sk(sk);
488 struct net *net = sock_net(sk);
489 struct inet_connection_sock *icsk = inet_csk(sk);
494 lockdep_sock_is_held(sk));
496 WARN_ON_ONCE(sk->sk_state != TCP_SYN_RECV &&
497 sk->sk_state != TCP_FIN_WAIT1);
498 tcp_fastopen_synack_timer(sk, req);
508 skb = tcp_rtx_queue_head(sk);
514 if (!tp->snd_wnd && !sock_flag(sk, SOCK_DEAD) &&
515 !((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))) {
521 struct inet_sock *inet = inet_sk(sk);
522 if (sk->sk_family == AF_INET) {
530 else if (sk->sk_family == AF_INET6) {
532 &sk->sk_v6_daddr,
538 if (tcp_rtx_probe0_timed_out(sk, skb)) {
539 tcp_write_err(sk);
542 tcp_enter_loss(sk);
543 tcp_retransmit_skb(sk, skb, 1);
544 __sk_dst_reset(sk);
548 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTS);
549 if (tcp_write_timeout(sk))
570 __NET_INC_STATS(sock_net(sk), mib_idx);
573 tcp_enter_loss(sk);
576 if (tcp_retransmit_skb(sk, tcp_rtx_queue_head(sk), 1) > 0) {
580 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
613 if (sk->sk_state == TCP_ESTABLISHED &&
619 tcp_rto_min(sk),
625 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
626 tcp_clamp_rto_to_user_timeout(sk), TCP_RTO_MAX);
627 if (retransmits_timed_out(sk, READ_ONCE(net->ipv4.sysctl_tcp_retries1) + 1, 0))
628 __sk_dst_reset(sk);
635 void tcp_write_timer_handler(struct sock *sk)
637 struct inet_connection_sock *icsk = inet_csk(sk);
640 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) ||
645 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout);
649 tcp_mstamp_refresh(tcp_sk(sk));
654 tcp_rack_reo_timeout(sk);
657 tcp_send_loss_probe(sk);
661 tcp_retransmit_timer(sk);
665 tcp_probe_timer(sk);
670 sk_mem_reclaim(sk);
677 struct sock *sk = &icsk->icsk_inet.sk;
679 bh_lock_sock(sk);
680 if (!sock_owned_by_user(sk)) {
681 tcp_write_timer_handler(sk);
684 if (!test_and_set_bit(TCP_WRITE_TIMER_DEFERRED, &sk->sk_tsq_flags))
685 sock_hold(sk);
687 bh_unlock_sock(sk);
688 sock_put(sk);
699 void tcp_set_keepalive(struct sock *sk, int val)
701 if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN))
704 if (val && !sock_flag(sk, SOCK_KEEPOPEN))
705 inet_csk_reset_keepalive_timer(sk, keepalive_time_when(tcp_sk(sk)));
707 inet_csk_delete_keepalive_timer(sk);
714 struct sock *sk = from_timer(sk, t, sk_timer);
715 struct inet_connection_sock *icsk = inet_csk(sk);
716 struct tcp_sock *tp = tcp_sk(sk);
720 bh_lock_sock(sk);
721 if (sock_owned_by_user(sk)) {
723 inet_csk_reset_keepalive_timer (sk, HZ/20);
727 if (sk->sk_state == TCP_LISTEN) {
733 if (sk->sk_state == TCP_FIN_WAIT2 && sock_flag(sk, SOCK_DEAD)) {
735 const int tmo = tcp_fin_time(sk) - TCP_TIMEWAIT_LEN;
738 tcp_time_wait(sk, TCP_FIN_WAIT2, tmo);
742 tcp_send_active_reset(sk, GFP_ATOMIC);
746 if (!sock_flag(sk, SOCK_KEEPOPEN) ||
747 ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_SYN_SENT)))
753 if (tp->packets_out || !tcp_write_queue_empty(sk))
767 tcp_send_active_reset(sk, GFP_ATOMIC);
768 tcp_write_err(sk);
771 if (tcp_write_wakeup(sk, LINUX_MIB_TCPKEEPALIVE) <= 0) {
785 sk_mem_reclaim(sk);
788 inet_csk_reset_keepalive_timer (sk, elapsed);
792 tcp_done(sk);
795 bh_unlock_sock(sk);
796 sock_put(sk);
802 struct sock *sk = (struct sock *)tp;
804 bh_lock_sock(sk);
805 if (!sock_owned_by_user(sk)) {
812 tcp_send_ack(sk);
816 &sk->sk_tsq_flags))
817 sock_hold(sk);
819 bh_unlock_sock(sk);
821 sock_put(sk);
826 void tcp_init_xmit_timers(struct sock *sk)
828 inet_csk_init_xmit_timers(sk, &tcp_write_timer, &tcp_delack_timer,
830 hrtimer_init(&tcp_sk(sk)->pacing_timer, CLOCK_MONOTONIC,
832 tcp_sk(sk)->pacing_timer.function = tcp_pace_kick;
834 hrtimer_init(&tcp_sk(sk)->compressed_ack_timer, CLOCK_MONOTONIC,
836 tcp_sk(sk)->compressed_ack_timer.function = tcp_compressed_ack_kick;