Lines Matching defs:tcp_sk
64 * accepted a non-mp-capable flow: sk is a tcp_sk,
93 msk->scaling_ratio = tcp_sk(ssock->sk)->scaling_ratio;
557 const struct tcp_sock *tp = tcp_sk(ssk);
659 tp = tcp_sk(ssk);
1147 if (unlikely(tcp_sk(ssk)->snd_wnd < mptcp_snd_wnd)) {
1148 tcp_sk(ssk)->snd_wnd = min_t(u64, U32_MAX, mptcp_snd_wnd);
1280 tcp_mark_push(tcp_sk(ssk), skb);
1287 tcp_mark_push(tcp_sk(ssk), skb);
1336 WRITE_ONCE(tcp_sk(ssk)->write_seq, tcp_sk(ssk)->write_seq + copy);
1399 u32 rcv_tstamp = READ_ONCE(tcp_sk(mptcp_subflow_tcp_sock(subflow))->rcv_tstamp);
1493 tcp_push(ssk, 0, info->mss_now, tcp_sk(ssk)->nonagle, info->size_goal);
1696 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle,
1989 tp = tcp_sk(mptcp_subflow_tcp_sock(subflow));
2040 tcp_sk(ssk)->window_clamp = window_clamp;
2544 struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow);
2547 slow = lock_sock_fast(tcp_sk);
2548 if (tcp_sk->sk_state != TCP_CLOSE) {
2549 tcp_send_active_reset(tcp_sk, GFP_ATOMIC);
2550 tcp_set_state(tcp_sk, TCP_CLOSE);
2552 unlock_sock_fast(tcp_sk, slow);
2642 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle,
2945 struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow);
2947 mptcp_subflow_shutdown(sk, tcp_sk, SEND_SHUTDOWN);
3272 msk->wnd_end = msk->snd_nxt + tcp_sk(ssk)->snd_wnd;
3316 const struct tcp_sock *tp = tcp_sk(ssk);
3353 tcp_sk(newsk)->is_mptcp = 0;
3495 struct tcp_sock *tp = tcp_sk(ssk);
3674 struct tcp_sock *tp = tcp_sk(msk->first);
3746 if (rcu_access_pointer(tcp_sk(ssk)->md5sig_info))