Lines Matching refs:ssk
577 void mptcp_subflow_process_delegated(struct sock *ssk, long actions);
634 void mptcp_subflow_shutdown(struct sock *sk, struct sock *ssk, int how);
635 void mptcp_close_ssk(struct sock *sk, struct sock *ssk,
637 void __mptcp_subflow_send_ack(struct sock *ssk);
638 void mptcp_subflow_reset(struct sock *ssk);
639 void mptcp_subflow_queue_clean(struct sock *sk, struct sock *ssk);
674 static inline bool __tcp_can_send(const struct sock *ssk)
677 return ((1 << inet_sk_state_load(ssk)) & (TCPF_ESTABLISHED | TCPF_CLOSE_WAIT));
693 void mptcp_subflow_drop_ctx(struct sock *ssk);
713 struct sock *ssk,
748 void mptcp_rcv_space_init(struct mptcp_sock *msk, const struct sock *ssk);
749 void mptcp_data_ready(struct sock *sk, struct sock *ssk);
765 void __mptcp_check_push(struct sock *sk, struct sock *ssk);
809 static inline void __mptcp_propagate_sndbuf(struct sock *sk, struct sock *ssk)
811 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
813 if (READ_ONCE(ssk->sk_sndbuf) != subflow->cached_sndbuf)
822 static inline void mptcp_propagate_sndbuf(struct sock *sk, struct sock *ssk)
824 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
826 if (likely(READ_ONCE(ssk->sk_sndbuf) == subflow->cached_sndbuf))
846 int mptcp_token_new_connect(struct sock *ssk);
871 void mptcp_pm_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk);
872 void mptcp_pm_nl_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk);
873 void mptcp_pm_new_connection(struct mptcp_sock *msk, const struct sock *ssk, int server_side);
874 void mptcp_pm_fully_established(struct mptcp_sock *msk, const struct sock *ssk);
879 void mptcp_pm_subflow_check_next(struct mptcp_sock *msk, const struct sock *ssk,
881 void mptcp_pm_add_addr_received(const struct sock *ssk,
936 const struct sock *ssk, gfp_t gfp);
937 void mptcp_event_addr_announced(const struct sock *ssk, const struct mptcp_addr_info *info);
939 void mptcp_event_pm_listener(const struct sock *ssk,
1044 void mptcp_sockopt_sync(struct mptcp_sock *msk, struct sock *ssk);
1045 void mptcp_sockopt_sync_locked(struct mptcp_sock *msk, struct sock *ssk);
1076 static inline void mptcp_do_fallback(struct sock *ssk)
1078 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
1084 if (READ_ONCE(msk->snd_data_fin_enable) && !(ssk->sk_shutdown & SEND_SHUTDOWN)) {
1085 gfp_t saved_allocation = ssk->sk_allocation;
1087 /* we are in a atomic (BH) scope, override ssk default for data
1090 ssk->sk_allocation = GFP_ATOMIC;
1091 ssk->sk_shutdown |= SEND_SHUTDOWN;
1092 tcp_shutdown(ssk, SEND_SHUTDOWN);
1093 ssk->sk_allocation = saved_allocation;