Lines Matching refs:ssk

595 void mptcp_subflow_process_delegated(struct sock *ssk, long actions);
653 void mptcp_subflow_shutdown(struct sock *sk, struct sock *ssk, int how);
654 void mptcp_close_ssk(struct sock *sk, struct sock *ssk,
656 void __mptcp_subflow_send_ack(struct sock *ssk);
657 void mptcp_subflow_reset(struct sock *ssk);
658 void mptcp_subflow_queue_clean(struct sock *sk, struct sock *ssk);
711 static inline bool __tcp_can_send(const struct sock *ssk)
714 return ((1 << inet_sk_state_load(ssk)) & (TCPF_ESTABLISHED | TCPF_CLOSE_WAIT));
730 void mptcp_subflow_drop_ctx(struct sock *ssk);
750 struct sock *ssk,
786 void mptcp_rcv_space_init(struct mptcp_sock *msk, const struct sock *ssk);
787 void mptcp_data_ready(struct sock *sk, struct sock *ssk);
803 void __mptcp_check_push(struct sock *sk, struct sock *ssk);
869 static inline void __mptcp_propagate_sndbuf(struct sock *sk, struct sock *ssk)
871 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
873 if (READ_ONCE(ssk->sk_sndbuf) != subflow->cached_sndbuf)
882 static inline void mptcp_propagate_sndbuf(struct sock *sk, struct sock *ssk)
884 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
886 if (likely(READ_ONCE(ssk->sk_sndbuf) == subflow->cached_sndbuf))
906 int mptcp_token_new_connect(struct sock *ssk);
931 void mptcp_pm_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk);
932 void mptcp_pm_nl_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk);
933 void mptcp_pm_new_connection(struct mptcp_sock *msk, const struct sock *ssk, int server_side);
934 void mptcp_pm_fully_established(struct mptcp_sock *msk, const struct sock *ssk);
941 void mptcp_pm_add_addr_received(const struct sock *ssk,
986 const struct sock *ssk, gfp_t gfp);
987 void mptcp_event_addr_announced(const struct sock *ssk, const struct mptcp_addr_info *info);
989 void mptcp_event_pm_listener(const struct sock *ssk,
1105 void mptcp_sockopt_sync(struct mptcp_sock *msk, struct sock *ssk);
1106 void mptcp_sockopt_sync_locked(struct mptcp_sock *msk, struct sock *ssk);
1139 struct sock *ssk = READ_ONCE(msk->first);
1141 return ssk && ((1 << inet_sk_state_load(ssk)) &
1146 static inline void mptcp_do_fallback(struct sock *ssk)
1148 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
1154 if (READ_ONCE(msk->snd_data_fin_enable) && !(ssk->sk_shutdown & SEND_SHUTDOWN)) {
1155 gfp_t saved_allocation = ssk->sk_allocation;
1157 /* we are in a atomic (BH) scope, override ssk default for data
1160 ssk->sk_allocation = GFP_ATOMIC;
1161 ssk->sk_shutdown |= SEND_SHUTDOWN;
1162 tcp_shutdown(ssk, SEND_SHUTDOWN);
1163 ssk->sk_allocation = saved_allocation;