Lines Matching defs:tcp_sk
78 msk->scaling_ratio = tcp_sk(ssock->sk)->scaling_ratio;
542 const struct tcp_sock *tp = tcp_sk(ssk);
644 tp = tcp_sk(ssk);
1134 if (unlikely(tcp_sk(ssk)->snd_wnd < mptcp_snd_wnd)) {
1135 tcp_sk(ssk)->snd_wnd = min_t(u64, U32_MAX, mptcp_snd_wnd);
1267 tcp_mark_push(tcp_sk(ssk), skb);
1274 tcp_mark_push(tcp_sk(ssk), skb);
1323 WRITE_ONCE(tcp_sk(ssk)->write_seq, tcp_sk(ssk)->write_seq + copy);
1386 u32 rcv_tstamp = READ_ONCE(tcp_sk(mptcp_subflow_tcp_sock(subflow))->rcv_tstamp);
1480 tcp_push(ssk, 0, info->mss_now, tcp_sk(ssk)->nonagle, info->size_goal);
1683 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle,
2008 tp = tcp_sk(mptcp_subflow_tcp_sock(subflow));
2059 tcp_sk(ssk)->window_clamp = window_clamp;
2563 struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow);
2566 slow = lock_sock_fast(tcp_sk);
2567 if (tcp_sk->sk_state != TCP_CLOSE) {
2568 tcp_send_active_reset(tcp_sk, GFP_ATOMIC);
2569 tcp_set_state(tcp_sk, TCP_CLOSE);
2571 unlock_sock_fast(tcp_sk, slow);
2661 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle,
2965 struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow);
2967 mptcp_subflow_shutdown(sk, tcp_sk, SEND_SHUTDOWN);
3286 WRITE_ONCE(msk->wnd_end, msk->snd_nxt + tcp_sk(ssk)->snd_wnd);
3330 const struct tcp_sock *tp = tcp_sk(ssk);
3468 struct tcp_sock *tp = tcp_sk(ssk);
3647 struct tcp_sock *tp = tcp_sk(msk->first);
3719 if (rcu_access_pointer(tcp_sk(ssk)->md5sig_info))
3908 tcp_sk(newsk)->is_mptcp = 0;
3948 * flow: sk is a tcp_sk, not an mptcp one.