Lines Matching defs:sk
5 static u32 tcp_rack_reo_wnd(const struct sock *sk)
7 const struct tcp_sock *tp = tcp_sk(sk);
13 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery)
17 !(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_recovery) &
58 static void tcp_rack_detect_loss(struct sock *sk, u32 *reo_timeout)
60 struct tcp_sock *tp = tcp_sk(sk);
65 reo_wnd = tcp_rack_reo_wnd(sk);
86 tcp_mark_skb_lost(sk, skb);
95 bool tcp_rack_mark_lost(struct sock *sk)
97 struct tcp_sock *tp = tcp_sk(sk);
105 tcp_rack_detect_loss(sk, &timeout);
108 inet_csk_reset_xmit_timer(sk, ICSK_TIME_REO_TIMEOUT,
109 timeout, inet_csk(sk)->icsk_rto);
149 void tcp_rack_reo_timeout(struct sock *sk)
151 struct tcp_sock *tp = tcp_sk(sk);
156 tcp_rack_detect_loss(sk, &timeout);
158 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) {
159 tcp_enter_recovery(sk, false);
160 if (!inet_csk(sk)->icsk_ca_ops->cong_control)
161 tcp_cwnd_reduction(sk, 1, tp->lost - lost, 0);
163 tcp_xmit_retransmit_queue(sk);
165 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS)
166 tcp_rearm_rto(sk);
187 void tcp_rack_update_reo_wnd(struct sock *sk, struct rate_sample *rs)
189 struct tcp_sock *tp = tcp_sk(sk);
191 if ((READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_recovery) &
217 void tcp_newreno_mark_lost(struct sock *sk, bool snd_una_advanced)
219 const u8 state = inet_csk(sk)->icsk_ca_state;
220 struct tcp_sock *tp = tcp_sk(sk);
224 struct sk_buff *skb = tcp_rtx_queue_head(sk);
232 tcp_fragment(sk, TCP_FRAG_IN_RTX_QUEUE, skb,
235 tcp_mark_skb_lost(sk, skb);