Lines Matching refs:icsk

28 	struct inet_connection_sock *icsk = inet_csk(sk);
33 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
35 return icsk->icsk_rto;
45 return min_t(u32, icsk->icsk_rto, msecs_to_jiffies(remaining));
50 struct inet_connection_sock *icsk = inet_csk(sk);
54 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
55 if (!user_timeout || !icsk->icsk_probes_tstamp)
58 elapsed = tcp_jiffies32 - icsk->icsk_probes_tstamp;
166 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk)
175 if (!icsk->icsk_mtup.enabled) {
176 icsk->icsk_mtup.enabled = 1;
177 icsk->icsk_mtup.probe_timestamp = tcp_jiffies32;
179 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1;
183 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss);
185 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
245 struct inet_connection_sock *icsk = inet_csk(sk);
252 if (icsk->icsk_retransmits)
255 retry_until = READ_ONCE(icsk->icsk_syn_retries) ? :
262 expired = icsk->icsk_retransmits >= max_retransmits;
266 tcp_mtu_probing(icsk, sk);
273 const bool alive = icsk->icsk_rto < TCP_RTO_MAX;
285 READ_ONCE(icsk->icsk_user_timeout));
290 icsk->icsk_retransmits,
291 icsk->icsk_rto, (int)expired);
310 struct inet_connection_sock *icsk = inet_csk(sk);
323 if (!(icsk->icsk_ack.pending & ICSK_ACK_TIMER))
326 if (time_after(icsk->icsk_ack.timeout, jiffies)) {
327 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout);
330 icsk->icsk_ack.pending &= ~ICSK_ACK_TIMER;
335 icsk->icsk_ack.ato = min_t(u32, icsk->icsk_ack.ato << 1, icsk->icsk_rto);
341 icsk->icsk_ack.ato = TCP_ATO_MIN;
361 struct inet_connection_sock *icsk =
362 from_timer(icsk, t, icsk_delack_timer);
363 struct sock *sk = &icsk->icsk_inet.sk;
380 struct inet_connection_sock *icsk = inet_csk(sk);
386 icsk->icsk_probes_out = 0;
387 icsk->icsk_probes_tstamp = 0;
399 if (!icsk->icsk_probes_tstamp) {
400 icsk->icsk_probes_tstamp = tcp_jiffies32;
402 u32 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
405 (s32)(tcp_jiffies32 - icsk->icsk_probes_tstamp) >=
411 const bool alive = inet_csk_rto_backoff(icsk, TCP_RTO_MAX) < TCP_RTO_MAX;
414 if (!alive && icsk->icsk_backoff >= max_probes)
420 if (icsk->icsk_probes_out >= max_probes) {
430 struct inet_connection_sock *icsk = inet_csk(sk);
433 if (!icsk->icsk_retransmits) {
437 icsk->icsk_retransmits++;
447 struct inet_connection_sock *icsk = inet_csk(sk);
456 max_retries = READ_ONCE(icsk->icsk_syn_retries) ? :
464 if (icsk->icsk_retransmits == 1)
510 struct inet_connection_sock *icsk = inet_csk(sk);
580 if (icsk->icsk_retransmits == 0) {
583 if (icsk->icsk_ca_state == TCP_CA_Recovery) {
588 } else if (icsk->icsk_ca_state == TCP_CA_Loss) {
590 } else if ((icsk->icsk_ca_state == TCP_CA_Disorder) ||
643 icsk->icsk_retransmits <= TCP_THIN_LINEAR_RETRIES) {
644 icsk->icsk_backoff = 0;
645 icsk->icsk_rto = clamp(__tcp_set_rto(tp),
654 icsk->icsk_backoff++;
655 icsk->icsk_rto = min(icsk->icsk_rto << 1, TCP_RTO_MAX);
669 struct inet_connection_sock *icsk = inet_csk(sk);
673 !icsk->icsk_pending)
676 if (time_after(icsk->icsk_timeout, jiffies)) {
677 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout);
682 event = icsk->icsk_pending;
692 icsk->icsk_pending = 0;
696 icsk->icsk_pending = 0;
704 struct inet_connection_sock *icsk =
705 from_timer(icsk, t, icsk_retransmit_timer);
706 struct sock *sk = &icsk->icsk_inet.sk;
744 struct inet_connection_sock *icsk = inet_csk(sk);
788 u32 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
795 icsk->icsk_probes_out > 0) ||
797 icsk->icsk_probes_out >= keepalive_probes(tp))) {
803 icsk->icsk_probes_out++;