• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/net/ipv4/

Lines Matching refs:icsk

125 	struct inet_connection_sock *icsk = inet_csk(sk);
126 const unsigned int lss = icsk->icsk_ack.last_seg_size;
129 icsk->icsk_ack.last_seg_size = 0;
135 if (len >= icsk->icsk_ack.rcv_mss) {
136 icsk->icsk_ack.rcv_mss = len;
157 icsk->icsk_ack.last_seg_size = len;
159 icsk->icsk_ack.rcv_mss = len;
163 if (icsk->icsk_ack.pending & ICSK_ACK_PUSHED)
164 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED2;
165 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED;
171 struct inet_connection_sock *icsk = inet_csk(sk);
172 unsigned quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss);
176 if (quickacks > icsk->icsk_ack.quick)
177 icsk->icsk_ack.quick = min(quickacks, TCP_MAX_QUICKACKS);
182 struct inet_connection_sock *icsk = inet_csk(sk);
184 icsk->icsk_ack.pingpong = 0;
185 icsk->icsk_ack.ato = TCP_ATO_MIN;
194 const struct inet_connection_sock *icsk = inet_csk(sk);
195 return icsk->icsk_ack.quick && !icsk->icsk_ack.pingpong;
338 struct inet_connection_sock *icsk = inet_csk(sk);
340 icsk->icsk_ack.quick = 0;
512 struct inet_connection_sock *icsk = inet_csk(sk);
523 if (!icsk->icsk_ack.ato) {
528 icsk->icsk_ack.ato = TCP_ATO_MIN;
530 int m = now - icsk->icsk_ack.lrcvtime;
534 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + TCP_ATO_MIN / 2;
535 } else if (m < icsk->icsk_ack.ato) {
536 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + m;
537 if (icsk->icsk_ack.ato > icsk->icsk_rto)
538 icsk->icsk_ack.ato = icsk->icsk_rto;
539 } else if (m > icsk->icsk_rto) {
547 icsk->icsk_ack.lrcvtime = now;
676 const struct inet_connection_sock *icsk = inet_csk(sk);
679 if (icsk->icsk_backoff || !tp->srtt) {
728 icsk->icsk_ca_state == TCP_CA_Open) {
773 const struct inet_connection_sock *icsk = inet_csk(sk);
777 if (icsk->icsk_ca_state < TCP_CA_CWR) {
780 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
945 const struct inet_connection_sock *icsk = inet_csk(sk);
1222 if (lost_retrans && icsk->icsk_ca_state == TCP_CA_Recovery) {
1256 if ((reord < tp->fackets_out) && icsk->icsk_ca_state != TCP_CA_Loss &&
1315 const struct inet_connection_sock *icsk = inet_csk(sk);
1319 if ((!tp->frto_counter && icsk->icsk_ca_state <= TCP_CA_Disorder) ||
1321 ((icsk->icsk_ca_state == TCP_CA_Loss || tp->frto_counter) &&
1322 !icsk->icsk_retransmits)) {
1337 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
1340 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
1366 ((1 << icsk->icsk_ca_state) & (TCPF_CA_Recovery|TCPF_CA_Loss))) &&
1458 const struct inet_connection_sock *icsk = inet_csk(sk);
1464 if (icsk->icsk_ca_state <= TCP_CA_Disorder || tp->snd_una == tp->high_seq ||
1465 (icsk->icsk_ca_state == TCP_CA_Loss && !icsk->icsk_retransmits)) {
1467 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
1523 struct inet_connection_sock *icsk = inet_csk(sk);
1527 icsk->icsk_retransmits++;
1530 icsk->icsk_rto, TCP_RTO_MAX);
1901 const struct inet_connection_sock *icsk = inet_csk(sk);
1903 if (icsk->icsk_ca_ops->undo_cwnd)
1904 tp->snd_cwnd = icsk->icsk_ca_ops->undo_cwnd(sk);
2067 struct inet_connection_sock *icsk = inet_csk(sk);
2069 icsk->icsk_mtup.search_high = icsk->icsk_mtup.probe_size - 1;
2070 icsk->icsk_mtup.probe_size = 0;
2076 struct inet_connection_sock *icsk = inet_csk(sk);
2081 icsk->icsk_mtup.probe_size;
2086 icsk->icsk_mtup.search_low = icsk->icsk_mtup.probe_size;
2087 icsk->icsk_mtup.probe_size = 0;
2088 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
2107 struct inet_connection_sock *icsk = inet_csk(sk);
2131 icsk->icsk_ca_state != TCP_CA_Open &&
2142 if (icsk->icsk_ca_state == TCP_CA_Open) {
2146 switch (icsk->icsk_ca_state) {
2148 icsk->icsk_retransmits = 0;
2184 switch (icsk->icsk_ca_state) {
2198 icsk->icsk_retransmits = 0;
2204 if (icsk->icsk_ca_state != TCP_CA_Open)
2215 if (icsk->icsk_ca_state == TCP_CA_Disorder)
2224 if (icsk->icsk_ca_state < TCP_CA_CWR &&
2225 icsk->icsk_mtup.probe_size &&
2246 if (icsk->icsk_ca_state < TCP_CA_CWR) {
2249 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
2326 const struct inet_connection_sock *icsk = inet_csk(sk);
2327 icsk->icsk_ca_ops->cong_avoid(sk, ack, rtt, in_flight, good);
2406 const struct inet_connection_sock *icsk = inet_csk(sk);
2446 if (icsk->icsk_mtup.probe_size) {
2503 const struct inet_connection_sock *icsk = inet_csk(sk);
2506 tp->lost_out, icsk->icsk_ca_state);
2511 tp->sacked_out, icsk->icsk_ca_state);
2516 tp->retrans_out, icsk->icsk_ca_state);
2528 struct inet_connection_sock *icsk = inet_csk(sk);
2534 icsk->icsk_backoff = 0;
2541 min(icsk->icsk_rto << icsk->icsk_backoff, TCP_RTO_MAX),
2749 struct inet_connection_sock *icsk = inet_csk(sk);
2769 if (icsk->icsk_ca_state < TCP_CA_CWR)
2771 else if (icsk->icsk_ca_state == TCP_CA_Loss)
2839 icsk->icsk_probes_out = 0;
4400 struct inet_connection_sock *icsk = inet_csk(sk);
4493 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
4506 icsk->icsk_af_ops->rebuild_header(sk);
4533 icsk->icsk_accept_queue.rskq_defer_accept ||
4534 icsk->icsk_ack.pingpong) {
4543 icsk->icsk_ack.lrcvtime = tcp_time_stamp;
4544 icsk->icsk_ack.ato = TCP_ATO_MIN;
4604 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
4638 struct inet_connection_sock *icsk = inet_csk(sk);
4655 if (icsk->icsk_af_ops->conn_request(sk, skb) < 0)
4770 icsk->icsk_af_ops->rebuild_header(sk);