Searched refs:mss_cache (Results 1 - 19 of 19) sorted by relevance

/linux-master/net/ipv4/
H A Dtcp_westwood.c191 w->accounted += tp->mss_cache;
192 w->cumul_ack = tp->mss_cache;
195 if (w->cumul_ack > tp->mss_cache) {
199 w->cumul_ack = tp->mss_cache;
214 * in packets we use mss_cache). Rttmin is guaranteed to be >= 2
222 return max_t(u32, (w->bw_est * w->rtt_min) / tp->mss_cache, 2);
H A Dtcp_nv.c349 bytes_acked >= (ca->nv_min_cwnd - 1) * tp->mss_cache &&
355 ca->nv_min_cwnd * tp->mss_cache;
367 80000ULL * tp->mss_cache);
H A Dtcp_dctcp.c220 info->dctcp.dctcp_ab_ecn = tp->mss_cache *
222 info->dctcp.dctcp_ab_tot = tp->mss_cache *
H A Dtcp_rate.c199 tp->write_seq - tp->snd_nxt < tp->mss_cache &&
H A Dtcp_bbr.c246 unsigned int mss = tcp_sk(sk)->mss_cache;
315 segs = max_t(u32, bytes / tp->mss_cache, bbr_min_tso_segs(sk));
1116 bw = bw * tp->mss_cache * USEC_PER_SEC >> BW_SCALE;
H A Dtcp_output.c834 * advertised. But we subtract them from tp->mss_cache so that
1816 tp->mss_cache is current effective sending mss, including
1824 NOTE2. inet_csk(sk)->icsk_pmtu_cookie and tp->mss_cache
1843 tp->mss_cache = mss_now;
1861 mss_now = tp->mss_cache;
1871 /* The mss_cache is sized based on tp->tcp_header_len, which assumes
2240 cong_win = (tcp_snd_cwnd(tp) - in_flight) * tp->mss_cache;
2245 if (limit >= max_segs * tp->mss_cache)
2254 u32 chunk = min(tp->snd_wnd, tcp_snd_cwnd(tp) * tp->mss_cache);
2268 if (limit > tcp_max_tso_deferred_mss(tp) * tp->mss_cache)
[all...]
H A Dtcp_input.c425 per_mss = max_t(u32, tp->rx_opt.mss_clamp, tp->mss_cache) +
618 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache);
935 rate = (u64)tp->mss_cache * ((USEC_PER_SEC / 100) << 3);
1034 if (seq_len > tp->mss_cache)
1035 dup_segs = DIV_ROUND_UP(seq_len, tp->mss_cache);
1069 const u32 mss = tp->mss_cache;
2767 val = (u64)tcp_snd_cwnd(tp) * tcp_mss_to_mtu(sk, tp->mss_cache);
2931 tp->snd_una + tp->reordering * tp->mss_cache);
3395 (tp->snd_una - prior_snd_una) < tp->mss_cache &&
3454 sample.in_flight = tp->mss_cache *
[all...]
H A Dtcp.c402 rate64 = (u64)rate * tp->mss_cache * USEC_PER_SEC;
446 tp->mss_cache = TCP_MSS_DEFAULT;
3829 info->tcpi_snd_mss = tp->mss_cache;
4023 val = tp->mss_cache;
4616 CACHELINE_ASSERT_GROUP_MEMBER(struct tcp_sock, tcp_sock_read_txrx, mss_cache);
/linux-master/tools/testing/selftests/bpf/progs/
H A Dtcp_ca_write_sk_pacing.c46 ((tp->snd_cwnd * tp->mss_cache * USEC_PER_SEC) << 3) /
H A Dtest_sock_fields.c95 dst->mss_cache = src->mss_cache;
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_qp_tx.c674 c_tx->tcp_seglen = tp->mss_cache * tp->gso_segs;
677 tp->mss_cache *
680 c_tx->tcp_seglen = tp->mss_cache;
/linux-master/tools/testing/selftests/bpf/prog_tests/
H A Dsock_fields.c86 "snd_nxt:%u snd:una:%u mss_cache:%u ecn_flags:%u "
93 tp->rcv_nxt, tp->snd_nxt, tp->snd_una, tp->mss_cache,
/linux-master/include/linux/
H A Dtcp.h218 u32 mss_cache; /* Cached effective mss, not including SACKS */ member in struct:tcp_sock
/linux-master/include/uapi/linux/
H A Dbpf.h6345 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member in struct:bpf_tcp_sock
6784 __u32 mss_cache; member in struct:bpf_sock_ops
/linux-master/tools/include/uapi/linux/
H A Dbpf.h6345 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member in struct:bpf_tcp_sock
6784 __u32 mss_cache; member in struct:bpf_sock_ops
/linux-master/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_io.c192 FLOWC_PARAM(MSS, tp->mss_cache);
H A Dchtls_cm.c153 tp->mss_cache = tp->rx_opt.mss_clamp;
/linux-master/net/ipv6/
H A Dtcp_ipv6.c361 if (tcp_mtu_to_mss(sk, mtu) >= tcp_sk(sk)->mss_cache)
/linux-master/net/core/
H A Dfilter.c7125 case offsetof(struct bpf_tcp_sock, mss_cache):
7126 BPF_TCP_SOCK_GET_COMMON(mss_cache);
10580 case offsetof(struct bpf_sock_ops, mss_cache):
10581 SOCK_OPS_GET_TCP_SOCK_FIELD(mss_cache);

Completed in 598 milliseconds