/linux-master/include/uapi/linux/ |
H A D | bpf.h | 6351 __u32 snd_nxt; /* Next sequence we send */ member in struct:bpf_tcp_sock 6794 __u32 snd_nxt; member in struct:bpf_sock_ops
|
/linux-master/net/core/ |
H A D | filter.c | 7151 case offsetof(struct bpf_tcp_sock, snd_nxt): 7152 BPF_TCP_SOCK_GET_COMMON(snd_nxt); 10604 case offsetof(struct bpf_sock_ops, snd_nxt): 10605 SOCK_OPS_GET_TCP_SOCK_FIELD(snd_nxt);
|
/linux-master/net/ipv4/ |
H A D | tcp.c | 647 READ_ONCE(tp->snd_nxt); 3019 (tp->snd_nxt != tp->write_seq && 3878 info->tcpi_notsent_bytes = max_t(int, 0, tp->write_seq - tp->snd_nxt); 4015 max_t(int, 0, tp->write_seq - tp->snd_nxt)); 4687 CACHELINE_ASSERT_GROUP_MEMBER(struct tcp_sock, tcp_sock_write_txrx, snd_nxt);
|
H A D | tcp_cdg.c | 277 ca->rtt_seq = tp->snd_nxt; 359 ca->rtt_seq = tp->snd_nxt; 364 ca->rtt_seq = tp->snd_nxt; 383 ca->rtt_seq = tp->snd_nxt;
|
H A D | tcp_cubic.c | 124 ca->end_seq = tp->snd_nxt;
|
H A D | tcp_dctcp.c | 83 ca->next_seq = tp->snd_nxt;
|
H A D | tcp_illinois.c | 62 ca->end_seq = tp->snd_nxt;
|
H A D | tcp_input.c | 914 tp->rtt_seq = tp->snd_nxt; 925 tp->rtt_seq = tp->snd_nxt; 1218 * there's no guarantee that it will be before snd_nxt (n). The problem 1219 * happens when start_seq resides between end_seq wrap (e_w) and snd_nxt 1231 * similar case (end_seq after snd_nxt wrap) as earlier reversed check in 1232 * snd_nxt wrap -> snd_una region will then become "well defined", i.e., 1252 if (after(end_seq, tp->snd_nxt) || !before(start_seq, end_seq)) 1256 if (!before(start_seq, tp->snd_nxt)) 1862 state->reord = tp->snd_nxt; 2211 tp->high_seq = tp->snd_nxt; [all...] |
H A D | tcp_ipv4.c | 550 !between(seq, snd_una, tp->snd_nxt)) { 1090 tcp_sk(sk)->snd_nxt; 3322 READ_ONCE(tp->snd_nxt);
|
H A D | tcp_minisocks.c | 311 tcptw->tw_snd_nxt = tp->snd_nxt; 542 WRITE_ONCE(newtp->snd_nxt, seq);
|
H A D | tcp_nv.c | 338 ca->nv_rtt_start_seq = tp->snd_nxt; 354 ca->nv_rtt_start_seq = tp->snd_nxt +
|
H A D | tcp_output.c | 73 WRITE_ONCE(tp->snd_nxt, TCP_SKB_CB(skb)->end_seq); 101 if (!before(tcp_wnd_end(tp), tp->snd_nxt) || 103 ((tp->snd_nxt - tcp_wnd_end(tp)) < (1 << tp->rx_opt.rcv_wscale)))) 104 return tp->snd_nxt; 384 !before(TCP_SKB_CB(skb)->seq, tp->snd_nxt)) { 1395 } else if (after(tcb->seq + 0xFFFF, tp->snd_nxt)) { 1448 if (after(tcb->end_seq, tp->snd_nxt) || tcb->seq == tcb->end_seq) 1673 if (!before(tp->snd_nxt, TCP_SKB_CB(buff)->end_seq)) { 1931 tp->cwnd_usage_seq = tp->snd_nxt; 1966 !after(tp->snd_sml, tp->snd_nxt); [all...] |
H A D | tcp_rate.c | 199 tp->write_seq - tp->snd_nxt < tp->mss_cache &&
|
H A D | tcp_timer.c | 555 inet->inet_num, tp->snd_una, tp->snd_nxt, 563 inet->inet_num, tp->snd_una, tp->snd_nxt,
|
H A D | tcp_vegas.c | 80 vegas->beg_snd_nxt = tp->snd_nxt; 181 vegas->beg_snd_nxt = tp->snd_nxt;
|
H A D | tcp_yeah.c | 84 * of snd_una and snd_nxt at the beginning of the current RTT. More 178 yeah->vegas.beg_snd_nxt = tp->snd_nxt;
|
/linux-master/net/ipv6/ |
H A D | tcp_ipv6.c | 437 !between(seq, snd_una, tp->snd_nxt)) { 1281 tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt,
|
/linux-master/net/mptcp/ |
H A D | options.c | 1041 u64 new_wnd_end, new_snd_una, snd_nxt = READ_ONCE(msk->snd_nxt); local 1055 if (unlikely(after64(new_snd_una, snd_nxt))) 1064 if (after64(msk->wnd_end, snd_nxt)) 1135 __mptcp_snd_una_update(msk, READ_ONCE(msk->snd_nxt));
|
H A D | protocol.c | 1041 if (snd_una == msk->snd_nxt && snd_una == msk->write_seq) { 1298 if (snd_una != msk->snd_nxt || tcp_write_queue_tail(ssk)) { 1468 burst = min_t(int, MPTCP_SEND_BURST_SIZE, mptcp_wnd_end(msk) - msk->snd_nxt); 1499 /* snd_nxt_new can be smaller than snd_nxt in case mptcp 1508 if (likely(after64(snd_nxt_new, msk->snd_nxt))) { 1509 msk->bytes_sent += snd_nxt_new - msk->snd_nxt; 1510 WRITE_ONCE(msk->snd_nxt, snd_nxt_new); 1694 msk->snd_nxt + 1 == msk->write_seq) 1786 not_sent = msk->write_seq - msk->snd_nxt; 2364 msk->recovery_snd_nxt = msk->snd_nxt; [all...] |
H A D | protocol.h | 267 u64 snd_nxt; member in struct:mptcp_sock 428 if (msk->snd_una == msk->snd_nxt) 853 READ_ONCE(msk->write_seq) == READ_ONCE(msk->snd_nxt); 870 notsent_bytes = READ_ONCE(msk->write_seq) - READ_ONCE(msk->snd_nxt);
|
H A D | subflow.c | 457 WRITE_ONCE(msk->snd_nxt, msk->write_seq);
|
/linux-master/net/tipc/ |
H A D | link.c | 120 * @snd_nxt: next sequence number to be used 179 u16 snd_nxt; member in struct:tipc_link 377 rcv_l->acked = snd_l->snd_nxt - 1; 386 u16 ack = snd_l->snd_nxt - 1; 413 if (more(l->snd_nxt, l->rcv_nxt)) 414 gap = l->snd_nxt - l->rcv_nxt; 833 u16 bc_snt = l->bc_sndlink->snd_nxt - 1; 856 if (l->snd_nxt == l->checkpoint) { 860 l->checkpoint = l->snd_nxt; 996 l->snd_nxt [all...] |
H A D | name_distr.c | 106 msg_set_named_seqno(buf_msg(skb), nt->snd_nxt++); 135 msg_set_named_seqno(buf_msg(skb), nt->snd_nxt++); 217 seqno = nt->snd_nxt;
|
H A D | name_table.h | 102 * @snd_nxt: next sequence number to be used 111 u32 snd_nxt; member in struct:name_table
|
/linux-master/tools/include/uapi/linux/ |
H A D | bpf.h | 6351 __u32 snd_nxt; /* Next sequence we send */ member in struct:bpf_tcp_sock 6794 __u32 snd_nxt; member in struct:bpf_sock_ops
|