/xnu-2782.1.97/bsd/netinet/ |
H A D | tcp_cc.h | 104 void (*congestion_avd) (struct tcpcb *tp, struct tcphdr *th); 107 void (*ack_rcvd) (struct tcpcb *tp, struct tcphdr *th); 113 void (*post_fr) (struct tcpcb *tp, struct tcphdr *th); 122 int (*delay_ack)(struct tcpcb *tp, struct tcphdr *th); 140 extern int tcp_cc_delay_ack(struct tcpcb *tp, struct tcphdr *th); 141 extern void tcp_ccdbg_trace(struct tcpcb *tp, struct tcphdr *th,
|
H A D | tcp_debug.c | 110 tcp_trace(act, ostate, tp, ipgen, th, req) 114 struct tcphdr *th; 167 if (th) { 170 td->td_ti.ti_t = *th; 171 bzero((caddr_t)&td->td_ti6.th, sizeof(td->td_ti6.th)); 175 td->td_ti6.th = *th; 183 bzero((caddr_t)&td->td_ti6.th, sizeof(td->td_ti6.th)); [all...] |
H A D | tcp_input.c | 311 static void tcp_bad_rexmt_restore_state(struct tcpcb *tp, struct tcphdr *th); 313 struct tcphdr *th); 314 static void tcp_early_rexmt_check(struct tcpcb *tp, struct tcphdr *th); 315 static void tcp_bad_rexmt_check(struct tcpcb *tp, struct tcphdr *th, 355 #define DELAY_ACK(tp, th) \ 356 (CC_ALGO(tp)->delay_ack != NULL && CC_ALGO(tp)->delay_ack(tp, th)) 359 static void tcp_newreno_partial_ack(struct tcpcb *tp, struct tcphdr *th); 556 tcp_reass(struct tcpcb *tp, struct tcphdr *th, int *tlenp, struct mbuf *m, argument 575 * Call with th==0 after become established to 578 if (th 1243 tcp_detect_bad_rexmt(struct tcpcb *tp, struct tcphdr *th, struct tcpopt *to, u_int32_t rxtime) argument 1276 tcp_bad_rexmt_restore_state(struct tcpcb *tp, struct tcphdr *th) argument 1332 tcp_bad_rexmt_check(struct tcpcb *tp, struct tcphdr *th, struct tcpopt *to) argument 1390 tcp_early_rexmt_check(struct tcpcb *tp, struct tcphdr *th) argument 1490 register struct tcphdr *th; local 4694 tcp_compute_rtt(struct tcpcb *tp, struct tcpopt *to, struct tcphdr *th) argument 5558 tcp_input_checksum(int af, struct mbuf *m, struct tcphdr *th, int off, int tlen) argument [all...] |
H A D | tcp_debug.h | 86 struct tcphdr th; member in struct:tcp_debug::__anon396
|
H A D | tcp_newreno.c | 89 void tcp_newreno_congestion_avd(struct tcpcb *tp, struct tcphdr *th); 90 void tcp_newreno_ack_rcvd(struct tcpcb *tp, struct tcphdr *th); 92 void tcp_newreno_post_fr(struct tcpcb *tp, struct tcphdr *th); 95 int tcp_newreno_delay_ack(struct tcpcb *tp, struct tcphdr *th); 147 tcp_newreno_congestion_avd(struct tcpcb *tp, struct tcphdr *th) { argument 149 acked = BYTES_ACKED(th, tp); 165 tcp_newreno_ack_rcvd(struct tcpcb *tp, struct tcphdr *th) { argument 183 acked = BYTES_ACKED(th, tp); 239 tcp_newreno_post_fr(struct tcpcb *tp, struct tcphdr *th) { argument 242 ss = tp->snd_max - th 324 tcp_newreno_delay_ack(struct tcpcb *tp, struct tcphdr *th) argument [all...] |
H A D | tcp_cubic.c | 59 static void tcp_cubic_congestion_avd(struct tcpcb *tp, struct tcphdr *th); 60 static void tcp_cubic_ack_rcvd(struct tcpcb *tp, struct tcphdr *th); 62 static void tcp_cubic_post_fr(struct tcpcb *tp, struct tcphdr *th); 64 static int tcp_cubic_delay_ack(struct tcpcb *tp, struct tcphdr *th); 67 static uint32_t tcp_cubic_tcpwin(struct tcpcb *tp, struct tcphdr *th); 248 tcp_cubic_tcpwin(struct tcpcb *tp, struct tcphdr *th) argument 255 BYTES_ACKED(th, tp); 270 tcp_cubic_congestion_avd(struct tcpcb *tp, struct tcphdr *th) argument 274 tp->t_bytes_acked += BYTES_ACKED(th, tp); 285 tcp_win = tcp_cubic_tcpwin(tp, th); 319 tcp_cubic_ack_rcvd(struct tcpcb *tp, struct tcphdr *th) argument 422 tcp_cubic_post_fr(struct tcpcb *tp, struct tcphdr *th) argument 464 tcp_cubic_delay_ack(struct tcpcb *tp, struct tcphdr *th) argument [all...] |
H A D | tcp_ledbat.c | 61 void tcp_ledbat_congestion_avd(struct tcpcb *tp, struct tcphdr *th); 62 void tcp_ledbat_ack_rcvd(struct tcpcb *tp, struct tcphdr *th); 64 void tcp_ledbat_post_fr(struct tcpcb *tp, struct tcphdr *th); 67 int tcp_ledbat_delay_ack(struct tcpcb *tp, struct tcphdr *th); 223 tcp_ledbat_congestion_avd(struct tcpcb *tp, struct tcphdr *th) { argument 227 acked = BYTES_ACKED(th, tp); 241 tcp_ledbat_ack_rcvd(struct tcpcb *tp, struct tcphdr *th) { argument 259 acked = BYTES_ACKED(th, tp); 304 tcp_ledbat_post_fr(struct tcpcb *tp, struct tcphdr *th) { argument 307 ss = tp->snd_max - th 404 tcp_ledbat_delay_ack(struct tcpcb *tp, struct tcphdr *th) argument [all...] |
H A D | tcp_cc.c | 195 tcp_ccdbg_trace(struct tcpcb *tp, struct tcphdr *th, int32_t event) argument 198 #pragma unused(th) 255 struct tcpcb *, tp, struct tcphdr *, th, int32_t, event); 335 tcp_cc_delay_ack(struct tcpcb *tp, struct tcphdr *th) argument 338 if ((th->th_flags & ~TH_ACK)) 347 (th->th_flags & TH_PUSH) == 0 && 353 (th->th_flags & TH_PUSH) == 0 &&
|
H A D | mptcp_opt.c | 1048 mptcp_do_mpcapable_opt(struct tcpcb *tp, u_char *cp, struct tcphdr *th, argument 1078 if ((th->th_flags & (TH_SYN | TH_ACK)) == TH_SYN) { 1081 } else if ((th->th_flags & (TH_SYN | TH_ACK)) == (TH_SYN | TH_ACK)) { 1112 optlen, tp->t_state, th->th_sport, th->th_dport, 1116 } else if ((th->th_flags & TH_ACK) && 1211 mptcp_do_mpjoin_opt(struct tcpcb *tp, u_char *cp, struct tcphdr *th, int optlen) argument 1224 if ((th->th_flags & (TH_SYN | TH_ACK)) == TH_SYN) { 1268 } else if ((th->th_flags & (TH_SYN | TH_ACK)) == (TH_SYN | TH_ACK)) { 1310 } else if ((th 1712 mptcp_do_dss_opt(struct tcpcb *tp, u_char *cp, struct tcphdr *th, int optlen) argument 1742 mptcp_do_fastclose_opt(struct tcpcb *tp, u_char *cp, struct tcphdr *th) argument 1796 mptcp_do_mpfail_opt(struct tcpcb *tp, u_char *cp, struct tcphdr *th) argument 1836 tcp_do_mptcp_options(struct tcpcb *tp, u_char *cp, struct tcphdr *th, struct tcpopt *to, int optlen) argument [all...] |
H A D | tcp_output.c | 245 static struct mbuf* tcp_send_lroacks(struct tcpcb *tp, struct mbuf *m, struct tcphdr *th); 287 struct tcphdr *th; local 774 * 2. send buffer is filled to 7/8th with data (so we actually 1679 th = (struct tcphdr *)(void *)(ip6 + 1); 1680 tcp_fillheaders(tp, ip6, th); 1695 th = (struct tcphdr *)(void *)(ip + 1); 1697 tcp_fillheaders(tp, ip, th); 1731 th->th_seq = htonl(tp->snd_nxt); 1733 th->th_seq = htonl(tp->snd_max); 1735 th 2567 tcp_send_lroacks(struct tcpcb *tp, struct mbuf *m, struct tcphdr *th) argument [all...] |
H A D | tcp_subr.c | 603 tcp_respond(struct tcpcb *tp, void *ipgen, struct tcphdr *th, struct mbuf *m, argument 673 bcopy((caddr_t)th, (caddr_t)nth, sizeof(struct tcphdr)); 701 if (th != nth) { 707 nth->th_sport = th->th_sport; 708 nth->th_dport = th->th_dport; 775 tcp_trace(TA_OUTPUT, 0, tp, mtod(m, void *), th, 0); 1832 struct tcphdr th; local 1838 &th, sizeof (th)); 1839 inp = in_pcblookup_hash(&tcbinfo, faddr, th 1918 struct tcphdr th; local 2429 struct tcphdr *th; local [all...] |
H A D | tcp_sack.c | 392 tcp_sack_doack(struct tcpcb *tp, struct tcpopt *to, struct tcphdr *th, argument 398 tcp_seq old_snd_fack = 0, th_ack = th->th_ack; 638 tcp_sack_partialack(tp, th) 640 struct tcphdr *th; 647 if (((BYTES_ACKED(th, tp)) / tp->t_maxseg) > 2)
|
/xnu-2782.1.97/osfmk/ipc/ |
H A D | ipc_mqueue.c | 247 thread_t th; local 250 th = wait_queue_wakeup64_identity_locked( 257 if (th == THREAD_NULL) 266 if (th->ith_state != MACH_RCV_IN_PROGRESS) { 267 thread_unlock(th); 279 msize = ipc_kmsg_copyout_size(kmsg, th->map); 280 if (th->ith_msize < 281 (msize + REQUESTED_TRAILER_SIZE(thread_is_64bit(th), th->ith_option))) { 282 th [all...] |
/xnu-2782.1.97/osfmk/kern/ |
H A D | sched.h | 366 #define sched_run_incr(th) \ 369 #define sched_run_decr(th) \ 382 #define sched_share_incr(th) \ 387 #define sched_share_decr(th) \ 393 #define sched_background_incr(th) \ 398 #define sched_background_decr(th) \ 403 #define assert_thread_sched_count(th) \
|
H A D | bsd_kern.c | 78 void *get_bsdthreadtask_info(thread_t th) argument 80 return(th->task != TASK_NULL ? th->task->bsd_info : NULL); 94 void *get_bsdthread_info(thread_t th) argument 96 return(th->uthread); 102 int get_thread_lock_count(thread_t th); /* forced forward */ 103 int get_thread_lock_count(thread_t th) argument 105 return(th->mutex_count); 426 task_t get_threadtask(thread_t th) argument 428 return(th 535 get_thread_userstop( thread_t th) argument 565 thread_should_abort( thread_t th) argument 583 thread_t th = current_thread(); local [all...] |
H A D | thread.h | 220 uint16_t grrr_deficit; /* fixed point (1/1000th quantum) fractional deficit */ 554 #define thread_lock_init(th) simple_lock_init(&(th)->sched_lock, 0) 555 #define thread_lock(th) simple_lock(&(th)->sched_lock) 556 #define thread_unlock(th) simple_unlock(&(th)->sched_lock) 558 #define wake_lock_init(th) simple_lock_init(&(th)->wake_lock, 0) 559 #define wake_lock(th) simple_loc [all...] |
H A D | priority.c | 293 static int do_priority_computation(thread_t th) { argument 294 register int priority = th->priority /* start with base priority */ 295 - (th->sched_usage >> th->pri_shift);
|
/xnu-2782.1.97/osfmk/i386/ |
H A D | pmap.h | 611 #define PMAP_SWITCH_USER(th, new_map, my_cpu) { \ 615 PMAP_DEACTIVATE_MAP(th->map, th); \ 616 th->map = new_map; \ 617 PMAP_ACTIVATE_MAP(th->map, th); \ 619 inval_copy_windows(th); \ 622 #define PMAP_SWITCH_USER(th, new_map, my_cpu) { \ 626 PMAP_DEACTIVATE_MAP(th->map, th, my_cp [all...] |
/xnu-2782.1.97/bsd/net/ |
H A D | pf_norm.c | 2100 struct tcphdr *th = pd->hdr.tcp; local 2109 sxport.port = th->th_sport; 2110 dxport.port = th->th_dport; 2138 !pf_osfp_match(pf_osfp_fingerprint(pd, m, off, th), 2165 flags = th->th_flags; 2186 if (th->th_off < (sizeof (struct tcphdr) >> 2)) 2190 if (flags != th->th_flags || th->th_x2 != 0) { 2193 ov = *(u_int16_t *)(&th->th_ack + 1); 2194 th 2243 pf_normalize_tcp_init(struct mbuf *m, int off, struct pf_pdesc *pd, struct tcphdr *th, struct pf_state_peer *src, struct pf_state_peer *dst) argument 2338 pf_normalize_tcp_stateful(struct mbuf *m, int off, struct pf_pdesc *pd, u_short *reason, struct tcphdr *th, struct pf_state *state, struct pf_state_peer *src, struct pf_state_peer *dst, int *writeback) argument 2738 pf_normalize_tcpopt(struct pf_rule *r, int dir, struct pfi_kif *kif, struct pf_pdesc *pd, struct mbuf *m, struct tcphdr *th, int off, int *rewrptr) argument [all...] |
H A D | pf.c | 2321 struct tcphdr *th, struct pf_state_peer *dst) 2323 int hlen = (th->th_off << 2) - sizeof (*th), thoptlen = hlen; 2330 !pf_pull_hdr(m, off + sizeof (*th), opts, hlen, NULL, NULL, pd->af)) 2348 pf_change_a(&sack.start, &th->th_sum, 2351 pf_change_a(&sack.end, &th->th_sum, 2356 copyback = off + sizeof (*th) + thoptlen; 2371 m_copyback(m, off + sizeof (*th), thoptlen, opts); 2392 struct tcphdr *th = NULL; local 2477 th 2320 pf_modulate_sack(struct mbuf *m, int off, struct pf_pdesc *pd, struct tcphdr *th, struct pf_state_peer *dst) argument 4060 struct tcphdr *th = pd->hdr.tcp; local 5151 struct tcphdr *th = pd->hdr.tcp; local 5502 struct tcphdr *th; local 5850 struct tcphdr *th = pd->hdr.tcp; local 6778 struct tcphdr th; local 8216 struct tcphdr th; local 8807 struct tcphdr th; local [all...] |
H A D | pktap.c | 854 struct tcphdr th; local 857 sizeof(struct tcphdr), &th); 861 fport = th.th_sport; 862 lport = th.th_dport; 913 struct tcphdr th; local 916 sizeof(struct tcphdr), &th); 924 fport = th.th_sport; 925 lport = th.th_dport;
|
/xnu-2782.1.97/tools/lldbmacros/ |
H A D | net.py | 793 out_string += "\nAlloc: (thread " + hex(rtd.rtd_alloc.th) + "):\n" 804 out_string += "\nFree: (thread " + hex(rtd.rtd_free.th) + "):\n" 816 out_string += "\nHold [" + str(int(cnt)) + "] (thread " + hex(rtd.rtd_refhold[cnt].th) + "):\n" 830 out_string += "\nRelease [" + str(int(cnt)) + "] (thread " + hex(rtd.rtd_refrele[cnt].th) + "):\n" 847 out_string += "\nLock [" + str(int(cnt)) + "] (thread " + hex(rtd.rtd_lock[cnt].th) + "):\n" 861 out_string += "\nUnlock [" + str(int(cnt)) + "] (thread " + hex(rtd.rtd_unlock[cnt].th) + "):\n" 891 out_string += "\nAlloc: (thread " + hex(inifa.inifa_alloc.th) + "):\n" 902 out_string += "\nFree: (thread " + hex(inifa.inifa_free.th) + "):\n" 914 out_string += "\nHold [" + str(int(cnt)) + "] (thread " + hex(inifa.inifa_refhold[cnt].th) + "):\n" 928 out_string += "\nRelease [" + str(int(cnt)) + "] (thread " + hex(inifa.inifa_refrele[cnt].th) [all...] |
/xnu-2782.1.97/osfmk/kdp/ |
H A D | kdp_udp.c | 1571 struct corehdr *th = NULL; local 1601 th = create_panic_header(request, corename, (unsigned)length, panic_block); 1609 memset(th->th_data + length, 'Y', 1613 if (!kdp_machine_vm_read((mach_vm_address_t)(uintptr_t)panic_data, (caddr_t) th->th_data, length)) { 1615 memset((caddr_t) th->th_data, 'X', (size_t)length); 1618 if (!kdp_machine_vm_read((mach_vm_address_t)(uintptr_t)next_page, (caddr_t) th->th_data + (length - resid), resid)) { 1619 memset((caddr_t) th->th_data + (length - resid), 'X', (size_t)resid); 1626 *(uint64_t *) th->th_data = OSSwapHostToBigInt64((*(uint64_t *) panic_data)); 1628 *(unsigned int *) th->th_data = htonl(*(unsigned int *) panic_data); 1644 th [all...] |
/xnu-2782.1.97/tools/lldbmacros/core/ |
H A D | operating_system.py | 738 th = self.version.CreateValueFromExpression(str(th_ptr),'(struct thread *)' + str(th_ptr)) 739 thread_id = th.GetChildMemberWithName('thread_id').GetValueAsUnsigned() 744 'ptr' : th.GetValueAsUnsigned(), 745 'name' : hex(th.GetValueAsUnsigned()).rstrip('L'), 746 'queue' : hex(th.GetChildMemberWithName('wait_queue').GetValueAsUnsigned()).rstrip('L'), 789 th = processor_list_val.GetChildMemberWithName('active_thread') 790 th_id = th.GetChildMemberWithName('thread_id').GetValueAsUnsigned() 792 self.processors.append({'active_thread': th.GetValueAsUnsigned(), 'cpu_id': cpu_id}) 793 self.create_thread(th_id, th.GetValueAsUnsigned()) 813 for th i [all...] |
/xnu-2782.1.97/bsd/netinet6/ |
H A D | ip6_fw.c | 842 struct tcphdr th; member in struct:__anon404 851 ti.th = *tcp; 852 ti.th.th_seq = ntohl(ti.th.th_seq); 853 ti.th.th_ack = ntohl(ti.th.th_ack); 855 if (ti.th.th_flags & TH_ACK) { 857 seq = ti.th.th_ack; 860 ack = ti.th.th_seq; 863 - (ti.th [all...] |