Lines Matching refs:rxnet

30 	struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
33 return seq_list_start_head_rcu(&rxnet->calls, *_pos);
38 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
40 return seq_list_next_rcu(v, &rxnet->calls, pos);
53 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
59 if (v == &rxnet->calls) {
116 __acquires(rxnet->conn_lock)
118 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
120 read_lock(&rxnet->conn_lock);
121 return seq_list_start_head(&rxnet->conn_proc_list, *_pos);
127 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
129 return seq_list_next(v, &rxnet->conn_proc_list, pos);
133 __releases(rxnet->conn_lock)
135 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
137 read_unlock(&rxnet->conn_lock);
143 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
147 if (v == &rxnet->conn_proc_list) {
203 __acquires(rxnet->conn_lock)
205 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
207 read_lock(&rxnet->conn_lock);
208 return seq_list_start_head(&rxnet->bundle_proc_list, *_pos);
214 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
216 return seq_list_next(v, &rxnet->bundle_proc_list, pos);
220 __releases(rxnet->conn_lock)
222 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
224 read_unlock(&rxnet->conn_lock);
230 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
233 if (v == &rxnet->bundle_proc_list) {
318 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
320 unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash);
331 if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
342 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1);
353 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
355 unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash);
363 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos);
372 if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
381 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1);
433 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
445 return seq_hlist_start_rcu(&rxnet->local_endpoints, n - 1);
450 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
455 return seq_hlist_next_rcu(v, &rxnet->local_endpoints, _pos);
476 struct rxrpc_net *rxnet = rxrpc_net(seq_file_single_net(seq));
480 atomic_read(&rxnet->stat_tx_data_send),
481 atomic_read(&rxnet->stat_tx_data_send_frag),
482 atomic_read(&rxnet->stat_tx_data_send_fail));
485 atomic_read(&rxnet->stat_tx_data),
486 atomic_read(&rxnet->stat_tx_data_retrans),
487 atomic_read(&rxnet->stat_tx_data_underflow),
488 atomic_read(&rxnet->stat_tx_data_cwnd_reset));
491 atomic_read(&rxnet->stat_rx_data),
492 atomic_read(&rxnet->stat_rx_data_reqack),
493 atomic_read(&rxnet->stat_rx_data_jumbo));
496 atomic_read(&rxnet->stat_tx_ack_fill),
497 atomic_read(&rxnet->stat_tx_ack_send),
498 atomic_read(&rxnet->stat_tx_ack_skip));
501 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_REQUESTED]),
502 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_DUPLICATE]),
503 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_OUT_OF_SEQUENCE]),
504 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_EXCEEDS_WINDOW]),
505 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_NOSPACE]),
506 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_PING]),
507 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_PING_RESPONSE]),
508 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_DELAY]),
509 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_IDLE]));
512 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_REQUESTED]),
513 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_DUPLICATE]),
514 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_OUT_OF_SEQUENCE]),
515 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_EXCEEDS_WINDOW]),
516 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_NOSPACE]),
517 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_PING]),
518 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_PING_RESPONSE]),
519 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_DELAY]),
520 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_IDLE]));
523 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_ack_lost]),
524 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_already_on]),
525 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_more_rtt]),
526 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_old_rtt]));
529 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_no_srv_last]),
530 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_retrans]),
531 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_slow_start]),
532 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_small_txwin]));
539 atomic_read(&rxnet->stat_io_loop));
549 struct rxrpc_net *rxnet = rxrpc_net(seq_file_single_net(m));
554 atomic_set(&rxnet->stat_tx_data, 0);
555 atomic_set(&rxnet->stat_tx_data_retrans, 0);
556 atomic_set(&rxnet->stat_tx_data_underflow, 0);
557 atomic_set(&rxnet->stat_tx_data_cwnd_reset, 0);
558 atomic_set(&rxnet->stat_tx_data_send, 0);
559 atomic_set(&rxnet->stat_tx_data_send_frag, 0);
560 atomic_set(&rxnet->stat_tx_data_send_fail, 0);
561 atomic_set(&rxnet->stat_rx_data, 0);
562 atomic_set(&rxnet->stat_rx_data_reqack, 0);
563 atomic_set(&rxnet->stat_rx_data_jumbo, 0);
565 atomic_set(&rxnet->stat_tx_ack_fill, 0);
566 atomic_set(&rxnet->stat_tx_ack_send, 0);
567 atomic_set(&rxnet->stat_tx_ack_skip, 0);
568 memset(&rxnet->stat_tx_acks, 0, sizeof(rxnet->stat_tx_acks));
569 memset(&rxnet->stat_rx_acks, 0, sizeof(rxnet->stat_rx_acks));
571 memset(&rxnet->stat_why_req_ack, 0, sizeof(rxnet->stat_why_req_ack));
573 atomic_set(&rxnet->stat_io_loop, 0);