Lines Matching refs:rxnet

30 	struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
33 return seq_list_start_head_rcu(&rxnet->calls, *_pos);
38 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
40 return seq_list_next_rcu(v, &rxnet->calls, pos);
53 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
59 if (v == &rxnet->calls) {
118 __acquires(rxnet->conn_lock)
120 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
122 read_lock(&rxnet->conn_lock);
123 return seq_list_start_head(&rxnet->conn_proc_list, *_pos);
129 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
131 return seq_list_next(v, &rxnet->conn_proc_list, pos);
135 __releases(rxnet->conn_lock)
137 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
139 read_unlock(&rxnet->conn_lock);
145 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
149 if (v == &rxnet->conn_proc_list) {
244 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
246 unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash);
257 if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
268 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1);
279 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
281 unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash);
289 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos);
298 if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
307 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1);
359 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
371 return seq_hlist_start_rcu(&rxnet->local_endpoints, n - 1);
376 struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq));
381 return seq_hlist_next_rcu(v, &rxnet->local_endpoints, _pos);
402 struct rxrpc_net *rxnet = rxrpc_net(seq_file_single_net(seq));
406 atomic_read(&rxnet->stat_tx_data_send),
407 atomic_read(&rxnet->stat_tx_data_send_frag),
408 atomic_read(&rxnet->stat_tx_data_send_fail));
411 atomic_read(&rxnet->stat_tx_data),
412 atomic_read(&rxnet->stat_tx_data_retrans),
413 atomic_read(&rxnet->stat_tx_data_underflow),
414 atomic_read(&rxnet->stat_tx_data_cwnd_reset));
417 atomic_read(&rxnet->stat_rx_data),
418 atomic_read(&rxnet->stat_rx_data_reqack),
419 atomic_read(&rxnet->stat_rx_data_jumbo));
422 atomic_read(&rxnet->stat_tx_ack_fill),
423 atomic_read(&rxnet->stat_tx_ack_send),
424 atomic_read(&rxnet->stat_tx_ack_skip));
427 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_REQUESTED]),
428 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_DUPLICATE]),
429 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_OUT_OF_SEQUENCE]),
430 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_EXCEEDS_WINDOW]),
431 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_NOSPACE]),
432 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_PING]),
433 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_PING_RESPONSE]),
434 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_DELAY]),
435 atomic_read(&rxnet->stat_tx_acks[RXRPC_ACK_IDLE]));
438 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_REQUESTED]),
439 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_DUPLICATE]),
440 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_OUT_OF_SEQUENCE]),
441 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_EXCEEDS_WINDOW]),
442 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_NOSPACE]),
443 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_PING]),
444 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_PING_RESPONSE]),
445 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_DELAY]),
446 atomic_read(&rxnet->stat_rx_acks[RXRPC_ACK_IDLE]));
449 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_ack_lost]),
450 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_already_on]),
451 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_more_rtt]),
452 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_old_rtt]));
455 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_no_srv_last]),
456 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_retrans]),
457 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_slow_start]),
458 atomic_read(&rxnet->stat_why_req_ack[rxrpc_reqack_small_txwin]));
465 atomic_read(&rxnet->stat_io_loop));
475 struct rxrpc_net *rxnet = rxrpc_net(seq_file_single_net(m));
480 atomic_set(&rxnet->stat_tx_data, 0);
481 atomic_set(&rxnet->stat_tx_data_retrans, 0);
482 atomic_set(&rxnet->stat_tx_data_underflow, 0);
483 atomic_set(&rxnet->stat_tx_data_cwnd_reset, 0);
484 atomic_set(&rxnet->stat_tx_data_send, 0);
485 atomic_set(&rxnet->stat_tx_data_send_frag, 0);
486 atomic_set(&rxnet->stat_tx_data_send_fail, 0);
487 atomic_set(&rxnet->stat_rx_data, 0);
488 atomic_set(&rxnet->stat_rx_data_reqack, 0);
489 atomic_set(&rxnet->stat_rx_data_jumbo, 0);
491 atomic_set(&rxnet->stat_tx_ack_fill, 0);
492 atomic_set(&rxnet->stat_tx_ack_send, 0);
493 atomic_set(&rxnet->stat_tx_ack_skip, 0);
494 memset(&rxnet->stat_tx_acks, 0, sizeof(rxnet->stat_tx_acks));
495 memset(&rxnet->stat_rx_acks, 0, sizeof(rxnet->stat_rx_acks));
497 memset(&rxnet->stat_why_req_ack, 0, sizeof(rxnet->stat_why_req_ack));
499 atomic_set(&rxnet->stat_io_loop, 0);