rxnet 734 net/rxrpc/af_rxrpc.c struct rxrpc_net *rxnet; rxnet 774 net/rxrpc/af_rxrpc.c rxnet = rxrpc_net(sock_net(&rx->sk)); rxnet 775 net/rxrpc/af_rxrpc.c timer_reduce(&rxnet->peer_keepalive_timer, jiffies + 1); rxnet 270 net/rxrpc/ar-internal.h struct rxrpc_net *rxnet; /* The network ns in which this resides */ rxnet 563 net/rxrpc/ar-internal.h struct rxrpc_net *rxnet; /* Network namespace to which call belongs */ rxnet 38 net/rxrpc/call_accept.c struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk)); rxnet 80 net/rxrpc/call_accept.c conn = rxrpc_prealloc_service_connection(rxnet, gfp); rxnet 137 net/rxrpc/call_accept.c rxnet = call->rxnet; rxnet 138 net/rxrpc/call_accept.c write_lock(&rxnet->call_lock); rxnet 139 net/rxrpc/call_accept.c list_add_tail(&call->link, &rxnet->calls); rxnet 140 net/rxrpc/call_accept.c write_unlock(&rxnet->call_lock); rxnet 188 net/rxrpc/call_accept.c struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk)); rxnet 213 net/rxrpc/call_accept.c write_lock(&rxnet->conn_lock); rxnet 216 net/rxrpc/call_accept.c write_unlock(&rxnet->conn_lock); rxnet 218 net/rxrpc/call_accept.c if (atomic_dec_and_test(&rxnet->nr_conns)) rxnet 219 net/rxrpc/call_accept.c wake_up_var(&rxnet->nr_conns); rxnet 102 net/rxrpc/call_object.c struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk)); rxnet 155 net/rxrpc/call_object.c call->rxnet = rxnet; rxnet 156 net/rxrpc/call_object.c atomic_inc(&rxnet->nr_calls); rxnet 226 net/rxrpc/call_object.c struct rxrpc_net *rxnet; rxnet 278 net/rxrpc/call_object.c rxnet = call->rxnet; rxnet 279 net/rxrpc/call_object.c write_lock(&rxnet->call_lock); rxnet 280 net/rxrpc/call_object.c list_add_tail(&call->link, &rxnet->calls); rxnet 281 net/rxrpc/call_object.c write_unlock(&rxnet->call_lock); rxnet 539 net/rxrpc/call_object.c struct rxrpc_net *rxnet = call->rxnet; rxnet 554 net/rxrpc/call_object.c write_lock(&rxnet->call_lock); rxnet 556 net/rxrpc/call_object.c write_unlock(&rxnet->call_lock); rxnet 569 net/rxrpc/call_object.c struct rxrpc_net *rxnet = call->rxnet; rxnet 576 net/rxrpc/call_object.c if (atomic_dec_and_test(&rxnet->nr_calls)) rxnet 577 net/rxrpc/call_object.c wake_up_var(&rxnet->nr_calls); rxnet 621 net/rxrpc/call_object.c void rxrpc_destroy_all_calls(struct rxrpc_net *rxnet) rxnet 627 net/rxrpc/call_object.c if (!list_empty(&rxnet->calls)) { rxnet 628 net/rxrpc/call_object.c write_lock(&rxnet->call_lock); rxnet 630 net/rxrpc/call_object.c while (!list_empty(&rxnet->calls)) { rxnet 631 net/rxrpc/call_object.c call = list_entry(rxnet->calls.next, rxnet 643 net/rxrpc/call_object.c write_unlock(&rxnet->call_lock); rxnet 645 net/rxrpc/call_object.c write_lock(&rxnet->call_lock); rxnet 648 net/rxrpc/call_object.c write_unlock(&rxnet->call_lock); rxnet 651 net/rxrpc/call_object.c atomic_dec(&rxnet->nr_calls); rxnet 652 net/rxrpc/call_object.c wait_var_event(&rxnet->nr_calls, !atomic_read(&rxnet->nr_calls)); rxnet 103 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = conn->params.local->rxnet; rxnet 119 net/rxrpc/conn_client.c conn->proto.epoch = rxnet->epoch; rxnet 171 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = cp->local->rxnet; rxnet 205 net/rxrpc/conn_client.c atomic_inc(&rxnet->nr_conns); rxnet 206 net/rxrpc/conn_client.c write_lock(&rxnet->conn_lock); rxnet 207 net/rxrpc/conn_client.c list_add_tail(&conn->proc_link, &rxnet->conn_proc_list); rxnet 208 net/rxrpc/conn_client.c write_unlock(&rxnet->conn_lock); rxnet 237 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = conn->params.local->rxnet; rxnet 243 net/rxrpc/conn_client.c if (conn->proto.epoch != rxnet->epoch) rxnet 450 net/rxrpc/conn_client.c static void rxrpc_activate_conn(struct rxrpc_net *rxnet, rxnet 460 net/rxrpc/conn_client.c rxnet->nr_active_client_conns++; rxnet 461 net/rxrpc/conn_client.c list_move_tail(&conn->cache_link, &rxnet->active_client_conns); rxnet 476 net/rxrpc/conn_client.c static void rxrpc_animate_client_conn(struct rxrpc_net *rxnet, rxnet 487 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 489 net/rxrpc/conn_client.c nr_conns = rxnet->nr_client_conns; rxnet 492 net/rxrpc/conn_client.c rxnet->nr_client_conns = nr_conns + 1; rxnet 513 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 520 net/rxrpc/conn_client.c rxrpc_activate_conn(rxnet, conn); rxnet 527 net/rxrpc/conn_client.c list_move_tail(&conn->cache_link, &rxnet->waiting_client_conns); rxnet 700 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = cp->local->rxnet; rxnet 705 net/rxrpc/conn_client.c rxrpc_discard_expired_client_conns(&rxnet->client_conn_reaper); rxnet 706 net/rxrpc/conn_client.c rxrpc_cull_active_client_conns(rxnet); rxnet 712 net/rxrpc/conn_client.c rxrpc_animate_client_conn(rxnet, call->conn); rxnet 774 net/rxrpc/conn_client.c static void rxrpc_set_client_reap_timer(struct rxrpc_net *rxnet) rxnet 779 net/rxrpc/conn_client.c if (rxnet->live) rxnet 780 net/rxrpc/conn_client.c timer_reduce(&rxnet->client_conn_reap_timer, reap_at); rxnet 790 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = conn->params.local->rxnet; rxnet 819 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 873 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 889 net/rxrpc/conn_client.c rxnet->nr_active_client_conns--; rxnet 915 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 929 net/rxrpc/conn_client.c list_move_tail(&conn->cache_link, &rxnet->idle_client_conns); rxnet 930 net/rxrpc/conn_client.c if (rxnet->idle_client_conns.next == &conn->cache_link && rxnet 931 net/rxrpc/conn_client.c !rxnet->kill_all_client_conns) rxnet 932 net/rxrpc/conn_client.c rxrpc_set_client_reap_timer(rxnet); rxnet 949 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = local->rxnet; rxnet 968 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 969 net/rxrpc/conn_client.c nr_conns = --rxnet->nr_client_conns; rxnet 972 net/rxrpc/conn_client.c !list_empty(&rxnet->waiting_client_conns)) { rxnet 973 net/rxrpc/conn_client.c next = list_entry(rxnet->waiting_client_conns.next, rxnet 976 net/rxrpc/conn_client.c rxrpc_activate_conn(rxnet, next); rxnet 979 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 1015 net/rxrpc/conn_client.c static void rxrpc_cull_active_client_conns(struct rxrpc_net *rxnet) rxnet 1018 net/rxrpc/conn_client.c unsigned int nr_conns = rxnet->nr_client_conns; rxnet 1030 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 1031 net/rxrpc/conn_client.c nr_active = rxnet->nr_active_client_conns; rxnet 1034 net/rxrpc/conn_client.c ASSERT(!list_empty(&rxnet->active_client_conns)); rxnet 1035 net/rxrpc/conn_client.c conn = list_entry(rxnet->active_client_conns.next, rxnet 1048 net/rxrpc/conn_client.c &rxnet->waiting_client_conns); rxnet 1054 net/rxrpc/conn_client.c rxnet->nr_active_client_conns = nr_active; rxnet 1055 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 1070 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = rxnet 1077 net/rxrpc/conn_client.c if (list_empty(&rxnet->idle_client_conns)) { rxnet 1083 net/rxrpc/conn_client.c if (!spin_trylock(&rxnet->client_conn_discard_lock)) { rxnet 1091 net/rxrpc/conn_client.c nr_conns = rxnet->nr_client_conns; rxnet 1094 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 1096 net/rxrpc/conn_client.c if (list_empty(&rxnet->idle_client_conns)) rxnet 1099 net/rxrpc/conn_client.c conn = list_entry(rxnet->idle_client_conns.next, rxnet 1103 net/rxrpc/conn_client.c if (!rxnet->kill_all_client_conns) { rxnet 1128 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 1147 net/rxrpc/conn_client.c if (!rxnet->kill_all_client_conns) rxnet 1148 net/rxrpc/conn_client.c timer_reduce(&rxnet->client_conn_reap_timer, rxnet 1152 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 1153 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_discard_lock); rxnet 1161 net/rxrpc/conn_client.c void rxrpc_destroy_all_client_connections(struct rxrpc_net *rxnet) rxnet 1165 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 1166 net/rxrpc/conn_client.c rxnet->kill_all_client_conns = true; rxnet 1167 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 1169 net/rxrpc/conn_client.c del_timer_sync(&rxnet->client_conn_reap_timer); rxnet 1171 net/rxrpc/conn_client.c if (!rxrpc_queue_work(&rxnet->client_conn_reaper)) rxnet 1183 net/rxrpc/conn_client.c struct rxrpc_net *rxnet = local->rxnet; rxnet 1189 net/rxrpc/conn_client.c spin_lock(&rxnet->client_conn_cache_lock); rxnet 1190 net/rxrpc/conn_client.c nr_active = rxnet->nr_active_client_conns; rxnet 1192 net/rxrpc/conn_client.c list_for_each_entry_safe(conn, tmp, &rxnet->idle_client_conns, rxnet 1206 net/rxrpc/conn_client.c rxnet->nr_active_client_conns = nr_active; rxnet 1207 net/rxrpc/conn_client.c spin_unlock(&rxnet->client_conn_cache_lock); rxnet 235 net/rxrpc/conn_object.c struct rxrpc_net *rxnet = conn->params.local->rxnet; rxnet 243 net/rxrpc/conn_object.c write_lock(&rxnet->conn_lock); rxnet 245 net/rxrpc/conn_object.c write_unlock(&rxnet->conn_lock); rxnet 322 net/rxrpc/conn_object.c static void rxrpc_set_service_reap_timer(struct rxrpc_net *rxnet, rxnet 325 net/rxrpc/conn_object.c if (rxnet->live) rxnet 326 net/rxrpc/conn_object.c timer_reduce(&rxnet->service_conn_reap_timer, reap_at); rxnet 342 net/rxrpc/conn_object.c rxrpc_set_service_reap_timer(conn->params.local->rxnet, rxnet 368 net/rxrpc/conn_object.c if (atomic_dec_and_test(&conn->params.local->rxnet->nr_conns)) rxnet 369 net/rxrpc/conn_object.c wake_up_var(&conn->params.local->rxnet->nr_conns); rxnet 382 net/rxrpc/conn_object.c struct rxrpc_net *rxnet = rxnet 393 net/rxrpc/conn_object.c write_lock(&rxnet->conn_lock); rxnet 394 net/rxrpc/conn_object.c list_for_each_entry_safe(conn, _p, &rxnet->service_conns, link) { rxnet 401 net/rxrpc/conn_object.c if (rxnet->live && !conn->params.local->dead) { rxnet 432 net/rxrpc/conn_object.c write_unlock(&rxnet->conn_lock); rxnet 437 net/rxrpc/conn_object.c rxrpc_set_service_reap_timer(rxnet, earliest); rxnet 456 net/rxrpc/conn_object.c void rxrpc_destroy_all_connections(struct rxrpc_net *rxnet) rxnet 463 net/rxrpc/conn_object.c atomic_dec(&rxnet->nr_conns); rxnet 464 net/rxrpc/conn_object.c rxrpc_destroy_all_client_connections(rxnet); rxnet 466 net/rxrpc/conn_object.c del_timer_sync(&rxnet->service_conn_reap_timer); rxnet 467 net/rxrpc/conn_object.c rxrpc_queue_work(&rxnet->service_conn_reaper); rxnet 470 net/rxrpc/conn_object.c write_lock(&rxnet->conn_lock); rxnet 471 net/rxrpc/conn_object.c list_for_each_entry_safe(conn, _p, &rxnet->service_conns, link) { rxnet 476 net/rxrpc/conn_object.c write_unlock(&rxnet->conn_lock); rxnet 479 net/rxrpc/conn_object.c ASSERT(list_empty(&rxnet->conn_proc_list)); rxnet 484 net/rxrpc/conn_object.c wait_var_event(&rxnet->nr_conns, !atomic_read(&rxnet->nr_conns)); rxnet 119 net/rxrpc/conn_service.c struct rxrpc_connection *rxrpc_prealloc_service_connection(struct rxrpc_net *rxnet, rxnet 131 net/rxrpc/conn_service.c atomic_inc(&rxnet->nr_conns); rxnet 132 net/rxrpc/conn_service.c write_lock(&rxnet->conn_lock); rxnet 133 net/rxrpc/conn_service.c list_add_tail(&conn->link, &rxnet->service_conns); rxnet 134 net/rxrpc/conn_service.c list_add_tail(&conn->proc_link, &rxnet->conn_proc_list); rxnet 135 net/rxrpc/conn_service.c write_unlock(&rxnet->conn_lock); rxnet 74 net/rxrpc/local_object.c static struct rxrpc_local *rxrpc_alloc_local(struct rxrpc_net *rxnet, rxnet 83 net/rxrpc/local_object.c local->rxnet = rxnet; rxnet 225 net/rxrpc/local_object.c struct rxrpc_net *rxnet = rxrpc_net(net); rxnet 234 net/rxrpc/local_object.c mutex_lock(&rxnet->local_mutex); rxnet 236 net/rxrpc/local_object.c for (cursor = rxnet->local_endpoints.next; rxnet 237 net/rxrpc/local_object.c cursor != &rxnet->local_endpoints; rxnet 268 net/rxrpc/local_object.c local = rxrpc_alloc_local(rxnet, srx); rxnet 276 net/rxrpc/local_object.c if (cursor != &rxnet->local_endpoints) rxnet 283 net/rxrpc/local_object.c mutex_unlock(&rxnet->local_mutex); rxnet 294 net/rxrpc/local_object.c mutex_unlock(&rxnet->local_mutex); rxnet 301 net/rxrpc/local_object.c mutex_unlock(&rxnet->local_mutex); rxnet 413 net/rxrpc/local_object.c struct rxrpc_net *rxnet = local->rxnet; rxnet 419 net/rxrpc/local_object.c mutex_lock(&rxnet->local_mutex); rxnet 421 net/rxrpc/local_object.c mutex_unlock(&rxnet->local_mutex); rxnet 424 net/rxrpc/local_object.c rxrpc_service_connection_reaper(&rxnet->service_conn_reaper); rxnet 496 net/rxrpc/local_object.c void rxrpc_destroy_all_locals(struct rxrpc_net *rxnet) rxnet 504 net/rxrpc/local_object.c if (!list_empty(&rxnet->local_endpoints)) { rxnet 505 net/rxrpc/local_object.c mutex_lock(&rxnet->local_mutex); rxnet 506 net/rxrpc/local_object.c list_for_each_entry(local, &rxnet->local_endpoints, link) { rxnet 510 net/rxrpc/local_object.c mutex_unlock(&rxnet->local_mutex); rxnet 15 net/rxrpc/net_ns.c struct rxrpc_net *rxnet = rxnet 18 net/rxrpc/net_ns.c if (rxnet->live) rxnet 19 net/rxrpc/net_ns.c rxrpc_queue_work(&rxnet->client_conn_reaper); rxnet 24 net/rxrpc/net_ns.c struct rxrpc_net *rxnet = rxnet 27 net/rxrpc/net_ns.c if (rxnet->live) rxnet 28 net/rxrpc/net_ns.c rxrpc_queue_work(&rxnet->service_conn_reaper); rxnet 33 net/rxrpc/net_ns.c struct rxrpc_net *rxnet = rxnet 36 net/rxrpc/net_ns.c if (rxnet->live) rxnet 37 net/rxrpc/net_ns.c rxrpc_queue_work(&rxnet->peer_keepalive_work); rxnet 45 net/rxrpc/net_ns.c struct rxrpc_net *rxnet = rxrpc_net(net); rxnet 48 net/rxrpc/net_ns.c rxnet->live = true; rxnet 49 net/rxrpc/net_ns.c get_random_bytes(&rxnet->epoch, sizeof(rxnet->epoch)); rxnet 50 net/rxrpc/net_ns.c rxnet->epoch |= RXRPC_RANDOM_EPOCH; rxnet 52 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->calls); rxnet 53 net/rxrpc/net_ns.c rwlock_init(&rxnet->call_lock); rxnet 54 net/rxrpc/net_ns.c atomic_set(&rxnet->nr_calls, 1); rxnet 56 net/rxrpc/net_ns.c atomic_set(&rxnet->nr_conns, 1); rxnet 57 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->conn_proc_list); rxnet 58 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->service_conns); rxnet 59 net/rxrpc/net_ns.c rwlock_init(&rxnet->conn_lock); rxnet 60 net/rxrpc/net_ns.c INIT_WORK(&rxnet->service_conn_reaper, rxnet 62 net/rxrpc/net_ns.c timer_setup(&rxnet->service_conn_reap_timer, rxnet 65 net/rxrpc/net_ns.c rxnet->nr_client_conns = 0; rxnet 66 net/rxrpc/net_ns.c rxnet->nr_active_client_conns = 0; rxnet 67 net/rxrpc/net_ns.c rxnet->kill_all_client_conns = false; rxnet 68 net/rxrpc/net_ns.c spin_lock_init(&rxnet->client_conn_cache_lock); rxnet 69 net/rxrpc/net_ns.c spin_lock_init(&rxnet->client_conn_discard_lock); rxnet 70 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->waiting_client_conns); rxnet 71 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->active_client_conns); rxnet 72 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->idle_client_conns); rxnet 73 net/rxrpc/net_ns.c INIT_WORK(&rxnet->client_conn_reaper, rxnet 75 net/rxrpc/net_ns.c timer_setup(&rxnet->client_conn_reap_timer, rxnet 78 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->local_endpoints); rxnet 79 net/rxrpc/net_ns.c mutex_init(&rxnet->local_mutex); rxnet 81 net/rxrpc/net_ns.c hash_init(rxnet->peer_hash); rxnet 82 net/rxrpc/net_ns.c spin_lock_init(&rxnet->peer_hash_lock); rxnet 83 net/rxrpc/net_ns.c for (i = 0; i < ARRAY_SIZE(rxnet->peer_keepalive); i++) rxnet 84 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->peer_keepalive[i]); rxnet 85 net/rxrpc/net_ns.c INIT_LIST_HEAD(&rxnet->peer_keepalive_new); rxnet 86 net/rxrpc/net_ns.c timer_setup(&rxnet->peer_keepalive_timer, rxnet 88 net/rxrpc/net_ns.c INIT_WORK(&rxnet->peer_keepalive_work, rxrpc_peer_keepalive_worker); rxnet 89 net/rxrpc/net_ns.c rxnet->peer_keepalive_base = ktime_get_seconds(); rxnet 92 net/rxrpc/net_ns.c rxnet->proc_net = proc_net_mkdir(net, "rxrpc", net->proc_net); rxnet 93 net/rxrpc/net_ns.c if (!rxnet->proc_net) rxnet 96 net/rxrpc/net_ns.c proc_create_net("calls", 0444, rxnet->proc_net, &rxrpc_call_seq_ops, rxnet 98 net/rxrpc/net_ns.c proc_create_net("conns", 0444, rxnet->proc_net, rxnet 101 net/rxrpc/net_ns.c proc_create_net("peers", 0444, rxnet->proc_net, rxnet 107 net/rxrpc/net_ns.c rxnet->live = false; rxnet 116 net/rxrpc/net_ns.c struct rxrpc_net *rxnet = rxrpc_net(net); rxnet 118 net/rxrpc/net_ns.c rxnet->live = false; rxnet 119 net/rxrpc/net_ns.c del_timer_sync(&rxnet->peer_keepalive_timer); rxnet 120 net/rxrpc/net_ns.c cancel_work_sync(&rxnet->peer_keepalive_work); rxnet 121 net/rxrpc/net_ns.c rxrpc_destroy_all_calls(rxnet); rxnet 122 net/rxrpc/net_ns.c rxrpc_destroy_all_connections(rxnet); rxnet 123 net/rxrpc/net_ns.c rxrpc_destroy_all_peers(rxnet); rxnet 124 net/rxrpc/net_ns.c rxrpc_destroy_all_locals(rxnet); rxnet 125 net/rxrpc/net_ns.c proc_remove(rxnet->proc_net); rxnet 602 net/rxrpc/output.c whdr.epoch = htonl(peer->local->rxnet->epoch); rxnet 301 net/rxrpc/peer_event.c static void rxrpc_peer_keepalive_dispatch(struct rxrpc_net *rxnet, rxnet 307 net/rxrpc/peer_event.c const u8 mask = ARRAY_SIZE(rxnet->peer_keepalive) - 1; rxnet 311 net/rxrpc/peer_event.c spin_lock_bh(&rxnet->peer_hash_lock); rxnet 322 net/rxrpc/peer_event.c spin_unlock_bh(&rxnet->peer_hash_lock); rxnet 341 net/rxrpc/peer_event.c spin_lock_bh(&rxnet->peer_hash_lock); rxnet 343 net/rxrpc/peer_event.c &rxnet->peer_keepalive[slot & mask]); rxnet 349 net/rxrpc/peer_event.c spin_unlock_bh(&rxnet->peer_hash_lock); rxnet 357 net/rxrpc/peer_event.c struct rxrpc_net *rxnet = rxnet 359 net/rxrpc/peer_event.c const u8 mask = ARRAY_SIZE(rxnet->peer_keepalive) - 1; rxnet 365 net/rxrpc/peer_event.c base = rxnet->peer_keepalive_base; rxnet 366 net/rxrpc/peer_event.c cursor = rxnet->peer_keepalive_cursor; rxnet 369 net/rxrpc/peer_event.c if (!rxnet->live) rxnet 379 net/rxrpc/peer_event.c spin_lock_bh(&rxnet->peer_hash_lock); rxnet 380 net/rxrpc/peer_event.c list_splice_init(&rxnet->peer_keepalive_new, &collector); rxnet 382 net/rxrpc/peer_event.c stop = cursor + ARRAY_SIZE(rxnet->peer_keepalive); rxnet 384 net/rxrpc/peer_event.c list_splice_tail_init(&rxnet->peer_keepalive[cursor & mask], rxnet 391 net/rxrpc/peer_event.c spin_unlock_bh(&rxnet->peer_hash_lock); rxnet 393 net/rxrpc/peer_event.c rxnet->peer_keepalive_base = base; rxnet 394 net/rxrpc/peer_event.c rxnet->peer_keepalive_cursor = cursor; rxnet 395 net/rxrpc/peer_event.c rxrpc_peer_keepalive_dispatch(rxnet, &collector, base, cursor); rxnet 399 net/rxrpc/peer_event.c cursor = rxnet->peer_keepalive_cursor; rxnet 402 net/rxrpc/peer_event.c if (!list_empty(&rxnet->peer_keepalive[cursor & mask])) rxnet 412 net/rxrpc/peer_event.c if (rxnet->live) rxnet 413 net/rxrpc/peer_event.c timer_reduce(&rxnet->peer_keepalive_timer, jiffies + delay); rxnet 120 net/rxrpc/peer_object.c struct rxrpc_net *rxnet = local->rxnet; rxnet 122 net/rxrpc/peer_object.c hash_for_each_possible_rcu(rxnet->peer_hash, peer, hash_link, hash_key) { rxnet 308 net/rxrpc/peer_object.c struct rxrpc_net *rxnet = local->rxnet; rxnet 314 net/rxrpc/peer_object.c spin_lock(&rxnet->peer_hash_lock); rxnet 315 net/rxrpc/peer_object.c hash_add_rcu(rxnet->peer_hash, &peer->hash_link, hash_key); rxnet 316 net/rxrpc/peer_object.c list_add_tail(&peer->keepalive_link, &rxnet->peer_keepalive_new); rxnet 317 net/rxrpc/peer_object.c spin_unlock(&rxnet->peer_hash_lock); rxnet 328 net/rxrpc/peer_object.c struct rxrpc_net *rxnet = local->rxnet; rxnet 350 net/rxrpc/peer_object.c spin_lock_bh(&rxnet->peer_hash_lock); rxnet 357 net/rxrpc/peer_object.c hash_add_rcu(rxnet->peer_hash, rxnet 360 net/rxrpc/peer_object.c &rxnet->peer_keepalive_new); rxnet 363 net/rxrpc/peer_object.c spin_unlock_bh(&rxnet->peer_hash_lock); rxnet 412 net/rxrpc/peer_object.c struct rxrpc_net *rxnet = peer->local->rxnet; rxnet 416 net/rxrpc/peer_object.c spin_lock_bh(&rxnet->peer_hash_lock); rxnet 419 net/rxrpc/peer_object.c spin_unlock_bh(&rxnet->peer_hash_lock); rxnet 466 net/rxrpc/peer_object.c void rxrpc_destroy_all_peers(struct rxrpc_net *rxnet) rxnet 471 net/rxrpc/peer_object.c for (i = 0; i < HASH_SIZE(rxnet->peer_hash); i++) { rxnet 472 net/rxrpc/peer_object.c if (hlist_empty(&rxnet->peer_hash[i])) rxnet 475 net/rxrpc/peer_object.c hlist_for_each_entry(peer, &rxnet->peer_hash[i], hash_link) { rxnet 29 net/rxrpc/proc.c __acquires(rxnet->call_lock) rxnet 31 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 34 net/rxrpc/proc.c read_lock(&rxnet->call_lock); rxnet 35 net/rxrpc/proc.c return seq_list_start_head(&rxnet->calls, *_pos); rxnet 40 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 42 net/rxrpc/proc.c return seq_list_next(v, &rxnet->calls, pos); rxnet 46 net/rxrpc/proc.c __releases(rxnet->call_lock) rxnet 49 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 51 net/rxrpc/proc.c read_unlock(&rxnet->call_lock); rxnet 61 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 66 net/rxrpc/proc.c if (v == &rxnet->calls) { rxnet 133 net/rxrpc/proc.c __acquires(rxnet->conn_lock) rxnet 135 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 137 net/rxrpc/proc.c read_lock(&rxnet->conn_lock); rxnet 138 net/rxrpc/proc.c return seq_list_start_head(&rxnet->conn_proc_list, *_pos); rxnet 144 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 146 net/rxrpc/proc.c return seq_list_next(v, &rxnet->conn_proc_list, pos); rxnet 150 net/rxrpc/proc.c __releases(rxnet->conn_lock) rxnet 152 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 154 net/rxrpc/proc.c read_unlock(&rxnet->conn_lock); rxnet 160 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 163 net/rxrpc/proc.c if (v == &rxnet->conn_proc_list) { rxnet 255 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 257 net/rxrpc/proc.c unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash); rxnet 268 net/rxrpc/proc.c if (bucket >= HASH_SIZE(rxnet->peer_hash)) { rxnet 279 net/rxrpc/proc.c p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); rxnet 290 net/rxrpc/proc.c struct rxrpc_net *rxnet = rxrpc_net(seq_file_net(seq)); rxnet 292 net/rxrpc/proc.c unsigned int shift = 32 - HASH_BITS(rxnet->peer_hash); rxnet 300 net/rxrpc/proc.c p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); rxnet 309 net/rxrpc/proc.c if (bucket >= HASH_SIZE(rxnet->peer_hash)) { rxnet 318 net/rxrpc/proc.c p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1);