ilb 9357 kernel/sched/fair.c int ilb; ilb 9359 kernel/sched/fair.c for_each_cpu_and(ilb, nohz.idle_cpus_mask, ilb 9361 kernel/sched/fair.c if (idle_cpu(ilb)) ilb 9362 kernel/sched/fair.c return ilb; ilb 912 net/ipv4/inet_diag.c struct inet_listen_hashbucket *ilb; ilb 916 net/ipv4/inet_diag.c ilb = &hashinfo->listening_hash[i]; ilb 917 net/ipv4/inet_diag.c spin_lock(&ilb->lock); ilb 918 net/ipv4/inet_diag.c sk_nulls_for_each(sk, node, &ilb->nulls_head) { ilb 939 net/ipv4/inet_diag.c spin_unlock(&ilb->lock); ilb 946 net/ipv4/inet_diag.c spin_unlock(&ilb->lock); ilb 516 net/ipv4/inet_hashtables.c struct inet_listen_hashbucket *ilb) ilb 523 net/ipv4/inet_hashtables.c sk_nulls_for_each_rcu(sk2, node, &ilb->nulls_head) { ilb 541 net/ipv4/inet_hashtables.c struct inet_listen_hashbucket *ilb; ilb 549 net/ipv4/inet_hashtables.c ilb = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)]; ilb 551 net/ipv4/inet_hashtables.c spin_lock(&ilb->lock); ilb 553 net/ipv4/inet_hashtables.c err = inet_reuseport_add_sock(sk, ilb); ilb 559 net/ipv4/inet_hashtables.c __sk_nulls_add_node_tail_rcu(sk, &ilb->nulls_head); ilb 561 net/ipv4/inet_hashtables.c __sk_nulls_add_node_rcu(sk, &ilb->nulls_head); ilb 563 net/ipv4/inet_hashtables.c ilb->count++; ilb 567 net/ipv4/inet_hashtables.c spin_unlock(&ilb->lock); ilb 590 net/ipv4/inet_hashtables.c struct inet_listen_hashbucket *ilb = NULL; ilb 597 net/ipv4/inet_hashtables.c ilb = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)]; ilb 598 net/ipv4/inet_hashtables.c lock = &ilb->lock; ilb 608 net/ipv4/inet_hashtables.c if (ilb) { ilb 610 net/ipv4/inet_hashtables.c ilb->count--; ilb 2151 net/ipv4/tcp_ipv4.c struct inet_listen_hashbucket *ilb; ilb 2157 net/ipv4/tcp_ipv4.c ilb = &tcp_hashinfo.listening_hash[st->bucket]; ilb 2158 net/ipv4/tcp_ipv4.c spin_lock(&ilb->lock); ilb 2159 net/ipv4/tcp_ipv4.c sk = sk_nulls_head(&ilb->nulls_head); ilb 2163 net/ipv4/tcp_ipv4.c ilb = &tcp_hashinfo.listening_hash[st->bucket]; ilb 2175 net/ipv4/tcp_ipv4.c spin_unlock(&ilb->lock);