/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
H A D | ninet_hashtables.c | 193 struct inet_listen_hashbucket *ilb; in __ninet_hash() local 203 ilb = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)]; in __ninet_hash() 205 spin_lock(&ilb->lock); in __ninet_hash() 207 __sk_nulls_add_node_rcu(sk, &ilb->nulls_head); in __ninet_hash() 210 ilb->count++; in __ninet_hash() 214 spin_unlock(&ilb->lock); in __ninet_hash() 249 static void __ninet_unhash(struct sock *sk, struct inet_listen_hashbucket *ilb) in __ninet_unhash() argument 254 if (ilb) { in __ninet_unhash() 258 if (ilb->count) in __ninet_unhash() 259 ilb in __ninet_unhash() 273 struct inet_listen_hashbucket *ilb; ninet_unhash() local [all...] |
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | inet_hashtables.c | 615 struct inet_listen_hashbucket *ilb) in inet_reuseport_add_sock() 622 sk_nulls_for_each_rcu(sk2, node, &ilb->nulls_head) { in inet_reuseport_add_sock() 640 struct inet_listen_hashbucket *ilb; in __inet_hash() local 650 ilb = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)]; in __inet_hash() 652 spin_lock(&ilb->lock); in __inet_hash() 654 err = inet_reuseport_add_sock(sk, ilb); in __inet_hash() 660 __sk_nulls_add_node_tail_rcu(sk, &ilb->nulls_head); in __inet_hash() 662 __sk_nulls_add_node_rcu(sk, &ilb->nulls_head); in __inet_hash() 664 ilb->count++; in __inet_hash() 668 spin_unlock(&ilb in __inet_hash() 614 inet_reuseport_add_sock(struct sock *sk, struct inet_listen_hashbucket *ilb) inet_reuseport_add_sock() argument 685 __inet_unhash(struct sock *sk, struct inet_listen_hashbucket *ilb) __inet_unhash() argument 710 struct inet_listen_hashbucket *ilb; inet_unhash() local [all...] |
H A D | inet_diag.c | 1027 struct inet_listen_hashbucket *ilb; in inet_diag_dump_icsk() local 1031 ilb = &hashinfo->listening_hash[i]; in inet_diag_dump_icsk() 1032 spin_lock(&ilb->lock); in inet_diag_dump_icsk() 1033 sk_nulls_for_each(sk, node, &ilb->nulls_head) { in inet_diag_dump_icsk() 1058 spin_unlock(&ilb->lock); in inet_diag_dump_icsk() 1065 spin_unlock(&ilb->lock); in inet_diag_dump_icsk()
|
H A D | tcp_ipv4.c | 2293 struct inet_listen_hashbucket *ilb; in listening_get_next() local 2304 ilb = &tcp_hashinfo.listening_hash[st->bucket]; in listening_get_next() 2305 spin_lock(&ilb->lock); in listening_get_next() 2306 sk = sk_nulls_head(&ilb->nulls_head); in listening_get_next() 2310 ilb = &tcp_hashinfo.listening_hash[st->bucket]; in listening_get_next() 2323 spin_unlock(&ilb->lock); in listening_get_next()
|
/kernel/linux/linux-6.6/net/mptcp/ |
H A D | mptcp_diag.c | 90 struct inet_listen_hashbucket *ilb; in mptcp_diag_dump_listeners() local 95 ilb = &hinfo->lhash2[i]; in mptcp_diag_dump_listeners() 98 spin_lock(&ilb->lock); in mptcp_diag_dump_listeners() 99 sk_nulls_for_each(sk, node, &ilb->nulls_head) { in mptcp_diag_dump_listeners() 130 spin_unlock(&ilb->lock); in mptcp_diag_dump_listeners() 141 spin_unlock(&ilb->lock); in mptcp_diag_dump_listeners()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | inet_diag.c | 1032 struct inet_listen_hashbucket *ilb; in inet_diag_dump_icsk() local 1036 ilb = &hashinfo->lhash2[i]; in inet_diag_dump_icsk() 1038 spin_lock(&ilb->lock); in inet_diag_dump_icsk() 1039 sk_nulls_for_each(sk, node, &ilb->nulls_head) { in inet_diag_dump_icsk() 1064 spin_unlock(&ilb->lock); in inet_diag_dump_icsk() 1071 spin_unlock(&ilb->lock); in inet_diag_dump_icsk()
|
H A D | inet_hashtables.c | 711 struct inet_listen_hashbucket *ilb) in inet_reuseport_add_sock() 718 sk_nulls_for_each_rcu(sk2, node, &ilb->nulls_head) { in inet_reuseport_add_sock() 710 inet_reuseport_add_sock(struct sock *sk, struct inet_listen_hashbucket *ilb) inet_reuseport_add_sock() argument
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | fair.c | 12144 int ilb; in find_new_ilb() local 12149 for_each_cpu_and(ilb, nohz.idle_cpus_mask, hk_mask) { in find_new_ilb() 12151 if (cpu_isolated(ilb)) in find_new_ilb() 12154 if (ilb == smp_processor_id()) in find_new_ilb() 12157 if (idle_cpu(ilb)) in find_new_ilb() 12158 return ilb; in find_new_ilb() 12173 * Increase nohz.next_balance only when if full ilb is triggered but in kick_ilb()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | fair.c | 10760 int ilb; in find_new_ilb() local 10762 for_each_cpu_and(ilb, nohz.idle_cpus_mask, in find_new_ilb() 10764 if (cpu_isolated(ilb)) in find_new_ilb() 10767 if (idle_cpu(ilb)) in find_new_ilb() 10768 return ilb; in find_new_ilb() 10783 * Increase nohz.next_balance only when if full ilb is triggered but in kick_ilb() 11198 * need to be updated. Run the ilb locally as it is a good in nohz_newidle_balance() 11199 * candidate for ilb instead of waking up another idle CPU. in nohz_newidle_balance() 11200 * Kick an normal ilb if we failed to do the update. in nohz_newidle_balance()
|