Lines Matching defs:bucket
147 held not per host, but per port pair and TW bucket is used as state
150 If TW bucket has been already destroyed we fall back to VJ's scheme
157 * and releasing the bucket lock.
2265 /* Clean up a referenced TCP bind bucket. */
2285 * starting from bucket given in st->bucket; when st->bucket is zero the
2304 ilb = &tcp_hashinfo.listening_hash[st->bucket];
2310 ilb = &tcp_hashinfo.listening_hash[st->bucket];
2325 if (++st->bucket < INET_LHTABLE_SIZE)
2335 st->bucket = 0;
2348 return hlist_nulls_empty(&tcp_hashinfo.ehash[st->bucket].chain);
2352 * Get first established socket starting from bucket given in st->bucket.
2353 * If st->bucket is zero, the very first socket in the hash is returned.
2368 for (; st->bucket <= tcp_hashinfo.ehash_mask; ++st->bucket) {
2371 spinlock_t *lock = inet_ehash_lockp(&tcp_hashinfo, st->bucket);
2378 sk_nulls_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) {
2418 spin_unlock_bh(inet_ehash_lockp(&tcp_hashinfo, st->bucket));
2419 ++st->bucket;
2428 st->bucket = 0;
2457 int bucket = st->bucket;
2464 if (st->bucket >= INET_LHTABLE_SIZE)
2468 while (offset-- && rc && bucket == st->bucket)
2472 st->bucket = 0;
2476 if (st->bucket > tcp_hashinfo.ehash_mask)
2479 while (offset-- && rc && bucket == st->bucket)
2501 st->bucket = 0;
2526 st->bucket = 0;
2549 spin_unlock(&tcp_hashinfo.listening_hash[st->bucket].lock);
2553 spin_unlock_bh(inet_ehash_lockp(&tcp_hashinfo, st->bucket));