/kernel/linux/linux-5.10/arch/mips/netlogic/xlr/ |
H A D | fmn-config.c | 50 int bkt; in print_credit_config() local 54 for (bkt = 0; bkt < 16; bkt++) in print_credit_config() 56 xlr_board_fmn_config.bucket_size[(bkt * 8) + 0], in print_credit_config() 57 xlr_board_fmn_config.bucket_size[(bkt * 8) + 1], in print_credit_config() 58 xlr_board_fmn_config.bucket_size[(bkt * 8) + 2], in print_credit_config() 59 xlr_board_fmn_config.bucket_size[(bkt * 8) + 3], in print_credit_config() 60 xlr_board_fmn_config.bucket_size[(bkt * 8) + 4], in print_credit_config() 61 xlr_board_fmn_config.bucket_size[(bkt * in print_credit_config() 84 int bkt, n, total_credits, ncores; check_credit_distribution() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | hashtable.h | 122 * @bkt: integer to use as bucket loop cursor 126 #define hash_for_each(name, bkt, obj, member) \ 127 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 128 (bkt)++)\ 129 hlist_for_each_entry(obj, &name[bkt], member) 134 * @bkt: integer to use as bucket loop cursor 138 #define hash_for_each_rcu(name, bkt, obj, member) \ 139 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
H A D | rhashtable.h | 327 struct rhash_lock_head __rcu **bkt) in rht_lock() 330 bit_spin_lock(0, (unsigned long *)bkt); in rht_lock() 344 struct rhash_lock_head __rcu **bkt) in rht_unlock() 347 bit_spin_unlock(0, (unsigned long *)bkt); in rht_unlock() 352 struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) in __rht_ptr() 356 (unsigned long)RHT_NULLS_MARKER(bkt)); in __rht_ptr() 360 * Where 'bkt' is a bucket and might be locked: 367 struct rhash_lock_head __rcu *const *bkt) in rht_ptr_rcu() 369 return __rht_ptr(rcu_dereference(*bkt), bkt); in rht_ptr_rcu() 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument 343 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_unlock() argument 351 __rht_ptr( struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) __rht_ptr() argument 366 rht_ptr_rcu( struct rhash_lock_head __rcu *const *bkt) rht_ptr_rcu() argument 372 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument 380 rht_ptr_exclusive( struct rhash_lock_head __rcu *const *bkt) rht_ptr_exclusive() argument 386 rht_assign_locked(struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) rht_assign_locked() argument 394 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) rht_assign_unlock() argument 589 struct rhash_lock_head __rcu *const *bkt; __rhashtable_lookup() local 705 struct rhash_lock_head __rcu **bkt; __rhashtable_insert_fast() local 991 struct rhash_lock_head __rcu **bkt; __rhashtable_remove_fast_one() local 1143 struct rhash_lock_head __rcu **bkt; __rhashtable_replace_fast() local [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | hashtable.h | 122 * @bkt: integer to use as bucket loop cursor 126 #define hash_for_each(name, bkt, obj, member) \ 127 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 128 (bkt)++)\ 129 hlist_for_each_entry(obj, &name[bkt], member) 134 * @bkt: integer to use as bucket loop cursor 138 #define hash_for_each_rcu(name, bkt, obj, member) \ 139 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
H A D | rhashtable.h | 327 struct rhash_lock_head __rcu **bkt) in rht_lock() 332 bit_spin_lock(0, (unsigned long *)bkt); in rht_lock() 350 struct rhash_lock_head __rcu **bkt, in rht_unlock() 354 bit_spin_unlock(0, (unsigned long *)bkt); in rht_unlock() 359 struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) in __rht_ptr() 363 (unsigned long)RHT_NULLS_MARKER(bkt)); in __rht_ptr() 367 * Where 'bkt' is a bucket and might be locked: 374 struct rhash_lock_head __rcu *const *bkt) in rht_ptr_rcu() 376 return __rht_ptr(rcu_dereference(*bkt), bkt); in rht_ptr_rcu() 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument 349 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, unsigned long flags) rht_unlock() argument 358 __rht_ptr( struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) __rht_ptr() argument 373 rht_ptr_rcu( struct rhash_lock_head __rcu *const *bkt) rht_ptr_rcu() argument 379 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument 387 rht_ptr_exclusive( struct rhash_lock_head __rcu *const *bkt) rht_ptr_exclusive() argument 393 rht_assign_locked(struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) rht_assign_locked() argument 401 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj, unsigned long flags) rht_assign_unlock() argument 597 struct rhash_lock_head __rcu *const *bkt; __rhashtable_lookup() local 713 struct rhash_lock_head __rcu **bkt; __rhashtable_insert_fast() local 1000 struct rhash_lock_head __rcu **bkt; __rhashtable_remove_fast_one() local 1153 struct rhash_lock_head __rcu **bkt; __rhashtable_replace_fast() local [all...] |
/kernel/linux/linux-5.10/tools/include/linux/ |
H A D | hashtable.h | 101 * @bkt: integer to use as bucket loop cursor 105 #define hash_for_each(name, bkt, obj, member) \ 106 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 107 (bkt)++)\ 108 hlist_for_each_entry(obj, &name[bkt], member) 114 * @bkt: integer to use as bucket loop cursor 119 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ 120 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
/kernel/linux/linux-6.6/tools/include/linux/ |
H A D | hashtable.h | 101 * @bkt: integer to use as bucket loop cursor 105 #define hash_for_each(name, bkt, obj, member) \ 106 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 107 (bkt)++)\ 108 hlist_for_each_entry(obj, &name[bkt], member) 114 * @bkt: integer to use as bucket loop cursor 119 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ 120 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
/kernel/linux/linux-5.10/tools/lib/bpf/ |
H A D | hashmap.h | 153 * @bkt: integer used as a bucket loop cursor 155 #define hashmap__for_each_entry(map, cur, bkt) \ 156 for (bkt = 0; bkt < map->cap; bkt++) \ 157 for (cur = map->buckets[bkt]; cur; cur = cur->next) 165 * @bkt: integer used as a bucket loop cursor 167 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 168 for (bkt = 0; bkt < ma [all...] |
H A D | hashmap.c | 66 size_t bkt; in hashmap__clear() local 68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap__clear() 106 size_t h, bkt; in hashmap_grow() local 117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap_grow()
|
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | hashmap.h | 153 * @bkt: integer used as a bucket loop cursor 155 #define hashmap__for_each_entry(map, cur, bkt) \ 156 for (bkt = 0; bkt < map->cap; bkt++) \ 157 for (cur = map->buckets[bkt]; cur; cur = cur->next) 165 * @bkt: integer used as a bucket loop cursor 167 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 168 for (bkt = 0; bkt < ma [all...] |
/kernel/linux/linux-6.6/tools/lib/bpf/ |
H A D | hashmap.h | 166 * @bkt: integer used as a bucket loop cursor 168 #define hashmap__for_each_entry(map, cur, bkt) \ 169 for (bkt = 0; bkt < map->cap; bkt++) \ 170 for (cur = map->buckets[bkt]; cur; cur = cur->next) 178 * @bkt: integer used as a bucket loop cursor 180 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 181 for (bkt = 0; bkt < ma [all...] |
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | hashmap.h | 166 * @bkt: integer used as a bucket loop cursor 168 #define hashmap__for_each_entry(map, cur, bkt) \ 169 for (bkt = 0; bkt < map->cap; bkt++) \ 170 for (cur = map->buckets[bkt]; cur; cur = cur->next) 178 * @bkt: integer used as a bucket loop cursor 180 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 181 for (bkt = 0; bkt < ma [all...] |
H A D | expr.c | 86 size_t bkt; in ids__free() local 91 hashmap__for_each_entry(ids, cur, bkt) { in ids__free() 115 size_t bkt; in ids__union() local 133 hashmap__for_each_entry(ids2, cur, bkt) { in ids__union() 232 size_t bkt; in expr__subset_of_ids() local 235 hashmap__for_each_entry(needles->ids, cur, bkt) { in expr__subset_of_ids() 314 size_t bkt; in expr__ctx_clear() local 316 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_clear() 326 size_t bkt; in expr__ctx_free() local 332 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_free() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_internal.h | 280 #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ 281 for ((bkt) = 0, obj = NULL; \ 282 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 283 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) 289 #define vnic_hash_for_each(name, bkt, obj, member) \ 290 for ((bkt) = 0, obj = NULL; \ 291 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 292 hlist_for_each_entry(obj, &name[bkt], membe [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_internal.h | 280 #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ 281 for ((bkt) = 0, obj = NULL; \ 282 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 283 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) 289 #define vnic_hash_for_each(name, bkt, obj, member) \ 290 for ((bkt) = 0, obj = NULL; \ 291 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 292 hlist_for_each_entry(obj, &name[bkt], membe [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | hashtable_test.c | 68 int bkt; in hashtable_test_hash_add() local 80 hash_for_each(hash, bkt, x, node) { in hashtable_test_hash_add() 127 int bkt, i, j, count; in hashtable_test_hash_for_each() local 139 hash_for_each(hash, bkt, x, node) { in hashtable_test_hash_for_each() 157 int bkt, i, j, count; in hashtable_test_hash_for_each_safe() local 169 hash_for_each_safe(hash, bkt, tmp, x, node) { in hashtable_test_hash_for_each_safe() 190 int bkt, i, j, count; in hashtable_test_hash_for_each_possible() local 220 hash_for_each(hash, bkt, y, node) { in hashtable_test_hash_for_each_possible() 223 buckets[y->key] = bkt; in hashtable_test_hash_for_each_possible() 245 int bkt, in hashtable_test_hash_for_each_possible_safe() local [all...] |
H A D | rhashtable.c | 225 struct rhash_lock_head __rcu **bkt, in rhashtable_rehash_one() 241 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), in rhashtable_rehash_one() 270 rht_assign_locked(bkt, next); in rhashtable_rehash_one() 280 struct rhash_lock_head __rcu **bkt = rht_bucket_var(old_tbl, old_hash); in rhashtable_rehash_chain() local 284 if (!bkt) in rhashtable_rehash_chain() 286 flags = rht_lock(old_tbl, bkt); in rhashtable_rehash_chain() 288 while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) in rhashtable_rehash_chain() 293 rht_unlock(old_tbl, bkt, flags); in rhashtable_rehash_chain() 491 struct rhash_lock_head __rcu **bkt, in rhashtable_lookup_one() 504 rht_for_each_from(head, rht_ptr(bkt, tb in rhashtable_lookup_one() 224 rhashtable_rehash_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, unsigned int old_hash) rhashtable_rehash_one() argument 490 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument 541 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument 595 struct rhash_lock_head __rcu **bkt; rhashtable_try_insert() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_eswitch.c | 198 unsigned int bkt; in ice_eswitch_release_reprs() local 202 ice_for_each_vf(pf, bkt, vf) { in ice_eswitch_release_reprs() 229 unsigned int bkt; in ice_eswitch_setup_reprs() local 233 ice_for_each_vf(pf, bkt, vf) { in ice_eswitch_setup_reprs() 279 ice_for_each_vf(pf, bkt, vf) { in ice_eswitch_setup_reprs() 433 unsigned int bkt; in ice_eswitch_napi_del() local 437 ice_for_each_vf(pf, bkt, vf) in ice_eswitch_napi_del() 448 unsigned int bkt; in ice_eswitch_napi_enable() local 452 ice_for_each_vf(pf, bkt, vf) in ice_eswitch_napi_enable() 463 unsigned int bkt; in ice_eswitch_napi_disable() local 657 unsigned int bkt; ice_eswitch_start_all_tx_queues() local 677 unsigned int bkt; ice_eswitch_stop_all_tx_queues() local [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/prog_tests/ |
H A D | hashmap.c | 46 int err, bkt, found_cnt, i; in test_hashmap_generic() local 94 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 142 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 201 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 236 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 244 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 266 int err, bkt; in test_hashmap_multimap() local 304 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_multimap() 336 int bkt; in test_hashmap_empty() local 359 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_empty() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/prog_tests/ |
H A D | hashmap.c | 47 int err, bkt, found_cnt, i; in test_hashmap_generic() local 92 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 140 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 197 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 230 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 238 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 265 int err, i, bkt; in test_hashmap_ptr_iface() local 319 hashmap__for_each_entry(map, cur, bkt) { in test_hashmap_ptr_iface() 345 int err, bkt; in test_hashmap_multimap() local 382 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_multimap() 414 int bkt; test_hashmap_empty() local [all...] |
/kernel/linux/linux-5.10/drivers/s390/crypto/ |
H A D | ap_card.c | 80 int bkt; in request_count_store() local 85 hash_for_each(ap_queues, bkt, aq, hnode) in request_count_store() 99 int bkt; in requestq_count_show() local 106 hash_for_each(ap_queues, bkt, aq, hnode) in requestq_count_show() 118 int bkt; in pendingq_count_show() local 125 hash_for_each(ap_queues, bkt, aq, hnode) in pendingq_count_show()
|
/kernel/linux/linux-6.6/drivers/s390/crypto/ |
H A D | ap_card.c | 80 int bkt; in request_count_store() local 85 hash_for_each(ap_queues, bkt, aq, hnode) in request_count_store() 99 int bkt; in requestq_count_show() local 106 hash_for_each(ap_queues, bkt, aq, hnode) in requestq_count_show() 118 int bkt; in pendingq_count_show() local 125 hash_for_each(ap_queues, bkt, aq, hnode) in pendingq_count_show()
|
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | calipso.c | 203 u32 bkt; in calipso_cache_check() local 212 bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_check() 213 spin_lock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 214 list_for_each_entry(entry, &calipso_cache[bkt].list, list) { in calipso_cache_check() 224 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 239 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 244 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 268 u32 bkt; in calipso_cache_add() local 291 bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_add() 292 spin_lock_bh(&calipso_cache[bkt] in calipso_cache_add() [all...] |
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | calipso.c | 203 u32 bkt; in calipso_cache_check() local 212 bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_check() 213 spin_lock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 214 list_for_each_entry(entry, &calipso_cache[bkt].list, list) { in calipso_cache_check() 224 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 239 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 244 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 268 u32 bkt; in calipso_cache_add() local 291 bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_add() 292 spin_lock_bh(&calipso_cache[bkt] in calipso_cache_add() [all...] |
/kernel/linux/linux-5.10/lib/ |
H A D | rhashtable.c | 225 struct rhash_lock_head __rcu **bkt, in rhashtable_rehash_one() 240 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), in rhashtable_rehash_one() 268 rht_assign_locked(bkt, next); in rhashtable_rehash_one() 278 struct rhash_lock_head __rcu **bkt = rht_bucket_var(old_tbl, old_hash); in rhashtable_rehash_chain() local 281 if (!bkt) in rhashtable_rehash_chain() 283 rht_lock(old_tbl, bkt); in rhashtable_rehash_chain() 285 while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) in rhashtable_rehash_chain() 290 rht_unlock(old_tbl, bkt); in rhashtable_rehash_chain() 488 struct rhash_lock_head __rcu **bkt, in rhashtable_lookup_one() 501 rht_for_each_from(head, rht_ptr(bkt, tb in rhashtable_lookup_one() 224 rhashtable_rehash_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, unsigned int old_hash) rhashtable_rehash_one() argument 487 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument 538 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument 592 struct rhash_lock_head __rcu **bkt; rhashtable_try_insert() local [all...] |