/third_party/node/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/ |
H A D | mac_tool.py | 730 # Create empty hashmap buckets. 731 buckets = [None] * capacity 738 while buckets[key & capacity - 1] is not None: 740 buckets[key & capacity - 1] = (file, path) 743 for bucket in buckets: 757 for bucket in buckets:
|
/third_party/skia/third_party/externals/abseil-cpp/absl/random/ |
H A D | beta_distribution_test.cc | 279 bool SingleChiSquaredTest(double p, size_t samples, size_t buckets); 335 size_t buckets) { in SingleChiSquaredTest() 338 const double bucket_width = 1.0 / static_cast<double>(buckets); in SingleChiSquaredTest() 341 for (; i < buckets; ++i) { in SingleChiSquaredTest() 364 // Merge all remaining buckets. in SingleChiSquaredTest() 365 expected.push_back(static_cast<double>(buckets - i + 1) * bucket_width * in SingleChiSquaredTest() 367 // Make sure that we don't merge all the buckets, making this test in SingleChiSquaredTest() 334 SingleChiSquaredTest(double p, size_t samples, size_t buckets) SingleChiSquaredTest() argument
|
/kernel/linux/linux-6.6/tools/testing/selftests/net/forwarding/ |
H A D | router_mpath_nh_res.sh | 168 ip nexthop add id 103 group 101/102 type resilient buckets 512 \ 174 ip nexthop add id 106 group 104/105 type resilient buckets 512 \ 180 ip nexthop add id 203 group 201/202 type resilient buckets 512 \ 186 ip nexthop add id 206 group 204/205 type resilient buckets 512 \
|
/third_party/node/deps/brotli/c/enc/ |
H A D | encoder_dict.h | 33 const uint16_t* buckets; member
|
H A D | encoder_dict.c | 24 dict->buckets = kStaticDictionaryBuckets; in BrotliInitEncoderDictionary()
|
/third_party/skia/third_party/externals/brotli/c/enc/ |
H A D | encoder_dict.h | 33 const uint16_t* buckets; member
|
H A D | encoder_dict.c | 24 dict->buckets = kStaticDictionaryBuckets; in BrotliInitEncoderDictionary()
|
/kernel/linux/linux-5.10/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_main.c | 1277 for (neigh = rcu_dereference_bh(htbl->buckets[hash_val]); in ipoib_neigh_get() 1323 struct ipoib_neigh __rcu **np = &htbl->buckets[i]; in __ipoib_reap_neigh() 1402 for (neigh = rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1428 rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1430 rcu_assign_pointer(htbl->buckets[hash_val], neigh); in ipoib_neigh_alloc() 1487 np = &htbl->buckets[hash_val]; in ipoib_neigh_free() 1512 struct ipoib_neigh __rcu **buckets; in ipoib_neigh_hash_init() local 1521 buckets = kvcalloc(size, sizeof(*buckets), GFP_KERNEL); in ipoib_neigh_hash_init() 1522 if (!buckets) { in ipoib_neigh_hash_init() 1545 struct ipoib_neigh __rcu **buckets = htbl->buckets; neigh_hash_free_rcu() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_main.c | 1286 for (neigh = rcu_dereference_bh(htbl->buckets[hash_val]); in ipoib_neigh_get() 1332 struct ipoib_neigh __rcu **np = &htbl->buckets[i]; in __ipoib_reap_neigh() 1411 for (neigh = rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1437 rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1439 rcu_assign_pointer(htbl->buckets[hash_val], neigh); in ipoib_neigh_alloc() 1496 np = &htbl->buckets[hash_val]; in ipoib_neigh_free() 1521 struct ipoib_neigh __rcu **buckets; in ipoib_neigh_hash_init() local 1530 buckets = kvcalloc(size, sizeof(*buckets), GFP_KERNEL); in ipoib_neigh_hash_init() 1531 if (!buckets) { in ipoib_neigh_hash_init() 1554 struct ipoib_neigh __rcu **buckets = htbl->buckets; neigh_hash_free_rcu() local [all...] |
/kernel/linux/linux-5.10/drivers/lightnvm/ |
H A D | pblk-sysfs.c | 398 int buckets = pblk->min_write_pgs - 1; in pblk_sysfs_get_padding_dist() local 403 for (i = 0; i < (buckets + 1); i++) in pblk_sysfs_get_padding_dist() 411 for (i = 0; i < buckets; i++) in pblk_sysfs_get_padding_dist() 417 for (i = 0; i < buckets; i++) { in pblk_sysfs_get_padding_dist() 528 int buckets = pblk->min_write_pgs - 1; in pblk_sysfs_set_padding_dist() local 541 for (i = 0; i < buckets; i++) in pblk_sysfs_set_padding_dist()
|
/third_party/mesa3d/src/amd/registers/ |
H A D | regdb.py | 840 buckets = defaultdict(list) 842 buckets[RegisterDatabase.enum_key(enum)].append(name) 844 for bucket in buckets.values(): 858 buckets = defaultdict(list) 866 buckets[key].append((name, regtype.fields)) 868 for bucket in buckets.values():
|
/third_party/libunwind/libunwind/include/tdep-ia64/ |
H A D | script.h | 74 struct ia64_script buckets[IA64_UNW_CACHE_SIZE]; member
|
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_renderer_internal.h | 25 } buckets[27]; member 27 /* which buckets have shmems */
|
/kernel/linux/linux-5.10/net/openvswitch/ |
H A D | flow_table.c | 147 kvfree(ti->buckets); in __table_instance_destroy() 159 ti->buckets = kvmalloc_array(new_size, sizeof(struct hlist_head), in table_instance_alloc() 161 if (!ti->buckets) { in table_instance_alloc() 167 INIT_HLIST_HEAD(&ti->buckets[i]); in table_instance_alloc() 485 struct hlist_head *head = &ti->buckets[i]; in table_instance_flow_flush() 538 head = &ti->buckets[*bucket]; in ovs_flow_tbl_dump_next() 557 return &ti->buckets[hash & (ti->n_buckets - 1)]; in find_bucket() 590 struct hlist_head *head = &old->buckets[i]; in flow_table_copy_flows()
|
/kernel/linux/linux-5.10/drivers/message/fusion/ |
H A D | mptlan.c | 92 atomic_t buckets_out; /* number of unused buckets on IOC */ 106 int max_buckets_out; /* Max buckets to send to IOC */ 462 any buckets it still has. */ 506 dlprintk((KERN_INFO MYNAM ":lan_close: Posted %d buckets " in mpt_lan_close() 834 dioprintk((MYNAM "/receive_skb: %d buckets remaining\n", in mpt_lan_receive_skb() 840 dioprintk((KERN_INFO MYNAM "/receive_post_reply: %d buckets " in mpt_lan_receive_skb() 918 "IOC returned %d buckets, freeing them...\n", count)); in mpt_lan_receive_post_free() 948 /* dlprintk((KERN_INFO MYNAM "/receive_post_reply: freed %d buckets\n", in mpt_lan_receive_post_free() 951 /**/ dlprintk((KERN_INFO MYNAM "@receive_post_reply: %d buckets " in mpt_lan_receive_post_free() 1008 // dioprintk((KERN_INFO MYNAM ": %s/%s: Multiple buckets returne in mpt_lan_receive_post_reply() 1149 u32 curr, buckets, count, max; mpt_lan_post_receive_buckets() local [all...] |
/kernel/linux/linux-6.6/drivers/message/fusion/ |
H A D | mptlan.c | 89 atomic_t buckets_out; /* number of unused buckets on IOC */ 103 int max_buckets_out; /* Max buckets to send to IOC */ 459 any buckets it still has. */ 503 dlprintk((KERN_INFO MYNAM ":lan_close: Posted %d buckets " in mpt_lan_close() 832 dioprintk((MYNAM "/receive_skb: %d buckets remaining\n", in mpt_lan_receive_skb() 838 dioprintk((KERN_INFO MYNAM "/receive_post_reply: %d buckets " in mpt_lan_receive_skb() 920 "IOC returned %d buckets, freeing them...\n", count)); in mpt_lan_receive_post_free() 950 /* dlprintk((KERN_INFO MYNAM "/receive_post_reply: freed %d buckets\n", in mpt_lan_receive_post_free() 953 /**/ dlprintk((KERN_INFO MYNAM "@receive_post_reply: %d buckets " in mpt_lan_receive_post_free() 1010 // dioprintk((KERN_INFO MYNAM ": %s/%s: Multiple buckets returne in mpt_lan_receive_post_reply() 1151 u32 curr, buckets, count, max; mpt_lan_post_receive_buckets() local [all...] |
/kernel/linux/linux-6.6/net/openvswitch/ |
H A D | flow_table.c | 148 kvfree(ti->buckets); in __table_instance_destroy() 160 ti->buckets = kvmalloc_array(new_size, sizeof(struct hlist_head), in table_instance_alloc() 162 if (!ti->buckets) { in table_instance_alloc() 168 INIT_HLIST_HEAD(&ti->buckets[i]); in table_instance_alloc() 486 struct hlist_head *head = &ti->buckets[i]; in table_instance_flow_flush() 539 head = &ti->buckets[*bucket]; in ovs_flow_tbl_dump_next() 558 return &ti->buckets[hash & (ti->n_buckets - 1)]; in find_bucket() 591 struct hlist_head *head = &old->buckets[i]; in flow_table_copy_flows()
|
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | hashtab.c | 91 struct bucket *buckets; member 99 u32 n_buckets; /* number of hash buckets */ 140 INIT_HLIST_NULLS_HEAD(&htab->buckets[i].head, i); in htab_init_buckets() 142 raw_spin_lock_init(&htab->buckets[i].raw_lock); in htab_init_buckets() 144 spin_lock_init(&htab->buckets[i].lock); in htab_init_buckets() 481 htab->buckets = bpf_map_area_alloc(htab->n_buckets * in htab_map_alloc() 484 if (!htab->buckets) in htab_map_alloc() 514 bpf_map_area_free(htab->buckets); in htab_map_alloc() 529 return &htab->buckets[hash & (htab->n_buckets - 1)]; in __select_bucket() 753 /* iterate over buckets */ in htab_map_get_next_key() [all...] |
/third_party/node/deps/v8/tools/sanitizers/ |
H A D | sancov_merger.py | 95 Splits the sancov files into several buckets, so that each bucket can be 98 we might need to avoid splitting buckets of executables with few files. 108 # Chop files into buckets. 109 buckets = [files[i:i+n] for i in range(0, len(files), n)] 115 for i, b in enumerate(buckets)])
|
/third_party/mesa3d/src/panfrost/lib/ |
H A D | pan_props.c | 314 for (unsigned i = 0; i < ARRAY_SIZE(dev->bo_cache.buckets); ++i) in panfrost_open_device() 315 list_inithead(&dev->bo_cache.buckets[i]); in panfrost_open_device()
|
/third_party/skia/third_party/externals/freetype/src/cache/ |
H A D | ftccache.h | 77 ( ( cache )->buckets + \ 148 FTC_Node* buckets; member
|
/third_party/mesa3d/src/gallium/auxiliary/pipebuffer/ |
H A D | pb_cache.h | 51 /* The cache is divided into buckets for minimizing cache misses. 54 struct list_head *buckets; member
|
/kernel/linux/linux-5.10/drivers/md/bcache/ |
H A D | bcache.h | 42 * To do this, we first divide the cache device up into buckets. A bucket is the 47 * it. The gens and priorities for all the buckets are stored contiguously and 48 * packed on disk (in a linked list of buckets - aside from the superblock, all 49 * of bcache's metadata is stored in buckets). 56 * The generation is used for invalidating buckets. Each pointer also has an 8 62 * Bcache is entirely COW - we never write twice to a bucket, even buckets that 113 * (If buckets are really big we'll only use part of the bucket for a btree node 143 * Thus, the primary purpose of garbage collection is to find buckets to reuse. 145 * allocation can reuse buckets sooner when they've been mostly overwritten. 424 * When allocating new buckets, prio_writ 448 struct bucket *buckets; global() member [all...] |
/kernel/linux/linux-6.6/drivers/md/bcache/ |
H A D | bcache.h | 42 * To do this, we first divide the cache device up into buckets. A bucket is the 47 * it. The gens and priorities for all the buckets are stored contiguously and 48 * packed on disk (in a linked list of buckets - aside from the superblock, all 49 * of bcache's metadata is stored in buckets). 56 * The generation is used for invalidating buckets. Each pointer also has an 8 62 * Bcache is entirely COW - we never write twice to a bucket, even buckets that 113 * (If buckets are really big we'll only use part of the bucket for a btree node 143 * Thus, the primary purpose of garbage collection is to find buckets to reuse. 145 * allocation can reuse buckets sooner when they've been mostly overwritten. 432 * When allocating new buckets, prio_writ 456 struct bucket *buckets; global() member [all...] |
/third_party/node/deps/v8/src/logging/ |
H A D | counters.h | 70 // sample values. buckets is the maximum number of buckets 72 void* CreateHistogram(const char* name, int min, int max, size_t buckets) { in CreateHistogram() argument 74 return create_histogram_function_(name, min, max, buckets); in CreateHistogram() 630 void* CreateHistogram(const char* name, int min, int max, size_t buckets) { in CreateHistogram() argument 631 return stats_table_.CreateHistogram(name, min, max, buckets); in CreateHistogram()
|