/kernel/linux/linux-6.6/net/netfilter/ipvs/ |
H A D | ip_vs_lblcr.c | 56 * It is for garbage collection of stale IPVS lblcr entries, 68 * entries that haven't been touched for a day. 274 atomic_t entries; /* number of entries */ member 275 int max_size; /* maximum size of entries */ 276 struct timer_list periodic_timer; /* collect stale entries */ 336 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash() 393 * Flush all the entries of the specified table. 439 atomic_dec(&tbl->entries); in ip_vs_lblcr_full_check() 449 * It is used to collect stale entries whe [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | vf_palettegen.c | 50 struct color_ref *entries; member 301 refs[k++] = &node->entries[i]; in load_color_refs() 455 e = &node->entries[i]; in color_inc() 462 e = av_dynarray2_add((void**)&node->entries, &node->nb_entries, in color_inc() 463 sizeof(*node->entries), NULL); in color_inc() 540 av_freep(&s->histogram[i].entries); in filter_frame() 598 av_freep(&s->histogram[i].entries); in uninit()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | alloc.c | 203 struct list_head entries; member 231 INIT_LIST_HEAD(&zones->entries); in mlx4_zone_allocator_create() 291 if (!list_is_last(&entry->list, &zone_alloc->entries)) { in __mlx4_zone_remove_one_entry() 325 list_for_each_entry_safe(zone, tmp, &zone_alloc->entries, list) { in mlx4_zone_allocator_destroy() 361 list_for_each_entry_continue_reverse(it, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 375 list_for_each_entry_from(it, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 400 list_for_each_entry_from(curr_node, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 430 list_for_each_entry(zone, &zones->entries, list) { in __mlx4_find_zone_by_uid() 490 list_for_each_entry(zone, &zones->entries, list) { in __mlx4_find_zone_by_uid_unique()
|
/kernel/linux/linux-5.10/drivers/iommu/ |
H A D | iova.c | 488 if (fq->entries[idx].counter >= counter) in fq_ring_free() 492 iovad->entry_dtor(fq->entries[idx].data); in fq_ring_free() 495 fq->entries[idx].iova_pfn, in fq_ring_free() 496 fq->entries[idx].pages); in fq_ring_free() 516 * entries. in fq_destroy_all_entries() 526 iovad->entry_dtor(fq->entries[idx].data); in fq_destroy_all_entries() 560 * First remove all entries from the flush queue that have already been in queue_iova() 573 fq->entries[idx].iova_pfn = pfn; in queue_iova() 574 fq->entries[idx].pages = pages; in queue_iova() 575 fq->entries[id in queue_iova() [all...] |
/kernel/linux/linux-6.6/fs/xfs/scrub/ |
H A D | agheader.c | 623 /* Number of AGFL entries that the AGF claims are in use. */ 626 /* Number of AGFL entries that we found. */ 629 /* Buffer to hold AGFL entries for extent checking. */ 630 xfs_agblock_t *entries; member 664 sai->entries[sai->nr_entries++] = agbno; in xchk_agfl_block() 745 /* Allocate buffer to ensure uniqueness of AGFL entries. */ in xchk_agfl() 752 sai.entries = kvcalloc(sai.agflcount, sizeof(xfs_agblock_t), in xchk_agfl() 754 if (!sai.entries) { in xchk_agfl() 774 /* Sort entries, check for duplicates. */ in xchk_agfl() 775 sort(sai.entries, sa in xchk_agfl() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
H A D | alloc.c | 203 struct list_head entries; member 231 INIT_LIST_HEAD(&zones->entries); in mlx4_zone_allocator_create() 291 if (!list_is_last(&entry->list, &zone_alloc->entries)) { in __mlx4_zone_remove_one_entry() 325 list_for_each_entry_safe(zone, tmp, &zone_alloc->entries, list) { in mlx4_zone_allocator_destroy() 361 list_for_each_entry_continue_reverse(it, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 375 list_for_each_entry_from(it, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 400 list_for_each_entry_from(curr_node, &zone_alloc->entries, list) { in __mlx4_alloc_from_zone() 430 list_for_each_entry(zone, &zones->entries, list) { in __mlx4_find_zone_by_uid() 490 list_for_each_entry(zone, &zones->entries, list) { in __mlx4_find_zone_by_uid_unique()
|
/kernel/linux/linux-6.6/tools/perf/ui/browsers/ |
H A D | annotate.c | 26 struct rb_root entries; member 184 * to ab->offset[] entries that are set to NULL, so to make this code in annotate_browser__draw_current_jump() 252 struct rb_root *root = &browser->entries; in disasm_rb_tree__insert() 315 browser->entries = RB_ROOT; in annotate_browser__calc_percent() 348 browser->curr_hot = rb_last(&browser->entries); in annotate_browser__calc_percent() 358 list_for_each_entry_continue(it, browser->b.entries, node) { in annotate_browser__find_next_asm_line() 365 list_for_each_entry_continue_reverse(it, browser->b.entries, node) { in annotate_browser__find_next_asm_line() 767 nd = rb_last(&browser->entries); in annotate_browser__run() 775 nd = rb_first(&browser->entries); in annotate_browser__run() 996 browser.b.entries in symbol__tui_annotate() [all...] |
/third_party/skia/third_party/externals/spirv-tools/utils/ |
H A D | generate_grammar_tables.py | 478 entries = sorted(enum.get('enumerants', []), key=functor) 484 # across all those grammar entries. Preserve order. 486 for e in entries: 489 for e in entries: 498 entries = [' {}'.format(generate_enum_operand_kind_entry(e, extension_map)) 499 for e in entries] 502 '{entries}', '}};'] 503 entries = '\n'.join(template).format( 505 entries=',\n'.join(entries)) [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/utils/ |
H A D | generate_grammar_tables.py | 478 entries = sorted(enum.get('enumerants', []), key=functor) 484 # across all those grammar entries. Preserve order. 486 for e in entries: 489 for e in entries: 498 entries = [' {}'.format(generate_enum_operand_kind_entry(e, extension_map)) 499 for e in entries] 502 '{entries}', '}};'] 503 entries = '\n'.join(template).format( 505 entries=',\n'.join(entries)) [all...] |
/third_party/spirv-tools/utils/ |
H A D | generate_grammar_tables.py | 493 entries = sorted(enum.get('enumerants', []), key=functor) 499 # across all those grammar entries. Preserve order. 501 for e in entries: 504 for e in entries: 513 entries = [' {}'.format(generate_enum_operand_kind_entry(e, extension_map)) 514 for e in entries] 517 '{entries}', '}};'] 518 entries = '\n'.join(template).format( 520 entries=',\n'.join(entries)) [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
H A D | qib_mad.c | 603 /* 64 blocks of 32 16-bit P_Key entries */ in subn_get_pkeytable() 2135 struct ib_cc_congestion_entry_shadow *entries; in cc_get_congestion_setting() local 2139 entries = ppd->congestion_entries_shadow->entries; in cc_get_congestion_setting() 2145 p->entries[i].ccti_increase = entries[i].ccti_increase; in cc_get_congestion_setting() 2146 p->entries[i].ccti_timer = cpu_to_be16(entries[i].ccti_timer); in cc_get_congestion_setting() 2147 p->entries[i].trigger_threshold = entries[ in cc_get_congestion_setting() 2166 struct ib_cc_table_entry_shadow *entries; cc_get_congestion_control_table() local 2244 struct ib_cc_table_entry_shadow *entries; cc_set_congestion_control_table() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/qib/ |
H A D | qib_mad.c | 603 /* 64 blocks of 32 16-bit P_Key entries */ in subn_get_pkeytable() 2135 struct ib_cc_congestion_entry_shadow *entries; in cc_get_congestion_setting() local 2139 entries = ppd->congestion_entries_shadow->entries; in cc_get_congestion_setting() 2145 p->entries[i].ccti_increase = entries[i].ccti_increase; in cc_get_congestion_setting() 2146 p->entries[i].ccti_timer = cpu_to_be16(entries[i].ccti_timer); in cc_get_congestion_setting() 2147 p->entries[i].trigger_threshold = entries[ in cc_get_congestion_setting() 2166 struct ib_cc_table_entry_shadow *entries; cc_get_congestion_control_table() local 2244 struct ib_cc_table_entry_shadow *entries; cc_set_congestion_control_table() local [all...] |
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | cpuid.h | 18 struct kvm_cpuid_entry2 __user *entries, 22 struct kvm_cpuid_entry __user *entries); 25 struct kvm_cpuid_entry2 __user *entries); 28 struct kvm_cpuid_entry2 __user *entries);
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | msix.c | 63 struct hfi1_msix_entry *entries; in msix_initialize() local 84 entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), in msix_initialize() 86 if (!entries) { in msix_initialize() 91 dd->msix_info.msix_entries = entries; in msix_initialize()
|
/kernel/linux/linux-5.10/net/xdp/ |
H A D | xsk_queue.h | 94 * RESERVE entries PEEK in the ring for entries 96 * SUBMIT entries RELEASE entries 98 * The producer reserves one or more entries in the ring. It can then 99 * fill in these entries and finally submit them so that they can be 103 * any new entries. If so, the consumer can then read these entries 105 * so that the producer can use these slots to fill in new entries. 226 u32 entries in xskq_cons_has_entries() local [all...] |
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | unwind-libdw.c | 98 * Store all entries within entries array, 104 struct unwind_entry *e = &ui->entries[ui->idx++]; in entry() 241 ui = zalloc(sizeof(ui_buf) + sizeof(ui_buf.entries[0]) * max_stack); in unwind__get_entries() 277 err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0; in unwind__get_entries()
|
/kernel/linux/linux-5.10/mm/ |
H A D | page_ext.c | 77 int entries = ARRAY_SIZE(page_ext_ops); in invoke_need_callbacks() local 80 for (i = 0; i < entries; i++) { in invoke_need_callbacks() 94 int entries = ARRAY_SIZE(page_ext_ops); in invoke_init_callbacks() local 96 for (i = 0; i < entries; i++) { in invoke_init_callbacks()
|
/kernel/linux/linux-5.10/kernel/trace/ |
H A D | ring_buffer_benchmark.c | 234 unsigned long long entries; in ring_buffer_producer() local 299 entries = ring_buffer_entries(buffer); in ring_buffer_producer() 332 trace_printk("Entries: %lld\n", entries); in ring_buffer_producer() 333 trace_printk("Total: %lld\n", entries + overruns + read); in ring_buffer_producer()
|
/kernel/linux/linux-5.10/drivers/rpmsg/ |
H A D | qcom_glink_rpm.c | 44 struct rpm_toc_entry entries[]; member 212 dev_err(dev, "Invalid number of toc entries\n"); in glink_rpm_parse_toc() 217 id = le32_to_cpu(toc->entries[i].id); in glink_rpm_parse_toc() 218 offset = le32_to_cpu(toc->entries[i].offset); in glink_rpm_parse_toc() 219 size = le32_to_cpu(toc->entries[i].size); in glink_rpm_parse_toc()
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | cpuid.h | 21 struct kvm_cpuid_entry2 __user *entries, 25 struct kvm_cpuid_entry __user *entries); 28 struct kvm_cpuid_entry2 __user *entries); 31 struct kvm_cpuid_entry2 __user *entries);
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | msix.c | 20 struct hfi1_msix_entry *entries; in msix_initialize() local 41 entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), in msix_initialize() 43 if (!entries) { in msix_initialize() 48 dd->msix_info.msix_entries = entries; in msix_initialize()
|
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | unwind-libdw.c | 120 * Store all entries within entries array, 126 struct unwind_entry *e = &ui->entries[ui->idx++]; in entry() 279 ui = zalloc(sizeof(ui_buf) + sizeof(ui_buf.entries[0]) * max_stack); in unwind__get_entries() 315 err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0; in unwind__get_entries()
|
/kernel/linux/linux-6.6/kernel/trace/ |
H A D | ring_buffer_benchmark.c | 234 unsigned long long entries; in ring_buffer_producer() local 299 entries = ring_buffer_entries(buffer); in ring_buffer_producer() 332 trace_printk("Entries: %lld\n", entries); in ring_buffer_producer() 333 trace_printk("Total: %lld\n", entries + overruns + read); in ring_buffer_producer()
|
/kernel/linux/linux-6.6/mm/kasan/ |
H A D | common.c | 42 unsigned long entries[KASAN_STACK_DEPTH]; in kasan_save_stack() local 45 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in kasan_save_stack() 46 return __stack_depot_save(entries, nr_entries, flags, can_alloc); in kasan_save_stack()
|
/third_party/icu/icu4j/main/classes/core/src/com/ibm/icu/impl/ |
H A D | URLHandler.java | 228 Enumeration<JarEntry> entries = jarFile.entries(); in guide() 230 while (entries.hasMoreElements()) { in guide() 231 JarEntry entry = entries.nextElement(); in guide()
|