/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | cxgb4vf_main.c | 112 * list entries are 64-bit PCI DMA addresses. And since the state of 114 * always have at least one Egress Unit's worth of Free List entries 269 * Addresses are programmed to hash region, if tcam runs out of entries. 280 /* We ran out of TCAM entries. try programming hash region. */ in cxgb4vf_change_mac() 697 * mapping table has plenty of entries. 1961 * see partially corrupted log entries. But i9t's probably Good Enough(tm). 1987 /* skip over unused entries */ in mboxlog_show() 2131 * Return the number of "entries" in our "file". We group the multi-Queue 2146 int entries = sge_queue_entries(seq->private); in sge_queue_start() local 2148 return *pos < entries in sge_queue_start() 2157 int entries = sge_queue_entries(seq->private); sge_queue_next() local 2275 int entries = sge_qstats_entries(seq->private); sge_qstats_start() local 2286 int entries = sge_qstats_entries(seq->private); sge_qstats_next() local 2806 struct msix_entry entries[MSIX_ENTRIES]; enable_msix() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/display/dc/dcn30/ |
H A D | dcn30_resource.c | 2293 min_dram_speed_mts = dc->clk_mgr->bw_params->clk_table.entries[dc->clk_mgr->bw_params->clk_table.num_entries - 1].memclk_mhz * 16; in dcn30_calculate_wm_and_dlg_fp() 2482 if (bw_params->clk_table.entries[0].memclk_mhz) { in dcn30_update_bw_bounding_box() 2486 if (bw_params->clk_table.entries[i].dcfclk_mhz > max_dcfclk_mhz) in dcn30_update_bw_bounding_box() 2487 max_dcfclk_mhz = bw_params->clk_table.entries[i].dcfclk_mhz; in dcn30_update_bw_bounding_box() 2488 if (bw_params->clk_table.entries[i].dispclk_mhz > max_dispclk_mhz) in dcn30_update_bw_bounding_box() 2489 max_dispclk_mhz = bw_params->clk_table.entries[i].dispclk_mhz; in dcn30_update_bw_bounding_box() 2490 if (bw_params->clk_table.entries[i].dppclk_mhz > max_dppclk_mhz) in dcn30_update_bw_bounding_box() 2491 max_dppclk_mhz = bw_params->clk_table.entries[i].dppclk_mhz; in dcn30_update_bw_bounding_box() 2492 if (bw_params->clk_table.entries[i].phyclk_mhz > max_phyclk_mhz) in dcn30_update_bw_bounding_box() 2493 max_phyclk_mhz = bw_params->clk_table.entries[ in dcn30_update_bw_bounding_box() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_crat.c | 743 pr_debug("Added [%d] GPU cache entries\n", *num_of_entries); in kfd_fill_gpu_cache_info() 970 uint32_t entries = 0; in kfd_create_vcrat_image_cpu() local 1036 &entries, in kfd_create_vcrat_image_cpu() 1041 if (entries) { in kfd_create_vcrat_image_cpu() 1042 crat_table->length += (sub_type_hdr->length * entries); in kfd_create_vcrat_image_cpu() 1043 crat_table->total_entries += entries; in kfd_create_vcrat_image_cpu() 1046 sub_type_hdr->length * entries); in kfd_create_vcrat_image_cpu()
|
/kernel/linux/linux-6.6/arch/powerpc/kvm/ |
H A D | book3s.c | 1056 struct kvm_kernel_irq_routing_entry *entries, int gsi) in kvm_irq_map_gsi() 1058 entries->gsi = gsi; in kvm_irq_map_gsi() 1059 entries->type = KVM_IRQ_ROUTING_IRQCHIP; in kvm_irq_map_gsi() 1060 entries->set = kvmppc_book3s_set_irq; in kvm_irq_map_gsi() 1061 entries->irqchip.irqchip = 0; in kvm_irq_map_gsi() 1062 entries->irqchip.pin = gsi; in kvm_irq_map_gsi() 1055 kvm_irq_map_gsi(struct kvm *kvm, struct kvm_kernel_irq_routing_entry *entries, int gsi) kvm_irq_map_gsi() argument
|
/kernel/linux/linux-5.10/drivers/scsi/ |
H A D | qlogicpti.c | 821 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_map_queues() 916 /* Fill in first four sg entries: */ in load_cmd() 1388 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_sbus_probe() 1427 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_sbus_remove()
|
/kernel/linux/linux-6.6/drivers/usb/cdns3/ |
H A D | cdnsp-mem.c | 537 * 128 entries. in cdnsp_alloc_stream_ctx() 559 * stream context array entries must be a power of two. 1035 erst->entries = dma_alloc_coherent(pdev->dev, size, in cdnsp_alloc_erst() 1037 if (!erst->entries) in cdnsp_alloc_erst() 1044 entry = &erst->entries[val]; in cdnsp_alloc_erst() 1059 if (erst->entries) in cdnsp_free_erst() 1060 dma_free_coherent(dev, size, erst->entries, in cdnsp_free_erst() 1063 erst->entries = NULL; in cdnsp_free_erst() 1292 /* Set ERST count with the number of entries in the segment table. */ in cdnsp_mem_init()
|
/kernel/linux/linux-6.6/drivers/scsi/ |
H A D | qlogicpti.c | 821 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_map_queues() 917 /* Fill in first four sg entries: */ in load_cmd() 1387 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_sbus_probe() 1426 #define QSIZE(entries) (((entries) + 1) * QUEUE_ENTRY_LEN) in qpti_sbus_remove()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/ |
H A D | nouveau_svm.c | 49 u32 entries; member 731 /* Parse available fault buffer entries into a cache, and update in nouveau_svm_fault() 732 * the GET pointer so HW can reuse the entries. in nouveau_svm_fault() 746 if (++buffer->get == buffer->entries) in nouveau_svm_fault() 976 for (i = 0; buffer->fault[i] && i < buffer->entries; i++) in nouveau_svm_fault_buffer_dtor() 1004 buffer->entries = args.entries; in nouveau_svm_fault_buffer_ctor() 1014 buffer->fault = kvcalloc(buffer->entries, sizeof(*buffer->fault), GFP_KERNEL); in nouveau_svm_fault_buffer_ctor()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_span.c | 34 struct mlxsw_sp_span_entry entries[]; member 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 103 mlxsw_sp->span->entries[i].id = i; in mlxsw_sp_span_init() 858 if (!refcount_read(&mlxsw_sp->span->entries[i].ref_count)) { in mlxsw_sp_span_entry_create() 859 span_entry = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_create() 900 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_port() 921 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_id() 937 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_parms() 1036 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_respin_work()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_init.c | 571 __le32 entries = cpu_to_le32(directory->num_entries); in nx_get_table_desc() local 573 for (i = 0; i < entries; i++) { in nx_get_table_desc() 595 __le32 entries; in netxen_nic_validate_header() local 601 entries = cpu_to_le32(directory->num_entries); in netxen_nic_validate_header() 603 tab_size = cpu_to_le32(directory->findex) + (entries * entry_size); in netxen_nic_validate_header() 688 __le32 entries; in netxen_nic_validate_product_offs() local 697 entries = cpu_to_le32(ptab_descr->num_entries); in netxen_nic_validate_product_offs() 699 tab_size = cpu_to_le32(ptab_descr->findex) + (entries * entry_size); in netxen_nic_validate_product_offs() 705 for (i = 0; i < entries; i++) { in netxen_nic_validate_product_offs()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/enetc/ |
H A D | enetc_qos.c | 105 temp_entry = &admin_conf->entries[i]; in enetc_setup_taprio() 400 struct action_gate_entry entries[]; member 811 struct action_gate_entry *from = &sgi->entries[i]; in enetc_streamgate_hw_set() 1203 entries_size = struct_size(sgi, entries, entryg->gate.num_entries); in enetc_psfp_parse_clsflower() 1217 e = sgi->entries; in enetc_psfp_parse_clsflower() 1219 e[i].gate_state = entryg->gate.entries[i].gate_state; in enetc_psfp_parse_clsflower() 1220 e[i].interval = entryg->gate.entries[i].interval; in enetc_psfp_parse_clsflower() 1221 e[i].ipv = entryg->gate.entries[i].ipv; in enetc_psfp_parse_clsflower() 1222 e[i].maxoctets = entryg->gate.entries[i].maxoctets; in enetc_psfp_parse_clsflower()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/enetc/ |
H A D | enetc_qos.c | 101 temp_entry = &admin_conf->entries[i]; in enetc_setup_taprio() 446 struct action_gate_entry entries[]; member 814 struct action_gate_entry *from = &sgi->entries[i]; in enetc_streamgate_hw_set() 1245 entries_size = struct_size(sgi, entries, entryg->gate.num_entries); in enetc_psfp_parse_clsflower() 1259 e = sgi->entries; in enetc_psfp_parse_clsflower() 1261 e[i].gate_state = entryg->gate.entries[i].gate_state; in enetc_psfp_parse_clsflower() 1262 e[i].interval = entryg->gate.entries[i].interval; in enetc_psfp_parse_clsflower() 1263 e[i].ipv = entryg->gate.entries[i].ipv; in enetc_psfp_parse_clsflower() 1264 e[i].maxoctets = entryg->gate.entries[i].maxoctets; in enetc_psfp_parse_clsflower()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_span.c | 34 struct mlxsw_sp_span_entry entries[]; member 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 103 mlxsw_sp->span->entries[i].id = i; in mlxsw_sp_span_init() 860 if (!refcount_read(&mlxsw_sp->span->entries[i].ref_count)) { in mlxsw_sp_span_entry_create() 861 span_entry = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_create() 902 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_port() 923 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_id() 939 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_entry_find_by_parms() 1039 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span->entries[i]; in mlxsw_sp_span_respin_work()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_init.c | 569 __le32 entries = cpu_to_le32(directory->num_entries); in nx_get_table_desc() local 571 for (i = 0; i < entries; i++) { in nx_get_table_desc() 593 __le32 entries; in netxen_nic_validate_header() local 599 entries = cpu_to_le32(directory->num_entries); in netxen_nic_validate_header() 601 tab_size = cpu_to_le32(directory->findex) + (entries * entry_size); in netxen_nic_validate_header() 686 __le32 entries; in netxen_nic_validate_product_offs() local 695 entries = cpu_to_le32(ptab_descr->num_entries); in netxen_nic_validate_product_offs() 697 tab_size = cpu_to_le32(ptab_descr->findex) + (entries * entry_size); in netxen_nic_validate_product_offs() 703 for (i = 0; i < entries; i++) { in netxen_nic_validate_product_offs()
|
/kernel/linux/linux-5.10/mm/ |
H A D | memcontrol.c | 3780 * parent's id, and then move all entries from this cgroup's list_lrus in memcg_offline_kmem() 4264 for (; i >= 0 && unlikely(t->entries[i].threshold > usage); i--) in __mem_cgroup_threshold() 4265 eventfd_signal(t->entries[i].eventfd, 1); in __mem_cgroup_threshold() 4276 for (; i < t->size && unlikely(t->entries[i].threshold <= usage); i++) in __mem_cgroup_threshold() 4277 eventfd_signal(t->entries[i].eventfd, 1); in __mem_cgroup_threshold() 4362 new = kmalloc(struct_size(new, entries, size), GFP_KERNEL); in __mem_cgroup_usage_register_event() 4371 memcpy(new->entries, thresholds->primary->entries, in __mem_cgroup_usage_register_event() 4372 flex_array_size(new, entries, size - 1)); in __mem_cgroup_usage_register_event() 4375 new->entries[siz in __mem_cgroup_usage_register_event() 4429 int i, j, size, entries; __mem_cgroup_usage_unregister_event() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/display/ |
H A D | i9xx_wm.c | 550 int entries, wm_size; in intel_calculate_wm() local 558 entries = intel_wm_method1(pixel_rate, cpp, in intel_calculate_wm() 560 entries = DIV_ROUND_UP(entries, wm->cacheline_size) + in intel_calculate_wm() 562 DRM_DEBUG_KMS("FIFO entries required for mode: %d\n", entries); in intel_calculate_wm() 564 wm_size = fifo_size - entries; in intel_calculate_wm() 2013 /* Calc sr entries for one plane configs */ in i965_update_wm() 2026 int entries; in i965_update_wm() local 2028 entries in i965_update_wm() 2197 int entries; i9xx_update_wm() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | memcontrol.c | 4260 for (; i >= 0 && unlikely(t->entries[i].threshold > usage); i--) in __mem_cgroup_threshold() 4261 eventfd_signal(t->entries[i].eventfd, 1); in __mem_cgroup_threshold() 4272 for (; i < t->size && unlikely(t->entries[i].threshold <= usage); i++) in __mem_cgroup_threshold() 4273 eventfd_signal(t->entries[i].eventfd, 1); in __mem_cgroup_threshold() 4358 new = kmalloc(struct_size(new, entries, size), GFP_KERNEL); in __mem_cgroup_usage_register_event() 4367 memcpy(new->entries, thresholds->primary->entries, in __mem_cgroup_usage_register_event() 4368 flex_array_size(new, entries, size - 1)); in __mem_cgroup_usage_register_event() 4371 new->entries[size - 1].eventfd = eventfd; in __mem_cgroup_usage_register_event() 4372 new->entries[siz in __mem_cgroup_usage_register_event() 4425 int i, j, size, entries; __mem_cgroup_usage_unregister_event() local [all...] |
/kernel/linux/linux-6.6/drivers/net/dsa/sja1105/ |
H A D | sja1105_static_config.c | 1026 l2_fwd_params = tables[BLK_IDX_L2_FORWARDING_PARAMS].entries; in static_config_check_memory_size() 1032 vl_fwd_params = tables[BLK_IDX_VL_FORWARDING_PARAMS].entries; in static_config_check_memory_size() 1072 vl_lookup = tables[BLK_IDX_VL_LOOKUP].entries; in sja1105_static_config_check_valid() 1145 u8 *entry_ptr = table->entries; in sja1105_static_config_pack() 1178 /* Tables (headers and entries) */ in sja1105_static_config_get_length() 1910 kfree(config->tables[i].entries); in sja1105_static_config_free() 1919 u8 *entries = table->entries; in sja1105_table_delete_entry() local 1924 memmove(entries + i * entry_size, entries in sja1105_table_delete_entry() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
H A D | ni_dpm.c | 998 if (0xff01 == table->entries[i].v) { in ni_patch_single_dependency_table_based_on_leakage() 1001 table->entries[i].v = pi->max_vddc; in ni_patch_single_dependency_table_based_on_leakage() 1266 table->lowSMIO[i] |= cpu_to_be32(voltage_table->entries[i].smio_low); in ni_populate_smc_voltage_table() 1284 if (pi->max_vddc_in_table <= eg_pi->vddc_voltage_table.entries[i].value) { in ni_populate_smc_voltage_tables() 1308 if (value <= table->entries[i].value) { in ni_populate_voltage_value() 1310 voltage->value = cpu_to_be16(table->entries[i].value); in ni_populate_voltage_value() 1347 if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries && in ni_get_std_voltage_value() 1349 *std_voltage = rdev->pm.dpm.dyn_state.cac_leakage_table.entries[voltage->index].vddc; in ni_get_std_voltage_value() 3075 eg_pi->vddc_voltage_table.entries[j].value, in ni_init_driver_calculated_leakage_table() 3123 smc_leakage = leakage_table->entries[ in ni_init_simplified_leakage_table() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | ni_dpm.c | 997 if (0xff01 == table->entries[i].v) { in ni_patch_single_dependency_table_based_on_leakage() 1000 table->entries[i].v = pi->max_vddc; in ni_patch_single_dependency_table_based_on_leakage() 1265 table->lowSMIO[i] |= cpu_to_be32(voltage_table->entries[i].smio_low); in ni_populate_smc_voltage_table() 1283 if (pi->max_vddc_in_table <= eg_pi->vddc_voltage_table.entries[i].value) { in ni_populate_smc_voltage_tables() 1307 if (value <= table->entries[i].value) { in ni_populate_voltage_value() 1309 voltage->value = cpu_to_be16(table->entries[i].value); in ni_populate_voltage_value() 1346 if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries && in ni_get_std_voltage_value() 1348 *std_voltage = rdev->pm.dpm.dyn_state.cac_leakage_table.entries[voltage->index].vddc; in ni_get_std_voltage_value() 3076 eg_pi->vddc_voltage_table.entries[j].value, in ni_init_driver_calculated_leakage_table() 3124 smc_leakage = leakage_table->entries[ in ni_init_simplified_leakage_table() [all...] |
/kernel/linux/linux-5.10/drivers/firmware/efi/ |
H A D | efi.c | 244 LIST_HEAD(entries); in efivar_ssdt_load() 253 ret = efivar_init(efivar_ssdt_iter, &entries, true, &entries); in efivar_ssdt_load() 255 list_for_each_entry_safe(entry, aux, &entries, list) { in efivar_ssdt_load()
|
/kernel/linux/linux-5.10/arch/mips/alchemy/common/ |
H A D | dbdma.c | 390 u32 au1xxx_dbdma_ring_alloc(u32 chanid, int entries) in au1xxx_dbdma_ring_alloc() argument 414 desc_base = (u32)kmalloc_array(entries, sizeof(au1x_ddma_desc_t), in au1xxx_dbdma_ring_alloc() 425 i = entries * sizeof(au1x_ddma_desc_t); in au1xxx_dbdma_ring_alloc() 563 for (i = 0; i < entries; i++) { in au1xxx_dbdma_ring_alloc()
|
/kernel/linux/linux-6.6/arch/mips/alchemy/common/ |
H A D | dbdma.c | 391 u32 au1xxx_dbdma_ring_alloc(u32 chanid, int entries) in au1xxx_dbdma_ring_alloc() argument 415 desc_base = (u32)kmalloc_array(entries, sizeof(au1x_ddma_desc_t), in au1xxx_dbdma_ring_alloc() 426 i = entries * sizeof(au1x_ddma_desc_t); in au1xxx_dbdma_ring_alloc() 564 for (i = 0; i < entries; i++) { in au1xxx_dbdma_ring_alloc()
|
/kernel/linux/linux-5.10/drivers/net/wireless/intersil/p54/ |
H A D | fwio.c | 440 for (i = 0; i < priv->output_limit->entries; i++) { in p54_scan() 467 if (i == priv->output_limit->entries) in p54_scan() 471 for (i = 0; i < priv->curve_data->entries; i++) { in p54_scan() 495 if (i == priv->curve_data->entries) in p54_scan()
|
/kernel/linux/linux-5.10/drivers/tty/vt/ |
H A D | vt_ioctl.c | 500 return con_set_unimap(vc, tmp.entry_ct, tmp.entries); in do_unimap_ioctl() 505 tmp.entries); in do_unimap_ioctl() 1032 compat_caddr_t entries; member 1044 tmp_entries = compat_ptr(tmp.entries); in compat_unimap_ioctl()
|