/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | affinity.c | 137 int possible, curr_cpu, i, ht; in init_real_cpu_mask() local 155 curr_cpu = cpumask_first(&node_affinity.real_cpu_mask); in init_real_cpu_mask() 157 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 163 cpumask_clear_cpu(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 164 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 290 int curr_cpu; in per_cpu_affinity_get() local 312 for_each_cpu(curr_cpu, possible_cpumask) { in per_cpu_affinity_get() 313 cntr = *per_cpu_ptr(comp_vect_affinity, curr_cpu); in per_cpu_affinity_get() 330 int curr_cpu; per_cpu_affinity_put_max() local 541 int i, j, curr_cpu; global() variable 639 int curr_cpu, possible, i, ret; hfi1_dev_affinity_init() local 1010 int possible, curr_cpu, i; find_hw_thread_mask() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | affinity.c | 95 int possible, curr_cpu, i, ht; in init_real_cpu_mask() local 113 curr_cpu = cpumask_first(&node_affinity.real_cpu_mask); in init_real_cpu_mask() 115 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 121 cpumask_clear_cpu(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 122 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 246 int curr_cpu; in per_cpu_affinity_get() local 268 for_each_cpu(curr_cpu, possible_cpumask) { in per_cpu_affinity_get() 269 cntr = *per_cpu_ptr(comp_vect_affinity, curr_cpu); in per_cpu_affinity_get() 286 int curr_cpu; per_cpu_affinity_put_max() local 497 int i, j, curr_cpu; global() variable 595 int curr_cpu, possible, i, ret; hfi1_dev_affinity_init() local 966 int possible, curr_cpu, i; find_hw_thread_mask() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_trace.h | 77 __field(int, curr_cpu) 91 __entry->curr_cpu = get_cpu(); 98 TP_printk("i40e_napi_poll on dev %s q %s irq %d irq_mask %s curr_cpu %d " 102 __get_bitmask(irq_affinity), __entry->curr_cpu, __entry->budget,
|
/kernel/linux/linux-5.10/drivers/pci/controller/ |
H A D | pci-xgene-msi.c | 176 int curr_cpu; in xgene_msi_set_affinity() local 178 curr_cpu = hwirq_to_cpu(irqdata->hwirq); in xgene_msi_set_affinity() 179 if (curr_cpu == target_cpu) in xgene_msi_set_affinity()
|
H A D | pcie-iproc-msi.c | 211 int curr_cpu; in iproc_msi_irq_set_affinity() local 214 curr_cpu = hwirq_to_cpu(msi, data->hwirq); in iproc_msi_irq_set_affinity() 215 if (curr_cpu == target_cpu) in iproc_msi_irq_set_affinity()
|
/kernel/linux/linux-6.6/drivers/pci/controller/ |
H A D | pci-xgene-msi.c | 176 int curr_cpu; in xgene_msi_set_affinity() local 178 curr_cpu = hwirq_to_cpu(irqdata->hwirq); in xgene_msi_set_affinity() 179 if (curr_cpu == target_cpu) in xgene_msi_set_affinity()
|
H A D | pcie-iproc-msi.c | 211 int curr_cpu; in iproc_msi_irq_set_affinity() local 214 curr_cpu = hwirq_to_cpu(msi, data->hwirq); in iproc_msi_irq_set_affinity() 215 if (curr_cpu == target_cpu) in iproc_msi_irq_set_affinity()
|
/kernel/linux/linux-5.10/tools/perf/bench/ |
H A D | numa.c | 56 int curr_cpu; member 875 g->threads[task_nr].curr_cpu = cpu; in update_curr_cpu() 903 node = numa_node_of_cpu(td->curr_cpu); in count_process_nodes() 904 if (node < 0) /* curr_cpu was likely still -1 */ in count_process_nodes() 939 n = numa_node_of_cpu(td->curr_cpu); in count_node_processes() 1007 cpu = td->curr_cpu; in calc_convergence() 1239 this_cpu = g->threads[task_nr].curr_cpu; in worker_thread() 1340 td->curr_cpu = -1; in worker_process()
|
/kernel/linux/linux-6.6/drivers/vdpa/vdpa_user/ |
H A D | vduse_dev.c | 1095 int curr_cpu = vq->irq_effective_cpu; in vduse_vq_update_effective_cpu() local 1098 curr_cpu = cpumask_next(curr_cpu, &vq->irq_affinity); in vduse_vq_update_effective_cpu() 1099 if (cpu_online(curr_cpu)) in vduse_vq_update_effective_cpu() 1102 if (curr_cpu >= nr_cpu_ids) in vduse_vq_update_effective_cpu() 1103 curr_cpu = IRQ_UNBOUND; in vduse_vq_update_effective_cpu() 1106 vq->irq_effective_cpu = curr_cpu; in vduse_vq_update_effective_cpu()
|
/kernel/linux/linux-6.6/tools/perf/bench/ |
H A D | numa.c | 57 int curr_cpu; member 912 g->threads[task_nr].curr_cpu = cpu; in update_curr_cpu() 943 node = numa_node_of_cpu(td->curr_cpu); in count_process_nodes() 944 if (node < 0) /* curr_cpu was likely still -1 */ { in count_process_nodes() 982 n = numa_node_of_cpu(td->curr_cpu); in count_node_processes() 1052 cpu = td->curr_cpu; in calc_convergence() 1288 this_cpu = g->threads[task_nr].curr_cpu; in worker_thread() 1390 td->curr_cpu = -1; in worker_process()
|
/kernel/linux/linux-5.10/kernel/ |
H A D | relay.c | 660 unsigned int i, curr_cpu; in relay_late_setup_files() local 694 curr_cpu = get_cpu(); in relay_late_setup_files() 714 if (curr_cpu == i) { in relay_late_setup_files()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | relay.c | 572 unsigned int i, curr_cpu; in relay_late_setup_files() local 606 curr_cpu = get_cpu(); in relay_late_setup_files() 626 if (curr_cpu == i) { in relay_late_setup_files()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | core.c | 2185 int target_cpu, int curr_cpu) in migrate_swap() 2192 .src_cpu = curr_cpu, in migrate_swap() 6870 int curr_cpu = task_cpu(p); in migrate_task_to() local 6872 if (curr_cpu == target_cpu) in migrate_task_to() 6880 trace_sched_move_numa(p, curr_cpu, target_cpu); in migrate_task_to() 6881 return stop_one_cpu(curr_cpu, migration_cpu_stop, &arg); in migrate_task_to() 2184 migrate_swap(struct task_struct *cur, struct task_struct *p, int target_cpu, int curr_cpu) migrate_swap() argument
|
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | stat-display.c | 1484 struct perf_cpu curr_cpu = config->aggr_map->map[aggr_idx].cpu; in print_percore() local 1485 struct aggr_cpu_id core_id = aggr_cpu_id__core(curr_cpu, NULL); in print_percore()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | core.c | 3553 int target_cpu, int curr_cpu) in migrate_swap() 3560 .src_cpu = curr_cpu, in migrate_swap() 9591 int curr_cpu = task_cpu(p); in migrate_task_to() local 9593 if (curr_cpu == target_cpu) in migrate_task_to() 9601 trace_sched_move_numa(p, curr_cpu, target_cpu); in migrate_task_to() 9602 return stop_one_cpu(curr_cpu, migration_cpu_stop, &arg); in migrate_task_to() 3552 migrate_swap(struct task_struct *cur, struct task_struct *p, int target_cpu, int curr_cpu) migrate_swap() argument
|