/kernel/linux/linux-5.10/kernel/irq/ |
H A D | affinity.c | 48 masks = kcalloc(nr_node_ids, sizeof(cpumask_var_t), GFP_KERNEL); in alloc_node_to_cpumask() 52 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask() 70 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask() 137 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors() 156 sort(node_vectors, nr_node_ids, sizeof(node_vectors[0]), in alloc_nodes_vectors() 227 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors() 281 node_vectors = kcalloc(nr_node_ids, in __irq_build_affinity_masks() 291 for (i = 0; i < nr_node_ids; i++) { in __irq_build_affinity_masks()
|
/kernel/linux/linux-6.6/lib/ |
H A D | group_cpus.c | 50 masks = kcalloc(nr_node_ids, sizeof(cpumask_var_t), GFP_KERNEL); in alloc_node_to_cpumask() 54 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask() 72 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask() 139 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups() 158 sort(node_groups, nr_node_ids, sizeof(node_groups[0]), in alloc_nodes_groups() 229 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups() 280 node_groups = kcalloc(nr_node_ids, in __group_cpus_evenly() 289 for (i = 0; i < nr_node_ids; i++) { in __group_cpus_evenly()
|
/kernel/linux/linux-5.10/arch/arm64/mm/ |
H A D | numa.c | 53 if (WARN_ON(node < 0 || node >= nr_node_ids)) in cpumask_of_node() 105 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 106 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 110 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map() 116 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 282 size = nr_node_ids * nr_node_ids * sizeof(numa_distance[0]); in numa_alloc_distance() 291 numa_distance_cnt = nr_node_ids; in numa_alloc_distance()
|
/kernel/linux/linux-6.6/drivers/base/ |
H A D | arch_numa.c | 52 if (WARN_ON(node < 0 || node >= nr_node_ids)) in cpumask_of_node() 104 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 105 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 109 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map() 115 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 278 size = nr_node_ids * nr_node_ids * sizeof(numa_distance[0]); in numa_alloc_distance() 283 numa_distance_cnt = nr_node_ids; in numa_alloc_distance()
|
/kernel/linux/linux-6.6/scripts/gdb/linux/ |
H A D | slab.py | 183 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 185 nr_node_ids = 1 198 for i in range(0, nr_node_ids): 279 nr_node_ids = None 295 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 297 nr_node_ids = 1 303 for i in range(0, nr_node_ids):
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | numa.c | 113 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 114 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 118 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map() 122 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 896 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node() 898 "cpumask_of_node(%d): (unsigned)node >= nr_node_ids(%u)\n", in cpumask_of_node() 899 node, nr_node_ids); in cpumask_of_node()
|
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | numa.c | 114 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 115 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 119 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map() 123 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 903 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node() 905 "cpumask_of_node(%d): (unsigned)node >= nr_node_ids(%u)\n", in cpumask_of_node() 906 node, nr_node_ids); in cpumask_of_node()
|
/kernel/linux/linux-6.6/mm/ |
H A D | shrinker_debug.c | 54 count_per_node = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); in shrinker_debugfs_count_show() 130 if (nid < 0 || nid >= nr_node_ids) in shrinker_debugfs_scan_write()
|
H A D | list_lru.c | 343 mlru = kmalloc(struct_size(mlru, node, nr_node_ids), gfp); in memcg_init_list_lru_one() 571 lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL); in __list_lru_init()
|
/kernel/linux/linux-5.10/arch/loongarch/loongson64/ |
H A D | numa.c | 87 if (nr_node_ids >= 8) in setup_per_cpu_areas() 403 loongson_sysconf.nr_nodes = nr_node_ids; in init_numa_memory()
|
/kernel/linux/linux-6.6/arch/loongarch/kernel/ |
H A D | numa.c | 80 if (nr_node_ids >= 8) in setup_per_cpu_areas() 409 loongson_sysconf.nr_nodes = nr_node_ids; in init_numa_memory()
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | setup_percpu.c | 122 pr_info("NR_CPUS:%d nr_cpumask_bits:%d nr_cpu_ids:%u nr_node_ids:%u\n", in setup_per_cpu_areas() 123 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | topology.c | 1553 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn() 1555 for (j = 0; j < nr_node_ids; j++) in sched_numa_warn() 1647 for (i = 0; i < nr_node_ids; i++) { in sched_init_numa() 1648 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa() 1706 kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa() 1710 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa() 1781 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set() 1793 for (j = 0; j < nr_node_ids; j++) in sched_domains_numa_masks_clear()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | nodemask.h | 447 extern unsigned int nr_node_ids; 488 #define nr_node_ids 1U macro
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | nodemask.h | 455 extern unsigned int nr_node_ids; 497 #define nr_node_ids 1U macro
|
/kernel/linux/linux-5.10/arch/powerpc/mm/ |
H A D | numa.c | 81 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 82 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 90 dbg("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 189 if (nid == 0xffff || nid >= nr_node_ids) in __associativity_to_nid() 194 * Returns nid in the range [0..nr_node_ids], or -1 if no useful NUMA
|
/kernel/linux/linux-6.6/arch/powerpc/mm/ |
H A D | numa.c | 79 /* setup nr_node_ids if not done yet */ in setup_node_to_cpumask_map() 80 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 88 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 185 if (nid == 0xffff || nid >= nr_node_ids) in __associativity_to_nid() 190 * Returns nid in the range [0..nr_node_ids], or -1 if no useful NUMA
|
/kernel/linux/linux-5.10/arch/loongarch/kernel/ |
H A D | paravirt.c | 74 if (nr_node_ids < 8) in slow_virt_to_phys()
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | setup_percpu.c | 174 pr_info("NR_CPUS:%d nr_cpumask_bits:%d nr_cpu_ids:%u nr_node_ids:%u\n", in setup_per_cpu_areas() 175 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
|
/kernel/linux/linux-5.10/drivers/hv/ |
H A D | hv.c | 89 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | topology.c | 1733 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn() 1735 for (j = 0; j < nr_node_ids; j++) { in sched_numa_warn() 1912 masks[i] = kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa() 2043 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set() 2059 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_clear() 2184 if (node >= nr_node_ids || hops >= sched_domains_numa_levels) in sched_numa_hop_mask()
|
/kernel/linux/linux-6.6/drivers/hv/ |
H A D | hv.c | 112 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()
|
/kernel/linux/linux-5.10/mm/ |
H A D | ksm.c | 2967 buf = kcalloc(nr_node_ids + nr_node_ids, sizeof(*buf), in merge_across_nodes_store() 2974 root_unstable_tree = buf + nr_node_ids; in merge_across_nodes_store() 2981 ksm_nr_node_ids = knob ? 1 : nr_node_ids; in merge_across_nodes_store()
|
/kernel/linux/linux-6.6/arch/powerpc/platforms/pseries/ |
H A D | hotplug-cpu.c | 241 if (rc && nr_node_ids > 1) { in pseries_add_processor()
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
H A D | siw_main.c | 138 int i, num_nodes = nr_node_ids; in siw_init_cpulist()
|