| /kernel/linux/linux-5.10/include/linux/ |
| H A D | rcu_node_tree.h | 24 * Define shape of hierarchy based on NR_CPUS, CONFIG_RCU_FANOUT, and 52 #if NR_CPUS <= RCU_FANOUT_1 59 #elif NR_CPUS <= RCU_FANOUT_2 62 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 67 #elif NR_CPUS <= RCU_FANOUT_3 70 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2) 71 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 76 #elif NR_CPUS <= RCU_FANOUT_4 79 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_3) 80 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_ [all...] |
| H A D | cpumask.h | 8 * only nr_cpu_ids (<= NR_CPUS) bits are valid. 17 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; 36 #if NR_CPUS == 1 43 /* Assuming NR_CPUS is huge, a runtime limit is more efficient. Also, 47 #define nr_cpumask_bits ((unsigned int)NR_CPUS) 70 * all NR_CPUS bits set, otherwise it is just the set of CPUs that 81 * 1) UP arch's (NR_CPUS == 1, CONFIG_SMP not defined) hardcode 107 #if NR_CPUS > 1 138 #if defined(CONFIG_CPU_ISOLATION_OPT) && NR_CPUS > 1 148 #else /* !CONFIG_CPU_ISOLATION_OPT || NR_CPUS 847 bitmap_zero(cpumask_bits(&__cpu_possible_mask), NR_CPUS); reset_cpu_possible_mask() local [all...] |
| /kernel/linux/linux-6.6/include/linux/ |
| H A D | rcu_node_tree.h | 26 * Define shape of hierarchy based on NR_CPUS, CONFIG_RCU_FANOUT, and 54 #if NR_CPUS <= RCU_FANOUT_1 61 #elif NR_CPUS <= RCU_FANOUT_2 64 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 69 #elif NR_CPUS <= RCU_FANOUT_3 72 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2) 73 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 78 #elif NR_CPUS <= RCU_FANOUT_4 81 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_3) 82 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_ [all...] |
| H A D | cpumask.h | 8 * only nr_cpu_ids (<= NR_CPUS) bits are valid. 19 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; 38 #if (NR_CPUS == 1) || defined(CONFIG_FORCE_NR_CPUS) 39 #define nr_cpu_ids ((unsigned int)NR_CPUS) 46 #if (NR_CPUS == 1) || defined(CONFIG_FORCE_NR_CPUS) in set_nr_cpu_ids() 59 * the size is constant. So if NR_CPUS fits in one single word, we are 77 #if NR_CPUS <= BITS_PER_LONG 78 #define small_cpumask_bits ((unsigned int)NR_CPUS) 79 #define large_cpumask_bits ((unsigned int)NR_CPUS) 80 #elif NR_CPUS < 1021 bitmap_zero(cpumask_bits(&__cpu_possible_mask), NR_CPUS); reset_cpu_possible_mask() local [all...] |
| /kernel/linux/linux-5.10/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
| H A D | preempt.c | 9 /* Support NR_CPUS of at most 64 */ 23 * Simulate disabling preemption by locking a particular cpu. NR_CPUS 26 struct lock_impl cpu_preemption_locks[NR_CPUS] = { 28 #if (NR_CPUS - 1) & 1 31 #if (NR_CPUS - 1) & 2 34 #if (NR_CPUS - 1) & 4 37 #if (NR_CPUS - 1) & 8 40 #if (NR_CPUS - 1) & 16 43 #if (NR_CPUS - 1) & 32 67 assume(thread_cpu_id < NR_CPUS); in preempt_disable() [all...] |
| /kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
| H A D | irq.h | 41 extern void *critirq_ctx[NR_CPUS]; 42 extern void *dbgirq_ctx[NR_CPUS]; 43 extern void *mcheckirq_ctx[NR_CPUS]; 49 extern void *hardirq_ctx[NR_CPUS]; 50 extern void *softirq_ctx[NR_CPUS];
|
| /kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
| H A D | irq.h | 47 extern void *critirq_ctx[NR_CPUS]; 48 extern void *dbgirq_ctx[NR_CPUS]; 49 extern void *mcheckirq_ctx[NR_CPUS]; 55 extern void *hardirq_ctx[NR_CPUS]; 56 extern void *softirq_ctx[NR_CPUS];
|
| /kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_hashmap_lookup.c | 17 #define NR_CPUS 256 macro 18 #define CPU_MASK (NR_CPUS-1) 23 u32 __attribute__((__aligned__(8))) key[NR_CPUS]; 26 u64 __attribute__((__aligned__(256))) percpu_times_index[NR_CPUS]; 27 u64 __attribute__((__aligned__(256))) percpu_times[NR_CPUS][NR_SLOTS];
|
| /kernel/linux/linux-5.10/arch/ia64/kernel/ |
| H A D | err_inject.c | 43 static u64 call_start[NR_CPUS]; 44 static u64 phys_addr[NR_CPUS]; 45 static u64 err_type_info[NR_CPUS]; 46 static u64 err_struct_info[NR_CPUS]; 51 } __attribute__((__aligned__(16))) err_data_buffer[NR_CPUS]; 52 static s64 status[NR_CPUS]; 53 static u64 capabilities[NR_CPUS]; 54 static u64 resources[NR_CPUS];
|
| /kernel/linux/linux-6.6/arch/ia64/kernel/ |
| H A D | err_inject.c | 43 static u64 call_start[NR_CPUS]; 44 static u64 phys_addr[NR_CPUS]; 45 static u64 err_type_info[NR_CPUS]; 46 static u64 err_struct_info[NR_CPUS]; 51 } __attribute__((__aligned__(16))) err_data_buffer[NR_CPUS]; 52 static s64 status[NR_CPUS]; 53 static u64 capabilities[NR_CPUS]; 54 static u64 resources[NR_CPUS];
|
| /kernel/linux/linux-6.6/arch/arm64/kvm/hyp/nvhe/ |
| H A D | hyp-smp.c | 17 u64 __ro_after_init hyp_cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID }; 26 unsigned long __ro_after_init kvm_arm_hyp_percpu_base[NR_CPUS];
|
| /kernel/linux/linux-5.10/arch/riscv/kernel/ |
| H A D | cpu_ops.c | 16 const struct cpu_operations *cpu_ops[NR_CPUS] __ro_after_init; 18 void *__cpu_up_stack_pointer[NR_CPUS] __section(".data"); 19 void *__cpu_up_task_pointer[NR_CPUS] __section(".data");
|
| H A D | smp.c | 33 unsigned long __cpuid_to_hartid_map[NR_CPUS] = { 34 [0 ... NR_CPUS-1] = INVALID_HARTID 46 } ipi_data[NR_CPUS] __cacheline_aligned; 52 for (i = 0; i < NR_CPUS; i++) in riscv_hartid_to_cpuid()
|
| /kernel/linux/linux-5.10/arch/sparc/include/asm/ |
| H A D | topology_64.h | 53 extern cpumask_t cpu_core_map[NR_CPUS]; 54 extern cpumask_t cpu_core_sib_map[NR_CPUS]; 55 extern cpumask_t cpu_core_sib_cache_map[NR_CPUS];
|
| /kernel/linux/linux-6.6/arch/riscv/kernel/ |
| H A D | cpu_ops_spinwait.c | 17 void *__cpu_spinwait_stack_pointer[NR_CPUS] __section(".data"); 18 void *__cpu_spinwait_task_pointer[NR_CPUS] __section(".data"); 26 * The hartid must be less than NR_CPUS to avoid out-of-bound access in cpu_update_secondary_bootdata() 32 if (hartid == INVALID_HARTID || hartid >= (unsigned long) NR_CPUS) in cpu_update_secondary_bootdata()
|
| /kernel/linux/linux-6.6/arch/sparc/include/asm/ |
| H A D | topology_64.h | 53 extern cpumask_t cpu_core_map[NR_CPUS]; 54 extern cpumask_t cpu_core_sib_map[NR_CPUS]; 55 extern cpumask_t cpu_core_sib_cache_map[NR_CPUS];
|
| /kernel/linux/linux-5.10/arch/arm/mach-shmobile/ |
| H A D | headsmp.S | 118 cmp r1, #NR_CPUS 141 .space NR_CPUS * 4 144 .space NR_CPUS * 4 147 .space NR_CPUS * 4
|
| /kernel/linux/linux-6.6/arch/arm/mach-shmobile/ |
| H A D | headsmp.S | 118 cmp r1, #NR_CPUS 141 .space NR_CPUS * 4 144 .space NR_CPUS * 4 147 .space NR_CPUS * 4
|
| /kernel/linux/linux-5.10/arch/ia64/include/asm/native/ |
| H A D | irq.h | 14 #if (NR_VECTORS + 32 * NR_CPUS) < 1024 15 #define IA64_NATIVE_NR_IRQS (NR_VECTORS + 32 * NR_CPUS)
|
| /kernel/linux/linux-6.6/arch/ia64/include/asm/native/ |
| H A D | irq.h | 14 #if (NR_VECTORS + 32 * NR_CPUS) < 1024 15 #define IA64_NATIVE_NR_IRQS (NR_VECTORS + 32 * NR_CPUS)
|
| /kernel/linux/linux-5.10/arch/ia64/include/asm/ |
| H A D | smp.h | 50 int cpu_phys_id[NR_CPUS]; 55 extern cpumask_t cpu_core_map[NR_CPUS]; 74 for (i = 0; i < NR_CPUS; ++i) in cpu_logical_id()
|
| /kernel/linux/linux-6.6/arch/ia64/include/asm/ |
| H A D | smp.h | 50 int cpu_phys_id[NR_CPUS]; 55 extern cpumask_t cpu_core_map[NR_CPUS]; 74 for (i = 0; i < NR_CPUS; ++i) in cpu_logical_id()
|
| /kernel/linux/linux-5.10/arch/sh/kernel/ |
| H A D | irq.c | 65 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 66 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 68 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 69 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
|
| /kernel/linux/linux-6.6/arch/sh/kernel/ |
| H A D | irq.c | 66 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 67 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 69 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 70 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
|
| /kernel/linux/linux-5.10/arch/sparc/kernel/ |
| H A D | leon_smp.c | 54 extern volatile unsigned long cpu_callin_map[NR_CPUS]; 167 (unsigned int)nrcpu, (unsigned int)NR_CPUS, in leon_boot_cpus() 237 for (i = 0; i < NR_CPUS; i++) { in leon_smp_done() 368 unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */ 369 unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */ 380 register int high = NR_CPUS - 1; in leon_cross_call()
|