Lines Matching refs:cpumask
42 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, struct cpumask *groupmask)
293 static void perf_domain_debug(const struct cpumask *cpu_map, struct perf_domain *pd)
359 static bool build_perf_domains(const struct cpumask *cpu_map)
666 * the cpumask of the domain), this allows us to quickly tell if
897 static void build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask)
899 const struct cpumask *sg_span = sched_group_span(sg);
938 struct cpumask *sg_span;
958 struct cpumask *mask = sched_domains_tmpmask2;
960 struct cpumask *sg_span;
1009 const struct cpumask *span = sched_domain_span(sd);
1010 struct cpumask *covered = sched_domains_tmpmask;
1019 struct cpumask *sg_span;
1216 * covered by the given span, will set each group's ->cpumask correctly,
1225 const struct cpumask *span = sched_domain_span(sd);
1226 struct cpumask *covered;
1351 static void __sdt_free(const struct cpumask *cpu_map);
1352 static int __sdt_alloc(const struct cpumask *cpu_map);
1354 static void __free_domain_allocs(struct s_data *d, enum s_alloc what, const struct cpumask *cpu_map)
1373 static enum s_alloc __visit_domain_allocation_hell(struct s_data *d, const struct cpumask *cpu_map)
1425 static struct cpumask ***sched_domains_numa_masks;
1447 static struct sched_domain *sd_init(struct sched_domain_topology_level *tl, const struct cpumask *cpu_map,
1574 static const struct cpumask *sd_numa_mask(int cpu)
1756 struct cpumask *mask = kzalloc(cpumask_size(), GFP_KERNEL);
1849 * closest to @cpu from @cpumask.
1850 * cpumask: cpumask to find a cpu from
1855 int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
1870 static int __sdt_alloc(const struct cpumask *cpu_map)
1944 static void __sdt_free(const struct cpumask *cpu_map)
1984 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, const struct cpumask *cpu_map,
2013 static bool topology_span_sane(struct sched_domain_topology_level *tl, const struct cpumask *cpu_map, int cpu)
2051 static struct sched_domain_topology_level *asym_cpu_capacity_level(const struct cpumask *cpu_map)
2112 static int build_sched_domains(const struct cpumask *cpu_map, struct sched_domain_attr *attr)
2246 * cpumask) fails, then fallback to a single sched domain,
2247 * as determined by the single cpumask fallback_doms.
2292 int sched_init_domains(const struct cpumask *cpu_map)
2317 static void detach_destroy_domains(const struct cpumask *cpu_map)
2355 * not be load balanced. If the same cpumask appears both in the