Searched refs:sched_group_span (Results 1 - 8 of 8) sorted by relevance
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | topology.c | 50 if (group && !cpumask_test_cpu(cpu, sched_group_span(group))) { in sched_domain_debug_one() 77 if (!cpumask_weight(sched_group_span(group))) { in sched_domain_debug_one() 84 cpumask_intersects(groupmask, sched_group_span(group))) { in sched_domain_debug_one() 90 cpumask_or(groupmask, groupmask, sched_group_span(group)); in sched_domain_debug_one() 94 cpumask_pr_args(sched_group_span(group))); in sched_domain_debug_one() 97 !cpumask_equal(group_balance_mask(group), sched_group_span(group))) { in sched_domain_debug_one() 107 sched_group_span(group))) { in sched_domain_debug_one() 862 const struct cpumask *sg_span = sched_group_span(sg); in build_balance_mask() 908 sg_span = sched_group_span(sg); in build_group_from_child_sched_domain() 927 cpu = cpumask_first_and(sched_group_span(s in init_overlap_sched_group() [all...] |
H A D | fair.c | 6237 return cpumask_first(sched_group_span(group)); in find_idlest_group_cpu() 6240 for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { in find_idlest_group_cpu() 8741 for_each_cpu(cpu, sched_group_span(sdg)) { in update_group_capacity() 8761 sched_group_span(group); in update_group_capacity() 8972 local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(group)); in update_sg_lb_stats() 8974 for_each_cpu_and(i, sched_group_span(group), env->cpus) { in update_sg_lb_stats() 9268 for_each_cpu(i, sched_group_span(group)) { in update_sg_wakeup_stats() 9393 if (!cpumask_intersects(sched_group_span(group), in find_idlest_group() 9396 if (!cpumask_intersects(sched_group_span(group), in find_idlest_group() 9402 sched_group_span(grou in find_idlest_group() [all...] |
H A D | rt.c | 1766 if (!cpumask_intersects(lowest_mask, sched_group_span(sg))) in find_cas_cpu() 1771 sched_group_span(sg))) { in find_cas_cpu() 1827 cpumask_and(&search_cpu, lowest_mask, sched_group_span(sg_target)); in find_cas_cpu()
|
H A D | sched.h | 1633 static inline struct cpumask *sched_group_span(struct sched_group *sg) in sched_group_span() function 1652 return cpumask_first(sched_group_span(group)); in group_first_cpu()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | topology.c | 51 if (group && !cpumask_test_cpu(cpu, sched_group_span(group))) { in sched_domain_debug_one() 78 if (cpumask_empty(sched_group_span(group))) { in sched_domain_debug_one() 85 cpumask_intersects(groupmask, sched_group_span(group))) { in sched_domain_debug_one() 91 cpumask_or(groupmask, groupmask, sched_group_span(group)); in sched_domain_debug_one() 95 cpumask_pr_args(sched_group_span(group))); in sched_domain_debug_one() 98 !cpumask_equal(group_balance_mask(group), sched_group_span(group))) { in sched_domain_debug_one() 108 sched_group_span(group))) { in sched_domain_debug_one() 910 const struct cpumask *sg_span = sched_group_span(sg); in build_balance_mask() 956 sg_span = sched_group_span(sg); in build_group_from_child_sched_domain() 990 sg_span = sched_group_span(s in init_overlap_sched_group() [all...] |
H A D | fair.c | 7193 return cpumask_first(sched_group_span(group)); in find_idlest_group_cpu() 7196 for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { in find_idlest_group_cpu() 9817 for_each_cpu(cpu, sched_group_span(sdg)) { in update_group_capacity() 9837 sched_group_span(group); in update_group_capacity() 10151 for_each_cpu_and(i, sched_group_span(group), env->cpus) { in update_sg_lb_stats() 10487 for_each_cpu(i, sched_group_span(group)) { in update_sg_wakeup_stats() 10609 if (!cpumask_intersects(sched_group_span(group), in find_idlest_group() 10612 if (!cpumask_intersects(sched_group_span(group), in find_idlest_group() 10622 sched_group_span(group)); in find_idlest_group() 10720 idlest_cpu = cpumask_first(sched_group_span(idles in find_idlest_group() [all...] |
H A D | sched.h | 1352 static inline struct cpumask *sched_group_span(struct sched_group *sg); 1438 for_each_cpu_and(cpu, sched_group_span(group), p->cpus_ptr) { in sched_group_cookie_match() 2064 static inline struct cpumask *sched_group_span(struct sched_group *sg) in sched_group_span() function 2083 return cpumask_first(sched_group_span(group)); in group_first_cpu()
|
H A D | rt.c | 1940 if (!cpumask_intersects(lowest_mask, sched_group_span(sg))) in find_cas_cpu() 1945 sched_group_span(sg))) { in find_cas_cpu() 2001 cpumask_and(&search_cpu, lowest_mask, sched_group_span(sg_target)); in find_cas_cpu()
|
Completed in 53 milliseconds