Lines Matching refs:groups

44     struct sched_group *group = sd->groups;
57 printk(KERN_ERR "ERROR: domain->groups does not contain CPU%d\n", cpu);
74 printk(KERN_DEBUG "%*s groups:", level + 1, "");
106 if (group == sd->groups && sd->child && !cpumask_equal(sched_domain_span(sd->child), sched_group_span(group))) {
107 printk(KERN_ERR "ERROR: domain->groups does not match domain->child\n");
114 if (group != sd->groups) {
117 } while (group != sd->groups);
121 printk(KERN_ERR "ERROR: groups don't span domain->span\n");
181 /* Following flags need at least 2 groups */
182 if ((sd->flags & SD_DEGENERATE_GROUPS_MASK) && (sd->groups != sd->groups->next)) {
186 /* Following flags don't use groups */
206 /* Flags needing groups don't count if only 1 group in parent */
207 if (parent->groups == parent->groups->next) {
631 * overlapping domain, having private groups, only one. Iterate,
634 free_sched_groups(sd->groups, 1);
812 * We want to construct domains and groups to represent this. The way we go
819 * groups: {0-1,3},{1-3} {0-2},{0,2-3} {1-3},{0-1,3} {0,2-3},{0-2}
822 * groups: {0},{1},{3} {0},{1},{2} {1},{2},{3} {0},{2},{3}
832 * In order to minimize this overlap, we only build enough groups to cover the
833 * domain. For instance Node-0 NUMA-2 would only get groups: 0-1,3 and 1-3.
842 * more complicated. Consider for instance the groups of NODE-1 NUMA-2, both
843 * groups include the CPUs of Node-0, while those CPUs would not in fact ever
844 * end up at those groups (they would end up in group: 0-1,3).
880 * groups: {0-2},{1-3} {1-3},{0-2}
933 * statistics having the groups node local is of dubious benefit.
1058 * groups: {0-2},{1-3} {1-3},{0-2}
1061 * groups: {0-1},{1-3} {0-2},{2-3} {1-3},{0-1} {2-3},{0-2}
1064 * groups: {0},{1} {1},{2},{0} {2},{3},{1} {3},{2}
1068 * The NUMA-2 groups for nodes 0 and 3 are obviously buggered, as the
1094 sd->groups = first;
1159 * - The second is the sched_group_capacity; we want all identical groups
1215 * build_sched_groups will build a circular linked list of the groups
1255 sd->groups = first;
1261 * Initialize sched groups cpu_capacity.
1264 * distributing the load between different sched groups in a sched domain.
1265 * Typically cpu_capacity for all the groups in a sched domain will be same
1272 struct sched_group *sg = sd->groups;
1305 } while (sg != sd->groups);
1958 free_sched_groups(sd->groups, 0);
2164 /* Build the groups for the domains */
2289 * Set up scheduler domains and groups. For now this just excludes isolated