sched: Separate out build of SMT sched groups from __build_sched_domains
... to further strip down __build_sched_domains().
Signed-off-by: Andreas Herrmann <andreas.herrmann3@amd.com>
Cc: Peter Zijlstra <peterz@infradead.org>
LKML-Reference: <20090818105751.GH29515@alberich.amd.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
diff --git a/kernel/sched.c b/kernel/sched.c
index 2ecec06..43cfc6e5 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -8563,6 +8563,25 @@
return sd;
}
+static void build_sched_groups(struct s_data *d, enum sched_domain_level l,
+ const struct cpumask *cpu_map, int cpu)
+{
+ switch (l) {
+#ifdef CONFIG_SCHED_SMT
+ case SD_LV_SIBLING: /* set up CPU (sibling) groups */
+ cpumask_and(d->this_sibling_map, cpu_map,
+ topology_thread_cpumask(cpu));
+ if (cpu == cpumask_first(d->this_sibling_map))
+ init_sched_build_groups(d->this_sibling_map, cpu_map,
+ &cpu_to_cpu_group,
+ d->send_covered, d->tmpmask);
+ break;
+#endif
+ default:
+ break;
+ }
+}
+
/*
* Build sched domains for a given set of cpus and attach the sched domains
* to the individual cpus
@@ -8597,19 +8616,9 @@
sd = __build_smt_sched_domain(&d, cpu_map, attr, sd, i);
}
-#ifdef CONFIG_SCHED_SMT
- /* Set up CPU (sibling) groups */
for_each_cpu(i, cpu_map) {
- cpumask_and(d.this_sibling_map,
- topology_thread_cpumask(i), cpu_map);
- if (i != cpumask_first(d.this_sibling_map))
- continue;
-
- init_sched_build_groups(d.this_sibling_map, cpu_map,
- &cpu_to_cpu_group,
- d.send_covered, d.tmpmask);
+ build_sched_groups(&d, SD_LV_SIBLING, cpu_map, i);
}
-#endif
#ifdef CONFIG_SCHED_MC
/* Set up multi-core groups */