Lines Matching refs:sched_domain

34 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level,  in sched_domain_debug_one()
130 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
170 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
188 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
604 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain()
620 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu()
623 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu()
629 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains()
644 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
648 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
649 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
650 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
656 struct sched_domain *sd; in update_top_cache_domain()
687 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
690 struct sched_domain *tmp; in cpu_attach_domain()
694 struct sched_domain *parent = tmp->parent; in cpu_attach_domain()
746 struct sched_domain * __percpu *sd;
878 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) in build_balance_mask()
882 struct sched_domain *sibling; in build_balance_mask()
915 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain()
938 static void init_overlap_sched_group(struct sched_domain *sd, in init_overlap_sched_group()
966 static struct sched_domain *
967 find_descended_sibling(struct sched_domain *sd, struct sched_domain *sibling) in find_descended_sibling()
992 build_overlap_sched_groups(struct sched_domain *sd, int cpu) in build_overlap_sched_groups()
998 struct sched_domain *sibling; in build_overlap_sched_groups()
1159 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1160 struct sched_domain *child = sd->child; in get_group()
1203 build_sched_groups(struct sched_domain *sd, int cpu) in build_sched_groups()
1248 static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) in init_sched_groups_capacity()
1407 static void set_domain_attribute(struct sched_domain *sd, in set_domain_attribute()
1454 d->sd = alloc_percpu(struct sched_domain *); in __visit_domain_allocation_hell()
1469 static void claim_allocations(int cpu, struct sched_domain *sd) in claim_allocations()
1521 static struct sched_domain *
1524 struct sched_domain *child, int cpu) in sd_init()
1527 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in sd_init()
1546 *sd = (struct sched_domain){ in sd_init()
2024 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
2041 struct sched_domain *sd; in __sdt_alloc()
2046 sd = kzalloc_node(sizeof(struct sched_domain) + cpumask_size(), in __sdt_alloc()
2094 struct sched_domain *sd; in __sdt_free()
2121 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain()
2123 struct sched_domain *child, int cpu) in build_sched_domain()
2125 struct sched_domain *sd = sd_init(tl, cpu_map, child, cpu); in build_sched_domain()
2195 struct sched_domain *sd; in build_sched_domains()