Lines Matching refs:sdd

901 	struct sd_data *sdd = sd->private;  in build_balance_mask()  local
908 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask()
962 struct sd_data *sdd = sd->private; in init_overlap_sched_group() local
969 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group()
1017 struct sd_data *sdd = sd->private; in build_overlap_sched_groups() local
1029 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups()
1177 static struct sched_group *get_group(int cpu, struct sd_data *sdd) in get_group() argument
1179 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1187 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
1188 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
1226 struct sd_data *sdd = sd->private; in build_sched_groups() local
1242 sg = get_group(i, sdd); in build_sched_groups()
1491 struct sd_data *sdd = sd->private; in claim_allocations() local
1493 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations()
1494 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations()
1496 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations()
1497 *per_cpu_ptr(sdd->sds, cpu) = NULL; in claim_allocations()
1499 if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref)) in claim_allocations()
1500 *per_cpu_ptr(sdd->sg, cpu) = NULL; in claim_allocations()
1502 if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref)) in claim_allocations()
1503 *per_cpu_ptr(sdd->sgc, cpu) = NULL; in claim_allocations()
1544 struct sd_data *sdd = &tl->data; in sd_init() local
1545 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in sd_init()
1641 sd->shared = *per_cpu_ptr(sdd->sds, sd_id); in sd_init()
1646 sd->private = sdd; in sd_init()
2078 struct sd_data *sdd = &tl->data; in __sdt_alloc() local
2080 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
2081 if (!sdd->sd) in __sdt_alloc()
2084 sdd->sds = alloc_percpu(struct sched_domain_shared *); in __sdt_alloc()
2085 if (!sdd->sds) in __sdt_alloc()
2088 sdd->sg = alloc_percpu(struct sched_group *); in __sdt_alloc()
2089 if (!sdd->sg) in __sdt_alloc()
2092 sdd->sgc = alloc_percpu(struct sched_group_capacity *); in __sdt_alloc()
2093 if (!sdd->sgc) in __sdt_alloc()
2107 *per_cpu_ptr(sdd->sd, j) = sd; in __sdt_alloc()
2114 *per_cpu_ptr(sdd->sds, j) = sds; in __sdt_alloc()
2123 *per_cpu_ptr(sdd->sg, j) = sg; in __sdt_alloc()
2134 *per_cpu_ptr(sdd->sgc, j) = sgc; in __sdt_alloc()
2147 struct sd_data *sdd = &tl->data; in __sdt_free() local
2152 if (sdd->sd) { in __sdt_free()
2153 sd = *per_cpu_ptr(sdd->sd, j); in __sdt_free()
2156 kfree(*per_cpu_ptr(sdd->sd, j)); in __sdt_free()
2159 if (sdd->sds) in __sdt_free()
2160 kfree(*per_cpu_ptr(sdd->sds, j)); in __sdt_free()
2161 if (sdd->sg) in __sdt_free()
2162 kfree(*per_cpu_ptr(sdd->sg, j)); in __sdt_free()
2163 if (sdd->sgc) in __sdt_free()
2164 kfree(*per_cpu_ptr(sdd->sgc, j)); in __sdt_free()
2166 free_percpu(sdd->sd); in __sdt_free()
2167 sdd->sd = NULL; in __sdt_free()
2168 free_percpu(sdd->sds); in __sdt_free()
2169 sdd->sds = NULL; in __sdt_free()
2170 free_percpu(sdd->sg); in __sdt_free()
2171 sdd->sg = NULL; in __sdt_free()
2172 free_percpu(sdd->sgc); in __sdt_free()
2173 sdd->sgc = NULL; in __sdt_free()