1 /* sysfs.c: Toplogy sysfs support code for sparc64.
2 *
3 * Copyright (C) 2007 David S. Miller <davem@davemloft.net>
4 */
5 #include <linux/sched.h>
6 #include <linux/device.h>
7 #include <linux/cpu.h>
8 #include <linux/smp.h>
9 #include <linux/percpu.h>
10 #include <linux/init.h>
11
12 #include <asm/cpudata.h>
13 #include <asm/hypervisor.h>
14 #include <asm/spitfire.h>
15
16 static DEFINE_PER_CPU(struct hv_mmu_statistics, mmu_stats) __attribute__((aligned(64)));
17
18 #define SHOW_MMUSTAT_ULONG(NAME) \
19 static ssize_t show_##NAME(struct device *dev, \
20 struct device_attribute *attr, char *buf) \
21 { \
22 struct hv_mmu_statistics *p = &per_cpu(mmu_stats, dev->id); \
23 return sprintf(buf, "%lu\n", p->NAME); \
24 } \
25 static DEVICE_ATTR(NAME, 0444, show_##NAME, NULL)
26
27 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_8k_tte);
28 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_8k_tte);
29 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_64k_tte);
30 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_64k_tte);
31 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_4mb_tte);
32 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_4mb_tte);
33 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_256mb_tte);
34 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_256mb_tte);
35 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_8k_tte);
36 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_8k_tte);
37 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_64k_tte);
38 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_64k_tte);
39 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_4mb_tte);
40 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_4mb_tte);
41 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_256mb_tte);
42 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_256mb_tte);
43 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_8k_tte);
44 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_8k_tte);
45 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_64k_tte);
46 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_64k_tte);
47 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_4mb_tte);
48 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_4mb_tte);
49 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_256mb_tte);
50 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_256mb_tte);
51 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_8k_tte);
52 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_8k_tte);
53 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_64k_tte);
54 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_64k_tte);
55 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_4mb_tte);
56 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_4mb_tte);
57 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_256mb_tte);
58 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_256mb_tte);
59
60 static struct attribute *mmu_stat_attrs[] = {
61 &dev_attr_immu_tsb_hits_ctx0_8k_tte.attr,
62 &dev_attr_immu_tsb_ticks_ctx0_8k_tte.attr,
63 &dev_attr_immu_tsb_hits_ctx0_64k_tte.attr,
64 &dev_attr_immu_tsb_ticks_ctx0_64k_tte.attr,
65 &dev_attr_immu_tsb_hits_ctx0_4mb_tte.attr,
66 &dev_attr_immu_tsb_ticks_ctx0_4mb_tte.attr,
67 &dev_attr_immu_tsb_hits_ctx0_256mb_tte.attr,
68 &dev_attr_immu_tsb_ticks_ctx0_256mb_tte.attr,
69 &dev_attr_immu_tsb_hits_ctxnon0_8k_tte.attr,
70 &dev_attr_immu_tsb_ticks_ctxnon0_8k_tte.attr,
71 &dev_attr_immu_tsb_hits_ctxnon0_64k_tte.attr,
72 &dev_attr_immu_tsb_ticks_ctxnon0_64k_tte.attr,
73 &dev_attr_immu_tsb_hits_ctxnon0_4mb_tte.attr,
74 &dev_attr_immu_tsb_ticks_ctxnon0_4mb_tte.attr,
75 &dev_attr_immu_tsb_hits_ctxnon0_256mb_tte.attr,
76 &dev_attr_immu_tsb_ticks_ctxnon0_256mb_tte.attr,
77 &dev_attr_dmmu_tsb_hits_ctx0_8k_tte.attr,
78 &dev_attr_dmmu_tsb_ticks_ctx0_8k_tte.attr,
79 &dev_attr_dmmu_tsb_hits_ctx0_64k_tte.attr,
80 &dev_attr_dmmu_tsb_ticks_ctx0_64k_tte.attr,
81 &dev_attr_dmmu_tsb_hits_ctx0_4mb_tte.attr,
82 &dev_attr_dmmu_tsb_ticks_ctx0_4mb_tte.attr,
83 &dev_attr_dmmu_tsb_hits_ctx0_256mb_tte.attr,
84 &dev_attr_dmmu_tsb_ticks_ctx0_256mb_tte.attr,
85 &dev_attr_dmmu_tsb_hits_ctxnon0_8k_tte.attr,
86 &dev_attr_dmmu_tsb_ticks_ctxnon0_8k_tte.attr,
87 &dev_attr_dmmu_tsb_hits_ctxnon0_64k_tte.attr,
88 &dev_attr_dmmu_tsb_ticks_ctxnon0_64k_tte.attr,
89 &dev_attr_dmmu_tsb_hits_ctxnon0_4mb_tte.attr,
90 &dev_attr_dmmu_tsb_ticks_ctxnon0_4mb_tte.attr,
91 &dev_attr_dmmu_tsb_hits_ctxnon0_256mb_tte.attr,
92 &dev_attr_dmmu_tsb_ticks_ctxnon0_256mb_tte.attr,
93 NULL,
94 };
95
96 static struct attribute_group mmu_stat_group = {
97 .attrs = mmu_stat_attrs,
98 .name = "mmu_stats",
99 };
100
101 /* XXX convert to rusty's on_one_cpu */
run_on_cpu(unsigned long cpu,unsigned long (* func)(unsigned long),unsigned long arg)102 static unsigned long run_on_cpu(unsigned long cpu,
103 unsigned long (*func)(unsigned long),
104 unsigned long arg)
105 {
106 cpumask_t old_affinity;
107 unsigned long ret;
108
109 cpumask_copy(&old_affinity, tsk_cpus_allowed(current));
110 /* should return -EINVAL to userspace */
111 if (set_cpus_allowed_ptr(current, cpumask_of(cpu)))
112 return 0;
113
114 ret = func(arg);
115
116 set_cpus_allowed_ptr(current, &old_affinity);
117
118 return ret;
119 }
120
read_mmustat_enable(unsigned long junk)121 static unsigned long read_mmustat_enable(unsigned long junk)
122 {
123 unsigned long ra = 0;
124
125 sun4v_mmustat_info(&ra);
126
127 return ra != 0;
128 }
129
write_mmustat_enable(unsigned long val)130 static unsigned long write_mmustat_enable(unsigned long val)
131 {
132 unsigned long ra, orig_ra;
133
134 if (val)
135 ra = __pa(&per_cpu(mmu_stats, smp_processor_id()));
136 else
137 ra = 0UL;
138
139 return sun4v_mmustat_conf(ra, &orig_ra);
140 }
141
show_mmustat_enable(struct device * s,struct device_attribute * attr,char * buf)142 static ssize_t show_mmustat_enable(struct device *s,
143 struct device_attribute *attr, char *buf)
144 {
145 unsigned long val = run_on_cpu(s->id, read_mmustat_enable, 0);
146 return sprintf(buf, "%lx\n", val);
147 }
148
store_mmustat_enable(struct device * s,struct device_attribute * attr,const char * buf,size_t count)149 static ssize_t store_mmustat_enable(struct device *s,
150 struct device_attribute *attr, const char *buf,
151 size_t count)
152 {
153 unsigned long val, err;
154 int ret = sscanf(buf, "%ld", &val);
155
156 if (ret != 1)
157 return -EINVAL;
158
159 err = run_on_cpu(s->id, write_mmustat_enable, val);
160 if (err)
161 return -EIO;
162
163 return count;
164 }
165
166 static DEVICE_ATTR(mmustat_enable, 0644, show_mmustat_enable, store_mmustat_enable);
167
168 static int mmu_stats_supported;
169
register_mmu_stats(struct device * s)170 static int register_mmu_stats(struct device *s)
171 {
172 if (!mmu_stats_supported)
173 return 0;
174 device_create_file(s, &dev_attr_mmustat_enable);
175 return sysfs_create_group(&s->kobj, &mmu_stat_group);
176 }
177
178 #ifdef CONFIG_HOTPLUG_CPU
unregister_mmu_stats(struct device * s)179 static void unregister_mmu_stats(struct device *s)
180 {
181 if (!mmu_stats_supported)
182 return;
183 sysfs_remove_group(&s->kobj, &mmu_stat_group);
184 device_remove_file(s, &dev_attr_mmustat_enable);
185 }
186 #endif
187
188 #define SHOW_CPUDATA_ULONG_NAME(NAME, MEMBER) \
189 static ssize_t show_##NAME(struct device *dev, \
190 struct device_attribute *attr, char *buf) \
191 { \
192 cpuinfo_sparc *c = &cpu_data(dev->id); \
193 return sprintf(buf, "%lu\n", c->MEMBER); \
194 }
195
196 #define SHOW_CPUDATA_UINT_NAME(NAME, MEMBER) \
197 static ssize_t show_##NAME(struct device *dev, \
198 struct device_attribute *attr, char *buf) \
199 { \
200 cpuinfo_sparc *c = &cpu_data(dev->id); \
201 return sprintf(buf, "%u\n", c->MEMBER); \
202 }
203
204 SHOW_CPUDATA_ULONG_NAME(clock_tick, clock_tick);
205 SHOW_CPUDATA_UINT_NAME(l1_dcache_size, dcache_size);
206 SHOW_CPUDATA_UINT_NAME(l1_dcache_line_size, dcache_line_size);
207 SHOW_CPUDATA_UINT_NAME(l1_icache_size, icache_size);
208 SHOW_CPUDATA_UINT_NAME(l1_icache_line_size, icache_line_size);
209 SHOW_CPUDATA_UINT_NAME(l2_cache_size, ecache_size);
210 SHOW_CPUDATA_UINT_NAME(l2_cache_line_size, ecache_line_size);
211
212 static struct device_attribute cpu_core_attrs[] = {
213 __ATTR(clock_tick, 0444, show_clock_tick, NULL),
214 __ATTR(l1_dcache_size, 0444, show_l1_dcache_size, NULL),
215 __ATTR(l1_dcache_line_size, 0444, show_l1_dcache_line_size, NULL),
216 __ATTR(l1_icache_size, 0444, show_l1_icache_size, NULL),
217 __ATTR(l1_icache_line_size, 0444, show_l1_icache_line_size, NULL),
218 __ATTR(l2_cache_size, 0444, show_l2_cache_size, NULL),
219 __ATTR(l2_cache_line_size, 0444, show_l2_cache_line_size, NULL),
220 };
221
222 static DEFINE_PER_CPU(struct cpu, cpu_devices);
223
register_cpu_online(unsigned int cpu)224 static void register_cpu_online(unsigned int cpu)
225 {
226 struct cpu *c = &per_cpu(cpu_devices, cpu);
227 struct device *s = &c->dev;
228 int i;
229
230 for (i = 0; i < ARRAY_SIZE(cpu_core_attrs); i++)
231 device_create_file(s, &cpu_core_attrs[i]);
232
233 register_mmu_stats(s);
234 }
235
236 #ifdef CONFIG_HOTPLUG_CPU
unregister_cpu_online(unsigned int cpu)237 static void unregister_cpu_online(unsigned int cpu)
238 {
239 struct cpu *c = &per_cpu(cpu_devices, cpu);
240 struct device *s = &c->dev;
241 int i;
242
243 unregister_mmu_stats(s);
244 for (i = 0; i < ARRAY_SIZE(cpu_core_attrs); i++)
245 device_remove_file(s, &cpu_core_attrs[i]);
246 }
247 #endif
248
sysfs_cpu_notify(struct notifier_block * self,unsigned long action,void * hcpu)249 static int __cpuinit sysfs_cpu_notify(struct notifier_block *self,
250 unsigned long action, void *hcpu)
251 {
252 unsigned int cpu = (unsigned int)(long)hcpu;
253
254 switch (action) {
255 case CPU_ONLINE:
256 case CPU_ONLINE_FROZEN:
257 register_cpu_online(cpu);
258 break;
259 #ifdef CONFIG_HOTPLUG_CPU
260 case CPU_DEAD:
261 case CPU_DEAD_FROZEN:
262 unregister_cpu_online(cpu);
263 break;
264 #endif
265 }
266 return NOTIFY_OK;
267 }
268
269 static struct notifier_block __cpuinitdata sysfs_cpu_nb = {
270 .notifier_call = sysfs_cpu_notify,
271 };
272
check_mmu_stats(void)273 static void __init check_mmu_stats(void)
274 {
275 unsigned long dummy1, err;
276
277 if (tlb_type != hypervisor)
278 return;
279
280 err = sun4v_mmustat_info(&dummy1);
281 if (!err)
282 mmu_stats_supported = 1;
283 }
284
register_nodes(void)285 static void register_nodes(void)
286 {
287 #ifdef CONFIG_NUMA
288 int i;
289
290 for (i = 0; i < MAX_NUMNODES; i++)
291 register_one_node(i);
292 #endif
293 }
294
topology_init(void)295 static int __init topology_init(void)
296 {
297 int cpu;
298
299 register_nodes();
300
301 check_mmu_stats();
302
303 register_cpu_notifier(&sysfs_cpu_nb);
304
305 for_each_possible_cpu(cpu) {
306 struct cpu *c = &per_cpu(cpu_devices, cpu);
307
308 register_cpu(c, cpu);
309 if (cpu_online(cpu))
310 register_cpu_online(cpu);
311 }
312
313 return 0;
314 }
315
316 subsys_initcall(topology_init);
317