1 // SPDX-License-Identifier: GPL-2.0+
2 //
3 // Copyright 2019 Madhavan Srinivasan, IBM Corporation.
4 
5 #define pr_fmt(fmt)	"generic-compat-pmu: " fmt
6 
7 #include "isa207-common.h"
8 
9 /*
10  * Raw event encoding:
11  *
12  *        60        56        52        48        44        40        36        32
13  * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
14  *
15  *        28        24        20        16        12         8         4         0
16  * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
17  *                                 [ pmc ]                       [    pmcxsel    ]
18  */
19 
20 /*
21  * Event codes defined in ISA v3.0B
22  */
23 #define EVENT(_name, _code)	_name = _code,
24 
25 enum {
26 	/* Cycles, alternate code */
27 	EVENT(PM_CYC_ALT,			0x100f0)
28 	/* One or more instructions completed in a cycle */
29 	EVENT(PM_CYC_INST_CMPL,			0x100f2)
30 	/* Floating-point instruction completed */
31 	EVENT(PM_FLOP_CMPL,			0x100f4)
32 	/* Instruction ERAT/L1-TLB miss */
33 	EVENT(PM_L1_ITLB_MISS,			0x100f6)
34 	/* All instructions completed and none available */
35 	EVENT(PM_NO_INST_AVAIL,			0x100f8)
36 	/* A load-type instruction completed (ISA v3.0+) */
37 	EVENT(PM_LD_CMPL,			0x100fc)
38 	/* Instruction completed, alternate code (ISA v3.0+) */
39 	EVENT(PM_INST_CMPL_ALT,			0x100fe)
40 	/* A store-type instruction completed */
41 	EVENT(PM_ST_CMPL,			0x200f0)
42 	/* Instruction Dispatched */
43 	EVENT(PM_INST_DISP,			0x200f2)
44 	/* Run_cycles */
45 	EVENT(PM_RUN_CYC,			0x200f4)
46 	/* Data ERAT/L1-TLB miss/reload */
47 	EVENT(PM_L1_DTLB_RELOAD,		0x200f6)
48 	/* Taken branch completed */
49 	EVENT(PM_BR_TAKEN_CMPL,			0x200fa)
50 	/* Demand iCache Miss */
51 	EVENT(PM_L1_ICACHE_MISS,		0x200fc)
52 	/* L1 Dcache reload from memory */
53 	EVENT(PM_L1_RELOAD_FROM_MEM,		0x200fe)
54 	/* L1 Dcache store miss */
55 	EVENT(PM_ST_MISS_L1,			0x300f0)
56 	/* Alternate code for PM_INST_DISP */
57 	EVENT(PM_INST_DISP_ALT,			0x300f2)
58 	/* Branch direction or target mispredicted */
59 	EVENT(PM_BR_MISPREDICT,			0x300f6)
60 	/* Data TLB miss/reload */
61 	EVENT(PM_DTLB_MISS,			0x300fc)
62 	/* Demand LD - L3 Miss (not L2 hit and not L3 hit) */
63 	EVENT(PM_DATA_FROM_L3MISS,		0x300fe)
64 	/* L1 Dcache load miss */
65 	EVENT(PM_LD_MISS_L1,			0x400f0)
66 	/* Cycle when instruction(s) dispatched */
67 	EVENT(PM_CYC_INST_DISP,			0x400f2)
68 	/* Branch or branch target mispredicted */
69 	EVENT(PM_BR_MPRED_CMPL,			0x400f6)
70 	/* Instructions completed with run latch set */
71 	EVENT(PM_RUN_INST_CMPL,			0x400fa)
72 	/* Instruction TLB miss/reload */
73 	EVENT(PM_ITLB_MISS,			0x400fc)
74 	/* Load data not cached */
75 	EVENT(PM_LD_NOT_CACHED,			0x400fe)
76 	/* Instructions */
77 	EVENT(PM_INST_CMPL,			0x500fa)
78 	/* Cycles */
79 	EVENT(PM_CYC,				0x600f4)
80 };
81 
82 #undef EVENT
83 
84 /* Table of alternatives, sorted in increasing order of column 0 */
85 /* Note that in each row, column 0 must be the smallest */
86 static const unsigned int generic_event_alternatives[][MAX_ALT] = {
87 	{ PM_CYC_ALT,			PM_CYC },
88 	{ PM_INST_CMPL_ALT,		PM_INST_CMPL },
89 	{ PM_INST_DISP,			PM_INST_DISP_ALT },
90 };
91 
generic_get_alternatives(u64 event,unsigned int flags,u64 alt[])92 static int generic_get_alternatives(u64 event, unsigned int flags, u64 alt[])
93 {
94 	int num_alt = 0;
95 
96 	num_alt = isa207_get_alternatives(event, alt,
97 					  ARRAY_SIZE(generic_event_alternatives), flags,
98 					  generic_event_alternatives);
99 
100 	return num_alt;
101 }
102 
103 GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
104 GENERIC_EVENT_ATTR(instructions,		PM_INST_CMPL);
105 GENERIC_EVENT_ATTR(stalled-cycles-frontend,	PM_NO_INST_AVAIL);
106 GENERIC_EVENT_ATTR(branch-misses,		PM_BR_MPRED_CMPL);
107 GENERIC_EVENT_ATTR(cache-misses,		PM_LD_MISS_L1);
108 
109 CACHE_EVENT_ATTR(L1-dcache-load-misses,		PM_LD_MISS_L1);
110 CACHE_EVENT_ATTR(L1-dcache-store-misses,	PM_ST_MISS_L1);
111 CACHE_EVENT_ATTR(L1-icache-load-misses,		PM_L1_ICACHE_MISS);
112 CACHE_EVENT_ATTR(LLC-load-misses,		PM_DATA_FROM_L3MISS);
113 CACHE_EVENT_ATTR(branch-load-misses,		PM_BR_MPRED_CMPL);
114 CACHE_EVENT_ATTR(dTLB-load-misses,		PM_DTLB_MISS);
115 CACHE_EVENT_ATTR(iTLB-load-misses,		PM_ITLB_MISS);
116 
117 static struct attribute *generic_compat_events_attr[] = {
118 	GENERIC_EVENT_PTR(PM_CYC),
119 	GENERIC_EVENT_PTR(PM_INST_CMPL),
120 	GENERIC_EVENT_PTR(PM_NO_INST_AVAIL),
121 	GENERIC_EVENT_PTR(PM_BR_MPRED_CMPL),
122 	GENERIC_EVENT_PTR(PM_LD_MISS_L1),
123 	CACHE_EVENT_PTR(PM_LD_MISS_L1),
124 	CACHE_EVENT_PTR(PM_ST_MISS_L1),
125 	CACHE_EVENT_PTR(PM_L1_ICACHE_MISS),
126 	CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS),
127 	CACHE_EVENT_PTR(PM_BR_MPRED_CMPL),
128 	CACHE_EVENT_PTR(PM_DTLB_MISS),
129 	CACHE_EVENT_PTR(PM_ITLB_MISS),
130 	NULL
131 };
132 
133 static const struct attribute_group generic_compat_pmu_events_group = {
134 	.name = "events",
135 	.attrs = generic_compat_events_attr,
136 };
137 
138 PMU_FORMAT_ATTR(event,		"config:0-19");
139 PMU_FORMAT_ATTR(pmcxsel,	"config:0-7");
140 PMU_FORMAT_ATTR(pmc,		"config:16-19");
141 
142 static struct attribute *generic_compat_pmu_format_attr[] = {
143 	&format_attr_event.attr,
144 	&format_attr_pmcxsel.attr,
145 	&format_attr_pmc.attr,
146 	NULL,
147 };
148 
149 static const struct attribute_group generic_compat_pmu_format_group = {
150 	.name = "format",
151 	.attrs = generic_compat_pmu_format_attr,
152 };
153 
154 static const struct attribute_group *generic_compat_pmu_attr_groups[] = {
155 	&generic_compat_pmu_format_group,
156 	&generic_compat_pmu_events_group,
157 	NULL,
158 };
159 
160 static int compat_generic_events[] = {
161 	[PERF_COUNT_HW_CPU_CYCLES] =			PM_CYC,
162 	[PERF_COUNT_HW_INSTRUCTIONS] =			PM_INST_CMPL,
163 	[PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] =	PM_NO_INST_AVAIL,
164 	[PERF_COUNT_HW_BRANCH_MISSES] =			PM_BR_MPRED_CMPL,
165 	[PERF_COUNT_HW_CACHE_MISSES] =			PM_LD_MISS_L1,
166 };
167 
168 #define C(x)	PERF_COUNT_HW_CACHE_##x
169 
170 /*
171  * Table of generalized cache-related events.
172  * 0 means not supported, -1 means nonsensical, other values
173  * are event codes.
174  */
175 static u64 generic_compat_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
176 	[ C(L1D) ] = {
177 		[ C(OP_READ) ] = {
178 			[ C(RESULT_ACCESS) ] = 0,
179 			[ C(RESULT_MISS)   ] = PM_LD_MISS_L1,
180 		},
181 		[ C(OP_WRITE) ] = {
182 			[ C(RESULT_ACCESS) ] = 0,
183 			[ C(RESULT_MISS)   ] = PM_ST_MISS_L1,
184 		},
185 		[ C(OP_PREFETCH) ] = {
186 			[ C(RESULT_ACCESS) ] = 0,
187 			[ C(RESULT_MISS)   ] = 0,
188 		},
189 	},
190 	[ C(L1I) ] = {
191 		[ C(OP_READ) ] = {
192 			[ C(RESULT_ACCESS) ] = 0,
193 			[ C(RESULT_MISS)   ] = PM_L1_ICACHE_MISS,
194 		},
195 		[ C(OP_WRITE) ] = {
196 			[ C(RESULT_ACCESS) ] = 0,
197 			[ C(RESULT_MISS)   ] = -1,
198 		},
199 		[ C(OP_PREFETCH) ] = {
200 			[ C(RESULT_ACCESS) ] = 0,
201 			[ C(RESULT_MISS)   ] = 0,
202 		},
203 	},
204 	[ C(LL) ] = {
205 		[ C(OP_READ) ] = {
206 			[ C(RESULT_ACCESS) ] = 0,
207 			[ C(RESULT_MISS)   ] = PM_DATA_FROM_L3MISS,
208 		},
209 		[ C(OP_WRITE) ] = {
210 			[ C(RESULT_ACCESS) ] = 0,
211 			[ C(RESULT_MISS)   ] = 0,
212 		},
213 		[ C(OP_PREFETCH) ] = {
214 			[ C(RESULT_ACCESS) ] = 0,
215 			[ C(RESULT_MISS)   ] = 0,
216 		},
217 	},
218 	[ C(DTLB) ] = {
219 		[ C(OP_READ) ] = {
220 			[ C(RESULT_ACCESS) ] = 0,
221 			[ C(RESULT_MISS)   ] = PM_DTLB_MISS,
222 		},
223 		[ C(OP_WRITE) ] = {
224 			[ C(RESULT_ACCESS) ] = -1,
225 			[ C(RESULT_MISS)   ] = -1,
226 		},
227 		[ C(OP_PREFETCH) ] = {
228 			[ C(RESULT_ACCESS) ] = -1,
229 			[ C(RESULT_MISS)   ] = -1,
230 		},
231 	},
232 	[ C(ITLB) ] = {
233 		[ C(OP_READ) ] = {
234 			[ C(RESULT_ACCESS) ] = 0,
235 			[ C(RESULT_MISS)   ] = PM_ITLB_MISS,
236 		},
237 		[ C(OP_WRITE) ] = {
238 			[ C(RESULT_ACCESS) ] = -1,
239 			[ C(RESULT_MISS)   ] = -1,
240 		},
241 		[ C(OP_PREFETCH) ] = {
242 			[ C(RESULT_ACCESS) ] = -1,
243 			[ C(RESULT_MISS)   ] = -1,
244 		},
245 	},
246 	[ C(BPU) ] = {
247 		[ C(OP_READ) ] = {
248 			[ C(RESULT_ACCESS) ] = 0,
249 			[ C(RESULT_MISS)   ] = PM_BR_MPRED_CMPL,
250 		},
251 		[ C(OP_WRITE) ] = {
252 			[ C(RESULT_ACCESS) ] = -1,
253 			[ C(RESULT_MISS)   ] = -1,
254 		},
255 		[ C(OP_PREFETCH) ] = {
256 			[ C(RESULT_ACCESS) ] = -1,
257 			[ C(RESULT_MISS)   ] = -1,
258 		},
259 	},
260 	[ C(NODE) ] = {
261 		[ C(OP_READ) ] = {
262 			[ C(RESULT_ACCESS) ] = -1,
263 			[ C(RESULT_MISS)   ] = -1,
264 		},
265 		[ C(OP_WRITE) ] = {
266 			[ C(RESULT_ACCESS) ] = -1,
267 			[ C(RESULT_MISS)   ] = -1,
268 		},
269 		[ C(OP_PREFETCH) ] = {
270 			[ C(RESULT_ACCESS) ] = -1,
271 			[ C(RESULT_MISS)   ] = -1,
272 		},
273 	},
274 };
275 
276 #undef C
277 
278 /*
279  * We set MMCR0[CC5-6RUN] so we can use counters 5 and 6 for
280  * PM_INST_CMPL and PM_CYC.
281  */
generic_compute_mmcr(u64 event[],int n_ev,unsigned int hwc[],struct mmcr_regs * mmcr,struct perf_event * pevents[],u32 flags)282 static int generic_compute_mmcr(u64 event[], int n_ev,
283 				unsigned int hwc[], struct mmcr_regs *mmcr,
284 				struct perf_event *pevents[], u32 flags)
285 {
286 	int ret;
287 
288 	ret = isa207_compute_mmcr(event, n_ev, hwc, mmcr, pevents, flags);
289 	if (!ret)
290 		mmcr->mmcr0 |= MMCR0_C56RUN;
291 	return ret;
292 }
293 
294 static struct power_pmu generic_compat_pmu = {
295 	.name			= "GENERIC_COMPAT",
296 	.n_counter		= MAX_PMU_COUNTERS,
297 	.add_fields		= ISA207_ADD_FIELDS,
298 	.test_adder		= ISA207_TEST_ADDER,
299 	.compute_mmcr		= generic_compute_mmcr,
300 	.get_constraint		= isa207_get_constraint,
301 	.get_alternatives	= generic_get_alternatives,
302 	.disable_pmc		= isa207_disable_pmc,
303 	.flags			= PPMU_HAS_SIER | PPMU_ARCH_207S,
304 	.n_generic		= ARRAY_SIZE(compat_generic_events),
305 	.generic_events		= compat_generic_events,
306 	.cache_events		= &generic_compat_cache_events,
307 	.attr_groups		= generic_compat_pmu_attr_groups,
308 };
309 
init_generic_compat_pmu(void)310 int __init init_generic_compat_pmu(void)
311 {
312 	int rc = 0;
313 
314 	/*
315 	 * From ISA v2.07 on, PMU features are architected;
316 	 * we require >= v3.0 because (a) that has PM_LD_CMPL and
317 	 * PM_INST_CMPL_ALT, which v2.07 doesn't have, and
318 	 * (b) we don't expect any non-IBM Power ISA
319 	 * implementations that conform to v2.07 but not v3.0.
320 	 */
321 	if (!cpu_has_feature(CPU_FTR_ARCH_300))
322 		return -ENODEV;
323 
324 	rc = register_power_pmu(&generic_compat_pmu);
325 	if (rc)
326 		return rc;
327 
328 	/* Tell userspace that EBB is supported */
329 	cur_cpu_spec->cpu_user_features2 |= PPC_FEATURE2_EBB;
330 
331 	return 0;
332 }
333