arm: omap2+: add missing HWMOD_NO_IDLEST in 81xx hwmod data
[deliverable/linux.git] / arch / arm / kernel / perf_event_v7.c
1 /*
2 * ARMv7 Cortex-A8 and Cortex-A9 Performance Events handling code.
3 *
4 * ARMv7 support: Jean Pihet <jpihet@mvista.com>
5 * 2010 (c) MontaVista Software, LLC.
6 *
7 * Copied from ARMv6 code, with the low level code inspired
8 * by the ARMv7 Oprofile code.
9 *
10 * Cortex-A8 has up to 4 configurable performance counters and
11 * a single cycle counter.
12 * Cortex-A9 has up to 31 configurable performance counters and
13 * a single cycle counter.
14 *
15 * All counters can be enabled/disabled and IRQ masked separately. The cycle
16 * counter and all 4 performance counters together can be reset separately.
17 */
18
19 #ifdef CONFIG_CPU_V7
20
21 #include <asm/cp15.h>
22 #include <asm/cputype.h>
23 #include <asm/irq_regs.h>
24 #include <asm/vfp.h>
25 #include "../vfp/vfpinstr.h"
26
27 #include <linux/of.h>
28 #include <linux/perf/arm_pmu.h>
29 #include <linux/platform_device.h>
30
31 /*
32 * Common ARMv7 event types
33 *
34 * Note: An implementation may not be able to count all of these events
35 * but the encodings are considered to be `reserved' in the case that
36 * they are not available.
37 */
38 enum armv7_perf_types {
39 ARMV7_PERFCTR_PMNC_SW_INCR = 0x00,
40 ARMV7_PERFCTR_L1_ICACHE_REFILL = 0x01,
41 ARMV7_PERFCTR_ITLB_REFILL = 0x02,
42 ARMV7_PERFCTR_L1_DCACHE_REFILL = 0x03,
43 ARMV7_PERFCTR_L1_DCACHE_ACCESS = 0x04,
44 ARMV7_PERFCTR_DTLB_REFILL = 0x05,
45 ARMV7_PERFCTR_MEM_READ = 0x06,
46 ARMV7_PERFCTR_MEM_WRITE = 0x07,
47 ARMV7_PERFCTR_INSTR_EXECUTED = 0x08,
48 ARMV7_PERFCTR_EXC_TAKEN = 0x09,
49 ARMV7_PERFCTR_EXC_EXECUTED = 0x0A,
50 ARMV7_PERFCTR_CID_WRITE = 0x0B,
51
52 /*
53 * ARMV7_PERFCTR_PC_WRITE is equivalent to HW_BRANCH_INSTRUCTIONS.
54 * It counts:
55 * - all (taken) branch instructions,
56 * - instructions that explicitly write the PC,
57 * - exception generating instructions.
58 */
59 ARMV7_PERFCTR_PC_WRITE = 0x0C,
60 ARMV7_PERFCTR_PC_IMM_BRANCH = 0x0D,
61 ARMV7_PERFCTR_PC_PROC_RETURN = 0x0E,
62 ARMV7_PERFCTR_MEM_UNALIGNED_ACCESS = 0x0F,
63 ARMV7_PERFCTR_PC_BRANCH_MIS_PRED = 0x10,
64 ARMV7_PERFCTR_CLOCK_CYCLES = 0x11,
65 ARMV7_PERFCTR_PC_BRANCH_PRED = 0x12,
66
67 /* These events are defined by the PMUv2 supplement (ARM DDI 0457A). */
68 ARMV7_PERFCTR_MEM_ACCESS = 0x13,
69 ARMV7_PERFCTR_L1_ICACHE_ACCESS = 0x14,
70 ARMV7_PERFCTR_L1_DCACHE_WB = 0x15,
71 ARMV7_PERFCTR_L2_CACHE_ACCESS = 0x16,
72 ARMV7_PERFCTR_L2_CACHE_REFILL = 0x17,
73 ARMV7_PERFCTR_L2_CACHE_WB = 0x18,
74 ARMV7_PERFCTR_BUS_ACCESS = 0x19,
75 ARMV7_PERFCTR_MEM_ERROR = 0x1A,
76 ARMV7_PERFCTR_INSTR_SPEC = 0x1B,
77 ARMV7_PERFCTR_TTBR_WRITE = 0x1C,
78 ARMV7_PERFCTR_BUS_CYCLES = 0x1D,
79
80 ARMV7_PERFCTR_CPU_CYCLES = 0xFF
81 };
82
83 /* ARMv7 Cortex-A8 specific event types */
84 enum armv7_a8_perf_types {
85 ARMV7_A8_PERFCTR_L2_CACHE_ACCESS = 0x43,
86 ARMV7_A8_PERFCTR_L2_CACHE_REFILL = 0x44,
87 ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS = 0x50,
88 ARMV7_A8_PERFCTR_STALL_ISIDE = 0x56,
89 };
90
91 /* ARMv7 Cortex-A9 specific event types */
92 enum armv7_a9_perf_types {
93 ARMV7_A9_PERFCTR_INSTR_CORE_RENAME = 0x68,
94 ARMV7_A9_PERFCTR_STALL_ICACHE = 0x60,
95 ARMV7_A9_PERFCTR_STALL_DISPATCH = 0x66,
96 };
97
98 /* ARMv7 Cortex-A5 specific event types */
99 enum armv7_a5_perf_types {
100 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL = 0xc2,
101 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP = 0xc3,
102 };
103
104 /* ARMv7 Cortex-A15 specific event types */
105 enum armv7_a15_perf_types {
106 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ = 0x40,
107 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE = 0x41,
108 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ = 0x42,
109 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE = 0x43,
110
111 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ = 0x4C,
112 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE = 0x4D,
113
114 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ = 0x50,
115 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE = 0x51,
116 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ = 0x52,
117 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE = 0x53,
118
119 ARMV7_A15_PERFCTR_PC_WRITE_SPEC = 0x76,
120 };
121
122 /* ARMv7 Cortex-A12 specific event types */
123 enum armv7_a12_perf_types {
124 ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_READ = 0x40,
125 ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_WRITE = 0x41,
126
127 ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_READ = 0x50,
128 ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_WRITE = 0x51,
129
130 ARMV7_A12_PERFCTR_PC_WRITE_SPEC = 0x76,
131
132 ARMV7_A12_PERFCTR_PF_TLB_REFILL = 0xe7,
133 };
134
135 /* ARMv7 Krait specific event types */
136 enum krait_perf_types {
137 KRAIT_PMRESR0_GROUP0 = 0xcc,
138 KRAIT_PMRESR1_GROUP0 = 0xd0,
139 KRAIT_PMRESR2_GROUP0 = 0xd4,
140 KRAIT_VPMRESR0_GROUP0 = 0xd8,
141
142 KRAIT_PERFCTR_L1_ICACHE_ACCESS = 0x10011,
143 KRAIT_PERFCTR_L1_ICACHE_MISS = 0x10010,
144
145 KRAIT_PERFCTR_L1_ITLB_ACCESS = 0x12222,
146 KRAIT_PERFCTR_L1_DTLB_ACCESS = 0x12210,
147 };
148
149 /* ARMv7 Scorpion specific event types */
150 enum scorpion_perf_types {
151 SCORPION_LPM0_GROUP0 = 0x4c,
152 SCORPION_LPM1_GROUP0 = 0x50,
153 SCORPION_LPM2_GROUP0 = 0x54,
154 SCORPION_L2LPM_GROUP0 = 0x58,
155 SCORPION_VLPM_GROUP0 = 0x5c,
156
157 SCORPION_ICACHE_ACCESS = 0x10053,
158 SCORPION_ICACHE_MISS = 0x10052,
159
160 SCORPION_DTLB_ACCESS = 0x12013,
161 SCORPION_DTLB_MISS = 0x12012,
162
163 SCORPION_ITLB_MISS = 0x12021,
164 };
165
166 /*
167 * Cortex-A8 HW events mapping
168 *
169 * The hardware events that we support. We do support cache operations but
170 * we have harvard caches and no way to combine instruction and data
171 * accesses/misses in hardware.
172 */
173 static const unsigned armv7_a8_perf_map[PERF_COUNT_HW_MAX] = {
174 PERF_MAP_ALL_UNSUPPORTED,
175 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
176 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
177 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
178 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
179 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
180 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
181 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = ARMV7_A8_PERFCTR_STALL_ISIDE,
182 };
183
184 static const unsigned armv7_a8_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
185 [PERF_COUNT_HW_CACHE_OP_MAX]
186 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
187 PERF_CACHE_MAP_ALL_UNSUPPORTED,
188
189 /*
190 * The performance counters don't differentiate between read and write
191 * accesses/misses so this isn't strictly correct, but it's the best we
192 * can do. Writes and reads get combined.
193 */
194 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
195 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
196 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
197 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
198
199 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS,
200 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
201
202 [C(LL)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS,
203 [C(LL)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL,
204 [C(LL)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS,
205 [C(LL)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL,
206
207 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
208 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
209
210 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
211 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
212
213 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
214 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
215 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
216 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
217 };
218
219 /*
220 * Cortex-A9 HW events mapping
221 */
222 static const unsigned armv7_a9_perf_map[PERF_COUNT_HW_MAX] = {
223 PERF_MAP_ALL_UNSUPPORTED,
224 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
225 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_A9_PERFCTR_INSTR_CORE_RENAME,
226 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
227 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
228 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
229 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
230 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = ARMV7_A9_PERFCTR_STALL_ICACHE,
231 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = ARMV7_A9_PERFCTR_STALL_DISPATCH,
232 };
233
234 static const unsigned armv7_a9_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
235 [PERF_COUNT_HW_CACHE_OP_MAX]
236 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
237 PERF_CACHE_MAP_ALL_UNSUPPORTED,
238
239 /*
240 * The performance counters don't differentiate between read and write
241 * accesses/misses so this isn't strictly correct, but it's the best we
242 * can do. Writes and reads get combined.
243 */
244 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
245 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
246 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
247 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
248
249 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
250
251 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
252 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
253
254 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
255 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
256
257 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
258 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
259 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
260 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
261 };
262
263 /*
264 * Cortex-A5 HW events mapping
265 */
266 static const unsigned armv7_a5_perf_map[PERF_COUNT_HW_MAX] = {
267 PERF_MAP_ALL_UNSUPPORTED,
268 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
269 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
270 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
271 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
272 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
273 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
274 };
275
276 static const unsigned armv7_a5_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
277 [PERF_COUNT_HW_CACHE_OP_MAX]
278 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
279 PERF_CACHE_MAP_ALL_UNSUPPORTED,
280
281 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
282 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
283 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
284 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
285 [C(L1D)][C(OP_PREFETCH)][C(RESULT_ACCESS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL,
286 [C(L1D)][C(OP_PREFETCH)][C(RESULT_MISS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP,
287
288 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
289 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
290 /*
291 * The prefetch counters don't differentiate between the I side and the
292 * D side.
293 */
294 [C(L1I)][C(OP_PREFETCH)][C(RESULT_ACCESS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL,
295 [C(L1I)][C(OP_PREFETCH)][C(RESULT_MISS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP,
296
297 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
298 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
299
300 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
301 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
302
303 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
304 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
305 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
306 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
307 };
308
309 /*
310 * Cortex-A15 HW events mapping
311 */
312 static const unsigned armv7_a15_perf_map[PERF_COUNT_HW_MAX] = {
313 PERF_MAP_ALL_UNSUPPORTED,
314 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
315 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
316 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
317 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
318 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_A15_PERFCTR_PC_WRITE_SPEC,
319 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
320 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_BUS_CYCLES,
321 };
322
323 static const unsigned armv7_a15_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
324 [PERF_COUNT_HW_CACHE_OP_MAX]
325 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
326 PERF_CACHE_MAP_ALL_UNSUPPORTED,
327
328 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ,
329 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ,
330 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE,
331 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE,
332
333 /*
334 * Not all performance counters differentiate between read and write
335 * accesses/misses so we're not always strictly correct, but it's the
336 * best we can do. Writes and reads get combined in these cases.
337 */
338 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
339 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
340
341 [C(LL)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ,
342 [C(LL)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ,
343 [C(LL)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE,
344 [C(LL)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE,
345
346 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ,
347 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE,
348
349 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
350 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
351
352 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
353 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
354 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
355 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
356 };
357
358 /*
359 * Cortex-A7 HW events mapping
360 */
361 static const unsigned armv7_a7_perf_map[PERF_COUNT_HW_MAX] = {
362 PERF_MAP_ALL_UNSUPPORTED,
363 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
364 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
365 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
366 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
367 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
368 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
369 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_BUS_CYCLES,
370 };
371
372 static const unsigned armv7_a7_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
373 [PERF_COUNT_HW_CACHE_OP_MAX]
374 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
375 PERF_CACHE_MAP_ALL_UNSUPPORTED,
376
377 /*
378 * The performance counters don't differentiate between read and write
379 * accesses/misses so this isn't strictly correct, but it's the best we
380 * can do. Writes and reads get combined.
381 */
382 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
383 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
384 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
385 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
386
387 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
388 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
389
390 [C(LL)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L2_CACHE_ACCESS,
391 [C(LL)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
392 [C(LL)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L2_CACHE_ACCESS,
393 [C(LL)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
394
395 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
396 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
397
398 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
399 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
400
401 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
402 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
403 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
404 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
405 };
406
407 /*
408 * Cortex-A12 HW events mapping
409 */
410 static const unsigned armv7_a12_perf_map[PERF_COUNT_HW_MAX] = {
411 PERF_MAP_ALL_UNSUPPORTED,
412 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
413 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
414 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
415 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
416 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_A12_PERFCTR_PC_WRITE_SPEC,
417 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
418 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_BUS_CYCLES,
419 };
420
421 static const unsigned armv7_a12_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
422 [PERF_COUNT_HW_CACHE_OP_MAX]
423 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
424 PERF_CACHE_MAP_ALL_UNSUPPORTED,
425
426 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_READ,
427 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
428 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_WRITE,
429 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
430
431 /*
432 * Not all performance counters differentiate between read and write
433 * accesses/misses so we're not always strictly correct, but it's the
434 * best we can do. Writes and reads get combined in these cases.
435 */
436 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
437 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
438
439 [C(LL)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_READ,
440 [C(LL)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
441 [C(LL)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_WRITE,
442 [C(LL)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
443
444 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
445 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
446 [C(DTLB)][C(OP_PREFETCH)][C(RESULT_MISS)] = ARMV7_A12_PERFCTR_PF_TLB_REFILL,
447
448 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
449 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
450
451 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
452 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
453 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
454 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
455 };
456
457 /*
458 * Krait HW events mapping
459 */
460 static const unsigned krait_perf_map[PERF_COUNT_HW_MAX] = {
461 PERF_MAP_ALL_UNSUPPORTED,
462 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
463 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
464 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
465 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
466 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_CLOCK_CYCLES,
467 };
468
469 static const unsigned krait_perf_map_no_branch[PERF_COUNT_HW_MAX] = {
470 PERF_MAP_ALL_UNSUPPORTED,
471 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
472 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
473 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
474 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_CLOCK_CYCLES,
475 };
476
477 static const unsigned krait_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
478 [PERF_COUNT_HW_CACHE_OP_MAX]
479 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
480 PERF_CACHE_MAP_ALL_UNSUPPORTED,
481
482 /*
483 * The performance counters don't differentiate between read and write
484 * accesses/misses so this isn't strictly correct, but it's the best we
485 * can do. Writes and reads get combined.
486 */
487 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
488 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
489 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
490 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
491
492 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = KRAIT_PERFCTR_L1_ICACHE_ACCESS,
493 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = KRAIT_PERFCTR_L1_ICACHE_MISS,
494
495 [C(DTLB)][C(OP_READ)][C(RESULT_ACCESS)] = KRAIT_PERFCTR_L1_DTLB_ACCESS,
496 [C(DTLB)][C(OP_WRITE)][C(RESULT_ACCESS)] = KRAIT_PERFCTR_L1_DTLB_ACCESS,
497
498 [C(ITLB)][C(OP_READ)][C(RESULT_ACCESS)] = KRAIT_PERFCTR_L1_ITLB_ACCESS,
499 [C(ITLB)][C(OP_WRITE)][C(RESULT_ACCESS)] = KRAIT_PERFCTR_L1_ITLB_ACCESS,
500
501 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
502 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
503 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
504 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
505 };
506
507 /*
508 * Scorpion HW events mapping
509 */
510 static const unsigned scorpion_perf_map[PERF_COUNT_HW_MAX] = {
511 PERF_MAP_ALL_UNSUPPORTED,
512 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
513 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
514 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
515 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
516 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_CLOCK_CYCLES,
517 };
518
519 static const unsigned scorpion_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
520 [PERF_COUNT_HW_CACHE_OP_MAX]
521 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
522 PERF_CACHE_MAP_ALL_UNSUPPORTED,
523 /*
524 * The performance counters don't differentiate between read and write
525 * accesses/misses so this isn't strictly correct, but it's the best we
526 * can do. Writes and reads get combined.
527 */
528 [C(L1D)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
529 [C(L1D)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
530 [C(L1D)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
531 [C(L1D)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
532 [C(L1I)][C(OP_READ)][C(RESULT_ACCESS)] = SCORPION_ICACHE_ACCESS,
533 [C(L1I)][C(OP_READ)][C(RESULT_MISS)] = SCORPION_ICACHE_MISS,
534 /*
535 * Only ITLB misses and DTLB refills are supported. If users want the
536 * DTLB refills misses a raw counter must be used.
537 */
538 [C(DTLB)][C(OP_READ)][C(RESULT_ACCESS)] = SCORPION_DTLB_ACCESS,
539 [C(DTLB)][C(OP_READ)][C(RESULT_MISS)] = SCORPION_DTLB_MISS,
540 [C(DTLB)][C(OP_WRITE)][C(RESULT_ACCESS)] = SCORPION_DTLB_ACCESS,
541 [C(DTLB)][C(OP_WRITE)][C(RESULT_MISS)] = SCORPION_DTLB_MISS,
542 [C(ITLB)][C(OP_READ)][C(RESULT_MISS)] = SCORPION_ITLB_MISS,
543 [C(ITLB)][C(OP_WRITE)][C(RESULT_MISS)] = SCORPION_ITLB_MISS,
544 [C(BPU)][C(OP_READ)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
545 [C(BPU)][C(OP_READ)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
546 [C(BPU)][C(OP_WRITE)][C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
547 [C(BPU)][C(OP_WRITE)][C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
548 };
549
550 /*
551 * Perf Events' indices
552 */
553 #define ARMV7_IDX_CYCLE_COUNTER 0
554 #define ARMV7_IDX_COUNTER0 1
555 #define ARMV7_IDX_COUNTER_LAST(cpu_pmu) \
556 (ARMV7_IDX_CYCLE_COUNTER + cpu_pmu->num_events - 1)
557
558 #define ARMV7_MAX_COUNTERS 32
559 #define ARMV7_COUNTER_MASK (ARMV7_MAX_COUNTERS - 1)
560
561 /*
562 * ARMv7 low level PMNC access
563 */
564
565 /*
566 * Perf Event to low level counters mapping
567 */
568 #define ARMV7_IDX_TO_COUNTER(x) \
569 (((x) - ARMV7_IDX_COUNTER0) & ARMV7_COUNTER_MASK)
570
571 /*
572 * Per-CPU PMNC: config reg
573 */
574 #define ARMV7_PMNC_E (1 << 0) /* Enable all counters */
575 #define ARMV7_PMNC_P (1 << 1) /* Reset all counters */
576 #define ARMV7_PMNC_C (1 << 2) /* Cycle counter reset */
577 #define ARMV7_PMNC_D (1 << 3) /* CCNT counts every 64th cpu cycle */
578 #define ARMV7_PMNC_X (1 << 4) /* Export to ETM */
579 #define ARMV7_PMNC_DP (1 << 5) /* Disable CCNT if non-invasive debug*/
580 #define ARMV7_PMNC_N_SHIFT 11 /* Number of counters supported */
581 #define ARMV7_PMNC_N_MASK 0x1f
582 #define ARMV7_PMNC_MASK 0x3f /* Mask for writable bits */
583
584 /*
585 * FLAG: counters overflow flag status reg
586 */
587 #define ARMV7_FLAG_MASK 0xffffffff /* Mask for writable bits */
588 #define ARMV7_OVERFLOWED_MASK ARMV7_FLAG_MASK
589
590 /*
591 * PMXEVTYPER: Event selection reg
592 */
593 #define ARMV7_EVTYPE_MASK 0xc80000ff /* Mask for writable bits */
594 #define ARMV7_EVTYPE_EVENT 0xff /* Mask for EVENT bits */
595
596 /*
597 * Event filters for PMUv2
598 */
599 #define ARMV7_EXCLUDE_PL1 (1 << 31)
600 #define ARMV7_EXCLUDE_USER (1 << 30)
601 #define ARMV7_INCLUDE_HYP (1 << 27)
602
603 static inline u32 armv7_pmnc_read(void)
604 {
605 u32 val;
606 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r"(val));
607 return val;
608 }
609
610 static inline void armv7_pmnc_write(u32 val)
611 {
612 val &= ARMV7_PMNC_MASK;
613 isb();
614 asm volatile("mcr p15, 0, %0, c9, c12, 0" : : "r"(val));
615 }
616
617 static inline int armv7_pmnc_has_overflowed(u32 pmnc)
618 {
619 return pmnc & ARMV7_OVERFLOWED_MASK;
620 }
621
622 static inline int armv7_pmnc_counter_valid(struct arm_pmu *cpu_pmu, int idx)
623 {
624 return idx >= ARMV7_IDX_CYCLE_COUNTER &&
625 idx <= ARMV7_IDX_COUNTER_LAST(cpu_pmu);
626 }
627
628 static inline int armv7_pmnc_counter_has_overflowed(u32 pmnc, int idx)
629 {
630 return pmnc & BIT(ARMV7_IDX_TO_COUNTER(idx));
631 }
632
633 static inline void armv7_pmnc_select_counter(int idx)
634 {
635 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
636 asm volatile("mcr p15, 0, %0, c9, c12, 5" : : "r" (counter));
637 isb();
638 }
639
640 static inline u32 armv7pmu_read_counter(struct perf_event *event)
641 {
642 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
643 struct hw_perf_event *hwc = &event->hw;
644 int idx = hwc->idx;
645 u32 value = 0;
646
647 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
648 pr_err("CPU%u reading wrong counter %d\n",
649 smp_processor_id(), idx);
650 } else if (idx == ARMV7_IDX_CYCLE_COUNTER) {
651 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (value));
652 } else {
653 armv7_pmnc_select_counter(idx);
654 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (value));
655 }
656
657 return value;
658 }
659
660 static inline void armv7pmu_write_counter(struct perf_event *event, u32 value)
661 {
662 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
663 struct hw_perf_event *hwc = &event->hw;
664 int idx = hwc->idx;
665
666 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
667 pr_err("CPU%u writing wrong counter %d\n",
668 smp_processor_id(), idx);
669 } else if (idx == ARMV7_IDX_CYCLE_COUNTER) {
670 asm volatile("mcr p15, 0, %0, c9, c13, 0" : : "r" (value));
671 } else {
672 armv7_pmnc_select_counter(idx);
673 asm volatile("mcr p15, 0, %0, c9, c13, 2" : : "r" (value));
674 }
675 }
676
677 static inline void armv7_pmnc_write_evtsel(int idx, u32 val)
678 {
679 armv7_pmnc_select_counter(idx);
680 val &= ARMV7_EVTYPE_MASK;
681 asm volatile("mcr p15, 0, %0, c9, c13, 1" : : "r" (val));
682 }
683
684 static inline void armv7_pmnc_enable_counter(int idx)
685 {
686 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
687 asm volatile("mcr p15, 0, %0, c9, c12, 1" : : "r" (BIT(counter)));
688 }
689
690 static inline void armv7_pmnc_disable_counter(int idx)
691 {
692 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
693 asm volatile("mcr p15, 0, %0, c9, c12, 2" : : "r" (BIT(counter)));
694 }
695
696 static inline void armv7_pmnc_enable_intens(int idx)
697 {
698 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
699 asm volatile("mcr p15, 0, %0, c9, c14, 1" : : "r" (BIT(counter)));
700 }
701
702 static inline void armv7_pmnc_disable_intens(int idx)
703 {
704 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
705 asm volatile("mcr p15, 0, %0, c9, c14, 2" : : "r" (BIT(counter)));
706 isb();
707 /* Clear the overflow flag in case an interrupt is pending. */
708 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (BIT(counter)));
709 isb();
710 }
711
712 static inline u32 armv7_pmnc_getreset_flags(void)
713 {
714 u32 val;
715
716 /* Read */
717 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val));
718
719 /* Write to clear flags */
720 val &= ARMV7_FLAG_MASK;
721 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (val));
722
723 return val;
724 }
725
726 #ifdef DEBUG
727 static void armv7_pmnc_dump_regs(struct arm_pmu *cpu_pmu)
728 {
729 u32 val;
730 unsigned int cnt;
731
732 pr_info("PMNC registers dump:\n");
733
734 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r" (val));
735 pr_info("PMNC =0x%08x\n", val);
736
737 asm volatile("mrc p15, 0, %0, c9, c12, 1" : "=r" (val));
738 pr_info("CNTENS=0x%08x\n", val);
739
740 asm volatile("mrc p15, 0, %0, c9, c14, 1" : "=r" (val));
741 pr_info("INTENS=0x%08x\n", val);
742
743 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val));
744 pr_info("FLAGS =0x%08x\n", val);
745
746 asm volatile("mrc p15, 0, %0, c9, c12, 5" : "=r" (val));
747 pr_info("SELECT=0x%08x\n", val);
748
749 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (val));
750 pr_info("CCNT =0x%08x\n", val);
751
752 for (cnt = ARMV7_IDX_COUNTER0;
753 cnt <= ARMV7_IDX_COUNTER_LAST(cpu_pmu); cnt++) {
754 armv7_pmnc_select_counter(cnt);
755 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (val));
756 pr_info("CNT[%d] count =0x%08x\n",
757 ARMV7_IDX_TO_COUNTER(cnt), val);
758 asm volatile("mrc p15, 0, %0, c9, c13, 1" : "=r" (val));
759 pr_info("CNT[%d] evtsel=0x%08x\n",
760 ARMV7_IDX_TO_COUNTER(cnt), val);
761 }
762 }
763 #endif
764
765 static void armv7pmu_enable_event(struct perf_event *event)
766 {
767 unsigned long flags;
768 struct hw_perf_event *hwc = &event->hw;
769 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
770 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
771 int idx = hwc->idx;
772
773 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
774 pr_err("CPU%u enabling wrong PMNC counter IRQ enable %d\n",
775 smp_processor_id(), idx);
776 return;
777 }
778
779 /*
780 * Enable counter and interrupt, and set the counter to count
781 * the event that we're interested in.
782 */
783 raw_spin_lock_irqsave(&events->pmu_lock, flags);
784
785 /*
786 * Disable counter
787 */
788 armv7_pmnc_disable_counter(idx);
789
790 /*
791 * Set event (if destined for PMNx counters)
792 * We only need to set the event for the cycle counter if we
793 * have the ability to perform event filtering.
794 */
795 if (cpu_pmu->set_event_filter || idx != ARMV7_IDX_CYCLE_COUNTER)
796 armv7_pmnc_write_evtsel(idx, hwc->config_base);
797
798 /*
799 * Enable interrupt for this counter
800 */
801 armv7_pmnc_enable_intens(idx);
802
803 /*
804 * Enable counter
805 */
806 armv7_pmnc_enable_counter(idx);
807
808 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
809 }
810
811 static void armv7pmu_disable_event(struct perf_event *event)
812 {
813 unsigned long flags;
814 struct hw_perf_event *hwc = &event->hw;
815 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
816 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
817 int idx = hwc->idx;
818
819 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
820 pr_err("CPU%u disabling wrong PMNC counter IRQ enable %d\n",
821 smp_processor_id(), idx);
822 return;
823 }
824
825 /*
826 * Disable counter and interrupt
827 */
828 raw_spin_lock_irqsave(&events->pmu_lock, flags);
829
830 /*
831 * Disable counter
832 */
833 armv7_pmnc_disable_counter(idx);
834
835 /*
836 * Disable interrupt for this counter
837 */
838 armv7_pmnc_disable_intens(idx);
839
840 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
841 }
842
843 static irqreturn_t armv7pmu_handle_irq(int irq_num, void *dev)
844 {
845 u32 pmnc;
846 struct perf_sample_data data;
847 struct arm_pmu *cpu_pmu = (struct arm_pmu *)dev;
848 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events);
849 struct pt_regs *regs;
850 int idx;
851
852 /*
853 * Get and reset the IRQ flags
854 */
855 pmnc = armv7_pmnc_getreset_flags();
856
857 /*
858 * Did an overflow occur?
859 */
860 if (!armv7_pmnc_has_overflowed(pmnc))
861 return IRQ_NONE;
862
863 /*
864 * Handle the counter(s) overflow(s)
865 */
866 regs = get_irq_regs();
867
868 for (idx = 0; idx < cpu_pmu->num_events; ++idx) {
869 struct perf_event *event = cpuc->events[idx];
870 struct hw_perf_event *hwc;
871
872 /* Ignore if we don't have an event. */
873 if (!event)
874 continue;
875
876 /*
877 * We have a single interrupt for all counters. Check that
878 * each counter has overflowed before we process it.
879 */
880 if (!armv7_pmnc_counter_has_overflowed(pmnc, idx))
881 continue;
882
883 hwc = &event->hw;
884 armpmu_event_update(event);
885 perf_sample_data_init(&data, 0, hwc->last_period);
886 if (!armpmu_event_set_period(event))
887 continue;
888
889 if (perf_event_overflow(event, &data, regs))
890 cpu_pmu->disable(event);
891 }
892
893 /*
894 * Handle the pending perf events.
895 *
896 * Note: this call *must* be run with interrupts disabled. For
897 * platforms that can have the PMU interrupts raised as an NMI, this
898 * will not work.
899 */
900 irq_work_run();
901
902 return IRQ_HANDLED;
903 }
904
905 static void armv7pmu_start(struct arm_pmu *cpu_pmu)
906 {
907 unsigned long flags;
908 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
909
910 raw_spin_lock_irqsave(&events->pmu_lock, flags);
911 /* Enable all counters */
912 armv7_pmnc_write(armv7_pmnc_read() | ARMV7_PMNC_E);
913 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
914 }
915
916 static void armv7pmu_stop(struct arm_pmu *cpu_pmu)
917 {
918 unsigned long flags;
919 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
920
921 raw_spin_lock_irqsave(&events->pmu_lock, flags);
922 /* Disable all counters */
923 armv7_pmnc_write(armv7_pmnc_read() & ~ARMV7_PMNC_E);
924 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
925 }
926
927 static int armv7pmu_get_event_idx(struct pmu_hw_events *cpuc,
928 struct perf_event *event)
929 {
930 int idx;
931 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
932 struct hw_perf_event *hwc = &event->hw;
933 unsigned long evtype = hwc->config_base & ARMV7_EVTYPE_EVENT;
934
935 /* Always place a cycle counter into the cycle counter. */
936 if (evtype == ARMV7_PERFCTR_CPU_CYCLES) {
937 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask))
938 return -EAGAIN;
939
940 return ARMV7_IDX_CYCLE_COUNTER;
941 }
942
943 /*
944 * For anything other than a cycle counter, try and use
945 * the events counters
946 */
947 for (idx = ARMV7_IDX_COUNTER0; idx < cpu_pmu->num_events; ++idx) {
948 if (!test_and_set_bit(idx, cpuc->used_mask))
949 return idx;
950 }
951
952 /* The counters are all in use. */
953 return -EAGAIN;
954 }
955
956 /*
957 * Add an event filter to a given event. This will only work for PMUv2 PMUs.
958 */
959 static int armv7pmu_set_event_filter(struct hw_perf_event *event,
960 struct perf_event_attr *attr)
961 {
962 unsigned long config_base = 0;
963
964 if (attr->exclude_idle)
965 return -EPERM;
966 if (attr->exclude_user)
967 config_base |= ARMV7_EXCLUDE_USER;
968 if (attr->exclude_kernel)
969 config_base |= ARMV7_EXCLUDE_PL1;
970 if (!attr->exclude_hv)
971 config_base |= ARMV7_INCLUDE_HYP;
972
973 /*
974 * Install the filter into config_base as this is used to
975 * construct the event type.
976 */
977 event->config_base = config_base;
978
979 return 0;
980 }
981
982 static void armv7pmu_reset(void *info)
983 {
984 struct arm_pmu *cpu_pmu = (struct arm_pmu *)info;
985 u32 idx, nb_cnt = cpu_pmu->num_events;
986
987 /* The counter and interrupt enable registers are unknown at reset. */
988 for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) {
989 armv7_pmnc_disable_counter(idx);
990 armv7_pmnc_disable_intens(idx);
991 }
992
993 /* Initialize & Reset PMNC: C and P bits */
994 armv7_pmnc_write(ARMV7_PMNC_P | ARMV7_PMNC_C);
995 }
996
997 static int armv7_a8_map_event(struct perf_event *event)
998 {
999 return armpmu_map_event(event, &armv7_a8_perf_map,
1000 &armv7_a8_perf_cache_map, 0xFF);
1001 }
1002
1003 static int armv7_a9_map_event(struct perf_event *event)
1004 {
1005 return armpmu_map_event(event, &armv7_a9_perf_map,
1006 &armv7_a9_perf_cache_map, 0xFF);
1007 }
1008
1009 static int armv7_a5_map_event(struct perf_event *event)
1010 {
1011 return armpmu_map_event(event, &armv7_a5_perf_map,
1012 &armv7_a5_perf_cache_map, 0xFF);
1013 }
1014
1015 static int armv7_a15_map_event(struct perf_event *event)
1016 {
1017 return armpmu_map_event(event, &armv7_a15_perf_map,
1018 &armv7_a15_perf_cache_map, 0xFF);
1019 }
1020
1021 static int armv7_a7_map_event(struct perf_event *event)
1022 {
1023 return armpmu_map_event(event, &armv7_a7_perf_map,
1024 &armv7_a7_perf_cache_map, 0xFF);
1025 }
1026
1027 static int armv7_a12_map_event(struct perf_event *event)
1028 {
1029 return armpmu_map_event(event, &armv7_a12_perf_map,
1030 &armv7_a12_perf_cache_map, 0xFF);
1031 }
1032
1033 static int krait_map_event(struct perf_event *event)
1034 {
1035 return armpmu_map_event(event, &krait_perf_map,
1036 &krait_perf_cache_map, 0xFFFFF);
1037 }
1038
1039 static int krait_map_event_no_branch(struct perf_event *event)
1040 {
1041 return armpmu_map_event(event, &krait_perf_map_no_branch,
1042 &krait_perf_cache_map, 0xFFFFF);
1043 }
1044
1045 static int scorpion_map_event(struct perf_event *event)
1046 {
1047 return armpmu_map_event(event, &scorpion_perf_map,
1048 &scorpion_perf_cache_map, 0xFFFFF);
1049 }
1050
1051 static void armv7pmu_init(struct arm_pmu *cpu_pmu)
1052 {
1053 cpu_pmu->handle_irq = armv7pmu_handle_irq;
1054 cpu_pmu->enable = armv7pmu_enable_event;
1055 cpu_pmu->disable = armv7pmu_disable_event;
1056 cpu_pmu->read_counter = armv7pmu_read_counter;
1057 cpu_pmu->write_counter = armv7pmu_write_counter;
1058 cpu_pmu->get_event_idx = armv7pmu_get_event_idx;
1059 cpu_pmu->start = armv7pmu_start;
1060 cpu_pmu->stop = armv7pmu_stop;
1061 cpu_pmu->reset = armv7pmu_reset;
1062 cpu_pmu->max_period = (1LLU << 32) - 1;
1063 };
1064
1065 static void armv7_read_num_pmnc_events(void *info)
1066 {
1067 int *nb_cnt = info;
1068
1069 /* Read the nb of CNTx counters supported from PMNC */
1070 *nb_cnt = (armv7_pmnc_read() >> ARMV7_PMNC_N_SHIFT) & ARMV7_PMNC_N_MASK;
1071
1072 /* Add the CPU cycles counter */
1073 *nb_cnt += 1;
1074 }
1075
1076 static int armv7_probe_num_events(struct arm_pmu *arm_pmu)
1077 {
1078 return smp_call_function_any(&arm_pmu->supported_cpus,
1079 armv7_read_num_pmnc_events,
1080 &arm_pmu->num_events, 1);
1081 }
1082
1083 static int armv7_a8_pmu_init(struct arm_pmu *cpu_pmu)
1084 {
1085 armv7pmu_init(cpu_pmu);
1086 cpu_pmu->name = "armv7_cortex_a8";
1087 cpu_pmu->map_event = armv7_a8_map_event;
1088 return armv7_probe_num_events(cpu_pmu);
1089 }
1090
1091 static int armv7_a9_pmu_init(struct arm_pmu *cpu_pmu)
1092 {
1093 armv7pmu_init(cpu_pmu);
1094 cpu_pmu->name = "armv7_cortex_a9";
1095 cpu_pmu->map_event = armv7_a9_map_event;
1096 return armv7_probe_num_events(cpu_pmu);
1097 }
1098
1099 static int armv7_a5_pmu_init(struct arm_pmu *cpu_pmu)
1100 {
1101 armv7pmu_init(cpu_pmu);
1102 cpu_pmu->name = "armv7_cortex_a5";
1103 cpu_pmu->map_event = armv7_a5_map_event;
1104 return armv7_probe_num_events(cpu_pmu);
1105 }
1106
1107 static int armv7_a15_pmu_init(struct arm_pmu *cpu_pmu)
1108 {
1109 armv7pmu_init(cpu_pmu);
1110 cpu_pmu->name = "armv7_cortex_a15";
1111 cpu_pmu->map_event = armv7_a15_map_event;
1112 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1113 return armv7_probe_num_events(cpu_pmu);
1114 }
1115
1116 static int armv7_a7_pmu_init(struct arm_pmu *cpu_pmu)
1117 {
1118 armv7pmu_init(cpu_pmu);
1119 cpu_pmu->name = "armv7_cortex_a7";
1120 cpu_pmu->map_event = armv7_a7_map_event;
1121 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1122 return armv7_probe_num_events(cpu_pmu);
1123 }
1124
1125 static int armv7_a12_pmu_init(struct arm_pmu *cpu_pmu)
1126 {
1127 armv7pmu_init(cpu_pmu);
1128 cpu_pmu->name = "armv7_cortex_a12";
1129 cpu_pmu->map_event = armv7_a12_map_event;
1130 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1131 return armv7_probe_num_events(cpu_pmu);
1132 }
1133
1134 static int armv7_a17_pmu_init(struct arm_pmu *cpu_pmu)
1135 {
1136 int ret = armv7_a12_pmu_init(cpu_pmu);
1137 cpu_pmu->name = "armv7_cortex_a17";
1138 return ret;
1139 }
1140
1141 /*
1142 * Krait Performance Monitor Region Event Selection Register (PMRESRn)
1143 *
1144 * 31 30 24 16 8 0
1145 * +--------------------------------+
1146 * PMRESR0 | EN | CC | CC | CC | CC | N = 1, R = 0
1147 * +--------------------------------+
1148 * PMRESR1 | EN | CC | CC | CC | CC | N = 1, R = 1
1149 * +--------------------------------+
1150 * PMRESR2 | EN | CC | CC | CC | CC | N = 1, R = 2
1151 * +--------------------------------+
1152 * VPMRESR0 | EN | CC | CC | CC | CC | N = 2, R = ?
1153 * +--------------------------------+
1154 * EN | G=3 | G=2 | G=1 | G=0
1155 *
1156 * Event Encoding:
1157 *
1158 * hwc->config_base = 0xNRCCG
1159 *
1160 * N = prefix, 1 for Krait CPU (PMRESRn), 2 for Venum VFP (VPMRESR)
1161 * R = region register
1162 * CC = class of events the group G is choosing from
1163 * G = group or particular event
1164 *
1165 * Example: 0x12021 is a Krait CPU event in PMRESR2's group 1 with code 2
1166 *
1167 * A region (R) corresponds to a piece of the CPU (execution unit, instruction
1168 * unit, etc.) while the event code (CC) corresponds to a particular class of
1169 * events (interrupts for example). An event code is broken down into
1170 * groups (G) that can be mapped into the PMU (irq, fiqs, and irq+fiqs for
1171 * example).
1172 */
1173
1174 #define KRAIT_EVENT (1 << 16)
1175 #define VENUM_EVENT (2 << 16)
1176 #define KRAIT_EVENT_MASK (KRAIT_EVENT | VENUM_EVENT)
1177 #define PMRESRn_EN BIT(31)
1178
1179 #define EVENT_REGION(event) (((event) >> 12) & 0xf) /* R */
1180 #define EVENT_GROUP(event) ((event) & 0xf) /* G */
1181 #define EVENT_CODE(event) (((event) >> 4) & 0xff) /* CC */
1182 #define EVENT_VENUM(event) (!!(event & VENUM_EVENT)) /* N=2 */
1183 #define EVENT_CPU(event) (!!(event & KRAIT_EVENT)) /* N=1 */
1184
1185 static u32 krait_read_pmresrn(int n)
1186 {
1187 u32 val;
1188
1189 switch (n) {
1190 case 0:
1191 asm volatile("mrc p15, 1, %0, c9, c15, 0" : "=r" (val));
1192 break;
1193 case 1:
1194 asm volatile("mrc p15, 1, %0, c9, c15, 1" : "=r" (val));
1195 break;
1196 case 2:
1197 asm volatile("mrc p15, 1, %0, c9, c15, 2" : "=r" (val));
1198 break;
1199 default:
1200 BUG(); /* Should be validated in krait_pmu_get_event_idx() */
1201 }
1202
1203 return val;
1204 }
1205
1206 static void krait_write_pmresrn(int n, u32 val)
1207 {
1208 switch (n) {
1209 case 0:
1210 asm volatile("mcr p15, 1, %0, c9, c15, 0" : : "r" (val));
1211 break;
1212 case 1:
1213 asm volatile("mcr p15, 1, %0, c9, c15, 1" : : "r" (val));
1214 break;
1215 case 2:
1216 asm volatile("mcr p15, 1, %0, c9, c15, 2" : : "r" (val));
1217 break;
1218 default:
1219 BUG(); /* Should be validated in krait_pmu_get_event_idx() */
1220 }
1221 }
1222
1223 static u32 venum_read_pmresr(void)
1224 {
1225 u32 val;
1226 asm volatile("mrc p10, 7, %0, c11, c0, 0" : "=r" (val));
1227 return val;
1228 }
1229
1230 static void venum_write_pmresr(u32 val)
1231 {
1232 asm volatile("mcr p10, 7, %0, c11, c0, 0" : : "r" (val));
1233 }
1234
1235 static void venum_pre_pmresr(u32 *venum_orig_val, u32 *fp_orig_val)
1236 {
1237 u32 venum_new_val;
1238 u32 fp_new_val;
1239
1240 BUG_ON(preemptible());
1241 /* CPACR Enable CP10 and CP11 access */
1242 *venum_orig_val = get_copro_access();
1243 venum_new_val = *venum_orig_val | CPACC_SVC(10) | CPACC_SVC(11);
1244 set_copro_access(venum_new_val);
1245
1246 /* Enable FPEXC */
1247 *fp_orig_val = fmrx(FPEXC);
1248 fp_new_val = *fp_orig_val | FPEXC_EN;
1249 fmxr(FPEXC, fp_new_val);
1250 }
1251
1252 static void venum_post_pmresr(u32 venum_orig_val, u32 fp_orig_val)
1253 {
1254 BUG_ON(preemptible());
1255 /* Restore FPEXC */
1256 fmxr(FPEXC, fp_orig_val);
1257 isb();
1258 /* Restore CPACR */
1259 set_copro_access(venum_orig_val);
1260 }
1261
1262 static u32 krait_get_pmresrn_event(unsigned int region)
1263 {
1264 static const u32 pmresrn_table[] = { KRAIT_PMRESR0_GROUP0,
1265 KRAIT_PMRESR1_GROUP0,
1266 KRAIT_PMRESR2_GROUP0 };
1267 return pmresrn_table[region];
1268 }
1269
1270 static void krait_evt_setup(int idx, u32 config_base)
1271 {
1272 u32 val;
1273 u32 mask;
1274 u32 vval, fval;
1275 unsigned int region = EVENT_REGION(config_base);
1276 unsigned int group = EVENT_GROUP(config_base);
1277 unsigned int code = EVENT_CODE(config_base);
1278 unsigned int group_shift;
1279 bool venum_event = EVENT_VENUM(config_base);
1280
1281 group_shift = group * 8;
1282 mask = 0xff << group_shift;
1283
1284 /* Configure evtsel for the region and group */
1285 if (venum_event)
1286 val = KRAIT_VPMRESR0_GROUP0;
1287 else
1288 val = krait_get_pmresrn_event(region);
1289 val += group;
1290 /* Mix in mode-exclusion bits */
1291 val |= config_base & (ARMV7_EXCLUDE_USER | ARMV7_EXCLUDE_PL1);
1292 armv7_pmnc_write_evtsel(idx, val);
1293
1294 if (venum_event) {
1295 venum_pre_pmresr(&vval, &fval);
1296 val = venum_read_pmresr();
1297 val &= ~mask;
1298 val |= code << group_shift;
1299 val |= PMRESRn_EN;
1300 venum_write_pmresr(val);
1301 venum_post_pmresr(vval, fval);
1302 } else {
1303 val = krait_read_pmresrn(region);
1304 val &= ~mask;
1305 val |= code << group_shift;
1306 val |= PMRESRn_EN;
1307 krait_write_pmresrn(region, val);
1308 }
1309 }
1310
1311 static u32 clear_pmresrn_group(u32 val, int group)
1312 {
1313 u32 mask;
1314 int group_shift;
1315
1316 group_shift = group * 8;
1317 mask = 0xff << group_shift;
1318 val &= ~mask;
1319
1320 /* Don't clear enable bit if entire region isn't disabled */
1321 if (val & ~PMRESRn_EN)
1322 return val |= PMRESRn_EN;
1323
1324 return 0;
1325 }
1326
1327 static void krait_clearpmu(u32 config_base)
1328 {
1329 u32 val;
1330 u32 vval, fval;
1331 unsigned int region = EVENT_REGION(config_base);
1332 unsigned int group = EVENT_GROUP(config_base);
1333 bool venum_event = EVENT_VENUM(config_base);
1334
1335 if (venum_event) {
1336 venum_pre_pmresr(&vval, &fval);
1337 val = venum_read_pmresr();
1338 val = clear_pmresrn_group(val, group);
1339 venum_write_pmresr(val);
1340 venum_post_pmresr(vval, fval);
1341 } else {
1342 val = krait_read_pmresrn(region);
1343 val = clear_pmresrn_group(val, group);
1344 krait_write_pmresrn(region, val);
1345 }
1346 }
1347
1348 static void krait_pmu_disable_event(struct perf_event *event)
1349 {
1350 unsigned long flags;
1351 struct hw_perf_event *hwc = &event->hw;
1352 int idx = hwc->idx;
1353 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1354 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
1355
1356 /* Disable counter and interrupt */
1357 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1358
1359 /* Disable counter */
1360 armv7_pmnc_disable_counter(idx);
1361
1362 /*
1363 * Clear pmresr code (if destined for PMNx counters)
1364 */
1365 if (hwc->config_base & KRAIT_EVENT_MASK)
1366 krait_clearpmu(hwc->config_base);
1367
1368 /* Disable interrupt for this counter */
1369 armv7_pmnc_disable_intens(idx);
1370
1371 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1372 }
1373
1374 static void krait_pmu_enable_event(struct perf_event *event)
1375 {
1376 unsigned long flags;
1377 struct hw_perf_event *hwc = &event->hw;
1378 int idx = hwc->idx;
1379 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1380 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
1381
1382 /*
1383 * Enable counter and interrupt, and set the counter to count
1384 * the event that we're interested in.
1385 */
1386 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1387
1388 /* Disable counter */
1389 armv7_pmnc_disable_counter(idx);
1390
1391 /*
1392 * Set event (if destined for PMNx counters)
1393 * We set the event for the cycle counter because we
1394 * have the ability to perform event filtering.
1395 */
1396 if (hwc->config_base & KRAIT_EVENT_MASK)
1397 krait_evt_setup(idx, hwc->config_base);
1398 else
1399 armv7_pmnc_write_evtsel(idx, hwc->config_base);
1400
1401 /* Enable interrupt for this counter */
1402 armv7_pmnc_enable_intens(idx);
1403
1404 /* Enable counter */
1405 armv7_pmnc_enable_counter(idx);
1406
1407 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1408 }
1409
1410 static void krait_pmu_reset(void *info)
1411 {
1412 u32 vval, fval;
1413 struct arm_pmu *cpu_pmu = info;
1414 u32 idx, nb_cnt = cpu_pmu->num_events;
1415
1416 armv7pmu_reset(info);
1417
1418 /* Clear all pmresrs */
1419 krait_write_pmresrn(0, 0);
1420 krait_write_pmresrn(1, 0);
1421 krait_write_pmresrn(2, 0);
1422
1423 venum_pre_pmresr(&vval, &fval);
1424 venum_write_pmresr(0);
1425 venum_post_pmresr(vval, fval);
1426
1427 /* Reset PMxEVNCTCR to sane default */
1428 for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) {
1429 armv7_pmnc_select_counter(idx);
1430 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1431 }
1432
1433 }
1434
1435 static int krait_event_to_bit(struct perf_event *event, unsigned int region,
1436 unsigned int group)
1437 {
1438 int bit;
1439 struct hw_perf_event *hwc = &event->hw;
1440 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1441
1442 if (hwc->config_base & VENUM_EVENT)
1443 bit = KRAIT_VPMRESR0_GROUP0;
1444 else
1445 bit = krait_get_pmresrn_event(region);
1446 bit -= krait_get_pmresrn_event(0);
1447 bit += group;
1448 /*
1449 * Lower bits are reserved for use by the counters (see
1450 * armv7pmu_get_event_idx() for more info)
1451 */
1452 bit += ARMV7_IDX_COUNTER_LAST(cpu_pmu) + 1;
1453
1454 return bit;
1455 }
1456
1457 /*
1458 * We check for column exclusion constraints here.
1459 * Two events cant use the same group within a pmresr register.
1460 */
1461 static int krait_pmu_get_event_idx(struct pmu_hw_events *cpuc,
1462 struct perf_event *event)
1463 {
1464 int idx;
1465 int bit = -1;
1466 struct hw_perf_event *hwc = &event->hw;
1467 unsigned int region = EVENT_REGION(hwc->config_base);
1468 unsigned int code = EVENT_CODE(hwc->config_base);
1469 unsigned int group = EVENT_GROUP(hwc->config_base);
1470 bool venum_event = EVENT_VENUM(hwc->config_base);
1471 bool krait_event = EVENT_CPU(hwc->config_base);
1472
1473 if (venum_event || krait_event) {
1474 /* Ignore invalid events */
1475 if (group > 3 || region > 2)
1476 return -EINVAL;
1477 if (venum_event && (code & 0xe0))
1478 return -EINVAL;
1479
1480 bit = krait_event_to_bit(event, region, group);
1481 if (test_and_set_bit(bit, cpuc->used_mask))
1482 return -EAGAIN;
1483 }
1484
1485 idx = armv7pmu_get_event_idx(cpuc, event);
1486 if (idx < 0 && bit >= 0)
1487 clear_bit(bit, cpuc->used_mask);
1488
1489 return idx;
1490 }
1491
1492 static void krait_pmu_clear_event_idx(struct pmu_hw_events *cpuc,
1493 struct perf_event *event)
1494 {
1495 int bit;
1496 struct hw_perf_event *hwc = &event->hw;
1497 unsigned int region = EVENT_REGION(hwc->config_base);
1498 unsigned int group = EVENT_GROUP(hwc->config_base);
1499 bool venum_event = EVENT_VENUM(hwc->config_base);
1500 bool krait_event = EVENT_CPU(hwc->config_base);
1501
1502 if (venum_event || krait_event) {
1503 bit = krait_event_to_bit(event, region, group);
1504 clear_bit(bit, cpuc->used_mask);
1505 }
1506 }
1507
1508 static int krait_pmu_init(struct arm_pmu *cpu_pmu)
1509 {
1510 armv7pmu_init(cpu_pmu);
1511 cpu_pmu->name = "armv7_krait";
1512 /* Some early versions of Krait don't support PC write events */
1513 if (of_property_read_bool(cpu_pmu->plat_device->dev.of_node,
1514 "qcom,no-pc-write"))
1515 cpu_pmu->map_event = krait_map_event_no_branch;
1516 else
1517 cpu_pmu->map_event = krait_map_event;
1518 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1519 cpu_pmu->reset = krait_pmu_reset;
1520 cpu_pmu->enable = krait_pmu_enable_event;
1521 cpu_pmu->disable = krait_pmu_disable_event;
1522 cpu_pmu->get_event_idx = krait_pmu_get_event_idx;
1523 cpu_pmu->clear_event_idx = krait_pmu_clear_event_idx;
1524 return armv7_probe_num_events(cpu_pmu);
1525 }
1526
1527 /*
1528 * Scorpion Local Performance Monitor Register (LPMn)
1529 *
1530 * 31 30 24 16 8 0
1531 * +--------------------------------+
1532 * LPM0 | EN | CC | CC | CC | CC | N = 1, R = 0
1533 * +--------------------------------+
1534 * LPM1 | EN | CC | CC | CC | CC | N = 1, R = 1
1535 * +--------------------------------+
1536 * LPM2 | EN | CC | CC | CC | CC | N = 1, R = 2
1537 * +--------------------------------+
1538 * L2LPM | EN | CC | CC | CC | CC | N = 1, R = 3
1539 * +--------------------------------+
1540 * VLPM | EN | CC | CC | CC | CC | N = 2, R = ?
1541 * +--------------------------------+
1542 * EN | G=3 | G=2 | G=1 | G=0
1543 *
1544 *
1545 * Event Encoding:
1546 *
1547 * hwc->config_base = 0xNRCCG
1548 *
1549 * N = prefix, 1 for Scorpion CPU (LPMn/L2LPM), 2 for Venum VFP (VLPM)
1550 * R = region register
1551 * CC = class of events the group G is choosing from
1552 * G = group or particular event
1553 *
1554 * Example: 0x12021 is a Scorpion CPU event in LPM2's group 1 with code 2
1555 *
1556 * A region (R) corresponds to a piece of the CPU (execution unit, instruction
1557 * unit, etc.) while the event code (CC) corresponds to a particular class of
1558 * events (interrupts for example). An event code is broken down into
1559 * groups (G) that can be mapped into the PMU (irq, fiqs, and irq+fiqs for
1560 * example).
1561 */
1562
1563 static u32 scorpion_read_pmresrn(int n)
1564 {
1565 u32 val;
1566
1567 switch (n) {
1568 case 0:
1569 asm volatile("mrc p15, 0, %0, c15, c0, 0" : "=r" (val));
1570 break;
1571 case 1:
1572 asm volatile("mrc p15, 1, %0, c15, c0, 0" : "=r" (val));
1573 break;
1574 case 2:
1575 asm volatile("mrc p15, 2, %0, c15, c0, 0" : "=r" (val));
1576 break;
1577 case 3:
1578 asm volatile("mrc p15, 3, %0, c15, c2, 0" : "=r" (val));
1579 break;
1580 default:
1581 BUG(); /* Should be validated in scorpion_pmu_get_event_idx() */
1582 }
1583
1584 return val;
1585 }
1586
1587 static void scorpion_write_pmresrn(int n, u32 val)
1588 {
1589 switch (n) {
1590 case 0:
1591 asm volatile("mcr p15, 0, %0, c15, c0, 0" : : "r" (val));
1592 break;
1593 case 1:
1594 asm volatile("mcr p15, 1, %0, c15, c0, 0" : : "r" (val));
1595 break;
1596 case 2:
1597 asm volatile("mcr p15, 2, %0, c15, c0, 0" : : "r" (val));
1598 break;
1599 case 3:
1600 asm volatile("mcr p15, 3, %0, c15, c2, 0" : : "r" (val));
1601 break;
1602 default:
1603 BUG(); /* Should be validated in scorpion_pmu_get_event_idx() */
1604 }
1605 }
1606
1607 static u32 scorpion_get_pmresrn_event(unsigned int region)
1608 {
1609 static const u32 pmresrn_table[] = { SCORPION_LPM0_GROUP0,
1610 SCORPION_LPM1_GROUP0,
1611 SCORPION_LPM2_GROUP0,
1612 SCORPION_L2LPM_GROUP0 };
1613 return pmresrn_table[region];
1614 }
1615
1616 static void scorpion_evt_setup(int idx, u32 config_base)
1617 {
1618 u32 val;
1619 u32 mask;
1620 u32 vval, fval;
1621 unsigned int region = EVENT_REGION(config_base);
1622 unsigned int group = EVENT_GROUP(config_base);
1623 unsigned int code = EVENT_CODE(config_base);
1624 unsigned int group_shift;
1625 bool venum_event = EVENT_VENUM(config_base);
1626
1627 group_shift = group * 8;
1628 mask = 0xff << group_shift;
1629
1630 /* Configure evtsel for the region and group */
1631 if (venum_event)
1632 val = SCORPION_VLPM_GROUP0;
1633 else
1634 val = scorpion_get_pmresrn_event(region);
1635 val += group;
1636 /* Mix in mode-exclusion bits */
1637 val |= config_base & (ARMV7_EXCLUDE_USER | ARMV7_EXCLUDE_PL1);
1638 armv7_pmnc_write_evtsel(idx, val);
1639
1640 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1641
1642 if (venum_event) {
1643 venum_pre_pmresr(&vval, &fval);
1644 val = venum_read_pmresr();
1645 val &= ~mask;
1646 val |= code << group_shift;
1647 val |= PMRESRn_EN;
1648 venum_write_pmresr(val);
1649 venum_post_pmresr(vval, fval);
1650 } else {
1651 val = scorpion_read_pmresrn(region);
1652 val &= ~mask;
1653 val |= code << group_shift;
1654 val |= PMRESRn_EN;
1655 scorpion_write_pmresrn(region, val);
1656 }
1657 }
1658
1659 static void scorpion_clearpmu(u32 config_base)
1660 {
1661 u32 val;
1662 u32 vval, fval;
1663 unsigned int region = EVENT_REGION(config_base);
1664 unsigned int group = EVENT_GROUP(config_base);
1665 bool venum_event = EVENT_VENUM(config_base);
1666
1667 if (venum_event) {
1668 venum_pre_pmresr(&vval, &fval);
1669 val = venum_read_pmresr();
1670 val = clear_pmresrn_group(val, group);
1671 venum_write_pmresr(val);
1672 venum_post_pmresr(vval, fval);
1673 } else {
1674 val = scorpion_read_pmresrn(region);
1675 val = clear_pmresrn_group(val, group);
1676 scorpion_write_pmresrn(region, val);
1677 }
1678 }
1679
1680 static void scorpion_pmu_disable_event(struct perf_event *event)
1681 {
1682 unsigned long flags;
1683 struct hw_perf_event *hwc = &event->hw;
1684 int idx = hwc->idx;
1685 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1686 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
1687
1688 /* Disable counter and interrupt */
1689 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1690
1691 /* Disable counter */
1692 armv7_pmnc_disable_counter(idx);
1693
1694 /*
1695 * Clear pmresr code (if destined for PMNx counters)
1696 */
1697 if (hwc->config_base & KRAIT_EVENT_MASK)
1698 scorpion_clearpmu(hwc->config_base);
1699
1700 /* Disable interrupt for this counter */
1701 armv7_pmnc_disable_intens(idx);
1702
1703 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1704 }
1705
1706 static void scorpion_pmu_enable_event(struct perf_event *event)
1707 {
1708 unsigned long flags;
1709 struct hw_perf_event *hwc = &event->hw;
1710 int idx = hwc->idx;
1711 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1712 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events);
1713
1714 /*
1715 * Enable counter and interrupt, and set the counter to count
1716 * the event that we're interested in.
1717 */
1718 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1719
1720 /* Disable counter */
1721 armv7_pmnc_disable_counter(idx);
1722
1723 /*
1724 * Set event (if destined for PMNx counters)
1725 * We don't set the event for the cycle counter because we
1726 * don't have the ability to perform event filtering.
1727 */
1728 if (hwc->config_base & KRAIT_EVENT_MASK)
1729 scorpion_evt_setup(idx, hwc->config_base);
1730 else if (idx != ARMV7_IDX_CYCLE_COUNTER)
1731 armv7_pmnc_write_evtsel(idx, hwc->config_base);
1732
1733 /* Enable interrupt for this counter */
1734 armv7_pmnc_enable_intens(idx);
1735
1736 /* Enable counter */
1737 armv7_pmnc_enable_counter(idx);
1738
1739 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1740 }
1741
1742 static void scorpion_pmu_reset(void *info)
1743 {
1744 u32 vval, fval;
1745 struct arm_pmu *cpu_pmu = info;
1746 u32 idx, nb_cnt = cpu_pmu->num_events;
1747
1748 armv7pmu_reset(info);
1749
1750 /* Clear all pmresrs */
1751 scorpion_write_pmresrn(0, 0);
1752 scorpion_write_pmresrn(1, 0);
1753 scorpion_write_pmresrn(2, 0);
1754 scorpion_write_pmresrn(3, 0);
1755
1756 venum_pre_pmresr(&vval, &fval);
1757 venum_write_pmresr(0);
1758 venum_post_pmresr(vval, fval);
1759
1760 /* Reset PMxEVNCTCR to sane default */
1761 for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) {
1762 armv7_pmnc_select_counter(idx);
1763 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1764 }
1765 }
1766
1767 static int scorpion_event_to_bit(struct perf_event *event, unsigned int region,
1768 unsigned int group)
1769 {
1770 int bit;
1771 struct hw_perf_event *hwc = &event->hw;
1772 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1773
1774 if (hwc->config_base & VENUM_EVENT)
1775 bit = SCORPION_VLPM_GROUP0;
1776 else
1777 bit = scorpion_get_pmresrn_event(region);
1778 bit -= scorpion_get_pmresrn_event(0);
1779 bit += group;
1780 /*
1781 * Lower bits are reserved for use by the counters (see
1782 * armv7pmu_get_event_idx() for more info)
1783 */
1784 bit += ARMV7_IDX_COUNTER_LAST(cpu_pmu) + 1;
1785
1786 return bit;
1787 }
1788
1789 /*
1790 * We check for column exclusion constraints here.
1791 * Two events cant use the same group within a pmresr register.
1792 */
1793 static int scorpion_pmu_get_event_idx(struct pmu_hw_events *cpuc,
1794 struct perf_event *event)
1795 {
1796 int idx;
1797 int bit = -1;
1798 struct hw_perf_event *hwc = &event->hw;
1799 unsigned int region = EVENT_REGION(hwc->config_base);
1800 unsigned int group = EVENT_GROUP(hwc->config_base);
1801 bool venum_event = EVENT_VENUM(hwc->config_base);
1802 bool scorpion_event = EVENT_CPU(hwc->config_base);
1803
1804 if (venum_event || scorpion_event) {
1805 /* Ignore invalid events */
1806 if (group > 3 || region > 3)
1807 return -EINVAL;
1808
1809 bit = scorpion_event_to_bit(event, region, group);
1810 if (test_and_set_bit(bit, cpuc->used_mask))
1811 return -EAGAIN;
1812 }
1813
1814 idx = armv7pmu_get_event_idx(cpuc, event);
1815 if (idx < 0 && bit >= 0)
1816 clear_bit(bit, cpuc->used_mask);
1817
1818 return idx;
1819 }
1820
1821 static void scorpion_pmu_clear_event_idx(struct pmu_hw_events *cpuc,
1822 struct perf_event *event)
1823 {
1824 int bit;
1825 struct hw_perf_event *hwc = &event->hw;
1826 unsigned int region = EVENT_REGION(hwc->config_base);
1827 unsigned int group = EVENT_GROUP(hwc->config_base);
1828 bool venum_event = EVENT_VENUM(hwc->config_base);
1829 bool scorpion_event = EVENT_CPU(hwc->config_base);
1830
1831 if (venum_event || scorpion_event) {
1832 bit = scorpion_event_to_bit(event, region, group);
1833 clear_bit(bit, cpuc->used_mask);
1834 }
1835 }
1836
1837 static int scorpion_pmu_init(struct arm_pmu *cpu_pmu)
1838 {
1839 armv7pmu_init(cpu_pmu);
1840 cpu_pmu->name = "armv7_scorpion";
1841 cpu_pmu->map_event = scorpion_map_event;
1842 cpu_pmu->reset = scorpion_pmu_reset;
1843 cpu_pmu->enable = scorpion_pmu_enable_event;
1844 cpu_pmu->disable = scorpion_pmu_disable_event;
1845 cpu_pmu->get_event_idx = scorpion_pmu_get_event_idx;
1846 cpu_pmu->clear_event_idx = scorpion_pmu_clear_event_idx;
1847 return armv7_probe_num_events(cpu_pmu);
1848 }
1849
1850 static int scorpion_mp_pmu_init(struct arm_pmu *cpu_pmu)
1851 {
1852 armv7pmu_init(cpu_pmu);
1853 cpu_pmu->name = "armv7_scorpion_mp";
1854 cpu_pmu->map_event = scorpion_map_event;
1855 cpu_pmu->reset = scorpion_pmu_reset;
1856 cpu_pmu->enable = scorpion_pmu_enable_event;
1857 cpu_pmu->disable = scorpion_pmu_disable_event;
1858 cpu_pmu->get_event_idx = scorpion_pmu_get_event_idx;
1859 cpu_pmu->clear_event_idx = scorpion_pmu_clear_event_idx;
1860 return armv7_probe_num_events(cpu_pmu);
1861 }
1862
1863 static const struct of_device_id armv7_pmu_of_device_ids[] = {
1864 {.compatible = "arm,cortex-a17-pmu", .data = armv7_a17_pmu_init},
1865 {.compatible = "arm,cortex-a15-pmu", .data = armv7_a15_pmu_init},
1866 {.compatible = "arm,cortex-a12-pmu", .data = armv7_a12_pmu_init},
1867 {.compatible = "arm,cortex-a9-pmu", .data = armv7_a9_pmu_init},
1868 {.compatible = "arm,cortex-a8-pmu", .data = armv7_a8_pmu_init},
1869 {.compatible = "arm,cortex-a7-pmu", .data = armv7_a7_pmu_init},
1870 {.compatible = "arm,cortex-a5-pmu", .data = armv7_a5_pmu_init},
1871 {.compatible = "qcom,krait-pmu", .data = krait_pmu_init},
1872 {.compatible = "qcom,scorpion-pmu", .data = scorpion_pmu_init},
1873 {.compatible = "qcom,scorpion-mp-pmu", .data = scorpion_mp_pmu_init},
1874 {},
1875 };
1876
1877 static const struct pmu_probe_info armv7_pmu_probe_table[] = {
1878 ARM_PMU_PROBE(ARM_CPU_PART_CORTEX_A8, armv7_a8_pmu_init),
1879 ARM_PMU_PROBE(ARM_CPU_PART_CORTEX_A9, armv7_a9_pmu_init),
1880 { /* sentinel value */ }
1881 };
1882
1883
1884 static int armv7_pmu_device_probe(struct platform_device *pdev)
1885 {
1886 return arm_pmu_device_probe(pdev, armv7_pmu_of_device_ids,
1887 armv7_pmu_probe_table);
1888 }
1889
1890 static struct platform_driver armv7_pmu_driver = {
1891 .driver = {
1892 .name = "armv7-pmu",
1893 .of_match_table = armv7_pmu_of_device_ids,
1894 },
1895 .probe = armv7_pmu_device_probe,
1896 };
1897
1898 static int __init register_armv7_pmu_driver(void)
1899 {
1900 return platform_driver_register(&armv7_pmu_driver);
1901 }
1902 device_initcall(register_armv7_pmu_driver);
1903 #endif /* CONFIG_CPU_V7 */
This page took 0.073984 seconds and 5 git commands to generate.