2 * ARMv7 Cortex-A8 and Cortex-A9 Performance Events handling code.
4 * ARMv7 support: Jean Pihet <jpihet@mvista.com>
5 * 2010 (c) MontaVista Software, LLC.
7 * Copied from ARMv6 code, with the low level code inspired
8 * by the ARMv7 Oprofile code.
10 * Cortex-A8 has up to 4 configurable performance counters and
11 * a single cycle counter.
12 * Cortex-A9 has up to 31 configurable performance counters and
13 * a single cycle counter.
15 * All counters can be enabled/disabled and IRQ masked separately. The cycle
16 * counter and all 4 performance counters together can be reset separately.
23 #include "../vfp/vfpinstr.h"
26 * Common ARMv7 event types
28 * Note: An implementation may not be able to count all of these events
29 * but the encodings are considered to be `reserved' in the case that
30 * they are not available.
32 enum armv7_perf_types
{
33 ARMV7_PERFCTR_PMNC_SW_INCR
= 0x00,
34 ARMV7_PERFCTR_L1_ICACHE_REFILL
= 0x01,
35 ARMV7_PERFCTR_ITLB_REFILL
= 0x02,
36 ARMV7_PERFCTR_L1_DCACHE_REFILL
= 0x03,
37 ARMV7_PERFCTR_L1_DCACHE_ACCESS
= 0x04,
38 ARMV7_PERFCTR_DTLB_REFILL
= 0x05,
39 ARMV7_PERFCTR_MEM_READ
= 0x06,
40 ARMV7_PERFCTR_MEM_WRITE
= 0x07,
41 ARMV7_PERFCTR_INSTR_EXECUTED
= 0x08,
42 ARMV7_PERFCTR_EXC_TAKEN
= 0x09,
43 ARMV7_PERFCTR_EXC_EXECUTED
= 0x0A,
44 ARMV7_PERFCTR_CID_WRITE
= 0x0B,
47 * ARMV7_PERFCTR_PC_WRITE is equivalent to HW_BRANCH_INSTRUCTIONS.
49 * - all (taken) branch instructions,
50 * - instructions that explicitly write the PC,
51 * - exception generating instructions.
53 ARMV7_PERFCTR_PC_WRITE
= 0x0C,
54 ARMV7_PERFCTR_PC_IMM_BRANCH
= 0x0D,
55 ARMV7_PERFCTR_PC_PROC_RETURN
= 0x0E,
56 ARMV7_PERFCTR_MEM_UNALIGNED_ACCESS
= 0x0F,
57 ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
= 0x10,
58 ARMV7_PERFCTR_CLOCK_CYCLES
= 0x11,
59 ARMV7_PERFCTR_PC_BRANCH_PRED
= 0x12,
61 /* These events are defined by the PMUv2 supplement (ARM DDI 0457A). */
62 ARMV7_PERFCTR_MEM_ACCESS
= 0x13,
63 ARMV7_PERFCTR_L1_ICACHE_ACCESS
= 0x14,
64 ARMV7_PERFCTR_L1_DCACHE_WB
= 0x15,
65 ARMV7_PERFCTR_L2_CACHE_ACCESS
= 0x16,
66 ARMV7_PERFCTR_L2_CACHE_REFILL
= 0x17,
67 ARMV7_PERFCTR_L2_CACHE_WB
= 0x18,
68 ARMV7_PERFCTR_BUS_ACCESS
= 0x19,
69 ARMV7_PERFCTR_MEM_ERROR
= 0x1A,
70 ARMV7_PERFCTR_INSTR_SPEC
= 0x1B,
71 ARMV7_PERFCTR_TTBR_WRITE
= 0x1C,
72 ARMV7_PERFCTR_BUS_CYCLES
= 0x1D,
74 ARMV7_PERFCTR_CPU_CYCLES
= 0xFF
77 /* ARMv7 Cortex-A8 specific event types */
78 enum armv7_a8_perf_types
{
79 ARMV7_A8_PERFCTR_L2_CACHE_ACCESS
= 0x43,
80 ARMV7_A8_PERFCTR_L2_CACHE_REFILL
= 0x44,
81 ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS
= 0x50,
82 ARMV7_A8_PERFCTR_STALL_ISIDE
= 0x56,
85 /* ARMv7 Cortex-A9 specific event types */
86 enum armv7_a9_perf_types
{
87 ARMV7_A9_PERFCTR_INSTR_CORE_RENAME
= 0x68,
88 ARMV7_A9_PERFCTR_STALL_ICACHE
= 0x60,
89 ARMV7_A9_PERFCTR_STALL_DISPATCH
= 0x66,
92 /* ARMv7 Cortex-A5 specific event types */
93 enum armv7_a5_perf_types
{
94 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL
= 0xc2,
95 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP
= 0xc3,
98 /* ARMv7 Cortex-A15 specific event types */
99 enum armv7_a15_perf_types
{
100 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ
= 0x40,
101 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE
= 0x41,
102 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ
= 0x42,
103 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE
= 0x43,
105 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ
= 0x4C,
106 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE
= 0x4D,
108 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ
= 0x50,
109 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE
= 0x51,
110 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ
= 0x52,
111 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE
= 0x53,
113 ARMV7_A15_PERFCTR_PC_WRITE_SPEC
= 0x76,
116 /* ARMv7 Cortex-A12 specific event types */
117 enum armv7_a12_perf_types
{
118 ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_READ
= 0x40,
119 ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_WRITE
= 0x41,
121 ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_READ
= 0x50,
122 ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_WRITE
= 0x51,
124 ARMV7_A12_PERFCTR_PC_WRITE_SPEC
= 0x76,
126 ARMV7_A12_PERFCTR_PF_TLB_REFILL
= 0xe7,
129 /* ARMv7 Krait specific event types */
130 enum krait_perf_types
{
131 KRAIT_PMRESR0_GROUP0
= 0xcc,
132 KRAIT_PMRESR1_GROUP0
= 0xd0,
133 KRAIT_PMRESR2_GROUP0
= 0xd4,
134 KRAIT_VPMRESR0_GROUP0
= 0xd8,
136 KRAIT_PERFCTR_L1_ICACHE_ACCESS
= 0x10011,
137 KRAIT_PERFCTR_L1_ICACHE_MISS
= 0x10010,
139 KRAIT_PERFCTR_L1_ITLB_ACCESS
= 0x12222,
140 KRAIT_PERFCTR_L1_DTLB_ACCESS
= 0x12210,
143 /* ARMv7 Scorpion specific event types */
144 enum scorpion_perf_types
{
145 SCORPION_LPM0_GROUP0
= 0x4c,
146 SCORPION_LPM1_GROUP0
= 0x50,
147 SCORPION_LPM2_GROUP0
= 0x54,
148 SCORPION_L2LPM_GROUP0
= 0x58,
149 SCORPION_VLPM_GROUP0
= 0x5c,
151 SCORPION_ICACHE_ACCESS
= 0x10053,
152 SCORPION_ICACHE_MISS
= 0x10052,
154 SCORPION_DTLB_ACCESS
= 0x12013,
155 SCORPION_DTLB_MISS
= 0x12012,
157 SCORPION_ITLB_MISS
= 0x12021,
161 * Cortex-A8 HW events mapping
163 * The hardware events that we support. We do support cache operations but
164 * we have harvard caches and no way to combine instruction and data
165 * accesses/misses in hardware.
167 static const unsigned armv7_a8_perf_map
[PERF_COUNT_HW_MAX
] = {
168 PERF_MAP_ALL_UNSUPPORTED
,
169 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
170 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
171 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
172 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
173 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
174 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
175 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND
] = ARMV7_A8_PERFCTR_STALL_ISIDE
,
178 static const unsigned armv7_a8_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
179 [PERF_COUNT_HW_CACHE_OP_MAX
]
180 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
181 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
184 * The performance counters don't differentiate between read and write
185 * accesses/misses so this isn't strictly correct, but it's the best we
186 * can do. Writes and reads get combined.
188 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
189 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
190 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
191 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
193 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS
,
194 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
196 [C(LL
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS
,
197 [C(LL
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL
,
198 [C(LL
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS
,
199 [C(LL
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL
,
201 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
202 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
204 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
205 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
207 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
208 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
209 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
210 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
214 * Cortex-A9 HW events mapping
216 static const unsigned armv7_a9_perf_map
[PERF_COUNT_HW_MAX
] = {
217 PERF_MAP_ALL_UNSUPPORTED
,
218 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
219 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_A9_PERFCTR_INSTR_CORE_RENAME
,
220 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
221 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
222 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
223 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
224 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND
] = ARMV7_A9_PERFCTR_STALL_ICACHE
,
225 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND
] = ARMV7_A9_PERFCTR_STALL_DISPATCH
,
228 static const unsigned armv7_a9_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
229 [PERF_COUNT_HW_CACHE_OP_MAX
]
230 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
231 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
234 * The performance counters don't differentiate between read and write
235 * accesses/misses so this isn't strictly correct, but it's the best we
236 * can do. Writes and reads get combined.
238 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
239 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
240 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
241 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
243 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
245 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
246 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
248 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
249 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
251 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
252 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
253 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
254 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
258 * Cortex-A5 HW events mapping
260 static const unsigned armv7_a5_perf_map
[PERF_COUNT_HW_MAX
] = {
261 PERF_MAP_ALL_UNSUPPORTED
,
262 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
263 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
264 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
265 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
266 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
267 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
270 static const unsigned armv7_a5_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
271 [PERF_COUNT_HW_CACHE_OP_MAX
]
272 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
273 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
275 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
276 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
277 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
278 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
279 [C(L1D
)][C(OP_PREFETCH
)][C(RESULT_ACCESS
)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL
,
280 [C(L1D
)][C(OP_PREFETCH
)][C(RESULT_MISS
)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP
,
282 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS
,
283 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
285 * The prefetch counters don't differentiate between the I side and the
288 [C(L1I
)][C(OP_PREFETCH
)][C(RESULT_ACCESS
)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL
,
289 [C(L1I
)][C(OP_PREFETCH
)][C(RESULT_MISS
)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP
,
291 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
292 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
294 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
295 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
297 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
298 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
299 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
300 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
304 * Cortex-A15 HW events mapping
306 static const unsigned armv7_a15_perf_map
[PERF_COUNT_HW_MAX
] = {
307 PERF_MAP_ALL_UNSUPPORTED
,
308 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
309 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
310 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
311 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
312 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_A15_PERFCTR_PC_WRITE_SPEC
,
313 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
314 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_BUS_CYCLES
,
317 static const unsigned armv7_a15_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
318 [PERF_COUNT_HW_CACHE_OP_MAX
]
319 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
320 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
322 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ
,
323 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ
,
324 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE
,
325 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE
,
328 * Not all performance counters differentiate between read and write
329 * accesses/misses so we're not always strictly correct, but it's the
330 * best we can do. Writes and reads get combined in these cases.
332 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS
,
333 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
335 [C(LL
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ
,
336 [C(LL
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ
,
337 [C(LL
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE
,
338 [C(LL
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE
,
340 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ
,
341 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE
,
343 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
344 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
346 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
347 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
348 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
349 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
353 * Cortex-A7 HW events mapping
355 static const unsigned armv7_a7_perf_map
[PERF_COUNT_HW_MAX
] = {
356 PERF_MAP_ALL_UNSUPPORTED
,
357 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
358 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
359 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
360 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
361 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
362 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
363 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_BUS_CYCLES
,
366 static const unsigned armv7_a7_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
367 [PERF_COUNT_HW_CACHE_OP_MAX
]
368 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
369 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
372 * The performance counters don't differentiate between read and write
373 * accesses/misses so this isn't strictly correct, but it's the best we
374 * can do. Writes and reads get combined.
376 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
377 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
378 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
379 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
381 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS
,
382 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
384 [C(LL
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L2_CACHE_ACCESS
,
385 [C(LL
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L2_CACHE_REFILL
,
386 [C(LL
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L2_CACHE_ACCESS
,
387 [C(LL
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L2_CACHE_REFILL
,
389 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
390 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
392 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
393 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
395 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
396 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
397 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
398 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
402 * Cortex-A12 HW events mapping
404 static const unsigned armv7_a12_perf_map
[PERF_COUNT_HW_MAX
] = {
405 PERF_MAP_ALL_UNSUPPORTED
,
406 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
407 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
408 [PERF_COUNT_HW_CACHE_REFERENCES
] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
409 [PERF_COUNT_HW_CACHE_MISSES
] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
410 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_A12_PERFCTR_PC_WRITE_SPEC
,
411 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
412 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_BUS_CYCLES
,
415 static const unsigned armv7_a12_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
416 [PERF_COUNT_HW_CACHE_OP_MAX
]
417 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
418 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
420 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_READ
,
421 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
422 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_A12_PERFCTR_L1_DCACHE_ACCESS_WRITE
,
423 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
426 * Not all performance counters differentiate between read and write
427 * accesses/misses so we're not always strictly correct, but it's the
428 * best we can do. Writes and reads get combined in these cases.
430 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS
,
431 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_ICACHE_REFILL
,
433 [C(LL
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_READ
,
434 [C(LL
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L2_CACHE_REFILL
,
435 [C(LL
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_A12_PERFCTR_L2_CACHE_ACCESS_WRITE
,
436 [C(LL
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L2_CACHE_REFILL
,
438 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
439 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_DTLB_REFILL
,
440 [C(DTLB
)][C(OP_PREFETCH
)][C(RESULT_MISS
)] = ARMV7_A12_PERFCTR_PF_TLB_REFILL
,
442 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
443 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_ITLB_REFILL
,
445 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
446 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
447 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
448 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
452 * Krait HW events mapping
454 static const unsigned krait_perf_map
[PERF_COUNT_HW_MAX
] = {
455 PERF_MAP_ALL_UNSUPPORTED
,
456 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
457 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
458 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
459 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
460 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_CLOCK_CYCLES
,
463 static const unsigned krait_perf_map_no_branch
[PERF_COUNT_HW_MAX
] = {
464 PERF_MAP_ALL_UNSUPPORTED
,
465 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
466 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
467 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
468 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_CLOCK_CYCLES
,
471 static const unsigned krait_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
472 [PERF_COUNT_HW_CACHE_OP_MAX
]
473 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
474 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
477 * The performance counters don't differentiate between read and write
478 * accesses/misses so this isn't strictly correct, but it's the best we
479 * can do. Writes and reads get combined.
481 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
482 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
483 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
484 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
486 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = KRAIT_PERFCTR_L1_ICACHE_ACCESS
,
487 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = KRAIT_PERFCTR_L1_ICACHE_MISS
,
489 [C(DTLB
)][C(OP_READ
)][C(RESULT_ACCESS
)] = KRAIT_PERFCTR_L1_DTLB_ACCESS
,
490 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = KRAIT_PERFCTR_L1_DTLB_ACCESS
,
492 [C(ITLB
)][C(OP_READ
)][C(RESULT_ACCESS
)] = KRAIT_PERFCTR_L1_ITLB_ACCESS
,
493 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = KRAIT_PERFCTR_L1_ITLB_ACCESS
,
495 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
496 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
497 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
498 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
502 * Scorpion HW events mapping
504 static const unsigned scorpion_perf_map
[PERF_COUNT_HW_MAX
] = {
505 PERF_MAP_ALL_UNSUPPORTED
,
506 [PERF_COUNT_HW_CPU_CYCLES
] = ARMV7_PERFCTR_CPU_CYCLES
,
507 [PERF_COUNT_HW_INSTRUCTIONS
] = ARMV7_PERFCTR_INSTR_EXECUTED
,
508 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = ARMV7_PERFCTR_PC_WRITE
,
509 [PERF_COUNT_HW_BRANCH_MISSES
] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
510 [PERF_COUNT_HW_BUS_CYCLES
] = ARMV7_PERFCTR_CLOCK_CYCLES
,
513 static const unsigned scorpion_perf_cache_map
[PERF_COUNT_HW_CACHE_MAX
]
514 [PERF_COUNT_HW_CACHE_OP_MAX
]
515 [PERF_COUNT_HW_CACHE_RESULT_MAX
] = {
516 PERF_CACHE_MAP_ALL_UNSUPPORTED
,
518 * The performance counters don't differentiate between read and write
519 * accesses/misses so this isn't strictly correct, but it's the best we
520 * can do. Writes and reads get combined.
522 [C(L1D
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
523 [C(L1D
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
524 [C(L1D
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS
,
525 [C(L1D
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_L1_DCACHE_REFILL
,
526 [C(L1I
)][C(OP_READ
)][C(RESULT_ACCESS
)] = SCORPION_ICACHE_ACCESS
,
527 [C(L1I
)][C(OP_READ
)][C(RESULT_MISS
)] = SCORPION_ICACHE_MISS
,
529 * Only ITLB misses and DTLB refills are supported. If users want the
530 * DTLB refills misses a raw counter must be used.
532 [C(DTLB
)][C(OP_READ
)][C(RESULT_ACCESS
)] = SCORPION_DTLB_ACCESS
,
533 [C(DTLB
)][C(OP_READ
)][C(RESULT_MISS
)] = SCORPION_DTLB_MISS
,
534 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = SCORPION_DTLB_ACCESS
,
535 [C(DTLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = SCORPION_DTLB_MISS
,
536 [C(ITLB
)][C(OP_READ
)][C(RESULT_MISS
)] = SCORPION_ITLB_MISS
,
537 [C(ITLB
)][C(OP_WRITE
)][C(RESULT_MISS
)] = SCORPION_ITLB_MISS
,
538 [C(BPU
)][C(OP_READ
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
539 [C(BPU
)][C(OP_READ
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
540 [C(BPU
)][C(OP_WRITE
)][C(RESULT_ACCESS
)] = ARMV7_PERFCTR_PC_BRANCH_PRED
,
541 [C(BPU
)][C(OP_WRITE
)][C(RESULT_MISS
)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED
,
545 * Perf Events' indices
547 #define ARMV7_IDX_CYCLE_COUNTER 0
548 #define ARMV7_IDX_COUNTER0 1
549 #define ARMV7_IDX_COUNTER_LAST(cpu_pmu) \
550 (ARMV7_IDX_CYCLE_COUNTER + cpu_pmu->num_events - 1)
552 #define ARMV7_MAX_COUNTERS 32
553 #define ARMV7_COUNTER_MASK (ARMV7_MAX_COUNTERS - 1)
556 * ARMv7 low level PMNC access
560 * Perf Event to low level counters mapping
562 #define ARMV7_IDX_TO_COUNTER(x) \
563 (((x) - ARMV7_IDX_COUNTER0) & ARMV7_COUNTER_MASK)
566 * Per-CPU PMNC: config reg
568 #define ARMV7_PMNC_E (1 << 0) /* Enable all counters */
569 #define ARMV7_PMNC_P (1 << 1) /* Reset all counters */
570 #define ARMV7_PMNC_C (1 << 2) /* Cycle counter reset */
571 #define ARMV7_PMNC_D (1 << 3) /* CCNT counts every 64th cpu cycle */
572 #define ARMV7_PMNC_X (1 << 4) /* Export to ETM */
573 #define ARMV7_PMNC_DP (1 << 5) /* Disable CCNT if non-invasive debug*/
574 #define ARMV7_PMNC_N_SHIFT 11 /* Number of counters supported */
575 #define ARMV7_PMNC_N_MASK 0x1f
576 #define ARMV7_PMNC_MASK 0x3f /* Mask for writable bits */
579 * FLAG: counters overflow flag status reg
581 #define ARMV7_FLAG_MASK 0xffffffff /* Mask for writable bits */
582 #define ARMV7_OVERFLOWED_MASK ARMV7_FLAG_MASK
585 * PMXEVTYPER: Event selection reg
587 #define ARMV7_EVTYPE_MASK 0xc80000ff /* Mask for writable bits */
588 #define ARMV7_EVTYPE_EVENT 0xff /* Mask for EVENT bits */
591 * Event filters for PMUv2
593 #define ARMV7_EXCLUDE_PL1 (1 << 31)
594 #define ARMV7_EXCLUDE_USER (1 << 30)
595 #define ARMV7_INCLUDE_HYP (1 << 27)
597 static inline u32
armv7_pmnc_read(void)
600 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r"(val
));
604 static inline void armv7_pmnc_write(u32 val
)
606 val
&= ARMV7_PMNC_MASK
;
608 asm volatile("mcr p15, 0, %0, c9, c12, 0" : : "r"(val
));
611 static inline int armv7_pmnc_has_overflowed(u32 pmnc
)
613 return pmnc
& ARMV7_OVERFLOWED_MASK
;
616 static inline int armv7_pmnc_counter_valid(struct arm_pmu
*cpu_pmu
, int idx
)
618 return idx
>= ARMV7_IDX_CYCLE_COUNTER
&&
619 idx
<= ARMV7_IDX_COUNTER_LAST(cpu_pmu
);
622 static inline int armv7_pmnc_counter_has_overflowed(u32 pmnc
, int idx
)
624 return pmnc
& BIT(ARMV7_IDX_TO_COUNTER(idx
));
627 static inline void armv7_pmnc_select_counter(int idx
)
629 u32 counter
= ARMV7_IDX_TO_COUNTER(idx
);
630 asm volatile("mcr p15, 0, %0, c9, c12, 5" : : "r" (counter
));
634 static inline u32
armv7pmu_read_counter(struct perf_event
*event
)
636 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
637 struct hw_perf_event
*hwc
= &event
->hw
;
641 if (!armv7_pmnc_counter_valid(cpu_pmu
, idx
)) {
642 pr_err("CPU%u reading wrong counter %d\n",
643 smp_processor_id(), idx
);
644 } else if (idx
== ARMV7_IDX_CYCLE_COUNTER
) {
645 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (value
));
647 armv7_pmnc_select_counter(idx
);
648 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (value
));
654 static inline void armv7pmu_write_counter(struct perf_event
*event
, u32 value
)
656 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
657 struct hw_perf_event
*hwc
= &event
->hw
;
660 if (!armv7_pmnc_counter_valid(cpu_pmu
, idx
)) {
661 pr_err("CPU%u writing wrong counter %d\n",
662 smp_processor_id(), idx
);
663 } else if (idx
== ARMV7_IDX_CYCLE_COUNTER
) {
664 asm volatile("mcr p15, 0, %0, c9, c13, 0" : : "r" (value
));
666 armv7_pmnc_select_counter(idx
);
667 asm volatile("mcr p15, 0, %0, c9, c13, 2" : : "r" (value
));
671 static inline void armv7_pmnc_write_evtsel(int idx
, u32 val
)
673 armv7_pmnc_select_counter(idx
);
674 val
&= ARMV7_EVTYPE_MASK
;
675 asm volatile("mcr p15, 0, %0, c9, c13, 1" : : "r" (val
));
678 static inline void armv7_pmnc_enable_counter(int idx
)
680 u32 counter
= ARMV7_IDX_TO_COUNTER(idx
);
681 asm volatile("mcr p15, 0, %0, c9, c12, 1" : : "r" (BIT(counter
)));
684 static inline void armv7_pmnc_disable_counter(int idx
)
686 u32 counter
= ARMV7_IDX_TO_COUNTER(idx
);
687 asm volatile("mcr p15, 0, %0, c9, c12, 2" : : "r" (BIT(counter
)));
690 static inline void armv7_pmnc_enable_intens(int idx
)
692 u32 counter
= ARMV7_IDX_TO_COUNTER(idx
);
693 asm volatile("mcr p15, 0, %0, c9, c14, 1" : : "r" (BIT(counter
)));
696 static inline void armv7_pmnc_disable_intens(int idx
)
698 u32 counter
= ARMV7_IDX_TO_COUNTER(idx
);
699 asm volatile("mcr p15, 0, %0, c9, c14, 2" : : "r" (BIT(counter
)));
701 /* Clear the overflow flag in case an interrupt is pending. */
702 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (BIT(counter
)));
706 static inline u32
armv7_pmnc_getreset_flags(void)
711 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val
));
713 /* Write to clear flags */
714 val
&= ARMV7_FLAG_MASK
;
715 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (val
));
721 static void armv7_pmnc_dump_regs(struct arm_pmu
*cpu_pmu
)
726 pr_info("PMNC registers dump:\n");
728 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r" (val
));
729 pr_info("PMNC =0x%08x\n", val
);
731 asm volatile("mrc p15, 0, %0, c9, c12, 1" : "=r" (val
));
732 pr_info("CNTENS=0x%08x\n", val
);
734 asm volatile("mrc p15, 0, %0, c9, c14, 1" : "=r" (val
));
735 pr_info("INTENS=0x%08x\n", val
);
737 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val
));
738 pr_info("FLAGS =0x%08x\n", val
);
740 asm volatile("mrc p15, 0, %0, c9, c12, 5" : "=r" (val
));
741 pr_info("SELECT=0x%08x\n", val
);
743 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (val
));
744 pr_info("CCNT =0x%08x\n", val
);
746 for (cnt
= ARMV7_IDX_COUNTER0
;
747 cnt
<= ARMV7_IDX_COUNTER_LAST(cpu_pmu
); cnt
++) {
748 armv7_pmnc_select_counter(cnt
);
749 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (val
));
750 pr_info("CNT[%d] count =0x%08x\n",
751 ARMV7_IDX_TO_COUNTER(cnt
), val
);
752 asm volatile("mrc p15, 0, %0, c9, c13, 1" : "=r" (val
));
753 pr_info("CNT[%d] evtsel=0x%08x\n",
754 ARMV7_IDX_TO_COUNTER(cnt
), val
);
759 static void armv7pmu_enable_event(struct perf_event
*event
)
762 struct hw_perf_event
*hwc
= &event
->hw
;
763 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
764 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
767 if (!armv7_pmnc_counter_valid(cpu_pmu
, idx
)) {
768 pr_err("CPU%u enabling wrong PMNC counter IRQ enable %d\n",
769 smp_processor_id(), idx
);
774 * Enable counter and interrupt, and set the counter to count
775 * the event that we're interested in.
777 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
782 armv7_pmnc_disable_counter(idx
);
785 * Set event (if destined for PMNx counters)
786 * We only need to set the event for the cycle counter if we
787 * have the ability to perform event filtering.
789 if (cpu_pmu
->set_event_filter
|| idx
!= ARMV7_IDX_CYCLE_COUNTER
)
790 armv7_pmnc_write_evtsel(idx
, hwc
->config_base
);
793 * Enable interrupt for this counter
795 armv7_pmnc_enable_intens(idx
);
800 armv7_pmnc_enable_counter(idx
);
802 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
805 static void armv7pmu_disable_event(struct perf_event
*event
)
808 struct hw_perf_event
*hwc
= &event
->hw
;
809 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
810 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
813 if (!armv7_pmnc_counter_valid(cpu_pmu
, idx
)) {
814 pr_err("CPU%u disabling wrong PMNC counter IRQ enable %d\n",
815 smp_processor_id(), idx
);
820 * Disable counter and interrupt
822 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
827 armv7_pmnc_disable_counter(idx
);
830 * Disable interrupt for this counter
832 armv7_pmnc_disable_intens(idx
);
834 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
837 static irqreturn_t
armv7pmu_handle_irq(int irq_num
, void *dev
)
840 struct perf_sample_data data
;
841 struct arm_pmu
*cpu_pmu
= (struct arm_pmu
*)dev
;
842 struct pmu_hw_events
*cpuc
= this_cpu_ptr(cpu_pmu
->hw_events
);
843 struct pt_regs
*regs
;
847 * Get and reset the IRQ flags
849 pmnc
= armv7_pmnc_getreset_flags();
852 * Did an overflow occur?
854 if (!armv7_pmnc_has_overflowed(pmnc
))
858 * Handle the counter(s) overflow(s)
860 regs
= get_irq_regs();
862 for (idx
= 0; idx
< cpu_pmu
->num_events
; ++idx
) {
863 struct perf_event
*event
= cpuc
->events
[idx
];
864 struct hw_perf_event
*hwc
;
866 /* Ignore if we don't have an event. */
871 * We have a single interrupt for all counters. Check that
872 * each counter has overflowed before we process it.
874 if (!armv7_pmnc_counter_has_overflowed(pmnc
, idx
))
878 armpmu_event_update(event
);
879 perf_sample_data_init(&data
, 0, hwc
->last_period
);
880 if (!armpmu_event_set_period(event
))
883 if (perf_event_overflow(event
, &data
, regs
))
884 cpu_pmu
->disable(event
);
888 * Handle the pending perf events.
890 * Note: this call *must* be run with interrupts disabled. For
891 * platforms that can have the PMU interrupts raised as an NMI, this
899 static void armv7pmu_start(struct arm_pmu
*cpu_pmu
)
902 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
904 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
905 /* Enable all counters */
906 armv7_pmnc_write(armv7_pmnc_read() | ARMV7_PMNC_E
);
907 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
910 static void armv7pmu_stop(struct arm_pmu
*cpu_pmu
)
913 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
915 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
916 /* Disable all counters */
917 armv7_pmnc_write(armv7_pmnc_read() & ~ARMV7_PMNC_E
);
918 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
921 static int armv7pmu_get_event_idx(struct pmu_hw_events
*cpuc
,
922 struct perf_event
*event
)
925 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
926 struct hw_perf_event
*hwc
= &event
->hw
;
927 unsigned long evtype
= hwc
->config_base
& ARMV7_EVTYPE_EVENT
;
929 /* Always place a cycle counter into the cycle counter. */
930 if (evtype
== ARMV7_PERFCTR_CPU_CYCLES
) {
931 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER
, cpuc
->used_mask
))
934 return ARMV7_IDX_CYCLE_COUNTER
;
938 * For anything other than a cycle counter, try and use
939 * the events counters
941 for (idx
= ARMV7_IDX_COUNTER0
; idx
< cpu_pmu
->num_events
; ++idx
) {
942 if (!test_and_set_bit(idx
, cpuc
->used_mask
))
946 /* The counters are all in use. */
951 * Add an event filter to a given event. This will only work for PMUv2 PMUs.
953 static int armv7pmu_set_event_filter(struct hw_perf_event
*event
,
954 struct perf_event_attr
*attr
)
956 unsigned long config_base
= 0;
958 if (attr
->exclude_idle
)
960 if (attr
->exclude_user
)
961 config_base
|= ARMV7_EXCLUDE_USER
;
962 if (attr
->exclude_kernel
)
963 config_base
|= ARMV7_EXCLUDE_PL1
;
964 if (!attr
->exclude_hv
)
965 config_base
|= ARMV7_INCLUDE_HYP
;
968 * Install the filter into config_base as this is used to
969 * construct the event type.
971 event
->config_base
= config_base
;
976 static void armv7pmu_reset(void *info
)
978 struct arm_pmu
*cpu_pmu
= (struct arm_pmu
*)info
;
979 u32 idx
, nb_cnt
= cpu_pmu
->num_events
;
981 /* The counter and interrupt enable registers are unknown at reset. */
982 for (idx
= ARMV7_IDX_CYCLE_COUNTER
; idx
< nb_cnt
; ++idx
) {
983 armv7_pmnc_disable_counter(idx
);
984 armv7_pmnc_disable_intens(idx
);
987 /* Initialize & Reset PMNC: C and P bits */
988 armv7_pmnc_write(ARMV7_PMNC_P
| ARMV7_PMNC_C
);
991 static int armv7_a8_map_event(struct perf_event
*event
)
993 return armpmu_map_event(event
, &armv7_a8_perf_map
,
994 &armv7_a8_perf_cache_map
, 0xFF);
997 static int armv7_a9_map_event(struct perf_event
*event
)
999 return armpmu_map_event(event
, &armv7_a9_perf_map
,
1000 &armv7_a9_perf_cache_map
, 0xFF);
1003 static int armv7_a5_map_event(struct perf_event
*event
)
1005 return armpmu_map_event(event
, &armv7_a5_perf_map
,
1006 &armv7_a5_perf_cache_map
, 0xFF);
1009 static int armv7_a15_map_event(struct perf_event
*event
)
1011 return armpmu_map_event(event
, &armv7_a15_perf_map
,
1012 &armv7_a15_perf_cache_map
, 0xFF);
1015 static int armv7_a7_map_event(struct perf_event
*event
)
1017 return armpmu_map_event(event
, &armv7_a7_perf_map
,
1018 &armv7_a7_perf_cache_map
, 0xFF);
1021 static int armv7_a12_map_event(struct perf_event
*event
)
1023 return armpmu_map_event(event
, &armv7_a12_perf_map
,
1024 &armv7_a12_perf_cache_map
, 0xFF);
1027 static int krait_map_event(struct perf_event
*event
)
1029 return armpmu_map_event(event
, &krait_perf_map
,
1030 &krait_perf_cache_map
, 0xFFFFF);
1033 static int krait_map_event_no_branch(struct perf_event
*event
)
1035 return armpmu_map_event(event
, &krait_perf_map_no_branch
,
1036 &krait_perf_cache_map
, 0xFFFFF);
1039 static int scorpion_map_event(struct perf_event
*event
)
1041 return armpmu_map_event(event
, &scorpion_perf_map
,
1042 &scorpion_perf_cache_map
, 0xFFFFF);
1045 static void armv7pmu_init(struct arm_pmu
*cpu_pmu
)
1047 cpu_pmu
->handle_irq
= armv7pmu_handle_irq
;
1048 cpu_pmu
->enable
= armv7pmu_enable_event
;
1049 cpu_pmu
->disable
= armv7pmu_disable_event
;
1050 cpu_pmu
->read_counter
= armv7pmu_read_counter
;
1051 cpu_pmu
->write_counter
= armv7pmu_write_counter
;
1052 cpu_pmu
->get_event_idx
= armv7pmu_get_event_idx
;
1053 cpu_pmu
->start
= armv7pmu_start
;
1054 cpu_pmu
->stop
= armv7pmu_stop
;
1055 cpu_pmu
->reset
= armv7pmu_reset
;
1056 cpu_pmu
->max_period
= (1LLU << 32) - 1;
1059 static u32
armv7_read_num_pmnc_events(void)
1063 /* Read the nb of CNTx counters supported from PMNC */
1064 nb_cnt
= (armv7_pmnc_read() >> ARMV7_PMNC_N_SHIFT
) & ARMV7_PMNC_N_MASK
;
1066 /* Add the CPU cycles counter and return */
1070 static int armv7_a8_pmu_init(struct arm_pmu
*cpu_pmu
)
1072 armv7pmu_init(cpu_pmu
);
1073 cpu_pmu
->name
= "armv7_cortex_a8";
1074 cpu_pmu
->map_event
= armv7_a8_map_event
;
1075 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1079 static int armv7_a9_pmu_init(struct arm_pmu
*cpu_pmu
)
1081 armv7pmu_init(cpu_pmu
);
1082 cpu_pmu
->name
= "armv7_cortex_a9";
1083 cpu_pmu
->map_event
= armv7_a9_map_event
;
1084 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1088 static int armv7_a5_pmu_init(struct arm_pmu
*cpu_pmu
)
1090 armv7pmu_init(cpu_pmu
);
1091 cpu_pmu
->name
= "armv7_cortex_a5";
1092 cpu_pmu
->map_event
= armv7_a5_map_event
;
1093 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1097 static int armv7_a15_pmu_init(struct arm_pmu
*cpu_pmu
)
1099 armv7pmu_init(cpu_pmu
);
1100 cpu_pmu
->name
= "armv7_cortex_a15";
1101 cpu_pmu
->map_event
= armv7_a15_map_event
;
1102 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1103 cpu_pmu
->set_event_filter
= armv7pmu_set_event_filter
;
1107 static int armv7_a7_pmu_init(struct arm_pmu
*cpu_pmu
)
1109 armv7pmu_init(cpu_pmu
);
1110 cpu_pmu
->name
= "armv7_cortex_a7";
1111 cpu_pmu
->map_event
= armv7_a7_map_event
;
1112 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1113 cpu_pmu
->set_event_filter
= armv7pmu_set_event_filter
;
1117 static int armv7_a12_pmu_init(struct arm_pmu
*cpu_pmu
)
1119 armv7pmu_init(cpu_pmu
);
1120 cpu_pmu
->name
= "armv7_cortex_a12";
1121 cpu_pmu
->map_event
= armv7_a12_map_event
;
1122 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1123 cpu_pmu
->set_event_filter
= armv7pmu_set_event_filter
;
1127 static int armv7_a17_pmu_init(struct arm_pmu
*cpu_pmu
)
1129 armv7_a12_pmu_init(cpu_pmu
);
1130 cpu_pmu
->name
= "armv7_cortex_a17";
1135 * Krait Performance Monitor Region Event Selection Register (PMRESRn)
1138 * +--------------------------------+
1139 * PMRESR0 | EN | CC | CC | CC | CC | N = 1, R = 0
1140 * +--------------------------------+
1141 * PMRESR1 | EN | CC | CC | CC | CC | N = 1, R = 1
1142 * +--------------------------------+
1143 * PMRESR2 | EN | CC | CC | CC | CC | N = 1, R = 2
1144 * +--------------------------------+
1145 * VPMRESR0 | EN | CC | CC | CC | CC | N = 2, R = ?
1146 * +--------------------------------+
1147 * EN | G=3 | G=2 | G=1 | G=0
1151 * hwc->config_base = 0xNRCCG
1153 * N = prefix, 1 for Krait CPU (PMRESRn), 2 for Venum VFP (VPMRESR)
1154 * R = region register
1155 * CC = class of events the group G is choosing from
1156 * G = group or particular event
1158 * Example: 0x12021 is a Krait CPU event in PMRESR2's group 1 with code 2
1160 * A region (R) corresponds to a piece of the CPU (execution unit, instruction
1161 * unit, etc.) while the event code (CC) corresponds to a particular class of
1162 * events (interrupts for example). An event code is broken down into
1163 * groups (G) that can be mapped into the PMU (irq, fiqs, and irq+fiqs for
1167 #define KRAIT_EVENT (1 << 16)
1168 #define VENUM_EVENT (2 << 16)
1169 #define KRAIT_EVENT_MASK (KRAIT_EVENT | VENUM_EVENT)
1170 #define PMRESRn_EN BIT(31)
1172 #define EVENT_REGION(event) (((event) >> 12) & 0xf) /* R */
1173 #define EVENT_GROUP(event) ((event) & 0xf) /* G */
1174 #define EVENT_CODE(event) (((event) >> 4) & 0xff) /* CC */
1175 #define EVENT_VENUM(event) (!!(event & VENUM_EVENT)) /* N=2 */
1176 #define EVENT_CPU(event) (!!(event & KRAIT_EVENT)) /* N=1 */
1178 static u32
krait_read_pmresrn(int n
)
1184 asm volatile("mrc p15, 1, %0, c9, c15, 0" : "=r" (val
));
1187 asm volatile("mrc p15, 1, %0, c9, c15, 1" : "=r" (val
));
1190 asm volatile("mrc p15, 1, %0, c9, c15, 2" : "=r" (val
));
1193 BUG(); /* Should be validated in krait_pmu_get_event_idx() */
1199 static void krait_write_pmresrn(int n
, u32 val
)
1203 asm volatile("mcr p15, 1, %0, c9, c15, 0" : : "r" (val
));
1206 asm volatile("mcr p15, 1, %0, c9, c15, 1" : : "r" (val
));
1209 asm volatile("mcr p15, 1, %0, c9, c15, 2" : : "r" (val
));
1212 BUG(); /* Should be validated in krait_pmu_get_event_idx() */
1216 static u32
venum_read_pmresr(void)
1219 asm volatile("mrc p10, 7, %0, c11, c0, 0" : "=r" (val
));
1223 static void venum_write_pmresr(u32 val
)
1225 asm volatile("mcr p10, 7, %0, c11, c0, 0" : : "r" (val
));
1228 static void venum_pre_pmresr(u32
*venum_orig_val
, u32
*fp_orig_val
)
1233 BUG_ON(preemptible());
1234 /* CPACR Enable CP10 and CP11 access */
1235 *venum_orig_val
= get_copro_access();
1236 venum_new_val
= *venum_orig_val
| CPACC_SVC(10) | CPACC_SVC(11);
1237 set_copro_access(venum_new_val
);
1240 *fp_orig_val
= fmrx(FPEXC
);
1241 fp_new_val
= *fp_orig_val
| FPEXC_EN
;
1242 fmxr(FPEXC
, fp_new_val
);
1245 static void venum_post_pmresr(u32 venum_orig_val
, u32 fp_orig_val
)
1247 BUG_ON(preemptible());
1249 fmxr(FPEXC
, fp_orig_val
);
1252 set_copro_access(venum_orig_val
);
1255 static u32
krait_get_pmresrn_event(unsigned int region
)
1257 static const u32 pmresrn_table
[] = { KRAIT_PMRESR0_GROUP0
,
1258 KRAIT_PMRESR1_GROUP0
,
1259 KRAIT_PMRESR2_GROUP0
};
1260 return pmresrn_table
[region
];
1263 static void krait_evt_setup(int idx
, u32 config_base
)
1268 unsigned int region
= EVENT_REGION(config_base
);
1269 unsigned int group
= EVENT_GROUP(config_base
);
1270 unsigned int code
= EVENT_CODE(config_base
);
1271 unsigned int group_shift
;
1272 bool venum_event
= EVENT_VENUM(config_base
);
1274 group_shift
= group
* 8;
1275 mask
= 0xff << group_shift
;
1277 /* Configure evtsel for the region and group */
1279 val
= KRAIT_VPMRESR0_GROUP0
;
1281 val
= krait_get_pmresrn_event(region
);
1283 /* Mix in mode-exclusion bits */
1284 val
|= config_base
& (ARMV7_EXCLUDE_USER
| ARMV7_EXCLUDE_PL1
);
1285 armv7_pmnc_write_evtsel(idx
, val
);
1288 venum_pre_pmresr(&vval
, &fval
);
1289 val
= venum_read_pmresr();
1291 val
|= code
<< group_shift
;
1293 venum_write_pmresr(val
);
1294 venum_post_pmresr(vval
, fval
);
1296 val
= krait_read_pmresrn(region
);
1298 val
|= code
<< group_shift
;
1300 krait_write_pmresrn(region
, val
);
1304 static u32
clear_pmresrn_group(u32 val
, int group
)
1309 group_shift
= group
* 8;
1310 mask
= 0xff << group_shift
;
1313 /* Don't clear enable bit if entire region isn't disabled */
1314 if (val
& ~PMRESRn_EN
)
1315 return val
|= PMRESRn_EN
;
1320 static void krait_clearpmu(u32 config_base
)
1324 unsigned int region
= EVENT_REGION(config_base
);
1325 unsigned int group
= EVENT_GROUP(config_base
);
1326 bool venum_event
= EVENT_VENUM(config_base
);
1329 venum_pre_pmresr(&vval
, &fval
);
1330 val
= venum_read_pmresr();
1331 val
= clear_pmresrn_group(val
, group
);
1332 venum_write_pmresr(val
);
1333 venum_post_pmresr(vval
, fval
);
1335 val
= krait_read_pmresrn(region
);
1336 val
= clear_pmresrn_group(val
, group
);
1337 krait_write_pmresrn(region
, val
);
1341 static void krait_pmu_disable_event(struct perf_event
*event
)
1343 unsigned long flags
;
1344 struct hw_perf_event
*hwc
= &event
->hw
;
1346 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1347 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
1349 /* Disable counter and interrupt */
1350 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
1352 /* Disable counter */
1353 armv7_pmnc_disable_counter(idx
);
1356 * Clear pmresr code (if destined for PMNx counters)
1358 if (hwc
->config_base
& KRAIT_EVENT_MASK
)
1359 krait_clearpmu(hwc
->config_base
);
1361 /* Disable interrupt for this counter */
1362 armv7_pmnc_disable_intens(idx
);
1364 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
1367 static void krait_pmu_enable_event(struct perf_event
*event
)
1369 unsigned long flags
;
1370 struct hw_perf_event
*hwc
= &event
->hw
;
1372 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1373 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
1376 * Enable counter and interrupt, and set the counter to count
1377 * the event that we're interested in.
1379 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
1381 /* Disable counter */
1382 armv7_pmnc_disable_counter(idx
);
1385 * Set event (if destined for PMNx counters)
1386 * We set the event for the cycle counter because we
1387 * have the ability to perform event filtering.
1389 if (hwc
->config_base
& KRAIT_EVENT_MASK
)
1390 krait_evt_setup(idx
, hwc
->config_base
);
1392 armv7_pmnc_write_evtsel(idx
, hwc
->config_base
);
1394 /* Enable interrupt for this counter */
1395 armv7_pmnc_enable_intens(idx
);
1397 /* Enable counter */
1398 armv7_pmnc_enable_counter(idx
);
1400 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
1403 static void krait_pmu_reset(void *info
)
1406 struct arm_pmu
*cpu_pmu
= info
;
1407 u32 idx
, nb_cnt
= cpu_pmu
->num_events
;
1409 armv7pmu_reset(info
);
1411 /* Clear all pmresrs */
1412 krait_write_pmresrn(0, 0);
1413 krait_write_pmresrn(1, 0);
1414 krait_write_pmresrn(2, 0);
1416 venum_pre_pmresr(&vval
, &fval
);
1417 venum_write_pmresr(0);
1418 venum_post_pmresr(vval
, fval
);
1420 /* Reset PMxEVNCTCR to sane default */
1421 for (idx
= ARMV7_IDX_CYCLE_COUNTER
; idx
< nb_cnt
; ++idx
) {
1422 armv7_pmnc_select_counter(idx
);
1423 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1428 static int krait_event_to_bit(struct perf_event
*event
, unsigned int region
,
1432 struct hw_perf_event
*hwc
= &event
->hw
;
1433 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1435 if (hwc
->config_base
& VENUM_EVENT
)
1436 bit
= KRAIT_VPMRESR0_GROUP0
;
1438 bit
= krait_get_pmresrn_event(region
);
1439 bit
-= krait_get_pmresrn_event(0);
1442 * Lower bits are reserved for use by the counters (see
1443 * armv7pmu_get_event_idx() for more info)
1445 bit
+= ARMV7_IDX_COUNTER_LAST(cpu_pmu
) + 1;
1451 * We check for column exclusion constraints here.
1452 * Two events cant use the same group within a pmresr register.
1454 static int krait_pmu_get_event_idx(struct pmu_hw_events
*cpuc
,
1455 struct perf_event
*event
)
1459 struct hw_perf_event
*hwc
= &event
->hw
;
1460 unsigned int region
= EVENT_REGION(hwc
->config_base
);
1461 unsigned int code
= EVENT_CODE(hwc
->config_base
);
1462 unsigned int group
= EVENT_GROUP(hwc
->config_base
);
1463 bool venum_event
= EVENT_VENUM(hwc
->config_base
);
1464 bool krait_event
= EVENT_CPU(hwc
->config_base
);
1466 if (venum_event
|| krait_event
) {
1467 /* Ignore invalid events */
1468 if (group
> 3 || region
> 2)
1470 if (venum_event
&& (code
& 0xe0))
1473 bit
= krait_event_to_bit(event
, region
, group
);
1474 if (test_and_set_bit(bit
, cpuc
->used_mask
))
1478 idx
= armv7pmu_get_event_idx(cpuc
, event
);
1479 if (idx
< 0 && bit
>= 0)
1480 clear_bit(bit
, cpuc
->used_mask
);
1485 static void krait_pmu_clear_event_idx(struct pmu_hw_events
*cpuc
,
1486 struct perf_event
*event
)
1489 struct hw_perf_event
*hwc
= &event
->hw
;
1490 unsigned int region
= EVENT_REGION(hwc
->config_base
);
1491 unsigned int group
= EVENT_GROUP(hwc
->config_base
);
1492 bool venum_event
= EVENT_VENUM(hwc
->config_base
);
1493 bool krait_event
= EVENT_CPU(hwc
->config_base
);
1495 if (venum_event
|| krait_event
) {
1496 bit
= krait_event_to_bit(event
, region
, group
);
1497 clear_bit(bit
, cpuc
->used_mask
);
1501 static int krait_pmu_init(struct arm_pmu
*cpu_pmu
)
1503 armv7pmu_init(cpu_pmu
);
1504 cpu_pmu
->name
= "armv7_krait";
1505 /* Some early versions of Krait don't support PC write events */
1506 if (of_property_read_bool(cpu_pmu
->plat_device
->dev
.of_node
,
1507 "qcom,no-pc-write"))
1508 cpu_pmu
->map_event
= krait_map_event_no_branch
;
1510 cpu_pmu
->map_event
= krait_map_event
;
1511 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1512 cpu_pmu
->set_event_filter
= armv7pmu_set_event_filter
;
1513 cpu_pmu
->reset
= krait_pmu_reset
;
1514 cpu_pmu
->enable
= krait_pmu_enable_event
;
1515 cpu_pmu
->disable
= krait_pmu_disable_event
;
1516 cpu_pmu
->get_event_idx
= krait_pmu_get_event_idx
;
1517 cpu_pmu
->clear_event_idx
= krait_pmu_clear_event_idx
;
1522 * Scorpion Local Performance Monitor Register (LPMn)
1525 * +--------------------------------+
1526 * LPM0 | EN | CC | CC | CC | CC | N = 1, R = 0
1527 * +--------------------------------+
1528 * LPM1 | EN | CC | CC | CC | CC | N = 1, R = 1
1529 * +--------------------------------+
1530 * LPM2 | EN | CC | CC | CC | CC | N = 1, R = 2
1531 * +--------------------------------+
1532 * L2LPM | EN | CC | CC | CC | CC | N = 1, R = 3
1533 * +--------------------------------+
1534 * VLPM | EN | CC | CC | CC | CC | N = 2, R = ?
1535 * +--------------------------------+
1536 * EN | G=3 | G=2 | G=1 | G=0
1541 * hwc->config_base = 0xNRCCG
1543 * N = prefix, 1 for Scorpion CPU (LPMn/L2LPM), 2 for Venum VFP (VLPM)
1544 * R = region register
1545 * CC = class of events the group G is choosing from
1546 * G = group or particular event
1548 * Example: 0x12021 is a Scorpion CPU event in LPM2's group 1 with code 2
1550 * A region (R) corresponds to a piece of the CPU (execution unit, instruction
1551 * unit, etc.) while the event code (CC) corresponds to a particular class of
1552 * events (interrupts for example). An event code is broken down into
1553 * groups (G) that can be mapped into the PMU (irq, fiqs, and irq+fiqs for
1557 static u32
scorpion_read_pmresrn(int n
)
1563 asm volatile("mrc p15, 0, %0, c15, c0, 0" : "=r" (val
));
1566 asm volatile("mrc p15, 1, %0, c15, c0, 0" : "=r" (val
));
1569 asm volatile("mrc p15, 2, %0, c15, c0, 0" : "=r" (val
));
1572 asm volatile("mrc p15, 3, %0, c15, c2, 0" : "=r" (val
));
1575 BUG(); /* Should be validated in scorpion_pmu_get_event_idx() */
1581 static void scorpion_write_pmresrn(int n
, u32 val
)
1585 asm volatile("mcr p15, 0, %0, c15, c0, 0" : : "r" (val
));
1588 asm volatile("mcr p15, 1, %0, c15, c0, 0" : : "r" (val
));
1591 asm volatile("mcr p15, 2, %0, c15, c0, 0" : : "r" (val
));
1594 asm volatile("mcr p15, 3, %0, c15, c2, 0" : : "r" (val
));
1597 BUG(); /* Should be validated in scorpion_pmu_get_event_idx() */
1601 static u32
scorpion_get_pmresrn_event(unsigned int region
)
1603 static const u32 pmresrn_table
[] = { SCORPION_LPM0_GROUP0
,
1604 SCORPION_LPM1_GROUP0
,
1605 SCORPION_LPM2_GROUP0
,
1606 SCORPION_L2LPM_GROUP0
};
1607 return pmresrn_table
[region
];
1610 static void scorpion_evt_setup(int idx
, u32 config_base
)
1615 unsigned int region
= EVENT_REGION(config_base
);
1616 unsigned int group
= EVENT_GROUP(config_base
);
1617 unsigned int code
= EVENT_CODE(config_base
);
1618 unsigned int group_shift
;
1619 bool venum_event
= EVENT_VENUM(config_base
);
1621 group_shift
= group
* 8;
1622 mask
= 0xff << group_shift
;
1624 /* Configure evtsel for the region and group */
1626 val
= SCORPION_VLPM_GROUP0
;
1628 val
= scorpion_get_pmresrn_event(region
);
1630 /* Mix in mode-exclusion bits */
1631 val
|= config_base
& (ARMV7_EXCLUDE_USER
| ARMV7_EXCLUDE_PL1
);
1632 armv7_pmnc_write_evtsel(idx
, val
);
1634 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1637 venum_pre_pmresr(&vval
, &fval
);
1638 val
= venum_read_pmresr();
1640 val
|= code
<< group_shift
;
1642 venum_write_pmresr(val
);
1643 venum_post_pmresr(vval
, fval
);
1645 val
= scorpion_read_pmresrn(region
);
1647 val
|= code
<< group_shift
;
1649 scorpion_write_pmresrn(region
, val
);
1653 static void scorpion_clearpmu(u32 config_base
)
1657 unsigned int region
= EVENT_REGION(config_base
);
1658 unsigned int group
= EVENT_GROUP(config_base
);
1659 bool venum_event
= EVENT_VENUM(config_base
);
1662 venum_pre_pmresr(&vval
, &fval
);
1663 val
= venum_read_pmresr();
1664 val
= clear_pmresrn_group(val
, group
);
1665 venum_write_pmresr(val
);
1666 venum_post_pmresr(vval
, fval
);
1668 val
= scorpion_read_pmresrn(region
);
1669 val
= clear_pmresrn_group(val
, group
);
1670 scorpion_write_pmresrn(region
, val
);
1674 static void scorpion_pmu_disable_event(struct perf_event
*event
)
1676 unsigned long flags
;
1677 struct hw_perf_event
*hwc
= &event
->hw
;
1679 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1680 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
1682 /* Disable counter and interrupt */
1683 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
1685 /* Disable counter */
1686 armv7_pmnc_disable_counter(idx
);
1689 * Clear pmresr code (if destined for PMNx counters)
1691 if (hwc
->config_base
& KRAIT_EVENT_MASK
)
1692 scorpion_clearpmu(hwc
->config_base
);
1694 /* Disable interrupt for this counter */
1695 armv7_pmnc_disable_intens(idx
);
1697 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
1700 static void scorpion_pmu_enable_event(struct perf_event
*event
)
1702 unsigned long flags
;
1703 struct hw_perf_event
*hwc
= &event
->hw
;
1705 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1706 struct pmu_hw_events
*events
= this_cpu_ptr(cpu_pmu
->hw_events
);
1709 * Enable counter and interrupt, and set the counter to count
1710 * the event that we're interested in.
1712 raw_spin_lock_irqsave(&events
->pmu_lock
, flags
);
1714 /* Disable counter */
1715 armv7_pmnc_disable_counter(idx
);
1718 * Set event (if destined for PMNx counters)
1719 * We don't set the event for the cycle counter because we
1720 * don't have the ability to perform event filtering.
1722 if (hwc
->config_base
& KRAIT_EVENT_MASK
)
1723 scorpion_evt_setup(idx
, hwc
->config_base
);
1724 else if (idx
!= ARMV7_IDX_CYCLE_COUNTER
)
1725 armv7_pmnc_write_evtsel(idx
, hwc
->config_base
);
1727 /* Enable interrupt for this counter */
1728 armv7_pmnc_enable_intens(idx
);
1730 /* Enable counter */
1731 armv7_pmnc_enable_counter(idx
);
1733 raw_spin_unlock_irqrestore(&events
->pmu_lock
, flags
);
1736 static void scorpion_pmu_reset(void *info
)
1739 struct arm_pmu
*cpu_pmu
= info
;
1740 u32 idx
, nb_cnt
= cpu_pmu
->num_events
;
1742 armv7pmu_reset(info
);
1744 /* Clear all pmresrs */
1745 scorpion_write_pmresrn(0, 0);
1746 scorpion_write_pmresrn(1, 0);
1747 scorpion_write_pmresrn(2, 0);
1748 scorpion_write_pmresrn(3, 0);
1750 venum_pre_pmresr(&vval
, &fval
);
1751 venum_write_pmresr(0);
1752 venum_post_pmresr(vval
, fval
);
1754 /* Reset PMxEVNCTCR to sane default */
1755 for (idx
= ARMV7_IDX_CYCLE_COUNTER
; idx
< nb_cnt
; ++idx
) {
1756 armv7_pmnc_select_counter(idx
);
1757 asm volatile("mcr p15, 0, %0, c9, c15, 0" : : "r" (0));
1761 static int scorpion_event_to_bit(struct perf_event
*event
, unsigned int region
,
1765 struct hw_perf_event
*hwc
= &event
->hw
;
1766 struct arm_pmu
*cpu_pmu
= to_arm_pmu(event
->pmu
);
1768 if (hwc
->config_base
& VENUM_EVENT
)
1769 bit
= SCORPION_VLPM_GROUP0
;
1771 bit
= scorpion_get_pmresrn_event(region
);
1772 bit
-= scorpion_get_pmresrn_event(0);
1775 * Lower bits are reserved for use by the counters (see
1776 * armv7pmu_get_event_idx() for more info)
1778 bit
+= ARMV7_IDX_COUNTER_LAST(cpu_pmu
) + 1;
1784 * We check for column exclusion constraints here.
1785 * Two events cant use the same group within a pmresr register.
1787 static int scorpion_pmu_get_event_idx(struct pmu_hw_events
*cpuc
,
1788 struct perf_event
*event
)
1792 struct hw_perf_event
*hwc
= &event
->hw
;
1793 unsigned int region
= EVENT_REGION(hwc
->config_base
);
1794 unsigned int group
= EVENT_GROUP(hwc
->config_base
);
1795 bool venum_event
= EVENT_VENUM(hwc
->config_base
);
1796 bool scorpion_event
= EVENT_CPU(hwc
->config_base
);
1798 if (venum_event
|| scorpion_event
) {
1799 /* Ignore invalid events */
1800 if (group
> 3 || region
> 3)
1803 bit
= scorpion_event_to_bit(event
, region
, group
);
1804 if (test_and_set_bit(bit
, cpuc
->used_mask
))
1808 idx
= armv7pmu_get_event_idx(cpuc
, event
);
1809 if (idx
< 0 && bit
>= 0)
1810 clear_bit(bit
, cpuc
->used_mask
);
1815 static void scorpion_pmu_clear_event_idx(struct pmu_hw_events
*cpuc
,
1816 struct perf_event
*event
)
1819 struct hw_perf_event
*hwc
= &event
->hw
;
1820 unsigned int region
= EVENT_REGION(hwc
->config_base
);
1821 unsigned int group
= EVENT_GROUP(hwc
->config_base
);
1822 bool venum_event
= EVENT_VENUM(hwc
->config_base
);
1823 bool scorpion_event
= EVENT_CPU(hwc
->config_base
);
1825 if (venum_event
|| scorpion_event
) {
1826 bit
= scorpion_event_to_bit(event
, region
, group
);
1827 clear_bit(bit
, cpuc
->used_mask
);
1831 static int scorpion_pmu_init(struct arm_pmu
*cpu_pmu
)
1833 armv7pmu_init(cpu_pmu
);
1834 cpu_pmu
->name
= "armv7_scorpion";
1835 cpu_pmu
->map_event
= scorpion_map_event
;
1836 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1837 cpu_pmu
->reset
= scorpion_pmu_reset
;
1838 cpu_pmu
->enable
= scorpion_pmu_enable_event
;
1839 cpu_pmu
->disable
= scorpion_pmu_disable_event
;
1840 cpu_pmu
->get_event_idx
= scorpion_pmu_get_event_idx
;
1841 cpu_pmu
->clear_event_idx
= scorpion_pmu_clear_event_idx
;
1845 static int scorpion_mp_pmu_init(struct arm_pmu
*cpu_pmu
)
1847 armv7pmu_init(cpu_pmu
);
1848 cpu_pmu
->name
= "armv7_scorpion_mp";
1849 cpu_pmu
->map_event
= scorpion_map_event
;
1850 cpu_pmu
->num_events
= armv7_read_num_pmnc_events();
1851 cpu_pmu
->reset
= scorpion_pmu_reset
;
1852 cpu_pmu
->enable
= scorpion_pmu_enable_event
;
1853 cpu_pmu
->disable
= scorpion_pmu_disable_event
;
1854 cpu_pmu
->get_event_idx
= scorpion_pmu_get_event_idx
;
1855 cpu_pmu
->clear_event_idx
= scorpion_pmu_clear_event_idx
;
1859 static inline int armv7_a8_pmu_init(struct arm_pmu
*cpu_pmu
)
1864 static inline int armv7_a9_pmu_init(struct arm_pmu
*cpu_pmu
)
1869 static inline int armv7_a5_pmu_init(struct arm_pmu
*cpu_pmu
)
1874 static inline int armv7_a15_pmu_init(struct arm_pmu
*cpu_pmu
)
1879 static inline int armv7_a7_pmu_init(struct arm_pmu
*cpu_pmu
)
1884 static inline int armv7_a12_pmu_init(struct arm_pmu
*cpu_pmu
)
1889 static inline int armv7_a17_pmu_init(struct arm_pmu
*cpu_pmu
)
1894 static inline int krait_pmu_init(struct arm_pmu
*cpu_pmu
)
1899 static inline int scorpion_pmu_init(struct arm_pmu
*cpu_pmu
)
1904 static inline int scorpion_mp_pmu_init(struct arm_pmu
*cpu_pmu
)
1908 #endif /* CONFIG_CPU_V7 */