2 * @file op_model_xscale.c
3 * XScale Performance Monitor Driver
5 * @remark Copyright 2000-2004 Deepak Saxena <dsaxena@mvista.com>
6 * @remark Copyright 2000-2004 MontaVista Software Inc
7 * @remark Copyright 2004 Dave Jiang <dave.jiang@intel.com>
8 * @remark Copyright 2004 Intel Corporation
9 * @remark Copyright 2004 Zwane Mwaikambo <zwane@arm.linux.org.uk>
10 * @remark Copyright 2004 OProfile Authors
12 * @remark Read the file COPYING
14 * @author Zwane Mwaikambo
18 #include <linux/types.h>
19 #include <linux/errno.h>
20 #include <linux/sched.h>
21 #include <linux/oprofile.h>
22 #include <linux/interrupt.h>
24 #include <asm/system.h>
26 #include "op_counter.h"
27 #include "op_arm_model.h"
29 #define PMU_ENABLE 0x001 /* Enable counters */
30 #define PMN_RESET 0x002 /* Reset event counters */
31 #define CCNT_RESET 0x004 /* Reset clock counter */
32 #define PMU_RESET (CCNT_RESET | PMN_RESET)
33 #define PMU_CNT64 0x008 /* Make CCNT count every 64th cycle */
35 /* TODO do runtime detection */
36 #ifdef CONFIG_ARCH_IOP32X
37 #define XSCALE_PMU_IRQ IRQ_IOP32X_CORE_PMU
39 #ifdef CONFIG_ARCH_IOP33X
40 #define XSCALE_PMU_IRQ IRQ_IOP33X_CORE_PMU
42 #ifdef CONFIG_ARCH_PXA
43 #define XSCALE_PMU_IRQ IRQ_PMU
47 * Different types of events that can be counted by the XScale PMU
48 * as used by Oprofile userspace. Here primarily for documentation
52 #define EVT_ICACHE_MISS 0x00
53 #define EVT_ICACHE_NO_DELIVER 0x01
54 #define EVT_DATA_STALL 0x02
55 #define EVT_ITLB_MISS 0x03
56 #define EVT_DTLB_MISS 0x04
57 #define EVT_BRANCH 0x05
58 #define EVT_BRANCH_MISS 0x06
59 #define EVT_INSTRUCTION 0x07
60 #define EVT_DCACHE_FULL_STALL 0x08
61 #define EVT_DCACHE_FULL_STALL_CONTIG 0x09
62 #define EVT_DCACHE_ACCESS 0x0A
63 #define EVT_DCACHE_MISS 0x0B
64 #define EVT_DCACE_WRITE_BACK 0x0C
65 #define EVT_PC_CHANGED 0x0D
66 #define EVT_BCU_REQUEST 0x10
67 #define EVT_BCU_FULL 0x11
68 #define EVT_BCU_DRAIN 0x12
69 #define EVT_BCU_ECC_NO_ELOG 0x14
70 #define EVT_BCU_1_BIT_ERR 0x15
72 /* EVT_CCNT is not hardware defined */
74 #define EVT_UNUSED 0xFF
77 volatile unsigned long ovf
;
78 unsigned long reset_counter
;
81 enum { CCNT
, PMN0
, PMN1
, PMN2
, PMN3
, MAX_COUNTERS
};
83 static struct pmu_counter results
[MAX_COUNTERS
];
86 * There are two versions of the PMU in current XScale processors
87 * with differing register layouts and number of performance counters.
88 * e.g. IOP32x is xsc1 whilst IOP33x is xsc2.
89 * We detect which register layout to use in xscale_detect_pmu()
91 enum { PMU_XSC1
, PMU_XSC2
};
97 unsigned int int_enable
;
98 unsigned int cnt_ovf
[MAX_COUNTERS
];
99 unsigned int int_mask
[MAX_COUNTERS
];
102 static struct pmu_type pmu_parms
[] = {
105 .name
= "arm/xscale1",
107 .int_mask
= { [PMN0
] = 0x10, [PMN1
] = 0x20,
109 .cnt_ovf
= { [CCNT
] = 0x400, [PMN0
] = 0x100,
114 .name
= "arm/xscale2",
116 .int_mask
= { [CCNT
] = 0x01, [PMN0
] = 0x02,
117 [PMN1
] = 0x04, [PMN2
] = 0x08,
119 .cnt_ovf
= { [CCNT
] = 0x01, [PMN0
] = 0x02,
120 [PMN1
] = 0x04, [PMN2
] = 0x08,
125 static struct pmu_type
*pmu
;
127 static void write_pmnc(u32 val
)
129 if (pmu
->id
== PMU_XSC1
) {
130 /* upper 4bits and 7, 11 are write-as-0 */
132 __asm__
__volatile__ ("mcr p14, 0, %0, c0, c0, 0" : : "r" (val
));
134 /* bits 4-23 are write-as-0, 24-31 are write ignored */
136 __asm__
__volatile__ ("mcr p14, 0, %0, c0, c1, 0" : : "r" (val
));
140 static u32
read_pmnc(void)
144 if (pmu
->id
== PMU_XSC1
)
145 __asm__
__volatile__ ("mrc p14, 0, %0, c0, c0, 0" : "=r" (val
));
147 __asm__
__volatile__ ("mrc p14, 0, %0, c0, c1, 0" : "=r" (val
));
148 /* bits 1-2 and 4-23 are read-unpredictable */
155 static u32
__xsc1_read_counter(int counter
)
161 __asm__
__volatile__ ("mrc p14, 0, %0, c1, c0, 0" : "=r" (val
));
164 __asm__
__volatile__ ("mrc p14, 0, %0, c2, c0, 0" : "=r" (val
));
167 __asm__
__volatile__ ("mrc p14, 0, %0, c3, c0, 0" : "=r" (val
));
173 static u32
__xsc2_read_counter(int counter
)
179 __asm__
__volatile__ ("mrc p14, 0, %0, c1, c1, 0" : "=r" (val
));
182 __asm__
__volatile__ ("mrc p14, 0, %0, c0, c2, 0" : "=r" (val
));
185 __asm__
__volatile__ ("mrc p14, 0, %0, c1, c2, 0" : "=r" (val
));
188 __asm__
__volatile__ ("mrc p14, 0, %0, c2, c2, 0" : "=r" (val
));
191 __asm__
__volatile__ ("mrc p14, 0, %0, c3, c2, 0" : "=r" (val
));
197 static u32
read_counter(int counter
)
201 if (pmu
->id
== PMU_XSC1
)
202 val
= __xsc1_read_counter(counter
);
204 val
= __xsc2_read_counter(counter
);
209 static void __xsc1_write_counter(int counter
, u32 val
)
213 __asm__
__volatile__ ("mcr p14, 0, %0, c1, c0, 0" : : "r" (val
));
216 __asm__
__volatile__ ("mcr p14, 0, %0, c2, c0, 0" : : "r" (val
));
219 __asm__
__volatile__ ("mcr p14, 0, %0, c3, c0, 0" : : "r" (val
));
224 static void __xsc2_write_counter(int counter
, u32 val
)
228 __asm__
__volatile__ ("mcr p14, 0, %0, c1, c1, 0" : : "r" (val
));
231 __asm__
__volatile__ ("mcr p14, 0, %0, c0, c2, 0" : : "r" (val
));
234 __asm__
__volatile__ ("mcr p14, 0, %0, c1, c2, 0" : : "r" (val
));
237 __asm__
__volatile__ ("mcr p14, 0, %0, c2, c2, 0" : : "r" (val
));
240 __asm__
__volatile__ ("mcr p14, 0, %0, c3, c2, 0" : : "r" (val
));
245 static void write_counter(int counter
, u32 val
)
247 if (pmu
->id
== PMU_XSC1
)
248 __xsc1_write_counter(counter
, val
);
250 __xsc2_write_counter(counter
, val
);
253 static int xscale_setup_ctrs(void)
258 for (i
= CCNT
; i
< MAX_COUNTERS
; i
++) {
259 if (counter_config
[i
].enabled
)
262 counter_config
[i
].event
= EVT_UNUSED
;
267 pmnc
= (counter_config
[PMN1
].event
<< 20) | (counter_config
[PMN0
].event
<< 12);
268 pr_debug("xscale_setup_ctrs: pmnc: %#08x\n", pmnc
);
273 evtsel
= counter_config
[PMN0
].event
| (counter_config
[PMN1
].event
<< 8) |
274 (counter_config
[PMN2
].event
<< 16) | (counter_config
[PMN3
].event
<< 24);
276 pr_debug("xscale_setup_ctrs: evtsel %#08x\n", evtsel
);
277 __asm__
__volatile__ ("mcr p14, 0, %0, c8, c1, 0" : : "r" (evtsel
));
281 for (i
= CCNT
; i
< MAX_COUNTERS
; i
++) {
282 if (counter_config
[i
].event
== EVT_UNUSED
) {
283 counter_config
[i
].event
= 0;
284 pmu
->int_enable
&= ~pmu
->int_mask
[i
];
288 results
[i
].reset_counter
= counter_config
[i
].count
;
289 write_counter(i
, -(u32
)counter_config
[i
].count
);
290 pmu
->int_enable
|= pmu
->int_mask
[i
];
291 pr_debug("xscale_setup_ctrs: counter%d %#08x from %#08lx\n", i
,
292 read_counter(i
), counter_config
[i
].count
);
298 static void inline __xsc1_check_ctrs(void)
301 u32 pmnc
= read_pmnc();
303 /* NOTE: there's an A stepping errata that states if an overflow */
304 /* bit already exists and another occurs, the previous */
305 /* Overflow bit gets cleared. There's no workaround. */
306 /* Fixed in B stepping or later */
308 /* Write the value back to clear the overflow flags. Overflow */
309 /* flags remain in pmnc for use below */
310 write_pmnc(pmnc
& ~PMU_ENABLE
);
312 for (i
= CCNT
; i
<= PMN1
; i
++) {
313 if (!(pmu
->int_mask
[i
] & pmu
->int_enable
))
316 if (pmnc
& pmu
->cnt_ovf
[i
])
321 static void inline __xsc2_check_ctrs(void)
324 u32 flag
= 0, pmnc
= read_pmnc();
329 /* read overflow flag register */
330 __asm__
__volatile__ ("mrc p14, 0, %0, c5, c1, 0" : "=r" (flag
));
332 for (i
= CCNT
; i
<= PMN3
; i
++) {
333 if (!(pmu
->int_mask
[i
] & pmu
->int_enable
))
336 if (flag
& pmu
->cnt_ovf
[i
])
340 /* writeback clears overflow bits */
341 __asm__
__volatile__ ("mcr p14, 0, %0, c5, c1, 0" : : "r" (flag
));
344 static irqreturn_t
xscale_pmu_interrupt(int irq
, void *arg
, struct pt_regs
*regs
)
349 if (pmu
->id
== PMU_XSC1
)
354 for (i
= CCNT
; i
< MAX_COUNTERS
; i
++) {
358 write_counter(i
, -(u32
)results
[i
].reset_counter
);
359 oprofile_add_sample(regs
, i
);
363 pmnc
= read_pmnc() | PMU_ENABLE
;
369 static void xscale_pmu_stop(void)
371 u32 pmnc
= read_pmnc();
376 free_irq(XSCALE_PMU_IRQ
, results
);
379 static int xscale_pmu_start(void)
382 u32 pmnc
= read_pmnc();
384 ret
= request_irq(XSCALE_PMU_IRQ
, xscale_pmu_interrupt
, IRQF_DISABLED
,
385 "XScale PMU", (void *)results
);
388 printk(KERN_ERR
"oprofile: unable to request IRQ%d for XScale PMU\n",
393 if (pmu
->id
== PMU_XSC1
)
394 pmnc
|= pmu
->int_enable
;
396 __asm__
__volatile__ ("mcr p14, 0, %0, c4, c1, 0" : : "r" (pmu
->int_enable
));
402 pr_debug("xscale_pmu_start: pmnc: %#08x mask: %08x\n", pmnc
, pmu
->int_enable
);
406 static int xscale_detect_pmu(void)
411 id
= (read_cpuid(CPUID_ID
) >> 13) & 0x7;
415 pmu
= &pmu_parms
[PMU_XSC1
];
418 pmu
= &pmu_parms
[PMU_XSC2
];
426 op_xscale_spec
.name
= pmu
->name
;
427 op_xscale_spec
.num_counters
= pmu
->num_counters
;
428 pr_debug("xscale_detect_pmu: detected %s PMU\n", pmu
->name
);
434 struct op_arm_model_spec op_xscale_spec
= {
435 .init
= xscale_detect_pmu
,
436 .setup_ctrs
= xscale_setup_ctrs
,
437 .start
= xscale_pmu_start
,
438 .stop
= xscale_pmu_stop
,