2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 2004, 05, 06 by Ralf Baechle
7 * Copyright (C) 2005 by MIPS Technologies, Inc.
9 #include <linux/cpumask.h>
10 #include <linux/oprofile.h>
11 #include <linux/interrupt.h>
12 #include <linux/smp.h>
13 #include <asm/irq_regs.h>
18 #define M_PERFCTL_EVENT(event) (((event) << MIPS_PERFCTRL_EVENT_S) & \
20 #define M_PERFCTL_VPEID(vpe) ((vpe) << MIPS_PERFCTRL_VPEID_S)
22 #define M_COUNTER_OVERFLOW (1UL << 31)
24 static int (*save_perf_irq
)(void);
25 static int perfcount_irq
;
28 * XLR has only one set of counters per core. Designate the
29 * first hardware thread in the core for setup and init.
30 * Skip CPUs with non-zero hardware thread id (4 hwt per core)
32 #if defined(CONFIG_CPU_XLR) && defined(CONFIG_SMP)
33 #define oprofile_skip_cpu(c) ((cpu_logical_map(c) & 0x3) != 0)
35 #define oprofile_skip_cpu(c) 0
38 #ifdef CONFIG_MIPS_MT_SMP
39 static int cpu_has_mipsmt_pertccounters
;
40 #define WHAT (MIPS_PERFCTRL_MT_EN_VPE | \
41 M_PERFCTL_VPEID(cpu_vpe_id(¤t_cpu_data)))
42 #define vpe_id() (cpu_has_mipsmt_pertccounters ? \
43 0 : cpu_vpe_id(¤t_cpu_data))
46 * The number of bits to shift to convert between counters per core and
47 * counters per VPE. There is no reasonable interface atm to obtain the
48 * number of VPEs used by Linux and in the 34K this number is fixed to two
49 * anyways so we hardcore a few things here for the moment. The way it's
50 * done here will ensure that oprofile VSMP kernel will run right on a lesser
51 * core like a 24K also or with maxcpus=1.
53 static inline unsigned int vpe_shift(void)
55 if (num_possible_cpus() > 1)
66 static inline unsigned int vpe_shift(void)
73 static inline unsigned int counters_total_to_per_cpu(unsigned int counters
)
75 return counters
>> vpe_shift();
78 static inline unsigned int counters_per_cpu_to_total(unsigned int counters
)
80 return counters
<< vpe_shift();
83 #define __define_perf_accessors(r, n, np) \
85 static inline unsigned int r_c0_ ## r ## n(void) \
87 unsigned int cpu = vpe_id(); \
91 return read_c0_ ## r ## n(); \
93 return read_c0_ ## r ## np(); \
100 static inline void w_c0_ ## r ## n(unsigned int value) \
102 unsigned int cpu = vpe_id(); \
106 write_c0_ ## r ## n(value); \
109 write_c0_ ## r ## np(value); \
117 __define_perf_accessors(perfcntr, 0, 2)
118 __define_perf_accessors(perfcntr
, 1, 3)
119 __define_perf_accessors(perfcntr
, 2, 0)
120 __define_perf_accessors(perfcntr
, 3, 1)
122 __define_perf_accessors(perfctrl
, 0, 2)
123 __define_perf_accessors(perfctrl
, 1, 3)
124 __define_perf_accessors(perfctrl
, 2, 0)
125 __define_perf_accessors(perfctrl
, 3, 1)
127 struct op_mips_model op_model_mipsxx_ops
;
129 static struct mipsxx_register_config
{
130 unsigned int control
[4];
131 unsigned int counter
[4];
134 /* Compute all of the registers in preparation for enabling profiling. */
136 static void mipsxx_reg_setup(struct op_counter_config
*ctr
)
138 unsigned int counters
= op_model_mipsxx_ops
.num_counters
;
141 /* Compute the performance counter control word. */
142 for (i
= 0; i
< counters
; i
++) {
149 reg
.control
[i
] = M_PERFCTL_EVENT(ctr
[i
].event
) |
152 reg
.control
[i
] |= MIPS_PERFCTRL_K
;
154 reg
.control
[i
] |= MIPS_PERFCTRL_U
;
156 reg
.control
[i
] |= MIPS_PERFCTRL_EXL
;
157 if (boot_cpu_type() == CPU_XLR
)
158 reg
.control
[i
] |= XLR_PERFCTRL_ALLTHREADS
;
159 reg
.counter
[i
] = 0x80000000 - ctr
[i
].count
;
163 /* Program all of the registers in preparation for enabling profiling. */
165 static void mipsxx_cpu_setup(void *args
)
167 unsigned int counters
= op_model_mipsxx_ops
.num_counters
;
169 if (oprofile_skip_cpu(smp_processor_id()))
175 w_c0_perfcntr3(reg
.counter
[3]);
178 w_c0_perfcntr2(reg
.counter
[2]);
181 w_c0_perfcntr1(reg
.counter
[1]);
184 w_c0_perfcntr0(reg
.counter
[0]);
188 /* Start all counters on current CPU */
189 static void mipsxx_cpu_start(void *args
)
191 unsigned int counters
= op_model_mipsxx_ops
.num_counters
;
193 if (oprofile_skip_cpu(smp_processor_id()))
198 w_c0_perfctrl3(WHAT
| reg
.control
[3]);
200 w_c0_perfctrl2(WHAT
| reg
.control
[2]);
202 w_c0_perfctrl1(WHAT
| reg
.control
[1]);
204 w_c0_perfctrl0(WHAT
| reg
.control
[0]);
208 /* Stop all counters on current CPU */
209 static void mipsxx_cpu_stop(void *args
)
211 unsigned int counters
= op_model_mipsxx_ops
.num_counters
;
213 if (oprofile_skip_cpu(smp_processor_id()))
228 static int mipsxx_perfcount_handler(void)
230 unsigned int counters
= op_model_mipsxx_ops
.num_counters
;
231 unsigned int control
;
232 unsigned int counter
;
233 int handled
= IRQ_NONE
;
235 if (cpu_has_mips_r2
&& !(read_c0_cause() & CAUSEF_PCI
))
239 #define HANDLE_COUNTER(n) \
241 control = r_c0_perfctrl ## n(); \
242 counter = r_c0_perfcntr ## n(); \
243 if ((control & MIPS_PERFCTRL_IE) && \
244 (counter & M_COUNTER_OVERFLOW)) { \
245 oprofile_add_sample(get_irq_regs(), n); \
246 w_c0_perfcntr ## n(reg.counter[n]); \
247 handled = IRQ_HANDLED; \
258 static inline int __n_counters(void)
262 if (!(read_c0_perfctrl0() & MIPS_PERFCTRL_M
))
264 if (!(read_c0_perfctrl1() & MIPS_PERFCTRL_M
))
266 if (!(read_c0_perfctrl2() & MIPS_PERFCTRL_M
))
272 static inline int n_counters(void)
276 switch (current_cpu_type()) {
288 counters
= __n_counters();
294 static void reset_counters(void *arg
)
296 int counters
= (int)(long)arg
;
313 static irqreturn_t
mipsxx_perfcount_int(int irq
, void *dev_id
)
315 return mipsxx_perfcount_handler();
318 static int __init
mipsxx_init(void)
322 counters
= n_counters();
324 printk(KERN_ERR
"Oprofile: CPU has no performance counters\n");
328 #ifdef CONFIG_MIPS_MT_SMP
329 cpu_has_mipsmt_pertccounters
= read_c0_config7() & (1<<19);
330 if (!cpu_has_mipsmt_pertccounters
)
331 counters
= counters_total_to_per_cpu(counters
);
333 on_each_cpu(reset_counters
, (void *)(long)counters
, 1);
335 op_model_mipsxx_ops
.num_counters
= counters
;
336 switch (current_cpu_type()) {
338 op_model_mipsxx_ops
.cpu_type
= "mips/M14Kc";
342 op_model_mipsxx_ops
.cpu_type
= "mips/M14KEc";
346 op_model_mipsxx_ops
.cpu_type
= "mips/20K";
350 op_model_mipsxx_ops
.cpu_type
= "mips/24K";
354 op_model_mipsxx_ops
.cpu_type
= "mips/25K";
359 op_model_mipsxx_ops
.cpu_type
= "mips/34K";
364 op_model_mipsxx_ops
.cpu_type
= "mips/74K";
368 op_model_mipsxx_ops
.cpu_type
= "mips/interAptiv";
372 op_model_mipsxx_ops
.cpu_type
= "mips/proAptiv";
376 op_model_mipsxx_ops
.cpu_type
= "mips/P5600";
380 op_model_mipsxx_ops
.cpu_type
= "mips/I6400";
384 op_model_mipsxx_ops
.cpu_type
= "mips/M5150";
388 op_model_mipsxx_ops
.cpu_type
= "mips/5K";
392 if ((current_cpu_data
.processor_id
& 0xff) == 0x20)
393 op_model_mipsxx_ops
.cpu_type
= "mips/r10000-v2.x";
395 op_model_mipsxx_ops
.cpu_type
= "mips/r10000";
400 op_model_mipsxx_ops
.cpu_type
= "mips/r12000";
404 op_model_mipsxx_ops
.cpu_type
= "mips/r16000";
409 op_model_mipsxx_ops
.cpu_type
= "mips/sb1";
413 op_model_mipsxx_ops
.cpu_type
= "mips/loongson1";
417 op_model_mipsxx_ops
.cpu_type
= "mips/xlr";
421 printk(KERN_ERR
"Profiling unsupported for this CPU\n");
426 save_perf_irq
= perf_irq
;
427 perf_irq
= mipsxx_perfcount_handler
;
429 if (get_c0_perfcount_int
)
430 perfcount_irq
= get_c0_perfcount_int();
431 else if (cp0_perfcount_irq
>= 0)
432 perfcount_irq
= MIPS_CPU_IRQ_BASE
+ cp0_perfcount_irq
;
436 if (perfcount_irq
>= 0)
437 return request_irq(perfcount_irq
, mipsxx_perfcount_int
,
438 IRQF_PERCPU
| IRQF_NOBALANCING
|
439 IRQF_NO_THREAD
| IRQF_NO_SUSPEND
|
441 "Perfcounter", save_perf_irq
);
446 static void mipsxx_exit(void)
448 int counters
= op_model_mipsxx_ops
.num_counters
;
450 if (perfcount_irq
>= 0)
451 free_irq(perfcount_irq
, save_perf_irq
);
453 counters
= counters_per_cpu_to_total(counters
);
454 on_each_cpu(reset_counters
, (void *)(long)counters
, 1);
456 perf_irq
= save_perf_irq
;
459 struct op_mips_model op_model_mipsxx_ops
= {
460 .reg_setup
= mipsxx_reg_setup
,
461 .cpu_setup
= mipsxx_cpu_setup
,
464 .cpu_start
= mipsxx_cpu_start
,
465 .cpu_stop
= mipsxx_cpu_stop
,