13 #include <asm/irq_regs.h>
17 #define M_PERFCTL_EXL (1UL << 0)
18 #define M_PERFCTL_KERNEL (1UL << 1)
19 #define M_PERFCTL_SUPERVISOR (1UL << 2)
20 #define M_PERFCTL_USER (1UL << 3)
21 #define M_PERFCTL_INTERRUPT_ENABLE (1UL << 4)
22 #define M_PERFCTL_EVENT(event) (((event) & 0x3ff) << 5)
23 #define M_PERFCTL_VPEID(vpe) ((vpe) << 16)
24 #define M_PERFCTL_MT_EN(filter) ((filter) << 20)
25 #define M_TC_EN_ALL M_PERFCTL_MT_EN(0)
26 #define M_TC_EN_VPE M_PERFCTL_MT_EN(1)
27 #define M_TC_EN_TC M_PERFCTL_MT_EN(2)
28 #define M_PERFCTL_TCID(tcid) ((tcid) << 22)
29 #define M_PERFCTL_WIDE (1UL << 30)
30 #define M_PERFCTL_MORE (1UL << 31)
32 #define M_COUNTER_OVERFLOW (1UL << 31)
34 static int (*save_perf_irq)(
void);
36 #ifdef CONFIG_MIPS_MT_SMP
37 static int cpu_has_mipsmt_pertccounters;
38 #define WHAT (M_TC_EN_VPE | \
39 M_PERFCTL_VPEID(cpu_data[smp_processor_id()].vpe_id))
40 #define vpe_id() (cpu_has_mipsmt_pertccounters ? \
41 0 : cpu_data[smp_processor_id()].vpe_id)
51 static inline unsigned int vpe_shift(
void)
64 static inline unsigned int vpe_shift(
void)
71 static inline unsigned int counters_total_to_per_cpu(
unsigned int counters)
73 return counters >> vpe_shift();
76 static inline unsigned int counters_per_cpu_to_total(
unsigned int counters)
78 return counters << vpe_shift();
81 #define __define_perf_accessors(r, n, np) \
83 static inline unsigned int r_c0_ ## r ## n(void) \
85 unsigned int cpu = vpe_id(); \
89 return read_c0_ ## r ## n(); \
91 return read_c0_ ## r ## np(); \
98 static inline void w_c0_ ## r ## n(unsigned int value) \
100 unsigned int cpu = vpe_id(); \
104 write_c0_ ## r ## n(value); \
107 write_c0_ ## r ## np(value); \
127 static
struct mipsxx_register_config {
140 for (i = 0; i < counters; i++) {
161 static void mipsxx_cpu_setup(
void *args)
168 w_c0_perfcntr3(
reg.counter[3]);
171 w_c0_perfcntr2(
reg.counter[2]);
174 w_c0_perfcntr1(
reg.counter[1]);
177 w_c0_perfcntr0(
reg.counter[0]);
182 static void mipsxx_cpu_start(
void *args)
188 w_c0_perfctrl3(
WHAT |
reg.control[3]);
190 w_c0_perfctrl2(
WHAT |
reg.control[2]);
192 w_c0_perfctrl1(
WHAT |
reg.control[1]);
194 w_c0_perfctrl0(
WHAT |
reg.control[0]);
199 static void mipsxx_cpu_stop(
void *args)
215 static int mipsxx_perfcount_handler(
void)
226 #define HANDLE_COUNTER(n) \
228 control = r_c0_perfctrl ## n(); \
229 counter = r_c0_perfcntr ## n(); \
230 if ((control & M_PERFCTL_INTERRUPT_ENABLE) && \
231 (counter & M_COUNTER_OVERFLOW)) { \
232 oprofile_add_sample(get_irq_regs(), n); \
233 w_c0_perfcntr ## n(reg.counter[n]); \
234 handled = IRQ_HANDLED; \
245 #define M_CONFIG1_PC (1 << 4)
247 static inline int __n_counters(
void)
261 static inline int n_counters(
void)
276 counters = __n_counters();
282 static void reset_counters(
void *
arg)
284 int counters = (
int)(
long)
arg;
303 return mipsxx_perfcount_handler();
306 static int __init mipsxx_init(
void)
310 counters = n_counters();
316 #ifdef CONFIG_MIPS_MT_SMP
318 if (!cpu_has_mipsmt_pertccounters)
319 counters = counters_total_to_per_cpu(counters);
321 on_each_cpu(reset_counters, (
void *)(
long)counters, 1);
382 perf_irq = mipsxx_perfcount_handler;
386 0,
"Perfcounter", save_perf_irq);
391 static void mipsxx_exit(
void)
398 counters = counters_per_cpu_to_total(counters);
399 on_each_cpu(reset_counters, (
void *)(
long)counters, 1);
405 .reg_setup = mipsxx_reg_setup,
406 .cpu_setup = mipsxx_cpu_setup,
409 .cpu_start = mipsxx_cpu_start,
410 .cpu_stop = mipsxx_cpu_stop,