1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef __ASMARM_ARCH_TIMER_H
3 #define __ASMARM_ARCH_TIMER_H
5 #include <asm/barrier.h>
8 #include <linux/clocksource.h>
9 #include <linux/init.h>
10 #include <linux/types.h>
12 #include <clocksource/arm_arch_timer.h>
14 #ifdef CONFIG_ARM_ARCH_TIMER
15 /* 32bit ARM doesn't know anything about timer errata... */
16 #define has_erratum_handler(h) (false)
17 #define erratum_handler(h) (arch_timer_##h)
19 int arch_timer_arch_init(void);
22 * These register accessors are marked inline so the compiler can
23 * nicely work out which register we want, and chuck away the rest of
24 * the code. At least it does so with a recent GCC (4.6.3).
26 static __always_inline
27 void arch_timer_reg_write_cp15(int access
, enum arch_timer_reg reg
, u32 val
)
29 if (access
== ARCH_TIMER_PHYS_ACCESS
) {
31 case ARCH_TIMER_REG_CTRL
:
32 asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" (val
));
34 case ARCH_TIMER_REG_TVAL
:
35 asm volatile("mcr p15, 0, %0, c14, c2, 0" : : "r" (val
));
38 } else if (access
== ARCH_TIMER_VIRT_ACCESS
) {
40 case ARCH_TIMER_REG_CTRL
:
41 asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" (val
));
43 case ARCH_TIMER_REG_TVAL
:
44 asm volatile("mcr p15, 0, %0, c14, c3, 0" : : "r" (val
));
52 static __always_inline
53 u32
arch_timer_reg_read_cp15(int access
, enum arch_timer_reg reg
)
57 if (access
== ARCH_TIMER_PHYS_ACCESS
) {
59 case ARCH_TIMER_REG_CTRL
:
60 asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val
));
62 case ARCH_TIMER_REG_TVAL
:
63 asm volatile("mrc p15, 0, %0, c14, c2, 0" : "=r" (val
));
66 } else if (access
== ARCH_TIMER_VIRT_ACCESS
) {
68 case ARCH_TIMER_REG_CTRL
:
69 asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val
));
71 case ARCH_TIMER_REG_TVAL
:
72 asm volatile("mrc p15, 0, %0, c14, c3, 0" : "=r" (val
));
80 static inline u32
arch_timer_get_cntfrq(void)
83 asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val
));
87 static inline u64
__arch_counter_get_cntpct(void)
92 asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval
));
96 static inline u64
__arch_counter_get_cntpct_stable(void)
98 return __arch_counter_get_cntpct();
101 static inline u64
__arch_counter_get_cntvct(void)
106 asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval
));
110 static inline u64
__arch_counter_get_cntvct_stable(void)
112 return __arch_counter_get_cntvct();
115 static inline u32
arch_timer_get_cntkctl(void)
118 asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl
));
122 static inline void arch_timer_set_cntkctl(u32 cntkctl
)
124 asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl
));
128 static inline void arch_timer_set_evtstrm_feature(void)
130 elf_hwcap
|= HWCAP_EVTSTRM
;
133 static inline bool arch_timer_have_evtstrm_feature(void)
135 return elf_hwcap
& HWCAP_EVTSTRM
;