2 * sched_clock.h: support for extending counters to full 64-bit ns counter
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
8 #ifndef ASM_SCHED_CLOCK
9 #define ASM_SCHED_CLOCK
11 #include <linux/kernel.h>
12 #include <linux/types.h>
22 #define DEFINE_CLOCK_DATA(name) struct clock_data name
24 static inline u64
cyc_to_ns(u64 cyc
, u32 mult
, u32 shift
)
26 return (cyc
* mult
) >> shift
;
30 * Atomically update the sched_clock epoch. Your update callback will
31 * be called from a timer before the counter wraps - read the current
32 * counter value, and call this function to safely move the epochs
33 * forward. Only use this from the update callback.
35 static inline void update_sched_clock(struct clock_data
*cd
, u32 cyc
, u32 mask
)
38 u64 ns
= cd
->epoch_ns
+
39 cyc_to_ns((cyc
- cd
->epoch_cyc
) & mask
, cd
->mult
, cd
->shift
);
42 * Write epoch_cyc and epoch_ns in a way that the update is
43 * detectable in cyc_to_fixed_sched_clock().
45 raw_local_irq_save(flags
);
50 cd
->epoch_cyc_copy
= cyc
;
51 raw_local_irq_restore(flags
);
55 * If your clock rate is known at compile time, using this will allow
56 * you to optimize the mult/shift loads away. This is paired with
57 * init_fixed_sched_clock() to ensure that your mult/shift are correct.
59 static inline unsigned long long cyc_to_fixed_sched_clock(struct clock_data
*cd
,
60 u32 cyc
, u32 mask
, u32 mult
, u32 shift
)
66 * Load the epoch_cyc and epoch_ns atomically. We do this by
67 * ensuring that we always write epoch_cyc, epoch_ns and
68 * epoch_cyc_copy in strict order, and read them in strict order.
69 * If epoch_cyc and epoch_cyc_copy are not equal, then we're in
70 * the middle of an update, and we should repeat the load.
73 epoch_cyc
= cd
->epoch_cyc
;
75 epoch_ns
= cd
->epoch_ns
;
77 } while (epoch_cyc
!= cd
->epoch_cyc_copy
);
79 return epoch_ns
+ cyc_to_ns((cyc
- epoch_cyc
) & mask
, mult
, shift
);
83 * Otherwise, you need to use this, which will obtain the mult/shift
84 * from the clock_data structure. Use init_sched_clock() with this.
86 static inline unsigned long long cyc_to_sched_clock(struct clock_data
*cd
,
89 return cyc_to_fixed_sched_clock(cd
, cyc
, mask
, cd
->mult
, cd
->shift
);
93 * Initialize the clock data - calculate the appropriate multiplier
94 * and shift. Also setup a timer to ensure that the epoch is refreshed
95 * at the appropriate time interval, which will call your update
98 void init_sched_clock(struct clock_data
*, void (*)(void),
99 unsigned int, unsigned long);
102 * Use this initialization function rather than init_sched_clock() if
103 * you're using cyc_to_fixed_sched_clock, which will warn if your
104 * constants are incorrect.
106 static inline void init_fixed_sched_clock(struct clock_data
*cd
,
107 void (*update
)(void), unsigned int bits
, unsigned long rate
,
110 init_sched_clock(cd
, update
, bits
, rate
);
111 if (cd
->mult
!= mult
|| cd
->shift
!= shift
) {
112 pr_crit("sched_clock: wrong multiply/shift: %u>>%u vs calculated %u>>%u\n"
113 "sched_clock: fix multiply/shift to avoid scheduler hiccups\n",
114 mult
, shift
, cd
->mult
, cd
->shift
);
118 extern void sched_clock_postinit(void);