blob: 653f0e2b6ca9a62a450b876c3cf902ea392430d1 [file] [log] [blame]
john stultz734efb42006-06-26 00:25:05 -07001/* linux/include/linux/clocksource.h
2 *
3 * This file contains the structure definitions for clocksources.
4 *
5 * If you are not a clocksource, or timekeeping code, you should
6 * not be including this file!
7 */
8#ifndef _LINUX_CLOCKSOURCE_H
9#define _LINUX_CLOCKSOURCE_H
10
11#include <linux/types.h>
12#include <linux/timex.h>
13#include <linux/time.h>
14#include <linux/list.h>
Eric Dumazet329c8d82007-05-08 00:27:57 -070015#include <linux/cache.h>
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -080016#include <linux/timer.h>
Martin Schwidefskyf1b82742009-08-14 15:47:21 +020017#include <linux/init.h>
john stultz734efb42006-06-26 00:25:05 -070018#include <asm/div64.h>
19#include <asm/io.h>
20
21/* clocksource cycle base type */
22typedef u64 cycle_t;
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -080023struct clocksource;
Thomas Gleixner09ac3692013-04-25 20:31:44 +000024struct module;
john stultz734efb42006-06-26 00:25:05 -070025
H. Peter Anvinae7bd112011-07-21 13:34:05 -070026#ifdef CONFIG_ARCH_CLOCKSOURCE_DATA
Andy Lutomirski433bd802011-07-13 09:24:13 -040027#include <asm/clocksource.h>
H. Peter Anvinae7bd112011-07-21 13:34:05 -070028#endif
Andy Lutomirski433bd802011-07-13 09:24:13 -040029
john stultz734efb42006-06-26 00:25:05 -070030/**
Patrick Ohlya038a352009-02-12 05:03:34 +000031 * struct cyclecounter - hardware abstraction for a free running counter
32 * Provides completely state-free accessors to the underlying hardware.
33 * Depending on which hardware it reads, the cycle counter may wrap
34 * around quickly. Locking rules (if necessary) have to be defined
35 * by the implementor and user of specific instances of this API.
36 *
37 * @read: returns the current cycle value
38 * @mask: bitmask for two's complement
39 * subtraction of non 64 bit counters,
40 * see CLOCKSOURCE_MASK() helper macro
41 * @mult: cycle to nanosecond multiplier
42 * @shift: cycle to nanosecond divisor (power of two)
43 */
44struct cyclecounter {
45 cycle_t (*read)(const struct cyclecounter *cc);
46 cycle_t mask;
47 u32 mult;
48 u32 shift;
49};
50
51/**
52 * struct timecounter - layer above a %struct cyclecounter which counts nanoseconds
53 * Contains the state needed by timecounter_read() to detect
54 * cycle counter wrap around. Initialize with
55 * timecounter_init(). Also used to convert cycle counts into the
56 * corresponding nanosecond counts with timecounter_cyc2time(). Users
57 * of this code are responsible for initializing the underlying
58 * cycle counter hardware, locking issues and reading the time
59 * more often than the cycle counter wraps around. The nanosecond
60 * counter will only wrap around after ~585 years.
61 *
62 * @cc: the cycle counter used by this instance
63 * @cycle_last: most recent cycle counter value seen by
64 * timecounter_read()
65 * @nsec: continuously increasing count
66 */
67struct timecounter {
68 const struct cyclecounter *cc;
69 cycle_t cycle_last;
70 u64 nsec;
71};
72
73/**
74 * cyclecounter_cyc2ns - converts cycle counter cycles to nanoseconds
Kusanagi Kouichib1b73d02011-12-19 18:13:19 +090075 * @cc: Pointer to cycle counter.
Patrick Ohlya038a352009-02-12 05:03:34 +000076 * @cycles: Cycles
77 *
78 * XXX - This could use some mult_lxl_ll() asm optimization. Same code
79 * as in cyc2ns, but with unsigned result.
80 */
81static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc,
82 cycle_t cycles)
83{
84 u64 ret = (u64)cycles;
85 ret = (ret * cc->mult) >> cc->shift;
86 return ret;
87}
88
89/**
90 * timecounter_init - initialize a time counter
91 * @tc: Pointer to time counter which is to be initialized/reset
92 * @cc: A cycle counter, ready to be used.
93 * @start_tstamp: Arbitrary initial time stamp.
94 *
95 * After this call the current cycle register (roughly) corresponds to
96 * the initial time stamp. Every call to timecounter_read() increments
97 * the time stamp counter by the number of elapsed nanoseconds.
98 */
99extern void timecounter_init(struct timecounter *tc,
100 const struct cyclecounter *cc,
101 u64 start_tstamp);
102
103/**
104 * timecounter_read - return nanoseconds elapsed since timecounter_init()
105 * plus the initial time stamp
106 * @tc: Pointer to time counter.
107 *
108 * In other words, keeps track of time since the same epoch as
109 * the function which generated the initial time stamp.
110 */
111extern u64 timecounter_read(struct timecounter *tc);
112
113/**
114 * timecounter_cyc2time - convert a cycle counter to same
115 * time base as values returned by
116 * timecounter_read()
117 * @tc: Pointer to time counter.
Kusanagi Kouichib1b73d02011-12-19 18:13:19 +0900118 * @cycle_tstamp: a value returned by tc->cc->read()
Patrick Ohlya038a352009-02-12 05:03:34 +0000119 *
120 * Cycle counts that are converted correctly as long as they
121 * fall into the interval [-1/2 max cycle count, +1/2 max cycle count],
122 * with "max cycle count" == cs->mask+1.
123 *
124 * This allows conversion of cycle counter values which were generated
125 * in the past.
126 */
127extern u64 timecounter_cyc2time(struct timecounter *tc,
128 cycle_t cycle_tstamp);
129
130/**
john stultz734efb42006-06-26 00:25:05 -0700131 * struct clocksource - hardware abstraction for a free running counter
132 * Provides mostly state-free accessors to the underlying hardware.
Patrick Ohlya038a352009-02-12 05:03:34 +0000133 * This is the structure used for system time.
john stultz734efb42006-06-26 00:25:05 -0700134 *
135 * @name: ptr to clocksource name
136 * @list: list head for registration
137 * @rating: rating value for selection (higher is better)
138 * To avoid rating inflation the following
139 * list should give you a guide as to how
140 * to assign your clocksource a rating
141 * 1-99: Unfit for real use
142 * Only available for bootup and testing purposes.
143 * 100-199: Base level usability.
144 * Functional for real use, but not desired.
145 * 200-299: Good.
146 * A correct and usable clocksource.
147 * 300-399: Desired.
148 * A reasonably fast and accurate clocksource.
149 * 400-499: Perfect
150 * The ideal clocksource. A must-use where
151 * available.
Magnus Damm8e196082009-04-21 12:24:00 -0700152 * @read: returns a cycle value, passes clocksource as argument
Magnus Damm4614e6a2009-04-21 12:24:02 -0700153 * @enable: optional function to enable the clocksource
154 * @disable: optional function to disable the clocksource
john stultz734efb42006-06-26 00:25:05 -0700155 * @mask: bitmask for two's complement
156 * subtraction of non 64 bit counters
Martin Schwidefsky0a544192009-08-14 15:47:28 +0200157 * @mult: cycle to nanosecond multiplier
john stultz734efb42006-06-26 00:25:05 -0700158 * @shift: cycle to nanosecond divisor (power of two)
Jon Hunter98962462009-08-18 12:45:10 -0500159 * @max_idle_ns: max idle time permitted by the clocksource (nsecs)
Kusanagi Kouichib1b73d02011-12-19 18:13:19 +0900160 * @maxadj: maximum adjustment value to mult (~11%)
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800161 * @flags: flags describing special properties
Andy Lutomirski433bd802011-07-13 09:24:13 -0400162 * @archdata: arch-specific data
Magnus Dammc54a42b2010-02-02 14:41:41 -0800163 * @suspend: suspend function for the clocksource, if necessary
Thomas Gleixnerb52f52a2007-05-09 02:35:15 -0700164 * @resume: resume function for the clocksource, if necessary
Thomas Gleixner09ac3692013-04-25 20:31:44 +0000165 * @owner: module reference, must be set by clocksource in modules
john stultz734efb42006-06-26 00:25:05 -0700166 */
167struct clocksource {
Eric Dumazet329c8d82007-05-08 00:27:57 -0700168 /*
Thomas Gleixner369db4c2011-05-18 21:33:40 +0000169 * Hotpath data, fits in a single cache line when the
170 * clocksource itself is cacheline aligned.
Eric Dumazet329c8d82007-05-08 00:27:57 -0700171 */
Magnus Damm8e196082009-04-21 12:24:00 -0700172 cycle_t (*read)(struct clocksource *cs);
john stultz734efb42006-06-26 00:25:05 -0700173 cycle_t mask;
174 u32 mult;
175 u32 shift;
Jon Hunter98962462009-08-18 12:45:10 -0500176 u64 max_idle_ns;
John Stultzd65670a2011-10-31 17:06:35 -0400177 u32 maxadj;
H. Peter Anvinae7bd112011-07-21 13:34:05 -0700178#ifdef CONFIG_ARCH_CLOCKSOURCE_DATA
Andy Lutomirski433bd802011-07-13 09:24:13 -0400179 struct arch_clocksource_data archdata;
Tony Luck0aa366f2007-07-20 11:22:30 -0700180#endif
Andy Lutomirski433bd802011-07-13 09:24:13 -0400181
Thomas Gleixner369db4c2011-05-18 21:33:40 +0000182 const char *name;
183 struct list_head list;
184 int rating;
Thomas Gleixner369db4c2011-05-18 21:33:40 +0000185 int (*enable)(struct clocksource *cs);
186 void (*disable)(struct clocksource *cs);
187 unsigned long flags;
188 void (*suspend)(struct clocksource *cs);
189 void (*resume)(struct clocksource *cs);
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800190
Kusanagi Kouichib1b73d02011-12-19 18:13:19 +0900191 /* private: */
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800192#ifdef CONFIG_CLOCKSOURCE_WATCHDOG
193 /* Watchdog related data, used by the framework */
194 struct list_head wd_list;
Thomas Gleixnerb5199512011-06-16 16:22:08 +0200195 cycle_t cs_last;
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800196 cycle_t wd_last;
197#endif
Thomas Gleixner09ac3692013-04-25 20:31:44 +0000198 struct module *owner;
Thomas Gleixner369db4c2011-05-18 21:33:40 +0000199} ____cacheline_aligned;
john stultz734efb42006-06-26 00:25:05 -0700200
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800201/*
202 * Clock source flags bits::
203 */
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800204#define CLOCK_SOURCE_IS_CONTINUOUS 0x01
205#define CLOCK_SOURCE_MUST_VERIFY 0x02
206
207#define CLOCK_SOURCE_WATCHDOG 0x10
208#define CLOCK_SOURCE_VALID_FOR_HRES 0x20
Martin Schwidefskyc55c87c2009-08-14 15:47:25 +0200209#define CLOCK_SOURCE_UNSTABLE 0x40
Feng Tang5caf4632013-03-12 11:56:46 +0800210#define CLOCK_SOURCE_SUSPEND_NONSTOP 0x80
Thomas Gleixner332962f2013-07-04 22:46:45 +0200211#define CLOCK_SOURCE_RESELECT 0x100
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800212
Jim Cromie7f9f3032006-06-26 00:25:15 -0700213/* simplify initialization of mask field */
Atsushi Nemoto1d76c262008-01-30 13:30:01 +0100214#define CLOCKSOURCE_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
john stultz734efb42006-06-26 00:25:05 -0700215
216/**
217 * clocksource_khz2mult - calculates mult from khz and shift
218 * @khz: Clocksource frequency in KHz
219 * @shift_constant: Clocksource shift factor
220 *
221 * Helper functions that converts a khz counter frequency to a timsource
222 * multiplier, given the clocksource shift value
223 */
224static inline u32 clocksource_khz2mult(u32 khz, u32 shift_constant)
225{
226 /* khz = cyc/(Million ns)
227 * mult/2^shift = ns/cyc
228 * mult = ns/cyc * 2^shift
229 * mult = 1Million/khz * 2^shift
230 * mult = 1000000 * 2^shift / khz
231 * mult = (1000000<<shift) / khz
232 */
233 u64 tmp = ((u64)1000000) << shift_constant;
234
235 tmp += khz/2; /* round for do_div */
236 do_div(tmp, khz);
237
238 return (u32)tmp;
239}
240
241/**
242 * clocksource_hz2mult - calculates mult from hz and shift
243 * @hz: Clocksource frequency in Hz
244 * @shift_constant: Clocksource shift factor
245 *
246 * Helper functions that converts a hz counter
247 * frequency to a timsource multiplier, given the
248 * clocksource shift value
249 */
250static inline u32 clocksource_hz2mult(u32 hz, u32 shift_constant)
251{
252 /* hz = cyc/(Billion ns)
253 * mult/2^shift = ns/cyc
254 * mult = ns/cyc * 2^shift
255 * mult = 1Billion/hz * 2^shift
256 * mult = 1000000000 * 2^shift / hz
257 * mult = (1000000000<<shift) / hz
258 */
259 u64 tmp = ((u64)1000000000) << shift_constant;
260
261 tmp += hz/2; /* round for do_div */
262 do_div(tmp, hz);
263
264 return (u32)tmp;
265}
266
267/**
Martin Schwidefsky155ec602009-08-14 15:47:26 +0200268 * clocksource_cyc2ns - converts clocksource cycles to nanoseconds
Kusanagi Kouichib1b73d02011-12-19 18:13:19 +0900269 * @cycles: cycles
270 * @mult: cycle to nanosecond multiplier
271 * @shift: cycle to nanosecond divisor (power of two)
john stultz734efb42006-06-26 00:25:05 -0700272 *
Martin Schwidefsky155ec602009-08-14 15:47:26 +0200273 * Converts cycles to nanoseconds, using the given mult and shift.
john stultz734efb42006-06-26 00:25:05 -0700274 *
275 * XXX - This could use some mult_lxl_ll() asm optimization
276 */
Martin Schwidefsky155ec602009-08-14 15:47:26 +0200277static inline s64 clocksource_cyc2ns(cycle_t cycles, u32 mult, u32 shift)
john stultz734efb42006-06-26 00:25:05 -0700278{
Martin Schwidefsky155ec602009-08-14 15:47:26 +0200279 return ((u64) cycles * mult) >> shift;
john stultz5eb6d202006-06-26 00:25:07 -0700280}
281
282
Thomas Gleixner92c7e002007-02-16 01:27:33 -0800283extern int clocksource_register(struct clocksource*);
Thomas Gleixnera89c7ed2013-04-25 20:31:46 +0000284extern int clocksource_unregister(struct clocksource*);
Jason Wessel7c3078b2008-02-15 14:55:54 -0600285extern void clocksource_touch_watchdog(void);
Thomas Gleixner92c7e002007-02-16 01:27:33 -0800286extern struct clocksource* clocksource_get_next(void);
287extern void clocksource_change_rating(struct clocksource *cs, int rating);
Magnus Dammc54a42b2010-02-02 14:41:41 -0800288extern void clocksource_suspend(void);
Thomas Gleixnerb52f52a2007-05-09 02:35:15 -0700289extern void clocksource_resume(void);
Martin Schwidefskyf1b82742009-08-14 15:47:21 +0200290extern struct clocksource * __init __weak clocksource_default_clock(void);
Thomas Gleixner7285dd72009-08-28 20:25:24 +0200291extern void clocksource_mark_unstable(struct clocksource *cs);
john stultz734efb42006-06-26 00:25:05 -0700292
Stephen Boyd87d8b9e2013-07-18 16:21:14 -0700293extern u64
294clocks_calc_max_nsecs(u32 mult, u32 shift, u32 maxadj, u64 mask);
Thomas Gleixner7d2f9442009-11-11 14:05:29 +0000295extern void
296clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 minsec);
297
John Stultzd7e81c22010-05-07 18:07:38 -0700298/*
299 * Don't call __clocksource_register_scale directly, use
300 * clocksource_register_hz/khz
301 */
302extern int
303__clocksource_register_scale(struct clocksource *cs, u32 scale, u32 freq);
John Stultz852db462010-07-13 17:56:28 -0700304extern void
305__clocksource_updatefreq_scale(struct clocksource *cs, u32 scale, u32 freq);
John Stultzd7e81c22010-05-07 18:07:38 -0700306
307static inline int clocksource_register_hz(struct clocksource *cs, u32 hz)
308{
309 return __clocksource_register_scale(cs, 1, hz);
310}
311
312static inline int clocksource_register_khz(struct clocksource *cs, u32 khz)
313{
314 return __clocksource_register_scale(cs, 1000, khz);
315}
316
John Stultz852db462010-07-13 17:56:28 -0700317static inline void __clocksource_updatefreq_hz(struct clocksource *cs, u32 hz)
318{
319 __clocksource_updatefreq_scale(cs, 1, hz);
320}
321
322static inline void __clocksource_updatefreq_khz(struct clocksource *cs, u32 khz)
323{
324 __clocksource_updatefreq_scale(cs, 1000, khz);
325}
John Stultzd7e81c22010-05-07 18:07:38 -0700326
john stultzacc9a9d2007-02-16 01:28:17 -0800327
Thomas Gleixnerba919d12013-04-25 20:31:44 +0000328extern int timekeeping_notify(struct clocksource *clock);
Martin Schwidefsky75c51582009-08-14 15:47:30 +0200329
Russell King442c8172011-05-08 14:06:52 +0100330extern cycle_t clocksource_mmio_readl_up(struct clocksource *);
331extern cycle_t clocksource_mmio_readl_down(struct clocksource *);
332extern cycle_t clocksource_mmio_readw_up(struct clocksource *);
333extern cycle_t clocksource_mmio_readw_down(struct clocksource *);
334
335extern int clocksource_mmio_init(void __iomem *, const char *,
336 unsigned long, int, unsigned, cycle_t (*)(struct clocksource *));
337
Russell King8c414ff2011-05-08 18:50:20 +0100338extern int clocksource_i8253_init(void);
339
Rob Herring54196cc2014-05-08 16:09:24 -0500340#define CLOCKSOURCE_OF_DECLARE(name, compat, fn) \
341 OF_DECLARE_1(clksrc, name, compat, fn)
342
Stephen Warrenae278a92012-11-19 16:41:20 -0700343#ifdef CONFIG_CLKSRC_OF
344extern void clocksource_of_init(void);
Stephen Warrene1d7ef12013-01-30 10:49:30 -0700345#else
Rob Herringe0c25362013-03-10 21:52:53 -0500346static inline void clocksource_of_init(void) {}
Stephen Warrenae278a92012-11-19 16:41:20 -0700347#endif
348
john stultz734efb42006-06-26 00:25:05 -0700349#endif /* _LINUX_CLOCKSOURCE_H */