blob: 19ad43af62d07aad5bdcc55f311572bd66480cda [file] [log] [blame]
john stultz734efb42006-06-26 00:25:05 -07001/* linux/include/linux/clocksource.h
2 *
3 * This file contains the structure definitions for clocksources.
4 *
5 * If you are not a clocksource, or timekeeping code, you should
6 * not be including this file!
7 */
8#ifndef _LINUX_CLOCKSOURCE_H
9#define _LINUX_CLOCKSOURCE_H
10
11#include <linux/types.h>
12#include <linux/timex.h>
13#include <linux/time.h>
14#include <linux/list.h>
Eric Dumazet329c8d82007-05-08 00:27:57 -070015#include <linux/cache.h>
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -080016#include <linux/timer.h>
Martin Schwidefskyf1b82742009-08-14 15:47:21 +020017#include <linux/init.h>
john stultz734efb42006-06-26 00:25:05 -070018#include <asm/div64.h>
19#include <asm/io.h>
20
21/* clocksource cycle base type */
22typedef u64 cycle_t;
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -080023struct clocksource;
john stultz734efb42006-06-26 00:25:05 -070024
25/**
Patrick Ohlya038a352009-02-12 05:03:34 +000026 * struct cyclecounter - hardware abstraction for a free running counter
27 * Provides completely state-free accessors to the underlying hardware.
28 * Depending on which hardware it reads, the cycle counter may wrap
29 * around quickly. Locking rules (if necessary) have to be defined
30 * by the implementor and user of specific instances of this API.
31 *
32 * @read: returns the current cycle value
33 * @mask: bitmask for two's complement
34 * subtraction of non 64 bit counters,
35 * see CLOCKSOURCE_MASK() helper macro
36 * @mult: cycle to nanosecond multiplier
37 * @shift: cycle to nanosecond divisor (power of two)
38 */
39struct cyclecounter {
40 cycle_t (*read)(const struct cyclecounter *cc);
41 cycle_t mask;
42 u32 mult;
43 u32 shift;
44};
45
46/**
47 * struct timecounter - layer above a %struct cyclecounter which counts nanoseconds
48 * Contains the state needed by timecounter_read() to detect
49 * cycle counter wrap around. Initialize with
50 * timecounter_init(). Also used to convert cycle counts into the
51 * corresponding nanosecond counts with timecounter_cyc2time(). Users
52 * of this code are responsible for initializing the underlying
53 * cycle counter hardware, locking issues and reading the time
54 * more often than the cycle counter wraps around. The nanosecond
55 * counter will only wrap around after ~585 years.
56 *
57 * @cc: the cycle counter used by this instance
58 * @cycle_last: most recent cycle counter value seen by
59 * timecounter_read()
60 * @nsec: continuously increasing count
61 */
62struct timecounter {
63 const struct cyclecounter *cc;
64 cycle_t cycle_last;
65 u64 nsec;
66};
67
68/**
69 * cyclecounter_cyc2ns - converts cycle counter cycles to nanoseconds
70 * @tc: Pointer to cycle counter.
71 * @cycles: Cycles
72 *
73 * XXX - This could use some mult_lxl_ll() asm optimization. Same code
74 * as in cyc2ns, but with unsigned result.
75 */
76static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc,
77 cycle_t cycles)
78{
79 u64 ret = (u64)cycles;
80 ret = (ret * cc->mult) >> cc->shift;
81 return ret;
82}
83
84/**
85 * timecounter_init - initialize a time counter
86 * @tc: Pointer to time counter which is to be initialized/reset
87 * @cc: A cycle counter, ready to be used.
88 * @start_tstamp: Arbitrary initial time stamp.
89 *
90 * After this call the current cycle register (roughly) corresponds to
91 * the initial time stamp. Every call to timecounter_read() increments
92 * the time stamp counter by the number of elapsed nanoseconds.
93 */
94extern void timecounter_init(struct timecounter *tc,
95 const struct cyclecounter *cc,
96 u64 start_tstamp);
97
98/**
99 * timecounter_read - return nanoseconds elapsed since timecounter_init()
100 * plus the initial time stamp
101 * @tc: Pointer to time counter.
102 *
103 * In other words, keeps track of time since the same epoch as
104 * the function which generated the initial time stamp.
105 */
106extern u64 timecounter_read(struct timecounter *tc);
107
108/**
109 * timecounter_cyc2time - convert a cycle counter to same
110 * time base as values returned by
111 * timecounter_read()
112 * @tc: Pointer to time counter.
113 * @cycle: a value returned by tc->cc->read()
114 *
115 * Cycle counts that are converted correctly as long as they
116 * fall into the interval [-1/2 max cycle count, +1/2 max cycle count],
117 * with "max cycle count" == cs->mask+1.
118 *
119 * This allows conversion of cycle counter values which were generated
120 * in the past.
121 */
122extern u64 timecounter_cyc2time(struct timecounter *tc,
123 cycle_t cycle_tstamp);
124
125/**
john stultz734efb42006-06-26 00:25:05 -0700126 * struct clocksource - hardware abstraction for a free running counter
127 * Provides mostly state-free accessors to the underlying hardware.
Patrick Ohlya038a352009-02-12 05:03:34 +0000128 * This is the structure used for system time.
john stultz734efb42006-06-26 00:25:05 -0700129 *
130 * @name: ptr to clocksource name
131 * @list: list head for registration
132 * @rating: rating value for selection (higher is better)
133 * To avoid rating inflation the following
134 * list should give you a guide as to how
135 * to assign your clocksource a rating
136 * 1-99: Unfit for real use
137 * Only available for bootup and testing purposes.
138 * 100-199: Base level usability.
139 * Functional for real use, but not desired.
140 * 200-299: Good.
141 * A correct and usable clocksource.
142 * 300-399: Desired.
143 * A reasonably fast and accurate clocksource.
144 * 400-499: Perfect
145 * The ideal clocksource. A must-use where
146 * available.
Magnus Damm8e196082009-04-21 12:24:00 -0700147 * @read: returns a cycle value, passes clocksource as argument
Magnus Damm4614e6a2009-04-21 12:24:02 -0700148 * @enable: optional function to enable the clocksource
149 * @disable: optional function to disable the clocksource
john stultz734efb42006-06-26 00:25:05 -0700150 * @mask: bitmask for two's complement
151 * subtraction of non 64 bit counters
John Stultz1aa5dfb2008-08-20 16:37:28 -0700152 * @mult: cycle to nanosecond multiplier (adjusted by NTP)
153 * @mult_orig: cycle to nanosecond multiplier (unadjusted by NTP)
john stultz734efb42006-06-26 00:25:05 -0700154 * @shift: cycle to nanosecond divisor (power of two)
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800155 * @flags: flags describing special properties
john stultzacc9a9d2007-02-16 01:28:17 -0800156 * @vread: vsyscall based read
Thomas Gleixnerb52f52a2007-05-09 02:35:15 -0700157 * @resume: resume function for the clocksource, if necessary
Roman Zippel19923c12006-06-26 00:25:18 -0700158 * @cycle_interval: Used internally by timekeeping core, please ignore.
159 * @xtime_interval: Used internally by timekeeping core, please ignore.
john stultz734efb42006-06-26 00:25:05 -0700160 */
161struct clocksource {
Eric Dumazet329c8d82007-05-08 00:27:57 -0700162 /*
163 * First part of structure is read mostly
164 */
john stultz734efb42006-06-26 00:25:05 -0700165 char *name;
166 struct list_head list;
167 int rating;
Magnus Damm8e196082009-04-21 12:24:00 -0700168 cycle_t (*read)(struct clocksource *cs);
Magnus Damm4614e6a2009-04-21 12:24:02 -0700169 int (*enable)(struct clocksource *cs);
170 void (*disable)(struct clocksource *cs);
john stultz734efb42006-06-26 00:25:05 -0700171 cycle_t mask;
172 u32 mult;
John Stultz1aa5dfb2008-08-20 16:37:28 -0700173 u32 mult_orig;
john stultz734efb42006-06-26 00:25:05 -0700174 u32 shift;
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800175 unsigned long flags;
john stultzacc9a9d2007-02-16 01:28:17 -0800176 cycle_t (*vread)(void);
Thomas Gleixnerb52f52a2007-05-09 02:35:15 -0700177 void (*resume)(void);
Tony Luck0aa366f2007-07-20 11:22:30 -0700178#ifdef CONFIG_IA64
179 void *fsys_mmio; /* used by fsyscall asm code */
180#define CLKSRC_FSYS_MMIO_SET(mmio, addr) ((mmio) = (addr))
181#else
182#define CLKSRC_FSYS_MMIO_SET(mmio, addr) do { } while (0)
183#endif
john stultz734efb42006-06-26 00:25:05 -0700184
185 /* timekeeping specific data, ignore */
Eric Dumazet329c8d82007-05-08 00:27:57 -0700186 cycle_t cycle_interval;
187 u64 xtime_interval;
John Stultz2d422442008-08-20 16:37:30 -0700188 u32 raw_interval;
Eric Dumazet329c8d82007-05-08 00:27:57 -0700189 /*
190 * Second part is written at each timer interrupt
191 * Keep it in a different cache line to dirty no
192 * more than one cache line.
193 */
194 cycle_t cycle_last ____cacheline_aligned_in_smp;
195 u64 xtime_nsec;
Roman Zippel19923c12006-06-26 00:25:18 -0700196 s64 error;
John Stultz2d422442008-08-20 16:37:30 -0700197 struct timespec raw_time;
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800198
199#ifdef CONFIG_CLOCKSOURCE_WATCHDOG
200 /* Watchdog related data, used by the framework */
201 struct list_head wd_list;
202 cycle_t wd_last;
203#endif
john stultz734efb42006-06-26 00:25:05 -0700204};
205
Roman Zippel7dffa3c2008-05-01 04:34:41 -0700206extern struct clocksource *clock; /* current clocksource */
207
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800208/*
209 * Clock source flags bits::
210 */
Thomas Gleixner5d8b34f2007-02-16 01:27:43 -0800211#define CLOCK_SOURCE_IS_CONTINUOUS 0x01
212#define CLOCK_SOURCE_MUST_VERIFY 0x02
213
214#define CLOCK_SOURCE_WATCHDOG 0x10
215#define CLOCK_SOURCE_VALID_FOR_HRES 0x20
Martin Schwidefskyc55c87c2009-08-14 15:47:25 +0200216#define CLOCK_SOURCE_UNSTABLE 0x40
Thomas Gleixner73b08d22007-02-16 01:27:36 -0800217
Jim Cromie7f9f3032006-06-26 00:25:15 -0700218/* simplify initialization of mask field */
Atsushi Nemoto1d76c262008-01-30 13:30:01 +0100219#define CLOCKSOURCE_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
john stultz734efb42006-06-26 00:25:05 -0700220
221/**
222 * clocksource_khz2mult - calculates mult from khz and shift
223 * @khz: Clocksource frequency in KHz
224 * @shift_constant: Clocksource shift factor
225 *
226 * Helper functions that converts a khz counter frequency to a timsource
227 * multiplier, given the clocksource shift value
228 */
229static inline u32 clocksource_khz2mult(u32 khz, u32 shift_constant)
230{
231 /* khz = cyc/(Million ns)
232 * mult/2^shift = ns/cyc
233 * mult = ns/cyc * 2^shift
234 * mult = 1Million/khz * 2^shift
235 * mult = 1000000 * 2^shift / khz
236 * mult = (1000000<<shift) / khz
237 */
238 u64 tmp = ((u64)1000000) << shift_constant;
239
240 tmp += khz/2; /* round for do_div */
241 do_div(tmp, khz);
242
243 return (u32)tmp;
244}
245
246/**
247 * clocksource_hz2mult - calculates mult from hz and shift
248 * @hz: Clocksource frequency in Hz
249 * @shift_constant: Clocksource shift factor
250 *
251 * Helper functions that converts a hz counter
252 * frequency to a timsource multiplier, given the
253 * clocksource shift value
254 */
255static inline u32 clocksource_hz2mult(u32 hz, u32 shift_constant)
256{
257 /* hz = cyc/(Billion ns)
258 * mult/2^shift = ns/cyc
259 * mult = ns/cyc * 2^shift
260 * mult = 1Billion/hz * 2^shift
261 * mult = 1000000000 * 2^shift / hz
262 * mult = (1000000000<<shift) / hz
263 */
264 u64 tmp = ((u64)1000000000) << shift_constant;
265
266 tmp += hz/2; /* round for do_div */
267 do_div(tmp, hz);
268
269 return (u32)tmp;
270}
271
272/**
john stultz734efb42006-06-26 00:25:05 -0700273 * cyc2ns - converts clocksource cycles to nanoseconds
274 * @cs: Pointer to clocksource
275 * @cycles: Cycles
276 *
277 * Uses the clocksource and ntp ajdustment to convert cycle_ts to nanoseconds.
278 *
279 * XXX - This could use some mult_lxl_ll() asm optimization
280 */
281static inline s64 cyc2ns(struct clocksource *cs, cycle_t cycles)
282{
283 u64 ret = (u64)cycles;
284 ret = (ret * cs->mult) >> cs->shift;
285 return ret;
286}
287
288/**
john stultza2752542006-06-26 00:25:14 -0700289 * clocksource_calculate_interval - Calculates a clocksource interval struct
john stultz734efb42006-06-26 00:25:05 -0700290 *
291 * @c: Pointer to clocksource.
292 * @length_nsec: Desired interval length in nanoseconds.
293 *
294 * Calculates a fixed cycle/nsec interval for a given clocksource/adjustment
295 * pair and interval request.
296 *
297 * Unless you're the timekeeping code, you should not be using this!
298 */
john stultza2752542006-06-26 00:25:14 -0700299static inline void clocksource_calculate_interval(struct clocksource *c,
Daniel Walkerf5f1a242006-12-10 02:21:33 -0800300 unsigned long length_nsec)
john stultz734efb42006-06-26 00:25:05 -0700301{
302 u64 tmp;
303
John Stultz1aa5dfb2008-08-20 16:37:28 -0700304 /* Do the ns -> cycle conversion first, using original mult */
john stultz734efb42006-06-26 00:25:05 -0700305 tmp = length_nsec;
306 tmp <<= c->shift;
John Stultz1aa5dfb2008-08-20 16:37:28 -0700307 tmp += c->mult_orig/2;
308 do_div(tmp, c->mult_orig);
john stultz734efb42006-06-26 00:25:05 -0700309
Roman Zippel19923c12006-06-26 00:25:18 -0700310 c->cycle_interval = (cycle_t)tmp;
311 if (c->cycle_interval == 0)
312 c->cycle_interval = 1;
john stultz734efb42006-06-26 00:25:05 -0700313
John Stultz1aa5dfb2008-08-20 16:37:28 -0700314 /* Go back from cycles -> shifted ns, this time use ntp adjused mult */
Roman Zippel19923c12006-06-26 00:25:18 -0700315 c->xtime_interval = (u64)c->cycle_interval * c->mult;
John Stultz2d422442008-08-20 16:37:30 -0700316 c->raw_interval = ((u64)c->cycle_interval * c->mult_orig) >> c->shift;
john stultz5eb6d202006-06-26 00:25:07 -0700317}
318
319
john stultz734efb42006-06-26 00:25:05 -0700320/* used to install a new clocksource */
Thomas Gleixner92c7e002007-02-16 01:27:33 -0800321extern int clocksource_register(struct clocksource*);
Thomas Gleixner4713e22c2008-01-30 13:30:02 +0100322extern void clocksource_unregister(struct clocksource*);
Jason Wessel7c3078b2008-02-15 14:55:54 -0600323extern void clocksource_touch_watchdog(void);
Thomas Gleixner92c7e002007-02-16 01:27:33 -0800324extern struct clocksource* clocksource_get_next(void);
325extern void clocksource_change_rating(struct clocksource *cs, int rating);
Thomas Gleixnerb52f52a2007-05-09 02:35:15 -0700326extern void clocksource_resume(void);
Martin Schwidefskyf1b82742009-08-14 15:47:21 +0200327extern struct clocksource * __init __weak clocksource_default_clock(void);
john stultz734efb42006-06-26 00:25:05 -0700328
john stultzacc9a9d2007-02-16 01:28:17 -0800329#ifdef CONFIG_GENERIC_TIME_VSYSCALL
330extern void update_vsyscall(struct timespec *ts, struct clocksource *c);
Tony Breeds2c622142007-10-18 03:04:57 -0700331extern void update_vsyscall_tz(void);
john stultzacc9a9d2007-02-16 01:28:17 -0800332#else
333static inline void update_vsyscall(struct timespec *ts, struct clocksource *c)
334{
335}
Tony Breeds2c622142007-10-18 03:04:57 -0700336
337static inline void update_vsyscall_tz(void)
338{
339}
john stultzacc9a9d2007-02-16 01:28:17 -0800340#endif
341
john stultz734efb42006-06-26 00:25:05 -0700342#endif /* _LINUX_CLOCKSOURCE_H */