Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Xen time implementation. |
| 3 | * |
| 4 | * This is implemented in terms of a clocksource driver which uses |
| 5 | * the hypervisor clock as a nanosecond timebase, and a clockevent |
| 6 | * driver which uses the hypervisor's timer mechanism. |
| 7 | * |
| 8 | * Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007 |
| 9 | */ |
| 10 | #include <linux/kernel.h> |
| 11 | #include <linux/interrupt.h> |
| 12 | #include <linux/clocksource.h> |
| 13 | #include <linux/clockchips.h> |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 14 | #include <linux/kernel_stat.h> |
Jeremy Fitzhardinge | f595ec9 | 2008-06-12 10:47:56 +0200 | [diff] [blame] | 15 | #include <linux/math64.h> |
Tejun Heo | 5a0e3ad | 2010-03-24 17:04:11 +0900 | [diff] [blame] | 16 | #include <linux/gfp.h> |
Konrad Rzeszutek Wilk | c9d76a2 | 2013-06-04 17:09:36 -0400 | [diff] [blame] | 17 | #include <linux/slab.h> |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 18 | #include <linux/pvclock_gtod.h> |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 19 | |
Gerd Hoffmann | 1c7b67f | 2008-06-03 16:17:30 +0200 | [diff] [blame] | 20 | #include <asm/pvclock.h> |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 21 | #include <asm/xen/hypervisor.h> |
| 22 | #include <asm/xen/hypercall.h> |
| 23 | |
| 24 | #include <xen/events.h> |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 25 | #include <xen/features.h> |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 26 | #include <xen/interface/xen.h> |
| 27 | #include <xen/interface/vcpu.h> |
| 28 | |
| 29 | #include "xen-ops.h" |
| 30 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 31 | /* Xen may fire a timer up to this many ns early */ |
| 32 | #define TIMER_SLOP 100000 |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 33 | #define NS_PER_TICK (1000000000LL / HZ) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 34 | |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 35 | /* runstate info updated by Xen */ |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 36 | static DEFINE_PER_CPU(struct vcpu_runstate_info, xen_runstate); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 37 | |
| 38 | /* snapshots of runstate info */ |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 39 | static DEFINE_PER_CPU(struct vcpu_runstate_info, xen_runstate_snapshot); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 40 | |
Laszlo Ersek | 0b0c002 | 2011-10-18 22:42:59 +0200 | [diff] [blame] | 41 | /* unused ns of stolen time */ |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 42 | static DEFINE_PER_CPU(u64, xen_residual_stolen); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 43 | |
| 44 | /* return an consistent snapshot of 64-bit time/counter value */ |
| 45 | static u64 get64(const u64 *p) |
| 46 | { |
| 47 | u64 ret; |
| 48 | |
| 49 | if (BITS_PER_LONG < 64) { |
| 50 | u32 *p32 = (u32 *)p; |
| 51 | u32 h, l; |
| 52 | |
| 53 | /* |
| 54 | * Read high then low, and then make sure high is |
| 55 | * still the same; this will only loop if low wraps |
| 56 | * and carries into high. |
| 57 | * XXX some clean way to make this endian-proof? |
| 58 | */ |
| 59 | do { |
| 60 | h = p32[1]; |
| 61 | barrier(); |
| 62 | l = p32[0]; |
| 63 | barrier(); |
| 64 | } while (p32[1] != h); |
| 65 | |
| 66 | ret = (((u64)h) << 32) | l; |
| 67 | } else |
| 68 | ret = *p; |
| 69 | |
| 70 | return ret; |
| 71 | } |
| 72 | |
| 73 | /* |
| 74 | * Runstate accounting |
| 75 | */ |
| 76 | static void get_runstate_snapshot(struct vcpu_runstate_info *res) |
| 77 | { |
| 78 | u64 state_time; |
| 79 | struct vcpu_runstate_info *state; |
| 80 | |
Jeremy Fitzhardinge | f120f13 | 2007-07-17 18:37:06 -0700 | [diff] [blame] | 81 | BUG_ON(preemptible()); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 82 | |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 83 | state = &__get_cpu_var(xen_runstate); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 84 | |
| 85 | /* |
| 86 | * The runstate info is always updated by the hypervisor on |
| 87 | * the current CPU, so there's no need to use anything |
| 88 | * stronger than a compiler barrier when fetching it. |
| 89 | */ |
| 90 | do { |
| 91 | state_time = get64(&state->state_entry_time); |
| 92 | barrier(); |
| 93 | *res = *state; |
| 94 | barrier(); |
| 95 | } while (get64(&state->state_entry_time) != state_time); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 96 | } |
| 97 | |
Jeremy Fitzhardinge | f0d7339 | 2007-10-16 11:51:30 -0700 | [diff] [blame] | 98 | /* return true when a vcpu could run but has no real cpu to run on */ |
| 99 | bool xen_vcpu_stolen(int vcpu) |
| 100 | { |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 101 | return per_cpu(xen_runstate, vcpu).state == RUNSTATE_runnable; |
Jeremy Fitzhardinge | f0d7339 | 2007-10-16 11:51:30 -0700 | [diff] [blame] | 102 | } |
| 103 | |
Ian Campbell | be01292 | 2009-11-21 08:35:55 +0800 | [diff] [blame] | 104 | void xen_setup_runstate_info(int cpu) |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 105 | { |
| 106 | struct vcpu_register_runstate_memory_area area; |
| 107 | |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 108 | area.addr.v = &per_cpu(xen_runstate, cpu); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 109 | |
| 110 | if (HYPERVISOR_vcpu_op(VCPUOP_register_runstate_memory_area, |
| 111 | cpu, &area)) |
| 112 | BUG(); |
| 113 | } |
| 114 | |
| 115 | static void do_stolen_accounting(void) |
| 116 | { |
| 117 | struct vcpu_runstate_info state; |
| 118 | struct vcpu_runstate_info *snap; |
Laszlo Ersek | 0b0c002 | 2011-10-18 22:42:59 +0200 | [diff] [blame] | 119 | s64 runnable, offline, stolen; |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 120 | cputime_t ticks; |
| 121 | |
| 122 | get_runstate_snapshot(&state); |
| 123 | |
| 124 | WARN_ON(state.state != RUNSTATE_running); |
| 125 | |
Tejun Heo | c6e22f9 | 2009-10-29 22:34:13 +0900 | [diff] [blame] | 126 | snap = &__get_cpu_var(xen_runstate_snapshot); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 127 | |
| 128 | /* work out how much time the VCPU has not been runn*ing* */ |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 129 | runnable = state.time[RUNSTATE_runnable] - snap->time[RUNSTATE_runnable]; |
| 130 | offline = state.time[RUNSTATE_offline] - snap->time[RUNSTATE_offline]; |
| 131 | |
| 132 | *snap = state; |
| 133 | |
| 134 | /* Add the appropriate number of ticks of stolen time, |
Martin Schwidefsky | 79741dd | 2008-12-31 15:11:38 +0100 | [diff] [blame] | 135 | including any left-overs from last time. */ |
Christoph Lameter | 780f36d | 2010-12-06 11:16:29 -0600 | [diff] [blame] | 136 | stolen = runnable + offline + __this_cpu_read(xen_residual_stolen); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 137 | |
| 138 | if (stolen < 0) |
| 139 | stolen = 0; |
| 140 | |
Jeremy Fitzhardinge | f595ec9 | 2008-06-12 10:47:56 +0200 | [diff] [blame] | 141 | ticks = iter_div_u64_rem(stolen, NS_PER_TICK, &stolen); |
Christoph Lameter | 780f36d | 2010-12-06 11:16:29 -0600 | [diff] [blame] | 142 | __this_cpu_write(xen_residual_stolen, stolen); |
Martin Schwidefsky | 79741dd | 2008-12-31 15:11:38 +0100 | [diff] [blame] | 143 | account_steal_ticks(ticks); |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 144 | } |
| 145 | |
Alok Kataria | e93ef94 | 2008-07-01 11:43:36 -0700 | [diff] [blame] | 146 | /* Get the TSC speed from Xen */ |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 147 | static unsigned long xen_tsc_khz(void) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 148 | { |
Glauber Costa | 3807f34 | 2008-07-28 11:47:52 -0300 | [diff] [blame] | 149 | struct pvclock_vcpu_time_info *info = |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 150 | &HYPERVISOR_shared_info->vcpu_info[0].time; |
| 151 | |
Glauber Costa | 3807f34 | 2008-07-28 11:47:52 -0300 | [diff] [blame] | 152 | return pvclock_tsc_khz(info); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 153 | } |
| 154 | |
Jeremy Fitzhardinge | ee7686b | 2008-08-21 13:17:56 -0700 | [diff] [blame] | 155 | cycle_t xen_clocksource_read(void) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 156 | { |
Gerd Hoffmann | 1c7b67f | 2008-06-03 16:17:30 +0200 | [diff] [blame] | 157 | struct pvclock_vcpu_time_info *src; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 158 | cycle_t ret; |
| 159 | |
Jeremy Fitzhardinge | f1c3962 | 2011-08-24 09:54:24 -0700 | [diff] [blame] | 160 | preempt_disable_notrace(); |
| 161 | src = &__get_cpu_var(xen_vcpu)->time; |
Gerd Hoffmann | 1c7b67f | 2008-06-03 16:17:30 +0200 | [diff] [blame] | 162 | ret = pvclock_clocksource_read(src); |
Jeremy Fitzhardinge | f1c3962 | 2011-08-24 09:54:24 -0700 | [diff] [blame] | 163 | preempt_enable_notrace(); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 164 | return ret; |
| 165 | } |
| 166 | |
Magnus Damm | 8e19608 | 2009-04-21 12:24:00 -0700 | [diff] [blame] | 167 | static cycle_t xen_clocksource_get_cycles(struct clocksource *cs) |
| 168 | { |
| 169 | return xen_clocksource_read(); |
| 170 | } |
| 171 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 172 | static void xen_read_wallclock(struct timespec *ts) |
| 173 | { |
Gerd Hoffmann | 1c7b67f | 2008-06-03 16:17:30 +0200 | [diff] [blame] | 174 | struct shared_info *s = HYPERVISOR_shared_info; |
| 175 | struct pvclock_wall_clock *wall_clock = &(s->wc); |
| 176 | struct pvclock_vcpu_time_info *vcpu_time; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 177 | |
Gerd Hoffmann | 1c7b67f | 2008-06-03 16:17:30 +0200 | [diff] [blame] | 178 | vcpu_time = &get_cpu_var(xen_vcpu)->time; |
| 179 | pvclock_read_wallclock(wall_clock, vcpu_time, ts); |
| 180 | put_cpu_var(xen_vcpu); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 181 | } |
| 182 | |
David Vrabel | 3565184 | 2013-05-13 18:56:06 +0100 | [diff] [blame] | 183 | static void xen_get_wallclock(struct timespec *now) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 184 | { |
David Vrabel | 3565184 | 2013-05-13 18:56:06 +0100 | [diff] [blame] | 185 | xen_read_wallclock(now); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 186 | } |
| 187 | |
David Vrabel | 3565184 | 2013-05-13 18:56:06 +0100 | [diff] [blame] | 188 | static int xen_set_wallclock(const struct timespec *now) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 189 | { |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 190 | return -1; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 191 | } |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 192 | |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 193 | static int xen_pvclock_gtod_notify(struct notifier_block *nb, |
| 194 | unsigned long was_set, void *priv) |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 195 | { |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 196 | /* Protected by the calling core code serialization */ |
| 197 | static struct timespec next_sync; |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 198 | |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 199 | struct xen_platform_op op; |
| 200 | struct timespec now; |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 201 | |
| 202 | now = __current_kernel_time(); |
| 203 | |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 204 | /* |
| 205 | * We only take the expensive HV call when the clock was set |
| 206 | * or when the 11 minutes RTC synchronization time elapsed. |
| 207 | */ |
| 208 | if (!was_set && timespec_compare(&now, &next_sync) < 0) |
| 209 | return NOTIFY_OK; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 210 | |
Magnus Damm | 8e19608 | 2009-04-21 12:24:00 -0700 | [diff] [blame] | 211 | op.cmd = XENPF_settime; |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 212 | op.u.settime.secs = now.tv_sec; |
| 213 | op.u.settime.nsecs = now.tv_nsec; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 214 | op.u.settime.system_time = xen_clocksource_read(); |
| 215 | |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 216 | (void)HYPERVISOR_dom0_op(&op); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 217 | |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 218 | /* |
| 219 | * Move the next drift compensation time 11 minutes |
| 220 | * ahead. That's emulating the sync_cmos_clock() update for |
| 221 | * the hardware RTC. |
| 222 | */ |
| 223 | next_sync = now; |
| 224 | next_sync.tv_sec += 11 * 60; |
| 225 | |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 226 | return NOTIFY_OK; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 227 | } |
| 228 | |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 229 | static struct notifier_block xen_pvclock_gtod_notifier = { |
| 230 | .notifier_call = xen_pvclock_gtod_notify, |
| 231 | }; |
| 232 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 233 | static struct clocksource xen_clocksource __read_mostly = { |
| 234 | .name = "xen", |
| 235 | .rating = 400, |
| 236 | .read = xen_clocksource_get_cycles, |
| 237 | .mask = ~0, |
| 238 | .flags = CLOCK_SOURCE_IS_CONTINUOUS, |
| 239 | }; |
| 240 | |
| 241 | /* |
| 242 | Xen clockevent implementation |
| 243 | |
| 244 | Xen has two clockevent implementations: |
| 245 | |
| 246 | The old timer_op one works with all released versions of Xen prior |
| 247 | to version 3.0.4. This version of the hypervisor provides a |
| 248 | single-shot timer with nanosecond resolution. However, sharing the |
| 249 | same event channel is a 100Hz tick which is delivered while the |
| 250 | vcpu is running. We don't care about or use this tick, but it will |
| 251 | cause the core time code to think the timer fired too soon, and |
| 252 | will end up resetting it each time. It could be filtered, but |
| 253 | doing so has complications when the ktime clocksource is not yet |
| 254 | the xen clocksource (ie, at boot time). |
| 255 | |
| 256 | The new vcpu_op-based timer interface allows the tick timer period |
| 257 | to be changed or turned off. The tick timer is not useful as a |
| 258 | periodic timer because events are only delivered to running vcpus. |
| 259 | The one-shot timer can report when a timeout is in the past, so |
| 260 | set_next_event is capable of returning -ETIME when appropriate. |
| 261 | This interface is used when available. |
| 262 | */ |
| 263 | |
| 264 | |
| 265 | /* |
| 266 | Get a hypervisor absolute time. In theory we could maintain an |
| 267 | offset between the kernel's time and the hypervisor's time, and |
| 268 | apply that to a kernel's absolute timeout. Unfortunately the |
| 269 | hypervisor and kernel times can drift even if the kernel is using |
| 270 | the Xen clocksource, because ntp can warp the kernel's clocksource. |
| 271 | */ |
| 272 | static s64 get_abs_timeout(unsigned long delta) |
| 273 | { |
| 274 | return xen_clocksource_read() + delta; |
| 275 | } |
| 276 | |
| 277 | static void xen_timerop_set_mode(enum clock_event_mode mode, |
| 278 | struct clock_event_device *evt) |
| 279 | { |
| 280 | switch (mode) { |
| 281 | case CLOCK_EVT_MODE_PERIODIC: |
Thomas Gleixner | 18de5bc | 2007-07-21 04:37:34 -0700 | [diff] [blame] | 282 | /* unsupported */ |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 283 | WARN_ON(1); |
| 284 | break; |
| 285 | |
| 286 | case CLOCK_EVT_MODE_ONESHOT: |
| 287 | case CLOCK_EVT_MODE_RESUME: |
| 288 | break; |
| 289 | |
| 290 | case CLOCK_EVT_MODE_UNUSED: |
| 291 | case CLOCK_EVT_MODE_SHUTDOWN: |
| 292 | HYPERVISOR_set_timer_op(0); /* cancel timeout */ |
| 293 | break; |
| 294 | } |
| 295 | } |
| 296 | |
| 297 | static int xen_timerop_set_next_event(unsigned long delta, |
| 298 | struct clock_event_device *evt) |
| 299 | { |
| 300 | WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); |
| 301 | |
| 302 | if (HYPERVISOR_set_timer_op(get_abs_timeout(delta)) < 0) |
| 303 | BUG(); |
| 304 | |
| 305 | /* We may have missed the deadline, but there's no real way of |
| 306 | knowing for sure. If the event was in the past, then we'll |
| 307 | get an immediate interrupt. */ |
| 308 | |
| 309 | return 0; |
| 310 | } |
| 311 | |
| 312 | static const struct clock_event_device xen_timerop_clockevent = { |
| 313 | .name = "xen", |
| 314 | .features = CLOCK_EVT_FEAT_ONESHOT, |
| 315 | |
| 316 | .max_delta_ns = 0xffffffff, |
| 317 | .min_delta_ns = TIMER_SLOP, |
| 318 | |
| 319 | .mult = 1, |
| 320 | .shift = 0, |
| 321 | .rating = 500, |
| 322 | |
| 323 | .set_mode = xen_timerop_set_mode, |
| 324 | .set_next_event = xen_timerop_set_next_event, |
| 325 | }; |
| 326 | |
| 327 | |
| 328 | |
| 329 | static void xen_vcpuop_set_mode(enum clock_event_mode mode, |
| 330 | struct clock_event_device *evt) |
| 331 | { |
| 332 | int cpu = smp_processor_id(); |
| 333 | |
| 334 | switch (mode) { |
| 335 | case CLOCK_EVT_MODE_PERIODIC: |
| 336 | WARN_ON(1); /* unsupported */ |
| 337 | break; |
| 338 | |
| 339 | case CLOCK_EVT_MODE_ONESHOT: |
| 340 | if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) |
| 341 | BUG(); |
| 342 | break; |
| 343 | |
| 344 | case CLOCK_EVT_MODE_UNUSED: |
Thomas Gleixner | 18de5bc | 2007-07-21 04:37:34 -0700 | [diff] [blame] | 345 | case CLOCK_EVT_MODE_SHUTDOWN: |
| 346 | if (HYPERVISOR_vcpu_op(VCPUOP_stop_singleshot_timer, cpu, NULL) || |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 347 | HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) |
| 348 | BUG(); |
| 349 | break; |
| 350 | case CLOCK_EVT_MODE_RESUME: |
| 351 | break; |
| 352 | } |
| 353 | } |
| 354 | |
| 355 | static int xen_vcpuop_set_next_event(unsigned long delta, |
| 356 | struct clock_event_device *evt) |
| 357 | { |
| 358 | int cpu = smp_processor_id(); |
| 359 | struct vcpu_set_singleshot_timer single; |
| 360 | int ret; |
| 361 | |
| 362 | WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); |
| 363 | |
| 364 | single.timeout_abs_ns = get_abs_timeout(delta); |
| 365 | single.flags = VCPU_SSHOTTMR_future; |
| 366 | |
| 367 | ret = HYPERVISOR_vcpu_op(VCPUOP_set_singleshot_timer, cpu, &single); |
| 368 | |
| 369 | BUG_ON(ret != 0 && ret != -ETIME); |
| 370 | |
| 371 | return ret; |
| 372 | } |
| 373 | |
| 374 | static const struct clock_event_device xen_vcpuop_clockevent = { |
| 375 | .name = "xen", |
| 376 | .features = CLOCK_EVT_FEAT_ONESHOT, |
| 377 | |
| 378 | .max_delta_ns = 0xffffffff, |
| 379 | .min_delta_ns = TIMER_SLOP, |
| 380 | |
| 381 | .mult = 1, |
| 382 | .shift = 0, |
| 383 | .rating = 500, |
| 384 | |
| 385 | .set_mode = xen_vcpuop_set_mode, |
| 386 | .set_next_event = xen_vcpuop_set_next_event, |
| 387 | }; |
| 388 | |
| 389 | static const struct clock_event_device *xen_clockevent = |
| 390 | &xen_timerop_clockevent; |
Konrad Rzeszutek Wilk | 31620a1 | 2013-06-04 17:06:36 -0400 | [diff] [blame] | 391 | |
| 392 | struct xen_clock_event_device { |
| 393 | struct clock_event_device evt; |
| 394 | char *name; |
| 395 | }; |
| 396 | static DEFINE_PER_CPU(struct xen_clock_event_device, xen_clock_events) = { .evt.irq = -1 }; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 397 | |
| 398 | static irqreturn_t xen_timer_interrupt(int irq, void *dev_id) |
| 399 | { |
Konrad Rzeszutek Wilk | 31620a1 | 2013-06-04 17:06:36 -0400 | [diff] [blame] | 400 | struct clock_event_device *evt = &__get_cpu_var(xen_clock_events).evt; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 401 | irqreturn_t ret; |
| 402 | |
| 403 | ret = IRQ_NONE; |
| 404 | if (evt->event_handler) { |
| 405 | evt->event_handler(evt); |
| 406 | ret = IRQ_HANDLED; |
| 407 | } |
| 408 | |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 409 | do_stolen_accounting(); |
| 410 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 411 | return ret; |
| 412 | } |
| 413 | |
Konrad Rzeszutek Wilk | 09e99da | 2013-06-04 17:13:29 -0400 | [diff] [blame] | 414 | void xen_teardown_timer(int cpu) |
| 415 | { |
| 416 | struct clock_event_device *evt; |
| 417 | BUG_ON(cpu == 0); |
| 418 | evt = &per_cpu(xen_clock_events, cpu).evt; |
| 419 | |
| 420 | if (evt->irq >= 0) { |
| 421 | unbind_from_irqhandler(evt->irq, NULL); |
| 422 | evt->irq = -1; |
| 423 | kfree(per_cpu(xen_clock_events, cpu).name); |
| 424 | per_cpu(xen_clock_events, cpu).name = NULL; |
| 425 | } |
| 426 | } |
| 427 | |
Jeremy Fitzhardinge | f87e4ca | 2007-07-17 18:37:06 -0700 | [diff] [blame] | 428 | void xen_setup_timer(int cpu) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 429 | { |
Konrad Rzeszutek Wilk | c9d76a2 | 2013-06-04 17:09:36 -0400 | [diff] [blame] | 430 | char *name; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 431 | struct clock_event_device *evt; |
| 432 | int irq; |
| 433 | |
Konrad Rzeszutek Wilk | 31620a1 | 2013-06-04 17:06:36 -0400 | [diff] [blame] | 434 | evt = &per_cpu(xen_clock_events, cpu).evt; |
Konrad Rzeszutek Wilk | ef35a4e | 2013-04-08 21:05:15 -0400 | [diff] [blame] | 435 | WARN(evt->irq >= 0, "IRQ%d for CPU%d is already allocated\n", evt->irq, cpu); |
Konrad Rzeszutek Wilk | 09e99da | 2013-06-04 17:13:29 -0400 | [diff] [blame] | 436 | if (evt->irq >= 0) |
| 437 | xen_teardown_timer(cpu); |
Konrad Rzeszutek Wilk | ef35a4e | 2013-04-08 21:05:15 -0400 | [diff] [blame] | 438 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 439 | printk(KERN_INFO "installing Xen timer for CPU %d\n", cpu); |
| 440 | |
| 441 | name = kasprintf(GFP_KERNEL, "timer%d", cpu); |
| 442 | if (!name) |
| 443 | name = "<timer kasprintf failed>"; |
| 444 | |
| 445 | irq = bind_virq_to_irqhandler(VIRQ_TIMER, cpu, xen_timer_interrupt, |
Michael Opdenacker | 9d71cee | 2013-09-07 08:46:49 +0200 | [diff] [blame^] | 446 | IRQF_PERCPU|IRQF_NOBALANCING|IRQF_TIMER| |
Ian Campbell | f611f2d | 2011-02-08 14:03:31 +0000 | [diff] [blame] | 447 | IRQF_FORCE_RESUME, |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 448 | name, NULL); |
| 449 | |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 450 | memcpy(evt, xen_clockevent, sizeof(*evt)); |
| 451 | |
Rusty Russell | 320ab2b | 2008-12-13 21:20:26 +1030 | [diff] [blame] | 452 | evt->cpumask = cpumask_of(cpu); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 453 | evt->irq = irq; |
Konrad Rzeszutek Wilk | c9d76a2 | 2013-06-04 17:09:36 -0400 | [diff] [blame] | 454 | per_cpu(xen_clock_events, cpu).name = name; |
Jeremy Fitzhardinge | f87e4ca | 2007-07-17 18:37:06 -0700 | [diff] [blame] | 455 | } |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 456 | |
Alex Nixon | d68d82a | 2008-08-22 11:52:15 +0100 | [diff] [blame] | 457 | |
Jeremy Fitzhardinge | f87e4ca | 2007-07-17 18:37:06 -0700 | [diff] [blame] | 458 | void xen_setup_cpu_clockevents(void) |
| 459 | { |
| 460 | BUG_ON(preemptible()); |
| 461 | |
Konrad Rzeszutek Wilk | 31620a1 | 2013-06-04 17:06:36 -0400 | [diff] [blame] | 462 | clockevents_register_device(&__get_cpu_var(xen_clock_events).evt); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 463 | } |
| 464 | |
Jeremy Fitzhardinge | d07af1f | 2008-05-31 01:33:03 +0100 | [diff] [blame] | 465 | void xen_timer_resume(void) |
| 466 | { |
| 467 | int cpu; |
| 468 | |
Jeremy Fitzhardinge | e7a3481c | 2010-10-25 16:53:46 -0700 | [diff] [blame] | 469 | pvclock_resume(); |
| 470 | |
Jeremy Fitzhardinge | d07af1f | 2008-05-31 01:33:03 +0100 | [diff] [blame] | 471 | if (xen_clockevent != &xen_vcpuop_clockevent) |
| 472 | return; |
| 473 | |
| 474 | for_each_online_cpu(cpu) { |
| 475 | if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) |
| 476 | BUG(); |
| 477 | } |
| 478 | } |
| 479 | |
Daniel Kiper | fb6ce5d | 2011-05-04 20:18:45 +0200 | [diff] [blame] | 480 | static const struct pv_time_ops xen_time_ops __initconst = { |
Jeremy Fitzhardinge | ca50a5f | 2010-08-04 14:49:16 -0700 | [diff] [blame] | 481 | .sched_clock = xen_clocksource_read, |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 482 | }; |
| 483 | |
Daniel Kiper | fb6ce5d | 2011-05-04 20:18:45 +0200 | [diff] [blame] | 484 | static void __init xen_time_init(void) |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 485 | { |
| 486 | int cpu = smp_processor_id(); |
John Stultz | c450725 | 2010-03-11 14:04:47 -0800 | [diff] [blame] | 487 | struct timespec tp; |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 488 | |
John Stultz | b01cc1b | 2010-04-26 19:03:05 -0700 | [diff] [blame] | 489 | clocksource_register_hz(&xen_clocksource, NSEC_PER_SEC); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 490 | |
| 491 | if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL) == 0) { |
Jeremy Fitzhardinge | f91a8b4 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 492 | /* Successfully turned off 100Hz tick, so we have the |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 493 | vcpuop-based timer interface */ |
| 494 | printk(KERN_DEBUG "Xen: using vcpuop timer interface\n"); |
| 495 | xen_clockevent = &xen_vcpuop_clockevent; |
| 496 | } |
| 497 | |
| 498 | /* Set initial system time with full resolution */ |
John Stultz | c450725 | 2010-03-11 14:04:47 -0800 | [diff] [blame] | 499 | xen_read_wallclock(&tp); |
| 500 | do_settimeofday(&tp); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 501 | |
Andi Kleen | 404ee5b | 2008-01-30 13:33:20 +0100 | [diff] [blame] | 502 | setup_force_cpu_cap(X86_FEATURE_TSC); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 503 | |
Ian Campbell | be01292 | 2009-11-21 08:35:55 +0800 | [diff] [blame] | 504 | xen_setup_runstate_info(cpu); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 505 | xen_setup_timer(cpu); |
Jeremy Fitzhardinge | f87e4ca | 2007-07-17 18:37:06 -0700 | [diff] [blame] | 506 | xen_setup_cpu_clockevents(); |
David Vrabel | 5584880 | 2013-06-27 11:35:47 +0100 | [diff] [blame] | 507 | |
| 508 | if (xen_initial_domain()) |
| 509 | pvclock_gtod_register_notifier(&xen_pvclock_gtod_notifier); |
Jeremy Fitzhardinge | 15c8473 | 2007-07-17 18:37:05 -0700 | [diff] [blame] | 510 | } |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 511 | |
Daniel Kiper | fb6ce5d | 2011-05-04 20:18:45 +0200 | [diff] [blame] | 512 | void __init xen_init_time_ops(void) |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 513 | { |
| 514 | pv_time_ops = xen_time_ops; |
| 515 | |
| 516 | x86_init.timers.timer_init = xen_time_init; |
| 517 | x86_init.timers.setup_percpu_clockev = x86_init_noop; |
| 518 | x86_cpuinit.setup_percpu_clockev = x86_init_noop; |
| 519 | |
| 520 | x86_platform.calibrate_tsc = xen_tsc_khz; |
| 521 | x86_platform.get_wallclock = xen_get_wallclock; |
David Vrabel | 47433b8 | 2013-06-27 11:35:48 +0100 | [diff] [blame] | 522 | /* Dom0 uses the native method to set the hardware RTC. */ |
| 523 | if (!xen_initial_domain()) |
| 524 | x86_platform.set_wallclock = xen_set_wallclock; |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 525 | } |
| 526 | |
Stefano Stabellini | ca65f9f | 2010-07-29 14:37:48 +0100 | [diff] [blame] | 527 | #ifdef CONFIG_XEN_PVHVM |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 528 | static void xen_hvm_setup_cpu_clockevents(void) |
| 529 | { |
| 530 | int cpu = smp_processor_id(); |
| 531 | xen_setup_runstate_info(cpu); |
Konrad Rzeszutek Wilk | 7918c92 | 2013-04-16 15:18:00 -0400 | [diff] [blame] | 532 | /* |
| 533 | * xen_setup_timer(cpu) - snprintf is bad in atomic context. Hence |
| 534 | * doing it xen_hvm_cpu_notify (which gets called by smp_init during |
| 535 | * early bootup and also during CPU hotplug events). |
| 536 | */ |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 537 | xen_setup_cpu_clockevents(); |
| 538 | } |
| 539 | |
Daniel Kiper | fb6ce5d | 2011-05-04 20:18:45 +0200 | [diff] [blame] | 540 | void __init xen_hvm_init_time_ops(void) |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 541 | { |
| 542 | /* vector callback is needed otherwise we cannot receive interrupts |
Stefano Stabellini | 31e7e93 | 2010-10-01 17:35:46 +0100 | [diff] [blame] | 543 | * on cpu > 0 and at this point we don't know how many cpus are |
| 544 | * available */ |
| 545 | if (!xen_have_vector_callback) |
Stefano Stabellini | 409771d | 2010-05-14 12:48:19 +0100 | [diff] [blame] | 546 | return; |
| 547 | if (!xen_feature(XENFEAT_hvm_safe_pvclock)) { |
| 548 | printk(KERN_INFO "Xen doesn't support pvclock on HVM," |
| 549 | "disable pv timer\n"); |
| 550 | return; |
| 551 | } |
| 552 | |
| 553 | pv_time_ops = xen_time_ops; |
| 554 | x86_init.timers.setup_percpu_clockev = xen_time_init; |
| 555 | x86_cpuinit.setup_percpu_clockev = xen_hvm_setup_cpu_clockevents; |
| 556 | |
| 557 | x86_platform.calibrate_tsc = xen_tsc_khz; |
| 558 | x86_platform.get_wallclock = xen_get_wallclock; |
| 559 | x86_platform.set_wallclock = xen_set_wallclock; |
| 560 | } |
Stefano Stabellini | ca65f9f | 2010-07-29 14:37:48 +0100 | [diff] [blame] | 561 | #endif |