Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 1 | /* |
| 2 | * DaVinci timer subsystem |
| 3 | * |
| 4 | * Author: Kevin Hilman, MontaVista Software, Inc. <source@mvista.com> |
| 5 | * |
| 6 | * 2007 (c) MontaVista Software, Inc. This file is licensed under |
| 7 | * the terms of the GNU General Public License version 2. This program |
| 8 | * is licensed "as is" without any warranty of any kind, whether express |
| 9 | * or implied. |
| 10 | */ |
| 11 | #include <linux/kernel.h> |
| 12 | #include <linux/init.h> |
| 13 | #include <linux/types.h> |
| 14 | #include <linux/interrupt.h> |
| 15 | #include <linux/clocksource.h> |
| 16 | #include <linux/clockchips.h> |
Russell King | fced80c | 2008-09-06 12:10:45 +0100 | [diff] [blame] | 17 | #include <linux/io.h> |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 18 | #include <linux/clk.h> |
| 19 | #include <linux/err.h> |
Kevin Hilman | fb63138 | 2009-04-29 16:23:59 -0700 | [diff] [blame] | 20 | #include <linux/platform_device.h> |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 21 | |
Russell King | a09e64f | 2008-08-05 16:14:15 +0100 | [diff] [blame] | 22 | #include <mach/hardware.h> |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 23 | #include <asm/mach/irq.h> |
| 24 | #include <asm/mach/time.h> |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 25 | #include <mach/cputype.h> |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 26 | #include <mach/time.h> |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 27 | #include "clock.h" |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 28 | |
| 29 | static struct clock_event_device clockevent_davinci; |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 30 | static unsigned int davinci_clock_tick_rate; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 31 | |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 32 | /* |
| 33 | * This driver configures the 2 64-bit count-up timers as 4 independent |
| 34 | * 32-bit count-up timers used as follows: |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 35 | */ |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 36 | |
| 37 | enum { |
| 38 | TID_CLOCKEVENT, |
| 39 | TID_CLOCKSOURCE, |
| 40 | }; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 41 | |
| 42 | /* Timer register offsets */ |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 43 | #define PID12 0x0 |
| 44 | #define TIM12 0x10 |
| 45 | #define TIM34 0x14 |
| 46 | #define PRD12 0x18 |
| 47 | #define PRD34 0x1c |
| 48 | #define TCR 0x20 |
| 49 | #define TGCR 0x24 |
| 50 | #define WDTCR 0x28 |
| 51 | |
| 52 | /* Offsets of the 8 compare registers */ |
| 53 | #define CMP12_0 0x60 |
| 54 | #define CMP12_1 0x64 |
| 55 | #define CMP12_2 0x68 |
| 56 | #define CMP12_3 0x6c |
| 57 | #define CMP12_4 0x70 |
| 58 | #define CMP12_5 0x74 |
| 59 | #define CMP12_6 0x78 |
| 60 | #define CMP12_7 0x7c |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 61 | |
| 62 | /* Timer register bitfields */ |
| 63 | #define TCR_ENAMODE_DISABLE 0x0 |
| 64 | #define TCR_ENAMODE_ONESHOT 0x1 |
| 65 | #define TCR_ENAMODE_PERIODIC 0x2 |
| 66 | #define TCR_ENAMODE_MASK 0x3 |
| 67 | |
| 68 | #define TGCR_TIMMODE_SHIFT 2 |
| 69 | #define TGCR_TIMMODE_64BIT_GP 0x0 |
| 70 | #define TGCR_TIMMODE_32BIT_UNCHAINED 0x1 |
| 71 | #define TGCR_TIMMODE_64BIT_WDOG 0x2 |
| 72 | #define TGCR_TIMMODE_32BIT_CHAINED 0x3 |
| 73 | |
| 74 | #define TGCR_TIM12RS_SHIFT 0 |
| 75 | #define TGCR_TIM34RS_SHIFT 1 |
| 76 | #define TGCR_RESET 0x0 |
| 77 | #define TGCR_UNRESET 0x1 |
| 78 | #define TGCR_RESET_MASK 0x3 |
| 79 | |
| 80 | #define WDTCR_WDEN_SHIFT 14 |
| 81 | #define WDTCR_WDEN_DISABLE 0x0 |
| 82 | #define WDTCR_WDEN_ENABLE 0x1 |
| 83 | #define WDTCR_WDKEY_SHIFT 16 |
| 84 | #define WDTCR_WDKEY_SEQ0 0xa5c6 |
| 85 | #define WDTCR_WDKEY_SEQ1 0xda7e |
| 86 | |
| 87 | struct timer_s { |
| 88 | char *name; |
| 89 | unsigned int id; |
| 90 | unsigned long period; |
| 91 | unsigned long opts; |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 92 | unsigned long flags; |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 93 | void __iomem *base; |
| 94 | unsigned long tim_off; |
| 95 | unsigned long prd_off; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 96 | unsigned long enamode_shift; |
| 97 | struct irqaction irqaction; |
| 98 | }; |
| 99 | static struct timer_s timers[]; |
| 100 | |
| 101 | /* values for 'opts' field of struct timer_s */ |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 102 | #define TIMER_OPTS_DISABLED 0x01 |
| 103 | #define TIMER_OPTS_ONESHOT 0x02 |
| 104 | #define TIMER_OPTS_PERIODIC 0x04 |
| 105 | #define TIMER_OPTS_STATE_MASK 0x07 |
| 106 | |
| 107 | #define TIMER_OPTS_USE_COMPARE 0x80000000 |
| 108 | #define USING_COMPARE(t) ((t)->opts & TIMER_OPTS_USE_COMPARE) |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 109 | |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 110 | static char *id_to_name[] = { |
| 111 | [T0_BOT] = "timer0_0", |
| 112 | [T0_TOP] = "timer0_1", |
| 113 | [T1_BOT] = "timer1_0", |
| 114 | [T1_TOP] = "timer1_1", |
| 115 | }; |
| 116 | |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 117 | static int timer32_config(struct timer_s *t) |
| 118 | { |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 119 | u32 tcr; |
Mark A. Greer | 5570078 | 2009-04-15 12:42:06 -0700 | [diff] [blame] | 120 | struct davinci_soc_info *soc_info = &davinci_soc_info; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 121 | |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 122 | if (USING_COMPARE(t)) { |
| 123 | struct davinci_timer_instance *dtip = |
| 124 | soc_info->timer_info->timers; |
| 125 | int event_timer = ID_TO_TIMER(timers[TID_CLOCKEVENT].id); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 126 | |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 127 | /* |
| 128 | * Next interrupt should be the current time reg value plus |
| 129 | * the new period (using 32-bit unsigned addition/wrapping |
| 130 | * to 0 on overflow). This assumes that the clocksource |
| 131 | * is setup to count to 2^32-1 before wrapping around to 0. |
| 132 | */ |
| 133 | __raw_writel(__raw_readl(t->base + t->tim_off) + t->period, |
| 134 | t->base + dtip[event_timer].cmp_off); |
| 135 | } else { |
| 136 | tcr = __raw_readl(t->base + TCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 137 | |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 138 | /* disable timer */ |
| 139 | tcr &= ~(TCR_ENAMODE_MASK << t->enamode_shift); |
| 140 | __raw_writel(tcr, t->base + TCR); |
| 141 | |
| 142 | /* reset counter to zero, set new period */ |
| 143 | __raw_writel(0, t->base + t->tim_off); |
| 144 | __raw_writel(t->period, t->base + t->prd_off); |
| 145 | |
| 146 | /* Set enable mode */ |
| 147 | if (t->opts & TIMER_OPTS_ONESHOT) |
| 148 | tcr |= TCR_ENAMODE_ONESHOT << t->enamode_shift; |
| 149 | else if (t->opts & TIMER_OPTS_PERIODIC) |
| 150 | tcr |= TCR_ENAMODE_PERIODIC << t->enamode_shift; |
| 151 | |
| 152 | __raw_writel(tcr, t->base + TCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 153 | } |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 154 | return 0; |
| 155 | } |
| 156 | |
| 157 | static inline u32 timer32_read(struct timer_s *t) |
| 158 | { |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 159 | return __raw_readl(t->base + t->tim_off); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 160 | } |
| 161 | |
| 162 | static irqreturn_t timer_interrupt(int irq, void *dev_id) |
| 163 | { |
| 164 | struct clock_event_device *evt = &clockevent_davinci; |
| 165 | |
| 166 | evt->event_handler(evt); |
| 167 | return IRQ_HANDLED; |
| 168 | } |
| 169 | |
| 170 | /* called when 32-bit counter wraps */ |
| 171 | static irqreturn_t freerun_interrupt(int irq, void *dev_id) |
| 172 | { |
| 173 | return IRQ_HANDLED; |
| 174 | } |
| 175 | |
| 176 | static struct timer_s timers[] = { |
| 177 | [TID_CLOCKEVENT] = { |
| 178 | .name = "clockevent", |
| 179 | .opts = TIMER_OPTS_DISABLED, |
| 180 | .irqaction = { |
| 181 | .flags = IRQF_DISABLED | IRQF_TIMER, |
| 182 | .handler = timer_interrupt, |
| 183 | } |
| 184 | }, |
| 185 | [TID_CLOCKSOURCE] = { |
| 186 | .name = "free-run counter", |
| 187 | .period = ~0, |
| 188 | .opts = TIMER_OPTS_PERIODIC, |
| 189 | .irqaction = { |
| 190 | .flags = IRQF_DISABLED | IRQF_TIMER, |
| 191 | .handler = freerun_interrupt, |
| 192 | } |
| 193 | }, |
| 194 | }; |
| 195 | |
| 196 | static void __init timer_init(void) |
| 197 | { |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 198 | struct davinci_soc_info *soc_info = &davinci_soc_info; |
| 199 | struct davinci_timer_instance *dtip = soc_info->timer_info->timers; |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 200 | void __iomem *base[2]; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 201 | int i; |
| 202 | |
| 203 | /* Global init of each 64-bit timer as a whole */ |
| 204 | for(i=0; i<2; i++) { |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 205 | u32 tgcr; |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 206 | |
| 207 | base[i] = ioremap(dtip[i].base, SZ_4K); |
| 208 | if (WARN_ON(!base[i])) |
| 209 | continue; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 210 | |
| 211 | /* Disabled, Internal clock source */ |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 212 | __raw_writel(0, base[i] + TCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 213 | |
| 214 | /* reset both timers, no pre-scaler for timer34 */ |
| 215 | tgcr = 0; |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 216 | __raw_writel(tgcr, base[i] + TGCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 217 | |
| 218 | /* Set both timers to unchained 32-bit */ |
| 219 | tgcr = TGCR_TIMMODE_32BIT_UNCHAINED << TGCR_TIMMODE_SHIFT; |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 220 | __raw_writel(tgcr, base[i] + TGCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 221 | |
| 222 | /* Unreset timers */ |
| 223 | tgcr |= (TGCR_UNRESET << TGCR_TIM12RS_SHIFT) | |
| 224 | (TGCR_UNRESET << TGCR_TIM34RS_SHIFT); |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 225 | __raw_writel(tgcr, base[i] + TGCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 226 | |
| 227 | /* Init both counters to zero */ |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 228 | __raw_writel(0, base[i] + TIM12); |
| 229 | __raw_writel(0, base[i] + TIM34); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 230 | } |
| 231 | |
| 232 | /* Init of each timer as a 32-bit timer */ |
| 233 | for (i=0; i< ARRAY_SIZE(timers); i++) { |
| 234 | struct timer_s *t = &timers[i]; |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 235 | int timer = ID_TO_TIMER(t->id); |
| 236 | u32 irq; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 237 | |
Cyril Chemparathy | 1bcd38a | 2010-05-07 17:06:35 -0400 | [diff] [blame] | 238 | t->base = base[timer]; |
| 239 | if (!t->base) |
| 240 | continue; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 241 | |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 242 | if (IS_TIMER_BOT(t->id)) { |
| 243 | t->enamode_shift = 6; |
| 244 | t->tim_off = TIM12; |
| 245 | t->prd_off = PRD12; |
| 246 | irq = dtip[timer].bottom_irq; |
| 247 | } else { |
| 248 | t->enamode_shift = 22; |
| 249 | t->tim_off = TIM34; |
| 250 | t->prd_off = PRD34; |
| 251 | irq = dtip[timer].top_irq; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 252 | } |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 253 | |
| 254 | /* Register interrupt */ |
| 255 | t->irqaction.name = t->name; |
| 256 | t->irqaction.dev_id = (void *)t; |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 257 | |
| 258 | if (t->irqaction.handler != NULL) { |
| 259 | irq = USING_COMPARE(t) ? dtip[i].cmp_irq : irq; |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 260 | setup_irq(irq, &t->irqaction); |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 261 | } |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 262 | } |
| 263 | } |
| 264 | |
| 265 | /* |
| 266 | * clocksource |
| 267 | */ |
Magnus Damm | 8e19608 | 2009-04-21 12:24:00 -0700 | [diff] [blame] | 268 | static cycle_t read_cycles(struct clocksource *cs) |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 269 | { |
| 270 | struct timer_s *t = &timers[TID_CLOCKSOURCE]; |
| 271 | |
| 272 | return (cycles_t)timer32_read(t); |
| 273 | } |
| 274 | |
| 275 | static struct clocksource clocksource_davinci = { |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 276 | .rating = 300, |
| 277 | .read = read_cycles, |
| 278 | .mask = CLOCKSOURCE_MASK(32), |
| 279 | .shift = 24, |
| 280 | .flags = CLOCK_SOURCE_IS_CONTINUOUS, |
| 281 | }; |
| 282 | |
| 283 | /* |
| 284 | * clockevent |
| 285 | */ |
| 286 | static int davinci_set_next_event(unsigned long cycles, |
| 287 | struct clock_event_device *evt) |
| 288 | { |
| 289 | struct timer_s *t = &timers[TID_CLOCKEVENT]; |
| 290 | |
| 291 | t->period = cycles; |
| 292 | timer32_config(t); |
| 293 | return 0; |
| 294 | } |
| 295 | |
| 296 | static void davinci_set_mode(enum clock_event_mode mode, |
| 297 | struct clock_event_device *evt) |
| 298 | { |
| 299 | struct timer_s *t = &timers[TID_CLOCKEVENT]; |
| 300 | |
| 301 | switch (mode) { |
| 302 | case CLOCK_EVT_MODE_PERIODIC: |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 303 | t->period = davinci_clock_tick_rate / (HZ); |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 304 | t->opts &= ~TIMER_OPTS_STATE_MASK; |
| 305 | t->opts |= TIMER_OPTS_PERIODIC; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 306 | timer32_config(t); |
| 307 | break; |
| 308 | case CLOCK_EVT_MODE_ONESHOT: |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 309 | t->opts &= ~TIMER_OPTS_STATE_MASK; |
| 310 | t->opts |= TIMER_OPTS_ONESHOT; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 311 | break; |
| 312 | case CLOCK_EVT_MODE_UNUSED: |
| 313 | case CLOCK_EVT_MODE_SHUTDOWN: |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 314 | t->opts &= ~TIMER_OPTS_STATE_MASK; |
| 315 | t->opts |= TIMER_OPTS_DISABLED; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 316 | break; |
Thomas Gleixner | 18de5bc | 2007-07-21 04:37:34 -0700 | [diff] [blame] | 317 | case CLOCK_EVT_MODE_RESUME: |
| 318 | break; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 319 | } |
| 320 | } |
| 321 | |
| 322 | static struct clock_event_device clockevent_davinci = { |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 323 | .features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT, |
| 324 | .shift = 32, |
| 325 | .set_next_event = davinci_set_next_event, |
| 326 | .set_mode = davinci_set_mode, |
| 327 | }; |
| 328 | |
| 329 | |
| 330 | static void __init davinci_timer_init(void) |
| 331 | { |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 332 | struct clk *timer_clk; |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 333 | struct davinci_soc_info *soc_info = &davinci_soc_info; |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 334 | unsigned int clockevent_id; |
| 335 | unsigned int clocksource_id; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 336 | static char err[] __initdata = KERN_ERR |
| 337 | "%s: can't register clocksource!\n"; |
Kevin Hilman | d99c387 | 2010-03-11 14:57:35 -0800 | [diff] [blame] | 338 | int i; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 339 | |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 340 | clockevent_id = soc_info->timer_info->clockevent_id; |
| 341 | clocksource_id = soc_info->timer_info->clocksource_id; |
| 342 | |
| 343 | timers[TID_CLOCKEVENT].id = clockevent_id; |
| 344 | timers[TID_CLOCKSOURCE].id = clocksource_id; |
| 345 | |
| 346 | /* |
| 347 | * If using same timer for both clock events & clocksource, |
| 348 | * a compare register must be used to generate an event interrupt. |
| 349 | * This is equivalent to a oneshot timer only (not periodic). |
| 350 | */ |
| 351 | if (clockevent_id == clocksource_id) { |
| 352 | struct davinci_timer_instance *dtip = |
| 353 | soc_info->timer_info->timers; |
| 354 | int event_timer = ID_TO_TIMER(clockevent_id); |
| 355 | |
| 356 | /* Only bottom timers can use compare regs */ |
| 357 | if (IS_TIMER_TOP(clockevent_id)) |
| 358 | pr_warning("davinci_timer_init: Invalid use" |
| 359 | " of system timers. Results unpredictable.\n"); |
| 360 | else if ((dtip[event_timer].cmp_off == 0) |
| 361 | || (dtip[event_timer].cmp_irq == 0)) |
| 362 | pr_warning("davinci_timer_init: Invalid timer instance" |
| 363 | " setup. Results unpredictable.\n"); |
| 364 | else { |
| 365 | timers[TID_CLOCKEVENT].opts |= TIMER_OPTS_USE_COMPARE; |
| 366 | clockevent_davinci.features = CLOCK_EVT_FEAT_ONESHOT; |
| 367 | } |
| 368 | } |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 369 | |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 370 | timer_clk = clk_get(NULL, "timer0"); |
| 371 | BUG_ON(IS_ERR(timer_clk)); |
| 372 | clk_enable(timer_clk); |
| 373 | |
Cyril Chemparathy | 8ca2e59 | 2010-03-25 17:43:45 -0400 | [diff] [blame] | 374 | /* init timer hw */ |
| 375 | timer_init(); |
| 376 | |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 377 | davinci_clock_tick_rate = clk_get_rate(timer_clk); |
| 378 | |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 379 | /* setup clocksource */ |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 380 | clocksource_davinci.name = id_to_name[clocksource_id]; |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 381 | clocksource_davinci.mult = |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 382 | clocksource_khz2mult(davinci_clock_tick_rate/1000, |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 383 | clocksource_davinci.shift); |
| 384 | if (clocksource_register(&clocksource_davinci)) |
| 385 | printk(err, clocksource_davinci.name); |
| 386 | |
| 387 | /* setup clockevent */ |
Mark A. Greer | f64691b | 2009-04-15 12:40:11 -0700 | [diff] [blame] | 388 | clockevent_davinci.name = id_to_name[timers[TID_CLOCKEVENT].id]; |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 389 | clockevent_davinci.mult = div_sc(davinci_clock_tick_rate, NSEC_PER_SEC, |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 390 | clockevent_davinci.shift); |
| 391 | clockevent_davinci.max_delta_ns = |
| 392 | clockevent_delta2ns(0xfffffffe, &clockevent_davinci); |
Mark A. Greer | 3abd5ac | 2009-04-15 12:41:54 -0700 | [diff] [blame] | 393 | clockevent_davinci.min_delta_ns = 50000; /* 50 usec */ |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 394 | |
Rusty Russell | 320ab2b | 2008-12-13 21:20:26 +1030 | [diff] [blame] | 395 | clockevent_davinci.cpumask = cpumask_of(0); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 396 | clockevents_register_device(&clockevent_davinci); |
Kevin Hilman | d99c387 | 2010-03-11 14:57:35 -0800 | [diff] [blame] | 397 | |
| 398 | for (i=0; i< ARRAY_SIZE(timers); i++) |
| 399 | timer32_config(&timers[i]); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 400 | } |
| 401 | |
| 402 | struct sys_timer davinci_timer = { |
| 403 | .init = davinci_timer_init, |
| 404 | }; |
| 405 | |
| 406 | |
| 407 | /* reset board using watchdog timer */ |
Cyril Chemparathy | c78a5bc | 2010-05-01 18:38:28 -0400 | [diff] [blame] | 408 | void davinci_watchdog_reset(struct platform_device *pdev) |
Kevin Hilman | fb63138 | 2009-04-29 16:23:59 -0700 | [diff] [blame] | 409 | { |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 410 | u32 tgcr, wdtcr; |
Cyril Chemparathy | c78a5bc | 2010-05-01 18:38:28 -0400 | [diff] [blame] | 411 | void __iomem *base; |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 412 | struct clk *wd_clk; |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 413 | |
Cyril Chemparathy | c78a5bc | 2010-05-01 18:38:28 -0400 | [diff] [blame] | 414 | base = ioremap(pdev->resource[0].start, SZ_4K); |
| 415 | if (WARN_ON(!base)) |
| 416 | return; |
| 417 | |
Kevin Hilman | 5fcd294 | 2009-06-03 12:24:50 -0700 | [diff] [blame] | 418 | wd_clk = clk_get(&pdev->dev, NULL); |
Kevin Hilman | e609900 | 2009-04-14 07:06:37 -0500 | [diff] [blame] | 419 | if (WARN_ON(IS_ERR(wd_clk))) |
| 420 | return; |
| 421 | clk_enable(wd_clk); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 422 | |
| 423 | /* disable, internal clock source */ |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 424 | __raw_writel(0, base + TCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 425 | |
| 426 | /* reset timer, set mode to 64-bit watchdog, and unreset */ |
| 427 | tgcr = 0; |
David Griego | a23f7dc | 2009-06-01 11:41:54 -0700 | [diff] [blame] | 428 | __raw_writel(tgcr, base + TGCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 429 | tgcr = TGCR_TIMMODE_64BIT_WDOG << TGCR_TIMMODE_SHIFT; |
| 430 | tgcr |= (TGCR_UNRESET << TGCR_TIM12RS_SHIFT) | |
| 431 | (TGCR_UNRESET << TGCR_TIM34RS_SHIFT); |
David Griego | a23f7dc | 2009-06-01 11:41:54 -0700 | [diff] [blame] | 432 | __raw_writel(tgcr, base + TGCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 433 | |
| 434 | /* clear counter and period regs */ |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 435 | __raw_writel(0, base + TIM12); |
| 436 | __raw_writel(0, base + TIM34); |
| 437 | __raw_writel(0, base + PRD12); |
| 438 | __raw_writel(0, base + PRD34); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 439 | |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 440 | /* put watchdog in pre-active state */ |
David Griego | a23f7dc | 2009-06-01 11:41:54 -0700 | [diff] [blame] | 441 | wdtcr = __raw_readl(base + WDTCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 442 | wdtcr = (WDTCR_WDKEY_SEQ0 << WDTCR_WDKEY_SHIFT) | |
| 443 | (WDTCR_WDEN_ENABLE << WDTCR_WDEN_SHIFT); |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 444 | __raw_writel(wdtcr, base + WDTCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 445 | |
| 446 | /* put watchdog in active state */ |
| 447 | wdtcr = (WDTCR_WDKEY_SEQ1 << WDTCR_WDKEY_SHIFT) | |
| 448 | (WDTCR_WDEN_ENABLE << WDTCR_WDEN_SHIFT); |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 449 | __raw_writel(wdtcr, base + WDTCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 450 | |
| 451 | /* write an invalid value to the WDKEY field to trigger |
| 452 | * a watchdog reset */ |
| 453 | wdtcr = 0x00004000; |
Kevin Hilman | f5c122d | 2009-04-14 07:04:16 -0500 | [diff] [blame] | 454 | __raw_writel(wdtcr, base + WDTCR); |
Kevin Hilman | 7c6337e | 2007-04-30 19:37:19 +0100 | [diff] [blame] | 455 | } |