Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | /* |
| 2 | * linux/arch/arm/mach-pxa/irq.c |
| 3 | * |
eric miao | e3630db | 2008-03-04 11:42:26 +0800 | [diff] [blame] | 4 | * Generic PXA IRQ handling |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 5 | * |
| 6 | * Author: Nicolas Pitre |
| 7 | * Created: Jun 15, 2001 |
| 8 | * Copyright: MontaVista Software Inc. |
| 9 | * |
| 10 | * This program is free software; you can redistribute it and/or modify |
| 11 | * it under the terms of the GNU General Public License version 2 as |
| 12 | * published by the Free Software Foundation. |
| 13 | */ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 14 | #include <linux/init.h> |
| 15 | #include <linux/module.h> |
| 16 | #include <linux/interrupt.h> |
Rafael J. Wysocki | 2eaa03b | 2011-04-22 22:03:11 +0200 | [diff] [blame] | 17 | #include <linux/syscore_ops.h> |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 18 | #include <linux/io.h> |
| 19 | #include <linux/irq.h> |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 20 | #include <linux/of_address.h> |
| 21 | #include <linux/of_irq.h> |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 22 | |
Jamie Iles | 5a567d7 | 2011-10-08 11:20:42 +0100 | [diff] [blame] | 23 | #include <asm/exception.h> |
| 24 | |
Russell King | a09e64f | 2008-08-05 16:14:15 +0100 | [diff] [blame] | 25 | #include <mach/hardware.h> |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 26 | #include <mach/irqs.h> |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 27 | |
| 28 | #include "generic.h" |
| 29 | |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 30 | #define ICIP (0x000) |
| 31 | #define ICMR (0x004) |
| 32 | #define ICLR (0x008) |
| 33 | #define ICFR (0x00c) |
| 34 | #define ICPR (0x010) |
| 35 | #define ICCR (0x014) |
| 36 | #define ICHP (0x018) |
| 37 | #define IPR(i) (((i) < 32) ? (0x01c + ((i) << 2)) : \ |
| 38 | ((i) < 64) ? (0x0b0 + (((i) - 32) << 2)) : \ |
| 39 | (0x144 + (((i) - 64) << 2))) |
Eric Miao | a551e4f | 2011-04-27 22:48:05 +0800 | [diff] [blame] | 40 | #define ICHP_VAL_IRQ (1 << 31) |
| 41 | #define ICHP_IRQ(i) (((i) >> 16) & 0x7fff) |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 42 | #define IPR_VALID (1 << 31) |
| 43 | #define IRQ_BIT(n) (((n) - PXA_IRQ(0)) & 0x1f) |
| 44 | |
| 45 | #define MAX_INTERNAL_IRQS 128 |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 46 | |
| 47 | /* |
| 48 | * This is for peripheral IRQs internal to the PXA chip. |
| 49 | */ |
| 50 | |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 51 | static void __iomem *pxa_irq_base; |
eric miao | f6fb7af | 2008-03-04 13:53:05 +0800 | [diff] [blame] | 52 | static int pxa_internal_irq_nr; |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 53 | static bool cpu_has_ipr; |
Haojian Zhuang | bb71bdd | 2010-11-17 19:03:36 +0800 | [diff] [blame] | 54 | |
Eric Miao | a1015a1 | 2011-01-12 16:42:24 -0600 | [diff] [blame] | 55 | static inline void __iomem *irq_base(int i) |
| 56 | { |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 57 | static unsigned long phys_base_offset[] = { |
| 58 | 0x0, |
| 59 | 0x9c, |
| 60 | 0x130, |
Eric Miao | a1015a1 | 2011-01-12 16:42:24 -0600 | [diff] [blame] | 61 | }; |
| 62 | |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 63 | return pxa_irq_base + phys_base_offset[i]; |
Eric Miao | a1015a1 | 2011-01-12 16:42:24 -0600 | [diff] [blame] | 64 | } |
| 65 | |
Eric Miao | 5d284e3 | 2011-04-27 22:48:04 +0800 | [diff] [blame] | 66 | void pxa_mask_irq(struct irq_data *d) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 67 | { |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 68 | void __iomem *base = irq_data_get_irq_chip_data(d); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 69 | uint32_t icmr = __raw_readl(base + ICMR); |
| 70 | |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 71 | icmr &= ~(1 << IRQ_BIT(d->irq)); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 72 | __raw_writel(icmr, base + ICMR); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 73 | } |
| 74 | |
Eric Miao | 5d284e3 | 2011-04-27 22:48:04 +0800 | [diff] [blame] | 75 | void pxa_unmask_irq(struct irq_data *d) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 76 | { |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 77 | void __iomem *base = irq_data_get_irq_chip_data(d); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 78 | uint32_t icmr = __raw_readl(base + ICMR); |
| 79 | |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 80 | icmr |= 1 << IRQ_BIT(d->irq); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 81 | __raw_writel(icmr, base + ICMR); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 82 | } |
| 83 | |
eric miao | f6fb7af | 2008-03-04 13:53:05 +0800 | [diff] [blame] | 84 | static struct irq_chip pxa_internal_irq_chip = { |
David Brownell | 38c677c | 2006-08-01 22:26:25 +0100 | [diff] [blame] | 85 | .name = "SC", |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 86 | .irq_ack = pxa_mask_irq, |
| 87 | .irq_mask = pxa_mask_irq, |
| 88 | .irq_unmask = pxa_unmask_irq, |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 89 | }; |
| 90 | |
Eric Miao | a551e4f | 2011-04-27 22:48:05 +0800 | [diff] [blame] | 91 | asmlinkage void __exception_irq_entry icip_handle_irq(struct pt_regs *regs) |
| 92 | { |
| 93 | uint32_t icip, icmr, mask; |
| 94 | |
| 95 | do { |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 96 | icip = __raw_readl(pxa_irq_base + ICIP); |
| 97 | icmr = __raw_readl(pxa_irq_base + ICMR); |
Eric Miao | a551e4f | 2011-04-27 22:48:05 +0800 | [diff] [blame] | 98 | mask = icip & icmr; |
| 99 | |
| 100 | if (mask == 0) |
| 101 | break; |
| 102 | |
| 103 | handle_IRQ(PXA_IRQ(fls(mask) - 1), regs); |
| 104 | } while (1); |
| 105 | } |
| 106 | |
| 107 | asmlinkage void __exception_irq_entry ichp_handle_irq(struct pt_regs *regs) |
| 108 | { |
| 109 | uint32_t ichp; |
| 110 | |
| 111 | do { |
| 112 | __asm__ __volatile__("mrc p6, 0, %0, c5, c0, 0\n": "=r"(ichp)); |
| 113 | |
| 114 | if ((ichp & ICHP_VAL_IRQ) == 0) |
| 115 | break; |
| 116 | |
| 117 | handle_IRQ(PXA_IRQ(ICHP_IRQ(ichp)), regs); |
| 118 | } while (1); |
| 119 | } |
| 120 | |
Haojian Zhuang | 157d264 | 2011-10-17 20:37:52 +0800 | [diff] [blame] | 121 | void __init pxa_init_irq(int irq_nr, int (*fn)(struct irq_data *, unsigned int)) |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 122 | { |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 123 | int irq, i, n; |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 124 | |
Haojian Zhuang | c482ae4 | 2009-11-02 14:02:21 -0500 | [diff] [blame] | 125 | BUG_ON(irq_nr > MAX_INTERNAL_IRQS); |
| 126 | |
eric miao | f6fb7af | 2008-03-04 13:53:05 +0800 | [diff] [blame] | 127 | pxa_internal_irq_nr = irq_nr; |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 128 | cpu_has_ipr = !cpu_is_pxa25x(); |
| 129 | pxa_irq_base = io_p2v(0x40d00000); |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 130 | |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 131 | for (n = 0; n < irq_nr; n += 32) { |
Marek Vasut | 1b624fb | 2011-01-10 23:53:12 +0100 | [diff] [blame] | 132 | void __iomem *base = irq_base(n >> 5); |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 133 | |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 134 | __raw_writel(0, base + ICMR); /* disable all IRQs */ |
| 135 | __raw_writel(0, base + ICLR); /* all IRQs are IRQ, not FIQ */ |
| 136 | for (i = n; (i < (n + 32)) && (i < irq_nr); i++) { |
| 137 | /* initialize interrupt priority */ |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 138 | if (cpu_has_ipr) |
| 139 | __raw_writel(i | IPR_VALID, pxa_irq_base + IPR(i)); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 140 | |
| 141 | irq = PXA_IRQ(i); |
Thomas Gleixner | f38c02f | 2011-03-24 13:35:09 +0100 | [diff] [blame] | 142 | irq_set_chip_and_handler(irq, &pxa_internal_irq_chip, |
| 143 | handle_level_irq); |
Thomas Gleixner | 9323f261 | 2011-03-24 13:29:39 +0100 | [diff] [blame] | 144 | irq_set_chip_data(irq, base); |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 145 | set_irq_flags(irq, IRQF_VALID); |
| 146 | } |
Haojian Zhuang | d2c3706 | 2009-08-19 19:49:31 +0800 | [diff] [blame] | 147 | } |
| 148 | |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 149 | /* only unmasked interrupts kick us out of idle */ |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 150 | __raw_writel(1, irq_base(0) + ICCR); |
Eric Miao | 53665a5 | 2007-06-06 06:36:04 +0100 | [diff] [blame] | 151 | |
Lennert Buytenhek | a3f4c92 | 2010-11-29 11:18:26 +0100 | [diff] [blame] | 152 | pxa_internal_irq_chip.irq_set_wake = fn; |
eric miao | c95530c | 2007-08-29 10:22:17 +0100 | [diff] [blame] | 153 | } |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 154 | |
| 155 | #ifdef CONFIG_PM |
Haojian Zhuang | c482ae4 | 2009-11-02 14:02:21 -0500 | [diff] [blame] | 156 | static unsigned long saved_icmr[MAX_INTERNAL_IRQS/32]; |
| 157 | static unsigned long saved_ipr[MAX_INTERNAL_IRQS]; |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 158 | |
Rafael J. Wysocki | 2eaa03b | 2011-04-22 22:03:11 +0200 | [diff] [blame] | 159 | static int pxa_irq_suspend(void) |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 160 | { |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 161 | int i; |
eric miao | f6fb7af | 2008-03-04 13:53:05 +0800 | [diff] [blame] | 162 | |
Marek Vasut | 1b624fb | 2011-01-10 23:53:12 +0100 | [diff] [blame] | 163 | for (i = 0; i < pxa_internal_irq_nr / 32; i++) { |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 164 | void __iomem *base = irq_base(i); |
| 165 | |
| 166 | saved_icmr[i] = __raw_readl(base + ICMR); |
| 167 | __raw_writel(0, base + ICMR); |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 168 | } |
Eric Miao | c70f5a6 | 2010-01-11 20:39:37 +0800 | [diff] [blame] | 169 | |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 170 | if (cpu_has_ipr) { |
Eric Miao | c70f5a6 | 2010-01-11 20:39:37 +0800 | [diff] [blame] | 171 | for (i = 0; i < pxa_internal_irq_nr; i++) |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 172 | saved_ipr[i] = __raw_readl(pxa_irq_base + IPR(i)); |
Eric Miao | c70f5a6 | 2010-01-11 20:39:37 +0800 | [diff] [blame] | 173 | } |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 174 | |
| 175 | return 0; |
| 176 | } |
| 177 | |
Rafael J. Wysocki | 2eaa03b | 2011-04-22 22:03:11 +0200 | [diff] [blame] | 178 | static void pxa_irq_resume(void) |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 179 | { |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 180 | int i; |
eric miao | f6fb7af | 2008-03-04 13:53:05 +0800 | [diff] [blame] | 181 | |
Marek Vasut | 1b624fb | 2011-01-10 23:53:12 +0100 | [diff] [blame] | 182 | for (i = 0; i < pxa_internal_irq_nr / 32; i++) { |
Haojian Zhuang | a79a9ad | 2010-11-24 11:54:22 +0800 | [diff] [blame] | 183 | void __iomem *base = irq_base(i); |
| 184 | |
| 185 | __raw_writel(saved_icmr[i], base + ICMR); |
| 186 | __raw_writel(0, base + ICLR); |
| 187 | } |
| 188 | |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 189 | if (cpu_has_ipr) |
Eric Miao | c70f5a6 | 2010-01-11 20:39:37 +0800 | [diff] [blame] | 190 | for (i = 0; i < pxa_internal_irq_nr; i++) |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 191 | __raw_writel(saved_ipr[i], pxa_irq_base + IPR(i)); |
Eric Miao | c70f5a6 | 2010-01-11 20:39:37 +0800 | [diff] [blame] | 192 | |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 193 | __raw_writel(1, pxa_irq_base + ICCR); |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 194 | } |
| 195 | #else |
| 196 | #define pxa_irq_suspend NULL |
| 197 | #define pxa_irq_resume NULL |
| 198 | #endif |
| 199 | |
Rafael J. Wysocki | 2eaa03b | 2011-04-22 22:03:11 +0200 | [diff] [blame] | 200 | struct syscore_ops pxa_irq_syscore_ops = { |
eric miao | c0165504 | 2008-01-28 23:00:02 +0000 | [diff] [blame] | 201 | .suspend = pxa_irq_suspend, |
| 202 | .resume = pxa_irq_resume, |
| 203 | }; |
Daniel Mack | 089d036 | 2012-07-22 19:50:22 +0200 | [diff] [blame] | 204 | |
| 205 | #ifdef CONFIG_OF |
| 206 | static struct irq_domain *pxa_irq_domain; |
| 207 | |
| 208 | static int pxa_irq_map(struct irq_domain *h, unsigned int virq, |
| 209 | irq_hw_number_t hw) |
| 210 | { |
| 211 | void __iomem *base = irq_base(hw / 32); |
| 212 | |
| 213 | /* initialize interrupt priority */ |
| 214 | if (cpu_has_ipr) |
| 215 | __raw_writel(hw | IPR_VALID, pxa_irq_base + IPR(hw)); |
| 216 | |
| 217 | irq_set_chip_and_handler(hw, &pxa_internal_irq_chip, |
| 218 | handle_level_irq); |
| 219 | irq_set_chip_data(hw, base); |
| 220 | set_irq_flags(hw, IRQF_VALID); |
| 221 | |
| 222 | return 0; |
| 223 | } |
| 224 | |
| 225 | static struct irq_domain_ops pxa_irq_ops = { |
| 226 | .map = pxa_irq_map, |
| 227 | .xlate = irq_domain_xlate_onecell, |
| 228 | }; |
| 229 | |
| 230 | static const struct of_device_id intc_ids[] __initconst = { |
| 231 | { .compatible = "marvell,pxa-intc", }, |
| 232 | {} |
| 233 | }; |
| 234 | |
| 235 | void __init pxa_dt_irq_init(int (*fn)(struct irq_data *, unsigned int)) |
| 236 | { |
| 237 | struct device_node *node; |
| 238 | const struct of_device_id *of_id; |
| 239 | struct pxa_intc_conf *conf; |
| 240 | struct resource res; |
| 241 | int n, ret; |
| 242 | |
| 243 | node = of_find_matching_node(NULL, intc_ids); |
| 244 | if (!node) { |
| 245 | pr_err("Failed to find interrupt controller in arch-pxa\n"); |
| 246 | return; |
| 247 | } |
| 248 | of_id = of_match_node(intc_ids, node); |
| 249 | conf = of_id->data; |
| 250 | |
| 251 | ret = of_property_read_u32(node, "marvell,intc-nr-irqs", |
| 252 | &pxa_internal_irq_nr); |
| 253 | if (ret) { |
| 254 | pr_err("Not found marvell,intc-nr-irqs property\n"); |
| 255 | return; |
| 256 | } |
| 257 | |
| 258 | ret = of_address_to_resource(node, 0, &res); |
| 259 | if (ret < 0) { |
| 260 | pr_err("No registers defined for node\n"); |
| 261 | return; |
| 262 | } |
| 263 | pxa_irq_base = io_p2v(res.start); |
| 264 | |
| 265 | if (of_find_property(node, "marvell,intc-priority", NULL)) |
| 266 | cpu_has_ipr = 1; |
| 267 | |
| 268 | ret = irq_alloc_descs(-1, 0, pxa_internal_irq_nr, 0); |
| 269 | if (ret < 0) { |
| 270 | pr_err("Failed to allocate IRQ numbers\n"); |
| 271 | return; |
| 272 | } |
| 273 | |
| 274 | pxa_irq_domain = irq_domain_add_legacy(node, pxa_internal_irq_nr, 0, 0, |
| 275 | &pxa_irq_ops, NULL); |
| 276 | if (!pxa_irq_domain) |
| 277 | panic("Unable to add PXA IRQ domain\n"); |
| 278 | |
| 279 | irq_set_default_host(pxa_irq_domain); |
| 280 | |
| 281 | for (n = 0; n < pxa_internal_irq_nr; n += 32) { |
| 282 | void __iomem *base = irq_base(n >> 5); |
| 283 | |
| 284 | __raw_writel(0, base + ICMR); /* disable all IRQs */ |
| 285 | __raw_writel(0, base + ICLR); /* all IRQs are IRQ, not FIQ */ |
| 286 | } |
| 287 | |
| 288 | /* only unmasked interrupts kick us out of idle */ |
| 289 | __raw_writel(1, irq_base(0) + ICCR); |
| 290 | |
| 291 | pxa_internal_irq_chip.irq_set_wake = fn; |
| 292 | } |
| 293 | #endif /* CONFIG_OF */ |