Haojian Zhuang | 8e4bebe | 2014-08-07 18:51:34 +0800 | [diff] [blame] | 1 | /* |
| 2 | * Hisilicon HiP04 INTC |
| 3 | * |
| 4 | * Copyright (C) 2002-2014 ARM Limited. |
| 5 | * Copyright (c) 2013-2014 Hisilicon Ltd. |
| 6 | * Copyright (c) 2013-2014 Linaro Ltd. |
| 7 | * |
| 8 | * This program is free software; you can redistribute it and/or modify |
| 9 | * it under the terms of the GNU General Public License version 2 as |
| 10 | * published by the Free Software Foundation. |
| 11 | * |
| 12 | * Interrupt architecture for the HIP04 INTC: |
| 13 | * |
| 14 | * o There is one Interrupt Distributor, which receives interrupts |
| 15 | * from system devices and sends them to the Interrupt Controllers. |
| 16 | * |
| 17 | * o There is one CPU Interface per CPU, which sends interrupts sent |
| 18 | * by the Distributor, and interrupts generated locally, to the |
| 19 | * associated CPU. The base address of the CPU interface is usually |
| 20 | * aliased so that the same address points to different chips depending |
| 21 | * on the CPU it is accessed from. |
| 22 | * |
| 23 | * Note that IRQs 0-31 are special - they are local to each CPU. |
| 24 | * As such, the enable set/clear, pending set/clear and active bit |
| 25 | * registers are banked per-cpu for these sources. |
| 26 | */ |
| 27 | |
| 28 | #include <linux/init.h> |
| 29 | #include <linux/kernel.h> |
| 30 | #include <linux/err.h> |
| 31 | #include <linux/module.h> |
| 32 | #include <linux/list.h> |
| 33 | #include <linux/smp.h> |
| 34 | #include <linux/cpu.h> |
| 35 | #include <linux/cpu_pm.h> |
| 36 | #include <linux/cpumask.h> |
| 37 | #include <linux/io.h> |
| 38 | #include <linux/of.h> |
| 39 | #include <linux/of_address.h> |
| 40 | #include <linux/of_irq.h> |
| 41 | #include <linux/irqdomain.h> |
| 42 | #include <linux/interrupt.h> |
| 43 | #include <linux/slab.h> |
| 44 | #include <linux/irqchip/arm-gic.h> |
| 45 | |
| 46 | #include <asm/irq.h> |
| 47 | #include <asm/exception.h> |
| 48 | #include <asm/smp_plat.h> |
| 49 | |
| 50 | #include "irq-gic-common.h" |
| 51 | #include "irqchip.h" |
| 52 | |
| 53 | #define HIP04_MAX_IRQS 510 |
| 54 | |
| 55 | struct hip04_irq_data { |
| 56 | void __iomem *dist_base; |
| 57 | void __iomem *cpu_base; |
| 58 | struct irq_domain *domain; |
| 59 | unsigned int nr_irqs; |
| 60 | }; |
| 61 | |
| 62 | static DEFINE_RAW_SPINLOCK(irq_controller_lock); |
| 63 | |
| 64 | /* |
| 65 | * The GIC mapping of CPU interfaces does not necessarily match |
| 66 | * the logical CPU numbering. Let's use a mapping as returned |
| 67 | * by the GIC itself. |
| 68 | */ |
| 69 | #define NR_HIP04_CPU_IF 16 |
| 70 | static u16 hip04_cpu_map[NR_HIP04_CPU_IF] __read_mostly; |
| 71 | |
| 72 | static struct hip04_irq_data hip04_data __read_mostly; |
| 73 | |
| 74 | static inline void __iomem *hip04_dist_base(struct irq_data *d) |
| 75 | { |
| 76 | struct hip04_irq_data *hip04_data = irq_data_get_irq_chip_data(d); |
| 77 | return hip04_data->dist_base; |
| 78 | } |
| 79 | |
| 80 | static inline void __iomem *hip04_cpu_base(struct irq_data *d) |
| 81 | { |
| 82 | struct hip04_irq_data *hip04_data = irq_data_get_irq_chip_data(d); |
| 83 | return hip04_data->cpu_base; |
| 84 | } |
| 85 | |
| 86 | static inline unsigned int hip04_irq(struct irq_data *d) |
| 87 | { |
| 88 | return d->hwirq; |
| 89 | } |
| 90 | |
| 91 | /* |
| 92 | * Routines to acknowledge, disable and enable interrupts |
| 93 | */ |
| 94 | static void hip04_mask_irq(struct irq_data *d) |
| 95 | { |
| 96 | u32 mask = 1 << (hip04_irq(d) % 32); |
| 97 | |
| 98 | raw_spin_lock(&irq_controller_lock); |
| 99 | writel_relaxed(mask, hip04_dist_base(d) + GIC_DIST_ENABLE_CLEAR + |
| 100 | (hip04_irq(d) / 32) * 4); |
| 101 | raw_spin_unlock(&irq_controller_lock); |
| 102 | } |
| 103 | |
| 104 | static void hip04_unmask_irq(struct irq_data *d) |
| 105 | { |
| 106 | u32 mask = 1 << (hip04_irq(d) % 32); |
| 107 | |
| 108 | raw_spin_lock(&irq_controller_lock); |
| 109 | writel_relaxed(mask, hip04_dist_base(d) + GIC_DIST_ENABLE_SET + |
| 110 | (hip04_irq(d) / 32) * 4); |
| 111 | raw_spin_unlock(&irq_controller_lock); |
| 112 | } |
| 113 | |
| 114 | static void hip04_eoi_irq(struct irq_data *d) |
| 115 | { |
| 116 | writel_relaxed(hip04_irq(d), hip04_cpu_base(d) + GIC_CPU_EOI); |
| 117 | } |
| 118 | |
| 119 | static int hip04_irq_set_type(struct irq_data *d, unsigned int type) |
| 120 | { |
| 121 | void __iomem *base = hip04_dist_base(d); |
| 122 | unsigned int irq = hip04_irq(d); |
| 123 | |
| 124 | /* Interrupt configuration for SGIs can't be changed */ |
| 125 | if (irq < 16) |
| 126 | return -EINVAL; |
| 127 | |
| 128 | if (type != IRQ_TYPE_LEVEL_HIGH && type != IRQ_TYPE_EDGE_RISING) |
| 129 | return -EINVAL; |
| 130 | |
| 131 | raw_spin_lock(&irq_controller_lock); |
| 132 | |
| 133 | gic_configure_irq(irq, type, base, NULL); |
| 134 | |
| 135 | raw_spin_unlock(&irq_controller_lock); |
| 136 | |
| 137 | return 0; |
| 138 | } |
| 139 | |
| 140 | #ifdef CONFIG_SMP |
| 141 | static int hip04_irq_set_affinity(struct irq_data *d, |
| 142 | const struct cpumask *mask_val, |
| 143 | bool force) |
| 144 | { |
| 145 | void __iomem *reg; |
| 146 | unsigned int cpu, shift = (hip04_irq(d) % 2) * 16; |
| 147 | u32 val, mask, bit; |
| 148 | |
| 149 | if (!force) |
| 150 | cpu = cpumask_any_and(mask_val, cpu_online_mask); |
| 151 | else |
| 152 | cpu = cpumask_first(mask_val); |
| 153 | |
| 154 | if (cpu >= NR_HIP04_CPU_IF || cpu >= nr_cpu_ids) |
| 155 | return -EINVAL; |
| 156 | |
| 157 | raw_spin_lock(&irq_controller_lock); |
| 158 | reg = hip04_dist_base(d) + GIC_DIST_TARGET + ((hip04_irq(d) * 2) & ~3); |
| 159 | mask = 0xffff << shift; |
| 160 | bit = hip04_cpu_map[cpu] << shift; |
| 161 | val = readl_relaxed(reg) & ~mask; |
| 162 | writel_relaxed(val | bit, reg); |
| 163 | raw_spin_unlock(&irq_controller_lock); |
| 164 | |
| 165 | return IRQ_SET_MASK_OK; |
| 166 | } |
| 167 | #endif |
| 168 | |
| 169 | static void __exception_irq_entry hip04_handle_irq(struct pt_regs *regs) |
| 170 | { |
| 171 | u32 irqstat, irqnr; |
| 172 | void __iomem *cpu_base = hip04_data.cpu_base; |
| 173 | |
| 174 | do { |
| 175 | irqstat = readl_relaxed(cpu_base + GIC_CPU_INTACK); |
| 176 | irqnr = irqstat & GICC_IAR_INT_ID_MASK; |
| 177 | |
| 178 | if (likely(irqnr > 15 && irqnr <= HIP04_MAX_IRQS)) { |
| 179 | irqnr = irq_find_mapping(hip04_data.domain, irqnr); |
| 180 | handle_IRQ(irqnr, regs); |
| 181 | continue; |
| 182 | } |
| 183 | if (irqnr < 16) { |
| 184 | writel_relaxed(irqstat, cpu_base + GIC_CPU_EOI); |
| 185 | #ifdef CONFIG_SMP |
| 186 | handle_IPI(irqnr, regs); |
| 187 | #endif |
| 188 | continue; |
| 189 | } |
| 190 | break; |
| 191 | } while (1); |
| 192 | } |
| 193 | |
| 194 | static struct irq_chip hip04_irq_chip = { |
| 195 | .name = "HIP04 INTC", |
| 196 | .irq_mask = hip04_mask_irq, |
| 197 | .irq_unmask = hip04_unmask_irq, |
| 198 | .irq_eoi = hip04_eoi_irq, |
| 199 | .irq_set_type = hip04_irq_set_type, |
| 200 | #ifdef CONFIG_SMP |
| 201 | .irq_set_affinity = hip04_irq_set_affinity, |
| 202 | #endif |
| 203 | }; |
| 204 | |
| 205 | static u16 hip04_get_cpumask(struct hip04_irq_data *intc) |
| 206 | { |
| 207 | void __iomem *base = intc->dist_base; |
| 208 | u32 mask, i; |
| 209 | |
| 210 | for (i = mask = 0; i < 32; i += 2) { |
| 211 | mask = readl_relaxed(base + GIC_DIST_TARGET + i * 2); |
| 212 | mask |= mask >> 16; |
| 213 | if (mask) |
| 214 | break; |
| 215 | } |
| 216 | |
| 217 | if (!mask) |
| 218 | pr_crit("GIC CPU mask not found - kernel will fail to boot.\n"); |
| 219 | |
| 220 | return mask; |
| 221 | } |
| 222 | |
| 223 | static void __init hip04_irq_dist_init(struct hip04_irq_data *intc) |
| 224 | { |
| 225 | unsigned int i; |
| 226 | u32 cpumask; |
| 227 | unsigned int nr_irqs = intc->nr_irqs; |
| 228 | void __iomem *base = intc->dist_base; |
| 229 | |
| 230 | writel_relaxed(0, base + GIC_DIST_CTRL); |
| 231 | |
| 232 | /* |
| 233 | * Set all global interrupts to this CPU only. |
| 234 | */ |
| 235 | cpumask = hip04_get_cpumask(intc); |
| 236 | cpumask |= cpumask << 16; |
| 237 | for (i = 32; i < nr_irqs; i += 2) |
| 238 | writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3)); |
| 239 | |
| 240 | gic_dist_config(base, nr_irqs, NULL); |
| 241 | |
| 242 | writel_relaxed(1, base + GIC_DIST_CTRL); |
| 243 | } |
| 244 | |
| 245 | static void hip04_irq_cpu_init(struct hip04_irq_data *intc) |
| 246 | { |
| 247 | void __iomem *dist_base = intc->dist_base; |
| 248 | void __iomem *base = intc->cpu_base; |
| 249 | unsigned int cpu_mask, cpu = smp_processor_id(); |
| 250 | int i; |
| 251 | |
| 252 | /* |
| 253 | * Get what the GIC says our CPU mask is. |
| 254 | */ |
| 255 | BUG_ON(cpu >= NR_HIP04_CPU_IF); |
| 256 | cpu_mask = hip04_get_cpumask(intc); |
| 257 | hip04_cpu_map[cpu] = cpu_mask; |
| 258 | |
| 259 | /* |
| 260 | * Clear our mask from the other map entries in case they're |
| 261 | * still undefined. |
| 262 | */ |
| 263 | for (i = 0; i < NR_HIP04_CPU_IF; i++) |
| 264 | if (i != cpu) |
| 265 | hip04_cpu_map[i] &= ~cpu_mask; |
| 266 | |
| 267 | gic_cpu_config(dist_base, NULL); |
| 268 | |
| 269 | writel_relaxed(0xf0, base + GIC_CPU_PRIMASK); |
| 270 | writel_relaxed(1, base + GIC_CPU_CTRL); |
| 271 | } |
| 272 | |
| 273 | #ifdef CONFIG_SMP |
| 274 | static void hip04_raise_softirq(const struct cpumask *mask, unsigned int irq) |
| 275 | { |
| 276 | int cpu; |
| 277 | unsigned long flags, map = 0; |
| 278 | |
| 279 | raw_spin_lock_irqsave(&irq_controller_lock, flags); |
| 280 | |
| 281 | /* Convert our logical CPU mask into a physical one. */ |
| 282 | for_each_cpu(cpu, mask) |
| 283 | map |= hip04_cpu_map[cpu]; |
| 284 | |
| 285 | /* |
| 286 | * Ensure that stores to Normal memory are visible to the |
| 287 | * other CPUs before they observe us issuing the IPI. |
| 288 | */ |
| 289 | dmb(ishst); |
| 290 | |
| 291 | /* this always happens on GIC0 */ |
| 292 | writel_relaxed(map << 8 | irq, hip04_data.dist_base + GIC_DIST_SOFTINT); |
| 293 | |
| 294 | raw_spin_unlock_irqrestore(&irq_controller_lock, flags); |
| 295 | } |
| 296 | #endif |
| 297 | |
| 298 | static int hip04_irq_domain_map(struct irq_domain *d, unsigned int irq, |
| 299 | irq_hw_number_t hw) |
| 300 | { |
| 301 | if (hw < 32) { |
| 302 | irq_set_percpu_devid(irq); |
| 303 | irq_set_chip_and_handler(irq, &hip04_irq_chip, |
| 304 | handle_percpu_devid_irq); |
| 305 | set_irq_flags(irq, IRQF_VALID | IRQF_NOAUTOEN); |
| 306 | } else { |
| 307 | irq_set_chip_and_handler(irq, &hip04_irq_chip, |
| 308 | handle_fasteoi_irq); |
| 309 | set_irq_flags(irq, IRQF_VALID | IRQF_PROBE); |
| 310 | } |
| 311 | irq_set_chip_data(irq, d->host_data); |
| 312 | return 0; |
| 313 | } |
| 314 | |
| 315 | static int hip04_irq_domain_xlate(struct irq_domain *d, |
| 316 | struct device_node *controller, |
| 317 | const u32 *intspec, unsigned int intsize, |
| 318 | unsigned long *out_hwirq, |
| 319 | unsigned int *out_type) |
| 320 | { |
| 321 | unsigned long ret = 0; |
| 322 | |
| 323 | if (d->of_node != controller) |
| 324 | return -EINVAL; |
| 325 | if (intsize < 3) |
| 326 | return -EINVAL; |
| 327 | |
| 328 | /* Get the interrupt number and add 16 to skip over SGIs */ |
| 329 | *out_hwirq = intspec[1] + 16; |
| 330 | |
| 331 | /* For SPIs, we need to add 16 more to get the irq ID number */ |
| 332 | if (!intspec[0]) |
| 333 | *out_hwirq += 16; |
| 334 | |
| 335 | *out_type = intspec[2] & IRQ_TYPE_SENSE_MASK; |
| 336 | |
| 337 | return ret; |
| 338 | } |
| 339 | |
| 340 | #ifdef CONFIG_SMP |
| 341 | static int hip04_irq_secondary_init(struct notifier_block *nfb, |
| 342 | unsigned long action, |
| 343 | void *hcpu) |
| 344 | { |
| 345 | if (action == CPU_STARTING || action == CPU_STARTING_FROZEN) |
| 346 | hip04_irq_cpu_init(&hip04_data); |
| 347 | return NOTIFY_OK; |
| 348 | } |
| 349 | |
| 350 | /* |
| 351 | * Notifier for enabling the INTC CPU interface. Set an arbitrarily high |
| 352 | * priority because the GIC needs to be up before the ARM generic timers. |
| 353 | */ |
| 354 | static struct notifier_block hip04_irq_cpu_notifier = { |
| 355 | .notifier_call = hip04_irq_secondary_init, |
| 356 | .priority = 100, |
| 357 | }; |
| 358 | #endif |
| 359 | |
| 360 | static const struct irq_domain_ops hip04_irq_domain_ops = { |
| 361 | .map = hip04_irq_domain_map, |
| 362 | .xlate = hip04_irq_domain_xlate, |
| 363 | }; |
| 364 | |
| 365 | static int __init |
| 366 | hip04_of_init(struct device_node *node, struct device_node *parent) |
| 367 | { |
| 368 | irq_hw_number_t hwirq_base = 16; |
| 369 | int nr_irqs, irq_base, i; |
| 370 | |
| 371 | if (WARN_ON(!node)) |
| 372 | return -ENODEV; |
| 373 | |
| 374 | hip04_data.dist_base = of_iomap(node, 0); |
| 375 | WARN(!hip04_data.dist_base, "fail to map hip04 intc dist registers\n"); |
| 376 | |
| 377 | hip04_data.cpu_base = of_iomap(node, 1); |
| 378 | WARN(!hip04_data.cpu_base, "unable to map hip04 intc cpu registers\n"); |
| 379 | |
| 380 | /* |
| 381 | * Initialize the CPU interface map to all CPUs. |
| 382 | * It will be refined as each CPU probes its ID. |
| 383 | */ |
| 384 | for (i = 0; i < NR_HIP04_CPU_IF; i++) |
| 385 | hip04_cpu_map[i] = 0xff; |
| 386 | |
| 387 | /* |
| 388 | * Find out how many interrupts are supported. |
| 389 | * The HIP04 INTC only supports up to 510 interrupt sources. |
| 390 | */ |
| 391 | nr_irqs = readl_relaxed(hip04_data.dist_base + GIC_DIST_CTR) & 0x1f; |
| 392 | nr_irqs = (nr_irqs + 1) * 32; |
| 393 | if (nr_irqs > HIP04_MAX_IRQS) |
| 394 | nr_irqs = HIP04_MAX_IRQS; |
| 395 | hip04_data.nr_irqs = nr_irqs; |
| 396 | |
| 397 | nr_irqs -= hwirq_base; /* calculate # of irqs to allocate */ |
| 398 | |
| 399 | irq_base = irq_alloc_descs(-1, hwirq_base, nr_irqs, numa_node_id()); |
| 400 | if (IS_ERR_VALUE(irq_base)) { |
| 401 | pr_err("failed to allocate IRQ numbers\n"); |
| 402 | return -EINVAL; |
| 403 | } |
| 404 | |
| 405 | hip04_data.domain = irq_domain_add_legacy(node, nr_irqs, irq_base, |
| 406 | hwirq_base, |
| 407 | &hip04_irq_domain_ops, |
| 408 | &hip04_data); |
| 409 | |
| 410 | if (WARN_ON(!hip04_data.domain)) |
| 411 | return -EINVAL; |
| 412 | |
| 413 | #ifdef CONFIG_SMP |
| 414 | set_smp_cross_call(hip04_raise_softirq); |
| 415 | register_cpu_notifier(&hip04_irq_cpu_notifier); |
| 416 | #endif |
| 417 | set_handle_irq(hip04_handle_irq); |
| 418 | |
| 419 | hip04_irq_dist_init(&hip04_data); |
| 420 | hip04_irq_cpu_init(&hip04_data); |
| 421 | |
| 422 | return 0; |
| 423 | } |
| 424 | IRQCHIP_DECLARE(hip04_intc, "hisilicon,hip04-intc", hip04_of_init); |