blob: c4849904f0139b4cb8284e51f3efa47798ac534e [file] [log] [blame]
Ralf Baechle41c594a2006-04-05 09:45:45 +01001/*
2 * Malta Platform-specific hooks for SMP operation
3 */
Ralf Baechle45a98eb2007-08-06 16:32:20 +01004#include <linux/irq.h>
Ralf Baechle57a20502007-03-04 18:27:34 +00005#include <linux/init.h>
Ralf Baechle41c594a2006-04-05 09:45:45 +01006
Ralf Baechle57a20502007-03-04 18:27:34 +00007#include <asm/mipsregs.h>
8#include <asm/mipsmtregs.h>
9#include <asm/smtc.h>
Ralf Baechle41c594a2006-04-05 09:45:45 +010010#include <asm/smtc_ipi.h>
Ralf Baechle41c594a2006-04-05 09:45:45 +010011
12/* VPE/SMP Prototype implements platform interfaces directly */
Ralf Baechle41c594a2006-04-05 09:45:45 +010013
14/*
15 * Cause the specified action to be performed on a targeted "CPU"
16 */
17
Ralf Baechle87353d82007-11-19 12:23:51 +000018static void msmtc_send_ipi_single(int cpu, unsigned int action)
Ralf Baechle41c594a2006-04-05 09:45:45 +010019{
Ralf Baechle57a20502007-03-04 18:27:34 +000020 /* "CPU" may be TC of same VPE, VPE of same CPU, or different CPU */
Ralf Baechle41c594a2006-04-05 09:45:45 +010021 smtc_send_ipi(cpu, LINUX_SMP_IPI, action);
Ralf Baechle41c594a2006-04-05 09:45:45 +010022}
23
Rusty Russell48a048f2009-09-24 09:34:44 -060024static void msmtc_send_ipi_mask(const struct cpumask *mask, unsigned int action)
Ralf Baechle41c594a2006-04-05 09:45:45 +010025{
Ralf Baechle87353d82007-11-19 12:23:51 +000026 unsigned int i;
27
Rusty Russell48a048f2009-09-24 09:34:44 -060028 for_each_cpu(i, mask)
Ralf Baechle87353d82007-11-19 12:23:51 +000029 msmtc_send_ipi_single(i, action);
Ralf Baechle41c594a2006-04-05 09:45:45 +010030}
31
32/*
33 * Post-config but pre-boot cleanup entry point
34 */
Paul Gortmaker078a55f2013-06-18 13:38:59 +000035static void msmtc_init_secondary(void)
Ralf Baechle41c594a2006-04-05 09:45:45 +010036{
Ralf Baechle41c594a2006-04-05 09:45:45 +010037 int myvpe;
38
39 /* Don't enable Malta I/O interrupts (IP2) for secondary VPEs */
40 myvpe = read_c0_tcbind() & TCBIND_CURVPE;
41 if (myvpe != 0) {
42 /* Ideally, this should be done only once per VPE, but... */
Ralf Baechledc0366b2007-08-01 19:42:37 +010043 clear_c0_status(ST0_IM);
44 set_c0_status((0x100 << cp0_compare_irq)
45 | (0x100 << MIPS_CPU_IPI_IRQ));
46 if (cp0_perfcount_irq >= 0)
47 set_c0_status(0x100 << cp0_perfcount_irq);
Ralf Baechle41c594a2006-04-05 09:45:45 +010048 }
49
Ralf Baechle87353d82007-11-19 12:23:51 +000050 smtc_init_secondary();
51}
52
53/*
54 * Platform "CPU" startup hook
55 */
Paul Gortmaker078a55f2013-06-18 13:38:59 +000056static void msmtc_boot_secondary(int cpu, struct task_struct *idle)
Ralf Baechle87353d82007-11-19 12:23:51 +000057{
58 smtc_boot_secondary(cpu, idle);
59}
60
61/*
62 * SMP initialization finalization entry point
63 */
Paul Gortmaker078a55f2013-06-18 13:38:59 +000064static void msmtc_smp_finish(void)
Ralf Baechle87353d82007-11-19 12:23:51 +000065{
66 smtc_smp_finish();
67}
68
69/*
70 * Hook for after all CPUs are online
71 */
72
73static void msmtc_cpus_done(void)
74{
Ralf Baechle41c594a2006-04-05 09:45:45 +010075}
76
77/*
78 * Platform SMP pre-initialization
79 *
80 * As noted above, we can assume a single CPU for now
81 * but it may be multithreaded.
82 */
83
Ralf Baechle87353d82007-11-19 12:23:51 +000084static void __init msmtc_smp_setup(void)
Ralf Baechle41c594a2006-04-05 09:45:45 +010085{
Kevin D. Kissell8531a352008-09-09 21:48:52 +020086 /*
87 * we won't get the definitive value until
88 * we've run smtc_prepare_cpus later, but
89 * we would appear to need an upper bound now.
90 */
91 smp_num_siblings = smtc_build_cpu_map(0);
Ralf Baechle41c594a2006-04-05 09:45:45 +010092}
93
Ralf Baechle87353d82007-11-19 12:23:51 +000094static void __init msmtc_prepare_cpus(unsigned int max_cpus)
Ralf Baechle41c594a2006-04-05 09:45:45 +010095{
Kevin D. Kissell8531a352008-09-09 21:48:52 +020096 smtc_prepare_cpus(max_cpus);
Ralf Baechle41c594a2006-04-05 09:45:45 +010097}
98
Ralf Baechle87353d82007-11-19 12:23:51 +000099struct plat_smp_ops msmtc_smp_ops = {
100 .send_ipi_single = msmtc_send_ipi_single,
101 .send_ipi_mask = msmtc_send_ipi_mask,
102 .init_secondary = msmtc_init_secondary,
103 .smp_finish = msmtc_smp_finish,
104 .cpus_done = msmtc_cpus_done,
105 .boot_secondary = msmtc_boot_secondary,
106 .smp_setup = msmtc_smp_setup,
107 .prepare_cpus = msmtc_prepare_cpus,
108};
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200109
110#ifdef CONFIG_MIPS_MT_SMTC_IRQAFF
111/*
112 * IRQ affinity hook
113 */
114
115
Thomas Gleixner7c8d9482011-03-23 21:08:57 +0000116int plat_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity,
117 bool force)
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200118{
Mike Travise65e49d2009-01-12 15:27:13 -0800119 cpumask_t tmask;
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200120 int cpu = 0;
121 void smtc_set_irq_affinity(unsigned int irq, cpumask_t aff);
122
123 /*
124 * On the legacy Malta development board, all I/O interrupts
125 * are routed through the 8259 and combined in a single signal
126 * to the CPU daughterboard, and on the CoreFPGA2/3 34K models,
127 * that signal is brought to IP2 of both VPEs. To avoid racing
128 * concurrent interrupt service events, IP2 is enabled only on
Ralf Baechle70342282013-01-22 12:59:30 +0100129 * one VPE, by convention VPE0. So long as no bits are ever
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200130 * cleared in the affinity mask, there will never be any
131 * interrupt forwarding. But as soon as a program or operator
132 * sets affinity for one of the related IRQs, we need to make
Lucas De Marchi25985ed2011-03-30 22:57:33 -0300133 * sure that we don't ever try to forward across the VPE boundary,
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200134 * at least not until we engineer a system where the interrupt
135 * _ack() or _end() function can somehow know that it corresponds
136 * to an interrupt taken on another VPE, and perform the appropriate
137 * restoration of Status.IM state using MFTR/MTTR instead of the
138 * normal local behavior. We also ensure that no attempt will
139 * be made to forward to an offline "CPU".
140 */
141
Mike Travise65e49d2009-01-12 15:27:13 -0800142 cpumask_copy(&tmask, affinity);
Rusty Russell0de26522008-12-13 21:20:26 +1030143 for_each_cpu(cpu, affinity) {
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200144 if ((cpu_data[cpu].vpe_id != 0) || !cpu_online(cpu))
145 cpu_clear(cpu, tmask);
146 }
Thomas Gleixner7c8d9482011-03-23 21:08:57 +0000147 cpumask_copy(d->affinity, &tmask);
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200148
149 if (cpus_empty(tmask))
150 /*
151 * We could restore a default mask here, but the
152 * runtime code can anyway deal with the null set
153 */
154 printk(KERN_WARNING
Ralf Baechle34ed9502011-05-28 00:57:13 +0100155 "IRQ affinity leaves no legal CPU for IRQ %d\n", d->irq);
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200156
157 /* Do any generic SMTC IRQ affinity setup */
Thomas Gleixner7c8d9482011-03-23 21:08:57 +0000158 smtc_set_irq_affinity(d->irq, tmask);
Yinghai Lud5dedd42009-04-27 17:59:21 -0700159
Thomas Gleixner7c8d9482011-03-23 21:08:57 +0000160 return IRQ_SET_MASK_OK_NOCOPY;
Kevin D. Kissellf571eff2007-08-03 19:38:03 +0200161}
162#endif /* CONFIG_MIPS_MT_SMTC_IRQAFF */