2 * Malta Platform-specific hooks for SMP operation
5 #include <linux/init.h>
7 #include <asm/mipsregs.h>
8 #include <asm/mipsmtregs.h>
10 #include <asm/smtc_ipi.h>
12 /* VPE/SMP Prototype implements platform interfaces directly */
15 * Cause the specified action to be performed on a targeted "CPU"
18 static void msmtc_send_ipi_single(int cpu, unsigned int action)
20 /* "CPU" may be TC of same VPE, VPE of same CPU, or different CPU */
21 smtc_send_ipi(cpu, LINUX_SMP_IPI, action);
24 static void msmtc_send_ipi_mask(cpumask_t mask, unsigned int action)
28 for_each_cpu_mask(i, mask)
29 msmtc_send_ipi_single(i, action);
33 * Post-config but pre-boot cleanup entry point
35 static void __cpuinit msmtc_init_secondary(void)
37 void smtc_init_secondary(void);
40 /* Don't enable Malta I/O interrupts (IP2) for secondary VPEs */
41 myvpe = read_c0_tcbind() & TCBIND_CURVPE;
43 /* Ideally, this should be done only once per VPE, but... */
44 clear_c0_status(ST0_IM);
45 set_c0_status((0x100 << cp0_compare_irq)
46 | (0x100 << MIPS_CPU_IPI_IRQ));
47 if (cp0_perfcount_irq >= 0)
48 set_c0_status(0x100 << cp0_perfcount_irq);
51 smtc_init_secondary();
55 * Platform "CPU" startup hook
57 static void __cpuinit msmtc_boot_secondary(int cpu, struct task_struct *idle)
59 smtc_boot_secondary(cpu, idle);
63 * SMP initialization finalization entry point
65 static void __cpuinit msmtc_smp_finish(void)
71 * Hook for after all CPUs are online
74 static void msmtc_cpus_done(void)
79 * Platform SMP pre-initialization
81 * As noted above, we can assume a single CPU for now
82 * but it may be multithreaded.
85 static void __init msmtc_smp_setup(void)
87 mipsmt_build_cpu_map(0);
90 static void __init msmtc_prepare_cpus(unsigned int max_cpus)
92 mipsmt_prepare_cpus();
95 struct plat_smp_ops msmtc_smp_ops = {
96 .send_ipi_single = msmtc_send_ipi_single,
97 .send_ipi_mask = msmtc_send_ipi_mask,
98 .init_secondary = msmtc_init_secondary,
99 .smp_finish = msmtc_smp_finish,
100 .cpus_done = msmtc_cpus_done,
101 .boot_secondary = msmtc_boot_secondary,
102 .smp_setup = msmtc_smp_setup,
103 .prepare_cpus = msmtc_prepare_cpus,
106 #ifdef CONFIG_MIPS_MT_SMTC_IRQAFF
112 void plat_set_irq_affinity(unsigned int irq, cpumask_t affinity)
114 cpumask_t tmask = affinity;
116 void smtc_set_irq_affinity(unsigned int irq, cpumask_t aff);
119 * On the legacy Malta development board, all I/O interrupts
120 * are routed through the 8259 and combined in a single signal
121 * to the CPU daughterboard, and on the CoreFPGA2/3 34K models,
122 * that signal is brought to IP2 of both VPEs. To avoid racing
123 * concurrent interrupt service events, IP2 is enabled only on
124 * one VPE, by convention VPE0. So long as no bits are ever
125 * cleared in the affinity mask, there will never be any
126 * interrupt forwarding. But as soon as a program or operator
127 * sets affinity for one of the related IRQs, we need to make
128 * sure that we don't ever try to forward across the VPE boundry,
129 * at least not until we engineer a system where the interrupt
130 * _ack() or _end() function can somehow know that it corresponds
131 * to an interrupt taken on another VPE, and perform the appropriate
132 * restoration of Status.IM state using MFTR/MTTR instead of the
133 * normal local behavior. We also ensure that no attempt will
134 * be made to forward to an offline "CPU".
137 for_each_cpu_mask(cpu, affinity) {
138 if ((cpu_data[cpu].vpe_id != 0) || !cpu_online(cpu))
139 cpu_clear(cpu, tmask);
141 irq_desc[irq].affinity = tmask;
143 if (cpus_empty(tmask))
145 * We could restore a default mask here, but the
146 * runtime code can anyway deal with the null set
149 "IRQ affinity leaves no legal CPU for IRQ %d\n", irq);
151 /* Do any generic SMTC IRQ affinity setup */
152 smtc_set_irq_affinity(irq, tmask);
154 #endif /* CONFIG_MIPS_MT_SMTC_IRQAFF */