1 /* $Id: sun4d_irq.c,v 1.29 2001/12/11 04:55:51 davem Exp $
2 * arch/sparc/kernel/sun4d_irq.c:
3 * SS1000/SC2000 interrupt handling.
5 * Copyright (C) 1997,1998 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
6 * Heavily based on arch/sparc/kernel/irq.c.
9 #include <linux/errno.h>
10 #include <linux/linkage.h>
11 #include <linux/kernel_stat.h>
12 #include <linux/signal.h>
13 #include <linux/sched.h>
14 #include <linux/ptrace.h>
15 #include <linux/interrupt.h>
16 #include <linux/slab.h>
17 #include <linux/random.h>
18 #include <linux/init.h>
19 #include <linux/smp.h>
20 #include <linux/smp_lock.h>
21 #include <linux/spinlock.h>
22 #include <linux/seq_file.h>
24 #include <asm/ptrace.h>
25 #include <asm/processor.h>
26 #include <asm/system.h>
29 #include <asm/vaddrs.h>
30 #include <asm/timer.h>
31 #include <asm/openprom.h>
32 #include <asm/oplib.h>
33 #include <asm/traps.h>
36 #include <asm/pgalloc.h>
37 #include <asm/pgtable.h>
40 #include <asm/cacheflush.h>
42 /* If you trust current SCSI layer to handle different SCSI IRQs, enable this. I don't trust it... -jj */
43 /* #define DISTRIBUTE_IRQS */
45 struct sun4d_timer_regs *sun4d_timers;
48 #define MAX_STATIC_ALLOC 4
49 extern struct irqaction static_irqaction[MAX_STATIC_ALLOC];
50 extern int static_irq_count;
51 unsigned char cpu_leds[32];
53 unsigned char sbus_tid[32];
56 static struct irqaction *irq_action[NR_IRQS];
57 extern spinlock_t irq_action_lock;
60 struct irqaction *action;
61 /* For SMP this needs to be extended */
64 static int pil_to_sbus[] = {
65 0, 0, 1, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0, 0,
68 static int sbus_to_pil[] = {
69 0, 2, 3, 5, 7, 9, 11, 13,
74 DEFINE_SPINLOCK(sun4d_imsk_lock);
77 int show_sun4d_interrupts(struct seq_file *p, void *v)
79 int i = *(loff_t *) v, j = 0, k = 0, sbusl;
80 struct irqaction * action;
86 spin_lock_irqsave(&irq_action_lock, flags);
88 sbusl = pil_to_sbus[i];
90 action = *(i + irq_action);
94 for (j = 0; j < nsbi; j++) {
95 for (k = 0; k < 4; k++)
96 if ((action = sbus_actions [(j << 5) + (sbusl << 2) + k].action))
101 found_it: seq_printf(p, "%3d: ", i);
103 seq_printf(p, "%10u ", kstat_irqs(i));
105 for_each_online_cpu(x)
106 seq_printf(p, "%10u ",
107 kstat_cpu(cpu_logical_map(x)).irqs[i]);
109 seq_printf(p, "%c %s",
110 (action->flags & SA_INTERRUPT) ? '+' : ' ',
112 action = action->next;
114 for (; action; action = action->next) {
115 seq_printf(p, ",%s %s",
116 (action->flags & SA_INTERRUPT) ? " +" : "",
122 action = sbus_actions [(j << 5) + (sbusl << 2) + k].action;
125 if (j == nsbi) break;
127 action = sbus_actions [(j << 5) + (sbusl << 2)].action;
133 spin_unlock_irqrestore(&irq_action_lock, flags);
137 void sun4d_free_irq(unsigned int irq, void *dev_id)
139 struct irqaction *action, **actionp;
140 struct irqaction *tmp = NULL;
143 spin_lock_irqsave(&irq_action_lock, flags);
145 actionp = irq + irq_action;
147 actionp = &(sbus_actions[irq - (1 << 5)].action);
150 printk("Trying to free free IRQ%d\n",irq);
154 for (; action; action = action->next) {
155 if (action->dev_id == dev_id)
160 printk("Trying to free free shared IRQ%d\n",irq);
163 } else if (action->flags & SA_SHIRQ) {
164 printk("Trying to free shared IRQ%d with NULL device ID\n", irq);
167 if (action->flags & SA_STATIC_ALLOC)
169 /* This interrupt is marked as specially allocated
170 * so it is a bad idea to free it.
172 printk("Attempt to free statically allocated IRQ%d (%s)\n",
178 tmp->next = action->next;
180 *actionp = action->next;
182 spin_unlock_irqrestore(&irq_action_lock, flags);
184 synchronize_irq(irq);
186 spin_lock_irqsave(&irq_action_lock, flags);
194 spin_unlock_irqrestore(&irq_action_lock, flags);
197 extern void unexpected_irq(int, void *, struct pt_regs *);
199 void sun4d_handler_irq(int irq, struct pt_regs * regs)
201 struct irqaction * action;
202 int cpu = smp_processor_id();
203 /* SBUS IRQ level (1 - 7) */
204 int sbusl = pil_to_sbus[irq];
206 /* FIXME: Is this necessary?? */
209 cc_set_iclr(1 << irq);
212 kstat_cpu(cpu).irqs[irq]++;
214 action = *(irq + irq_action);
216 unexpected_irq(irq, NULL, regs);
218 action->handler(irq, action->dev_id, regs);
219 action = action->next;
222 int bus_mask = bw_get_intr_mask(sbusl) & 0x3ffff;
224 struct sbus_action *actionp;
226 int sbil = (sbusl << 2);
228 bw_clear_intr_mask(sbusl, bus_mask);
230 /* Loop for each pending SBI */
231 for (sbino = 0; bus_mask; sbino++, bus_mask >>= 1)
233 mask = acquire_sbi(SBI2DEVID(sbino), 0xf << sbil);
234 mask &= (0xf << sbil);
235 actionp = sbus_actions + (sbino << 5) + (sbil);
236 /* Loop for each pending SBI slot */
237 for (slot = (1 << sbil); mask; slot <<= 1, actionp++)
240 action = actionp->action;
243 unexpected_irq(irq, NULL, regs);
245 action->handler(irq, action->dev_id, regs);
246 action = action->next;
248 release_sbi(SBI2DEVID(sbino), slot);
255 unsigned int sun4d_build_irq(struct sbus_dev *sdev, int irq)
257 int sbusl = pil_to_sbus[irq];
260 return ((sdev->bus->board + 1) << 5) + (sbusl << 2) + sdev->slot;
265 unsigned int sun4d_sbint_to_irq(struct sbus_dev *sdev, unsigned int sbint)
267 if (sbint >= sizeof(sbus_to_pil)) {
268 printk(KERN_ERR "%s: bogus SBINT %d\n", sdev->prom_name, sbint);
271 return sun4d_build_irq(sdev, sbus_to_pil[sbint]);
274 int sun4d_request_irq(unsigned int irq,
275 irqreturn_t (*handler)(int, void *, struct pt_regs *),
276 unsigned long irqflags, const char * devname, void *dev_id)
278 struct irqaction *action, *tmp = NULL, **actionp;
282 if(irq > 14 && irq < (1 << 5)) {
292 spin_lock_irqsave(&irq_action_lock, flags);
295 actionp = &(sbus_actions[irq - (1 << 5)].action);
297 actionp = irq + irq_action;
301 if ((action->flags & SA_SHIRQ) && (irqflags & SA_SHIRQ)) {
302 for (tmp = action; tmp->next; tmp = tmp->next);
307 if ((action->flags & SA_INTERRUPT) ^ (irqflags & SA_INTERRUPT)) {
308 printk("Attempt to mix fast and slow interrupts on IRQ%d denied\n", irq);
312 action = NULL; /* Or else! */
315 /* If this is flagged as statically allocated then we use our
316 * private struct which is never freed.
318 if (irqflags & SA_STATIC_ALLOC) {
319 if (static_irq_count < MAX_STATIC_ALLOC)
320 action = &static_irqaction[static_irq_count++];
322 printk("Request for IRQ%d (%s) SA_STATIC_ALLOC failed using kmalloc\n", irq, devname);
326 action = (struct irqaction *)kmalloc(sizeof(struct irqaction),
334 action->handler = handler;
335 action->flags = irqflags;
336 cpus_clear(action->mask);
337 action->name = devname;
339 action->dev_id = dev_id;
350 spin_unlock_irqrestore(&irq_action_lock, flags);
355 static void sun4d_disable_irq(unsigned int irq)
358 int tid = sbus_tid[(irq >> 5) - 1];
362 if (irq < NR_IRQS) return;
364 spin_lock_irqsave(&sun4d_imsk_lock, flags);
365 cc_set_imsk_other(tid, cc_get_imsk_other(tid) | (1 << sbus_to_pil[(irq >> 2) & 7]));
366 spin_unlock_irqrestore(&sun4d_imsk_lock, flags);
368 cc_set_imsk(cc_get_imsk() | (1 << sbus_to_pil[(irq >> 2) & 7]));
372 static void sun4d_enable_irq(unsigned int irq)
375 int tid = sbus_tid[(irq >> 5) - 1];
379 if (irq < NR_IRQS) return;
381 spin_lock_irqsave(&sun4d_imsk_lock, flags);
382 cc_set_imsk_other(tid, cc_get_imsk_other(tid) & ~(1 << sbus_to_pil[(irq >> 2) & 7]));
383 spin_unlock_irqrestore(&sun4d_imsk_lock, flags);
385 cc_set_imsk(cc_get_imsk() & ~(1 << sbus_to_pil[(irq >> 2) & 7]));
390 static void sun4d_set_cpu_int(int cpu, int level)
392 sun4d_send_ipi(cpu, level);
395 static void sun4d_clear_ipi(int cpu, int level)
399 static void sun4d_set_udt(int cpu)
403 /* Setup IRQ distribution scheme. */
404 void __init sun4d_distribute_irqs(void)
406 #ifdef DISTRIBUTE_IRQS
407 struct sbus_bus *sbus;
408 unsigned long sbus_serving_map;
410 sbus_serving_map = cpu_present_map;
411 for_each_sbus(sbus) {
412 if ((sbus->board * 2) == boot_cpu_id && (cpu_present_map & (1 << (sbus->board * 2 + 1))))
413 sbus_tid[sbus->board] = (sbus->board * 2 + 1);
414 else if (cpu_present_map & (1 << (sbus->board * 2)))
415 sbus_tid[sbus->board] = (sbus->board * 2);
416 else if (cpu_present_map & (1 << (sbus->board * 2 + 1)))
417 sbus_tid[sbus->board] = (sbus->board * 2 + 1);
419 sbus_tid[sbus->board] = 0xff;
420 if (sbus_tid[sbus->board] != 0xff)
421 sbus_serving_map &= ~(1 << sbus_tid[sbus->board]);
424 if (sbus_tid[sbus->board] == 0xff) {
427 if (!sbus_serving_map)
428 sbus_serving_map = cpu_present_map;
429 while (!(sbus_serving_map & (1 << i)))
431 sbus_tid[sbus->board] = i;
432 sbus_serving_map &= ~(1 << i);
434 for_each_sbus(sbus) {
435 printk("sbus%d IRQs directed to CPU%d\n", sbus->board, sbus_tid[sbus->board]);
436 set_sbi_tid(sbus->devid, sbus_tid[sbus->board] << 3);
439 struct sbus_bus *sbus;
440 int cpuid = cpu_logical_map(1);
443 cpuid = cpu_logical_map(0);
444 for_each_sbus(sbus) {
445 sbus_tid[sbus->board] = cpuid;
446 set_sbi_tid(sbus->devid, cpuid << 3);
448 printk("All sbus IRQs directed to CPU%d\n", cpuid);
453 static void sun4d_clear_clock_irq(void)
455 volatile unsigned int clear_intr;
456 clear_intr = sun4d_timers->l10_timer_limit;
459 static void sun4d_clear_profile_irq(int cpu)
461 bw_get_prof_limit(cpu);
464 static void sun4d_load_profile_irq(int cpu, unsigned int limit)
466 bw_set_prof_limit(cpu, limit);
469 static void __init sun4d_init_timers(irqreturn_t (*counter_fn)(int, void *, struct pt_regs *))
476 /* Map the User Timer registers. */
477 memset(&r, 0, sizeof(r));
479 r.start = CSR_BASE(boot_cpu_id)+BW_TIMER_LIMIT;
481 r.start = CSR_BASE(0)+BW_TIMER_LIMIT;
484 sun4d_timers = (struct sun4d_timer_regs *) sbus_ioremap(&r, 0,
485 PAGE_SIZE, "user timer");
487 sun4d_timers->l10_timer_limit = (((1000000/HZ) + 1) << 10);
488 master_l10_counter = &sun4d_timers->l10_cur_count;
489 master_l10_limit = &sun4d_timers->l10_timer_limit;
491 irq = request_irq(TIMER_IRQ,
493 (SA_INTERRUPT | SA_STATIC_ALLOC),
496 prom_printf("time_init: unable to attach IRQ%d\n",TIMER_IRQ);
500 /* Enable user timer free run for CPU 0 in BW */
501 /* bw_set_ctrl(0, bw_get_ctrl(0) | BW_CTRL_USER_TIMER); */
504 while (!cpu_find_by_instance(cpu, NULL, &mid)) {
505 sun4d_load_profile_irq(mid >> 3, 0);
512 extern unsigned long lvl14_save[4];
513 struct tt_entry *trap_table = &sparc_ttable[SP_TRAP_IRQ1 + (14 - 1)];
514 extern unsigned int real_irq_entry[], smp4d_ticker[];
515 extern unsigned int patchme_maybe_smp_msg[];
517 /* Adjust so that we jump directly to smp4d_ticker */
518 lvl14_save[2] += smp4d_ticker - real_irq_entry;
520 /* For SMP we use the level 14 ticker, however the bootup code
521 * has copied the firmwares level 14 vector into boot cpu's
522 * trap table, we must fix this now or we get squashed.
524 local_irq_save(flags);
525 patchme_maybe_smp_msg[0] = 0x01000000; /* NOP out the branch */
526 trap_table->inst_one = lvl14_save[0];
527 trap_table->inst_two = lvl14_save[1];
528 trap_table->inst_three = lvl14_save[2];
529 trap_table->inst_four = lvl14_save[3];
530 local_flush_cache_all();
531 local_irq_restore(flags);
536 void __init sun4d_init_sbi_irq(void)
538 struct sbus_bus *sbus;
544 sbus_actions = (struct sbus_action *)kmalloc (nsbi * 8 * 4 * sizeof(struct sbus_action), GFP_ATOMIC);
545 memset (sbus_actions, 0, (nsbi * 8 * 4 * sizeof(struct sbus_action)));
546 for_each_sbus(sbus) {
548 extern unsigned char boot_cpu_id;
550 set_sbi_tid(sbus->devid, boot_cpu_id << 3);
551 sbus_tid[sbus->board] = boot_cpu_id;
553 /* Get rid of pending irqs from PROM */
554 mask = acquire_sbi(sbus->devid, 0xffffffff);
556 printk ("Clearing pending IRQs %08x on SBI %d\n", mask, sbus->board);
557 release_sbi(sbus->devid, mask);
562 void __init sun4d_init_IRQ(void)
566 BTFIXUPSET_CALL(sbint_to_irq, sun4d_sbint_to_irq, BTFIXUPCALL_NORM);
567 BTFIXUPSET_CALL(enable_irq, sun4d_enable_irq, BTFIXUPCALL_NORM);
568 BTFIXUPSET_CALL(disable_irq, sun4d_disable_irq, BTFIXUPCALL_NORM);
569 BTFIXUPSET_CALL(clear_clock_irq, sun4d_clear_clock_irq, BTFIXUPCALL_NORM);
570 BTFIXUPSET_CALL(clear_profile_irq, sun4d_clear_profile_irq, BTFIXUPCALL_NORM);
571 BTFIXUPSET_CALL(load_profile_irq, sun4d_load_profile_irq, BTFIXUPCALL_NORM);
572 sparc_init_timers = sun4d_init_timers;
574 BTFIXUPSET_CALL(set_cpu_int, sun4d_set_cpu_int, BTFIXUPCALL_NORM);
575 BTFIXUPSET_CALL(clear_cpu_int, sun4d_clear_ipi, BTFIXUPCALL_NOP);
576 BTFIXUPSET_CALL(set_irq_udt, sun4d_set_udt, BTFIXUPCALL_NOP);
578 /* Cannot enable interrupts until OBP ticker is disabled. */