2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 1994 - 2000, 2001, 2003 Ralf Baechle
7 * Copyright (C) 1999, 2000 Silicon Graphics, Inc.
8 * Copyright (C) 2001 MIPS Technologies, Inc.
9 * Copyright (C) 2002 Maciej W. Rozycki
11 #include <linux/init.h>
14 #include <asm/asmmacro.h>
15 #include <asm/cacheops.h>
16 #include <asm/irqflags.h>
17 #include <asm/regdef.h>
18 #include <asm/fpregdef.h>
19 #include <asm/mipsregs.h>
20 #include <asm/stackframe.h>
24 #define PANIC_PIC(msg) \
37 NESTED(except_vec0_generic, 0, sp)
38 PANIC_PIC("Exception vector 0 called")
39 END(except_vec0_generic)
41 NESTED(except_vec1_generic, 0, sp)
42 PANIC_PIC("Exception vector 1 called")
43 END(except_vec1_generic)
46 * General exception vector for all other CPUs.
48 * Be careful when changing this, it has to be at most 128 bytes
49 * to fit into space reserved for the exception handler.
51 NESTED(except_vec3_generic, 0, sp)
54 #if R5432_CP0_INTERRUPT_WAR
62 PTR_L k0, exception_handlers(k1)
65 END(except_vec3_generic)
68 * General exception handler for CPUs with virtual coherency exception.
70 * Be careful when changing this, it has to be at most 256 (as a special
71 * exception) bytes to fit into space reserved for the exception handler.
73 NESTED(except_vec3_r4000, 0, sp)
83 beq k1, k0, handle_vced
85 beq k1, k0, handle_vcei
90 PTR_L k0, exception_handlers(k1)
94 * Big shit, we now may have two dirty primary cache lines for the same
95 * physical address. We can safely invalidate the line pointed to by
96 * c0_badvaddr because after return from this exception handler the
97 * load / store will be re-executed.
100 MFC0 k0, CP0_BADVADDR
101 li k1, -4 # Is this ...
102 and k0, k1 # ... really needed?
104 cache Index_Store_Tag_D, (k0)
105 cache Hit_Writeback_Inv_SD, (k0)
106 #ifdef CONFIG_PROC_FS
107 PTR_LA k0, vced_count
115 MFC0 k0, CP0_BADVADDR
116 cache Hit_Writeback_Inv_SD, (k0) # also cleans pi
117 #ifdef CONFIG_PROC_FS
118 PTR_LA k0, vcei_count
125 END(except_vec3_r4000)
130 NESTED(handle_int, PT_SIZE, sp)
135 LONG_L s0, TI_REGS($28)
136 LONG_S sp, TI_REGS($28)
137 PTR_LA ra, ret_from_irq
144 * Special interrupt vector for MIPS64 ISA & embedded MIPS processors.
145 * This is a dedicated interrupt exception vector which reduces the
146 * interrupt processing overhead. The jump instruction will be replaced
147 * at the initialization time.
149 * Be careful when changing this, it has to be at most 128 bytes
150 * to fit into space reserved for the exception handler.
152 NESTED(except_vec4, 0, sp)
153 1: j 1b /* Dummy, will be replaced */
157 * EJTAG debug exception handler.
158 * The EJTAG debug exception entry point is 0xbfc00480, which
159 * normally is in the boot PROM, so the boot PROM must do a
160 * unconditional jump to this vector.
162 NESTED(except_vec_ejtag_debug, 0, sp)
163 j ejtag_debug_handler
164 END(except_vec_ejtag_debug)
169 * Vectored interrupt handler.
170 * This prototype is copied to ebase + n*IntCtl.VS and patched
171 * to invoke the handler
173 NESTED(except_vec_vi, 0, sp)
178 #ifdef CONFIG_MIPS_MT_SMTC
180 * To keep from blindly blocking *all* interrupts
181 * during service by SMTC kernel, we also want to
182 * pass the IM value to be cleared.
184 FEXPORT(except_vec_vi_mori)
186 #endif /* CONFIG_MIPS_MT_SMTC */
187 FEXPORT(except_vec_vi_lui)
188 lui v0, 0 /* Patched */
189 j except_vec_vi_handler
190 FEXPORT(except_vec_vi_ori)
191 ori v0, 0 /* Patched */
194 EXPORT(except_vec_vi_end)
197 * Common Vectored Interrupt code
198 * Complete the register saves and invoke the handler which is passed in $v0
200 NESTED(except_vec_vi_handler, 0, sp)
203 #ifdef CONFIG_MIPS_MT_SMTC
205 * SMTC has an interesting problem that interrupts are level-triggered,
206 * and the CLI macro will clear EXL, potentially causing a duplicate
207 * interrupt service invocation. So we need to clear the associated
208 * IM bit of Status prior to doing CLI, and restore it after the
209 * service routine has been invoked - we must assume that the
210 * service routine will have cleared the state, and any active
211 * level represents a new or otherwised unserviced event...
215 mfc0 t2, CP0_TCCONTEXT
217 mtc0 t0, CP0_TCCONTEXT
221 #endif /* CONFIG_MIPS_MT_SMTC */
223 #ifdef CONFIG_TRACE_IRQFLAGS
225 #ifdef CONFIG_MIPS_MT_SMTC
229 #ifdef CONFIG_MIPS_MT_SMTC
235 LONG_L s0, TI_REGS($28)
236 LONG_S sp, TI_REGS($28)
237 PTR_LA ra, ret_from_irq
239 END(except_vec_vi_handler)
242 * EJTAG debug exception handler.
244 NESTED(ejtag_debug_handler, PT_SIZE, sp)
250 sll k0, k0, 30 # Check for SDBBP.
251 bgez k0, ejtag_return
253 PTR_LA k0, ejtag_debug_buffer
257 jal ejtag_exception_handler
259 PTR_LA k0, ejtag_debug_buffer
267 END(ejtag_debug_handler)
270 * This buffer is reserved for the use of the EJTAG debug
274 EXPORT(ejtag_debug_buffer)
281 * NMI debug exception handler for MIPS reference boards.
282 * The NMI debug exception entry point is 0xbfc00000, which
283 * normally is in the boot PROM, so the boot PROM must do a
284 * unconditional jump to this vector.
286 NESTED(except_vec_nmi, 0, sp)
292 NESTED(nmi_handler, PT_SIZE, sp)
297 jal nmi_exception_handler
304 .macro __build_clear_none
307 .macro __build_clear_sti
312 .macro __build_clear_cli
317 .macro __build_clear_fpe
326 .macro __build_clear_ade
327 MFC0 t0, CP0_BADVADDR
328 PTR_S t0, PT_BVADDR(sp)
332 .macro __BUILD_silent exception
335 /* Gas tries to parse the PRINT argument as a string containing
336 string escapes and emits bogus warnings if it believes to
337 recognize an unknown escape code. So make the arguments
338 start with an n and gas will believe \n is ok ... */
339 .macro __BUILD_verbose nexception
340 LONG_L a1, PT_EPC(sp)
342 PRINT("Got \nexception at %08lx\012")
345 PRINT("Got \nexception at %016lx\012")
349 .macro __BUILD_count exception
350 LONG_L t0,exception_count_\exception
352 LONG_S t0,exception_count_\exception
353 .comm exception_count\exception, 8, 8
356 .macro __BUILD_HANDLER exception handler clear verbose ext
358 NESTED(handle_\exception, PT_SIZE, sp)
361 FEXPORT(handle_\exception\ext)
364 __BUILD_\verbose \exception
366 PTR_LA ra, ret_from_exception
368 END(handle_\exception)
371 .macro BUILD_HANDLER exception handler clear verbose
372 __BUILD_HANDLER \exception \handler \clear \verbose _int
375 BUILD_HANDLER adel ade ade silent /* #4 */
376 BUILD_HANDLER ades ade ade silent /* #5 */
377 BUILD_HANDLER ibe be cli silent /* #6 */
378 BUILD_HANDLER dbe be cli silent /* #7 */
379 BUILD_HANDLER bp bp sti silent /* #9 */
380 BUILD_HANDLER ri ri sti silent /* #10 */
381 BUILD_HANDLER cpu cpu sti silent /* #11 */
382 BUILD_HANDLER ov ov sti silent /* #12 */
383 BUILD_HANDLER tr tr sti silent /* #13 */
384 BUILD_HANDLER fpe fpe fpe silent /* #15 */
385 BUILD_HANDLER mdmx mdmx sti silent /* #22 */
386 BUILD_HANDLER watch watch sti verbose /* #23 */
387 BUILD_HANDLER mcheck mcheck cli verbose /* #24 */
388 BUILD_HANDLER mt mt sti silent /* #25 */
389 BUILD_HANDLER dsp dsp sti silent /* #26 */
390 BUILD_HANDLER reserved reserved sti verbose /* others */
393 LEAF(handle_ri_rdhwr_vivt)
394 #ifdef CONFIG_MIPS_MT_SMTC
395 PANIC_PIC("handle_ri_rdhwr_vivt called")
400 /* check if TLB contains a entry for EPC */
402 andi k1, 0xff /* ASID_MASK */
404 PTR_SRL k0, PAGE_SHIFT + 1
405 PTR_SLL k0, PAGE_SHIFT + 1
413 bltz k1, handle_ri /* slow path */
416 END(handle_ri_rdhwr_vivt)
418 LEAF(handle_ri_rdhwr)
422 /* 0x7c03e83b: rdhwr v1,$29 */
428 bne k0, k1, handle_ri /* if not ours */
429 /* The insn is rdhwr. No need to check CAUSE.BD here. */
430 get_saved_sp /* k1 := current_thread_info */
433 #if defined(CONFIG_CPU_R3000) || defined(CONFIG_CPU_TX39XX)
435 xori k1, _THREAD_MASK
436 LONG_L v1, TI_TP_VALUE(k1)
441 LONG_ADDIU k0, 4 /* stall on $k0 */
443 /* I hope three instructions between MTC0 and ERET are enough... */
445 xori k1, _THREAD_MASK
446 LONG_L v1, TI_TP_VALUE(k1)
455 /* A temporary overflow handler used by check_daddi(). */
459 BUILD_HANDLER daddi_ov daddi_ov none silent /* #12 */