2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 2003, 2004 Ralf Baechle <ralf@linux-mips.org>
7 * Copyright (C) MIPS Technologies, Inc.
8 * written by Ralf Baechle <ralf@linux-mips.org>
10 #ifndef _ASM_HAZARDS_H
11 #define _ASM_HAZARDS_H
15 #define ASMMACRO(name, code...) .macro name; code; .endm
18 #define ASMMACRO(name, code...) \
19 __asm__(".macro " #name "; " #code "; .endm"); \
21 static inline void name(void) \
23 __asm__ __volatile__ (#name); \
39 #if defined(CONFIG_CPU_MIPSR2)
42 * MIPSR2 defines ehb for hazard avoidance
45 ASMMACRO(mtc0_tlbw_hazard,
48 ASMMACRO(tlbw_use_hazard,
51 ASMMACRO(tlb_probe_hazard,
54 ASMMACRO(irq_enable_hazard,
56 ASMMACRO(irq_disable_hazard,
59 ASMMACRO(back_to_back_c0_hazard,
63 * gcc has a tradition of misscompiling the previous construct using the
64 * address of a label as argument to inline assembler. Gas otoh has the
65 * annoying difference between la and dla which are only usable for 32-bit
66 * rsp. 64-bit code, so can't be used without conditional compilation.
67 * The alterantive is switching the assembler to 64-bit code which happens
68 * to work right even for 32-bit code ...
70 #define instruction_hazard() \
74 __asm__ __volatile__( \
83 #elif defined(CONFIG_CPU_R10000)
86 * R10000 rocks - all hazards handled in hardware, so this becomes a nobrainer.
89 ASMMACRO(mtc0_tlbw_hazard,
91 ASMMACRO(tlbw_use_hazard,
93 ASMMACRO(tlb_probe_hazard,
95 ASMMACRO(irq_enable_hazard,
97 ASMMACRO(irq_disable_hazard,
99 ASMMACRO(back_to_back_c0_hazard,
101 #define instruction_hazard() do { } while (0)
103 #elif defined(CONFIG_CPU_RM9000)
106 * RM9000 hazards. When the JTLB is updated by tlbwi or tlbwr, a subsequent
107 * use of the JTLB for instructions should not occur for 4 cpu cycles and use
108 * for data translations should not occur for 3 cpu cycles.
111 ASMMACRO(mtc0_tlbw_hazard,
112 _ssnop; _ssnop; _ssnop; _ssnop
114 ASMMACRO(tlbw_use_hazard,
115 _ssnop; _ssnop; _ssnop; _ssnop
117 ASMMACRO(tlb_probe_hazard,
118 _ssnop; _ssnop; _ssnop; _ssnop
120 ASMMACRO(irq_enable_hazard,
122 ASMMACRO(irq_disable_hazard,
124 ASMMACRO(back_to_back_c0_hazard,
126 #define instruction_hazard() do { } while (0)
128 #elif defined(CONFIG_CPU_SB1)
131 * Mostly like R4000 for historic reasons
133 ASMMACRO(mtc0_tlbw_hazard,
135 ASMMACRO(tlbw_use_hazard,
137 ASMMACRO(tlb_probe_hazard,
139 ASMMACRO(irq_enable_hazard,
141 ASMMACRO(irq_disable_hazard,
142 _ssnop; _ssnop; _ssnop
144 ASMMACRO(back_to_back_c0_hazard,
146 #define instruction_hazard() do { } while (0)
151 * Finally the catchall case for all other processors including R4000, R4400,
152 * R4600, R4700, R5000, RM7000, NEC VR41xx etc.
154 * The taken branch will result in a two cycle penalty for the two killed
155 * instructions on R4000 / R4400. Other processors only have a single cycle
156 * hazard so this is nice trick to have an optimal code for a range of
159 ASMMACRO(mtc0_tlbw_hazard,
162 ASMMACRO(tlbw_use_hazard,
165 ASMMACRO(tlb_probe_hazard,
168 ASMMACRO(irq_enable_hazard,
170 ASMMACRO(irq_disable_hazard,
173 ASMMACRO(back_to_back_c0_hazard,
174 _ssnop; _ssnop; _ssnop;
176 #define instruction_hazard() do { } while (0)
180 #endif /* _ASM_HAZARDS_H */