1 #ifndef __ARCH_S390_PERCPU__
2 #define __ARCH_S390_PERCPU__
4 #include <linux/compiler.h>
5 #include <asm/lowcore.h>
7 #define __GENERIC_PER_CPU
10 * s390 uses its own implementation for per cpu data, the offset of
11 * the cpu local data area is cached in the cpu's lowcore memory.
12 * For 64 bit module code s390 forces the use of a GOT slot for the
13 * address of the per cpu variable. This is needed because the module
14 * may be more than 4G above the per cpu area.
16 #if defined(__s390x__) && defined(MODULE)
18 #define __reloc_hide(var,offset) \
19 (*({ unsigned long *__ptr; \
20 asm ( "larl %0,per_cpu__"#var"@GOTENT" \
21 : "=a" (__ptr) : "X" (per_cpu__##var) ); \
22 (typeof(&per_cpu__##var))((*__ptr) + (offset)); }))
26 #define __reloc_hide(var, offset) \
27 (*({ unsigned long __ptr; \
28 asm ( "" : "=a" (__ptr) : "0" (&per_cpu__##var) ); \
29 (typeof(&per_cpu__##var)) (__ptr + (offset)); }))
35 extern unsigned long __per_cpu_offset[NR_CPUS];
37 /* Separate out the type, so (int[3], foo) works. */
38 #define DEFINE_PER_CPU(type, name) \
39 __attribute__((__section__(".data.percpu"))) \
40 __typeof__(type) per_cpu__##name
42 #define __get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset)
43 #define per_cpu(var,cpu) __reloc_hide(var,__per_cpu_offset[cpu])
45 /* A macro to avoid #include hell... */
46 #define percpu_modcopy(pcpudst, src, size) \
49 for_each_possible_cpu(__i) \
50 memcpy((pcpudst)+__per_cpu_offset[__i], \
56 #define DEFINE_PER_CPU(type, name) \
57 __typeof__(type) per_cpu__##name
59 #define __get_cpu_var(var) __reloc_hide(var,0)
60 #define per_cpu(var,cpu) __reloc_hide(var,0)
64 #define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name
66 #define EXPORT_PER_CPU_SYMBOL(var) EXPORT_SYMBOL(per_cpu__##var)
67 #define EXPORT_PER_CPU_SYMBOL_GPL(var) EXPORT_SYMBOL_GPL(per_cpu__##var)
69 #endif /* __ARCH_S390_PERCPU__ */