1 #ifndef _ARCH_SPARC64_LOCAL_H
2 #define _ARCH_SPARC64_LOCAL_H
4 #include <linux/percpu.h>
5 #include <asm/atomic.h>
7 typedef atomic64_t local_t;
9 #define LOCAL_INIT(i) ATOMIC64_INIT(i)
10 #define local_read(v) atomic64_read(v)
11 #define local_set(v,i) atomic64_set(v,i)
13 #define local_inc(v) atomic64_inc(v)
14 #define local_dec(v) atomic64_dec(v)
15 #define local_add(i, v) atomic64_add(i, v)
16 #define local_sub(i, v) atomic64_sub(i, v)
18 #define __local_inc(v) ((v)->counter++)
19 #define __local_dec(v) ((v)->counter--)
20 #define __local_add(i,v) ((v)->counter+=(i))
21 #define __local_sub(i,v) ((v)->counter-=(i))
23 /* Use these for per-cpu local_t variables: on some archs they are
24 * much more efficient than these naive implementations. Note they take
25 * a variable, not an address.
27 #define cpu_local_read(v) local_read(&__get_cpu_var(v))
28 #define cpu_local_set(v, i) local_set(&__get_cpu_var(v), (i))
30 #define cpu_local_inc(v) local_inc(&__get_cpu_var(v))
31 #define cpu_local_dec(v) local_dec(&__get_cpu_var(v))
32 #define cpu_local_add(i, v) local_add((i), &__get_cpu_var(v))
33 #define cpu_local_sub(i, v) local_sub((i), &__get_cpu_var(v))
35 #define __cpu_local_inc(v) __local_inc(&__get_cpu_var(v))
36 #define __cpu_local_dec(v) __local_dec(&__get_cpu_var(v))
37 #define __cpu_local_add(i, v) __local_add((i), &__get_cpu_var(v))
38 #define __cpu_local_sub(i, v) __local_sub((i), &__get_cpu_var(v))
40 #endif /* _ARCH_SPARC64_LOCAL_H */