1 #ifndef __ARCH_X86_64_ATOMIC__
2 #define __ARCH_X86_64_ATOMIC__
4 #include <asm/alternative.h>
5 #include <asm/cmpxchg.h>
7 /* atomic_t should be 32 bit signed type */
10 * Atomic operations that C can't guarantee us. Useful for
11 * resource counting etc..
15 #define LOCK "lock ; "
21 * Make sure gcc doesn't try to be clever and move things around
22 * on us. We need to use _exactly_ the address the user gave us,
23 * not some alias that contains the same information.
25 typedef struct { int counter; } atomic_t;
27 #define ATOMIC_INIT(i) { (i) }
30 * atomic_read - read atomic variable
31 * @v: pointer of type atomic_t
33 * Atomically reads the value of @v.
35 #define atomic_read(v) ((v)->counter)
38 * atomic_set - set atomic variable
39 * @v: pointer of type atomic_t
42 * Atomically sets the value of @v to @i.
44 #define atomic_set(v,i) (((v)->counter) = (i))
47 * atomic_add - add integer to atomic variable
48 * @i: integer value to add
49 * @v: pointer of type atomic_t
51 * Atomically adds @i to @v.
53 static __inline__ void atomic_add(int i, atomic_t *v)
56 LOCK_PREFIX "addl %1,%0"
58 :"ir" (i), "m" (v->counter));
62 * atomic_sub - subtract the atomic variable
63 * @i: integer value to subtract
64 * @v: pointer of type atomic_t
66 * Atomically subtracts @i from @v.
68 static __inline__ void atomic_sub(int i, atomic_t *v)
71 LOCK_PREFIX "subl %1,%0"
73 :"ir" (i), "m" (v->counter));
77 * atomic_sub_and_test - subtract value from variable and test result
78 * @i: integer value to subtract
79 * @v: pointer of type atomic_t
81 * Atomically subtracts @i from @v and returns
82 * true if the result is zero, or false for all
85 static __inline__ int atomic_sub_and_test(int i, atomic_t *v)
90 LOCK_PREFIX "subl %2,%0; sete %1"
91 :"=m" (v->counter), "=qm" (c)
92 :"ir" (i), "m" (v->counter) : "memory");
97 * atomic_inc - increment atomic variable
98 * @v: pointer of type atomic_t
100 * Atomically increments @v by 1.
102 static __inline__ void atomic_inc(atomic_t *v)
104 __asm__ __volatile__(
105 LOCK_PREFIX "incl %0"
111 * atomic_dec - decrement atomic variable
112 * @v: pointer of type atomic_t
114 * Atomically decrements @v by 1.
116 static __inline__ void atomic_dec(atomic_t *v)
118 __asm__ __volatile__(
119 LOCK_PREFIX "decl %0"
125 * atomic_dec_and_test - decrement and test
126 * @v: pointer of type atomic_t
128 * Atomically decrements @v by 1 and
129 * returns true if the result is 0, or false for all other
132 static __inline__ int atomic_dec_and_test(atomic_t *v)
136 __asm__ __volatile__(
137 LOCK_PREFIX "decl %0; sete %1"
138 :"=m" (v->counter), "=qm" (c)
139 :"m" (v->counter) : "memory");
144 * atomic_inc_and_test - increment and test
145 * @v: pointer of type atomic_t
147 * Atomically increments @v by 1
148 * and returns true if the result is zero, or false for all
151 static __inline__ int atomic_inc_and_test(atomic_t *v)
155 __asm__ __volatile__(
156 LOCK_PREFIX "incl %0; sete %1"
157 :"=m" (v->counter), "=qm" (c)
158 :"m" (v->counter) : "memory");
163 * atomic_add_negative - add and test if negative
164 * @i: integer value to add
165 * @v: pointer of type atomic_t
167 * Atomically adds @i to @v and returns true
168 * if the result is negative, or false when
169 * result is greater than or equal to zero.
171 static __inline__ int atomic_add_negative(int i, atomic_t *v)
175 __asm__ __volatile__(
176 LOCK_PREFIX "addl %2,%0; sets %1"
177 :"=m" (v->counter), "=qm" (c)
178 :"ir" (i), "m" (v->counter) : "memory");
183 * atomic_add_return - add and return
184 * @i: integer value to add
185 * @v: pointer of type atomic_t
187 * Atomically adds @i to @v and returns @i + @v
189 static __inline__ int atomic_add_return(int i, atomic_t *v)
192 __asm__ __volatile__(
193 LOCK_PREFIX "xaddl %0, %1"
194 :"+r" (i), "+m" (v->counter)
199 static __inline__ int atomic_sub_return(int i, atomic_t *v)
201 return atomic_add_return(-i,v);
204 #define atomic_inc_return(v) (atomic_add_return(1,v))
205 #define atomic_dec_return(v) (atomic_sub_return(1,v))
207 /* An 64bit atomic type */
209 typedef struct { volatile long counter; } atomic64_t;
211 #define ATOMIC64_INIT(i) { (i) }
214 * atomic64_read - read atomic64 variable
215 * @v: pointer of type atomic64_t
217 * Atomically reads the value of @v.
218 * Doesn't imply a read memory barrier.
220 #define atomic64_read(v) ((v)->counter)
223 * atomic64_set - set atomic64 variable
224 * @v: pointer to type atomic64_t
227 * Atomically sets the value of @v to @i.
229 #define atomic64_set(v,i) (((v)->counter) = (i))
232 * atomic64_add - add integer to atomic64 variable
233 * @i: integer value to add
234 * @v: pointer to type atomic64_t
236 * Atomically adds @i to @v.
238 static __inline__ void atomic64_add(long i, atomic64_t *v)
240 __asm__ __volatile__(
241 LOCK_PREFIX "addq %1,%0"
243 :"ir" (i), "m" (v->counter));
247 * atomic64_sub - subtract the atomic64 variable
248 * @i: integer value to subtract
249 * @v: pointer to type atomic64_t
251 * Atomically subtracts @i from @v.
253 static __inline__ void atomic64_sub(long i, atomic64_t *v)
255 __asm__ __volatile__(
256 LOCK_PREFIX "subq %1,%0"
258 :"ir" (i), "m" (v->counter));
262 * atomic64_sub_and_test - subtract value from variable and test result
263 * @i: integer value to subtract
264 * @v: pointer to type atomic64_t
266 * Atomically subtracts @i from @v and returns
267 * true if the result is zero, or false for all
270 static __inline__ int atomic64_sub_and_test(long i, atomic64_t *v)
274 __asm__ __volatile__(
275 LOCK_PREFIX "subq %2,%0; sete %1"
276 :"=m" (v->counter), "=qm" (c)
277 :"ir" (i), "m" (v->counter) : "memory");
282 * atomic64_inc - increment atomic64 variable
283 * @v: pointer to type atomic64_t
285 * Atomically increments @v by 1.
287 static __inline__ void atomic64_inc(atomic64_t *v)
289 __asm__ __volatile__(
290 LOCK_PREFIX "incq %0"
296 * atomic64_dec - decrement atomic64 variable
297 * @v: pointer to type atomic64_t
299 * Atomically decrements @v by 1.
301 static __inline__ void atomic64_dec(atomic64_t *v)
303 __asm__ __volatile__(
304 LOCK_PREFIX "decq %0"
310 * atomic64_dec_and_test - decrement and test
311 * @v: pointer to type atomic64_t
313 * Atomically decrements @v by 1 and
314 * returns true if the result is 0, or false for all other
317 static __inline__ int atomic64_dec_and_test(atomic64_t *v)
321 __asm__ __volatile__(
322 LOCK_PREFIX "decq %0; sete %1"
323 :"=m" (v->counter), "=qm" (c)
324 :"m" (v->counter) : "memory");
329 * atomic64_inc_and_test - increment and test
330 * @v: pointer to type atomic64_t
332 * Atomically increments @v by 1
333 * and returns true if the result is zero, or false for all
336 static __inline__ int atomic64_inc_and_test(atomic64_t *v)
340 __asm__ __volatile__(
341 LOCK_PREFIX "incq %0; sete %1"
342 :"=m" (v->counter), "=qm" (c)
343 :"m" (v->counter) : "memory");
348 * atomic64_add_negative - add and test if negative
349 * @i: integer value to add
350 * @v: pointer to type atomic64_t
352 * Atomically adds @i to @v and returns true
353 * if the result is negative, or false when
354 * result is greater than or equal to zero.
356 static __inline__ int atomic64_add_negative(long i, atomic64_t *v)
360 __asm__ __volatile__(
361 LOCK_PREFIX "addq %2,%0; sets %1"
362 :"=m" (v->counter), "=qm" (c)
363 :"ir" (i), "m" (v->counter) : "memory");
368 * atomic64_add_return - add and return
369 * @i: integer value to add
370 * @v: pointer to type atomic64_t
372 * Atomically adds @i to @v and returns @i + @v
374 static __inline__ long atomic64_add_return(long i, atomic64_t *v)
377 __asm__ __volatile__(
378 LOCK_PREFIX "xaddq %0, %1;"
379 :"+r" (i), "+m" (v->counter)
384 static __inline__ long atomic64_sub_return(long i, atomic64_t *v)
386 return atomic64_add_return(-i,v);
389 #define atomic64_inc_return(v) (atomic64_add_return(1,v))
390 #define atomic64_dec_return(v) (atomic64_sub_return(1,v))
392 #define atomic64_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new))
393 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new))
395 #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new))
396 #define atomic_xchg(v, new) (xchg(&((v)->counter), new))
399 * atomic_add_unless - add unless the number is a given value
400 * @v: pointer of type atomic_t
401 * @a: the amount to add to v...
402 * @u: ...unless v is equal to u.
404 * Atomically adds @a to @v, so long as it was not @u.
405 * Returns non-zero if @v was not @u, and zero otherwise.
407 static __inline__ int atomic_add_unless(atomic_t *v, int a, int u)
412 if (unlikely(c == (u)))
414 old = atomic_cmpxchg((v), c, c + (a));
415 if (likely(old == c))
422 #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0)
425 * atomic64_add_unless - add unless the number is a given value
426 * @v: pointer of type atomic64_t
427 * @a: the amount to add to v...
428 * @u: ...unless v is equal to u.
430 * Atomically adds @a to @v, so long as it was not @u.
431 * Returns non-zero if @v was not @u, and zero otherwise.
433 static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u)
436 c = atomic64_read(v);
438 if (unlikely(c == (u)))
440 old = atomic64_cmpxchg((v), c, c + (a));
441 if (likely(old == c))
448 #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0)
450 /* These are x86-specific, used by some header files */
451 #define atomic_clear_mask(mask, addr) \
452 __asm__ __volatile__(LOCK_PREFIX "andl %0,%1" \
453 : : "r" (~(mask)),"m" (*addr) : "memory")
455 #define atomic_set_mask(mask, addr) \
456 __asm__ __volatile__(LOCK_PREFIX "orl %0,%1" \
457 : : "r" ((unsigned)mask),"m" (*(addr)) : "memory")
459 /* Atomic operations are already serializing on x86 */
460 #define smp_mb__before_atomic_dec() barrier()
461 #define smp_mb__after_atomic_dec() barrier()
462 #define smp_mb__before_atomic_inc() barrier()
463 #define smp_mb__after_atomic_inc() barrier()
465 #include <asm-generic/atomic.h>