Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | #include <linux/module.h> |
| 2 | #include <linux/spinlock.h> |
| 3 | #include <asm/atomic.h> |
| 4 | |
| 5 | /* |
| 6 | * This is an architecture-neutral, but slow, |
| 7 | * implementation of the notion of "decrement |
| 8 | * a reference count, and return locked if it |
| 9 | * decremented to zero". |
| 10 | * |
| 11 | * NOTE NOTE NOTE! This is _not_ equivalent to |
| 12 | * |
| 13 | * if (atomic_dec_and_test(&atomic)) { |
| 14 | * spin_lock(&lock); |
| 15 | * return 1; |
| 16 | * } |
| 17 | * return 0; |
| 18 | * |
| 19 | * because the spin-lock and the decrement must be |
| 20 | * "atomic". |
| 21 | * |
| 22 | * This slow version gets the spinlock unconditionally, |
| 23 | * and releases it if it isn't needed. Architectures |
| 24 | * are encouraged to come up with better approaches, |
| 25 | * this is trivially done efficiently using a load-locked |
| 26 | * store-conditional approach, for example. |
| 27 | */ |
| 28 | |
| 29 | #ifndef ATOMIC_DEC_AND_LOCK |
| 30 | int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) |
| 31 | { |
| 32 | spin_lock(lock); |
| 33 | if (atomic_dec_and_test(atomic)) |
| 34 | return 1; |
| 35 | spin_unlock(lock); |
| 36 | return 0; |
| 37 | } |
| 38 | |
| 39 | EXPORT_SYMBOL(_atomic_dec_and_lock); |
| 40 | #endif |