| #include <linux/module.h> |
| #include <linux/spinlock.h> |
| #include <asm/atomic.h> |
| |
| /* |
| * This is an architecture-neutral, but slow, |
| * implementation of the notion of "decrement |
| * a reference count, and return locked if it |
| * decremented to zero". |
| * |
| * NOTE NOTE NOTE! This is _not_ equivalent to |
| * |
| * if (atomic_dec_and_test(&atomic)) { |
| * spin_lock(&lock); |
| * return 1; |
| * } |
| * return 0; |
| * |
| * because the spin-lock and the decrement must be |
| * "atomic". |
| * |
| * This slow version gets the spinlock unconditionally, |
| * and releases it if it isn't needed. Architectures |
| * are encouraged to come up with better approaches, |
| * this is trivially done efficiently using a load-locked |
| * store-conditional approach, for example. |
| */ |
| int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) |
| { |
| spin_lock(lock); |
| if (atomic_dec_and_test(atomic)) |
| return 1; |
| spin_unlock(lock); |
| return 0; |
| } |
| |
| EXPORT_SYMBOL(_atomic_dec_and_lock); |