blob: 43eaa6e742e06f3a1f77dffca9d6eccf61fa29d4 [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001#ifndef __ASM_SPINLOCK_H
2#define __ASM_SPINLOCK_H
3
4#include <asm/system.h>
Ingo Molnarfb1c8f92005-09-10 00:25:56 -07005#include <asm/processor.h>
6#include <asm/spinlock_types.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -07007
8/* Note that PA-RISC has to use `1' to mean unlocked and `0' to mean locked
9 * since it only has load-and-zero. Moreover, at least on some PA processors,
10 * the semaphore address has to be 16-byte aligned.
11 */
12
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070013static inline int __raw_spin_is_locked(raw_spinlock_t *x)
Linus Torvalds1da177e2005-04-16 15:20:36 -070014{
15 volatile unsigned int *a = __ldcw_align(x);
16 return *a == 0;
17}
18
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070019#define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock)
20#define __raw_spin_unlock_wait(x) \
21 do { cpu_relax(); } while (__raw_spin_is_locked(x))
Linus Torvalds1da177e2005-04-16 15:20:36 -070022
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070023static inline void __raw_spin_lock(raw_spinlock_t *x)
Linus Torvalds1da177e2005-04-16 15:20:36 -070024{
25 volatile unsigned int *a;
26
27 mb();
28 a = __ldcw_align(x);
29 while (__ldcw(a) == 0)
30 while (*a == 0);
31 mb();
32}
33
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070034static inline void __raw_spin_unlock(raw_spinlock_t *x)
Linus Torvalds1da177e2005-04-16 15:20:36 -070035{
36 volatile unsigned int *a;
37 mb();
38 a = __ldcw_align(x);
39 *a = 1;
40 mb();
41}
42
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070043static inline int __raw_spin_trylock(raw_spinlock_t *x)
Linus Torvalds1da177e2005-04-16 15:20:36 -070044{
45 volatile unsigned int *a;
46 int ret;
47
48 mb();
49 a = __ldcw_align(x);
50 ret = __ldcw(a) != 0;
51 mb();
52
53 return ret;
54}
Linus Torvalds1da177e2005-04-16 15:20:36 -070055
56/*
57 * Read-write spinlocks, allowing multiple readers
58 * but only one writer.
59 */
Linus Torvalds1da177e2005-04-16 15:20:36 -070060
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070061#define __raw_read_trylock(lock) generic__raw_read_trylock(lock)
Linus Torvalds1da177e2005-04-16 15:20:36 -070062
63/* read_lock, read_unlock are pretty straightforward. Of course it somehow
64 * sucks we end up saving/restoring flags twice for read_lock_irqsave aso. */
65
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070066static __inline__ void __raw_read_lock(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -070067{
68 unsigned long flags;
69 local_irq_save(flags);
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070070 __raw_spin_lock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -070071
72 rw->counter++;
73
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070074 __raw_spin_unlock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -070075 local_irq_restore(flags);
76}
Linus Torvalds1da177e2005-04-16 15:20:36 -070077
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070078static __inline__ void __raw_read_unlock(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -070079{
80 unsigned long flags;
81 local_irq_save(flags);
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070082 __raw_spin_lock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -070083
84 rw->counter--;
85
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070086 __raw_spin_unlock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -070087 local_irq_restore(flags);
88}
89
90/* write_lock is less trivial. We optimistically grab the lock and check
91 * if we surprised any readers. If so we release the lock and wait till
92 * they're all gone before trying again
93 *
94 * Also note that we don't use the _irqsave / _irqrestore suffixes here.
95 * If we're called with interrupts enabled and we've got readers (or other
96 * writers) in interrupt handlers someone fucked up and we'd dead-lock
97 * sooner or later anyway. prumpf */
98
Ingo Molnarfb1c8f92005-09-10 00:25:56 -070099static __inline__ void __raw_write_lock(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700100{
101retry:
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700102 __raw_spin_lock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700103
104 if(rw->counter != 0) {
105 /* this basically never happens */
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700106 __raw_spin_unlock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700107
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700108 while (rw->counter != 0)
109 cpu_relax();
Linus Torvalds1da177e2005-04-16 15:20:36 -0700110
111 goto retry;
112 }
113
114 /* got it. now leave without unlocking */
115 rw->counter = -1; /* remember we are locked */
116}
Linus Torvalds1da177e2005-04-16 15:20:36 -0700117
118/* write_unlock is absolutely trivial - we don't have to wait for anything */
119
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700120static __inline__ void __raw_write_unlock(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700121{
122 rw->counter = 0;
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700123 __raw_spin_unlock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700124}
125
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700126static __inline__ int __raw_write_trylock(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700127{
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700128 __raw_spin_lock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700129 if (rw->counter != 0) {
130 /* this basically never happens */
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700131 __raw_spin_unlock(&rw->lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700132
133 return 0;
134 }
135
136 /* got it. now leave without unlocking */
137 rw->counter = -1; /* remember we are locked */
138 return 1;
139}
Linus Torvalds1da177e2005-04-16 15:20:36 -0700140
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700141static __inline__ int __raw_is_read_locked(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700142{
143 return rw->counter > 0;
144}
145
Ingo Molnarfb1c8f92005-09-10 00:25:56 -0700146static __inline__ int __raw_is_write_locked(raw_rwlock_t *rw)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700147{
148 return rw->counter < 0;
149}
150
151#endif /* __ASM_SPINLOCK_H */