| /* |
| * Copyright (c) 2003 Fabrice Bellard |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, see <http://www.gnu.org/licenses/> |
| */ |
| |
| /* Locking primitives. Most of this code should be redundant - |
| system emulation doesn't need/use locking, NPTL userspace uses |
| pthread mutexes, and non-NPTL userspace isn't threadsafe anyway. |
| In either case a spinlock is probably the wrong kind of lock. |
| Spinlocks are only good if you know annother CPU has the lock and is |
| likely to release it soon. In environments where you have more threads |
| than physical CPUs (the extreme case being a single CPU host) a spinlock |
| simply wastes CPU until the OS decides to preempt it. */ |
| #if defined(CONFIG_USE_NPTL) |
| |
| #include <pthread.h> |
| #define spin_lock pthread_mutex_lock |
| #define spin_unlock pthread_mutex_unlock |
| #define spinlock_t pthread_mutex_t |
| #define SPIN_LOCK_UNLOCKED PTHREAD_MUTEX_INITIALIZER |
| |
| #else |
| |
| #if defined(__hppa__) |
| |
| typedef int spinlock_t[4]; |
| |
| #define SPIN_LOCK_UNLOCKED { 1, 1, 1, 1 } |
| |
| static inline void resetlock (spinlock_t *p) |
| { |
| (*p)[0] = (*p)[1] = (*p)[2] = (*p)[3] = 1; |
| } |
| |
| #else |
| |
| typedef int spinlock_t; |
| |
| #define SPIN_LOCK_UNLOCKED 0 |
| |
| static inline void resetlock (spinlock_t *p) |
| { |
| *p = SPIN_LOCK_UNLOCKED; |
| } |
| |
| #endif |
| |
| #if defined(_ARCH_PPC) |
| static inline int testandset (int *p) |
| { |
| int ret; |
| __asm__ __volatile__ ( |
| " lwarx %0,0,%1\n" |
| " xor. %0,%3,%0\n" |
| " bne $+12\n" |
| " stwcx. %2,0,%1\n" |
| " bne- $-16\n" |
| : "=&r" (ret) |
| : "r" (p), "r" (1), "r" (0) |
| : "cr0", "memory"); |
| return ret; |
| } |
| #elif defined(__i386__) |
| static inline int testandset (int *p) |
| { |
| long int readval = 0; |
| |
| __asm__ __volatile__ ("lock; cmpxchgl %2, %0" |
| : "+m" (*p), "+a" (readval) |
| : "r" (1) |
| : "cc"); |
| return readval; |
| } |
| #elif defined(__x86_64__) |
| static inline int testandset (int *p) |
| { |
| long int readval = 0; |
| |
| __asm__ __volatile__ ("lock; cmpxchgl %2, %0" |
| : "+m" (*p), "+a" (readval) |
| : "r" (1) |
| : "cc"); |
| return readval; |
| } |
| #elif defined(__s390__) |
| static inline int testandset (int *p) |
| { |
| int ret; |
| |
| __asm__ __volatile__ ("0: cs %0,%1,0(%2)\n" |
| " jl 0b" |
| : "=&d" (ret) |
| : "r" (1), "a" (p), "0" (*p) |
| : "cc", "memory" ); |
| return ret; |
| } |
| #elif defined(__alpha__) |
| static inline int testandset (int *p) |
| { |
| int ret; |
| unsigned long one; |
| |
| __asm__ __volatile__ ("0: mov 1,%2\n" |
| " ldl_l %0,%1\n" |
| " stl_c %2,%1\n" |
| " beq %2,1f\n" |
| ".subsection 2\n" |
| "1: br 0b\n" |
| ".previous" |
| : "=r" (ret), "=m" (*p), "=r" (one) |
| : "m" (*p)); |
| return ret; |
| } |
| #elif defined(__sparc__) |
| static inline int testandset (int *p) |
| { |
| int ret; |
| |
| __asm__ __volatile__("ldstub [%1], %0" |
| : "=r" (ret) |
| : "r" (p) |
| : "memory"); |
| |
| return (ret ? 1 : 0); |
| } |
| #elif defined(__arm__) |
| static inline int testandset (int *spinlock) |
| { |
| register unsigned int ret; |
| __asm__ __volatile__("swp %0, %1, [%2]" |
| : "=r"(ret) |
| : "0"(1), "r"(spinlock)); |
| |
| return ret; |
| } |
| #elif defined(__mc68000) |
| static inline int testandset (int *p) |
| { |
| char ret; |
| __asm__ __volatile__("tas %1; sne %0" |
| : "=r" (ret) |
| : "m" (p) |
| : "cc","memory"); |
| return ret; |
| } |
| #elif defined(__hppa__) |
| |
| /* Because malloc only guarantees 8-byte alignment for malloc'd data, |
| and GCC only guarantees 8-byte alignment for stack locals, we can't |
| be assured of 16-byte alignment for atomic lock data even if we |
| specify "__attribute ((aligned(16)))" in the type declaration. So, |
| we use a struct containing an array of four ints for the atomic lock |
| type and dynamically select the 16-byte aligned int from the array |
| for the semaphore. */ |
| #define __PA_LDCW_ALIGNMENT 16 |
| static inline void *ldcw_align (void *p) { |
| unsigned long a = (unsigned long)p; |
| a = (a + __PA_LDCW_ALIGNMENT - 1) & ~(__PA_LDCW_ALIGNMENT - 1); |
| return (void *)a; |
| } |
| |
| static inline int testandset (spinlock_t *p) |
| { |
| unsigned int ret; |
| p = ldcw_align(p); |
| __asm__ __volatile__("ldcw 0(%1),%0" |
| : "=r" (ret) |
| : "r" (p) |
| : "memory" ); |
| return !ret; |
| } |
| |
| #elif defined(__ia64) |
| |
| #include <ia64intrin.h> |
| |
| static inline int testandset (int *p) |
| { |
| return __sync_lock_test_and_set (p, 1); |
| } |
| #elif defined(__mips__) |
| static inline int testandset (int *p) |
| { |
| int ret; |
| |
| __asm__ __volatile__ ( |
| " .set push \n" |
| " .set noat \n" |
| " .set mips2 \n" |
| "1: li $1, 1 \n" |
| " ll %0, %1 \n" |
| " sc $1, %1 \n" |
| " beqz $1, 1b \n" |
| " .set pop " |
| : "=r" (ret), "+R" (*p) |
| : |
| : "memory"); |
| |
| return ret; |
| } |
| #else |
| #error unimplemented CPU support |
| #endif |
| |
| #if defined(CONFIG_USER_ONLY) |
| static inline void spin_lock(spinlock_t *lock) |
| { |
| while (testandset(lock)); |
| } |
| |
| static inline void spin_unlock(spinlock_t *lock) |
| { |
| resetlock(lock); |
| } |
| |
| static inline int spin_trylock(spinlock_t *lock) |
| { |
| return !testandset(lock); |
| } |
| #else |
| static inline void spin_lock(spinlock_t *lock) |
| { |
| } |
| |
| static inline void spin_unlock(spinlock_t *lock) |
| { |
| } |
| |
| static inline int spin_trylock(spinlock_t *lock) |
| { |
| return 1; |
| } |
| #endif |
| |
| #endif |