| /* |
| * Copyright (C) 2010 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "Dalvik.h" |
| |
| #include <cutils/atomic.h> |
| |
| #if defined(__arm__) |
| #include <machine/cpu-features.h> |
| #endif |
| |
| /*****************************************************************************/ |
| |
| #if defined(HAVE_MACOSX_IPC) |
| #define NEED_MAC_QUASI_ATOMICS 1 |
| |
| #elif defined(__i386__) || defined(__x86_64__) |
| #define NEED_PTHREADS_QUASI_ATOMICS 1 |
| |
| #elif defined(__mips__) |
| #define NEED_PTHREADS_QUASI_ATOMICS 1 |
| |
| #elif defined(__arm__) |
| |
| // TODO: Clang can not process our inline assembly at the moment. |
| #if defined(__ARM_HAVE_LDREXD) && !defined(__clang__) |
| #define NEED_ARM_LDREXD_QUASI_ATOMICS 1 |
| #else |
| #define NEED_PTHREADS_QUASI_ATOMICS 1 |
| #endif |
| |
| #else |
| #error "Unsupported atomic operations for this platform" |
| #endif |
| |
| /*****************************************************************************/ |
| |
| #if NEED_ARM_LDREXD_QUASI_ATOMICS |
| |
| static inline int64_t dvmQuasiAtomicSwap64Body(int64_t newvalue, |
| volatile int64_t* addr) |
| { |
| int64_t prev; |
| int status; |
| do { |
| __asm__ __volatile__ ("@ dvmQuasiAtomicSwap64\n" |
| "ldrexd %0, %H0, [%3]\n" |
| "strexd %1, %4, %H4, [%3]" |
| : "=&r" (prev), "=&r" (status), "+m"(*addr) |
| : "r" (addr), "r" (newvalue) |
| : "cc"); |
| } while (__builtin_expect(status != 0, 0)); |
| return prev; |
| } |
| |
| int64_t dvmQuasiAtomicSwap64(int64_t newvalue, volatile int64_t* addr) |
| { |
| return dvmQuasiAtomicSwap64Body(newvalue, addr); |
| } |
| |
| int64_t dvmQuasiAtomicSwap64Sync(int64_t newvalue, volatile int64_t* addr) |
| { |
| int64_t prev; |
| ANDROID_MEMBAR_STORE(); |
| prev = dvmQuasiAtomicSwap64Body(newvalue, addr); |
| ANDROID_MEMBAR_FULL(); |
| return prev; |
| } |
| |
| int dvmQuasiAtomicCas64(int64_t oldvalue, int64_t newvalue, |
| volatile int64_t* addr) |
| { |
| int64_t prev; |
| int status; |
| do { |
| __asm__ __volatile__ ("@ dvmQuasiAtomicCas64\n" |
| "ldrexd %0, %H0, [%3]\n" |
| "mov %1, #0\n" |
| "teq %0, %4\n" |
| "teqeq %H0, %H4\n" |
| "strexdeq %1, %5, %H5, [%3]" |
| : "=&r" (prev), "=&r" (status), "+m"(*addr) |
| : "r" (addr), "Ir" (oldvalue), "r" (newvalue) |
| : "cc"); |
| } while (__builtin_expect(status != 0, 0)); |
| return prev != oldvalue; |
| } |
| |
| int64_t dvmQuasiAtomicRead64(volatile const int64_t* addr) |
| { |
| int64_t value; |
| __asm__ __volatile__ ("@ dvmQuasiAtomicRead64\n" |
| "ldrexd %0, %H0, [%1]" |
| : "=&r" (value) |
| : "r" (addr)); |
| return value; |
| } |
| #endif |
| |
| /*****************************************************************************/ |
| |
| #if NEED_MAC_QUASI_ATOMICS |
| |
| #include <libkern/OSAtomic.h> |
| |
| int dvmQuasiAtomicCas64(int64_t oldvalue, int64_t newvalue, |
| volatile int64_t* addr) |
| { |
| return OSAtomicCompareAndSwap64Barrier(oldvalue, newvalue, |
| (int64_t*)addr) == 0; |
| } |
| |
| |
| static inline int64_t dvmQuasiAtomicSwap64Body(int64_t value, |
| volatile int64_t* addr) |
| { |
| int64_t oldValue; |
| do { |
| oldValue = *addr; |
| } while (dvmQuasiAtomicCas64(oldValue, value, addr)); |
| return oldValue; |
| } |
| |
| int64_t dvmQuasiAtomicSwap64(int64_t value, volatile int64_t* addr) |
| { |
| return dvmQuasiAtomicSwap64Body(value, addr); |
| } |
| |
| int64_t dvmQuasiAtomicSwap64Sync(int64_t value, volatile int64_t* addr) |
| { |
| int64_t oldValue; |
| ANDROID_MEMBAR_STORE(); |
| oldValue = dvmQuasiAtomicSwap64Body(value, addr); |
| /* TUNING: barriers can be avoided on some architectures */ |
| ANDROID_MEMBAR_FULL(); |
| return oldValue; |
| } |
| |
| int64_t dvmQuasiAtomicRead64(volatile const int64_t* addr) |
| { |
| return OSAtomicAdd64Barrier(0, addr); |
| } |
| #endif |
| |
| /*****************************************************************************/ |
| |
| #if NEED_PTHREADS_QUASI_ATOMICS |
| |
| // In the absence of a better implementation, we implement the 64-bit atomic |
| // operations through mutex locking. |
| |
| // another twist is that we use a small array of mutexes to dispatch |
| // the contention locks from different memory addresses |
| |
| #include <pthread.h> |
| |
| static const size_t kSwapLockCount = 32; |
| static pthread_mutex_t* gSwapLocks[kSwapLockCount]; |
| |
| void dvmQuasiAtomicsStartup() { |
| for (size_t i = 0; i < kSwapLockCount; ++i) { |
| pthread_mutex_t* m = new pthread_mutex_t; |
| dvmInitMutex(m); |
| gSwapLocks[i] = m; |
| } |
| } |
| |
| void dvmQuasiAtomicsShutdown() { |
| for (size_t i = 0; i < kSwapLockCount; ++i) { |
| pthread_mutex_t* m = gSwapLocks[i]; |
| gSwapLocks[i] = NULL; |
| if (m != NULL) { |
| dvmDestroyMutex(m); |
| } |
| delete m; |
| } |
| } |
| |
| static inline pthread_mutex_t* GetSwapLock(const volatile int64_t* addr) { |
| return gSwapLocks[((unsigned)(void*)(addr) >> 3U) % kSwapLockCount]; |
| } |
| |
| int64_t dvmQuasiAtomicSwap64(int64_t value, volatile int64_t* addr) |
| { |
| int64_t oldValue; |
| pthread_mutex_t* lock = GetSwapLock(addr); |
| |
| pthread_mutex_lock(lock); |
| |
| oldValue = *addr; |
| *addr = value; |
| |
| pthread_mutex_unlock(lock); |
| return oldValue; |
| } |
| |
| /* Same as dvmQuasiAtomicSwap64 - mutex handles barrier */ |
| int64_t dvmQuasiAtomicSwap64Sync(int64_t value, volatile int64_t* addr) |
| { |
| return dvmQuasiAtomicSwap64(value, addr); |
| } |
| |
| int dvmQuasiAtomicCas64(int64_t oldvalue, int64_t newvalue, |
| volatile int64_t* addr) |
| { |
| int result; |
| pthread_mutex_t* lock = GetSwapLock(addr); |
| |
| pthread_mutex_lock(lock); |
| |
| if (*addr == oldvalue) { |
| *addr = newvalue; |
| result = 0; |
| } else { |
| result = 1; |
| } |
| pthread_mutex_unlock(lock); |
| return result; |
| } |
| |
| int64_t dvmQuasiAtomicRead64(volatile const int64_t* addr) |
| { |
| int64_t result; |
| pthread_mutex_t* lock = GetSwapLock(addr); |
| |
| pthread_mutex_lock(lock); |
| result = *addr; |
| pthread_mutex_unlock(lock); |
| return result; |
| } |
| |
| #else |
| |
| // The other implementations don't need any special setup. |
| void dvmQuasiAtomicsStartup() {} |
| void dvmQuasiAtomicsShutdown() {} |
| |
| #endif /*NEED_PTHREADS_QUASI_ATOMICS*/ |