Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (C) 2008 The Android Open Source Project |
| 3 | * |
| 4 | * Licensed under the Apache License, Version 2.0 (the "License"); |
| 5 | * you may not use this file except in compliance with the License. |
| 6 | * You may obtain a copy of the License at |
| 7 | * |
| 8 | * http://www.apache.org/licenses/LICENSE-2.0 |
| 9 | * |
| 10 | * Unless required by applicable law or agreed to in writing, software |
| 11 | * distributed under the License is distributed on an "AS IS" BASIS, |
| 12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 13 | * See the License for the specific language governing permissions and |
| 14 | * limitations under the License. |
| 15 | */ |
| 16 | |
Brian Carlstrom | fc0e321 | 2013-07-17 14:40:12 -0700 | [diff] [blame] | 17 | #ifndef ART_RUNTIME_ATOMIC_H_ |
| 18 | #define ART_RUNTIME_ATOMIC_H_ |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 19 | |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 20 | #include <stdint.h> |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 21 | #include <atomic> |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 22 | #include <limits> |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 23 | #include <vector> |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 24 | |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 25 | #include "arch/instruction_set.h" |
Ian Rogers | a984454 | 2014-04-21 17:01:02 -0700 | [diff] [blame] | 26 | #include "base/logging.h" |
Elliott Hughes | 7616005 | 2012-12-12 16:31:20 -0800 | [diff] [blame] | 27 | #include "base/macros.h" |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 28 | |
| 29 | namespace art { |
| 30 | |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 31 | class Mutex; |
| 32 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 33 | // QuasiAtomic encapsulates two separate facilities that we are |
| 34 | // trying to move away from: "quasiatomic" 64 bit operations |
| 35 | // and custom memory fences. For the time being, they remain |
| 36 | // exposed. Clients should be converted to use either class Atomic |
| 37 | // below whenever possible, and should eventually use C++11 atomics. |
| 38 | // The two facilities that do not have a good C++11 analog are |
| 39 | // ThreadFenceForConstructor and Atomic::*JavaData. |
| 40 | // |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 41 | // NOTE: Two "quasiatomic" operations on the exact same memory address |
| 42 | // are guaranteed to operate atomically with respect to each other, |
| 43 | // but no guarantees are made about quasiatomic operations mixed with |
| 44 | // non-quasiatomic operations on the same address, nor about |
| 45 | // quasiatomic operations that are performed on partially-overlapping |
| 46 | // memory. |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 47 | class QuasiAtomic { |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 48 | static constexpr bool NeedSwapMutexes(InstructionSet isa) { |
| 49 | // TODO - mips64 still need this for Cas64 ??? |
| 50 | return (isa == kMips) || (isa == kMips64); |
| 51 | } |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 52 | |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 53 | public: |
| 54 | static void Startup(); |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 55 | |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 56 | static void Shutdown(); |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 57 | |
Ian Rogers | 9adbff5 | 2013-01-23 18:19:03 -0800 | [diff] [blame] | 58 | // Reads the 64-bit value at "addr" without tearing. |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 59 | static int64_t Read64(volatile const int64_t* addr) { |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 60 | if (!NeedSwapMutexes(kRuntimeISA)) { |
Ian Rogers | a984454 | 2014-04-21 17:01:02 -0700 | [diff] [blame] | 61 | int64_t value; |
| 62 | #if defined(__LP64__) |
| 63 | value = *addr; |
| 64 | #else |
| 65 | #if defined(__arm__) |
| 66 | #if defined(__ARM_FEATURE_LPAE) |
| 67 | // With LPAE support (such as Cortex-A15) then ldrd is defined not to tear. |
| 68 | __asm__ __volatile__("@ QuasiAtomic::Read64\n" |
| 69 | "ldrd %0, %H0, %1" |
| 70 | : "=r" (value) |
| 71 | : "m" (*addr)); |
| 72 | #else |
| 73 | // Exclusive loads are defined not to tear, clearing the exclusive state isn't necessary. |
| 74 | __asm__ __volatile__("@ QuasiAtomic::Read64\n" |
| 75 | "ldrexd %0, %H0, %1" |
| 76 | : "=r" (value) |
| 77 | : "Q" (*addr)); |
| 78 | #endif |
| 79 | #elif defined(__i386__) |
| 80 | __asm__ __volatile__( |
| 81 | "movq %1, %0\n" |
| 82 | : "=x" (value) |
| 83 | : "m" (*addr)); |
| 84 | #else |
| 85 | LOG(FATAL) << "Unsupported architecture"; |
| 86 | #endif |
| 87 | #endif // defined(__LP64__) |
| 88 | return value; |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 89 | } else { |
| 90 | return SwapMutexRead64(addr); |
| 91 | } |
| 92 | } |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 93 | |
Ian Rogers | 9adbff5 | 2013-01-23 18:19:03 -0800 | [diff] [blame] | 94 | // Writes to the 64-bit value at "addr" without tearing. |
Ian Rogers | a984454 | 2014-04-21 17:01:02 -0700 | [diff] [blame] | 95 | static void Write64(volatile int64_t* addr, int64_t value) { |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 96 | if (!NeedSwapMutexes(kRuntimeISA)) { |
Ian Rogers | a984454 | 2014-04-21 17:01:02 -0700 | [diff] [blame] | 97 | #if defined(__LP64__) |
| 98 | *addr = value; |
| 99 | #else |
| 100 | #if defined(__arm__) |
| 101 | #if defined(__ARM_FEATURE_LPAE) |
| 102 | // If we know that ARM architecture has LPAE (such as Cortex-A15) strd is defined not to tear. |
| 103 | __asm__ __volatile__("@ QuasiAtomic::Write64\n" |
| 104 | "strd %1, %H1, %0" |
| 105 | : "=m"(*addr) |
| 106 | : "r" (value)); |
| 107 | #else |
| 108 | // The write is done as a swap so that the cache-line is in the exclusive state for the store. |
| 109 | int64_t prev; |
| 110 | int status; |
| 111 | do { |
| 112 | __asm__ __volatile__("@ QuasiAtomic::Write64\n" |
| 113 | "ldrexd %0, %H0, %2\n" |
| 114 | "strexd %1, %3, %H3, %2" |
| 115 | : "=&r" (prev), "=&r" (status), "+Q"(*addr) |
| 116 | : "r" (value) |
| 117 | : "cc"); |
| 118 | } while (UNLIKELY(status != 0)); |
| 119 | #endif |
| 120 | #elif defined(__i386__) |
| 121 | __asm__ __volatile__( |
| 122 | "movq %1, %0" |
| 123 | : "=m" (*addr) |
| 124 | : "x" (value)); |
| 125 | #else |
| 126 | LOG(FATAL) << "Unsupported architecture"; |
| 127 | #endif |
| 128 | #endif // defined(__LP64__) |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 129 | } else { |
Ian Rogers | a984454 | 2014-04-21 17:01:02 -0700 | [diff] [blame] | 130 | SwapMutexWrite64(addr, value); |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 131 | } |
| 132 | } |
Ian Rogers | 9adbff5 | 2013-01-23 18:19:03 -0800 | [diff] [blame] | 133 | |
| 134 | // Atomically compare the value at "addr" to "old_value", if equal replace it with "new_value" |
| 135 | // and return true. Otherwise, don't swap, and return false. |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 136 | // This is fully ordered, i.e. it has C++11 memory_order_seq_cst |
| 137 | // semantics (assuming all other accesses use a mutex if this one does). |
| 138 | // This has "strong" semantics; if it fails then it is guaranteed that |
| 139 | // at some point during the execution of Cas64, *addr was not equal to |
| 140 | // old_value. |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 141 | static bool Cas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) { |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 142 | if (!NeedSwapMutexes(kRuntimeISA)) { |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 143 | return __sync_bool_compare_and_swap(addr, old_value, new_value); |
| 144 | } else { |
| 145 | return SwapMutexCas64(old_value, new_value, addr); |
| 146 | } |
| 147 | } |
Ian Rogers | 9adbff5 | 2013-01-23 18:19:03 -0800 | [diff] [blame] | 148 | |
| 149 | // Does the architecture provide reasonable atomic long operations or do we fall back on mutexes? |
Andreas Gampe | 0866f4e | 2016-02-22 10:03:12 -0800 | [diff] [blame] | 150 | static bool LongAtomicsUseMutexes(InstructionSet isa) { |
| 151 | return NeedSwapMutexes(isa); |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 152 | } |
| 153 | |
Hans Boehm | a1ec065 | 2014-06-06 17:13:03 -0700 | [diff] [blame] | 154 | static void ThreadFenceAcquire() { |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 155 | std::atomic_thread_fence(std::memory_order_acquire); |
| 156 | } |
| 157 | |
Hans Boehm | a1ec065 | 2014-06-06 17:13:03 -0700 | [diff] [blame] | 158 | static void ThreadFenceRelease() { |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 159 | std::atomic_thread_fence(std::memory_order_release); |
| 160 | } |
| 161 | |
| 162 | static void ThreadFenceForConstructor() { |
| 163 | #if defined(__aarch64__) |
| 164 | __asm__ __volatile__("dmb ishst" : : : "memory"); |
| 165 | #else |
| 166 | std::atomic_thread_fence(std::memory_order_release); |
| 167 | #endif |
| 168 | } |
| 169 | |
| 170 | static void ThreadFenceSequentiallyConsistent() { |
| 171 | std::atomic_thread_fence(std::memory_order_seq_cst); |
| 172 | } |
| 173 | |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 174 | private: |
Ian Rogers | b122a4b | 2013-11-19 18:00:50 -0800 | [diff] [blame] | 175 | static Mutex* GetSwapMutex(const volatile int64_t* addr); |
| 176 | static int64_t SwapMutexRead64(volatile const int64_t* addr); |
| 177 | static void SwapMutexWrite64(volatile int64_t* addr, int64_t val); |
| 178 | static bool SwapMutexCas64(int64_t old_value, int64_t new_value, volatile int64_t* addr); |
| 179 | |
| 180 | // We stripe across a bunch of different mutexes to reduce contention. |
| 181 | static constexpr size_t kSwapMutexCount = 32; |
| 182 | static std::vector<Mutex*>* gSwapMutexes; |
| 183 | |
Elliott Hughes | 7c6169d | 2012-05-02 16:11:48 -0700 | [diff] [blame] | 184 | DISALLOW_COPY_AND_ASSIGN(QuasiAtomic); |
| 185 | }; |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 186 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 187 | template<typename T> |
Dan Albert | aab0f86 | 2014-08-11 16:38:02 -0700 | [diff] [blame] | 188 | class PACKED(sizeof(T)) Atomic : public std::atomic<T> { |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 189 | public: |
Dan Albert | 6a3f8d9 | 2014-08-12 11:48:34 -0700 | [diff] [blame] | 190 | Atomic<T>() : std::atomic<T>(0) { } |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 191 | |
| 192 | explicit Atomic<T>(T value) : std::atomic<T>(value) { } |
| 193 | |
| 194 | // Load from memory without ordering or synchronization constraints. |
| 195 | T LoadRelaxed() const { |
| 196 | return this->load(std::memory_order_relaxed); |
| 197 | } |
| 198 | |
Calin Juravle | 4d77b6a | 2015-12-01 18:38:09 +0000 | [diff] [blame] | 199 | // Load from memory with acquire ordering. |
| 200 | T LoadAcquire() const { |
| 201 | return this->load(std::memory_order_acquire); |
| 202 | } |
| 203 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 204 | // Word tearing allowed, but may race. |
| 205 | // TODO: Optimize? |
| 206 | // There has been some discussion of eventually disallowing word |
| 207 | // tearing for Java data loads. |
| 208 | T LoadJavaData() const { |
| 209 | return this->load(std::memory_order_relaxed); |
| 210 | } |
| 211 | |
| 212 | // Load from memory with a total ordering. |
| 213 | // Corresponds exactly to a Java volatile load. |
| 214 | T LoadSequentiallyConsistent() const { |
| 215 | return this->load(std::memory_order_seq_cst); |
| 216 | } |
| 217 | |
| 218 | // Store to memory without ordering or synchronization constraints. |
| 219 | void StoreRelaxed(T desired) { |
| 220 | this->store(desired, std::memory_order_relaxed); |
| 221 | } |
| 222 | |
| 223 | // Word tearing allowed, but may race. |
| 224 | void StoreJavaData(T desired) { |
| 225 | this->store(desired, std::memory_order_relaxed); |
| 226 | } |
| 227 | |
| 228 | // Store to memory with release ordering. |
| 229 | void StoreRelease(T desired) { |
| 230 | this->store(desired, std::memory_order_release); |
| 231 | } |
| 232 | |
| 233 | // Store to memory with a total ordering. |
| 234 | void StoreSequentiallyConsistent(T desired) { |
| 235 | this->store(desired, std::memory_order_seq_cst); |
| 236 | } |
| 237 | |
Richard Uhler | caaa2b0 | 2017-02-01 09:54:17 +0000 | [diff] [blame] | 238 | // Atomically replace the value with desired value. |
| 239 | T ExchangeRelaxed(T desired_value) { |
| 240 | return this->exchange(desired_value, std::memory_order_relaxed); |
| 241 | } |
| 242 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 243 | // Atomically replace the value with desired value if it matches the expected value. |
| 244 | // Participates in total ordering of atomic operations. |
| 245 | bool CompareExchangeStrongSequentiallyConsistent(T expected_value, T desired_value) { |
| 246 | return this->compare_exchange_strong(expected_value, desired_value, std::memory_order_seq_cst); |
| 247 | } |
| 248 | |
| 249 | // The same, except it may fail spuriously. |
| 250 | bool CompareExchangeWeakSequentiallyConsistent(T expected_value, T desired_value) { |
| 251 | return this->compare_exchange_weak(expected_value, desired_value, std::memory_order_seq_cst); |
| 252 | } |
| 253 | |
| 254 | // Atomically replace the value with desired value if it matches the expected value. Doesn't |
| 255 | // imply ordering or synchronization constraints. |
| 256 | bool CompareExchangeStrongRelaxed(T expected_value, T desired_value) { |
| 257 | return this->compare_exchange_strong(expected_value, desired_value, std::memory_order_relaxed); |
| 258 | } |
| 259 | |
Mathieu Chartier | fdd513d | 2017-06-01 11:26:50 -0700 | [diff] [blame] | 260 | // Atomically replace the value with desired value if it matches the expected value. Prior writes |
| 261 | // to other memory locations become visible to the threads that do a consume or an acquire on the |
| 262 | // same location. |
| 263 | bool CompareExchangeStrongRelease(T expected_value, T desired_value) { |
| 264 | return this->compare_exchange_strong(expected_value, desired_value, std::memory_order_release); |
| 265 | } |
| 266 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 267 | // The same, except it may fail spuriously. |
| 268 | bool CompareExchangeWeakRelaxed(T expected_value, T desired_value) { |
| 269 | return this->compare_exchange_weak(expected_value, desired_value, std::memory_order_relaxed); |
| 270 | } |
| 271 | |
| 272 | // Atomically replace the value with desired value if it matches the expected value. Prior writes |
| 273 | // made to other memory locations by the thread that did the release become visible in this |
| 274 | // thread. |
| 275 | bool CompareExchangeWeakAcquire(T expected_value, T desired_value) { |
| 276 | return this->compare_exchange_weak(expected_value, desired_value, std::memory_order_acquire); |
| 277 | } |
| 278 | |
| 279 | // Atomically replace the value with desired value if it matches the expected value. prior writes |
| 280 | // to other memory locations become visible to the threads that do a consume or an acquire on the |
| 281 | // same location. |
| 282 | bool CompareExchangeWeakRelease(T expected_value, T desired_value) { |
| 283 | return this->compare_exchange_weak(expected_value, desired_value, std::memory_order_release); |
| 284 | } |
| 285 | |
| 286 | T FetchAndAddSequentiallyConsistent(const T value) { |
| 287 | return this->fetch_add(value, std::memory_order_seq_cst); // Return old_value. |
| 288 | } |
| 289 | |
Hans Boehm | b0171b9 | 2016-01-28 17:19:15 -0800 | [diff] [blame] | 290 | T FetchAndAddRelaxed(const T value) { |
| 291 | return this->fetch_add(value, std::memory_order_relaxed); // Return old_value. |
| 292 | } |
| 293 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 294 | T FetchAndSubSequentiallyConsistent(const T value) { |
| 295 | return this->fetch_sub(value, std::memory_order_seq_cst); // Return old value. |
| 296 | } |
| 297 | |
Richard Uhler | caaa2b0 | 2017-02-01 09:54:17 +0000 | [diff] [blame] | 298 | T FetchAndSubRelaxed(const T value) { |
| 299 | return this->fetch_sub(value, std::memory_order_relaxed); // Return old value. |
| 300 | } |
| 301 | |
Ian Rogers | 8c1b5f7 | 2014-07-09 22:02:36 -0700 | [diff] [blame] | 302 | T FetchAndOrSequentiallyConsistent(const T value) { |
| 303 | return this->fetch_or(value, std::memory_order_seq_cst); // Return old_value. |
| 304 | } |
| 305 | |
| 306 | T FetchAndAndSequentiallyConsistent(const T value) { |
| 307 | return this->fetch_and(value, std::memory_order_seq_cst); // Return old_value. |
| 308 | } |
| 309 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 310 | volatile T* Address() { |
| 311 | return reinterpret_cast<T*>(this); |
| 312 | } |
| 313 | |
| 314 | static T MaxValue() { |
| 315 | return std::numeric_limits<T>::max(); |
| 316 | } |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 317 | }; |
| 318 | |
Hans Boehm | 3035961 | 2014-05-21 17:46:23 -0700 | [diff] [blame] | 319 | typedef Atomic<int32_t> AtomicInteger; |
| 320 | |
Andreas Gampe | 575e78c | 2014-11-03 23:41:03 -0800 | [diff] [blame] | 321 | static_assert(sizeof(AtomicInteger) == sizeof(int32_t), "Weird AtomicInteger size"); |
| 322 | static_assert(alignof(AtomicInteger) == alignof(int32_t), |
| 323 | "AtomicInteger alignment differs from that of underlyingtype"); |
| 324 | static_assert(sizeof(Atomic<int64_t>) == sizeof(int64_t), "Weird Atomic<int64> size"); |
Dan Albert | aab0f86 | 2014-08-11 16:38:02 -0700 | [diff] [blame] | 325 | |
| 326 | // Assert the alignment of 64-bit integers is 64-bit. This isn't true on certain 32-bit |
| 327 | // architectures (e.g. x86-32) but we know that 64-bit integers here are arranged to be 8-byte |
| 328 | // aligned. |
Hans Boehm | 2f4a2ed | 2014-06-06 18:17:43 -0700 | [diff] [blame] | 329 | #if defined(__LP64__) |
Andreas Gampe | 575e78c | 2014-11-03 23:41:03 -0800 | [diff] [blame] | 330 | static_assert(alignof(Atomic<int64_t>) == alignof(int64_t), |
| 331 | "Atomic<int64> alignment differs from that of underlying type"); |
Hans Boehm | 2f4a2ed | 2014-06-06 18:17:43 -0700 | [diff] [blame] | 332 | #endif |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 333 | |
Elliott Hughes | 5ea047b | 2011-09-13 14:38:18 -0700 | [diff] [blame] | 334 | } // namespace art |
| 335 | |
Brian Carlstrom | fc0e321 | 2013-07-17 14:40:12 -0700 | [diff] [blame] | 336 | #endif // ART_RUNTIME_ATOMIC_H_ |