blob: e2057ed0731d3bf4715a458cf6e0aece595620a7 [file] [log] [blame]
Ben Murdochb0fe1622011-05-05 13:52:32 +01001// Copyright 2010 the V8 project authors. All rights reserved.
2// Redistribution and use in source and binary forms, with or without
3// modification, are permitted provided that the following conditions are
4// met:
5//
6// * Redistributions of source code must retain the above copyright
7// notice, this list of conditions and the following disclaimer.
8// * Redistributions in binary form must reproduce the above
9// copyright notice, this list of conditions and the following
10// disclaimer in the documentation and/or other materials provided
11// with the distribution.
12// * Neither the name of Google Inc. nor the names of its
13// contributors may be used to endorse or promote products derived
14// from this software without specific prior written permission.
15//
16// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
28// The routines exported by this module are subtle. If you use them, even if
29// you get the code right, it will depend on careful reasoning about atomicity
30// and memory ordering; it will be less readable, and harder to maintain. If
31// you plan to use these routines, you should have a good reason, such as solid
32// evidence that performance would otherwise suffer, or there being no
33// alternative. You should assume only properties explicitly guaranteed by the
34// specifications in this file. You are almost certainly _not_ writing code
35// just for the x86; if you assume x86 semantics, x86 hardware bugs and
36// implementations on other archtectures will cause your code to break. If you
37// do not know what you are doing, avoid these routines, and use a Mutex.
38//
39// It is incorrect to make direct assignments to/from an atomic variable.
40// You should use one of the Load or Store routines. The NoBarrier
41// versions are provided when no barriers are needed:
42// NoBarrier_Store()
43// NoBarrier_Load()
44// Although there are currently no compiler enforcement, you are encouraged
45// to use these.
46//
47
48#ifndef V8_ATOMICOPS_H_
49#define V8_ATOMICOPS_H_
50
51#include "../include/v8.h"
52#include "globals.h"
53
54namespace v8 {
55namespace internal {
56
57typedef int32_t Atomic32;
58#ifdef V8_HOST_ARCH_64_BIT
59// We need to be able to go between Atomic64 and AtomicWord implicitly. This
60// means Atomic64 and AtomicWord should be the same type on 64-bit.
61#if defined(__APPLE__)
62// MacOS is an exception to the implicit conversion rule above,
63// because it uses long for intptr_t.
64typedef int64_t Atomic64;
65#else
66typedef intptr_t Atomic64;
67#endif
68#endif
69
70// Use AtomicWord for a machine-sized pointer. It will use the Atomic32 or
71// Atomic64 routines below, depending on your architecture.
72typedef intptr_t AtomicWord;
73
74// Atomically execute:
75// result = *ptr;
76// if (*ptr == old_value)
77// *ptr = new_value;
78// return result;
79//
80// I.e., replace "*ptr" with "new_value" if "*ptr" used to be "old_value".
81// Always return the old value of "*ptr"
82//
83// This routine implies no memory barriers.
84Atomic32 NoBarrier_CompareAndSwap(volatile Atomic32* ptr,
85 Atomic32 old_value,
86 Atomic32 new_value);
87
88// Atomically store new_value into *ptr, returning the previous value held in
89// *ptr. This routine implies no memory barriers.
90Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, Atomic32 new_value);
91
92// Atomically increment *ptr by "increment". Returns the new value of
93// *ptr with the increment applied. This routine implies no memory barriers.
94Atomic32 NoBarrier_AtomicIncrement(volatile Atomic32* ptr, Atomic32 increment);
95
96Atomic32 Barrier_AtomicIncrement(volatile Atomic32* ptr,
97 Atomic32 increment);
98
99// These following lower-level operations are typically useful only to people
100// implementing higher-level synchronization operations like spinlocks,
101// mutexes, and condition-variables. They combine CompareAndSwap(), a load, or
102// a store with appropriate memory-ordering instructions. "Acquire" operations
103// ensure that no later memory access can be reordered ahead of the operation.
104// "Release" operations ensure that no previous memory access can be reordered
105// after the operation. "Barrier" operations have both "Acquire" and "Release"
106// semantics. A MemoryBarrier() has "Barrier" semantics, but does no memory
107// access.
108Atomic32 Acquire_CompareAndSwap(volatile Atomic32* ptr,
109 Atomic32 old_value,
110 Atomic32 new_value);
111Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr,
112 Atomic32 old_value,
113 Atomic32 new_value);
114
115void MemoryBarrier();
116void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value);
117void Acquire_Store(volatile Atomic32* ptr, Atomic32 value);
118void Release_Store(volatile Atomic32* ptr, Atomic32 value);
119
120Atomic32 NoBarrier_Load(volatile const Atomic32* ptr);
121Atomic32 Acquire_Load(volatile const Atomic32* ptr);
122Atomic32 Release_Load(volatile const Atomic32* ptr);
123
124// 64-bit atomic operations (only available on 64-bit processors).
125#ifdef V8_HOST_ARCH_64_BIT
126Atomic64 NoBarrier_CompareAndSwap(volatile Atomic64* ptr,
127 Atomic64 old_value,
128 Atomic64 new_value);
129Atomic64 NoBarrier_AtomicExchange(volatile Atomic64* ptr, Atomic64 new_value);
130Atomic64 NoBarrier_AtomicIncrement(volatile Atomic64* ptr, Atomic64 increment);
131Atomic64 Barrier_AtomicIncrement(volatile Atomic64* ptr, Atomic64 increment);
132
133Atomic64 Acquire_CompareAndSwap(volatile Atomic64* ptr,
134 Atomic64 old_value,
135 Atomic64 new_value);
136Atomic64 Release_CompareAndSwap(volatile Atomic64* ptr,
137 Atomic64 old_value,
138 Atomic64 new_value);
139void NoBarrier_Store(volatile Atomic64* ptr, Atomic64 value);
140void Acquire_Store(volatile Atomic64* ptr, Atomic64 value);
141void Release_Store(volatile Atomic64* ptr, Atomic64 value);
142Atomic64 NoBarrier_Load(volatile const Atomic64* ptr);
143Atomic64 Acquire_Load(volatile const Atomic64* ptr);
144Atomic64 Release_Load(volatile const Atomic64* ptr);
145#endif // V8_HOST_ARCH_64_BIT
146
147} } // namespace v8::internal
148
149// Include our platform specific implementation.
150#if defined(_MSC_VER) && \
151 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64))
152#include "atomicops_internals_x86_msvc.h"
153#elif defined(__APPLE__) && \
154 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64))
155#include "atomicops_internals_x86_macosx.h"
156#elif defined(__GNUC__) && \
157 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64))
158#include "atomicops_internals_x86_gcc.h"
159#elif defined(__GNUC__) && defined(V8_HOST_ARCH_ARM)
160#include "atomicops_internals_arm_gcc.h"
Steve Block44f0eee2011-05-26 01:26:41 +0100161#elif defined(__GNUC__) && defined(V8_HOST_ARCH_MIPS)
162#include "atomicops_internals_mips_gcc.h"
Ben Murdochb0fe1622011-05-05 13:52:32 +0100163#else
164#error "Atomic operations are not supported on your platform"
165#endif
166
167#endif // V8_ATOMICOPS_H_