blob: 7d72016083fa1cb0b2cf74c44d7f8ae80f7c6c7a [file] [log] [blame]
Antoine Pitrou074e5ed2009-11-10 19:50:40 +00001/*
2 * Implementation of the Global Interpreter Lock (GIL).
3 */
4
5#include <stdlib.h>
6#include <errno.h>
7
8
9/* First some general settings */
10
11/* microseconds (the Python API uses seconds, though) */
12#define DEFAULT_INTERVAL 5000
13static unsigned long gil_interval = DEFAULT_INTERVAL;
14#define INTERVAL (gil_interval >= 1 ? gil_interval : 1)
15
16/* Enable if you want to force the switching of threads at least every `gil_interval` */
17#undef FORCE_SWITCHING
18#define FORCE_SWITCHING
19
20
21/*
22 Notes about the implementation:
23
24 - The GIL is just a boolean variable (gil_locked) whose access is protected
25 by a mutex (gil_mutex), and whose changes are signalled by a condition
26 variable (gil_cond). gil_mutex is taken for short periods of time,
27 and therefore mostly uncontended.
28
29 - In the GIL-holding thread, the main loop (PyEval_EvalFrameEx) must be
30 able to release the GIL on demand by another thread. A volatile boolean
31 variable (gil_drop_request) is used for that purpose, which is checked
32 at every turn of the eval loop. That variable is set after a wait of
33 `interval` microseconds on `gil_cond` has timed out.
34
35 [Actually, another volatile boolean variable (eval_breaker) is used
36 which ORs several conditions into one. Volatile booleans are
37 sufficient as inter-thread signalling means since Python is run
38 on cache-coherent architectures only.]
39
40 - A thread wanting to take the GIL will first let pass a given amount of
41 time (`interval` microseconds) before setting gil_drop_request. This
42 encourages a defined switching period, but doesn't enforce it since
43 opcodes can take an arbitrary time to execute.
44
45 The `interval` value is available for the user to read and modify
46 using the Python API `sys.{get,set}switchinterval()`.
47
48 - When a thread releases the GIL and gil_drop_request is set, that thread
49 ensures that another GIL-awaiting thread gets scheduled.
50 It does so by waiting on a condition variable (switch_cond) until
51 the value of gil_last_holder is changed to something else than its
52 own thread state pointer, indicating that another thread was able to
53 take the GIL.
54
55 This is meant to prohibit the latency-adverse behaviour on multi-core
56 machines where one thread would speculatively release the GIL, but still
57 run and end up being the first to re-acquire it, making the "timeslices"
58 much longer than expected.
59 (Note: this mechanism is enabled with FORCE_SWITCHING above)
60*/
61
62#ifndef _POSIX_THREADS
63/* This means pthreads are not implemented in libc headers, hence the macro
64 not present in unistd.h. But they still can be implemented as an external
65 library (e.g. gnu pth in pthread emulation) */
66# ifdef HAVE_PTHREAD_H
67# include <pthread.h> /* _POSIX_THREADS */
68# endif
69#endif
70
71
72#ifdef _POSIX_THREADS
73
74/*
75 * POSIX support
76 */
77
78#include <pthread.h>
79
80#define ADD_MICROSECONDS(tv, interval) \
81do { \
82 tv.tv_usec += (long) interval; \
83 tv.tv_sec += tv.tv_usec / 1000000; \
84 tv.tv_usec %= 1000000; \
85} while (0)
86
87/* We assume all modern POSIX systems have gettimeofday() */
88#ifdef GETTIMEOFDAY_NO_TZ
89#define GETTIMEOFDAY(ptv) gettimeofday(ptv)
90#else
91#define GETTIMEOFDAY(ptv) gettimeofday(ptv, (struct timezone *)NULL)
92#endif
93
94#define MUTEX_T pthread_mutex_t
95#define MUTEX_INIT(mut) \
96 if (pthread_mutex_init(&mut, NULL)) { \
97 Py_FatalError("pthread_mutex_init(" #mut ") failed"); };
98#define MUTEX_LOCK(mut) \
99 if (pthread_mutex_lock(&mut)) { \
100 Py_FatalError("pthread_mutex_lock(" #mut ") failed"); };
101#define MUTEX_UNLOCK(mut) \
102 if (pthread_mutex_unlock(&mut)) { \
103 Py_FatalError("pthread_mutex_unlock(" #mut ") failed"); };
104
105#define COND_T pthread_cond_t
106#define COND_INIT(cond) \
107 if (pthread_cond_init(&cond, NULL)) { \
108 Py_FatalError("pthread_cond_init(" #cond ") failed"); };
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000109#define COND_SIGNAL(cond) \
110 if (pthread_cond_signal(&cond)) { \
111 Py_FatalError("pthread_cond_signal(" #cond ") failed"); };
112#define COND_WAIT(cond, mut) \
113 if (pthread_cond_wait(&cond, &mut)) { \
114 Py_FatalError("pthread_cond_wait(" #cond ") failed"); };
115#define COND_TIMED_WAIT(cond, mut, microseconds, timeout_result) \
116 { \
117 int r; \
118 struct timespec ts; \
119 struct timeval deadline; \
120 \
121 GETTIMEOFDAY(&deadline); \
122 ADD_MICROSECONDS(deadline, microseconds); \
123 ts.tv_sec = deadline.tv_sec; \
124 ts.tv_nsec = deadline.tv_usec * 1000; \
125 \
126 r = pthread_cond_timedwait(&cond, &mut, &ts); \
127 if (r == ETIMEDOUT) \
128 timeout_result = 1; \
129 else if (r) \
130 Py_FatalError("pthread_cond_timedwait(" #cond ") failed"); \
131 else \
132 timeout_result = 0; \
133 } \
134
135#elif defined(NT_THREADS)
136
137/*
138 * Windows (2000 and later, as well as (hopefully) CE) support
139 */
140
141#include <windows.h>
142
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000143#define MUTEX_T CRITICAL_SECTION
144#define MUTEX_INIT(mut) do { \
145 if (!(InitializeCriticalSectionAndSpinCount(&(mut), 4000))) \
146 Py_FatalError("CreateMutex(" #mut ") failed"); \
147} while (0)
148#define MUTEX_FINI(mut) \
149 DeleteCriticalSection(&(mut))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000150#define MUTEX_LOCK(mut) \
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000151 EnterCriticalSection(&(mut))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000152#define MUTEX_UNLOCK(mut) \
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000153 LeaveCriticalSection(&(mut))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000154
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000155/* We emulate condition variables with a semaphore.
156 We use a Semaphore rather than an auto-reset event, because although
157 an auto-resent event might appear to solve the lost-wakeup bug (race
158 condition between releasing the outer lock and waiting) because it
159 maintains state even though a wait hasn't happened, there is still
160 a lost wakeup problem if more than one thread are interrupted in the
161 critical place. A semaphore solves that.
162 Because it is ok to signal a condition variable with no one
163 waiting, we need to keep track of the number of
164 waiting threads. Otherwise, the semaphore's state could rise
165 without bound.
Antoine Pitroucf4cabb2009-11-11 18:11:36 +0000166
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000167 Generic emulations of the pthread_cond_* API using
Antoine Pitroucf4cabb2009-11-11 18:11:36 +0000168 Win32 functions can be found on the Web.
169 The following read can be edificating (or not):
170 http://www.cse.wustl.edu/~schmidt/win32-cv-1.html
171*/
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000172typedef struct COND_T
173{
174 HANDLE sem; /* the semaphore */
175 int n_waiting; /* how many are unreleased */
176} COND_T;
177
178__inline static void _cond_init(COND_T *cond)
179{
180 /* A semaphore with a large max value, The positive value
181 * is only needed to catch those "lost wakeup" events and
182 * race conditions when a timed wait elapses.
183 */
184 if (!(cond->sem = CreateSemaphore(NULL, 0, 1000, NULL)))
185 Py_FatalError("CreateSemaphore() failed");
186 cond->n_waiting = 0;
187}
188
189__inline static void _cond_fini(COND_T *cond)
190{
191 BOOL ok = CloseHandle(cond->sem);
192 if (!ok)
193 Py_FatalError("CloseHandle() failed");
194}
195
196__inline static void _cond_wait(COND_T *cond, MUTEX_T *mut)
197{
198 ++cond->n_waiting;
199 MUTEX_UNLOCK(*mut);
200 /* "lost wakeup bug" would occur if the caller were interrupted here,
201 * but we are safe because we are using a semaphore wich has an internal
202 * count.
203 */
204 if (WaitForSingleObject(cond->sem, INFINITE) == WAIT_FAILED)
205 Py_FatalError("WaitForSingleObject() failed");
206 MUTEX_LOCK(*mut);
207}
208
209__inline static int _cond_timed_wait(COND_T *cond, MUTEX_T *mut,
210 int us)
211{
212 DWORD r;
213 ++cond->n_waiting;
214 MUTEX_UNLOCK(*mut);
215 r = WaitForSingleObject(cond->sem, us / 1000);
216 if (r == WAIT_FAILED)
217 Py_FatalError("WaitForSingleObject() failed");
218 MUTEX_LOCK(*mut);
219 if (r == WAIT_TIMEOUT)
220 --cond->n_waiting;
221 /* Here we have a benign race condition with _cond_signal. If the
222 * wait operation has timed out, but before we can acquire the
223 * mutex again to decrement n_waiting, a thread holding the mutex
224 * still sees a positive n_waiting value and may call
225 * ReleaseSemaphore and decrement n_waiting.
226 * This will cause n_waiting to be decremented twice.
227 * This is benign, though, because ReleaseSemaphore will also have
228 * been called, leaving the semaphore state positive. We may
229 * thus end up with semaphore in state 1, and n_waiting == -1, and
230 * the next time someone calls _cond_wait(), that thread will
231 * pass right through, decrementing the semaphore state and
232 * incrementing n_waiting, thus correcting the extra _cond_signal.
233 */
234 return r == WAIT_TIMEOUT;
235}
236
237__inline static void _cond_signal(COND_T *cond) {
238 /* NOTE: This must be called with the mutex held */
239 if (cond->n_waiting > 0) {
240 if (!ReleaseSemaphore(cond->sem, 1, NULL))
241 Py_FatalError("ReleaseSemaphore() failed");
242 --cond->n_waiting;
243 }
244}
245
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000246#define COND_INIT(cond) \
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000247 _cond_init(&(cond))
248#define COND_FINI(cond) \
249 _cond_fini(&(cond))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000250#define COND_SIGNAL(cond) \
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000251 _cond_signal(&(cond))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000252#define COND_WAIT(cond, mut) \
Antoine Pitroue1dd1742010-08-10 13:48:51 +0000253 _cond_wait(&(cond), &(mut))
254#define COND_TIMED_WAIT(cond, mut, us, timeout_result) do { \
255 (timeout_result) = _cond_timed_wait(&(cond), &(mut), us); \
256} while (0)
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000257
258#else
259
260#error You need either a POSIX-compatible or a Windows system!
261
262#endif /* _POSIX_THREADS, NT_THREADS */
263
264
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000265/* Whether the GIL is already taken (-1 if uninitialized). This is atomic
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000266 because it can be read without any lock taken in ceval.c. */
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000267static _Py_atomic_int gil_locked = {-1};
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000268/* Number of GIL switches since the beginning. */
269static unsigned long gil_switch_number = 0;
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000270/* Last PyThreadState holding / having held the GIL. This helps us know
271 whether anyone else was scheduled after we dropped the GIL. */
272static _Py_atomic_address gil_last_holder = {NULL};
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000273
274/* This condition variable allows one or several threads to wait until
275 the GIL is released. In addition, the mutex also protects the above
276 variables. */
277static COND_T gil_cond;
278static MUTEX_T gil_mutex;
279
280#ifdef FORCE_SWITCHING
281/* This condition variable helps the GIL-releasing thread wait for
282 a GIL-awaiting thread to be scheduled and take the GIL. */
283static COND_T switch_cond;
284static MUTEX_T switch_mutex;
285#endif
286
287
288static int gil_created(void)
289{
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000290 return _Py_atomic_load_explicit(&gil_locked, _Py_memory_order_acquire) >= 0;
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000291}
292
293static void create_gil(void)
294{
295 MUTEX_INIT(gil_mutex);
296#ifdef FORCE_SWITCHING
297 MUTEX_INIT(switch_mutex);
298#endif
299 COND_INIT(gil_cond);
300#ifdef FORCE_SWITCHING
301 COND_INIT(switch_cond);
302#endif
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000303 _Py_atomic_store_relaxed(&gil_last_holder, NULL);
304 _Py_ANNOTATE_RWLOCK_CREATE(&gil_locked);
305 _Py_atomic_store_explicit(&gil_locked, 0, _Py_memory_order_release);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000306}
307
308static void recreate_gil(void)
309{
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000310 _Py_ANNOTATE_RWLOCK_DESTROY(&gil_locked);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000311 create_gil();
312}
313
314static void drop_gil(PyThreadState *tstate)
315{
316 /* NOTE: tstate is allowed to be NULL. */
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000317 if (!_Py_atomic_load_relaxed(&gil_locked))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000318 Py_FatalError("drop_gil: GIL is not locked");
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000319 if (tstate != NULL &&
320 tstate != _Py_atomic_load_relaxed(&gil_last_holder))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000321 Py_FatalError("drop_gil: wrong thread state");
322
323 MUTEX_LOCK(gil_mutex);
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000324 _Py_ANNOTATE_RWLOCK_RELEASED(&gil_locked, /*is_write=*/1);
325 _Py_atomic_store_relaxed(&gil_locked, 0);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000326 COND_SIGNAL(gil_cond);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000327 MUTEX_UNLOCK(gil_mutex);
328
329#ifdef FORCE_SWITCHING
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000330 if (_Py_atomic_load_relaxed(&gil_drop_request) && tstate != NULL) {
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000331 MUTEX_LOCK(switch_mutex);
332 /* Not switched yet => wait */
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000333 if (_Py_atomic_load_relaxed(&gil_last_holder) == tstate) {
Antoine Pitroua1d23322009-11-12 22:56:02 +0000334 RESET_GIL_DROP_REQUEST();
Antoine Pitroucf4cabb2009-11-11 18:11:36 +0000335 /* NOTE: if COND_WAIT does not atomically start waiting when
336 releasing the mutex, another thread can run through, take
337 the GIL and drop it again, and reset the condition
Antoine Pitroua1d23322009-11-12 22:56:02 +0000338 before we even had a chance to wait for it. */
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000339 COND_WAIT(switch_cond, switch_mutex);
Antoine Pitroua1d23322009-11-12 22:56:02 +0000340 }
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000341 MUTEX_UNLOCK(switch_mutex);
342 }
343#endif
344}
345
346static void take_gil(PyThreadState *tstate)
347{
348 int err;
349 if (tstate == NULL)
350 Py_FatalError("take_gil: NULL tstate");
351
352 err = errno;
353 MUTEX_LOCK(gil_mutex);
354
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000355 if (!_Py_atomic_load_relaxed(&gil_locked))
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000356 goto _ready;
357
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000358 while (_Py_atomic_load_relaxed(&gil_locked)) {
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000359 int timed_out = 0;
360 unsigned long saved_switchnum;
361
362 saved_switchnum = gil_switch_number;
363 COND_TIMED_WAIT(gil_cond, gil_mutex, INTERVAL, timed_out);
364 /* If we timed out and no switch occurred in the meantime, it is time
365 to ask the GIL-holding thread to drop it. */
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000366 if (timed_out &&
367 _Py_atomic_load_relaxed(&gil_locked) &&
368 gil_switch_number == saved_switchnum) {
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000369 SET_GIL_DROP_REQUEST();
370 }
371 }
372_ready:
373#ifdef FORCE_SWITCHING
374 /* This mutex must be taken before modifying gil_last_holder (see drop_gil()). */
375 MUTEX_LOCK(switch_mutex);
376#endif
377 /* We now hold the GIL */
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000378 _Py_atomic_store_relaxed(&gil_locked, 1);
379 _Py_ANNOTATE_RWLOCK_ACQUIRED(&gil_locked, /*is_write=*/1);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000380
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000381 if (tstate != _Py_atomic_load_relaxed(&gil_last_holder)) {
382 _Py_atomic_store_relaxed(&gil_last_holder, tstate);
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000383 ++gil_switch_number;
384 }
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000385
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000386#ifdef FORCE_SWITCHING
387 COND_SIGNAL(switch_cond);
388 MUTEX_UNLOCK(switch_mutex);
389#endif
Jeffrey Yasskin39370832010-05-03 19:29:34 +0000390 if (_Py_atomic_load_relaxed(&gil_drop_request)) {
Antoine Pitrou074e5ed2009-11-10 19:50:40 +0000391 RESET_GIL_DROP_REQUEST();
392 }
393 if (tstate->async_exc != NULL) {
394 _PyEval_SignalAsyncExc();
395 }
396
397 MUTEX_UNLOCK(gil_mutex);
398 errno = err;
399}
400
401void _PyEval_SetSwitchInterval(unsigned long microseconds)
402{
403 gil_interval = microseconds;
404}
405
406unsigned long _PyEval_GetSwitchInterval()
407{
408 return gil_interval;
409}