blob: d69ac4ecc88b9c0d6ff4d5f97cf0fa9d89b5fcdc [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001#ifndef _LINUX_WAIT_H
2#define _LINUX_WAIT_H
Ingo Molnarfb869b62013-10-04 10:24:49 +02003/*
4 * Linux wait queue related types and methods
5 */
Linus Torvalds1da177e2005-04-16 15:20:36 -07006#include <linux/list.h>
7#include <linux/stddef.h>
8#include <linux/spinlock.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -07009#include <asm/current.h>
David Howells607ca462012-10-13 10:46:48 +010010#include <uapi/linux/wait.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070011
12typedef struct __wait_queue wait_queue_t;
Peter Zijlstra7d478722009-09-14 19:55:44 +020013typedef int (*wait_queue_func_t)(wait_queue_t *wait, unsigned mode, int flags, void *key);
14int default_wake_function(wait_queue_t *wait, unsigned mode, int flags, void *key);
Linus Torvalds1da177e2005-04-16 15:20:36 -070015
Peter Zijlstra61ada522014-09-24 10:18:47 +020016/* __wait_queue::flags */
17#define WQ_FLAG_EXCLUSIVE 0x01
18#define WQ_FLAG_WOKEN 0x02
19
Linus Torvalds1da177e2005-04-16 15:20:36 -070020struct __wait_queue {
Ingo Molnarfb869b62013-10-04 10:24:49 +020021 unsigned int flags;
Ingo Molnarfb869b62013-10-04 10:24:49 +020022 void *private;
23 wait_queue_func_t func;
24 struct list_head task_list;
Linus Torvalds1da177e2005-04-16 15:20:36 -070025};
26
27struct wait_bit_key {
Ingo Molnarfb869b62013-10-04 10:24:49 +020028 void *flags;
29 int bit_nr;
30#define WAIT_ATOMIC_T_BIT_NR -1
NeilBrowncbbce822014-09-25 13:55:19 +100031 unsigned long timeout;
Linus Torvalds1da177e2005-04-16 15:20:36 -070032};
33
34struct wait_bit_queue {
Ingo Molnarfb869b62013-10-04 10:24:49 +020035 struct wait_bit_key key;
36 wait_queue_t wait;
Linus Torvalds1da177e2005-04-16 15:20:36 -070037};
38
39struct __wait_queue_head {
Ingo Molnarfb869b62013-10-04 10:24:49 +020040 spinlock_t lock;
41 struct list_head task_list;
Linus Torvalds1da177e2005-04-16 15:20:36 -070042};
43typedef struct __wait_queue_head wait_queue_head_t;
44
Tim Schmielau8c65b4a2005-11-07 00:59:43 -080045struct task_struct;
Linus Torvalds1da177e2005-04-16 15:20:36 -070046
47/*
48 * Macros for declaration and initialisaton of the datatypes
49 */
50
51#define __WAITQUEUE_INITIALIZER(name, tsk) { \
Benjamin LaHaisec43dc2f2005-06-23 00:10:27 -070052 .private = tsk, \
Linus Torvalds1da177e2005-04-16 15:20:36 -070053 .func = default_wake_function, \
54 .task_list = { NULL, NULL } }
55
56#define DECLARE_WAITQUEUE(name, tsk) \
57 wait_queue_t name = __WAITQUEUE_INITIALIZER(name, tsk)
58
59#define __WAIT_QUEUE_HEAD_INITIALIZER(name) { \
Ingo Molnare4d91912006-07-03 00:24:34 -070060 .lock = __SPIN_LOCK_UNLOCKED(name.lock), \
Linus Torvalds1da177e2005-04-16 15:20:36 -070061 .task_list = { &(name).task_list, &(name).task_list } }
62
63#define DECLARE_WAIT_QUEUE_HEAD(name) \
64 wait_queue_head_t name = __WAIT_QUEUE_HEAD_INITIALIZER(name)
65
66#define __WAIT_BIT_KEY_INITIALIZER(word, bit) \
67 { .flags = word, .bit_nr = bit, }
68
David Howellscb655372013-05-10 19:50:26 +010069#define __WAIT_ATOMIC_T_KEY_INITIALIZER(p) \
70 { .flags = p, .bit_nr = WAIT_ATOMIC_T_BIT_NR, }
71
Peter Zijlstraf07fdec2011-12-13 13:20:54 +010072extern void __init_waitqueue_head(wait_queue_head_t *q, const char *name, struct lock_class_key *);
Peter Zijlstra2fc39112009-08-10 12:33:05 +010073
74#define init_waitqueue_head(q) \
75 do { \
76 static struct lock_class_key __key; \
77 \
Peter Zijlstraf07fdec2011-12-13 13:20:54 +010078 __init_waitqueue_head((q), #q, &__key); \
Peter Zijlstra2fc39112009-08-10 12:33:05 +010079 } while (0)
Linus Torvalds1da177e2005-04-16 15:20:36 -070080
Peter Zijlstra7259f0d2006-10-29 22:46:36 -080081#ifdef CONFIG_LOCKDEP
82# define __WAIT_QUEUE_HEAD_INIT_ONSTACK(name) \
83 ({ init_waitqueue_head(&name); name; })
84# define DECLARE_WAIT_QUEUE_HEAD_ONSTACK(name) \
85 wait_queue_head_t name = __WAIT_QUEUE_HEAD_INIT_ONSTACK(name)
86#else
87# define DECLARE_WAIT_QUEUE_HEAD_ONSTACK(name) DECLARE_WAIT_QUEUE_HEAD(name)
88#endif
89
Linus Torvalds1da177e2005-04-16 15:20:36 -070090static inline void init_waitqueue_entry(wait_queue_t *q, struct task_struct *p)
91{
Ingo Molnarfb869b62013-10-04 10:24:49 +020092 q->flags = 0;
93 q->private = p;
94 q->func = default_wake_function;
Linus Torvalds1da177e2005-04-16 15:20:36 -070095}
96
Ingo Molnarfb869b62013-10-04 10:24:49 +020097static inline void
98init_waitqueue_func_entry(wait_queue_t *q, wait_queue_func_t func)
Linus Torvalds1da177e2005-04-16 15:20:36 -070099{
Ingo Molnarfb869b62013-10-04 10:24:49 +0200100 q->flags = 0;
101 q->private = NULL;
102 q->func = func;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700103}
104
105static inline int waitqueue_active(wait_queue_head_t *q)
106{
107 return !list_empty(&q->task_list);
108}
109
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800110extern void add_wait_queue(wait_queue_head_t *q, wait_queue_t *wait);
111extern void add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t *wait);
112extern void remove_wait_queue(wait_queue_head_t *q, wait_queue_t *wait);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700113
114static inline void __add_wait_queue(wait_queue_head_t *head, wait_queue_t *new)
115{
116 list_add(&new->task_list, &head->task_list);
117}
118
119/*
120 * Used for wake-one threads:
121 */
Ingo Molnarfb869b62013-10-04 10:24:49 +0200122static inline void
123__add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t *wait)
Changli Gaoa93d2f12010-05-07 14:33:26 +0800124{
125 wait->flags |= WQ_FLAG_EXCLUSIVE;
126 __add_wait_queue(q, wait);
127}
128
Linus Torvalds1da177e2005-04-16 15:20:36 -0700129static inline void __add_wait_queue_tail(wait_queue_head_t *head,
Changli Gaoa93d2f12010-05-07 14:33:26 +0800130 wait_queue_t *new)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700131{
132 list_add_tail(&new->task_list, &head->task_list);
133}
134
Ingo Molnarfb869b62013-10-04 10:24:49 +0200135static inline void
136__add_wait_queue_tail_exclusive(wait_queue_head_t *q, wait_queue_t *wait)
Changli Gaoa93d2f12010-05-07 14:33:26 +0800137{
138 wait->flags |= WQ_FLAG_EXCLUSIVE;
139 __add_wait_queue_tail(q, wait);
140}
141
Ingo Molnarfb869b62013-10-04 10:24:49 +0200142static inline void
143__remove_wait_queue(wait_queue_head_t *head, wait_queue_t *old)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700144{
145 list_del(&old->task_list);
146}
147
NeilBrownc1221322014-07-07 15:16:04 +1000148typedef int wait_bit_action_f(struct wait_bit_key *);
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800149void __wake_up(wait_queue_head_t *q, unsigned int mode, int nr, void *key);
Davide Libenzi4ede8162009-03-31 15:24:20 -0700150void __wake_up_locked_key(wait_queue_head_t *q, unsigned int mode, void *key);
Ingo Molnarfb869b62013-10-04 10:24:49 +0200151void __wake_up_sync_key(wait_queue_head_t *q, unsigned int mode, int nr, void *key);
Thomas Gleixner63b20012011-12-01 00:04:00 +0100152void __wake_up_locked(wait_queue_head_t *q, unsigned int mode, int nr);
Davide Libenzi4ede8162009-03-31 15:24:20 -0700153void __wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr);
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800154void __wake_up_bit(wait_queue_head_t *, void *, int);
NeilBrownc1221322014-07-07 15:16:04 +1000155int __wait_on_bit(wait_queue_head_t *, struct wait_bit_queue *, wait_bit_action_f *, unsigned);
156int __wait_on_bit_lock(wait_queue_head_t *, struct wait_bit_queue *, wait_bit_action_f *, unsigned);
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800157void wake_up_bit(void *, int);
David Howellscb655372013-05-10 19:50:26 +0100158void wake_up_atomic_t(atomic_t *);
NeilBrownc1221322014-07-07 15:16:04 +1000159int out_of_line_wait_on_bit(void *, int, wait_bit_action_f *, unsigned);
NeilBrowncbbce822014-09-25 13:55:19 +1000160int out_of_line_wait_on_bit_timeout(void *, int, wait_bit_action_f *, unsigned, unsigned long);
NeilBrownc1221322014-07-07 15:16:04 +1000161int out_of_line_wait_on_bit_lock(void *, int, wait_bit_action_f *, unsigned);
David Howellscb655372013-05-10 19:50:26 +0100162int out_of_line_wait_on_atomic_t(atomic_t *, int (*)(atomic_t *), unsigned);
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800163wait_queue_head_t *bit_waitqueue(void *, int);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700164
Matthew Wilcoxe64d66c2007-12-06 17:34:36 -0500165#define wake_up(x) __wake_up(x, TASK_NORMAL, 1, NULL)
166#define wake_up_nr(x, nr) __wake_up(x, TASK_NORMAL, nr, NULL)
167#define wake_up_all(x) __wake_up(x, TASK_NORMAL, 0, NULL)
Thomas Gleixner63b20012011-12-01 00:04:00 +0100168#define wake_up_locked(x) __wake_up_locked((x), TASK_NORMAL, 1)
169#define wake_up_all_locked(x) __wake_up_locked((x), TASK_NORMAL, 0)
Matthew Wilcoxe64d66c2007-12-06 17:34:36 -0500170
Linus Torvalds1da177e2005-04-16 15:20:36 -0700171#define wake_up_interruptible(x) __wake_up(x, TASK_INTERRUPTIBLE, 1, NULL)
172#define wake_up_interruptible_nr(x, nr) __wake_up(x, TASK_INTERRUPTIBLE, nr, NULL)
173#define wake_up_interruptible_all(x) __wake_up(x, TASK_INTERRUPTIBLE, 0, NULL)
Matthew Wilcoxe64d66c2007-12-06 17:34:36 -0500174#define wake_up_interruptible_sync(x) __wake_up_sync((x), TASK_INTERRUPTIBLE, 1)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700175
Peter Zijlstra0ccf8312008-02-04 22:27:20 -0800176/*
Davide Libenzic0da3772009-03-31 15:24:20 -0700177 * Wakeup macros to be used to report events to the targets.
Peter Zijlstra0ccf8312008-02-04 22:27:20 -0800178 */
Ingo Molnarfb869b62013-10-04 10:24:49 +0200179#define wake_up_poll(x, m) \
Davide Libenzic0da3772009-03-31 15:24:20 -0700180 __wake_up(x, TASK_NORMAL, 1, (void *) (m))
Ingo Molnarfb869b62013-10-04 10:24:49 +0200181#define wake_up_locked_poll(x, m) \
Davide Libenzic0da3772009-03-31 15:24:20 -0700182 __wake_up_locked_key((x), TASK_NORMAL, (void *) (m))
Ingo Molnarfb869b62013-10-04 10:24:49 +0200183#define wake_up_interruptible_poll(x, m) \
Davide Libenzic0da3772009-03-31 15:24:20 -0700184 __wake_up(x, TASK_INTERRUPTIBLE, 1, (void *) (m))
185#define wake_up_interruptible_sync_poll(x, m) \
186 __wake_up_sync_key((x), TASK_INTERRUPTIBLE, 1, (void *) (m))
Peter Zijlstra0ccf8312008-02-04 22:27:20 -0800187
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200188#define ___wait_cond_timeout(condition) \
Peter Zijlstra2953ef22013-10-02 11:22:19 +0200189({ \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200190 bool __cond = (condition); \
191 if (__cond && !__ret) \
192 __ret = 1; \
193 __cond || !__ret; \
Peter Zijlstra2953ef22013-10-02 11:22:19 +0200194})
195
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200196#define ___wait_is_interruptible(state) \
197 (!__builtin_constant_p(state) || \
198 state == TASK_INTERRUPTIBLE || state == TASK_KILLABLE) \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200199
Peter Zijlstra8b322012014-04-18 15:07:17 -0700200/*
201 * The below macro ___wait_event() has an explicit shadow of the __ret
202 * variable when used from the wait_event_*() macros.
203 *
204 * This is so that both can use the ___wait_cond_timeout() construct
205 * to wrap the condition.
206 *
207 * The type inconsistency of the wait_event_*() __ret variable is also
208 * on purpose; we use long where we can return timeout values and int
209 * otherwise.
210 */
211
Peter Zijlstra41a14312013-10-02 11:22:21 +0200212#define ___wait_event(wq, condition, state, exclusive, ret, cmd) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200213({ \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200214 __label__ __out; \
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200215 wait_queue_t __wait; \
Peter Zijlstra8b322012014-04-18 15:07:17 -0700216 long __ret = ret; /* explicit shadow */ \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200217 \
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200218 INIT_LIST_HEAD(&__wait.task_list); \
219 if (exclusive) \
220 __wait.flags = WQ_FLAG_EXCLUSIVE; \
221 else \
222 __wait.flags = 0; \
223 \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200224 for (;;) { \
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200225 long __int = prepare_to_wait_event(&wq, &__wait, state);\
Peter Zijlstra41a14312013-10-02 11:22:21 +0200226 \
227 if (condition) \
228 break; \
229 \
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200230 if (___wait_is_interruptible(state) && __int) { \
231 __ret = __int; \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200232 if (exclusive) { \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200233 abort_exclusive_wait(&wq, &__wait, \
234 state, NULL); \
Peter Zijlstra41a14312013-10-02 11:22:21 +0200235 goto __out; \
236 } \
237 break; \
238 } \
239 \
240 cmd; \
241 } \
242 finish_wait(&wq, &__wait); \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200243__out: __ret; \
244})
Peter Zijlstra41a14312013-10-02 11:22:21 +0200245
Ingo Molnarfb869b62013-10-04 10:24:49 +0200246#define __wait_event(wq, condition) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200247 (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \
248 schedule())
Linus Torvalds1da177e2005-04-16 15:20:36 -0700249
250/**
251 * wait_event - sleep until a condition gets true
252 * @wq: the waitqueue to wait on
253 * @condition: a C expression for the event to wait for
254 *
255 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
256 * @condition evaluates to true. The @condition is checked each time
257 * the waitqueue @wq is woken up.
258 *
259 * wake_up() has to be called after changing any variable that could
260 * change the result of the wait condition.
261 */
Ingo Molnarfb869b62013-10-04 10:24:49 +0200262#define wait_event(wq, condition) \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700263do { \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200264 might_sleep(); \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200265 if (condition) \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700266 break; \
267 __wait_event(wq, condition); \
268} while (0)
269
Peter Zijlstra2c561242015-02-03 12:55:31 +0100270#define __io_wait_event(wq, condition) \
271 (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \
272 io_schedule())
273
274/*
275 * io_wait_event() -- like wait_event() but with io_schedule()
276 */
277#define io_wait_event(wq, condition) \
278do { \
279 might_sleep(); \
280 if (condition) \
281 break; \
282 __io_wait_event(wq, condition); \
283} while (0)
284
Peter Zijlstra36df04b2014-10-29 12:21:57 +0100285#define __wait_event_freezable(wq, condition) \
286 ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 0, 0, \
287 schedule(); try_to_freeze())
288
289/**
290 * wait_event - sleep (or freeze) until a condition gets true
291 * @wq: the waitqueue to wait on
292 * @condition: a C expression for the event to wait for
293 *
294 * The process is put to sleep (TASK_INTERRUPTIBLE -- so as not to contribute
295 * to system load) until the @condition evaluates to true. The
296 * @condition is checked each time the waitqueue @wq is woken up.
297 *
298 * wake_up() has to be called after changing any variable that could
299 * change the result of the wait condition.
300 */
301#define wait_event_freezable(wq, condition) \
302({ \
303 int __ret = 0; \
304 might_sleep(); \
305 if (!(condition)) \
306 __ret = __wait_event_freezable(wq, condition); \
307 __ret; \
308})
309
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200310#define __wait_event_timeout(wq, condition, timeout) \
311 ___wait_event(wq, ___wait_cond_timeout(condition), \
312 TASK_UNINTERRUPTIBLE, 0, timeout, \
313 __ret = schedule_timeout(__ret))
Linus Torvalds1da177e2005-04-16 15:20:36 -0700314
315/**
316 * wait_event_timeout - sleep until a condition gets true or a timeout elapses
317 * @wq: the waitqueue to wait on
318 * @condition: a C expression for the event to wait for
319 * @timeout: timeout, in jiffies
320 *
321 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
322 * @condition evaluates to true. The @condition is checked each time
323 * the waitqueue @wq is woken up.
324 *
325 * wake_up() has to be called after changing any variable that could
326 * change the result of the wait condition.
327 *
Scot Doyle6b44f512014-08-24 17:12:27 +0000328 * Returns:
329 * 0 if the @condition evaluated to %false after the @timeout elapsed,
330 * 1 if the @condition evaluated to %true after the @timeout elapsed,
331 * or the remaining jiffies (at least 1) if the @condition evaluated
332 * to %true before the @timeout elapsed.
Linus Torvalds1da177e2005-04-16 15:20:36 -0700333 */
334#define wait_event_timeout(wq, condition, timeout) \
335({ \
336 long __ret = timeout; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200337 might_sleep(); \
Oleg Nesterov89229152013-10-07 20:31:06 +0200338 if (!___wait_cond_timeout(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200339 __ret = __wait_event_timeout(wq, condition, timeout); \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700340 __ret; \
341})
342
Peter Zijlstra36df04b2014-10-29 12:21:57 +0100343#define __wait_event_freezable_timeout(wq, condition, timeout) \
344 ___wait_event(wq, ___wait_cond_timeout(condition), \
345 TASK_INTERRUPTIBLE, 0, timeout, \
346 __ret = schedule_timeout(__ret); try_to_freeze())
347
348/*
349 * like wait_event_timeout() -- except it uses TASK_INTERRUPTIBLE to avoid
350 * increasing load and is freezable.
351 */
352#define wait_event_freezable_timeout(wq, condition, timeout) \
353({ \
354 long __ret = timeout; \
355 might_sleep(); \
356 if (!___wait_cond_timeout(condition)) \
357 __ret = __wait_event_freezable_timeout(wq, condition, timeout); \
358 __ret; \
359})
360
Shaohua Li82e06c82013-11-14 15:16:16 +1100361#define __wait_event_cmd(wq, condition, cmd1, cmd2) \
362 (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \
363 cmd1; schedule(); cmd2)
364
365/**
366 * wait_event_cmd - sleep until a condition gets true
367 * @wq: the waitqueue to wait on
368 * @condition: a C expression for the event to wait for
Masanari Iidaf434f7a2014-01-22 01:22:06 +0900369 * @cmd1: the command will be executed before sleep
370 * @cmd2: the command will be executed after sleep
Shaohua Li82e06c82013-11-14 15:16:16 +1100371 *
372 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
373 * @condition evaluates to true. The @condition is checked each time
374 * the waitqueue @wq is woken up.
375 *
376 * wake_up() has to be called after changing any variable that could
377 * change the result of the wait condition.
378 */
379#define wait_event_cmd(wq, condition, cmd1, cmd2) \
380do { \
381 if (condition) \
382 break; \
383 __wait_event_cmd(wq, condition, cmd1, cmd2); \
384} while (0)
385
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200386#define __wait_event_interruptible(wq, condition) \
387 ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 0, 0, \
Peter Zijlstraf13f4c42013-10-02 11:22:24 +0200388 schedule())
Linus Torvalds1da177e2005-04-16 15:20:36 -0700389
390/**
391 * wait_event_interruptible - sleep until a condition gets true
392 * @wq: the waitqueue to wait on
393 * @condition: a C expression for the event to wait for
394 *
395 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
396 * @condition evaluates to true or a signal is received.
397 * The @condition is checked each time the waitqueue @wq is woken up.
398 *
399 * wake_up() has to be called after changing any variable that could
400 * change the result of the wait condition.
401 *
402 * The function will return -ERESTARTSYS if it was interrupted by a
403 * signal and 0 if @condition evaluated to true.
404 */
405#define wait_event_interruptible(wq, condition) \
406({ \
407 int __ret = 0; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200408 might_sleep(); \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700409 if (!(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200410 __ret = __wait_event_interruptible(wq, condition); \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700411 __ret; \
412})
413
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200414#define __wait_event_interruptible_timeout(wq, condition, timeout) \
415 ___wait_event(wq, ___wait_cond_timeout(condition), \
416 TASK_INTERRUPTIBLE, 0, timeout, \
417 __ret = schedule_timeout(__ret))
Linus Torvalds1da177e2005-04-16 15:20:36 -0700418
419/**
420 * wait_event_interruptible_timeout - sleep until a condition gets true or a timeout elapses
421 * @wq: the waitqueue to wait on
422 * @condition: a C expression for the event to wait for
423 * @timeout: timeout, in jiffies
424 *
425 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
426 * @condition evaluates to true or a signal is received.
427 * The @condition is checked each time the waitqueue @wq is woken up.
428 *
429 * wake_up() has to be called after changing any variable that could
430 * change the result of the wait condition.
431 *
Imre Deak4c663cf2013-05-24 15:55:09 -0700432 * Returns:
Scot Doyle6b44f512014-08-24 17:12:27 +0000433 * 0 if the @condition evaluated to %false after the @timeout elapsed,
434 * 1 if the @condition evaluated to %true after the @timeout elapsed,
435 * the remaining jiffies (at least 1) if the @condition evaluated
436 * to %true before the @timeout elapsed, or -%ERESTARTSYS if it was
437 * interrupted by a signal.
Linus Torvalds1da177e2005-04-16 15:20:36 -0700438 */
439#define wait_event_interruptible_timeout(wq, condition, timeout) \
440({ \
441 long __ret = timeout; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200442 might_sleep(); \
Oleg Nesterov89229152013-10-07 20:31:06 +0200443 if (!___wait_cond_timeout(condition)) \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200444 __ret = __wait_event_interruptible_timeout(wq, \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200445 condition, timeout); \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700446 __ret; \
447})
448
Kent Overstreet774a08b2013-05-07 16:18:43 -0700449#define __wait_event_hrtimeout(wq, condition, timeout, state) \
450({ \
451 int __ret = 0; \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700452 struct hrtimer_sleeper __t; \
453 \
454 hrtimer_init_on_stack(&__t.timer, CLOCK_MONOTONIC, \
455 HRTIMER_MODE_REL); \
456 hrtimer_init_sleeper(&__t, current); \
457 if ((timeout).tv64 != KTIME_MAX) \
458 hrtimer_start_range_ns(&__t.timer, timeout, \
459 current->timer_slack_ns, \
460 HRTIMER_MODE_REL); \
461 \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200462 __ret = ___wait_event(wq, condition, state, 0, 0, \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700463 if (!__t.task) { \
464 __ret = -ETIME; \
465 break; \
466 } \
Peter Zijlstraebdc1952013-10-02 11:22:32 +0200467 schedule()); \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700468 \
469 hrtimer_cancel(&__t.timer); \
470 destroy_hrtimer_on_stack(&__t.timer); \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700471 __ret; \
472})
473
474/**
475 * wait_event_hrtimeout - sleep until a condition gets true or a timeout elapses
476 * @wq: the waitqueue to wait on
477 * @condition: a C expression for the event to wait for
478 * @timeout: timeout, as a ktime_t
479 *
480 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
481 * @condition evaluates to true or a signal is received.
482 * The @condition is checked each time the waitqueue @wq is woken up.
483 *
484 * wake_up() has to be called after changing any variable that could
485 * change the result of the wait condition.
486 *
487 * The function returns 0 if @condition became true, or -ETIME if the timeout
488 * elapsed.
489 */
490#define wait_event_hrtimeout(wq, condition, timeout) \
491({ \
492 int __ret = 0; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200493 might_sleep(); \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700494 if (!(condition)) \
495 __ret = __wait_event_hrtimeout(wq, condition, timeout, \
496 TASK_UNINTERRUPTIBLE); \
497 __ret; \
498})
499
500/**
501 * wait_event_interruptible_hrtimeout - sleep until a condition gets true or a timeout elapses
502 * @wq: the waitqueue to wait on
503 * @condition: a C expression for the event to wait for
504 * @timeout: timeout, as a ktime_t
505 *
506 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
507 * @condition evaluates to true or a signal is received.
508 * The @condition is checked each time the waitqueue @wq is woken up.
509 *
510 * wake_up() has to be called after changing any variable that could
511 * change the result of the wait condition.
512 *
513 * The function returns 0 if @condition became true, -ERESTARTSYS if it was
514 * interrupted by a signal, or -ETIME if the timeout elapsed.
515 */
516#define wait_event_interruptible_hrtimeout(wq, condition, timeout) \
517({ \
518 long __ret = 0; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200519 might_sleep(); \
Kent Overstreet774a08b2013-05-07 16:18:43 -0700520 if (!(condition)) \
521 __ret = __wait_event_hrtimeout(wq, condition, timeout, \
522 TASK_INTERRUPTIBLE); \
523 __ret; \
524})
525
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200526#define __wait_event_interruptible_exclusive(wq, condition) \
527 ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 1, 0, \
Peter Zijlstra48c25212013-10-02 11:22:26 +0200528 schedule())
Linus Torvalds1da177e2005-04-16 15:20:36 -0700529
530#define wait_event_interruptible_exclusive(wq, condition) \
531({ \
532 int __ret = 0; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200533 might_sleep(); \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700534 if (!(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200535 __ret = __wait_event_interruptible_exclusive(wq, condition);\
Linus Torvalds1da177e2005-04-16 15:20:36 -0700536 __ret; \
537})
538
Michal Nazarewicz22c43c82010-05-05 12:53:11 +0200539
Peter Zijlstra36df04b2014-10-29 12:21:57 +0100540#define __wait_event_freezable_exclusive(wq, condition) \
541 ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 1, 0, \
542 schedule(); try_to_freeze())
543
544#define wait_event_freezable_exclusive(wq, condition) \
545({ \
546 int __ret = 0; \
547 might_sleep(); \
548 if (!(condition)) \
549 __ret = __wait_event_freezable_exclusive(wq, condition);\
550 __ret; \
551})
552
553
Michal Nazarewicz22c43c82010-05-05 12:53:11 +0200554#define __wait_event_interruptible_locked(wq, condition, exclusive, irq) \
555({ \
556 int __ret = 0; \
557 DEFINE_WAIT(__wait); \
558 if (exclusive) \
559 __wait.flags |= WQ_FLAG_EXCLUSIVE; \
560 do { \
561 if (likely(list_empty(&__wait.task_list))) \
562 __add_wait_queue_tail(&(wq), &__wait); \
563 set_current_state(TASK_INTERRUPTIBLE); \
564 if (signal_pending(current)) { \
565 __ret = -ERESTARTSYS; \
566 break; \
567 } \
568 if (irq) \
569 spin_unlock_irq(&(wq).lock); \
570 else \
571 spin_unlock(&(wq).lock); \
572 schedule(); \
573 if (irq) \
574 spin_lock_irq(&(wq).lock); \
575 else \
576 spin_lock(&(wq).lock); \
577 } while (!(condition)); \
578 __remove_wait_queue(&(wq), &__wait); \
579 __set_current_state(TASK_RUNNING); \
580 __ret; \
581})
582
583
584/**
585 * wait_event_interruptible_locked - sleep until a condition gets true
586 * @wq: the waitqueue to wait on
587 * @condition: a C expression for the event to wait for
588 *
589 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
590 * @condition evaluates to true or a signal is received.
591 * The @condition is checked each time the waitqueue @wq is woken up.
592 *
593 * It must be called with wq.lock being held. This spinlock is
594 * unlocked while sleeping but @condition testing is done while lock
595 * is held and when this macro exits the lock is held.
596 *
597 * The lock is locked/unlocked using spin_lock()/spin_unlock()
598 * functions which must match the way they are locked/unlocked outside
599 * of this macro.
600 *
601 * wake_up_locked() has to be called after changing any variable that could
602 * change the result of the wait condition.
603 *
604 * The function will return -ERESTARTSYS if it was interrupted by a
605 * signal and 0 if @condition evaluated to true.
606 */
607#define wait_event_interruptible_locked(wq, condition) \
608 ((condition) \
609 ? 0 : __wait_event_interruptible_locked(wq, condition, 0, 0))
610
611/**
612 * wait_event_interruptible_locked_irq - sleep until a condition gets true
613 * @wq: the waitqueue to wait on
614 * @condition: a C expression for the event to wait for
615 *
616 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
617 * @condition evaluates to true or a signal is received.
618 * The @condition is checked each time the waitqueue @wq is woken up.
619 *
620 * It must be called with wq.lock being held. This spinlock is
621 * unlocked while sleeping but @condition testing is done while lock
622 * is held and when this macro exits the lock is held.
623 *
624 * The lock is locked/unlocked using spin_lock_irq()/spin_unlock_irq()
625 * functions which must match the way they are locked/unlocked outside
626 * of this macro.
627 *
628 * wake_up_locked() has to be called after changing any variable that could
629 * change the result of the wait condition.
630 *
631 * The function will return -ERESTARTSYS if it was interrupted by a
632 * signal and 0 if @condition evaluated to true.
633 */
634#define wait_event_interruptible_locked_irq(wq, condition) \
635 ((condition) \
636 ? 0 : __wait_event_interruptible_locked(wq, condition, 0, 1))
637
638/**
639 * wait_event_interruptible_exclusive_locked - sleep exclusively until a condition gets true
640 * @wq: the waitqueue to wait on
641 * @condition: a C expression for the event to wait for
642 *
643 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
644 * @condition evaluates to true or a signal is received.
645 * The @condition is checked each time the waitqueue @wq is woken up.
646 *
647 * It must be called with wq.lock being held. This spinlock is
648 * unlocked while sleeping but @condition testing is done while lock
649 * is held and when this macro exits the lock is held.
650 *
651 * The lock is locked/unlocked using spin_lock()/spin_unlock()
652 * functions which must match the way they are locked/unlocked outside
653 * of this macro.
654 *
655 * The process is put on the wait queue with an WQ_FLAG_EXCLUSIVE flag
656 * set thus when other process waits process on the list if this
657 * process is awaken further processes are not considered.
658 *
659 * wake_up_locked() has to be called after changing any variable that could
660 * change the result of the wait condition.
661 *
662 * The function will return -ERESTARTSYS if it was interrupted by a
663 * signal and 0 if @condition evaluated to true.
664 */
665#define wait_event_interruptible_exclusive_locked(wq, condition) \
666 ((condition) \
667 ? 0 : __wait_event_interruptible_locked(wq, condition, 1, 0))
668
669/**
670 * wait_event_interruptible_exclusive_locked_irq - sleep until a condition gets true
671 * @wq: the waitqueue to wait on
672 * @condition: a C expression for the event to wait for
673 *
674 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
675 * @condition evaluates to true or a signal is received.
676 * The @condition is checked each time the waitqueue @wq is woken up.
677 *
678 * It must be called with wq.lock being held. This spinlock is
679 * unlocked while sleeping but @condition testing is done while lock
680 * is held and when this macro exits the lock is held.
681 *
682 * The lock is locked/unlocked using spin_lock_irq()/spin_unlock_irq()
683 * functions which must match the way they are locked/unlocked outside
684 * of this macro.
685 *
686 * The process is put on the wait queue with an WQ_FLAG_EXCLUSIVE flag
687 * set thus when other process waits process on the list if this
688 * process is awaken further processes are not considered.
689 *
690 * wake_up_locked() has to be called after changing any variable that could
691 * change the result of the wait condition.
692 *
693 * The function will return -ERESTARTSYS if it was interrupted by a
694 * signal and 0 if @condition evaluated to true.
695 */
696#define wait_event_interruptible_exclusive_locked_irq(wq, condition) \
697 ((condition) \
698 ? 0 : __wait_event_interruptible_locked(wq, condition, 1, 1))
699
700
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200701#define __wait_event_killable(wq, condition) \
702 ___wait_event(wq, condition, TASK_KILLABLE, 0, 0, schedule())
Matthew Wilcox1411d5a2007-12-06 12:00:00 -0500703
704/**
705 * wait_event_killable - sleep until a condition gets true
706 * @wq: the waitqueue to wait on
707 * @condition: a C expression for the event to wait for
708 *
709 * The process is put to sleep (TASK_KILLABLE) until the
710 * @condition evaluates to true or a signal is received.
711 * The @condition is checked each time the waitqueue @wq is woken up.
712 *
713 * wake_up() has to be called after changing any variable that could
714 * change the result of the wait condition.
715 *
716 * The function will return -ERESTARTSYS if it was interrupted by a
717 * signal and 0 if @condition evaluated to true.
718 */
719#define wait_event_killable(wq, condition) \
720({ \
721 int __ret = 0; \
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200722 might_sleep(); \
Matthew Wilcox1411d5a2007-12-06 12:00:00 -0500723 if (!(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200724 __ret = __wait_event_killable(wq, condition); \
Matthew Wilcox1411d5a2007-12-06 12:00:00 -0500725 __ret; \
726})
727
Lukas Czernereed8c022012-11-30 11:42:40 +0100728
729#define __wait_event_lock_irq(wq, condition, lock, cmd) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200730 (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \
731 spin_unlock_irq(&lock); \
732 cmd; \
733 schedule(); \
734 spin_lock_irq(&lock))
Lukas Czernereed8c022012-11-30 11:42:40 +0100735
736/**
737 * wait_event_lock_irq_cmd - sleep until a condition gets true. The
738 * condition is checked under the lock. This
739 * is expected to be called with the lock
740 * taken.
741 * @wq: the waitqueue to wait on
742 * @condition: a C expression for the event to wait for
743 * @lock: a locked spinlock_t, which will be released before cmd
744 * and schedule() and reacquired afterwards.
745 * @cmd: a command which is invoked outside the critical section before
746 * sleep
747 *
748 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
749 * @condition evaluates to true. The @condition is checked each time
750 * the waitqueue @wq is woken up.
751 *
752 * wake_up() has to be called after changing any variable that could
753 * change the result of the wait condition.
754 *
755 * This is supposed to be called while holding the lock. The lock is
756 * dropped before invoking the cmd and going to sleep and is reacquired
757 * afterwards.
758 */
759#define wait_event_lock_irq_cmd(wq, condition, lock, cmd) \
760do { \
761 if (condition) \
762 break; \
763 __wait_event_lock_irq(wq, condition, lock, cmd); \
764} while (0)
765
766/**
767 * wait_event_lock_irq - sleep until a condition gets true. The
768 * condition is checked under the lock. This
769 * is expected to be called with the lock
770 * taken.
771 * @wq: the waitqueue to wait on
772 * @condition: a C expression for the event to wait for
773 * @lock: a locked spinlock_t, which will be released before schedule()
774 * and reacquired afterwards.
775 *
776 * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the
777 * @condition evaluates to true. The @condition is checked each time
778 * the waitqueue @wq is woken up.
779 *
780 * wake_up() has to be called after changing any variable that could
781 * change the result of the wait condition.
782 *
783 * This is supposed to be called while holding the lock. The lock is
784 * dropped before going to sleep and is reacquired afterwards.
785 */
786#define wait_event_lock_irq(wq, condition, lock) \
787do { \
788 if (condition) \
789 break; \
790 __wait_event_lock_irq(wq, condition, lock, ); \
791} while (0)
792
793
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200794#define __wait_event_interruptible_lock_irq(wq, condition, lock, cmd) \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200795 ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 0, 0, \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200796 spin_unlock_irq(&lock); \
797 cmd; \
798 schedule(); \
Peter Zijlstra8fbd88f2013-10-02 11:22:28 +0200799 spin_lock_irq(&lock))
Lukas Czernereed8c022012-11-30 11:42:40 +0100800
801/**
802 * wait_event_interruptible_lock_irq_cmd - sleep until a condition gets true.
803 * The condition is checked under the lock. This is expected to
804 * be called with the lock taken.
805 * @wq: the waitqueue to wait on
806 * @condition: a C expression for the event to wait for
807 * @lock: a locked spinlock_t, which will be released before cmd and
808 * schedule() and reacquired afterwards.
809 * @cmd: a command which is invoked outside the critical section before
810 * sleep
811 *
812 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
813 * @condition evaluates to true or a signal is received. The @condition is
814 * checked each time the waitqueue @wq is woken up.
815 *
816 * wake_up() has to be called after changing any variable that could
817 * change the result of the wait condition.
818 *
819 * This is supposed to be called while holding the lock. The lock is
820 * dropped before invoking the cmd and going to sleep and is reacquired
821 * afterwards.
822 *
823 * The macro will return -ERESTARTSYS if it was interrupted by a signal
824 * and 0 if @condition evaluated to true.
825 */
826#define wait_event_interruptible_lock_irq_cmd(wq, condition, lock, cmd) \
827({ \
828 int __ret = 0; \
Lukas Czernereed8c022012-11-30 11:42:40 +0100829 if (!(condition)) \
Ingo Molnarfb869b62013-10-04 10:24:49 +0200830 __ret = __wait_event_interruptible_lock_irq(wq, \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200831 condition, lock, cmd); \
Lukas Czernereed8c022012-11-30 11:42:40 +0100832 __ret; \
833})
834
835/**
836 * wait_event_interruptible_lock_irq - sleep until a condition gets true.
837 * The condition is checked under the lock. This is expected
838 * to be called with the lock taken.
839 * @wq: the waitqueue to wait on
840 * @condition: a C expression for the event to wait for
841 * @lock: a locked spinlock_t, which will be released before schedule()
842 * and reacquired afterwards.
843 *
844 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
845 * @condition evaluates to true or signal is received. The @condition is
846 * checked each time the waitqueue @wq is woken up.
847 *
848 * wake_up() has to be called after changing any variable that could
849 * change the result of the wait condition.
850 *
851 * This is supposed to be called while holding the lock. The lock is
852 * dropped before going to sleep and is reacquired afterwards.
853 *
854 * The macro will return -ERESTARTSYS if it was interrupted by a signal
855 * and 0 if @condition evaluated to true.
856 */
857#define wait_event_interruptible_lock_irq(wq, condition, lock) \
858({ \
859 int __ret = 0; \
Lukas Czernereed8c022012-11-30 11:42:40 +0100860 if (!(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200861 __ret = __wait_event_interruptible_lock_irq(wq, \
Thierry Reding92ec1182013-10-23 13:40:55 +0200862 condition, lock,); \
Lukas Czernereed8c022012-11-30 11:42:40 +0100863 __ret; \
864})
865
Ingo Molnarfb869b62013-10-04 10:24:49 +0200866#define __wait_event_interruptible_lock_irq_timeout(wq, condition, \
867 lock, timeout) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200868 ___wait_event(wq, ___wait_cond_timeout(condition), \
Heiko Carstens7d716452013-10-31 12:48:14 +0100869 TASK_INTERRUPTIBLE, 0, timeout, \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200870 spin_unlock_irq(&lock); \
871 __ret = schedule_timeout(__ret); \
Peter Zijlstraa1dc6852013-10-02 11:22:29 +0200872 spin_lock_irq(&lock));
Martin Peschked79ff142013-08-22 17:45:36 +0200873
874/**
Ingo Molnarfb869b62013-10-04 10:24:49 +0200875 * wait_event_interruptible_lock_irq_timeout - sleep until a condition gets
876 * true or a timeout elapses. The condition is checked under
877 * the lock. This is expected to be called with the lock taken.
Martin Peschked79ff142013-08-22 17:45:36 +0200878 * @wq: the waitqueue to wait on
879 * @condition: a C expression for the event to wait for
880 * @lock: a locked spinlock_t, which will be released before schedule()
881 * and reacquired afterwards.
882 * @timeout: timeout, in jiffies
883 *
884 * The process is put to sleep (TASK_INTERRUPTIBLE) until the
885 * @condition evaluates to true or signal is received. The @condition is
886 * checked each time the waitqueue @wq is woken up.
887 *
888 * wake_up() has to be called after changing any variable that could
889 * change the result of the wait condition.
890 *
891 * This is supposed to be called while holding the lock. The lock is
892 * dropped before going to sleep and is reacquired afterwards.
893 *
894 * The function returns 0 if the @timeout elapsed, -ERESTARTSYS if it
895 * was interrupted by a signal, and the remaining jiffies otherwise
896 * if the condition evaluated to true before the timeout elapsed.
897 */
898#define wait_event_interruptible_lock_irq_timeout(wq, condition, lock, \
899 timeout) \
900({ \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200901 long __ret = timeout; \
Oleg Nesterov89229152013-10-07 20:31:06 +0200902 if (!___wait_cond_timeout(condition)) \
Peter Zijlstra35a2af92013-10-02 11:22:33 +0200903 __ret = __wait_event_interruptible_lock_irq_timeout( \
904 wq, condition, lock, timeout); \
Martin Peschked79ff142013-08-22 17:45:36 +0200905 __ret; \
906})
907
Linus Torvalds1da177e2005-04-16 15:20:36 -0700908/*
909 * Waitqueues which are removed from the waitqueue_head at wakeup time
910 */
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800911void prepare_to_wait(wait_queue_head_t *q, wait_queue_t *wait, int state);
912void prepare_to_wait_exclusive(wait_queue_head_t *q, wait_queue_t *wait, int state);
Oleg Nesterovc2d81642013-10-07 18:18:24 +0200913long prepare_to_wait_event(wait_queue_head_t *q, wait_queue_t *wait, int state);
Harvey Harrisonb3c97522008-02-13 15:03:15 -0800914void finish_wait(wait_queue_head_t *q, wait_queue_t *wait);
Ingo Molnarfb869b62013-10-04 10:24:49 +0200915void abort_exclusive_wait(wait_queue_head_t *q, wait_queue_t *wait, unsigned int mode, void *key);
Peter Zijlstra61ada522014-09-24 10:18:47 +0200916long wait_woken(wait_queue_t *wait, unsigned mode, long timeout);
917int woken_wake_function(wait_queue_t *wait, unsigned mode, int sync, void *key);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700918int autoremove_wake_function(wait_queue_t *wait, unsigned mode, int sync, void *key);
919int wake_bit_function(wait_queue_t *wait, unsigned mode, int sync, void *key);
920
Eric Dumazetbf368e42009-04-28 02:24:21 -0700921#define DEFINE_WAIT_FUNC(name, function) \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700922 wait_queue_t name = { \
Benjamin LaHaisec43dc2f2005-06-23 00:10:27 -0700923 .private = current, \
Eric Dumazetbf368e42009-04-28 02:24:21 -0700924 .func = function, \
blaisorblade@yahoo.it7e43c842005-05-25 01:31:42 +0200925 .task_list = LIST_HEAD_INIT((name).task_list), \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700926 }
927
Eric Dumazetbf368e42009-04-28 02:24:21 -0700928#define DEFINE_WAIT(name) DEFINE_WAIT_FUNC(name, autoremove_wake_function)
929
Linus Torvalds1da177e2005-04-16 15:20:36 -0700930#define DEFINE_WAIT_BIT(name, word, bit) \
931 struct wait_bit_queue name = { \
932 .key = __WAIT_BIT_KEY_INITIALIZER(word, bit), \
933 .wait = { \
Benjamin LaHaisec43dc2f2005-06-23 00:10:27 -0700934 .private = current, \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700935 .func = wake_bit_function, \
936 .task_list = \
937 LIST_HEAD_INIT((name).wait.task_list), \
938 }, \
939 }
940
941#define init_wait(wait) \
942 do { \
Benjamin LaHaisec43dc2f2005-06-23 00:10:27 -0700943 (wait)->private = current; \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700944 (wait)->func = autoremove_wake_function; \
945 INIT_LIST_HEAD(&(wait)->task_list); \
Evgeny Kuznetsov231d0ae2010-10-05 12:47:57 +0400946 (wait)->flags = 0; \
Linus Torvalds1da177e2005-04-16 15:20:36 -0700947 } while (0)
948
NeilBrown74316202014-07-07 15:16:04 +1000949
NeilBrownc1221322014-07-07 15:16:04 +1000950extern int bit_wait(struct wait_bit_key *);
951extern int bit_wait_io(struct wait_bit_key *);
NeilBrowncbbce822014-09-25 13:55:19 +1000952extern int bit_wait_timeout(struct wait_bit_key *);
953extern int bit_wait_io_timeout(struct wait_bit_key *);
NeilBrown74316202014-07-07 15:16:04 +1000954
Linus Torvalds1da177e2005-04-16 15:20:36 -0700955/**
956 * wait_on_bit - wait for a bit to be cleared
957 * @word: the word being waited on, a kernel virtual address
958 * @bit: the bit of the word being waited on
Linus Torvalds1da177e2005-04-16 15:20:36 -0700959 * @mode: the task state to sleep in
960 *
961 * There is a standard hashed waitqueue table for generic use. This
962 * is the part of the hashtable's accessor API that waits on a bit.
963 * For instance, if one were to have waiters on a bitflag, one would
964 * call wait_on_bit() in threads waiting for the bit to clear.
965 * One uses wait_on_bit() where one is waiting for the bit to clear,
966 * but has no intention of setting it.
NeilBrown74316202014-07-07 15:16:04 +1000967 * Returned value will be zero if the bit was cleared, or non-zero
968 * if the process received a signal and the mode permitted wakeup
969 * on that signal.
Linus Torvalds1da177e2005-04-16 15:20:36 -0700970 */
Ingo Molnarfb869b62013-10-04 10:24:49 +0200971static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -0700972wait_on_bit(unsigned long *word, int bit, unsigned mode)
NeilBrown74316202014-07-07 15:16:04 +1000973{
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200974 might_sleep();
NeilBrown74316202014-07-07 15:16:04 +1000975 if (!test_bit(bit, word))
976 return 0;
977 return out_of_line_wait_on_bit(word, bit,
978 bit_wait,
979 mode);
980}
981
982/**
983 * wait_on_bit_io - wait for a bit to be cleared
984 * @word: the word being waited on, a kernel virtual address
985 * @bit: the bit of the word being waited on
986 * @mode: the task state to sleep in
987 *
988 * Use the standard hashed waitqueue table to wait for a bit
989 * to be cleared. This is similar to wait_on_bit(), but calls
990 * io_schedule() instead of schedule() for the actual waiting.
991 *
992 * Returned value will be zero if the bit was cleared, or non-zero
993 * if the process received a signal and the mode permitted wakeup
994 * on that signal.
995 */
996static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -0700997wait_on_bit_io(unsigned long *word, int bit, unsigned mode)
NeilBrown74316202014-07-07 15:16:04 +1000998{
Peter Zijlstrae22b8862014-09-24 10:18:48 +0200999 might_sleep();
NeilBrown74316202014-07-07 15:16:04 +10001000 if (!test_bit(bit, word))
1001 return 0;
1002 return out_of_line_wait_on_bit(word, bit,
1003 bit_wait_io,
1004 mode);
1005}
1006
1007/**
Johan Hedberg44fc0e52015-01-30 13:14:36 +02001008 * wait_on_bit_timeout - wait for a bit to be cleared or a timeout elapses
1009 * @word: the word being waited on, a kernel virtual address
1010 * @bit: the bit of the word being waited on
1011 * @mode: the task state to sleep in
1012 * @timeout: timeout, in jiffies
1013 *
1014 * Use the standard hashed waitqueue table to wait for a bit
1015 * to be cleared. This is similar to wait_on_bit(), except also takes a
1016 * timeout parameter.
1017 *
1018 * Returned value will be zero if the bit was cleared before the
1019 * @timeout elapsed, or non-zero if the @timeout elapsed or process
1020 * received a signal and the mode permitted wakeup on that signal.
1021 */
1022static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -07001023wait_on_bit_timeout(unsigned long *word, int bit, unsigned mode,
1024 unsigned long timeout)
Johan Hedberg44fc0e52015-01-30 13:14:36 +02001025{
1026 might_sleep();
1027 if (!test_bit(bit, word))
1028 return 0;
1029 return out_of_line_wait_on_bit_timeout(word, bit,
1030 bit_wait_timeout,
1031 mode, timeout);
1032}
1033
1034/**
NeilBrown74316202014-07-07 15:16:04 +10001035 * wait_on_bit_action - wait for a bit to be cleared
1036 * @word: the word being waited on, a kernel virtual address
1037 * @bit: the bit of the word being waited on
1038 * @action: the function used to sleep, which may take special actions
1039 * @mode: the task state to sleep in
1040 *
1041 * Use the standard hashed waitqueue table to wait for a bit
1042 * to be cleared, and allow the waiting action to be specified.
1043 * This is like wait_on_bit() but allows fine control of how the waiting
1044 * is done.
1045 *
1046 * Returned value will be zero if the bit was cleared, or non-zero
1047 * if the process received a signal and the mode permitted wakeup
1048 * on that signal.
1049 */
1050static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -07001051wait_on_bit_action(unsigned long *word, int bit, wait_bit_action_f *action,
1052 unsigned mode)
Linus Torvalds1da177e2005-04-16 15:20:36 -07001053{
Peter Zijlstrae22b8862014-09-24 10:18:48 +02001054 might_sleep();
Linus Torvalds1da177e2005-04-16 15:20:36 -07001055 if (!test_bit(bit, word))
1056 return 0;
1057 return out_of_line_wait_on_bit(word, bit, action, mode);
1058}
1059
1060/**
1061 * wait_on_bit_lock - wait for a bit to be cleared, when wanting to set it
1062 * @word: the word being waited on, a kernel virtual address
1063 * @bit: the bit of the word being waited on
Linus Torvalds1da177e2005-04-16 15:20:36 -07001064 * @mode: the task state to sleep in
1065 *
1066 * There is a standard hashed waitqueue table for generic use. This
1067 * is the part of the hashtable's accessor API that waits on a bit
1068 * when one intends to set it, for instance, trying to lock bitflags.
1069 * For instance, if one were to have waiters trying to set bitflag
1070 * and waiting for it to clear before setting it, one would call
1071 * wait_on_bit() in threads waiting to be able to set the bit.
1072 * One uses wait_on_bit_lock() where one is waiting for the bit to
1073 * clear with the intention of setting it, and when done, clearing it.
NeilBrown74316202014-07-07 15:16:04 +10001074 *
1075 * Returns zero if the bit was (eventually) found to be clear and was
1076 * set. Returns non-zero if a signal was delivered to the process and
1077 * the @mode allows that signal to wake the process.
Linus Torvalds1da177e2005-04-16 15:20:36 -07001078 */
Ingo Molnarfb869b62013-10-04 10:24:49 +02001079static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -07001080wait_on_bit_lock(unsigned long *word, int bit, unsigned mode)
NeilBrown74316202014-07-07 15:16:04 +10001081{
Peter Zijlstrae22b8862014-09-24 10:18:48 +02001082 might_sleep();
NeilBrown74316202014-07-07 15:16:04 +10001083 if (!test_and_set_bit(bit, word))
1084 return 0;
1085 return out_of_line_wait_on_bit_lock(word, bit, bit_wait, mode);
1086}
1087
1088/**
1089 * wait_on_bit_lock_io - wait for a bit to be cleared, when wanting to set it
1090 * @word: the word being waited on, a kernel virtual address
1091 * @bit: the bit of the word being waited on
1092 * @mode: the task state to sleep in
1093 *
1094 * Use the standard hashed waitqueue table to wait for a bit
1095 * to be cleared and then to atomically set it. This is similar
1096 * to wait_on_bit(), but calls io_schedule() instead of schedule()
1097 * for the actual waiting.
1098 *
1099 * Returns zero if the bit was (eventually) found to be clear and was
1100 * set. Returns non-zero if a signal was delivered to the process and
1101 * the @mode allows that signal to wake the process.
1102 */
1103static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -07001104wait_on_bit_lock_io(unsigned long *word, int bit, unsigned mode)
NeilBrown74316202014-07-07 15:16:04 +10001105{
Peter Zijlstrae22b8862014-09-24 10:18:48 +02001106 might_sleep();
NeilBrown74316202014-07-07 15:16:04 +10001107 if (!test_and_set_bit(bit, word))
1108 return 0;
1109 return out_of_line_wait_on_bit_lock(word, bit, bit_wait_io, mode);
1110}
1111
1112/**
1113 * wait_on_bit_lock_action - wait for a bit to be cleared, when wanting to set it
1114 * @word: the word being waited on, a kernel virtual address
1115 * @bit: the bit of the word being waited on
1116 * @action: the function used to sleep, which may take special actions
1117 * @mode: the task state to sleep in
1118 *
1119 * Use the standard hashed waitqueue table to wait for a bit
1120 * to be cleared and then to set it, and allow the waiting action
1121 * to be specified.
1122 * This is like wait_on_bit() but allows fine control of how the waiting
1123 * is done.
1124 *
1125 * Returns zero if the bit was (eventually) found to be clear and was
1126 * set. Returns non-zero if a signal was delivered to the process and
1127 * the @mode allows that signal to wake the process.
1128 */
1129static inline int
Palmer Dabbelt7e605982015-04-30 21:19:56 -07001130wait_on_bit_lock_action(unsigned long *word, int bit, wait_bit_action_f *action,
1131 unsigned mode)
Linus Torvalds1da177e2005-04-16 15:20:36 -07001132{
Peter Zijlstrae22b8862014-09-24 10:18:48 +02001133 might_sleep();
Linus Torvalds1da177e2005-04-16 15:20:36 -07001134 if (!test_and_set_bit(bit, word))
1135 return 0;
1136 return out_of_line_wait_on_bit_lock(word, bit, action, mode);
1137}
David Howellscb655372013-05-10 19:50:26 +01001138
1139/**
1140 * wait_on_atomic_t - Wait for an atomic_t to become 0
1141 * @val: The atomic value being waited on, a kernel virtual address
1142 * @action: the function used to sleep, which may take special actions
1143 * @mode: the task state to sleep in
1144 *
1145 * Wait for an atomic_t to become 0. We abuse the bit-wait waitqueue table for
1146 * the purpose of getting a waitqueue, but we set the key to a bit number
1147 * outside of the target 'word'.
1148 */
1149static inline
1150int wait_on_atomic_t(atomic_t *val, int (*action)(atomic_t *), unsigned mode)
1151{
Peter Zijlstrae22b8862014-09-24 10:18:48 +02001152 might_sleep();
David Howellscb655372013-05-10 19:50:26 +01001153 if (atomic_read(val) == 0)
1154 return 0;
1155 return out_of_line_wait_on_atomic_t(val, action, mode);
1156}
Ingo Molnarfb869b62013-10-04 10:24:49 +02001157
1158#endif /* _LINUX_WAIT_H */