blob: 94a62c0d4ade651b8c4e2a5c7cbb8ff26110738f [file] [log] [blame]
Matthew Wilcox64ac24e2008-03-07 21:55:58 -05001/*
2 * Copyright (c) 2008 Intel Corporation
3 * Author: Matthew Wilcox <willy@linux.intel.com>
4 *
5 * Distributed under the terms of the GNU GPL, version 2
Matthew Wilcox714493c2008-04-11 15:23:52 -04006 *
7 * This file implements counting semaphores.
8 * A counting semaphore may be acquired 'n' times before sleeping.
9 * See mutex.c for single-acquisition sleeping locks which enforce
10 * rules which allow code to be debugged more easily.
11 */
12
13/*
14 * Some notes on the implementation:
15 *
16 * The spinlock controls access to the other members of the semaphore.
17 * down_trylock() and up() can be called from interrupt context, so we
18 * have to disable interrupts when taking the lock. It turns out various
19 * parts of the kernel expect to be able to use down() on a semaphore in
20 * interrupt context when they know it will succeed, so we have to use
21 * irqsave variants for down(), down_interruptible() and down_killable()
22 * too.
23 *
24 * The ->count variable represents how many more tasks can acquire this
25 * semaphore. If it's zero, there may be tasks waiting on the wait_list.
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050026 */
27
28#include <linux/compiler.h>
29#include <linux/kernel.h>
30#include <linux/module.h>
31#include <linux/sched.h>
32#include <linux/semaphore.h>
33#include <linux/spinlock.h>
Ingo Molnar74f4e362008-05-12 21:21:15 +020034#include <linux/ftrace.h>
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050035
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050036static noinline void __down(struct semaphore *sem);
37static noinline int __down_interruptible(struct semaphore *sem);
Matthew Wilcoxf06d9682008-03-14 13:19:33 -040038static noinline int __down_killable(struct semaphore *sem);
Matthew Wilcoxf1241c82008-03-14 13:43:13 -040039static noinline int __down_timeout(struct semaphore *sem, long jiffies);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050040static noinline void __up(struct semaphore *sem);
41
Matthew Wilcox714493c2008-04-11 15:23:52 -040042/**
43 * down - acquire the semaphore
44 * @sem: the semaphore to be acquired
45 *
46 * Acquires the semaphore. If no more tasks are allowed to acquire the
47 * semaphore, calling this function will put the task to sleep until the
48 * semaphore is released.
49 *
50 * Use of this function is deprecated, please use down_interruptible() or
51 * down_killable() instead.
52 */
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050053void down(struct semaphore *sem)
54{
55 unsigned long flags;
56
57 spin_lock_irqsave(&sem->lock, flags);
Linus Torvalds00b41ec2008-05-10 20:43:22 -070058 if (likely(sem->count > 0))
59 sem->count--;
60 else
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050061 __down(sem);
62 spin_unlock_irqrestore(&sem->lock, flags);
63}
64EXPORT_SYMBOL(down);
65
Matthew Wilcox714493c2008-04-11 15:23:52 -040066/**
67 * down_interruptible - acquire the semaphore unless interrupted
68 * @sem: the semaphore to be acquired
69 *
70 * Attempts to acquire the semaphore. If no more tasks are allowed to
71 * acquire the semaphore, calling this function will put the task to sleep.
72 * If the sleep is interrupted by a signal, this function will return -EINTR.
73 * If the semaphore is successfully acquired, this function returns 0.
74 */
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050075int down_interruptible(struct semaphore *sem)
76{
77 unsigned long flags;
78 int result = 0;
79
80 spin_lock_irqsave(&sem->lock, flags);
Linus Torvalds00b41ec2008-05-10 20:43:22 -070081 if (likely(sem->count > 0))
Ingo Molnarbf726ea2008-05-08 11:53:48 +020082 sem->count--;
Linus Torvalds00b41ec2008-05-10 20:43:22 -070083 else
84 result = __down_interruptible(sem);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -050085 spin_unlock_irqrestore(&sem->lock, flags);
86
87 return result;
88}
89EXPORT_SYMBOL(down_interruptible);
90
Matthew Wilcox714493c2008-04-11 15:23:52 -040091/**
92 * down_killable - acquire the semaphore unless killed
93 * @sem: the semaphore to be acquired
94 *
95 * Attempts to acquire the semaphore. If no more tasks are allowed to
96 * acquire the semaphore, calling this function will put the task to sleep.
97 * If the sleep is interrupted by a fatal signal, this function will return
98 * -EINTR. If the semaphore is successfully acquired, this function returns
99 * 0.
100 */
Matthew Wilcoxf06d9682008-03-14 13:19:33 -0400101int down_killable(struct semaphore *sem)
102{
103 unsigned long flags;
104 int result = 0;
105
106 spin_lock_irqsave(&sem->lock, flags);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700107 if (likely(sem->count > 0))
Ingo Molnarbf726ea2008-05-08 11:53:48 +0200108 sem->count--;
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700109 else
110 result = __down_killable(sem);
Matthew Wilcoxf06d9682008-03-14 13:19:33 -0400111 spin_unlock_irqrestore(&sem->lock, flags);
112
113 return result;
114}
115EXPORT_SYMBOL(down_killable);
116
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500117/**
118 * down_trylock - try to acquire the semaphore, without waiting
119 * @sem: the semaphore to be acquired
120 *
121 * Try to acquire the semaphore atomically. Returns 0 if the mutex has
Matthew Wilcox714493c2008-04-11 15:23:52 -0400122 * been acquired successfully or 1 if it it cannot be acquired.
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500123 *
124 * NOTE: This return value is inverted from both spin_trylock and
125 * mutex_trylock! Be careful about this when converting code.
126 *
127 * Unlike mutex_trylock, this function can be used from interrupt context,
128 * and the semaphore can be released by any task or interrupt.
129 */
130int down_trylock(struct semaphore *sem)
131{
132 unsigned long flags;
133 int count;
134
135 spin_lock_irqsave(&sem->lock, flags);
136 count = sem->count - 1;
137 if (likely(count >= 0))
138 sem->count = count;
139 spin_unlock_irqrestore(&sem->lock, flags);
140
141 return (count < 0);
142}
143EXPORT_SYMBOL(down_trylock);
144
Matthew Wilcox714493c2008-04-11 15:23:52 -0400145/**
146 * down_timeout - acquire the semaphore within a specified time
147 * @sem: the semaphore to be acquired
148 * @jiffies: how long to wait before failing
149 *
150 * Attempts to acquire the semaphore. If no more tasks are allowed to
151 * acquire the semaphore, calling this function will put the task to sleep.
152 * If the semaphore is not released within the specified number of jiffies,
153 * this function returns -ETIME. It returns 0 if the semaphore was acquired.
154 */
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400155int down_timeout(struct semaphore *sem, long jiffies)
156{
157 unsigned long flags;
158 int result = 0;
159
160 spin_lock_irqsave(&sem->lock, flags);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700161 if (likely(sem->count > 0))
Ingo Molnarbf726ea2008-05-08 11:53:48 +0200162 sem->count--;
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700163 else
164 result = __down_timeout(sem, jiffies);
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400165 spin_unlock_irqrestore(&sem->lock, flags);
166
167 return result;
168}
169EXPORT_SYMBOL(down_timeout);
170
Matthew Wilcox714493c2008-04-11 15:23:52 -0400171/**
172 * up - release the semaphore
173 * @sem: the semaphore to release
174 *
175 * Release the semaphore. Unlike mutexes, up() may be called from any
176 * context and even by tasks which have never called down().
177 */
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500178void up(struct semaphore *sem)
179{
180 unsigned long flags;
181
182 spin_lock_irqsave(&sem->lock, flags);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700183 if (likely(list_empty(&sem->wait_list)))
184 sem->count++;
185 else
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500186 __up(sem);
187 spin_unlock_irqrestore(&sem->lock, flags);
188}
189EXPORT_SYMBOL(up);
190
191/* Functions for the contended case */
192
193struct semaphore_waiter {
194 struct list_head list;
195 struct task_struct *task;
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700196 int up;
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500197};
198
199/*
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400200 * Because this function is inlined, the 'state' parameter will be
201 * constant, and thus optimised away by the compiler. Likewise the
202 * 'timeout' parameter for the cases without timeouts.
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500203 */
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400204static inline int __sched __down_common(struct semaphore *sem, long state,
205 long timeout)
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500206{
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500207 struct task_struct *task = current;
208 struct semaphore_waiter waiter;
209
Ingo Molnarbf726ea2008-05-08 11:53:48 +0200210 list_add_tail(&waiter.list, &sem->wait_list);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700211 waiter.task = task;
212 waiter.up = 0;
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500213
214 for (;;) {
Oleg Nesterov5b2becc2008-08-05 13:01:13 -0700215 if (signal_pending_state(state, task))
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700216 goto interrupted;
217 if (timeout <= 0)
218 goto timed_out;
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500219 __set_task_state(task, state);
220 spin_unlock_irq(&sem->lock);
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400221 timeout = schedule_timeout(timeout);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500222 spin_lock_irq(&sem->lock);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700223 if (waiter.up)
224 return 0;
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500225 }
226
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700227 timed_out:
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400228 list_del(&waiter.list);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700229 return -ETIME;
230
231 interrupted:
232 list_del(&waiter.list);
233 return -EINTR;
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500234}
235
236static noinline void __sched __down(struct semaphore *sem)
237{
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400238 __down_common(sem, TASK_UNINTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500239}
240
241static noinline int __sched __down_interruptible(struct semaphore *sem)
242{
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400243 return __down_common(sem, TASK_INTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500244}
245
Matthew Wilcoxf06d9682008-03-14 13:19:33 -0400246static noinline int __sched __down_killable(struct semaphore *sem)
247{
Matthew Wilcoxf1241c82008-03-14 13:43:13 -0400248 return __down_common(sem, TASK_KILLABLE, MAX_SCHEDULE_TIMEOUT);
249}
250
251static noinline int __sched __down_timeout(struct semaphore *sem, long jiffies)
252{
253 return __down_common(sem, TASK_UNINTERRUPTIBLE, jiffies);
Matthew Wilcoxf06d9682008-03-14 13:19:33 -0400254}
255
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500256static noinline void __sched __up(struct semaphore *sem)
257{
Matthew Wilcoxb17170b2008-03-14 14:35:22 -0400258 struct semaphore_waiter *waiter = list_first_entry(&sem->wait_list,
259 struct semaphore_waiter, list);
Linus Torvalds00b41ec2008-05-10 20:43:22 -0700260 list_del(&waiter->list);
261 waiter->up = 1;
Matthew Wilcoxb17170b2008-03-14 14:35:22 -0400262 wake_up_process(waiter->task);
Matthew Wilcox64ac24e2008-03-07 21:55:58 -0500263}