blob: 738e8a8cb811ddd86097e2a1f180cec96057e6ea [file] [log] [blame]
David Vrabel8cc13a02008-09-17 16:34:09 +01001/*
2 * UWB reservation management.
3 *
4 * Copyright (C) 2008 Cambridge Silicon Radio Ltd.
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License version
8 * 2 as published by the Free Software Foundation.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 * You should have received a copy of the GNU General Public License
16 * along with this program. If not, see <http://www.gnu.org/licenses/>.
17 */
David Vrabel8cc13a02008-09-17 16:34:09 +010018#include <linux/kernel.h>
19#include <linux/uwb.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090020#include <linux/slab.h>
Stefano Panella5b377172008-12-12 13:00:06 +000021#include <linux/random.h>
Paul Gortmaker475c0a62011-07-10 13:18:02 -040022#include <linux/export.h>
David Vrabel8cc13a02008-09-17 16:34:09 +010023
24#include "uwb-internal.h"
25
26static void uwb_rsv_timer(unsigned long arg);
27
28static const char *rsv_states[] = {
Stefano Panella5b377172008-12-12 13:00:06 +000029 [UWB_RSV_STATE_NONE] = "none ",
30 [UWB_RSV_STATE_O_INITIATED] = "o initiated ",
31 [UWB_RSV_STATE_O_PENDING] = "o pending ",
32 [UWB_RSV_STATE_O_MODIFIED] = "o modified ",
33 [UWB_RSV_STATE_O_ESTABLISHED] = "o established ",
34 [UWB_RSV_STATE_O_TO_BE_MOVED] = "o to be moved ",
35 [UWB_RSV_STATE_O_MOVE_EXPANDING] = "o move expanding",
36 [UWB_RSV_STATE_O_MOVE_COMBINING] = "o move combining",
37 [UWB_RSV_STATE_O_MOVE_REDUCING] = "o move reducing ",
38 [UWB_RSV_STATE_T_ACCEPTED] = "t accepted ",
39 [UWB_RSV_STATE_T_CONFLICT] = "t conflict ",
40 [UWB_RSV_STATE_T_PENDING] = "t pending ",
41 [UWB_RSV_STATE_T_DENIED] = "t denied ",
42 [UWB_RSV_STATE_T_RESIZED] = "t resized ",
43 [UWB_RSV_STATE_T_EXPANDING_ACCEPTED] = "t expanding acc ",
44 [UWB_RSV_STATE_T_EXPANDING_CONFLICT] = "t expanding conf",
45 [UWB_RSV_STATE_T_EXPANDING_PENDING] = "t expanding pend",
46 [UWB_RSV_STATE_T_EXPANDING_DENIED] = "t expanding den ",
David Vrabel8cc13a02008-09-17 16:34:09 +010047};
48
49static const char *rsv_types[] = {
50 [UWB_DRP_TYPE_ALIEN_BP] = "alien-bp",
51 [UWB_DRP_TYPE_HARD] = "hard",
52 [UWB_DRP_TYPE_SOFT] = "soft",
53 [UWB_DRP_TYPE_PRIVATE] = "private",
54 [UWB_DRP_TYPE_PCA] = "pca",
55};
56
Stefano Panella5b377172008-12-12 13:00:06 +000057bool uwb_rsv_has_two_drp_ies(struct uwb_rsv *rsv)
58{
59 static const bool has_two_drp_ies[] = {
60 [UWB_RSV_STATE_O_INITIATED] = false,
61 [UWB_RSV_STATE_O_PENDING] = false,
62 [UWB_RSV_STATE_O_MODIFIED] = false,
63 [UWB_RSV_STATE_O_ESTABLISHED] = false,
64 [UWB_RSV_STATE_O_TO_BE_MOVED] = false,
65 [UWB_RSV_STATE_O_MOVE_COMBINING] = false,
66 [UWB_RSV_STATE_O_MOVE_REDUCING] = false,
67 [UWB_RSV_STATE_O_MOVE_EXPANDING] = true,
68 [UWB_RSV_STATE_T_ACCEPTED] = false,
69 [UWB_RSV_STATE_T_CONFLICT] = false,
70 [UWB_RSV_STATE_T_PENDING] = false,
71 [UWB_RSV_STATE_T_DENIED] = false,
72 [UWB_RSV_STATE_T_RESIZED] = false,
73 [UWB_RSV_STATE_T_EXPANDING_ACCEPTED] = true,
74 [UWB_RSV_STATE_T_EXPANDING_CONFLICT] = true,
75 [UWB_RSV_STATE_T_EXPANDING_PENDING] = true,
76 [UWB_RSV_STATE_T_EXPANDING_DENIED] = true,
77 };
78
79 return has_two_drp_ies[rsv->state];
80}
81
David Vrabel8cc13a02008-09-17 16:34:09 +010082/**
83 * uwb_rsv_state_str - return a string for a reservation state
84 * @state: the reservation state.
85 */
86const char *uwb_rsv_state_str(enum uwb_rsv_state state)
87{
88 if (state < UWB_RSV_STATE_NONE || state >= UWB_RSV_STATE_LAST)
89 return "unknown";
90 return rsv_states[state];
91}
92EXPORT_SYMBOL_GPL(uwb_rsv_state_str);
93
94/**
95 * uwb_rsv_type_str - return a string for a reservation type
96 * @type: the reservation type
97 */
98const char *uwb_rsv_type_str(enum uwb_drp_type type)
99{
100 if (type < UWB_DRP_TYPE_ALIEN_BP || type > UWB_DRP_TYPE_PCA)
101 return "invalid";
102 return rsv_types[type];
103}
104EXPORT_SYMBOL_GPL(uwb_rsv_type_str);
105
Stefano Panella5b377172008-12-12 13:00:06 +0000106void uwb_rsv_dump(char *text, struct uwb_rsv *rsv)
David Vrabel8cc13a02008-09-17 16:34:09 +0100107{
108 struct device *dev = &rsv->rc->uwb_dev.dev;
109 struct uwb_dev_addr devaddr;
110 char owner[UWB_ADDR_STRSIZE], target[UWB_ADDR_STRSIZE];
111
112 uwb_dev_addr_print(owner, sizeof(owner), &rsv->owner->dev_addr);
113 if (rsv->target.type == UWB_RSV_TARGET_DEV)
114 devaddr = rsv->target.dev->dev_addr;
115 else
116 devaddr = rsv->target.devaddr;
117 uwb_dev_addr_print(target, sizeof(target), &devaddr);
118
David Vrabel2226b1c2009-01-06 17:55:32 +0000119 dev_dbg(dev, "rsv %s %s -> %s: %s\n",
120 text, owner, target, uwb_rsv_state_str(rsv->state));
David Vrabel8cc13a02008-09-17 16:34:09 +0100121}
122
David Vrabelcae1c112008-10-27 15:22:46 +0000123static void uwb_rsv_release(struct kref *kref)
124{
125 struct uwb_rsv *rsv = container_of(kref, struct uwb_rsv, kref);
126
127 kfree(rsv);
128}
129
Stefano Panella5b377172008-12-12 13:00:06 +0000130void uwb_rsv_get(struct uwb_rsv *rsv)
David Vrabelcae1c112008-10-27 15:22:46 +0000131{
132 kref_get(&rsv->kref);
133}
134
Stefano Panella5b377172008-12-12 13:00:06 +0000135void uwb_rsv_put(struct uwb_rsv *rsv)
David Vrabelcae1c112008-10-27 15:22:46 +0000136{
137 kref_put(&rsv->kref, uwb_rsv_release);
138}
139
David Vrabel8cc13a02008-09-17 16:34:09 +0100140/*
141 * Get a free stream index for a reservation.
142 *
143 * If the target is a DevAddr (e.g., a WUSB cluster reservation) then
144 * the stream is allocated from a pool of per-RC stream indexes,
145 * otherwise a unique stream index for the target is selected.
146 */
147static int uwb_rsv_get_stream(struct uwb_rsv *rsv)
148{
149 struct uwb_rc *rc = rsv->rc;
Stefano Panella5b377172008-12-12 13:00:06 +0000150 struct device *dev = &rc->uwb_dev.dev;
David Vrabel8cc13a02008-09-17 16:34:09 +0100151 unsigned long *streams_bm;
152 int stream;
153
154 switch (rsv->target.type) {
155 case UWB_RSV_TARGET_DEV:
156 streams_bm = rsv->target.dev->streams;
157 break;
158 case UWB_RSV_TARGET_DEVADDR:
159 streams_bm = rc->uwb_dev.streams;
160 break;
161 default:
162 return -EINVAL;
163 }
164
165 stream = find_first_zero_bit(streams_bm, UWB_NUM_STREAMS);
166 if (stream >= UWB_NUM_STREAMS)
167 return -EBUSY;
168
169 rsv->stream = stream;
170 set_bit(stream, streams_bm);
171
Stefano Panella5b377172008-12-12 13:00:06 +0000172 dev_dbg(dev, "get stream %d\n", rsv->stream);
173
David Vrabel8cc13a02008-09-17 16:34:09 +0100174 return 0;
175}
176
177static void uwb_rsv_put_stream(struct uwb_rsv *rsv)
178{
179 struct uwb_rc *rc = rsv->rc;
Stefano Panella5b377172008-12-12 13:00:06 +0000180 struct device *dev = &rc->uwb_dev.dev;
David Vrabel8cc13a02008-09-17 16:34:09 +0100181 unsigned long *streams_bm;
182
183 switch (rsv->target.type) {
184 case UWB_RSV_TARGET_DEV:
185 streams_bm = rsv->target.dev->streams;
186 break;
187 case UWB_RSV_TARGET_DEVADDR:
188 streams_bm = rc->uwb_dev.streams;
189 break;
190 default:
191 return;
192 }
193
194 clear_bit(rsv->stream, streams_bm);
Stefano Panella5b377172008-12-12 13:00:06 +0000195
196 dev_dbg(dev, "put stream %d\n", rsv->stream);
David Vrabel8cc13a02008-09-17 16:34:09 +0100197}
198
Stefano Panella5b377172008-12-12 13:00:06 +0000199void uwb_rsv_backoff_win_timer(unsigned long arg)
David Vrabel8cc13a02008-09-17 16:34:09 +0100200{
Stefano Panella5b377172008-12-12 13:00:06 +0000201 struct uwb_drp_backoff_win *bow = (struct uwb_drp_backoff_win *)arg;
202 struct uwb_rc *rc = container_of(bow, struct uwb_rc, bow);
203 struct device *dev = &rc->uwb_dev.dev;
David Vrabel8cc13a02008-09-17 16:34:09 +0100204
Stefano Panella5b377172008-12-12 13:00:06 +0000205 bow->can_reserve_extra_mases = true;
206 if (bow->total_expired <= 4) {
207 bow->total_expired++;
208 } else {
209 /* after 4 backoff window has expired we can exit from
210 * the backoff procedure */
211 bow->total_expired = 0;
212 bow->window = UWB_DRP_BACKOFF_WIN_MIN >> 1;
David Vrabel8cc13a02008-09-17 16:34:09 +0100213 }
Stefano Panella5b377172008-12-12 13:00:06 +0000214 dev_dbg(dev, "backoff_win_timer total_expired=%d, n=%d\n: ", bow->total_expired, bow->n);
215
216 /* try to relocate all the "to be moved" relocations */
217 uwb_rsv_handle_drp_avail_change(rc);
David Vrabel8cc13a02008-09-17 16:34:09 +0100218}
219
Stefano Panella5b377172008-12-12 13:00:06 +0000220void uwb_rsv_backoff_win_increment(struct uwb_rc *rc)
David Vrabel8cc13a02008-09-17 16:34:09 +0100221{
Stefano Panella5b377172008-12-12 13:00:06 +0000222 struct uwb_drp_backoff_win *bow = &rc->bow;
223 struct device *dev = &rc->uwb_dev.dev;
224 unsigned timeout_us;
David Vrabel8cc13a02008-09-17 16:34:09 +0100225
Stefano Panella5b377172008-12-12 13:00:06 +0000226 dev_dbg(dev, "backoff_win_increment: window=%d\n", bow->window);
David Vrabel8cc13a02008-09-17 16:34:09 +0100227
Stefano Panella5b377172008-12-12 13:00:06 +0000228 bow->can_reserve_extra_mases = false;
David Vrabel8cc13a02008-09-17 16:34:09 +0100229
Stefano Panella5b377172008-12-12 13:00:06 +0000230 if((bow->window << 1) == UWB_DRP_BACKOFF_WIN_MAX)
231 return;
David Vrabel8cc13a02008-09-17 16:34:09 +0100232
Stefano Panella5b377172008-12-12 13:00:06 +0000233 bow->window <<= 1;
Akinobu Mita78782022013-04-29 16:21:34 -0700234 bow->n = prandom_u32() & (bow->window - 1);
Stefano Panella5b377172008-12-12 13:00:06 +0000235 dev_dbg(dev, "new_window=%d, n=%d\n: ", bow->window, bow->n);
David Vrabel8cc13a02008-09-17 16:34:09 +0100236
Stefano Panella5b377172008-12-12 13:00:06 +0000237 /* reset the timer associated variables */
238 timeout_us = bow->n * UWB_SUPERFRAME_LENGTH_US;
239 bow->total_expired = 0;
240 mod_timer(&bow->timer, jiffies + usecs_to_jiffies(timeout_us));
David Vrabel8cc13a02008-09-17 16:34:09 +0100241}
242
243static void uwb_rsv_stroke_timer(struct uwb_rsv *rsv)
244{
245 int sframes = UWB_MAX_LOST_BEACONS;
246
247 /*
248 * Multicast reservations can become established within 1
249 * super frame and should not be terminated if no response is
250 * received.
251 */
252 if (rsv->is_multicast) {
Stefano Panella5b377172008-12-12 13:00:06 +0000253 if (rsv->state == UWB_RSV_STATE_O_INITIATED
254 || rsv->state == UWB_RSV_STATE_O_MOVE_EXPANDING
255 || rsv->state == UWB_RSV_STATE_O_MOVE_COMBINING
256 || rsv->state == UWB_RSV_STATE_O_MOVE_REDUCING)
David Vrabel8cc13a02008-09-17 16:34:09 +0100257 sframes = 1;
258 if (rsv->state == UWB_RSV_STATE_O_ESTABLISHED)
259 sframes = 0;
Stefano Panella5b377172008-12-12 13:00:06 +0000260
David Vrabel8cc13a02008-09-17 16:34:09 +0100261 }
262
David Vrabel8cc13a02008-09-17 16:34:09 +0100263 if (sframes > 0) {
264 /*
265 * Add an additional 2 superframes to account for the
266 * time to send the SET DRP IE command.
267 */
268 unsigned timeout_us = (sframes + 2) * UWB_SUPERFRAME_LENGTH_US;
269 mod_timer(&rsv->timer, jiffies + usecs_to_jiffies(timeout_us));
270 } else
271 del_timer(&rsv->timer);
272}
273
274/*
275 * Update a reservations state, and schedule an update of the
276 * transmitted DRP IEs.
277 */
278static void uwb_rsv_state_update(struct uwb_rsv *rsv,
279 enum uwb_rsv_state new_state)
280{
281 rsv->state = new_state;
282 rsv->ie_valid = false;
283
Stefano Panella5b377172008-12-12 13:00:06 +0000284 uwb_rsv_dump("SU", rsv);
David Vrabel8cc13a02008-09-17 16:34:09 +0100285
286 uwb_rsv_stroke_timer(rsv);
287 uwb_rsv_sched_update(rsv->rc);
288}
289
290static void uwb_rsv_callback(struct uwb_rsv *rsv)
291{
292 if (rsv->callback)
293 rsv->callback(rsv);
294}
295
296void uwb_rsv_set_state(struct uwb_rsv *rsv, enum uwb_rsv_state new_state)
297{
Stefano Panella5b377172008-12-12 13:00:06 +0000298 struct uwb_rsv_move *mv = &rsv->mv;
299
David Vrabel8cc13a02008-09-17 16:34:09 +0100300 if (rsv->state == new_state) {
301 switch (rsv->state) {
302 case UWB_RSV_STATE_O_ESTABLISHED:
Stefano Panella5b377172008-12-12 13:00:06 +0000303 case UWB_RSV_STATE_O_MOVE_EXPANDING:
304 case UWB_RSV_STATE_O_MOVE_COMBINING:
305 case UWB_RSV_STATE_O_MOVE_REDUCING:
David Vrabel8cc13a02008-09-17 16:34:09 +0100306 case UWB_RSV_STATE_T_ACCEPTED:
Stefano Panella5b377172008-12-12 13:00:06 +0000307 case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
308 case UWB_RSV_STATE_T_RESIZED:
David Vrabel8cc13a02008-09-17 16:34:09 +0100309 case UWB_RSV_STATE_NONE:
310 uwb_rsv_stroke_timer(rsv);
311 break;
312 default:
313 /* Expecting a state transition so leave timer
314 as-is. */
315 break;
316 }
317 return;
318 }
319
Stefano Panella5b377172008-12-12 13:00:06 +0000320 uwb_rsv_dump("SC", rsv);
321
David Vrabel8cc13a02008-09-17 16:34:09 +0100322 switch (new_state) {
323 case UWB_RSV_STATE_NONE:
David Vrabel8cc13a02008-09-17 16:34:09 +0100324 uwb_rsv_state_update(rsv, UWB_RSV_STATE_NONE);
325 uwb_rsv_callback(rsv);
326 break;
327 case UWB_RSV_STATE_O_INITIATED:
328 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_INITIATED);
329 break;
330 case UWB_RSV_STATE_O_PENDING:
331 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_PENDING);
332 break;
Stefano Panella5b377172008-12-12 13:00:06 +0000333 case UWB_RSV_STATE_O_MODIFIED:
334 /* in the companion there are the MASes to drop */
335 bitmap_andnot(rsv->mas.bm, rsv->mas.bm, mv->companion_mas.bm, UWB_NUM_MAS);
336 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MODIFIED);
337 break;
David Vrabel8cc13a02008-09-17 16:34:09 +0100338 case UWB_RSV_STATE_O_ESTABLISHED:
Stefano Panella5b377172008-12-12 13:00:06 +0000339 if (rsv->state == UWB_RSV_STATE_O_MODIFIED
340 || rsv->state == UWB_RSV_STATE_O_MOVE_REDUCING) {
341 uwb_drp_avail_release(rsv->rc, &mv->companion_mas);
342 rsv->needs_release_companion_mas = false;
343 }
David Vrabel8cc13a02008-09-17 16:34:09 +0100344 uwb_drp_avail_reserve(rsv->rc, &rsv->mas);
345 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_ESTABLISHED);
346 uwb_rsv_callback(rsv);
347 break;
Stefano Panella5b377172008-12-12 13:00:06 +0000348 case UWB_RSV_STATE_O_MOVE_EXPANDING:
349 rsv->needs_release_companion_mas = true;
350 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_EXPANDING);
351 break;
352 case UWB_RSV_STATE_O_MOVE_COMBINING:
353 rsv->needs_release_companion_mas = false;
354 uwb_drp_avail_reserve(rsv->rc, &mv->companion_mas);
355 bitmap_or(rsv->mas.bm, rsv->mas.bm, mv->companion_mas.bm, UWB_NUM_MAS);
356 rsv->mas.safe += mv->companion_mas.safe;
357 rsv->mas.unsafe += mv->companion_mas.unsafe;
358 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_COMBINING);
359 break;
360 case UWB_RSV_STATE_O_MOVE_REDUCING:
361 bitmap_andnot(mv->companion_mas.bm, rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS);
362 rsv->needs_release_companion_mas = true;
363 rsv->mas.safe = mv->final_mas.safe;
364 rsv->mas.unsafe = mv->final_mas.unsafe;
365 bitmap_copy(rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS);
366 bitmap_copy(rsv->mas.unsafe_bm, mv->final_mas.unsafe_bm, UWB_NUM_MAS);
367 uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_REDUCING);
368 break;
David Vrabel8cc13a02008-09-17 16:34:09 +0100369 case UWB_RSV_STATE_T_ACCEPTED:
Stefano Panella5b377172008-12-12 13:00:06 +0000370 case UWB_RSV_STATE_T_RESIZED:
371 rsv->needs_release_companion_mas = false;
David Vrabel8cc13a02008-09-17 16:34:09 +0100372 uwb_drp_avail_reserve(rsv->rc, &rsv->mas);
373 uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_ACCEPTED);
374 uwb_rsv_callback(rsv);
375 break;
376 case UWB_RSV_STATE_T_DENIED:
377 uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_DENIED);
378 break;
Stefano Panella5b377172008-12-12 13:00:06 +0000379 case UWB_RSV_STATE_T_CONFLICT:
380 uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_CONFLICT);
381 break;
382 case UWB_RSV_STATE_T_PENDING:
383 uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_PENDING);
384 break;
385 case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
386 rsv->needs_release_companion_mas = true;
387 uwb_drp_avail_reserve(rsv->rc, &mv->companion_mas);
388 uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_EXPANDING_ACCEPTED);
389 break;
David Vrabel8cc13a02008-09-17 16:34:09 +0100390 default:
391 dev_err(&rsv->rc->uwb_dev.dev, "unhandled state: %s (%d)\n",
392 uwb_rsv_state_str(new_state), new_state);
393 }
394}
395
Stefano Panella5b377172008-12-12 13:00:06 +0000396static void uwb_rsv_handle_timeout_work(struct work_struct *work)
397{
398 struct uwb_rsv *rsv = container_of(work, struct uwb_rsv,
399 handle_timeout_work);
400 struct uwb_rc *rc = rsv->rc;
401
402 mutex_lock(&rc->rsvs_mutex);
403
404 uwb_rsv_dump("TO", rsv);
405
406 switch (rsv->state) {
407 case UWB_RSV_STATE_O_INITIATED:
408 if (rsv->is_multicast) {
409 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_ESTABLISHED);
410 goto unlock;
411 }
412 break;
413 case UWB_RSV_STATE_O_MOVE_EXPANDING:
414 if (rsv->is_multicast) {
415 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_COMBINING);
416 goto unlock;
417 }
418 break;
419 case UWB_RSV_STATE_O_MOVE_COMBINING:
420 if (rsv->is_multicast) {
421 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_REDUCING);
422 goto unlock;
423 }
424 break;
425 case UWB_RSV_STATE_O_MOVE_REDUCING:
426 if (rsv->is_multicast) {
427 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_ESTABLISHED);
428 goto unlock;
429 }
430 break;
431 case UWB_RSV_STATE_O_ESTABLISHED:
432 if (rsv->is_multicast)
433 goto unlock;
434 break;
435 case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
436 /*
437 * The time out could be for the main or of the
438 * companion DRP, assume it's for the companion and
439 * drop that first. A further time out is required to
440 * drop the main.
441 */
442 uwb_rsv_set_state(rsv, UWB_RSV_STATE_T_ACCEPTED);
443 uwb_drp_avail_release(rsv->rc, &rsv->mv.companion_mas);
444 goto unlock;
445 default:
446 break;
447 }
448
449 uwb_rsv_remove(rsv);
450
451unlock:
452 mutex_unlock(&rc->rsvs_mutex);
453}
454
David Vrabel8cc13a02008-09-17 16:34:09 +0100455static struct uwb_rsv *uwb_rsv_alloc(struct uwb_rc *rc)
456{
457 struct uwb_rsv *rsv;
458
459 rsv = kzalloc(sizeof(struct uwb_rsv), GFP_KERNEL);
460 if (!rsv)
461 return NULL;
462
463 INIT_LIST_HEAD(&rsv->rc_node);
464 INIT_LIST_HEAD(&rsv->pal_node);
David Vrabelcae1c112008-10-27 15:22:46 +0000465 kref_init(&rsv->kref);
David Vrabel8cc13a02008-09-17 16:34:09 +0100466 init_timer(&rsv->timer);
467 rsv->timer.function = uwb_rsv_timer;
468 rsv->timer.data = (unsigned long)rsv;
469
470 rsv->rc = rc;
Stefano Panella5b377172008-12-12 13:00:06 +0000471 INIT_WORK(&rsv->handle_timeout_work, uwb_rsv_handle_timeout_work);
David Vrabel8cc13a02008-09-17 16:34:09 +0100472
473 return rsv;
474}
475
David Vrabel8cc13a02008-09-17 16:34:09 +0100476/**
477 * uwb_rsv_create - allocate and initialize a UWB reservation structure
478 * @rc: the radio controller
479 * @cb: callback to use when the reservation completes or terminates
480 * @pal_priv: data private to the PAL to be passed in the callback
481 *
482 * The callback is called when the state of the reservation changes from:
483 *
484 * - pending to accepted
485 * - pending to denined
486 * - accepted to terminated
487 * - pending to terminated
488 */
489struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb, void *pal_priv)
490{
491 struct uwb_rsv *rsv;
492
493 rsv = uwb_rsv_alloc(rc);
494 if (!rsv)
495 return NULL;
496
497 rsv->callback = cb;
498 rsv->pal_priv = pal_priv;
499
500 return rsv;
501}
502EXPORT_SYMBOL_GPL(uwb_rsv_create);
503
504void uwb_rsv_remove(struct uwb_rsv *rsv)
505{
Stefano Panella5b377172008-12-12 13:00:06 +0000506 uwb_rsv_dump("RM", rsv);
507
David Vrabel8cc13a02008-09-17 16:34:09 +0100508 if (rsv->state != UWB_RSV_STATE_NONE)
509 uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
Stefano Panella5b377172008-12-12 13:00:06 +0000510
511 if (rsv->needs_release_companion_mas)
512 uwb_drp_avail_release(rsv->rc, &rsv->mv.companion_mas);
513 uwb_drp_avail_release(rsv->rc, &rsv->mas);
514
515 if (uwb_rsv_is_owner(rsv))
516 uwb_rsv_put_stream(rsv);
David Vrabel2226b1c2009-01-06 17:55:32 +0000517
David Vrabelcae1c112008-10-27 15:22:46 +0000518 uwb_dev_put(rsv->owner);
519 if (rsv->target.type == UWB_RSV_TARGET_DEV)
520 uwb_dev_put(rsv->target.dev);
521
522 list_del_init(&rsv->rc_node);
523 uwb_rsv_put(rsv);
David Vrabel8cc13a02008-09-17 16:34:09 +0100524}
525
526/**
527 * uwb_rsv_destroy - free a UWB reservation structure
528 * @rsv: the reservation to free
529 *
David Vrabelcae1c112008-10-27 15:22:46 +0000530 * The reservation must already be terminated.
David Vrabel8cc13a02008-09-17 16:34:09 +0100531 */
532void uwb_rsv_destroy(struct uwb_rsv *rsv)
533{
David Vrabelcae1c112008-10-27 15:22:46 +0000534 uwb_rsv_put(rsv);
David Vrabel8cc13a02008-09-17 16:34:09 +0100535}
536EXPORT_SYMBOL_GPL(uwb_rsv_destroy);
537
538/**
539 * usb_rsv_establish - start a reservation establishment
540 * @rsv: the reservation
541 *
542 * The PAL should fill in @rsv's owner, target, type, max_mas,
Stefano Panella5b377172008-12-12 13:00:06 +0000543 * min_mas, max_interval and is_multicast fields. If the target is a
David Vrabel8cc13a02008-09-17 16:34:09 +0100544 * uwb_dev it must be referenced.
545 *
546 * The reservation's callback will be called when the reservation is
547 * accepted, denied or times out.
548 */
549int uwb_rsv_establish(struct uwb_rsv *rsv)
550{
551 struct uwb_rc *rc = rsv->rc;
Stefano Panella5b377172008-12-12 13:00:06 +0000552 struct uwb_mas_bm available;
David Vrabel8cc13a02008-09-17 16:34:09 +0100553 int ret;
554
555 mutex_lock(&rc->rsvs_mutex);
David Vrabel8cc13a02008-09-17 16:34:09 +0100556 ret = uwb_rsv_get_stream(rsv);
557 if (ret)
558 goto out;
559
Akinobu Mita78782022013-04-29 16:21:34 -0700560 rsv->tiebreaker = prandom_u32() & 1;
Stefano Panella5b377172008-12-12 13:00:06 +0000561 /* get available mas bitmap */
562 uwb_drp_available(rc, &available);
563
564 ret = uwb_rsv_find_best_allocation(rsv, &available, &rsv->mas);
565 if (ret == UWB_RSV_ALLOC_NOT_FOUND) {
566 ret = -EBUSY;
567 uwb_rsv_put_stream(rsv);
568 goto out;
569 }
570
571 ret = uwb_drp_avail_reserve_pending(rc, &rsv->mas);
572 if (ret != 0) {
David Vrabel8cc13a02008-09-17 16:34:09 +0100573 uwb_rsv_put_stream(rsv);
574 goto out;
575 }
576
David Vrabelcae1c112008-10-27 15:22:46 +0000577 uwb_rsv_get(rsv);
David Vrabel8cc13a02008-09-17 16:34:09 +0100578 list_add_tail(&rsv->rc_node, &rc->reservations);
579 rsv->owner = &rc->uwb_dev;
580 uwb_dev_get(rsv->owner);
581 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_INITIATED);
582out:
583 mutex_unlock(&rc->rsvs_mutex);
584 return ret;
585}
586EXPORT_SYMBOL_GPL(uwb_rsv_establish);
587
588/**
589 * uwb_rsv_modify - modify an already established reservation
590 * @rsv: the reservation to modify
591 * @max_mas: new maximum MAS to reserve
592 * @min_mas: new minimum MAS to reserve
Stefano Panella5b377172008-12-12 13:00:06 +0000593 * @max_interval: new max_interval to use
David Vrabel8cc13a02008-09-17 16:34:09 +0100594 *
595 * FIXME: implement this once there are PALs that use it.
596 */
Stefano Panella5b377172008-12-12 13:00:06 +0000597int uwb_rsv_modify(struct uwb_rsv *rsv, int max_mas, int min_mas, int max_interval)
David Vrabel8cc13a02008-09-17 16:34:09 +0100598{
599 return -ENOSYS;
600}
601EXPORT_SYMBOL_GPL(uwb_rsv_modify);
602
Stefano Panella5b377172008-12-12 13:00:06 +0000603/*
604 * move an already established reservation (rc->rsvs_mutex must to be
605 * taken when tis function is called)
606 */
607int uwb_rsv_try_move(struct uwb_rsv *rsv, struct uwb_mas_bm *available)
608{
609 struct uwb_rc *rc = rsv->rc;
610 struct uwb_drp_backoff_win *bow = &rc->bow;
611 struct device *dev = &rc->uwb_dev.dev;
612 struct uwb_rsv_move *mv;
613 int ret = 0;
614
615 if (bow->can_reserve_extra_mases == false)
616 return -EBUSY;
617
618 mv = &rsv->mv;
619
620 if (uwb_rsv_find_best_allocation(rsv, available, &mv->final_mas) == UWB_RSV_ALLOC_FOUND) {
621
622 if (!bitmap_equal(rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS)) {
623 /* We want to move the reservation */
624 bitmap_andnot(mv->companion_mas.bm, mv->final_mas.bm, rsv->mas.bm, UWB_NUM_MAS);
625 uwb_drp_avail_reserve_pending(rc, &mv->companion_mas);
626 uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_EXPANDING);
627 }
628 } else {
629 dev_dbg(dev, "new allocation not found\n");
630 }
631
632 return ret;
633}
634
635/* It will try to move every reservation in state O_ESTABLISHED giving
636 * to the MAS allocator algorithm an availability that is the real one
637 * plus the allocation already established from the reservation. */
638void uwb_rsv_handle_drp_avail_change(struct uwb_rc *rc)
639{
640 struct uwb_drp_backoff_win *bow = &rc->bow;
641 struct uwb_rsv *rsv;
642 struct uwb_mas_bm mas;
643
644 if (bow->can_reserve_extra_mases == false)
645 return;
646
647 list_for_each_entry(rsv, &rc->reservations, rc_node) {
648 if (rsv->state == UWB_RSV_STATE_O_ESTABLISHED ||
649 rsv->state == UWB_RSV_STATE_O_TO_BE_MOVED) {
650 uwb_drp_available(rc, &mas);
651 bitmap_or(mas.bm, mas.bm, rsv->mas.bm, UWB_NUM_MAS);
652 uwb_rsv_try_move(rsv, &mas);
653 }
654 }
655
656}
657
David Vrabel8cc13a02008-09-17 16:34:09 +0100658/**
659 * uwb_rsv_terminate - terminate an established reservation
660 * @rsv: the reservation to terminate
661 *
662 * A reservation is terminated by removing the DRP IE from the beacon,
663 * the other end will consider the reservation to be terminated when
664 * it does not see the DRP IE for at least mMaxLostBeacons.
665 *
666 * If applicable, the reference to the target uwb_dev will be released.
667 */
668void uwb_rsv_terminate(struct uwb_rsv *rsv)
669{
670 struct uwb_rc *rc = rsv->rc;
671
672 mutex_lock(&rc->rsvs_mutex);
673
David Vrabel671e4702008-12-12 13:23:24 +0000674 if (rsv->state != UWB_RSV_STATE_NONE)
675 uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
David Vrabel8cc13a02008-09-17 16:34:09 +0100676
677 mutex_unlock(&rc->rsvs_mutex);
678}
679EXPORT_SYMBOL_GPL(uwb_rsv_terminate);
680
681/**
682 * uwb_rsv_accept - accept a new reservation from a peer
683 * @rsv: the reservation
684 * @cb: call back for reservation changes
685 * @pal_priv: data to be passed in the above call back
686 *
687 * Reservation requests from peers are denied unless a PAL accepts it
688 * by calling this function.
David Vrabelcae1c112008-10-27 15:22:46 +0000689 *
690 * The PAL call uwb_rsv_destroy() for all accepted reservations before
691 * calling uwb_pal_unregister().
David Vrabel8cc13a02008-09-17 16:34:09 +0100692 */
693void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv)
694{
David Vrabelcae1c112008-10-27 15:22:46 +0000695 uwb_rsv_get(rsv);
696
David Vrabel8cc13a02008-09-17 16:34:09 +0100697 rsv->callback = cb;
698 rsv->pal_priv = pal_priv;
699 rsv->state = UWB_RSV_STATE_T_ACCEPTED;
700}
701EXPORT_SYMBOL_GPL(uwb_rsv_accept);
702
703/*
704 * Is a received DRP IE for this reservation?
705 */
706static bool uwb_rsv_match(struct uwb_rsv *rsv, struct uwb_dev *src,
707 struct uwb_ie_drp *drp_ie)
708{
709 struct uwb_dev_addr *rsv_src;
710 int stream;
711
712 stream = uwb_ie_drp_stream_index(drp_ie);
713
714 if (rsv->stream != stream)
715 return false;
716
717 switch (rsv->target.type) {
718 case UWB_RSV_TARGET_DEVADDR:
719 return rsv->stream == stream;
720 case UWB_RSV_TARGET_DEV:
721 if (uwb_ie_drp_owner(drp_ie))
722 rsv_src = &rsv->owner->dev_addr;
723 else
724 rsv_src = &rsv->target.dev->dev_addr;
725 return uwb_dev_addr_cmp(&src->dev_addr, rsv_src) == 0;
726 }
727 return false;
728}
729
730static struct uwb_rsv *uwb_rsv_new_target(struct uwb_rc *rc,
731 struct uwb_dev *src,
732 struct uwb_ie_drp *drp_ie)
733{
734 struct uwb_rsv *rsv;
735 struct uwb_pal *pal;
736 enum uwb_rsv_state state;
737
738 rsv = uwb_rsv_alloc(rc);
739 if (!rsv)
740 return NULL;
741
742 rsv->rc = rc;
743 rsv->owner = src;
744 uwb_dev_get(rsv->owner);
745 rsv->target.type = UWB_RSV_TARGET_DEV;
746 rsv->target.dev = &rc->uwb_dev;
Stefano Panella5b377172008-12-12 13:00:06 +0000747 uwb_dev_get(&rc->uwb_dev);
David Vrabel8cc13a02008-09-17 16:34:09 +0100748 rsv->type = uwb_ie_drp_type(drp_ie);
749 rsv->stream = uwb_ie_drp_stream_index(drp_ie);
David Vrabel8cc13a02008-09-17 16:34:09 +0100750 uwb_drp_ie_to_bm(&rsv->mas, drp_ie);
751
752 /*
753 * See if any PALs are interested in this reservation. If not,
754 * deny the request.
755 */
756 rsv->state = UWB_RSV_STATE_T_DENIED;
David Vrabel6fae35f2008-11-17 15:53:42 +0000757 mutex_lock(&rc->uwb_dev.mutex);
David Vrabel8cc13a02008-09-17 16:34:09 +0100758 list_for_each_entry(pal, &rc->pals, node) {
759 if (pal->new_rsv)
David Vrabele17be2b2008-11-17 15:24:14 +0000760 pal->new_rsv(pal, rsv);
David Vrabel8cc13a02008-09-17 16:34:09 +0100761 if (rsv->state == UWB_RSV_STATE_T_ACCEPTED)
762 break;
763 }
David Vrabel6fae35f2008-11-17 15:53:42 +0000764 mutex_unlock(&rc->uwb_dev.mutex);
David Vrabel8cc13a02008-09-17 16:34:09 +0100765
766 list_add_tail(&rsv->rc_node, &rc->reservations);
767 state = rsv->state;
768 rsv->state = UWB_RSV_STATE_NONE;
Stefano Panella5b377172008-12-12 13:00:06 +0000769
770 /* FIXME: do something sensible here */
771 if (state == UWB_RSV_STATE_T_ACCEPTED
772 && uwb_drp_avail_reserve_pending(rc, &rsv->mas) == -EBUSY) {
773 /* FIXME: do something sensible here */
774 } else {
775 uwb_rsv_set_state(rsv, state);
776 }
David Vrabel8cc13a02008-09-17 16:34:09 +0100777
778 return rsv;
779}
780
781/**
Stefano Panella5b377172008-12-12 13:00:06 +0000782 * uwb_rsv_get_usable_mas - get the bitmap of the usable MAS of a reservations
783 * @rsv: the reservation.
784 * @mas: returns the available MAS.
785 *
786 * The usable MAS of a reservation may be less than the negotiated MAS
787 * if alien BPs are present.
788 */
789void uwb_rsv_get_usable_mas(struct uwb_rsv *rsv, struct uwb_mas_bm *mas)
790{
791 bitmap_zero(mas->bm, UWB_NUM_MAS);
792 bitmap_andnot(mas->bm, rsv->mas.bm, rsv->rc->cnflt_alien_bitmap.bm, UWB_NUM_MAS);
793}
794EXPORT_SYMBOL_GPL(uwb_rsv_get_usable_mas);
795
796/**
David Vrabel8cc13a02008-09-17 16:34:09 +0100797 * uwb_rsv_find - find a reservation for a received DRP IE.
798 * @rc: the radio controller
799 * @src: source of the DRP IE
800 * @drp_ie: the DRP IE
801 *
802 * If the reservation cannot be found and the DRP IE is from a peer
803 * attempting to establish a new reservation, create a new reservation
804 * and add it to the list.
805 */
806struct uwb_rsv *uwb_rsv_find(struct uwb_rc *rc, struct uwb_dev *src,
807 struct uwb_ie_drp *drp_ie)
808{
809 struct uwb_rsv *rsv;
810
811 list_for_each_entry(rsv, &rc->reservations, rc_node) {
812 if (uwb_rsv_match(rsv, src, drp_ie))
813 return rsv;
814 }
815
816 if (uwb_ie_drp_owner(drp_ie))
817 return uwb_rsv_new_target(rc, src, drp_ie);
818
819 return NULL;
820}
821
822/*
823 * Go through all the reservations and check for timeouts and (if
824 * necessary) update their DRP IEs.
825 *
826 * FIXME: look at building the SET_DRP_IE command here rather than
827 * having to rescan the list in uwb_rc_send_all_drp_ie().
828 */
829static bool uwb_rsv_update_all(struct uwb_rc *rc)
830{
831 struct uwb_rsv *rsv, *t;
832 bool ie_updated = false;
833
834 list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
David Vrabel8cc13a02008-09-17 16:34:09 +0100835 if (!rsv->ie_valid) {
836 uwb_drp_ie_update(rsv);
837 ie_updated = true;
838 }
839 }
840
841 return ie_updated;
842}
843
Stefano Panella5b377172008-12-12 13:00:06 +0000844void uwb_rsv_queue_update(struct uwb_rc *rc)
845{
846 unsigned long delay_us = UWB_MAS_LENGTH_US * UWB_MAS_PER_ZONE;
847
848 queue_delayed_work(rc->rsv_workq, &rc->rsv_update_work, usecs_to_jiffies(delay_us));
849}
850
851/**
852 * uwb_rsv_sched_update - schedule an update of the DRP IEs
853 * @rc: the radio controller.
854 *
855 * To improve performance and ensure correctness with [ECMA-368] the
856 * number of SET-DRP-IE commands that are done are limited.
857 *
858 * DRP IEs update come from two sources: DRP events from the hardware
859 * which all occur at the beginning of the superframe ('syncronous'
860 * events) and reservation establishment/termination requests from
861 * PALs or timers ('asynchronous' events).
862 *
863 * A delayed work ensures that all the synchronous events result in
864 * one SET-DRP-IE command.
865 *
866 * Additional logic (the set_drp_ie_pending and rsv_updated_postponed
867 * flags) will prevent an asynchrous event starting a SET-DRP-IE
868 * command if one is currently awaiting a response.
869 *
870 * FIXME: this does leave a window where an asynchrous event can delay
871 * the SET-DRP-IE for a synchronous event by one superframe.
872 */
David Vrabel8cc13a02008-09-17 16:34:09 +0100873void uwb_rsv_sched_update(struct uwb_rc *rc)
874{
tpugliese31388872013-05-20 15:09:05 -0500875 spin_lock_irq(&rc->rsvs_lock);
Stefano Panella5b377172008-12-12 13:00:06 +0000876 if (!delayed_work_pending(&rc->rsv_update_work)) {
877 if (rc->set_drp_ie_pending > 0) {
878 rc->set_drp_ie_pending++;
879 goto unlock;
880 }
881 uwb_rsv_queue_update(rc);
882 }
883unlock:
tpugliese31388872013-05-20 15:09:05 -0500884 spin_unlock_irq(&rc->rsvs_lock);
David Vrabel8cc13a02008-09-17 16:34:09 +0100885}
886
887/*
888 * Update DRP IEs and, if necessary, the DRP Availability IE and send
889 * the updated IEs to the radio controller.
890 */
891static void uwb_rsv_update_work(struct work_struct *work)
892{
Stefano Panella5b377172008-12-12 13:00:06 +0000893 struct uwb_rc *rc = container_of(work, struct uwb_rc,
894 rsv_update_work.work);
David Vrabel8cc13a02008-09-17 16:34:09 +0100895 bool ie_updated;
896
897 mutex_lock(&rc->rsvs_mutex);
898
899 ie_updated = uwb_rsv_update_all(rc);
900
901 if (!rc->drp_avail.ie_valid) {
902 uwb_drp_avail_ie_update(rc);
903 ie_updated = true;
904 }
905
Stefano Panella5b377172008-12-12 13:00:06 +0000906 if (ie_updated && (rc->set_drp_ie_pending == 0))
David Vrabel8cc13a02008-09-17 16:34:09 +0100907 uwb_rc_send_all_drp_ie(rc);
908
909 mutex_unlock(&rc->rsvs_mutex);
910}
911
Stefano Panella5b377172008-12-12 13:00:06 +0000912static void uwb_rsv_alien_bp_work(struct work_struct *work)
913{
914 struct uwb_rc *rc = container_of(work, struct uwb_rc,
915 rsv_alien_bp_work.work);
916 struct uwb_rsv *rsv;
917
918 mutex_lock(&rc->rsvs_mutex);
919
920 list_for_each_entry(rsv, &rc->reservations, rc_node) {
921 if (rsv->type != UWB_DRP_TYPE_ALIEN_BP) {
922 rsv->callback(rsv);
923 }
924 }
925
926 mutex_unlock(&rc->rsvs_mutex);
927}
928
David Vrabel8cc13a02008-09-17 16:34:09 +0100929static void uwb_rsv_timer(unsigned long arg)
930{
931 struct uwb_rsv *rsv = (struct uwb_rsv *)arg;
932
Stefano Panella5b377172008-12-12 13:00:06 +0000933 queue_work(rsv->rc->rsv_workq, &rsv->handle_timeout_work);
David Vrabel8cc13a02008-09-17 16:34:09 +0100934}
935
David Vrabel307ba6d2008-11-07 17:37:33 +0000936/**
937 * uwb_rsv_remove_all - remove all reservations
938 * @rc: the radio controller
939 *
940 * A DRP IE update is not done.
941 */
942void uwb_rsv_remove_all(struct uwb_rc *rc)
943{
944 struct uwb_rsv *rsv, *t;
945
946 mutex_lock(&rc->rsvs_mutex);
947 list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
David Vrabel2226b1c2009-01-06 17:55:32 +0000948 if (rsv->state != UWB_RSV_STATE_NONE)
949 uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
950 del_timer_sync(&rsv->timer);
David Vrabel307ba6d2008-11-07 17:37:33 +0000951 }
Stefano Panella5b377172008-12-12 13:00:06 +0000952 /* Cancel any postponed update. */
953 rc->set_drp_ie_pending = 0;
David Vrabel307ba6d2008-11-07 17:37:33 +0000954 mutex_unlock(&rc->rsvs_mutex);
955
Stefano Panella5b377172008-12-12 13:00:06 +0000956 cancel_delayed_work_sync(&rc->rsv_update_work);
David Vrabel2226b1c2009-01-06 17:55:32 +0000957 flush_workqueue(rc->rsv_workq);
958
959 mutex_lock(&rc->rsvs_mutex);
960 list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
961 uwb_rsv_remove(rsv);
962 }
963 mutex_unlock(&rc->rsvs_mutex);
David Vrabel307ba6d2008-11-07 17:37:33 +0000964}
965
David Vrabel8cc13a02008-09-17 16:34:09 +0100966void uwb_rsv_init(struct uwb_rc *rc)
967{
968 INIT_LIST_HEAD(&rc->reservations);
Stefano Panella5b377172008-12-12 13:00:06 +0000969 INIT_LIST_HEAD(&rc->cnflt_alien_list);
David Vrabel8cc13a02008-09-17 16:34:09 +0100970 mutex_init(&rc->rsvs_mutex);
Stefano Panella5b377172008-12-12 13:00:06 +0000971 spin_lock_init(&rc->rsvs_lock);
972 INIT_DELAYED_WORK(&rc->rsv_update_work, uwb_rsv_update_work);
973 INIT_DELAYED_WORK(&rc->rsv_alien_bp_work, uwb_rsv_alien_bp_work);
974 rc->bow.can_reserve_extra_mases = true;
975 rc->bow.total_expired = 0;
976 rc->bow.window = UWB_DRP_BACKOFF_WIN_MIN >> 1;
977 init_timer(&rc->bow.timer);
978 rc->bow.timer.function = uwb_rsv_backoff_win_timer;
979 rc->bow.timer.data = (unsigned long)&rc->bow;
David Vrabel8cc13a02008-09-17 16:34:09 +0100980
981 bitmap_complement(rc->uwb_dev.streams, rc->uwb_dev.streams, UWB_NUM_STREAMS);
982}
983
984int uwb_rsv_setup(struct uwb_rc *rc)
985{
986 char name[16];
987
988 snprintf(name, sizeof(name), "%s_rsvd", dev_name(&rc->uwb_dev.dev));
989 rc->rsv_workq = create_singlethread_workqueue(name);
990 if (rc->rsv_workq == NULL)
991 return -ENOMEM;
992
993 return 0;
994}
995
996void uwb_rsv_cleanup(struct uwb_rc *rc)
997{
David Vrabel307ba6d2008-11-07 17:37:33 +0000998 uwb_rsv_remove_all(rc);
David Vrabel8cc13a02008-09-17 16:34:09 +0100999 destroy_workqueue(rc->rsv_workq);
1000}