blob: 9968b1c7c03a21c99b6497f109d014475965430c [file] [log] [blame]
Johan Hedberg0857dd32014-12-19 13:40:20 +02001/*
2 BlueZ - Bluetooth protocol stack for Linux
3
4 Copyright (C) 2014 Intel Corporation
5
6 This program is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License version 2 as
8 published by the Free Software Foundation;
9
10 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
11 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
12 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
13 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
14 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
15 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
16 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
17 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
18
19 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
20 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
21 SOFTWARE IS DISCLAIMED.
22*/
23
Johan Hedbergb1a89172015-11-25 16:15:42 +020024#include <asm/unaligned.h>
25
Johan Hedberg0857dd32014-12-19 13:40:20 +020026#include <net/bluetooth/bluetooth.h>
27#include <net/bluetooth/hci_core.h>
Johan Hedbergf2252572015-11-18 12:49:20 +020028#include <net/bluetooth/mgmt.h>
Johan Hedberg0857dd32014-12-19 13:40:20 +020029
30#include "smp.h"
31#include "hci_request.h"
32
Johan Hedbergbe91cd02015-11-10 09:44:54 +020033#define HCI_REQ_DONE 0
34#define HCI_REQ_PEND 1
35#define HCI_REQ_CANCELED 2
36
Johan Hedberg0857dd32014-12-19 13:40:20 +020037void hci_req_init(struct hci_request *req, struct hci_dev *hdev)
38{
39 skb_queue_head_init(&req->cmd_q);
40 req->hdev = hdev;
41 req->err = 0;
42}
43
Johan Hedberge62144872015-04-02 13:41:08 +030044static int req_run(struct hci_request *req, hci_req_complete_t complete,
45 hci_req_complete_skb_t complete_skb)
Johan Hedberg0857dd32014-12-19 13:40:20 +020046{
47 struct hci_dev *hdev = req->hdev;
48 struct sk_buff *skb;
49 unsigned long flags;
50
51 BT_DBG("length %u", skb_queue_len(&req->cmd_q));
52
53 /* If an error occurred during request building, remove all HCI
54 * commands queued on the HCI request queue.
55 */
56 if (req->err) {
57 skb_queue_purge(&req->cmd_q);
58 return req->err;
59 }
60
61 /* Do not allow empty requests */
62 if (skb_queue_empty(&req->cmd_q))
63 return -ENODATA;
64
65 skb = skb_peek_tail(&req->cmd_q);
Johan Hedberg44d27132015-11-05 09:31:40 +020066 if (complete) {
67 bt_cb(skb)->hci.req_complete = complete;
68 } else if (complete_skb) {
69 bt_cb(skb)->hci.req_complete_skb = complete_skb;
70 bt_cb(skb)->hci.req_flags |= HCI_REQ_SKB;
71 }
Johan Hedberg0857dd32014-12-19 13:40:20 +020072
73 spin_lock_irqsave(&hdev->cmd_q.lock, flags);
74 skb_queue_splice_tail(&req->cmd_q, &hdev->cmd_q);
75 spin_unlock_irqrestore(&hdev->cmd_q.lock, flags);
76
77 queue_work(hdev->workqueue, &hdev->cmd_work);
78
79 return 0;
80}
81
Johan Hedberge62144872015-04-02 13:41:08 +030082int hci_req_run(struct hci_request *req, hci_req_complete_t complete)
83{
84 return req_run(req, complete, NULL);
85}
86
87int hci_req_run_skb(struct hci_request *req, hci_req_complete_skb_t complete)
88{
89 return req_run(req, NULL, complete);
90}
91
Johan Hedbergbe91cd02015-11-10 09:44:54 +020092static void hci_req_sync_complete(struct hci_dev *hdev, u8 result, u16 opcode,
93 struct sk_buff *skb)
94{
95 BT_DBG("%s result 0x%2.2x", hdev->name, result);
96
97 if (hdev->req_status == HCI_REQ_PEND) {
98 hdev->req_result = result;
99 hdev->req_status = HCI_REQ_DONE;
100 if (skb)
101 hdev->req_skb = skb_get(skb);
102 wake_up_interruptible(&hdev->req_wait_q);
103 }
104}
105
Johan Hedbergb5044302015-11-10 09:44:55 +0200106void hci_req_sync_cancel(struct hci_dev *hdev, int err)
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200107{
108 BT_DBG("%s err 0x%2.2x", hdev->name, err);
109
110 if (hdev->req_status == HCI_REQ_PEND) {
111 hdev->req_result = err;
112 hdev->req_status = HCI_REQ_CANCELED;
113 wake_up_interruptible(&hdev->req_wait_q);
114 }
115}
116
117struct sk_buff *__hci_cmd_sync_ev(struct hci_dev *hdev, u16 opcode, u32 plen,
118 const void *param, u8 event, u32 timeout)
119{
120 DECLARE_WAITQUEUE(wait, current);
121 struct hci_request req;
122 struct sk_buff *skb;
123 int err = 0;
124
125 BT_DBG("%s", hdev->name);
126
127 hci_req_init(&req, hdev);
128
129 hci_req_add_ev(&req, opcode, plen, param, event);
130
131 hdev->req_status = HCI_REQ_PEND;
132
133 add_wait_queue(&hdev->req_wait_q, &wait);
134 set_current_state(TASK_INTERRUPTIBLE);
135
136 err = hci_req_run_skb(&req, hci_req_sync_complete);
137 if (err < 0) {
138 remove_wait_queue(&hdev->req_wait_q, &wait);
139 set_current_state(TASK_RUNNING);
140 return ERR_PTR(err);
141 }
142
143 schedule_timeout(timeout);
144
145 remove_wait_queue(&hdev->req_wait_q, &wait);
146
147 if (signal_pending(current))
148 return ERR_PTR(-EINTR);
149
150 switch (hdev->req_status) {
151 case HCI_REQ_DONE:
152 err = -bt_to_errno(hdev->req_result);
153 break;
154
155 case HCI_REQ_CANCELED:
156 err = -hdev->req_result;
157 break;
158
159 default:
160 err = -ETIMEDOUT;
161 break;
162 }
163
164 hdev->req_status = hdev->req_result = 0;
165 skb = hdev->req_skb;
166 hdev->req_skb = NULL;
167
168 BT_DBG("%s end: err %d", hdev->name, err);
169
170 if (err < 0) {
171 kfree_skb(skb);
172 return ERR_PTR(err);
173 }
174
175 if (!skb)
176 return ERR_PTR(-ENODATA);
177
178 return skb;
179}
180EXPORT_SYMBOL(__hci_cmd_sync_ev);
181
182struct sk_buff *__hci_cmd_sync(struct hci_dev *hdev, u16 opcode, u32 plen,
183 const void *param, u32 timeout)
184{
185 return __hci_cmd_sync_ev(hdev, opcode, plen, param, 0, timeout);
186}
187EXPORT_SYMBOL(__hci_cmd_sync);
188
189/* Execute request and wait for completion. */
Johan Hedberga1d01db2015-11-11 08:11:25 +0200190int __hci_req_sync(struct hci_dev *hdev, int (*func)(struct hci_request *req,
191 unsigned long opt),
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200192 unsigned long opt, u32 timeout, u8 *hci_status)
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200193{
194 struct hci_request req;
195 DECLARE_WAITQUEUE(wait, current);
196 int err = 0;
197
198 BT_DBG("%s start", hdev->name);
199
200 hci_req_init(&req, hdev);
201
202 hdev->req_status = HCI_REQ_PEND;
203
Johan Hedberga1d01db2015-11-11 08:11:25 +0200204 err = func(&req, opt);
205 if (err) {
206 if (hci_status)
207 *hci_status = HCI_ERROR_UNSPECIFIED;
208 return err;
209 }
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200210
211 add_wait_queue(&hdev->req_wait_q, &wait);
212 set_current_state(TASK_INTERRUPTIBLE);
213
214 err = hci_req_run_skb(&req, hci_req_sync_complete);
215 if (err < 0) {
216 hdev->req_status = 0;
217
218 remove_wait_queue(&hdev->req_wait_q, &wait);
219 set_current_state(TASK_RUNNING);
220
221 /* ENODATA means the HCI request command queue is empty.
222 * This can happen when a request with conditionals doesn't
223 * trigger any commands to be sent. This is normal behavior
224 * and should not trigger an error return.
225 */
Johan Hedberg568f44f2015-11-23 14:40:47 +0200226 if (err == -ENODATA) {
227 if (hci_status)
228 *hci_status = 0;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200229 return 0;
Johan Hedberg568f44f2015-11-23 14:40:47 +0200230 }
231
232 if (hci_status)
233 *hci_status = HCI_ERROR_UNSPECIFIED;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200234
235 return err;
236 }
237
238 schedule_timeout(timeout);
239
240 remove_wait_queue(&hdev->req_wait_q, &wait);
241
242 if (signal_pending(current))
243 return -EINTR;
244
245 switch (hdev->req_status) {
246 case HCI_REQ_DONE:
247 err = -bt_to_errno(hdev->req_result);
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200248 if (hci_status)
249 *hci_status = hdev->req_result;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200250 break;
251
252 case HCI_REQ_CANCELED:
253 err = -hdev->req_result;
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200254 if (hci_status)
255 *hci_status = HCI_ERROR_UNSPECIFIED;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200256 break;
257
258 default:
259 err = -ETIMEDOUT;
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200260 if (hci_status)
261 *hci_status = HCI_ERROR_UNSPECIFIED;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200262 break;
263 }
264
Frederic Dalleau9afee942016-08-23 07:59:19 +0200265 kfree_skb(hdev->req_skb);
266 hdev->req_skb = NULL;
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200267 hdev->req_status = hdev->req_result = 0;
268
269 BT_DBG("%s end: err %d", hdev->name, err);
270
271 return err;
272}
273
Johan Hedberga1d01db2015-11-11 08:11:25 +0200274int hci_req_sync(struct hci_dev *hdev, int (*req)(struct hci_request *req,
275 unsigned long opt),
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200276 unsigned long opt, u32 timeout, u8 *hci_status)
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200277{
278 int ret;
279
280 if (!test_bit(HCI_UP, &hdev->flags))
281 return -ENETDOWN;
282
283 /* Serialize all requests */
Johan Hedbergb5044302015-11-10 09:44:55 +0200284 hci_req_sync_lock(hdev);
Johan Hedberg4ebeee22015-11-11 08:11:19 +0200285 ret = __hci_req_sync(hdev, req, opt, timeout, hci_status);
Johan Hedbergb5044302015-11-10 09:44:55 +0200286 hci_req_sync_unlock(hdev);
Johan Hedbergbe91cd02015-11-10 09:44:54 +0200287
288 return ret;
289}
290
Johan Hedberg0857dd32014-12-19 13:40:20 +0200291struct sk_buff *hci_prepare_cmd(struct hci_dev *hdev, u16 opcode, u32 plen,
292 const void *param)
293{
294 int len = HCI_COMMAND_HDR_SIZE + plen;
295 struct hci_command_hdr *hdr;
296 struct sk_buff *skb;
297
298 skb = bt_skb_alloc(len, GFP_ATOMIC);
299 if (!skb)
300 return NULL;
301
302 hdr = (struct hci_command_hdr *) skb_put(skb, HCI_COMMAND_HDR_SIZE);
303 hdr->opcode = cpu_to_le16(opcode);
304 hdr->plen = plen;
305
306 if (plen)
307 memcpy(skb_put(skb, plen), param, plen);
308
309 BT_DBG("skb len %d", skb->len);
310
Marcel Holtmannd79f34e2015-11-05 07:10:00 +0100311 hci_skb_pkt_type(skb) = HCI_COMMAND_PKT;
312 hci_skb_opcode(skb) = opcode;
Johan Hedberg0857dd32014-12-19 13:40:20 +0200313
314 return skb;
315}
316
317/* Queue a command to an asynchronous HCI request */
318void hci_req_add_ev(struct hci_request *req, u16 opcode, u32 plen,
319 const void *param, u8 event)
320{
321 struct hci_dev *hdev = req->hdev;
322 struct sk_buff *skb;
323
324 BT_DBG("%s opcode 0x%4.4x plen %d", hdev->name, opcode, plen);
325
326 /* If an error occurred during request building, there is no point in
327 * queueing the HCI command. We can simply return.
328 */
329 if (req->err)
330 return;
331
332 skb = hci_prepare_cmd(hdev, opcode, plen, param);
333 if (!skb) {
334 BT_ERR("%s no memory for command (opcode 0x%4.4x)",
335 hdev->name, opcode);
336 req->err = -ENOMEM;
337 return;
338 }
339
340 if (skb_queue_empty(&req->cmd_q))
Johan Hedberg44d27132015-11-05 09:31:40 +0200341 bt_cb(skb)->hci.req_flags |= HCI_REQ_START;
Johan Hedberg0857dd32014-12-19 13:40:20 +0200342
Marcel Holtmann242c0eb2015-10-25 22:45:53 +0100343 bt_cb(skb)->hci.req_event = event;
Johan Hedberg0857dd32014-12-19 13:40:20 +0200344
345 skb_queue_tail(&req->cmd_q, skb);
346}
347
348void hci_req_add(struct hci_request *req, u16 opcode, u32 plen,
349 const void *param)
350{
351 hci_req_add_ev(req, opcode, plen, param, 0);
352}
353
Johan Hedbergbf943cb2015-11-25 16:15:43 +0200354void __hci_req_write_fast_connectable(struct hci_request *req, bool enable)
355{
356 struct hci_dev *hdev = req->hdev;
357 struct hci_cp_write_page_scan_activity acp;
358 u8 type;
359
360 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
361 return;
362
363 if (hdev->hci_ver < BLUETOOTH_VER_1_2)
364 return;
365
366 if (enable) {
367 type = PAGE_SCAN_TYPE_INTERLACED;
368
369 /* 160 msec page scan interval */
370 acp.interval = cpu_to_le16(0x0100);
371 } else {
372 type = PAGE_SCAN_TYPE_STANDARD; /* default */
373
374 /* default 1.28 sec page scan */
375 acp.interval = cpu_to_le16(0x0800);
376 }
377
378 acp.window = cpu_to_le16(0x0012);
379
380 if (__cpu_to_le16(hdev->page_scan_interval) != acp.interval ||
381 __cpu_to_le16(hdev->page_scan_window) != acp.window)
382 hci_req_add(req, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY,
383 sizeof(acp), &acp);
384
385 if (hdev->page_scan_type != type)
386 hci_req_add(req, HCI_OP_WRITE_PAGE_SCAN_TYPE, 1, &type);
387}
388
Johan Hedberg196a5e92015-11-22 18:55:44 +0200389/* This function controls the background scanning based on hdev->pend_le_conns
390 * list. If there are pending LE connection we start the background scanning,
391 * otherwise we stop it.
392 *
393 * This function requires the caller holds hdev->lock.
394 */
395static void __hci_update_background_scan(struct hci_request *req)
396{
397 struct hci_dev *hdev = req->hdev;
398
399 if (!test_bit(HCI_UP, &hdev->flags) ||
400 test_bit(HCI_INIT, &hdev->flags) ||
401 hci_dev_test_flag(hdev, HCI_SETUP) ||
402 hci_dev_test_flag(hdev, HCI_CONFIG) ||
403 hci_dev_test_flag(hdev, HCI_AUTO_OFF) ||
404 hci_dev_test_flag(hdev, HCI_UNREGISTER))
405 return;
406
407 /* No point in doing scanning if LE support hasn't been enabled */
408 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
409 return;
410
411 /* If discovery is active don't interfere with it */
412 if (hdev->discovery.state != DISCOVERY_STOPPED)
413 return;
414
415 /* Reset RSSI and UUID filters when starting background scanning
416 * since these filters are meant for service discovery only.
417 *
418 * The Start Discovery and Start Service Discovery operations
419 * ensure to set proper values for RSSI threshold and UUID
420 * filter list. So it is safe to just reset them here.
421 */
422 hci_discovery_filter_clear(hdev);
423
424 if (list_empty(&hdev->pend_le_conns) &&
425 list_empty(&hdev->pend_le_reports)) {
426 /* If there is no pending LE connections or devices
427 * to be scanned for, we should stop the background
428 * scanning.
429 */
430
431 /* If controller is not scanning we are done. */
432 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
433 return;
434
435 hci_req_add_le_scan_disable(req);
436
437 BT_DBG("%s stopping background scanning", hdev->name);
438 } else {
439 /* If there is at least one pending LE connection, we should
440 * keep the background scan running.
441 */
442
443 /* If controller is connecting, we should not start scanning
444 * since some controllers are not able to scan and connect at
445 * the same time.
446 */
447 if (hci_lookup_le_connect(hdev))
448 return;
449
450 /* If controller is currently scanning, we stop it to ensure we
451 * don't miss any advertising (due to duplicates filter).
452 */
453 if (hci_dev_test_flag(hdev, HCI_LE_SCAN))
454 hci_req_add_le_scan_disable(req);
455
456 hci_req_add_le_passive_scan(req);
457
458 BT_DBG("%s starting background scanning", hdev->name);
459 }
460}
461
Johan Hedberg00cf5042015-11-25 16:15:41 +0200462void __hci_req_update_name(struct hci_request *req)
463{
464 struct hci_dev *hdev = req->hdev;
465 struct hci_cp_write_local_name cp;
466
467 memcpy(cp.name, hdev->dev_name, sizeof(cp.name));
468
469 hci_req_add(req, HCI_OP_WRITE_LOCAL_NAME, sizeof(cp), &cp);
470}
471
Johan Hedbergb1a89172015-11-25 16:15:42 +0200472#define PNP_INFO_SVCLASS_ID 0x1200
473
474static u8 *create_uuid16_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
475{
476 u8 *ptr = data, *uuids_start = NULL;
477 struct bt_uuid *uuid;
478
479 if (len < 4)
480 return ptr;
481
482 list_for_each_entry(uuid, &hdev->uuids, list) {
483 u16 uuid16;
484
485 if (uuid->size != 16)
486 continue;
487
488 uuid16 = get_unaligned_le16(&uuid->uuid[12]);
489 if (uuid16 < 0x1100)
490 continue;
491
492 if (uuid16 == PNP_INFO_SVCLASS_ID)
493 continue;
494
495 if (!uuids_start) {
496 uuids_start = ptr;
497 uuids_start[0] = 1;
498 uuids_start[1] = EIR_UUID16_ALL;
499 ptr += 2;
500 }
501
502 /* Stop if not enough space to put next UUID */
503 if ((ptr - data) + sizeof(u16) > len) {
504 uuids_start[1] = EIR_UUID16_SOME;
505 break;
506 }
507
508 *ptr++ = (uuid16 & 0x00ff);
509 *ptr++ = (uuid16 & 0xff00) >> 8;
510 uuids_start[0] += sizeof(uuid16);
511 }
512
513 return ptr;
514}
515
516static u8 *create_uuid32_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
517{
518 u8 *ptr = data, *uuids_start = NULL;
519 struct bt_uuid *uuid;
520
521 if (len < 6)
522 return ptr;
523
524 list_for_each_entry(uuid, &hdev->uuids, list) {
525 if (uuid->size != 32)
526 continue;
527
528 if (!uuids_start) {
529 uuids_start = ptr;
530 uuids_start[0] = 1;
531 uuids_start[1] = EIR_UUID32_ALL;
532 ptr += 2;
533 }
534
535 /* Stop if not enough space to put next UUID */
536 if ((ptr - data) + sizeof(u32) > len) {
537 uuids_start[1] = EIR_UUID32_SOME;
538 break;
539 }
540
541 memcpy(ptr, &uuid->uuid[12], sizeof(u32));
542 ptr += sizeof(u32);
543 uuids_start[0] += sizeof(u32);
544 }
545
546 return ptr;
547}
548
549static u8 *create_uuid128_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
550{
551 u8 *ptr = data, *uuids_start = NULL;
552 struct bt_uuid *uuid;
553
554 if (len < 18)
555 return ptr;
556
557 list_for_each_entry(uuid, &hdev->uuids, list) {
558 if (uuid->size != 128)
559 continue;
560
561 if (!uuids_start) {
562 uuids_start = ptr;
563 uuids_start[0] = 1;
564 uuids_start[1] = EIR_UUID128_ALL;
565 ptr += 2;
566 }
567
568 /* Stop if not enough space to put next UUID */
569 if ((ptr - data) + 16 > len) {
570 uuids_start[1] = EIR_UUID128_SOME;
571 break;
572 }
573
574 memcpy(ptr, uuid->uuid, 16);
575 ptr += 16;
576 uuids_start[0] += 16;
577 }
578
579 return ptr;
580}
581
582static void create_eir(struct hci_dev *hdev, u8 *data)
583{
584 u8 *ptr = data;
585 size_t name_len;
586
587 name_len = strlen(hdev->dev_name);
588
589 if (name_len > 0) {
590 /* EIR Data type */
591 if (name_len > 48) {
592 name_len = 48;
593 ptr[1] = EIR_NAME_SHORT;
594 } else
595 ptr[1] = EIR_NAME_COMPLETE;
596
597 /* EIR Data length */
598 ptr[0] = name_len + 1;
599
600 memcpy(ptr + 2, hdev->dev_name, name_len);
601
602 ptr += (name_len + 2);
603 }
604
605 if (hdev->inq_tx_power != HCI_TX_POWER_INVALID) {
606 ptr[0] = 2;
607 ptr[1] = EIR_TX_POWER;
608 ptr[2] = (u8) hdev->inq_tx_power;
609
610 ptr += 3;
611 }
612
613 if (hdev->devid_source > 0) {
614 ptr[0] = 9;
615 ptr[1] = EIR_DEVICE_ID;
616
617 put_unaligned_le16(hdev->devid_source, ptr + 2);
618 put_unaligned_le16(hdev->devid_vendor, ptr + 4);
619 put_unaligned_le16(hdev->devid_product, ptr + 6);
620 put_unaligned_le16(hdev->devid_version, ptr + 8);
621
622 ptr += 10;
623 }
624
625 ptr = create_uuid16_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
626 ptr = create_uuid32_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
627 ptr = create_uuid128_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
628}
629
630void __hci_req_update_eir(struct hci_request *req)
631{
632 struct hci_dev *hdev = req->hdev;
633 struct hci_cp_write_eir cp;
634
635 if (!hdev_is_powered(hdev))
636 return;
637
638 if (!lmp_ext_inq_capable(hdev))
639 return;
640
641 if (!hci_dev_test_flag(hdev, HCI_SSP_ENABLED))
642 return;
643
644 if (hci_dev_test_flag(hdev, HCI_SERVICE_CACHE))
645 return;
646
647 memset(&cp, 0, sizeof(cp));
648
649 create_eir(hdev, cp.data);
650
651 if (memcmp(cp.data, hdev->eir, sizeof(cp.data)) == 0)
652 return;
653
654 memcpy(hdev->eir, cp.data, sizeof(cp.data));
655
656 hci_req_add(req, HCI_OP_WRITE_EIR, sizeof(cp), &cp);
657}
658
Johan Hedberg0857dd32014-12-19 13:40:20 +0200659void hci_req_add_le_scan_disable(struct hci_request *req)
660{
661 struct hci_cp_le_set_scan_enable cp;
662
663 memset(&cp, 0, sizeof(cp));
664 cp.enable = LE_SCAN_DISABLE;
665 hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
666}
667
668static void add_to_white_list(struct hci_request *req,
669 struct hci_conn_params *params)
670{
671 struct hci_cp_le_add_to_white_list cp;
672
673 cp.bdaddr_type = params->addr_type;
674 bacpy(&cp.bdaddr, &params->addr);
675
676 hci_req_add(req, HCI_OP_LE_ADD_TO_WHITE_LIST, sizeof(cp), &cp);
677}
678
679static u8 update_white_list(struct hci_request *req)
680{
681 struct hci_dev *hdev = req->hdev;
682 struct hci_conn_params *params;
683 struct bdaddr_list *b;
684 uint8_t white_list_entries = 0;
685
686 /* Go through the current white list programmed into the
687 * controller one by one and check if that address is still
688 * in the list of pending connections or list of devices to
689 * report. If not present in either list, then queue the
690 * command to remove it from the controller.
691 */
692 list_for_each_entry(b, &hdev->le_white_list, list) {
Johan Hedbergcff10ce2016-01-26 14:31:31 -0500693 /* If the device is neither in pend_le_conns nor
694 * pend_le_reports then remove it from the whitelist.
695 */
696 if (!hci_pend_le_action_lookup(&hdev->pend_le_conns,
697 &b->bdaddr, b->bdaddr_type) &&
698 !hci_pend_le_action_lookup(&hdev->pend_le_reports,
699 &b->bdaddr, b->bdaddr_type)) {
700 struct hci_cp_le_del_from_white_list cp;
Johan Hedberg0857dd32014-12-19 13:40:20 +0200701
Johan Hedbergcff10ce2016-01-26 14:31:31 -0500702 cp.bdaddr_type = b->bdaddr_type;
703 bacpy(&cp.bdaddr, &b->bdaddr);
704
705 hci_req_add(req, HCI_OP_LE_DEL_FROM_WHITE_LIST,
706 sizeof(cp), &cp);
Johan Hedberg0857dd32014-12-19 13:40:20 +0200707 continue;
708 }
709
Johan Hedbergcff10ce2016-01-26 14:31:31 -0500710 if (hci_find_irk_by_addr(hdev, &b->bdaddr, b->bdaddr_type)) {
711 /* White list can not be used with RPAs */
712 return 0x00;
713 }
Johan Hedberg0857dd32014-12-19 13:40:20 +0200714
Johan Hedbergcff10ce2016-01-26 14:31:31 -0500715 white_list_entries++;
Johan Hedberg0857dd32014-12-19 13:40:20 +0200716 }
717
718 /* Since all no longer valid white list entries have been
719 * removed, walk through the list of pending connections
720 * and ensure that any new device gets programmed into
721 * the controller.
722 *
723 * If the list of the devices is larger than the list of
724 * available white list entries in the controller, then
725 * just abort and return filer policy value to not use the
726 * white list.
727 */
728 list_for_each_entry(params, &hdev->pend_le_conns, action) {
729 if (hci_bdaddr_list_lookup(&hdev->le_white_list,
730 &params->addr, params->addr_type))
731 continue;
732
733 if (white_list_entries >= hdev->le_white_list_size) {
734 /* Select filter policy to accept all advertising */
735 return 0x00;
736 }
737
738 if (hci_find_irk_by_addr(hdev, &params->addr,
739 params->addr_type)) {
740 /* White list can not be used with RPAs */
741 return 0x00;
742 }
743
744 white_list_entries++;
745 add_to_white_list(req, params);
746 }
747
748 /* After adding all new pending connections, walk through
749 * the list of pending reports and also add these to the
750 * white list if there is still space.
751 */
752 list_for_each_entry(params, &hdev->pend_le_reports, action) {
753 if (hci_bdaddr_list_lookup(&hdev->le_white_list,
754 &params->addr, params->addr_type))
755 continue;
756
757 if (white_list_entries >= hdev->le_white_list_size) {
758 /* Select filter policy to accept all advertising */
759 return 0x00;
760 }
761
762 if (hci_find_irk_by_addr(hdev, &params->addr,
763 params->addr_type)) {
764 /* White list can not be used with RPAs */
765 return 0x00;
766 }
767
768 white_list_entries++;
769 add_to_white_list(req, params);
770 }
771
772 /* Select filter policy to use white list */
773 return 0x01;
774}
775
Johan Hedberg82a37ad2016-03-09 17:30:34 +0200776static bool scan_use_rpa(struct hci_dev *hdev)
777{
778 return hci_dev_test_flag(hdev, HCI_PRIVACY);
779}
780
Johan Hedberg0857dd32014-12-19 13:40:20 +0200781void hci_req_add_le_passive_scan(struct hci_request *req)
782{
783 struct hci_cp_le_set_scan_param param_cp;
784 struct hci_cp_le_set_scan_enable enable_cp;
785 struct hci_dev *hdev = req->hdev;
786 u8 own_addr_type;
787 u8 filter_policy;
788
789 /* Set require_privacy to false since no SCAN_REQ are send
790 * during passive scanning. Not using an non-resolvable address
791 * here is important so that peer devices using direct
792 * advertising with our address will be correctly reported
793 * by the controller.
794 */
Johan Hedberg82a37ad2016-03-09 17:30:34 +0200795 if (hci_update_random_address(req, false, scan_use_rpa(hdev),
796 &own_addr_type))
Johan Hedberg0857dd32014-12-19 13:40:20 +0200797 return;
798
799 /* Adding or removing entries from the white list must
800 * happen before enabling scanning. The controller does
801 * not allow white list modification while scanning.
802 */
803 filter_policy = update_white_list(req);
804
805 /* When the controller is using random resolvable addresses and
806 * with that having LE privacy enabled, then controllers with
807 * Extended Scanner Filter Policies support can now enable support
808 * for handling directed advertising.
809 *
810 * So instead of using filter polices 0x00 (no whitelist)
811 * and 0x01 (whitelist enabled) use the new filter policies
812 * 0x02 (no whitelist) and 0x03 (whitelist enabled).
813 */
Marcel Holtmannd7a5a112015-03-13 02:11:00 -0700814 if (hci_dev_test_flag(hdev, HCI_PRIVACY) &&
Johan Hedberg0857dd32014-12-19 13:40:20 +0200815 (hdev->le_features[0] & HCI_LE_EXT_SCAN_POLICY))
816 filter_policy |= 0x02;
817
818 memset(&param_cp, 0, sizeof(param_cp));
819 param_cp.type = LE_SCAN_PASSIVE;
820 param_cp.interval = cpu_to_le16(hdev->le_scan_interval);
821 param_cp.window = cpu_to_le16(hdev->le_scan_window);
822 param_cp.own_address_type = own_addr_type;
823 param_cp.filter_policy = filter_policy;
824 hci_req_add(req, HCI_OP_LE_SET_SCAN_PARAM, sizeof(param_cp),
825 &param_cp);
826
827 memset(&enable_cp, 0, sizeof(enable_cp));
828 enable_cp.enable = LE_SCAN_ENABLE;
829 enable_cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
830 hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(enable_cp),
831 &enable_cp);
832}
833
Johan Hedbergf2252572015-11-18 12:49:20 +0200834static u8 get_cur_adv_instance_scan_rsp_len(struct hci_dev *hdev)
835{
Johan Hedbergcab054a2015-11-30 11:21:45 +0200836 u8 instance = hdev->cur_adv_instance;
Johan Hedbergf2252572015-11-18 12:49:20 +0200837 struct adv_info *adv_instance;
838
839 /* Ignore instance 0 */
840 if (instance == 0x00)
841 return 0;
842
843 adv_instance = hci_find_adv_instance(hdev, instance);
844 if (!adv_instance)
845 return 0;
846
847 /* TODO: Take into account the "appearance" and "local-name" flags here.
848 * These are currently being ignored as they are not supported.
849 */
850 return adv_instance->scan_rsp_len;
851}
852
853void __hci_req_disable_advertising(struct hci_request *req)
854{
855 u8 enable = 0x00;
856
857 hci_req_add(req, HCI_OP_LE_SET_ADV_ENABLE, sizeof(enable), &enable);
858}
859
860static u32 get_adv_instance_flags(struct hci_dev *hdev, u8 instance)
861{
862 u32 flags;
863 struct adv_info *adv_instance;
864
865 if (instance == 0x00) {
866 /* Instance 0 always manages the "Tx Power" and "Flags"
867 * fields
868 */
869 flags = MGMT_ADV_FLAG_TX_POWER | MGMT_ADV_FLAG_MANAGED_FLAGS;
870
871 /* For instance 0, the HCI_ADVERTISING_CONNECTABLE setting
872 * corresponds to the "connectable" instance flag.
873 */
874 if (hci_dev_test_flag(hdev, HCI_ADVERTISING_CONNECTABLE))
875 flags |= MGMT_ADV_FLAG_CONNECTABLE;
876
Johan Hedberg6a19cc82016-03-11 09:56:32 +0200877 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE))
878 flags |= MGMT_ADV_FLAG_LIMITED_DISCOV;
879 else if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
Johan Hedbergd43efbd2016-03-09 17:30:33 +0200880 flags |= MGMT_ADV_FLAG_DISCOV;
881
Johan Hedbergf2252572015-11-18 12:49:20 +0200882 return flags;
883 }
884
885 adv_instance = hci_find_adv_instance(hdev, instance);
886
887 /* Return 0 when we got an invalid instance identifier. */
888 if (!adv_instance)
889 return 0;
890
891 return adv_instance->flags;
892}
893
Johan Hedberg82a37ad2016-03-09 17:30:34 +0200894static bool adv_use_rpa(struct hci_dev *hdev, uint32_t flags)
895{
896 /* If privacy is not enabled don't use RPA */
897 if (!hci_dev_test_flag(hdev, HCI_PRIVACY))
898 return false;
899
900 /* If basic privacy mode is enabled use RPA */
901 if (!hci_dev_test_flag(hdev, HCI_LIMITED_PRIVACY))
902 return true;
903
904 /* If limited privacy mode is enabled don't use RPA if we're
905 * both discoverable and bondable.
906 */
907 if ((flags & MGMT_ADV_FLAG_DISCOV) &&
908 hci_dev_test_flag(hdev, HCI_BONDABLE))
909 return false;
910
911 /* We're neither bondable nor discoverable in the limited
912 * privacy mode, therefore use RPA.
913 */
914 return true;
915}
916
Johan Hedbergf2252572015-11-18 12:49:20 +0200917void __hci_req_enable_advertising(struct hci_request *req)
918{
919 struct hci_dev *hdev = req->hdev;
920 struct hci_cp_le_set_adv_param cp;
921 u8 own_addr_type, enable = 0x01;
922 bool connectable;
Johan Hedbergf2252572015-11-18 12:49:20 +0200923 u32 flags;
924
925 if (hci_conn_num(hdev, LE_LINK) > 0)
926 return;
927
928 if (hci_dev_test_flag(hdev, HCI_LE_ADV))
929 __hci_req_disable_advertising(req);
930
931 /* Clear the HCI_LE_ADV bit temporarily so that the
932 * hci_update_random_address knows that it's safe to go ahead
933 * and write a new random address. The flag will be set back on
934 * as soon as the SET_ADV_ENABLE HCI command completes.
935 */
936 hci_dev_clear_flag(hdev, HCI_LE_ADV);
937
Johan Hedbergcab054a2015-11-30 11:21:45 +0200938 flags = get_adv_instance_flags(hdev, hdev->cur_adv_instance);
Johan Hedbergf2252572015-11-18 12:49:20 +0200939
940 /* If the "connectable" instance flag was not set, then choose between
941 * ADV_IND and ADV_NONCONN_IND based on the global connectable setting.
942 */
943 connectable = (flags & MGMT_ADV_FLAG_CONNECTABLE) ||
944 mgmt_get_connectable(hdev);
945
946 /* Set require_privacy to true only when non-connectable
947 * advertising is used. In that case it is fine to use a
948 * non-resolvable private address.
949 */
Johan Hedberg82a37ad2016-03-09 17:30:34 +0200950 if (hci_update_random_address(req, !connectable,
951 adv_use_rpa(hdev, flags),
952 &own_addr_type) < 0)
Johan Hedbergf2252572015-11-18 12:49:20 +0200953 return;
954
955 memset(&cp, 0, sizeof(cp));
956 cp.min_interval = cpu_to_le16(hdev->le_adv_min_interval);
957 cp.max_interval = cpu_to_le16(hdev->le_adv_max_interval);
958
959 if (connectable)
960 cp.type = LE_ADV_IND;
961 else if (get_cur_adv_instance_scan_rsp_len(hdev))
962 cp.type = LE_ADV_SCAN_IND;
963 else
964 cp.type = LE_ADV_NONCONN_IND;
965
966 cp.own_address_type = own_addr_type;
967 cp.channel_map = hdev->le_adv_channel_map;
968
969 hci_req_add(req, HCI_OP_LE_SET_ADV_PARAM, sizeof(cp), &cp);
970
971 hci_req_add(req, HCI_OP_LE_SET_ADV_ENABLE, sizeof(enable), &enable);
972}
973
974static u8 create_default_scan_rsp_data(struct hci_dev *hdev, u8 *ptr)
975{
976 u8 ad_len = 0;
977 size_t name_len;
978
979 name_len = strlen(hdev->dev_name);
980 if (name_len > 0) {
981 size_t max_len = HCI_MAX_AD_LENGTH - ad_len - 2;
982
983 if (name_len > max_len) {
984 name_len = max_len;
985 ptr[1] = EIR_NAME_SHORT;
986 } else
987 ptr[1] = EIR_NAME_COMPLETE;
988
989 ptr[0] = name_len + 1;
990
991 memcpy(ptr + 2, hdev->dev_name, name_len);
992
993 ad_len += (name_len + 2);
994 ptr += (name_len + 2);
995 }
996
997 return ad_len;
998}
999
1000static u8 create_instance_scan_rsp_data(struct hci_dev *hdev, u8 instance,
1001 u8 *ptr)
1002{
1003 struct adv_info *adv_instance;
1004
1005 adv_instance = hci_find_adv_instance(hdev, instance);
1006 if (!adv_instance)
1007 return 0;
1008
1009 /* TODO: Set the appropriate entries based on advertising instance flags
1010 * here once flags other than 0 are supported.
1011 */
1012 memcpy(ptr, adv_instance->scan_rsp_data,
1013 adv_instance->scan_rsp_len);
1014
1015 return adv_instance->scan_rsp_len;
1016}
1017
Johan Hedbergcab054a2015-11-30 11:21:45 +02001018void __hci_req_update_scan_rsp_data(struct hci_request *req, u8 instance)
Johan Hedbergf2252572015-11-18 12:49:20 +02001019{
1020 struct hci_dev *hdev = req->hdev;
1021 struct hci_cp_le_set_scan_rsp_data cp;
1022 u8 len;
1023
1024 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1025 return;
1026
1027 memset(&cp, 0, sizeof(cp));
1028
1029 if (instance)
1030 len = create_instance_scan_rsp_data(hdev, instance, cp.data);
1031 else
1032 len = create_default_scan_rsp_data(hdev, cp.data);
1033
1034 if (hdev->scan_rsp_data_len == len &&
1035 !memcmp(cp.data, hdev->scan_rsp_data, len))
1036 return;
1037
1038 memcpy(hdev->scan_rsp_data, cp.data, sizeof(cp.data));
1039 hdev->scan_rsp_data_len = len;
1040
1041 cp.length = len;
1042
1043 hci_req_add(req, HCI_OP_LE_SET_SCAN_RSP_DATA, sizeof(cp), &cp);
1044}
1045
Johan Hedbergf2252572015-11-18 12:49:20 +02001046static u8 create_instance_adv_data(struct hci_dev *hdev, u8 instance, u8 *ptr)
1047{
1048 struct adv_info *adv_instance = NULL;
1049 u8 ad_len = 0, flags = 0;
1050 u32 instance_flags;
1051
1052 /* Return 0 when the current instance identifier is invalid. */
1053 if (instance) {
1054 adv_instance = hci_find_adv_instance(hdev, instance);
1055 if (!adv_instance)
1056 return 0;
1057 }
1058
1059 instance_flags = get_adv_instance_flags(hdev, instance);
1060
1061 /* The Add Advertising command allows userspace to set both the general
1062 * and limited discoverable flags.
1063 */
1064 if (instance_flags & MGMT_ADV_FLAG_DISCOV)
1065 flags |= LE_AD_GENERAL;
1066
1067 if (instance_flags & MGMT_ADV_FLAG_LIMITED_DISCOV)
1068 flags |= LE_AD_LIMITED;
1069
Johan Hedbergf18ba582016-04-06 13:09:05 +03001070 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1071 flags |= LE_AD_NO_BREDR;
1072
Johan Hedbergf2252572015-11-18 12:49:20 +02001073 if (flags || (instance_flags & MGMT_ADV_FLAG_MANAGED_FLAGS)) {
1074 /* If a discovery flag wasn't provided, simply use the global
1075 * settings.
1076 */
1077 if (!flags)
1078 flags |= mgmt_get_adv_discov_flags(hdev);
1079
Johan Hedbergf2252572015-11-18 12:49:20 +02001080 /* If flags would still be empty, then there is no need to
1081 * include the "Flags" AD field".
1082 */
1083 if (flags) {
1084 ptr[0] = 0x02;
1085 ptr[1] = EIR_FLAGS;
1086 ptr[2] = flags;
1087
1088 ad_len += 3;
1089 ptr += 3;
1090 }
1091 }
1092
1093 if (adv_instance) {
1094 memcpy(ptr, adv_instance->adv_data,
1095 adv_instance->adv_data_len);
1096 ad_len += adv_instance->adv_data_len;
1097 ptr += adv_instance->adv_data_len;
1098 }
1099
1100 /* Provide Tx Power only if we can provide a valid value for it */
1101 if (hdev->adv_tx_power != HCI_TX_POWER_INVALID &&
1102 (instance_flags & MGMT_ADV_FLAG_TX_POWER)) {
1103 ptr[0] = 0x02;
1104 ptr[1] = EIR_TX_POWER;
1105 ptr[2] = (u8)hdev->adv_tx_power;
1106
1107 ad_len += 3;
1108 ptr += 3;
1109 }
1110
1111 return ad_len;
1112}
1113
Johan Hedbergcab054a2015-11-30 11:21:45 +02001114void __hci_req_update_adv_data(struct hci_request *req, u8 instance)
Johan Hedbergf2252572015-11-18 12:49:20 +02001115{
1116 struct hci_dev *hdev = req->hdev;
1117 struct hci_cp_le_set_adv_data cp;
1118 u8 len;
1119
1120 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1121 return;
1122
1123 memset(&cp, 0, sizeof(cp));
1124
1125 len = create_instance_adv_data(hdev, instance, cp.data);
1126
1127 /* There's nothing to do if the data hasn't changed */
1128 if (hdev->adv_data_len == len &&
1129 memcmp(cp.data, hdev->adv_data, len) == 0)
1130 return;
1131
1132 memcpy(hdev->adv_data, cp.data, sizeof(cp.data));
1133 hdev->adv_data_len = len;
1134
1135 cp.length = len;
1136
1137 hci_req_add(req, HCI_OP_LE_SET_ADV_DATA, sizeof(cp), &cp);
1138}
1139
Johan Hedbergcab054a2015-11-30 11:21:45 +02001140int hci_req_update_adv_data(struct hci_dev *hdev, u8 instance)
Johan Hedbergf2252572015-11-18 12:49:20 +02001141{
1142 struct hci_request req;
1143
1144 hci_req_init(&req, hdev);
1145 __hci_req_update_adv_data(&req, instance);
1146
1147 return hci_req_run(&req, NULL);
1148}
1149
1150static void adv_enable_complete(struct hci_dev *hdev, u8 status, u16 opcode)
1151{
1152 BT_DBG("%s status %u", hdev->name, status);
1153}
1154
1155void hci_req_reenable_advertising(struct hci_dev *hdev)
1156{
1157 struct hci_request req;
Johan Hedbergf2252572015-11-18 12:49:20 +02001158
1159 if (!hci_dev_test_flag(hdev, HCI_ADVERTISING) &&
Johan Hedberg17fd08f2015-11-26 12:15:59 +02001160 list_empty(&hdev->adv_instances))
Johan Hedbergf2252572015-11-18 12:49:20 +02001161 return;
1162
Johan Hedbergf2252572015-11-18 12:49:20 +02001163 hci_req_init(&req, hdev);
1164
Johan Hedbergcab054a2015-11-30 11:21:45 +02001165 if (hdev->cur_adv_instance) {
1166 __hci_req_schedule_adv_instance(&req, hdev->cur_adv_instance,
1167 true);
Johan Hedbergf2252572015-11-18 12:49:20 +02001168 } else {
Johan Hedbergcab054a2015-11-30 11:21:45 +02001169 __hci_req_update_adv_data(&req, 0x00);
1170 __hci_req_update_scan_rsp_data(&req, 0x00);
Johan Hedbergf2252572015-11-18 12:49:20 +02001171 __hci_req_enable_advertising(&req);
1172 }
1173
1174 hci_req_run(&req, adv_enable_complete);
1175}
1176
1177static void adv_timeout_expire(struct work_struct *work)
1178{
1179 struct hci_dev *hdev = container_of(work, struct hci_dev,
1180 adv_instance_expire.work);
1181
1182 struct hci_request req;
1183 u8 instance;
1184
1185 BT_DBG("%s", hdev->name);
1186
1187 hci_dev_lock(hdev);
1188
1189 hdev->adv_instance_timeout = 0;
1190
Johan Hedbergcab054a2015-11-30 11:21:45 +02001191 instance = hdev->cur_adv_instance;
Johan Hedbergf2252572015-11-18 12:49:20 +02001192 if (instance == 0x00)
1193 goto unlock;
1194
1195 hci_req_init(&req, hdev);
1196
Johan Hedberg37d3a1f2016-08-28 20:53:34 +03001197 hci_req_clear_adv_instance(hdev, NULL, &req, instance, false);
Johan Hedbergf2252572015-11-18 12:49:20 +02001198
1199 if (list_empty(&hdev->adv_instances))
1200 __hci_req_disable_advertising(&req);
1201
Johan Hedberg550a8ca2015-11-27 11:11:52 +02001202 hci_req_run(&req, NULL);
Johan Hedbergf2252572015-11-18 12:49:20 +02001203
1204unlock:
1205 hci_dev_unlock(hdev);
1206}
1207
1208int __hci_req_schedule_adv_instance(struct hci_request *req, u8 instance,
1209 bool force)
1210{
1211 struct hci_dev *hdev = req->hdev;
1212 struct adv_info *adv_instance = NULL;
1213 u16 timeout;
1214
1215 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
Johan Hedberg17fd08f2015-11-26 12:15:59 +02001216 list_empty(&hdev->adv_instances))
Johan Hedbergf2252572015-11-18 12:49:20 +02001217 return -EPERM;
1218
1219 if (hdev->adv_instance_timeout)
1220 return -EBUSY;
1221
1222 adv_instance = hci_find_adv_instance(hdev, instance);
1223 if (!adv_instance)
1224 return -ENOENT;
1225
1226 /* A zero timeout means unlimited advertising. As long as there is
1227 * only one instance, duration should be ignored. We still set a timeout
1228 * in case further instances are being added later on.
1229 *
1230 * If the remaining lifetime of the instance is more than the duration
1231 * then the timeout corresponds to the duration, otherwise it will be
1232 * reduced to the remaining instance lifetime.
1233 */
1234 if (adv_instance->timeout == 0 ||
1235 adv_instance->duration <= adv_instance->remaining_time)
1236 timeout = adv_instance->duration;
1237 else
1238 timeout = adv_instance->remaining_time;
1239
1240 /* The remaining time is being reduced unless the instance is being
1241 * advertised without time limit.
1242 */
1243 if (adv_instance->timeout)
1244 adv_instance->remaining_time =
1245 adv_instance->remaining_time - timeout;
1246
1247 hdev->adv_instance_timeout = timeout;
1248 queue_delayed_work(hdev->req_workqueue,
1249 &hdev->adv_instance_expire,
1250 msecs_to_jiffies(timeout * 1000));
1251
1252 /* If we're just re-scheduling the same instance again then do not
1253 * execute any HCI commands. This happens when a single instance is
1254 * being advertised.
1255 */
1256 if (!force && hdev->cur_adv_instance == instance &&
1257 hci_dev_test_flag(hdev, HCI_LE_ADV))
1258 return 0;
1259
1260 hdev->cur_adv_instance = instance;
Johan Hedbergcab054a2015-11-30 11:21:45 +02001261 __hci_req_update_adv_data(req, instance);
1262 __hci_req_update_scan_rsp_data(req, instance);
Johan Hedbergf2252572015-11-18 12:49:20 +02001263 __hci_req_enable_advertising(req);
1264
1265 return 0;
1266}
1267
1268static void cancel_adv_timeout(struct hci_dev *hdev)
1269{
1270 if (hdev->adv_instance_timeout) {
1271 hdev->adv_instance_timeout = 0;
1272 cancel_delayed_work(&hdev->adv_instance_expire);
1273 }
1274}
1275
1276/* For a single instance:
1277 * - force == true: The instance will be removed even when its remaining
1278 * lifetime is not zero.
1279 * - force == false: the instance will be deactivated but kept stored unless
1280 * the remaining lifetime is zero.
1281 *
1282 * For instance == 0x00:
1283 * - force == true: All instances will be removed regardless of their timeout
1284 * setting.
1285 * - force == false: Only instances that have a timeout will be removed.
1286 */
Johan Hedberg37d3a1f2016-08-28 20:53:34 +03001287void hci_req_clear_adv_instance(struct hci_dev *hdev, struct sock *sk,
1288 struct hci_request *req, u8 instance,
1289 bool force)
Johan Hedbergf2252572015-11-18 12:49:20 +02001290{
1291 struct adv_info *adv_instance, *n, *next_instance = NULL;
1292 int err;
1293 u8 rem_inst;
1294
1295 /* Cancel any timeout concerning the removed instance(s). */
1296 if (!instance || hdev->cur_adv_instance == instance)
1297 cancel_adv_timeout(hdev);
1298
1299 /* Get the next instance to advertise BEFORE we remove
1300 * the current one. This can be the same instance again
1301 * if there is only one instance.
1302 */
1303 if (instance && hdev->cur_adv_instance == instance)
1304 next_instance = hci_get_next_instance(hdev, instance);
1305
1306 if (instance == 0x00) {
1307 list_for_each_entry_safe(adv_instance, n, &hdev->adv_instances,
1308 list) {
1309 if (!(force || adv_instance->timeout))
1310 continue;
1311
1312 rem_inst = adv_instance->instance;
1313 err = hci_remove_adv_instance(hdev, rem_inst);
1314 if (!err)
Johan Hedberg37d3a1f2016-08-28 20:53:34 +03001315 mgmt_advertising_removed(sk, hdev, rem_inst);
Johan Hedbergf2252572015-11-18 12:49:20 +02001316 }
Johan Hedbergf2252572015-11-18 12:49:20 +02001317 } else {
1318 adv_instance = hci_find_adv_instance(hdev, instance);
1319
1320 if (force || (adv_instance && adv_instance->timeout &&
1321 !adv_instance->remaining_time)) {
1322 /* Don't advertise a removed instance. */
1323 if (next_instance &&
1324 next_instance->instance == instance)
1325 next_instance = NULL;
1326
1327 err = hci_remove_adv_instance(hdev, instance);
1328 if (!err)
Johan Hedberg37d3a1f2016-08-28 20:53:34 +03001329 mgmt_advertising_removed(sk, hdev, instance);
Johan Hedbergf2252572015-11-18 12:49:20 +02001330 }
1331 }
1332
Johan Hedbergf2252572015-11-18 12:49:20 +02001333 if (!req || !hdev_is_powered(hdev) ||
1334 hci_dev_test_flag(hdev, HCI_ADVERTISING))
1335 return;
1336
1337 if (next_instance)
1338 __hci_req_schedule_adv_instance(req, next_instance->instance,
1339 false);
1340}
1341
Johan Hedberg0857dd32014-12-19 13:40:20 +02001342static void set_random_addr(struct hci_request *req, bdaddr_t *rpa)
1343{
1344 struct hci_dev *hdev = req->hdev;
1345
1346 /* If we're advertising or initiating an LE connection we can't
1347 * go ahead and change the random address at this time. This is
1348 * because the eventual initiator address used for the
1349 * subsequently created connection will be undefined (some
1350 * controllers use the new address and others the one we had
1351 * when the operation started).
1352 *
1353 * In this kind of scenario skip the update and let the random
1354 * address be updated at the next cycle.
1355 */
Marcel Holtmannd7a5a112015-03-13 02:11:00 -07001356 if (hci_dev_test_flag(hdev, HCI_LE_ADV) ||
Jakub Pawlowskie7d9ab72015-08-07 20:22:52 +02001357 hci_lookup_le_connect(hdev)) {
Johan Hedberg0857dd32014-12-19 13:40:20 +02001358 BT_DBG("Deferring random address update");
Marcel Holtmanna1536da2015-03-13 02:11:01 -07001359 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED);
Johan Hedberg0857dd32014-12-19 13:40:20 +02001360 return;
1361 }
1362
1363 hci_req_add(req, HCI_OP_LE_SET_RANDOM_ADDR, 6, rpa);
1364}
1365
1366int hci_update_random_address(struct hci_request *req, bool require_privacy,
Johan Hedberg82a37ad2016-03-09 17:30:34 +02001367 bool use_rpa, u8 *own_addr_type)
Johan Hedberg0857dd32014-12-19 13:40:20 +02001368{
1369 struct hci_dev *hdev = req->hdev;
1370 int err;
1371
1372 /* If privacy is enabled use a resolvable private address. If
1373 * current RPA has expired or there is something else than
1374 * the current RPA in use, then generate a new one.
1375 */
Johan Hedberg82a37ad2016-03-09 17:30:34 +02001376 if (use_rpa) {
Johan Hedberg0857dd32014-12-19 13:40:20 +02001377 int to;
1378
1379 *own_addr_type = ADDR_LE_DEV_RANDOM;
1380
Marcel Holtmanna69d8922015-03-13 02:11:05 -07001381 if (!hci_dev_test_and_clear_flag(hdev, HCI_RPA_EXPIRED) &&
Johan Hedberg0857dd32014-12-19 13:40:20 +02001382 !bacmp(&hdev->random_addr, &hdev->rpa))
1383 return 0;
1384
1385 err = smp_generate_rpa(hdev, hdev->irk, &hdev->rpa);
1386 if (err < 0) {
1387 BT_ERR("%s failed to generate new RPA", hdev->name);
1388 return err;
1389 }
1390
1391 set_random_addr(req, &hdev->rpa);
1392
1393 to = msecs_to_jiffies(hdev->rpa_timeout * 1000);
1394 queue_delayed_work(hdev->workqueue, &hdev->rpa_expired, to);
1395
1396 return 0;
1397 }
1398
1399 /* In case of required privacy without resolvable private address,
1400 * use an non-resolvable private address. This is useful for active
1401 * scanning and non-connectable advertising.
1402 */
1403 if (require_privacy) {
1404 bdaddr_t nrpa;
1405
1406 while (true) {
1407 /* The non-resolvable private address is generated
1408 * from random six bytes with the two most significant
1409 * bits cleared.
1410 */
1411 get_random_bytes(&nrpa, 6);
1412 nrpa.b[5] &= 0x3f;
1413
1414 /* The non-resolvable private address shall not be
1415 * equal to the public address.
1416 */
1417 if (bacmp(&hdev->bdaddr, &nrpa))
1418 break;
1419 }
1420
1421 *own_addr_type = ADDR_LE_DEV_RANDOM;
1422 set_random_addr(req, &nrpa);
1423 return 0;
1424 }
1425
1426 /* If forcing static address is in use or there is no public
1427 * address use the static address as random address (but skip
1428 * the HCI command if the current random address is already the
1429 * static one.
Marcel Holtmann50b5b952014-12-19 23:05:35 +01001430 *
1431 * In case BR/EDR has been disabled on a dual-mode controller
1432 * and a static address has been configured, then use that
1433 * address instead of the public BR/EDR address.
Johan Hedberg0857dd32014-12-19 13:40:20 +02001434 */
Marcel Holtmannb7cb93e2015-03-13 10:20:35 -07001435 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) ||
Marcel Holtmann50b5b952014-12-19 23:05:35 +01001436 !bacmp(&hdev->bdaddr, BDADDR_ANY) ||
Marcel Holtmannd7a5a112015-03-13 02:11:00 -07001437 (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) &&
Marcel Holtmann50b5b952014-12-19 23:05:35 +01001438 bacmp(&hdev->static_addr, BDADDR_ANY))) {
Johan Hedberg0857dd32014-12-19 13:40:20 +02001439 *own_addr_type = ADDR_LE_DEV_RANDOM;
1440 if (bacmp(&hdev->static_addr, &hdev->random_addr))
1441 hci_req_add(req, HCI_OP_LE_SET_RANDOM_ADDR, 6,
1442 &hdev->static_addr);
1443 return 0;
1444 }
1445
1446 /* Neither privacy nor static address is being used so use a
1447 * public address.
1448 */
1449 *own_addr_type = ADDR_LE_DEV_PUBLIC;
1450
1451 return 0;
1452}
Johan Hedberg2cf22212014-12-19 22:26:00 +02001453
Johan Hedberg405a2612014-12-19 23:18:22 +02001454static bool disconnected_whitelist_entries(struct hci_dev *hdev)
1455{
1456 struct bdaddr_list *b;
1457
1458 list_for_each_entry(b, &hdev->whitelist, list) {
1459 struct hci_conn *conn;
1460
1461 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &b->bdaddr);
1462 if (!conn)
1463 return true;
1464
1465 if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
1466 return true;
1467 }
1468
1469 return false;
1470}
1471
Johan Hedberg01b1cb82015-11-16 12:52:21 +02001472void __hci_req_update_scan(struct hci_request *req)
Johan Hedberg405a2612014-12-19 23:18:22 +02001473{
1474 struct hci_dev *hdev = req->hdev;
1475 u8 scan;
1476
Marcel Holtmannd7a5a112015-03-13 02:11:00 -07001477 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
Johan Hedberg405a2612014-12-19 23:18:22 +02001478 return;
1479
1480 if (!hdev_is_powered(hdev))
1481 return;
1482
1483 if (mgmt_powering_down(hdev))
1484 return;
1485
Marcel Holtmannd7a5a112015-03-13 02:11:00 -07001486 if (hci_dev_test_flag(hdev, HCI_CONNECTABLE) ||
Johan Hedberg405a2612014-12-19 23:18:22 +02001487 disconnected_whitelist_entries(hdev))
1488 scan = SCAN_PAGE;
1489 else
1490 scan = SCAN_DISABLED;
1491
Marcel Holtmannd7a5a112015-03-13 02:11:00 -07001492 if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
Johan Hedberg405a2612014-12-19 23:18:22 +02001493 scan |= SCAN_INQUIRY;
1494
Johan Hedberg01b1cb82015-11-16 12:52:21 +02001495 if (test_bit(HCI_PSCAN, &hdev->flags) == !!(scan & SCAN_PAGE) &&
1496 test_bit(HCI_ISCAN, &hdev->flags) == !!(scan & SCAN_INQUIRY))
1497 return;
1498
Johan Hedberg405a2612014-12-19 23:18:22 +02001499 hci_req_add(req, HCI_OP_WRITE_SCAN_ENABLE, 1, &scan);
1500}
1501
Johan Hedberg01b1cb82015-11-16 12:52:21 +02001502static int update_scan(struct hci_request *req, unsigned long opt)
Johan Hedberg405a2612014-12-19 23:18:22 +02001503{
Johan Hedberg01b1cb82015-11-16 12:52:21 +02001504 hci_dev_lock(req->hdev);
1505 __hci_req_update_scan(req);
1506 hci_dev_unlock(req->hdev);
1507 return 0;
1508}
Johan Hedberg405a2612014-12-19 23:18:22 +02001509
Johan Hedberg01b1cb82015-11-16 12:52:21 +02001510static void scan_update_work(struct work_struct *work)
1511{
1512 struct hci_dev *hdev = container_of(work, struct hci_dev, scan_update);
1513
1514 hci_req_sync(hdev, update_scan, 0, HCI_CMD_TIMEOUT, NULL);
Johan Hedberg405a2612014-12-19 23:18:22 +02001515}
1516
Johan Hedberg53c0ba72015-11-22 16:43:43 +03001517static int connectable_update(struct hci_request *req, unsigned long opt)
1518{
1519 struct hci_dev *hdev = req->hdev;
1520
1521 hci_dev_lock(hdev);
1522
1523 __hci_req_update_scan(req);
1524
1525 /* If BR/EDR is not enabled and we disable advertising as a
1526 * by-product of disabling connectable, we need to update the
1527 * advertising flags.
1528 */
1529 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
Johan Hedbergcab054a2015-11-30 11:21:45 +02001530 __hci_req_update_adv_data(req, hdev->cur_adv_instance);
Johan Hedberg53c0ba72015-11-22 16:43:43 +03001531
1532 /* Update the advertising parameters if necessary */
1533 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
Johan Hedberg17fd08f2015-11-26 12:15:59 +02001534 !list_empty(&hdev->adv_instances))
Johan Hedberg53c0ba72015-11-22 16:43:43 +03001535 __hci_req_enable_advertising(req);
1536
1537 __hci_update_background_scan(req);
1538
1539 hci_dev_unlock(hdev);
1540
1541 return 0;
1542}
1543
1544static void connectable_update_work(struct work_struct *work)
1545{
1546 struct hci_dev *hdev = container_of(work, struct hci_dev,
1547 connectable_update);
1548 u8 status;
1549
1550 hci_req_sync(hdev, connectable_update, 0, HCI_CMD_TIMEOUT, &status);
1551 mgmt_set_connectable_complete(hdev, status);
1552}
1553
Johan Hedberg14bf5ea2015-11-22 19:00:22 +02001554static u8 get_service_classes(struct hci_dev *hdev)
1555{
1556 struct bt_uuid *uuid;
1557 u8 val = 0;
1558
1559 list_for_each_entry(uuid, &hdev->uuids, list)
1560 val |= uuid->svc_hint;
1561
1562 return val;
1563}
1564
1565void __hci_req_update_class(struct hci_request *req)
1566{
1567 struct hci_dev *hdev = req->hdev;
1568 u8 cod[3];
1569
1570 BT_DBG("%s", hdev->name);
1571
1572 if (!hdev_is_powered(hdev))
1573 return;
1574
1575 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1576 return;
1577
1578 if (hci_dev_test_flag(hdev, HCI_SERVICE_CACHE))
1579 return;
1580
1581 cod[0] = hdev->minor_class;
1582 cod[1] = hdev->major_class;
1583 cod[2] = get_service_classes(hdev);
1584
1585 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE))
1586 cod[1] |= 0x20;
1587
1588 if (memcmp(cod, hdev->dev_class, 3) == 0)
1589 return;
1590
1591 hci_req_add(req, HCI_OP_WRITE_CLASS_OF_DEV, sizeof(cod), cod);
1592}
1593
Johan Hedbergaed1a882015-11-22 17:24:44 +03001594static void write_iac(struct hci_request *req)
1595{
1596 struct hci_dev *hdev = req->hdev;
1597 struct hci_cp_write_current_iac_lap cp;
1598
1599 if (!hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
1600 return;
1601
1602 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE)) {
1603 /* Limited discoverable mode */
1604 cp.num_iac = min_t(u8, hdev->num_iac, 2);
1605 cp.iac_lap[0] = 0x00; /* LIAC */
1606 cp.iac_lap[1] = 0x8b;
1607 cp.iac_lap[2] = 0x9e;
1608 cp.iac_lap[3] = 0x33; /* GIAC */
1609 cp.iac_lap[4] = 0x8b;
1610 cp.iac_lap[5] = 0x9e;
1611 } else {
1612 /* General discoverable mode */
1613 cp.num_iac = 1;
1614 cp.iac_lap[0] = 0x33; /* GIAC */
1615 cp.iac_lap[1] = 0x8b;
1616 cp.iac_lap[2] = 0x9e;
1617 }
1618
1619 hci_req_add(req, HCI_OP_WRITE_CURRENT_IAC_LAP,
1620 (cp.num_iac * 3) + 1, &cp);
1621}
1622
1623static int discoverable_update(struct hci_request *req, unsigned long opt)
1624{
1625 struct hci_dev *hdev = req->hdev;
1626
1627 hci_dev_lock(hdev);
1628
1629 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) {
1630 write_iac(req);
1631 __hci_req_update_scan(req);
1632 __hci_req_update_class(req);
1633 }
1634
1635 /* Advertising instances don't use the global discoverable setting, so
1636 * only update AD if advertising was enabled using Set Advertising.
1637 */
Johan Hedberg82a37ad2016-03-09 17:30:34 +02001638 if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) {
Johan Hedbergcab054a2015-11-30 11:21:45 +02001639 __hci_req_update_adv_data(req, 0x00);
Johan Hedbergaed1a882015-11-22 17:24:44 +03001640
Johan Hedberg82a37ad2016-03-09 17:30:34 +02001641 /* Discoverable mode affects the local advertising
1642 * address in limited privacy mode.
1643 */
1644 if (hci_dev_test_flag(hdev, HCI_LIMITED_PRIVACY))
1645 __hci_req_enable_advertising(req);
1646 }
1647
Johan Hedbergaed1a882015-11-22 17:24:44 +03001648 hci_dev_unlock(hdev);
1649
1650 return 0;
1651}
1652
1653static void discoverable_update_work(struct work_struct *work)
1654{
1655 struct hci_dev *hdev = container_of(work, struct hci_dev,
1656 discoverable_update);
1657 u8 status;
1658
1659 hci_req_sync(hdev, discoverable_update, 0, HCI_CMD_TIMEOUT, &status);
1660 mgmt_set_discoverable_complete(hdev, status);
1661}
1662
Johan Hedbergdcc0f0d2015-10-22 10:49:37 +03001663void __hci_abort_conn(struct hci_request *req, struct hci_conn *conn,
1664 u8 reason)
1665{
1666 switch (conn->state) {
1667 case BT_CONNECTED:
1668 case BT_CONFIG:
1669 if (conn->type == AMP_LINK) {
1670 struct hci_cp_disconn_phy_link cp;
1671
1672 cp.phy_handle = HCI_PHY_HANDLE(conn->handle);
1673 cp.reason = reason;
1674 hci_req_add(req, HCI_OP_DISCONN_PHY_LINK, sizeof(cp),
1675 &cp);
1676 } else {
1677 struct hci_cp_disconnect dc;
1678
1679 dc.handle = cpu_to_le16(conn->handle);
1680 dc.reason = reason;
1681 hci_req_add(req, HCI_OP_DISCONNECT, sizeof(dc), &dc);
1682 }
1683
1684 conn->state = BT_DISCONN;
1685
1686 break;
1687 case BT_CONNECT:
1688 if (conn->type == LE_LINK) {
1689 if (test_bit(HCI_CONN_SCANNING, &conn->flags))
1690 break;
1691 hci_req_add(req, HCI_OP_LE_CREATE_CONN_CANCEL,
1692 0, NULL);
1693 } else if (conn->type == ACL_LINK) {
1694 if (req->hdev->hci_ver < BLUETOOTH_VER_1_2)
1695 break;
1696 hci_req_add(req, HCI_OP_CREATE_CONN_CANCEL,
1697 6, &conn->dst);
1698 }
1699 break;
1700 case BT_CONNECT2:
1701 if (conn->type == ACL_LINK) {
1702 struct hci_cp_reject_conn_req rej;
1703
1704 bacpy(&rej.bdaddr, &conn->dst);
1705 rej.reason = reason;
1706
1707 hci_req_add(req, HCI_OP_REJECT_CONN_REQ,
1708 sizeof(rej), &rej);
1709 } else if (conn->type == SCO_LINK || conn->type == ESCO_LINK) {
1710 struct hci_cp_reject_sync_conn_req rej;
1711
1712 bacpy(&rej.bdaddr, &conn->dst);
1713
1714 /* SCO rejection has its own limited set of
1715 * allowed error values (0x0D-0x0F) which isn't
1716 * compatible with most values passed to this
1717 * function. To be safe hard-code one of the
1718 * values that's suitable for SCO.
1719 */
1720 rej.reason = HCI_ERROR_REMOTE_LOW_RESOURCES;
1721
1722 hci_req_add(req, HCI_OP_REJECT_SYNC_CONN_REQ,
1723 sizeof(rej), &rej);
1724 }
1725 break;
1726 default:
1727 conn->state = BT_CLOSED;
1728 break;
1729 }
1730}
1731
1732static void abort_conn_complete(struct hci_dev *hdev, u8 status, u16 opcode)
1733{
1734 if (status)
1735 BT_DBG("Failed to abort connection: status 0x%2.2x", status);
1736}
1737
1738int hci_abort_conn(struct hci_conn *conn, u8 reason)
1739{
1740 struct hci_request req;
1741 int err;
1742
1743 hci_req_init(&req, conn->hdev);
1744
1745 __hci_abort_conn(&req, conn, reason);
1746
1747 err = hci_req_run(&req, abort_conn_complete);
1748 if (err && err != -ENODATA) {
1749 BT_ERR("Failed to run HCI request: err %d", err);
1750 return err;
1751 }
1752
1753 return 0;
1754}
Johan Hedberg5fc16cc2015-11-11 08:11:16 +02001755
Johan Hedberga1d01db2015-11-11 08:11:25 +02001756static int update_bg_scan(struct hci_request *req, unsigned long opt)
Johan Hedberg2e93e532015-11-11 08:11:17 +02001757{
1758 hci_dev_lock(req->hdev);
1759 __hci_update_background_scan(req);
1760 hci_dev_unlock(req->hdev);
Johan Hedberga1d01db2015-11-11 08:11:25 +02001761 return 0;
Johan Hedberg2e93e532015-11-11 08:11:17 +02001762}
1763
1764static void bg_scan_update(struct work_struct *work)
1765{
1766 struct hci_dev *hdev = container_of(work, struct hci_dev,
1767 bg_scan_update);
Johan Hedberg84235d22015-11-11 08:11:20 +02001768 struct hci_conn *conn;
1769 u8 status;
1770 int err;
Johan Hedberg2e93e532015-11-11 08:11:17 +02001771
Johan Hedberg84235d22015-11-11 08:11:20 +02001772 err = hci_req_sync(hdev, update_bg_scan, 0, HCI_CMD_TIMEOUT, &status);
1773 if (!err)
1774 return;
1775
1776 hci_dev_lock(hdev);
1777
1778 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
1779 if (conn)
1780 hci_le_conn_failed(conn, status);
1781
1782 hci_dev_unlock(hdev);
Johan Hedberg2e93e532015-11-11 08:11:17 +02001783}
1784
Johan Hedberga1d01db2015-11-11 08:11:25 +02001785static int le_scan_disable(struct hci_request *req, unsigned long opt)
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001786{
1787 hci_req_add_le_scan_disable(req);
Johan Hedberga1d01db2015-11-11 08:11:25 +02001788 return 0;
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001789}
1790
Johan Hedbergf4a2cb42015-11-11 12:24:22 +02001791static int bredr_inquiry(struct hci_request *req, unsigned long opt)
1792{
1793 u8 length = opt;
Johan Hedberg78b781c2016-01-05 13:19:32 +02001794 const u8 giac[3] = { 0x33, 0x8b, 0x9e };
1795 const u8 liac[3] = { 0x00, 0x8b, 0x9e };
Johan Hedbergf4a2cb42015-11-11 12:24:22 +02001796 struct hci_cp_inquiry cp;
1797
1798 BT_DBG("%s", req->hdev->name);
1799
1800 hci_dev_lock(req->hdev);
1801 hci_inquiry_cache_flush(req->hdev);
1802 hci_dev_unlock(req->hdev);
1803
1804 memset(&cp, 0, sizeof(cp));
Johan Hedberg78b781c2016-01-05 13:19:32 +02001805
1806 if (req->hdev->discovery.limited)
1807 memcpy(&cp.lap, liac, sizeof(cp.lap));
1808 else
1809 memcpy(&cp.lap, giac, sizeof(cp.lap));
1810
Johan Hedbergf4a2cb42015-11-11 12:24:22 +02001811 cp.length = length;
1812
1813 hci_req_add(req, HCI_OP_INQUIRY, sizeof(cp), &cp);
1814
1815 return 0;
1816}
1817
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001818static void le_scan_disable_work(struct work_struct *work)
1819{
1820 struct hci_dev *hdev = container_of(work, struct hci_dev,
1821 le_scan_disable.work);
1822 u8 status;
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001823
1824 BT_DBG("%s", hdev->name);
1825
Johan Hedbergf4a2cb42015-11-11 12:24:22 +02001826 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001827 return;
1828
Johan Hedbergf4a2cb42015-11-11 12:24:22 +02001829 cancel_delayed_work(&hdev->le_scan_restart);
1830
1831 hci_req_sync(hdev, le_scan_disable, 0, HCI_CMD_TIMEOUT, &status);
1832 if (status) {
1833 BT_ERR("Failed to disable LE scan: status 0x%02x", status);
1834 return;
1835 }
1836
1837 hdev->discovery.scan_start = 0;
1838
1839 /* If we were running LE only scan, change discovery state. If
1840 * we were running both LE and BR/EDR inquiry simultaneously,
1841 * and BR/EDR inquiry is already finished, stop discovery,
1842 * otherwise BR/EDR inquiry will stop discovery when finished.
1843 * If we will resolve remote device name, do not change
1844 * discovery state.
1845 */
1846
1847 if (hdev->discovery.type == DISCOV_TYPE_LE)
1848 goto discov_stopped;
1849
1850 if (hdev->discovery.type != DISCOV_TYPE_INTERLEAVED)
1851 return;
1852
1853 if (test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY, &hdev->quirks)) {
1854 if (!test_bit(HCI_INQUIRY, &hdev->flags) &&
1855 hdev->discovery.state != DISCOVERY_RESOLVING)
1856 goto discov_stopped;
1857
1858 return;
1859 }
1860
1861 hci_req_sync(hdev, bredr_inquiry, DISCOV_INTERLEAVED_INQUIRY_LEN,
1862 HCI_CMD_TIMEOUT, &status);
1863 if (status) {
1864 BT_ERR("Inquiry failed: status 0x%02x", status);
1865 goto discov_stopped;
1866 }
1867
1868 return;
1869
1870discov_stopped:
1871 hci_dev_lock(hdev);
1872 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1873 hci_dev_unlock(hdev);
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001874}
1875
Johan Hedberg3dfe5902015-11-11 12:24:23 +02001876static int le_scan_restart(struct hci_request *req, unsigned long opt)
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001877{
Johan Hedberg3dfe5902015-11-11 12:24:23 +02001878 struct hci_dev *hdev = req->hdev;
1879 struct hci_cp_le_set_scan_enable cp;
1880
1881 /* If controller is not scanning we are done. */
1882 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
1883 return 0;
1884
1885 hci_req_add_le_scan_disable(req);
1886
1887 memset(&cp, 0, sizeof(cp));
1888 cp.enable = LE_SCAN_ENABLE;
1889 cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
1890 hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1891
1892 return 0;
1893}
1894
1895static void le_scan_restart_work(struct work_struct *work)
1896{
1897 struct hci_dev *hdev = container_of(work, struct hci_dev,
1898 le_scan_restart.work);
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001899 unsigned long timeout, duration, scan_start, now;
Johan Hedberg3dfe5902015-11-11 12:24:23 +02001900 u8 status;
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001901
1902 BT_DBG("%s", hdev->name);
1903
Johan Hedberg3dfe5902015-11-11 12:24:23 +02001904 hci_req_sync(hdev, le_scan_restart, 0, HCI_CMD_TIMEOUT, &status);
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02001905 if (status) {
1906 BT_ERR("Failed to restart LE scan: status %d", status);
1907 return;
1908 }
1909
1910 hci_dev_lock(hdev);
1911
1912 if (!test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks) ||
1913 !hdev->discovery.scan_start)
1914 goto unlock;
1915
1916 /* When the scan was started, hdev->le_scan_disable has been queued
1917 * after duration from scan_start. During scan restart this job
1918 * has been canceled, and we need to queue it again after proper
1919 * timeout, to make sure that scan does not run indefinitely.
1920 */
1921 duration = hdev->discovery.scan_duration;
1922 scan_start = hdev->discovery.scan_start;
1923 now = jiffies;
1924 if (now - scan_start <= duration) {
1925 int elapsed;
1926
1927 if (now >= scan_start)
1928 elapsed = now - scan_start;
1929 else
1930 elapsed = ULONG_MAX - scan_start + now;
1931
1932 timeout = duration - elapsed;
1933 } else {
1934 timeout = 0;
1935 }
1936
1937 queue_delayed_work(hdev->req_workqueue,
1938 &hdev->le_scan_disable, timeout);
1939
1940unlock:
1941 hci_dev_unlock(hdev);
1942}
1943
Johan Hedberge68f0722015-11-11 08:30:30 +02001944static void disable_advertising(struct hci_request *req)
1945{
1946 u8 enable = 0x00;
1947
1948 hci_req_add(req, HCI_OP_LE_SET_ADV_ENABLE, sizeof(enable), &enable);
1949}
1950
1951static int active_scan(struct hci_request *req, unsigned long opt)
1952{
1953 uint16_t interval = opt;
1954 struct hci_dev *hdev = req->hdev;
1955 struct hci_cp_le_set_scan_param param_cp;
1956 struct hci_cp_le_set_scan_enable enable_cp;
1957 u8 own_addr_type;
1958 int err;
1959
1960 BT_DBG("%s", hdev->name);
1961
1962 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) {
1963 hci_dev_lock(hdev);
1964
1965 /* Don't let discovery abort an outgoing connection attempt
1966 * that's using directed advertising.
1967 */
1968 if (hci_lookup_le_connect(hdev)) {
1969 hci_dev_unlock(hdev);
1970 return -EBUSY;
1971 }
1972
1973 cancel_adv_timeout(hdev);
1974 hci_dev_unlock(hdev);
1975
1976 disable_advertising(req);
1977 }
1978
1979 /* If controller is scanning, it means the background scanning is
1980 * running. Thus, we should temporarily stop it in order to set the
1981 * discovery scanning parameters.
1982 */
1983 if (hci_dev_test_flag(hdev, HCI_LE_SCAN))
1984 hci_req_add_le_scan_disable(req);
1985
1986 /* All active scans will be done with either a resolvable private
1987 * address (when privacy feature has been enabled) or non-resolvable
1988 * private address.
1989 */
Johan Hedberg82a37ad2016-03-09 17:30:34 +02001990 err = hci_update_random_address(req, true, scan_use_rpa(hdev),
1991 &own_addr_type);
Johan Hedberge68f0722015-11-11 08:30:30 +02001992 if (err < 0)
1993 own_addr_type = ADDR_LE_DEV_PUBLIC;
1994
1995 memset(&param_cp, 0, sizeof(param_cp));
1996 param_cp.type = LE_SCAN_ACTIVE;
1997 param_cp.interval = cpu_to_le16(interval);
1998 param_cp.window = cpu_to_le16(DISCOV_LE_SCAN_WIN);
1999 param_cp.own_address_type = own_addr_type;
2000
2001 hci_req_add(req, HCI_OP_LE_SET_SCAN_PARAM, sizeof(param_cp),
2002 &param_cp);
2003
2004 memset(&enable_cp, 0, sizeof(enable_cp));
2005 enable_cp.enable = LE_SCAN_ENABLE;
2006 enable_cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
2007
2008 hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(enable_cp),
2009 &enable_cp);
2010
2011 return 0;
2012}
2013
2014static int interleaved_discov(struct hci_request *req, unsigned long opt)
2015{
2016 int err;
2017
2018 BT_DBG("%s", req->hdev->name);
2019
2020 err = active_scan(req, opt);
2021 if (err)
2022 return err;
2023
Johan Hedberg7df26b52015-11-11 12:24:21 +02002024 return bredr_inquiry(req, DISCOV_BREDR_INQUIRY_LEN);
Johan Hedberge68f0722015-11-11 08:30:30 +02002025}
2026
2027static void start_discovery(struct hci_dev *hdev, u8 *status)
2028{
2029 unsigned long timeout;
2030
2031 BT_DBG("%s type %u", hdev->name, hdev->discovery.type);
2032
2033 switch (hdev->discovery.type) {
2034 case DISCOV_TYPE_BREDR:
2035 if (!hci_dev_test_flag(hdev, HCI_INQUIRY))
Johan Hedberg7df26b52015-11-11 12:24:21 +02002036 hci_req_sync(hdev, bredr_inquiry,
2037 DISCOV_BREDR_INQUIRY_LEN, HCI_CMD_TIMEOUT,
Johan Hedberge68f0722015-11-11 08:30:30 +02002038 status);
2039 return;
2040 case DISCOV_TYPE_INTERLEAVED:
2041 /* When running simultaneous discovery, the LE scanning time
2042 * should occupy the whole discovery time sine BR/EDR inquiry
2043 * and LE scanning are scheduled by the controller.
2044 *
2045 * For interleaving discovery in comparison, BR/EDR inquiry
2046 * and LE scanning are done sequentially with separate
2047 * timeouts.
2048 */
2049 if (test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY,
2050 &hdev->quirks)) {
2051 timeout = msecs_to_jiffies(DISCOV_LE_TIMEOUT);
2052 /* During simultaneous discovery, we double LE scan
2053 * interval. We must leave some time for the controller
2054 * to do BR/EDR inquiry.
2055 */
2056 hci_req_sync(hdev, interleaved_discov,
2057 DISCOV_LE_SCAN_INT * 2, HCI_CMD_TIMEOUT,
2058 status);
2059 break;
2060 }
2061
2062 timeout = msecs_to_jiffies(hdev->discov_interleaved_timeout);
2063 hci_req_sync(hdev, active_scan, DISCOV_LE_SCAN_INT,
2064 HCI_CMD_TIMEOUT, status);
2065 break;
2066 case DISCOV_TYPE_LE:
2067 timeout = msecs_to_jiffies(DISCOV_LE_TIMEOUT);
2068 hci_req_sync(hdev, active_scan, DISCOV_LE_SCAN_INT,
2069 HCI_CMD_TIMEOUT, status);
2070 break;
2071 default:
2072 *status = HCI_ERROR_UNSPECIFIED;
2073 return;
2074 }
2075
2076 if (*status)
2077 return;
2078
2079 BT_DBG("%s timeout %u ms", hdev->name, jiffies_to_msecs(timeout));
2080
2081 /* When service discovery is used and the controller has a
2082 * strict duplicate filter, it is important to remember the
2083 * start and duration of the scan. This is required for
2084 * restarting scanning during the discovery phase.
2085 */
2086 if (test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks) &&
2087 hdev->discovery.result_filtering) {
2088 hdev->discovery.scan_start = jiffies;
2089 hdev->discovery.scan_duration = timeout;
2090 }
2091
2092 queue_delayed_work(hdev->req_workqueue, &hdev->le_scan_disable,
2093 timeout);
2094}
2095
Johan Hedberg2154d3f2015-11-11 08:30:45 +02002096bool hci_req_stop_discovery(struct hci_request *req)
2097{
2098 struct hci_dev *hdev = req->hdev;
2099 struct discovery_state *d = &hdev->discovery;
2100 struct hci_cp_remote_name_req_cancel cp;
2101 struct inquiry_entry *e;
2102 bool ret = false;
2103
2104 BT_DBG("%s state %u", hdev->name, hdev->discovery.state);
2105
2106 if (d->state == DISCOVERY_FINDING || d->state == DISCOVERY_STOPPING) {
2107 if (test_bit(HCI_INQUIRY, &hdev->flags))
2108 hci_req_add(req, HCI_OP_INQUIRY_CANCEL, 0, NULL);
2109
2110 if (hci_dev_test_flag(hdev, HCI_LE_SCAN)) {
2111 cancel_delayed_work(&hdev->le_scan_disable);
2112 hci_req_add_le_scan_disable(req);
2113 }
2114
2115 ret = true;
2116 } else {
2117 /* Passive scanning */
2118 if (hci_dev_test_flag(hdev, HCI_LE_SCAN)) {
2119 hci_req_add_le_scan_disable(req);
2120 ret = true;
2121 }
2122 }
2123
2124 /* No further actions needed for LE-only discovery */
2125 if (d->type == DISCOV_TYPE_LE)
2126 return ret;
2127
2128 if (d->state == DISCOVERY_RESOLVING || d->state == DISCOVERY_STOPPING) {
2129 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY,
2130 NAME_PENDING);
2131 if (!e)
2132 return ret;
2133
2134 bacpy(&cp.bdaddr, &e->data.bdaddr);
2135 hci_req_add(req, HCI_OP_REMOTE_NAME_REQ_CANCEL, sizeof(cp),
2136 &cp);
2137 ret = true;
2138 }
2139
2140 return ret;
2141}
2142
2143static int stop_discovery(struct hci_request *req, unsigned long opt)
2144{
2145 hci_dev_lock(req->hdev);
2146 hci_req_stop_discovery(req);
2147 hci_dev_unlock(req->hdev);
2148
2149 return 0;
2150}
2151
Johan Hedberge68f0722015-11-11 08:30:30 +02002152static void discov_update(struct work_struct *work)
2153{
2154 struct hci_dev *hdev = container_of(work, struct hci_dev,
2155 discov_update);
2156 u8 status = 0;
2157
2158 switch (hdev->discovery.state) {
2159 case DISCOVERY_STARTING:
2160 start_discovery(hdev, &status);
2161 mgmt_start_discovery_complete(hdev, status);
2162 if (status)
2163 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2164 else
2165 hci_discovery_set_state(hdev, DISCOVERY_FINDING);
2166 break;
Johan Hedberg2154d3f2015-11-11 08:30:45 +02002167 case DISCOVERY_STOPPING:
2168 hci_req_sync(hdev, stop_discovery, 0, HCI_CMD_TIMEOUT, &status);
2169 mgmt_stop_discovery_complete(hdev, status);
2170 if (!status)
2171 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2172 break;
Johan Hedberge68f0722015-11-11 08:30:30 +02002173 case DISCOVERY_STOPPED:
2174 default:
2175 return;
2176 }
2177}
2178
Johan Hedbergc366f552015-11-23 15:43:06 +02002179static void discov_off(struct work_struct *work)
2180{
2181 struct hci_dev *hdev = container_of(work, struct hci_dev,
2182 discov_off.work);
2183
2184 BT_DBG("%s", hdev->name);
2185
2186 hci_dev_lock(hdev);
2187
2188 /* When discoverable timeout triggers, then just make sure
2189 * the limited discoverable flag is cleared. Even in the case
2190 * of a timeout triggered from general discoverable, it is
2191 * safe to unconditionally clear the flag.
2192 */
2193 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
2194 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
2195 hdev->discov_timeout = 0;
2196
2197 hci_dev_unlock(hdev);
2198
2199 hci_req_sync(hdev, discoverable_update, 0, HCI_CMD_TIMEOUT, NULL);
2200 mgmt_new_settings(hdev);
2201}
2202
Johan Hedberg2ff13892015-11-25 16:15:44 +02002203static int powered_update_hci(struct hci_request *req, unsigned long opt)
2204{
2205 struct hci_dev *hdev = req->hdev;
Johan Hedberg2ff13892015-11-25 16:15:44 +02002206 u8 link_sec;
2207
2208 hci_dev_lock(hdev);
2209
2210 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED) &&
2211 !lmp_host_ssp_capable(hdev)) {
2212 u8 mode = 0x01;
2213
2214 hci_req_add(req, HCI_OP_WRITE_SSP_MODE, sizeof(mode), &mode);
2215
2216 if (bredr_sc_enabled(hdev) && !lmp_host_sc_capable(hdev)) {
2217 u8 support = 0x01;
2218
2219 hci_req_add(req, HCI_OP_WRITE_SC_SUPPORT,
2220 sizeof(support), &support);
2221 }
2222 }
2223
2224 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED) &&
2225 lmp_bredr_capable(hdev)) {
2226 struct hci_cp_write_le_host_supported cp;
2227
2228 cp.le = 0x01;
2229 cp.simul = 0x00;
2230
2231 /* Check first if we already have the right
2232 * host state (host features set)
2233 */
2234 if (cp.le != lmp_host_le_capable(hdev) ||
2235 cp.simul != lmp_host_le_br_capable(hdev))
2236 hci_req_add(req, HCI_OP_WRITE_LE_HOST_SUPPORTED,
2237 sizeof(cp), &cp);
2238 }
2239
Johan Hedbergd6b7e2c2015-11-30 11:21:44 +02002240 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) {
Johan Hedberg2ff13892015-11-25 16:15:44 +02002241 /* Make sure the controller has a good default for
2242 * advertising data. This also applies to the case
2243 * where BR/EDR was toggled during the AUTO_OFF phase.
2244 */
Johan Hedbergd6b7e2c2015-11-30 11:21:44 +02002245 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
2246 list_empty(&hdev->adv_instances)) {
2247 __hci_req_update_adv_data(req, 0x00);
2248 __hci_req_update_scan_rsp_data(req, 0x00);
Johan Hedberg2ff13892015-11-25 16:15:44 +02002249
Johan Hedbergd6b7e2c2015-11-30 11:21:44 +02002250 if (hci_dev_test_flag(hdev, HCI_ADVERTISING))
2251 __hci_req_enable_advertising(req);
2252 } else if (!list_empty(&hdev->adv_instances)) {
2253 struct adv_info *adv_instance;
2254
Johan Hedberg2ff13892015-11-25 16:15:44 +02002255 adv_instance = list_first_entry(&hdev->adv_instances,
2256 struct adv_info, list);
Johan Hedberg2ff13892015-11-25 16:15:44 +02002257 __hci_req_schedule_adv_instance(req,
Johan Hedbergd6b7e2c2015-11-30 11:21:44 +02002258 adv_instance->instance,
Johan Hedberg2ff13892015-11-25 16:15:44 +02002259 true);
Johan Hedbergd6b7e2c2015-11-30 11:21:44 +02002260 }
Johan Hedberg2ff13892015-11-25 16:15:44 +02002261 }
2262
2263 link_sec = hci_dev_test_flag(hdev, HCI_LINK_SECURITY);
2264 if (link_sec != test_bit(HCI_AUTH, &hdev->flags))
2265 hci_req_add(req, HCI_OP_WRITE_AUTH_ENABLE,
2266 sizeof(link_sec), &link_sec);
2267
2268 if (lmp_bredr_capable(hdev)) {
2269 if (hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE))
2270 __hci_req_write_fast_connectable(req, true);
2271 else
2272 __hci_req_write_fast_connectable(req, false);
2273 __hci_req_update_scan(req);
2274 __hci_req_update_class(req);
2275 __hci_req_update_name(req);
2276 __hci_req_update_eir(req);
2277 }
2278
2279 hci_dev_unlock(hdev);
2280 return 0;
2281}
2282
2283int __hci_req_hci_power_on(struct hci_dev *hdev)
2284{
2285 /* Register the available SMP channels (BR/EDR and LE) only when
2286 * successfully powering on the controller. This late
2287 * registration is required so that LE SMP can clearly decide if
2288 * the public address or static address is used.
2289 */
2290 smp_register(hdev);
2291
2292 return __hci_req_sync(hdev, powered_update_hci, 0, HCI_CMD_TIMEOUT,
2293 NULL);
2294}
2295
Johan Hedberg5fc16cc2015-11-11 08:11:16 +02002296void hci_request_setup(struct hci_dev *hdev)
2297{
Johan Hedberge68f0722015-11-11 08:30:30 +02002298 INIT_WORK(&hdev->discov_update, discov_update);
Johan Hedberg2e93e532015-11-11 08:11:17 +02002299 INIT_WORK(&hdev->bg_scan_update, bg_scan_update);
Johan Hedberg01b1cb82015-11-16 12:52:21 +02002300 INIT_WORK(&hdev->scan_update, scan_update_work);
Johan Hedberg53c0ba72015-11-22 16:43:43 +03002301 INIT_WORK(&hdev->connectable_update, connectable_update_work);
Johan Hedbergaed1a882015-11-22 17:24:44 +03002302 INIT_WORK(&hdev->discoverable_update, discoverable_update_work);
Johan Hedbergc366f552015-11-23 15:43:06 +02002303 INIT_DELAYED_WORK(&hdev->discov_off, discov_off);
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02002304 INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work);
2305 INIT_DELAYED_WORK(&hdev->le_scan_restart, le_scan_restart_work);
Johan Hedbergf2252572015-11-18 12:49:20 +02002306 INIT_DELAYED_WORK(&hdev->adv_instance_expire, adv_timeout_expire);
Johan Hedberg5fc16cc2015-11-11 08:11:16 +02002307}
2308
2309void hci_request_cancel_all(struct hci_dev *hdev)
2310{
Johan Hedberg7df0f732015-11-12 15:15:00 +02002311 hci_req_sync_cancel(hdev, ENODEV);
2312
Johan Hedberge68f0722015-11-11 08:30:30 +02002313 cancel_work_sync(&hdev->discov_update);
Johan Hedberg2e93e532015-11-11 08:11:17 +02002314 cancel_work_sync(&hdev->bg_scan_update);
Johan Hedberg01b1cb82015-11-16 12:52:21 +02002315 cancel_work_sync(&hdev->scan_update);
Johan Hedberg53c0ba72015-11-22 16:43:43 +03002316 cancel_work_sync(&hdev->connectable_update);
Johan Hedbergaed1a882015-11-22 17:24:44 +03002317 cancel_work_sync(&hdev->discoverable_update);
Johan Hedbergc366f552015-11-23 15:43:06 +02002318 cancel_delayed_work_sync(&hdev->discov_off);
Johan Hedberg7c1fbed2015-11-11 08:11:23 +02002319 cancel_delayed_work_sync(&hdev->le_scan_disable);
2320 cancel_delayed_work_sync(&hdev->le_scan_restart);
Johan Hedbergf2252572015-11-18 12:49:20 +02002321
2322 if (hdev->adv_instance_timeout) {
2323 cancel_delayed_work_sync(&hdev->adv_instance_expire);
2324 hdev->adv_instance_timeout = 0;
2325 }
Johan Hedberg5fc16cc2015-11-11 08:11:16 +02002326}