Bluetooth: Update HCI timeouts constants to use msecs_to_jiffies
[deliverable/linux.git] / net / bluetooth / hci_core.c
CommitLineData
8e87d142 1/*
1da177e4
LT
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
590051de 4 Copyright (C) 2011 ProFUSION Embedded Systems
1da177e4
LT
5
6 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
7
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License version 2 as
10 published by the Free Software Foundation;
11
12 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
13 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
14 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
15 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
8e87d142
YH
16 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
17 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
18 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
1da177e4
LT
19 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
20
8e87d142
YH
21 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
22 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
1da177e4
LT
23 SOFTWARE IS DISCLAIMED.
24*/
25
26/* Bluetooth HCI core. */
27
8c520a59 28#include <linux/export.h>
3df92b31 29#include <linux/idr.h>
1da177e4 30
8c520a59 31#include <linux/rfkill.h>
1da177e4
LT
32
33#include <net/bluetooth/bluetooth.h>
34#include <net/bluetooth/hci_core.h>
35
ab81cbf9
JH
36#define AUTO_OFF_TIMEOUT 2000
37
b78752cc 38static void hci_rx_work(struct work_struct *work);
c347b765 39static void hci_cmd_work(struct work_struct *work);
3eff45ea 40static void hci_tx_work(struct work_struct *work);
1da177e4 41
1da177e4
LT
42/* HCI device list */
43LIST_HEAD(hci_dev_list);
44DEFINE_RWLOCK(hci_dev_list_lock);
45
46/* HCI callback list */
47LIST_HEAD(hci_cb_list);
48DEFINE_RWLOCK(hci_cb_list_lock);
49
3df92b31
SL
50/* HCI ID Numbering */
51static DEFINE_IDA(hci_index_ida);
52
1da177e4
LT
53/* ---- HCI notifications ---- */
54
6516455d 55static void hci_notify(struct hci_dev *hdev, int event)
1da177e4 56{
040030ef 57 hci_sock_dev_event(hdev, event);
1da177e4
LT
58}
59
60/* ---- HCI requests ---- */
61
23bb5763 62void hci_req_complete(struct hci_dev *hdev, __u16 cmd, int result)
1da177e4 63{
23bb5763
JH
64 BT_DBG("%s command 0x%04x result 0x%2.2x", hdev->name, cmd, result);
65
a5040efa
JH
66 /* If this is the init phase check if the completed command matches
67 * the last init command, and if not just return.
68 */
75fb0e32
JH
69 if (test_bit(HCI_INIT, &hdev->flags) && hdev->init_last_cmd != cmd) {
70 struct hci_command_hdr *sent = (void *) hdev->sent_cmd->data;
1036b890 71 u16 opcode = __le16_to_cpu(sent->opcode);
75fb0e32
JH
72 struct sk_buff *skb;
73
74 /* Some CSR based controllers generate a spontaneous
75 * reset complete event during init and any pending
76 * command will never be completed. In such a case we
77 * need to resend whatever was the last sent
78 * command.
79 */
80
1036b890 81 if (cmd != HCI_OP_RESET || opcode == HCI_OP_RESET)
75fb0e32
JH
82 return;
83
84 skb = skb_clone(hdev->sent_cmd, GFP_ATOMIC);
85 if (skb) {
86 skb_queue_head(&hdev->cmd_q, skb);
87 queue_work(hdev->workqueue, &hdev->cmd_work);
88 }
89
23bb5763 90 return;
75fb0e32 91 }
1da177e4
LT
92
93 if (hdev->req_status == HCI_REQ_PEND) {
94 hdev->req_result = result;
95 hdev->req_status = HCI_REQ_DONE;
96 wake_up_interruptible(&hdev->req_wait_q);
97 }
98}
99
100static void hci_req_cancel(struct hci_dev *hdev, int err)
101{
102 BT_DBG("%s err 0x%2.2x", hdev->name, err);
103
104 if (hdev->req_status == HCI_REQ_PEND) {
105 hdev->req_result = err;
106 hdev->req_status = HCI_REQ_CANCELED;
107 wake_up_interruptible(&hdev->req_wait_q);
108 }
109}
110
111/* Execute request and wait for completion. */
a8c5fb1a
GP
112static int __hci_request(struct hci_dev *hdev,
113 void (*req)(struct hci_dev *hdev, unsigned long opt),
114 unsigned long opt, __u32 timeout)
1da177e4
LT
115{
116 DECLARE_WAITQUEUE(wait, current);
117 int err = 0;
118
119 BT_DBG("%s start", hdev->name);
120
121 hdev->req_status = HCI_REQ_PEND;
122
123 add_wait_queue(&hdev->req_wait_q, &wait);
124 set_current_state(TASK_INTERRUPTIBLE);
125
126 req(hdev, opt);
127 schedule_timeout(timeout);
128
129 remove_wait_queue(&hdev->req_wait_q, &wait);
130
131 if (signal_pending(current))
132 return -EINTR;
133
134 switch (hdev->req_status) {
135 case HCI_REQ_DONE:
e175072f 136 err = -bt_to_errno(hdev->req_result);
1da177e4
LT
137 break;
138
139 case HCI_REQ_CANCELED:
140 err = -hdev->req_result;
141 break;
142
143 default:
144 err = -ETIMEDOUT;
145 break;
3ff50b79 146 }
1da177e4 147
a5040efa 148 hdev->req_status = hdev->req_result = 0;
1da177e4
LT
149
150 BT_DBG("%s end: err %d", hdev->name, err);
151
152 return err;
153}
154
6039aa73
GP
155static int hci_request(struct hci_dev *hdev,
156 void (*req)(struct hci_dev *hdev, unsigned long opt),
157 unsigned long opt, __u32 timeout)
1da177e4
LT
158{
159 int ret;
160
7c6a329e
MH
161 if (!test_bit(HCI_UP, &hdev->flags))
162 return -ENETDOWN;
163
1da177e4
LT
164 /* Serialize all requests */
165 hci_req_lock(hdev);
166 ret = __hci_request(hdev, req, opt, timeout);
167 hci_req_unlock(hdev);
168
169 return ret;
170}
171
172static void hci_reset_req(struct hci_dev *hdev, unsigned long opt)
173{
174 BT_DBG("%s %ld", hdev->name, opt);
175
176 /* Reset device */
f630cf0d 177 set_bit(HCI_RESET, &hdev->flags);
a9de9248 178 hci_send_cmd(hdev, HCI_OP_RESET, 0, NULL);
1da177e4
LT
179}
180
e61ef499 181static void bredr_init(struct hci_dev *hdev)
1da177e4 182{
b0916ea0 183 struct hci_cp_delete_stored_link_key cp;
1ebb9252 184 __le16 param;
89f2783d 185 __u8 flt_type;
1da177e4 186
2455a3ea
AE
187 hdev->flow_ctl_mode = HCI_FLOW_CTL_MODE_PACKET_BASED;
188
1da177e4
LT
189 /* Mandatory initialization */
190
191 /* Reset */
a6c511c6 192 if (!test_bit(HCI_QUIRK_RESET_ON_CLOSE, &hdev->quirks)) {
e61ef499
AE
193 set_bit(HCI_RESET, &hdev->flags);
194 hci_send_cmd(hdev, HCI_OP_RESET, 0, NULL);
f630cf0d 195 }
1da177e4
LT
196
197 /* Read Local Supported Features */
a9de9248 198 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_FEATURES, 0, NULL);
1da177e4 199
1143e5a6 200 /* Read Local Version */
a9de9248 201 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL);
1143e5a6 202
1da177e4 203 /* Read Buffer Size (ACL mtu, max pkt, etc.) */
a9de9248 204 hci_send_cmd(hdev, HCI_OP_READ_BUFFER_SIZE, 0, NULL);
1da177e4 205
1da177e4 206 /* Read BD Address */
a9de9248
MH
207 hci_send_cmd(hdev, HCI_OP_READ_BD_ADDR, 0, NULL);
208
209 /* Read Class of Device */
210 hci_send_cmd(hdev, HCI_OP_READ_CLASS_OF_DEV, 0, NULL);
211
212 /* Read Local Name */
213 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_NAME, 0, NULL);
1da177e4
LT
214
215 /* Read Voice Setting */
a9de9248 216 hci_send_cmd(hdev, HCI_OP_READ_VOICE_SETTING, 0, NULL);
1da177e4
LT
217
218 /* Optional initialization */
219
220 /* Clear Event Filters */
89f2783d 221 flt_type = HCI_FLT_CLEAR_ALL;
a9de9248 222 hci_send_cmd(hdev, HCI_OP_SET_EVENT_FLT, 1, &flt_type);
1da177e4 223
1da177e4 224 /* Connection accept timeout ~20 secs */
82781e63 225 param = __constant_cpu_to_le16(0x7d00);
a9de9248 226 hci_send_cmd(hdev, HCI_OP_WRITE_CA_TIMEOUT, 2, &param);
b0916ea0
JH
227
228 bacpy(&cp.bdaddr, BDADDR_ANY);
229 cp.delete_all = 1;
230 hci_send_cmd(hdev, HCI_OP_DELETE_STORED_LINK_KEY, sizeof(cp), &cp);
1da177e4
LT
231}
232
e61ef499
AE
233static void amp_init(struct hci_dev *hdev)
234{
2455a3ea
AE
235 hdev->flow_ctl_mode = HCI_FLOW_CTL_MODE_BLOCK_BASED;
236
e61ef499
AE
237 /* Reset */
238 hci_send_cmd(hdev, HCI_OP_RESET, 0, NULL);
239
240 /* Read Local Version */
241 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL);
6bcbc489
AE
242
243 /* Read Local AMP Info */
244 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_AMP_INFO, 0, NULL);
e61ef499
AE
245}
246
247static void hci_init_req(struct hci_dev *hdev, unsigned long opt)
248{
249 struct sk_buff *skb;
250
251 BT_DBG("%s %ld", hdev->name, opt);
252
253 /* Driver initialization */
254
255 /* Special commands */
256 while ((skb = skb_dequeue(&hdev->driver_init))) {
257 bt_cb(skb)->pkt_type = HCI_COMMAND_PKT;
258 skb->dev = (void *) hdev;
259
260 skb_queue_tail(&hdev->cmd_q, skb);
261 queue_work(hdev->workqueue, &hdev->cmd_work);
262 }
263 skb_queue_purge(&hdev->driver_init);
264
265 switch (hdev->dev_type) {
266 case HCI_BREDR:
267 bredr_init(hdev);
268 break;
269
270 case HCI_AMP:
271 amp_init(hdev);
272 break;
273
274 default:
275 BT_ERR("Unknown device type %d", hdev->dev_type);
276 break;
277 }
278
279}
280
6ed58ec5
VT
281static void hci_le_init_req(struct hci_dev *hdev, unsigned long opt)
282{
283 BT_DBG("%s", hdev->name);
284
285 /* Read LE buffer size */
286 hci_send_cmd(hdev, HCI_OP_LE_READ_BUFFER_SIZE, 0, NULL);
287}
288
1da177e4
LT
289static void hci_scan_req(struct hci_dev *hdev, unsigned long opt)
290{
291 __u8 scan = opt;
292
293 BT_DBG("%s %x", hdev->name, scan);
294
295 /* Inquiry and Page scans */
a9de9248 296 hci_send_cmd(hdev, HCI_OP_WRITE_SCAN_ENABLE, 1, &scan);
1da177e4
LT
297}
298
299static void hci_auth_req(struct hci_dev *hdev, unsigned long opt)
300{
301 __u8 auth = opt;
302
303 BT_DBG("%s %x", hdev->name, auth);
304
305 /* Authentication */
a9de9248 306 hci_send_cmd(hdev, HCI_OP_WRITE_AUTH_ENABLE, 1, &auth);
1da177e4
LT
307}
308
309static void hci_encrypt_req(struct hci_dev *hdev, unsigned long opt)
310{
311 __u8 encrypt = opt;
312
313 BT_DBG("%s %x", hdev->name, encrypt);
314
e4e8e37c 315 /* Encryption */
a9de9248 316 hci_send_cmd(hdev, HCI_OP_WRITE_ENCRYPT_MODE, 1, &encrypt);
1da177e4
LT
317}
318
e4e8e37c
MH
319static void hci_linkpol_req(struct hci_dev *hdev, unsigned long opt)
320{
321 __le16 policy = cpu_to_le16(opt);
322
a418b893 323 BT_DBG("%s %x", hdev->name, policy);
e4e8e37c
MH
324
325 /* Default link policy */
326 hci_send_cmd(hdev, HCI_OP_WRITE_DEF_LINK_POLICY, 2, &policy);
327}
328
8e87d142 329/* Get HCI device by index.
1da177e4
LT
330 * Device is held on return. */
331struct hci_dev *hci_dev_get(int index)
332{
8035ded4 333 struct hci_dev *hdev = NULL, *d;
1da177e4
LT
334
335 BT_DBG("%d", index);
336
337 if (index < 0)
338 return NULL;
339
340 read_lock(&hci_dev_list_lock);
8035ded4 341 list_for_each_entry(d, &hci_dev_list, list) {
1da177e4
LT
342 if (d->id == index) {
343 hdev = hci_dev_hold(d);
344 break;
345 }
346 }
347 read_unlock(&hci_dev_list_lock);
348 return hdev;
349}
1da177e4
LT
350
351/* ---- Inquiry support ---- */
ff9ef578 352
30dc78e1
JH
353bool hci_discovery_active(struct hci_dev *hdev)
354{
355 struct discovery_state *discov = &hdev->discovery;
356
6fbe195d 357 switch (discov->state) {
343f935b 358 case DISCOVERY_FINDING:
6fbe195d 359 case DISCOVERY_RESOLVING:
30dc78e1
JH
360 return true;
361
6fbe195d
AG
362 default:
363 return false;
364 }
30dc78e1
JH
365}
366
ff9ef578
JH
367void hci_discovery_set_state(struct hci_dev *hdev, int state)
368{
369 BT_DBG("%s state %u -> %u", hdev->name, hdev->discovery.state, state);
370
371 if (hdev->discovery.state == state)
372 return;
373
374 switch (state) {
375 case DISCOVERY_STOPPED:
7b99b659
AG
376 if (hdev->discovery.state != DISCOVERY_STARTING)
377 mgmt_discovering(hdev, 0);
ff9ef578
JH
378 break;
379 case DISCOVERY_STARTING:
380 break;
343f935b 381 case DISCOVERY_FINDING:
ff9ef578
JH
382 mgmt_discovering(hdev, 1);
383 break;
30dc78e1
JH
384 case DISCOVERY_RESOLVING:
385 break;
ff9ef578
JH
386 case DISCOVERY_STOPPING:
387 break;
388 }
389
390 hdev->discovery.state = state;
391}
392
1da177e4
LT
393static void inquiry_cache_flush(struct hci_dev *hdev)
394{
30883512 395 struct discovery_state *cache = &hdev->discovery;
b57c1a56 396 struct inquiry_entry *p, *n;
1da177e4 397
561aafbc
JH
398 list_for_each_entry_safe(p, n, &cache->all, all) {
399 list_del(&p->all);
b57c1a56 400 kfree(p);
1da177e4 401 }
561aafbc
JH
402
403 INIT_LIST_HEAD(&cache->unknown);
404 INIT_LIST_HEAD(&cache->resolve);
1da177e4
LT
405}
406
a8c5fb1a
GP
407struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev,
408 bdaddr_t *bdaddr)
1da177e4 409{
30883512 410 struct discovery_state *cache = &hdev->discovery;
1da177e4
LT
411 struct inquiry_entry *e;
412
413 BT_DBG("cache %p, %s", cache, batostr(bdaddr));
414
561aafbc
JH
415 list_for_each_entry(e, &cache->all, all) {
416 if (!bacmp(&e->data.bdaddr, bdaddr))
417 return e;
418 }
419
420 return NULL;
421}
422
423struct inquiry_entry *hci_inquiry_cache_lookup_unknown(struct hci_dev *hdev,
04124681 424 bdaddr_t *bdaddr)
561aafbc 425{
30883512 426 struct discovery_state *cache = &hdev->discovery;
561aafbc
JH
427 struct inquiry_entry *e;
428
429 BT_DBG("cache %p, %s", cache, batostr(bdaddr));
430
431 list_for_each_entry(e, &cache->unknown, list) {
1da177e4 432 if (!bacmp(&e->data.bdaddr, bdaddr))
b57c1a56
JH
433 return e;
434 }
435
436 return NULL;
1da177e4
LT
437}
438
30dc78e1 439struct inquiry_entry *hci_inquiry_cache_lookup_resolve(struct hci_dev *hdev,
04124681
GP
440 bdaddr_t *bdaddr,
441 int state)
30dc78e1
JH
442{
443 struct discovery_state *cache = &hdev->discovery;
444 struct inquiry_entry *e;
445
446 BT_DBG("cache %p bdaddr %s state %d", cache, batostr(bdaddr), state);
447
448 list_for_each_entry(e, &cache->resolve, list) {
449 if (!bacmp(bdaddr, BDADDR_ANY) && e->name_state == state)
450 return e;
451 if (!bacmp(&e->data.bdaddr, bdaddr))
452 return e;
453 }
454
455 return NULL;
456}
457
a3d4e20a 458void hci_inquiry_cache_update_resolve(struct hci_dev *hdev,
04124681 459 struct inquiry_entry *ie)
a3d4e20a
JH
460{
461 struct discovery_state *cache = &hdev->discovery;
462 struct list_head *pos = &cache->resolve;
463 struct inquiry_entry *p;
464
465 list_del(&ie->list);
466
467 list_for_each_entry(p, &cache->resolve, list) {
468 if (p->name_state != NAME_PENDING &&
a8c5fb1a 469 abs(p->data.rssi) >= abs(ie->data.rssi))
a3d4e20a
JH
470 break;
471 pos = &p->list;
472 }
473
474 list_add(&ie->list, pos);
475}
476
3175405b 477bool hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data,
04124681 478 bool name_known, bool *ssp)
1da177e4 479{
30883512 480 struct discovery_state *cache = &hdev->discovery;
70f23020 481 struct inquiry_entry *ie;
1da177e4
LT
482
483 BT_DBG("cache %p, %s", cache, batostr(&data->bdaddr));
484
388fc8fa
JH
485 if (ssp)
486 *ssp = data->ssp_mode;
487
70f23020 488 ie = hci_inquiry_cache_lookup(hdev, &data->bdaddr);
a3d4e20a 489 if (ie) {
388fc8fa
JH
490 if (ie->data.ssp_mode && ssp)
491 *ssp = true;
492
a3d4e20a 493 if (ie->name_state == NAME_NEEDED &&
a8c5fb1a 494 data->rssi != ie->data.rssi) {
a3d4e20a
JH
495 ie->data.rssi = data->rssi;
496 hci_inquiry_cache_update_resolve(hdev, ie);
497 }
498
561aafbc 499 goto update;
a3d4e20a 500 }
561aafbc
JH
501
502 /* Entry not in the cache. Add new one. */
503 ie = kzalloc(sizeof(struct inquiry_entry), GFP_ATOMIC);
504 if (!ie)
3175405b 505 return false;
561aafbc
JH
506
507 list_add(&ie->all, &cache->all);
508
509 if (name_known) {
510 ie->name_state = NAME_KNOWN;
511 } else {
512 ie->name_state = NAME_NOT_KNOWN;
513 list_add(&ie->list, &cache->unknown);
514 }
70f23020 515
561aafbc
JH
516update:
517 if (name_known && ie->name_state != NAME_KNOWN &&
a8c5fb1a 518 ie->name_state != NAME_PENDING) {
561aafbc
JH
519 ie->name_state = NAME_KNOWN;
520 list_del(&ie->list);
1da177e4
LT
521 }
522
70f23020
AE
523 memcpy(&ie->data, data, sizeof(*data));
524 ie->timestamp = jiffies;
1da177e4 525 cache->timestamp = jiffies;
3175405b
JH
526
527 if (ie->name_state == NAME_NOT_KNOWN)
528 return false;
529
530 return true;
1da177e4
LT
531}
532
533static int inquiry_cache_dump(struct hci_dev *hdev, int num, __u8 *buf)
534{
30883512 535 struct discovery_state *cache = &hdev->discovery;
1da177e4
LT
536 struct inquiry_info *info = (struct inquiry_info *) buf;
537 struct inquiry_entry *e;
538 int copied = 0;
539
561aafbc 540 list_for_each_entry(e, &cache->all, all) {
1da177e4 541 struct inquiry_data *data = &e->data;
b57c1a56
JH
542
543 if (copied >= num)
544 break;
545
1da177e4
LT
546 bacpy(&info->bdaddr, &data->bdaddr);
547 info->pscan_rep_mode = data->pscan_rep_mode;
548 info->pscan_period_mode = data->pscan_period_mode;
549 info->pscan_mode = data->pscan_mode;
550 memcpy(info->dev_class, data->dev_class, 3);
551 info->clock_offset = data->clock_offset;
b57c1a56 552
1da177e4 553 info++;
b57c1a56 554 copied++;
1da177e4
LT
555 }
556
557 BT_DBG("cache %p, copied %d", cache, copied);
558 return copied;
559}
560
561static void hci_inq_req(struct hci_dev *hdev, unsigned long opt)
562{
563 struct hci_inquiry_req *ir = (struct hci_inquiry_req *) opt;
564 struct hci_cp_inquiry cp;
565
566 BT_DBG("%s", hdev->name);
567
568 if (test_bit(HCI_INQUIRY, &hdev->flags))
569 return;
570
571 /* Start Inquiry */
572 memcpy(&cp.lap, &ir->lap, 3);
573 cp.length = ir->length;
574 cp.num_rsp = ir->num_rsp;
a9de9248 575 hci_send_cmd(hdev, HCI_OP_INQUIRY, sizeof(cp), &cp);
1da177e4
LT
576}
577
578int hci_inquiry(void __user *arg)
579{
580 __u8 __user *ptr = arg;
581 struct hci_inquiry_req ir;
582 struct hci_dev *hdev;
583 int err = 0, do_inquiry = 0, max_rsp;
584 long timeo;
585 __u8 *buf;
586
587 if (copy_from_user(&ir, ptr, sizeof(ir)))
588 return -EFAULT;
589
5a08ecce
AE
590 hdev = hci_dev_get(ir.dev_id);
591 if (!hdev)
1da177e4
LT
592 return -ENODEV;
593
09fd0de5 594 hci_dev_lock(hdev);
8e87d142 595 if (inquiry_cache_age(hdev) > INQUIRY_CACHE_AGE_MAX ||
a8c5fb1a 596 inquiry_cache_empty(hdev) || ir.flags & IREQ_CACHE_FLUSH) {
1da177e4
LT
597 inquiry_cache_flush(hdev);
598 do_inquiry = 1;
599 }
09fd0de5 600 hci_dev_unlock(hdev);
1da177e4 601
04837f64 602 timeo = ir.length * msecs_to_jiffies(2000);
70f23020
AE
603
604 if (do_inquiry) {
605 err = hci_request(hdev, hci_inq_req, (unsigned long)&ir, timeo);
606 if (err < 0)
607 goto done;
608 }
1da177e4 609
8fc9ced3
GP
610 /* for unlimited number of responses we will use buffer with
611 * 255 entries
612 */
1da177e4
LT
613 max_rsp = (ir.num_rsp == 0) ? 255 : ir.num_rsp;
614
615 /* cache_dump can't sleep. Therefore we allocate temp buffer and then
616 * copy it to the user space.
617 */
01df8c31 618 buf = kmalloc(sizeof(struct inquiry_info) * max_rsp, GFP_KERNEL);
70f23020 619 if (!buf) {
1da177e4
LT
620 err = -ENOMEM;
621 goto done;
622 }
623
09fd0de5 624 hci_dev_lock(hdev);
1da177e4 625 ir.num_rsp = inquiry_cache_dump(hdev, max_rsp, buf);
09fd0de5 626 hci_dev_unlock(hdev);
1da177e4
LT
627
628 BT_DBG("num_rsp %d", ir.num_rsp);
629
630 if (!copy_to_user(ptr, &ir, sizeof(ir))) {
631 ptr += sizeof(ir);
632 if (copy_to_user(ptr, buf, sizeof(struct inquiry_info) *
a8c5fb1a 633 ir.num_rsp))
1da177e4 634 err = -EFAULT;
8e87d142 635 } else
1da177e4
LT
636 err = -EFAULT;
637
638 kfree(buf);
639
640done:
641 hci_dev_put(hdev);
642 return err;
643}
644
645/* ---- HCI ioctl helpers ---- */
646
647int hci_dev_open(__u16 dev)
648{
649 struct hci_dev *hdev;
650 int ret = 0;
651
5a08ecce
AE
652 hdev = hci_dev_get(dev);
653 if (!hdev)
1da177e4
LT
654 return -ENODEV;
655
656 BT_DBG("%s %p", hdev->name, hdev);
657
658 hci_req_lock(hdev);
659
94324962
JH
660 if (test_bit(HCI_UNREGISTER, &hdev->dev_flags)) {
661 ret = -ENODEV;
662 goto done;
663 }
664
611b30f7
MH
665 if (hdev->rfkill && rfkill_blocked(hdev->rfkill)) {
666 ret = -ERFKILL;
667 goto done;
668 }
669
1da177e4
LT
670 if (test_bit(HCI_UP, &hdev->flags)) {
671 ret = -EALREADY;
672 goto done;
673 }
674
675 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks))
676 set_bit(HCI_RAW, &hdev->flags);
677
07e3b94a
AE
678 /* Treat all non BR/EDR controllers as raw devices if
679 enable_hs is not set */
680 if (hdev->dev_type != HCI_BREDR && !enable_hs)
943da25d
MH
681 set_bit(HCI_RAW, &hdev->flags);
682
1da177e4
LT
683 if (hdev->open(hdev)) {
684 ret = -EIO;
685 goto done;
686 }
687
688 if (!test_bit(HCI_RAW, &hdev->flags)) {
689 atomic_set(&hdev->cmd_cnt, 1);
690 set_bit(HCI_INIT, &hdev->flags);
a5040efa 691 hdev->init_last_cmd = 0;
1da177e4 692
5f246e89 693 ret = __hci_request(hdev, hci_init_req, 0, HCI_INIT_TIMEOUT);
1da177e4 694
eead27da 695 if (lmp_host_le_capable(hdev))
6ed58ec5 696 ret = __hci_request(hdev, hci_le_init_req, 0,
5f246e89 697 HCI_INIT_TIMEOUT);
6ed58ec5 698
1da177e4
LT
699 clear_bit(HCI_INIT, &hdev->flags);
700 }
701
702 if (!ret) {
703 hci_dev_hold(hdev);
704 set_bit(HCI_UP, &hdev->flags);
705 hci_notify(hdev, HCI_DEV_UP);
a8b2d5c2 706 if (!test_bit(HCI_SETUP, &hdev->dev_flags)) {
09fd0de5 707 hci_dev_lock(hdev);
744cf19e 708 mgmt_powered(hdev, 1);
09fd0de5 709 hci_dev_unlock(hdev);
56e5cb86 710 }
8e87d142 711 } else {
1da177e4 712 /* Init failed, cleanup */
3eff45ea 713 flush_work(&hdev->tx_work);
c347b765 714 flush_work(&hdev->cmd_work);
b78752cc 715 flush_work(&hdev->rx_work);
1da177e4
LT
716
717 skb_queue_purge(&hdev->cmd_q);
718 skb_queue_purge(&hdev->rx_q);
719
720 if (hdev->flush)
721 hdev->flush(hdev);
722
723 if (hdev->sent_cmd) {
724 kfree_skb(hdev->sent_cmd);
725 hdev->sent_cmd = NULL;
726 }
727
728 hdev->close(hdev);
729 hdev->flags = 0;
730 }
731
732done:
733 hci_req_unlock(hdev);
734 hci_dev_put(hdev);
735 return ret;
736}
737
738static int hci_dev_do_close(struct hci_dev *hdev)
739{
740 BT_DBG("%s %p", hdev->name, hdev);
741
28b75a89
AG
742 cancel_work_sync(&hdev->le_scan);
743
1da177e4
LT
744 hci_req_cancel(hdev, ENODEV);
745 hci_req_lock(hdev);
746
747 if (!test_and_clear_bit(HCI_UP, &hdev->flags)) {
b79f44c1 748 del_timer_sync(&hdev->cmd_timer);
1da177e4
LT
749 hci_req_unlock(hdev);
750 return 0;
751 }
752
3eff45ea
GP
753 /* Flush RX and TX works */
754 flush_work(&hdev->tx_work);
b78752cc 755 flush_work(&hdev->rx_work);
1da177e4 756
16ab91ab 757 if (hdev->discov_timeout > 0) {
e0f9309f 758 cancel_delayed_work(&hdev->discov_off);
16ab91ab 759 hdev->discov_timeout = 0;
5e5282bb 760 clear_bit(HCI_DISCOVERABLE, &hdev->dev_flags);
16ab91ab
JH
761 }
762
a8b2d5c2 763 if (test_and_clear_bit(HCI_SERVICE_CACHE, &hdev->dev_flags))
7d78525d
JH
764 cancel_delayed_work(&hdev->service_cache);
765
7ba8b4be
AG
766 cancel_delayed_work_sync(&hdev->le_scan_disable);
767
09fd0de5 768 hci_dev_lock(hdev);
1da177e4
LT
769 inquiry_cache_flush(hdev);
770 hci_conn_hash_flush(hdev);
09fd0de5 771 hci_dev_unlock(hdev);
1da177e4
LT
772
773 hci_notify(hdev, HCI_DEV_DOWN);
774
775 if (hdev->flush)
776 hdev->flush(hdev);
777
778 /* Reset device */
779 skb_queue_purge(&hdev->cmd_q);
780 atomic_set(&hdev->cmd_cnt, 1);
8af59467 781 if (!test_bit(HCI_RAW, &hdev->flags) &&
a6c511c6 782 test_bit(HCI_QUIRK_RESET_ON_CLOSE, &hdev->quirks)) {
1da177e4 783 set_bit(HCI_INIT, &hdev->flags);
5f246e89 784 __hci_request(hdev, hci_reset_req, 0, HCI_CMD_TIMEOUT);
1da177e4
LT
785 clear_bit(HCI_INIT, &hdev->flags);
786 }
787
c347b765
GP
788 /* flush cmd work */
789 flush_work(&hdev->cmd_work);
1da177e4
LT
790
791 /* Drop queues */
792 skb_queue_purge(&hdev->rx_q);
793 skb_queue_purge(&hdev->cmd_q);
794 skb_queue_purge(&hdev->raw_q);
795
796 /* Drop last sent command */
797 if (hdev->sent_cmd) {
b79f44c1 798 del_timer_sync(&hdev->cmd_timer);
1da177e4
LT
799 kfree_skb(hdev->sent_cmd);
800 hdev->sent_cmd = NULL;
801 }
802
803 /* After this point our queues are empty
804 * and no tasks are scheduled. */
805 hdev->close(hdev);
806
8ee56540
MH
807 if (!test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags)) {
808 hci_dev_lock(hdev);
809 mgmt_powered(hdev, 0);
810 hci_dev_unlock(hdev);
811 }
5add6af8 812
1da177e4
LT
813 /* Clear flags */
814 hdev->flags = 0;
815
e59fda8d 816 memset(hdev->eir, 0, sizeof(hdev->eir));
09b3c3fb 817 memset(hdev->dev_class, 0, sizeof(hdev->dev_class));
e59fda8d 818
1da177e4
LT
819 hci_req_unlock(hdev);
820
821 hci_dev_put(hdev);
822 return 0;
823}
824
825int hci_dev_close(__u16 dev)
826{
827 struct hci_dev *hdev;
828 int err;
829
70f23020
AE
830 hdev = hci_dev_get(dev);
831 if (!hdev)
1da177e4 832 return -ENODEV;
8ee56540
MH
833
834 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
835 cancel_delayed_work(&hdev->power_off);
836
1da177e4 837 err = hci_dev_do_close(hdev);
8ee56540 838
1da177e4
LT
839 hci_dev_put(hdev);
840 return err;
841}
842
843int hci_dev_reset(__u16 dev)
844{
845 struct hci_dev *hdev;
846 int ret = 0;
847
70f23020
AE
848 hdev = hci_dev_get(dev);
849 if (!hdev)
1da177e4
LT
850 return -ENODEV;
851
852 hci_req_lock(hdev);
1da177e4
LT
853
854 if (!test_bit(HCI_UP, &hdev->flags))
855 goto done;
856
857 /* Drop queues */
858 skb_queue_purge(&hdev->rx_q);
859 skb_queue_purge(&hdev->cmd_q);
860
09fd0de5 861 hci_dev_lock(hdev);
1da177e4
LT
862 inquiry_cache_flush(hdev);
863 hci_conn_hash_flush(hdev);
09fd0de5 864 hci_dev_unlock(hdev);
1da177e4
LT
865
866 if (hdev->flush)
867 hdev->flush(hdev);
868
8e87d142 869 atomic_set(&hdev->cmd_cnt, 1);
6ed58ec5 870 hdev->acl_cnt = 0; hdev->sco_cnt = 0; hdev->le_cnt = 0;
1da177e4
LT
871
872 if (!test_bit(HCI_RAW, &hdev->flags))
5f246e89 873 ret = __hci_request(hdev, hci_reset_req, 0, HCI_INIT_TIMEOUT);
1da177e4
LT
874
875done:
1da177e4
LT
876 hci_req_unlock(hdev);
877 hci_dev_put(hdev);
878 return ret;
879}
880
881int hci_dev_reset_stat(__u16 dev)
882{
883 struct hci_dev *hdev;
884 int ret = 0;
885
70f23020
AE
886 hdev = hci_dev_get(dev);
887 if (!hdev)
1da177e4
LT
888 return -ENODEV;
889
890 memset(&hdev->stat, 0, sizeof(struct hci_dev_stats));
891
892 hci_dev_put(hdev);
893
894 return ret;
895}
896
897int hci_dev_cmd(unsigned int cmd, void __user *arg)
898{
899 struct hci_dev *hdev;
900 struct hci_dev_req dr;
901 int err = 0;
902
903 if (copy_from_user(&dr, arg, sizeof(dr)))
904 return -EFAULT;
905
70f23020
AE
906 hdev = hci_dev_get(dr.dev_id);
907 if (!hdev)
1da177e4
LT
908 return -ENODEV;
909
910 switch (cmd) {
911 case HCISETAUTH:
04837f64 912 err = hci_request(hdev, hci_auth_req, dr.dev_opt,
5f246e89 913 HCI_INIT_TIMEOUT);
1da177e4
LT
914 break;
915
916 case HCISETENCRYPT:
917 if (!lmp_encrypt_capable(hdev)) {
918 err = -EOPNOTSUPP;
919 break;
920 }
921
922 if (!test_bit(HCI_AUTH, &hdev->flags)) {
923 /* Auth must be enabled first */
04837f64 924 err = hci_request(hdev, hci_auth_req, dr.dev_opt,
5f246e89 925 HCI_INIT_TIMEOUT);
1da177e4
LT
926 if (err)
927 break;
928 }
929
04837f64 930 err = hci_request(hdev, hci_encrypt_req, dr.dev_opt,
5f246e89 931 HCI_INIT_TIMEOUT);
1da177e4
LT
932 break;
933
934 case HCISETSCAN:
04837f64 935 err = hci_request(hdev, hci_scan_req, dr.dev_opt,
5f246e89 936 HCI_INIT_TIMEOUT);
1da177e4
LT
937 break;
938
1da177e4 939 case HCISETLINKPOL:
e4e8e37c 940 err = hci_request(hdev, hci_linkpol_req, dr.dev_opt,
5f246e89 941 HCI_INIT_TIMEOUT);
1da177e4
LT
942 break;
943
944 case HCISETLINKMODE:
e4e8e37c
MH
945 hdev->link_mode = ((__u16) dr.dev_opt) &
946 (HCI_LM_MASTER | HCI_LM_ACCEPT);
947 break;
948
949 case HCISETPTYPE:
950 hdev->pkt_type = (__u16) dr.dev_opt;
1da177e4
LT
951 break;
952
953 case HCISETACLMTU:
e4e8e37c
MH
954 hdev->acl_mtu = *((__u16 *) &dr.dev_opt + 1);
955 hdev->acl_pkts = *((__u16 *) &dr.dev_opt + 0);
1da177e4
LT
956 break;
957
958 case HCISETSCOMTU:
e4e8e37c
MH
959 hdev->sco_mtu = *((__u16 *) &dr.dev_opt + 1);
960 hdev->sco_pkts = *((__u16 *) &dr.dev_opt + 0);
1da177e4
LT
961 break;
962
963 default:
964 err = -EINVAL;
965 break;
966 }
e4e8e37c 967
1da177e4
LT
968 hci_dev_put(hdev);
969 return err;
970}
971
972int hci_get_dev_list(void __user *arg)
973{
8035ded4 974 struct hci_dev *hdev;
1da177e4
LT
975 struct hci_dev_list_req *dl;
976 struct hci_dev_req *dr;
1da177e4
LT
977 int n = 0, size, err;
978 __u16 dev_num;
979
980 if (get_user(dev_num, (__u16 __user *) arg))
981 return -EFAULT;
982
983 if (!dev_num || dev_num > (PAGE_SIZE * 2) / sizeof(*dr))
984 return -EINVAL;
985
986 size = sizeof(*dl) + dev_num * sizeof(*dr);
987
70f23020
AE
988 dl = kzalloc(size, GFP_KERNEL);
989 if (!dl)
1da177e4
LT
990 return -ENOMEM;
991
992 dr = dl->dev_req;
993
f20d09d5 994 read_lock(&hci_dev_list_lock);
8035ded4 995 list_for_each_entry(hdev, &hci_dev_list, list) {
a8b2d5c2 996 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
e0f9309f 997 cancel_delayed_work(&hdev->power_off);
c542a06c 998
a8b2d5c2
JH
999 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1000 set_bit(HCI_PAIRABLE, &hdev->dev_flags);
c542a06c 1001
1da177e4
LT
1002 (dr + n)->dev_id = hdev->id;
1003 (dr + n)->dev_opt = hdev->flags;
c542a06c 1004
1da177e4
LT
1005 if (++n >= dev_num)
1006 break;
1007 }
f20d09d5 1008 read_unlock(&hci_dev_list_lock);
1da177e4
LT
1009
1010 dl->dev_num = n;
1011 size = sizeof(*dl) + n * sizeof(*dr);
1012
1013 err = copy_to_user(arg, dl, size);
1014 kfree(dl);
1015
1016 return err ? -EFAULT : 0;
1017}
1018
1019int hci_get_dev_info(void __user *arg)
1020{
1021 struct hci_dev *hdev;
1022 struct hci_dev_info di;
1023 int err = 0;
1024
1025 if (copy_from_user(&di, arg, sizeof(di)))
1026 return -EFAULT;
1027
70f23020
AE
1028 hdev = hci_dev_get(di.dev_id);
1029 if (!hdev)
1da177e4
LT
1030 return -ENODEV;
1031
a8b2d5c2 1032 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
3243553f 1033 cancel_delayed_work_sync(&hdev->power_off);
ab81cbf9 1034
a8b2d5c2
JH
1035 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1036 set_bit(HCI_PAIRABLE, &hdev->dev_flags);
c542a06c 1037
1da177e4
LT
1038 strcpy(di.name, hdev->name);
1039 di.bdaddr = hdev->bdaddr;
943da25d 1040 di.type = (hdev->bus & 0x0f) | (hdev->dev_type << 4);
1da177e4
LT
1041 di.flags = hdev->flags;
1042 di.pkt_type = hdev->pkt_type;
1043 di.acl_mtu = hdev->acl_mtu;
1044 di.acl_pkts = hdev->acl_pkts;
1045 di.sco_mtu = hdev->sco_mtu;
1046 di.sco_pkts = hdev->sco_pkts;
1047 di.link_policy = hdev->link_policy;
1048 di.link_mode = hdev->link_mode;
1049
1050 memcpy(&di.stat, &hdev->stat, sizeof(di.stat));
1051 memcpy(&di.features, &hdev->features, sizeof(di.features));
1052
1053 if (copy_to_user(arg, &di, sizeof(di)))
1054 err = -EFAULT;
1055
1056 hci_dev_put(hdev);
1057
1058 return err;
1059}
1060
1061/* ---- Interface to HCI drivers ---- */
1062
611b30f7
MH
1063static int hci_rfkill_set_block(void *data, bool blocked)
1064{
1065 struct hci_dev *hdev = data;
1066
1067 BT_DBG("%p name %s blocked %d", hdev, hdev->name, blocked);
1068
1069 if (!blocked)
1070 return 0;
1071
1072 hci_dev_do_close(hdev);
1073
1074 return 0;
1075}
1076
1077static const struct rfkill_ops hci_rfkill_ops = {
1078 .set_block = hci_rfkill_set_block,
1079};
1080
ab81cbf9
JH
1081static void hci_power_on(struct work_struct *work)
1082{
1083 struct hci_dev *hdev = container_of(work, struct hci_dev, power_on);
1084
1085 BT_DBG("%s", hdev->name);
1086
1087 if (hci_dev_open(hdev->id) < 0)
1088 return;
1089
a8b2d5c2 1090 if (test_bit(HCI_AUTO_OFF, &hdev->dev_flags))
80b7ab33 1091 schedule_delayed_work(&hdev->power_off,
a8c5fb1a 1092 msecs_to_jiffies(AUTO_OFF_TIMEOUT));
ab81cbf9 1093
a8b2d5c2 1094 if (test_and_clear_bit(HCI_SETUP, &hdev->dev_flags))
744cf19e 1095 mgmt_index_added(hdev);
ab81cbf9
JH
1096}
1097
1098static void hci_power_off(struct work_struct *work)
1099{
3243553f 1100 struct hci_dev *hdev = container_of(work, struct hci_dev,
a8c5fb1a 1101 power_off.work);
ab81cbf9
JH
1102
1103 BT_DBG("%s", hdev->name);
1104
8ee56540 1105 hci_dev_do_close(hdev);
ab81cbf9
JH
1106}
1107
16ab91ab
JH
1108static void hci_discov_off(struct work_struct *work)
1109{
1110 struct hci_dev *hdev;
1111 u8 scan = SCAN_PAGE;
1112
1113 hdev = container_of(work, struct hci_dev, discov_off.work);
1114
1115 BT_DBG("%s", hdev->name);
1116
09fd0de5 1117 hci_dev_lock(hdev);
16ab91ab
JH
1118
1119 hci_send_cmd(hdev, HCI_OP_WRITE_SCAN_ENABLE, sizeof(scan), &scan);
1120
1121 hdev->discov_timeout = 0;
1122
09fd0de5 1123 hci_dev_unlock(hdev);
16ab91ab
JH
1124}
1125
2aeb9a1a
JH
1126int hci_uuids_clear(struct hci_dev *hdev)
1127{
1128 struct list_head *p, *n;
1129
1130 list_for_each_safe(p, n, &hdev->uuids) {
1131 struct bt_uuid *uuid;
1132
1133 uuid = list_entry(p, struct bt_uuid, list);
1134
1135 list_del(p);
1136 kfree(uuid);
1137 }
1138
1139 return 0;
1140}
1141
55ed8ca1
JH
1142int hci_link_keys_clear(struct hci_dev *hdev)
1143{
1144 struct list_head *p, *n;
1145
1146 list_for_each_safe(p, n, &hdev->link_keys) {
1147 struct link_key *key;
1148
1149 key = list_entry(p, struct link_key, list);
1150
1151 list_del(p);
1152 kfree(key);
1153 }
1154
1155 return 0;
1156}
1157
b899efaf
VCG
1158int hci_smp_ltks_clear(struct hci_dev *hdev)
1159{
1160 struct smp_ltk *k, *tmp;
1161
1162 list_for_each_entry_safe(k, tmp, &hdev->long_term_keys, list) {
1163 list_del(&k->list);
1164 kfree(k);
1165 }
1166
1167 return 0;
1168}
1169
55ed8ca1
JH
1170struct link_key *hci_find_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr)
1171{
8035ded4 1172 struct link_key *k;
55ed8ca1 1173
8035ded4 1174 list_for_each_entry(k, &hdev->link_keys, list)
55ed8ca1
JH
1175 if (bacmp(bdaddr, &k->bdaddr) == 0)
1176 return k;
55ed8ca1
JH
1177
1178 return NULL;
1179}
1180
745c0ce3 1181static bool hci_persistent_key(struct hci_dev *hdev, struct hci_conn *conn,
a8c5fb1a 1182 u8 key_type, u8 old_key_type)
d25e28ab
JH
1183{
1184 /* Legacy key */
1185 if (key_type < 0x03)
745c0ce3 1186 return true;
d25e28ab
JH
1187
1188 /* Debug keys are insecure so don't store them persistently */
1189 if (key_type == HCI_LK_DEBUG_COMBINATION)
745c0ce3 1190 return false;
d25e28ab
JH
1191
1192 /* Changed combination key and there's no previous one */
1193 if (key_type == HCI_LK_CHANGED_COMBINATION && old_key_type == 0xff)
745c0ce3 1194 return false;
d25e28ab
JH
1195
1196 /* Security mode 3 case */
1197 if (!conn)
745c0ce3 1198 return true;
d25e28ab
JH
1199
1200 /* Neither local nor remote side had no-bonding as requirement */
1201 if (conn->auth_type > 0x01 && conn->remote_auth > 0x01)
745c0ce3 1202 return true;
d25e28ab
JH
1203
1204 /* Local side had dedicated bonding as requirement */
1205 if (conn->auth_type == 0x02 || conn->auth_type == 0x03)
745c0ce3 1206 return true;
d25e28ab
JH
1207
1208 /* Remote side had dedicated bonding as requirement */
1209 if (conn->remote_auth == 0x02 || conn->remote_auth == 0x03)
745c0ce3 1210 return true;
d25e28ab
JH
1211
1212 /* If none of the above criteria match, then don't store the key
1213 * persistently */
745c0ce3 1214 return false;
d25e28ab
JH
1215}
1216
c9839a11 1217struct smp_ltk *hci_find_ltk(struct hci_dev *hdev, __le16 ediv, u8 rand[8])
75d262c2 1218{
c9839a11 1219 struct smp_ltk *k;
75d262c2 1220
c9839a11
VCG
1221 list_for_each_entry(k, &hdev->long_term_keys, list) {
1222 if (k->ediv != ediv ||
a8c5fb1a 1223 memcmp(rand, k->rand, sizeof(k->rand)))
75d262c2
VCG
1224 continue;
1225
c9839a11 1226 return k;
75d262c2
VCG
1227 }
1228
1229 return NULL;
1230}
75d262c2 1231
c9839a11 1232struct smp_ltk *hci_find_ltk_by_addr(struct hci_dev *hdev, bdaddr_t *bdaddr,
04124681 1233 u8 addr_type)
75d262c2 1234{
c9839a11 1235 struct smp_ltk *k;
75d262c2 1236
c9839a11
VCG
1237 list_for_each_entry(k, &hdev->long_term_keys, list)
1238 if (addr_type == k->bdaddr_type &&
a8c5fb1a 1239 bacmp(bdaddr, &k->bdaddr) == 0)
75d262c2
VCG
1240 return k;
1241
1242 return NULL;
1243}
75d262c2 1244
d25e28ab 1245int hci_add_link_key(struct hci_dev *hdev, struct hci_conn *conn, int new_key,
04124681 1246 bdaddr_t *bdaddr, u8 *val, u8 type, u8 pin_len)
55ed8ca1
JH
1247{
1248 struct link_key *key, *old_key;
745c0ce3
VA
1249 u8 old_key_type;
1250 bool persistent;
55ed8ca1
JH
1251
1252 old_key = hci_find_link_key(hdev, bdaddr);
1253 if (old_key) {
1254 old_key_type = old_key->type;
1255 key = old_key;
1256 } else {
12adcf3a 1257 old_key_type = conn ? conn->key_type : 0xff;
55ed8ca1
JH
1258 key = kzalloc(sizeof(*key), GFP_ATOMIC);
1259 if (!key)
1260 return -ENOMEM;
1261 list_add(&key->list, &hdev->link_keys);
1262 }
1263
1264 BT_DBG("%s key for %s type %u", hdev->name, batostr(bdaddr), type);
1265
d25e28ab
JH
1266 /* Some buggy controller combinations generate a changed
1267 * combination key for legacy pairing even when there's no
1268 * previous key */
1269 if (type == HCI_LK_CHANGED_COMBINATION &&
a8c5fb1a 1270 (!conn || conn->remote_auth == 0xff) && old_key_type == 0xff) {
d25e28ab 1271 type = HCI_LK_COMBINATION;
655fe6ec
JH
1272 if (conn)
1273 conn->key_type = type;
1274 }
d25e28ab 1275
55ed8ca1 1276 bacpy(&key->bdaddr, bdaddr);
9b3b4460 1277 memcpy(key->val, val, HCI_LINK_KEY_SIZE);
55ed8ca1
JH
1278 key->pin_len = pin_len;
1279
b6020ba0 1280 if (type == HCI_LK_CHANGED_COMBINATION)
55ed8ca1 1281 key->type = old_key_type;
4748fed2
JH
1282 else
1283 key->type = type;
1284
4df378a1
JH
1285 if (!new_key)
1286 return 0;
1287
1288 persistent = hci_persistent_key(hdev, conn, type, old_key_type);
1289
744cf19e 1290 mgmt_new_link_key(hdev, key, persistent);
4df378a1 1291
6ec5bcad
VA
1292 if (conn)
1293 conn->flush_key = !persistent;
55ed8ca1
JH
1294
1295 return 0;
1296}
1297
c9839a11 1298int hci_add_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 addr_type, u8 type,
9a006657 1299 int new_key, u8 authenticated, u8 tk[16], u8 enc_size, __le16
04124681 1300 ediv, u8 rand[8])
75d262c2 1301{
c9839a11 1302 struct smp_ltk *key, *old_key;
75d262c2 1303
c9839a11
VCG
1304 if (!(type & HCI_SMP_STK) && !(type & HCI_SMP_LTK))
1305 return 0;
75d262c2 1306
c9839a11
VCG
1307 old_key = hci_find_ltk_by_addr(hdev, bdaddr, addr_type);
1308 if (old_key)
75d262c2 1309 key = old_key;
c9839a11
VCG
1310 else {
1311 key = kzalloc(sizeof(*key), GFP_ATOMIC);
75d262c2
VCG
1312 if (!key)
1313 return -ENOMEM;
c9839a11 1314 list_add(&key->list, &hdev->long_term_keys);
75d262c2
VCG
1315 }
1316
75d262c2 1317 bacpy(&key->bdaddr, bdaddr);
c9839a11
VCG
1318 key->bdaddr_type = addr_type;
1319 memcpy(key->val, tk, sizeof(key->val));
1320 key->authenticated = authenticated;
1321 key->ediv = ediv;
1322 key->enc_size = enc_size;
1323 key->type = type;
1324 memcpy(key->rand, rand, sizeof(key->rand));
75d262c2 1325
c9839a11
VCG
1326 if (!new_key)
1327 return 0;
75d262c2 1328
261cc5aa
VCG
1329 if (type & HCI_SMP_LTK)
1330 mgmt_new_ltk(hdev, key, 1);
1331
75d262c2
VCG
1332 return 0;
1333}
1334
55ed8ca1
JH
1335int hci_remove_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr)
1336{
1337 struct link_key *key;
1338
1339 key = hci_find_link_key(hdev, bdaddr);
1340 if (!key)
1341 return -ENOENT;
1342
1343 BT_DBG("%s removing %s", hdev->name, batostr(bdaddr));
1344
1345 list_del(&key->list);
1346 kfree(key);
1347
1348 return 0;
1349}
1350
b899efaf
VCG
1351int hci_remove_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr)
1352{
1353 struct smp_ltk *k, *tmp;
1354
1355 list_for_each_entry_safe(k, tmp, &hdev->long_term_keys, list) {
1356 if (bacmp(bdaddr, &k->bdaddr))
1357 continue;
1358
1359 BT_DBG("%s removing %s", hdev->name, batostr(bdaddr));
1360
1361 list_del(&k->list);
1362 kfree(k);
1363 }
1364
1365 return 0;
1366}
1367
6bd32326
VT
1368/* HCI command timer function */
1369static void hci_cmd_timer(unsigned long arg)
1370{
1371 struct hci_dev *hdev = (void *) arg;
1372
1373 BT_ERR("%s command tx timeout", hdev->name);
1374 atomic_set(&hdev->cmd_cnt, 1);
c347b765 1375 queue_work(hdev->workqueue, &hdev->cmd_work);
6bd32326
VT
1376}
1377
2763eda6 1378struct oob_data *hci_find_remote_oob_data(struct hci_dev *hdev,
04124681 1379 bdaddr_t *bdaddr)
2763eda6
SJ
1380{
1381 struct oob_data *data;
1382
1383 list_for_each_entry(data, &hdev->remote_oob_data, list)
1384 if (bacmp(bdaddr, &data->bdaddr) == 0)
1385 return data;
1386
1387 return NULL;
1388}
1389
1390int hci_remove_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr)
1391{
1392 struct oob_data *data;
1393
1394 data = hci_find_remote_oob_data(hdev, bdaddr);
1395 if (!data)
1396 return -ENOENT;
1397
1398 BT_DBG("%s removing %s", hdev->name, batostr(bdaddr));
1399
1400 list_del(&data->list);
1401 kfree(data);
1402
1403 return 0;
1404}
1405
1406int hci_remote_oob_data_clear(struct hci_dev *hdev)
1407{
1408 struct oob_data *data, *n;
1409
1410 list_for_each_entry_safe(data, n, &hdev->remote_oob_data, list) {
1411 list_del(&data->list);
1412 kfree(data);
1413 }
1414
1415 return 0;
1416}
1417
1418int hci_add_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 *hash,
04124681 1419 u8 *randomizer)
2763eda6
SJ
1420{
1421 struct oob_data *data;
1422
1423 data = hci_find_remote_oob_data(hdev, bdaddr);
1424
1425 if (!data) {
1426 data = kmalloc(sizeof(*data), GFP_ATOMIC);
1427 if (!data)
1428 return -ENOMEM;
1429
1430 bacpy(&data->bdaddr, bdaddr);
1431 list_add(&data->list, &hdev->remote_oob_data);
1432 }
1433
1434 memcpy(data->hash, hash, sizeof(data->hash));
1435 memcpy(data->randomizer, randomizer, sizeof(data->randomizer));
1436
1437 BT_DBG("%s for %s", hdev->name, batostr(bdaddr));
1438
1439 return 0;
1440}
1441
04124681 1442struct bdaddr_list *hci_blacklist_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr)
b2a66aad 1443{
8035ded4 1444 struct bdaddr_list *b;
b2a66aad 1445
8035ded4 1446 list_for_each_entry(b, &hdev->blacklist, list)
b2a66aad
AJ
1447 if (bacmp(bdaddr, &b->bdaddr) == 0)
1448 return b;
b2a66aad
AJ
1449
1450 return NULL;
1451}
1452
1453int hci_blacklist_clear(struct hci_dev *hdev)
1454{
1455 struct list_head *p, *n;
1456
1457 list_for_each_safe(p, n, &hdev->blacklist) {
1458 struct bdaddr_list *b;
1459
1460 b = list_entry(p, struct bdaddr_list, list);
1461
1462 list_del(p);
1463 kfree(b);
1464 }
1465
1466 return 0;
1467}
1468
88c1fe4b 1469int hci_blacklist_add(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type)
b2a66aad
AJ
1470{
1471 struct bdaddr_list *entry;
b2a66aad
AJ
1472
1473 if (bacmp(bdaddr, BDADDR_ANY) == 0)
1474 return -EBADF;
1475
5e762444
AJ
1476 if (hci_blacklist_lookup(hdev, bdaddr))
1477 return -EEXIST;
b2a66aad
AJ
1478
1479 entry = kzalloc(sizeof(struct bdaddr_list), GFP_KERNEL);
5e762444
AJ
1480 if (!entry)
1481 return -ENOMEM;
b2a66aad
AJ
1482
1483 bacpy(&entry->bdaddr, bdaddr);
1484
1485 list_add(&entry->list, &hdev->blacklist);
1486
88c1fe4b 1487 return mgmt_device_blocked(hdev, bdaddr, type);
b2a66aad
AJ
1488}
1489
88c1fe4b 1490int hci_blacklist_del(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type)
b2a66aad
AJ
1491{
1492 struct bdaddr_list *entry;
b2a66aad 1493
1ec918ce 1494 if (bacmp(bdaddr, BDADDR_ANY) == 0)
5e762444 1495 return hci_blacklist_clear(hdev);
b2a66aad
AJ
1496
1497 entry = hci_blacklist_lookup(hdev, bdaddr);
1ec918ce 1498 if (!entry)
5e762444 1499 return -ENOENT;
b2a66aad
AJ
1500
1501 list_del(&entry->list);
1502 kfree(entry);
1503
88c1fe4b 1504 return mgmt_device_unblocked(hdev, bdaddr, type);
b2a66aad
AJ
1505}
1506
7ba8b4be
AG
1507static void le_scan_param_req(struct hci_dev *hdev, unsigned long opt)
1508{
1509 struct le_scan_params *param = (struct le_scan_params *) opt;
1510 struct hci_cp_le_set_scan_param cp;
1511
1512 memset(&cp, 0, sizeof(cp));
1513 cp.type = param->type;
1514 cp.interval = cpu_to_le16(param->interval);
1515 cp.window = cpu_to_le16(param->window);
1516
1517 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_PARAM, sizeof(cp), &cp);
1518}
1519
1520static void le_scan_enable_req(struct hci_dev *hdev, unsigned long opt)
1521{
1522 struct hci_cp_le_set_scan_enable cp;
1523
1524 memset(&cp, 0, sizeof(cp));
1525 cp.enable = 1;
0431a43c 1526 cp.filter_dup = 1;
7ba8b4be
AG
1527
1528 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1529}
1530
1531static int hci_do_le_scan(struct hci_dev *hdev, u8 type, u16 interval,
04124681 1532 u16 window, int timeout)
7ba8b4be
AG
1533{
1534 long timeo = msecs_to_jiffies(3000);
1535 struct le_scan_params param;
1536 int err;
1537
1538 BT_DBG("%s", hdev->name);
1539
1540 if (test_bit(HCI_LE_SCAN, &hdev->dev_flags))
1541 return -EINPROGRESS;
1542
1543 param.type = type;
1544 param.interval = interval;
1545 param.window = window;
1546
1547 hci_req_lock(hdev);
1548
1549 err = __hci_request(hdev, le_scan_param_req, (unsigned long) &param,
04124681 1550 timeo);
7ba8b4be
AG
1551 if (!err)
1552 err = __hci_request(hdev, le_scan_enable_req, 0, timeo);
1553
1554 hci_req_unlock(hdev);
1555
1556 if (err < 0)
1557 return err;
1558
1559 schedule_delayed_work(&hdev->le_scan_disable,
04124681 1560 msecs_to_jiffies(timeout));
7ba8b4be
AG
1561
1562 return 0;
1563}
1564
7dbfac1d
AG
1565int hci_cancel_le_scan(struct hci_dev *hdev)
1566{
1567 BT_DBG("%s", hdev->name);
1568
1569 if (!test_bit(HCI_LE_SCAN, &hdev->dev_flags))
1570 return -EALREADY;
1571
1572 if (cancel_delayed_work(&hdev->le_scan_disable)) {
1573 struct hci_cp_le_set_scan_enable cp;
1574
1575 /* Send HCI command to disable LE Scan */
1576 memset(&cp, 0, sizeof(cp));
1577 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1578 }
1579
1580 return 0;
1581}
1582
7ba8b4be
AG
1583static void le_scan_disable_work(struct work_struct *work)
1584{
1585 struct hci_dev *hdev = container_of(work, struct hci_dev,
04124681 1586 le_scan_disable.work);
7ba8b4be
AG
1587 struct hci_cp_le_set_scan_enable cp;
1588
1589 BT_DBG("%s", hdev->name);
1590
1591 memset(&cp, 0, sizeof(cp));
1592
1593 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1594}
1595
28b75a89
AG
1596static void le_scan_work(struct work_struct *work)
1597{
1598 struct hci_dev *hdev = container_of(work, struct hci_dev, le_scan);
1599 struct le_scan_params *param = &hdev->le_scan_params;
1600
1601 BT_DBG("%s", hdev->name);
1602
04124681
GP
1603 hci_do_le_scan(hdev, param->type, param->interval, param->window,
1604 param->timeout);
28b75a89
AG
1605}
1606
1607int hci_le_scan(struct hci_dev *hdev, u8 type, u16 interval, u16 window,
04124681 1608 int timeout)
28b75a89
AG
1609{
1610 struct le_scan_params *param = &hdev->le_scan_params;
1611
1612 BT_DBG("%s", hdev->name);
1613
1614 if (work_busy(&hdev->le_scan))
1615 return -EINPROGRESS;
1616
1617 param->type = type;
1618 param->interval = interval;
1619 param->window = window;
1620 param->timeout = timeout;
1621
1622 queue_work(system_long_wq, &hdev->le_scan);
1623
1624 return 0;
1625}
1626
9be0dab7
DH
1627/* Alloc HCI device */
1628struct hci_dev *hci_alloc_dev(void)
1629{
1630 struct hci_dev *hdev;
1631
1632 hdev = kzalloc(sizeof(struct hci_dev), GFP_KERNEL);
1633 if (!hdev)
1634 return NULL;
1635
b1b813d4
DH
1636 hdev->pkt_type = (HCI_DM1 | HCI_DH1 | HCI_HV1);
1637 hdev->esco_type = (ESCO_HV1);
1638 hdev->link_mode = (HCI_LM_ACCEPT);
1639 hdev->io_capability = 0x03; /* No Input No Output */
1640
b1b813d4
DH
1641 hdev->sniff_max_interval = 800;
1642 hdev->sniff_min_interval = 80;
1643
1644 mutex_init(&hdev->lock);
1645 mutex_init(&hdev->req_lock);
1646
1647 INIT_LIST_HEAD(&hdev->mgmt_pending);
1648 INIT_LIST_HEAD(&hdev->blacklist);
1649 INIT_LIST_HEAD(&hdev->uuids);
1650 INIT_LIST_HEAD(&hdev->link_keys);
1651 INIT_LIST_HEAD(&hdev->long_term_keys);
1652 INIT_LIST_HEAD(&hdev->remote_oob_data);
b1b813d4
DH
1653
1654 INIT_WORK(&hdev->rx_work, hci_rx_work);
1655 INIT_WORK(&hdev->cmd_work, hci_cmd_work);
1656 INIT_WORK(&hdev->tx_work, hci_tx_work);
1657 INIT_WORK(&hdev->power_on, hci_power_on);
1658 INIT_WORK(&hdev->le_scan, le_scan_work);
1659
b1b813d4
DH
1660 INIT_DELAYED_WORK(&hdev->power_off, hci_power_off);
1661 INIT_DELAYED_WORK(&hdev->discov_off, hci_discov_off);
1662 INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work);
1663
9be0dab7 1664 skb_queue_head_init(&hdev->driver_init);
b1b813d4
DH
1665 skb_queue_head_init(&hdev->rx_q);
1666 skb_queue_head_init(&hdev->cmd_q);
1667 skb_queue_head_init(&hdev->raw_q);
1668
1669 init_waitqueue_head(&hdev->req_wait_q);
1670
1671 setup_timer(&hdev->cmd_timer, hci_cmd_timer, (unsigned long) hdev);
1672
b1b813d4
DH
1673 hci_init_sysfs(hdev);
1674 discovery_init(hdev);
1675 hci_conn_hash_init(hdev);
9be0dab7
DH
1676
1677 return hdev;
1678}
1679EXPORT_SYMBOL(hci_alloc_dev);
1680
1681/* Free HCI device */
1682void hci_free_dev(struct hci_dev *hdev)
1683{
1684 skb_queue_purge(&hdev->driver_init);
1685
1686 /* will free via device release */
1687 put_device(&hdev->dev);
1688}
1689EXPORT_SYMBOL(hci_free_dev);
1690
1da177e4
LT
1691/* Register HCI device */
1692int hci_register_dev(struct hci_dev *hdev)
1693{
b1b813d4 1694 int id, error;
1da177e4 1695
010666a1 1696 if (!hdev->open || !hdev->close)
1da177e4
LT
1697 return -EINVAL;
1698
08add513
MM
1699 /* Do not allow HCI_AMP devices to register at index 0,
1700 * so the index can be used as the AMP controller ID.
1701 */
3df92b31
SL
1702 switch (hdev->dev_type) {
1703 case HCI_BREDR:
1704 id = ida_simple_get(&hci_index_ida, 0, 0, GFP_KERNEL);
1705 break;
1706 case HCI_AMP:
1707 id = ida_simple_get(&hci_index_ida, 1, 0, GFP_KERNEL);
1708 break;
1709 default:
1710 return -EINVAL;
1da177e4 1711 }
8e87d142 1712
3df92b31
SL
1713 if (id < 0)
1714 return id;
1715
1da177e4
LT
1716 sprintf(hdev->name, "hci%d", id);
1717 hdev->id = id;
2d8b3a11
AE
1718
1719 BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus);
1720
3df92b31
SL
1721 write_lock(&hci_dev_list_lock);
1722 list_add(&hdev->list, &hci_dev_list);
f20d09d5 1723 write_unlock(&hci_dev_list_lock);
1da177e4 1724
32845eb1 1725 hdev->workqueue = alloc_workqueue(hdev->name, WQ_HIGHPRI | WQ_UNBOUND |
a8c5fb1a 1726 WQ_MEM_RECLAIM, 1);
33ca954d
DH
1727 if (!hdev->workqueue) {
1728 error = -ENOMEM;
1729 goto err;
1730 }
f48fd9c8 1731
33ca954d
DH
1732 error = hci_add_sysfs(hdev);
1733 if (error < 0)
1734 goto err_wqueue;
1da177e4 1735
611b30f7 1736 hdev->rfkill = rfkill_alloc(hdev->name, &hdev->dev,
a8c5fb1a
GP
1737 RFKILL_TYPE_BLUETOOTH, &hci_rfkill_ops,
1738 hdev);
611b30f7
MH
1739 if (hdev->rfkill) {
1740 if (rfkill_register(hdev->rfkill) < 0) {
1741 rfkill_destroy(hdev->rfkill);
1742 hdev->rfkill = NULL;
1743 }
1744 }
1745
a8b2d5c2
JH
1746 set_bit(HCI_AUTO_OFF, &hdev->dev_flags);
1747 set_bit(HCI_SETUP, &hdev->dev_flags);
7f971041 1748 schedule_work(&hdev->power_on);
ab81cbf9 1749
1da177e4 1750 hci_notify(hdev, HCI_DEV_REG);
dc946bd8 1751 hci_dev_hold(hdev);
1da177e4
LT
1752
1753 return id;
f48fd9c8 1754
33ca954d
DH
1755err_wqueue:
1756 destroy_workqueue(hdev->workqueue);
1757err:
3df92b31 1758 ida_simple_remove(&hci_index_ida, hdev->id);
f20d09d5 1759 write_lock(&hci_dev_list_lock);
f48fd9c8 1760 list_del(&hdev->list);
f20d09d5 1761 write_unlock(&hci_dev_list_lock);
f48fd9c8 1762
33ca954d 1763 return error;
1da177e4
LT
1764}
1765EXPORT_SYMBOL(hci_register_dev);
1766
1767/* Unregister HCI device */
59735631 1768void hci_unregister_dev(struct hci_dev *hdev)
1da177e4 1769{
3df92b31 1770 int i, id;
ef222013 1771
c13854ce 1772 BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus);
1da177e4 1773
94324962
JH
1774 set_bit(HCI_UNREGISTER, &hdev->dev_flags);
1775
3df92b31
SL
1776 id = hdev->id;
1777
f20d09d5 1778 write_lock(&hci_dev_list_lock);
1da177e4 1779 list_del(&hdev->list);
f20d09d5 1780 write_unlock(&hci_dev_list_lock);
1da177e4
LT
1781
1782 hci_dev_do_close(hdev);
1783
cd4c5391 1784 for (i = 0; i < NUM_REASSEMBLY; i++)
ef222013
MH
1785 kfree_skb(hdev->reassembly[i]);
1786
ab81cbf9 1787 if (!test_bit(HCI_INIT, &hdev->flags) &&
a8c5fb1a 1788 !test_bit(HCI_SETUP, &hdev->dev_flags)) {
09fd0de5 1789 hci_dev_lock(hdev);
744cf19e 1790 mgmt_index_removed(hdev);
09fd0de5 1791 hci_dev_unlock(hdev);
56e5cb86 1792 }
ab81cbf9 1793
2e58ef3e
JH
1794 /* mgmt_index_removed should take care of emptying the
1795 * pending list */
1796 BUG_ON(!list_empty(&hdev->mgmt_pending));
1797
1da177e4
LT
1798 hci_notify(hdev, HCI_DEV_UNREG);
1799
611b30f7
MH
1800 if (hdev->rfkill) {
1801 rfkill_unregister(hdev->rfkill);
1802 rfkill_destroy(hdev->rfkill);
1803 }
1804
ce242970 1805 hci_del_sysfs(hdev);
147e2d59 1806
f48fd9c8
MH
1807 destroy_workqueue(hdev->workqueue);
1808
09fd0de5 1809 hci_dev_lock(hdev);
e2e0cacb 1810 hci_blacklist_clear(hdev);
2aeb9a1a 1811 hci_uuids_clear(hdev);
55ed8ca1 1812 hci_link_keys_clear(hdev);
b899efaf 1813 hci_smp_ltks_clear(hdev);
2763eda6 1814 hci_remote_oob_data_clear(hdev);
09fd0de5 1815 hci_dev_unlock(hdev);
e2e0cacb 1816
dc946bd8 1817 hci_dev_put(hdev);
3df92b31
SL
1818
1819 ida_simple_remove(&hci_index_ida, id);
1da177e4
LT
1820}
1821EXPORT_SYMBOL(hci_unregister_dev);
1822
1823/* Suspend HCI device */
1824int hci_suspend_dev(struct hci_dev *hdev)
1825{
1826 hci_notify(hdev, HCI_DEV_SUSPEND);
1827 return 0;
1828}
1829EXPORT_SYMBOL(hci_suspend_dev);
1830
1831/* Resume HCI device */
1832int hci_resume_dev(struct hci_dev *hdev)
1833{
1834 hci_notify(hdev, HCI_DEV_RESUME);
1835 return 0;
1836}
1837EXPORT_SYMBOL(hci_resume_dev);
1838
76bca880
MH
1839/* Receive frame from HCI drivers */
1840int hci_recv_frame(struct sk_buff *skb)
1841{
1842 struct hci_dev *hdev = (struct hci_dev *) skb->dev;
1843 if (!hdev || (!test_bit(HCI_UP, &hdev->flags)
a8c5fb1a 1844 && !test_bit(HCI_INIT, &hdev->flags))) {
76bca880
MH
1845 kfree_skb(skb);
1846 return -ENXIO;
1847 }
1848
1849 /* Incomming skb */
1850 bt_cb(skb)->incoming = 1;
1851
1852 /* Time stamp */
1853 __net_timestamp(skb);
1854
76bca880 1855 skb_queue_tail(&hdev->rx_q, skb);
b78752cc 1856 queue_work(hdev->workqueue, &hdev->rx_work);
c78ae283 1857
76bca880
MH
1858 return 0;
1859}
1860EXPORT_SYMBOL(hci_recv_frame);
1861
33e882a5 1862static int hci_reassembly(struct hci_dev *hdev, int type, void *data,
a8c5fb1a 1863 int count, __u8 index)
33e882a5
SS
1864{
1865 int len = 0;
1866 int hlen = 0;
1867 int remain = count;
1868 struct sk_buff *skb;
1869 struct bt_skb_cb *scb;
1870
1871 if ((type < HCI_ACLDATA_PKT || type > HCI_EVENT_PKT) ||
a8c5fb1a 1872 index >= NUM_REASSEMBLY)
33e882a5
SS
1873 return -EILSEQ;
1874
1875 skb = hdev->reassembly[index];
1876
1877 if (!skb) {
1878 switch (type) {
1879 case HCI_ACLDATA_PKT:
1880 len = HCI_MAX_FRAME_SIZE;
1881 hlen = HCI_ACL_HDR_SIZE;
1882 break;
1883 case HCI_EVENT_PKT:
1884 len = HCI_MAX_EVENT_SIZE;
1885 hlen = HCI_EVENT_HDR_SIZE;
1886 break;
1887 case HCI_SCODATA_PKT:
1888 len = HCI_MAX_SCO_SIZE;
1889 hlen = HCI_SCO_HDR_SIZE;
1890 break;
1891 }
1892
1e429f38 1893 skb = bt_skb_alloc(len, GFP_ATOMIC);
33e882a5
SS
1894 if (!skb)
1895 return -ENOMEM;
1896
1897 scb = (void *) skb->cb;
1898 scb->expect = hlen;
1899 scb->pkt_type = type;
1900
1901 skb->dev = (void *) hdev;
1902 hdev->reassembly[index] = skb;
1903 }
1904
1905 while (count) {
1906 scb = (void *) skb->cb;
89bb46d0 1907 len = min_t(uint, scb->expect, count);
33e882a5
SS
1908
1909 memcpy(skb_put(skb, len), data, len);
1910
1911 count -= len;
1912 data += len;
1913 scb->expect -= len;
1914 remain = count;
1915
1916 switch (type) {
1917 case HCI_EVENT_PKT:
1918 if (skb->len == HCI_EVENT_HDR_SIZE) {
1919 struct hci_event_hdr *h = hci_event_hdr(skb);
1920 scb->expect = h->plen;
1921
1922 if (skb_tailroom(skb) < scb->expect) {
1923 kfree_skb(skb);
1924 hdev->reassembly[index] = NULL;
1925 return -ENOMEM;
1926 }
1927 }
1928 break;
1929
1930 case HCI_ACLDATA_PKT:
1931 if (skb->len == HCI_ACL_HDR_SIZE) {
1932 struct hci_acl_hdr *h = hci_acl_hdr(skb);
1933 scb->expect = __le16_to_cpu(h->dlen);
1934
1935 if (skb_tailroom(skb) < scb->expect) {
1936 kfree_skb(skb);
1937 hdev->reassembly[index] = NULL;
1938 return -ENOMEM;
1939 }
1940 }
1941 break;
1942
1943 case HCI_SCODATA_PKT:
1944 if (skb->len == HCI_SCO_HDR_SIZE) {
1945 struct hci_sco_hdr *h = hci_sco_hdr(skb);
1946 scb->expect = h->dlen;
1947
1948 if (skb_tailroom(skb) < scb->expect) {
1949 kfree_skb(skb);
1950 hdev->reassembly[index] = NULL;
1951 return -ENOMEM;
1952 }
1953 }
1954 break;
1955 }
1956
1957 if (scb->expect == 0) {
1958 /* Complete frame */
1959
1960 bt_cb(skb)->pkt_type = type;
1961 hci_recv_frame(skb);
1962
1963 hdev->reassembly[index] = NULL;
1964 return remain;
1965 }
1966 }
1967
1968 return remain;
1969}
1970
ef222013
MH
1971int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count)
1972{
f39a3c06
SS
1973 int rem = 0;
1974
ef222013
MH
1975 if (type < HCI_ACLDATA_PKT || type > HCI_EVENT_PKT)
1976 return -EILSEQ;
1977
da5f6c37 1978 while (count) {
1e429f38 1979 rem = hci_reassembly(hdev, type, data, count, type - 1);
f39a3c06
SS
1980 if (rem < 0)
1981 return rem;
ef222013 1982
f39a3c06
SS
1983 data += (count - rem);
1984 count = rem;
f81c6224 1985 }
ef222013 1986
f39a3c06 1987 return rem;
ef222013
MH
1988}
1989EXPORT_SYMBOL(hci_recv_fragment);
1990
99811510
SS
1991#define STREAM_REASSEMBLY 0
1992
1993int hci_recv_stream_fragment(struct hci_dev *hdev, void *data, int count)
1994{
1995 int type;
1996 int rem = 0;
1997
da5f6c37 1998 while (count) {
99811510
SS
1999 struct sk_buff *skb = hdev->reassembly[STREAM_REASSEMBLY];
2000
2001 if (!skb) {
2002 struct { char type; } *pkt;
2003
2004 /* Start of the frame */
2005 pkt = data;
2006 type = pkt->type;
2007
2008 data++;
2009 count--;
2010 } else
2011 type = bt_cb(skb)->pkt_type;
2012
1e429f38 2013 rem = hci_reassembly(hdev, type, data, count,
a8c5fb1a 2014 STREAM_REASSEMBLY);
99811510
SS
2015 if (rem < 0)
2016 return rem;
2017
2018 data += (count - rem);
2019 count = rem;
f81c6224 2020 }
99811510
SS
2021
2022 return rem;
2023}
2024EXPORT_SYMBOL(hci_recv_stream_fragment);
2025
1da177e4
LT
2026/* ---- Interface to upper protocols ---- */
2027
1da177e4
LT
2028int hci_register_cb(struct hci_cb *cb)
2029{
2030 BT_DBG("%p name %s", cb, cb->name);
2031
f20d09d5 2032 write_lock(&hci_cb_list_lock);
1da177e4 2033 list_add(&cb->list, &hci_cb_list);
f20d09d5 2034 write_unlock(&hci_cb_list_lock);
1da177e4
LT
2035
2036 return 0;
2037}
2038EXPORT_SYMBOL(hci_register_cb);
2039
2040int hci_unregister_cb(struct hci_cb *cb)
2041{
2042 BT_DBG("%p name %s", cb, cb->name);
2043
f20d09d5 2044 write_lock(&hci_cb_list_lock);
1da177e4 2045 list_del(&cb->list);
f20d09d5 2046 write_unlock(&hci_cb_list_lock);
1da177e4
LT
2047
2048 return 0;
2049}
2050EXPORT_SYMBOL(hci_unregister_cb);
2051
2052static int hci_send_frame(struct sk_buff *skb)
2053{
2054 struct hci_dev *hdev = (struct hci_dev *) skb->dev;
2055
2056 if (!hdev) {
2057 kfree_skb(skb);
2058 return -ENODEV;
2059 }
2060
0d48d939 2061 BT_DBG("%s type %d len %d", hdev->name, bt_cb(skb)->pkt_type, skb->len);
1da177e4 2062
cd82e61c
MH
2063 /* Time stamp */
2064 __net_timestamp(skb);
1da177e4 2065
cd82e61c
MH
2066 /* Send copy to monitor */
2067 hci_send_to_monitor(hdev, skb);
2068
2069 if (atomic_read(&hdev->promisc)) {
2070 /* Send copy to the sockets */
470fe1b5 2071 hci_send_to_sock(hdev, skb);
1da177e4
LT
2072 }
2073
2074 /* Get rid of skb owner, prior to sending to the driver. */
2075 skb_orphan(skb);
2076
2077 return hdev->send(skb);
2078}
2079
2080/* Send HCI command */
a9de9248 2081int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param)
1da177e4
LT
2082{
2083 int len = HCI_COMMAND_HDR_SIZE + plen;
2084 struct hci_command_hdr *hdr;
2085 struct sk_buff *skb;
2086
a9de9248 2087 BT_DBG("%s opcode 0x%x plen %d", hdev->name, opcode, plen);
1da177e4
LT
2088
2089 skb = bt_skb_alloc(len, GFP_ATOMIC);
2090 if (!skb) {
ef222013 2091 BT_ERR("%s no memory for command", hdev->name);
1da177e4
LT
2092 return -ENOMEM;
2093 }
2094
2095 hdr = (struct hci_command_hdr *) skb_put(skb, HCI_COMMAND_HDR_SIZE);
a9de9248 2096 hdr->opcode = cpu_to_le16(opcode);
1da177e4
LT
2097 hdr->plen = plen;
2098
2099 if (plen)
2100 memcpy(skb_put(skb, plen), param, plen);
2101
2102 BT_DBG("skb len %d", skb->len);
2103
0d48d939 2104 bt_cb(skb)->pkt_type = HCI_COMMAND_PKT;
1da177e4 2105 skb->dev = (void *) hdev;
c78ae283 2106
a5040efa
JH
2107 if (test_bit(HCI_INIT, &hdev->flags))
2108 hdev->init_last_cmd = opcode;
2109
1da177e4 2110 skb_queue_tail(&hdev->cmd_q, skb);
c347b765 2111 queue_work(hdev->workqueue, &hdev->cmd_work);
1da177e4
LT
2112
2113 return 0;
2114}
1da177e4
LT
2115
2116/* Get data from the previously sent command */
a9de9248 2117void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode)
1da177e4
LT
2118{
2119 struct hci_command_hdr *hdr;
2120
2121 if (!hdev->sent_cmd)
2122 return NULL;
2123
2124 hdr = (void *) hdev->sent_cmd->data;
2125
a9de9248 2126 if (hdr->opcode != cpu_to_le16(opcode))
1da177e4
LT
2127 return NULL;
2128
a9de9248 2129 BT_DBG("%s opcode 0x%x", hdev->name, opcode);
1da177e4
LT
2130
2131 return hdev->sent_cmd->data + HCI_COMMAND_HDR_SIZE;
2132}
2133
2134/* Send ACL data */
2135static void hci_add_acl_hdr(struct sk_buff *skb, __u16 handle, __u16 flags)
2136{
2137 struct hci_acl_hdr *hdr;
2138 int len = skb->len;
2139
badff6d0
ACM
2140 skb_push(skb, HCI_ACL_HDR_SIZE);
2141 skb_reset_transport_header(skb);
9c70220b 2142 hdr = (struct hci_acl_hdr *)skb_transport_header(skb);
aca3192c
YH
2143 hdr->handle = cpu_to_le16(hci_handle_pack(handle, flags));
2144 hdr->dlen = cpu_to_le16(len);
1da177e4
LT
2145}
2146
73d80deb 2147static void hci_queue_acl(struct hci_conn *conn, struct sk_buff_head *queue,
a8c5fb1a 2148 struct sk_buff *skb, __u16 flags)
1da177e4
LT
2149{
2150 struct hci_dev *hdev = conn->hdev;
2151 struct sk_buff *list;
2152
087bfd99
GP
2153 skb->len = skb_headlen(skb);
2154 skb->data_len = 0;
2155
2156 bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT;
2157 hci_add_acl_hdr(skb, conn->handle, flags);
2158
70f23020
AE
2159 list = skb_shinfo(skb)->frag_list;
2160 if (!list) {
1da177e4
LT
2161 /* Non fragmented */
2162 BT_DBG("%s nonfrag skb %p len %d", hdev->name, skb, skb->len);
2163
73d80deb 2164 skb_queue_tail(queue, skb);
1da177e4
LT
2165 } else {
2166 /* Fragmented */
2167 BT_DBG("%s frag %p len %d", hdev->name, skb, skb->len);
2168
2169 skb_shinfo(skb)->frag_list = NULL;
2170
2171 /* Queue all fragments atomically */
af3e6359 2172 spin_lock(&queue->lock);
1da177e4 2173
73d80deb 2174 __skb_queue_tail(queue, skb);
e702112f
AE
2175
2176 flags &= ~ACL_START;
2177 flags |= ACL_CONT;
1da177e4
LT
2178 do {
2179 skb = list; list = list->next;
8e87d142 2180
1da177e4 2181 skb->dev = (void *) hdev;
0d48d939 2182 bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT;
e702112f 2183 hci_add_acl_hdr(skb, conn->handle, flags);
1da177e4
LT
2184
2185 BT_DBG("%s frag %p len %d", hdev->name, skb, skb->len);
2186
73d80deb 2187 __skb_queue_tail(queue, skb);
1da177e4
LT
2188 } while (list);
2189
af3e6359 2190 spin_unlock(&queue->lock);
1da177e4 2191 }
73d80deb
LAD
2192}
2193
2194void hci_send_acl(struct hci_chan *chan, struct sk_buff *skb, __u16 flags)
2195{
2196 struct hci_conn *conn = chan->conn;
2197 struct hci_dev *hdev = conn->hdev;
2198
2199 BT_DBG("%s chan %p flags 0x%x", hdev->name, chan, flags);
2200
2201 skb->dev = (void *) hdev;
73d80deb
LAD
2202
2203 hci_queue_acl(conn, &chan->data_q, skb, flags);
1da177e4 2204
3eff45ea 2205 queue_work(hdev->workqueue, &hdev->tx_work);
1da177e4 2206}
1da177e4
LT
2207
2208/* Send SCO data */
0d861d8b 2209void hci_send_sco(struct hci_conn *conn, struct sk_buff *skb)
1da177e4
LT
2210{
2211 struct hci_dev *hdev = conn->hdev;
2212 struct hci_sco_hdr hdr;
2213
2214 BT_DBG("%s len %d", hdev->name, skb->len);
2215
aca3192c 2216 hdr.handle = cpu_to_le16(conn->handle);
1da177e4
LT
2217 hdr.dlen = skb->len;
2218
badff6d0
ACM
2219 skb_push(skb, HCI_SCO_HDR_SIZE);
2220 skb_reset_transport_header(skb);
9c70220b 2221 memcpy(skb_transport_header(skb), &hdr, HCI_SCO_HDR_SIZE);
1da177e4
LT
2222
2223 skb->dev = (void *) hdev;
0d48d939 2224 bt_cb(skb)->pkt_type = HCI_SCODATA_PKT;
c78ae283 2225
1da177e4 2226 skb_queue_tail(&conn->data_q, skb);
3eff45ea 2227 queue_work(hdev->workqueue, &hdev->tx_work);
1da177e4 2228}
1da177e4
LT
2229
2230/* ---- HCI TX task (outgoing data) ---- */
2231
2232/* HCI Connection scheduler */
6039aa73
GP
2233static struct hci_conn *hci_low_sent(struct hci_dev *hdev, __u8 type,
2234 int *quote)
1da177e4
LT
2235{
2236 struct hci_conn_hash *h = &hdev->conn_hash;
8035ded4 2237 struct hci_conn *conn = NULL, *c;
abc5de8f 2238 unsigned int num = 0, min = ~0;
1da177e4 2239
8e87d142 2240 /* We don't have to lock device here. Connections are always
1da177e4 2241 * added and removed with TX task disabled. */
bf4c6325
GP
2242
2243 rcu_read_lock();
2244
2245 list_for_each_entry_rcu(c, &h->list, list) {
769be974 2246 if (c->type != type || skb_queue_empty(&c->data_q))
1da177e4 2247 continue;
769be974
MH
2248
2249 if (c->state != BT_CONNECTED && c->state != BT_CONFIG)
2250 continue;
2251
1da177e4
LT
2252 num++;
2253
2254 if (c->sent < min) {
2255 min = c->sent;
2256 conn = c;
2257 }
52087a79
LAD
2258
2259 if (hci_conn_num(hdev, type) == num)
2260 break;
1da177e4
LT
2261 }
2262
bf4c6325
GP
2263 rcu_read_unlock();
2264
1da177e4 2265 if (conn) {
6ed58ec5
VT
2266 int cnt, q;
2267
2268 switch (conn->type) {
2269 case ACL_LINK:
2270 cnt = hdev->acl_cnt;
2271 break;
2272 case SCO_LINK:
2273 case ESCO_LINK:
2274 cnt = hdev->sco_cnt;
2275 break;
2276 case LE_LINK:
2277 cnt = hdev->le_mtu ? hdev->le_cnt : hdev->acl_cnt;
2278 break;
2279 default:
2280 cnt = 0;
2281 BT_ERR("Unknown link type");
2282 }
2283
2284 q = cnt / num;
1da177e4
LT
2285 *quote = q ? q : 1;
2286 } else
2287 *quote = 0;
2288
2289 BT_DBG("conn %p quote %d", conn, *quote);
2290 return conn;
2291}
2292
6039aa73 2293static void hci_link_tx_to(struct hci_dev *hdev, __u8 type)
1da177e4
LT
2294{
2295 struct hci_conn_hash *h = &hdev->conn_hash;
8035ded4 2296 struct hci_conn *c;
1da177e4 2297
bae1f5d9 2298 BT_ERR("%s link tx timeout", hdev->name);
1da177e4 2299
bf4c6325
GP
2300 rcu_read_lock();
2301
1da177e4 2302 /* Kill stalled connections */
bf4c6325 2303 list_for_each_entry_rcu(c, &h->list, list) {
bae1f5d9
VT
2304 if (c->type == type && c->sent) {
2305 BT_ERR("%s killing stalled connection %s",
a8c5fb1a 2306 hdev->name, batostr(&c->dst));
7490c6c2 2307 hci_acl_disconn(c, HCI_ERROR_REMOTE_USER_TERM);
1da177e4
LT
2308 }
2309 }
bf4c6325
GP
2310
2311 rcu_read_unlock();
1da177e4
LT
2312}
2313
6039aa73
GP
2314static struct hci_chan *hci_chan_sent(struct hci_dev *hdev, __u8 type,
2315 int *quote)
1da177e4 2316{
73d80deb
LAD
2317 struct hci_conn_hash *h = &hdev->conn_hash;
2318 struct hci_chan *chan = NULL;
abc5de8f 2319 unsigned int num = 0, min = ~0, cur_prio = 0;
1da177e4 2320 struct hci_conn *conn;
73d80deb
LAD
2321 int cnt, q, conn_num = 0;
2322
2323 BT_DBG("%s", hdev->name);
2324
bf4c6325
GP
2325 rcu_read_lock();
2326
2327 list_for_each_entry_rcu(conn, &h->list, list) {
73d80deb
LAD
2328 struct hci_chan *tmp;
2329
2330 if (conn->type != type)
2331 continue;
2332
2333 if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
2334 continue;
2335
2336 conn_num++;
2337
8192edef 2338 list_for_each_entry_rcu(tmp, &conn->chan_list, list) {
73d80deb
LAD
2339 struct sk_buff *skb;
2340
2341 if (skb_queue_empty(&tmp->data_q))
2342 continue;
2343
2344 skb = skb_peek(&tmp->data_q);
2345 if (skb->priority < cur_prio)
2346 continue;
2347
2348 if (skb->priority > cur_prio) {
2349 num = 0;
2350 min = ~0;
2351 cur_prio = skb->priority;
2352 }
2353
2354 num++;
2355
2356 if (conn->sent < min) {
2357 min = conn->sent;
2358 chan = tmp;
2359 }
2360 }
2361
2362 if (hci_conn_num(hdev, type) == conn_num)
2363 break;
2364 }
2365
bf4c6325
GP
2366 rcu_read_unlock();
2367
73d80deb
LAD
2368 if (!chan)
2369 return NULL;
2370
2371 switch (chan->conn->type) {
2372 case ACL_LINK:
2373 cnt = hdev->acl_cnt;
2374 break;
2375 case SCO_LINK:
2376 case ESCO_LINK:
2377 cnt = hdev->sco_cnt;
2378 break;
2379 case LE_LINK:
2380 cnt = hdev->le_mtu ? hdev->le_cnt : hdev->acl_cnt;
2381 break;
2382 default:
2383 cnt = 0;
2384 BT_ERR("Unknown link type");
2385 }
2386
2387 q = cnt / num;
2388 *quote = q ? q : 1;
2389 BT_DBG("chan %p quote %d", chan, *quote);
2390 return chan;
2391}
2392
02b20f0b
LAD
2393static void hci_prio_recalculate(struct hci_dev *hdev, __u8 type)
2394{
2395 struct hci_conn_hash *h = &hdev->conn_hash;
2396 struct hci_conn *conn;
2397 int num = 0;
2398
2399 BT_DBG("%s", hdev->name);
2400
bf4c6325
GP
2401 rcu_read_lock();
2402
2403 list_for_each_entry_rcu(conn, &h->list, list) {
02b20f0b
LAD
2404 struct hci_chan *chan;
2405
2406 if (conn->type != type)
2407 continue;
2408
2409 if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
2410 continue;
2411
2412 num++;
2413
8192edef 2414 list_for_each_entry_rcu(chan, &conn->chan_list, list) {
02b20f0b
LAD
2415 struct sk_buff *skb;
2416
2417 if (chan->sent) {
2418 chan->sent = 0;
2419 continue;
2420 }
2421
2422 if (skb_queue_empty(&chan->data_q))
2423 continue;
2424
2425 skb = skb_peek(&chan->data_q);
2426 if (skb->priority >= HCI_PRIO_MAX - 1)
2427 continue;
2428
2429 skb->priority = HCI_PRIO_MAX - 1;
2430
2431 BT_DBG("chan %p skb %p promoted to %d", chan, skb,
a8c5fb1a 2432 skb->priority);
02b20f0b
LAD
2433 }
2434
2435 if (hci_conn_num(hdev, type) == num)
2436 break;
2437 }
bf4c6325
GP
2438
2439 rcu_read_unlock();
2440
02b20f0b
LAD
2441}
2442
b71d385a
AE
2443static inline int __get_blocks(struct hci_dev *hdev, struct sk_buff *skb)
2444{
2445 /* Calculate count of blocks used by this packet */
2446 return DIV_ROUND_UP(skb->len - HCI_ACL_HDR_SIZE, hdev->block_len);
2447}
2448
6039aa73 2449static void __check_timeout(struct hci_dev *hdev, unsigned int cnt)
73d80deb 2450{
1da177e4
LT
2451 if (!test_bit(HCI_RAW, &hdev->flags)) {
2452 /* ACL tx timeout must be longer than maximum
2453 * link supervision timeout (40.9 seconds) */
63d2bc1b 2454 if (!cnt && time_after(jiffies, hdev->acl_last_tx +
5f246e89 2455 HCI_ACL_TX_TIMEOUT))
bae1f5d9 2456 hci_link_tx_to(hdev, ACL_LINK);
1da177e4 2457 }
63d2bc1b 2458}
1da177e4 2459
6039aa73 2460static void hci_sched_acl_pkt(struct hci_dev *hdev)
63d2bc1b
AE
2461{
2462 unsigned int cnt = hdev->acl_cnt;
2463 struct hci_chan *chan;
2464 struct sk_buff *skb;
2465 int quote;
2466
2467 __check_timeout(hdev, cnt);
04837f64 2468
73d80deb 2469 while (hdev->acl_cnt &&
a8c5fb1a 2470 (chan = hci_chan_sent(hdev, ACL_LINK, &quote))) {
ec1cce24
LAD
2471 u32 priority = (skb_peek(&chan->data_q))->priority;
2472 while (quote-- && (skb = skb_peek(&chan->data_q))) {
73d80deb 2473 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2474 skb->len, skb->priority);
73d80deb 2475
ec1cce24
LAD
2476 /* Stop if priority has changed */
2477 if (skb->priority < priority)
2478 break;
2479
2480 skb = skb_dequeue(&chan->data_q);
2481
73d80deb 2482 hci_conn_enter_active_mode(chan->conn,
04124681 2483 bt_cb(skb)->force_active);
04837f64 2484
1da177e4
LT
2485 hci_send_frame(skb);
2486 hdev->acl_last_tx = jiffies;
2487
2488 hdev->acl_cnt--;
73d80deb
LAD
2489 chan->sent++;
2490 chan->conn->sent++;
1da177e4
LT
2491 }
2492 }
02b20f0b
LAD
2493
2494 if (cnt != hdev->acl_cnt)
2495 hci_prio_recalculate(hdev, ACL_LINK);
1da177e4
LT
2496}
2497
6039aa73 2498static void hci_sched_acl_blk(struct hci_dev *hdev)
b71d385a 2499{
63d2bc1b 2500 unsigned int cnt = hdev->block_cnt;
b71d385a
AE
2501 struct hci_chan *chan;
2502 struct sk_buff *skb;
2503 int quote;
b71d385a 2504
63d2bc1b 2505 __check_timeout(hdev, cnt);
b71d385a
AE
2506
2507 while (hdev->block_cnt > 0 &&
a8c5fb1a 2508 (chan = hci_chan_sent(hdev, ACL_LINK, &quote))) {
b71d385a
AE
2509 u32 priority = (skb_peek(&chan->data_q))->priority;
2510 while (quote > 0 && (skb = skb_peek(&chan->data_q))) {
2511 int blocks;
2512
2513 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2514 skb->len, skb->priority);
b71d385a
AE
2515
2516 /* Stop if priority has changed */
2517 if (skb->priority < priority)
2518 break;
2519
2520 skb = skb_dequeue(&chan->data_q);
2521
2522 blocks = __get_blocks(hdev, skb);
2523 if (blocks > hdev->block_cnt)
2524 return;
2525
2526 hci_conn_enter_active_mode(chan->conn,
a8c5fb1a 2527 bt_cb(skb)->force_active);
b71d385a
AE
2528
2529 hci_send_frame(skb);
2530 hdev->acl_last_tx = jiffies;
2531
2532 hdev->block_cnt -= blocks;
2533 quote -= blocks;
2534
2535 chan->sent += blocks;
2536 chan->conn->sent += blocks;
2537 }
2538 }
2539
2540 if (cnt != hdev->block_cnt)
2541 hci_prio_recalculate(hdev, ACL_LINK);
2542}
2543
6039aa73 2544static void hci_sched_acl(struct hci_dev *hdev)
b71d385a
AE
2545{
2546 BT_DBG("%s", hdev->name);
2547
2548 if (!hci_conn_num(hdev, ACL_LINK))
2549 return;
2550
2551 switch (hdev->flow_ctl_mode) {
2552 case HCI_FLOW_CTL_MODE_PACKET_BASED:
2553 hci_sched_acl_pkt(hdev);
2554 break;
2555
2556 case HCI_FLOW_CTL_MODE_BLOCK_BASED:
2557 hci_sched_acl_blk(hdev);
2558 break;
2559 }
2560}
2561
1da177e4 2562/* Schedule SCO */
6039aa73 2563static void hci_sched_sco(struct hci_dev *hdev)
1da177e4
LT
2564{
2565 struct hci_conn *conn;
2566 struct sk_buff *skb;
2567 int quote;
2568
2569 BT_DBG("%s", hdev->name);
2570
52087a79
LAD
2571 if (!hci_conn_num(hdev, SCO_LINK))
2572 return;
2573
1da177e4
LT
2574 while (hdev->sco_cnt && (conn = hci_low_sent(hdev, SCO_LINK, &quote))) {
2575 while (quote-- && (skb = skb_dequeue(&conn->data_q))) {
2576 BT_DBG("skb %p len %d", skb, skb->len);
2577 hci_send_frame(skb);
2578
2579 conn->sent++;
2580 if (conn->sent == ~0)
2581 conn->sent = 0;
2582 }
2583 }
2584}
2585
6039aa73 2586static void hci_sched_esco(struct hci_dev *hdev)
b6a0dc82
MH
2587{
2588 struct hci_conn *conn;
2589 struct sk_buff *skb;
2590 int quote;
2591
2592 BT_DBG("%s", hdev->name);
2593
52087a79
LAD
2594 if (!hci_conn_num(hdev, ESCO_LINK))
2595 return;
2596
8fc9ced3
GP
2597 while (hdev->sco_cnt && (conn = hci_low_sent(hdev, ESCO_LINK,
2598 &quote))) {
b6a0dc82
MH
2599 while (quote-- && (skb = skb_dequeue(&conn->data_q))) {
2600 BT_DBG("skb %p len %d", skb, skb->len);
2601 hci_send_frame(skb);
2602
2603 conn->sent++;
2604 if (conn->sent == ~0)
2605 conn->sent = 0;
2606 }
2607 }
2608}
2609
6039aa73 2610static void hci_sched_le(struct hci_dev *hdev)
6ed58ec5 2611{
73d80deb 2612 struct hci_chan *chan;
6ed58ec5 2613 struct sk_buff *skb;
02b20f0b 2614 int quote, cnt, tmp;
6ed58ec5
VT
2615
2616 BT_DBG("%s", hdev->name);
2617
52087a79
LAD
2618 if (!hci_conn_num(hdev, LE_LINK))
2619 return;
2620
6ed58ec5
VT
2621 if (!test_bit(HCI_RAW, &hdev->flags)) {
2622 /* LE tx timeout must be longer than maximum
2623 * link supervision timeout (40.9 seconds) */
bae1f5d9 2624 if (!hdev->le_cnt && hdev->le_pkts &&
a8c5fb1a 2625 time_after(jiffies, hdev->le_last_tx + HZ * 45))
bae1f5d9 2626 hci_link_tx_to(hdev, LE_LINK);
6ed58ec5
VT
2627 }
2628
2629 cnt = hdev->le_pkts ? hdev->le_cnt : hdev->acl_cnt;
02b20f0b 2630 tmp = cnt;
73d80deb 2631 while (cnt && (chan = hci_chan_sent(hdev, LE_LINK, &quote))) {
ec1cce24
LAD
2632 u32 priority = (skb_peek(&chan->data_q))->priority;
2633 while (quote-- && (skb = skb_peek(&chan->data_q))) {
73d80deb 2634 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2635 skb->len, skb->priority);
6ed58ec5 2636
ec1cce24
LAD
2637 /* Stop if priority has changed */
2638 if (skb->priority < priority)
2639 break;
2640
2641 skb = skb_dequeue(&chan->data_q);
2642
6ed58ec5
VT
2643 hci_send_frame(skb);
2644 hdev->le_last_tx = jiffies;
2645
2646 cnt--;
73d80deb
LAD
2647 chan->sent++;
2648 chan->conn->sent++;
6ed58ec5
VT
2649 }
2650 }
73d80deb 2651
6ed58ec5
VT
2652 if (hdev->le_pkts)
2653 hdev->le_cnt = cnt;
2654 else
2655 hdev->acl_cnt = cnt;
02b20f0b
LAD
2656
2657 if (cnt != tmp)
2658 hci_prio_recalculate(hdev, LE_LINK);
6ed58ec5
VT
2659}
2660
3eff45ea 2661static void hci_tx_work(struct work_struct *work)
1da177e4 2662{
3eff45ea 2663 struct hci_dev *hdev = container_of(work, struct hci_dev, tx_work);
1da177e4
LT
2664 struct sk_buff *skb;
2665
6ed58ec5 2666 BT_DBG("%s acl %d sco %d le %d", hdev->name, hdev->acl_cnt,
a8c5fb1a 2667 hdev->sco_cnt, hdev->le_cnt);
1da177e4
LT
2668
2669 /* Schedule queues and send stuff to HCI driver */
2670
2671 hci_sched_acl(hdev);
2672
2673 hci_sched_sco(hdev);
2674
b6a0dc82
MH
2675 hci_sched_esco(hdev);
2676
6ed58ec5
VT
2677 hci_sched_le(hdev);
2678
1da177e4
LT
2679 /* Send next queued raw (unknown type) packet */
2680 while ((skb = skb_dequeue(&hdev->raw_q)))
2681 hci_send_frame(skb);
1da177e4
LT
2682}
2683
25985edc 2684/* ----- HCI RX task (incoming data processing) ----- */
1da177e4
LT
2685
2686/* ACL data packet */
6039aa73 2687static void hci_acldata_packet(struct hci_dev *hdev, struct sk_buff *skb)
1da177e4
LT
2688{
2689 struct hci_acl_hdr *hdr = (void *) skb->data;
2690 struct hci_conn *conn;
2691 __u16 handle, flags;
2692
2693 skb_pull(skb, HCI_ACL_HDR_SIZE);
2694
2695 handle = __le16_to_cpu(hdr->handle);
2696 flags = hci_flags(handle);
2697 handle = hci_handle(handle);
2698
a8c5fb1a
GP
2699 BT_DBG("%s len %d handle 0x%x flags 0x%x", hdev->name, skb->len,
2700 handle, flags);
1da177e4
LT
2701
2702 hdev->stat.acl_rx++;
2703
2704 hci_dev_lock(hdev);
2705 conn = hci_conn_hash_lookup_handle(hdev, handle);
2706 hci_dev_unlock(hdev);
8e87d142 2707
1da177e4 2708 if (conn) {
65983fc7 2709 hci_conn_enter_active_mode(conn, BT_POWER_FORCE_ACTIVE_OFF);
04837f64 2710
671267bf
JH
2711 hci_dev_lock(hdev);
2712 if (test_bit(HCI_MGMT, &hdev->dev_flags) &&
2713 !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2714 mgmt_device_connected(hdev, &conn->dst, conn->type,
2715 conn->dst_type, 0, NULL, 0,
2716 conn->dev_class);
2717 hci_dev_unlock(hdev);
2718
1da177e4 2719 /* Send to upper protocol */
686ebf28
UF
2720 l2cap_recv_acldata(conn, skb, flags);
2721 return;
1da177e4 2722 } else {
8e87d142 2723 BT_ERR("%s ACL packet for unknown connection handle %d",
a8c5fb1a 2724 hdev->name, handle);
1da177e4
LT
2725 }
2726
2727 kfree_skb(skb);
2728}
2729
2730/* SCO data packet */
6039aa73 2731static void hci_scodata_packet(struct hci_dev *hdev, struct sk_buff *skb)
1da177e4
LT
2732{
2733 struct hci_sco_hdr *hdr = (void *) skb->data;
2734 struct hci_conn *conn;
2735 __u16 handle;
2736
2737 skb_pull(skb, HCI_SCO_HDR_SIZE);
2738
2739 handle = __le16_to_cpu(hdr->handle);
2740
2741 BT_DBG("%s len %d handle 0x%x", hdev->name, skb->len, handle);
2742
2743 hdev->stat.sco_rx++;
2744
2745 hci_dev_lock(hdev);
2746 conn = hci_conn_hash_lookup_handle(hdev, handle);
2747 hci_dev_unlock(hdev);
2748
2749 if (conn) {
1da177e4 2750 /* Send to upper protocol */
686ebf28
UF
2751 sco_recv_scodata(conn, skb);
2752 return;
1da177e4 2753 } else {
8e87d142 2754 BT_ERR("%s SCO packet for unknown connection handle %d",
a8c5fb1a 2755 hdev->name, handle);
1da177e4
LT
2756 }
2757
2758 kfree_skb(skb);
2759}
2760
b78752cc 2761static void hci_rx_work(struct work_struct *work)
1da177e4 2762{
b78752cc 2763 struct hci_dev *hdev = container_of(work, struct hci_dev, rx_work);
1da177e4
LT
2764 struct sk_buff *skb;
2765
2766 BT_DBG("%s", hdev->name);
2767
1da177e4 2768 while ((skb = skb_dequeue(&hdev->rx_q))) {
cd82e61c
MH
2769 /* Send copy to monitor */
2770 hci_send_to_monitor(hdev, skb);
2771
1da177e4
LT
2772 if (atomic_read(&hdev->promisc)) {
2773 /* Send copy to the sockets */
470fe1b5 2774 hci_send_to_sock(hdev, skb);
1da177e4
LT
2775 }
2776
2777 if (test_bit(HCI_RAW, &hdev->flags)) {
2778 kfree_skb(skb);
2779 continue;
2780 }
2781
2782 if (test_bit(HCI_INIT, &hdev->flags)) {
2783 /* Don't process data packets in this states. */
0d48d939 2784 switch (bt_cb(skb)->pkt_type) {
1da177e4
LT
2785 case HCI_ACLDATA_PKT:
2786 case HCI_SCODATA_PKT:
2787 kfree_skb(skb);
2788 continue;
3ff50b79 2789 }
1da177e4
LT
2790 }
2791
2792 /* Process frame */
0d48d939 2793 switch (bt_cb(skb)->pkt_type) {
1da177e4 2794 case HCI_EVENT_PKT:
b78752cc 2795 BT_DBG("%s Event packet", hdev->name);
1da177e4
LT
2796 hci_event_packet(hdev, skb);
2797 break;
2798
2799 case HCI_ACLDATA_PKT:
2800 BT_DBG("%s ACL data packet", hdev->name);
2801 hci_acldata_packet(hdev, skb);
2802 break;
2803
2804 case HCI_SCODATA_PKT:
2805 BT_DBG("%s SCO data packet", hdev->name);
2806 hci_scodata_packet(hdev, skb);
2807 break;
2808
2809 default:
2810 kfree_skb(skb);
2811 break;
2812 }
2813 }
1da177e4
LT
2814}
2815
c347b765 2816static void hci_cmd_work(struct work_struct *work)
1da177e4 2817{
c347b765 2818 struct hci_dev *hdev = container_of(work, struct hci_dev, cmd_work);
1da177e4
LT
2819 struct sk_buff *skb;
2820
2821 BT_DBG("%s cmd %d", hdev->name, atomic_read(&hdev->cmd_cnt));
2822
1da177e4 2823 /* Send queued commands */
5a08ecce
AE
2824 if (atomic_read(&hdev->cmd_cnt)) {
2825 skb = skb_dequeue(&hdev->cmd_q);
2826 if (!skb)
2827 return;
2828
7585b97a 2829 kfree_skb(hdev->sent_cmd);
1da177e4 2830
70f23020
AE
2831 hdev->sent_cmd = skb_clone(skb, GFP_ATOMIC);
2832 if (hdev->sent_cmd) {
1da177e4
LT
2833 atomic_dec(&hdev->cmd_cnt);
2834 hci_send_frame(skb);
7bdb8a5c
SJ
2835 if (test_bit(HCI_RESET, &hdev->flags))
2836 del_timer(&hdev->cmd_timer);
2837 else
2838 mod_timer(&hdev->cmd_timer,
5f246e89 2839 jiffies + HCI_CMD_TIMEOUT);
1da177e4
LT
2840 } else {
2841 skb_queue_head(&hdev->cmd_q, skb);
c347b765 2842 queue_work(hdev->workqueue, &hdev->cmd_work);
1da177e4
LT
2843 }
2844 }
2845}
2519a1fc
AG
2846
2847int hci_do_inquiry(struct hci_dev *hdev, u8 length)
2848{
2849 /* General inquiry access code (GIAC) */
2850 u8 lap[3] = { 0x33, 0x8b, 0x9e };
2851 struct hci_cp_inquiry cp;
2852
2853 BT_DBG("%s", hdev->name);
2854
2855 if (test_bit(HCI_INQUIRY, &hdev->flags))
2856 return -EINPROGRESS;
2857
4663262c
JH
2858 inquiry_cache_flush(hdev);
2859
2519a1fc
AG
2860 memset(&cp, 0, sizeof(cp));
2861 memcpy(&cp.lap, lap, sizeof(cp.lap));
2862 cp.length = length;
2863
2864 return hci_send_cmd(hdev, HCI_OP_INQUIRY, sizeof(cp), &cp);
2865}
023d5049
AG
2866
2867int hci_cancel_inquiry(struct hci_dev *hdev)
2868{
2869 BT_DBG("%s", hdev->name);
2870
2871 if (!test_bit(HCI_INQUIRY, &hdev->flags))
7537e5c3 2872 return -EALREADY;
023d5049
AG
2873
2874 return hci_send_cmd(hdev, HCI_OP_INQUIRY_CANCEL, 0, NULL);
2875}
31f7956c
AG
2876
2877u8 bdaddr_to_le(u8 bdaddr_type)
2878{
2879 switch (bdaddr_type) {
2880 case BDADDR_LE_PUBLIC:
2881 return ADDR_LE_DEV_PUBLIC;
2882
2883 default:
2884 /* Fallback to LE Random address type */
2885 return ADDR_LE_DEV_RANDOM;
2886 }
2887}
This page took 1.005769 seconds and 5 git commands to generate.