Bluetooth: hci-uart-ath: Correct the comments in this driver
[deliverable/linux.git] / net / bluetooth / hci_event.c
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 /* Bluetooth HCI event handling. */
26
27 #include <asm/unaligned.h>
28
29 #include <net/bluetooth/bluetooth.h>
30 #include <net/bluetooth/hci_core.h>
31 #include <net/bluetooth/mgmt.h>
32
33 #include "a2mp.h"
34 #include "amp.h"
35 #include "smp.h"
36
37 /* Handle HCI Event packets */
38
39 static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb)
40 {
41 __u8 status = *((__u8 *) skb->data);
42
43 BT_DBG("%s status 0x%2.2x", hdev->name, status);
44
45 if (status)
46 return;
47
48 clear_bit(HCI_INQUIRY, &hdev->flags);
49 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
50 wake_up_bit(&hdev->flags, HCI_INQUIRY);
51
52 hci_dev_lock(hdev);
53 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
54 hci_dev_unlock(hdev);
55
56 hci_conn_check_pending(hdev);
57 }
58
59 static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
60 {
61 __u8 status = *((__u8 *) skb->data);
62
63 BT_DBG("%s status 0x%2.2x", hdev->name, status);
64
65 if (status)
66 return;
67
68 set_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
69 }
70
71 static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
72 {
73 __u8 status = *((__u8 *) skb->data);
74
75 BT_DBG("%s status 0x%2.2x", hdev->name, status);
76
77 if (status)
78 return;
79
80 clear_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
81
82 hci_conn_check_pending(hdev);
83 }
84
85 static void hci_cc_remote_name_req_cancel(struct hci_dev *hdev,
86 struct sk_buff *skb)
87 {
88 BT_DBG("%s", hdev->name);
89 }
90
91 static void hci_cc_role_discovery(struct hci_dev *hdev, struct sk_buff *skb)
92 {
93 struct hci_rp_role_discovery *rp = (void *) skb->data;
94 struct hci_conn *conn;
95
96 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
97
98 if (rp->status)
99 return;
100
101 hci_dev_lock(hdev);
102
103 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
104 if (conn)
105 conn->role = rp->role;
106
107 hci_dev_unlock(hdev);
108 }
109
110 static void hci_cc_read_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
111 {
112 struct hci_rp_read_link_policy *rp = (void *) skb->data;
113 struct hci_conn *conn;
114
115 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
116
117 if (rp->status)
118 return;
119
120 hci_dev_lock(hdev);
121
122 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
123 if (conn)
124 conn->link_policy = __le16_to_cpu(rp->policy);
125
126 hci_dev_unlock(hdev);
127 }
128
129 static void hci_cc_write_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
130 {
131 struct hci_rp_write_link_policy *rp = (void *) skb->data;
132 struct hci_conn *conn;
133 void *sent;
134
135 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
136
137 if (rp->status)
138 return;
139
140 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LINK_POLICY);
141 if (!sent)
142 return;
143
144 hci_dev_lock(hdev);
145
146 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
147 if (conn)
148 conn->link_policy = get_unaligned_le16(sent + 2);
149
150 hci_dev_unlock(hdev);
151 }
152
153 static void hci_cc_read_def_link_policy(struct hci_dev *hdev,
154 struct sk_buff *skb)
155 {
156 struct hci_rp_read_def_link_policy *rp = (void *) skb->data;
157
158 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
159
160 if (rp->status)
161 return;
162
163 hdev->link_policy = __le16_to_cpu(rp->policy);
164 }
165
166 static void hci_cc_write_def_link_policy(struct hci_dev *hdev,
167 struct sk_buff *skb)
168 {
169 __u8 status = *((__u8 *) skb->data);
170 void *sent;
171
172 BT_DBG("%s status 0x%2.2x", hdev->name, status);
173
174 if (status)
175 return;
176
177 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_DEF_LINK_POLICY);
178 if (!sent)
179 return;
180
181 hdev->link_policy = get_unaligned_le16(sent);
182 }
183
184 static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb)
185 {
186 __u8 status = *((__u8 *) skb->data);
187
188 BT_DBG("%s status 0x%2.2x", hdev->name, status);
189
190 clear_bit(HCI_RESET, &hdev->flags);
191
192 if (status)
193 return;
194
195 /* Reset all non-persistent flags */
196 hdev->dev_flags &= ~HCI_PERSISTENT_MASK;
197
198 hdev->discovery.state = DISCOVERY_STOPPED;
199 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
200 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
201
202 memset(hdev->adv_data, 0, sizeof(hdev->adv_data));
203 hdev->adv_data_len = 0;
204
205 memset(hdev->scan_rsp_data, 0, sizeof(hdev->scan_rsp_data));
206 hdev->scan_rsp_data_len = 0;
207
208 hdev->le_scan_type = LE_SCAN_PASSIVE;
209
210 hdev->ssp_debug_mode = 0;
211
212 hci_bdaddr_list_clear(&hdev->le_white_list);
213 }
214
215 static void hci_cc_write_local_name(struct hci_dev *hdev, struct sk_buff *skb)
216 {
217 __u8 status = *((__u8 *) skb->data);
218 void *sent;
219
220 BT_DBG("%s status 0x%2.2x", hdev->name, status);
221
222 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LOCAL_NAME);
223 if (!sent)
224 return;
225
226 hci_dev_lock(hdev);
227
228 if (test_bit(HCI_MGMT, &hdev->dev_flags))
229 mgmt_set_local_name_complete(hdev, sent, status);
230 else if (!status)
231 memcpy(hdev->dev_name, sent, HCI_MAX_NAME_LENGTH);
232
233 hci_dev_unlock(hdev);
234 }
235
236 static void hci_cc_read_local_name(struct hci_dev *hdev, struct sk_buff *skb)
237 {
238 struct hci_rp_read_local_name *rp = (void *) skb->data;
239
240 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
241
242 if (rp->status)
243 return;
244
245 if (test_bit(HCI_SETUP, &hdev->dev_flags))
246 memcpy(hdev->dev_name, rp->name, HCI_MAX_NAME_LENGTH);
247 }
248
249 static void hci_cc_write_auth_enable(struct hci_dev *hdev, struct sk_buff *skb)
250 {
251 __u8 status = *((__u8 *) skb->data);
252 void *sent;
253
254 BT_DBG("%s status 0x%2.2x", hdev->name, status);
255
256 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_AUTH_ENABLE);
257 if (!sent)
258 return;
259
260 if (!status) {
261 __u8 param = *((__u8 *) sent);
262
263 if (param == AUTH_ENABLED)
264 set_bit(HCI_AUTH, &hdev->flags);
265 else
266 clear_bit(HCI_AUTH, &hdev->flags);
267 }
268
269 if (test_bit(HCI_MGMT, &hdev->dev_flags))
270 mgmt_auth_enable_complete(hdev, status);
271 }
272
273 static void hci_cc_write_encrypt_mode(struct hci_dev *hdev, struct sk_buff *skb)
274 {
275 __u8 status = *((__u8 *) skb->data);
276 __u8 param;
277 void *sent;
278
279 BT_DBG("%s status 0x%2.2x", hdev->name, status);
280
281 if (status)
282 return;
283
284 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_ENCRYPT_MODE);
285 if (!sent)
286 return;
287
288 param = *((__u8 *) sent);
289
290 if (param)
291 set_bit(HCI_ENCRYPT, &hdev->flags);
292 else
293 clear_bit(HCI_ENCRYPT, &hdev->flags);
294 }
295
296 static void hci_cc_write_scan_enable(struct hci_dev *hdev, struct sk_buff *skb)
297 {
298 __u8 status = *((__u8 *) skb->data);
299 __u8 param;
300 void *sent;
301
302 BT_DBG("%s status 0x%2.2x", hdev->name, status);
303
304 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SCAN_ENABLE);
305 if (!sent)
306 return;
307
308 param = *((__u8 *) sent);
309
310 hci_dev_lock(hdev);
311
312 if (status) {
313 hdev->discov_timeout = 0;
314 goto done;
315 }
316
317 if (param & SCAN_INQUIRY)
318 set_bit(HCI_ISCAN, &hdev->flags);
319 else
320 clear_bit(HCI_ISCAN, &hdev->flags);
321
322 if (param & SCAN_PAGE)
323 set_bit(HCI_PSCAN, &hdev->flags);
324 else
325 clear_bit(HCI_PSCAN, &hdev->flags);
326
327 done:
328 hci_dev_unlock(hdev);
329 }
330
331 static void hci_cc_read_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
332 {
333 struct hci_rp_read_class_of_dev *rp = (void *) skb->data;
334
335 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
336
337 if (rp->status)
338 return;
339
340 memcpy(hdev->dev_class, rp->dev_class, 3);
341
342 BT_DBG("%s class 0x%.2x%.2x%.2x", hdev->name,
343 hdev->dev_class[2], hdev->dev_class[1], hdev->dev_class[0]);
344 }
345
346 static void hci_cc_write_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
347 {
348 __u8 status = *((__u8 *) skb->data);
349 void *sent;
350
351 BT_DBG("%s status 0x%2.2x", hdev->name, status);
352
353 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_CLASS_OF_DEV);
354 if (!sent)
355 return;
356
357 hci_dev_lock(hdev);
358
359 if (status == 0)
360 memcpy(hdev->dev_class, sent, 3);
361
362 if (test_bit(HCI_MGMT, &hdev->dev_flags))
363 mgmt_set_class_of_dev_complete(hdev, sent, status);
364
365 hci_dev_unlock(hdev);
366 }
367
368 static void hci_cc_read_voice_setting(struct hci_dev *hdev, struct sk_buff *skb)
369 {
370 struct hci_rp_read_voice_setting *rp = (void *) skb->data;
371 __u16 setting;
372
373 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
374
375 if (rp->status)
376 return;
377
378 setting = __le16_to_cpu(rp->voice_setting);
379
380 if (hdev->voice_setting == setting)
381 return;
382
383 hdev->voice_setting = setting;
384
385 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
386
387 if (hdev->notify)
388 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
389 }
390
391 static void hci_cc_write_voice_setting(struct hci_dev *hdev,
392 struct sk_buff *skb)
393 {
394 __u8 status = *((__u8 *) skb->data);
395 __u16 setting;
396 void *sent;
397
398 BT_DBG("%s status 0x%2.2x", hdev->name, status);
399
400 if (status)
401 return;
402
403 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_VOICE_SETTING);
404 if (!sent)
405 return;
406
407 setting = get_unaligned_le16(sent);
408
409 if (hdev->voice_setting == setting)
410 return;
411
412 hdev->voice_setting = setting;
413
414 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
415
416 if (hdev->notify)
417 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
418 }
419
420 static void hci_cc_read_num_supported_iac(struct hci_dev *hdev,
421 struct sk_buff *skb)
422 {
423 struct hci_rp_read_num_supported_iac *rp = (void *) skb->data;
424
425 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
426
427 if (rp->status)
428 return;
429
430 hdev->num_iac = rp->num_iac;
431
432 BT_DBG("%s num iac %d", hdev->name, hdev->num_iac);
433 }
434
435 static void hci_cc_write_ssp_mode(struct hci_dev *hdev, struct sk_buff *skb)
436 {
437 __u8 status = *((__u8 *) skb->data);
438 struct hci_cp_write_ssp_mode *sent;
439
440 BT_DBG("%s status 0x%2.2x", hdev->name, status);
441
442 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_MODE);
443 if (!sent)
444 return;
445
446 if (!status) {
447 if (sent->mode)
448 hdev->features[1][0] |= LMP_HOST_SSP;
449 else
450 hdev->features[1][0] &= ~LMP_HOST_SSP;
451 }
452
453 if (test_bit(HCI_MGMT, &hdev->dev_flags))
454 mgmt_ssp_enable_complete(hdev, sent->mode, status);
455 else if (!status) {
456 if (sent->mode)
457 set_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
458 else
459 clear_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
460 }
461 }
462
463 static void hci_cc_write_sc_support(struct hci_dev *hdev, struct sk_buff *skb)
464 {
465 u8 status = *((u8 *) skb->data);
466 struct hci_cp_write_sc_support *sent;
467
468 BT_DBG("%s status 0x%2.2x", hdev->name, status);
469
470 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SC_SUPPORT);
471 if (!sent)
472 return;
473
474 if (!status) {
475 if (sent->support)
476 hdev->features[1][0] |= LMP_HOST_SC;
477 else
478 hdev->features[1][0] &= ~LMP_HOST_SC;
479 }
480
481 if (test_bit(HCI_MGMT, &hdev->dev_flags))
482 mgmt_sc_enable_complete(hdev, sent->support, status);
483 else if (!status) {
484 if (sent->support)
485 set_bit(HCI_SC_ENABLED, &hdev->dev_flags);
486 else
487 clear_bit(HCI_SC_ENABLED, &hdev->dev_flags);
488 }
489 }
490
491 static void hci_cc_read_local_version(struct hci_dev *hdev, struct sk_buff *skb)
492 {
493 struct hci_rp_read_local_version *rp = (void *) skb->data;
494
495 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
496
497 if (rp->status)
498 return;
499
500 if (test_bit(HCI_SETUP, &hdev->dev_flags)) {
501 hdev->hci_ver = rp->hci_ver;
502 hdev->hci_rev = __le16_to_cpu(rp->hci_rev);
503 hdev->lmp_ver = rp->lmp_ver;
504 hdev->manufacturer = __le16_to_cpu(rp->manufacturer);
505 hdev->lmp_subver = __le16_to_cpu(rp->lmp_subver);
506 }
507 }
508
509 static void hci_cc_read_local_commands(struct hci_dev *hdev,
510 struct sk_buff *skb)
511 {
512 struct hci_rp_read_local_commands *rp = (void *) skb->data;
513
514 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
515
516 if (rp->status)
517 return;
518
519 if (test_bit(HCI_SETUP, &hdev->dev_flags))
520 memcpy(hdev->commands, rp->commands, sizeof(hdev->commands));
521 }
522
523 static void hci_cc_read_local_features(struct hci_dev *hdev,
524 struct sk_buff *skb)
525 {
526 struct hci_rp_read_local_features *rp = (void *) skb->data;
527
528 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
529
530 if (rp->status)
531 return;
532
533 memcpy(hdev->features, rp->features, 8);
534
535 /* Adjust default settings according to features
536 * supported by device. */
537
538 if (hdev->features[0][0] & LMP_3SLOT)
539 hdev->pkt_type |= (HCI_DM3 | HCI_DH3);
540
541 if (hdev->features[0][0] & LMP_5SLOT)
542 hdev->pkt_type |= (HCI_DM5 | HCI_DH5);
543
544 if (hdev->features[0][1] & LMP_HV2) {
545 hdev->pkt_type |= (HCI_HV2);
546 hdev->esco_type |= (ESCO_HV2);
547 }
548
549 if (hdev->features[0][1] & LMP_HV3) {
550 hdev->pkt_type |= (HCI_HV3);
551 hdev->esco_type |= (ESCO_HV3);
552 }
553
554 if (lmp_esco_capable(hdev))
555 hdev->esco_type |= (ESCO_EV3);
556
557 if (hdev->features[0][4] & LMP_EV4)
558 hdev->esco_type |= (ESCO_EV4);
559
560 if (hdev->features[0][4] & LMP_EV5)
561 hdev->esco_type |= (ESCO_EV5);
562
563 if (hdev->features[0][5] & LMP_EDR_ESCO_2M)
564 hdev->esco_type |= (ESCO_2EV3);
565
566 if (hdev->features[0][5] & LMP_EDR_ESCO_3M)
567 hdev->esco_type |= (ESCO_3EV3);
568
569 if (hdev->features[0][5] & LMP_EDR_3S_ESCO)
570 hdev->esco_type |= (ESCO_2EV5 | ESCO_3EV5);
571 }
572
573 static void hci_cc_read_local_ext_features(struct hci_dev *hdev,
574 struct sk_buff *skb)
575 {
576 struct hci_rp_read_local_ext_features *rp = (void *) skb->data;
577
578 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
579
580 if (rp->status)
581 return;
582
583 if (hdev->max_page < rp->max_page)
584 hdev->max_page = rp->max_page;
585
586 if (rp->page < HCI_MAX_PAGES)
587 memcpy(hdev->features[rp->page], rp->features, 8);
588 }
589
590 static void hci_cc_read_flow_control_mode(struct hci_dev *hdev,
591 struct sk_buff *skb)
592 {
593 struct hci_rp_read_flow_control_mode *rp = (void *) skb->data;
594
595 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
596
597 if (rp->status)
598 return;
599
600 hdev->flow_ctl_mode = rp->mode;
601 }
602
603 static void hci_cc_read_buffer_size(struct hci_dev *hdev, struct sk_buff *skb)
604 {
605 struct hci_rp_read_buffer_size *rp = (void *) skb->data;
606
607 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
608
609 if (rp->status)
610 return;
611
612 hdev->acl_mtu = __le16_to_cpu(rp->acl_mtu);
613 hdev->sco_mtu = rp->sco_mtu;
614 hdev->acl_pkts = __le16_to_cpu(rp->acl_max_pkt);
615 hdev->sco_pkts = __le16_to_cpu(rp->sco_max_pkt);
616
617 if (test_bit(HCI_QUIRK_FIXUP_BUFFER_SIZE, &hdev->quirks)) {
618 hdev->sco_mtu = 64;
619 hdev->sco_pkts = 8;
620 }
621
622 hdev->acl_cnt = hdev->acl_pkts;
623 hdev->sco_cnt = hdev->sco_pkts;
624
625 BT_DBG("%s acl mtu %d:%d sco mtu %d:%d", hdev->name, hdev->acl_mtu,
626 hdev->acl_pkts, hdev->sco_mtu, hdev->sco_pkts);
627 }
628
629 static void hci_cc_read_bd_addr(struct hci_dev *hdev, struct sk_buff *skb)
630 {
631 struct hci_rp_read_bd_addr *rp = (void *) skb->data;
632
633 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
634
635 if (rp->status)
636 return;
637
638 if (test_bit(HCI_INIT, &hdev->flags))
639 bacpy(&hdev->bdaddr, &rp->bdaddr);
640
641 if (test_bit(HCI_SETUP, &hdev->dev_flags))
642 bacpy(&hdev->setup_addr, &rp->bdaddr);
643 }
644
645 static void hci_cc_read_page_scan_activity(struct hci_dev *hdev,
646 struct sk_buff *skb)
647 {
648 struct hci_rp_read_page_scan_activity *rp = (void *) skb->data;
649
650 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
651
652 if (rp->status)
653 return;
654
655 if (test_bit(HCI_INIT, &hdev->flags)) {
656 hdev->page_scan_interval = __le16_to_cpu(rp->interval);
657 hdev->page_scan_window = __le16_to_cpu(rp->window);
658 }
659 }
660
661 static void hci_cc_write_page_scan_activity(struct hci_dev *hdev,
662 struct sk_buff *skb)
663 {
664 u8 status = *((u8 *) skb->data);
665 struct hci_cp_write_page_scan_activity *sent;
666
667 BT_DBG("%s status 0x%2.2x", hdev->name, status);
668
669 if (status)
670 return;
671
672 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY);
673 if (!sent)
674 return;
675
676 hdev->page_scan_interval = __le16_to_cpu(sent->interval);
677 hdev->page_scan_window = __le16_to_cpu(sent->window);
678 }
679
680 static void hci_cc_read_page_scan_type(struct hci_dev *hdev,
681 struct sk_buff *skb)
682 {
683 struct hci_rp_read_page_scan_type *rp = (void *) skb->data;
684
685 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
686
687 if (rp->status)
688 return;
689
690 if (test_bit(HCI_INIT, &hdev->flags))
691 hdev->page_scan_type = rp->type;
692 }
693
694 static void hci_cc_write_page_scan_type(struct hci_dev *hdev,
695 struct sk_buff *skb)
696 {
697 u8 status = *((u8 *) skb->data);
698 u8 *type;
699
700 BT_DBG("%s status 0x%2.2x", hdev->name, status);
701
702 if (status)
703 return;
704
705 type = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_TYPE);
706 if (type)
707 hdev->page_scan_type = *type;
708 }
709
710 static void hci_cc_read_data_block_size(struct hci_dev *hdev,
711 struct sk_buff *skb)
712 {
713 struct hci_rp_read_data_block_size *rp = (void *) skb->data;
714
715 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
716
717 if (rp->status)
718 return;
719
720 hdev->block_mtu = __le16_to_cpu(rp->max_acl_len);
721 hdev->block_len = __le16_to_cpu(rp->block_len);
722 hdev->num_blocks = __le16_to_cpu(rp->num_blocks);
723
724 hdev->block_cnt = hdev->num_blocks;
725
726 BT_DBG("%s blk mtu %d cnt %d len %d", hdev->name, hdev->block_mtu,
727 hdev->block_cnt, hdev->block_len);
728 }
729
730 static void hci_cc_read_clock(struct hci_dev *hdev, struct sk_buff *skb)
731 {
732 struct hci_rp_read_clock *rp = (void *) skb->data;
733 struct hci_cp_read_clock *cp;
734 struct hci_conn *conn;
735
736 BT_DBG("%s", hdev->name);
737
738 if (skb->len < sizeof(*rp))
739 return;
740
741 if (rp->status)
742 return;
743
744 hci_dev_lock(hdev);
745
746 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_CLOCK);
747 if (!cp)
748 goto unlock;
749
750 if (cp->which == 0x00) {
751 hdev->clock = le32_to_cpu(rp->clock);
752 goto unlock;
753 }
754
755 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
756 if (conn) {
757 conn->clock = le32_to_cpu(rp->clock);
758 conn->clock_accuracy = le16_to_cpu(rp->accuracy);
759 }
760
761 unlock:
762 hci_dev_unlock(hdev);
763 }
764
765 static void hci_cc_read_local_amp_info(struct hci_dev *hdev,
766 struct sk_buff *skb)
767 {
768 struct hci_rp_read_local_amp_info *rp = (void *) skb->data;
769
770 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
771
772 if (rp->status)
773 goto a2mp_rsp;
774
775 hdev->amp_status = rp->amp_status;
776 hdev->amp_total_bw = __le32_to_cpu(rp->total_bw);
777 hdev->amp_max_bw = __le32_to_cpu(rp->max_bw);
778 hdev->amp_min_latency = __le32_to_cpu(rp->min_latency);
779 hdev->amp_max_pdu = __le32_to_cpu(rp->max_pdu);
780 hdev->amp_type = rp->amp_type;
781 hdev->amp_pal_cap = __le16_to_cpu(rp->pal_cap);
782 hdev->amp_assoc_size = __le16_to_cpu(rp->max_assoc_size);
783 hdev->amp_be_flush_to = __le32_to_cpu(rp->be_flush_to);
784 hdev->amp_max_flush_to = __le32_to_cpu(rp->max_flush_to);
785
786 a2mp_rsp:
787 a2mp_send_getinfo_rsp(hdev);
788 }
789
790 static void hci_cc_read_local_amp_assoc(struct hci_dev *hdev,
791 struct sk_buff *skb)
792 {
793 struct hci_rp_read_local_amp_assoc *rp = (void *) skb->data;
794 struct amp_assoc *assoc = &hdev->loc_assoc;
795 size_t rem_len, frag_len;
796
797 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
798
799 if (rp->status)
800 goto a2mp_rsp;
801
802 frag_len = skb->len - sizeof(*rp);
803 rem_len = __le16_to_cpu(rp->rem_len);
804
805 if (rem_len > frag_len) {
806 BT_DBG("frag_len %zu rem_len %zu", frag_len, rem_len);
807
808 memcpy(assoc->data + assoc->offset, rp->frag, frag_len);
809 assoc->offset += frag_len;
810
811 /* Read other fragments */
812 amp_read_loc_assoc_frag(hdev, rp->phy_handle);
813
814 return;
815 }
816
817 memcpy(assoc->data + assoc->offset, rp->frag, rem_len);
818 assoc->len = assoc->offset + rem_len;
819 assoc->offset = 0;
820
821 a2mp_rsp:
822 /* Send A2MP Rsp when all fragments are received */
823 a2mp_send_getampassoc_rsp(hdev, rp->status);
824 a2mp_send_create_phy_link_req(hdev, rp->status);
825 }
826
827 static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev,
828 struct sk_buff *skb)
829 {
830 struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data;
831
832 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
833
834 if (rp->status)
835 return;
836
837 hdev->inq_tx_power = rp->tx_power;
838 }
839
840 static void hci_cc_pin_code_reply(struct hci_dev *hdev, struct sk_buff *skb)
841 {
842 struct hci_rp_pin_code_reply *rp = (void *) skb->data;
843 struct hci_cp_pin_code_reply *cp;
844 struct hci_conn *conn;
845
846 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
847
848 hci_dev_lock(hdev);
849
850 if (test_bit(HCI_MGMT, &hdev->dev_flags))
851 mgmt_pin_code_reply_complete(hdev, &rp->bdaddr, rp->status);
852
853 if (rp->status)
854 goto unlock;
855
856 cp = hci_sent_cmd_data(hdev, HCI_OP_PIN_CODE_REPLY);
857 if (!cp)
858 goto unlock;
859
860 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
861 if (conn)
862 conn->pin_length = cp->pin_len;
863
864 unlock:
865 hci_dev_unlock(hdev);
866 }
867
868 static void hci_cc_pin_code_neg_reply(struct hci_dev *hdev, struct sk_buff *skb)
869 {
870 struct hci_rp_pin_code_neg_reply *rp = (void *) skb->data;
871
872 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
873
874 hci_dev_lock(hdev);
875
876 if (test_bit(HCI_MGMT, &hdev->dev_flags))
877 mgmt_pin_code_neg_reply_complete(hdev, &rp->bdaddr,
878 rp->status);
879
880 hci_dev_unlock(hdev);
881 }
882
883 static void hci_cc_le_read_buffer_size(struct hci_dev *hdev,
884 struct sk_buff *skb)
885 {
886 struct hci_rp_le_read_buffer_size *rp = (void *) skb->data;
887
888 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
889
890 if (rp->status)
891 return;
892
893 hdev->le_mtu = __le16_to_cpu(rp->le_mtu);
894 hdev->le_pkts = rp->le_max_pkt;
895
896 hdev->le_cnt = hdev->le_pkts;
897
898 BT_DBG("%s le mtu %d:%d", hdev->name, hdev->le_mtu, hdev->le_pkts);
899 }
900
901 static void hci_cc_le_read_local_features(struct hci_dev *hdev,
902 struct sk_buff *skb)
903 {
904 struct hci_rp_le_read_local_features *rp = (void *) skb->data;
905
906 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
907
908 if (rp->status)
909 return;
910
911 memcpy(hdev->le_features, rp->features, 8);
912 }
913
914 static void hci_cc_le_read_adv_tx_power(struct hci_dev *hdev,
915 struct sk_buff *skb)
916 {
917 struct hci_rp_le_read_adv_tx_power *rp = (void *) skb->data;
918
919 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
920
921 if (rp->status)
922 return;
923
924 hdev->adv_tx_power = rp->tx_power;
925 }
926
927 static void hci_cc_user_confirm_reply(struct hci_dev *hdev, struct sk_buff *skb)
928 {
929 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
930
931 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
932
933 hci_dev_lock(hdev);
934
935 if (test_bit(HCI_MGMT, &hdev->dev_flags))
936 mgmt_user_confirm_reply_complete(hdev, &rp->bdaddr, ACL_LINK, 0,
937 rp->status);
938
939 hci_dev_unlock(hdev);
940 }
941
942 static void hci_cc_user_confirm_neg_reply(struct hci_dev *hdev,
943 struct sk_buff *skb)
944 {
945 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
946
947 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
948
949 hci_dev_lock(hdev);
950
951 if (test_bit(HCI_MGMT, &hdev->dev_flags))
952 mgmt_user_confirm_neg_reply_complete(hdev, &rp->bdaddr,
953 ACL_LINK, 0, rp->status);
954
955 hci_dev_unlock(hdev);
956 }
957
958 static void hci_cc_user_passkey_reply(struct hci_dev *hdev, struct sk_buff *skb)
959 {
960 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
961
962 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
963
964 hci_dev_lock(hdev);
965
966 if (test_bit(HCI_MGMT, &hdev->dev_flags))
967 mgmt_user_passkey_reply_complete(hdev, &rp->bdaddr, ACL_LINK,
968 0, rp->status);
969
970 hci_dev_unlock(hdev);
971 }
972
973 static void hci_cc_user_passkey_neg_reply(struct hci_dev *hdev,
974 struct sk_buff *skb)
975 {
976 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
977
978 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
979
980 hci_dev_lock(hdev);
981
982 if (test_bit(HCI_MGMT, &hdev->dev_flags))
983 mgmt_user_passkey_neg_reply_complete(hdev, &rp->bdaddr,
984 ACL_LINK, 0, rp->status);
985
986 hci_dev_unlock(hdev);
987 }
988
989 static void hci_cc_read_local_oob_data(struct hci_dev *hdev,
990 struct sk_buff *skb)
991 {
992 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
993
994 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
995
996 hci_dev_lock(hdev);
997 mgmt_read_local_oob_data_complete(hdev, rp->hash, rp->randomizer,
998 NULL, NULL, rp->status);
999 hci_dev_unlock(hdev);
1000 }
1001
1002 static void hci_cc_read_local_oob_ext_data(struct hci_dev *hdev,
1003 struct sk_buff *skb)
1004 {
1005 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
1006
1007 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1008
1009 hci_dev_lock(hdev);
1010 mgmt_read_local_oob_data_complete(hdev, rp->hash192, rp->randomizer192,
1011 rp->hash256, rp->randomizer256,
1012 rp->status);
1013 hci_dev_unlock(hdev);
1014 }
1015
1016
1017 static void hci_cc_le_set_random_addr(struct hci_dev *hdev, struct sk_buff *skb)
1018 {
1019 __u8 status = *((__u8 *) skb->data);
1020 bdaddr_t *sent;
1021
1022 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1023
1024 if (status)
1025 return;
1026
1027 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_RANDOM_ADDR);
1028 if (!sent)
1029 return;
1030
1031 hci_dev_lock(hdev);
1032
1033 bacpy(&hdev->random_addr, sent);
1034
1035 hci_dev_unlock(hdev);
1036 }
1037
1038 static void hci_cc_le_set_adv_enable(struct hci_dev *hdev, struct sk_buff *skb)
1039 {
1040 __u8 *sent, status = *((__u8 *) skb->data);
1041
1042 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1043
1044 if (status)
1045 return;
1046
1047 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_ENABLE);
1048 if (!sent)
1049 return;
1050
1051 hci_dev_lock(hdev);
1052
1053 /* If we're doing connection initiation as peripheral. Set a
1054 * timeout in case something goes wrong.
1055 */
1056 if (*sent) {
1057 struct hci_conn *conn;
1058
1059 set_bit(HCI_LE_ADV, &hdev->dev_flags);
1060
1061 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
1062 if (conn)
1063 queue_delayed_work(hdev->workqueue,
1064 &conn->le_conn_timeout,
1065 conn->conn_timeout);
1066 } else {
1067 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
1068 }
1069
1070 hci_dev_unlock(hdev);
1071 }
1072
1073 static void hci_cc_le_set_scan_param(struct hci_dev *hdev, struct sk_buff *skb)
1074 {
1075 struct hci_cp_le_set_scan_param *cp;
1076 __u8 status = *((__u8 *) skb->data);
1077
1078 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1079
1080 if (status)
1081 return;
1082
1083 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_PARAM);
1084 if (!cp)
1085 return;
1086
1087 hci_dev_lock(hdev);
1088
1089 hdev->le_scan_type = cp->type;
1090
1091 hci_dev_unlock(hdev);
1092 }
1093
1094 static bool has_pending_adv_report(struct hci_dev *hdev)
1095 {
1096 struct discovery_state *d = &hdev->discovery;
1097
1098 return bacmp(&d->last_adv_addr, BDADDR_ANY);
1099 }
1100
1101 static void clear_pending_adv_report(struct hci_dev *hdev)
1102 {
1103 struct discovery_state *d = &hdev->discovery;
1104
1105 bacpy(&d->last_adv_addr, BDADDR_ANY);
1106 d->last_adv_data_len = 0;
1107 }
1108
1109 static void store_pending_adv_report(struct hci_dev *hdev, bdaddr_t *bdaddr,
1110 u8 bdaddr_type, s8 rssi, u32 flags,
1111 u8 *data, u8 len)
1112 {
1113 struct discovery_state *d = &hdev->discovery;
1114
1115 bacpy(&d->last_adv_addr, bdaddr);
1116 d->last_adv_addr_type = bdaddr_type;
1117 d->last_adv_rssi = rssi;
1118 d->last_adv_flags = flags;
1119 memcpy(d->last_adv_data, data, len);
1120 d->last_adv_data_len = len;
1121 }
1122
1123 static void hci_cc_le_set_scan_enable(struct hci_dev *hdev,
1124 struct sk_buff *skb)
1125 {
1126 struct hci_cp_le_set_scan_enable *cp;
1127 __u8 status = *((__u8 *) skb->data);
1128
1129 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1130
1131 if (status)
1132 return;
1133
1134 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_ENABLE);
1135 if (!cp)
1136 return;
1137
1138 switch (cp->enable) {
1139 case LE_SCAN_ENABLE:
1140 set_bit(HCI_LE_SCAN, &hdev->dev_flags);
1141 if (hdev->le_scan_type == LE_SCAN_ACTIVE)
1142 clear_pending_adv_report(hdev);
1143 break;
1144
1145 case LE_SCAN_DISABLE:
1146 /* We do this here instead of when setting DISCOVERY_STOPPED
1147 * since the latter would potentially require waiting for
1148 * inquiry to stop too.
1149 */
1150 if (has_pending_adv_report(hdev)) {
1151 struct discovery_state *d = &hdev->discovery;
1152
1153 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
1154 d->last_adv_addr_type, NULL,
1155 d->last_adv_rssi, d->last_adv_flags,
1156 d->last_adv_data,
1157 d->last_adv_data_len, NULL, 0);
1158 }
1159
1160 /* Cancel this timer so that we don't try to disable scanning
1161 * when it's already disabled.
1162 */
1163 cancel_delayed_work(&hdev->le_scan_disable);
1164
1165 clear_bit(HCI_LE_SCAN, &hdev->dev_flags);
1166
1167 /* The HCI_LE_SCAN_INTERRUPTED flag indicates that we
1168 * interrupted scanning due to a connect request. Mark
1169 * therefore discovery as stopped. If this was not
1170 * because of a connect request advertising might have
1171 * been disabled because of active scanning, so
1172 * re-enable it again if necessary.
1173 */
1174 if (test_and_clear_bit(HCI_LE_SCAN_INTERRUPTED,
1175 &hdev->dev_flags))
1176 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1177 else if (!test_bit(HCI_LE_ADV, &hdev->dev_flags) &&
1178 hdev->discovery.state == DISCOVERY_FINDING)
1179 mgmt_reenable_advertising(hdev);
1180
1181 break;
1182
1183 default:
1184 BT_ERR("Used reserved LE_Scan_Enable param %d", cp->enable);
1185 break;
1186 }
1187 }
1188
1189 static void hci_cc_le_read_white_list_size(struct hci_dev *hdev,
1190 struct sk_buff *skb)
1191 {
1192 struct hci_rp_le_read_white_list_size *rp = (void *) skb->data;
1193
1194 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1195
1196 if (rp->status)
1197 return;
1198
1199 hdev->le_white_list_size = rp->size;
1200 }
1201
1202 static void hci_cc_le_clear_white_list(struct hci_dev *hdev,
1203 struct sk_buff *skb)
1204 {
1205 __u8 status = *((__u8 *) skb->data);
1206
1207 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1208
1209 if (status)
1210 return;
1211
1212 hci_bdaddr_list_clear(&hdev->le_white_list);
1213 }
1214
1215 static void hci_cc_le_add_to_white_list(struct hci_dev *hdev,
1216 struct sk_buff *skb)
1217 {
1218 struct hci_cp_le_add_to_white_list *sent;
1219 __u8 status = *((__u8 *) skb->data);
1220
1221 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1222
1223 if (status)
1224 return;
1225
1226 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_WHITE_LIST);
1227 if (!sent)
1228 return;
1229
1230 hci_bdaddr_list_add(&hdev->le_white_list, &sent->bdaddr,
1231 sent->bdaddr_type);
1232 }
1233
1234 static void hci_cc_le_del_from_white_list(struct hci_dev *hdev,
1235 struct sk_buff *skb)
1236 {
1237 struct hci_cp_le_del_from_white_list *sent;
1238 __u8 status = *((__u8 *) skb->data);
1239
1240 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1241
1242 if (status)
1243 return;
1244
1245 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_WHITE_LIST);
1246 if (!sent)
1247 return;
1248
1249 hci_bdaddr_list_del(&hdev->le_white_list, &sent->bdaddr,
1250 sent->bdaddr_type);
1251 }
1252
1253 static void hci_cc_le_read_supported_states(struct hci_dev *hdev,
1254 struct sk_buff *skb)
1255 {
1256 struct hci_rp_le_read_supported_states *rp = (void *) skb->data;
1257
1258 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1259
1260 if (rp->status)
1261 return;
1262
1263 memcpy(hdev->le_states, rp->le_states, 8);
1264 }
1265
1266 static void hci_cc_write_le_host_supported(struct hci_dev *hdev,
1267 struct sk_buff *skb)
1268 {
1269 struct hci_cp_write_le_host_supported *sent;
1270 __u8 status = *((__u8 *) skb->data);
1271
1272 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1273
1274 if (status)
1275 return;
1276
1277 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED);
1278 if (!sent)
1279 return;
1280
1281 if (sent->le) {
1282 hdev->features[1][0] |= LMP_HOST_LE;
1283 set_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1284 } else {
1285 hdev->features[1][0] &= ~LMP_HOST_LE;
1286 clear_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1287 clear_bit(HCI_ADVERTISING, &hdev->dev_flags);
1288 }
1289
1290 if (sent->simul)
1291 hdev->features[1][0] |= LMP_HOST_LE_BREDR;
1292 else
1293 hdev->features[1][0] &= ~LMP_HOST_LE_BREDR;
1294 }
1295
1296 static void hci_cc_set_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1297 {
1298 struct hci_cp_le_set_adv_param *cp;
1299 u8 status = *((u8 *) skb->data);
1300
1301 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1302
1303 if (status)
1304 return;
1305
1306 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_PARAM);
1307 if (!cp)
1308 return;
1309
1310 hci_dev_lock(hdev);
1311 hdev->adv_addr_type = cp->own_address_type;
1312 hci_dev_unlock(hdev);
1313 }
1314
1315 static void hci_cc_write_remote_amp_assoc(struct hci_dev *hdev,
1316 struct sk_buff *skb)
1317 {
1318 struct hci_rp_write_remote_amp_assoc *rp = (void *) skb->data;
1319
1320 BT_DBG("%s status 0x%2.2x phy_handle 0x%2.2x",
1321 hdev->name, rp->status, rp->phy_handle);
1322
1323 if (rp->status)
1324 return;
1325
1326 amp_write_rem_assoc_continue(hdev, rp->phy_handle);
1327 }
1328
1329 static void hci_cc_read_rssi(struct hci_dev *hdev, struct sk_buff *skb)
1330 {
1331 struct hci_rp_read_rssi *rp = (void *) skb->data;
1332 struct hci_conn *conn;
1333
1334 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1335
1336 if (rp->status)
1337 return;
1338
1339 hci_dev_lock(hdev);
1340
1341 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1342 if (conn)
1343 conn->rssi = rp->rssi;
1344
1345 hci_dev_unlock(hdev);
1346 }
1347
1348 static void hci_cc_read_tx_power(struct hci_dev *hdev, struct sk_buff *skb)
1349 {
1350 struct hci_cp_read_tx_power *sent;
1351 struct hci_rp_read_tx_power *rp = (void *) skb->data;
1352 struct hci_conn *conn;
1353
1354 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1355
1356 if (rp->status)
1357 return;
1358
1359 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_TX_POWER);
1360 if (!sent)
1361 return;
1362
1363 hci_dev_lock(hdev);
1364
1365 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1366 if (!conn)
1367 goto unlock;
1368
1369 switch (sent->type) {
1370 case 0x00:
1371 conn->tx_power = rp->tx_power;
1372 break;
1373 case 0x01:
1374 conn->max_tx_power = rp->tx_power;
1375 break;
1376 }
1377
1378 unlock:
1379 hci_dev_unlock(hdev);
1380 }
1381
1382 static void hci_cs_inquiry(struct hci_dev *hdev, __u8 status)
1383 {
1384 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1385
1386 if (status) {
1387 hci_conn_check_pending(hdev);
1388 return;
1389 }
1390
1391 set_bit(HCI_INQUIRY, &hdev->flags);
1392 }
1393
1394 static void hci_cs_create_conn(struct hci_dev *hdev, __u8 status)
1395 {
1396 struct hci_cp_create_conn *cp;
1397 struct hci_conn *conn;
1398
1399 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1400
1401 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_CONN);
1402 if (!cp)
1403 return;
1404
1405 hci_dev_lock(hdev);
1406
1407 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1408
1409 BT_DBG("%s bdaddr %pMR hcon %p", hdev->name, &cp->bdaddr, conn);
1410
1411 if (status) {
1412 if (conn && conn->state == BT_CONNECT) {
1413 if (status != 0x0c || conn->attempt > 2) {
1414 conn->state = BT_CLOSED;
1415 hci_proto_connect_cfm(conn, status);
1416 hci_conn_del(conn);
1417 } else
1418 conn->state = BT_CONNECT2;
1419 }
1420 } else {
1421 if (!conn) {
1422 conn = hci_conn_add(hdev, ACL_LINK, &cp->bdaddr,
1423 HCI_ROLE_MASTER);
1424 if (!conn)
1425 BT_ERR("No memory for new connection");
1426 }
1427 }
1428
1429 hci_dev_unlock(hdev);
1430 }
1431
1432 static void hci_cs_add_sco(struct hci_dev *hdev, __u8 status)
1433 {
1434 struct hci_cp_add_sco *cp;
1435 struct hci_conn *acl, *sco;
1436 __u16 handle;
1437
1438 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1439
1440 if (!status)
1441 return;
1442
1443 cp = hci_sent_cmd_data(hdev, HCI_OP_ADD_SCO);
1444 if (!cp)
1445 return;
1446
1447 handle = __le16_to_cpu(cp->handle);
1448
1449 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1450
1451 hci_dev_lock(hdev);
1452
1453 acl = hci_conn_hash_lookup_handle(hdev, handle);
1454 if (acl) {
1455 sco = acl->link;
1456 if (sco) {
1457 sco->state = BT_CLOSED;
1458
1459 hci_proto_connect_cfm(sco, status);
1460 hci_conn_del(sco);
1461 }
1462 }
1463
1464 hci_dev_unlock(hdev);
1465 }
1466
1467 static void hci_cs_auth_requested(struct hci_dev *hdev, __u8 status)
1468 {
1469 struct hci_cp_auth_requested *cp;
1470 struct hci_conn *conn;
1471
1472 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1473
1474 if (!status)
1475 return;
1476
1477 cp = hci_sent_cmd_data(hdev, HCI_OP_AUTH_REQUESTED);
1478 if (!cp)
1479 return;
1480
1481 hci_dev_lock(hdev);
1482
1483 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1484 if (conn) {
1485 if (conn->state == BT_CONFIG) {
1486 hci_proto_connect_cfm(conn, status);
1487 hci_conn_drop(conn);
1488 }
1489 }
1490
1491 hci_dev_unlock(hdev);
1492 }
1493
1494 static void hci_cs_set_conn_encrypt(struct hci_dev *hdev, __u8 status)
1495 {
1496 struct hci_cp_set_conn_encrypt *cp;
1497 struct hci_conn *conn;
1498
1499 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1500
1501 if (!status)
1502 return;
1503
1504 cp = hci_sent_cmd_data(hdev, HCI_OP_SET_CONN_ENCRYPT);
1505 if (!cp)
1506 return;
1507
1508 hci_dev_lock(hdev);
1509
1510 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1511 if (conn) {
1512 if (conn->state == BT_CONFIG) {
1513 hci_proto_connect_cfm(conn, status);
1514 hci_conn_drop(conn);
1515 }
1516 }
1517
1518 hci_dev_unlock(hdev);
1519 }
1520
1521 static int hci_outgoing_auth_needed(struct hci_dev *hdev,
1522 struct hci_conn *conn)
1523 {
1524 if (conn->state != BT_CONFIG || !conn->out)
1525 return 0;
1526
1527 if (conn->pending_sec_level == BT_SECURITY_SDP)
1528 return 0;
1529
1530 /* Only request authentication for SSP connections or non-SSP
1531 * devices with sec_level MEDIUM or HIGH or if MITM protection
1532 * is requested.
1533 */
1534 if (!hci_conn_ssp_enabled(conn) && !(conn->auth_type & 0x01) &&
1535 conn->pending_sec_level != BT_SECURITY_FIPS &&
1536 conn->pending_sec_level != BT_SECURITY_HIGH &&
1537 conn->pending_sec_level != BT_SECURITY_MEDIUM)
1538 return 0;
1539
1540 return 1;
1541 }
1542
1543 static int hci_resolve_name(struct hci_dev *hdev,
1544 struct inquiry_entry *e)
1545 {
1546 struct hci_cp_remote_name_req cp;
1547
1548 memset(&cp, 0, sizeof(cp));
1549
1550 bacpy(&cp.bdaddr, &e->data.bdaddr);
1551 cp.pscan_rep_mode = e->data.pscan_rep_mode;
1552 cp.pscan_mode = e->data.pscan_mode;
1553 cp.clock_offset = e->data.clock_offset;
1554
1555 return hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
1556 }
1557
1558 static bool hci_resolve_next_name(struct hci_dev *hdev)
1559 {
1560 struct discovery_state *discov = &hdev->discovery;
1561 struct inquiry_entry *e;
1562
1563 if (list_empty(&discov->resolve))
1564 return false;
1565
1566 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1567 if (!e)
1568 return false;
1569
1570 if (hci_resolve_name(hdev, e) == 0) {
1571 e->name_state = NAME_PENDING;
1572 return true;
1573 }
1574
1575 return false;
1576 }
1577
1578 static void hci_check_pending_name(struct hci_dev *hdev, struct hci_conn *conn,
1579 bdaddr_t *bdaddr, u8 *name, u8 name_len)
1580 {
1581 struct discovery_state *discov = &hdev->discovery;
1582 struct inquiry_entry *e;
1583
1584 if (conn && conn->state == BT_CONFIG &&
1585 !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
1586 mgmt_device_connected(hdev, conn, 0, name, name_len);
1587
1588 if (discov->state == DISCOVERY_STOPPED)
1589 return;
1590
1591 if (discov->state == DISCOVERY_STOPPING)
1592 goto discov_complete;
1593
1594 if (discov->state != DISCOVERY_RESOLVING)
1595 return;
1596
1597 e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
1598 /* If the device was not found in a list of found devices names of which
1599 * are pending. there is no need to continue resolving a next name as it
1600 * will be done upon receiving another Remote Name Request Complete
1601 * Event */
1602 if (!e)
1603 return;
1604
1605 list_del(&e->list);
1606 if (name) {
1607 e->name_state = NAME_KNOWN;
1608 mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00,
1609 e->data.rssi, name, name_len);
1610 } else {
1611 e->name_state = NAME_NOT_KNOWN;
1612 }
1613
1614 if (hci_resolve_next_name(hdev))
1615 return;
1616
1617 discov_complete:
1618 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1619 }
1620
1621 static void hci_cs_remote_name_req(struct hci_dev *hdev, __u8 status)
1622 {
1623 struct hci_cp_remote_name_req *cp;
1624 struct hci_conn *conn;
1625
1626 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1627
1628 /* If successful wait for the name req complete event before
1629 * checking for the need to do authentication */
1630 if (!status)
1631 return;
1632
1633 cp = hci_sent_cmd_data(hdev, HCI_OP_REMOTE_NAME_REQ);
1634 if (!cp)
1635 return;
1636
1637 hci_dev_lock(hdev);
1638
1639 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1640
1641 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1642 hci_check_pending_name(hdev, conn, &cp->bdaddr, NULL, 0);
1643
1644 if (!conn)
1645 goto unlock;
1646
1647 if (!hci_outgoing_auth_needed(hdev, conn))
1648 goto unlock;
1649
1650 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
1651 struct hci_cp_auth_requested auth_cp;
1652
1653 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
1654
1655 auth_cp.handle = __cpu_to_le16(conn->handle);
1656 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED,
1657 sizeof(auth_cp), &auth_cp);
1658 }
1659
1660 unlock:
1661 hci_dev_unlock(hdev);
1662 }
1663
1664 static void hci_cs_read_remote_features(struct hci_dev *hdev, __u8 status)
1665 {
1666 struct hci_cp_read_remote_features *cp;
1667 struct hci_conn *conn;
1668
1669 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1670
1671 if (!status)
1672 return;
1673
1674 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_FEATURES);
1675 if (!cp)
1676 return;
1677
1678 hci_dev_lock(hdev);
1679
1680 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1681 if (conn) {
1682 if (conn->state == BT_CONFIG) {
1683 hci_proto_connect_cfm(conn, status);
1684 hci_conn_drop(conn);
1685 }
1686 }
1687
1688 hci_dev_unlock(hdev);
1689 }
1690
1691 static void hci_cs_read_remote_ext_features(struct hci_dev *hdev, __u8 status)
1692 {
1693 struct hci_cp_read_remote_ext_features *cp;
1694 struct hci_conn *conn;
1695
1696 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1697
1698 if (!status)
1699 return;
1700
1701 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES);
1702 if (!cp)
1703 return;
1704
1705 hci_dev_lock(hdev);
1706
1707 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1708 if (conn) {
1709 if (conn->state == BT_CONFIG) {
1710 hci_proto_connect_cfm(conn, status);
1711 hci_conn_drop(conn);
1712 }
1713 }
1714
1715 hci_dev_unlock(hdev);
1716 }
1717
1718 static void hci_cs_setup_sync_conn(struct hci_dev *hdev, __u8 status)
1719 {
1720 struct hci_cp_setup_sync_conn *cp;
1721 struct hci_conn *acl, *sco;
1722 __u16 handle;
1723
1724 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1725
1726 if (!status)
1727 return;
1728
1729 cp = hci_sent_cmd_data(hdev, HCI_OP_SETUP_SYNC_CONN);
1730 if (!cp)
1731 return;
1732
1733 handle = __le16_to_cpu(cp->handle);
1734
1735 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1736
1737 hci_dev_lock(hdev);
1738
1739 acl = hci_conn_hash_lookup_handle(hdev, handle);
1740 if (acl) {
1741 sco = acl->link;
1742 if (sco) {
1743 sco->state = BT_CLOSED;
1744
1745 hci_proto_connect_cfm(sco, status);
1746 hci_conn_del(sco);
1747 }
1748 }
1749
1750 hci_dev_unlock(hdev);
1751 }
1752
1753 static void hci_cs_sniff_mode(struct hci_dev *hdev, __u8 status)
1754 {
1755 struct hci_cp_sniff_mode *cp;
1756 struct hci_conn *conn;
1757
1758 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1759
1760 if (!status)
1761 return;
1762
1763 cp = hci_sent_cmd_data(hdev, HCI_OP_SNIFF_MODE);
1764 if (!cp)
1765 return;
1766
1767 hci_dev_lock(hdev);
1768
1769 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1770 if (conn) {
1771 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1772
1773 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1774 hci_sco_setup(conn, status);
1775 }
1776
1777 hci_dev_unlock(hdev);
1778 }
1779
1780 static void hci_cs_exit_sniff_mode(struct hci_dev *hdev, __u8 status)
1781 {
1782 struct hci_cp_exit_sniff_mode *cp;
1783 struct hci_conn *conn;
1784
1785 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1786
1787 if (!status)
1788 return;
1789
1790 cp = hci_sent_cmd_data(hdev, HCI_OP_EXIT_SNIFF_MODE);
1791 if (!cp)
1792 return;
1793
1794 hci_dev_lock(hdev);
1795
1796 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1797 if (conn) {
1798 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1799
1800 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1801 hci_sco_setup(conn, status);
1802 }
1803
1804 hci_dev_unlock(hdev);
1805 }
1806
1807 static void hci_cs_disconnect(struct hci_dev *hdev, u8 status)
1808 {
1809 struct hci_cp_disconnect *cp;
1810 struct hci_conn *conn;
1811
1812 if (!status)
1813 return;
1814
1815 cp = hci_sent_cmd_data(hdev, HCI_OP_DISCONNECT);
1816 if (!cp)
1817 return;
1818
1819 hci_dev_lock(hdev);
1820
1821 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1822 if (conn)
1823 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
1824 conn->dst_type, status);
1825
1826 hci_dev_unlock(hdev);
1827 }
1828
1829 static void hci_cs_create_phylink(struct hci_dev *hdev, u8 status)
1830 {
1831 struct hci_cp_create_phy_link *cp;
1832
1833 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1834
1835 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_PHY_LINK);
1836 if (!cp)
1837 return;
1838
1839 hci_dev_lock(hdev);
1840
1841 if (status) {
1842 struct hci_conn *hcon;
1843
1844 hcon = hci_conn_hash_lookup_handle(hdev, cp->phy_handle);
1845 if (hcon)
1846 hci_conn_del(hcon);
1847 } else {
1848 amp_write_remote_assoc(hdev, cp->phy_handle);
1849 }
1850
1851 hci_dev_unlock(hdev);
1852 }
1853
1854 static void hci_cs_accept_phylink(struct hci_dev *hdev, u8 status)
1855 {
1856 struct hci_cp_accept_phy_link *cp;
1857
1858 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1859
1860 if (status)
1861 return;
1862
1863 cp = hci_sent_cmd_data(hdev, HCI_OP_ACCEPT_PHY_LINK);
1864 if (!cp)
1865 return;
1866
1867 amp_write_remote_assoc(hdev, cp->phy_handle);
1868 }
1869
1870 static void hci_cs_le_create_conn(struct hci_dev *hdev, u8 status)
1871 {
1872 struct hci_cp_le_create_conn *cp;
1873 struct hci_conn *conn;
1874
1875 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1876
1877 /* All connection failure handling is taken care of by the
1878 * hci_le_conn_failed function which is triggered by the HCI
1879 * request completion callbacks used for connecting.
1880 */
1881 if (status)
1882 return;
1883
1884 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_CREATE_CONN);
1885 if (!cp)
1886 return;
1887
1888 hci_dev_lock(hdev);
1889
1890 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->peer_addr);
1891 if (!conn)
1892 goto unlock;
1893
1894 /* Store the initiator and responder address information which
1895 * is needed for SMP. These values will not change during the
1896 * lifetime of the connection.
1897 */
1898 conn->init_addr_type = cp->own_address_type;
1899 if (cp->own_address_type == ADDR_LE_DEV_RANDOM)
1900 bacpy(&conn->init_addr, &hdev->random_addr);
1901 else
1902 bacpy(&conn->init_addr, &hdev->bdaddr);
1903
1904 conn->resp_addr_type = cp->peer_addr_type;
1905 bacpy(&conn->resp_addr, &cp->peer_addr);
1906
1907 /* We don't want the connection attempt to stick around
1908 * indefinitely since LE doesn't have a page timeout concept
1909 * like BR/EDR. Set a timer for any connection that doesn't use
1910 * the white list for connecting.
1911 */
1912 if (cp->filter_policy == HCI_LE_USE_PEER_ADDR)
1913 queue_delayed_work(conn->hdev->workqueue,
1914 &conn->le_conn_timeout,
1915 conn->conn_timeout);
1916
1917 unlock:
1918 hci_dev_unlock(hdev);
1919 }
1920
1921 static void hci_cs_le_start_enc(struct hci_dev *hdev, u8 status)
1922 {
1923 struct hci_cp_le_start_enc *cp;
1924 struct hci_conn *conn;
1925
1926 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1927
1928 if (!status)
1929 return;
1930
1931 hci_dev_lock(hdev);
1932
1933 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_START_ENC);
1934 if (!cp)
1935 goto unlock;
1936
1937 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1938 if (!conn)
1939 goto unlock;
1940
1941 if (conn->state != BT_CONNECTED)
1942 goto unlock;
1943
1944 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
1945 hci_conn_drop(conn);
1946
1947 unlock:
1948 hci_dev_unlock(hdev);
1949 }
1950
1951 static void hci_cs_switch_role(struct hci_dev *hdev, u8 status)
1952 {
1953 struct hci_cp_switch_role *cp;
1954 struct hci_conn *conn;
1955
1956 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1957
1958 if (!status)
1959 return;
1960
1961 cp = hci_sent_cmd_data(hdev, HCI_OP_SWITCH_ROLE);
1962 if (!cp)
1963 return;
1964
1965 hci_dev_lock(hdev);
1966
1967 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1968 if (conn)
1969 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
1970
1971 hci_dev_unlock(hdev);
1972 }
1973
1974 static void hci_inquiry_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
1975 {
1976 __u8 status = *((__u8 *) skb->data);
1977 struct discovery_state *discov = &hdev->discovery;
1978 struct inquiry_entry *e;
1979
1980 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1981
1982 hci_conn_check_pending(hdev);
1983
1984 if (!test_and_clear_bit(HCI_INQUIRY, &hdev->flags))
1985 return;
1986
1987 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
1988 wake_up_bit(&hdev->flags, HCI_INQUIRY);
1989
1990 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1991 return;
1992
1993 hci_dev_lock(hdev);
1994
1995 if (discov->state != DISCOVERY_FINDING)
1996 goto unlock;
1997
1998 if (list_empty(&discov->resolve)) {
1999 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2000 goto unlock;
2001 }
2002
2003 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
2004 if (e && hci_resolve_name(hdev, e) == 0) {
2005 e->name_state = NAME_PENDING;
2006 hci_discovery_set_state(hdev, DISCOVERY_RESOLVING);
2007 } else {
2008 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2009 }
2010
2011 unlock:
2012 hci_dev_unlock(hdev);
2013 }
2014
2015 static void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff *skb)
2016 {
2017 struct inquiry_data data;
2018 struct inquiry_info *info = (void *) (skb->data + 1);
2019 int num_rsp = *((__u8 *) skb->data);
2020
2021 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
2022
2023 if (!num_rsp)
2024 return;
2025
2026 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
2027 return;
2028
2029 hci_dev_lock(hdev);
2030
2031 for (; num_rsp; num_rsp--, info++) {
2032 u32 flags;
2033
2034 bacpy(&data.bdaddr, &info->bdaddr);
2035 data.pscan_rep_mode = info->pscan_rep_mode;
2036 data.pscan_period_mode = info->pscan_period_mode;
2037 data.pscan_mode = info->pscan_mode;
2038 memcpy(data.dev_class, info->dev_class, 3);
2039 data.clock_offset = info->clock_offset;
2040 data.rssi = 0x00;
2041 data.ssp_mode = 0x00;
2042
2043 flags = hci_inquiry_cache_update(hdev, &data, false);
2044
2045 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
2046 info->dev_class, 0, flags, NULL, 0, NULL, 0);
2047 }
2048
2049 hci_dev_unlock(hdev);
2050 }
2051
2052 static void hci_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2053 {
2054 struct hci_ev_conn_complete *ev = (void *) skb->data;
2055 struct hci_conn *conn;
2056
2057 BT_DBG("%s", hdev->name);
2058
2059 hci_dev_lock(hdev);
2060
2061 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
2062 if (!conn) {
2063 if (ev->link_type != SCO_LINK)
2064 goto unlock;
2065
2066 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
2067 if (!conn)
2068 goto unlock;
2069
2070 conn->type = SCO_LINK;
2071 }
2072
2073 if (!ev->status) {
2074 conn->handle = __le16_to_cpu(ev->handle);
2075
2076 if (conn->type == ACL_LINK) {
2077 conn->state = BT_CONFIG;
2078 hci_conn_hold(conn);
2079
2080 if (!conn->out && !hci_conn_ssp_enabled(conn) &&
2081 !hci_find_link_key(hdev, &ev->bdaddr))
2082 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
2083 else
2084 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2085 } else
2086 conn->state = BT_CONNECTED;
2087
2088 hci_conn_add_sysfs(conn);
2089
2090 if (test_bit(HCI_AUTH, &hdev->flags))
2091 set_bit(HCI_CONN_AUTH, &conn->flags);
2092
2093 if (test_bit(HCI_ENCRYPT, &hdev->flags))
2094 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2095
2096 /* Get remote features */
2097 if (conn->type == ACL_LINK) {
2098 struct hci_cp_read_remote_features cp;
2099 cp.handle = ev->handle;
2100 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_FEATURES,
2101 sizeof(cp), &cp);
2102
2103 hci_update_page_scan(hdev, NULL);
2104 }
2105
2106 /* Set packet type for incoming connection */
2107 if (!conn->out && hdev->hci_ver < BLUETOOTH_VER_2_0) {
2108 struct hci_cp_change_conn_ptype cp;
2109 cp.handle = ev->handle;
2110 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2111 hci_send_cmd(hdev, HCI_OP_CHANGE_CONN_PTYPE, sizeof(cp),
2112 &cp);
2113 }
2114 } else {
2115 conn->state = BT_CLOSED;
2116 if (conn->type == ACL_LINK)
2117 mgmt_connect_failed(hdev, &conn->dst, conn->type,
2118 conn->dst_type, ev->status);
2119 }
2120
2121 if (conn->type == ACL_LINK)
2122 hci_sco_setup(conn, ev->status);
2123
2124 if (ev->status) {
2125 hci_proto_connect_cfm(conn, ev->status);
2126 hci_conn_del(conn);
2127 } else if (ev->link_type != ACL_LINK)
2128 hci_proto_connect_cfm(conn, ev->status);
2129
2130 unlock:
2131 hci_dev_unlock(hdev);
2132
2133 hci_conn_check_pending(hdev);
2134 }
2135
2136 static void hci_reject_conn(struct hci_dev *hdev, bdaddr_t *bdaddr)
2137 {
2138 struct hci_cp_reject_conn_req cp;
2139
2140 bacpy(&cp.bdaddr, bdaddr);
2141 cp.reason = HCI_ERROR_REJ_BAD_ADDR;
2142 hci_send_cmd(hdev, HCI_OP_REJECT_CONN_REQ, sizeof(cp), &cp);
2143 }
2144
2145 static void hci_conn_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
2146 {
2147 struct hci_ev_conn_request *ev = (void *) skb->data;
2148 int mask = hdev->link_mode;
2149 struct inquiry_entry *ie;
2150 struct hci_conn *conn;
2151 __u8 flags = 0;
2152
2153 BT_DBG("%s bdaddr %pMR type 0x%x", hdev->name, &ev->bdaddr,
2154 ev->link_type);
2155
2156 mask |= hci_proto_connect_ind(hdev, &ev->bdaddr, ev->link_type,
2157 &flags);
2158
2159 if (!(mask & HCI_LM_ACCEPT)) {
2160 hci_reject_conn(hdev, &ev->bdaddr);
2161 return;
2162 }
2163
2164 if (hci_bdaddr_list_lookup(&hdev->blacklist, &ev->bdaddr,
2165 BDADDR_BREDR)) {
2166 hci_reject_conn(hdev, &ev->bdaddr);
2167 return;
2168 }
2169
2170 if (!test_bit(HCI_CONNECTABLE, &hdev->dev_flags) &&
2171 !hci_bdaddr_list_lookup(&hdev->whitelist, &ev->bdaddr,
2172 BDADDR_BREDR)) {
2173 hci_reject_conn(hdev, &ev->bdaddr);
2174 return;
2175 }
2176
2177 /* Connection accepted */
2178
2179 hci_dev_lock(hdev);
2180
2181 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
2182 if (ie)
2183 memcpy(ie->data.dev_class, ev->dev_class, 3);
2184
2185 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type,
2186 &ev->bdaddr);
2187 if (!conn) {
2188 conn = hci_conn_add(hdev, ev->link_type, &ev->bdaddr,
2189 HCI_ROLE_SLAVE);
2190 if (!conn) {
2191 BT_ERR("No memory for new connection");
2192 hci_dev_unlock(hdev);
2193 return;
2194 }
2195 }
2196
2197 memcpy(conn->dev_class, ev->dev_class, 3);
2198
2199 hci_dev_unlock(hdev);
2200
2201 if (ev->link_type == ACL_LINK ||
2202 (!(flags & HCI_PROTO_DEFER) && !lmp_esco_capable(hdev))) {
2203 struct hci_cp_accept_conn_req cp;
2204 conn->state = BT_CONNECT;
2205
2206 bacpy(&cp.bdaddr, &ev->bdaddr);
2207
2208 if (lmp_rswitch_capable(hdev) && (mask & HCI_LM_MASTER))
2209 cp.role = 0x00; /* Become master */
2210 else
2211 cp.role = 0x01; /* Remain slave */
2212
2213 hci_send_cmd(hdev, HCI_OP_ACCEPT_CONN_REQ, sizeof(cp), &cp);
2214 } else if (!(flags & HCI_PROTO_DEFER)) {
2215 struct hci_cp_accept_sync_conn_req cp;
2216 conn->state = BT_CONNECT;
2217
2218 bacpy(&cp.bdaddr, &ev->bdaddr);
2219 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2220
2221 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
2222 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
2223 cp.max_latency = cpu_to_le16(0xffff);
2224 cp.content_format = cpu_to_le16(hdev->voice_setting);
2225 cp.retrans_effort = 0xff;
2226
2227 hci_send_cmd(hdev, HCI_OP_ACCEPT_SYNC_CONN_REQ, sizeof(cp),
2228 &cp);
2229 } else {
2230 conn->state = BT_CONNECT2;
2231 hci_proto_connect_cfm(conn, 0);
2232 }
2233 }
2234
2235 static u8 hci_to_mgmt_reason(u8 err)
2236 {
2237 switch (err) {
2238 case HCI_ERROR_CONNECTION_TIMEOUT:
2239 return MGMT_DEV_DISCONN_TIMEOUT;
2240 case HCI_ERROR_REMOTE_USER_TERM:
2241 case HCI_ERROR_REMOTE_LOW_RESOURCES:
2242 case HCI_ERROR_REMOTE_POWER_OFF:
2243 return MGMT_DEV_DISCONN_REMOTE;
2244 case HCI_ERROR_LOCAL_HOST_TERM:
2245 return MGMT_DEV_DISCONN_LOCAL_HOST;
2246 default:
2247 return MGMT_DEV_DISCONN_UNKNOWN;
2248 }
2249 }
2250
2251 static void hci_disconn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2252 {
2253 struct hci_ev_disconn_complete *ev = (void *) skb->data;
2254 u8 reason = hci_to_mgmt_reason(ev->reason);
2255 struct hci_conn_params *params;
2256 struct hci_conn *conn;
2257 bool mgmt_connected;
2258 u8 type;
2259
2260 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2261
2262 hci_dev_lock(hdev);
2263
2264 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2265 if (!conn)
2266 goto unlock;
2267
2268 if (ev->status) {
2269 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2270 conn->dst_type, ev->status);
2271 goto unlock;
2272 }
2273
2274 conn->state = BT_CLOSED;
2275
2276 mgmt_connected = test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags);
2277 mgmt_device_disconnected(hdev, &conn->dst, conn->type, conn->dst_type,
2278 reason, mgmt_connected);
2279
2280 if (conn->type == ACL_LINK) {
2281 if (test_bit(HCI_CONN_FLUSH_KEY, &conn->flags))
2282 hci_remove_link_key(hdev, &conn->dst);
2283
2284 hci_update_page_scan(hdev, NULL);
2285 }
2286
2287 params = hci_conn_params_lookup(hdev, &conn->dst, conn->dst_type);
2288 if (params) {
2289 switch (params->auto_connect) {
2290 case HCI_AUTO_CONN_LINK_LOSS:
2291 if (ev->reason != HCI_ERROR_CONNECTION_TIMEOUT)
2292 break;
2293 /* Fall through */
2294
2295 case HCI_AUTO_CONN_DIRECT:
2296 case HCI_AUTO_CONN_ALWAYS:
2297 list_del_init(&params->action);
2298 list_add(&params->action, &hdev->pend_le_conns);
2299 hci_update_background_scan(hdev);
2300 break;
2301
2302 default:
2303 break;
2304 }
2305 }
2306
2307 type = conn->type;
2308
2309 hci_proto_disconn_cfm(conn, ev->reason);
2310 hci_conn_del(conn);
2311
2312 /* Re-enable advertising if necessary, since it might
2313 * have been disabled by the connection. From the
2314 * HCI_LE_Set_Advertise_Enable command description in
2315 * the core specification (v4.0):
2316 * "The Controller shall continue advertising until the Host
2317 * issues an LE_Set_Advertise_Enable command with
2318 * Advertising_Enable set to 0x00 (Advertising is disabled)
2319 * or until a connection is created or until the Advertising
2320 * is timed out due to Directed Advertising."
2321 */
2322 if (type == LE_LINK)
2323 mgmt_reenable_advertising(hdev);
2324
2325 unlock:
2326 hci_dev_unlock(hdev);
2327 }
2328
2329 static void hci_auth_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2330 {
2331 struct hci_ev_auth_complete *ev = (void *) skb->data;
2332 struct hci_conn *conn;
2333
2334 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2335
2336 hci_dev_lock(hdev);
2337
2338 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2339 if (!conn)
2340 goto unlock;
2341
2342 if (!ev->status) {
2343 if (!hci_conn_ssp_enabled(conn) &&
2344 test_bit(HCI_CONN_REAUTH_PEND, &conn->flags)) {
2345 BT_INFO("re-auth of legacy device is not possible.");
2346 } else {
2347 set_bit(HCI_CONN_AUTH, &conn->flags);
2348 conn->sec_level = conn->pending_sec_level;
2349 }
2350 } else {
2351 mgmt_auth_failed(conn, ev->status);
2352 }
2353
2354 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2355 clear_bit(HCI_CONN_REAUTH_PEND, &conn->flags);
2356
2357 if (conn->state == BT_CONFIG) {
2358 if (!ev->status && hci_conn_ssp_enabled(conn)) {
2359 struct hci_cp_set_conn_encrypt cp;
2360 cp.handle = ev->handle;
2361 cp.encrypt = 0x01;
2362 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2363 &cp);
2364 } else {
2365 conn->state = BT_CONNECTED;
2366 hci_proto_connect_cfm(conn, ev->status);
2367 hci_conn_drop(conn);
2368 }
2369 } else {
2370 hci_auth_cfm(conn, ev->status);
2371
2372 hci_conn_hold(conn);
2373 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2374 hci_conn_drop(conn);
2375 }
2376
2377 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags)) {
2378 if (!ev->status) {
2379 struct hci_cp_set_conn_encrypt cp;
2380 cp.handle = ev->handle;
2381 cp.encrypt = 0x01;
2382 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2383 &cp);
2384 } else {
2385 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2386 hci_encrypt_cfm(conn, ev->status, 0x00);
2387 }
2388 }
2389
2390 unlock:
2391 hci_dev_unlock(hdev);
2392 }
2393
2394 static void hci_remote_name_evt(struct hci_dev *hdev, struct sk_buff *skb)
2395 {
2396 struct hci_ev_remote_name *ev = (void *) skb->data;
2397 struct hci_conn *conn;
2398
2399 BT_DBG("%s", hdev->name);
2400
2401 hci_conn_check_pending(hdev);
2402
2403 hci_dev_lock(hdev);
2404
2405 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2406
2407 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
2408 goto check_auth;
2409
2410 if (ev->status == 0)
2411 hci_check_pending_name(hdev, conn, &ev->bdaddr, ev->name,
2412 strnlen(ev->name, HCI_MAX_NAME_LENGTH));
2413 else
2414 hci_check_pending_name(hdev, conn, &ev->bdaddr, NULL, 0);
2415
2416 check_auth:
2417 if (!conn)
2418 goto unlock;
2419
2420 if (!hci_outgoing_auth_needed(hdev, conn))
2421 goto unlock;
2422
2423 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
2424 struct hci_cp_auth_requested cp;
2425
2426 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
2427
2428 cp.handle = __cpu_to_le16(conn->handle);
2429 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED, sizeof(cp), &cp);
2430 }
2431
2432 unlock:
2433 hci_dev_unlock(hdev);
2434 }
2435
2436 static void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2437 {
2438 struct hci_ev_encrypt_change *ev = (void *) skb->data;
2439 struct hci_conn *conn;
2440
2441 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2442
2443 hci_dev_lock(hdev);
2444
2445 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2446 if (!conn)
2447 goto unlock;
2448
2449 if (!ev->status) {
2450 if (ev->encrypt) {
2451 /* Encryption implies authentication */
2452 set_bit(HCI_CONN_AUTH, &conn->flags);
2453 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2454 conn->sec_level = conn->pending_sec_level;
2455
2456 /* P-256 authentication key implies FIPS */
2457 if (conn->key_type == HCI_LK_AUTH_COMBINATION_P256)
2458 set_bit(HCI_CONN_FIPS, &conn->flags);
2459
2460 if ((conn->type == ACL_LINK && ev->encrypt == 0x02) ||
2461 conn->type == LE_LINK)
2462 set_bit(HCI_CONN_AES_CCM, &conn->flags);
2463 } else {
2464 clear_bit(HCI_CONN_ENCRYPT, &conn->flags);
2465 clear_bit(HCI_CONN_AES_CCM, &conn->flags);
2466 }
2467 }
2468
2469 /* We should disregard the current RPA and generate a new one
2470 * whenever the encryption procedure fails.
2471 */
2472 if (ev->status && conn->type == LE_LINK)
2473 set_bit(HCI_RPA_EXPIRED, &hdev->dev_flags);
2474
2475 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2476
2477 if (ev->status && conn->state == BT_CONNECTED) {
2478 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2479 hci_conn_drop(conn);
2480 goto unlock;
2481 }
2482
2483 if (conn->state == BT_CONFIG) {
2484 if (!ev->status)
2485 conn->state = BT_CONNECTED;
2486
2487 /* In Secure Connections Only mode, do not allow any
2488 * connections that are not encrypted with AES-CCM
2489 * using a P-256 authenticated combination key.
2490 */
2491 if (test_bit(HCI_SC_ONLY, &hdev->dev_flags) &&
2492 (!test_bit(HCI_CONN_AES_CCM, &conn->flags) ||
2493 conn->key_type != HCI_LK_AUTH_COMBINATION_P256)) {
2494 hci_proto_connect_cfm(conn, HCI_ERROR_AUTH_FAILURE);
2495 hci_conn_drop(conn);
2496 goto unlock;
2497 }
2498
2499 hci_proto_connect_cfm(conn, ev->status);
2500 hci_conn_drop(conn);
2501 } else
2502 hci_encrypt_cfm(conn, ev->status, ev->encrypt);
2503
2504 unlock:
2505 hci_dev_unlock(hdev);
2506 }
2507
2508 static void hci_change_link_key_complete_evt(struct hci_dev *hdev,
2509 struct sk_buff *skb)
2510 {
2511 struct hci_ev_change_link_key_complete *ev = (void *) skb->data;
2512 struct hci_conn *conn;
2513
2514 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2515
2516 hci_dev_lock(hdev);
2517
2518 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2519 if (conn) {
2520 if (!ev->status)
2521 set_bit(HCI_CONN_SECURE, &conn->flags);
2522
2523 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2524
2525 hci_key_change_cfm(conn, ev->status);
2526 }
2527
2528 hci_dev_unlock(hdev);
2529 }
2530
2531 static void hci_remote_features_evt(struct hci_dev *hdev,
2532 struct sk_buff *skb)
2533 {
2534 struct hci_ev_remote_features *ev = (void *) skb->data;
2535 struct hci_conn *conn;
2536
2537 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2538
2539 hci_dev_lock(hdev);
2540
2541 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2542 if (!conn)
2543 goto unlock;
2544
2545 if (!ev->status)
2546 memcpy(conn->features[0], ev->features, 8);
2547
2548 if (conn->state != BT_CONFIG)
2549 goto unlock;
2550
2551 if (!ev->status && lmp_ssp_capable(hdev) && lmp_ssp_capable(conn)) {
2552 struct hci_cp_read_remote_ext_features cp;
2553 cp.handle = ev->handle;
2554 cp.page = 0x01;
2555 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES,
2556 sizeof(cp), &cp);
2557 goto unlock;
2558 }
2559
2560 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
2561 struct hci_cp_remote_name_req cp;
2562 memset(&cp, 0, sizeof(cp));
2563 bacpy(&cp.bdaddr, &conn->dst);
2564 cp.pscan_rep_mode = 0x02;
2565 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
2566 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2567 mgmt_device_connected(hdev, conn, 0, NULL, 0);
2568
2569 if (!hci_outgoing_auth_needed(hdev, conn)) {
2570 conn->state = BT_CONNECTED;
2571 hci_proto_connect_cfm(conn, ev->status);
2572 hci_conn_drop(conn);
2573 }
2574
2575 unlock:
2576 hci_dev_unlock(hdev);
2577 }
2578
2579 static void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2580 {
2581 struct hci_ev_cmd_complete *ev = (void *) skb->data;
2582 u8 status = skb->data[sizeof(*ev)];
2583 __u16 opcode;
2584
2585 skb_pull(skb, sizeof(*ev));
2586
2587 opcode = __le16_to_cpu(ev->opcode);
2588
2589 switch (opcode) {
2590 case HCI_OP_INQUIRY_CANCEL:
2591 hci_cc_inquiry_cancel(hdev, skb);
2592 break;
2593
2594 case HCI_OP_PERIODIC_INQ:
2595 hci_cc_periodic_inq(hdev, skb);
2596 break;
2597
2598 case HCI_OP_EXIT_PERIODIC_INQ:
2599 hci_cc_exit_periodic_inq(hdev, skb);
2600 break;
2601
2602 case HCI_OP_REMOTE_NAME_REQ_CANCEL:
2603 hci_cc_remote_name_req_cancel(hdev, skb);
2604 break;
2605
2606 case HCI_OP_ROLE_DISCOVERY:
2607 hci_cc_role_discovery(hdev, skb);
2608 break;
2609
2610 case HCI_OP_READ_LINK_POLICY:
2611 hci_cc_read_link_policy(hdev, skb);
2612 break;
2613
2614 case HCI_OP_WRITE_LINK_POLICY:
2615 hci_cc_write_link_policy(hdev, skb);
2616 break;
2617
2618 case HCI_OP_READ_DEF_LINK_POLICY:
2619 hci_cc_read_def_link_policy(hdev, skb);
2620 break;
2621
2622 case HCI_OP_WRITE_DEF_LINK_POLICY:
2623 hci_cc_write_def_link_policy(hdev, skb);
2624 break;
2625
2626 case HCI_OP_RESET:
2627 hci_cc_reset(hdev, skb);
2628 break;
2629
2630 case HCI_OP_WRITE_LOCAL_NAME:
2631 hci_cc_write_local_name(hdev, skb);
2632 break;
2633
2634 case HCI_OP_READ_LOCAL_NAME:
2635 hci_cc_read_local_name(hdev, skb);
2636 break;
2637
2638 case HCI_OP_WRITE_AUTH_ENABLE:
2639 hci_cc_write_auth_enable(hdev, skb);
2640 break;
2641
2642 case HCI_OP_WRITE_ENCRYPT_MODE:
2643 hci_cc_write_encrypt_mode(hdev, skb);
2644 break;
2645
2646 case HCI_OP_WRITE_SCAN_ENABLE:
2647 hci_cc_write_scan_enable(hdev, skb);
2648 break;
2649
2650 case HCI_OP_READ_CLASS_OF_DEV:
2651 hci_cc_read_class_of_dev(hdev, skb);
2652 break;
2653
2654 case HCI_OP_WRITE_CLASS_OF_DEV:
2655 hci_cc_write_class_of_dev(hdev, skb);
2656 break;
2657
2658 case HCI_OP_READ_VOICE_SETTING:
2659 hci_cc_read_voice_setting(hdev, skb);
2660 break;
2661
2662 case HCI_OP_WRITE_VOICE_SETTING:
2663 hci_cc_write_voice_setting(hdev, skb);
2664 break;
2665
2666 case HCI_OP_READ_NUM_SUPPORTED_IAC:
2667 hci_cc_read_num_supported_iac(hdev, skb);
2668 break;
2669
2670 case HCI_OP_WRITE_SSP_MODE:
2671 hci_cc_write_ssp_mode(hdev, skb);
2672 break;
2673
2674 case HCI_OP_WRITE_SC_SUPPORT:
2675 hci_cc_write_sc_support(hdev, skb);
2676 break;
2677
2678 case HCI_OP_READ_LOCAL_VERSION:
2679 hci_cc_read_local_version(hdev, skb);
2680 break;
2681
2682 case HCI_OP_READ_LOCAL_COMMANDS:
2683 hci_cc_read_local_commands(hdev, skb);
2684 break;
2685
2686 case HCI_OP_READ_LOCAL_FEATURES:
2687 hci_cc_read_local_features(hdev, skb);
2688 break;
2689
2690 case HCI_OP_READ_LOCAL_EXT_FEATURES:
2691 hci_cc_read_local_ext_features(hdev, skb);
2692 break;
2693
2694 case HCI_OP_READ_BUFFER_SIZE:
2695 hci_cc_read_buffer_size(hdev, skb);
2696 break;
2697
2698 case HCI_OP_READ_BD_ADDR:
2699 hci_cc_read_bd_addr(hdev, skb);
2700 break;
2701
2702 case HCI_OP_READ_PAGE_SCAN_ACTIVITY:
2703 hci_cc_read_page_scan_activity(hdev, skb);
2704 break;
2705
2706 case HCI_OP_WRITE_PAGE_SCAN_ACTIVITY:
2707 hci_cc_write_page_scan_activity(hdev, skb);
2708 break;
2709
2710 case HCI_OP_READ_PAGE_SCAN_TYPE:
2711 hci_cc_read_page_scan_type(hdev, skb);
2712 break;
2713
2714 case HCI_OP_WRITE_PAGE_SCAN_TYPE:
2715 hci_cc_write_page_scan_type(hdev, skb);
2716 break;
2717
2718 case HCI_OP_READ_DATA_BLOCK_SIZE:
2719 hci_cc_read_data_block_size(hdev, skb);
2720 break;
2721
2722 case HCI_OP_READ_FLOW_CONTROL_MODE:
2723 hci_cc_read_flow_control_mode(hdev, skb);
2724 break;
2725
2726 case HCI_OP_READ_LOCAL_AMP_INFO:
2727 hci_cc_read_local_amp_info(hdev, skb);
2728 break;
2729
2730 case HCI_OP_READ_CLOCK:
2731 hci_cc_read_clock(hdev, skb);
2732 break;
2733
2734 case HCI_OP_READ_LOCAL_AMP_ASSOC:
2735 hci_cc_read_local_amp_assoc(hdev, skb);
2736 break;
2737
2738 case HCI_OP_READ_INQ_RSP_TX_POWER:
2739 hci_cc_read_inq_rsp_tx_power(hdev, skb);
2740 break;
2741
2742 case HCI_OP_PIN_CODE_REPLY:
2743 hci_cc_pin_code_reply(hdev, skb);
2744 break;
2745
2746 case HCI_OP_PIN_CODE_NEG_REPLY:
2747 hci_cc_pin_code_neg_reply(hdev, skb);
2748 break;
2749
2750 case HCI_OP_READ_LOCAL_OOB_DATA:
2751 hci_cc_read_local_oob_data(hdev, skb);
2752 break;
2753
2754 case HCI_OP_READ_LOCAL_OOB_EXT_DATA:
2755 hci_cc_read_local_oob_ext_data(hdev, skb);
2756 break;
2757
2758 case HCI_OP_LE_READ_BUFFER_SIZE:
2759 hci_cc_le_read_buffer_size(hdev, skb);
2760 break;
2761
2762 case HCI_OP_LE_READ_LOCAL_FEATURES:
2763 hci_cc_le_read_local_features(hdev, skb);
2764 break;
2765
2766 case HCI_OP_LE_READ_ADV_TX_POWER:
2767 hci_cc_le_read_adv_tx_power(hdev, skb);
2768 break;
2769
2770 case HCI_OP_USER_CONFIRM_REPLY:
2771 hci_cc_user_confirm_reply(hdev, skb);
2772 break;
2773
2774 case HCI_OP_USER_CONFIRM_NEG_REPLY:
2775 hci_cc_user_confirm_neg_reply(hdev, skb);
2776 break;
2777
2778 case HCI_OP_USER_PASSKEY_REPLY:
2779 hci_cc_user_passkey_reply(hdev, skb);
2780 break;
2781
2782 case HCI_OP_USER_PASSKEY_NEG_REPLY:
2783 hci_cc_user_passkey_neg_reply(hdev, skb);
2784 break;
2785
2786 case HCI_OP_LE_SET_RANDOM_ADDR:
2787 hci_cc_le_set_random_addr(hdev, skb);
2788 break;
2789
2790 case HCI_OP_LE_SET_ADV_ENABLE:
2791 hci_cc_le_set_adv_enable(hdev, skb);
2792 break;
2793
2794 case HCI_OP_LE_SET_SCAN_PARAM:
2795 hci_cc_le_set_scan_param(hdev, skb);
2796 break;
2797
2798 case HCI_OP_LE_SET_SCAN_ENABLE:
2799 hci_cc_le_set_scan_enable(hdev, skb);
2800 break;
2801
2802 case HCI_OP_LE_READ_WHITE_LIST_SIZE:
2803 hci_cc_le_read_white_list_size(hdev, skb);
2804 break;
2805
2806 case HCI_OP_LE_CLEAR_WHITE_LIST:
2807 hci_cc_le_clear_white_list(hdev, skb);
2808 break;
2809
2810 case HCI_OP_LE_ADD_TO_WHITE_LIST:
2811 hci_cc_le_add_to_white_list(hdev, skb);
2812 break;
2813
2814 case HCI_OP_LE_DEL_FROM_WHITE_LIST:
2815 hci_cc_le_del_from_white_list(hdev, skb);
2816 break;
2817
2818 case HCI_OP_LE_READ_SUPPORTED_STATES:
2819 hci_cc_le_read_supported_states(hdev, skb);
2820 break;
2821
2822 case HCI_OP_WRITE_LE_HOST_SUPPORTED:
2823 hci_cc_write_le_host_supported(hdev, skb);
2824 break;
2825
2826 case HCI_OP_LE_SET_ADV_PARAM:
2827 hci_cc_set_adv_param(hdev, skb);
2828 break;
2829
2830 case HCI_OP_WRITE_REMOTE_AMP_ASSOC:
2831 hci_cc_write_remote_amp_assoc(hdev, skb);
2832 break;
2833
2834 case HCI_OP_READ_RSSI:
2835 hci_cc_read_rssi(hdev, skb);
2836 break;
2837
2838 case HCI_OP_READ_TX_POWER:
2839 hci_cc_read_tx_power(hdev, skb);
2840 break;
2841
2842 default:
2843 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2844 break;
2845 }
2846
2847 if (opcode != HCI_OP_NOP)
2848 cancel_delayed_work(&hdev->cmd_timer);
2849
2850 hci_req_cmd_complete(hdev, opcode, status);
2851
2852 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2853 atomic_set(&hdev->cmd_cnt, 1);
2854 if (!skb_queue_empty(&hdev->cmd_q))
2855 queue_work(hdev->workqueue, &hdev->cmd_work);
2856 }
2857 }
2858
2859 static void hci_cmd_status_evt(struct hci_dev *hdev, struct sk_buff *skb)
2860 {
2861 struct hci_ev_cmd_status *ev = (void *) skb->data;
2862 __u16 opcode;
2863
2864 skb_pull(skb, sizeof(*ev));
2865
2866 opcode = __le16_to_cpu(ev->opcode);
2867
2868 switch (opcode) {
2869 case HCI_OP_INQUIRY:
2870 hci_cs_inquiry(hdev, ev->status);
2871 break;
2872
2873 case HCI_OP_CREATE_CONN:
2874 hci_cs_create_conn(hdev, ev->status);
2875 break;
2876
2877 case HCI_OP_DISCONNECT:
2878 hci_cs_disconnect(hdev, ev->status);
2879 break;
2880
2881 case HCI_OP_ADD_SCO:
2882 hci_cs_add_sco(hdev, ev->status);
2883 break;
2884
2885 case HCI_OP_AUTH_REQUESTED:
2886 hci_cs_auth_requested(hdev, ev->status);
2887 break;
2888
2889 case HCI_OP_SET_CONN_ENCRYPT:
2890 hci_cs_set_conn_encrypt(hdev, ev->status);
2891 break;
2892
2893 case HCI_OP_REMOTE_NAME_REQ:
2894 hci_cs_remote_name_req(hdev, ev->status);
2895 break;
2896
2897 case HCI_OP_READ_REMOTE_FEATURES:
2898 hci_cs_read_remote_features(hdev, ev->status);
2899 break;
2900
2901 case HCI_OP_READ_REMOTE_EXT_FEATURES:
2902 hci_cs_read_remote_ext_features(hdev, ev->status);
2903 break;
2904
2905 case HCI_OP_SETUP_SYNC_CONN:
2906 hci_cs_setup_sync_conn(hdev, ev->status);
2907 break;
2908
2909 case HCI_OP_CREATE_PHY_LINK:
2910 hci_cs_create_phylink(hdev, ev->status);
2911 break;
2912
2913 case HCI_OP_ACCEPT_PHY_LINK:
2914 hci_cs_accept_phylink(hdev, ev->status);
2915 break;
2916
2917 case HCI_OP_SNIFF_MODE:
2918 hci_cs_sniff_mode(hdev, ev->status);
2919 break;
2920
2921 case HCI_OP_EXIT_SNIFF_MODE:
2922 hci_cs_exit_sniff_mode(hdev, ev->status);
2923 break;
2924
2925 case HCI_OP_SWITCH_ROLE:
2926 hci_cs_switch_role(hdev, ev->status);
2927 break;
2928
2929 case HCI_OP_LE_CREATE_CONN:
2930 hci_cs_le_create_conn(hdev, ev->status);
2931 break;
2932
2933 case HCI_OP_LE_START_ENC:
2934 hci_cs_le_start_enc(hdev, ev->status);
2935 break;
2936
2937 default:
2938 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2939 break;
2940 }
2941
2942 if (opcode != HCI_OP_NOP)
2943 cancel_delayed_work(&hdev->cmd_timer);
2944
2945 if (ev->status ||
2946 (hdev->sent_cmd && !bt_cb(hdev->sent_cmd)->req.event))
2947 hci_req_cmd_complete(hdev, opcode, ev->status);
2948
2949 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2950 atomic_set(&hdev->cmd_cnt, 1);
2951 if (!skb_queue_empty(&hdev->cmd_q))
2952 queue_work(hdev->workqueue, &hdev->cmd_work);
2953 }
2954 }
2955
2956 static void hci_hardware_error_evt(struct hci_dev *hdev, struct sk_buff *skb)
2957 {
2958 struct hci_ev_hardware_error *ev = (void *) skb->data;
2959
2960 BT_ERR("%s hardware error 0x%2.2x", hdev->name, ev->code);
2961 }
2962
2963 static void hci_role_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2964 {
2965 struct hci_ev_role_change *ev = (void *) skb->data;
2966 struct hci_conn *conn;
2967
2968 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2969
2970 hci_dev_lock(hdev);
2971
2972 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2973 if (conn) {
2974 if (!ev->status)
2975 conn->role = ev->role;
2976
2977 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
2978
2979 hci_role_switch_cfm(conn, ev->status, ev->role);
2980 }
2981
2982 hci_dev_unlock(hdev);
2983 }
2984
2985 static void hci_num_comp_pkts_evt(struct hci_dev *hdev, struct sk_buff *skb)
2986 {
2987 struct hci_ev_num_comp_pkts *ev = (void *) skb->data;
2988 int i;
2989
2990 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_PACKET_BASED) {
2991 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
2992 return;
2993 }
2994
2995 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
2996 ev->num_hndl * sizeof(struct hci_comp_pkts_info)) {
2997 BT_DBG("%s bad parameters", hdev->name);
2998 return;
2999 }
3000
3001 BT_DBG("%s num_hndl %d", hdev->name, ev->num_hndl);
3002
3003 for (i = 0; i < ev->num_hndl; i++) {
3004 struct hci_comp_pkts_info *info = &ev->handles[i];
3005 struct hci_conn *conn;
3006 __u16 handle, count;
3007
3008 handle = __le16_to_cpu(info->handle);
3009 count = __le16_to_cpu(info->count);
3010
3011 conn = hci_conn_hash_lookup_handle(hdev, handle);
3012 if (!conn)
3013 continue;
3014
3015 conn->sent -= count;
3016
3017 switch (conn->type) {
3018 case ACL_LINK:
3019 hdev->acl_cnt += count;
3020 if (hdev->acl_cnt > hdev->acl_pkts)
3021 hdev->acl_cnt = hdev->acl_pkts;
3022 break;
3023
3024 case LE_LINK:
3025 if (hdev->le_pkts) {
3026 hdev->le_cnt += count;
3027 if (hdev->le_cnt > hdev->le_pkts)
3028 hdev->le_cnt = hdev->le_pkts;
3029 } else {
3030 hdev->acl_cnt += count;
3031 if (hdev->acl_cnt > hdev->acl_pkts)
3032 hdev->acl_cnt = hdev->acl_pkts;
3033 }
3034 break;
3035
3036 case SCO_LINK:
3037 hdev->sco_cnt += count;
3038 if (hdev->sco_cnt > hdev->sco_pkts)
3039 hdev->sco_cnt = hdev->sco_pkts;
3040 break;
3041
3042 default:
3043 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3044 break;
3045 }
3046 }
3047
3048 queue_work(hdev->workqueue, &hdev->tx_work);
3049 }
3050
3051 static struct hci_conn *__hci_conn_lookup_handle(struct hci_dev *hdev,
3052 __u16 handle)
3053 {
3054 struct hci_chan *chan;
3055
3056 switch (hdev->dev_type) {
3057 case HCI_BREDR:
3058 return hci_conn_hash_lookup_handle(hdev, handle);
3059 case HCI_AMP:
3060 chan = hci_chan_lookup_handle(hdev, handle);
3061 if (chan)
3062 return chan->conn;
3063 break;
3064 default:
3065 BT_ERR("%s unknown dev_type %d", hdev->name, hdev->dev_type);
3066 break;
3067 }
3068
3069 return NULL;
3070 }
3071
3072 static void hci_num_comp_blocks_evt(struct hci_dev *hdev, struct sk_buff *skb)
3073 {
3074 struct hci_ev_num_comp_blocks *ev = (void *) skb->data;
3075 int i;
3076
3077 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_BLOCK_BASED) {
3078 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
3079 return;
3080 }
3081
3082 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
3083 ev->num_hndl * sizeof(struct hci_comp_blocks_info)) {
3084 BT_DBG("%s bad parameters", hdev->name);
3085 return;
3086 }
3087
3088 BT_DBG("%s num_blocks %d num_hndl %d", hdev->name, ev->num_blocks,
3089 ev->num_hndl);
3090
3091 for (i = 0; i < ev->num_hndl; i++) {
3092 struct hci_comp_blocks_info *info = &ev->handles[i];
3093 struct hci_conn *conn = NULL;
3094 __u16 handle, block_count;
3095
3096 handle = __le16_to_cpu(info->handle);
3097 block_count = __le16_to_cpu(info->blocks);
3098
3099 conn = __hci_conn_lookup_handle(hdev, handle);
3100 if (!conn)
3101 continue;
3102
3103 conn->sent -= block_count;
3104
3105 switch (conn->type) {
3106 case ACL_LINK:
3107 case AMP_LINK:
3108 hdev->block_cnt += block_count;
3109 if (hdev->block_cnt > hdev->num_blocks)
3110 hdev->block_cnt = hdev->num_blocks;
3111 break;
3112
3113 default:
3114 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3115 break;
3116 }
3117 }
3118
3119 queue_work(hdev->workqueue, &hdev->tx_work);
3120 }
3121
3122 static void hci_mode_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3123 {
3124 struct hci_ev_mode_change *ev = (void *) skb->data;
3125 struct hci_conn *conn;
3126
3127 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3128
3129 hci_dev_lock(hdev);
3130
3131 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3132 if (conn) {
3133 conn->mode = ev->mode;
3134
3135 if (!test_and_clear_bit(HCI_CONN_MODE_CHANGE_PEND,
3136 &conn->flags)) {
3137 if (conn->mode == HCI_CM_ACTIVE)
3138 set_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3139 else
3140 clear_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3141 }
3142
3143 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
3144 hci_sco_setup(conn, ev->status);
3145 }
3146
3147 hci_dev_unlock(hdev);
3148 }
3149
3150 static void hci_pin_code_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3151 {
3152 struct hci_ev_pin_code_req *ev = (void *) skb->data;
3153 struct hci_conn *conn;
3154
3155 BT_DBG("%s", hdev->name);
3156
3157 hci_dev_lock(hdev);
3158
3159 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3160 if (!conn)
3161 goto unlock;
3162
3163 if (conn->state == BT_CONNECTED) {
3164 hci_conn_hold(conn);
3165 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
3166 hci_conn_drop(conn);
3167 }
3168
3169 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags) &&
3170 !test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags)) {
3171 hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3172 sizeof(ev->bdaddr), &ev->bdaddr);
3173 } else if (test_bit(HCI_MGMT, &hdev->dev_flags)) {
3174 u8 secure;
3175
3176 if (conn->pending_sec_level == BT_SECURITY_HIGH)
3177 secure = 1;
3178 else
3179 secure = 0;
3180
3181 mgmt_pin_code_request(hdev, &ev->bdaddr, secure);
3182 }
3183
3184 unlock:
3185 hci_dev_unlock(hdev);
3186 }
3187
3188 static void hci_link_key_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3189 {
3190 struct hci_ev_link_key_req *ev = (void *) skb->data;
3191 struct hci_cp_link_key_reply cp;
3192 struct hci_conn *conn;
3193 struct link_key *key;
3194
3195 BT_DBG("%s", hdev->name);
3196
3197 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3198 return;
3199
3200 hci_dev_lock(hdev);
3201
3202 key = hci_find_link_key(hdev, &ev->bdaddr);
3203 if (!key) {
3204 BT_DBG("%s link key not found for %pMR", hdev->name,
3205 &ev->bdaddr);
3206 goto not_found;
3207 }
3208
3209 BT_DBG("%s found key type %u for %pMR", hdev->name, key->type,
3210 &ev->bdaddr);
3211
3212 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3213 if (conn) {
3214 if ((key->type == HCI_LK_UNAUTH_COMBINATION_P192 ||
3215 key->type == HCI_LK_UNAUTH_COMBINATION_P256) &&
3216 conn->auth_type != 0xff && (conn->auth_type & 0x01)) {
3217 BT_DBG("%s ignoring unauthenticated key", hdev->name);
3218 goto not_found;
3219 }
3220
3221 if (key->type == HCI_LK_COMBINATION && key->pin_len < 16 &&
3222 (conn->pending_sec_level == BT_SECURITY_HIGH ||
3223 conn->pending_sec_level == BT_SECURITY_FIPS)) {
3224 BT_DBG("%s ignoring key unauthenticated for high security",
3225 hdev->name);
3226 goto not_found;
3227 }
3228
3229 conn->key_type = key->type;
3230 conn->pin_length = key->pin_len;
3231 }
3232
3233 bacpy(&cp.bdaddr, &ev->bdaddr);
3234 memcpy(cp.link_key, key->val, HCI_LINK_KEY_SIZE);
3235
3236 hci_send_cmd(hdev, HCI_OP_LINK_KEY_REPLY, sizeof(cp), &cp);
3237
3238 hci_dev_unlock(hdev);
3239
3240 return;
3241
3242 not_found:
3243 hci_send_cmd(hdev, HCI_OP_LINK_KEY_NEG_REPLY, 6, &ev->bdaddr);
3244 hci_dev_unlock(hdev);
3245 }
3246
3247 static void hci_link_key_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3248 {
3249 struct hci_ev_link_key_notify *ev = (void *) skb->data;
3250 struct hci_conn *conn;
3251 struct link_key *key;
3252 bool persistent;
3253 u8 pin_len = 0;
3254
3255 BT_DBG("%s", hdev->name);
3256
3257 hci_dev_lock(hdev);
3258
3259 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3260 if (conn) {
3261 hci_conn_hold(conn);
3262 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3263 pin_len = conn->pin_length;
3264
3265 if (ev->key_type != HCI_LK_CHANGED_COMBINATION)
3266 conn->key_type = ev->key_type;
3267
3268 hci_conn_drop(conn);
3269 }
3270
3271 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3272 goto unlock;
3273
3274 key = hci_add_link_key(hdev, conn, &ev->bdaddr, ev->link_key,
3275 ev->key_type, pin_len, &persistent);
3276 if (!key)
3277 goto unlock;
3278
3279 mgmt_new_link_key(hdev, key, persistent);
3280
3281 /* Keep debug keys around only if the HCI_KEEP_DEBUG_KEYS flag
3282 * is set. If it's not set simply remove the key from the kernel
3283 * list (we've still notified user space about it but with
3284 * store_hint being 0).
3285 */
3286 if (key->type == HCI_LK_DEBUG_COMBINATION &&
3287 !test_bit(HCI_KEEP_DEBUG_KEYS, &hdev->dev_flags)) {
3288 list_del(&key->list);
3289 kfree(key);
3290 } else if (conn) {
3291 if (persistent)
3292 clear_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3293 else
3294 set_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3295 }
3296
3297 unlock:
3298 hci_dev_unlock(hdev);
3299 }
3300
3301 static void hci_clock_offset_evt(struct hci_dev *hdev, struct sk_buff *skb)
3302 {
3303 struct hci_ev_clock_offset *ev = (void *) skb->data;
3304 struct hci_conn *conn;
3305
3306 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3307
3308 hci_dev_lock(hdev);
3309
3310 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3311 if (conn && !ev->status) {
3312 struct inquiry_entry *ie;
3313
3314 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3315 if (ie) {
3316 ie->data.clock_offset = ev->clock_offset;
3317 ie->timestamp = jiffies;
3318 }
3319 }
3320
3321 hci_dev_unlock(hdev);
3322 }
3323
3324 static void hci_pkt_type_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3325 {
3326 struct hci_ev_pkt_type_change *ev = (void *) skb->data;
3327 struct hci_conn *conn;
3328
3329 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3330
3331 hci_dev_lock(hdev);
3332
3333 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3334 if (conn && !ev->status)
3335 conn->pkt_type = __le16_to_cpu(ev->pkt_type);
3336
3337 hci_dev_unlock(hdev);
3338 }
3339
3340 static void hci_pscan_rep_mode_evt(struct hci_dev *hdev, struct sk_buff *skb)
3341 {
3342 struct hci_ev_pscan_rep_mode *ev = (void *) skb->data;
3343 struct inquiry_entry *ie;
3344
3345 BT_DBG("%s", hdev->name);
3346
3347 hci_dev_lock(hdev);
3348
3349 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3350 if (ie) {
3351 ie->data.pscan_rep_mode = ev->pscan_rep_mode;
3352 ie->timestamp = jiffies;
3353 }
3354
3355 hci_dev_unlock(hdev);
3356 }
3357
3358 static void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev,
3359 struct sk_buff *skb)
3360 {
3361 struct inquiry_data data;
3362 int num_rsp = *((__u8 *) skb->data);
3363
3364 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3365
3366 if (!num_rsp)
3367 return;
3368
3369 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3370 return;
3371
3372 hci_dev_lock(hdev);
3373
3374 if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) {
3375 struct inquiry_info_with_rssi_and_pscan_mode *info;
3376 info = (void *) (skb->data + 1);
3377
3378 for (; num_rsp; num_rsp--, info++) {
3379 u32 flags;
3380
3381 bacpy(&data.bdaddr, &info->bdaddr);
3382 data.pscan_rep_mode = info->pscan_rep_mode;
3383 data.pscan_period_mode = info->pscan_period_mode;
3384 data.pscan_mode = info->pscan_mode;
3385 memcpy(data.dev_class, info->dev_class, 3);
3386 data.clock_offset = info->clock_offset;
3387 data.rssi = info->rssi;
3388 data.ssp_mode = 0x00;
3389
3390 flags = hci_inquiry_cache_update(hdev, &data, false);
3391
3392 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3393 info->dev_class, info->rssi,
3394 flags, NULL, 0, NULL, 0);
3395 }
3396 } else {
3397 struct inquiry_info_with_rssi *info = (void *) (skb->data + 1);
3398
3399 for (; num_rsp; num_rsp--, info++) {
3400 u32 flags;
3401
3402 bacpy(&data.bdaddr, &info->bdaddr);
3403 data.pscan_rep_mode = info->pscan_rep_mode;
3404 data.pscan_period_mode = info->pscan_period_mode;
3405 data.pscan_mode = 0x00;
3406 memcpy(data.dev_class, info->dev_class, 3);
3407 data.clock_offset = info->clock_offset;
3408 data.rssi = info->rssi;
3409 data.ssp_mode = 0x00;
3410
3411 flags = hci_inquiry_cache_update(hdev, &data, false);
3412
3413 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3414 info->dev_class, info->rssi,
3415 flags, NULL, 0, NULL, 0);
3416 }
3417 }
3418
3419 hci_dev_unlock(hdev);
3420 }
3421
3422 static void hci_remote_ext_features_evt(struct hci_dev *hdev,
3423 struct sk_buff *skb)
3424 {
3425 struct hci_ev_remote_ext_features *ev = (void *) skb->data;
3426 struct hci_conn *conn;
3427
3428 BT_DBG("%s", hdev->name);
3429
3430 hci_dev_lock(hdev);
3431
3432 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3433 if (!conn)
3434 goto unlock;
3435
3436 if (ev->page < HCI_MAX_PAGES)
3437 memcpy(conn->features[ev->page], ev->features, 8);
3438
3439 if (!ev->status && ev->page == 0x01) {
3440 struct inquiry_entry *ie;
3441
3442 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3443 if (ie)
3444 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3445
3446 if (ev->features[0] & LMP_HOST_SSP) {
3447 set_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3448 } else {
3449 /* It is mandatory by the Bluetooth specification that
3450 * Extended Inquiry Results are only used when Secure
3451 * Simple Pairing is enabled, but some devices violate
3452 * this.
3453 *
3454 * To make these devices work, the internal SSP
3455 * enabled flag needs to be cleared if the remote host
3456 * features do not indicate SSP support */
3457 clear_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3458 }
3459
3460 if (ev->features[0] & LMP_HOST_SC)
3461 set_bit(HCI_CONN_SC_ENABLED, &conn->flags);
3462 }
3463
3464 if (conn->state != BT_CONFIG)
3465 goto unlock;
3466
3467 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
3468 struct hci_cp_remote_name_req cp;
3469 memset(&cp, 0, sizeof(cp));
3470 bacpy(&cp.bdaddr, &conn->dst);
3471 cp.pscan_rep_mode = 0x02;
3472 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
3473 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
3474 mgmt_device_connected(hdev, conn, 0, NULL, 0);
3475
3476 if (!hci_outgoing_auth_needed(hdev, conn)) {
3477 conn->state = BT_CONNECTED;
3478 hci_proto_connect_cfm(conn, ev->status);
3479 hci_conn_drop(conn);
3480 }
3481
3482 unlock:
3483 hci_dev_unlock(hdev);
3484 }
3485
3486 static void hci_sync_conn_complete_evt(struct hci_dev *hdev,
3487 struct sk_buff *skb)
3488 {
3489 struct hci_ev_sync_conn_complete *ev = (void *) skb->data;
3490 struct hci_conn *conn;
3491
3492 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3493
3494 hci_dev_lock(hdev);
3495
3496 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
3497 if (!conn) {
3498 if (ev->link_type == ESCO_LINK)
3499 goto unlock;
3500
3501 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
3502 if (!conn)
3503 goto unlock;
3504
3505 conn->type = SCO_LINK;
3506 }
3507
3508 switch (ev->status) {
3509 case 0x00:
3510 conn->handle = __le16_to_cpu(ev->handle);
3511 conn->state = BT_CONNECTED;
3512
3513 hci_conn_add_sysfs(conn);
3514 break;
3515
3516 case 0x10: /* Connection Accept Timeout */
3517 case 0x0d: /* Connection Rejected due to Limited Resources */
3518 case 0x11: /* Unsupported Feature or Parameter Value */
3519 case 0x1c: /* SCO interval rejected */
3520 case 0x1a: /* Unsupported Remote Feature */
3521 case 0x1f: /* Unspecified error */
3522 case 0x20: /* Unsupported LMP Parameter value */
3523 if (conn->out) {
3524 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
3525 (hdev->esco_type & EDR_ESCO_MASK);
3526 if (hci_setup_sync(conn, conn->link->handle))
3527 goto unlock;
3528 }
3529 /* fall through */
3530
3531 default:
3532 conn->state = BT_CLOSED;
3533 break;
3534 }
3535
3536 hci_proto_connect_cfm(conn, ev->status);
3537 if (ev->status)
3538 hci_conn_del(conn);
3539
3540 unlock:
3541 hci_dev_unlock(hdev);
3542 }
3543
3544 static inline size_t eir_get_length(u8 *eir, size_t eir_len)
3545 {
3546 size_t parsed = 0;
3547
3548 while (parsed < eir_len) {
3549 u8 field_len = eir[0];
3550
3551 if (field_len == 0)
3552 return parsed;
3553
3554 parsed += field_len + 1;
3555 eir += field_len + 1;
3556 }
3557
3558 return eir_len;
3559 }
3560
3561 static void hci_extended_inquiry_result_evt(struct hci_dev *hdev,
3562 struct sk_buff *skb)
3563 {
3564 struct inquiry_data data;
3565 struct extended_inquiry_info *info = (void *) (skb->data + 1);
3566 int num_rsp = *((__u8 *) skb->data);
3567 size_t eir_len;
3568
3569 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3570
3571 if (!num_rsp)
3572 return;
3573
3574 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3575 return;
3576
3577 hci_dev_lock(hdev);
3578
3579 for (; num_rsp; num_rsp--, info++) {
3580 u32 flags;
3581 bool name_known;
3582
3583 bacpy(&data.bdaddr, &info->bdaddr);
3584 data.pscan_rep_mode = info->pscan_rep_mode;
3585 data.pscan_period_mode = info->pscan_period_mode;
3586 data.pscan_mode = 0x00;
3587 memcpy(data.dev_class, info->dev_class, 3);
3588 data.clock_offset = info->clock_offset;
3589 data.rssi = info->rssi;
3590 data.ssp_mode = 0x01;
3591
3592 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3593 name_known = eir_has_data_type(info->data,
3594 sizeof(info->data),
3595 EIR_NAME_COMPLETE);
3596 else
3597 name_known = true;
3598
3599 flags = hci_inquiry_cache_update(hdev, &data, name_known);
3600
3601 eir_len = eir_get_length(info->data, sizeof(info->data));
3602
3603 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3604 info->dev_class, info->rssi,
3605 flags, info->data, eir_len, NULL, 0);
3606 }
3607
3608 hci_dev_unlock(hdev);
3609 }
3610
3611 static void hci_key_refresh_complete_evt(struct hci_dev *hdev,
3612 struct sk_buff *skb)
3613 {
3614 struct hci_ev_key_refresh_complete *ev = (void *) skb->data;
3615 struct hci_conn *conn;
3616
3617 BT_DBG("%s status 0x%2.2x handle 0x%4.4x", hdev->name, ev->status,
3618 __le16_to_cpu(ev->handle));
3619
3620 hci_dev_lock(hdev);
3621
3622 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3623 if (!conn)
3624 goto unlock;
3625
3626 /* For BR/EDR the necessary steps are taken through the
3627 * auth_complete event.
3628 */
3629 if (conn->type != LE_LINK)
3630 goto unlock;
3631
3632 if (!ev->status)
3633 conn->sec_level = conn->pending_sec_level;
3634
3635 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
3636
3637 if (ev->status && conn->state == BT_CONNECTED) {
3638 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
3639 hci_conn_drop(conn);
3640 goto unlock;
3641 }
3642
3643 if (conn->state == BT_CONFIG) {
3644 if (!ev->status)
3645 conn->state = BT_CONNECTED;
3646
3647 hci_proto_connect_cfm(conn, ev->status);
3648 hci_conn_drop(conn);
3649 } else {
3650 hci_auth_cfm(conn, ev->status);
3651
3652 hci_conn_hold(conn);
3653 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3654 hci_conn_drop(conn);
3655 }
3656
3657 unlock:
3658 hci_dev_unlock(hdev);
3659 }
3660
3661 static u8 hci_get_auth_req(struct hci_conn *conn)
3662 {
3663 /* If remote requests no-bonding follow that lead */
3664 if (conn->remote_auth == HCI_AT_NO_BONDING ||
3665 conn->remote_auth == HCI_AT_NO_BONDING_MITM)
3666 return conn->remote_auth | (conn->auth_type & 0x01);
3667
3668 /* If both remote and local have enough IO capabilities, require
3669 * MITM protection
3670 */
3671 if (conn->remote_cap != HCI_IO_NO_INPUT_OUTPUT &&
3672 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT)
3673 return conn->remote_auth | 0x01;
3674
3675 /* No MITM protection possible so ignore remote requirement */
3676 return (conn->remote_auth & ~0x01) | (conn->auth_type & 0x01);
3677 }
3678
3679 static void hci_io_capa_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3680 {
3681 struct hci_ev_io_capa_request *ev = (void *) skb->data;
3682 struct hci_conn *conn;
3683
3684 BT_DBG("%s", hdev->name);
3685
3686 hci_dev_lock(hdev);
3687
3688 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3689 if (!conn)
3690 goto unlock;
3691
3692 hci_conn_hold(conn);
3693
3694 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3695 goto unlock;
3696
3697 /* Allow pairing if we're pairable, the initiators of the
3698 * pairing or if the remote is not requesting bonding.
3699 */
3700 if (test_bit(HCI_BONDABLE, &hdev->dev_flags) ||
3701 test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags) ||
3702 (conn->remote_auth & ~0x01) == HCI_AT_NO_BONDING) {
3703 struct hci_cp_io_capability_reply cp;
3704
3705 bacpy(&cp.bdaddr, &ev->bdaddr);
3706 /* Change the IO capability from KeyboardDisplay
3707 * to DisplayYesNo as it is not supported by BT spec. */
3708 cp.capability = (conn->io_capability == 0x04) ?
3709 HCI_IO_DISPLAY_YESNO : conn->io_capability;
3710
3711 /* If we are initiators, there is no remote information yet */
3712 if (conn->remote_auth == 0xff) {
3713 /* Request MITM protection if our IO caps allow it
3714 * except for the no-bonding case.
3715 */
3716 if (conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3717 conn->auth_type != HCI_AT_NO_BONDING)
3718 conn->auth_type |= 0x01;
3719 } else {
3720 conn->auth_type = hci_get_auth_req(conn);
3721 }
3722
3723 /* If we're not bondable, force one of the non-bondable
3724 * authentication requirement values.
3725 */
3726 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags))
3727 conn->auth_type &= HCI_AT_NO_BONDING_MITM;
3728
3729 cp.authentication = conn->auth_type;
3730
3731 if (hci_find_remote_oob_data(hdev, &conn->dst) &&
3732 (conn->out || test_bit(HCI_CONN_REMOTE_OOB, &conn->flags)))
3733 cp.oob_data = 0x01;
3734 else
3735 cp.oob_data = 0x00;
3736
3737 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_REPLY,
3738 sizeof(cp), &cp);
3739 } else {
3740 struct hci_cp_io_capability_neg_reply cp;
3741
3742 bacpy(&cp.bdaddr, &ev->bdaddr);
3743 cp.reason = HCI_ERROR_PAIRING_NOT_ALLOWED;
3744
3745 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_NEG_REPLY,
3746 sizeof(cp), &cp);
3747 }
3748
3749 unlock:
3750 hci_dev_unlock(hdev);
3751 }
3752
3753 static void hci_io_capa_reply_evt(struct hci_dev *hdev, struct sk_buff *skb)
3754 {
3755 struct hci_ev_io_capa_reply *ev = (void *) skb->data;
3756 struct hci_conn *conn;
3757
3758 BT_DBG("%s", hdev->name);
3759
3760 hci_dev_lock(hdev);
3761
3762 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3763 if (!conn)
3764 goto unlock;
3765
3766 conn->remote_cap = ev->capability;
3767 conn->remote_auth = ev->authentication;
3768 if (ev->oob_data)
3769 set_bit(HCI_CONN_REMOTE_OOB, &conn->flags);
3770
3771 unlock:
3772 hci_dev_unlock(hdev);
3773 }
3774
3775 static void hci_user_confirm_request_evt(struct hci_dev *hdev,
3776 struct sk_buff *skb)
3777 {
3778 struct hci_ev_user_confirm_req *ev = (void *) skb->data;
3779 int loc_mitm, rem_mitm, confirm_hint = 0;
3780 struct hci_conn *conn;
3781
3782 BT_DBG("%s", hdev->name);
3783
3784 hci_dev_lock(hdev);
3785
3786 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3787 goto unlock;
3788
3789 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3790 if (!conn)
3791 goto unlock;
3792
3793 loc_mitm = (conn->auth_type & 0x01);
3794 rem_mitm = (conn->remote_auth & 0x01);
3795
3796 /* If we require MITM but the remote device can't provide that
3797 * (it has NoInputNoOutput) then reject the confirmation
3798 * request. We check the security level here since it doesn't
3799 * necessarily match conn->auth_type.
3800 */
3801 if (conn->pending_sec_level > BT_SECURITY_MEDIUM &&
3802 conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) {
3803 BT_DBG("Rejecting request: remote device can't provide MITM");
3804 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_NEG_REPLY,
3805 sizeof(ev->bdaddr), &ev->bdaddr);
3806 goto unlock;
3807 }
3808
3809 /* If no side requires MITM protection; auto-accept */
3810 if ((!loc_mitm || conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) &&
3811 (!rem_mitm || conn->io_capability == HCI_IO_NO_INPUT_OUTPUT)) {
3812
3813 /* If we're not the initiators request authorization to
3814 * proceed from user space (mgmt_user_confirm with
3815 * confirm_hint set to 1). The exception is if neither
3816 * side had MITM or if the local IO capability is
3817 * NoInputNoOutput, in which case we do auto-accept
3818 */
3819 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) &&
3820 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3821 (loc_mitm || rem_mitm)) {
3822 BT_DBG("Confirming auto-accept as acceptor");
3823 confirm_hint = 1;
3824 goto confirm;
3825 }
3826
3827 BT_DBG("Auto-accept of user confirmation with %ums delay",
3828 hdev->auto_accept_delay);
3829
3830 if (hdev->auto_accept_delay > 0) {
3831 int delay = msecs_to_jiffies(hdev->auto_accept_delay);
3832 queue_delayed_work(conn->hdev->workqueue,
3833 &conn->auto_accept_work, delay);
3834 goto unlock;
3835 }
3836
3837 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY,
3838 sizeof(ev->bdaddr), &ev->bdaddr);
3839 goto unlock;
3840 }
3841
3842 confirm:
3843 mgmt_user_confirm_request(hdev, &ev->bdaddr, ACL_LINK, 0,
3844 le32_to_cpu(ev->passkey), confirm_hint);
3845
3846 unlock:
3847 hci_dev_unlock(hdev);
3848 }
3849
3850 static void hci_user_passkey_request_evt(struct hci_dev *hdev,
3851 struct sk_buff *skb)
3852 {
3853 struct hci_ev_user_passkey_req *ev = (void *) skb->data;
3854
3855 BT_DBG("%s", hdev->name);
3856
3857 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3858 mgmt_user_passkey_request(hdev, &ev->bdaddr, ACL_LINK, 0);
3859 }
3860
3861 static void hci_user_passkey_notify_evt(struct hci_dev *hdev,
3862 struct sk_buff *skb)
3863 {
3864 struct hci_ev_user_passkey_notify *ev = (void *) skb->data;
3865 struct hci_conn *conn;
3866
3867 BT_DBG("%s", hdev->name);
3868
3869 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3870 if (!conn)
3871 return;
3872
3873 conn->passkey_notify = __le32_to_cpu(ev->passkey);
3874 conn->passkey_entered = 0;
3875
3876 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3877 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3878 conn->dst_type, conn->passkey_notify,
3879 conn->passkey_entered);
3880 }
3881
3882 static void hci_keypress_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3883 {
3884 struct hci_ev_keypress_notify *ev = (void *) skb->data;
3885 struct hci_conn *conn;
3886
3887 BT_DBG("%s", hdev->name);
3888
3889 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3890 if (!conn)
3891 return;
3892
3893 switch (ev->type) {
3894 case HCI_KEYPRESS_STARTED:
3895 conn->passkey_entered = 0;
3896 return;
3897
3898 case HCI_KEYPRESS_ENTERED:
3899 conn->passkey_entered++;
3900 break;
3901
3902 case HCI_KEYPRESS_ERASED:
3903 conn->passkey_entered--;
3904 break;
3905
3906 case HCI_KEYPRESS_CLEARED:
3907 conn->passkey_entered = 0;
3908 break;
3909
3910 case HCI_KEYPRESS_COMPLETED:
3911 return;
3912 }
3913
3914 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3915 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3916 conn->dst_type, conn->passkey_notify,
3917 conn->passkey_entered);
3918 }
3919
3920 static void hci_simple_pair_complete_evt(struct hci_dev *hdev,
3921 struct sk_buff *skb)
3922 {
3923 struct hci_ev_simple_pair_complete *ev = (void *) skb->data;
3924 struct hci_conn *conn;
3925
3926 BT_DBG("%s", hdev->name);
3927
3928 hci_dev_lock(hdev);
3929
3930 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3931 if (!conn)
3932 goto unlock;
3933
3934 /* Reset the authentication requirement to unknown */
3935 conn->remote_auth = 0xff;
3936
3937 /* To avoid duplicate auth_failed events to user space we check
3938 * the HCI_CONN_AUTH_PEND flag which will be set if we
3939 * initiated the authentication. A traditional auth_complete
3940 * event gets always produced as initiator and is also mapped to
3941 * the mgmt_auth_failed event */
3942 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) && ev->status)
3943 mgmt_auth_failed(conn, ev->status);
3944
3945 hci_conn_drop(conn);
3946
3947 unlock:
3948 hci_dev_unlock(hdev);
3949 }
3950
3951 static void hci_remote_host_features_evt(struct hci_dev *hdev,
3952 struct sk_buff *skb)
3953 {
3954 struct hci_ev_remote_host_features *ev = (void *) skb->data;
3955 struct inquiry_entry *ie;
3956 struct hci_conn *conn;
3957
3958 BT_DBG("%s", hdev->name);
3959
3960 hci_dev_lock(hdev);
3961
3962 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3963 if (conn)
3964 memcpy(conn->features[1], ev->features, 8);
3965
3966 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3967 if (ie)
3968 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3969
3970 hci_dev_unlock(hdev);
3971 }
3972
3973 static void hci_remote_oob_data_request_evt(struct hci_dev *hdev,
3974 struct sk_buff *skb)
3975 {
3976 struct hci_ev_remote_oob_data_request *ev = (void *) skb->data;
3977 struct oob_data *data;
3978
3979 BT_DBG("%s", hdev->name);
3980
3981 hci_dev_lock(hdev);
3982
3983 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3984 goto unlock;
3985
3986 data = hci_find_remote_oob_data(hdev, &ev->bdaddr);
3987 if (data) {
3988 if (test_bit(HCI_SC_ENABLED, &hdev->dev_flags)) {
3989 struct hci_cp_remote_oob_ext_data_reply cp;
3990
3991 bacpy(&cp.bdaddr, &ev->bdaddr);
3992 memcpy(cp.hash192, data->hash192, sizeof(cp.hash192));
3993 memcpy(cp.randomizer192, data->randomizer192,
3994 sizeof(cp.randomizer192));
3995 memcpy(cp.hash256, data->hash256, sizeof(cp.hash256));
3996 memcpy(cp.randomizer256, data->randomizer256,
3997 sizeof(cp.randomizer256));
3998
3999 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_EXT_DATA_REPLY,
4000 sizeof(cp), &cp);
4001 } else {
4002 struct hci_cp_remote_oob_data_reply cp;
4003
4004 bacpy(&cp.bdaddr, &ev->bdaddr);
4005 memcpy(cp.hash, data->hash192, sizeof(cp.hash));
4006 memcpy(cp.randomizer, data->randomizer192,
4007 sizeof(cp.randomizer));
4008
4009 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_REPLY,
4010 sizeof(cp), &cp);
4011 }
4012 } else {
4013 struct hci_cp_remote_oob_data_neg_reply cp;
4014
4015 bacpy(&cp.bdaddr, &ev->bdaddr);
4016 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_NEG_REPLY,
4017 sizeof(cp), &cp);
4018 }
4019
4020 unlock:
4021 hci_dev_unlock(hdev);
4022 }
4023
4024 static void hci_phy_link_complete_evt(struct hci_dev *hdev,
4025 struct sk_buff *skb)
4026 {
4027 struct hci_ev_phy_link_complete *ev = (void *) skb->data;
4028 struct hci_conn *hcon, *bredr_hcon;
4029
4030 BT_DBG("%s handle 0x%2.2x status 0x%2.2x", hdev->name, ev->phy_handle,
4031 ev->status);
4032
4033 hci_dev_lock(hdev);
4034
4035 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4036 if (!hcon) {
4037 hci_dev_unlock(hdev);
4038 return;
4039 }
4040
4041 if (ev->status) {
4042 hci_conn_del(hcon);
4043 hci_dev_unlock(hdev);
4044 return;
4045 }
4046
4047 bredr_hcon = hcon->amp_mgr->l2cap_conn->hcon;
4048
4049 hcon->state = BT_CONNECTED;
4050 bacpy(&hcon->dst, &bredr_hcon->dst);
4051
4052 hci_conn_hold(hcon);
4053 hcon->disc_timeout = HCI_DISCONN_TIMEOUT;
4054 hci_conn_drop(hcon);
4055
4056 hci_conn_add_sysfs(hcon);
4057
4058 amp_physical_cfm(bredr_hcon, hcon);
4059
4060 hci_dev_unlock(hdev);
4061 }
4062
4063 static void hci_loglink_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4064 {
4065 struct hci_ev_logical_link_complete *ev = (void *) skb->data;
4066 struct hci_conn *hcon;
4067 struct hci_chan *hchan;
4068 struct amp_mgr *mgr;
4069
4070 BT_DBG("%s log_handle 0x%4.4x phy_handle 0x%2.2x status 0x%2.2x",
4071 hdev->name, le16_to_cpu(ev->handle), ev->phy_handle,
4072 ev->status);
4073
4074 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4075 if (!hcon)
4076 return;
4077
4078 /* Create AMP hchan */
4079 hchan = hci_chan_create(hcon);
4080 if (!hchan)
4081 return;
4082
4083 hchan->handle = le16_to_cpu(ev->handle);
4084
4085 BT_DBG("hcon %p mgr %p hchan %p", hcon, hcon->amp_mgr, hchan);
4086
4087 mgr = hcon->amp_mgr;
4088 if (mgr && mgr->bredr_chan) {
4089 struct l2cap_chan *bredr_chan = mgr->bredr_chan;
4090
4091 l2cap_chan_lock(bredr_chan);
4092
4093 bredr_chan->conn->mtu = hdev->block_mtu;
4094 l2cap_logical_cfm(bredr_chan, hchan, 0);
4095 hci_conn_hold(hcon);
4096
4097 l2cap_chan_unlock(bredr_chan);
4098 }
4099 }
4100
4101 static void hci_disconn_loglink_complete_evt(struct hci_dev *hdev,
4102 struct sk_buff *skb)
4103 {
4104 struct hci_ev_disconn_logical_link_complete *ev = (void *) skb->data;
4105 struct hci_chan *hchan;
4106
4107 BT_DBG("%s log handle 0x%4.4x status 0x%2.2x", hdev->name,
4108 le16_to_cpu(ev->handle), ev->status);
4109
4110 if (ev->status)
4111 return;
4112
4113 hci_dev_lock(hdev);
4114
4115 hchan = hci_chan_lookup_handle(hdev, le16_to_cpu(ev->handle));
4116 if (!hchan)
4117 goto unlock;
4118
4119 amp_destroy_logical_link(hchan, ev->reason);
4120
4121 unlock:
4122 hci_dev_unlock(hdev);
4123 }
4124
4125 static void hci_disconn_phylink_complete_evt(struct hci_dev *hdev,
4126 struct sk_buff *skb)
4127 {
4128 struct hci_ev_disconn_phy_link_complete *ev = (void *) skb->data;
4129 struct hci_conn *hcon;
4130
4131 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4132
4133 if (ev->status)
4134 return;
4135
4136 hci_dev_lock(hdev);
4137
4138 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4139 if (hcon) {
4140 hcon->state = BT_CLOSED;
4141 hci_conn_del(hcon);
4142 }
4143
4144 hci_dev_unlock(hdev);
4145 }
4146
4147 static void hci_le_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4148 {
4149 struct hci_ev_le_conn_complete *ev = (void *) skb->data;
4150 struct hci_conn_params *params;
4151 struct hci_conn *conn;
4152 struct smp_irk *irk;
4153 u8 addr_type;
4154
4155 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4156
4157 hci_dev_lock(hdev);
4158
4159 /* All controllers implicitly stop advertising in the event of a
4160 * connection, so ensure that the state bit is cleared.
4161 */
4162 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
4163
4164 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
4165 if (!conn) {
4166 conn = hci_conn_add(hdev, LE_LINK, &ev->bdaddr, ev->role);
4167 if (!conn) {
4168 BT_ERR("No memory for new connection");
4169 goto unlock;
4170 }
4171
4172 conn->dst_type = ev->bdaddr_type;
4173
4174 /* If we didn't have a hci_conn object previously
4175 * but we're in master role this must be something
4176 * initiated using a white list. Since white list based
4177 * connections are not "first class citizens" we don't
4178 * have full tracking of them. Therefore, we go ahead
4179 * with a "best effort" approach of determining the
4180 * initiator address based on the HCI_PRIVACY flag.
4181 */
4182 if (conn->out) {
4183 conn->resp_addr_type = ev->bdaddr_type;
4184 bacpy(&conn->resp_addr, &ev->bdaddr);
4185 if (test_bit(HCI_PRIVACY, &hdev->dev_flags)) {
4186 conn->init_addr_type = ADDR_LE_DEV_RANDOM;
4187 bacpy(&conn->init_addr, &hdev->rpa);
4188 } else {
4189 hci_copy_identity_address(hdev,
4190 &conn->init_addr,
4191 &conn->init_addr_type);
4192 }
4193 }
4194 } else {
4195 cancel_delayed_work(&conn->le_conn_timeout);
4196 }
4197
4198 if (!conn->out) {
4199 /* Set the responder (our side) address type based on
4200 * the advertising address type.
4201 */
4202 conn->resp_addr_type = hdev->adv_addr_type;
4203 if (hdev->adv_addr_type == ADDR_LE_DEV_RANDOM)
4204 bacpy(&conn->resp_addr, &hdev->random_addr);
4205 else
4206 bacpy(&conn->resp_addr, &hdev->bdaddr);
4207
4208 conn->init_addr_type = ev->bdaddr_type;
4209 bacpy(&conn->init_addr, &ev->bdaddr);
4210
4211 /* For incoming connections, set the default minimum
4212 * and maximum connection interval. They will be used
4213 * to check if the parameters are in range and if not
4214 * trigger the connection update procedure.
4215 */
4216 conn->le_conn_min_interval = hdev->le_conn_min_interval;
4217 conn->le_conn_max_interval = hdev->le_conn_max_interval;
4218 }
4219
4220 /* Lookup the identity address from the stored connection
4221 * address and address type.
4222 *
4223 * When establishing connections to an identity address, the
4224 * connection procedure will store the resolvable random
4225 * address first. Now if it can be converted back into the
4226 * identity address, start using the identity address from
4227 * now on.
4228 */
4229 irk = hci_get_irk(hdev, &conn->dst, conn->dst_type);
4230 if (irk) {
4231 bacpy(&conn->dst, &irk->bdaddr);
4232 conn->dst_type = irk->addr_type;
4233 }
4234
4235 if (ev->status) {
4236 hci_le_conn_failed(conn, ev->status);
4237 goto unlock;
4238 }
4239
4240 if (conn->dst_type == ADDR_LE_DEV_PUBLIC)
4241 addr_type = BDADDR_LE_PUBLIC;
4242 else
4243 addr_type = BDADDR_LE_RANDOM;
4244
4245 /* Drop the connection if the device is blocked */
4246 if (hci_bdaddr_list_lookup(&hdev->blacklist, &conn->dst, addr_type)) {
4247 hci_conn_drop(conn);
4248 goto unlock;
4249 }
4250
4251 if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4252 mgmt_device_connected(hdev, conn, 0, NULL, 0);
4253
4254 conn->sec_level = BT_SECURITY_LOW;
4255 conn->handle = __le16_to_cpu(ev->handle);
4256 conn->state = BT_CONNECTED;
4257
4258 conn->le_conn_interval = le16_to_cpu(ev->interval);
4259 conn->le_conn_latency = le16_to_cpu(ev->latency);
4260 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4261
4262 hci_conn_add_sysfs(conn);
4263
4264 hci_proto_connect_cfm(conn, ev->status);
4265
4266 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, &conn->dst,
4267 conn->dst_type);
4268 if (params) {
4269 list_del_init(&params->action);
4270 if (params->conn) {
4271 hci_conn_drop(params->conn);
4272 hci_conn_put(params->conn);
4273 params->conn = NULL;
4274 }
4275 }
4276
4277 unlock:
4278 hci_update_background_scan(hdev);
4279 hci_dev_unlock(hdev);
4280 }
4281
4282 static void hci_le_conn_update_complete_evt(struct hci_dev *hdev,
4283 struct sk_buff *skb)
4284 {
4285 struct hci_ev_le_conn_update_complete *ev = (void *) skb->data;
4286 struct hci_conn *conn;
4287
4288 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4289
4290 if (ev->status)
4291 return;
4292
4293 hci_dev_lock(hdev);
4294
4295 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4296 if (conn) {
4297 conn->le_conn_interval = le16_to_cpu(ev->interval);
4298 conn->le_conn_latency = le16_to_cpu(ev->latency);
4299 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4300 }
4301
4302 hci_dev_unlock(hdev);
4303 }
4304
4305 /* This function requires the caller holds hdev->lock */
4306 static struct hci_conn *check_pending_le_conn(struct hci_dev *hdev,
4307 bdaddr_t *addr,
4308 u8 addr_type, u8 adv_type)
4309 {
4310 struct hci_conn *conn;
4311 struct hci_conn_params *params;
4312
4313 /* If the event is not connectable don't proceed further */
4314 if (adv_type != LE_ADV_IND && adv_type != LE_ADV_DIRECT_IND)
4315 return NULL;
4316
4317 /* Ignore if the device is blocked */
4318 if (hci_bdaddr_list_lookup(&hdev->blacklist, addr, addr_type))
4319 return NULL;
4320
4321 /* Most controller will fail if we try to create new connections
4322 * while we have an existing one in slave role.
4323 */
4324 if (hdev->conn_hash.le_num_slave > 0)
4325 return NULL;
4326
4327 /* If we're not connectable only connect devices that we have in
4328 * our pend_le_conns list.
4329 */
4330 params = hci_pend_le_action_lookup(&hdev->pend_le_conns,
4331 addr, addr_type);
4332 if (!params)
4333 return NULL;
4334
4335 switch (params->auto_connect) {
4336 case HCI_AUTO_CONN_DIRECT:
4337 /* Only devices advertising with ADV_DIRECT_IND are
4338 * triggering a connection attempt. This is allowing
4339 * incoming connections from slave devices.
4340 */
4341 if (adv_type != LE_ADV_DIRECT_IND)
4342 return NULL;
4343 break;
4344 case HCI_AUTO_CONN_ALWAYS:
4345 /* Devices advertising with ADV_IND or ADV_DIRECT_IND
4346 * are triggering a connection attempt. This means
4347 * that incoming connectioms from slave device are
4348 * accepted and also outgoing connections to slave
4349 * devices are established when found.
4350 */
4351 break;
4352 default:
4353 return NULL;
4354 }
4355
4356 conn = hci_connect_le(hdev, addr, addr_type, BT_SECURITY_LOW,
4357 HCI_LE_AUTOCONN_TIMEOUT, HCI_ROLE_MASTER);
4358 if (!IS_ERR(conn)) {
4359 /* Store the pointer since we don't really have any
4360 * other owner of the object besides the params that
4361 * triggered it. This way we can abort the connection if
4362 * the parameters get removed and keep the reference
4363 * count consistent once the connection is established.
4364 */
4365 params->conn = hci_conn_get(conn);
4366 return conn;
4367 }
4368
4369 switch (PTR_ERR(conn)) {
4370 case -EBUSY:
4371 /* If hci_connect() returns -EBUSY it means there is already
4372 * an LE connection attempt going on. Since controllers don't
4373 * support more than one connection attempt at the time, we
4374 * don't consider this an error case.
4375 */
4376 break;
4377 default:
4378 BT_DBG("Failed to connect: err %ld", PTR_ERR(conn));
4379 return NULL;
4380 }
4381
4382 return NULL;
4383 }
4384
4385 static void process_adv_report(struct hci_dev *hdev, u8 type, bdaddr_t *bdaddr,
4386 u8 bdaddr_type, s8 rssi, u8 *data, u8 len)
4387 {
4388 struct discovery_state *d = &hdev->discovery;
4389 struct smp_irk *irk;
4390 struct hci_conn *conn;
4391 bool match;
4392 u32 flags;
4393
4394 /* Check if we need to convert to identity address */
4395 irk = hci_get_irk(hdev, bdaddr, bdaddr_type);
4396 if (irk) {
4397 bdaddr = &irk->bdaddr;
4398 bdaddr_type = irk->addr_type;
4399 }
4400
4401 /* Check if we have been requested to connect to this device */
4402 conn = check_pending_le_conn(hdev, bdaddr, bdaddr_type, type);
4403 if (conn && type == LE_ADV_IND) {
4404 /* Store report for later inclusion by
4405 * mgmt_device_connected
4406 */
4407 memcpy(conn->le_adv_data, data, len);
4408 conn->le_adv_data_len = len;
4409 }
4410
4411 /* Passive scanning shouldn't trigger any device found events,
4412 * except for devices marked as CONN_REPORT for which we do send
4413 * device found events.
4414 */
4415 if (hdev->le_scan_type == LE_SCAN_PASSIVE) {
4416 if (type == LE_ADV_DIRECT_IND)
4417 return;
4418
4419 if (!hci_pend_le_action_lookup(&hdev->pend_le_reports,
4420 bdaddr, bdaddr_type))
4421 return;
4422
4423 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND)
4424 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4425 else
4426 flags = 0;
4427 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4428 rssi, flags, data, len, NULL, 0);
4429 return;
4430 }
4431
4432 /* When receiving non-connectable or scannable undirected
4433 * advertising reports, this means that the remote device is
4434 * not connectable and then clearly indicate this in the
4435 * device found event.
4436 *
4437 * When receiving a scan response, then there is no way to
4438 * know if the remote device is connectable or not. However
4439 * since scan responses are merged with a previously seen
4440 * advertising report, the flags field from that report
4441 * will be used.
4442 *
4443 * In the really unlikely case that a controller get confused
4444 * and just sends a scan response event, then it is marked as
4445 * not connectable as well.
4446 */
4447 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND ||
4448 type == LE_ADV_SCAN_RSP)
4449 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4450 else
4451 flags = 0;
4452
4453 /* If there's nothing pending either store the data from this
4454 * event or send an immediate device found event if the data
4455 * should not be stored for later.
4456 */
4457 if (!has_pending_adv_report(hdev)) {
4458 /* If the report will trigger a SCAN_REQ store it for
4459 * later merging.
4460 */
4461 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4462 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4463 rssi, flags, data, len);
4464 return;
4465 }
4466
4467 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4468 rssi, flags, data, len, NULL, 0);
4469 return;
4470 }
4471
4472 /* Check if the pending report is for the same device as the new one */
4473 match = (!bacmp(bdaddr, &d->last_adv_addr) &&
4474 bdaddr_type == d->last_adv_addr_type);
4475
4476 /* If the pending data doesn't match this report or this isn't a
4477 * scan response (e.g. we got a duplicate ADV_IND) then force
4478 * sending of the pending data.
4479 */
4480 if (type != LE_ADV_SCAN_RSP || !match) {
4481 /* Send out whatever is in the cache, but skip duplicates */
4482 if (!match)
4483 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4484 d->last_adv_addr_type, NULL,
4485 d->last_adv_rssi, d->last_adv_flags,
4486 d->last_adv_data,
4487 d->last_adv_data_len, NULL, 0);
4488
4489 /* If the new report will trigger a SCAN_REQ store it for
4490 * later merging.
4491 */
4492 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4493 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4494 rssi, flags, data, len);
4495 return;
4496 }
4497
4498 /* The advertising reports cannot be merged, so clear
4499 * the pending report and send out a device found event.
4500 */
4501 clear_pending_adv_report(hdev);
4502 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4503 rssi, flags, data, len, NULL, 0);
4504 return;
4505 }
4506
4507 /* If we get here we've got a pending ADV_IND or ADV_SCAN_IND and
4508 * the new event is a SCAN_RSP. We can therefore proceed with
4509 * sending a merged device found event.
4510 */
4511 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4512 d->last_adv_addr_type, NULL, rssi, d->last_adv_flags,
4513 d->last_adv_data, d->last_adv_data_len, data, len);
4514 clear_pending_adv_report(hdev);
4515 }
4516
4517 static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
4518 {
4519 u8 num_reports = skb->data[0];
4520 void *ptr = &skb->data[1];
4521
4522 hci_dev_lock(hdev);
4523
4524 while (num_reports--) {
4525 struct hci_ev_le_advertising_info *ev = ptr;
4526 s8 rssi;
4527
4528 rssi = ev->data[ev->length];
4529 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
4530 ev->bdaddr_type, rssi, ev->data, ev->length);
4531
4532 ptr += sizeof(*ev) + ev->length + 1;
4533 }
4534
4535 hci_dev_unlock(hdev);
4536 }
4537
4538 static void hci_le_ltk_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
4539 {
4540 struct hci_ev_le_ltk_req *ev = (void *) skb->data;
4541 struct hci_cp_le_ltk_reply cp;
4542 struct hci_cp_le_ltk_neg_reply neg;
4543 struct hci_conn *conn;
4544 struct smp_ltk *ltk;
4545
4546 BT_DBG("%s handle 0x%4.4x", hdev->name, __le16_to_cpu(ev->handle));
4547
4548 hci_dev_lock(hdev);
4549
4550 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4551 if (conn == NULL)
4552 goto not_found;
4553
4554 ltk = hci_find_ltk(hdev, ev->ediv, ev->rand, conn->role);
4555 if (ltk == NULL)
4556 goto not_found;
4557
4558 memcpy(cp.ltk, ltk->val, sizeof(ltk->val));
4559 cp.handle = cpu_to_le16(conn->handle);
4560
4561 conn->pending_sec_level = smp_ltk_sec_level(ltk);
4562
4563 conn->enc_key_size = ltk->enc_size;
4564
4565 hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp);
4566
4567 /* Ref. Bluetooth Core SPEC pages 1975 and 2004. STK is a
4568 * temporary key used to encrypt a connection following
4569 * pairing. It is used during the Encrypted Session Setup to
4570 * distribute the keys. Later, security can be re-established
4571 * using a distributed LTK.
4572 */
4573 if (ltk->type == SMP_STK) {
4574 set_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4575 list_del(&ltk->list);
4576 kfree(ltk);
4577 } else {
4578 clear_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4579 }
4580
4581 hci_dev_unlock(hdev);
4582
4583 return;
4584
4585 not_found:
4586 neg.handle = ev->handle;
4587 hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(neg), &neg);
4588 hci_dev_unlock(hdev);
4589 }
4590
4591 static void send_conn_param_neg_reply(struct hci_dev *hdev, u16 handle,
4592 u8 reason)
4593 {
4594 struct hci_cp_le_conn_param_req_neg_reply cp;
4595
4596 cp.handle = cpu_to_le16(handle);
4597 cp.reason = reason;
4598
4599 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_NEG_REPLY, sizeof(cp),
4600 &cp);
4601 }
4602
4603 static void hci_le_remote_conn_param_req_evt(struct hci_dev *hdev,
4604 struct sk_buff *skb)
4605 {
4606 struct hci_ev_le_remote_conn_param_req *ev = (void *) skb->data;
4607 struct hci_cp_le_conn_param_req_reply cp;
4608 struct hci_conn *hcon;
4609 u16 handle, min, max, latency, timeout;
4610
4611 handle = le16_to_cpu(ev->handle);
4612 min = le16_to_cpu(ev->interval_min);
4613 max = le16_to_cpu(ev->interval_max);
4614 latency = le16_to_cpu(ev->latency);
4615 timeout = le16_to_cpu(ev->timeout);
4616
4617 hcon = hci_conn_hash_lookup_handle(hdev, handle);
4618 if (!hcon || hcon->state != BT_CONNECTED)
4619 return send_conn_param_neg_reply(hdev, handle,
4620 HCI_ERROR_UNKNOWN_CONN_ID);
4621
4622 if (hci_check_conn_params(min, max, latency, timeout))
4623 return send_conn_param_neg_reply(hdev, handle,
4624 HCI_ERROR_INVALID_LL_PARAMS);
4625
4626 if (hcon->role == HCI_ROLE_MASTER) {
4627 struct hci_conn_params *params;
4628 u8 store_hint;
4629
4630 hci_dev_lock(hdev);
4631
4632 params = hci_conn_params_lookup(hdev, &hcon->dst,
4633 hcon->dst_type);
4634 if (params) {
4635 params->conn_min_interval = min;
4636 params->conn_max_interval = max;
4637 params->conn_latency = latency;
4638 params->supervision_timeout = timeout;
4639 store_hint = 0x01;
4640 } else{
4641 store_hint = 0x00;
4642 }
4643
4644 hci_dev_unlock(hdev);
4645
4646 mgmt_new_conn_param(hdev, &hcon->dst, hcon->dst_type,
4647 store_hint, min, max, latency, timeout);
4648 }
4649
4650 cp.handle = ev->handle;
4651 cp.interval_min = ev->interval_min;
4652 cp.interval_max = ev->interval_max;
4653 cp.latency = ev->latency;
4654 cp.timeout = ev->timeout;
4655 cp.min_ce_len = 0;
4656 cp.max_ce_len = 0;
4657
4658 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_REPLY, sizeof(cp), &cp);
4659 }
4660
4661 static void hci_le_meta_evt(struct hci_dev *hdev, struct sk_buff *skb)
4662 {
4663 struct hci_ev_le_meta *le_ev = (void *) skb->data;
4664
4665 skb_pull(skb, sizeof(*le_ev));
4666
4667 switch (le_ev->subevent) {
4668 case HCI_EV_LE_CONN_COMPLETE:
4669 hci_le_conn_complete_evt(hdev, skb);
4670 break;
4671
4672 case HCI_EV_LE_CONN_UPDATE_COMPLETE:
4673 hci_le_conn_update_complete_evt(hdev, skb);
4674 break;
4675
4676 case HCI_EV_LE_ADVERTISING_REPORT:
4677 hci_le_adv_report_evt(hdev, skb);
4678 break;
4679
4680 case HCI_EV_LE_LTK_REQ:
4681 hci_le_ltk_request_evt(hdev, skb);
4682 break;
4683
4684 case HCI_EV_LE_REMOTE_CONN_PARAM_REQ:
4685 hci_le_remote_conn_param_req_evt(hdev, skb);
4686 break;
4687
4688 default:
4689 break;
4690 }
4691 }
4692
4693 static void hci_chan_selected_evt(struct hci_dev *hdev, struct sk_buff *skb)
4694 {
4695 struct hci_ev_channel_selected *ev = (void *) skb->data;
4696 struct hci_conn *hcon;
4697
4698 BT_DBG("%s handle 0x%2.2x", hdev->name, ev->phy_handle);
4699
4700 skb_pull(skb, sizeof(*ev));
4701
4702 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4703 if (!hcon)
4704 return;
4705
4706 amp_read_loc_assoc_final_data(hdev, hcon);
4707 }
4708
4709 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb)
4710 {
4711 struct hci_event_hdr *hdr = (void *) skb->data;
4712 __u8 event = hdr->evt;
4713
4714 hci_dev_lock(hdev);
4715
4716 /* Received events are (currently) only needed when a request is
4717 * ongoing so avoid unnecessary memory allocation.
4718 */
4719 if (hci_req_pending(hdev)) {
4720 kfree_skb(hdev->recv_evt);
4721 hdev->recv_evt = skb_clone(skb, GFP_KERNEL);
4722 }
4723
4724 hci_dev_unlock(hdev);
4725
4726 skb_pull(skb, HCI_EVENT_HDR_SIZE);
4727
4728 if (hdev->sent_cmd && bt_cb(hdev->sent_cmd)->req.event == event) {
4729 struct hci_command_hdr *cmd_hdr = (void *) hdev->sent_cmd->data;
4730 u16 opcode = __le16_to_cpu(cmd_hdr->opcode);
4731
4732 hci_req_cmd_complete(hdev, opcode, 0);
4733 }
4734
4735 switch (event) {
4736 case HCI_EV_INQUIRY_COMPLETE:
4737 hci_inquiry_complete_evt(hdev, skb);
4738 break;
4739
4740 case HCI_EV_INQUIRY_RESULT:
4741 hci_inquiry_result_evt(hdev, skb);
4742 break;
4743
4744 case HCI_EV_CONN_COMPLETE:
4745 hci_conn_complete_evt(hdev, skb);
4746 break;
4747
4748 case HCI_EV_CONN_REQUEST:
4749 hci_conn_request_evt(hdev, skb);
4750 break;
4751
4752 case HCI_EV_DISCONN_COMPLETE:
4753 hci_disconn_complete_evt(hdev, skb);
4754 break;
4755
4756 case HCI_EV_AUTH_COMPLETE:
4757 hci_auth_complete_evt(hdev, skb);
4758 break;
4759
4760 case HCI_EV_REMOTE_NAME:
4761 hci_remote_name_evt(hdev, skb);
4762 break;
4763
4764 case HCI_EV_ENCRYPT_CHANGE:
4765 hci_encrypt_change_evt(hdev, skb);
4766 break;
4767
4768 case HCI_EV_CHANGE_LINK_KEY_COMPLETE:
4769 hci_change_link_key_complete_evt(hdev, skb);
4770 break;
4771
4772 case HCI_EV_REMOTE_FEATURES:
4773 hci_remote_features_evt(hdev, skb);
4774 break;
4775
4776 case HCI_EV_CMD_COMPLETE:
4777 hci_cmd_complete_evt(hdev, skb);
4778 break;
4779
4780 case HCI_EV_CMD_STATUS:
4781 hci_cmd_status_evt(hdev, skb);
4782 break;
4783
4784 case HCI_EV_HARDWARE_ERROR:
4785 hci_hardware_error_evt(hdev, skb);
4786 break;
4787
4788 case HCI_EV_ROLE_CHANGE:
4789 hci_role_change_evt(hdev, skb);
4790 break;
4791
4792 case HCI_EV_NUM_COMP_PKTS:
4793 hci_num_comp_pkts_evt(hdev, skb);
4794 break;
4795
4796 case HCI_EV_MODE_CHANGE:
4797 hci_mode_change_evt(hdev, skb);
4798 break;
4799
4800 case HCI_EV_PIN_CODE_REQ:
4801 hci_pin_code_request_evt(hdev, skb);
4802 break;
4803
4804 case HCI_EV_LINK_KEY_REQ:
4805 hci_link_key_request_evt(hdev, skb);
4806 break;
4807
4808 case HCI_EV_LINK_KEY_NOTIFY:
4809 hci_link_key_notify_evt(hdev, skb);
4810 break;
4811
4812 case HCI_EV_CLOCK_OFFSET:
4813 hci_clock_offset_evt(hdev, skb);
4814 break;
4815
4816 case HCI_EV_PKT_TYPE_CHANGE:
4817 hci_pkt_type_change_evt(hdev, skb);
4818 break;
4819
4820 case HCI_EV_PSCAN_REP_MODE:
4821 hci_pscan_rep_mode_evt(hdev, skb);
4822 break;
4823
4824 case HCI_EV_INQUIRY_RESULT_WITH_RSSI:
4825 hci_inquiry_result_with_rssi_evt(hdev, skb);
4826 break;
4827
4828 case HCI_EV_REMOTE_EXT_FEATURES:
4829 hci_remote_ext_features_evt(hdev, skb);
4830 break;
4831
4832 case HCI_EV_SYNC_CONN_COMPLETE:
4833 hci_sync_conn_complete_evt(hdev, skb);
4834 break;
4835
4836 case HCI_EV_EXTENDED_INQUIRY_RESULT:
4837 hci_extended_inquiry_result_evt(hdev, skb);
4838 break;
4839
4840 case HCI_EV_KEY_REFRESH_COMPLETE:
4841 hci_key_refresh_complete_evt(hdev, skb);
4842 break;
4843
4844 case HCI_EV_IO_CAPA_REQUEST:
4845 hci_io_capa_request_evt(hdev, skb);
4846 break;
4847
4848 case HCI_EV_IO_CAPA_REPLY:
4849 hci_io_capa_reply_evt(hdev, skb);
4850 break;
4851
4852 case HCI_EV_USER_CONFIRM_REQUEST:
4853 hci_user_confirm_request_evt(hdev, skb);
4854 break;
4855
4856 case HCI_EV_USER_PASSKEY_REQUEST:
4857 hci_user_passkey_request_evt(hdev, skb);
4858 break;
4859
4860 case HCI_EV_USER_PASSKEY_NOTIFY:
4861 hci_user_passkey_notify_evt(hdev, skb);
4862 break;
4863
4864 case HCI_EV_KEYPRESS_NOTIFY:
4865 hci_keypress_notify_evt(hdev, skb);
4866 break;
4867
4868 case HCI_EV_SIMPLE_PAIR_COMPLETE:
4869 hci_simple_pair_complete_evt(hdev, skb);
4870 break;
4871
4872 case HCI_EV_REMOTE_HOST_FEATURES:
4873 hci_remote_host_features_evt(hdev, skb);
4874 break;
4875
4876 case HCI_EV_LE_META:
4877 hci_le_meta_evt(hdev, skb);
4878 break;
4879
4880 case HCI_EV_CHANNEL_SELECTED:
4881 hci_chan_selected_evt(hdev, skb);
4882 break;
4883
4884 case HCI_EV_REMOTE_OOB_DATA_REQUEST:
4885 hci_remote_oob_data_request_evt(hdev, skb);
4886 break;
4887
4888 case HCI_EV_PHY_LINK_COMPLETE:
4889 hci_phy_link_complete_evt(hdev, skb);
4890 break;
4891
4892 case HCI_EV_LOGICAL_LINK_COMPLETE:
4893 hci_loglink_complete_evt(hdev, skb);
4894 break;
4895
4896 case HCI_EV_DISCONN_LOGICAL_LINK_COMPLETE:
4897 hci_disconn_loglink_complete_evt(hdev, skb);
4898 break;
4899
4900 case HCI_EV_DISCONN_PHY_LINK_COMPLETE:
4901 hci_disconn_phylink_complete_evt(hdev, skb);
4902 break;
4903
4904 case HCI_EV_NUM_COMP_BLOCKS:
4905 hci_num_comp_blocks_evt(hdev, skb);
4906 break;
4907
4908 default:
4909 BT_DBG("%s event 0x%2.2x", hdev->name, event);
4910 break;
4911 }
4912
4913 kfree_skb(skb);
4914 hdev->stat.evt_rx++;
4915 }
This page took 0.225506 seconds and 6 git commands to generate.