Bluetooth: Fix sending Read Remote Extended Features command
[deliverable/linux.git] / net / bluetooth / hci_event.c
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 /* Bluetooth HCI event handling. */
26
27 #include <asm/unaligned.h>
28
29 #include <net/bluetooth/bluetooth.h>
30 #include <net/bluetooth/hci_core.h>
31 #include <net/bluetooth/mgmt.h>
32
33 #include "hci_request.h"
34 #include "hci_debugfs.h"
35 #include "a2mp.h"
36 #include "amp.h"
37 #include "smp.h"
38
39 #define ZERO_KEY "\x00\x00\x00\x00\x00\x00\x00\x00" \
40 "\x00\x00\x00\x00\x00\x00\x00\x00"
41
42 /* Handle HCI Event packets */
43
44 static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb)
45 {
46 __u8 status = *((__u8 *) skb->data);
47
48 BT_DBG("%s status 0x%2.2x", hdev->name, status);
49
50 if (status)
51 return;
52
53 clear_bit(HCI_INQUIRY, &hdev->flags);
54 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
55 wake_up_bit(&hdev->flags, HCI_INQUIRY);
56
57 hci_dev_lock(hdev);
58 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
59 hci_dev_unlock(hdev);
60
61 hci_conn_check_pending(hdev);
62 }
63
64 static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
65 {
66 __u8 status = *((__u8 *) skb->data);
67
68 BT_DBG("%s status 0x%2.2x", hdev->name, status);
69
70 if (status)
71 return;
72
73 set_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
74 }
75
76 static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
77 {
78 __u8 status = *((__u8 *) skb->data);
79
80 BT_DBG("%s status 0x%2.2x", hdev->name, status);
81
82 if (status)
83 return;
84
85 clear_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
86
87 hci_conn_check_pending(hdev);
88 }
89
90 static void hci_cc_remote_name_req_cancel(struct hci_dev *hdev,
91 struct sk_buff *skb)
92 {
93 BT_DBG("%s", hdev->name);
94 }
95
96 static void hci_cc_role_discovery(struct hci_dev *hdev, struct sk_buff *skb)
97 {
98 struct hci_rp_role_discovery *rp = (void *) skb->data;
99 struct hci_conn *conn;
100
101 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
102
103 if (rp->status)
104 return;
105
106 hci_dev_lock(hdev);
107
108 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
109 if (conn)
110 conn->role = rp->role;
111
112 hci_dev_unlock(hdev);
113 }
114
115 static void hci_cc_read_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
116 {
117 struct hci_rp_read_link_policy *rp = (void *) skb->data;
118 struct hci_conn *conn;
119
120 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
121
122 if (rp->status)
123 return;
124
125 hci_dev_lock(hdev);
126
127 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
128 if (conn)
129 conn->link_policy = __le16_to_cpu(rp->policy);
130
131 hci_dev_unlock(hdev);
132 }
133
134 static void hci_cc_write_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
135 {
136 struct hci_rp_write_link_policy *rp = (void *) skb->data;
137 struct hci_conn *conn;
138 void *sent;
139
140 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
141
142 if (rp->status)
143 return;
144
145 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LINK_POLICY);
146 if (!sent)
147 return;
148
149 hci_dev_lock(hdev);
150
151 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
152 if (conn)
153 conn->link_policy = get_unaligned_le16(sent + 2);
154
155 hci_dev_unlock(hdev);
156 }
157
158 static void hci_cc_read_def_link_policy(struct hci_dev *hdev,
159 struct sk_buff *skb)
160 {
161 struct hci_rp_read_def_link_policy *rp = (void *) skb->data;
162
163 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
164
165 if (rp->status)
166 return;
167
168 hdev->link_policy = __le16_to_cpu(rp->policy);
169 }
170
171 static void hci_cc_write_def_link_policy(struct hci_dev *hdev,
172 struct sk_buff *skb)
173 {
174 __u8 status = *((__u8 *) skb->data);
175 void *sent;
176
177 BT_DBG("%s status 0x%2.2x", hdev->name, status);
178
179 if (status)
180 return;
181
182 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_DEF_LINK_POLICY);
183 if (!sent)
184 return;
185
186 hdev->link_policy = get_unaligned_le16(sent);
187 }
188
189 static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb)
190 {
191 __u8 status = *((__u8 *) skb->data);
192
193 BT_DBG("%s status 0x%2.2x", hdev->name, status);
194
195 clear_bit(HCI_RESET, &hdev->flags);
196
197 if (status)
198 return;
199
200 /* Reset all non-persistent flags */
201 hdev->dev_flags &= ~HCI_PERSISTENT_MASK;
202
203 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
204
205 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
206 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
207
208 memset(hdev->adv_data, 0, sizeof(hdev->adv_data));
209 hdev->adv_data_len = 0;
210
211 memset(hdev->scan_rsp_data, 0, sizeof(hdev->scan_rsp_data));
212 hdev->scan_rsp_data_len = 0;
213
214 hdev->le_scan_type = LE_SCAN_PASSIVE;
215
216 hdev->ssp_debug_mode = 0;
217
218 hci_bdaddr_list_clear(&hdev->le_white_list);
219 }
220
221 static void hci_cc_read_stored_link_key(struct hci_dev *hdev,
222 struct sk_buff *skb)
223 {
224 struct hci_rp_read_stored_link_key *rp = (void *)skb->data;
225 struct hci_cp_read_stored_link_key *sent;
226
227 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
228
229 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_STORED_LINK_KEY);
230 if (!sent)
231 return;
232
233 if (!rp->status && sent->read_all == 0x01) {
234 hdev->stored_max_keys = rp->max_keys;
235 hdev->stored_num_keys = rp->num_keys;
236 }
237 }
238
239 static void hci_cc_delete_stored_link_key(struct hci_dev *hdev,
240 struct sk_buff *skb)
241 {
242 struct hci_rp_delete_stored_link_key *rp = (void *)skb->data;
243
244 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
245
246 if (rp->status)
247 return;
248
249 if (rp->num_keys <= hdev->stored_num_keys)
250 hdev->stored_num_keys -= rp->num_keys;
251 else
252 hdev->stored_num_keys = 0;
253 }
254
255 static void hci_cc_write_local_name(struct hci_dev *hdev, struct sk_buff *skb)
256 {
257 __u8 status = *((__u8 *) skb->data);
258 void *sent;
259
260 BT_DBG("%s status 0x%2.2x", hdev->name, status);
261
262 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LOCAL_NAME);
263 if (!sent)
264 return;
265
266 hci_dev_lock(hdev);
267
268 if (test_bit(HCI_MGMT, &hdev->dev_flags))
269 mgmt_set_local_name_complete(hdev, sent, status);
270 else if (!status)
271 memcpy(hdev->dev_name, sent, HCI_MAX_NAME_LENGTH);
272
273 hci_dev_unlock(hdev);
274 }
275
276 static void hci_cc_read_local_name(struct hci_dev *hdev, struct sk_buff *skb)
277 {
278 struct hci_rp_read_local_name *rp = (void *) skb->data;
279
280 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
281
282 if (rp->status)
283 return;
284
285 if (test_bit(HCI_SETUP, &hdev->dev_flags) ||
286 test_bit(HCI_CONFIG, &hdev->dev_flags))
287 memcpy(hdev->dev_name, rp->name, HCI_MAX_NAME_LENGTH);
288 }
289
290 static void hci_cc_write_auth_enable(struct hci_dev *hdev, struct sk_buff *skb)
291 {
292 __u8 status = *((__u8 *) skb->data);
293 void *sent;
294
295 BT_DBG("%s status 0x%2.2x", hdev->name, status);
296
297 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_AUTH_ENABLE);
298 if (!sent)
299 return;
300
301 hci_dev_lock(hdev);
302
303 if (!status) {
304 __u8 param = *((__u8 *) sent);
305
306 if (param == AUTH_ENABLED)
307 set_bit(HCI_AUTH, &hdev->flags);
308 else
309 clear_bit(HCI_AUTH, &hdev->flags);
310 }
311
312 if (test_bit(HCI_MGMT, &hdev->dev_flags))
313 mgmt_auth_enable_complete(hdev, status);
314
315 hci_dev_unlock(hdev);
316 }
317
318 static void hci_cc_write_encrypt_mode(struct hci_dev *hdev, struct sk_buff *skb)
319 {
320 __u8 status = *((__u8 *) skb->data);
321 __u8 param;
322 void *sent;
323
324 BT_DBG("%s status 0x%2.2x", hdev->name, status);
325
326 if (status)
327 return;
328
329 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_ENCRYPT_MODE);
330 if (!sent)
331 return;
332
333 param = *((__u8 *) sent);
334
335 if (param)
336 set_bit(HCI_ENCRYPT, &hdev->flags);
337 else
338 clear_bit(HCI_ENCRYPT, &hdev->flags);
339 }
340
341 static void hci_cc_write_scan_enable(struct hci_dev *hdev, struct sk_buff *skb)
342 {
343 __u8 status = *((__u8 *) skb->data);
344 __u8 param;
345 void *sent;
346
347 BT_DBG("%s status 0x%2.2x", hdev->name, status);
348
349 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SCAN_ENABLE);
350 if (!sent)
351 return;
352
353 param = *((__u8 *) sent);
354
355 hci_dev_lock(hdev);
356
357 if (status) {
358 hdev->discov_timeout = 0;
359 goto done;
360 }
361
362 if (param & SCAN_INQUIRY)
363 set_bit(HCI_ISCAN, &hdev->flags);
364 else
365 clear_bit(HCI_ISCAN, &hdev->flags);
366
367 if (param & SCAN_PAGE)
368 set_bit(HCI_PSCAN, &hdev->flags);
369 else
370 clear_bit(HCI_PSCAN, &hdev->flags);
371
372 done:
373 hci_dev_unlock(hdev);
374 }
375
376 static void hci_cc_read_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
377 {
378 struct hci_rp_read_class_of_dev *rp = (void *) skb->data;
379
380 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
381
382 if (rp->status)
383 return;
384
385 memcpy(hdev->dev_class, rp->dev_class, 3);
386
387 BT_DBG("%s class 0x%.2x%.2x%.2x", hdev->name,
388 hdev->dev_class[2], hdev->dev_class[1], hdev->dev_class[0]);
389 }
390
391 static void hci_cc_write_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
392 {
393 __u8 status = *((__u8 *) skb->data);
394 void *sent;
395
396 BT_DBG("%s status 0x%2.2x", hdev->name, status);
397
398 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_CLASS_OF_DEV);
399 if (!sent)
400 return;
401
402 hci_dev_lock(hdev);
403
404 if (status == 0)
405 memcpy(hdev->dev_class, sent, 3);
406
407 if (test_bit(HCI_MGMT, &hdev->dev_flags))
408 mgmt_set_class_of_dev_complete(hdev, sent, status);
409
410 hci_dev_unlock(hdev);
411 }
412
413 static void hci_cc_read_voice_setting(struct hci_dev *hdev, struct sk_buff *skb)
414 {
415 struct hci_rp_read_voice_setting *rp = (void *) skb->data;
416 __u16 setting;
417
418 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
419
420 if (rp->status)
421 return;
422
423 setting = __le16_to_cpu(rp->voice_setting);
424
425 if (hdev->voice_setting == setting)
426 return;
427
428 hdev->voice_setting = setting;
429
430 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
431
432 if (hdev->notify)
433 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
434 }
435
436 static void hci_cc_write_voice_setting(struct hci_dev *hdev,
437 struct sk_buff *skb)
438 {
439 __u8 status = *((__u8 *) skb->data);
440 __u16 setting;
441 void *sent;
442
443 BT_DBG("%s status 0x%2.2x", hdev->name, status);
444
445 if (status)
446 return;
447
448 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_VOICE_SETTING);
449 if (!sent)
450 return;
451
452 setting = get_unaligned_le16(sent);
453
454 if (hdev->voice_setting == setting)
455 return;
456
457 hdev->voice_setting = setting;
458
459 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
460
461 if (hdev->notify)
462 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
463 }
464
465 static void hci_cc_read_num_supported_iac(struct hci_dev *hdev,
466 struct sk_buff *skb)
467 {
468 struct hci_rp_read_num_supported_iac *rp = (void *) skb->data;
469
470 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
471
472 if (rp->status)
473 return;
474
475 hdev->num_iac = rp->num_iac;
476
477 BT_DBG("%s num iac %d", hdev->name, hdev->num_iac);
478 }
479
480 static void hci_cc_write_ssp_mode(struct hci_dev *hdev, struct sk_buff *skb)
481 {
482 __u8 status = *((__u8 *) skb->data);
483 struct hci_cp_write_ssp_mode *sent;
484
485 BT_DBG("%s status 0x%2.2x", hdev->name, status);
486
487 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_MODE);
488 if (!sent)
489 return;
490
491 hci_dev_lock(hdev);
492
493 if (!status) {
494 if (sent->mode)
495 hdev->features[1][0] |= LMP_HOST_SSP;
496 else
497 hdev->features[1][0] &= ~LMP_HOST_SSP;
498 }
499
500 if (test_bit(HCI_MGMT, &hdev->dev_flags))
501 mgmt_ssp_enable_complete(hdev, sent->mode, status);
502 else if (!status) {
503 if (sent->mode)
504 set_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
505 else
506 clear_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
507 }
508
509 hci_dev_unlock(hdev);
510 }
511
512 static void hci_cc_write_sc_support(struct hci_dev *hdev, struct sk_buff *skb)
513 {
514 u8 status = *((u8 *) skb->data);
515 struct hci_cp_write_sc_support *sent;
516
517 BT_DBG("%s status 0x%2.2x", hdev->name, status);
518
519 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SC_SUPPORT);
520 if (!sent)
521 return;
522
523 hci_dev_lock(hdev);
524
525 if (!status) {
526 if (sent->support)
527 hdev->features[1][0] |= LMP_HOST_SC;
528 else
529 hdev->features[1][0] &= ~LMP_HOST_SC;
530 }
531
532 if (!test_bit(HCI_MGMT, &hdev->dev_flags) && !status) {
533 if (sent->support)
534 set_bit(HCI_SC_ENABLED, &hdev->dev_flags);
535 else
536 clear_bit(HCI_SC_ENABLED, &hdev->dev_flags);
537 }
538
539 hci_dev_unlock(hdev);
540 }
541
542 static void hci_cc_read_local_version(struct hci_dev *hdev, struct sk_buff *skb)
543 {
544 struct hci_rp_read_local_version *rp = (void *) skb->data;
545
546 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
547
548 if (rp->status)
549 return;
550
551 if (test_bit(HCI_SETUP, &hdev->dev_flags) ||
552 test_bit(HCI_CONFIG, &hdev->dev_flags)) {
553 hdev->hci_ver = rp->hci_ver;
554 hdev->hci_rev = __le16_to_cpu(rp->hci_rev);
555 hdev->lmp_ver = rp->lmp_ver;
556 hdev->manufacturer = __le16_to_cpu(rp->manufacturer);
557 hdev->lmp_subver = __le16_to_cpu(rp->lmp_subver);
558 }
559 }
560
561 static void hci_cc_read_local_commands(struct hci_dev *hdev,
562 struct sk_buff *skb)
563 {
564 struct hci_rp_read_local_commands *rp = (void *) skb->data;
565
566 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
567
568 if (rp->status)
569 return;
570
571 if (test_bit(HCI_SETUP, &hdev->dev_flags) ||
572 test_bit(HCI_CONFIG, &hdev->dev_flags))
573 memcpy(hdev->commands, rp->commands, sizeof(hdev->commands));
574 }
575
576 static void hci_cc_read_local_features(struct hci_dev *hdev,
577 struct sk_buff *skb)
578 {
579 struct hci_rp_read_local_features *rp = (void *) skb->data;
580
581 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
582
583 if (rp->status)
584 return;
585
586 memcpy(hdev->features, rp->features, 8);
587
588 /* Adjust default settings according to features
589 * supported by device. */
590
591 if (hdev->features[0][0] & LMP_3SLOT)
592 hdev->pkt_type |= (HCI_DM3 | HCI_DH3);
593
594 if (hdev->features[0][0] & LMP_5SLOT)
595 hdev->pkt_type |= (HCI_DM5 | HCI_DH5);
596
597 if (hdev->features[0][1] & LMP_HV2) {
598 hdev->pkt_type |= (HCI_HV2);
599 hdev->esco_type |= (ESCO_HV2);
600 }
601
602 if (hdev->features[0][1] & LMP_HV3) {
603 hdev->pkt_type |= (HCI_HV3);
604 hdev->esco_type |= (ESCO_HV3);
605 }
606
607 if (lmp_esco_capable(hdev))
608 hdev->esco_type |= (ESCO_EV3);
609
610 if (hdev->features[0][4] & LMP_EV4)
611 hdev->esco_type |= (ESCO_EV4);
612
613 if (hdev->features[0][4] & LMP_EV5)
614 hdev->esco_type |= (ESCO_EV5);
615
616 if (hdev->features[0][5] & LMP_EDR_ESCO_2M)
617 hdev->esco_type |= (ESCO_2EV3);
618
619 if (hdev->features[0][5] & LMP_EDR_ESCO_3M)
620 hdev->esco_type |= (ESCO_3EV3);
621
622 if (hdev->features[0][5] & LMP_EDR_3S_ESCO)
623 hdev->esco_type |= (ESCO_2EV5 | ESCO_3EV5);
624 }
625
626 static void hci_cc_read_local_ext_features(struct hci_dev *hdev,
627 struct sk_buff *skb)
628 {
629 struct hci_rp_read_local_ext_features *rp = (void *) skb->data;
630
631 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
632
633 if (rp->status)
634 return;
635
636 if (hdev->max_page < rp->max_page)
637 hdev->max_page = rp->max_page;
638
639 if (rp->page < HCI_MAX_PAGES)
640 memcpy(hdev->features[rp->page], rp->features, 8);
641 }
642
643 static void hci_cc_read_flow_control_mode(struct hci_dev *hdev,
644 struct sk_buff *skb)
645 {
646 struct hci_rp_read_flow_control_mode *rp = (void *) skb->data;
647
648 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
649
650 if (rp->status)
651 return;
652
653 hdev->flow_ctl_mode = rp->mode;
654 }
655
656 static void hci_cc_read_buffer_size(struct hci_dev *hdev, struct sk_buff *skb)
657 {
658 struct hci_rp_read_buffer_size *rp = (void *) skb->data;
659
660 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
661
662 if (rp->status)
663 return;
664
665 hdev->acl_mtu = __le16_to_cpu(rp->acl_mtu);
666 hdev->sco_mtu = rp->sco_mtu;
667 hdev->acl_pkts = __le16_to_cpu(rp->acl_max_pkt);
668 hdev->sco_pkts = __le16_to_cpu(rp->sco_max_pkt);
669
670 if (test_bit(HCI_QUIRK_FIXUP_BUFFER_SIZE, &hdev->quirks)) {
671 hdev->sco_mtu = 64;
672 hdev->sco_pkts = 8;
673 }
674
675 hdev->acl_cnt = hdev->acl_pkts;
676 hdev->sco_cnt = hdev->sco_pkts;
677
678 BT_DBG("%s acl mtu %d:%d sco mtu %d:%d", hdev->name, hdev->acl_mtu,
679 hdev->acl_pkts, hdev->sco_mtu, hdev->sco_pkts);
680 }
681
682 static void hci_cc_read_bd_addr(struct hci_dev *hdev, struct sk_buff *skb)
683 {
684 struct hci_rp_read_bd_addr *rp = (void *) skb->data;
685
686 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
687
688 if (rp->status)
689 return;
690
691 if (test_bit(HCI_INIT, &hdev->flags))
692 bacpy(&hdev->bdaddr, &rp->bdaddr);
693
694 if (test_bit(HCI_SETUP, &hdev->dev_flags))
695 bacpy(&hdev->setup_addr, &rp->bdaddr);
696 }
697
698 static void hci_cc_read_page_scan_activity(struct hci_dev *hdev,
699 struct sk_buff *skb)
700 {
701 struct hci_rp_read_page_scan_activity *rp = (void *) skb->data;
702
703 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
704
705 if (rp->status)
706 return;
707
708 if (test_bit(HCI_INIT, &hdev->flags)) {
709 hdev->page_scan_interval = __le16_to_cpu(rp->interval);
710 hdev->page_scan_window = __le16_to_cpu(rp->window);
711 }
712 }
713
714 static void hci_cc_write_page_scan_activity(struct hci_dev *hdev,
715 struct sk_buff *skb)
716 {
717 u8 status = *((u8 *) skb->data);
718 struct hci_cp_write_page_scan_activity *sent;
719
720 BT_DBG("%s status 0x%2.2x", hdev->name, status);
721
722 if (status)
723 return;
724
725 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY);
726 if (!sent)
727 return;
728
729 hdev->page_scan_interval = __le16_to_cpu(sent->interval);
730 hdev->page_scan_window = __le16_to_cpu(sent->window);
731 }
732
733 static void hci_cc_read_page_scan_type(struct hci_dev *hdev,
734 struct sk_buff *skb)
735 {
736 struct hci_rp_read_page_scan_type *rp = (void *) skb->data;
737
738 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
739
740 if (rp->status)
741 return;
742
743 if (test_bit(HCI_INIT, &hdev->flags))
744 hdev->page_scan_type = rp->type;
745 }
746
747 static void hci_cc_write_page_scan_type(struct hci_dev *hdev,
748 struct sk_buff *skb)
749 {
750 u8 status = *((u8 *) skb->data);
751 u8 *type;
752
753 BT_DBG("%s status 0x%2.2x", hdev->name, status);
754
755 if (status)
756 return;
757
758 type = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_TYPE);
759 if (type)
760 hdev->page_scan_type = *type;
761 }
762
763 static void hci_cc_read_data_block_size(struct hci_dev *hdev,
764 struct sk_buff *skb)
765 {
766 struct hci_rp_read_data_block_size *rp = (void *) skb->data;
767
768 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
769
770 if (rp->status)
771 return;
772
773 hdev->block_mtu = __le16_to_cpu(rp->max_acl_len);
774 hdev->block_len = __le16_to_cpu(rp->block_len);
775 hdev->num_blocks = __le16_to_cpu(rp->num_blocks);
776
777 hdev->block_cnt = hdev->num_blocks;
778
779 BT_DBG("%s blk mtu %d cnt %d len %d", hdev->name, hdev->block_mtu,
780 hdev->block_cnt, hdev->block_len);
781 }
782
783 static void hci_cc_read_clock(struct hci_dev *hdev, struct sk_buff *skb)
784 {
785 struct hci_rp_read_clock *rp = (void *) skb->data;
786 struct hci_cp_read_clock *cp;
787 struct hci_conn *conn;
788
789 BT_DBG("%s", hdev->name);
790
791 if (skb->len < sizeof(*rp))
792 return;
793
794 if (rp->status)
795 return;
796
797 hci_dev_lock(hdev);
798
799 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_CLOCK);
800 if (!cp)
801 goto unlock;
802
803 if (cp->which == 0x00) {
804 hdev->clock = le32_to_cpu(rp->clock);
805 goto unlock;
806 }
807
808 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
809 if (conn) {
810 conn->clock = le32_to_cpu(rp->clock);
811 conn->clock_accuracy = le16_to_cpu(rp->accuracy);
812 }
813
814 unlock:
815 hci_dev_unlock(hdev);
816 }
817
818 static void hci_cc_read_local_amp_info(struct hci_dev *hdev,
819 struct sk_buff *skb)
820 {
821 struct hci_rp_read_local_amp_info *rp = (void *) skb->data;
822
823 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
824
825 if (rp->status)
826 goto a2mp_rsp;
827
828 hdev->amp_status = rp->amp_status;
829 hdev->amp_total_bw = __le32_to_cpu(rp->total_bw);
830 hdev->amp_max_bw = __le32_to_cpu(rp->max_bw);
831 hdev->amp_min_latency = __le32_to_cpu(rp->min_latency);
832 hdev->amp_max_pdu = __le32_to_cpu(rp->max_pdu);
833 hdev->amp_type = rp->amp_type;
834 hdev->amp_pal_cap = __le16_to_cpu(rp->pal_cap);
835 hdev->amp_assoc_size = __le16_to_cpu(rp->max_assoc_size);
836 hdev->amp_be_flush_to = __le32_to_cpu(rp->be_flush_to);
837 hdev->amp_max_flush_to = __le32_to_cpu(rp->max_flush_to);
838
839 a2mp_rsp:
840 a2mp_send_getinfo_rsp(hdev);
841 }
842
843 static void hci_cc_read_local_amp_assoc(struct hci_dev *hdev,
844 struct sk_buff *skb)
845 {
846 struct hci_rp_read_local_amp_assoc *rp = (void *) skb->data;
847 struct amp_assoc *assoc = &hdev->loc_assoc;
848 size_t rem_len, frag_len;
849
850 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
851
852 if (rp->status)
853 goto a2mp_rsp;
854
855 frag_len = skb->len - sizeof(*rp);
856 rem_len = __le16_to_cpu(rp->rem_len);
857
858 if (rem_len > frag_len) {
859 BT_DBG("frag_len %zu rem_len %zu", frag_len, rem_len);
860
861 memcpy(assoc->data + assoc->offset, rp->frag, frag_len);
862 assoc->offset += frag_len;
863
864 /* Read other fragments */
865 amp_read_loc_assoc_frag(hdev, rp->phy_handle);
866
867 return;
868 }
869
870 memcpy(assoc->data + assoc->offset, rp->frag, rem_len);
871 assoc->len = assoc->offset + rem_len;
872 assoc->offset = 0;
873
874 a2mp_rsp:
875 /* Send A2MP Rsp when all fragments are received */
876 a2mp_send_getampassoc_rsp(hdev, rp->status);
877 a2mp_send_create_phy_link_req(hdev, rp->status);
878 }
879
880 static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev,
881 struct sk_buff *skb)
882 {
883 struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data;
884
885 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
886
887 if (rp->status)
888 return;
889
890 hdev->inq_tx_power = rp->tx_power;
891 }
892
893 static void hci_cc_pin_code_reply(struct hci_dev *hdev, struct sk_buff *skb)
894 {
895 struct hci_rp_pin_code_reply *rp = (void *) skb->data;
896 struct hci_cp_pin_code_reply *cp;
897 struct hci_conn *conn;
898
899 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
900
901 hci_dev_lock(hdev);
902
903 if (test_bit(HCI_MGMT, &hdev->dev_flags))
904 mgmt_pin_code_reply_complete(hdev, &rp->bdaddr, rp->status);
905
906 if (rp->status)
907 goto unlock;
908
909 cp = hci_sent_cmd_data(hdev, HCI_OP_PIN_CODE_REPLY);
910 if (!cp)
911 goto unlock;
912
913 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
914 if (conn)
915 conn->pin_length = cp->pin_len;
916
917 unlock:
918 hci_dev_unlock(hdev);
919 }
920
921 static void hci_cc_pin_code_neg_reply(struct hci_dev *hdev, struct sk_buff *skb)
922 {
923 struct hci_rp_pin_code_neg_reply *rp = (void *) skb->data;
924
925 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
926
927 hci_dev_lock(hdev);
928
929 if (test_bit(HCI_MGMT, &hdev->dev_flags))
930 mgmt_pin_code_neg_reply_complete(hdev, &rp->bdaddr,
931 rp->status);
932
933 hci_dev_unlock(hdev);
934 }
935
936 static void hci_cc_le_read_buffer_size(struct hci_dev *hdev,
937 struct sk_buff *skb)
938 {
939 struct hci_rp_le_read_buffer_size *rp = (void *) skb->data;
940
941 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
942
943 if (rp->status)
944 return;
945
946 hdev->le_mtu = __le16_to_cpu(rp->le_mtu);
947 hdev->le_pkts = rp->le_max_pkt;
948
949 hdev->le_cnt = hdev->le_pkts;
950
951 BT_DBG("%s le mtu %d:%d", hdev->name, hdev->le_mtu, hdev->le_pkts);
952 }
953
954 static void hci_cc_le_read_local_features(struct hci_dev *hdev,
955 struct sk_buff *skb)
956 {
957 struct hci_rp_le_read_local_features *rp = (void *) skb->data;
958
959 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
960
961 if (rp->status)
962 return;
963
964 memcpy(hdev->le_features, rp->features, 8);
965 }
966
967 static void hci_cc_le_read_adv_tx_power(struct hci_dev *hdev,
968 struct sk_buff *skb)
969 {
970 struct hci_rp_le_read_adv_tx_power *rp = (void *) skb->data;
971
972 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
973
974 if (rp->status)
975 return;
976
977 hdev->adv_tx_power = rp->tx_power;
978 }
979
980 static void hci_cc_user_confirm_reply(struct hci_dev *hdev, struct sk_buff *skb)
981 {
982 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
983
984 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
985
986 hci_dev_lock(hdev);
987
988 if (test_bit(HCI_MGMT, &hdev->dev_flags))
989 mgmt_user_confirm_reply_complete(hdev, &rp->bdaddr, ACL_LINK, 0,
990 rp->status);
991
992 hci_dev_unlock(hdev);
993 }
994
995 static void hci_cc_user_confirm_neg_reply(struct hci_dev *hdev,
996 struct sk_buff *skb)
997 {
998 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
999
1000 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1001
1002 hci_dev_lock(hdev);
1003
1004 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1005 mgmt_user_confirm_neg_reply_complete(hdev, &rp->bdaddr,
1006 ACL_LINK, 0, rp->status);
1007
1008 hci_dev_unlock(hdev);
1009 }
1010
1011 static void hci_cc_user_passkey_reply(struct hci_dev *hdev, struct sk_buff *skb)
1012 {
1013 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
1014
1015 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1016
1017 hci_dev_lock(hdev);
1018
1019 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1020 mgmt_user_passkey_reply_complete(hdev, &rp->bdaddr, ACL_LINK,
1021 0, rp->status);
1022
1023 hci_dev_unlock(hdev);
1024 }
1025
1026 static void hci_cc_user_passkey_neg_reply(struct hci_dev *hdev,
1027 struct sk_buff *skb)
1028 {
1029 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
1030
1031 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1032
1033 hci_dev_lock(hdev);
1034
1035 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1036 mgmt_user_passkey_neg_reply_complete(hdev, &rp->bdaddr,
1037 ACL_LINK, 0, rp->status);
1038
1039 hci_dev_unlock(hdev);
1040 }
1041
1042 static void hci_cc_read_local_oob_data(struct hci_dev *hdev,
1043 struct sk_buff *skb)
1044 {
1045 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
1046
1047 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1048
1049 hci_dev_lock(hdev);
1050 mgmt_read_local_oob_data_complete(hdev, rp->hash, rp->rand, NULL, NULL,
1051 rp->status);
1052 hci_dev_unlock(hdev);
1053 }
1054
1055 static void hci_cc_read_local_oob_ext_data(struct hci_dev *hdev,
1056 struct sk_buff *skb)
1057 {
1058 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
1059
1060 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1061
1062 hci_dev_lock(hdev);
1063 mgmt_read_local_oob_data_complete(hdev, rp->hash192, rp->rand192,
1064 rp->hash256, rp->rand256,
1065 rp->status);
1066 hci_dev_unlock(hdev);
1067 }
1068
1069
1070 static void hci_cc_le_set_random_addr(struct hci_dev *hdev, struct sk_buff *skb)
1071 {
1072 __u8 status = *((__u8 *) skb->data);
1073 bdaddr_t *sent;
1074
1075 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1076
1077 if (status)
1078 return;
1079
1080 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_RANDOM_ADDR);
1081 if (!sent)
1082 return;
1083
1084 hci_dev_lock(hdev);
1085
1086 bacpy(&hdev->random_addr, sent);
1087
1088 hci_dev_unlock(hdev);
1089 }
1090
1091 static void hci_cc_le_set_adv_enable(struct hci_dev *hdev, struct sk_buff *skb)
1092 {
1093 __u8 *sent, status = *((__u8 *) skb->data);
1094
1095 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1096
1097 if (status)
1098 return;
1099
1100 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_ENABLE);
1101 if (!sent)
1102 return;
1103
1104 hci_dev_lock(hdev);
1105
1106 /* If we're doing connection initiation as peripheral. Set a
1107 * timeout in case something goes wrong.
1108 */
1109 if (*sent) {
1110 struct hci_conn *conn;
1111
1112 set_bit(HCI_LE_ADV, &hdev->dev_flags);
1113
1114 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
1115 if (conn)
1116 queue_delayed_work(hdev->workqueue,
1117 &conn->le_conn_timeout,
1118 conn->conn_timeout);
1119 } else {
1120 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
1121 }
1122
1123 hci_dev_unlock(hdev);
1124 }
1125
1126 static void hci_cc_le_set_scan_param(struct hci_dev *hdev, struct sk_buff *skb)
1127 {
1128 struct hci_cp_le_set_scan_param *cp;
1129 __u8 status = *((__u8 *) skb->data);
1130
1131 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1132
1133 if (status)
1134 return;
1135
1136 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_PARAM);
1137 if (!cp)
1138 return;
1139
1140 hci_dev_lock(hdev);
1141
1142 hdev->le_scan_type = cp->type;
1143
1144 hci_dev_unlock(hdev);
1145 }
1146
1147 static bool has_pending_adv_report(struct hci_dev *hdev)
1148 {
1149 struct discovery_state *d = &hdev->discovery;
1150
1151 return bacmp(&d->last_adv_addr, BDADDR_ANY);
1152 }
1153
1154 static void clear_pending_adv_report(struct hci_dev *hdev)
1155 {
1156 struct discovery_state *d = &hdev->discovery;
1157
1158 bacpy(&d->last_adv_addr, BDADDR_ANY);
1159 d->last_adv_data_len = 0;
1160 }
1161
1162 static void store_pending_adv_report(struct hci_dev *hdev, bdaddr_t *bdaddr,
1163 u8 bdaddr_type, s8 rssi, u32 flags,
1164 u8 *data, u8 len)
1165 {
1166 struct discovery_state *d = &hdev->discovery;
1167
1168 bacpy(&d->last_adv_addr, bdaddr);
1169 d->last_adv_addr_type = bdaddr_type;
1170 d->last_adv_rssi = rssi;
1171 d->last_adv_flags = flags;
1172 memcpy(d->last_adv_data, data, len);
1173 d->last_adv_data_len = len;
1174 }
1175
1176 static void hci_cc_le_set_scan_enable(struct hci_dev *hdev,
1177 struct sk_buff *skb)
1178 {
1179 struct hci_cp_le_set_scan_enable *cp;
1180 __u8 status = *((__u8 *) skb->data);
1181
1182 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1183
1184 if (status)
1185 return;
1186
1187 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_ENABLE);
1188 if (!cp)
1189 return;
1190
1191 hci_dev_lock(hdev);
1192
1193 switch (cp->enable) {
1194 case LE_SCAN_ENABLE:
1195 set_bit(HCI_LE_SCAN, &hdev->dev_flags);
1196 if (hdev->le_scan_type == LE_SCAN_ACTIVE)
1197 clear_pending_adv_report(hdev);
1198 break;
1199
1200 case LE_SCAN_DISABLE:
1201 /* We do this here instead of when setting DISCOVERY_STOPPED
1202 * since the latter would potentially require waiting for
1203 * inquiry to stop too.
1204 */
1205 if (has_pending_adv_report(hdev)) {
1206 struct discovery_state *d = &hdev->discovery;
1207
1208 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
1209 d->last_adv_addr_type, NULL,
1210 d->last_adv_rssi, d->last_adv_flags,
1211 d->last_adv_data,
1212 d->last_adv_data_len, NULL, 0);
1213 }
1214
1215 /* Cancel this timer so that we don't try to disable scanning
1216 * when it's already disabled.
1217 */
1218 cancel_delayed_work(&hdev->le_scan_disable);
1219
1220 clear_bit(HCI_LE_SCAN, &hdev->dev_flags);
1221
1222 /* The HCI_LE_SCAN_INTERRUPTED flag indicates that we
1223 * interrupted scanning due to a connect request. Mark
1224 * therefore discovery as stopped. If this was not
1225 * because of a connect request advertising might have
1226 * been disabled because of active scanning, so
1227 * re-enable it again if necessary.
1228 */
1229 if (test_and_clear_bit(HCI_LE_SCAN_INTERRUPTED,
1230 &hdev->dev_flags))
1231 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1232 else if (!test_bit(HCI_LE_ADV, &hdev->dev_flags) &&
1233 hdev->discovery.state == DISCOVERY_FINDING)
1234 mgmt_reenable_advertising(hdev);
1235
1236 break;
1237
1238 default:
1239 BT_ERR("Used reserved LE_Scan_Enable param %d", cp->enable);
1240 break;
1241 }
1242
1243 hci_dev_unlock(hdev);
1244 }
1245
1246 static void hci_cc_le_read_white_list_size(struct hci_dev *hdev,
1247 struct sk_buff *skb)
1248 {
1249 struct hci_rp_le_read_white_list_size *rp = (void *) skb->data;
1250
1251 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1252
1253 if (rp->status)
1254 return;
1255
1256 hdev->le_white_list_size = rp->size;
1257 }
1258
1259 static void hci_cc_le_clear_white_list(struct hci_dev *hdev,
1260 struct sk_buff *skb)
1261 {
1262 __u8 status = *((__u8 *) skb->data);
1263
1264 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1265
1266 if (status)
1267 return;
1268
1269 hci_bdaddr_list_clear(&hdev->le_white_list);
1270 }
1271
1272 static void hci_cc_le_add_to_white_list(struct hci_dev *hdev,
1273 struct sk_buff *skb)
1274 {
1275 struct hci_cp_le_add_to_white_list *sent;
1276 __u8 status = *((__u8 *) skb->data);
1277
1278 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1279
1280 if (status)
1281 return;
1282
1283 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_WHITE_LIST);
1284 if (!sent)
1285 return;
1286
1287 hci_bdaddr_list_add(&hdev->le_white_list, &sent->bdaddr,
1288 sent->bdaddr_type);
1289 }
1290
1291 static void hci_cc_le_del_from_white_list(struct hci_dev *hdev,
1292 struct sk_buff *skb)
1293 {
1294 struct hci_cp_le_del_from_white_list *sent;
1295 __u8 status = *((__u8 *) skb->data);
1296
1297 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1298
1299 if (status)
1300 return;
1301
1302 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_WHITE_LIST);
1303 if (!sent)
1304 return;
1305
1306 hci_bdaddr_list_del(&hdev->le_white_list, &sent->bdaddr,
1307 sent->bdaddr_type);
1308 }
1309
1310 static void hci_cc_le_read_supported_states(struct hci_dev *hdev,
1311 struct sk_buff *skb)
1312 {
1313 struct hci_rp_le_read_supported_states *rp = (void *) skb->data;
1314
1315 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1316
1317 if (rp->status)
1318 return;
1319
1320 memcpy(hdev->le_states, rp->le_states, 8);
1321 }
1322
1323 static void hci_cc_le_read_def_data_len(struct hci_dev *hdev,
1324 struct sk_buff *skb)
1325 {
1326 struct hci_rp_le_read_def_data_len *rp = (void *) skb->data;
1327
1328 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1329
1330 if (rp->status)
1331 return;
1332
1333 hdev->le_def_tx_len = le16_to_cpu(rp->tx_len);
1334 hdev->le_def_tx_time = le16_to_cpu(rp->tx_time);
1335 }
1336
1337 static void hci_cc_le_write_def_data_len(struct hci_dev *hdev,
1338 struct sk_buff *skb)
1339 {
1340 struct hci_cp_le_write_def_data_len *sent;
1341 __u8 status = *((__u8 *) skb->data);
1342
1343 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1344
1345 if (status)
1346 return;
1347
1348 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_WRITE_DEF_DATA_LEN);
1349 if (!sent)
1350 return;
1351
1352 hdev->le_def_tx_len = le16_to_cpu(sent->tx_len);
1353 hdev->le_def_tx_time = le16_to_cpu(sent->tx_time);
1354 }
1355
1356 static void hci_cc_le_read_max_data_len(struct hci_dev *hdev,
1357 struct sk_buff *skb)
1358 {
1359 struct hci_rp_le_read_max_data_len *rp = (void *) skb->data;
1360
1361 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1362
1363 if (rp->status)
1364 return;
1365
1366 hdev->le_max_tx_len = le16_to_cpu(rp->tx_len);
1367 hdev->le_max_tx_time = le16_to_cpu(rp->tx_time);
1368 hdev->le_max_rx_len = le16_to_cpu(rp->rx_len);
1369 hdev->le_max_rx_time = le16_to_cpu(rp->rx_time);
1370 }
1371
1372 static void hci_cc_write_le_host_supported(struct hci_dev *hdev,
1373 struct sk_buff *skb)
1374 {
1375 struct hci_cp_write_le_host_supported *sent;
1376 __u8 status = *((__u8 *) skb->data);
1377
1378 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1379
1380 if (status)
1381 return;
1382
1383 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED);
1384 if (!sent)
1385 return;
1386
1387 hci_dev_lock(hdev);
1388
1389 if (sent->le) {
1390 hdev->features[1][0] |= LMP_HOST_LE;
1391 set_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1392 } else {
1393 hdev->features[1][0] &= ~LMP_HOST_LE;
1394 clear_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1395 clear_bit(HCI_ADVERTISING, &hdev->dev_flags);
1396 }
1397
1398 if (sent->simul)
1399 hdev->features[1][0] |= LMP_HOST_LE_BREDR;
1400 else
1401 hdev->features[1][0] &= ~LMP_HOST_LE_BREDR;
1402
1403 hci_dev_unlock(hdev);
1404 }
1405
1406 static void hci_cc_set_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1407 {
1408 struct hci_cp_le_set_adv_param *cp;
1409 u8 status = *((u8 *) skb->data);
1410
1411 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1412
1413 if (status)
1414 return;
1415
1416 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_PARAM);
1417 if (!cp)
1418 return;
1419
1420 hci_dev_lock(hdev);
1421 hdev->adv_addr_type = cp->own_address_type;
1422 hci_dev_unlock(hdev);
1423 }
1424
1425 static void hci_cc_write_remote_amp_assoc(struct hci_dev *hdev,
1426 struct sk_buff *skb)
1427 {
1428 struct hci_rp_write_remote_amp_assoc *rp = (void *) skb->data;
1429
1430 BT_DBG("%s status 0x%2.2x phy_handle 0x%2.2x",
1431 hdev->name, rp->status, rp->phy_handle);
1432
1433 if (rp->status)
1434 return;
1435
1436 amp_write_rem_assoc_continue(hdev, rp->phy_handle);
1437 }
1438
1439 static void hci_cc_read_rssi(struct hci_dev *hdev, struct sk_buff *skb)
1440 {
1441 struct hci_rp_read_rssi *rp = (void *) skb->data;
1442 struct hci_conn *conn;
1443
1444 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1445
1446 if (rp->status)
1447 return;
1448
1449 hci_dev_lock(hdev);
1450
1451 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1452 if (conn)
1453 conn->rssi = rp->rssi;
1454
1455 hci_dev_unlock(hdev);
1456 }
1457
1458 static void hci_cc_read_tx_power(struct hci_dev *hdev, struct sk_buff *skb)
1459 {
1460 struct hci_cp_read_tx_power *sent;
1461 struct hci_rp_read_tx_power *rp = (void *) skb->data;
1462 struct hci_conn *conn;
1463
1464 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1465
1466 if (rp->status)
1467 return;
1468
1469 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_TX_POWER);
1470 if (!sent)
1471 return;
1472
1473 hci_dev_lock(hdev);
1474
1475 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1476 if (!conn)
1477 goto unlock;
1478
1479 switch (sent->type) {
1480 case 0x00:
1481 conn->tx_power = rp->tx_power;
1482 break;
1483 case 0x01:
1484 conn->max_tx_power = rp->tx_power;
1485 break;
1486 }
1487
1488 unlock:
1489 hci_dev_unlock(hdev);
1490 }
1491
1492 static void hci_cs_inquiry(struct hci_dev *hdev, __u8 status)
1493 {
1494 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1495
1496 if (status) {
1497 hci_conn_check_pending(hdev);
1498 return;
1499 }
1500
1501 set_bit(HCI_INQUIRY, &hdev->flags);
1502 }
1503
1504 static void hci_cs_create_conn(struct hci_dev *hdev, __u8 status)
1505 {
1506 struct hci_cp_create_conn *cp;
1507 struct hci_conn *conn;
1508
1509 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1510
1511 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_CONN);
1512 if (!cp)
1513 return;
1514
1515 hci_dev_lock(hdev);
1516
1517 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1518
1519 BT_DBG("%s bdaddr %pMR hcon %p", hdev->name, &cp->bdaddr, conn);
1520
1521 if (status) {
1522 if (conn && conn->state == BT_CONNECT) {
1523 if (status != 0x0c || conn->attempt > 2) {
1524 conn->state = BT_CLOSED;
1525 hci_proto_connect_cfm(conn, status);
1526 hci_conn_del(conn);
1527 } else
1528 conn->state = BT_CONNECT2;
1529 }
1530 } else {
1531 if (!conn) {
1532 conn = hci_conn_add(hdev, ACL_LINK, &cp->bdaddr,
1533 HCI_ROLE_MASTER);
1534 if (!conn)
1535 BT_ERR("No memory for new connection");
1536 }
1537 }
1538
1539 hci_dev_unlock(hdev);
1540 }
1541
1542 static void hci_cs_add_sco(struct hci_dev *hdev, __u8 status)
1543 {
1544 struct hci_cp_add_sco *cp;
1545 struct hci_conn *acl, *sco;
1546 __u16 handle;
1547
1548 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1549
1550 if (!status)
1551 return;
1552
1553 cp = hci_sent_cmd_data(hdev, HCI_OP_ADD_SCO);
1554 if (!cp)
1555 return;
1556
1557 handle = __le16_to_cpu(cp->handle);
1558
1559 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1560
1561 hci_dev_lock(hdev);
1562
1563 acl = hci_conn_hash_lookup_handle(hdev, handle);
1564 if (acl) {
1565 sco = acl->link;
1566 if (sco) {
1567 sco->state = BT_CLOSED;
1568
1569 hci_proto_connect_cfm(sco, status);
1570 hci_conn_del(sco);
1571 }
1572 }
1573
1574 hci_dev_unlock(hdev);
1575 }
1576
1577 static void hci_cs_auth_requested(struct hci_dev *hdev, __u8 status)
1578 {
1579 struct hci_cp_auth_requested *cp;
1580 struct hci_conn *conn;
1581
1582 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1583
1584 if (!status)
1585 return;
1586
1587 cp = hci_sent_cmd_data(hdev, HCI_OP_AUTH_REQUESTED);
1588 if (!cp)
1589 return;
1590
1591 hci_dev_lock(hdev);
1592
1593 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1594 if (conn) {
1595 if (conn->state == BT_CONFIG) {
1596 hci_proto_connect_cfm(conn, status);
1597 hci_conn_drop(conn);
1598 }
1599 }
1600
1601 hci_dev_unlock(hdev);
1602 }
1603
1604 static void hci_cs_set_conn_encrypt(struct hci_dev *hdev, __u8 status)
1605 {
1606 struct hci_cp_set_conn_encrypt *cp;
1607 struct hci_conn *conn;
1608
1609 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1610
1611 if (!status)
1612 return;
1613
1614 cp = hci_sent_cmd_data(hdev, HCI_OP_SET_CONN_ENCRYPT);
1615 if (!cp)
1616 return;
1617
1618 hci_dev_lock(hdev);
1619
1620 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1621 if (conn) {
1622 if (conn->state == BT_CONFIG) {
1623 hci_proto_connect_cfm(conn, status);
1624 hci_conn_drop(conn);
1625 }
1626 }
1627
1628 hci_dev_unlock(hdev);
1629 }
1630
1631 static int hci_outgoing_auth_needed(struct hci_dev *hdev,
1632 struct hci_conn *conn)
1633 {
1634 if (conn->state != BT_CONFIG || !conn->out)
1635 return 0;
1636
1637 if (conn->pending_sec_level == BT_SECURITY_SDP)
1638 return 0;
1639
1640 /* Only request authentication for SSP connections or non-SSP
1641 * devices with sec_level MEDIUM or HIGH or if MITM protection
1642 * is requested.
1643 */
1644 if (!hci_conn_ssp_enabled(conn) && !(conn->auth_type & 0x01) &&
1645 conn->pending_sec_level != BT_SECURITY_FIPS &&
1646 conn->pending_sec_level != BT_SECURITY_HIGH &&
1647 conn->pending_sec_level != BT_SECURITY_MEDIUM)
1648 return 0;
1649
1650 return 1;
1651 }
1652
1653 static int hci_resolve_name(struct hci_dev *hdev,
1654 struct inquiry_entry *e)
1655 {
1656 struct hci_cp_remote_name_req cp;
1657
1658 memset(&cp, 0, sizeof(cp));
1659
1660 bacpy(&cp.bdaddr, &e->data.bdaddr);
1661 cp.pscan_rep_mode = e->data.pscan_rep_mode;
1662 cp.pscan_mode = e->data.pscan_mode;
1663 cp.clock_offset = e->data.clock_offset;
1664
1665 return hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
1666 }
1667
1668 static bool hci_resolve_next_name(struct hci_dev *hdev)
1669 {
1670 struct discovery_state *discov = &hdev->discovery;
1671 struct inquiry_entry *e;
1672
1673 if (list_empty(&discov->resolve))
1674 return false;
1675
1676 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1677 if (!e)
1678 return false;
1679
1680 if (hci_resolve_name(hdev, e) == 0) {
1681 e->name_state = NAME_PENDING;
1682 return true;
1683 }
1684
1685 return false;
1686 }
1687
1688 static void hci_check_pending_name(struct hci_dev *hdev, struct hci_conn *conn,
1689 bdaddr_t *bdaddr, u8 *name, u8 name_len)
1690 {
1691 struct discovery_state *discov = &hdev->discovery;
1692 struct inquiry_entry *e;
1693
1694 /* Update the mgmt connected state if necessary. Be careful with
1695 * conn objects that exist but are not (yet) connected however.
1696 * Only those in BT_CONFIG or BT_CONNECTED states can be
1697 * considered connected.
1698 */
1699 if (conn &&
1700 (conn->state == BT_CONFIG || conn->state == BT_CONNECTED) &&
1701 !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
1702 mgmt_device_connected(hdev, conn, 0, name, name_len);
1703
1704 if (discov->state == DISCOVERY_STOPPED)
1705 return;
1706
1707 if (discov->state == DISCOVERY_STOPPING)
1708 goto discov_complete;
1709
1710 if (discov->state != DISCOVERY_RESOLVING)
1711 return;
1712
1713 e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
1714 /* If the device was not found in a list of found devices names of which
1715 * are pending. there is no need to continue resolving a next name as it
1716 * will be done upon receiving another Remote Name Request Complete
1717 * Event */
1718 if (!e)
1719 return;
1720
1721 list_del(&e->list);
1722 if (name) {
1723 e->name_state = NAME_KNOWN;
1724 mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00,
1725 e->data.rssi, name, name_len);
1726 } else {
1727 e->name_state = NAME_NOT_KNOWN;
1728 }
1729
1730 if (hci_resolve_next_name(hdev))
1731 return;
1732
1733 discov_complete:
1734 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1735 }
1736
1737 static void hci_cs_remote_name_req(struct hci_dev *hdev, __u8 status)
1738 {
1739 struct hci_cp_remote_name_req *cp;
1740 struct hci_conn *conn;
1741
1742 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1743
1744 /* If successful wait for the name req complete event before
1745 * checking for the need to do authentication */
1746 if (!status)
1747 return;
1748
1749 cp = hci_sent_cmd_data(hdev, HCI_OP_REMOTE_NAME_REQ);
1750 if (!cp)
1751 return;
1752
1753 hci_dev_lock(hdev);
1754
1755 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1756
1757 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1758 hci_check_pending_name(hdev, conn, &cp->bdaddr, NULL, 0);
1759
1760 if (!conn)
1761 goto unlock;
1762
1763 if (!hci_outgoing_auth_needed(hdev, conn))
1764 goto unlock;
1765
1766 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
1767 struct hci_cp_auth_requested auth_cp;
1768
1769 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
1770
1771 auth_cp.handle = __cpu_to_le16(conn->handle);
1772 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED,
1773 sizeof(auth_cp), &auth_cp);
1774 }
1775
1776 unlock:
1777 hci_dev_unlock(hdev);
1778 }
1779
1780 static void hci_cs_read_remote_features(struct hci_dev *hdev, __u8 status)
1781 {
1782 struct hci_cp_read_remote_features *cp;
1783 struct hci_conn *conn;
1784
1785 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1786
1787 if (!status)
1788 return;
1789
1790 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_FEATURES);
1791 if (!cp)
1792 return;
1793
1794 hci_dev_lock(hdev);
1795
1796 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1797 if (conn) {
1798 if (conn->state == BT_CONFIG) {
1799 hci_proto_connect_cfm(conn, status);
1800 hci_conn_drop(conn);
1801 }
1802 }
1803
1804 hci_dev_unlock(hdev);
1805 }
1806
1807 static void hci_cs_read_remote_ext_features(struct hci_dev *hdev, __u8 status)
1808 {
1809 struct hci_cp_read_remote_ext_features *cp;
1810 struct hci_conn *conn;
1811
1812 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1813
1814 if (!status)
1815 return;
1816
1817 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES);
1818 if (!cp)
1819 return;
1820
1821 hci_dev_lock(hdev);
1822
1823 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1824 if (conn) {
1825 if (conn->state == BT_CONFIG) {
1826 hci_proto_connect_cfm(conn, status);
1827 hci_conn_drop(conn);
1828 }
1829 }
1830
1831 hci_dev_unlock(hdev);
1832 }
1833
1834 static void hci_cs_setup_sync_conn(struct hci_dev *hdev, __u8 status)
1835 {
1836 struct hci_cp_setup_sync_conn *cp;
1837 struct hci_conn *acl, *sco;
1838 __u16 handle;
1839
1840 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1841
1842 if (!status)
1843 return;
1844
1845 cp = hci_sent_cmd_data(hdev, HCI_OP_SETUP_SYNC_CONN);
1846 if (!cp)
1847 return;
1848
1849 handle = __le16_to_cpu(cp->handle);
1850
1851 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1852
1853 hci_dev_lock(hdev);
1854
1855 acl = hci_conn_hash_lookup_handle(hdev, handle);
1856 if (acl) {
1857 sco = acl->link;
1858 if (sco) {
1859 sco->state = BT_CLOSED;
1860
1861 hci_proto_connect_cfm(sco, status);
1862 hci_conn_del(sco);
1863 }
1864 }
1865
1866 hci_dev_unlock(hdev);
1867 }
1868
1869 static void hci_cs_sniff_mode(struct hci_dev *hdev, __u8 status)
1870 {
1871 struct hci_cp_sniff_mode *cp;
1872 struct hci_conn *conn;
1873
1874 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1875
1876 if (!status)
1877 return;
1878
1879 cp = hci_sent_cmd_data(hdev, HCI_OP_SNIFF_MODE);
1880 if (!cp)
1881 return;
1882
1883 hci_dev_lock(hdev);
1884
1885 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1886 if (conn) {
1887 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1888
1889 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1890 hci_sco_setup(conn, status);
1891 }
1892
1893 hci_dev_unlock(hdev);
1894 }
1895
1896 static void hci_cs_exit_sniff_mode(struct hci_dev *hdev, __u8 status)
1897 {
1898 struct hci_cp_exit_sniff_mode *cp;
1899 struct hci_conn *conn;
1900
1901 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1902
1903 if (!status)
1904 return;
1905
1906 cp = hci_sent_cmd_data(hdev, HCI_OP_EXIT_SNIFF_MODE);
1907 if (!cp)
1908 return;
1909
1910 hci_dev_lock(hdev);
1911
1912 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1913 if (conn) {
1914 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1915
1916 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1917 hci_sco_setup(conn, status);
1918 }
1919
1920 hci_dev_unlock(hdev);
1921 }
1922
1923 static void hci_cs_disconnect(struct hci_dev *hdev, u8 status)
1924 {
1925 struct hci_cp_disconnect *cp;
1926 struct hci_conn *conn;
1927
1928 if (!status)
1929 return;
1930
1931 cp = hci_sent_cmd_data(hdev, HCI_OP_DISCONNECT);
1932 if (!cp)
1933 return;
1934
1935 hci_dev_lock(hdev);
1936
1937 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1938 if (conn)
1939 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
1940 conn->dst_type, status);
1941
1942 hci_dev_unlock(hdev);
1943 }
1944
1945 static void hci_cs_create_phylink(struct hci_dev *hdev, u8 status)
1946 {
1947 struct hci_cp_create_phy_link *cp;
1948
1949 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1950
1951 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_PHY_LINK);
1952 if (!cp)
1953 return;
1954
1955 hci_dev_lock(hdev);
1956
1957 if (status) {
1958 struct hci_conn *hcon;
1959
1960 hcon = hci_conn_hash_lookup_handle(hdev, cp->phy_handle);
1961 if (hcon)
1962 hci_conn_del(hcon);
1963 } else {
1964 amp_write_remote_assoc(hdev, cp->phy_handle);
1965 }
1966
1967 hci_dev_unlock(hdev);
1968 }
1969
1970 static void hci_cs_accept_phylink(struct hci_dev *hdev, u8 status)
1971 {
1972 struct hci_cp_accept_phy_link *cp;
1973
1974 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1975
1976 if (status)
1977 return;
1978
1979 cp = hci_sent_cmd_data(hdev, HCI_OP_ACCEPT_PHY_LINK);
1980 if (!cp)
1981 return;
1982
1983 amp_write_remote_assoc(hdev, cp->phy_handle);
1984 }
1985
1986 static void hci_cs_le_create_conn(struct hci_dev *hdev, u8 status)
1987 {
1988 struct hci_cp_le_create_conn *cp;
1989 struct hci_conn *conn;
1990
1991 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1992
1993 /* All connection failure handling is taken care of by the
1994 * hci_le_conn_failed function which is triggered by the HCI
1995 * request completion callbacks used for connecting.
1996 */
1997 if (status)
1998 return;
1999
2000 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_CREATE_CONN);
2001 if (!cp)
2002 return;
2003
2004 hci_dev_lock(hdev);
2005
2006 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->peer_addr);
2007 if (!conn)
2008 goto unlock;
2009
2010 /* Store the initiator and responder address information which
2011 * is needed for SMP. These values will not change during the
2012 * lifetime of the connection.
2013 */
2014 conn->init_addr_type = cp->own_address_type;
2015 if (cp->own_address_type == ADDR_LE_DEV_RANDOM)
2016 bacpy(&conn->init_addr, &hdev->random_addr);
2017 else
2018 bacpy(&conn->init_addr, &hdev->bdaddr);
2019
2020 conn->resp_addr_type = cp->peer_addr_type;
2021 bacpy(&conn->resp_addr, &cp->peer_addr);
2022
2023 /* We don't want the connection attempt to stick around
2024 * indefinitely since LE doesn't have a page timeout concept
2025 * like BR/EDR. Set a timer for any connection that doesn't use
2026 * the white list for connecting.
2027 */
2028 if (cp->filter_policy == HCI_LE_USE_PEER_ADDR)
2029 queue_delayed_work(conn->hdev->workqueue,
2030 &conn->le_conn_timeout,
2031 conn->conn_timeout);
2032
2033 unlock:
2034 hci_dev_unlock(hdev);
2035 }
2036
2037 static void hci_cs_le_start_enc(struct hci_dev *hdev, u8 status)
2038 {
2039 struct hci_cp_le_start_enc *cp;
2040 struct hci_conn *conn;
2041
2042 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2043
2044 if (!status)
2045 return;
2046
2047 hci_dev_lock(hdev);
2048
2049 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_START_ENC);
2050 if (!cp)
2051 goto unlock;
2052
2053 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2054 if (!conn)
2055 goto unlock;
2056
2057 if (conn->state != BT_CONNECTED)
2058 goto unlock;
2059
2060 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2061 hci_conn_drop(conn);
2062
2063 unlock:
2064 hci_dev_unlock(hdev);
2065 }
2066
2067 static void hci_cs_switch_role(struct hci_dev *hdev, u8 status)
2068 {
2069 struct hci_cp_switch_role *cp;
2070 struct hci_conn *conn;
2071
2072 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2073
2074 if (!status)
2075 return;
2076
2077 cp = hci_sent_cmd_data(hdev, HCI_OP_SWITCH_ROLE);
2078 if (!cp)
2079 return;
2080
2081 hci_dev_lock(hdev);
2082
2083 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
2084 if (conn)
2085 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
2086
2087 hci_dev_unlock(hdev);
2088 }
2089
2090 static void hci_inquiry_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2091 {
2092 __u8 status = *((__u8 *) skb->data);
2093 struct discovery_state *discov = &hdev->discovery;
2094 struct inquiry_entry *e;
2095
2096 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2097
2098 hci_conn_check_pending(hdev);
2099
2100 if (!test_and_clear_bit(HCI_INQUIRY, &hdev->flags))
2101 return;
2102
2103 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
2104 wake_up_bit(&hdev->flags, HCI_INQUIRY);
2105
2106 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
2107 return;
2108
2109 hci_dev_lock(hdev);
2110
2111 if (discov->state != DISCOVERY_FINDING)
2112 goto unlock;
2113
2114 if (list_empty(&discov->resolve)) {
2115 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2116 goto unlock;
2117 }
2118
2119 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
2120 if (e && hci_resolve_name(hdev, e) == 0) {
2121 e->name_state = NAME_PENDING;
2122 hci_discovery_set_state(hdev, DISCOVERY_RESOLVING);
2123 } else {
2124 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2125 }
2126
2127 unlock:
2128 hci_dev_unlock(hdev);
2129 }
2130
2131 static void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff *skb)
2132 {
2133 struct inquiry_data data;
2134 struct inquiry_info *info = (void *) (skb->data + 1);
2135 int num_rsp = *((__u8 *) skb->data);
2136
2137 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
2138
2139 if (!num_rsp)
2140 return;
2141
2142 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
2143 return;
2144
2145 hci_dev_lock(hdev);
2146
2147 for (; num_rsp; num_rsp--, info++) {
2148 u32 flags;
2149
2150 bacpy(&data.bdaddr, &info->bdaddr);
2151 data.pscan_rep_mode = info->pscan_rep_mode;
2152 data.pscan_period_mode = info->pscan_period_mode;
2153 data.pscan_mode = info->pscan_mode;
2154 memcpy(data.dev_class, info->dev_class, 3);
2155 data.clock_offset = info->clock_offset;
2156 data.rssi = HCI_RSSI_INVALID;
2157 data.ssp_mode = 0x00;
2158
2159 flags = hci_inquiry_cache_update(hdev, &data, false);
2160
2161 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
2162 info->dev_class, HCI_RSSI_INVALID,
2163 flags, NULL, 0, NULL, 0);
2164 }
2165
2166 hci_dev_unlock(hdev);
2167 }
2168
2169 static void hci_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2170 {
2171 struct hci_ev_conn_complete *ev = (void *) skb->data;
2172 struct hci_conn *conn;
2173
2174 BT_DBG("%s", hdev->name);
2175
2176 hci_dev_lock(hdev);
2177
2178 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
2179 if (!conn) {
2180 if (ev->link_type != SCO_LINK)
2181 goto unlock;
2182
2183 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
2184 if (!conn)
2185 goto unlock;
2186
2187 conn->type = SCO_LINK;
2188 }
2189
2190 if (!ev->status) {
2191 conn->handle = __le16_to_cpu(ev->handle);
2192
2193 if (conn->type == ACL_LINK) {
2194 conn->state = BT_CONFIG;
2195 hci_conn_hold(conn);
2196
2197 if (!conn->out && !hci_conn_ssp_enabled(conn) &&
2198 !hci_find_link_key(hdev, &ev->bdaddr))
2199 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
2200 else
2201 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2202 } else
2203 conn->state = BT_CONNECTED;
2204
2205 hci_debugfs_create_conn(conn);
2206 hci_conn_add_sysfs(conn);
2207
2208 if (test_bit(HCI_AUTH, &hdev->flags))
2209 set_bit(HCI_CONN_AUTH, &conn->flags);
2210
2211 if (test_bit(HCI_ENCRYPT, &hdev->flags))
2212 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2213
2214 /* Get remote features */
2215 if (conn->type == ACL_LINK) {
2216 struct hci_cp_read_remote_features cp;
2217 cp.handle = ev->handle;
2218 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_FEATURES,
2219 sizeof(cp), &cp);
2220
2221 hci_update_page_scan(hdev);
2222 }
2223
2224 /* Set packet type for incoming connection */
2225 if (!conn->out && hdev->hci_ver < BLUETOOTH_VER_2_0) {
2226 struct hci_cp_change_conn_ptype cp;
2227 cp.handle = ev->handle;
2228 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2229 hci_send_cmd(hdev, HCI_OP_CHANGE_CONN_PTYPE, sizeof(cp),
2230 &cp);
2231 }
2232 } else {
2233 conn->state = BT_CLOSED;
2234 if (conn->type == ACL_LINK)
2235 mgmt_connect_failed(hdev, &conn->dst, conn->type,
2236 conn->dst_type, ev->status);
2237 }
2238
2239 if (conn->type == ACL_LINK)
2240 hci_sco_setup(conn, ev->status);
2241
2242 if (ev->status) {
2243 hci_proto_connect_cfm(conn, ev->status);
2244 hci_conn_del(conn);
2245 } else if (ev->link_type != ACL_LINK)
2246 hci_proto_connect_cfm(conn, ev->status);
2247
2248 unlock:
2249 hci_dev_unlock(hdev);
2250
2251 hci_conn_check_pending(hdev);
2252 }
2253
2254 static void hci_reject_conn(struct hci_dev *hdev, bdaddr_t *bdaddr)
2255 {
2256 struct hci_cp_reject_conn_req cp;
2257
2258 bacpy(&cp.bdaddr, bdaddr);
2259 cp.reason = HCI_ERROR_REJ_BAD_ADDR;
2260 hci_send_cmd(hdev, HCI_OP_REJECT_CONN_REQ, sizeof(cp), &cp);
2261 }
2262
2263 static void hci_conn_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
2264 {
2265 struct hci_ev_conn_request *ev = (void *) skb->data;
2266 int mask = hdev->link_mode;
2267 struct inquiry_entry *ie;
2268 struct hci_conn *conn;
2269 __u8 flags = 0;
2270
2271 BT_DBG("%s bdaddr %pMR type 0x%x", hdev->name, &ev->bdaddr,
2272 ev->link_type);
2273
2274 mask |= hci_proto_connect_ind(hdev, &ev->bdaddr, ev->link_type,
2275 &flags);
2276
2277 if (!(mask & HCI_LM_ACCEPT)) {
2278 hci_reject_conn(hdev, &ev->bdaddr);
2279 return;
2280 }
2281
2282 if (hci_bdaddr_list_lookup(&hdev->blacklist, &ev->bdaddr,
2283 BDADDR_BREDR)) {
2284 hci_reject_conn(hdev, &ev->bdaddr);
2285 return;
2286 }
2287
2288 /* Require HCI_CONNECTABLE or a whitelist entry to accept the
2289 * connection. These features are only touched through mgmt so
2290 * only do the checks if HCI_MGMT is set.
2291 */
2292 if (test_bit(HCI_MGMT, &hdev->dev_flags) &&
2293 !test_bit(HCI_CONNECTABLE, &hdev->dev_flags) &&
2294 !hci_bdaddr_list_lookup(&hdev->whitelist, &ev->bdaddr,
2295 BDADDR_BREDR)) {
2296 hci_reject_conn(hdev, &ev->bdaddr);
2297 return;
2298 }
2299
2300 /* Connection accepted */
2301
2302 hci_dev_lock(hdev);
2303
2304 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
2305 if (ie)
2306 memcpy(ie->data.dev_class, ev->dev_class, 3);
2307
2308 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type,
2309 &ev->bdaddr);
2310 if (!conn) {
2311 conn = hci_conn_add(hdev, ev->link_type, &ev->bdaddr,
2312 HCI_ROLE_SLAVE);
2313 if (!conn) {
2314 BT_ERR("No memory for new connection");
2315 hci_dev_unlock(hdev);
2316 return;
2317 }
2318 }
2319
2320 memcpy(conn->dev_class, ev->dev_class, 3);
2321
2322 hci_dev_unlock(hdev);
2323
2324 if (ev->link_type == ACL_LINK ||
2325 (!(flags & HCI_PROTO_DEFER) && !lmp_esco_capable(hdev))) {
2326 struct hci_cp_accept_conn_req cp;
2327 conn->state = BT_CONNECT;
2328
2329 bacpy(&cp.bdaddr, &ev->bdaddr);
2330
2331 if (lmp_rswitch_capable(hdev) && (mask & HCI_LM_MASTER))
2332 cp.role = 0x00; /* Become master */
2333 else
2334 cp.role = 0x01; /* Remain slave */
2335
2336 hci_send_cmd(hdev, HCI_OP_ACCEPT_CONN_REQ, sizeof(cp), &cp);
2337 } else if (!(flags & HCI_PROTO_DEFER)) {
2338 struct hci_cp_accept_sync_conn_req cp;
2339 conn->state = BT_CONNECT;
2340
2341 bacpy(&cp.bdaddr, &ev->bdaddr);
2342 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2343
2344 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
2345 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
2346 cp.max_latency = cpu_to_le16(0xffff);
2347 cp.content_format = cpu_to_le16(hdev->voice_setting);
2348 cp.retrans_effort = 0xff;
2349
2350 hci_send_cmd(hdev, HCI_OP_ACCEPT_SYNC_CONN_REQ, sizeof(cp),
2351 &cp);
2352 } else {
2353 conn->state = BT_CONNECT2;
2354 hci_proto_connect_cfm(conn, 0);
2355 }
2356 }
2357
2358 static u8 hci_to_mgmt_reason(u8 err)
2359 {
2360 switch (err) {
2361 case HCI_ERROR_CONNECTION_TIMEOUT:
2362 return MGMT_DEV_DISCONN_TIMEOUT;
2363 case HCI_ERROR_REMOTE_USER_TERM:
2364 case HCI_ERROR_REMOTE_LOW_RESOURCES:
2365 case HCI_ERROR_REMOTE_POWER_OFF:
2366 return MGMT_DEV_DISCONN_REMOTE;
2367 case HCI_ERROR_LOCAL_HOST_TERM:
2368 return MGMT_DEV_DISCONN_LOCAL_HOST;
2369 default:
2370 return MGMT_DEV_DISCONN_UNKNOWN;
2371 }
2372 }
2373
2374 static void hci_disconn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2375 {
2376 struct hci_ev_disconn_complete *ev = (void *) skb->data;
2377 u8 reason = hci_to_mgmt_reason(ev->reason);
2378 struct hci_conn_params *params;
2379 struct hci_conn *conn;
2380 bool mgmt_connected;
2381 u8 type;
2382
2383 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2384
2385 hci_dev_lock(hdev);
2386
2387 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2388 if (!conn)
2389 goto unlock;
2390
2391 if (ev->status) {
2392 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2393 conn->dst_type, ev->status);
2394 goto unlock;
2395 }
2396
2397 conn->state = BT_CLOSED;
2398
2399 mgmt_connected = test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags);
2400 mgmt_device_disconnected(hdev, &conn->dst, conn->type, conn->dst_type,
2401 reason, mgmt_connected);
2402
2403 if (conn->type == ACL_LINK) {
2404 if (test_bit(HCI_CONN_FLUSH_KEY, &conn->flags))
2405 hci_remove_link_key(hdev, &conn->dst);
2406
2407 hci_update_page_scan(hdev);
2408 }
2409
2410 params = hci_conn_params_lookup(hdev, &conn->dst, conn->dst_type);
2411 if (params) {
2412 switch (params->auto_connect) {
2413 case HCI_AUTO_CONN_LINK_LOSS:
2414 if (ev->reason != HCI_ERROR_CONNECTION_TIMEOUT)
2415 break;
2416 /* Fall through */
2417
2418 case HCI_AUTO_CONN_DIRECT:
2419 case HCI_AUTO_CONN_ALWAYS:
2420 list_del_init(&params->action);
2421 list_add(&params->action, &hdev->pend_le_conns);
2422 hci_update_background_scan(hdev);
2423 break;
2424
2425 default:
2426 break;
2427 }
2428 }
2429
2430 type = conn->type;
2431
2432 hci_proto_disconn_cfm(conn, ev->reason);
2433 hci_conn_del(conn);
2434
2435 /* Re-enable advertising if necessary, since it might
2436 * have been disabled by the connection. From the
2437 * HCI_LE_Set_Advertise_Enable command description in
2438 * the core specification (v4.0):
2439 * "The Controller shall continue advertising until the Host
2440 * issues an LE_Set_Advertise_Enable command with
2441 * Advertising_Enable set to 0x00 (Advertising is disabled)
2442 * or until a connection is created or until the Advertising
2443 * is timed out due to Directed Advertising."
2444 */
2445 if (type == LE_LINK)
2446 mgmt_reenable_advertising(hdev);
2447
2448 unlock:
2449 hci_dev_unlock(hdev);
2450 }
2451
2452 static void hci_auth_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2453 {
2454 struct hci_ev_auth_complete *ev = (void *) skb->data;
2455 struct hci_conn *conn;
2456
2457 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2458
2459 hci_dev_lock(hdev);
2460
2461 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2462 if (!conn)
2463 goto unlock;
2464
2465 if (!ev->status) {
2466 if (!hci_conn_ssp_enabled(conn) &&
2467 test_bit(HCI_CONN_REAUTH_PEND, &conn->flags)) {
2468 BT_INFO("re-auth of legacy device is not possible.");
2469 } else {
2470 set_bit(HCI_CONN_AUTH, &conn->flags);
2471 conn->sec_level = conn->pending_sec_level;
2472 }
2473 } else {
2474 mgmt_auth_failed(conn, ev->status);
2475 }
2476
2477 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2478 clear_bit(HCI_CONN_REAUTH_PEND, &conn->flags);
2479
2480 if (conn->state == BT_CONFIG) {
2481 if (!ev->status && hci_conn_ssp_enabled(conn)) {
2482 struct hci_cp_set_conn_encrypt cp;
2483 cp.handle = ev->handle;
2484 cp.encrypt = 0x01;
2485 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2486 &cp);
2487 } else {
2488 conn->state = BT_CONNECTED;
2489 hci_proto_connect_cfm(conn, ev->status);
2490 hci_conn_drop(conn);
2491 }
2492 } else {
2493 hci_auth_cfm(conn, ev->status);
2494
2495 hci_conn_hold(conn);
2496 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2497 hci_conn_drop(conn);
2498 }
2499
2500 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags)) {
2501 if (!ev->status) {
2502 struct hci_cp_set_conn_encrypt cp;
2503 cp.handle = ev->handle;
2504 cp.encrypt = 0x01;
2505 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2506 &cp);
2507 } else {
2508 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2509 hci_encrypt_cfm(conn, ev->status, 0x00);
2510 }
2511 }
2512
2513 unlock:
2514 hci_dev_unlock(hdev);
2515 }
2516
2517 static void hci_remote_name_evt(struct hci_dev *hdev, struct sk_buff *skb)
2518 {
2519 struct hci_ev_remote_name *ev = (void *) skb->data;
2520 struct hci_conn *conn;
2521
2522 BT_DBG("%s", hdev->name);
2523
2524 hci_conn_check_pending(hdev);
2525
2526 hci_dev_lock(hdev);
2527
2528 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2529
2530 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
2531 goto check_auth;
2532
2533 if (ev->status == 0)
2534 hci_check_pending_name(hdev, conn, &ev->bdaddr, ev->name,
2535 strnlen(ev->name, HCI_MAX_NAME_LENGTH));
2536 else
2537 hci_check_pending_name(hdev, conn, &ev->bdaddr, NULL, 0);
2538
2539 check_auth:
2540 if (!conn)
2541 goto unlock;
2542
2543 if (!hci_outgoing_auth_needed(hdev, conn))
2544 goto unlock;
2545
2546 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
2547 struct hci_cp_auth_requested cp;
2548
2549 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
2550
2551 cp.handle = __cpu_to_le16(conn->handle);
2552 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED, sizeof(cp), &cp);
2553 }
2554
2555 unlock:
2556 hci_dev_unlock(hdev);
2557 }
2558
2559 static void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2560 {
2561 struct hci_ev_encrypt_change *ev = (void *) skb->data;
2562 struct hci_conn *conn;
2563
2564 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2565
2566 hci_dev_lock(hdev);
2567
2568 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2569 if (!conn)
2570 goto unlock;
2571
2572 if (!ev->status) {
2573 if (ev->encrypt) {
2574 /* Encryption implies authentication */
2575 set_bit(HCI_CONN_AUTH, &conn->flags);
2576 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2577 conn->sec_level = conn->pending_sec_level;
2578
2579 /* P-256 authentication key implies FIPS */
2580 if (conn->key_type == HCI_LK_AUTH_COMBINATION_P256)
2581 set_bit(HCI_CONN_FIPS, &conn->flags);
2582
2583 if ((conn->type == ACL_LINK && ev->encrypt == 0x02) ||
2584 conn->type == LE_LINK)
2585 set_bit(HCI_CONN_AES_CCM, &conn->flags);
2586 } else {
2587 clear_bit(HCI_CONN_ENCRYPT, &conn->flags);
2588 clear_bit(HCI_CONN_AES_CCM, &conn->flags);
2589 }
2590 }
2591
2592 /* We should disregard the current RPA and generate a new one
2593 * whenever the encryption procedure fails.
2594 */
2595 if (ev->status && conn->type == LE_LINK)
2596 set_bit(HCI_RPA_EXPIRED, &hdev->dev_flags);
2597
2598 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2599
2600 if (ev->status && conn->state == BT_CONNECTED) {
2601 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2602 hci_conn_drop(conn);
2603 goto unlock;
2604 }
2605
2606 if (conn->state == BT_CONFIG) {
2607 if (!ev->status)
2608 conn->state = BT_CONNECTED;
2609
2610 /* In Secure Connections Only mode, do not allow any
2611 * connections that are not encrypted with AES-CCM
2612 * using a P-256 authenticated combination key.
2613 */
2614 if (test_bit(HCI_SC_ONLY, &hdev->dev_flags) &&
2615 (!test_bit(HCI_CONN_AES_CCM, &conn->flags) ||
2616 conn->key_type != HCI_LK_AUTH_COMBINATION_P256)) {
2617 hci_proto_connect_cfm(conn, HCI_ERROR_AUTH_FAILURE);
2618 hci_conn_drop(conn);
2619 goto unlock;
2620 }
2621
2622 hci_proto_connect_cfm(conn, ev->status);
2623 hci_conn_drop(conn);
2624 } else
2625 hci_encrypt_cfm(conn, ev->status, ev->encrypt);
2626
2627 unlock:
2628 hci_dev_unlock(hdev);
2629 }
2630
2631 static void hci_change_link_key_complete_evt(struct hci_dev *hdev,
2632 struct sk_buff *skb)
2633 {
2634 struct hci_ev_change_link_key_complete *ev = (void *) skb->data;
2635 struct hci_conn *conn;
2636
2637 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2638
2639 hci_dev_lock(hdev);
2640
2641 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2642 if (conn) {
2643 if (!ev->status)
2644 set_bit(HCI_CONN_SECURE, &conn->flags);
2645
2646 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2647
2648 hci_key_change_cfm(conn, ev->status);
2649 }
2650
2651 hci_dev_unlock(hdev);
2652 }
2653
2654 static void hci_remote_features_evt(struct hci_dev *hdev,
2655 struct sk_buff *skb)
2656 {
2657 struct hci_ev_remote_features *ev = (void *) skb->data;
2658 struct hci_conn *conn;
2659
2660 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2661
2662 hci_dev_lock(hdev);
2663
2664 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2665 if (!conn)
2666 goto unlock;
2667
2668 if (!ev->status)
2669 memcpy(conn->features[0], ev->features, 8);
2670
2671 if (conn->state != BT_CONFIG)
2672 goto unlock;
2673
2674 if (!ev->status && lmp_ext_feat_capable(hdev) &&
2675 lmp_ext_feat_capable(conn)) {
2676 struct hci_cp_read_remote_ext_features cp;
2677 cp.handle = ev->handle;
2678 cp.page = 0x01;
2679 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES,
2680 sizeof(cp), &cp);
2681 goto unlock;
2682 }
2683
2684 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
2685 struct hci_cp_remote_name_req cp;
2686 memset(&cp, 0, sizeof(cp));
2687 bacpy(&cp.bdaddr, &conn->dst);
2688 cp.pscan_rep_mode = 0x02;
2689 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
2690 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2691 mgmt_device_connected(hdev, conn, 0, NULL, 0);
2692
2693 if (!hci_outgoing_auth_needed(hdev, conn)) {
2694 conn->state = BT_CONNECTED;
2695 hci_proto_connect_cfm(conn, ev->status);
2696 hci_conn_drop(conn);
2697 }
2698
2699 unlock:
2700 hci_dev_unlock(hdev);
2701 }
2702
2703 static void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2704 {
2705 struct hci_ev_cmd_complete *ev = (void *) skb->data;
2706 u8 status = skb->data[sizeof(*ev)];
2707 __u16 opcode;
2708
2709 skb_pull(skb, sizeof(*ev));
2710
2711 opcode = __le16_to_cpu(ev->opcode);
2712
2713 switch (opcode) {
2714 case HCI_OP_INQUIRY_CANCEL:
2715 hci_cc_inquiry_cancel(hdev, skb);
2716 break;
2717
2718 case HCI_OP_PERIODIC_INQ:
2719 hci_cc_periodic_inq(hdev, skb);
2720 break;
2721
2722 case HCI_OP_EXIT_PERIODIC_INQ:
2723 hci_cc_exit_periodic_inq(hdev, skb);
2724 break;
2725
2726 case HCI_OP_REMOTE_NAME_REQ_CANCEL:
2727 hci_cc_remote_name_req_cancel(hdev, skb);
2728 break;
2729
2730 case HCI_OP_ROLE_DISCOVERY:
2731 hci_cc_role_discovery(hdev, skb);
2732 break;
2733
2734 case HCI_OP_READ_LINK_POLICY:
2735 hci_cc_read_link_policy(hdev, skb);
2736 break;
2737
2738 case HCI_OP_WRITE_LINK_POLICY:
2739 hci_cc_write_link_policy(hdev, skb);
2740 break;
2741
2742 case HCI_OP_READ_DEF_LINK_POLICY:
2743 hci_cc_read_def_link_policy(hdev, skb);
2744 break;
2745
2746 case HCI_OP_WRITE_DEF_LINK_POLICY:
2747 hci_cc_write_def_link_policy(hdev, skb);
2748 break;
2749
2750 case HCI_OP_RESET:
2751 hci_cc_reset(hdev, skb);
2752 break;
2753
2754 case HCI_OP_READ_STORED_LINK_KEY:
2755 hci_cc_read_stored_link_key(hdev, skb);
2756 break;
2757
2758 case HCI_OP_DELETE_STORED_LINK_KEY:
2759 hci_cc_delete_stored_link_key(hdev, skb);
2760 break;
2761
2762 case HCI_OP_WRITE_LOCAL_NAME:
2763 hci_cc_write_local_name(hdev, skb);
2764 break;
2765
2766 case HCI_OP_READ_LOCAL_NAME:
2767 hci_cc_read_local_name(hdev, skb);
2768 break;
2769
2770 case HCI_OP_WRITE_AUTH_ENABLE:
2771 hci_cc_write_auth_enable(hdev, skb);
2772 break;
2773
2774 case HCI_OP_WRITE_ENCRYPT_MODE:
2775 hci_cc_write_encrypt_mode(hdev, skb);
2776 break;
2777
2778 case HCI_OP_WRITE_SCAN_ENABLE:
2779 hci_cc_write_scan_enable(hdev, skb);
2780 break;
2781
2782 case HCI_OP_READ_CLASS_OF_DEV:
2783 hci_cc_read_class_of_dev(hdev, skb);
2784 break;
2785
2786 case HCI_OP_WRITE_CLASS_OF_DEV:
2787 hci_cc_write_class_of_dev(hdev, skb);
2788 break;
2789
2790 case HCI_OP_READ_VOICE_SETTING:
2791 hci_cc_read_voice_setting(hdev, skb);
2792 break;
2793
2794 case HCI_OP_WRITE_VOICE_SETTING:
2795 hci_cc_write_voice_setting(hdev, skb);
2796 break;
2797
2798 case HCI_OP_READ_NUM_SUPPORTED_IAC:
2799 hci_cc_read_num_supported_iac(hdev, skb);
2800 break;
2801
2802 case HCI_OP_WRITE_SSP_MODE:
2803 hci_cc_write_ssp_mode(hdev, skb);
2804 break;
2805
2806 case HCI_OP_WRITE_SC_SUPPORT:
2807 hci_cc_write_sc_support(hdev, skb);
2808 break;
2809
2810 case HCI_OP_READ_LOCAL_VERSION:
2811 hci_cc_read_local_version(hdev, skb);
2812 break;
2813
2814 case HCI_OP_READ_LOCAL_COMMANDS:
2815 hci_cc_read_local_commands(hdev, skb);
2816 break;
2817
2818 case HCI_OP_READ_LOCAL_FEATURES:
2819 hci_cc_read_local_features(hdev, skb);
2820 break;
2821
2822 case HCI_OP_READ_LOCAL_EXT_FEATURES:
2823 hci_cc_read_local_ext_features(hdev, skb);
2824 break;
2825
2826 case HCI_OP_READ_BUFFER_SIZE:
2827 hci_cc_read_buffer_size(hdev, skb);
2828 break;
2829
2830 case HCI_OP_READ_BD_ADDR:
2831 hci_cc_read_bd_addr(hdev, skb);
2832 break;
2833
2834 case HCI_OP_READ_PAGE_SCAN_ACTIVITY:
2835 hci_cc_read_page_scan_activity(hdev, skb);
2836 break;
2837
2838 case HCI_OP_WRITE_PAGE_SCAN_ACTIVITY:
2839 hci_cc_write_page_scan_activity(hdev, skb);
2840 break;
2841
2842 case HCI_OP_READ_PAGE_SCAN_TYPE:
2843 hci_cc_read_page_scan_type(hdev, skb);
2844 break;
2845
2846 case HCI_OP_WRITE_PAGE_SCAN_TYPE:
2847 hci_cc_write_page_scan_type(hdev, skb);
2848 break;
2849
2850 case HCI_OP_READ_DATA_BLOCK_SIZE:
2851 hci_cc_read_data_block_size(hdev, skb);
2852 break;
2853
2854 case HCI_OP_READ_FLOW_CONTROL_MODE:
2855 hci_cc_read_flow_control_mode(hdev, skb);
2856 break;
2857
2858 case HCI_OP_READ_LOCAL_AMP_INFO:
2859 hci_cc_read_local_amp_info(hdev, skb);
2860 break;
2861
2862 case HCI_OP_READ_CLOCK:
2863 hci_cc_read_clock(hdev, skb);
2864 break;
2865
2866 case HCI_OP_READ_LOCAL_AMP_ASSOC:
2867 hci_cc_read_local_amp_assoc(hdev, skb);
2868 break;
2869
2870 case HCI_OP_READ_INQ_RSP_TX_POWER:
2871 hci_cc_read_inq_rsp_tx_power(hdev, skb);
2872 break;
2873
2874 case HCI_OP_PIN_CODE_REPLY:
2875 hci_cc_pin_code_reply(hdev, skb);
2876 break;
2877
2878 case HCI_OP_PIN_CODE_NEG_REPLY:
2879 hci_cc_pin_code_neg_reply(hdev, skb);
2880 break;
2881
2882 case HCI_OP_READ_LOCAL_OOB_DATA:
2883 hci_cc_read_local_oob_data(hdev, skb);
2884 break;
2885
2886 case HCI_OP_READ_LOCAL_OOB_EXT_DATA:
2887 hci_cc_read_local_oob_ext_data(hdev, skb);
2888 break;
2889
2890 case HCI_OP_LE_READ_BUFFER_SIZE:
2891 hci_cc_le_read_buffer_size(hdev, skb);
2892 break;
2893
2894 case HCI_OP_LE_READ_LOCAL_FEATURES:
2895 hci_cc_le_read_local_features(hdev, skb);
2896 break;
2897
2898 case HCI_OP_LE_READ_ADV_TX_POWER:
2899 hci_cc_le_read_adv_tx_power(hdev, skb);
2900 break;
2901
2902 case HCI_OP_USER_CONFIRM_REPLY:
2903 hci_cc_user_confirm_reply(hdev, skb);
2904 break;
2905
2906 case HCI_OP_USER_CONFIRM_NEG_REPLY:
2907 hci_cc_user_confirm_neg_reply(hdev, skb);
2908 break;
2909
2910 case HCI_OP_USER_PASSKEY_REPLY:
2911 hci_cc_user_passkey_reply(hdev, skb);
2912 break;
2913
2914 case HCI_OP_USER_PASSKEY_NEG_REPLY:
2915 hci_cc_user_passkey_neg_reply(hdev, skb);
2916 break;
2917
2918 case HCI_OP_LE_SET_RANDOM_ADDR:
2919 hci_cc_le_set_random_addr(hdev, skb);
2920 break;
2921
2922 case HCI_OP_LE_SET_ADV_ENABLE:
2923 hci_cc_le_set_adv_enable(hdev, skb);
2924 break;
2925
2926 case HCI_OP_LE_SET_SCAN_PARAM:
2927 hci_cc_le_set_scan_param(hdev, skb);
2928 break;
2929
2930 case HCI_OP_LE_SET_SCAN_ENABLE:
2931 hci_cc_le_set_scan_enable(hdev, skb);
2932 break;
2933
2934 case HCI_OP_LE_READ_WHITE_LIST_SIZE:
2935 hci_cc_le_read_white_list_size(hdev, skb);
2936 break;
2937
2938 case HCI_OP_LE_CLEAR_WHITE_LIST:
2939 hci_cc_le_clear_white_list(hdev, skb);
2940 break;
2941
2942 case HCI_OP_LE_ADD_TO_WHITE_LIST:
2943 hci_cc_le_add_to_white_list(hdev, skb);
2944 break;
2945
2946 case HCI_OP_LE_DEL_FROM_WHITE_LIST:
2947 hci_cc_le_del_from_white_list(hdev, skb);
2948 break;
2949
2950 case HCI_OP_LE_READ_SUPPORTED_STATES:
2951 hci_cc_le_read_supported_states(hdev, skb);
2952 break;
2953
2954 case HCI_OP_LE_READ_DEF_DATA_LEN:
2955 hci_cc_le_read_def_data_len(hdev, skb);
2956 break;
2957
2958 case HCI_OP_LE_WRITE_DEF_DATA_LEN:
2959 hci_cc_le_write_def_data_len(hdev, skb);
2960 break;
2961
2962 case HCI_OP_LE_READ_MAX_DATA_LEN:
2963 hci_cc_le_read_max_data_len(hdev, skb);
2964 break;
2965
2966 case HCI_OP_WRITE_LE_HOST_SUPPORTED:
2967 hci_cc_write_le_host_supported(hdev, skb);
2968 break;
2969
2970 case HCI_OP_LE_SET_ADV_PARAM:
2971 hci_cc_set_adv_param(hdev, skb);
2972 break;
2973
2974 case HCI_OP_WRITE_REMOTE_AMP_ASSOC:
2975 hci_cc_write_remote_amp_assoc(hdev, skb);
2976 break;
2977
2978 case HCI_OP_READ_RSSI:
2979 hci_cc_read_rssi(hdev, skb);
2980 break;
2981
2982 case HCI_OP_READ_TX_POWER:
2983 hci_cc_read_tx_power(hdev, skb);
2984 break;
2985
2986 default:
2987 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2988 break;
2989 }
2990
2991 if (opcode != HCI_OP_NOP)
2992 cancel_delayed_work(&hdev->cmd_timer);
2993
2994 hci_req_cmd_complete(hdev, opcode, status);
2995
2996 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2997 atomic_set(&hdev->cmd_cnt, 1);
2998 if (!skb_queue_empty(&hdev->cmd_q))
2999 queue_work(hdev->workqueue, &hdev->cmd_work);
3000 }
3001 }
3002
3003 static void hci_cmd_status_evt(struct hci_dev *hdev, struct sk_buff *skb)
3004 {
3005 struct hci_ev_cmd_status *ev = (void *) skb->data;
3006 __u16 opcode;
3007
3008 skb_pull(skb, sizeof(*ev));
3009
3010 opcode = __le16_to_cpu(ev->opcode);
3011
3012 switch (opcode) {
3013 case HCI_OP_INQUIRY:
3014 hci_cs_inquiry(hdev, ev->status);
3015 break;
3016
3017 case HCI_OP_CREATE_CONN:
3018 hci_cs_create_conn(hdev, ev->status);
3019 break;
3020
3021 case HCI_OP_DISCONNECT:
3022 hci_cs_disconnect(hdev, ev->status);
3023 break;
3024
3025 case HCI_OP_ADD_SCO:
3026 hci_cs_add_sco(hdev, ev->status);
3027 break;
3028
3029 case HCI_OP_AUTH_REQUESTED:
3030 hci_cs_auth_requested(hdev, ev->status);
3031 break;
3032
3033 case HCI_OP_SET_CONN_ENCRYPT:
3034 hci_cs_set_conn_encrypt(hdev, ev->status);
3035 break;
3036
3037 case HCI_OP_REMOTE_NAME_REQ:
3038 hci_cs_remote_name_req(hdev, ev->status);
3039 break;
3040
3041 case HCI_OP_READ_REMOTE_FEATURES:
3042 hci_cs_read_remote_features(hdev, ev->status);
3043 break;
3044
3045 case HCI_OP_READ_REMOTE_EXT_FEATURES:
3046 hci_cs_read_remote_ext_features(hdev, ev->status);
3047 break;
3048
3049 case HCI_OP_SETUP_SYNC_CONN:
3050 hci_cs_setup_sync_conn(hdev, ev->status);
3051 break;
3052
3053 case HCI_OP_CREATE_PHY_LINK:
3054 hci_cs_create_phylink(hdev, ev->status);
3055 break;
3056
3057 case HCI_OP_ACCEPT_PHY_LINK:
3058 hci_cs_accept_phylink(hdev, ev->status);
3059 break;
3060
3061 case HCI_OP_SNIFF_MODE:
3062 hci_cs_sniff_mode(hdev, ev->status);
3063 break;
3064
3065 case HCI_OP_EXIT_SNIFF_MODE:
3066 hci_cs_exit_sniff_mode(hdev, ev->status);
3067 break;
3068
3069 case HCI_OP_SWITCH_ROLE:
3070 hci_cs_switch_role(hdev, ev->status);
3071 break;
3072
3073 case HCI_OP_LE_CREATE_CONN:
3074 hci_cs_le_create_conn(hdev, ev->status);
3075 break;
3076
3077 case HCI_OP_LE_START_ENC:
3078 hci_cs_le_start_enc(hdev, ev->status);
3079 break;
3080
3081 default:
3082 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
3083 break;
3084 }
3085
3086 if (opcode != HCI_OP_NOP)
3087 cancel_delayed_work(&hdev->cmd_timer);
3088
3089 if (ev->status ||
3090 (hdev->sent_cmd && !bt_cb(hdev->sent_cmd)->req.event))
3091 hci_req_cmd_complete(hdev, opcode, ev->status);
3092
3093 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
3094 atomic_set(&hdev->cmd_cnt, 1);
3095 if (!skb_queue_empty(&hdev->cmd_q))
3096 queue_work(hdev->workqueue, &hdev->cmd_work);
3097 }
3098 }
3099
3100 static void hci_hardware_error_evt(struct hci_dev *hdev, struct sk_buff *skb)
3101 {
3102 struct hci_ev_hardware_error *ev = (void *) skb->data;
3103
3104 hdev->hw_error_code = ev->code;
3105
3106 queue_work(hdev->req_workqueue, &hdev->error_reset);
3107 }
3108
3109 static void hci_role_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3110 {
3111 struct hci_ev_role_change *ev = (void *) skb->data;
3112 struct hci_conn *conn;
3113
3114 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3115
3116 hci_dev_lock(hdev);
3117
3118 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3119 if (conn) {
3120 if (!ev->status)
3121 conn->role = ev->role;
3122
3123 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
3124
3125 hci_role_switch_cfm(conn, ev->status, ev->role);
3126 }
3127
3128 hci_dev_unlock(hdev);
3129 }
3130
3131 static void hci_num_comp_pkts_evt(struct hci_dev *hdev, struct sk_buff *skb)
3132 {
3133 struct hci_ev_num_comp_pkts *ev = (void *) skb->data;
3134 int i;
3135
3136 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_PACKET_BASED) {
3137 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
3138 return;
3139 }
3140
3141 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
3142 ev->num_hndl * sizeof(struct hci_comp_pkts_info)) {
3143 BT_DBG("%s bad parameters", hdev->name);
3144 return;
3145 }
3146
3147 BT_DBG("%s num_hndl %d", hdev->name, ev->num_hndl);
3148
3149 for (i = 0; i < ev->num_hndl; i++) {
3150 struct hci_comp_pkts_info *info = &ev->handles[i];
3151 struct hci_conn *conn;
3152 __u16 handle, count;
3153
3154 handle = __le16_to_cpu(info->handle);
3155 count = __le16_to_cpu(info->count);
3156
3157 conn = hci_conn_hash_lookup_handle(hdev, handle);
3158 if (!conn)
3159 continue;
3160
3161 conn->sent -= count;
3162
3163 switch (conn->type) {
3164 case ACL_LINK:
3165 hdev->acl_cnt += count;
3166 if (hdev->acl_cnt > hdev->acl_pkts)
3167 hdev->acl_cnt = hdev->acl_pkts;
3168 break;
3169
3170 case LE_LINK:
3171 if (hdev->le_pkts) {
3172 hdev->le_cnt += count;
3173 if (hdev->le_cnt > hdev->le_pkts)
3174 hdev->le_cnt = hdev->le_pkts;
3175 } else {
3176 hdev->acl_cnt += count;
3177 if (hdev->acl_cnt > hdev->acl_pkts)
3178 hdev->acl_cnt = hdev->acl_pkts;
3179 }
3180 break;
3181
3182 case SCO_LINK:
3183 hdev->sco_cnt += count;
3184 if (hdev->sco_cnt > hdev->sco_pkts)
3185 hdev->sco_cnt = hdev->sco_pkts;
3186 break;
3187
3188 default:
3189 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3190 break;
3191 }
3192 }
3193
3194 queue_work(hdev->workqueue, &hdev->tx_work);
3195 }
3196
3197 static struct hci_conn *__hci_conn_lookup_handle(struct hci_dev *hdev,
3198 __u16 handle)
3199 {
3200 struct hci_chan *chan;
3201
3202 switch (hdev->dev_type) {
3203 case HCI_BREDR:
3204 return hci_conn_hash_lookup_handle(hdev, handle);
3205 case HCI_AMP:
3206 chan = hci_chan_lookup_handle(hdev, handle);
3207 if (chan)
3208 return chan->conn;
3209 break;
3210 default:
3211 BT_ERR("%s unknown dev_type %d", hdev->name, hdev->dev_type);
3212 break;
3213 }
3214
3215 return NULL;
3216 }
3217
3218 static void hci_num_comp_blocks_evt(struct hci_dev *hdev, struct sk_buff *skb)
3219 {
3220 struct hci_ev_num_comp_blocks *ev = (void *) skb->data;
3221 int i;
3222
3223 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_BLOCK_BASED) {
3224 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
3225 return;
3226 }
3227
3228 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
3229 ev->num_hndl * sizeof(struct hci_comp_blocks_info)) {
3230 BT_DBG("%s bad parameters", hdev->name);
3231 return;
3232 }
3233
3234 BT_DBG("%s num_blocks %d num_hndl %d", hdev->name, ev->num_blocks,
3235 ev->num_hndl);
3236
3237 for (i = 0; i < ev->num_hndl; i++) {
3238 struct hci_comp_blocks_info *info = &ev->handles[i];
3239 struct hci_conn *conn = NULL;
3240 __u16 handle, block_count;
3241
3242 handle = __le16_to_cpu(info->handle);
3243 block_count = __le16_to_cpu(info->blocks);
3244
3245 conn = __hci_conn_lookup_handle(hdev, handle);
3246 if (!conn)
3247 continue;
3248
3249 conn->sent -= block_count;
3250
3251 switch (conn->type) {
3252 case ACL_LINK:
3253 case AMP_LINK:
3254 hdev->block_cnt += block_count;
3255 if (hdev->block_cnt > hdev->num_blocks)
3256 hdev->block_cnt = hdev->num_blocks;
3257 break;
3258
3259 default:
3260 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3261 break;
3262 }
3263 }
3264
3265 queue_work(hdev->workqueue, &hdev->tx_work);
3266 }
3267
3268 static void hci_mode_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3269 {
3270 struct hci_ev_mode_change *ev = (void *) skb->data;
3271 struct hci_conn *conn;
3272
3273 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3274
3275 hci_dev_lock(hdev);
3276
3277 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3278 if (conn) {
3279 conn->mode = ev->mode;
3280
3281 if (!test_and_clear_bit(HCI_CONN_MODE_CHANGE_PEND,
3282 &conn->flags)) {
3283 if (conn->mode == HCI_CM_ACTIVE)
3284 set_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3285 else
3286 clear_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3287 }
3288
3289 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
3290 hci_sco_setup(conn, ev->status);
3291 }
3292
3293 hci_dev_unlock(hdev);
3294 }
3295
3296 static void hci_pin_code_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3297 {
3298 struct hci_ev_pin_code_req *ev = (void *) skb->data;
3299 struct hci_conn *conn;
3300
3301 BT_DBG("%s", hdev->name);
3302
3303 hci_dev_lock(hdev);
3304
3305 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3306 if (!conn)
3307 goto unlock;
3308
3309 if (conn->state == BT_CONNECTED) {
3310 hci_conn_hold(conn);
3311 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
3312 hci_conn_drop(conn);
3313 }
3314
3315 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags) &&
3316 !test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags)) {
3317 hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3318 sizeof(ev->bdaddr), &ev->bdaddr);
3319 } else if (test_bit(HCI_MGMT, &hdev->dev_flags)) {
3320 u8 secure;
3321
3322 if (conn->pending_sec_level == BT_SECURITY_HIGH)
3323 secure = 1;
3324 else
3325 secure = 0;
3326
3327 mgmt_pin_code_request(hdev, &ev->bdaddr, secure);
3328 }
3329
3330 unlock:
3331 hci_dev_unlock(hdev);
3332 }
3333
3334 static void conn_set_key(struct hci_conn *conn, u8 key_type, u8 pin_len)
3335 {
3336 if (key_type == HCI_LK_CHANGED_COMBINATION)
3337 return;
3338
3339 conn->pin_length = pin_len;
3340 conn->key_type = key_type;
3341
3342 switch (key_type) {
3343 case HCI_LK_LOCAL_UNIT:
3344 case HCI_LK_REMOTE_UNIT:
3345 case HCI_LK_DEBUG_COMBINATION:
3346 return;
3347 case HCI_LK_COMBINATION:
3348 if (pin_len == 16)
3349 conn->pending_sec_level = BT_SECURITY_HIGH;
3350 else
3351 conn->pending_sec_level = BT_SECURITY_MEDIUM;
3352 break;
3353 case HCI_LK_UNAUTH_COMBINATION_P192:
3354 case HCI_LK_UNAUTH_COMBINATION_P256:
3355 conn->pending_sec_level = BT_SECURITY_MEDIUM;
3356 break;
3357 case HCI_LK_AUTH_COMBINATION_P192:
3358 conn->pending_sec_level = BT_SECURITY_HIGH;
3359 break;
3360 case HCI_LK_AUTH_COMBINATION_P256:
3361 conn->pending_sec_level = BT_SECURITY_FIPS;
3362 break;
3363 }
3364 }
3365
3366 static void hci_link_key_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3367 {
3368 struct hci_ev_link_key_req *ev = (void *) skb->data;
3369 struct hci_cp_link_key_reply cp;
3370 struct hci_conn *conn;
3371 struct link_key *key;
3372
3373 BT_DBG("%s", hdev->name);
3374
3375 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3376 return;
3377
3378 hci_dev_lock(hdev);
3379
3380 key = hci_find_link_key(hdev, &ev->bdaddr);
3381 if (!key) {
3382 BT_DBG("%s link key not found for %pMR", hdev->name,
3383 &ev->bdaddr);
3384 goto not_found;
3385 }
3386
3387 BT_DBG("%s found key type %u for %pMR", hdev->name, key->type,
3388 &ev->bdaddr);
3389
3390 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3391 if (conn) {
3392 clear_bit(HCI_CONN_NEW_LINK_KEY, &conn->flags);
3393
3394 if ((key->type == HCI_LK_UNAUTH_COMBINATION_P192 ||
3395 key->type == HCI_LK_UNAUTH_COMBINATION_P256) &&
3396 conn->auth_type != 0xff && (conn->auth_type & 0x01)) {
3397 BT_DBG("%s ignoring unauthenticated key", hdev->name);
3398 goto not_found;
3399 }
3400
3401 if (key->type == HCI_LK_COMBINATION && key->pin_len < 16 &&
3402 (conn->pending_sec_level == BT_SECURITY_HIGH ||
3403 conn->pending_sec_level == BT_SECURITY_FIPS)) {
3404 BT_DBG("%s ignoring key unauthenticated for high security",
3405 hdev->name);
3406 goto not_found;
3407 }
3408
3409 conn_set_key(conn, key->type, key->pin_len);
3410 }
3411
3412 bacpy(&cp.bdaddr, &ev->bdaddr);
3413 memcpy(cp.link_key, key->val, HCI_LINK_KEY_SIZE);
3414
3415 hci_send_cmd(hdev, HCI_OP_LINK_KEY_REPLY, sizeof(cp), &cp);
3416
3417 hci_dev_unlock(hdev);
3418
3419 return;
3420
3421 not_found:
3422 hci_send_cmd(hdev, HCI_OP_LINK_KEY_NEG_REPLY, 6, &ev->bdaddr);
3423 hci_dev_unlock(hdev);
3424 }
3425
3426 static void hci_link_key_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3427 {
3428 struct hci_ev_link_key_notify *ev = (void *) skb->data;
3429 struct hci_conn *conn;
3430 struct link_key *key;
3431 bool persistent;
3432 u8 pin_len = 0;
3433
3434 BT_DBG("%s", hdev->name);
3435
3436 hci_dev_lock(hdev);
3437
3438 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3439 if (!conn)
3440 goto unlock;
3441
3442 hci_conn_hold(conn);
3443 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3444 hci_conn_drop(conn);
3445
3446 set_bit(HCI_CONN_NEW_LINK_KEY, &conn->flags);
3447 conn_set_key(conn, ev->key_type, conn->pin_length);
3448
3449 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3450 goto unlock;
3451
3452 key = hci_add_link_key(hdev, conn, &ev->bdaddr, ev->link_key,
3453 ev->key_type, pin_len, &persistent);
3454 if (!key)
3455 goto unlock;
3456
3457 /* Update connection information since adding the key will have
3458 * fixed up the type in the case of changed combination keys.
3459 */
3460 if (ev->key_type == HCI_LK_CHANGED_COMBINATION)
3461 conn_set_key(conn, key->type, key->pin_len);
3462
3463 mgmt_new_link_key(hdev, key, persistent);
3464
3465 /* Keep debug keys around only if the HCI_KEEP_DEBUG_KEYS flag
3466 * is set. If it's not set simply remove the key from the kernel
3467 * list (we've still notified user space about it but with
3468 * store_hint being 0).
3469 */
3470 if (key->type == HCI_LK_DEBUG_COMBINATION &&
3471 !test_bit(HCI_KEEP_DEBUG_KEYS, &hdev->dev_flags)) {
3472 list_del_rcu(&key->list);
3473 kfree_rcu(key, rcu);
3474 goto unlock;
3475 }
3476
3477 if (persistent)
3478 clear_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3479 else
3480 set_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3481
3482 unlock:
3483 hci_dev_unlock(hdev);
3484 }
3485
3486 static void hci_clock_offset_evt(struct hci_dev *hdev, struct sk_buff *skb)
3487 {
3488 struct hci_ev_clock_offset *ev = (void *) skb->data;
3489 struct hci_conn *conn;
3490
3491 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3492
3493 hci_dev_lock(hdev);
3494
3495 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3496 if (conn && !ev->status) {
3497 struct inquiry_entry *ie;
3498
3499 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3500 if (ie) {
3501 ie->data.clock_offset = ev->clock_offset;
3502 ie->timestamp = jiffies;
3503 }
3504 }
3505
3506 hci_dev_unlock(hdev);
3507 }
3508
3509 static void hci_pkt_type_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3510 {
3511 struct hci_ev_pkt_type_change *ev = (void *) skb->data;
3512 struct hci_conn *conn;
3513
3514 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3515
3516 hci_dev_lock(hdev);
3517
3518 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3519 if (conn && !ev->status)
3520 conn->pkt_type = __le16_to_cpu(ev->pkt_type);
3521
3522 hci_dev_unlock(hdev);
3523 }
3524
3525 static void hci_pscan_rep_mode_evt(struct hci_dev *hdev, struct sk_buff *skb)
3526 {
3527 struct hci_ev_pscan_rep_mode *ev = (void *) skb->data;
3528 struct inquiry_entry *ie;
3529
3530 BT_DBG("%s", hdev->name);
3531
3532 hci_dev_lock(hdev);
3533
3534 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3535 if (ie) {
3536 ie->data.pscan_rep_mode = ev->pscan_rep_mode;
3537 ie->timestamp = jiffies;
3538 }
3539
3540 hci_dev_unlock(hdev);
3541 }
3542
3543 static void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev,
3544 struct sk_buff *skb)
3545 {
3546 struct inquiry_data data;
3547 int num_rsp = *((__u8 *) skb->data);
3548
3549 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3550
3551 if (!num_rsp)
3552 return;
3553
3554 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3555 return;
3556
3557 hci_dev_lock(hdev);
3558
3559 if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) {
3560 struct inquiry_info_with_rssi_and_pscan_mode *info;
3561 info = (void *) (skb->data + 1);
3562
3563 for (; num_rsp; num_rsp--, info++) {
3564 u32 flags;
3565
3566 bacpy(&data.bdaddr, &info->bdaddr);
3567 data.pscan_rep_mode = info->pscan_rep_mode;
3568 data.pscan_period_mode = info->pscan_period_mode;
3569 data.pscan_mode = info->pscan_mode;
3570 memcpy(data.dev_class, info->dev_class, 3);
3571 data.clock_offset = info->clock_offset;
3572 data.rssi = info->rssi;
3573 data.ssp_mode = 0x00;
3574
3575 flags = hci_inquiry_cache_update(hdev, &data, false);
3576
3577 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3578 info->dev_class, info->rssi,
3579 flags, NULL, 0, NULL, 0);
3580 }
3581 } else {
3582 struct inquiry_info_with_rssi *info = (void *) (skb->data + 1);
3583
3584 for (; num_rsp; num_rsp--, info++) {
3585 u32 flags;
3586
3587 bacpy(&data.bdaddr, &info->bdaddr);
3588 data.pscan_rep_mode = info->pscan_rep_mode;
3589 data.pscan_period_mode = info->pscan_period_mode;
3590 data.pscan_mode = 0x00;
3591 memcpy(data.dev_class, info->dev_class, 3);
3592 data.clock_offset = info->clock_offset;
3593 data.rssi = info->rssi;
3594 data.ssp_mode = 0x00;
3595
3596 flags = hci_inquiry_cache_update(hdev, &data, false);
3597
3598 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3599 info->dev_class, info->rssi,
3600 flags, NULL, 0, NULL, 0);
3601 }
3602 }
3603
3604 hci_dev_unlock(hdev);
3605 }
3606
3607 static void hci_remote_ext_features_evt(struct hci_dev *hdev,
3608 struct sk_buff *skb)
3609 {
3610 struct hci_ev_remote_ext_features *ev = (void *) skb->data;
3611 struct hci_conn *conn;
3612
3613 BT_DBG("%s", hdev->name);
3614
3615 hci_dev_lock(hdev);
3616
3617 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3618 if (!conn)
3619 goto unlock;
3620
3621 if (ev->page < HCI_MAX_PAGES)
3622 memcpy(conn->features[ev->page], ev->features, 8);
3623
3624 if (!ev->status && ev->page == 0x01) {
3625 struct inquiry_entry *ie;
3626
3627 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3628 if (ie)
3629 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3630
3631 if (ev->features[0] & LMP_HOST_SSP) {
3632 set_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3633 } else {
3634 /* It is mandatory by the Bluetooth specification that
3635 * Extended Inquiry Results are only used when Secure
3636 * Simple Pairing is enabled, but some devices violate
3637 * this.
3638 *
3639 * To make these devices work, the internal SSP
3640 * enabled flag needs to be cleared if the remote host
3641 * features do not indicate SSP support */
3642 clear_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3643 }
3644
3645 if (ev->features[0] & LMP_HOST_SC)
3646 set_bit(HCI_CONN_SC_ENABLED, &conn->flags);
3647 }
3648
3649 if (conn->state != BT_CONFIG)
3650 goto unlock;
3651
3652 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
3653 struct hci_cp_remote_name_req cp;
3654 memset(&cp, 0, sizeof(cp));
3655 bacpy(&cp.bdaddr, &conn->dst);
3656 cp.pscan_rep_mode = 0x02;
3657 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
3658 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
3659 mgmt_device_connected(hdev, conn, 0, NULL, 0);
3660
3661 if (!hci_outgoing_auth_needed(hdev, conn)) {
3662 conn->state = BT_CONNECTED;
3663 hci_proto_connect_cfm(conn, ev->status);
3664 hci_conn_drop(conn);
3665 }
3666
3667 unlock:
3668 hci_dev_unlock(hdev);
3669 }
3670
3671 static void hci_sync_conn_complete_evt(struct hci_dev *hdev,
3672 struct sk_buff *skb)
3673 {
3674 struct hci_ev_sync_conn_complete *ev = (void *) skb->data;
3675 struct hci_conn *conn;
3676
3677 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3678
3679 hci_dev_lock(hdev);
3680
3681 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
3682 if (!conn) {
3683 if (ev->link_type == ESCO_LINK)
3684 goto unlock;
3685
3686 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
3687 if (!conn)
3688 goto unlock;
3689
3690 conn->type = SCO_LINK;
3691 }
3692
3693 switch (ev->status) {
3694 case 0x00:
3695 conn->handle = __le16_to_cpu(ev->handle);
3696 conn->state = BT_CONNECTED;
3697
3698 hci_debugfs_create_conn(conn);
3699 hci_conn_add_sysfs(conn);
3700 break;
3701
3702 case 0x10: /* Connection Accept Timeout */
3703 case 0x0d: /* Connection Rejected due to Limited Resources */
3704 case 0x11: /* Unsupported Feature or Parameter Value */
3705 case 0x1c: /* SCO interval rejected */
3706 case 0x1a: /* Unsupported Remote Feature */
3707 case 0x1f: /* Unspecified error */
3708 case 0x20: /* Unsupported LMP Parameter value */
3709 if (conn->out) {
3710 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
3711 (hdev->esco_type & EDR_ESCO_MASK);
3712 if (hci_setup_sync(conn, conn->link->handle))
3713 goto unlock;
3714 }
3715 /* fall through */
3716
3717 default:
3718 conn->state = BT_CLOSED;
3719 break;
3720 }
3721
3722 hci_proto_connect_cfm(conn, ev->status);
3723 if (ev->status)
3724 hci_conn_del(conn);
3725
3726 unlock:
3727 hci_dev_unlock(hdev);
3728 }
3729
3730 static inline size_t eir_get_length(u8 *eir, size_t eir_len)
3731 {
3732 size_t parsed = 0;
3733
3734 while (parsed < eir_len) {
3735 u8 field_len = eir[0];
3736
3737 if (field_len == 0)
3738 return parsed;
3739
3740 parsed += field_len + 1;
3741 eir += field_len + 1;
3742 }
3743
3744 return eir_len;
3745 }
3746
3747 static void hci_extended_inquiry_result_evt(struct hci_dev *hdev,
3748 struct sk_buff *skb)
3749 {
3750 struct inquiry_data data;
3751 struct extended_inquiry_info *info = (void *) (skb->data + 1);
3752 int num_rsp = *((__u8 *) skb->data);
3753 size_t eir_len;
3754
3755 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3756
3757 if (!num_rsp)
3758 return;
3759
3760 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3761 return;
3762
3763 hci_dev_lock(hdev);
3764
3765 for (; num_rsp; num_rsp--, info++) {
3766 u32 flags;
3767 bool name_known;
3768
3769 bacpy(&data.bdaddr, &info->bdaddr);
3770 data.pscan_rep_mode = info->pscan_rep_mode;
3771 data.pscan_period_mode = info->pscan_period_mode;
3772 data.pscan_mode = 0x00;
3773 memcpy(data.dev_class, info->dev_class, 3);
3774 data.clock_offset = info->clock_offset;
3775 data.rssi = info->rssi;
3776 data.ssp_mode = 0x01;
3777
3778 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3779 name_known = eir_has_data_type(info->data,
3780 sizeof(info->data),
3781 EIR_NAME_COMPLETE);
3782 else
3783 name_known = true;
3784
3785 flags = hci_inquiry_cache_update(hdev, &data, name_known);
3786
3787 eir_len = eir_get_length(info->data, sizeof(info->data));
3788
3789 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3790 info->dev_class, info->rssi,
3791 flags, info->data, eir_len, NULL, 0);
3792 }
3793
3794 hci_dev_unlock(hdev);
3795 }
3796
3797 static void hci_key_refresh_complete_evt(struct hci_dev *hdev,
3798 struct sk_buff *skb)
3799 {
3800 struct hci_ev_key_refresh_complete *ev = (void *) skb->data;
3801 struct hci_conn *conn;
3802
3803 BT_DBG("%s status 0x%2.2x handle 0x%4.4x", hdev->name, ev->status,
3804 __le16_to_cpu(ev->handle));
3805
3806 hci_dev_lock(hdev);
3807
3808 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3809 if (!conn)
3810 goto unlock;
3811
3812 /* For BR/EDR the necessary steps are taken through the
3813 * auth_complete event.
3814 */
3815 if (conn->type != LE_LINK)
3816 goto unlock;
3817
3818 if (!ev->status)
3819 conn->sec_level = conn->pending_sec_level;
3820
3821 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
3822
3823 if (ev->status && conn->state == BT_CONNECTED) {
3824 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
3825 hci_conn_drop(conn);
3826 goto unlock;
3827 }
3828
3829 if (conn->state == BT_CONFIG) {
3830 if (!ev->status)
3831 conn->state = BT_CONNECTED;
3832
3833 hci_proto_connect_cfm(conn, ev->status);
3834 hci_conn_drop(conn);
3835 } else {
3836 hci_auth_cfm(conn, ev->status);
3837
3838 hci_conn_hold(conn);
3839 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3840 hci_conn_drop(conn);
3841 }
3842
3843 unlock:
3844 hci_dev_unlock(hdev);
3845 }
3846
3847 static u8 hci_get_auth_req(struct hci_conn *conn)
3848 {
3849 /* If remote requests no-bonding follow that lead */
3850 if (conn->remote_auth == HCI_AT_NO_BONDING ||
3851 conn->remote_auth == HCI_AT_NO_BONDING_MITM)
3852 return conn->remote_auth | (conn->auth_type & 0x01);
3853
3854 /* If both remote and local have enough IO capabilities, require
3855 * MITM protection
3856 */
3857 if (conn->remote_cap != HCI_IO_NO_INPUT_OUTPUT &&
3858 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT)
3859 return conn->remote_auth | 0x01;
3860
3861 /* No MITM protection possible so ignore remote requirement */
3862 return (conn->remote_auth & ~0x01) | (conn->auth_type & 0x01);
3863 }
3864
3865 static u8 bredr_oob_data_present(struct hci_conn *conn)
3866 {
3867 struct hci_dev *hdev = conn->hdev;
3868 struct oob_data *data;
3869
3870 data = hci_find_remote_oob_data(hdev, &conn->dst, BDADDR_BREDR);
3871 if (!data)
3872 return 0x00;
3873
3874 /* When Secure Connections Only mode is enabled, then the P-256
3875 * values are required. If they are not available, then do not
3876 * declare that OOB data is present.
3877 */
3878 if (bredr_sc_enabled(hdev) &&
3879 test_bit(HCI_SC_ONLY, &hdev->dev_flags) &&
3880 (!memcmp(data->rand256, ZERO_KEY, 16) ||
3881 !memcmp(data->hash256, ZERO_KEY, 16)))
3882 return 0x00;
3883
3884 if (conn->out || test_bit(HCI_CONN_REMOTE_OOB, &conn->flags))
3885 return 0x01;
3886
3887 return 0x00;
3888 }
3889
3890 static void hci_io_capa_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3891 {
3892 struct hci_ev_io_capa_request *ev = (void *) skb->data;
3893 struct hci_conn *conn;
3894
3895 BT_DBG("%s", hdev->name);
3896
3897 hci_dev_lock(hdev);
3898
3899 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3900 if (!conn)
3901 goto unlock;
3902
3903 hci_conn_hold(conn);
3904
3905 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3906 goto unlock;
3907
3908 /* Allow pairing if we're pairable, the initiators of the
3909 * pairing or if the remote is not requesting bonding.
3910 */
3911 if (test_bit(HCI_BONDABLE, &hdev->dev_flags) ||
3912 test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags) ||
3913 (conn->remote_auth & ~0x01) == HCI_AT_NO_BONDING) {
3914 struct hci_cp_io_capability_reply cp;
3915
3916 bacpy(&cp.bdaddr, &ev->bdaddr);
3917 /* Change the IO capability from KeyboardDisplay
3918 * to DisplayYesNo as it is not supported by BT spec. */
3919 cp.capability = (conn->io_capability == 0x04) ?
3920 HCI_IO_DISPLAY_YESNO : conn->io_capability;
3921
3922 /* If we are initiators, there is no remote information yet */
3923 if (conn->remote_auth == 0xff) {
3924 /* Request MITM protection if our IO caps allow it
3925 * except for the no-bonding case.
3926 */
3927 if (conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3928 conn->auth_type != HCI_AT_NO_BONDING)
3929 conn->auth_type |= 0x01;
3930 } else {
3931 conn->auth_type = hci_get_auth_req(conn);
3932 }
3933
3934 /* If we're not bondable, force one of the non-bondable
3935 * authentication requirement values.
3936 */
3937 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags))
3938 conn->auth_type &= HCI_AT_NO_BONDING_MITM;
3939
3940 cp.authentication = conn->auth_type;
3941 cp.oob_data = bredr_oob_data_present(conn);
3942
3943 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_REPLY,
3944 sizeof(cp), &cp);
3945 } else {
3946 struct hci_cp_io_capability_neg_reply cp;
3947
3948 bacpy(&cp.bdaddr, &ev->bdaddr);
3949 cp.reason = HCI_ERROR_PAIRING_NOT_ALLOWED;
3950
3951 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_NEG_REPLY,
3952 sizeof(cp), &cp);
3953 }
3954
3955 unlock:
3956 hci_dev_unlock(hdev);
3957 }
3958
3959 static void hci_io_capa_reply_evt(struct hci_dev *hdev, struct sk_buff *skb)
3960 {
3961 struct hci_ev_io_capa_reply *ev = (void *) skb->data;
3962 struct hci_conn *conn;
3963
3964 BT_DBG("%s", hdev->name);
3965
3966 hci_dev_lock(hdev);
3967
3968 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3969 if (!conn)
3970 goto unlock;
3971
3972 conn->remote_cap = ev->capability;
3973 conn->remote_auth = ev->authentication;
3974 if (ev->oob_data)
3975 set_bit(HCI_CONN_REMOTE_OOB, &conn->flags);
3976
3977 unlock:
3978 hci_dev_unlock(hdev);
3979 }
3980
3981 static void hci_user_confirm_request_evt(struct hci_dev *hdev,
3982 struct sk_buff *skb)
3983 {
3984 struct hci_ev_user_confirm_req *ev = (void *) skb->data;
3985 int loc_mitm, rem_mitm, confirm_hint = 0;
3986 struct hci_conn *conn;
3987
3988 BT_DBG("%s", hdev->name);
3989
3990 hci_dev_lock(hdev);
3991
3992 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3993 goto unlock;
3994
3995 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3996 if (!conn)
3997 goto unlock;
3998
3999 loc_mitm = (conn->auth_type & 0x01);
4000 rem_mitm = (conn->remote_auth & 0x01);
4001
4002 /* If we require MITM but the remote device can't provide that
4003 * (it has NoInputNoOutput) then reject the confirmation
4004 * request. We check the security level here since it doesn't
4005 * necessarily match conn->auth_type.
4006 */
4007 if (conn->pending_sec_level > BT_SECURITY_MEDIUM &&
4008 conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) {
4009 BT_DBG("Rejecting request: remote device can't provide MITM");
4010 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_NEG_REPLY,
4011 sizeof(ev->bdaddr), &ev->bdaddr);
4012 goto unlock;
4013 }
4014
4015 /* If no side requires MITM protection; auto-accept */
4016 if ((!loc_mitm || conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) &&
4017 (!rem_mitm || conn->io_capability == HCI_IO_NO_INPUT_OUTPUT)) {
4018
4019 /* If we're not the initiators request authorization to
4020 * proceed from user space (mgmt_user_confirm with
4021 * confirm_hint set to 1). The exception is if neither
4022 * side had MITM or if the local IO capability is
4023 * NoInputNoOutput, in which case we do auto-accept
4024 */
4025 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) &&
4026 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
4027 (loc_mitm || rem_mitm)) {
4028 BT_DBG("Confirming auto-accept as acceptor");
4029 confirm_hint = 1;
4030 goto confirm;
4031 }
4032
4033 BT_DBG("Auto-accept of user confirmation with %ums delay",
4034 hdev->auto_accept_delay);
4035
4036 if (hdev->auto_accept_delay > 0) {
4037 int delay = msecs_to_jiffies(hdev->auto_accept_delay);
4038 queue_delayed_work(conn->hdev->workqueue,
4039 &conn->auto_accept_work, delay);
4040 goto unlock;
4041 }
4042
4043 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY,
4044 sizeof(ev->bdaddr), &ev->bdaddr);
4045 goto unlock;
4046 }
4047
4048 confirm:
4049 mgmt_user_confirm_request(hdev, &ev->bdaddr, ACL_LINK, 0,
4050 le32_to_cpu(ev->passkey), confirm_hint);
4051
4052 unlock:
4053 hci_dev_unlock(hdev);
4054 }
4055
4056 static void hci_user_passkey_request_evt(struct hci_dev *hdev,
4057 struct sk_buff *skb)
4058 {
4059 struct hci_ev_user_passkey_req *ev = (void *) skb->data;
4060
4061 BT_DBG("%s", hdev->name);
4062
4063 if (test_bit(HCI_MGMT, &hdev->dev_flags))
4064 mgmt_user_passkey_request(hdev, &ev->bdaddr, ACL_LINK, 0);
4065 }
4066
4067 static void hci_user_passkey_notify_evt(struct hci_dev *hdev,
4068 struct sk_buff *skb)
4069 {
4070 struct hci_ev_user_passkey_notify *ev = (void *) skb->data;
4071 struct hci_conn *conn;
4072
4073 BT_DBG("%s", hdev->name);
4074
4075 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4076 if (!conn)
4077 return;
4078
4079 conn->passkey_notify = __le32_to_cpu(ev->passkey);
4080 conn->passkey_entered = 0;
4081
4082 if (test_bit(HCI_MGMT, &hdev->dev_flags))
4083 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
4084 conn->dst_type, conn->passkey_notify,
4085 conn->passkey_entered);
4086 }
4087
4088 static void hci_keypress_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
4089 {
4090 struct hci_ev_keypress_notify *ev = (void *) skb->data;
4091 struct hci_conn *conn;
4092
4093 BT_DBG("%s", hdev->name);
4094
4095 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4096 if (!conn)
4097 return;
4098
4099 switch (ev->type) {
4100 case HCI_KEYPRESS_STARTED:
4101 conn->passkey_entered = 0;
4102 return;
4103
4104 case HCI_KEYPRESS_ENTERED:
4105 conn->passkey_entered++;
4106 break;
4107
4108 case HCI_KEYPRESS_ERASED:
4109 conn->passkey_entered--;
4110 break;
4111
4112 case HCI_KEYPRESS_CLEARED:
4113 conn->passkey_entered = 0;
4114 break;
4115
4116 case HCI_KEYPRESS_COMPLETED:
4117 return;
4118 }
4119
4120 if (test_bit(HCI_MGMT, &hdev->dev_flags))
4121 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
4122 conn->dst_type, conn->passkey_notify,
4123 conn->passkey_entered);
4124 }
4125
4126 static void hci_simple_pair_complete_evt(struct hci_dev *hdev,
4127 struct sk_buff *skb)
4128 {
4129 struct hci_ev_simple_pair_complete *ev = (void *) skb->data;
4130 struct hci_conn *conn;
4131
4132 BT_DBG("%s", hdev->name);
4133
4134 hci_dev_lock(hdev);
4135
4136 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4137 if (!conn)
4138 goto unlock;
4139
4140 /* Reset the authentication requirement to unknown */
4141 conn->remote_auth = 0xff;
4142
4143 /* To avoid duplicate auth_failed events to user space we check
4144 * the HCI_CONN_AUTH_PEND flag which will be set if we
4145 * initiated the authentication. A traditional auth_complete
4146 * event gets always produced as initiator and is also mapped to
4147 * the mgmt_auth_failed event */
4148 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) && ev->status)
4149 mgmt_auth_failed(conn, ev->status);
4150
4151 hci_conn_drop(conn);
4152
4153 unlock:
4154 hci_dev_unlock(hdev);
4155 }
4156
4157 static void hci_remote_host_features_evt(struct hci_dev *hdev,
4158 struct sk_buff *skb)
4159 {
4160 struct hci_ev_remote_host_features *ev = (void *) skb->data;
4161 struct inquiry_entry *ie;
4162 struct hci_conn *conn;
4163
4164 BT_DBG("%s", hdev->name);
4165
4166 hci_dev_lock(hdev);
4167
4168 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4169 if (conn)
4170 memcpy(conn->features[1], ev->features, 8);
4171
4172 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
4173 if (ie)
4174 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
4175
4176 hci_dev_unlock(hdev);
4177 }
4178
4179 static void hci_remote_oob_data_request_evt(struct hci_dev *hdev,
4180 struct sk_buff *skb)
4181 {
4182 struct hci_ev_remote_oob_data_request *ev = (void *) skb->data;
4183 struct oob_data *data;
4184
4185 BT_DBG("%s", hdev->name);
4186
4187 hci_dev_lock(hdev);
4188
4189 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
4190 goto unlock;
4191
4192 data = hci_find_remote_oob_data(hdev, &ev->bdaddr, BDADDR_BREDR);
4193 if (!data) {
4194 struct hci_cp_remote_oob_data_neg_reply cp;
4195
4196 bacpy(&cp.bdaddr, &ev->bdaddr);
4197 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_NEG_REPLY,
4198 sizeof(cp), &cp);
4199 goto unlock;
4200 }
4201
4202 if (bredr_sc_enabled(hdev)) {
4203 struct hci_cp_remote_oob_ext_data_reply cp;
4204
4205 bacpy(&cp.bdaddr, &ev->bdaddr);
4206 if (test_bit(HCI_SC_ONLY, &hdev->dev_flags)) {
4207 memset(cp.hash192, 0, sizeof(cp.hash192));
4208 memset(cp.rand192, 0, sizeof(cp.rand192));
4209 } else {
4210 memcpy(cp.hash192, data->hash192, sizeof(cp.hash192));
4211 memcpy(cp.rand192, data->rand192, sizeof(cp.rand192));
4212 }
4213 memcpy(cp.hash256, data->hash256, sizeof(cp.hash256));
4214 memcpy(cp.rand256, data->rand256, sizeof(cp.rand256));
4215
4216 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_EXT_DATA_REPLY,
4217 sizeof(cp), &cp);
4218 } else {
4219 struct hci_cp_remote_oob_data_reply cp;
4220
4221 bacpy(&cp.bdaddr, &ev->bdaddr);
4222 memcpy(cp.hash, data->hash192, sizeof(cp.hash));
4223 memcpy(cp.rand, data->rand192, sizeof(cp.rand));
4224
4225 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_REPLY,
4226 sizeof(cp), &cp);
4227 }
4228
4229 unlock:
4230 hci_dev_unlock(hdev);
4231 }
4232
4233 static void hci_phy_link_complete_evt(struct hci_dev *hdev,
4234 struct sk_buff *skb)
4235 {
4236 struct hci_ev_phy_link_complete *ev = (void *) skb->data;
4237 struct hci_conn *hcon, *bredr_hcon;
4238
4239 BT_DBG("%s handle 0x%2.2x status 0x%2.2x", hdev->name, ev->phy_handle,
4240 ev->status);
4241
4242 hci_dev_lock(hdev);
4243
4244 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4245 if (!hcon) {
4246 hci_dev_unlock(hdev);
4247 return;
4248 }
4249
4250 if (ev->status) {
4251 hci_conn_del(hcon);
4252 hci_dev_unlock(hdev);
4253 return;
4254 }
4255
4256 bredr_hcon = hcon->amp_mgr->l2cap_conn->hcon;
4257
4258 hcon->state = BT_CONNECTED;
4259 bacpy(&hcon->dst, &bredr_hcon->dst);
4260
4261 hci_conn_hold(hcon);
4262 hcon->disc_timeout = HCI_DISCONN_TIMEOUT;
4263 hci_conn_drop(hcon);
4264
4265 hci_debugfs_create_conn(hcon);
4266 hci_conn_add_sysfs(hcon);
4267
4268 amp_physical_cfm(bredr_hcon, hcon);
4269
4270 hci_dev_unlock(hdev);
4271 }
4272
4273 static void hci_loglink_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4274 {
4275 struct hci_ev_logical_link_complete *ev = (void *) skb->data;
4276 struct hci_conn *hcon;
4277 struct hci_chan *hchan;
4278 struct amp_mgr *mgr;
4279
4280 BT_DBG("%s log_handle 0x%4.4x phy_handle 0x%2.2x status 0x%2.2x",
4281 hdev->name, le16_to_cpu(ev->handle), ev->phy_handle,
4282 ev->status);
4283
4284 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4285 if (!hcon)
4286 return;
4287
4288 /* Create AMP hchan */
4289 hchan = hci_chan_create(hcon);
4290 if (!hchan)
4291 return;
4292
4293 hchan->handle = le16_to_cpu(ev->handle);
4294
4295 BT_DBG("hcon %p mgr %p hchan %p", hcon, hcon->amp_mgr, hchan);
4296
4297 mgr = hcon->amp_mgr;
4298 if (mgr && mgr->bredr_chan) {
4299 struct l2cap_chan *bredr_chan = mgr->bredr_chan;
4300
4301 l2cap_chan_lock(bredr_chan);
4302
4303 bredr_chan->conn->mtu = hdev->block_mtu;
4304 l2cap_logical_cfm(bredr_chan, hchan, 0);
4305 hci_conn_hold(hcon);
4306
4307 l2cap_chan_unlock(bredr_chan);
4308 }
4309 }
4310
4311 static void hci_disconn_loglink_complete_evt(struct hci_dev *hdev,
4312 struct sk_buff *skb)
4313 {
4314 struct hci_ev_disconn_logical_link_complete *ev = (void *) skb->data;
4315 struct hci_chan *hchan;
4316
4317 BT_DBG("%s log handle 0x%4.4x status 0x%2.2x", hdev->name,
4318 le16_to_cpu(ev->handle), ev->status);
4319
4320 if (ev->status)
4321 return;
4322
4323 hci_dev_lock(hdev);
4324
4325 hchan = hci_chan_lookup_handle(hdev, le16_to_cpu(ev->handle));
4326 if (!hchan)
4327 goto unlock;
4328
4329 amp_destroy_logical_link(hchan, ev->reason);
4330
4331 unlock:
4332 hci_dev_unlock(hdev);
4333 }
4334
4335 static void hci_disconn_phylink_complete_evt(struct hci_dev *hdev,
4336 struct sk_buff *skb)
4337 {
4338 struct hci_ev_disconn_phy_link_complete *ev = (void *) skb->data;
4339 struct hci_conn *hcon;
4340
4341 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4342
4343 if (ev->status)
4344 return;
4345
4346 hci_dev_lock(hdev);
4347
4348 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4349 if (hcon) {
4350 hcon->state = BT_CLOSED;
4351 hci_conn_del(hcon);
4352 }
4353
4354 hci_dev_unlock(hdev);
4355 }
4356
4357 static void hci_le_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4358 {
4359 struct hci_ev_le_conn_complete *ev = (void *) skb->data;
4360 struct hci_conn_params *params;
4361 struct hci_conn *conn;
4362 struct smp_irk *irk;
4363 u8 addr_type;
4364
4365 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4366
4367 hci_dev_lock(hdev);
4368
4369 /* All controllers implicitly stop advertising in the event of a
4370 * connection, so ensure that the state bit is cleared.
4371 */
4372 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
4373
4374 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
4375 if (!conn) {
4376 conn = hci_conn_add(hdev, LE_LINK, &ev->bdaddr, ev->role);
4377 if (!conn) {
4378 BT_ERR("No memory for new connection");
4379 goto unlock;
4380 }
4381
4382 conn->dst_type = ev->bdaddr_type;
4383
4384 /* If we didn't have a hci_conn object previously
4385 * but we're in master role this must be something
4386 * initiated using a white list. Since white list based
4387 * connections are not "first class citizens" we don't
4388 * have full tracking of them. Therefore, we go ahead
4389 * with a "best effort" approach of determining the
4390 * initiator address based on the HCI_PRIVACY flag.
4391 */
4392 if (conn->out) {
4393 conn->resp_addr_type = ev->bdaddr_type;
4394 bacpy(&conn->resp_addr, &ev->bdaddr);
4395 if (test_bit(HCI_PRIVACY, &hdev->dev_flags)) {
4396 conn->init_addr_type = ADDR_LE_DEV_RANDOM;
4397 bacpy(&conn->init_addr, &hdev->rpa);
4398 } else {
4399 hci_copy_identity_address(hdev,
4400 &conn->init_addr,
4401 &conn->init_addr_type);
4402 }
4403 }
4404 } else {
4405 cancel_delayed_work(&conn->le_conn_timeout);
4406 }
4407
4408 if (!conn->out) {
4409 /* Set the responder (our side) address type based on
4410 * the advertising address type.
4411 */
4412 conn->resp_addr_type = hdev->adv_addr_type;
4413 if (hdev->adv_addr_type == ADDR_LE_DEV_RANDOM)
4414 bacpy(&conn->resp_addr, &hdev->random_addr);
4415 else
4416 bacpy(&conn->resp_addr, &hdev->bdaddr);
4417
4418 conn->init_addr_type = ev->bdaddr_type;
4419 bacpy(&conn->init_addr, &ev->bdaddr);
4420
4421 /* For incoming connections, set the default minimum
4422 * and maximum connection interval. They will be used
4423 * to check if the parameters are in range and if not
4424 * trigger the connection update procedure.
4425 */
4426 conn->le_conn_min_interval = hdev->le_conn_min_interval;
4427 conn->le_conn_max_interval = hdev->le_conn_max_interval;
4428 }
4429
4430 /* Lookup the identity address from the stored connection
4431 * address and address type.
4432 *
4433 * When establishing connections to an identity address, the
4434 * connection procedure will store the resolvable random
4435 * address first. Now if it can be converted back into the
4436 * identity address, start using the identity address from
4437 * now on.
4438 */
4439 irk = hci_get_irk(hdev, &conn->dst, conn->dst_type);
4440 if (irk) {
4441 bacpy(&conn->dst, &irk->bdaddr);
4442 conn->dst_type = irk->addr_type;
4443 }
4444
4445 if (ev->status) {
4446 hci_le_conn_failed(conn, ev->status);
4447 goto unlock;
4448 }
4449
4450 if (conn->dst_type == ADDR_LE_DEV_PUBLIC)
4451 addr_type = BDADDR_LE_PUBLIC;
4452 else
4453 addr_type = BDADDR_LE_RANDOM;
4454
4455 /* Drop the connection if the device is blocked */
4456 if (hci_bdaddr_list_lookup(&hdev->blacklist, &conn->dst, addr_type)) {
4457 hci_conn_drop(conn);
4458 goto unlock;
4459 }
4460
4461 if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4462 mgmt_device_connected(hdev, conn, 0, NULL, 0);
4463
4464 conn->sec_level = BT_SECURITY_LOW;
4465 conn->handle = __le16_to_cpu(ev->handle);
4466 conn->state = BT_CONNECTED;
4467
4468 conn->le_conn_interval = le16_to_cpu(ev->interval);
4469 conn->le_conn_latency = le16_to_cpu(ev->latency);
4470 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4471
4472 hci_debugfs_create_conn(conn);
4473 hci_conn_add_sysfs(conn);
4474
4475 hci_proto_connect_cfm(conn, ev->status);
4476
4477 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, &conn->dst,
4478 conn->dst_type);
4479 if (params) {
4480 list_del_init(&params->action);
4481 if (params->conn) {
4482 hci_conn_drop(params->conn);
4483 hci_conn_put(params->conn);
4484 params->conn = NULL;
4485 }
4486 }
4487
4488 unlock:
4489 hci_update_background_scan(hdev);
4490 hci_dev_unlock(hdev);
4491 }
4492
4493 static void hci_le_conn_update_complete_evt(struct hci_dev *hdev,
4494 struct sk_buff *skb)
4495 {
4496 struct hci_ev_le_conn_update_complete *ev = (void *) skb->data;
4497 struct hci_conn *conn;
4498
4499 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4500
4501 if (ev->status)
4502 return;
4503
4504 hci_dev_lock(hdev);
4505
4506 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4507 if (conn) {
4508 conn->le_conn_interval = le16_to_cpu(ev->interval);
4509 conn->le_conn_latency = le16_to_cpu(ev->latency);
4510 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4511 }
4512
4513 hci_dev_unlock(hdev);
4514 }
4515
4516 /* This function requires the caller holds hdev->lock */
4517 static struct hci_conn *check_pending_le_conn(struct hci_dev *hdev,
4518 bdaddr_t *addr,
4519 u8 addr_type, u8 adv_type)
4520 {
4521 struct hci_conn *conn;
4522 struct hci_conn_params *params;
4523
4524 /* If the event is not connectable don't proceed further */
4525 if (adv_type != LE_ADV_IND && adv_type != LE_ADV_DIRECT_IND)
4526 return NULL;
4527
4528 /* Ignore if the device is blocked */
4529 if (hci_bdaddr_list_lookup(&hdev->blacklist, addr, addr_type))
4530 return NULL;
4531
4532 /* Most controller will fail if we try to create new connections
4533 * while we have an existing one in slave role.
4534 */
4535 if (hdev->conn_hash.le_num_slave > 0)
4536 return NULL;
4537
4538 /* If we're not connectable only connect devices that we have in
4539 * our pend_le_conns list.
4540 */
4541 params = hci_pend_le_action_lookup(&hdev->pend_le_conns,
4542 addr, addr_type);
4543 if (!params)
4544 return NULL;
4545
4546 switch (params->auto_connect) {
4547 case HCI_AUTO_CONN_DIRECT:
4548 /* Only devices advertising with ADV_DIRECT_IND are
4549 * triggering a connection attempt. This is allowing
4550 * incoming connections from slave devices.
4551 */
4552 if (adv_type != LE_ADV_DIRECT_IND)
4553 return NULL;
4554 break;
4555 case HCI_AUTO_CONN_ALWAYS:
4556 /* Devices advertising with ADV_IND or ADV_DIRECT_IND
4557 * are triggering a connection attempt. This means
4558 * that incoming connectioms from slave device are
4559 * accepted and also outgoing connections to slave
4560 * devices are established when found.
4561 */
4562 break;
4563 default:
4564 return NULL;
4565 }
4566
4567 conn = hci_connect_le(hdev, addr, addr_type, BT_SECURITY_LOW,
4568 HCI_LE_AUTOCONN_TIMEOUT, HCI_ROLE_MASTER);
4569 if (!IS_ERR(conn)) {
4570 /* Store the pointer since we don't really have any
4571 * other owner of the object besides the params that
4572 * triggered it. This way we can abort the connection if
4573 * the parameters get removed and keep the reference
4574 * count consistent once the connection is established.
4575 */
4576 params->conn = hci_conn_get(conn);
4577 return conn;
4578 }
4579
4580 switch (PTR_ERR(conn)) {
4581 case -EBUSY:
4582 /* If hci_connect() returns -EBUSY it means there is already
4583 * an LE connection attempt going on. Since controllers don't
4584 * support more than one connection attempt at the time, we
4585 * don't consider this an error case.
4586 */
4587 break;
4588 default:
4589 BT_DBG("Failed to connect: err %ld", PTR_ERR(conn));
4590 return NULL;
4591 }
4592
4593 return NULL;
4594 }
4595
4596 static void process_adv_report(struct hci_dev *hdev, u8 type, bdaddr_t *bdaddr,
4597 u8 bdaddr_type, bdaddr_t *direct_addr,
4598 u8 direct_addr_type, s8 rssi, u8 *data, u8 len)
4599 {
4600 struct discovery_state *d = &hdev->discovery;
4601 struct smp_irk *irk;
4602 struct hci_conn *conn;
4603 bool match;
4604 u32 flags;
4605
4606 /* If the direct address is present, then this report is from
4607 * a LE Direct Advertising Report event. In that case it is
4608 * important to see if the address is matching the local
4609 * controller address.
4610 */
4611 if (direct_addr) {
4612 /* Only resolvable random addresses are valid for these
4613 * kind of reports and others can be ignored.
4614 */
4615 if (!hci_bdaddr_is_rpa(direct_addr, direct_addr_type))
4616 return;
4617
4618 /* If the controller is not using resolvable random
4619 * addresses, then this report can be ignored.
4620 */
4621 if (!test_bit(HCI_PRIVACY, &hdev->dev_flags))
4622 return;
4623
4624 /* If the local IRK of the controller does not match
4625 * with the resolvable random address provided, then
4626 * this report can be ignored.
4627 */
4628 if (!smp_irk_matches(hdev, hdev->irk, direct_addr))
4629 return;
4630 }
4631
4632 /* Check if we need to convert to identity address */
4633 irk = hci_get_irk(hdev, bdaddr, bdaddr_type);
4634 if (irk) {
4635 bdaddr = &irk->bdaddr;
4636 bdaddr_type = irk->addr_type;
4637 }
4638
4639 /* Check if we have been requested to connect to this device */
4640 conn = check_pending_le_conn(hdev, bdaddr, bdaddr_type, type);
4641 if (conn && type == LE_ADV_IND) {
4642 /* Store report for later inclusion by
4643 * mgmt_device_connected
4644 */
4645 memcpy(conn->le_adv_data, data, len);
4646 conn->le_adv_data_len = len;
4647 }
4648
4649 /* Passive scanning shouldn't trigger any device found events,
4650 * except for devices marked as CONN_REPORT for which we do send
4651 * device found events.
4652 */
4653 if (hdev->le_scan_type == LE_SCAN_PASSIVE) {
4654 if (type == LE_ADV_DIRECT_IND)
4655 return;
4656
4657 if (!hci_pend_le_action_lookup(&hdev->pend_le_reports,
4658 bdaddr, bdaddr_type))
4659 return;
4660
4661 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND)
4662 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4663 else
4664 flags = 0;
4665 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4666 rssi, flags, data, len, NULL, 0);
4667 return;
4668 }
4669
4670 /* When receiving non-connectable or scannable undirected
4671 * advertising reports, this means that the remote device is
4672 * not connectable and then clearly indicate this in the
4673 * device found event.
4674 *
4675 * When receiving a scan response, then there is no way to
4676 * know if the remote device is connectable or not. However
4677 * since scan responses are merged with a previously seen
4678 * advertising report, the flags field from that report
4679 * will be used.
4680 *
4681 * In the really unlikely case that a controller get confused
4682 * and just sends a scan response event, then it is marked as
4683 * not connectable as well.
4684 */
4685 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND ||
4686 type == LE_ADV_SCAN_RSP)
4687 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4688 else
4689 flags = 0;
4690
4691 /* If there's nothing pending either store the data from this
4692 * event or send an immediate device found event if the data
4693 * should not be stored for later.
4694 */
4695 if (!has_pending_adv_report(hdev)) {
4696 /* If the report will trigger a SCAN_REQ store it for
4697 * later merging.
4698 */
4699 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4700 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4701 rssi, flags, data, len);
4702 return;
4703 }
4704
4705 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4706 rssi, flags, data, len, NULL, 0);
4707 return;
4708 }
4709
4710 /* Check if the pending report is for the same device as the new one */
4711 match = (!bacmp(bdaddr, &d->last_adv_addr) &&
4712 bdaddr_type == d->last_adv_addr_type);
4713
4714 /* If the pending data doesn't match this report or this isn't a
4715 * scan response (e.g. we got a duplicate ADV_IND) then force
4716 * sending of the pending data.
4717 */
4718 if (type != LE_ADV_SCAN_RSP || !match) {
4719 /* Send out whatever is in the cache, but skip duplicates */
4720 if (!match)
4721 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4722 d->last_adv_addr_type, NULL,
4723 d->last_adv_rssi, d->last_adv_flags,
4724 d->last_adv_data,
4725 d->last_adv_data_len, NULL, 0);
4726
4727 /* If the new report will trigger a SCAN_REQ store it for
4728 * later merging.
4729 */
4730 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4731 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4732 rssi, flags, data, len);
4733 return;
4734 }
4735
4736 /* The advertising reports cannot be merged, so clear
4737 * the pending report and send out a device found event.
4738 */
4739 clear_pending_adv_report(hdev);
4740 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4741 rssi, flags, data, len, NULL, 0);
4742 return;
4743 }
4744
4745 /* If we get here we've got a pending ADV_IND or ADV_SCAN_IND and
4746 * the new event is a SCAN_RSP. We can therefore proceed with
4747 * sending a merged device found event.
4748 */
4749 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4750 d->last_adv_addr_type, NULL, rssi, d->last_adv_flags,
4751 d->last_adv_data, d->last_adv_data_len, data, len);
4752 clear_pending_adv_report(hdev);
4753 }
4754
4755 static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
4756 {
4757 u8 num_reports = skb->data[0];
4758 void *ptr = &skb->data[1];
4759
4760 hci_dev_lock(hdev);
4761
4762 while (num_reports--) {
4763 struct hci_ev_le_advertising_info *ev = ptr;
4764 s8 rssi;
4765
4766 rssi = ev->data[ev->length];
4767 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
4768 ev->bdaddr_type, NULL, 0, rssi,
4769 ev->data, ev->length);
4770
4771 ptr += sizeof(*ev) + ev->length + 1;
4772 }
4773
4774 hci_dev_unlock(hdev);
4775 }
4776
4777 static void hci_le_ltk_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
4778 {
4779 struct hci_ev_le_ltk_req *ev = (void *) skb->data;
4780 struct hci_cp_le_ltk_reply cp;
4781 struct hci_cp_le_ltk_neg_reply neg;
4782 struct hci_conn *conn;
4783 struct smp_ltk *ltk;
4784
4785 BT_DBG("%s handle 0x%4.4x", hdev->name, __le16_to_cpu(ev->handle));
4786
4787 hci_dev_lock(hdev);
4788
4789 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4790 if (conn == NULL)
4791 goto not_found;
4792
4793 ltk = hci_find_ltk(hdev, &conn->dst, conn->dst_type, conn->role);
4794 if (!ltk)
4795 goto not_found;
4796
4797 if (smp_ltk_is_sc(ltk)) {
4798 /* With SC both EDiv and Rand are set to zero */
4799 if (ev->ediv || ev->rand)
4800 goto not_found;
4801 } else {
4802 /* For non-SC keys check that EDiv and Rand match */
4803 if (ev->ediv != ltk->ediv || ev->rand != ltk->rand)
4804 goto not_found;
4805 }
4806
4807 memcpy(cp.ltk, ltk->val, sizeof(ltk->val));
4808 cp.handle = cpu_to_le16(conn->handle);
4809
4810 conn->pending_sec_level = smp_ltk_sec_level(ltk);
4811
4812 conn->enc_key_size = ltk->enc_size;
4813
4814 hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp);
4815
4816 /* Ref. Bluetooth Core SPEC pages 1975 and 2004. STK is a
4817 * temporary key used to encrypt a connection following
4818 * pairing. It is used during the Encrypted Session Setup to
4819 * distribute the keys. Later, security can be re-established
4820 * using a distributed LTK.
4821 */
4822 if (ltk->type == SMP_STK) {
4823 set_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4824 list_del_rcu(&ltk->list);
4825 kfree_rcu(ltk, rcu);
4826 } else {
4827 clear_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4828 }
4829
4830 hci_dev_unlock(hdev);
4831
4832 return;
4833
4834 not_found:
4835 neg.handle = ev->handle;
4836 hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(neg), &neg);
4837 hci_dev_unlock(hdev);
4838 }
4839
4840 static void send_conn_param_neg_reply(struct hci_dev *hdev, u16 handle,
4841 u8 reason)
4842 {
4843 struct hci_cp_le_conn_param_req_neg_reply cp;
4844
4845 cp.handle = cpu_to_le16(handle);
4846 cp.reason = reason;
4847
4848 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_NEG_REPLY, sizeof(cp),
4849 &cp);
4850 }
4851
4852 static void hci_le_remote_conn_param_req_evt(struct hci_dev *hdev,
4853 struct sk_buff *skb)
4854 {
4855 struct hci_ev_le_remote_conn_param_req *ev = (void *) skb->data;
4856 struct hci_cp_le_conn_param_req_reply cp;
4857 struct hci_conn *hcon;
4858 u16 handle, min, max, latency, timeout;
4859
4860 handle = le16_to_cpu(ev->handle);
4861 min = le16_to_cpu(ev->interval_min);
4862 max = le16_to_cpu(ev->interval_max);
4863 latency = le16_to_cpu(ev->latency);
4864 timeout = le16_to_cpu(ev->timeout);
4865
4866 hcon = hci_conn_hash_lookup_handle(hdev, handle);
4867 if (!hcon || hcon->state != BT_CONNECTED)
4868 return send_conn_param_neg_reply(hdev, handle,
4869 HCI_ERROR_UNKNOWN_CONN_ID);
4870
4871 if (hci_check_conn_params(min, max, latency, timeout))
4872 return send_conn_param_neg_reply(hdev, handle,
4873 HCI_ERROR_INVALID_LL_PARAMS);
4874
4875 if (hcon->role == HCI_ROLE_MASTER) {
4876 struct hci_conn_params *params;
4877 u8 store_hint;
4878
4879 hci_dev_lock(hdev);
4880
4881 params = hci_conn_params_lookup(hdev, &hcon->dst,
4882 hcon->dst_type);
4883 if (params) {
4884 params->conn_min_interval = min;
4885 params->conn_max_interval = max;
4886 params->conn_latency = latency;
4887 params->supervision_timeout = timeout;
4888 store_hint = 0x01;
4889 } else{
4890 store_hint = 0x00;
4891 }
4892
4893 hci_dev_unlock(hdev);
4894
4895 mgmt_new_conn_param(hdev, &hcon->dst, hcon->dst_type,
4896 store_hint, min, max, latency, timeout);
4897 }
4898
4899 cp.handle = ev->handle;
4900 cp.interval_min = ev->interval_min;
4901 cp.interval_max = ev->interval_max;
4902 cp.latency = ev->latency;
4903 cp.timeout = ev->timeout;
4904 cp.min_ce_len = 0;
4905 cp.max_ce_len = 0;
4906
4907 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_REPLY, sizeof(cp), &cp);
4908 }
4909
4910 static void hci_le_direct_adv_report_evt(struct hci_dev *hdev,
4911 struct sk_buff *skb)
4912 {
4913 u8 num_reports = skb->data[0];
4914 void *ptr = &skb->data[1];
4915
4916 hci_dev_lock(hdev);
4917
4918 while (num_reports--) {
4919 struct hci_ev_le_direct_adv_info *ev = ptr;
4920
4921 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
4922 ev->bdaddr_type, &ev->direct_addr,
4923 ev->direct_addr_type, ev->rssi, NULL, 0);
4924
4925 ptr += sizeof(*ev);
4926 }
4927
4928 hci_dev_unlock(hdev);
4929 }
4930
4931 static void hci_le_meta_evt(struct hci_dev *hdev, struct sk_buff *skb)
4932 {
4933 struct hci_ev_le_meta *le_ev = (void *) skb->data;
4934
4935 skb_pull(skb, sizeof(*le_ev));
4936
4937 switch (le_ev->subevent) {
4938 case HCI_EV_LE_CONN_COMPLETE:
4939 hci_le_conn_complete_evt(hdev, skb);
4940 break;
4941
4942 case HCI_EV_LE_CONN_UPDATE_COMPLETE:
4943 hci_le_conn_update_complete_evt(hdev, skb);
4944 break;
4945
4946 case HCI_EV_LE_ADVERTISING_REPORT:
4947 hci_le_adv_report_evt(hdev, skb);
4948 break;
4949
4950 case HCI_EV_LE_LTK_REQ:
4951 hci_le_ltk_request_evt(hdev, skb);
4952 break;
4953
4954 case HCI_EV_LE_REMOTE_CONN_PARAM_REQ:
4955 hci_le_remote_conn_param_req_evt(hdev, skb);
4956 break;
4957
4958 case HCI_EV_LE_DIRECT_ADV_REPORT:
4959 hci_le_direct_adv_report_evt(hdev, skb);
4960 break;
4961
4962 default:
4963 break;
4964 }
4965 }
4966
4967 static void hci_chan_selected_evt(struct hci_dev *hdev, struct sk_buff *skb)
4968 {
4969 struct hci_ev_channel_selected *ev = (void *) skb->data;
4970 struct hci_conn *hcon;
4971
4972 BT_DBG("%s handle 0x%2.2x", hdev->name, ev->phy_handle);
4973
4974 skb_pull(skb, sizeof(*ev));
4975
4976 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4977 if (!hcon)
4978 return;
4979
4980 amp_read_loc_assoc_final_data(hdev, hcon);
4981 }
4982
4983 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb)
4984 {
4985 struct hci_event_hdr *hdr = (void *) skb->data;
4986 __u8 event = hdr->evt;
4987
4988 hci_dev_lock(hdev);
4989
4990 /* Received events are (currently) only needed when a request is
4991 * ongoing so avoid unnecessary memory allocation.
4992 */
4993 if (hci_req_pending(hdev)) {
4994 kfree_skb(hdev->recv_evt);
4995 hdev->recv_evt = skb_clone(skb, GFP_KERNEL);
4996 }
4997
4998 hci_dev_unlock(hdev);
4999
5000 skb_pull(skb, HCI_EVENT_HDR_SIZE);
5001
5002 if (hdev->sent_cmd && bt_cb(hdev->sent_cmd)->req.event == event) {
5003 struct hci_command_hdr *cmd_hdr = (void *) hdev->sent_cmd->data;
5004 u16 opcode = __le16_to_cpu(cmd_hdr->opcode);
5005
5006 hci_req_cmd_complete(hdev, opcode, 0);
5007 }
5008
5009 switch (event) {
5010 case HCI_EV_INQUIRY_COMPLETE:
5011 hci_inquiry_complete_evt(hdev, skb);
5012 break;
5013
5014 case HCI_EV_INQUIRY_RESULT:
5015 hci_inquiry_result_evt(hdev, skb);
5016 break;
5017
5018 case HCI_EV_CONN_COMPLETE:
5019 hci_conn_complete_evt(hdev, skb);
5020 break;
5021
5022 case HCI_EV_CONN_REQUEST:
5023 hci_conn_request_evt(hdev, skb);
5024 break;
5025
5026 case HCI_EV_DISCONN_COMPLETE:
5027 hci_disconn_complete_evt(hdev, skb);
5028 break;
5029
5030 case HCI_EV_AUTH_COMPLETE:
5031 hci_auth_complete_evt(hdev, skb);
5032 break;
5033
5034 case HCI_EV_REMOTE_NAME:
5035 hci_remote_name_evt(hdev, skb);
5036 break;
5037
5038 case HCI_EV_ENCRYPT_CHANGE:
5039 hci_encrypt_change_evt(hdev, skb);
5040 break;
5041
5042 case HCI_EV_CHANGE_LINK_KEY_COMPLETE:
5043 hci_change_link_key_complete_evt(hdev, skb);
5044 break;
5045
5046 case HCI_EV_REMOTE_FEATURES:
5047 hci_remote_features_evt(hdev, skb);
5048 break;
5049
5050 case HCI_EV_CMD_COMPLETE:
5051 hci_cmd_complete_evt(hdev, skb);
5052 break;
5053
5054 case HCI_EV_CMD_STATUS:
5055 hci_cmd_status_evt(hdev, skb);
5056 break;
5057
5058 case HCI_EV_HARDWARE_ERROR:
5059 hci_hardware_error_evt(hdev, skb);
5060 break;
5061
5062 case HCI_EV_ROLE_CHANGE:
5063 hci_role_change_evt(hdev, skb);
5064 break;
5065
5066 case HCI_EV_NUM_COMP_PKTS:
5067 hci_num_comp_pkts_evt(hdev, skb);
5068 break;
5069
5070 case HCI_EV_MODE_CHANGE:
5071 hci_mode_change_evt(hdev, skb);
5072 break;
5073
5074 case HCI_EV_PIN_CODE_REQ:
5075 hci_pin_code_request_evt(hdev, skb);
5076 break;
5077
5078 case HCI_EV_LINK_KEY_REQ:
5079 hci_link_key_request_evt(hdev, skb);
5080 break;
5081
5082 case HCI_EV_LINK_KEY_NOTIFY:
5083 hci_link_key_notify_evt(hdev, skb);
5084 break;
5085
5086 case HCI_EV_CLOCK_OFFSET:
5087 hci_clock_offset_evt(hdev, skb);
5088 break;
5089
5090 case HCI_EV_PKT_TYPE_CHANGE:
5091 hci_pkt_type_change_evt(hdev, skb);
5092 break;
5093
5094 case HCI_EV_PSCAN_REP_MODE:
5095 hci_pscan_rep_mode_evt(hdev, skb);
5096 break;
5097
5098 case HCI_EV_INQUIRY_RESULT_WITH_RSSI:
5099 hci_inquiry_result_with_rssi_evt(hdev, skb);
5100 break;
5101
5102 case HCI_EV_REMOTE_EXT_FEATURES:
5103 hci_remote_ext_features_evt(hdev, skb);
5104 break;
5105
5106 case HCI_EV_SYNC_CONN_COMPLETE:
5107 hci_sync_conn_complete_evt(hdev, skb);
5108 break;
5109
5110 case HCI_EV_EXTENDED_INQUIRY_RESULT:
5111 hci_extended_inquiry_result_evt(hdev, skb);
5112 break;
5113
5114 case HCI_EV_KEY_REFRESH_COMPLETE:
5115 hci_key_refresh_complete_evt(hdev, skb);
5116 break;
5117
5118 case HCI_EV_IO_CAPA_REQUEST:
5119 hci_io_capa_request_evt(hdev, skb);
5120 break;
5121
5122 case HCI_EV_IO_CAPA_REPLY:
5123 hci_io_capa_reply_evt(hdev, skb);
5124 break;
5125
5126 case HCI_EV_USER_CONFIRM_REQUEST:
5127 hci_user_confirm_request_evt(hdev, skb);
5128 break;
5129
5130 case HCI_EV_USER_PASSKEY_REQUEST:
5131 hci_user_passkey_request_evt(hdev, skb);
5132 break;
5133
5134 case HCI_EV_USER_PASSKEY_NOTIFY:
5135 hci_user_passkey_notify_evt(hdev, skb);
5136 break;
5137
5138 case HCI_EV_KEYPRESS_NOTIFY:
5139 hci_keypress_notify_evt(hdev, skb);
5140 break;
5141
5142 case HCI_EV_SIMPLE_PAIR_COMPLETE:
5143 hci_simple_pair_complete_evt(hdev, skb);
5144 break;
5145
5146 case HCI_EV_REMOTE_HOST_FEATURES:
5147 hci_remote_host_features_evt(hdev, skb);
5148 break;
5149
5150 case HCI_EV_LE_META:
5151 hci_le_meta_evt(hdev, skb);
5152 break;
5153
5154 case HCI_EV_CHANNEL_SELECTED:
5155 hci_chan_selected_evt(hdev, skb);
5156 break;
5157
5158 case HCI_EV_REMOTE_OOB_DATA_REQUEST:
5159 hci_remote_oob_data_request_evt(hdev, skb);
5160 break;
5161
5162 case HCI_EV_PHY_LINK_COMPLETE:
5163 hci_phy_link_complete_evt(hdev, skb);
5164 break;
5165
5166 case HCI_EV_LOGICAL_LINK_COMPLETE:
5167 hci_loglink_complete_evt(hdev, skb);
5168 break;
5169
5170 case HCI_EV_DISCONN_LOGICAL_LINK_COMPLETE:
5171 hci_disconn_loglink_complete_evt(hdev, skb);
5172 break;
5173
5174 case HCI_EV_DISCONN_PHY_LINK_COMPLETE:
5175 hci_disconn_phylink_complete_evt(hdev, skb);
5176 break;
5177
5178 case HCI_EV_NUM_COMP_BLOCKS:
5179 hci_num_comp_blocks_evt(hdev, skb);
5180 break;
5181
5182 default:
5183 BT_DBG("%s event 0x%2.2x", hdev->name, event);
5184 break;
5185 }
5186
5187 kfree_skb(skb);
5188 hdev->stat.evt_rx++;
5189 }
This page took 0.131146 seconds and 6 git commands to generate.