Bluetooth: Fix SMP security level when we have no IO capabilities
[deliverable/linux.git] / net / bluetooth / hci_event.c
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 /* Bluetooth HCI event handling. */
26
27 #include <asm/unaligned.h>
28
29 #include <net/bluetooth/bluetooth.h>
30 #include <net/bluetooth/hci_core.h>
31 #include <net/bluetooth/mgmt.h>
32
33 #include "a2mp.h"
34 #include "amp.h"
35 #include "smp.h"
36
37 /* Handle HCI Event packets */
38
39 static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb)
40 {
41 __u8 status = *((__u8 *) skb->data);
42
43 BT_DBG("%s status 0x%2.2x", hdev->name, status);
44
45 if (status)
46 return;
47
48 clear_bit(HCI_INQUIRY, &hdev->flags);
49 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
50 wake_up_bit(&hdev->flags, HCI_INQUIRY);
51
52 hci_dev_lock(hdev);
53 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
54 hci_dev_unlock(hdev);
55
56 hci_conn_check_pending(hdev);
57 }
58
59 static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
60 {
61 __u8 status = *((__u8 *) skb->data);
62
63 BT_DBG("%s status 0x%2.2x", hdev->name, status);
64
65 if (status)
66 return;
67
68 set_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
69 }
70
71 static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
72 {
73 __u8 status = *((__u8 *) skb->data);
74
75 BT_DBG("%s status 0x%2.2x", hdev->name, status);
76
77 if (status)
78 return;
79
80 clear_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
81
82 hci_conn_check_pending(hdev);
83 }
84
85 static void hci_cc_remote_name_req_cancel(struct hci_dev *hdev,
86 struct sk_buff *skb)
87 {
88 BT_DBG("%s", hdev->name);
89 }
90
91 static void hci_cc_role_discovery(struct hci_dev *hdev, struct sk_buff *skb)
92 {
93 struct hci_rp_role_discovery *rp = (void *) skb->data;
94 struct hci_conn *conn;
95
96 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
97
98 if (rp->status)
99 return;
100
101 hci_dev_lock(hdev);
102
103 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
104 if (conn)
105 conn->role = rp->role;
106
107 hci_dev_unlock(hdev);
108 }
109
110 static void hci_cc_read_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
111 {
112 struct hci_rp_read_link_policy *rp = (void *) skb->data;
113 struct hci_conn *conn;
114
115 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
116
117 if (rp->status)
118 return;
119
120 hci_dev_lock(hdev);
121
122 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
123 if (conn)
124 conn->link_policy = __le16_to_cpu(rp->policy);
125
126 hci_dev_unlock(hdev);
127 }
128
129 static void hci_cc_write_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
130 {
131 struct hci_rp_write_link_policy *rp = (void *) skb->data;
132 struct hci_conn *conn;
133 void *sent;
134
135 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
136
137 if (rp->status)
138 return;
139
140 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LINK_POLICY);
141 if (!sent)
142 return;
143
144 hci_dev_lock(hdev);
145
146 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
147 if (conn)
148 conn->link_policy = get_unaligned_le16(sent + 2);
149
150 hci_dev_unlock(hdev);
151 }
152
153 static void hci_cc_read_def_link_policy(struct hci_dev *hdev,
154 struct sk_buff *skb)
155 {
156 struct hci_rp_read_def_link_policy *rp = (void *) skb->data;
157
158 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
159
160 if (rp->status)
161 return;
162
163 hdev->link_policy = __le16_to_cpu(rp->policy);
164 }
165
166 static void hci_cc_write_def_link_policy(struct hci_dev *hdev,
167 struct sk_buff *skb)
168 {
169 __u8 status = *((__u8 *) skb->data);
170 void *sent;
171
172 BT_DBG("%s status 0x%2.2x", hdev->name, status);
173
174 if (status)
175 return;
176
177 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_DEF_LINK_POLICY);
178 if (!sent)
179 return;
180
181 hdev->link_policy = get_unaligned_le16(sent);
182 }
183
184 static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb)
185 {
186 __u8 status = *((__u8 *) skb->data);
187
188 BT_DBG("%s status 0x%2.2x", hdev->name, status);
189
190 clear_bit(HCI_RESET, &hdev->flags);
191
192 /* Reset all non-persistent flags */
193 hdev->dev_flags &= ~HCI_PERSISTENT_MASK;
194
195 hdev->discovery.state = DISCOVERY_STOPPED;
196 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
197 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
198
199 memset(hdev->adv_data, 0, sizeof(hdev->adv_data));
200 hdev->adv_data_len = 0;
201
202 memset(hdev->scan_rsp_data, 0, sizeof(hdev->scan_rsp_data));
203 hdev->scan_rsp_data_len = 0;
204
205 hdev->le_scan_type = LE_SCAN_PASSIVE;
206
207 hdev->ssp_debug_mode = 0;
208 }
209
210 static void hci_cc_write_local_name(struct hci_dev *hdev, struct sk_buff *skb)
211 {
212 __u8 status = *((__u8 *) skb->data);
213 void *sent;
214
215 BT_DBG("%s status 0x%2.2x", hdev->name, status);
216
217 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LOCAL_NAME);
218 if (!sent)
219 return;
220
221 hci_dev_lock(hdev);
222
223 if (test_bit(HCI_MGMT, &hdev->dev_flags))
224 mgmt_set_local_name_complete(hdev, sent, status);
225 else if (!status)
226 memcpy(hdev->dev_name, sent, HCI_MAX_NAME_LENGTH);
227
228 hci_dev_unlock(hdev);
229 }
230
231 static void hci_cc_read_local_name(struct hci_dev *hdev, struct sk_buff *skb)
232 {
233 struct hci_rp_read_local_name *rp = (void *) skb->data;
234
235 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
236
237 if (rp->status)
238 return;
239
240 if (test_bit(HCI_SETUP, &hdev->dev_flags))
241 memcpy(hdev->dev_name, rp->name, HCI_MAX_NAME_LENGTH);
242 }
243
244 static void hci_cc_write_auth_enable(struct hci_dev *hdev, struct sk_buff *skb)
245 {
246 __u8 status = *((__u8 *) skb->data);
247 void *sent;
248
249 BT_DBG("%s status 0x%2.2x", hdev->name, status);
250
251 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_AUTH_ENABLE);
252 if (!sent)
253 return;
254
255 if (!status) {
256 __u8 param = *((__u8 *) sent);
257
258 if (param == AUTH_ENABLED)
259 set_bit(HCI_AUTH, &hdev->flags);
260 else
261 clear_bit(HCI_AUTH, &hdev->flags);
262 }
263
264 if (test_bit(HCI_MGMT, &hdev->dev_flags))
265 mgmt_auth_enable_complete(hdev, status);
266 }
267
268 static void hci_cc_write_encrypt_mode(struct hci_dev *hdev, struct sk_buff *skb)
269 {
270 __u8 status = *((__u8 *) skb->data);
271 __u8 param;
272 void *sent;
273
274 BT_DBG("%s status 0x%2.2x", hdev->name, status);
275
276 if (status)
277 return;
278
279 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_ENCRYPT_MODE);
280 if (!sent)
281 return;
282
283 param = *((__u8 *) sent);
284
285 if (param)
286 set_bit(HCI_ENCRYPT, &hdev->flags);
287 else
288 clear_bit(HCI_ENCRYPT, &hdev->flags);
289 }
290
291 static void hci_cc_write_scan_enable(struct hci_dev *hdev, struct sk_buff *skb)
292 {
293 __u8 status = *((__u8 *) skb->data);
294 __u8 param;
295 void *sent;
296
297 BT_DBG("%s status 0x%2.2x", hdev->name, status);
298
299 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SCAN_ENABLE);
300 if (!sent)
301 return;
302
303 param = *((__u8 *) sent);
304
305 hci_dev_lock(hdev);
306
307 if (status) {
308 hdev->discov_timeout = 0;
309 goto done;
310 }
311
312 if (param & SCAN_INQUIRY)
313 set_bit(HCI_ISCAN, &hdev->flags);
314 else
315 clear_bit(HCI_ISCAN, &hdev->flags);
316
317 if (param & SCAN_PAGE)
318 set_bit(HCI_PSCAN, &hdev->flags);
319 else
320 clear_bit(HCI_PSCAN, &hdev->flags);
321
322 done:
323 hci_dev_unlock(hdev);
324 }
325
326 static void hci_cc_read_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
327 {
328 struct hci_rp_read_class_of_dev *rp = (void *) skb->data;
329
330 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
331
332 if (rp->status)
333 return;
334
335 memcpy(hdev->dev_class, rp->dev_class, 3);
336
337 BT_DBG("%s class 0x%.2x%.2x%.2x", hdev->name,
338 hdev->dev_class[2], hdev->dev_class[1], hdev->dev_class[0]);
339 }
340
341 static void hci_cc_write_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
342 {
343 __u8 status = *((__u8 *) skb->data);
344 void *sent;
345
346 BT_DBG("%s status 0x%2.2x", hdev->name, status);
347
348 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_CLASS_OF_DEV);
349 if (!sent)
350 return;
351
352 hci_dev_lock(hdev);
353
354 if (status == 0)
355 memcpy(hdev->dev_class, sent, 3);
356
357 if (test_bit(HCI_MGMT, &hdev->dev_flags))
358 mgmt_set_class_of_dev_complete(hdev, sent, status);
359
360 hci_dev_unlock(hdev);
361 }
362
363 static void hci_cc_read_voice_setting(struct hci_dev *hdev, struct sk_buff *skb)
364 {
365 struct hci_rp_read_voice_setting *rp = (void *) skb->data;
366 __u16 setting;
367
368 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
369
370 if (rp->status)
371 return;
372
373 setting = __le16_to_cpu(rp->voice_setting);
374
375 if (hdev->voice_setting == setting)
376 return;
377
378 hdev->voice_setting = setting;
379
380 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
381
382 if (hdev->notify)
383 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
384 }
385
386 static void hci_cc_write_voice_setting(struct hci_dev *hdev,
387 struct sk_buff *skb)
388 {
389 __u8 status = *((__u8 *) skb->data);
390 __u16 setting;
391 void *sent;
392
393 BT_DBG("%s status 0x%2.2x", hdev->name, status);
394
395 if (status)
396 return;
397
398 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_VOICE_SETTING);
399 if (!sent)
400 return;
401
402 setting = get_unaligned_le16(sent);
403
404 if (hdev->voice_setting == setting)
405 return;
406
407 hdev->voice_setting = setting;
408
409 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
410
411 if (hdev->notify)
412 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
413 }
414
415 static void hci_cc_read_num_supported_iac(struct hci_dev *hdev,
416 struct sk_buff *skb)
417 {
418 struct hci_rp_read_num_supported_iac *rp = (void *) skb->data;
419
420 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
421
422 if (rp->status)
423 return;
424
425 hdev->num_iac = rp->num_iac;
426
427 BT_DBG("%s num iac %d", hdev->name, hdev->num_iac);
428 }
429
430 static void hci_cc_write_ssp_mode(struct hci_dev *hdev, struct sk_buff *skb)
431 {
432 __u8 status = *((__u8 *) skb->data);
433 struct hci_cp_write_ssp_mode *sent;
434
435 BT_DBG("%s status 0x%2.2x", hdev->name, status);
436
437 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_MODE);
438 if (!sent)
439 return;
440
441 if (!status) {
442 if (sent->mode)
443 hdev->features[1][0] |= LMP_HOST_SSP;
444 else
445 hdev->features[1][0] &= ~LMP_HOST_SSP;
446 }
447
448 if (test_bit(HCI_MGMT, &hdev->dev_flags))
449 mgmt_ssp_enable_complete(hdev, sent->mode, status);
450 else if (!status) {
451 if (sent->mode)
452 set_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
453 else
454 clear_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
455 }
456 }
457
458 static void hci_cc_write_sc_support(struct hci_dev *hdev, struct sk_buff *skb)
459 {
460 u8 status = *((u8 *) skb->data);
461 struct hci_cp_write_sc_support *sent;
462
463 BT_DBG("%s status 0x%2.2x", hdev->name, status);
464
465 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SC_SUPPORT);
466 if (!sent)
467 return;
468
469 if (!status) {
470 if (sent->support)
471 hdev->features[1][0] |= LMP_HOST_SC;
472 else
473 hdev->features[1][0] &= ~LMP_HOST_SC;
474 }
475
476 if (test_bit(HCI_MGMT, &hdev->dev_flags))
477 mgmt_sc_enable_complete(hdev, sent->support, status);
478 else if (!status) {
479 if (sent->support)
480 set_bit(HCI_SC_ENABLED, &hdev->dev_flags);
481 else
482 clear_bit(HCI_SC_ENABLED, &hdev->dev_flags);
483 }
484 }
485
486 static void hci_cc_read_local_version(struct hci_dev *hdev, struct sk_buff *skb)
487 {
488 struct hci_rp_read_local_version *rp = (void *) skb->data;
489
490 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
491
492 if (rp->status)
493 return;
494
495 if (test_bit(HCI_SETUP, &hdev->dev_flags)) {
496 hdev->hci_ver = rp->hci_ver;
497 hdev->hci_rev = __le16_to_cpu(rp->hci_rev);
498 hdev->lmp_ver = rp->lmp_ver;
499 hdev->manufacturer = __le16_to_cpu(rp->manufacturer);
500 hdev->lmp_subver = __le16_to_cpu(rp->lmp_subver);
501 }
502 }
503
504 static void hci_cc_read_local_commands(struct hci_dev *hdev,
505 struct sk_buff *skb)
506 {
507 struct hci_rp_read_local_commands *rp = (void *) skb->data;
508
509 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
510
511 if (rp->status)
512 return;
513
514 if (test_bit(HCI_SETUP, &hdev->dev_flags))
515 memcpy(hdev->commands, rp->commands, sizeof(hdev->commands));
516 }
517
518 static void hci_cc_read_local_features(struct hci_dev *hdev,
519 struct sk_buff *skb)
520 {
521 struct hci_rp_read_local_features *rp = (void *) skb->data;
522
523 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
524
525 if (rp->status)
526 return;
527
528 memcpy(hdev->features, rp->features, 8);
529
530 /* Adjust default settings according to features
531 * supported by device. */
532
533 if (hdev->features[0][0] & LMP_3SLOT)
534 hdev->pkt_type |= (HCI_DM3 | HCI_DH3);
535
536 if (hdev->features[0][0] & LMP_5SLOT)
537 hdev->pkt_type |= (HCI_DM5 | HCI_DH5);
538
539 if (hdev->features[0][1] & LMP_HV2) {
540 hdev->pkt_type |= (HCI_HV2);
541 hdev->esco_type |= (ESCO_HV2);
542 }
543
544 if (hdev->features[0][1] & LMP_HV3) {
545 hdev->pkt_type |= (HCI_HV3);
546 hdev->esco_type |= (ESCO_HV3);
547 }
548
549 if (lmp_esco_capable(hdev))
550 hdev->esco_type |= (ESCO_EV3);
551
552 if (hdev->features[0][4] & LMP_EV4)
553 hdev->esco_type |= (ESCO_EV4);
554
555 if (hdev->features[0][4] & LMP_EV5)
556 hdev->esco_type |= (ESCO_EV5);
557
558 if (hdev->features[0][5] & LMP_EDR_ESCO_2M)
559 hdev->esco_type |= (ESCO_2EV3);
560
561 if (hdev->features[0][5] & LMP_EDR_ESCO_3M)
562 hdev->esco_type |= (ESCO_3EV3);
563
564 if (hdev->features[0][5] & LMP_EDR_3S_ESCO)
565 hdev->esco_type |= (ESCO_2EV5 | ESCO_3EV5);
566 }
567
568 static void hci_cc_read_local_ext_features(struct hci_dev *hdev,
569 struct sk_buff *skb)
570 {
571 struct hci_rp_read_local_ext_features *rp = (void *) skb->data;
572
573 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
574
575 if (rp->status)
576 return;
577
578 if (hdev->max_page < rp->max_page)
579 hdev->max_page = rp->max_page;
580
581 if (rp->page < HCI_MAX_PAGES)
582 memcpy(hdev->features[rp->page], rp->features, 8);
583 }
584
585 static void hci_cc_read_flow_control_mode(struct hci_dev *hdev,
586 struct sk_buff *skb)
587 {
588 struct hci_rp_read_flow_control_mode *rp = (void *) skb->data;
589
590 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
591
592 if (rp->status)
593 return;
594
595 hdev->flow_ctl_mode = rp->mode;
596 }
597
598 static void hci_cc_read_buffer_size(struct hci_dev *hdev, struct sk_buff *skb)
599 {
600 struct hci_rp_read_buffer_size *rp = (void *) skb->data;
601
602 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
603
604 if (rp->status)
605 return;
606
607 hdev->acl_mtu = __le16_to_cpu(rp->acl_mtu);
608 hdev->sco_mtu = rp->sco_mtu;
609 hdev->acl_pkts = __le16_to_cpu(rp->acl_max_pkt);
610 hdev->sco_pkts = __le16_to_cpu(rp->sco_max_pkt);
611
612 if (test_bit(HCI_QUIRK_FIXUP_BUFFER_SIZE, &hdev->quirks)) {
613 hdev->sco_mtu = 64;
614 hdev->sco_pkts = 8;
615 }
616
617 hdev->acl_cnt = hdev->acl_pkts;
618 hdev->sco_cnt = hdev->sco_pkts;
619
620 BT_DBG("%s acl mtu %d:%d sco mtu %d:%d", hdev->name, hdev->acl_mtu,
621 hdev->acl_pkts, hdev->sco_mtu, hdev->sco_pkts);
622 }
623
624 static void hci_cc_read_bd_addr(struct hci_dev *hdev, struct sk_buff *skb)
625 {
626 struct hci_rp_read_bd_addr *rp = (void *) skb->data;
627
628 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
629
630 if (rp->status)
631 return;
632
633 if (test_bit(HCI_INIT, &hdev->flags))
634 bacpy(&hdev->bdaddr, &rp->bdaddr);
635
636 if (test_bit(HCI_SETUP, &hdev->dev_flags))
637 bacpy(&hdev->setup_addr, &rp->bdaddr);
638 }
639
640 static void hci_cc_read_page_scan_activity(struct hci_dev *hdev,
641 struct sk_buff *skb)
642 {
643 struct hci_rp_read_page_scan_activity *rp = (void *) skb->data;
644
645 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
646
647 if (rp->status)
648 return;
649
650 if (test_bit(HCI_INIT, &hdev->flags)) {
651 hdev->page_scan_interval = __le16_to_cpu(rp->interval);
652 hdev->page_scan_window = __le16_to_cpu(rp->window);
653 }
654 }
655
656 static void hci_cc_write_page_scan_activity(struct hci_dev *hdev,
657 struct sk_buff *skb)
658 {
659 u8 status = *((u8 *) skb->data);
660 struct hci_cp_write_page_scan_activity *sent;
661
662 BT_DBG("%s status 0x%2.2x", hdev->name, status);
663
664 if (status)
665 return;
666
667 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY);
668 if (!sent)
669 return;
670
671 hdev->page_scan_interval = __le16_to_cpu(sent->interval);
672 hdev->page_scan_window = __le16_to_cpu(sent->window);
673 }
674
675 static void hci_cc_read_page_scan_type(struct hci_dev *hdev,
676 struct sk_buff *skb)
677 {
678 struct hci_rp_read_page_scan_type *rp = (void *) skb->data;
679
680 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
681
682 if (rp->status)
683 return;
684
685 if (test_bit(HCI_INIT, &hdev->flags))
686 hdev->page_scan_type = rp->type;
687 }
688
689 static void hci_cc_write_page_scan_type(struct hci_dev *hdev,
690 struct sk_buff *skb)
691 {
692 u8 status = *((u8 *) skb->data);
693 u8 *type;
694
695 BT_DBG("%s status 0x%2.2x", hdev->name, status);
696
697 if (status)
698 return;
699
700 type = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_TYPE);
701 if (type)
702 hdev->page_scan_type = *type;
703 }
704
705 static void hci_cc_read_data_block_size(struct hci_dev *hdev,
706 struct sk_buff *skb)
707 {
708 struct hci_rp_read_data_block_size *rp = (void *) skb->data;
709
710 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
711
712 if (rp->status)
713 return;
714
715 hdev->block_mtu = __le16_to_cpu(rp->max_acl_len);
716 hdev->block_len = __le16_to_cpu(rp->block_len);
717 hdev->num_blocks = __le16_to_cpu(rp->num_blocks);
718
719 hdev->block_cnt = hdev->num_blocks;
720
721 BT_DBG("%s blk mtu %d cnt %d len %d", hdev->name, hdev->block_mtu,
722 hdev->block_cnt, hdev->block_len);
723 }
724
725 static void hci_cc_read_clock(struct hci_dev *hdev, struct sk_buff *skb)
726 {
727 struct hci_rp_read_clock *rp = (void *) skb->data;
728 struct hci_cp_read_clock *cp;
729 struct hci_conn *conn;
730
731 BT_DBG("%s", hdev->name);
732
733 if (skb->len < sizeof(*rp))
734 return;
735
736 if (rp->status)
737 return;
738
739 hci_dev_lock(hdev);
740
741 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_CLOCK);
742 if (!cp)
743 goto unlock;
744
745 if (cp->which == 0x00) {
746 hdev->clock = le32_to_cpu(rp->clock);
747 goto unlock;
748 }
749
750 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
751 if (conn) {
752 conn->clock = le32_to_cpu(rp->clock);
753 conn->clock_accuracy = le16_to_cpu(rp->accuracy);
754 }
755
756 unlock:
757 hci_dev_unlock(hdev);
758 }
759
760 static void hci_cc_read_local_amp_info(struct hci_dev *hdev,
761 struct sk_buff *skb)
762 {
763 struct hci_rp_read_local_amp_info *rp = (void *) skb->data;
764
765 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
766
767 if (rp->status)
768 goto a2mp_rsp;
769
770 hdev->amp_status = rp->amp_status;
771 hdev->amp_total_bw = __le32_to_cpu(rp->total_bw);
772 hdev->amp_max_bw = __le32_to_cpu(rp->max_bw);
773 hdev->amp_min_latency = __le32_to_cpu(rp->min_latency);
774 hdev->amp_max_pdu = __le32_to_cpu(rp->max_pdu);
775 hdev->amp_type = rp->amp_type;
776 hdev->amp_pal_cap = __le16_to_cpu(rp->pal_cap);
777 hdev->amp_assoc_size = __le16_to_cpu(rp->max_assoc_size);
778 hdev->amp_be_flush_to = __le32_to_cpu(rp->be_flush_to);
779 hdev->amp_max_flush_to = __le32_to_cpu(rp->max_flush_to);
780
781 a2mp_rsp:
782 a2mp_send_getinfo_rsp(hdev);
783 }
784
785 static void hci_cc_read_local_amp_assoc(struct hci_dev *hdev,
786 struct sk_buff *skb)
787 {
788 struct hci_rp_read_local_amp_assoc *rp = (void *) skb->data;
789 struct amp_assoc *assoc = &hdev->loc_assoc;
790 size_t rem_len, frag_len;
791
792 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
793
794 if (rp->status)
795 goto a2mp_rsp;
796
797 frag_len = skb->len - sizeof(*rp);
798 rem_len = __le16_to_cpu(rp->rem_len);
799
800 if (rem_len > frag_len) {
801 BT_DBG("frag_len %zu rem_len %zu", frag_len, rem_len);
802
803 memcpy(assoc->data + assoc->offset, rp->frag, frag_len);
804 assoc->offset += frag_len;
805
806 /* Read other fragments */
807 amp_read_loc_assoc_frag(hdev, rp->phy_handle);
808
809 return;
810 }
811
812 memcpy(assoc->data + assoc->offset, rp->frag, rem_len);
813 assoc->len = assoc->offset + rem_len;
814 assoc->offset = 0;
815
816 a2mp_rsp:
817 /* Send A2MP Rsp when all fragments are received */
818 a2mp_send_getampassoc_rsp(hdev, rp->status);
819 a2mp_send_create_phy_link_req(hdev, rp->status);
820 }
821
822 static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev,
823 struct sk_buff *skb)
824 {
825 struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data;
826
827 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
828
829 if (rp->status)
830 return;
831
832 hdev->inq_tx_power = rp->tx_power;
833 }
834
835 static void hci_cc_pin_code_reply(struct hci_dev *hdev, struct sk_buff *skb)
836 {
837 struct hci_rp_pin_code_reply *rp = (void *) skb->data;
838 struct hci_cp_pin_code_reply *cp;
839 struct hci_conn *conn;
840
841 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
842
843 hci_dev_lock(hdev);
844
845 if (test_bit(HCI_MGMT, &hdev->dev_flags))
846 mgmt_pin_code_reply_complete(hdev, &rp->bdaddr, rp->status);
847
848 if (rp->status)
849 goto unlock;
850
851 cp = hci_sent_cmd_data(hdev, HCI_OP_PIN_CODE_REPLY);
852 if (!cp)
853 goto unlock;
854
855 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
856 if (conn)
857 conn->pin_length = cp->pin_len;
858
859 unlock:
860 hci_dev_unlock(hdev);
861 }
862
863 static void hci_cc_pin_code_neg_reply(struct hci_dev *hdev, struct sk_buff *skb)
864 {
865 struct hci_rp_pin_code_neg_reply *rp = (void *) skb->data;
866
867 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
868
869 hci_dev_lock(hdev);
870
871 if (test_bit(HCI_MGMT, &hdev->dev_flags))
872 mgmt_pin_code_neg_reply_complete(hdev, &rp->bdaddr,
873 rp->status);
874
875 hci_dev_unlock(hdev);
876 }
877
878 static void hci_cc_le_read_buffer_size(struct hci_dev *hdev,
879 struct sk_buff *skb)
880 {
881 struct hci_rp_le_read_buffer_size *rp = (void *) skb->data;
882
883 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
884
885 if (rp->status)
886 return;
887
888 hdev->le_mtu = __le16_to_cpu(rp->le_mtu);
889 hdev->le_pkts = rp->le_max_pkt;
890
891 hdev->le_cnt = hdev->le_pkts;
892
893 BT_DBG("%s le mtu %d:%d", hdev->name, hdev->le_mtu, hdev->le_pkts);
894 }
895
896 static void hci_cc_le_read_local_features(struct hci_dev *hdev,
897 struct sk_buff *skb)
898 {
899 struct hci_rp_le_read_local_features *rp = (void *) skb->data;
900
901 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
902
903 if (rp->status)
904 return;
905
906 memcpy(hdev->le_features, rp->features, 8);
907 }
908
909 static void hci_cc_le_read_adv_tx_power(struct hci_dev *hdev,
910 struct sk_buff *skb)
911 {
912 struct hci_rp_le_read_adv_tx_power *rp = (void *) skb->data;
913
914 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
915
916 if (rp->status)
917 return;
918
919 hdev->adv_tx_power = rp->tx_power;
920 }
921
922 static void hci_cc_user_confirm_reply(struct hci_dev *hdev, struct sk_buff *skb)
923 {
924 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
925
926 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
927
928 hci_dev_lock(hdev);
929
930 if (test_bit(HCI_MGMT, &hdev->dev_flags))
931 mgmt_user_confirm_reply_complete(hdev, &rp->bdaddr, ACL_LINK, 0,
932 rp->status);
933
934 hci_dev_unlock(hdev);
935 }
936
937 static void hci_cc_user_confirm_neg_reply(struct hci_dev *hdev,
938 struct sk_buff *skb)
939 {
940 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
941
942 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
943
944 hci_dev_lock(hdev);
945
946 if (test_bit(HCI_MGMT, &hdev->dev_flags))
947 mgmt_user_confirm_neg_reply_complete(hdev, &rp->bdaddr,
948 ACL_LINK, 0, rp->status);
949
950 hci_dev_unlock(hdev);
951 }
952
953 static void hci_cc_user_passkey_reply(struct hci_dev *hdev, struct sk_buff *skb)
954 {
955 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
956
957 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
958
959 hci_dev_lock(hdev);
960
961 if (test_bit(HCI_MGMT, &hdev->dev_flags))
962 mgmt_user_passkey_reply_complete(hdev, &rp->bdaddr, ACL_LINK,
963 0, rp->status);
964
965 hci_dev_unlock(hdev);
966 }
967
968 static void hci_cc_user_passkey_neg_reply(struct hci_dev *hdev,
969 struct sk_buff *skb)
970 {
971 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
972
973 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
974
975 hci_dev_lock(hdev);
976
977 if (test_bit(HCI_MGMT, &hdev->dev_flags))
978 mgmt_user_passkey_neg_reply_complete(hdev, &rp->bdaddr,
979 ACL_LINK, 0, rp->status);
980
981 hci_dev_unlock(hdev);
982 }
983
984 static void hci_cc_read_local_oob_data(struct hci_dev *hdev,
985 struct sk_buff *skb)
986 {
987 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
988
989 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
990
991 hci_dev_lock(hdev);
992 mgmt_read_local_oob_data_complete(hdev, rp->hash, rp->randomizer,
993 NULL, NULL, rp->status);
994 hci_dev_unlock(hdev);
995 }
996
997 static void hci_cc_read_local_oob_ext_data(struct hci_dev *hdev,
998 struct sk_buff *skb)
999 {
1000 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
1001
1002 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1003
1004 hci_dev_lock(hdev);
1005 mgmt_read_local_oob_data_complete(hdev, rp->hash192, rp->randomizer192,
1006 rp->hash256, rp->randomizer256,
1007 rp->status);
1008 hci_dev_unlock(hdev);
1009 }
1010
1011
1012 static void hci_cc_le_set_random_addr(struct hci_dev *hdev, struct sk_buff *skb)
1013 {
1014 __u8 status = *((__u8 *) skb->data);
1015 bdaddr_t *sent;
1016
1017 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1018
1019 if (status)
1020 return;
1021
1022 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_RANDOM_ADDR);
1023 if (!sent)
1024 return;
1025
1026 hci_dev_lock(hdev);
1027
1028 bacpy(&hdev->random_addr, sent);
1029
1030 hci_dev_unlock(hdev);
1031 }
1032
1033 static void hci_cc_le_set_adv_enable(struct hci_dev *hdev, struct sk_buff *skb)
1034 {
1035 __u8 *sent, status = *((__u8 *) skb->data);
1036
1037 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1038
1039 if (status)
1040 return;
1041
1042 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_ENABLE);
1043 if (!sent)
1044 return;
1045
1046 hci_dev_lock(hdev);
1047
1048 /* If we're doing connection initation as peripheral. Set a
1049 * timeout in case something goes wrong.
1050 */
1051 if (*sent) {
1052 struct hci_conn *conn;
1053
1054 set_bit(HCI_LE_ADV, &hdev->dev_flags);
1055
1056 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
1057 if (conn)
1058 queue_delayed_work(hdev->workqueue,
1059 &conn->le_conn_timeout,
1060 conn->conn_timeout);
1061 } else {
1062 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
1063 }
1064
1065 hci_dev_unlock(hdev);
1066 }
1067
1068 static void hci_cc_le_set_scan_param(struct hci_dev *hdev, struct sk_buff *skb)
1069 {
1070 struct hci_cp_le_set_scan_param *cp;
1071 __u8 status = *((__u8 *) skb->data);
1072
1073 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1074
1075 if (status)
1076 return;
1077
1078 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_PARAM);
1079 if (!cp)
1080 return;
1081
1082 hci_dev_lock(hdev);
1083
1084 hdev->le_scan_type = cp->type;
1085
1086 hci_dev_unlock(hdev);
1087 }
1088
1089 static bool has_pending_adv_report(struct hci_dev *hdev)
1090 {
1091 struct discovery_state *d = &hdev->discovery;
1092
1093 return bacmp(&d->last_adv_addr, BDADDR_ANY);
1094 }
1095
1096 static void clear_pending_adv_report(struct hci_dev *hdev)
1097 {
1098 struct discovery_state *d = &hdev->discovery;
1099
1100 bacpy(&d->last_adv_addr, BDADDR_ANY);
1101 d->last_adv_data_len = 0;
1102 }
1103
1104 static void store_pending_adv_report(struct hci_dev *hdev, bdaddr_t *bdaddr,
1105 u8 bdaddr_type, s8 rssi, u32 flags,
1106 u8 *data, u8 len)
1107 {
1108 struct discovery_state *d = &hdev->discovery;
1109
1110 bacpy(&d->last_adv_addr, bdaddr);
1111 d->last_adv_addr_type = bdaddr_type;
1112 d->last_adv_rssi = rssi;
1113 d->last_adv_flags = flags;
1114 memcpy(d->last_adv_data, data, len);
1115 d->last_adv_data_len = len;
1116 }
1117
1118 static void hci_cc_le_set_scan_enable(struct hci_dev *hdev,
1119 struct sk_buff *skb)
1120 {
1121 struct hci_cp_le_set_scan_enable *cp;
1122 __u8 status = *((__u8 *) skb->data);
1123
1124 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1125
1126 if (status)
1127 return;
1128
1129 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_ENABLE);
1130 if (!cp)
1131 return;
1132
1133 switch (cp->enable) {
1134 case LE_SCAN_ENABLE:
1135 set_bit(HCI_LE_SCAN, &hdev->dev_flags);
1136 if (hdev->le_scan_type == LE_SCAN_ACTIVE)
1137 clear_pending_adv_report(hdev);
1138 break;
1139
1140 case LE_SCAN_DISABLE:
1141 /* We do this here instead of when setting DISCOVERY_STOPPED
1142 * since the latter would potentially require waiting for
1143 * inquiry to stop too.
1144 */
1145 if (has_pending_adv_report(hdev)) {
1146 struct discovery_state *d = &hdev->discovery;
1147
1148 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
1149 d->last_adv_addr_type, NULL,
1150 d->last_adv_rssi, d->last_adv_flags,
1151 d->last_adv_data,
1152 d->last_adv_data_len, NULL, 0);
1153 }
1154
1155 /* Cancel this timer so that we don't try to disable scanning
1156 * when it's already disabled.
1157 */
1158 cancel_delayed_work(&hdev->le_scan_disable);
1159
1160 clear_bit(HCI_LE_SCAN, &hdev->dev_flags);
1161
1162 /* The HCI_LE_SCAN_INTERRUPTED flag indicates that we
1163 * interrupted scanning due to a connect request. Mark
1164 * therefore discovery as stopped. If this was not
1165 * because of a connect request advertising might have
1166 * been disabled because of active scanning, so
1167 * re-enable it again if necessary.
1168 */
1169 if (test_and_clear_bit(HCI_LE_SCAN_INTERRUPTED,
1170 &hdev->dev_flags))
1171 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1172 else if (!test_bit(HCI_LE_ADV, &hdev->dev_flags) &&
1173 hdev->discovery.state == DISCOVERY_FINDING)
1174 mgmt_reenable_advertising(hdev);
1175
1176 break;
1177
1178 default:
1179 BT_ERR("Used reserved LE_Scan_Enable param %d", cp->enable);
1180 break;
1181 }
1182 }
1183
1184 static void hci_cc_le_read_white_list_size(struct hci_dev *hdev,
1185 struct sk_buff *skb)
1186 {
1187 struct hci_rp_le_read_white_list_size *rp = (void *) skb->data;
1188
1189 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1190
1191 if (rp->status)
1192 return;
1193
1194 hdev->le_white_list_size = rp->size;
1195 }
1196
1197 static void hci_cc_le_clear_white_list(struct hci_dev *hdev,
1198 struct sk_buff *skb)
1199 {
1200 __u8 status = *((__u8 *) skb->data);
1201
1202 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1203
1204 if (status)
1205 return;
1206
1207 hci_bdaddr_list_clear(&hdev->le_white_list);
1208 }
1209
1210 static void hci_cc_le_add_to_white_list(struct hci_dev *hdev,
1211 struct sk_buff *skb)
1212 {
1213 struct hci_cp_le_add_to_white_list *sent;
1214 __u8 status = *((__u8 *) skb->data);
1215
1216 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1217
1218 if (status)
1219 return;
1220
1221 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_WHITE_LIST);
1222 if (!sent)
1223 return;
1224
1225 hci_bdaddr_list_add(&hdev->le_white_list, &sent->bdaddr,
1226 sent->bdaddr_type);
1227 }
1228
1229 static void hci_cc_le_del_from_white_list(struct hci_dev *hdev,
1230 struct sk_buff *skb)
1231 {
1232 struct hci_cp_le_del_from_white_list *sent;
1233 __u8 status = *((__u8 *) skb->data);
1234
1235 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1236
1237 if (status)
1238 return;
1239
1240 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_WHITE_LIST);
1241 if (!sent)
1242 return;
1243
1244 hci_bdaddr_list_del(&hdev->le_white_list, &sent->bdaddr,
1245 sent->bdaddr_type);
1246 }
1247
1248 static void hci_cc_le_read_supported_states(struct hci_dev *hdev,
1249 struct sk_buff *skb)
1250 {
1251 struct hci_rp_le_read_supported_states *rp = (void *) skb->data;
1252
1253 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1254
1255 if (rp->status)
1256 return;
1257
1258 memcpy(hdev->le_states, rp->le_states, 8);
1259 }
1260
1261 static void hci_cc_write_le_host_supported(struct hci_dev *hdev,
1262 struct sk_buff *skb)
1263 {
1264 struct hci_cp_write_le_host_supported *sent;
1265 __u8 status = *((__u8 *) skb->data);
1266
1267 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1268
1269 if (status)
1270 return;
1271
1272 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED);
1273 if (!sent)
1274 return;
1275
1276 if (sent->le) {
1277 hdev->features[1][0] |= LMP_HOST_LE;
1278 set_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1279 } else {
1280 hdev->features[1][0] &= ~LMP_HOST_LE;
1281 clear_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1282 clear_bit(HCI_ADVERTISING, &hdev->dev_flags);
1283 }
1284
1285 if (sent->simul)
1286 hdev->features[1][0] |= LMP_HOST_LE_BREDR;
1287 else
1288 hdev->features[1][0] &= ~LMP_HOST_LE_BREDR;
1289 }
1290
1291 static void hci_cc_set_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1292 {
1293 struct hci_cp_le_set_adv_param *cp;
1294 u8 status = *((u8 *) skb->data);
1295
1296 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1297
1298 if (status)
1299 return;
1300
1301 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_PARAM);
1302 if (!cp)
1303 return;
1304
1305 hci_dev_lock(hdev);
1306 hdev->adv_addr_type = cp->own_address_type;
1307 hci_dev_unlock(hdev);
1308 }
1309
1310 static void hci_cc_write_remote_amp_assoc(struct hci_dev *hdev,
1311 struct sk_buff *skb)
1312 {
1313 struct hci_rp_write_remote_amp_assoc *rp = (void *) skb->data;
1314
1315 BT_DBG("%s status 0x%2.2x phy_handle 0x%2.2x",
1316 hdev->name, rp->status, rp->phy_handle);
1317
1318 if (rp->status)
1319 return;
1320
1321 amp_write_rem_assoc_continue(hdev, rp->phy_handle);
1322 }
1323
1324 static void hci_cc_read_rssi(struct hci_dev *hdev, struct sk_buff *skb)
1325 {
1326 struct hci_rp_read_rssi *rp = (void *) skb->data;
1327 struct hci_conn *conn;
1328
1329 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1330
1331 if (rp->status)
1332 return;
1333
1334 hci_dev_lock(hdev);
1335
1336 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1337 if (conn)
1338 conn->rssi = rp->rssi;
1339
1340 hci_dev_unlock(hdev);
1341 }
1342
1343 static void hci_cc_read_tx_power(struct hci_dev *hdev, struct sk_buff *skb)
1344 {
1345 struct hci_cp_read_tx_power *sent;
1346 struct hci_rp_read_tx_power *rp = (void *) skb->data;
1347 struct hci_conn *conn;
1348
1349 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1350
1351 if (rp->status)
1352 return;
1353
1354 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_TX_POWER);
1355 if (!sent)
1356 return;
1357
1358 hci_dev_lock(hdev);
1359
1360 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1361 if (!conn)
1362 goto unlock;
1363
1364 switch (sent->type) {
1365 case 0x00:
1366 conn->tx_power = rp->tx_power;
1367 break;
1368 case 0x01:
1369 conn->max_tx_power = rp->tx_power;
1370 break;
1371 }
1372
1373 unlock:
1374 hci_dev_unlock(hdev);
1375 }
1376
1377 static void hci_cs_inquiry(struct hci_dev *hdev, __u8 status)
1378 {
1379 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1380
1381 if (status) {
1382 hci_conn_check_pending(hdev);
1383 return;
1384 }
1385
1386 set_bit(HCI_INQUIRY, &hdev->flags);
1387 }
1388
1389 static void hci_cs_create_conn(struct hci_dev *hdev, __u8 status)
1390 {
1391 struct hci_cp_create_conn *cp;
1392 struct hci_conn *conn;
1393
1394 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1395
1396 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_CONN);
1397 if (!cp)
1398 return;
1399
1400 hci_dev_lock(hdev);
1401
1402 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1403
1404 BT_DBG("%s bdaddr %pMR hcon %p", hdev->name, &cp->bdaddr, conn);
1405
1406 if (status) {
1407 if (conn && conn->state == BT_CONNECT) {
1408 if (status != 0x0c || conn->attempt > 2) {
1409 conn->state = BT_CLOSED;
1410 hci_proto_connect_cfm(conn, status);
1411 hci_conn_del(conn);
1412 } else
1413 conn->state = BT_CONNECT2;
1414 }
1415 } else {
1416 if (!conn) {
1417 conn = hci_conn_add(hdev, ACL_LINK, &cp->bdaddr,
1418 HCI_ROLE_MASTER);
1419 if (!conn)
1420 BT_ERR("No memory for new connection");
1421 }
1422 }
1423
1424 hci_dev_unlock(hdev);
1425 }
1426
1427 static void hci_cs_add_sco(struct hci_dev *hdev, __u8 status)
1428 {
1429 struct hci_cp_add_sco *cp;
1430 struct hci_conn *acl, *sco;
1431 __u16 handle;
1432
1433 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1434
1435 if (!status)
1436 return;
1437
1438 cp = hci_sent_cmd_data(hdev, HCI_OP_ADD_SCO);
1439 if (!cp)
1440 return;
1441
1442 handle = __le16_to_cpu(cp->handle);
1443
1444 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1445
1446 hci_dev_lock(hdev);
1447
1448 acl = hci_conn_hash_lookup_handle(hdev, handle);
1449 if (acl) {
1450 sco = acl->link;
1451 if (sco) {
1452 sco->state = BT_CLOSED;
1453
1454 hci_proto_connect_cfm(sco, status);
1455 hci_conn_del(sco);
1456 }
1457 }
1458
1459 hci_dev_unlock(hdev);
1460 }
1461
1462 static void hci_cs_auth_requested(struct hci_dev *hdev, __u8 status)
1463 {
1464 struct hci_cp_auth_requested *cp;
1465 struct hci_conn *conn;
1466
1467 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1468
1469 if (!status)
1470 return;
1471
1472 cp = hci_sent_cmd_data(hdev, HCI_OP_AUTH_REQUESTED);
1473 if (!cp)
1474 return;
1475
1476 hci_dev_lock(hdev);
1477
1478 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1479 if (conn) {
1480 if (conn->state == BT_CONFIG) {
1481 hci_proto_connect_cfm(conn, status);
1482 hci_conn_drop(conn);
1483 }
1484 }
1485
1486 hci_dev_unlock(hdev);
1487 }
1488
1489 static void hci_cs_set_conn_encrypt(struct hci_dev *hdev, __u8 status)
1490 {
1491 struct hci_cp_set_conn_encrypt *cp;
1492 struct hci_conn *conn;
1493
1494 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1495
1496 if (!status)
1497 return;
1498
1499 cp = hci_sent_cmd_data(hdev, HCI_OP_SET_CONN_ENCRYPT);
1500 if (!cp)
1501 return;
1502
1503 hci_dev_lock(hdev);
1504
1505 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1506 if (conn) {
1507 if (conn->state == BT_CONFIG) {
1508 hci_proto_connect_cfm(conn, status);
1509 hci_conn_drop(conn);
1510 }
1511 }
1512
1513 hci_dev_unlock(hdev);
1514 }
1515
1516 static int hci_outgoing_auth_needed(struct hci_dev *hdev,
1517 struct hci_conn *conn)
1518 {
1519 if (conn->state != BT_CONFIG || !conn->out)
1520 return 0;
1521
1522 if (conn->pending_sec_level == BT_SECURITY_SDP)
1523 return 0;
1524
1525 /* Only request authentication for SSP connections or non-SSP
1526 * devices with sec_level MEDIUM or HIGH or if MITM protection
1527 * is requested.
1528 */
1529 if (!hci_conn_ssp_enabled(conn) && !(conn->auth_type & 0x01) &&
1530 conn->pending_sec_level != BT_SECURITY_FIPS &&
1531 conn->pending_sec_level != BT_SECURITY_HIGH &&
1532 conn->pending_sec_level != BT_SECURITY_MEDIUM)
1533 return 0;
1534
1535 return 1;
1536 }
1537
1538 static int hci_resolve_name(struct hci_dev *hdev,
1539 struct inquiry_entry *e)
1540 {
1541 struct hci_cp_remote_name_req cp;
1542
1543 memset(&cp, 0, sizeof(cp));
1544
1545 bacpy(&cp.bdaddr, &e->data.bdaddr);
1546 cp.pscan_rep_mode = e->data.pscan_rep_mode;
1547 cp.pscan_mode = e->data.pscan_mode;
1548 cp.clock_offset = e->data.clock_offset;
1549
1550 return hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
1551 }
1552
1553 static bool hci_resolve_next_name(struct hci_dev *hdev)
1554 {
1555 struct discovery_state *discov = &hdev->discovery;
1556 struct inquiry_entry *e;
1557
1558 if (list_empty(&discov->resolve))
1559 return false;
1560
1561 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1562 if (!e)
1563 return false;
1564
1565 if (hci_resolve_name(hdev, e) == 0) {
1566 e->name_state = NAME_PENDING;
1567 return true;
1568 }
1569
1570 return false;
1571 }
1572
1573 static void hci_check_pending_name(struct hci_dev *hdev, struct hci_conn *conn,
1574 bdaddr_t *bdaddr, u8 *name, u8 name_len)
1575 {
1576 struct discovery_state *discov = &hdev->discovery;
1577 struct inquiry_entry *e;
1578
1579 if (conn && !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
1580 mgmt_device_connected(hdev, bdaddr, ACL_LINK, 0x00, 0, name,
1581 name_len, conn->dev_class);
1582
1583 if (discov->state == DISCOVERY_STOPPED)
1584 return;
1585
1586 if (discov->state == DISCOVERY_STOPPING)
1587 goto discov_complete;
1588
1589 if (discov->state != DISCOVERY_RESOLVING)
1590 return;
1591
1592 e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
1593 /* If the device was not found in a list of found devices names of which
1594 * are pending. there is no need to continue resolving a next name as it
1595 * will be done upon receiving another Remote Name Request Complete
1596 * Event */
1597 if (!e)
1598 return;
1599
1600 list_del(&e->list);
1601 if (name) {
1602 e->name_state = NAME_KNOWN;
1603 mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00,
1604 e->data.rssi, name, name_len);
1605 } else {
1606 e->name_state = NAME_NOT_KNOWN;
1607 }
1608
1609 if (hci_resolve_next_name(hdev))
1610 return;
1611
1612 discov_complete:
1613 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1614 }
1615
1616 static void hci_cs_remote_name_req(struct hci_dev *hdev, __u8 status)
1617 {
1618 struct hci_cp_remote_name_req *cp;
1619 struct hci_conn *conn;
1620
1621 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1622
1623 /* If successful wait for the name req complete event before
1624 * checking for the need to do authentication */
1625 if (!status)
1626 return;
1627
1628 cp = hci_sent_cmd_data(hdev, HCI_OP_REMOTE_NAME_REQ);
1629 if (!cp)
1630 return;
1631
1632 hci_dev_lock(hdev);
1633
1634 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1635
1636 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1637 hci_check_pending_name(hdev, conn, &cp->bdaddr, NULL, 0);
1638
1639 if (!conn)
1640 goto unlock;
1641
1642 if (!hci_outgoing_auth_needed(hdev, conn))
1643 goto unlock;
1644
1645 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
1646 struct hci_cp_auth_requested auth_cp;
1647
1648 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
1649
1650 auth_cp.handle = __cpu_to_le16(conn->handle);
1651 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED,
1652 sizeof(auth_cp), &auth_cp);
1653 }
1654
1655 unlock:
1656 hci_dev_unlock(hdev);
1657 }
1658
1659 static void hci_cs_read_remote_features(struct hci_dev *hdev, __u8 status)
1660 {
1661 struct hci_cp_read_remote_features *cp;
1662 struct hci_conn *conn;
1663
1664 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1665
1666 if (!status)
1667 return;
1668
1669 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_FEATURES);
1670 if (!cp)
1671 return;
1672
1673 hci_dev_lock(hdev);
1674
1675 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1676 if (conn) {
1677 if (conn->state == BT_CONFIG) {
1678 hci_proto_connect_cfm(conn, status);
1679 hci_conn_drop(conn);
1680 }
1681 }
1682
1683 hci_dev_unlock(hdev);
1684 }
1685
1686 static void hci_cs_read_remote_ext_features(struct hci_dev *hdev, __u8 status)
1687 {
1688 struct hci_cp_read_remote_ext_features *cp;
1689 struct hci_conn *conn;
1690
1691 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1692
1693 if (!status)
1694 return;
1695
1696 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES);
1697 if (!cp)
1698 return;
1699
1700 hci_dev_lock(hdev);
1701
1702 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1703 if (conn) {
1704 if (conn->state == BT_CONFIG) {
1705 hci_proto_connect_cfm(conn, status);
1706 hci_conn_drop(conn);
1707 }
1708 }
1709
1710 hci_dev_unlock(hdev);
1711 }
1712
1713 static void hci_cs_setup_sync_conn(struct hci_dev *hdev, __u8 status)
1714 {
1715 struct hci_cp_setup_sync_conn *cp;
1716 struct hci_conn *acl, *sco;
1717 __u16 handle;
1718
1719 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1720
1721 if (!status)
1722 return;
1723
1724 cp = hci_sent_cmd_data(hdev, HCI_OP_SETUP_SYNC_CONN);
1725 if (!cp)
1726 return;
1727
1728 handle = __le16_to_cpu(cp->handle);
1729
1730 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1731
1732 hci_dev_lock(hdev);
1733
1734 acl = hci_conn_hash_lookup_handle(hdev, handle);
1735 if (acl) {
1736 sco = acl->link;
1737 if (sco) {
1738 sco->state = BT_CLOSED;
1739
1740 hci_proto_connect_cfm(sco, status);
1741 hci_conn_del(sco);
1742 }
1743 }
1744
1745 hci_dev_unlock(hdev);
1746 }
1747
1748 static void hci_cs_sniff_mode(struct hci_dev *hdev, __u8 status)
1749 {
1750 struct hci_cp_sniff_mode *cp;
1751 struct hci_conn *conn;
1752
1753 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1754
1755 if (!status)
1756 return;
1757
1758 cp = hci_sent_cmd_data(hdev, HCI_OP_SNIFF_MODE);
1759 if (!cp)
1760 return;
1761
1762 hci_dev_lock(hdev);
1763
1764 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1765 if (conn) {
1766 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1767
1768 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1769 hci_sco_setup(conn, status);
1770 }
1771
1772 hci_dev_unlock(hdev);
1773 }
1774
1775 static void hci_cs_exit_sniff_mode(struct hci_dev *hdev, __u8 status)
1776 {
1777 struct hci_cp_exit_sniff_mode *cp;
1778 struct hci_conn *conn;
1779
1780 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1781
1782 if (!status)
1783 return;
1784
1785 cp = hci_sent_cmd_data(hdev, HCI_OP_EXIT_SNIFF_MODE);
1786 if (!cp)
1787 return;
1788
1789 hci_dev_lock(hdev);
1790
1791 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1792 if (conn) {
1793 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1794
1795 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1796 hci_sco_setup(conn, status);
1797 }
1798
1799 hci_dev_unlock(hdev);
1800 }
1801
1802 static void hci_cs_disconnect(struct hci_dev *hdev, u8 status)
1803 {
1804 struct hci_cp_disconnect *cp;
1805 struct hci_conn *conn;
1806
1807 if (!status)
1808 return;
1809
1810 cp = hci_sent_cmd_data(hdev, HCI_OP_DISCONNECT);
1811 if (!cp)
1812 return;
1813
1814 hci_dev_lock(hdev);
1815
1816 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1817 if (conn)
1818 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
1819 conn->dst_type, status);
1820
1821 hci_dev_unlock(hdev);
1822 }
1823
1824 static void hci_cs_create_phylink(struct hci_dev *hdev, u8 status)
1825 {
1826 struct hci_cp_create_phy_link *cp;
1827
1828 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1829
1830 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_PHY_LINK);
1831 if (!cp)
1832 return;
1833
1834 hci_dev_lock(hdev);
1835
1836 if (status) {
1837 struct hci_conn *hcon;
1838
1839 hcon = hci_conn_hash_lookup_handle(hdev, cp->phy_handle);
1840 if (hcon)
1841 hci_conn_del(hcon);
1842 } else {
1843 amp_write_remote_assoc(hdev, cp->phy_handle);
1844 }
1845
1846 hci_dev_unlock(hdev);
1847 }
1848
1849 static void hci_cs_accept_phylink(struct hci_dev *hdev, u8 status)
1850 {
1851 struct hci_cp_accept_phy_link *cp;
1852
1853 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1854
1855 if (status)
1856 return;
1857
1858 cp = hci_sent_cmd_data(hdev, HCI_OP_ACCEPT_PHY_LINK);
1859 if (!cp)
1860 return;
1861
1862 amp_write_remote_assoc(hdev, cp->phy_handle);
1863 }
1864
1865 static void hci_cs_le_create_conn(struct hci_dev *hdev, u8 status)
1866 {
1867 struct hci_cp_le_create_conn *cp;
1868 struct hci_conn *conn;
1869
1870 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1871
1872 /* All connection failure handling is taken care of by the
1873 * hci_le_conn_failed function which is triggered by the HCI
1874 * request completion callbacks used for connecting.
1875 */
1876 if (status)
1877 return;
1878
1879 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_CREATE_CONN);
1880 if (!cp)
1881 return;
1882
1883 hci_dev_lock(hdev);
1884
1885 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->peer_addr);
1886 if (!conn)
1887 goto unlock;
1888
1889 /* Store the initiator and responder address information which
1890 * is needed for SMP. These values will not change during the
1891 * lifetime of the connection.
1892 */
1893 conn->init_addr_type = cp->own_address_type;
1894 if (cp->own_address_type == ADDR_LE_DEV_RANDOM)
1895 bacpy(&conn->init_addr, &hdev->random_addr);
1896 else
1897 bacpy(&conn->init_addr, &hdev->bdaddr);
1898
1899 conn->resp_addr_type = cp->peer_addr_type;
1900 bacpy(&conn->resp_addr, &cp->peer_addr);
1901
1902 /* We don't want the connection attempt to stick around
1903 * indefinitely since LE doesn't have a page timeout concept
1904 * like BR/EDR. Set a timer for any connection that doesn't use
1905 * the white list for connecting.
1906 */
1907 if (cp->filter_policy == HCI_LE_USE_PEER_ADDR)
1908 queue_delayed_work(conn->hdev->workqueue,
1909 &conn->le_conn_timeout,
1910 conn->conn_timeout);
1911
1912 unlock:
1913 hci_dev_unlock(hdev);
1914 }
1915
1916 static void hci_cs_le_start_enc(struct hci_dev *hdev, u8 status)
1917 {
1918 struct hci_cp_le_start_enc *cp;
1919 struct hci_conn *conn;
1920
1921 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1922
1923 if (!status)
1924 return;
1925
1926 hci_dev_lock(hdev);
1927
1928 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_START_ENC);
1929 if (!cp)
1930 goto unlock;
1931
1932 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1933 if (!conn)
1934 goto unlock;
1935
1936 if (conn->state != BT_CONNECTED)
1937 goto unlock;
1938
1939 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
1940 hci_conn_drop(conn);
1941
1942 unlock:
1943 hci_dev_unlock(hdev);
1944 }
1945
1946 static void hci_inquiry_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
1947 {
1948 __u8 status = *((__u8 *) skb->data);
1949 struct discovery_state *discov = &hdev->discovery;
1950 struct inquiry_entry *e;
1951
1952 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1953
1954 hci_conn_check_pending(hdev);
1955
1956 if (!test_and_clear_bit(HCI_INQUIRY, &hdev->flags))
1957 return;
1958
1959 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
1960 wake_up_bit(&hdev->flags, HCI_INQUIRY);
1961
1962 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1963 return;
1964
1965 hci_dev_lock(hdev);
1966
1967 if (discov->state != DISCOVERY_FINDING)
1968 goto unlock;
1969
1970 if (list_empty(&discov->resolve)) {
1971 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1972 goto unlock;
1973 }
1974
1975 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1976 if (e && hci_resolve_name(hdev, e) == 0) {
1977 e->name_state = NAME_PENDING;
1978 hci_discovery_set_state(hdev, DISCOVERY_RESOLVING);
1979 } else {
1980 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1981 }
1982
1983 unlock:
1984 hci_dev_unlock(hdev);
1985 }
1986
1987 static void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff *skb)
1988 {
1989 struct inquiry_data data;
1990 struct inquiry_info *info = (void *) (skb->data + 1);
1991 int num_rsp = *((__u8 *) skb->data);
1992
1993 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
1994
1995 if (!num_rsp)
1996 return;
1997
1998 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
1999 return;
2000
2001 hci_dev_lock(hdev);
2002
2003 for (; num_rsp; num_rsp--, info++) {
2004 u32 flags;
2005
2006 bacpy(&data.bdaddr, &info->bdaddr);
2007 data.pscan_rep_mode = info->pscan_rep_mode;
2008 data.pscan_period_mode = info->pscan_period_mode;
2009 data.pscan_mode = info->pscan_mode;
2010 memcpy(data.dev_class, info->dev_class, 3);
2011 data.clock_offset = info->clock_offset;
2012 data.rssi = 0x00;
2013 data.ssp_mode = 0x00;
2014
2015 flags = hci_inquiry_cache_update(hdev, &data, false);
2016
2017 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
2018 info->dev_class, 0, flags, NULL, 0, NULL, 0);
2019 }
2020
2021 hci_dev_unlock(hdev);
2022 }
2023
2024 static void hci_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2025 {
2026 struct hci_ev_conn_complete *ev = (void *) skb->data;
2027 struct hci_conn *conn;
2028
2029 BT_DBG("%s", hdev->name);
2030
2031 hci_dev_lock(hdev);
2032
2033 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
2034 if (!conn) {
2035 if (ev->link_type != SCO_LINK)
2036 goto unlock;
2037
2038 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
2039 if (!conn)
2040 goto unlock;
2041
2042 conn->type = SCO_LINK;
2043 }
2044
2045 if (!ev->status) {
2046 conn->handle = __le16_to_cpu(ev->handle);
2047
2048 if (conn->type == ACL_LINK) {
2049 conn->state = BT_CONFIG;
2050 hci_conn_hold(conn);
2051
2052 if (!conn->out && !hci_conn_ssp_enabled(conn) &&
2053 !hci_find_link_key(hdev, &ev->bdaddr))
2054 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
2055 else
2056 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2057 } else
2058 conn->state = BT_CONNECTED;
2059
2060 hci_conn_add_sysfs(conn);
2061
2062 if (test_bit(HCI_AUTH, &hdev->flags))
2063 set_bit(HCI_CONN_AUTH, &conn->flags);
2064
2065 if (test_bit(HCI_ENCRYPT, &hdev->flags))
2066 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2067
2068 /* Get remote features */
2069 if (conn->type == ACL_LINK) {
2070 struct hci_cp_read_remote_features cp;
2071 cp.handle = ev->handle;
2072 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_FEATURES,
2073 sizeof(cp), &cp);
2074
2075 hci_update_page_scan(hdev, NULL);
2076 }
2077
2078 /* Set packet type for incoming connection */
2079 if (!conn->out && hdev->hci_ver < BLUETOOTH_VER_2_0) {
2080 struct hci_cp_change_conn_ptype cp;
2081 cp.handle = ev->handle;
2082 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2083 hci_send_cmd(hdev, HCI_OP_CHANGE_CONN_PTYPE, sizeof(cp),
2084 &cp);
2085 }
2086 } else {
2087 conn->state = BT_CLOSED;
2088 if (conn->type == ACL_LINK)
2089 mgmt_connect_failed(hdev, &conn->dst, conn->type,
2090 conn->dst_type, ev->status);
2091 }
2092
2093 if (conn->type == ACL_LINK)
2094 hci_sco_setup(conn, ev->status);
2095
2096 if (ev->status) {
2097 hci_proto_connect_cfm(conn, ev->status);
2098 hci_conn_del(conn);
2099 } else if (ev->link_type != ACL_LINK)
2100 hci_proto_connect_cfm(conn, ev->status);
2101
2102 unlock:
2103 hci_dev_unlock(hdev);
2104
2105 hci_conn_check_pending(hdev);
2106 }
2107
2108 static void hci_reject_conn(struct hci_dev *hdev, bdaddr_t *bdaddr)
2109 {
2110 struct hci_cp_reject_conn_req cp;
2111
2112 bacpy(&cp.bdaddr, bdaddr);
2113 cp.reason = HCI_ERROR_REJ_BAD_ADDR;
2114 hci_send_cmd(hdev, HCI_OP_REJECT_CONN_REQ, sizeof(cp), &cp);
2115 }
2116
2117 static void hci_conn_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
2118 {
2119 struct hci_ev_conn_request *ev = (void *) skb->data;
2120 int mask = hdev->link_mode;
2121 struct inquiry_entry *ie;
2122 struct hci_conn *conn;
2123 __u8 flags = 0;
2124
2125 BT_DBG("%s bdaddr %pMR type 0x%x", hdev->name, &ev->bdaddr,
2126 ev->link_type);
2127
2128 mask |= hci_proto_connect_ind(hdev, &ev->bdaddr, ev->link_type,
2129 &flags);
2130
2131 if (!(mask & HCI_LM_ACCEPT)) {
2132 hci_reject_conn(hdev, &ev->bdaddr);
2133 return;
2134 }
2135
2136 if (hci_bdaddr_list_lookup(&hdev->blacklist, &ev->bdaddr,
2137 BDADDR_BREDR)) {
2138 hci_reject_conn(hdev, &ev->bdaddr);
2139 return;
2140 }
2141
2142 if (!test_bit(HCI_CONNECTABLE, &hdev->dev_flags) &&
2143 !hci_bdaddr_list_lookup(&hdev->whitelist, &ev->bdaddr,
2144 BDADDR_BREDR)) {
2145 hci_reject_conn(hdev, &ev->bdaddr);
2146 return;
2147 }
2148
2149 /* Connection accepted */
2150
2151 hci_dev_lock(hdev);
2152
2153 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
2154 if (ie)
2155 memcpy(ie->data.dev_class, ev->dev_class, 3);
2156
2157 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type,
2158 &ev->bdaddr);
2159 if (!conn) {
2160 conn = hci_conn_add(hdev, ev->link_type, &ev->bdaddr,
2161 HCI_ROLE_SLAVE);
2162 if (!conn) {
2163 BT_ERR("No memory for new connection");
2164 hci_dev_unlock(hdev);
2165 return;
2166 }
2167 }
2168
2169 memcpy(conn->dev_class, ev->dev_class, 3);
2170
2171 hci_dev_unlock(hdev);
2172
2173 if (ev->link_type == ACL_LINK ||
2174 (!(flags & HCI_PROTO_DEFER) && !lmp_esco_capable(hdev))) {
2175 struct hci_cp_accept_conn_req cp;
2176 conn->state = BT_CONNECT;
2177
2178 bacpy(&cp.bdaddr, &ev->bdaddr);
2179
2180 if (lmp_rswitch_capable(hdev) && (mask & HCI_LM_MASTER))
2181 cp.role = 0x00; /* Become master */
2182 else
2183 cp.role = 0x01; /* Remain slave */
2184
2185 hci_send_cmd(hdev, HCI_OP_ACCEPT_CONN_REQ, sizeof(cp), &cp);
2186 } else if (!(flags & HCI_PROTO_DEFER)) {
2187 struct hci_cp_accept_sync_conn_req cp;
2188 conn->state = BT_CONNECT;
2189
2190 bacpy(&cp.bdaddr, &ev->bdaddr);
2191 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2192
2193 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
2194 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
2195 cp.max_latency = cpu_to_le16(0xffff);
2196 cp.content_format = cpu_to_le16(hdev->voice_setting);
2197 cp.retrans_effort = 0xff;
2198
2199 hci_send_cmd(hdev, HCI_OP_ACCEPT_SYNC_CONN_REQ, sizeof(cp),
2200 &cp);
2201 } else {
2202 conn->state = BT_CONNECT2;
2203 hci_proto_connect_cfm(conn, 0);
2204 }
2205 }
2206
2207 static u8 hci_to_mgmt_reason(u8 err)
2208 {
2209 switch (err) {
2210 case HCI_ERROR_CONNECTION_TIMEOUT:
2211 return MGMT_DEV_DISCONN_TIMEOUT;
2212 case HCI_ERROR_REMOTE_USER_TERM:
2213 case HCI_ERROR_REMOTE_LOW_RESOURCES:
2214 case HCI_ERROR_REMOTE_POWER_OFF:
2215 return MGMT_DEV_DISCONN_REMOTE;
2216 case HCI_ERROR_LOCAL_HOST_TERM:
2217 return MGMT_DEV_DISCONN_LOCAL_HOST;
2218 default:
2219 return MGMT_DEV_DISCONN_UNKNOWN;
2220 }
2221 }
2222
2223 static void hci_disconn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2224 {
2225 struct hci_ev_disconn_complete *ev = (void *) skb->data;
2226 u8 reason = hci_to_mgmt_reason(ev->reason);
2227 struct hci_conn_params *params;
2228 struct hci_conn *conn;
2229 bool mgmt_connected;
2230 u8 type;
2231
2232 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2233
2234 hci_dev_lock(hdev);
2235
2236 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2237 if (!conn)
2238 goto unlock;
2239
2240 if (ev->status) {
2241 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2242 conn->dst_type, ev->status);
2243 goto unlock;
2244 }
2245
2246 conn->state = BT_CLOSED;
2247
2248 mgmt_connected = test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags);
2249 mgmt_device_disconnected(hdev, &conn->dst, conn->type, conn->dst_type,
2250 reason, mgmt_connected);
2251
2252 if (conn->type == ACL_LINK) {
2253 if (test_bit(HCI_CONN_FLUSH_KEY, &conn->flags))
2254 hci_remove_link_key(hdev, &conn->dst);
2255
2256 hci_update_page_scan(hdev, NULL);
2257 }
2258
2259 params = hci_conn_params_lookup(hdev, &conn->dst, conn->dst_type);
2260 if (params) {
2261 switch (params->auto_connect) {
2262 case HCI_AUTO_CONN_LINK_LOSS:
2263 if (ev->reason != HCI_ERROR_CONNECTION_TIMEOUT)
2264 break;
2265 /* Fall through */
2266
2267 case HCI_AUTO_CONN_DIRECT:
2268 case HCI_AUTO_CONN_ALWAYS:
2269 list_del_init(&params->action);
2270 list_add(&params->action, &hdev->pend_le_conns);
2271 hci_update_background_scan(hdev);
2272 break;
2273
2274 default:
2275 break;
2276 }
2277 }
2278
2279 type = conn->type;
2280
2281 hci_proto_disconn_cfm(conn, ev->reason);
2282 hci_conn_del(conn);
2283
2284 /* Re-enable advertising if necessary, since it might
2285 * have been disabled by the connection. From the
2286 * HCI_LE_Set_Advertise_Enable command description in
2287 * the core specification (v4.0):
2288 * "The Controller shall continue advertising until the Host
2289 * issues an LE_Set_Advertise_Enable command with
2290 * Advertising_Enable set to 0x00 (Advertising is disabled)
2291 * or until a connection is created or until the Advertising
2292 * is timed out due to Directed Advertising."
2293 */
2294 if (type == LE_LINK)
2295 mgmt_reenable_advertising(hdev);
2296
2297 unlock:
2298 hci_dev_unlock(hdev);
2299 }
2300
2301 static void hci_auth_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2302 {
2303 struct hci_ev_auth_complete *ev = (void *) skb->data;
2304 struct hci_conn *conn;
2305
2306 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2307
2308 hci_dev_lock(hdev);
2309
2310 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2311 if (!conn)
2312 goto unlock;
2313
2314 if (!ev->status) {
2315 if (!hci_conn_ssp_enabled(conn) &&
2316 test_bit(HCI_CONN_REAUTH_PEND, &conn->flags)) {
2317 BT_INFO("re-auth of legacy device is not possible.");
2318 } else {
2319 set_bit(HCI_CONN_AUTH, &conn->flags);
2320 conn->sec_level = conn->pending_sec_level;
2321 }
2322 } else {
2323 mgmt_auth_failed(conn, ev->status);
2324 }
2325
2326 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2327 clear_bit(HCI_CONN_REAUTH_PEND, &conn->flags);
2328
2329 if (conn->state == BT_CONFIG) {
2330 if (!ev->status && hci_conn_ssp_enabled(conn)) {
2331 struct hci_cp_set_conn_encrypt cp;
2332 cp.handle = ev->handle;
2333 cp.encrypt = 0x01;
2334 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2335 &cp);
2336 } else {
2337 conn->state = BT_CONNECTED;
2338 hci_proto_connect_cfm(conn, ev->status);
2339 hci_conn_drop(conn);
2340 }
2341 } else {
2342 hci_auth_cfm(conn, ev->status);
2343
2344 hci_conn_hold(conn);
2345 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2346 hci_conn_drop(conn);
2347 }
2348
2349 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags)) {
2350 if (!ev->status) {
2351 struct hci_cp_set_conn_encrypt cp;
2352 cp.handle = ev->handle;
2353 cp.encrypt = 0x01;
2354 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2355 &cp);
2356 } else {
2357 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2358 hci_encrypt_cfm(conn, ev->status, 0x00);
2359 }
2360 }
2361
2362 unlock:
2363 hci_dev_unlock(hdev);
2364 }
2365
2366 static void hci_remote_name_evt(struct hci_dev *hdev, struct sk_buff *skb)
2367 {
2368 struct hci_ev_remote_name *ev = (void *) skb->data;
2369 struct hci_conn *conn;
2370
2371 BT_DBG("%s", hdev->name);
2372
2373 hci_conn_check_pending(hdev);
2374
2375 hci_dev_lock(hdev);
2376
2377 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2378
2379 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
2380 goto check_auth;
2381
2382 if (ev->status == 0)
2383 hci_check_pending_name(hdev, conn, &ev->bdaddr, ev->name,
2384 strnlen(ev->name, HCI_MAX_NAME_LENGTH));
2385 else
2386 hci_check_pending_name(hdev, conn, &ev->bdaddr, NULL, 0);
2387
2388 check_auth:
2389 if (!conn)
2390 goto unlock;
2391
2392 if (!hci_outgoing_auth_needed(hdev, conn))
2393 goto unlock;
2394
2395 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
2396 struct hci_cp_auth_requested cp;
2397
2398 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
2399
2400 cp.handle = __cpu_to_le16(conn->handle);
2401 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED, sizeof(cp), &cp);
2402 }
2403
2404 unlock:
2405 hci_dev_unlock(hdev);
2406 }
2407
2408 static void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2409 {
2410 struct hci_ev_encrypt_change *ev = (void *) skb->data;
2411 struct hci_conn *conn;
2412
2413 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2414
2415 hci_dev_lock(hdev);
2416
2417 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2418 if (!conn)
2419 goto unlock;
2420
2421 if (!ev->status) {
2422 if (ev->encrypt) {
2423 /* Encryption implies authentication */
2424 set_bit(HCI_CONN_AUTH, &conn->flags);
2425 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2426 conn->sec_level = conn->pending_sec_level;
2427
2428 /* P-256 authentication key implies FIPS */
2429 if (conn->key_type == HCI_LK_AUTH_COMBINATION_P256)
2430 set_bit(HCI_CONN_FIPS, &conn->flags);
2431
2432 if ((conn->type == ACL_LINK && ev->encrypt == 0x02) ||
2433 conn->type == LE_LINK)
2434 set_bit(HCI_CONN_AES_CCM, &conn->flags);
2435 } else {
2436 clear_bit(HCI_CONN_ENCRYPT, &conn->flags);
2437 clear_bit(HCI_CONN_AES_CCM, &conn->flags);
2438 }
2439 }
2440
2441 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2442
2443 if (ev->status && conn->state == BT_CONNECTED) {
2444 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2445 hci_conn_drop(conn);
2446 goto unlock;
2447 }
2448
2449 if (conn->state == BT_CONFIG) {
2450 if (!ev->status)
2451 conn->state = BT_CONNECTED;
2452
2453 /* In Secure Connections Only mode, do not allow any
2454 * connections that are not encrypted with AES-CCM
2455 * using a P-256 authenticated combination key.
2456 */
2457 if (test_bit(HCI_SC_ONLY, &hdev->dev_flags) &&
2458 (!test_bit(HCI_CONN_AES_CCM, &conn->flags) ||
2459 conn->key_type != HCI_LK_AUTH_COMBINATION_P256)) {
2460 hci_proto_connect_cfm(conn, HCI_ERROR_AUTH_FAILURE);
2461 hci_conn_drop(conn);
2462 goto unlock;
2463 }
2464
2465 hci_proto_connect_cfm(conn, ev->status);
2466 hci_conn_drop(conn);
2467 } else
2468 hci_encrypt_cfm(conn, ev->status, ev->encrypt);
2469
2470 unlock:
2471 hci_dev_unlock(hdev);
2472 }
2473
2474 static void hci_change_link_key_complete_evt(struct hci_dev *hdev,
2475 struct sk_buff *skb)
2476 {
2477 struct hci_ev_change_link_key_complete *ev = (void *) skb->data;
2478 struct hci_conn *conn;
2479
2480 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2481
2482 hci_dev_lock(hdev);
2483
2484 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2485 if (conn) {
2486 if (!ev->status)
2487 set_bit(HCI_CONN_SECURE, &conn->flags);
2488
2489 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2490
2491 hci_key_change_cfm(conn, ev->status);
2492 }
2493
2494 hci_dev_unlock(hdev);
2495 }
2496
2497 static void hci_remote_features_evt(struct hci_dev *hdev,
2498 struct sk_buff *skb)
2499 {
2500 struct hci_ev_remote_features *ev = (void *) skb->data;
2501 struct hci_conn *conn;
2502
2503 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2504
2505 hci_dev_lock(hdev);
2506
2507 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2508 if (!conn)
2509 goto unlock;
2510
2511 if (!ev->status)
2512 memcpy(conn->features[0], ev->features, 8);
2513
2514 if (conn->state != BT_CONFIG)
2515 goto unlock;
2516
2517 if (!ev->status && lmp_ssp_capable(hdev) && lmp_ssp_capable(conn)) {
2518 struct hci_cp_read_remote_ext_features cp;
2519 cp.handle = ev->handle;
2520 cp.page = 0x01;
2521 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES,
2522 sizeof(cp), &cp);
2523 goto unlock;
2524 }
2525
2526 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
2527 struct hci_cp_remote_name_req cp;
2528 memset(&cp, 0, sizeof(cp));
2529 bacpy(&cp.bdaddr, &conn->dst);
2530 cp.pscan_rep_mode = 0x02;
2531 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
2532 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2533 mgmt_device_connected(hdev, &conn->dst, conn->type,
2534 conn->dst_type, 0, NULL, 0,
2535 conn->dev_class);
2536
2537 if (!hci_outgoing_auth_needed(hdev, conn)) {
2538 conn->state = BT_CONNECTED;
2539 hci_proto_connect_cfm(conn, ev->status);
2540 hci_conn_drop(conn);
2541 }
2542
2543 unlock:
2544 hci_dev_unlock(hdev);
2545 }
2546
2547 static void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2548 {
2549 struct hci_ev_cmd_complete *ev = (void *) skb->data;
2550 u8 status = skb->data[sizeof(*ev)];
2551 __u16 opcode;
2552
2553 skb_pull(skb, sizeof(*ev));
2554
2555 opcode = __le16_to_cpu(ev->opcode);
2556
2557 switch (opcode) {
2558 case HCI_OP_INQUIRY_CANCEL:
2559 hci_cc_inquiry_cancel(hdev, skb);
2560 break;
2561
2562 case HCI_OP_PERIODIC_INQ:
2563 hci_cc_periodic_inq(hdev, skb);
2564 break;
2565
2566 case HCI_OP_EXIT_PERIODIC_INQ:
2567 hci_cc_exit_periodic_inq(hdev, skb);
2568 break;
2569
2570 case HCI_OP_REMOTE_NAME_REQ_CANCEL:
2571 hci_cc_remote_name_req_cancel(hdev, skb);
2572 break;
2573
2574 case HCI_OP_ROLE_DISCOVERY:
2575 hci_cc_role_discovery(hdev, skb);
2576 break;
2577
2578 case HCI_OP_READ_LINK_POLICY:
2579 hci_cc_read_link_policy(hdev, skb);
2580 break;
2581
2582 case HCI_OP_WRITE_LINK_POLICY:
2583 hci_cc_write_link_policy(hdev, skb);
2584 break;
2585
2586 case HCI_OP_READ_DEF_LINK_POLICY:
2587 hci_cc_read_def_link_policy(hdev, skb);
2588 break;
2589
2590 case HCI_OP_WRITE_DEF_LINK_POLICY:
2591 hci_cc_write_def_link_policy(hdev, skb);
2592 break;
2593
2594 case HCI_OP_RESET:
2595 hci_cc_reset(hdev, skb);
2596 break;
2597
2598 case HCI_OP_WRITE_LOCAL_NAME:
2599 hci_cc_write_local_name(hdev, skb);
2600 break;
2601
2602 case HCI_OP_READ_LOCAL_NAME:
2603 hci_cc_read_local_name(hdev, skb);
2604 break;
2605
2606 case HCI_OP_WRITE_AUTH_ENABLE:
2607 hci_cc_write_auth_enable(hdev, skb);
2608 break;
2609
2610 case HCI_OP_WRITE_ENCRYPT_MODE:
2611 hci_cc_write_encrypt_mode(hdev, skb);
2612 break;
2613
2614 case HCI_OP_WRITE_SCAN_ENABLE:
2615 hci_cc_write_scan_enable(hdev, skb);
2616 break;
2617
2618 case HCI_OP_READ_CLASS_OF_DEV:
2619 hci_cc_read_class_of_dev(hdev, skb);
2620 break;
2621
2622 case HCI_OP_WRITE_CLASS_OF_DEV:
2623 hci_cc_write_class_of_dev(hdev, skb);
2624 break;
2625
2626 case HCI_OP_READ_VOICE_SETTING:
2627 hci_cc_read_voice_setting(hdev, skb);
2628 break;
2629
2630 case HCI_OP_WRITE_VOICE_SETTING:
2631 hci_cc_write_voice_setting(hdev, skb);
2632 break;
2633
2634 case HCI_OP_READ_NUM_SUPPORTED_IAC:
2635 hci_cc_read_num_supported_iac(hdev, skb);
2636 break;
2637
2638 case HCI_OP_WRITE_SSP_MODE:
2639 hci_cc_write_ssp_mode(hdev, skb);
2640 break;
2641
2642 case HCI_OP_WRITE_SC_SUPPORT:
2643 hci_cc_write_sc_support(hdev, skb);
2644 break;
2645
2646 case HCI_OP_READ_LOCAL_VERSION:
2647 hci_cc_read_local_version(hdev, skb);
2648 break;
2649
2650 case HCI_OP_READ_LOCAL_COMMANDS:
2651 hci_cc_read_local_commands(hdev, skb);
2652 break;
2653
2654 case HCI_OP_READ_LOCAL_FEATURES:
2655 hci_cc_read_local_features(hdev, skb);
2656 break;
2657
2658 case HCI_OP_READ_LOCAL_EXT_FEATURES:
2659 hci_cc_read_local_ext_features(hdev, skb);
2660 break;
2661
2662 case HCI_OP_READ_BUFFER_SIZE:
2663 hci_cc_read_buffer_size(hdev, skb);
2664 break;
2665
2666 case HCI_OP_READ_BD_ADDR:
2667 hci_cc_read_bd_addr(hdev, skb);
2668 break;
2669
2670 case HCI_OP_READ_PAGE_SCAN_ACTIVITY:
2671 hci_cc_read_page_scan_activity(hdev, skb);
2672 break;
2673
2674 case HCI_OP_WRITE_PAGE_SCAN_ACTIVITY:
2675 hci_cc_write_page_scan_activity(hdev, skb);
2676 break;
2677
2678 case HCI_OP_READ_PAGE_SCAN_TYPE:
2679 hci_cc_read_page_scan_type(hdev, skb);
2680 break;
2681
2682 case HCI_OP_WRITE_PAGE_SCAN_TYPE:
2683 hci_cc_write_page_scan_type(hdev, skb);
2684 break;
2685
2686 case HCI_OP_READ_DATA_BLOCK_SIZE:
2687 hci_cc_read_data_block_size(hdev, skb);
2688 break;
2689
2690 case HCI_OP_READ_FLOW_CONTROL_MODE:
2691 hci_cc_read_flow_control_mode(hdev, skb);
2692 break;
2693
2694 case HCI_OP_READ_LOCAL_AMP_INFO:
2695 hci_cc_read_local_amp_info(hdev, skb);
2696 break;
2697
2698 case HCI_OP_READ_CLOCK:
2699 hci_cc_read_clock(hdev, skb);
2700 break;
2701
2702 case HCI_OP_READ_LOCAL_AMP_ASSOC:
2703 hci_cc_read_local_amp_assoc(hdev, skb);
2704 break;
2705
2706 case HCI_OP_READ_INQ_RSP_TX_POWER:
2707 hci_cc_read_inq_rsp_tx_power(hdev, skb);
2708 break;
2709
2710 case HCI_OP_PIN_CODE_REPLY:
2711 hci_cc_pin_code_reply(hdev, skb);
2712 break;
2713
2714 case HCI_OP_PIN_CODE_NEG_REPLY:
2715 hci_cc_pin_code_neg_reply(hdev, skb);
2716 break;
2717
2718 case HCI_OP_READ_LOCAL_OOB_DATA:
2719 hci_cc_read_local_oob_data(hdev, skb);
2720 break;
2721
2722 case HCI_OP_READ_LOCAL_OOB_EXT_DATA:
2723 hci_cc_read_local_oob_ext_data(hdev, skb);
2724 break;
2725
2726 case HCI_OP_LE_READ_BUFFER_SIZE:
2727 hci_cc_le_read_buffer_size(hdev, skb);
2728 break;
2729
2730 case HCI_OP_LE_READ_LOCAL_FEATURES:
2731 hci_cc_le_read_local_features(hdev, skb);
2732 break;
2733
2734 case HCI_OP_LE_READ_ADV_TX_POWER:
2735 hci_cc_le_read_adv_tx_power(hdev, skb);
2736 break;
2737
2738 case HCI_OP_USER_CONFIRM_REPLY:
2739 hci_cc_user_confirm_reply(hdev, skb);
2740 break;
2741
2742 case HCI_OP_USER_CONFIRM_NEG_REPLY:
2743 hci_cc_user_confirm_neg_reply(hdev, skb);
2744 break;
2745
2746 case HCI_OP_USER_PASSKEY_REPLY:
2747 hci_cc_user_passkey_reply(hdev, skb);
2748 break;
2749
2750 case HCI_OP_USER_PASSKEY_NEG_REPLY:
2751 hci_cc_user_passkey_neg_reply(hdev, skb);
2752 break;
2753
2754 case HCI_OP_LE_SET_RANDOM_ADDR:
2755 hci_cc_le_set_random_addr(hdev, skb);
2756 break;
2757
2758 case HCI_OP_LE_SET_ADV_ENABLE:
2759 hci_cc_le_set_adv_enable(hdev, skb);
2760 break;
2761
2762 case HCI_OP_LE_SET_SCAN_PARAM:
2763 hci_cc_le_set_scan_param(hdev, skb);
2764 break;
2765
2766 case HCI_OP_LE_SET_SCAN_ENABLE:
2767 hci_cc_le_set_scan_enable(hdev, skb);
2768 break;
2769
2770 case HCI_OP_LE_READ_WHITE_LIST_SIZE:
2771 hci_cc_le_read_white_list_size(hdev, skb);
2772 break;
2773
2774 case HCI_OP_LE_CLEAR_WHITE_LIST:
2775 hci_cc_le_clear_white_list(hdev, skb);
2776 break;
2777
2778 case HCI_OP_LE_ADD_TO_WHITE_LIST:
2779 hci_cc_le_add_to_white_list(hdev, skb);
2780 break;
2781
2782 case HCI_OP_LE_DEL_FROM_WHITE_LIST:
2783 hci_cc_le_del_from_white_list(hdev, skb);
2784 break;
2785
2786 case HCI_OP_LE_READ_SUPPORTED_STATES:
2787 hci_cc_le_read_supported_states(hdev, skb);
2788 break;
2789
2790 case HCI_OP_WRITE_LE_HOST_SUPPORTED:
2791 hci_cc_write_le_host_supported(hdev, skb);
2792 break;
2793
2794 case HCI_OP_LE_SET_ADV_PARAM:
2795 hci_cc_set_adv_param(hdev, skb);
2796 break;
2797
2798 case HCI_OP_WRITE_REMOTE_AMP_ASSOC:
2799 hci_cc_write_remote_amp_assoc(hdev, skb);
2800 break;
2801
2802 case HCI_OP_READ_RSSI:
2803 hci_cc_read_rssi(hdev, skb);
2804 break;
2805
2806 case HCI_OP_READ_TX_POWER:
2807 hci_cc_read_tx_power(hdev, skb);
2808 break;
2809
2810 default:
2811 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2812 break;
2813 }
2814
2815 if (opcode != HCI_OP_NOP)
2816 cancel_delayed_work(&hdev->cmd_timer);
2817
2818 hci_req_cmd_complete(hdev, opcode, status);
2819
2820 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2821 atomic_set(&hdev->cmd_cnt, 1);
2822 if (!skb_queue_empty(&hdev->cmd_q))
2823 queue_work(hdev->workqueue, &hdev->cmd_work);
2824 }
2825 }
2826
2827 static void hci_cmd_status_evt(struct hci_dev *hdev, struct sk_buff *skb)
2828 {
2829 struct hci_ev_cmd_status *ev = (void *) skb->data;
2830 __u16 opcode;
2831
2832 skb_pull(skb, sizeof(*ev));
2833
2834 opcode = __le16_to_cpu(ev->opcode);
2835
2836 switch (opcode) {
2837 case HCI_OP_INQUIRY:
2838 hci_cs_inquiry(hdev, ev->status);
2839 break;
2840
2841 case HCI_OP_CREATE_CONN:
2842 hci_cs_create_conn(hdev, ev->status);
2843 break;
2844
2845 case HCI_OP_ADD_SCO:
2846 hci_cs_add_sco(hdev, ev->status);
2847 break;
2848
2849 case HCI_OP_AUTH_REQUESTED:
2850 hci_cs_auth_requested(hdev, ev->status);
2851 break;
2852
2853 case HCI_OP_SET_CONN_ENCRYPT:
2854 hci_cs_set_conn_encrypt(hdev, ev->status);
2855 break;
2856
2857 case HCI_OP_REMOTE_NAME_REQ:
2858 hci_cs_remote_name_req(hdev, ev->status);
2859 break;
2860
2861 case HCI_OP_READ_REMOTE_FEATURES:
2862 hci_cs_read_remote_features(hdev, ev->status);
2863 break;
2864
2865 case HCI_OP_READ_REMOTE_EXT_FEATURES:
2866 hci_cs_read_remote_ext_features(hdev, ev->status);
2867 break;
2868
2869 case HCI_OP_SETUP_SYNC_CONN:
2870 hci_cs_setup_sync_conn(hdev, ev->status);
2871 break;
2872
2873 case HCI_OP_SNIFF_MODE:
2874 hci_cs_sniff_mode(hdev, ev->status);
2875 break;
2876
2877 case HCI_OP_EXIT_SNIFF_MODE:
2878 hci_cs_exit_sniff_mode(hdev, ev->status);
2879 break;
2880
2881 case HCI_OP_DISCONNECT:
2882 hci_cs_disconnect(hdev, ev->status);
2883 break;
2884
2885 case HCI_OP_CREATE_PHY_LINK:
2886 hci_cs_create_phylink(hdev, ev->status);
2887 break;
2888
2889 case HCI_OP_ACCEPT_PHY_LINK:
2890 hci_cs_accept_phylink(hdev, ev->status);
2891 break;
2892
2893 case HCI_OP_LE_CREATE_CONN:
2894 hci_cs_le_create_conn(hdev, ev->status);
2895 break;
2896
2897 case HCI_OP_LE_START_ENC:
2898 hci_cs_le_start_enc(hdev, ev->status);
2899 break;
2900
2901 default:
2902 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2903 break;
2904 }
2905
2906 if (opcode != HCI_OP_NOP)
2907 cancel_delayed_work(&hdev->cmd_timer);
2908
2909 if (ev->status ||
2910 (hdev->sent_cmd && !bt_cb(hdev->sent_cmd)->req.event))
2911 hci_req_cmd_complete(hdev, opcode, ev->status);
2912
2913 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2914 atomic_set(&hdev->cmd_cnt, 1);
2915 if (!skb_queue_empty(&hdev->cmd_q))
2916 queue_work(hdev->workqueue, &hdev->cmd_work);
2917 }
2918 }
2919
2920 static void hci_role_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2921 {
2922 struct hci_ev_role_change *ev = (void *) skb->data;
2923 struct hci_conn *conn;
2924
2925 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2926
2927 hci_dev_lock(hdev);
2928
2929 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2930 if (conn) {
2931 if (!ev->status)
2932 conn->role = ev->role;
2933
2934 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
2935
2936 hci_role_switch_cfm(conn, ev->status, ev->role);
2937 }
2938
2939 hci_dev_unlock(hdev);
2940 }
2941
2942 static void hci_num_comp_pkts_evt(struct hci_dev *hdev, struct sk_buff *skb)
2943 {
2944 struct hci_ev_num_comp_pkts *ev = (void *) skb->data;
2945 int i;
2946
2947 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_PACKET_BASED) {
2948 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
2949 return;
2950 }
2951
2952 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
2953 ev->num_hndl * sizeof(struct hci_comp_pkts_info)) {
2954 BT_DBG("%s bad parameters", hdev->name);
2955 return;
2956 }
2957
2958 BT_DBG("%s num_hndl %d", hdev->name, ev->num_hndl);
2959
2960 for (i = 0; i < ev->num_hndl; i++) {
2961 struct hci_comp_pkts_info *info = &ev->handles[i];
2962 struct hci_conn *conn;
2963 __u16 handle, count;
2964
2965 handle = __le16_to_cpu(info->handle);
2966 count = __le16_to_cpu(info->count);
2967
2968 conn = hci_conn_hash_lookup_handle(hdev, handle);
2969 if (!conn)
2970 continue;
2971
2972 conn->sent -= count;
2973
2974 switch (conn->type) {
2975 case ACL_LINK:
2976 hdev->acl_cnt += count;
2977 if (hdev->acl_cnt > hdev->acl_pkts)
2978 hdev->acl_cnt = hdev->acl_pkts;
2979 break;
2980
2981 case LE_LINK:
2982 if (hdev->le_pkts) {
2983 hdev->le_cnt += count;
2984 if (hdev->le_cnt > hdev->le_pkts)
2985 hdev->le_cnt = hdev->le_pkts;
2986 } else {
2987 hdev->acl_cnt += count;
2988 if (hdev->acl_cnt > hdev->acl_pkts)
2989 hdev->acl_cnt = hdev->acl_pkts;
2990 }
2991 break;
2992
2993 case SCO_LINK:
2994 hdev->sco_cnt += count;
2995 if (hdev->sco_cnt > hdev->sco_pkts)
2996 hdev->sco_cnt = hdev->sco_pkts;
2997 break;
2998
2999 default:
3000 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3001 break;
3002 }
3003 }
3004
3005 queue_work(hdev->workqueue, &hdev->tx_work);
3006 }
3007
3008 static struct hci_conn *__hci_conn_lookup_handle(struct hci_dev *hdev,
3009 __u16 handle)
3010 {
3011 struct hci_chan *chan;
3012
3013 switch (hdev->dev_type) {
3014 case HCI_BREDR:
3015 return hci_conn_hash_lookup_handle(hdev, handle);
3016 case HCI_AMP:
3017 chan = hci_chan_lookup_handle(hdev, handle);
3018 if (chan)
3019 return chan->conn;
3020 break;
3021 default:
3022 BT_ERR("%s unknown dev_type %d", hdev->name, hdev->dev_type);
3023 break;
3024 }
3025
3026 return NULL;
3027 }
3028
3029 static void hci_num_comp_blocks_evt(struct hci_dev *hdev, struct sk_buff *skb)
3030 {
3031 struct hci_ev_num_comp_blocks *ev = (void *) skb->data;
3032 int i;
3033
3034 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_BLOCK_BASED) {
3035 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
3036 return;
3037 }
3038
3039 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
3040 ev->num_hndl * sizeof(struct hci_comp_blocks_info)) {
3041 BT_DBG("%s bad parameters", hdev->name);
3042 return;
3043 }
3044
3045 BT_DBG("%s num_blocks %d num_hndl %d", hdev->name, ev->num_blocks,
3046 ev->num_hndl);
3047
3048 for (i = 0; i < ev->num_hndl; i++) {
3049 struct hci_comp_blocks_info *info = &ev->handles[i];
3050 struct hci_conn *conn = NULL;
3051 __u16 handle, block_count;
3052
3053 handle = __le16_to_cpu(info->handle);
3054 block_count = __le16_to_cpu(info->blocks);
3055
3056 conn = __hci_conn_lookup_handle(hdev, handle);
3057 if (!conn)
3058 continue;
3059
3060 conn->sent -= block_count;
3061
3062 switch (conn->type) {
3063 case ACL_LINK:
3064 case AMP_LINK:
3065 hdev->block_cnt += block_count;
3066 if (hdev->block_cnt > hdev->num_blocks)
3067 hdev->block_cnt = hdev->num_blocks;
3068 break;
3069
3070 default:
3071 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3072 break;
3073 }
3074 }
3075
3076 queue_work(hdev->workqueue, &hdev->tx_work);
3077 }
3078
3079 static void hci_mode_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3080 {
3081 struct hci_ev_mode_change *ev = (void *) skb->data;
3082 struct hci_conn *conn;
3083
3084 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3085
3086 hci_dev_lock(hdev);
3087
3088 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3089 if (conn) {
3090 conn->mode = ev->mode;
3091
3092 if (!test_and_clear_bit(HCI_CONN_MODE_CHANGE_PEND,
3093 &conn->flags)) {
3094 if (conn->mode == HCI_CM_ACTIVE)
3095 set_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3096 else
3097 clear_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3098 }
3099
3100 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
3101 hci_sco_setup(conn, ev->status);
3102 }
3103
3104 hci_dev_unlock(hdev);
3105 }
3106
3107 static void hci_pin_code_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3108 {
3109 struct hci_ev_pin_code_req *ev = (void *) skb->data;
3110 struct hci_conn *conn;
3111
3112 BT_DBG("%s", hdev->name);
3113
3114 hci_dev_lock(hdev);
3115
3116 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3117 if (!conn)
3118 goto unlock;
3119
3120 if (conn->state == BT_CONNECTED) {
3121 hci_conn_hold(conn);
3122 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
3123 hci_conn_drop(conn);
3124 }
3125
3126 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags) &&
3127 !test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags)) {
3128 hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3129 sizeof(ev->bdaddr), &ev->bdaddr);
3130 } else if (test_bit(HCI_MGMT, &hdev->dev_flags)) {
3131 u8 secure;
3132
3133 if (conn->pending_sec_level == BT_SECURITY_HIGH)
3134 secure = 1;
3135 else
3136 secure = 0;
3137
3138 mgmt_pin_code_request(hdev, &ev->bdaddr, secure);
3139 }
3140
3141 unlock:
3142 hci_dev_unlock(hdev);
3143 }
3144
3145 static void hci_link_key_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3146 {
3147 struct hci_ev_link_key_req *ev = (void *) skb->data;
3148 struct hci_cp_link_key_reply cp;
3149 struct hci_conn *conn;
3150 struct link_key *key;
3151
3152 BT_DBG("%s", hdev->name);
3153
3154 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3155 return;
3156
3157 hci_dev_lock(hdev);
3158
3159 key = hci_find_link_key(hdev, &ev->bdaddr);
3160 if (!key) {
3161 BT_DBG("%s link key not found for %pMR", hdev->name,
3162 &ev->bdaddr);
3163 goto not_found;
3164 }
3165
3166 BT_DBG("%s found key type %u for %pMR", hdev->name, key->type,
3167 &ev->bdaddr);
3168
3169 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3170 if (conn) {
3171 if ((key->type == HCI_LK_UNAUTH_COMBINATION_P192 ||
3172 key->type == HCI_LK_UNAUTH_COMBINATION_P256) &&
3173 conn->auth_type != 0xff && (conn->auth_type & 0x01)) {
3174 BT_DBG("%s ignoring unauthenticated key", hdev->name);
3175 goto not_found;
3176 }
3177
3178 if (key->type == HCI_LK_COMBINATION && key->pin_len < 16 &&
3179 (conn->pending_sec_level == BT_SECURITY_HIGH ||
3180 conn->pending_sec_level == BT_SECURITY_FIPS)) {
3181 BT_DBG("%s ignoring key unauthenticated for high security",
3182 hdev->name);
3183 goto not_found;
3184 }
3185
3186 conn->key_type = key->type;
3187 conn->pin_length = key->pin_len;
3188 }
3189
3190 bacpy(&cp.bdaddr, &ev->bdaddr);
3191 memcpy(cp.link_key, key->val, HCI_LINK_KEY_SIZE);
3192
3193 hci_send_cmd(hdev, HCI_OP_LINK_KEY_REPLY, sizeof(cp), &cp);
3194
3195 hci_dev_unlock(hdev);
3196
3197 return;
3198
3199 not_found:
3200 hci_send_cmd(hdev, HCI_OP_LINK_KEY_NEG_REPLY, 6, &ev->bdaddr);
3201 hci_dev_unlock(hdev);
3202 }
3203
3204 static void hci_link_key_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3205 {
3206 struct hci_ev_link_key_notify *ev = (void *) skb->data;
3207 struct hci_conn *conn;
3208 struct link_key *key;
3209 bool persistent;
3210 u8 pin_len = 0;
3211
3212 BT_DBG("%s", hdev->name);
3213
3214 hci_dev_lock(hdev);
3215
3216 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3217 if (conn) {
3218 hci_conn_hold(conn);
3219 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3220 pin_len = conn->pin_length;
3221
3222 if (ev->key_type != HCI_LK_CHANGED_COMBINATION)
3223 conn->key_type = ev->key_type;
3224
3225 hci_conn_drop(conn);
3226 }
3227
3228 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3229 goto unlock;
3230
3231 key = hci_add_link_key(hdev, conn, &ev->bdaddr, ev->link_key,
3232 ev->key_type, pin_len, &persistent);
3233 if (!key)
3234 goto unlock;
3235
3236 mgmt_new_link_key(hdev, key, persistent);
3237
3238 /* Keep debug keys around only if the HCI_KEEP_DEBUG_KEYS flag
3239 * is set. If it's not set simply remove the key from the kernel
3240 * list (we've still notified user space about it but with
3241 * store_hint being 0).
3242 */
3243 if (key->type == HCI_LK_DEBUG_COMBINATION &&
3244 !test_bit(HCI_KEEP_DEBUG_KEYS, &hdev->dev_flags)) {
3245 list_del(&key->list);
3246 kfree(key);
3247 } else if (conn) {
3248 if (persistent)
3249 clear_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3250 else
3251 set_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3252 }
3253
3254 unlock:
3255 hci_dev_unlock(hdev);
3256 }
3257
3258 static void hci_clock_offset_evt(struct hci_dev *hdev, struct sk_buff *skb)
3259 {
3260 struct hci_ev_clock_offset *ev = (void *) skb->data;
3261 struct hci_conn *conn;
3262
3263 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3264
3265 hci_dev_lock(hdev);
3266
3267 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3268 if (conn && !ev->status) {
3269 struct inquiry_entry *ie;
3270
3271 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3272 if (ie) {
3273 ie->data.clock_offset = ev->clock_offset;
3274 ie->timestamp = jiffies;
3275 }
3276 }
3277
3278 hci_dev_unlock(hdev);
3279 }
3280
3281 static void hci_pkt_type_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3282 {
3283 struct hci_ev_pkt_type_change *ev = (void *) skb->data;
3284 struct hci_conn *conn;
3285
3286 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3287
3288 hci_dev_lock(hdev);
3289
3290 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3291 if (conn && !ev->status)
3292 conn->pkt_type = __le16_to_cpu(ev->pkt_type);
3293
3294 hci_dev_unlock(hdev);
3295 }
3296
3297 static void hci_pscan_rep_mode_evt(struct hci_dev *hdev, struct sk_buff *skb)
3298 {
3299 struct hci_ev_pscan_rep_mode *ev = (void *) skb->data;
3300 struct inquiry_entry *ie;
3301
3302 BT_DBG("%s", hdev->name);
3303
3304 hci_dev_lock(hdev);
3305
3306 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3307 if (ie) {
3308 ie->data.pscan_rep_mode = ev->pscan_rep_mode;
3309 ie->timestamp = jiffies;
3310 }
3311
3312 hci_dev_unlock(hdev);
3313 }
3314
3315 static void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev,
3316 struct sk_buff *skb)
3317 {
3318 struct inquiry_data data;
3319 int num_rsp = *((__u8 *) skb->data);
3320
3321 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3322
3323 if (!num_rsp)
3324 return;
3325
3326 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3327 return;
3328
3329 hci_dev_lock(hdev);
3330
3331 if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) {
3332 struct inquiry_info_with_rssi_and_pscan_mode *info;
3333 info = (void *) (skb->data + 1);
3334
3335 for (; num_rsp; num_rsp--, info++) {
3336 u32 flags;
3337
3338 bacpy(&data.bdaddr, &info->bdaddr);
3339 data.pscan_rep_mode = info->pscan_rep_mode;
3340 data.pscan_period_mode = info->pscan_period_mode;
3341 data.pscan_mode = info->pscan_mode;
3342 memcpy(data.dev_class, info->dev_class, 3);
3343 data.clock_offset = info->clock_offset;
3344 data.rssi = info->rssi;
3345 data.ssp_mode = 0x00;
3346
3347 flags = hci_inquiry_cache_update(hdev, &data, false);
3348
3349 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3350 info->dev_class, info->rssi,
3351 flags, NULL, 0, NULL, 0);
3352 }
3353 } else {
3354 struct inquiry_info_with_rssi *info = (void *) (skb->data + 1);
3355
3356 for (; num_rsp; num_rsp--, info++) {
3357 u32 flags;
3358
3359 bacpy(&data.bdaddr, &info->bdaddr);
3360 data.pscan_rep_mode = info->pscan_rep_mode;
3361 data.pscan_period_mode = info->pscan_period_mode;
3362 data.pscan_mode = 0x00;
3363 memcpy(data.dev_class, info->dev_class, 3);
3364 data.clock_offset = info->clock_offset;
3365 data.rssi = info->rssi;
3366 data.ssp_mode = 0x00;
3367
3368 flags = hci_inquiry_cache_update(hdev, &data, false);
3369
3370 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3371 info->dev_class, info->rssi,
3372 flags, NULL, 0, NULL, 0);
3373 }
3374 }
3375
3376 hci_dev_unlock(hdev);
3377 }
3378
3379 static void hci_remote_ext_features_evt(struct hci_dev *hdev,
3380 struct sk_buff *skb)
3381 {
3382 struct hci_ev_remote_ext_features *ev = (void *) skb->data;
3383 struct hci_conn *conn;
3384
3385 BT_DBG("%s", hdev->name);
3386
3387 hci_dev_lock(hdev);
3388
3389 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3390 if (!conn)
3391 goto unlock;
3392
3393 if (ev->page < HCI_MAX_PAGES)
3394 memcpy(conn->features[ev->page], ev->features, 8);
3395
3396 if (!ev->status && ev->page == 0x01) {
3397 struct inquiry_entry *ie;
3398
3399 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3400 if (ie)
3401 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3402
3403 if (ev->features[0] & LMP_HOST_SSP) {
3404 set_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3405 } else {
3406 /* It is mandatory by the Bluetooth specification that
3407 * Extended Inquiry Results are only used when Secure
3408 * Simple Pairing is enabled, but some devices violate
3409 * this.
3410 *
3411 * To make these devices work, the internal SSP
3412 * enabled flag needs to be cleared if the remote host
3413 * features do not indicate SSP support */
3414 clear_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3415 }
3416
3417 if (ev->features[0] & LMP_HOST_SC)
3418 set_bit(HCI_CONN_SC_ENABLED, &conn->flags);
3419 }
3420
3421 if (conn->state != BT_CONFIG)
3422 goto unlock;
3423
3424 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
3425 struct hci_cp_remote_name_req cp;
3426 memset(&cp, 0, sizeof(cp));
3427 bacpy(&cp.bdaddr, &conn->dst);
3428 cp.pscan_rep_mode = 0x02;
3429 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
3430 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
3431 mgmt_device_connected(hdev, &conn->dst, conn->type,
3432 conn->dst_type, 0, NULL, 0,
3433 conn->dev_class);
3434
3435 if (!hci_outgoing_auth_needed(hdev, conn)) {
3436 conn->state = BT_CONNECTED;
3437 hci_proto_connect_cfm(conn, ev->status);
3438 hci_conn_drop(conn);
3439 }
3440
3441 unlock:
3442 hci_dev_unlock(hdev);
3443 }
3444
3445 static void hci_sync_conn_complete_evt(struct hci_dev *hdev,
3446 struct sk_buff *skb)
3447 {
3448 struct hci_ev_sync_conn_complete *ev = (void *) skb->data;
3449 struct hci_conn *conn;
3450
3451 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3452
3453 hci_dev_lock(hdev);
3454
3455 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
3456 if (!conn) {
3457 if (ev->link_type == ESCO_LINK)
3458 goto unlock;
3459
3460 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
3461 if (!conn)
3462 goto unlock;
3463
3464 conn->type = SCO_LINK;
3465 }
3466
3467 switch (ev->status) {
3468 case 0x00:
3469 conn->handle = __le16_to_cpu(ev->handle);
3470 conn->state = BT_CONNECTED;
3471
3472 hci_conn_add_sysfs(conn);
3473 break;
3474
3475 case 0x10: /* Connection Accept Timeout */
3476 case 0x0d: /* Connection Rejected due to Limited Resources */
3477 case 0x11: /* Unsupported Feature or Parameter Value */
3478 case 0x1c: /* SCO interval rejected */
3479 case 0x1a: /* Unsupported Remote Feature */
3480 case 0x1f: /* Unspecified error */
3481 case 0x20: /* Unsupported LMP Parameter value */
3482 if (conn->out) {
3483 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
3484 (hdev->esco_type & EDR_ESCO_MASK);
3485 if (hci_setup_sync(conn, conn->link->handle))
3486 goto unlock;
3487 }
3488 /* fall through */
3489
3490 default:
3491 conn->state = BT_CLOSED;
3492 break;
3493 }
3494
3495 hci_proto_connect_cfm(conn, ev->status);
3496 if (ev->status)
3497 hci_conn_del(conn);
3498
3499 unlock:
3500 hci_dev_unlock(hdev);
3501 }
3502
3503 static inline size_t eir_get_length(u8 *eir, size_t eir_len)
3504 {
3505 size_t parsed = 0;
3506
3507 while (parsed < eir_len) {
3508 u8 field_len = eir[0];
3509
3510 if (field_len == 0)
3511 return parsed;
3512
3513 parsed += field_len + 1;
3514 eir += field_len + 1;
3515 }
3516
3517 return eir_len;
3518 }
3519
3520 static void hci_extended_inquiry_result_evt(struct hci_dev *hdev,
3521 struct sk_buff *skb)
3522 {
3523 struct inquiry_data data;
3524 struct extended_inquiry_info *info = (void *) (skb->data + 1);
3525 int num_rsp = *((__u8 *) skb->data);
3526 size_t eir_len;
3527
3528 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3529
3530 if (!num_rsp)
3531 return;
3532
3533 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3534 return;
3535
3536 hci_dev_lock(hdev);
3537
3538 for (; num_rsp; num_rsp--, info++) {
3539 u32 flags;
3540 bool name_known;
3541
3542 bacpy(&data.bdaddr, &info->bdaddr);
3543 data.pscan_rep_mode = info->pscan_rep_mode;
3544 data.pscan_period_mode = info->pscan_period_mode;
3545 data.pscan_mode = 0x00;
3546 memcpy(data.dev_class, info->dev_class, 3);
3547 data.clock_offset = info->clock_offset;
3548 data.rssi = info->rssi;
3549 data.ssp_mode = 0x01;
3550
3551 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3552 name_known = eir_has_data_type(info->data,
3553 sizeof(info->data),
3554 EIR_NAME_COMPLETE);
3555 else
3556 name_known = true;
3557
3558 flags = hci_inquiry_cache_update(hdev, &data, name_known);
3559
3560 eir_len = eir_get_length(info->data, sizeof(info->data));
3561
3562 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3563 info->dev_class, info->rssi,
3564 flags, info->data, eir_len, NULL, 0);
3565 }
3566
3567 hci_dev_unlock(hdev);
3568 }
3569
3570 static void hci_key_refresh_complete_evt(struct hci_dev *hdev,
3571 struct sk_buff *skb)
3572 {
3573 struct hci_ev_key_refresh_complete *ev = (void *) skb->data;
3574 struct hci_conn *conn;
3575
3576 BT_DBG("%s status 0x%2.2x handle 0x%4.4x", hdev->name, ev->status,
3577 __le16_to_cpu(ev->handle));
3578
3579 hci_dev_lock(hdev);
3580
3581 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3582 if (!conn)
3583 goto unlock;
3584
3585 /* For BR/EDR the necessary steps are taken through the
3586 * auth_complete event.
3587 */
3588 if (conn->type != LE_LINK)
3589 goto unlock;
3590
3591 if (!ev->status)
3592 conn->sec_level = conn->pending_sec_level;
3593
3594 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
3595
3596 if (ev->status && conn->state == BT_CONNECTED) {
3597 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
3598 hci_conn_drop(conn);
3599 goto unlock;
3600 }
3601
3602 if (conn->state == BT_CONFIG) {
3603 if (!ev->status)
3604 conn->state = BT_CONNECTED;
3605
3606 hci_proto_connect_cfm(conn, ev->status);
3607 hci_conn_drop(conn);
3608 } else {
3609 hci_auth_cfm(conn, ev->status);
3610
3611 hci_conn_hold(conn);
3612 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3613 hci_conn_drop(conn);
3614 }
3615
3616 unlock:
3617 hci_dev_unlock(hdev);
3618 }
3619
3620 static u8 hci_get_auth_req(struct hci_conn *conn)
3621 {
3622 /* If remote requests no-bonding follow that lead */
3623 if (conn->remote_auth == HCI_AT_NO_BONDING ||
3624 conn->remote_auth == HCI_AT_NO_BONDING_MITM)
3625 return conn->remote_auth | (conn->auth_type & 0x01);
3626
3627 /* If both remote and local have enough IO capabilities, require
3628 * MITM protection
3629 */
3630 if (conn->remote_cap != HCI_IO_NO_INPUT_OUTPUT &&
3631 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT)
3632 return conn->remote_auth | 0x01;
3633
3634 /* No MITM protection possible so ignore remote requirement */
3635 return (conn->remote_auth & ~0x01) | (conn->auth_type & 0x01);
3636 }
3637
3638 static void hci_io_capa_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3639 {
3640 struct hci_ev_io_capa_request *ev = (void *) skb->data;
3641 struct hci_conn *conn;
3642
3643 BT_DBG("%s", hdev->name);
3644
3645 hci_dev_lock(hdev);
3646
3647 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3648 if (!conn)
3649 goto unlock;
3650
3651 hci_conn_hold(conn);
3652
3653 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3654 goto unlock;
3655
3656 /* Allow pairing if we're pairable, the initiators of the
3657 * pairing or if the remote is not requesting bonding.
3658 */
3659 if (test_bit(HCI_BONDABLE, &hdev->dev_flags) ||
3660 test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags) ||
3661 (conn->remote_auth & ~0x01) == HCI_AT_NO_BONDING) {
3662 struct hci_cp_io_capability_reply cp;
3663
3664 bacpy(&cp.bdaddr, &ev->bdaddr);
3665 /* Change the IO capability from KeyboardDisplay
3666 * to DisplayYesNo as it is not supported by BT spec. */
3667 cp.capability = (conn->io_capability == 0x04) ?
3668 HCI_IO_DISPLAY_YESNO : conn->io_capability;
3669
3670 /* If we are initiators, there is no remote information yet */
3671 if (conn->remote_auth == 0xff) {
3672 /* Request MITM protection if our IO caps allow it
3673 * except for the no-bonding case.
3674 */
3675 if (conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3676 conn->auth_type != HCI_AT_NO_BONDING)
3677 conn->auth_type |= 0x01;
3678 } else {
3679 conn->auth_type = hci_get_auth_req(conn);
3680 }
3681
3682 /* If we're not bondable, force one of the non-bondable
3683 * authentication requirement values.
3684 */
3685 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags))
3686 conn->auth_type &= HCI_AT_NO_BONDING_MITM;
3687
3688 cp.authentication = conn->auth_type;
3689
3690 if (hci_find_remote_oob_data(hdev, &conn->dst) &&
3691 (conn->out || test_bit(HCI_CONN_REMOTE_OOB, &conn->flags)))
3692 cp.oob_data = 0x01;
3693 else
3694 cp.oob_data = 0x00;
3695
3696 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_REPLY,
3697 sizeof(cp), &cp);
3698 } else {
3699 struct hci_cp_io_capability_neg_reply cp;
3700
3701 bacpy(&cp.bdaddr, &ev->bdaddr);
3702 cp.reason = HCI_ERROR_PAIRING_NOT_ALLOWED;
3703
3704 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_NEG_REPLY,
3705 sizeof(cp), &cp);
3706 }
3707
3708 unlock:
3709 hci_dev_unlock(hdev);
3710 }
3711
3712 static void hci_io_capa_reply_evt(struct hci_dev *hdev, struct sk_buff *skb)
3713 {
3714 struct hci_ev_io_capa_reply *ev = (void *) skb->data;
3715 struct hci_conn *conn;
3716
3717 BT_DBG("%s", hdev->name);
3718
3719 hci_dev_lock(hdev);
3720
3721 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3722 if (!conn)
3723 goto unlock;
3724
3725 conn->remote_cap = ev->capability;
3726 conn->remote_auth = ev->authentication;
3727 if (ev->oob_data)
3728 set_bit(HCI_CONN_REMOTE_OOB, &conn->flags);
3729
3730 unlock:
3731 hci_dev_unlock(hdev);
3732 }
3733
3734 static void hci_user_confirm_request_evt(struct hci_dev *hdev,
3735 struct sk_buff *skb)
3736 {
3737 struct hci_ev_user_confirm_req *ev = (void *) skb->data;
3738 int loc_mitm, rem_mitm, confirm_hint = 0;
3739 struct hci_conn *conn;
3740
3741 BT_DBG("%s", hdev->name);
3742
3743 hci_dev_lock(hdev);
3744
3745 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3746 goto unlock;
3747
3748 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3749 if (!conn)
3750 goto unlock;
3751
3752 loc_mitm = (conn->auth_type & 0x01);
3753 rem_mitm = (conn->remote_auth & 0x01);
3754
3755 /* If we require MITM but the remote device can't provide that
3756 * (it has NoInputNoOutput) then reject the confirmation
3757 * request. We check the security level here since it doesn't
3758 * necessarily match conn->auth_type.
3759 */
3760 if (conn->pending_sec_level > BT_SECURITY_MEDIUM &&
3761 conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) {
3762 BT_DBG("Rejecting request: remote device can't provide MITM");
3763 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_NEG_REPLY,
3764 sizeof(ev->bdaddr), &ev->bdaddr);
3765 goto unlock;
3766 }
3767
3768 /* If no side requires MITM protection; auto-accept */
3769 if ((!loc_mitm || conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) &&
3770 (!rem_mitm || conn->io_capability == HCI_IO_NO_INPUT_OUTPUT)) {
3771
3772 /* If we're not the initiators request authorization to
3773 * proceed from user space (mgmt_user_confirm with
3774 * confirm_hint set to 1). The exception is if neither
3775 * side had MITM or if the local IO capability is
3776 * NoInputNoOutput, in which case we do auto-accept
3777 */
3778 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) &&
3779 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3780 (loc_mitm || rem_mitm)) {
3781 BT_DBG("Confirming auto-accept as acceptor");
3782 confirm_hint = 1;
3783 goto confirm;
3784 }
3785
3786 BT_DBG("Auto-accept of user confirmation with %ums delay",
3787 hdev->auto_accept_delay);
3788
3789 if (hdev->auto_accept_delay > 0) {
3790 int delay = msecs_to_jiffies(hdev->auto_accept_delay);
3791 queue_delayed_work(conn->hdev->workqueue,
3792 &conn->auto_accept_work, delay);
3793 goto unlock;
3794 }
3795
3796 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY,
3797 sizeof(ev->bdaddr), &ev->bdaddr);
3798 goto unlock;
3799 }
3800
3801 confirm:
3802 mgmt_user_confirm_request(hdev, &ev->bdaddr, ACL_LINK, 0,
3803 le32_to_cpu(ev->passkey), confirm_hint);
3804
3805 unlock:
3806 hci_dev_unlock(hdev);
3807 }
3808
3809 static void hci_user_passkey_request_evt(struct hci_dev *hdev,
3810 struct sk_buff *skb)
3811 {
3812 struct hci_ev_user_passkey_req *ev = (void *) skb->data;
3813
3814 BT_DBG("%s", hdev->name);
3815
3816 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3817 mgmt_user_passkey_request(hdev, &ev->bdaddr, ACL_LINK, 0);
3818 }
3819
3820 static void hci_user_passkey_notify_evt(struct hci_dev *hdev,
3821 struct sk_buff *skb)
3822 {
3823 struct hci_ev_user_passkey_notify *ev = (void *) skb->data;
3824 struct hci_conn *conn;
3825
3826 BT_DBG("%s", hdev->name);
3827
3828 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3829 if (!conn)
3830 return;
3831
3832 conn->passkey_notify = __le32_to_cpu(ev->passkey);
3833 conn->passkey_entered = 0;
3834
3835 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3836 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3837 conn->dst_type, conn->passkey_notify,
3838 conn->passkey_entered);
3839 }
3840
3841 static void hci_keypress_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3842 {
3843 struct hci_ev_keypress_notify *ev = (void *) skb->data;
3844 struct hci_conn *conn;
3845
3846 BT_DBG("%s", hdev->name);
3847
3848 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3849 if (!conn)
3850 return;
3851
3852 switch (ev->type) {
3853 case HCI_KEYPRESS_STARTED:
3854 conn->passkey_entered = 0;
3855 return;
3856
3857 case HCI_KEYPRESS_ENTERED:
3858 conn->passkey_entered++;
3859 break;
3860
3861 case HCI_KEYPRESS_ERASED:
3862 conn->passkey_entered--;
3863 break;
3864
3865 case HCI_KEYPRESS_CLEARED:
3866 conn->passkey_entered = 0;
3867 break;
3868
3869 case HCI_KEYPRESS_COMPLETED:
3870 return;
3871 }
3872
3873 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3874 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3875 conn->dst_type, conn->passkey_notify,
3876 conn->passkey_entered);
3877 }
3878
3879 static void hci_simple_pair_complete_evt(struct hci_dev *hdev,
3880 struct sk_buff *skb)
3881 {
3882 struct hci_ev_simple_pair_complete *ev = (void *) skb->data;
3883 struct hci_conn *conn;
3884
3885 BT_DBG("%s", hdev->name);
3886
3887 hci_dev_lock(hdev);
3888
3889 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3890 if (!conn)
3891 goto unlock;
3892
3893 /* Reset the authentication requirement to unknown */
3894 conn->remote_auth = 0xff;
3895
3896 /* To avoid duplicate auth_failed events to user space we check
3897 * the HCI_CONN_AUTH_PEND flag which will be set if we
3898 * initiated the authentication. A traditional auth_complete
3899 * event gets always produced as initiator and is also mapped to
3900 * the mgmt_auth_failed event */
3901 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) && ev->status)
3902 mgmt_auth_failed(conn, ev->status);
3903
3904 hci_conn_drop(conn);
3905
3906 unlock:
3907 hci_dev_unlock(hdev);
3908 }
3909
3910 static void hci_remote_host_features_evt(struct hci_dev *hdev,
3911 struct sk_buff *skb)
3912 {
3913 struct hci_ev_remote_host_features *ev = (void *) skb->data;
3914 struct inquiry_entry *ie;
3915 struct hci_conn *conn;
3916
3917 BT_DBG("%s", hdev->name);
3918
3919 hci_dev_lock(hdev);
3920
3921 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3922 if (conn)
3923 memcpy(conn->features[1], ev->features, 8);
3924
3925 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3926 if (ie)
3927 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3928
3929 hci_dev_unlock(hdev);
3930 }
3931
3932 static void hci_remote_oob_data_request_evt(struct hci_dev *hdev,
3933 struct sk_buff *skb)
3934 {
3935 struct hci_ev_remote_oob_data_request *ev = (void *) skb->data;
3936 struct oob_data *data;
3937
3938 BT_DBG("%s", hdev->name);
3939
3940 hci_dev_lock(hdev);
3941
3942 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3943 goto unlock;
3944
3945 data = hci_find_remote_oob_data(hdev, &ev->bdaddr);
3946 if (data) {
3947 if (test_bit(HCI_SC_ENABLED, &hdev->dev_flags)) {
3948 struct hci_cp_remote_oob_ext_data_reply cp;
3949
3950 bacpy(&cp.bdaddr, &ev->bdaddr);
3951 memcpy(cp.hash192, data->hash192, sizeof(cp.hash192));
3952 memcpy(cp.randomizer192, data->randomizer192,
3953 sizeof(cp.randomizer192));
3954 memcpy(cp.hash256, data->hash256, sizeof(cp.hash256));
3955 memcpy(cp.randomizer256, data->randomizer256,
3956 sizeof(cp.randomizer256));
3957
3958 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_EXT_DATA_REPLY,
3959 sizeof(cp), &cp);
3960 } else {
3961 struct hci_cp_remote_oob_data_reply cp;
3962
3963 bacpy(&cp.bdaddr, &ev->bdaddr);
3964 memcpy(cp.hash, data->hash192, sizeof(cp.hash));
3965 memcpy(cp.randomizer, data->randomizer192,
3966 sizeof(cp.randomizer));
3967
3968 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_REPLY,
3969 sizeof(cp), &cp);
3970 }
3971 } else {
3972 struct hci_cp_remote_oob_data_neg_reply cp;
3973
3974 bacpy(&cp.bdaddr, &ev->bdaddr);
3975 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_NEG_REPLY,
3976 sizeof(cp), &cp);
3977 }
3978
3979 unlock:
3980 hci_dev_unlock(hdev);
3981 }
3982
3983 static void hci_phy_link_complete_evt(struct hci_dev *hdev,
3984 struct sk_buff *skb)
3985 {
3986 struct hci_ev_phy_link_complete *ev = (void *) skb->data;
3987 struct hci_conn *hcon, *bredr_hcon;
3988
3989 BT_DBG("%s handle 0x%2.2x status 0x%2.2x", hdev->name, ev->phy_handle,
3990 ev->status);
3991
3992 hci_dev_lock(hdev);
3993
3994 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
3995 if (!hcon) {
3996 hci_dev_unlock(hdev);
3997 return;
3998 }
3999
4000 if (ev->status) {
4001 hci_conn_del(hcon);
4002 hci_dev_unlock(hdev);
4003 return;
4004 }
4005
4006 bredr_hcon = hcon->amp_mgr->l2cap_conn->hcon;
4007
4008 hcon->state = BT_CONNECTED;
4009 bacpy(&hcon->dst, &bredr_hcon->dst);
4010
4011 hci_conn_hold(hcon);
4012 hcon->disc_timeout = HCI_DISCONN_TIMEOUT;
4013 hci_conn_drop(hcon);
4014
4015 hci_conn_add_sysfs(hcon);
4016
4017 amp_physical_cfm(bredr_hcon, hcon);
4018
4019 hci_dev_unlock(hdev);
4020 }
4021
4022 static void hci_loglink_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4023 {
4024 struct hci_ev_logical_link_complete *ev = (void *) skb->data;
4025 struct hci_conn *hcon;
4026 struct hci_chan *hchan;
4027 struct amp_mgr *mgr;
4028
4029 BT_DBG("%s log_handle 0x%4.4x phy_handle 0x%2.2x status 0x%2.2x",
4030 hdev->name, le16_to_cpu(ev->handle), ev->phy_handle,
4031 ev->status);
4032
4033 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4034 if (!hcon)
4035 return;
4036
4037 /* Create AMP hchan */
4038 hchan = hci_chan_create(hcon);
4039 if (!hchan)
4040 return;
4041
4042 hchan->handle = le16_to_cpu(ev->handle);
4043
4044 BT_DBG("hcon %p mgr %p hchan %p", hcon, hcon->amp_mgr, hchan);
4045
4046 mgr = hcon->amp_mgr;
4047 if (mgr && mgr->bredr_chan) {
4048 struct l2cap_chan *bredr_chan = mgr->bredr_chan;
4049
4050 l2cap_chan_lock(bredr_chan);
4051
4052 bredr_chan->conn->mtu = hdev->block_mtu;
4053 l2cap_logical_cfm(bredr_chan, hchan, 0);
4054 hci_conn_hold(hcon);
4055
4056 l2cap_chan_unlock(bredr_chan);
4057 }
4058 }
4059
4060 static void hci_disconn_loglink_complete_evt(struct hci_dev *hdev,
4061 struct sk_buff *skb)
4062 {
4063 struct hci_ev_disconn_logical_link_complete *ev = (void *) skb->data;
4064 struct hci_chan *hchan;
4065
4066 BT_DBG("%s log handle 0x%4.4x status 0x%2.2x", hdev->name,
4067 le16_to_cpu(ev->handle), ev->status);
4068
4069 if (ev->status)
4070 return;
4071
4072 hci_dev_lock(hdev);
4073
4074 hchan = hci_chan_lookup_handle(hdev, le16_to_cpu(ev->handle));
4075 if (!hchan)
4076 goto unlock;
4077
4078 amp_destroy_logical_link(hchan, ev->reason);
4079
4080 unlock:
4081 hci_dev_unlock(hdev);
4082 }
4083
4084 static void hci_disconn_phylink_complete_evt(struct hci_dev *hdev,
4085 struct sk_buff *skb)
4086 {
4087 struct hci_ev_disconn_phy_link_complete *ev = (void *) skb->data;
4088 struct hci_conn *hcon;
4089
4090 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4091
4092 if (ev->status)
4093 return;
4094
4095 hci_dev_lock(hdev);
4096
4097 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4098 if (hcon) {
4099 hcon->state = BT_CLOSED;
4100 hci_conn_del(hcon);
4101 }
4102
4103 hci_dev_unlock(hdev);
4104 }
4105
4106 static void hci_le_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4107 {
4108 struct hci_ev_le_conn_complete *ev = (void *) skb->data;
4109 struct hci_conn_params *params;
4110 struct hci_conn *conn;
4111 struct smp_irk *irk;
4112 u8 addr_type;
4113
4114 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4115
4116 hci_dev_lock(hdev);
4117
4118 /* All controllers implicitly stop advertising in the event of a
4119 * connection, so ensure that the state bit is cleared.
4120 */
4121 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
4122
4123 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
4124 if (!conn) {
4125 conn = hci_conn_add(hdev, LE_LINK, &ev->bdaddr, ev->role);
4126 if (!conn) {
4127 BT_ERR("No memory for new connection");
4128 goto unlock;
4129 }
4130
4131 conn->dst_type = ev->bdaddr_type;
4132
4133 /* If we didn't have a hci_conn object previously
4134 * but we're in master role this must be something
4135 * initiated using a white list. Since white list based
4136 * connections are not "first class citizens" we don't
4137 * have full tracking of them. Therefore, we go ahead
4138 * with a "best effort" approach of determining the
4139 * initiator address based on the HCI_PRIVACY flag.
4140 */
4141 if (conn->out) {
4142 conn->resp_addr_type = ev->bdaddr_type;
4143 bacpy(&conn->resp_addr, &ev->bdaddr);
4144 if (test_bit(HCI_PRIVACY, &hdev->dev_flags)) {
4145 conn->init_addr_type = ADDR_LE_DEV_RANDOM;
4146 bacpy(&conn->init_addr, &hdev->rpa);
4147 } else {
4148 hci_copy_identity_address(hdev,
4149 &conn->init_addr,
4150 &conn->init_addr_type);
4151 }
4152 }
4153 } else {
4154 cancel_delayed_work(&conn->le_conn_timeout);
4155 }
4156
4157 if (!conn->out) {
4158 /* Set the responder (our side) address type based on
4159 * the advertising address type.
4160 */
4161 conn->resp_addr_type = hdev->adv_addr_type;
4162 if (hdev->adv_addr_type == ADDR_LE_DEV_RANDOM)
4163 bacpy(&conn->resp_addr, &hdev->random_addr);
4164 else
4165 bacpy(&conn->resp_addr, &hdev->bdaddr);
4166
4167 conn->init_addr_type = ev->bdaddr_type;
4168 bacpy(&conn->init_addr, &ev->bdaddr);
4169
4170 /* For incoming connections, set the default minimum
4171 * and maximum connection interval. They will be used
4172 * to check if the parameters are in range and if not
4173 * trigger the connection update procedure.
4174 */
4175 conn->le_conn_min_interval = hdev->le_conn_min_interval;
4176 conn->le_conn_max_interval = hdev->le_conn_max_interval;
4177 }
4178
4179 /* Lookup the identity address from the stored connection
4180 * address and address type.
4181 *
4182 * When establishing connections to an identity address, the
4183 * connection procedure will store the resolvable random
4184 * address first. Now if it can be converted back into the
4185 * identity address, start using the identity address from
4186 * now on.
4187 */
4188 irk = hci_get_irk(hdev, &conn->dst, conn->dst_type);
4189 if (irk) {
4190 bacpy(&conn->dst, &irk->bdaddr);
4191 conn->dst_type = irk->addr_type;
4192 }
4193
4194 if (ev->status) {
4195 hci_le_conn_failed(conn, ev->status);
4196 goto unlock;
4197 }
4198
4199 if (conn->dst_type == ADDR_LE_DEV_PUBLIC)
4200 addr_type = BDADDR_LE_PUBLIC;
4201 else
4202 addr_type = BDADDR_LE_RANDOM;
4203
4204 /* Drop the connection if the device is blocked */
4205 if (hci_bdaddr_list_lookup(&hdev->blacklist, &conn->dst, addr_type)) {
4206 hci_conn_drop(conn);
4207 goto unlock;
4208 }
4209
4210 if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4211 mgmt_device_connected(hdev, &conn->dst, conn->type,
4212 conn->dst_type, 0, NULL, 0, NULL);
4213
4214 conn->sec_level = BT_SECURITY_LOW;
4215 conn->handle = __le16_to_cpu(ev->handle);
4216 conn->state = BT_CONNECTED;
4217
4218 conn->le_conn_interval = le16_to_cpu(ev->interval);
4219 conn->le_conn_latency = le16_to_cpu(ev->latency);
4220 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4221
4222 hci_conn_add_sysfs(conn);
4223
4224 hci_proto_connect_cfm(conn, ev->status);
4225
4226 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, &conn->dst,
4227 conn->dst_type);
4228 if (params) {
4229 list_del_init(&params->action);
4230 if (params->conn) {
4231 hci_conn_drop(params->conn);
4232 hci_conn_put(params->conn);
4233 params->conn = NULL;
4234 }
4235 }
4236
4237 unlock:
4238 hci_update_background_scan(hdev);
4239 hci_dev_unlock(hdev);
4240 }
4241
4242 static void hci_le_conn_update_complete_evt(struct hci_dev *hdev,
4243 struct sk_buff *skb)
4244 {
4245 struct hci_ev_le_conn_update_complete *ev = (void *) skb->data;
4246 struct hci_conn *conn;
4247
4248 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4249
4250 if (ev->status)
4251 return;
4252
4253 hci_dev_lock(hdev);
4254
4255 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4256 if (conn) {
4257 conn->le_conn_interval = le16_to_cpu(ev->interval);
4258 conn->le_conn_latency = le16_to_cpu(ev->latency);
4259 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4260 }
4261
4262 hci_dev_unlock(hdev);
4263 }
4264
4265 /* This function requires the caller holds hdev->lock */
4266 static void check_pending_le_conn(struct hci_dev *hdev, bdaddr_t *addr,
4267 u8 addr_type, u8 adv_type)
4268 {
4269 struct hci_conn *conn;
4270 struct hci_conn_params *params;
4271
4272 /* If the event is not connectable don't proceed further */
4273 if (adv_type != LE_ADV_IND && adv_type != LE_ADV_DIRECT_IND)
4274 return;
4275
4276 /* Ignore if the device is blocked */
4277 if (hci_bdaddr_list_lookup(&hdev->blacklist, addr, addr_type))
4278 return;
4279
4280 /* Most controller will fail if we try to create new connections
4281 * while we have an existing one in slave role.
4282 */
4283 if (hdev->conn_hash.le_num_slave > 0)
4284 return;
4285
4286 /* If we're not connectable only connect devices that we have in
4287 * our pend_le_conns list.
4288 */
4289 params = hci_pend_le_action_lookup(&hdev->pend_le_conns,
4290 addr, addr_type);
4291 if (!params)
4292 return;
4293
4294 switch (params->auto_connect) {
4295 case HCI_AUTO_CONN_DIRECT:
4296 /* Only devices advertising with ADV_DIRECT_IND are
4297 * triggering a connection attempt. This is allowing
4298 * incoming connections from slave devices.
4299 */
4300 if (adv_type != LE_ADV_DIRECT_IND)
4301 return;
4302 break;
4303 case HCI_AUTO_CONN_ALWAYS:
4304 /* Devices advertising with ADV_IND or ADV_DIRECT_IND
4305 * are triggering a connection attempt. This means
4306 * that incoming connectioms from slave device are
4307 * accepted and also outgoing connections to slave
4308 * devices are established when found.
4309 */
4310 break;
4311 default:
4312 return;
4313 }
4314
4315 conn = hci_connect_le(hdev, addr, addr_type, BT_SECURITY_LOW,
4316 HCI_LE_AUTOCONN_TIMEOUT, HCI_ROLE_MASTER);
4317 if (!IS_ERR(conn)) {
4318 /* Store the pointer since we don't really have any
4319 * other owner of the object besides the params that
4320 * triggered it. This way we can abort the connection if
4321 * the parameters get removed and keep the reference
4322 * count consistent once the connection is established.
4323 */
4324 params->conn = hci_conn_get(conn);
4325 return;
4326 }
4327
4328 switch (PTR_ERR(conn)) {
4329 case -EBUSY:
4330 /* If hci_connect() returns -EBUSY it means there is already
4331 * an LE connection attempt going on. Since controllers don't
4332 * support more than one connection attempt at the time, we
4333 * don't consider this an error case.
4334 */
4335 break;
4336 default:
4337 BT_DBG("Failed to connect: err %ld", PTR_ERR(conn));
4338 }
4339 }
4340
4341 static void process_adv_report(struct hci_dev *hdev, u8 type, bdaddr_t *bdaddr,
4342 u8 bdaddr_type, s8 rssi, u8 *data, u8 len)
4343 {
4344 struct discovery_state *d = &hdev->discovery;
4345 struct smp_irk *irk;
4346 bool match;
4347 u32 flags;
4348
4349 /* Check if we need to convert to identity address */
4350 irk = hci_get_irk(hdev, bdaddr, bdaddr_type);
4351 if (irk) {
4352 bdaddr = &irk->bdaddr;
4353 bdaddr_type = irk->addr_type;
4354 }
4355
4356 /* Check if we have been requested to connect to this device */
4357 check_pending_le_conn(hdev, bdaddr, bdaddr_type, type);
4358
4359 /* Passive scanning shouldn't trigger any device found events,
4360 * except for devices marked as CONN_REPORT for which we do send
4361 * device found events.
4362 */
4363 if (hdev->le_scan_type == LE_SCAN_PASSIVE) {
4364 if (type == LE_ADV_DIRECT_IND)
4365 return;
4366
4367 if (!hci_pend_le_action_lookup(&hdev->pend_le_reports,
4368 bdaddr, bdaddr_type))
4369 return;
4370
4371 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND)
4372 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4373 else
4374 flags = 0;
4375 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4376 rssi, flags, data, len, NULL, 0);
4377 return;
4378 }
4379
4380 /* When receiving non-connectable or scannable undirected
4381 * advertising reports, this means that the remote device is
4382 * not connectable and then clearly indicate this in the
4383 * device found event.
4384 *
4385 * When receiving a scan response, then there is no way to
4386 * know if the remote device is connectable or not. However
4387 * since scan responses are merged with a previously seen
4388 * advertising report, the flags field from that report
4389 * will be used.
4390 *
4391 * In the really unlikely case that a controller get confused
4392 * and just sends a scan response event, then it is marked as
4393 * not connectable as well.
4394 */
4395 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND ||
4396 type == LE_ADV_SCAN_RSP)
4397 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4398 else
4399 flags = 0;
4400
4401 /* If there's nothing pending either store the data from this
4402 * event or send an immediate device found event if the data
4403 * should not be stored for later.
4404 */
4405 if (!has_pending_adv_report(hdev)) {
4406 /* If the report will trigger a SCAN_REQ store it for
4407 * later merging.
4408 */
4409 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4410 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4411 rssi, flags, data, len);
4412 return;
4413 }
4414
4415 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4416 rssi, flags, data, len, NULL, 0);
4417 return;
4418 }
4419
4420 /* Check if the pending report is for the same device as the new one */
4421 match = (!bacmp(bdaddr, &d->last_adv_addr) &&
4422 bdaddr_type == d->last_adv_addr_type);
4423
4424 /* If the pending data doesn't match this report or this isn't a
4425 * scan response (e.g. we got a duplicate ADV_IND) then force
4426 * sending of the pending data.
4427 */
4428 if (type != LE_ADV_SCAN_RSP || !match) {
4429 /* Send out whatever is in the cache, but skip duplicates */
4430 if (!match)
4431 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4432 d->last_adv_addr_type, NULL,
4433 d->last_adv_rssi, d->last_adv_flags,
4434 d->last_adv_data,
4435 d->last_adv_data_len, NULL, 0);
4436
4437 /* If the new report will trigger a SCAN_REQ store it for
4438 * later merging.
4439 */
4440 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4441 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4442 rssi, flags, data, len);
4443 return;
4444 }
4445
4446 /* The advertising reports cannot be merged, so clear
4447 * the pending report and send out a device found event.
4448 */
4449 clear_pending_adv_report(hdev);
4450 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4451 rssi, flags, data, len, NULL, 0);
4452 return;
4453 }
4454
4455 /* If we get here we've got a pending ADV_IND or ADV_SCAN_IND and
4456 * the new event is a SCAN_RSP. We can therefore proceed with
4457 * sending a merged device found event.
4458 */
4459 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4460 d->last_adv_addr_type, NULL, rssi, d->last_adv_flags,
4461 d->last_adv_data, d->last_adv_data_len, data, len);
4462 clear_pending_adv_report(hdev);
4463 }
4464
4465 static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
4466 {
4467 u8 num_reports = skb->data[0];
4468 void *ptr = &skb->data[1];
4469
4470 hci_dev_lock(hdev);
4471
4472 while (num_reports--) {
4473 struct hci_ev_le_advertising_info *ev = ptr;
4474 s8 rssi;
4475
4476 rssi = ev->data[ev->length];
4477 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
4478 ev->bdaddr_type, rssi, ev->data, ev->length);
4479
4480 ptr += sizeof(*ev) + ev->length + 1;
4481 }
4482
4483 hci_dev_unlock(hdev);
4484 }
4485
4486 static void hci_le_ltk_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
4487 {
4488 struct hci_ev_le_ltk_req *ev = (void *) skb->data;
4489 struct hci_cp_le_ltk_reply cp;
4490 struct hci_cp_le_ltk_neg_reply neg;
4491 struct hci_conn *conn;
4492 struct smp_ltk *ltk;
4493
4494 BT_DBG("%s handle 0x%4.4x", hdev->name, __le16_to_cpu(ev->handle));
4495
4496 hci_dev_lock(hdev);
4497
4498 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4499 if (conn == NULL)
4500 goto not_found;
4501
4502 ltk = hci_find_ltk(hdev, ev->ediv, ev->rand, conn->role);
4503 if (ltk == NULL)
4504 goto not_found;
4505
4506 memcpy(cp.ltk, ltk->val, sizeof(ltk->val));
4507 cp.handle = cpu_to_le16(conn->handle);
4508
4509 if (ltk->authenticated)
4510 conn->pending_sec_level = BT_SECURITY_HIGH;
4511 else
4512 conn->pending_sec_level = BT_SECURITY_MEDIUM;
4513
4514 conn->enc_key_size = ltk->enc_size;
4515
4516 hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp);
4517
4518 /* Ref. Bluetooth Core SPEC pages 1975 and 2004. STK is a
4519 * temporary key used to encrypt a connection following
4520 * pairing. It is used during the Encrypted Session Setup to
4521 * distribute the keys. Later, security can be re-established
4522 * using a distributed LTK.
4523 */
4524 if (ltk->type == SMP_STK) {
4525 set_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4526 list_del(&ltk->list);
4527 kfree(ltk);
4528 } else {
4529 clear_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4530 }
4531
4532 hci_dev_unlock(hdev);
4533
4534 return;
4535
4536 not_found:
4537 neg.handle = ev->handle;
4538 hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(neg), &neg);
4539 hci_dev_unlock(hdev);
4540 }
4541
4542 static void send_conn_param_neg_reply(struct hci_dev *hdev, u16 handle,
4543 u8 reason)
4544 {
4545 struct hci_cp_le_conn_param_req_neg_reply cp;
4546
4547 cp.handle = cpu_to_le16(handle);
4548 cp.reason = reason;
4549
4550 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_NEG_REPLY, sizeof(cp),
4551 &cp);
4552 }
4553
4554 static void hci_le_remote_conn_param_req_evt(struct hci_dev *hdev,
4555 struct sk_buff *skb)
4556 {
4557 struct hci_ev_le_remote_conn_param_req *ev = (void *) skb->data;
4558 struct hci_cp_le_conn_param_req_reply cp;
4559 struct hci_conn *hcon;
4560 u16 handle, min, max, latency, timeout;
4561
4562 handle = le16_to_cpu(ev->handle);
4563 min = le16_to_cpu(ev->interval_min);
4564 max = le16_to_cpu(ev->interval_max);
4565 latency = le16_to_cpu(ev->latency);
4566 timeout = le16_to_cpu(ev->timeout);
4567
4568 hcon = hci_conn_hash_lookup_handle(hdev, handle);
4569 if (!hcon || hcon->state != BT_CONNECTED)
4570 return send_conn_param_neg_reply(hdev, handle,
4571 HCI_ERROR_UNKNOWN_CONN_ID);
4572
4573 if (hci_check_conn_params(min, max, latency, timeout))
4574 return send_conn_param_neg_reply(hdev, handle,
4575 HCI_ERROR_INVALID_LL_PARAMS);
4576
4577 if (hcon->role == HCI_ROLE_MASTER) {
4578 struct hci_conn_params *params;
4579 u8 store_hint;
4580
4581 hci_dev_lock(hdev);
4582
4583 params = hci_conn_params_lookup(hdev, &hcon->dst,
4584 hcon->dst_type);
4585 if (params) {
4586 params->conn_min_interval = min;
4587 params->conn_max_interval = max;
4588 params->conn_latency = latency;
4589 params->supervision_timeout = timeout;
4590 store_hint = 0x01;
4591 } else{
4592 store_hint = 0x00;
4593 }
4594
4595 hci_dev_unlock(hdev);
4596
4597 mgmt_new_conn_param(hdev, &hcon->dst, hcon->dst_type,
4598 store_hint, min, max, latency, timeout);
4599 }
4600
4601 cp.handle = ev->handle;
4602 cp.interval_min = ev->interval_min;
4603 cp.interval_max = ev->interval_max;
4604 cp.latency = ev->latency;
4605 cp.timeout = ev->timeout;
4606 cp.min_ce_len = 0;
4607 cp.max_ce_len = 0;
4608
4609 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_REPLY, sizeof(cp), &cp);
4610 }
4611
4612 static void hci_le_meta_evt(struct hci_dev *hdev, struct sk_buff *skb)
4613 {
4614 struct hci_ev_le_meta *le_ev = (void *) skb->data;
4615
4616 skb_pull(skb, sizeof(*le_ev));
4617
4618 switch (le_ev->subevent) {
4619 case HCI_EV_LE_CONN_COMPLETE:
4620 hci_le_conn_complete_evt(hdev, skb);
4621 break;
4622
4623 case HCI_EV_LE_CONN_UPDATE_COMPLETE:
4624 hci_le_conn_update_complete_evt(hdev, skb);
4625 break;
4626
4627 case HCI_EV_LE_ADVERTISING_REPORT:
4628 hci_le_adv_report_evt(hdev, skb);
4629 break;
4630
4631 case HCI_EV_LE_LTK_REQ:
4632 hci_le_ltk_request_evt(hdev, skb);
4633 break;
4634
4635 case HCI_EV_LE_REMOTE_CONN_PARAM_REQ:
4636 hci_le_remote_conn_param_req_evt(hdev, skb);
4637 break;
4638
4639 default:
4640 break;
4641 }
4642 }
4643
4644 static void hci_chan_selected_evt(struct hci_dev *hdev, struct sk_buff *skb)
4645 {
4646 struct hci_ev_channel_selected *ev = (void *) skb->data;
4647 struct hci_conn *hcon;
4648
4649 BT_DBG("%s handle 0x%2.2x", hdev->name, ev->phy_handle);
4650
4651 skb_pull(skb, sizeof(*ev));
4652
4653 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4654 if (!hcon)
4655 return;
4656
4657 amp_read_loc_assoc_final_data(hdev, hcon);
4658 }
4659
4660 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb)
4661 {
4662 struct hci_event_hdr *hdr = (void *) skb->data;
4663 __u8 event = hdr->evt;
4664
4665 hci_dev_lock(hdev);
4666
4667 /* Received events are (currently) only needed when a request is
4668 * ongoing so avoid unnecessary memory allocation.
4669 */
4670 if (hci_req_pending(hdev)) {
4671 kfree_skb(hdev->recv_evt);
4672 hdev->recv_evt = skb_clone(skb, GFP_KERNEL);
4673 }
4674
4675 hci_dev_unlock(hdev);
4676
4677 skb_pull(skb, HCI_EVENT_HDR_SIZE);
4678
4679 if (hdev->sent_cmd && bt_cb(hdev->sent_cmd)->req.event == event) {
4680 struct hci_command_hdr *cmd_hdr = (void *) hdev->sent_cmd->data;
4681 u16 opcode = __le16_to_cpu(cmd_hdr->opcode);
4682
4683 hci_req_cmd_complete(hdev, opcode, 0);
4684 }
4685
4686 switch (event) {
4687 case HCI_EV_INQUIRY_COMPLETE:
4688 hci_inquiry_complete_evt(hdev, skb);
4689 break;
4690
4691 case HCI_EV_INQUIRY_RESULT:
4692 hci_inquiry_result_evt(hdev, skb);
4693 break;
4694
4695 case HCI_EV_CONN_COMPLETE:
4696 hci_conn_complete_evt(hdev, skb);
4697 break;
4698
4699 case HCI_EV_CONN_REQUEST:
4700 hci_conn_request_evt(hdev, skb);
4701 break;
4702
4703 case HCI_EV_DISCONN_COMPLETE:
4704 hci_disconn_complete_evt(hdev, skb);
4705 break;
4706
4707 case HCI_EV_AUTH_COMPLETE:
4708 hci_auth_complete_evt(hdev, skb);
4709 break;
4710
4711 case HCI_EV_REMOTE_NAME:
4712 hci_remote_name_evt(hdev, skb);
4713 break;
4714
4715 case HCI_EV_ENCRYPT_CHANGE:
4716 hci_encrypt_change_evt(hdev, skb);
4717 break;
4718
4719 case HCI_EV_CHANGE_LINK_KEY_COMPLETE:
4720 hci_change_link_key_complete_evt(hdev, skb);
4721 break;
4722
4723 case HCI_EV_REMOTE_FEATURES:
4724 hci_remote_features_evt(hdev, skb);
4725 break;
4726
4727 case HCI_EV_CMD_COMPLETE:
4728 hci_cmd_complete_evt(hdev, skb);
4729 break;
4730
4731 case HCI_EV_CMD_STATUS:
4732 hci_cmd_status_evt(hdev, skb);
4733 break;
4734
4735 case HCI_EV_ROLE_CHANGE:
4736 hci_role_change_evt(hdev, skb);
4737 break;
4738
4739 case HCI_EV_NUM_COMP_PKTS:
4740 hci_num_comp_pkts_evt(hdev, skb);
4741 break;
4742
4743 case HCI_EV_MODE_CHANGE:
4744 hci_mode_change_evt(hdev, skb);
4745 break;
4746
4747 case HCI_EV_PIN_CODE_REQ:
4748 hci_pin_code_request_evt(hdev, skb);
4749 break;
4750
4751 case HCI_EV_LINK_KEY_REQ:
4752 hci_link_key_request_evt(hdev, skb);
4753 break;
4754
4755 case HCI_EV_LINK_KEY_NOTIFY:
4756 hci_link_key_notify_evt(hdev, skb);
4757 break;
4758
4759 case HCI_EV_CLOCK_OFFSET:
4760 hci_clock_offset_evt(hdev, skb);
4761 break;
4762
4763 case HCI_EV_PKT_TYPE_CHANGE:
4764 hci_pkt_type_change_evt(hdev, skb);
4765 break;
4766
4767 case HCI_EV_PSCAN_REP_MODE:
4768 hci_pscan_rep_mode_evt(hdev, skb);
4769 break;
4770
4771 case HCI_EV_INQUIRY_RESULT_WITH_RSSI:
4772 hci_inquiry_result_with_rssi_evt(hdev, skb);
4773 break;
4774
4775 case HCI_EV_REMOTE_EXT_FEATURES:
4776 hci_remote_ext_features_evt(hdev, skb);
4777 break;
4778
4779 case HCI_EV_SYNC_CONN_COMPLETE:
4780 hci_sync_conn_complete_evt(hdev, skb);
4781 break;
4782
4783 case HCI_EV_EXTENDED_INQUIRY_RESULT:
4784 hci_extended_inquiry_result_evt(hdev, skb);
4785 break;
4786
4787 case HCI_EV_KEY_REFRESH_COMPLETE:
4788 hci_key_refresh_complete_evt(hdev, skb);
4789 break;
4790
4791 case HCI_EV_IO_CAPA_REQUEST:
4792 hci_io_capa_request_evt(hdev, skb);
4793 break;
4794
4795 case HCI_EV_IO_CAPA_REPLY:
4796 hci_io_capa_reply_evt(hdev, skb);
4797 break;
4798
4799 case HCI_EV_USER_CONFIRM_REQUEST:
4800 hci_user_confirm_request_evt(hdev, skb);
4801 break;
4802
4803 case HCI_EV_USER_PASSKEY_REQUEST:
4804 hci_user_passkey_request_evt(hdev, skb);
4805 break;
4806
4807 case HCI_EV_USER_PASSKEY_NOTIFY:
4808 hci_user_passkey_notify_evt(hdev, skb);
4809 break;
4810
4811 case HCI_EV_KEYPRESS_NOTIFY:
4812 hci_keypress_notify_evt(hdev, skb);
4813 break;
4814
4815 case HCI_EV_SIMPLE_PAIR_COMPLETE:
4816 hci_simple_pair_complete_evt(hdev, skb);
4817 break;
4818
4819 case HCI_EV_REMOTE_HOST_FEATURES:
4820 hci_remote_host_features_evt(hdev, skb);
4821 break;
4822
4823 case HCI_EV_LE_META:
4824 hci_le_meta_evt(hdev, skb);
4825 break;
4826
4827 case HCI_EV_CHANNEL_SELECTED:
4828 hci_chan_selected_evt(hdev, skb);
4829 break;
4830
4831 case HCI_EV_REMOTE_OOB_DATA_REQUEST:
4832 hci_remote_oob_data_request_evt(hdev, skb);
4833 break;
4834
4835 case HCI_EV_PHY_LINK_COMPLETE:
4836 hci_phy_link_complete_evt(hdev, skb);
4837 break;
4838
4839 case HCI_EV_LOGICAL_LINK_COMPLETE:
4840 hci_loglink_complete_evt(hdev, skb);
4841 break;
4842
4843 case HCI_EV_DISCONN_LOGICAL_LINK_COMPLETE:
4844 hci_disconn_loglink_complete_evt(hdev, skb);
4845 break;
4846
4847 case HCI_EV_DISCONN_PHY_LINK_COMPLETE:
4848 hci_disconn_phylink_complete_evt(hdev, skb);
4849 break;
4850
4851 case HCI_EV_NUM_COMP_BLOCKS:
4852 hci_num_comp_blocks_evt(hdev, skb);
4853 break;
4854
4855 default:
4856 BT_DBG("%s event 0x%2.2x", hdev->name, event);
4857 break;
4858 }
4859
4860 kfree_skb(skb);
4861 hdev->stat.evt_rx++;
4862 }
This page took 0.188081 seconds and 6 git commands to generate.