2f02ff0ed781f88085394acc7a582c2765091d5e
[deliverable/linux.git] / net / bluetooth / hci_event.c
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 /* Bluetooth HCI event handling. */
26
27 #include <asm/unaligned.h>
28
29 #include <net/bluetooth/bluetooth.h>
30 #include <net/bluetooth/hci_core.h>
31 #include <net/bluetooth/mgmt.h>
32
33 #include "a2mp.h"
34 #include "amp.h"
35 #include "smp.h"
36
37 /* Handle HCI Event packets */
38
39 static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb)
40 {
41 __u8 status = *((__u8 *) skb->data);
42
43 BT_DBG("%s status 0x%2.2x", hdev->name, status);
44
45 if (status)
46 return;
47
48 clear_bit(HCI_INQUIRY, &hdev->flags);
49 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
50 wake_up_bit(&hdev->flags, HCI_INQUIRY);
51
52 hci_dev_lock(hdev);
53 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
54 hci_dev_unlock(hdev);
55
56 hci_conn_check_pending(hdev);
57 }
58
59 static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
60 {
61 __u8 status = *((__u8 *) skb->data);
62
63 BT_DBG("%s status 0x%2.2x", hdev->name, status);
64
65 if (status)
66 return;
67
68 set_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
69 }
70
71 static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
72 {
73 __u8 status = *((__u8 *) skb->data);
74
75 BT_DBG("%s status 0x%2.2x", hdev->name, status);
76
77 if (status)
78 return;
79
80 clear_bit(HCI_PERIODIC_INQ, &hdev->dev_flags);
81
82 hci_conn_check_pending(hdev);
83 }
84
85 static void hci_cc_remote_name_req_cancel(struct hci_dev *hdev,
86 struct sk_buff *skb)
87 {
88 BT_DBG("%s", hdev->name);
89 }
90
91 static void hci_cc_role_discovery(struct hci_dev *hdev, struct sk_buff *skb)
92 {
93 struct hci_rp_role_discovery *rp = (void *) skb->data;
94 struct hci_conn *conn;
95
96 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
97
98 if (rp->status)
99 return;
100
101 hci_dev_lock(hdev);
102
103 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
104 if (conn)
105 conn->role = rp->role;
106
107 hci_dev_unlock(hdev);
108 }
109
110 static void hci_cc_read_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
111 {
112 struct hci_rp_read_link_policy *rp = (void *) skb->data;
113 struct hci_conn *conn;
114
115 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
116
117 if (rp->status)
118 return;
119
120 hci_dev_lock(hdev);
121
122 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
123 if (conn)
124 conn->link_policy = __le16_to_cpu(rp->policy);
125
126 hci_dev_unlock(hdev);
127 }
128
129 static void hci_cc_write_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
130 {
131 struct hci_rp_write_link_policy *rp = (void *) skb->data;
132 struct hci_conn *conn;
133 void *sent;
134
135 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
136
137 if (rp->status)
138 return;
139
140 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LINK_POLICY);
141 if (!sent)
142 return;
143
144 hci_dev_lock(hdev);
145
146 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
147 if (conn)
148 conn->link_policy = get_unaligned_le16(sent + 2);
149
150 hci_dev_unlock(hdev);
151 }
152
153 static void hci_cc_read_def_link_policy(struct hci_dev *hdev,
154 struct sk_buff *skb)
155 {
156 struct hci_rp_read_def_link_policy *rp = (void *) skb->data;
157
158 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
159
160 if (rp->status)
161 return;
162
163 hdev->link_policy = __le16_to_cpu(rp->policy);
164 }
165
166 static void hci_cc_write_def_link_policy(struct hci_dev *hdev,
167 struct sk_buff *skb)
168 {
169 __u8 status = *((__u8 *) skb->data);
170 void *sent;
171
172 BT_DBG("%s status 0x%2.2x", hdev->name, status);
173
174 if (status)
175 return;
176
177 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_DEF_LINK_POLICY);
178 if (!sent)
179 return;
180
181 hdev->link_policy = get_unaligned_le16(sent);
182 }
183
184 static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb)
185 {
186 __u8 status = *((__u8 *) skb->data);
187
188 BT_DBG("%s status 0x%2.2x", hdev->name, status);
189
190 clear_bit(HCI_RESET, &hdev->flags);
191
192 if (status)
193 return;
194
195 /* Reset all non-persistent flags */
196 hdev->dev_flags &= ~HCI_PERSISTENT_MASK;
197
198 hdev->discovery.state = DISCOVERY_STOPPED;
199 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
200 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
201
202 memset(hdev->adv_data, 0, sizeof(hdev->adv_data));
203 hdev->adv_data_len = 0;
204
205 memset(hdev->scan_rsp_data, 0, sizeof(hdev->scan_rsp_data));
206 hdev->scan_rsp_data_len = 0;
207
208 hdev->le_scan_type = LE_SCAN_PASSIVE;
209
210 hdev->ssp_debug_mode = 0;
211
212 hci_bdaddr_list_clear(&hdev->le_white_list);
213 }
214
215 static void hci_cc_write_local_name(struct hci_dev *hdev, struct sk_buff *skb)
216 {
217 __u8 status = *((__u8 *) skb->data);
218 void *sent;
219
220 BT_DBG("%s status 0x%2.2x", hdev->name, status);
221
222 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LOCAL_NAME);
223 if (!sent)
224 return;
225
226 hci_dev_lock(hdev);
227
228 if (test_bit(HCI_MGMT, &hdev->dev_flags))
229 mgmt_set_local_name_complete(hdev, sent, status);
230 else if (!status)
231 memcpy(hdev->dev_name, sent, HCI_MAX_NAME_LENGTH);
232
233 hci_dev_unlock(hdev);
234 }
235
236 static void hci_cc_read_local_name(struct hci_dev *hdev, struct sk_buff *skb)
237 {
238 struct hci_rp_read_local_name *rp = (void *) skb->data;
239
240 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
241
242 if (rp->status)
243 return;
244
245 if (test_bit(HCI_SETUP, &hdev->dev_flags))
246 memcpy(hdev->dev_name, rp->name, HCI_MAX_NAME_LENGTH);
247 }
248
249 static void hci_cc_write_auth_enable(struct hci_dev *hdev, struct sk_buff *skb)
250 {
251 __u8 status = *((__u8 *) skb->data);
252 void *sent;
253
254 BT_DBG("%s status 0x%2.2x", hdev->name, status);
255
256 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_AUTH_ENABLE);
257 if (!sent)
258 return;
259
260 if (!status) {
261 __u8 param = *((__u8 *) sent);
262
263 if (param == AUTH_ENABLED)
264 set_bit(HCI_AUTH, &hdev->flags);
265 else
266 clear_bit(HCI_AUTH, &hdev->flags);
267 }
268
269 if (test_bit(HCI_MGMT, &hdev->dev_flags))
270 mgmt_auth_enable_complete(hdev, status);
271 }
272
273 static void hci_cc_write_encrypt_mode(struct hci_dev *hdev, struct sk_buff *skb)
274 {
275 __u8 status = *((__u8 *) skb->data);
276 __u8 param;
277 void *sent;
278
279 BT_DBG("%s status 0x%2.2x", hdev->name, status);
280
281 if (status)
282 return;
283
284 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_ENCRYPT_MODE);
285 if (!sent)
286 return;
287
288 param = *((__u8 *) sent);
289
290 if (param)
291 set_bit(HCI_ENCRYPT, &hdev->flags);
292 else
293 clear_bit(HCI_ENCRYPT, &hdev->flags);
294 }
295
296 static void hci_cc_write_scan_enable(struct hci_dev *hdev, struct sk_buff *skb)
297 {
298 __u8 status = *((__u8 *) skb->data);
299 __u8 param;
300 void *sent;
301
302 BT_DBG("%s status 0x%2.2x", hdev->name, status);
303
304 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SCAN_ENABLE);
305 if (!sent)
306 return;
307
308 param = *((__u8 *) sent);
309
310 hci_dev_lock(hdev);
311
312 if (status) {
313 hdev->discov_timeout = 0;
314 goto done;
315 }
316
317 if (param & SCAN_INQUIRY)
318 set_bit(HCI_ISCAN, &hdev->flags);
319 else
320 clear_bit(HCI_ISCAN, &hdev->flags);
321
322 if (param & SCAN_PAGE)
323 set_bit(HCI_PSCAN, &hdev->flags);
324 else
325 clear_bit(HCI_PSCAN, &hdev->flags);
326
327 done:
328 hci_dev_unlock(hdev);
329 }
330
331 static void hci_cc_read_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
332 {
333 struct hci_rp_read_class_of_dev *rp = (void *) skb->data;
334
335 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
336
337 if (rp->status)
338 return;
339
340 memcpy(hdev->dev_class, rp->dev_class, 3);
341
342 BT_DBG("%s class 0x%.2x%.2x%.2x", hdev->name,
343 hdev->dev_class[2], hdev->dev_class[1], hdev->dev_class[0]);
344 }
345
346 static void hci_cc_write_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
347 {
348 __u8 status = *((__u8 *) skb->data);
349 void *sent;
350
351 BT_DBG("%s status 0x%2.2x", hdev->name, status);
352
353 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_CLASS_OF_DEV);
354 if (!sent)
355 return;
356
357 hci_dev_lock(hdev);
358
359 if (status == 0)
360 memcpy(hdev->dev_class, sent, 3);
361
362 if (test_bit(HCI_MGMT, &hdev->dev_flags))
363 mgmt_set_class_of_dev_complete(hdev, sent, status);
364
365 hci_dev_unlock(hdev);
366 }
367
368 static void hci_cc_read_voice_setting(struct hci_dev *hdev, struct sk_buff *skb)
369 {
370 struct hci_rp_read_voice_setting *rp = (void *) skb->data;
371 __u16 setting;
372
373 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
374
375 if (rp->status)
376 return;
377
378 setting = __le16_to_cpu(rp->voice_setting);
379
380 if (hdev->voice_setting == setting)
381 return;
382
383 hdev->voice_setting = setting;
384
385 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
386
387 if (hdev->notify)
388 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
389 }
390
391 static void hci_cc_write_voice_setting(struct hci_dev *hdev,
392 struct sk_buff *skb)
393 {
394 __u8 status = *((__u8 *) skb->data);
395 __u16 setting;
396 void *sent;
397
398 BT_DBG("%s status 0x%2.2x", hdev->name, status);
399
400 if (status)
401 return;
402
403 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_VOICE_SETTING);
404 if (!sent)
405 return;
406
407 setting = get_unaligned_le16(sent);
408
409 if (hdev->voice_setting == setting)
410 return;
411
412 hdev->voice_setting = setting;
413
414 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
415
416 if (hdev->notify)
417 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
418 }
419
420 static void hci_cc_read_num_supported_iac(struct hci_dev *hdev,
421 struct sk_buff *skb)
422 {
423 struct hci_rp_read_num_supported_iac *rp = (void *) skb->data;
424
425 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
426
427 if (rp->status)
428 return;
429
430 hdev->num_iac = rp->num_iac;
431
432 BT_DBG("%s num iac %d", hdev->name, hdev->num_iac);
433 }
434
435 static void hci_cc_write_ssp_mode(struct hci_dev *hdev, struct sk_buff *skb)
436 {
437 __u8 status = *((__u8 *) skb->data);
438 struct hci_cp_write_ssp_mode *sent;
439
440 BT_DBG("%s status 0x%2.2x", hdev->name, status);
441
442 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_MODE);
443 if (!sent)
444 return;
445
446 if (!status) {
447 if (sent->mode)
448 hdev->features[1][0] |= LMP_HOST_SSP;
449 else
450 hdev->features[1][0] &= ~LMP_HOST_SSP;
451 }
452
453 if (test_bit(HCI_MGMT, &hdev->dev_flags))
454 mgmt_ssp_enable_complete(hdev, sent->mode, status);
455 else if (!status) {
456 if (sent->mode)
457 set_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
458 else
459 clear_bit(HCI_SSP_ENABLED, &hdev->dev_flags);
460 }
461 }
462
463 static void hci_cc_write_sc_support(struct hci_dev *hdev, struct sk_buff *skb)
464 {
465 u8 status = *((u8 *) skb->data);
466 struct hci_cp_write_sc_support *sent;
467
468 BT_DBG("%s status 0x%2.2x", hdev->name, status);
469
470 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SC_SUPPORT);
471 if (!sent)
472 return;
473
474 if (!status) {
475 if (sent->support)
476 hdev->features[1][0] |= LMP_HOST_SC;
477 else
478 hdev->features[1][0] &= ~LMP_HOST_SC;
479 }
480
481 if (test_bit(HCI_MGMT, &hdev->dev_flags))
482 mgmt_sc_enable_complete(hdev, sent->support, status);
483 else if (!status) {
484 if (sent->support)
485 set_bit(HCI_SC_ENABLED, &hdev->dev_flags);
486 else
487 clear_bit(HCI_SC_ENABLED, &hdev->dev_flags);
488 }
489 }
490
491 static void hci_cc_read_local_version(struct hci_dev *hdev, struct sk_buff *skb)
492 {
493 struct hci_rp_read_local_version *rp = (void *) skb->data;
494
495 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
496
497 if (rp->status)
498 return;
499
500 if (test_bit(HCI_SETUP, &hdev->dev_flags)) {
501 hdev->hci_ver = rp->hci_ver;
502 hdev->hci_rev = __le16_to_cpu(rp->hci_rev);
503 hdev->lmp_ver = rp->lmp_ver;
504 hdev->manufacturer = __le16_to_cpu(rp->manufacturer);
505 hdev->lmp_subver = __le16_to_cpu(rp->lmp_subver);
506 }
507 }
508
509 static void hci_cc_read_local_commands(struct hci_dev *hdev,
510 struct sk_buff *skb)
511 {
512 struct hci_rp_read_local_commands *rp = (void *) skb->data;
513
514 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
515
516 if (rp->status)
517 return;
518
519 if (test_bit(HCI_SETUP, &hdev->dev_flags))
520 memcpy(hdev->commands, rp->commands, sizeof(hdev->commands));
521 }
522
523 static void hci_cc_read_local_features(struct hci_dev *hdev,
524 struct sk_buff *skb)
525 {
526 struct hci_rp_read_local_features *rp = (void *) skb->data;
527
528 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
529
530 if (rp->status)
531 return;
532
533 memcpy(hdev->features, rp->features, 8);
534
535 /* Adjust default settings according to features
536 * supported by device. */
537
538 if (hdev->features[0][0] & LMP_3SLOT)
539 hdev->pkt_type |= (HCI_DM3 | HCI_DH3);
540
541 if (hdev->features[0][0] & LMP_5SLOT)
542 hdev->pkt_type |= (HCI_DM5 | HCI_DH5);
543
544 if (hdev->features[0][1] & LMP_HV2) {
545 hdev->pkt_type |= (HCI_HV2);
546 hdev->esco_type |= (ESCO_HV2);
547 }
548
549 if (hdev->features[0][1] & LMP_HV3) {
550 hdev->pkt_type |= (HCI_HV3);
551 hdev->esco_type |= (ESCO_HV3);
552 }
553
554 if (lmp_esco_capable(hdev))
555 hdev->esco_type |= (ESCO_EV3);
556
557 if (hdev->features[0][4] & LMP_EV4)
558 hdev->esco_type |= (ESCO_EV4);
559
560 if (hdev->features[0][4] & LMP_EV5)
561 hdev->esco_type |= (ESCO_EV5);
562
563 if (hdev->features[0][5] & LMP_EDR_ESCO_2M)
564 hdev->esco_type |= (ESCO_2EV3);
565
566 if (hdev->features[0][5] & LMP_EDR_ESCO_3M)
567 hdev->esco_type |= (ESCO_3EV3);
568
569 if (hdev->features[0][5] & LMP_EDR_3S_ESCO)
570 hdev->esco_type |= (ESCO_2EV5 | ESCO_3EV5);
571 }
572
573 static void hci_cc_read_local_ext_features(struct hci_dev *hdev,
574 struct sk_buff *skb)
575 {
576 struct hci_rp_read_local_ext_features *rp = (void *) skb->data;
577
578 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
579
580 if (rp->status)
581 return;
582
583 if (hdev->max_page < rp->max_page)
584 hdev->max_page = rp->max_page;
585
586 if (rp->page < HCI_MAX_PAGES)
587 memcpy(hdev->features[rp->page], rp->features, 8);
588 }
589
590 static void hci_cc_read_flow_control_mode(struct hci_dev *hdev,
591 struct sk_buff *skb)
592 {
593 struct hci_rp_read_flow_control_mode *rp = (void *) skb->data;
594
595 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
596
597 if (rp->status)
598 return;
599
600 hdev->flow_ctl_mode = rp->mode;
601 }
602
603 static void hci_cc_read_buffer_size(struct hci_dev *hdev, struct sk_buff *skb)
604 {
605 struct hci_rp_read_buffer_size *rp = (void *) skb->data;
606
607 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
608
609 if (rp->status)
610 return;
611
612 hdev->acl_mtu = __le16_to_cpu(rp->acl_mtu);
613 hdev->sco_mtu = rp->sco_mtu;
614 hdev->acl_pkts = __le16_to_cpu(rp->acl_max_pkt);
615 hdev->sco_pkts = __le16_to_cpu(rp->sco_max_pkt);
616
617 if (test_bit(HCI_QUIRK_FIXUP_BUFFER_SIZE, &hdev->quirks)) {
618 hdev->sco_mtu = 64;
619 hdev->sco_pkts = 8;
620 }
621
622 hdev->acl_cnt = hdev->acl_pkts;
623 hdev->sco_cnt = hdev->sco_pkts;
624
625 BT_DBG("%s acl mtu %d:%d sco mtu %d:%d", hdev->name, hdev->acl_mtu,
626 hdev->acl_pkts, hdev->sco_mtu, hdev->sco_pkts);
627 }
628
629 static void hci_cc_read_bd_addr(struct hci_dev *hdev, struct sk_buff *skb)
630 {
631 struct hci_rp_read_bd_addr *rp = (void *) skb->data;
632
633 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
634
635 if (rp->status)
636 return;
637
638 if (test_bit(HCI_INIT, &hdev->flags))
639 bacpy(&hdev->bdaddr, &rp->bdaddr);
640
641 if (test_bit(HCI_SETUP, &hdev->dev_flags))
642 bacpy(&hdev->setup_addr, &rp->bdaddr);
643 }
644
645 static void hci_cc_read_page_scan_activity(struct hci_dev *hdev,
646 struct sk_buff *skb)
647 {
648 struct hci_rp_read_page_scan_activity *rp = (void *) skb->data;
649
650 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
651
652 if (rp->status)
653 return;
654
655 if (test_bit(HCI_INIT, &hdev->flags)) {
656 hdev->page_scan_interval = __le16_to_cpu(rp->interval);
657 hdev->page_scan_window = __le16_to_cpu(rp->window);
658 }
659 }
660
661 static void hci_cc_write_page_scan_activity(struct hci_dev *hdev,
662 struct sk_buff *skb)
663 {
664 u8 status = *((u8 *) skb->data);
665 struct hci_cp_write_page_scan_activity *sent;
666
667 BT_DBG("%s status 0x%2.2x", hdev->name, status);
668
669 if (status)
670 return;
671
672 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY);
673 if (!sent)
674 return;
675
676 hdev->page_scan_interval = __le16_to_cpu(sent->interval);
677 hdev->page_scan_window = __le16_to_cpu(sent->window);
678 }
679
680 static void hci_cc_read_page_scan_type(struct hci_dev *hdev,
681 struct sk_buff *skb)
682 {
683 struct hci_rp_read_page_scan_type *rp = (void *) skb->data;
684
685 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
686
687 if (rp->status)
688 return;
689
690 if (test_bit(HCI_INIT, &hdev->flags))
691 hdev->page_scan_type = rp->type;
692 }
693
694 static void hci_cc_write_page_scan_type(struct hci_dev *hdev,
695 struct sk_buff *skb)
696 {
697 u8 status = *((u8 *) skb->data);
698 u8 *type;
699
700 BT_DBG("%s status 0x%2.2x", hdev->name, status);
701
702 if (status)
703 return;
704
705 type = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_TYPE);
706 if (type)
707 hdev->page_scan_type = *type;
708 }
709
710 static void hci_cc_read_data_block_size(struct hci_dev *hdev,
711 struct sk_buff *skb)
712 {
713 struct hci_rp_read_data_block_size *rp = (void *) skb->data;
714
715 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
716
717 if (rp->status)
718 return;
719
720 hdev->block_mtu = __le16_to_cpu(rp->max_acl_len);
721 hdev->block_len = __le16_to_cpu(rp->block_len);
722 hdev->num_blocks = __le16_to_cpu(rp->num_blocks);
723
724 hdev->block_cnt = hdev->num_blocks;
725
726 BT_DBG("%s blk mtu %d cnt %d len %d", hdev->name, hdev->block_mtu,
727 hdev->block_cnt, hdev->block_len);
728 }
729
730 static void hci_cc_read_clock(struct hci_dev *hdev, struct sk_buff *skb)
731 {
732 struct hci_rp_read_clock *rp = (void *) skb->data;
733 struct hci_cp_read_clock *cp;
734 struct hci_conn *conn;
735
736 BT_DBG("%s", hdev->name);
737
738 if (skb->len < sizeof(*rp))
739 return;
740
741 if (rp->status)
742 return;
743
744 hci_dev_lock(hdev);
745
746 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_CLOCK);
747 if (!cp)
748 goto unlock;
749
750 if (cp->which == 0x00) {
751 hdev->clock = le32_to_cpu(rp->clock);
752 goto unlock;
753 }
754
755 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
756 if (conn) {
757 conn->clock = le32_to_cpu(rp->clock);
758 conn->clock_accuracy = le16_to_cpu(rp->accuracy);
759 }
760
761 unlock:
762 hci_dev_unlock(hdev);
763 }
764
765 static void hci_cc_read_local_amp_info(struct hci_dev *hdev,
766 struct sk_buff *skb)
767 {
768 struct hci_rp_read_local_amp_info *rp = (void *) skb->data;
769
770 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
771
772 if (rp->status)
773 goto a2mp_rsp;
774
775 hdev->amp_status = rp->amp_status;
776 hdev->amp_total_bw = __le32_to_cpu(rp->total_bw);
777 hdev->amp_max_bw = __le32_to_cpu(rp->max_bw);
778 hdev->amp_min_latency = __le32_to_cpu(rp->min_latency);
779 hdev->amp_max_pdu = __le32_to_cpu(rp->max_pdu);
780 hdev->amp_type = rp->amp_type;
781 hdev->amp_pal_cap = __le16_to_cpu(rp->pal_cap);
782 hdev->amp_assoc_size = __le16_to_cpu(rp->max_assoc_size);
783 hdev->amp_be_flush_to = __le32_to_cpu(rp->be_flush_to);
784 hdev->amp_max_flush_to = __le32_to_cpu(rp->max_flush_to);
785
786 a2mp_rsp:
787 a2mp_send_getinfo_rsp(hdev);
788 }
789
790 static void hci_cc_read_local_amp_assoc(struct hci_dev *hdev,
791 struct sk_buff *skb)
792 {
793 struct hci_rp_read_local_amp_assoc *rp = (void *) skb->data;
794 struct amp_assoc *assoc = &hdev->loc_assoc;
795 size_t rem_len, frag_len;
796
797 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
798
799 if (rp->status)
800 goto a2mp_rsp;
801
802 frag_len = skb->len - sizeof(*rp);
803 rem_len = __le16_to_cpu(rp->rem_len);
804
805 if (rem_len > frag_len) {
806 BT_DBG("frag_len %zu rem_len %zu", frag_len, rem_len);
807
808 memcpy(assoc->data + assoc->offset, rp->frag, frag_len);
809 assoc->offset += frag_len;
810
811 /* Read other fragments */
812 amp_read_loc_assoc_frag(hdev, rp->phy_handle);
813
814 return;
815 }
816
817 memcpy(assoc->data + assoc->offset, rp->frag, rem_len);
818 assoc->len = assoc->offset + rem_len;
819 assoc->offset = 0;
820
821 a2mp_rsp:
822 /* Send A2MP Rsp when all fragments are received */
823 a2mp_send_getampassoc_rsp(hdev, rp->status);
824 a2mp_send_create_phy_link_req(hdev, rp->status);
825 }
826
827 static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev,
828 struct sk_buff *skb)
829 {
830 struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data;
831
832 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
833
834 if (rp->status)
835 return;
836
837 hdev->inq_tx_power = rp->tx_power;
838 }
839
840 static void hci_cc_pin_code_reply(struct hci_dev *hdev, struct sk_buff *skb)
841 {
842 struct hci_rp_pin_code_reply *rp = (void *) skb->data;
843 struct hci_cp_pin_code_reply *cp;
844 struct hci_conn *conn;
845
846 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
847
848 hci_dev_lock(hdev);
849
850 if (test_bit(HCI_MGMT, &hdev->dev_flags))
851 mgmt_pin_code_reply_complete(hdev, &rp->bdaddr, rp->status);
852
853 if (rp->status)
854 goto unlock;
855
856 cp = hci_sent_cmd_data(hdev, HCI_OP_PIN_CODE_REPLY);
857 if (!cp)
858 goto unlock;
859
860 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
861 if (conn)
862 conn->pin_length = cp->pin_len;
863
864 unlock:
865 hci_dev_unlock(hdev);
866 }
867
868 static void hci_cc_pin_code_neg_reply(struct hci_dev *hdev, struct sk_buff *skb)
869 {
870 struct hci_rp_pin_code_neg_reply *rp = (void *) skb->data;
871
872 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
873
874 hci_dev_lock(hdev);
875
876 if (test_bit(HCI_MGMT, &hdev->dev_flags))
877 mgmt_pin_code_neg_reply_complete(hdev, &rp->bdaddr,
878 rp->status);
879
880 hci_dev_unlock(hdev);
881 }
882
883 static void hci_cc_le_read_buffer_size(struct hci_dev *hdev,
884 struct sk_buff *skb)
885 {
886 struct hci_rp_le_read_buffer_size *rp = (void *) skb->data;
887
888 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
889
890 if (rp->status)
891 return;
892
893 hdev->le_mtu = __le16_to_cpu(rp->le_mtu);
894 hdev->le_pkts = rp->le_max_pkt;
895
896 hdev->le_cnt = hdev->le_pkts;
897
898 BT_DBG("%s le mtu %d:%d", hdev->name, hdev->le_mtu, hdev->le_pkts);
899 }
900
901 static void hci_cc_le_read_local_features(struct hci_dev *hdev,
902 struct sk_buff *skb)
903 {
904 struct hci_rp_le_read_local_features *rp = (void *) skb->data;
905
906 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
907
908 if (rp->status)
909 return;
910
911 memcpy(hdev->le_features, rp->features, 8);
912 }
913
914 static void hci_cc_le_read_adv_tx_power(struct hci_dev *hdev,
915 struct sk_buff *skb)
916 {
917 struct hci_rp_le_read_adv_tx_power *rp = (void *) skb->data;
918
919 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
920
921 if (rp->status)
922 return;
923
924 hdev->adv_tx_power = rp->tx_power;
925 }
926
927 static void hci_cc_user_confirm_reply(struct hci_dev *hdev, struct sk_buff *skb)
928 {
929 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
930
931 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
932
933 hci_dev_lock(hdev);
934
935 if (test_bit(HCI_MGMT, &hdev->dev_flags))
936 mgmt_user_confirm_reply_complete(hdev, &rp->bdaddr, ACL_LINK, 0,
937 rp->status);
938
939 hci_dev_unlock(hdev);
940 }
941
942 static void hci_cc_user_confirm_neg_reply(struct hci_dev *hdev,
943 struct sk_buff *skb)
944 {
945 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
946
947 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
948
949 hci_dev_lock(hdev);
950
951 if (test_bit(HCI_MGMT, &hdev->dev_flags))
952 mgmt_user_confirm_neg_reply_complete(hdev, &rp->bdaddr,
953 ACL_LINK, 0, rp->status);
954
955 hci_dev_unlock(hdev);
956 }
957
958 static void hci_cc_user_passkey_reply(struct hci_dev *hdev, struct sk_buff *skb)
959 {
960 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
961
962 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
963
964 hci_dev_lock(hdev);
965
966 if (test_bit(HCI_MGMT, &hdev->dev_flags))
967 mgmt_user_passkey_reply_complete(hdev, &rp->bdaddr, ACL_LINK,
968 0, rp->status);
969
970 hci_dev_unlock(hdev);
971 }
972
973 static void hci_cc_user_passkey_neg_reply(struct hci_dev *hdev,
974 struct sk_buff *skb)
975 {
976 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
977
978 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
979
980 hci_dev_lock(hdev);
981
982 if (test_bit(HCI_MGMT, &hdev->dev_flags))
983 mgmt_user_passkey_neg_reply_complete(hdev, &rp->bdaddr,
984 ACL_LINK, 0, rp->status);
985
986 hci_dev_unlock(hdev);
987 }
988
989 static void hci_cc_read_local_oob_data(struct hci_dev *hdev,
990 struct sk_buff *skb)
991 {
992 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
993
994 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
995
996 hci_dev_lock(hdev);
997 mgmt_read_local_oob_data_complete(hdev, rp->hash, rp->randomizer,
998 NULL, NULL, rp->status);
999 hci_dev_unlock(hdev);
1000 }
1001
1002 static void hci_cc_read_local_oob_ext_data(struct hci_dev *hdev,
1003 struct sk_buff *skb)
1004 {
1005 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
1006
1007 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1008
1009 hci_dev_lock(hdev);
1010 mgmt_read_local_oob_data_complete(hdev, rp->hash192, rp->randomizer192,
1011 rp->hash256, rp->randomizer256,
1012 rp->status);
1013 hci_dev_unlock(hdev);
1014 }
1015
1016
1017 static void hci_cc_le_set_random_addr(struct hci_dev *hdev, struct sk_buff *skb)
1018 {
1019 __u8 status = *((__u8 *) skb->data);
1020 bdaddr_t *sent;
1021
1022 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1023
1024 if (status)
1025 return;
1026
1027 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_RANDOM_ADDR);
1028 if (!sent)
1029 return;
1030
1031 hci_dev_lock(hdev);
1032
1033 bacpy(&hdev->random_addr, sent);
1034
1035 hci_dev_unlock(hdev);
1036 }
1037
1038 static void hci_cc_le_set_adv_enable(struct hci_dev *hdev, struct sk_buff *skb)
1039 {
1040 __u8 *sent, status = *((__u8 *) skb->data);
1041
1042 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1043
1044 if (status)
1045 return;
1046
1047 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_ENABLE);
1048 if (!sent)
1049 return;
1050
1051 hci_dev_lock(hdev);
1052
1053 /* If we're doing connection initiation as peripheral. Set a
1054 * timeout in case something goes wrong.
1055 */
1056 if (*sent) {
1057 struct hci_conn *conn;
1058
1059 set_bit(HCI_LE_ADV, &hdev->dev_flags);
1060
1061 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
1062 if (conn)
1063 queue_delayed_work(hdev->workqueue,
1064 &conn->le_conn_timeout,
1065 conn->conn_timeout);
1066 } else {
1067 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
1068 }
1069
1070 hci_dev_unlock(hdev);
1071 }
1072
1073 static void hci_cc_le_set_scan_param(struct hci_dev *hdev, struct sk_buff *skb)
1074 {
1075 struct hci_cp_le_set_scan_param *cp;
1076 __u8 status = *((__u8 *) skb->data);
1077
1078 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1079
1080 if (status)
1081 return;
1082
1083 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_PARAM);
1084 if (!cp)
1085 return;
1086
1087 hci_dev_lock(hdev);
1088
1089 hdev->le_scan_type = cp->type;
1090
1091 hci_dev_unlock(hdev);
1092 }
1093
1094 static bool has_pending_adv_report(struct hci_dev *hdev)
1095 {
1096 struct discovery_state *d = &hdev->discovery;
1097
1098 return bacmp(&d->last_adv_addr, BDADDR_ANY);
1099 }
1100
1101 static void clear_pending_adv_report(struct hci_dev *hdev)
1102 {
1103 struct discovery_state *d = &hdev->discovery;
1104
1105 bacpy(&d->last_adv_addr, BDADDR_ANY);
1106 d->last_adv_data_len = 0;
1107 }
1108
1109 static void store_pending_adv_report(struct hci_dev *hdev, bdaddr_t *bdaddr,
1110 u8 bdaddr_type, s8 rssi, u32 flags,
1111 u8 *data, u8 len)
1112 {
1113 struct discovery_state *d = &hdev->discovery;
1114
1115 bacpy(&d->last_adv_addr, bdaddr);
1116 d->last_adv_addr_type = bdaddr_type;
1117 d->last_adv_rssi = rssi;
1118 d->last_adv_flags = flags;
1119 memcpy(d->last_adv_data, data, len);
1120 d->last_adv_data_len = len;
1121 }
1122
1123 static void hci_cc_le_set_scan_enable(struct hci_dev *hdev,
1124 struct sk_buff *skb)
1125 {
1126 struct hci_cp_le_set_scan_enable *cp;
1127 __u8 status = *((__u8 *) skb->data);
1128
1129 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1130
1131 if (status)
1132 return;
1133
1134 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_ENABLE);
1135 if (!cp)
1136 return;
1137
1138 switch (cp->enable) {
1139 case LE_SCAN_ENABLE:
1140 set_bit(HCI_LE_SCAN, &hdev->dev_flags);
1141 if (hdev->le_scan_type == LE_SCAN_ACTIVE)
1142 clear_pending_adv_report(hdev);
1143 break;
1144
1145 case LE_SCAN_DISABLE:
1146 /* We do this here instead of when setting DISCOVERY_STOPPED
1147 * since the latter would potentially require waiting for
1148 * inquiry to stop too.
1149 */
1150 if (has_pending_adv_report(hdev)) {
1151 struct discovery_state *d = &hdev->discovery;
1152
1153 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
1154 d->last_adv_addr_type, NULL,
1155 d->last_adv_rssi, d->last_adv_flags,
1156 d->last_adv_data,
1157 d->last_adv_data_len, NULL, 0);
1158 }
1159
1160 /* Cancel this timer so that we don't try to disable scanning
1161 * when it's already disabled.
1162 */
1163 cancel_delayed_work(&hdev->le_scan_disable);
1164
1165 clear_bit(HCI_LE_SCAN, &hdev->dev_flags);
1166
1167 /* The HCI_LE_SCAN_INTERRUPTED flag indicates that we
1168 * interrupted scanning due to a connect request. Mark
1169 * therefore discovery as stopped. If this was not
1170 * because of a connect request advertising might have
1171 * been disabled because of active scanning, so
1172 * re-enable it again if necessary.
1173 */
1174 if (test_and_clear_bit(HCI_LE_SCAN_INTERRUPTED,
1175 &hdev->dev_flags))
1176 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1177 else if (!test_bit(HCI_LE_ADV, &hdev->dev_flags) &&
1178 hdev->discovery.state == DISCOVERY_FINDING)
1179 mgmt_reenable_advertising(hdev);
1180
1181 break;
1182
1183 default:
1184 BT_ERR("Used reserved LE_Scan_Enable param %d", cp->enable);
1185 break;
1186 }
1187 }
1188
1189 static void hci_cc_le_read_white_list_size(struct hci_dev *hdev,
1190 struct sk_buff *skb)
1191 {
1192 struct hci_rp_le_read_white_list_size *rp = (void *) skb->data;
1193
1194 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1195
1196 if (rp->status)
1197 return;
1198
1199 hdev->le_white_list_size = rp->size;
1200 }
1201
1202 static void hci_cc_le_clear_white_list(struct hci_dev *hdev,
1203 struct sk_buff *skb)
1204 {
1205 __u8 status = *((__u8 *) skb->data);
1206
1207 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1208
1209 if (status)
1210 return;
1211
1212 hci_bdaddr_list_clear(&hdev->le_white_list);
1213 }
1214
1215 static void hci_cc_le_add_to_white_list(struct hci_dev *hdev,
1216 struct sk_buff *skb)
1217 {
1218 struct hci_cp_le_add_to_white_list *sent;
1219 __u8 status = *((__u8 *) skb->data);
1220
1221 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1222
1223 if (status)
1224 return;
1225
1226 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_WHITE_LIST);
1227 if (!sent)
1228 return;
1229
1230 hci_bdaddr_list_add(&hdev->le_white_list, &sent->bdaddr,
1231 sent->bdaddr_type);
1232 }
1233
1234 static void hci_cc_le_del_from_white_list(struct hci_dev *hdev,
1235 struct sk_buff *skb)
1236 {
1237 struct hci_cp_le_del_from_white_list *sent;
1238 __u8 status = *((__u8 *) skb->data);
1239
1240 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1241
1242 if (status)
1243 return;
1244
1245 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_WHITE_LIST);
1246 if (!sent)
1247 return;
1248
1249 hci_bdaddr_list_del(&hdev->le_white_list, &sent->bdaddr,
1250 sent->bdaddr_type);
1251 }
1252
1253 static void hci_cc_le_read_supported_states(struct hci_dev *hdev,
1254 struct sk_buff *skb)
1255 {
1256 struct hci_rp_le_read_supported_states *rp = (void *) skb->data;
1257
1258 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1259
1260 if (rp->status)
1261 return;
1262
1263 memcpy(hdev->le_states, rp->le_states, 8);
1264 }
1265
1266 static void hci_cc_write_le_host_supported(struct hci_dev *hdev,
1267 struct sk_buff *skb)
1268 {
1269 struct hci_cp_write_le_host_supported *sent;
1270 __u8 status = *((__u8 *) skb->data);
1271
1272 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1273
1274 if (status)
1275 return;
1276
1277 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED);
1278 if (!sent)
1279 return;
1280
1281 if (sent->le) {
1282 hdev->features[1][0] |= LMP_HOST_LE;
1283 set_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1284 } else {
1285 hdev->features[1][0] &= ~LMP_HOST_LE;
1286 clear_bit(HCI_LE_ENABLED, &hdev->dev_flags);
1287 clear_bit(HCI_ADVERTISING, &hdev->dev_flags);
1288 }
1289
1290 if (sent->simul)
1291 hdev->features[1][0] |= LMP_HOST_LE_BREDR;
1292 else
1293 hdev->features[1][0] &= ~LMP_HOST_LE_BREDR;
1294 }
1295
1296 static void hci_cc_set_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1297 {
1298 struct hci_cp_le_set_adv_param *cp;
1299 u8 status = *((u8 *) skb->data);
1300
1301 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1302
1303 if (status)
1304 return;
1305
1306 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_PARAM);
1307 if (!cp)
1308 return;
1309
1310 hci_dev_lock(hdev);
1311 hdev->adv_addr_type = cp->own_address_type;
1312 hci_dev_unlock(hdev);
1313 }
1314
1315 static void hci_cc_write_remote_amp_assoc(struct hci_dev *hdev,
1316 struct sk_buff *skb)
1317 {
1318 struct hci_rp_write_remote_amp_assoc *rp = (void *) skb->data;
1319
1320 BT_DBG("%s status 0x%2.2x phy_handle 0x%2.2x",
1321 hdev->name, rp->status, rp->phy_handle);
1322
1323 if (rp->status)
1324 return;
1325
1326 amp_write_rem_assoc_continue(hdev, rp->phy_handle);
1327 }
1328
1329 static void hci_cc_read_rssi(struct hci_dev *hdev, struct sk_buff *skb)
1330 {
1331 struct hci_rp_read_rssi *rp = (void *) skb->data;
1332 struct hci_conn *conn;
1333
1334 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1335
1336 if (rp->status)
1337 return;
1338
1339 hci_dev_lock(hdev);
1340
1341 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1342 if (conn)
1343 conn->rssi = rp->rssi;
1344
1345 hci_dev_unlock(hdev);
1346 }
1347
1348 static void hci_cc_read_tx_power(struct hci_dev *hdev, struct sk_buff *skb)
1349 {
1350 struct hci_cp_read_tx_power *sent;
1351 struct hci_rp_read_tx_power *rp = (void *) skb->data;
1352 struct hci_conn *conn;
1353
1354 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1355
1356 if (rp->status)
1357 return;
1358
1359 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_TX_POWER);
1360 if (!sent)
1361 return;
1362
1363 hci_dev_lock(hdev);
1364
1365 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1366 if (!conn)
1367 goto unlock;
1368
1369 switch (sent->type) {
1370 case 0x00:
1371 conn->tx_power = rp->tx_power;
1372 break;
1373 case 0x01:
1374 conn->max_tx_power = rp->tx_power;
1375 break;
1376 }
1377
1378 unlock:
1379 hci_dev_unlock(hdev);
1380 }
1381
1382 static void hci_cs_inquiry(struct hci_dev *hdev, __u8 status)
1383 {
1384 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1385
1386 if (status) {
1387 hci_conn_check_pending(hdev);
1388 return;
1389 }
1390
1391 set_bit(HCI_INQUIRY, &hdev->flags);
1392 }
1393
1394 static void hci_cs_create_conn(struct hci_dev *hdev, __u8 status)
1395 {
1396 struct hci_cp_create_conn *cp;
1397 struct hci_conn *conn;
1398
1399 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1400
1401 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_CONN);
1402 if (!cp)
1403 return;
1404
1405 hci_dev_lock(hdev);
1406
1407 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1408
1409 BT_DBG("%s bdaddr %pMR hcon %p", hdev->name, &cp->bdaddr, conn);
1410
1411 if (status) {
1412 if (conn && conn->state == BT_CONNECT) {
1413 if (status != 0x0c || conn->attempt > 2) {
1414 conn->state = BT_CLOSED;
1415 hci_proto_connect_cfm(conn, status);
1416 hci_conn_del(conn);
1417 } else
1418 conn->state = BT_CONNECT2;
1419 }
1420 } else {
1421 if (!conn) {
1422 conn = hci_conn_add(hdev, ACL_LINK, &cp->bdaddr,
1423 HCI_ROLE_MASTER);
1424 if (!conn)
1425 BT_ERR("No memory for new connection");
1426 }
1427 }
1428
1429 hci_dev_unlock(hdev);
1430 }
1431
1432 static void hci_cs_add_sco(struct hci_dev *hdev, __u8 status)
1433 {
1434 struct hci_cp_add_sco *cp;
1435 struct hci_conn *acl, *sco;
1436 __u16 handle;
1437
1438 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1439
1440 if (!status)
1441 return;
1442
1443 cp = hci_sent_cmd_data(hdev, HCI_OP_ADD_SCO);
1444 if (!cp)
1445 return;
1446
1447 handle = __le16_to_cpu(cp->handle);
1448
1449 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1450
1451 hci_dev_lock(hdev);
1452
1453 acl = hci_conn_hash_lookup_handle(hdev, handle);
1454 if (acl) {
1455 sco = acl->link;
1456 if (sco) {
1457 sco->state = BT_CLOSED;
1458
1459 hci_proto_connect_cfm(sco, status);
1460 hci_conn_del(sco);
1461 }
1462 }
1463
1464 hci_dev_unlock(hdev);
1465 }
1466
1467 static void hci_cs_auth_requested(struct hci_dev *hdev, __u8 status)
1468 {
1469 struct hci_cp_auth_requested *cp;
1470 struct hci_conn *conn;
1471
1472 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1473
1474 if (!status)
1475 return;
1476
1477 cp = hci_sent_cmd_data(hdev, HCI_OP_AUTH_REQUESTED);
1478 if (!cp)
1479 return;
1480
1481 hci_dev_lock(hdev);
1482
1483 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1484 if (conn) {
1485 if (conn->state == BT_CONFIG) {
1486 hci_proto_connect_cfm(conn, status);
1487 hci_conn_drop(conn);
1488 }
1489 }
1490
1491 hci_dev_unlock(hdev);
1492 }
1493
1494 static void hci_cs_set_conn_encrypt(struct hci_dev *hdev, __u8 status)
1495 {
1496 struct hci_cp_set_conn_encrypt *cp;
1497 struct hci_conn *conn;
1498
1499 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1500
1501 if (!status)
1502 return;
1503
1504 cp = hci_sent_cmd_data(hdev, HCI_OP_SET_CONN_ENCRYPT);
1505 if (!cp)
1506 return;
1507
1508 hci_dev_lock(hdev);
1509
1510 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1511 if (conn) {
1512 if (conn->state == BT_CONFIG) {
1513 hci_proto_connect_cfm(conn, status);
1514 hci_conn_drop(conn);
1515 }
1516 }
1517
1518 hci_dev_unlock(hdev);
1519 }
1520
1521 static int hci_outgoing_auth_needed(struct hci_dev *hdev,
1522 struct hci_conn *conn)
1523 {
1524 if (conn->state != BT_CONFIG || !conn->out)
1525 return 0;
1526
1527 if (conn->pending_sec_level == BT_SECURITY_SDP)
1528 return 0;
1529
1530 /* Only request authentication for SSP connections or non-SSP
1531 * devices with sec_level MEDIUM or HIGH or if MITM protection
1532 * is requested.
1533 */
1534 if (!hci_conn_ssp_enabled(conn) && !(conn->auth_type & 0x01) &&
1535 conn->pending_sec_level != BT_SECURITY_FIPS &&
1536 conn->pending_sec_level != BT_SECURITY_HIGH &&
1537 conn->pending_sec_level != BT_SECURITY_MEDIUM)
1538 return 0;
1539
1540 return 1;
1541 }
1542
1543 static int hci_resolve_name(struct hci_dev *hdev,
1544 struct inquiry_entry *e)
1545 {
1546 struct hci_cp_remote_name_req cp;
1547
1548 memset(&cp, 0, sizeof(cp));
1549
1550 bacpy(&cp.bdaddr, &e->data.bdaddr);
1551 cp.pscan_rep_mode = e->data.pscan_rep_mode;
1552 cp.pscan_mode = e->data.pscan_mode;
1553 cp.clock_offset = e->data.clock_offset;
1554
1555 return hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
1556 }
1557
1558 static bool hci_resolve_next_name(struct hci_dev *hdev)
1559 {
1560 struct discovery_state *discov = &hdev->discovery;
1561 struct inquiry_entry *e;
1562
1563 if (list_empty(&discov->resolve))
1564 return false;
1565
1566 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1567 if (!e)
1568 return false;
1569
1570 if (hci_resolve_name(hdev, e) == 0) {
1571 e->name_state = NAME_PENDING;
1572 return true;
1573 }
1574
1575 return false;
1576 }
1577
1578 static void hci_check_pending_name(struct hci_dev *hdev, struct hci_conn *conn,
1579 bdaddr_t *bdaddr, u8 *name, u8 name_len)
1580 {
1581 struct discovery_state *discov = &hdev->discovery;
1582 struct inquiry_entry *e;
1583
1584 if (conn && !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
1585 mgmt_device_connected(hdev, conn, 0, name, name_len);
1586
1587 if (discov->state == DISCOVERY_STOPPED)
1588 return;
1589
1590 if (discov->state == DISCOVERY_STOPPING)
1591 goto discov_complete;
1592
1593 if (discov->state != DISCOVERY_RESOLVING)
1594 return;
1595
1596 e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
1597 /* If the device was not found in a list of found devices names of which
1598 * are pending. there is no need to continue resolving a next name as it
1599 * will be done upon receiving another Remote Name Request Complete
1600 * Event */
1601 if (!e)
1602 return;
1603
1604 list_del(&e->list);
1605 if (name) {
1606 e->name_state = NAME_KNOWN;
1607 mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00,
1608 e->data.rssi, name, name_len);
1609 } else {
1610 e->name_state = NAME_NOT_KNOWN;
1611 }
1612
1613 if (hci_resolve_next_name(hdev))
1614 return;
1615
1616 discov_complete:
1617 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1618 }
1619
1620 static void hci_cs_remote_name_req(struct hci_dev *hdev, __u8 status)
1621 {
1622 struct hci_cp_remote_name_req *cp;
1623 struct hci_conn *conn;
1624
1625 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1626
1627 /* If successful wait for the name req complete event before
1628 * checking for the need to do authentication */
1629 if (!status)
1630 return;
1631
1632 cp = hci_sent_cmd_data(hdev, HCI_OP_REMOTE_NAME_REQ);
1633 if (!cp)
1634 return;
1635
1636 hci_dev_lock(hdev);
1637
1638 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1639
1640 if (test_bit(HCI_MGMT, &hdev->dev_flags))
1641 hci_check_pending_name(hdev, conn, &cp->bdaddr, NULL, 0);
1642
1643 if (!conn)
1644 goto unlock;
1645
1646 if (!hci_outgoing_auth_needed(hdev, conn))
1647 goto unlock;
1648
1649 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
1650 struct hci_cp_auth_requested auth_cp;
1651
1652 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
1653
1654 auth_cp.handle = __cpu_to_le16(conn->handle);
1655 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED,
1656 sizeof(auth_cp), &auth_cp);
1657 }
1658
1659 unlock:
1660 hci_dev_unlock(hdev);
1661 }
1662
1663 static void hci_cs_read_remote_features(struct hci_dev *hdev, __u8 status)
1664 {
1665 struct hci_cp_read_remote_features *cp;
1666 struct hci_conn *conn;
1667
1668 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1669
1670 if (!status)
1671 return;
1672
1673 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_FEATURES);
1674 if (!cp)
1675 return;
1676
1677 hci_dev_lock(hdev);
1678
1679 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1680 if (conn) {
1681 if (conn->state == BT_CONFIG) {
1682 hci_proto_connect_cfm(conn, status);
1683 hci_conn_drop(conn);
1684 }
1685 }
1686
1687 hci_dev_unlock(hdev);
1688 }
1689
1690 static void hci_cs_read_remote_ext_features(struct hci_dev *hdev, __u8 status)
1691 {
1692 struct hci_cp_read_remote_ext_features *cp;
1693 struct hci_conn *conn;
1694
1695 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1696
1697 if (!status)
1698 return;
1699
1700 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES);
1701 if (!cp)
1702 return;
1703
1704 hci_dev_lock(hdev);
1705
1706 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1707 if (conn) {
1708 if (conn->state == BT_CONFIG) {
1709 hci_proto_connect_cfm(conn, status);
1710 hci_conn_drop(conn);
1711 }
1712 }
1713
1714 hci_dev_unlock(hdev);
1715 }
1716
1717 static void hci_cs_setup_sync_conn(struct hci_dev *hdev, __u8 status)
1718 {
1719 struct hci_cp_setup_sync_conn *cp;
1720 struct hci_conn *acl, *sco;
1721 __u16 handle;
1722
1723 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1724
1725 if (!status)
1726 return;
1727
1728 cp = hci_sent_cmd_data(hdev, HCI_OP_SETUP_SYNC_CONN);
1729 if (!cp)
1730 return;
1731
1732 handle = __le16_to_cpu(cp->handle);
1733
1734 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1735
1736 hci_dev_lock(hdev);
1737
1738 acl = hci_conn_hash_lookup_handle(hdev, handle);
1739 if (acl) {
1740 sco = acl->link;
1741 if (sco) {
1742 sco->state = BT_CLOSED;
1743
1744 hci_proto_connect_cfm(sco, status);
1745 hci_conn_del(sco);
1746 }
1747 }
1748
1749 hci_dev_unlock(hdev);
1750 }
1751
1752 static void hci_cs_sniff_mode(struct hci_dev *hdev, __u8 status)
1753 {
1754 struct hci_cp_sniff_mode *cp;
1755 struct hci_conn *conn;
1756
1757 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1758
1759 if (!status)
1760 return;
1761
1762 cp = hci_sent_cmd_data(hdev, HCI_OP_SNIFF_MODE);
1763 if (!cp)
1764 return;
1765
1766 hci_dev_lock(hdev);
1767
1768 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1769 if (conn) {
1770 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1771
1772 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1773 hci_sco_setup(conn, status);
1774 }
1775
1776 hci_dev_unlock(hdev);
1777 }
1778
1779 static void hci_cs_exit_sniff_mode(struct hci_dev *hdev, __u8 status)
1780 {
1781 struct hci_cp_exit_sniff_mode *cp;
1782 struct hci_conn *conn;
1783
1784 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1785
1786 if (!status)
1787 return;
1788
1789 cp = hci_sent_cmd_data(hdev, HCI_OP_EXIT_SNIFF_MODE);
1790 if (!cp)
1791 return;
1792
1793 hci_dev_lock(hdev);
1794
1795 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1796 if (conn) {
1797 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
1798
1799 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
1800 hci_sco_setup(conn, status);
1801 }
1802
1803 hci_dev_unlock(hdev);
1804 }
1805
1806 static void hci_cs_disconnect(struct hci_dev *hdev, u8 status)
1807 {
1808 struct hci_cp_disconnect *cp;
1809 struct hci_conn *conn;
1810
1811 if (!status)
1812 return;
1813
1814 cp = hci_sent_cmd_data(hdev, HCI_OP_DISCONNECT);
1815 if (!cp)
1816 return;
1817
1818 hci_dev_lock(hdev);
1819
1820 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1821 if (conn)
1822 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
1823 conn->dst_type, status);
1824
1825 hci_dev_unlock(hdev);
1826 }
1827
1828 static void hci_cs_create_phylink(struct hci_dev *hdev, u8 status)
1829 {
1830 struct hci_cp_create_phy_link *cp;
1831
1832 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1833
1834 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_PHY_LINK);
1835 if (!cp)
1836 return;
1837
1838 hci_dev_lock(hdev);
1839
1840 if (status) {
1841 struct hci_conn *hcon;
1842
1843 hcon = hci_conn_hash_lookup_handle(hdev, cp->phy_handle);
1844 if (hcon)
1845 hci_conn_del(hcon);
1846 } else {
1847 amp_write_remote_assoc(hdev, cp->phy_handle);
1848 }
1849
1850 hci_dev_unlock(hdev);
1851 }
1852
1853 static void hci_cs_accept_phylink(struct hci_dev *hdev, u8 status)
1854 {
1855 struct hci_cp_accept_phy_link *cp;
1856
1857 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1858
1859 if (status)
1860 return;
1861
1862 cp = hci_sent_cmd_data(hdev, HCI_OP_ACCEPT_PHY_LINK);
1863 if (!cp)
1864 return;
1865
1866 amp_write_remote_assoc(hdev, cp->phy_handle);
1867 }
1868
1869 static void hci_cs_le_create_conn(struct hci_dev *hdev, u8 status)
1870 {
1871 struct hci_cp_le_create_conn *cp;
1872 struct hci_conn *conn;
1873
1874 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1875
1876 /* All connection failure handling is taken care of by the
1877 * hci_le_conn_failed function which is triggered by the HCI
1878 * request completion callbacks used for connecting.
1879 */
1880 if (status)
1881 return;
1882
1883 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_CREATE_CONN);
1884 if (!cp)
1885 return;
1886
1887 hci_dev_lock(hdev);
1888
1889 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->peer_addr);
1890 if (!conn)
1891 goto unlock;
1892
1893 /* Store the initiator and responder address information which
1894 * is needed for SMP. These values will not change during the
1895 * lifetime of the connection.
1896 */
1897 conn->init_addr_type = cp->own_address_type;
1898 if (cp->own_address_type == ADDR_LE_DEV_RANDOM)
1899 bacpy(&conn->init_addr, &hdev->random_addr);
1900 else
1901 bacpy(&conn->init_addr, &hdev->bdaddr);
1902
1903 conn->resp_addr_type = cp->peer_addr_type;
1904 bacpy(&conn->resp_addr, &cp->peer_addr);
1905
1906 /* We don't want the connection attempt to stick around
1907 * indefinitely since LE doesn't have a page timeout concept
1908 * like BR/EDR. Set a timer for any connection that doesn't use
1909 * the white list for connecting.
1910 */
1911 if (cp->filter_policy == HCI_LE_USE_PEER_ADDR)
1912 queue_delayed_work(conn->hdev->workqueue,
1913 &conn->le_conn_timeout,
1914 conn->conn_timeout);
1915
1916 unlock:
1917 hci_dev_unlock(hdev);
1918 }
1919
1920 static void hci_cs_le_start_enc(struct hci_dev *hdev, u8 status)
1921 {
1922 struct hci_cp_le_start_enc *cp;
1923 struct hci_conn *conn;
1924
1925 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1926
1927 if (!status)
1928 return;
1929
1930 hci_dev_lock(hdev);
1931
1932 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_START_ENC);
1933 if (!cp)
1934 goto unlock;
1935
1936 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1937 if (!conn)
1938 goto unlock;
1939
1940 if (conn->state != BT_CONNECTED)
1941 goto unlock;
1942
1943 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
1944 hci_conn_drop(conn);
1945
1946 unlock:
1947 hci_dev_unlock(hdev);
1948 }
1949
1950 static void hci_inquiry_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
1951 {
1952 __u8 status = *((__u8 *) skb->data);
1953 struct discovery_state *discov = &hdev->discovery;
1954 struct inquiry_entry *e;
1955
1956 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1957
1958 hci_conn_check_pending(hdev);
1959
1960 if (!test_and_clear_bit(HCI_INQUIRY, &hdev->flags))
1961 return;
1962
1963 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
1964 wake_up_bit(&hdev->flags, HCI_INQUIRY);
1965
1966 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1967 return;
1968
1969 hci_dev_lock(hdev);
1970
1971 if (discov->state != DISCOVERY_FINDING)
1972 goto unlock;
1973
1974 if (list_empty(&discov->resolve)) {
1975 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1976 goto unlock;
1977 }
1978
1979 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1980 if (e && hci_resolve_name(hdev, e) == 0) {
1981 e->name_state = NAME_PENDING;
1982 hci_discovery_set_state(hdev, DISCOVERY_RESOLVING);
1983 } else {
1984 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1985 }
1986
1987 unlock:
1988 hci_dev_unlock(hdev);
1989 }
1990
1991 static void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff *skb)
1992 {
1993 struct inquiry_data data;
1994 struct inquiry_info *info = (void *) (skb->data + 1);
1995 int num_rsp = *((__u8 *) skb->data);
1996
1997 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
1998
1999 if (!num_rsp)
2000 return;
2001
2002 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
2003 return;
2004
2005 hci_dev_lock(hdev);
2006
2007 for (; num_rsp; num_rsp--, info++) {
2008 u32 flags;
2009
2010 bacpy(&data.bdaddr, &info->bdaddr);
2011 data.pscan_rep_mode = info->pscan_rep_mode;
2012 data.pscan_period_mode = info->pscan_period_mode;
2013 data.pscan_mode = info->pscan_mode;
2014 memcpy(data.dev_class, info->dev_class, 3);
2015 data.clock_offset = info->clock_offset;
2016 data.rssi = 0x00;
2017 data.ssp_mode = 0x00;
2018
2019 flags = hci_inquiry_cache_update(hdev, &data, false);
2020
2021 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
2022 info->dev_class, 0, flags, NULL, 0, NULL, 0);
2023 }
2024
2025 hci_dev_unlock(hdev);
2026 }
2027
2028 static void hci_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2029 {
2030 struct hci_ev_conn_complete *ev = (void *) skb->data;
2031 struct hci_conn *conn;
2032
2033 BT_DBG("%s", hdev->name);
2034
2035 hci_dev_lock(hdev);
2036
2037 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
2038 if (!conn) {
2039 if (ev->link_type != SCO_LINK)
2040 goto unlock;
2041
2042 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
2043 if (!conn)
2044 goto unlock;
2045
2046 conn->type = SCO_LINK;
2047 }
2048
2049 if (!ev->status) {
2050 conn->handle = __le16_to_cpu(ev->handle);
2051
2052 if (conn->type == ACL_LINK) {
2053 conn->state = BT_CONFIG;
2054 hci_conn_hold(conn);
2055
2056 if (!conn->out && !hci_conn_ssp_enabled(conn) &&
2057 !hci_find_link_key(hdev, &ev->bdaddr))
2058 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
2059 else
2060 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2061 } else
2062 conn->state = BT_CONNECTED;
2063
2064 hci_conn_add_sysfs(conn);
2065
2066 if (test_bit(HCI_AUTH, &hdev->flags))
2067 set_bit(HCI_CONN_AUTH, &conn->flags);
2068
2069 if (test_bit(HCI_ENCRYPT, &hdev->flags))
2070 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2071
2072 /* Get remote features */
2073 if (conn->type == ACL_LINK) {
2074 struct hci_cp_read_remote_features cp;
2075 cp.handle = ev->handle;
2076 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_FEATURES,
2077 sizeof(cp), &cp);
2078
2079 hci_update_page_scan(hdev, NULL);
2080 }
2081
2082 /* Set packet type for incoming connection */
2083 if (!conn->out && hdev->hci_ver < BLUETOOTH_VER_2_0) {
2084 struct hci_cp_change_conn_ptype cp;
2085 cp.handle = ev->handle;
2086 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2087 hci_send_cmd(hdev, HCI_OP_CHANGE_CONN_PTYPE, sizeof(cp),
2088 &cp);
2089 }
2090 } else {
2091 conn->state = BT_CLOSED;
2092 if (conn->type == ACL_LINK)
2093 mgmt_connect_failed(hdev, &conn->dst, conn->type,
2094 conn->dst_type, ev->status);
2095 }
2096
2097 if (conn->type == ACL_LINK)
2098 hci_sco_setup(conn, ev->status);
2099
2100 if (ev->status) {
2101 hci_proto_connect_cfm(conn, ev->status);
2102 hci_conn_del(conn);
2103 } else if (ev->link_type != ACL_LINK)
2104 hci_proto_connect_cfm(conn, ev->status);
2105
2106 unlock:
2107 hci_dev_unlock(hdev);
2108
2109 hci_conn_check_pending(hdev);
2110 }
2111
2112 static void hci_reject_conn(struct hci_dev *hdev, bdaddr_t *bdaddr)
2113 {
2114 struct hci_cp_reject_conn_req cp;
2115
2116 bacpy(&cp.bdaddr, bdaddr);
2117 cp.reason = HCI_ERROR_REJ_BAD_ADDR;
2118 hci_send_cmd(hdev, HCI_OP_REJECT_CONN_REQ, sizeof(cp), &cp);
2119 }
2120
2121 static void hci_conn_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
2122 {
2123 struct hci_ev_conn_request *ev = (void *) skb->data;
2124 int mask = hdev->link_mode;
2125 struct inquiry_entry *ie;
2126 struct hci_conn *conn;
2127 __u8 flags = 0;
2128
2129 BT_DBG("%s bdaddr %pMR type 0x%x", hdev->name, &ev->bdaddr,
2130 ev->link_type);
2131
2132 mask |= hci_proto_connect_ind(hdev, &ev->bdaddr, ev->link_type,
2133 &flags);
2134
2135 if (!(mask & HCI_LM_ACCEPT)) {
2136 hci_reject_conn(hdev, &ev->bdaddr);
2137 return;
2138 }
2139
2140 if (hci_bdaddr_list_lookup(&hdev->blacklist, &ev->bdaddr,
2141 BDADDR_BREDR)) {
2142 hci_reject_conn(hdev, &ev->bdaddr);
2143 return;
2144 }
2145
2146 if (!test_bit(HCI_CONNECTABLE, &hdev->dev_flags) &&
2147 !hci_bdaddr_list_lookup(&hdev->whitelist, &ev->bdaddr,
2148 BDADDR_BREDR)) {
2149 hci_reject_conn(hdev, &ev->bdaddr);
2150 return;
2151 }
2152
2153 /* Connection accepted */
2154
2155 hci_dev_lock(hdev);
2156
2157 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
2158 if (ie)
2159 memcpy(ie->data.dev_class, ev->dev_class, 3);
2160
2161 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type,
2162 &ev->bdaddr);
2163 if (!conn) {
2164 conn = hci_conn_add(hdev, ev->link_type, &ev->bdaddr,
2165 HCI_ROLE_SLAVE);
2166 if (!conn) {
2167 BT_ERR("No memory for new connection");
2168 hci_dev_unlock(hdev);
2169 return;
2170 }
2171 }
2172
2173 memcpy(conn->dev_class, ev->dev_class, 3);
2174
2175 hci_dev_unlock(hdev);
2176
2177 if (ev->link_type == ACL_LINK ||
2178 (!(flags & HCI_PROTO_DEFER) && !lmp_esco_capable(hdev))) {
2179 struct hci_cp_accept_conn_req cp;
2180 conn->state = BT_CONNECT;
2181
2182 bacpy(&cp.bdaddr, &ev->bdaddr);
2183
2184 if (lmp_rswitch_capable(hdev) && (mask & HCI_LM_MASTER))
2185 cp.role = 0x00; /* Become master */
2186 else
2187 cp.role = 0x01; /* Remain slave */
2188
2189 hci_send_cmd(hdev, HCI_OP_ACCEPT_CONN_REQ, sizeof(cp), &cp);
2190 } else if (!(flags & HCI_PROTO_DEFER)) {
2191 struct hci_cp_accept_sync_conn_req cp;
2192 conn->state = BT_CONNECT;
2193
2194 bacpy(&cp.bdaddr, &ev->bdaddr);
2195 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2196
2197 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
2198 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
2199 cp.max_latency = cpu_to_le16(0xffff);
2200 cp.content_format = cpu_to_le16(hdev->voice_setting);
2201 cp.retrans_effort = 0xff;
2202
2203 hci_send_cmd(hdev, HCI_OP_ACCEPT_SYNC_CONN_REQ, sizeof(cp),
2204 &cp);
2205 } else {
2206 conn->state = BT_CONNECT2;
2207 hci_proto_connect_cfm(conn, 0);
2208 }
2209 }
2210
2211 static u8 hci_to_mgmt_reason(u8 err)
2212 {
2213 switch (err) {
2214 case HCI_ERROR_CONNECTION_TIMEOUT:
2215 return MGMT_DEV_DISCONN_TIMEOUT;
2216 case HCI_ERROR_REMOTE_USER_TERM:
2217 case HCI_ERROR_REMOTE_LOW_RESOURCES:
2218 case HCI_ERROR_REMOTE_POWER_OFF:
2219 return MGMT_DEV_DISCONN_REMOTE;
2220 case HCI_ERROR_LOCAL_HOST_TERM:
2221 return MGMT_DEV_DISCONN_LOCAL_HOST;
2222 default:
2223 return MGMT_DEV_DISCONN_UNKNOWN;
2224 }
2225 }
2226
2227 static void hci_disconn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2228 {
2229 struct hci_ev_disconn_complete *ev = (void *) skb->data;
2230 u8 reason = hci_to_mgmt_reason(ev->reason);
2231 struct hci_conn_params *params;
2232 struct hci_conn *conn;
2233 bool mgmt_connected;
2234 u8 type;
2235
2236 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2237
2238 hci_dev_lock(hdev);
2239
2240 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2241 if (!conn)
2242 goto unlock;
2243
2244 if (ev->status) {
2245 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2246 conn->dst_type, ev->status);
2247 goto unlock;
2248 }
2249
2250 conn->state = BT_CLOSED;
2251
2252 mgmt_connected = test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags);
2253 mgmt_device_disconnected(hdev, &conn->dst, conn->type, conn->dst_type,
2254 reason, mgmt_connected);
2255
2256 if (conn->type == ACL_LINK) {
2257 if (test_bit(HCI_CONN_FLUSH_KEY, &conn->flags))
2258 hci_remove_link_key(hdev, &conn->dst);
2259
2260 hci_update_page_scan(hdev, NULL);
2261 }
2262
2263 params = hci_conn_params_lookup(hdev, &conn->dst, conn->dst_type);
2264 if (params) {
2265 switch (params->auto_connect) {
2266 case HCI_AUTO_CONN_LINK_LOSS:
2267 if (ev->reason != HCI_ERROR_CONNECTION_TIMEOUT)
2268 break;
2269 /* Fall through */
2270
2271 case HCI_AUTO_CONN_DIRECT:
2272 case HCI_AUTO_CONN_ALWAYS:
2273 list_del_init(&params->action);
2274 list_add(&params->action, &hdev->pend_le_conns);
2275 hci_update_background_scan(hdev);
2276 break;
2277
2278 default:
2279 break;
2280 }
2281 }
2282
2283 type = conn->type;
2284
2285 hci_proto_disconn_cfm(conn, ev->reason);
2286 hci_conn_del(conn);
2287
2288 /* Re-enable advertising if necessary, since it might
2289 * have been disabled by the connection. From the
2290 * HCI_LE_Set_Advertise_Enable command description in
2291 * the core specification (v4.0):
2292 * "The Controller shall continue advertising until the Host
2293 * issues an LE_Set_Advertise_Enable command with
2294 * Advertising_Enable set to 0x00 (Advertising is disabled)
2295 * or until a connection is created or until the Advertising
2296 * is timed out due to Directed Advertising."
2297 */
2298 if (type == LE_LINK)
2299 mgmt_reenable_advertising(hdev);
2300
2301 unlock:
2302 hci_dev_unlock(hdev);
2303 }
2304
2305 static void hci_auth_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2306 {
2307 struct hci_ev_auth_complete *ev = (void *) skb->data;
2308 struct hci_conn *conn;
2309
2310 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2311
2312 hci_dev_lock(hdev);
2313
2314 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2315 if (!conn)
2316 goto unlock;
2317
2318 if (!ev->status) {
2319 if (!hci_conn_ssp_enabled(conn) &&
2320 test_bit(HCI_CONN_REAUTH_PEND, &conn->flags)) {
2321 BT_INFO("re-auth of legacy device is not possible.");
2322 } else {
2323 set_bit(HCI_CONN_AUTH, &conn->flags);
2324 conn->sec_level = conn->pending_sec_level;
2325 }
2326 } else {
2327 mgmt_auth_failed(conn, ev->status);
2328 }
2329
2330 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2331 clear_bit(HCI_CONN_REAUTH_PEND, &conn->flags);
2332
2333 if (conn->state == BT_CONFIG) {
2334 if (!ev->status && hci_conn_ssp_enabled(conn)) {
2335 struct hci_cp_set_conn_encrypt cp;
2336 cp.handle = ev->handle;
2337 cp.encrypt = 0x01;
2338 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2339 &cp);
2340 } else {
2341 conn->state = BT_CONNECTED;
2342 hci_proto_connect_cfm(conn, ev->status);
2343 hci_conn_drop(conn);
2344 }
2345 } else {
2346 hci_auth_cfm(conn, ev->status);
2347
2348 hci_conn_hold(conn);
2349 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2350 hci_conn_drop(conn);
2351 }
2352
2353 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags)) {
2354 if (!ev->status) {
2355 struct hci_cp_set_conn_encrypt cp;
2356 cp.handle = ev->handle;
2357 cp.encrypt = 0x01;
2358 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2359 &cp);
2360 } else {
2361 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2362 hci_encrypt_cfm(conn, ev->status, 0x00);
2363 }
2364 }
2365
2366 unlock:
2367 hci_dev_unlock(hdev);
2368 }
2369
2370 static void hci_remote_name_evt(struct hci_dev *hdev, struct sk_buff *skb)
2371 {
2372 struct hci_ev_remote_name *ev = (void *) skb->data;
2373 struct hci_conn *conn;
2374
2375 BT_DBG("%s", hdev->name);
2376
2377 hci_conn_check_pending(hdev);
2378
2379 hci_dev_lock(hdev);
2380
2381 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2382
2383 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
2384 goto check_auth;
2385
2386 if (ev->status == 0)
2387 hci_check_pending_name(hdev, conn, &ev->bdaddr, ev->name,
2388 strnlen(ev->name, HCI_MAX_NAME_LENGTH));
2389 else
2390 hci_check_pending_name(hdev, conn, &ev->bdaddr, NULL, 0);
2391
2392 check_auth:
2393 if (!conn)
2394 goto unlock;
2395
2396 if (!hci_outgoing_auth_needed(hdev, conn))
2397 goto unlock;
2398
2399 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
2400 struct hci_cp_auth_requested cp;
2401
2402 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
2403
2404 cp.handle = __cpu_to_le16(conn->handle);
2405 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED, sizeof(cp), &cp);
2406 }
2407
2408 unlock:
2409 hci_dev_unlock(hdev);
2410 }
2411
2412 static void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2413 {
2414 struct hci_ev_encrypt_change *ev = (void *) skb->data;
2415 struct hci_conn *conn;
2416
2417 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2418
2419 hci_dev_lock(hdev);
2420
2421 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2422 if (!conn)
2423 goto unlock;
2424
2425 if (!ev->status) {
2426 if (ev->encrypt) {
2427 /* Encryption implies authentication */
2428 set_bit(HCI_CONN_AUTH, &conn->flags);
2429 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2430 conn->sec_level = conn->pending_sec_level;
2431
2432 /* P-256 authentication key implies FIPS */
2433 if (conn->key_type == HCI_LK_AUTH_COMBINATION_P256)
2434 set_bit(HCI_CONN_FIPS, &conn->flags);
2435
2436 if ((conn->type == ACL_LINK && ev->encrypt == 0x02) ||
2437 conn->type == LE_LINK)
2438 set_bit(HCI_CONN_AES_CCM, &conn->flags);
2439 } else {
2440 clear_bit(HCI_CONN_ENCRYPT, &conn->flags);
2441 clear_bit(HCI_CONN_AES_CCM, &conn->flags);
2442 }
2443 }
2444
2445 /* We should disregard the current RPA and generate a new one
2446 * whenever the encryption procedure fails.
2447 */
2448 if (ev->status && conn->type == LE_LINK)
2449 set_bit(HCI_RPA_EXPIRED, &hdev->dev_flags);
2450
2451 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2452
2453 if (ev->status && conn->state == BT_CONNECTED) {
2454 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2455 hci_conn_drop(conn);
2456 goto unlock;
2457 }
2458
2459 if (conn->state == BT_CONFIG) {
2460 if (!ev->status)
2461 conn->state = BT_CONNECTED;
2462
2463 /* In Secure Connections Only mode, do not allow any
2464 * connections that are not encrypted with AES-CCM
2465 * using a P-256 authenticated combination key.
2466 */
2467 if (test_bit(HCI_SC_ONLY, &hdev->dev_flags) &&
2468 (!test_bit(HCI_CONN_AES_CCM, &conn->flags) ||
2469 conn->key_type != HCI_LK_AUTH_COMBINATION_P256)) {
2470 hci_proto_connect_cfm(conn, HCI_ERROR_AUTH_FAILURE);
2471 hci_conn_drop(conn);
2472 goto unlock;
2473 }
2474
2475 hci_proto_connect_cfm(conn, ev->status);
2476 hci_conn_drop(conn);
2477 } else
2478 hci_encrypt_cfm(conn, ev->status, ev->encrypt);
2479
2480 unlock:
2481 hci_dev_unlock(hdev);
2482 }
2483
2484 static void hci_change_link_key_complete_evt(struct hci_dev *hdev,
2485 struct sk_buff *skb)
2486 {
2487 struct hci_ev_change_link_key_complete *ev = (void *) skb->data;
2488 struct hci_conn *conn;
2489
2490 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2491
2492 hci_dev_lock(hdev);
2493
2494 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2495 if (conn) {
2496 if (!ev->status)
2497 set_bit(HCI_CONN_SECURE, &conn->flags);
2498
2499 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2500
2501 hci_key_change_cfm(conn, ev->status);
2502 }
2503
2504 hci_dev_unlock(hdev);
2505 }
2506
2507 static void hci_remote_features_evt(struct hci_dev *hdev,
2508 struct sk_buff *skb)
2509 {
2510 struct hci_ev_remote_features *ev = (void *) skb->data;
2511 struct hci_conn *conn;
2512
2513 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2514
2515 hci_dev_lock(hdev);
2516
2517 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2518 if (!conn)
2519 goto unlock;
2520
2521 if (!ev->status)
2522 memcpy(conn->features[0], ev->features, 8);
2523
2524 if (conn->state != BT_CONFIG)
2525 goto unlock;
2526
2527 if (!ev->status && lmp_ssp_capable(hdev) && lmp_ssp_capable(conn)) {
2528 struct hci_cp_read_remote_ext_features cp;
2529 cp.handle = ev->handle;
2530 cp.page = 0x01;
2531 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES,
2532 sizeof(cp), &cp);
2533 goto unlock;
2534 }
2535
2536 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
2537 struct hci_cp_remote_name_req cp;
2538 memset(&cp, 0, sizeof(cp));
2539 bacpy(&cp.bdaddr, &conn->dst);
2540 cp.pscan_rep_mode = 0x02;
2541 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
2542 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2543 mgmt_device_connected(hdev, conn, 0, NULL, 0);
2544
2545 if (!hci_outgoing_auth_needed(hdev, conn)) {
2546 conn->state = BT_CONNECTED;
2547 hci_proto_connect_cfm(conn, ev->status);
2548 hci_conn_drop(conn);
2549 }
2550
2551 unlock:
2552 hci_dev_unlock(hdev);
2553 }
2554
2555 static void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2556 {
2557 struct hci_ev_cmd_complete *ev = (void *) skb->data;
2558 u8 status = skb->data[sizeof(*ev)];
2559 __u16 opcode;
2560
2561 skb_pull(skb, sizeof(*ev));
2562
2563 opcode = __le16_to_cpu(ev->opcode);
2564
2565 switch (opcode) {
2566 case HCI_OP_INQUIRY_CANCEL:
2567 hci_cc_inquiry_cancel(hdev, skb);
2568 break;
2569
2570 case HCI_OP_PERIODIC_INQ:
2571 hci_cc_periodic_inq(hdev, skb);
2572 break;
2573
2574 case HCI_OP_EXIT_PERIODIC_INQ:
2575 hci_cc_exit_periodic_inq(hdev, skb);
2576 break;
2577
2578 case HCI_OP_REMOTE_NAME_REQ_CANCEL:
2579 hci_cc_remote_name_req_cancel(hdev, skb);
2580 break;
2581
2582 case HCI_OP_ROLE_DISCOVERY:
2583 hci_cc_role_discovery(hdev, skb);
2584 break;
2585
2586 case HCI_OP_READ_LINK_POLICY:
2587 hci_cc_read_link_policy(hdev, skb);
2588 break;
2589
2590 case HCI_OP_WRITE_LINK_POLICY:
2591 hci_cc_write_link_policy(hdev, skb);
2592 break;
2593
2594 case HCI_OP_READ_DEF_LINK_POLICY:
2595 hci_cc_read_def_link_policy(hdev, skb);
2596 break;
2597
2598 case HCI_OP_WRITE_DEF_LINK_POLICY:
2599 hci_cc_write_def_link_policy(hdev, skb);
2600 break;
2601
2602 case HCI_OP_RESET:
2603 hci_cc_reset(hdev, skb);
2604 break;
2605
2606 case HCI_OP_WRITE_LOCAL_NAME:
2607 hci_cc_write_local_name(hdev, skb);
2608 break;
2609
2610 case HCI_OP_READ_LOCAL_NAME:
2611 hci_cc_read_local_name(hdev, skb);
2612 break;
2613
2614 case HCI_OP_WRITE_AUTH_ENABLE:
2615 hci_cc_write_auth_enable(hdev, skb);
2616 break;
2617
2618 case HCI_OP_WRITE_ENCRYPT_MODE:
2619 hci_cc_write_encrypt_mode(hdev, skb);
2620 break;
2621
2622 case HCI_OP_WRITE_SCAN_ENABLE:
2623 hci_cc_write_scan_enable(hdev, skb);
2624 break;
2625
2626 case HCI_OP_READ_CLASS_OF_DEV:
2627 hci_cc_read_class_of_dev(hdev, skb);
2628 break;
2629
2630 case HCI_OP_WRITE_CLASS_OF_DEV:
2631 hci_cc_write_class_of_dev(hdev, skb);
2632 break;
2633
2634 case HCI_OP_READ_VOICE_SETTING:
2635 hci_cc_read_voice_setting(hdev, skb);
2636 break;
2637
2638 case HCI_OP_WRITE_VOICE_SETTING:
2639 hci_cc_write_voice_setting(hdev, skb);
2640 break;
2641
2642 case HCI_OP_READ_NUM_SUPPORTED_IAC:
2643 hci_cc_read_num_supported_iac(hdev, skb);
2644 break;
2645
2646 case HCI_OP_WRITE_SSP_MODE:
2647 hci_cc_write_ssp_mode(hdev, skb);
2648 break;
2649
2650 case HCI_OP_WRITE_SC_SUPPORT:
2651 hci_cc_write_sc_support(hdev, skb);
2652 break;
2653
2654 case HCI_OP_READ_LOCAL_VERSION:
2655 hci_cc_read_local_version(hdev, skb);
2656 break;
2657
2658 case HCI_OP_READ_LOCAL_COMMANDS:
2659 hci_cc_read_local_commands(hdev, skb);
2660 break;
2661
2662 case HCI_OP_READ_LOCAL_FEATURES:
2663 hci_cc_read_local_features(hdev, skb);
2664 break;
2665
2666 case HCI_OP_READ_LOCAL_EXT_FEATURES:
2667 hci_cc_read_local_ext_features(hdev, skb);
2668 break;
2669
2670 case HCI_OP_READ_BUFFER_SIZE:
2671 hci_cc_read_buffer_size(hdev, skb);
2672 break;
2673
2674 case HCI_OP_READ_BD_ADDR:
2675 hci_cc_read_bd_addr(hdev, skb);
2676 break;
2677
2678 case HCI_OP_READ_PAGE_SCAN_ACTIVITY:
2679 hci_cc_read_page_scan_activity(hdev, skb);
2680 break;
2681
2682 case HCI_OP_WRITE_PAGE_SCAN_ACTIVITY:
2683 hci_cc_write_page_scan_activity(hdev, skb);
2684 break;
2685
2686 case HCI_OP_READ_PAGE_SCAN_TYPE:
2687 hci_cc_read_page_scan_type(hdev, skb);
2688 break;
2689
2690 case HCI_OP_WRITE_PAGE_SCAN_TYPE:
2691 hci_cc_write_page_scan_type(hdev, skb);
2692 break;
2693
2694 case HCI_OP_READ_DATA_BLOCK_SIZE:
2695 hci_cc_read_data_block_size(hdev, skb);
2696 break;
2697
2698 case HCI_OP_READ_FLOW_CONTROL_MODE:
2699 hci_cc_read_flow_control_mode(hdev, skb);
2700 break;
2701
2702 case HCI_OP_READ_LOCAL_AMP_INFO:
2703 hci_cc_read_local_amp_info(hdev, skb);
2704 break;
2705
2706 case HCI_OP_READ_CLOCK:
2707 hci_cc_read_clock(hdev, skb);
2708 break;
2709
2710 case HCI_OP_READ_LOCAL_AMP_ASSOC:
2711 hci_cc_read_local_amp_assoc(hdev, skb);
2712 break;
2713
2714 case HCI_OP_READ_INQ_RSP_TX_POWER:
2715 hci_cc_read_inq_rsp_tx_power(hdev, skb);
2716 break;
2717
2718 case HCI_OP_PIN_CODE_REPLY:
2719 hci_cc_pin_code_reply(hdev, skb);
2720 break;
2721
2722 case HCI_OP_PIN_CODE_NEG_REPLY:
2723 hci_cc_pin_code_neg_reply(hdev, skb);
2724 break;
2725
2726 case HCI_OP_READ_LOCAL_OOB_DATA:
2727 hci_cc_read_local_oob_data(hdev, skb);
2728 break;
2729
2730 case HCI_OP_READ_LOCAL_OOB_EXT_DATA:
2731 hci_cc_read_local_oob_ext_data(hdev, skb);
2732 break;
2733
2734 case HCI_OP_LE_READ_BUFFER_SIZE:
2735 hci_cc_le_read_buffer_size(hdev, skb);
2736 break;
2737
2738 case HCI_OP_LE_READ_LOCAL_FEATURES:
2739 hci_cc_le_read_local_features(hdev, skb);
2740 break;
2741
2742 case HCI_OP_LE_READ_ADV_TX_POWER:
2743 hci_cc_le_read_adv_tx_power(hdev, skb);
2744 break;
2745
2746 case HCI_OP_USER_CONFIRM_REPLY:
2747 hci_cc_user_confirm_reply(hdev, skb);
2748 break;
2749
2750 case HCI_OP_USER_CONFIRM_NEG_REPLY:
2751 hci_cc_user_confirm_neg_reply(hdev, skb);
2752 break;
2753
2754 case HCI_OP_USER_PASSKEY_REPLY:
2755 hci_cc_user_passkey_reply(hdev, skb);
2756 break;
2757
2758 case HCI_OP_USER_PASSKEY_NEG_REPLY:
2759 hci_cc_user_passkey_neg_reply(hdev, skb);
2760 break;
2761
2762 case HCI_OP_LE_SET_RANDOM_ADDR:
2763 hci_cc_le_set_random_addr(hdev, skb);
2764 break;
2765
2766 case HCI_OP_LE_SET_ADV_ENABLE:
2767 hci_cc_le_set_adv_enable(hdev, skb);
2768 break;
2769
2770 case HCI_OP_LE_SET_SCAN_PARAM:
2771 hci_cc_le_set_scan_param(hdev, skb);
2772 break;
2773
2774 case HCI_OP_LE_SET_SCAN_ENABLE:
2775 hci_cc_le_set_scan_enable(hdev, skb);
2776 break;
2777
2778 case HCI_OP_LE_READ_WHITE_LIST_SIZE:
2779 hci_cc_le_read_white_list_size(hdev, skb);
2780 break;
2781
2782 case HCI_OP_LE_CLEAR_WHITE_LIST:
2783 hci_cc_le_clear_white_list(hdev, skb);
2784 break;
2785
2786 case HCI_OP_LE_ADD_TO_WHITE_LIST:
2787 hci_cc_le_add_to_white_list(hdev, skb);
2788 break;
2789
2790 case HCI_OP_LE_DEL_FROM_WHITE_LIST:
2791 hci_cc_le_del_from_white_list(hdev, skb);
2792 break;
2793
2794 case HCI_OP_LE_READ_SUPPORTED_STATES:
2795 hci_cc_le_read_supported_states(hdev, skb);
2796 break;
2797
2798 case HCI_OP_WRITE_LE_HOST_SUPPORTED:
2799 hci_cc_write_le_host_supported(hdev, skb);
2800 break;
2801
2802 case HCI_OP_LE_SET_ADV_PARAM:
2803 hci_cc_set_adv_param(hdev, skb);
2804 break;
2805
2806 case HCI_OP_WRITE_REMOTE_AMP_ASSOC:
2807 hci_cc_write_remote_amp_assoc(hdev, skb);
2808 break;
2809
2810 case HCI_OP_READ_RSSI:
2811 hci_cc_read_rssi(hdev, skb);
2812 break;
2813
2814 case HCI_OP_READ_TX_POWER:
2815 hci_cc_read_tx_power(hdev, skb);
2816 break;
2817
2818 default:
2819 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2820 break;
2821 }
2822
2823 if (opcode != HCI_OP_NOP)
2824 cancel_delayed_work(&hdev->cmd_timer);
2825
2826 hci_req_cmd_complete(hdev, opcode, status);
2827
2828 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2829 atomic_set(&hdev->cmd_cnt, 1);
2830 if (!skb_queue_empty(&hdev->cmd_q))
2831 queue_work(hdev->workqueue, &hdev->cmd_work);
2832 }
2833 }
2834
2835 static void hci_cmd_status_evt(struct hci_dev *hdev, struct sk_buff *skb)
2836 {
2837 struct hci_ev_cmd_status *ev = (void *) skb->data;
2838 __u16 opcode;
2839
2840 skb_pull(skb, sizeof(*ev));
2841
2842 opcode = __le16_to_cpu(ev->opcode);
2843
2844 switch (opcode) {
2845 case HCI_OP_INQUIRY:
2846 hci_cs_inquiry(hdev, ev->status);
2847 break;
2848
2849 case HCI_OP_CREATE_CONN:
2850 hci_cs_create_conn(hdev, ev->status);
2851 break;
2852
2853 case HCI_OP_ADD_SCO:
2854 hci_cs_add_sco(hdev, ev->status);
2855 break;
2856
2857 case HCI_OP_AUTH_REQUESTED:
2858 hci_cs_auth_requested(hdev, ev->status);
2859 break;
2860
2861 case HCI_OP_SET_CONN_ENCRYPT:
2862 hci_cs_set_conn_encrypt(hdev, ev->status);
2863 break;
2864
2865 case HCI_OP_REMOTE_NAME_REQ:
2866 hci_cs_remote_name_req(hdev, ev->status);
2867 break;
2868
2869 case HCI_OP_READ_REMOTE_FEATURES:
2870 hci_cs_read_remote_features(hdev, ev->status);
2871 break;
2872
2873 case HCI_OP_READ_REMOTE_EXT_FEATURES:
2874 hci_cs_read_remote_ext_features(hdev, ev->status);
2875 break;
2876
2877 case HCI_OP_SETUP_SYNC_CONN:
2878 hci_cs_setup_sync_conn(hdev, ev->status);
2879 break;
2880
2881 case HCI_OP_SNIFF_MODE:
2882 hci_cs_sniff_mode(hdev, ev->status);
2883 break;
2884
2885 case HCI_OP_EXIT_SNIFF_MODE:
2886 hci_cs_exit_sniff_mode(hdev, ev->status);
2887 break;
2888
2889 case HCI_OP_DISCONNECT:
2890 hci_cs_disconnect(hdev, ev->status);
2891 break;
2892
2893 case HCI_OP_CREATE_PHY_LINK:
2894 hci_cs_create_phylink(hdev, ev->status);
2895 break;
2896
2897 case HCI_OP_ACCEPT_PHY_LINK:
2898 hci_cs_accept_phylink(hdev, ev->status);
2899 break;
2900
2901 case HCI_OP_LE_CREATE_CONN:
2902 hci_cs_le_create_conn(hdev, ev->status);
2903 break;
2904
2905 case HCI_OP_LE_START_ENC:
2906 hci_cs_le_start_enc(hdev, ev->status);
2907 break;
2908
2909 default:
2910 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
2911 break;
2912 }
2913
2914 if (opcode != HCI_OP_NOP)
2915 cancel_delayed_work(&hdev->cmd_timer);
2916
2917 if (ev->status ||
2918 (hdev->sent_cmd && !bt_cb(hdev->sent_cmd)->req.event))
2919 hci_req_cmd_complete(hdev, opcode, ev->status);
2920
2921 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags)) {
2922 atomic_set(&hdev->cmd_cnt, 1);
2923 if (!skb_queue_empty(&hdev->cmd_q))
2924 queue_work(hdev->workqueue, &hdev->cmd_work);
2925 }
2926 }
2927
2928 static void hci_hardware_error_evt(struct hci_dev *hdev, struct sk_buff *skb)
2929 {
2930 struct hci_ev_hardware_error *ev = (void *) skb->data;
2931
2932 BT_ERR("%s hardware error 0x%2.2x", hdev->name, ev->code);
2933 }
2934
2935 static void hci_role_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2936 {
2937 struct hci_ev_role_change *ev = (void *) skb->data;
2938 struct hci_conn *conn;
2939
2940 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2941
2942 hci_dev_lock(hdev);
2943
2944 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2945 if (conn) {
2946 if (!ev->status)
2947 conn->role = ev->role;
2948
2949 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
2950
2951 hci_role_switch_cfm(conn, ev->status, ev->role);
2952 }
2953
2954 hci_dev_unlock(hdev);
2955 }
2956
2957 static void hci_num_comp_pkts_evt(struct hci_dev *hdev, struct sk_buff *skb)
2958 {
2959 struct hci_ev_num_comp_pkts *ev = (void *) skb->data;
2960 int i;
2961
2962 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_PACKET_BASED) {
2963 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
2964 return;
2965 }
2966
2967 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
2968 ev->num_hndl * sizeof(struct hci_comp_pkts_info)) {
2969 BT_DBG("%s bad parameters", hdev->name);
2970 return;
2971 }
2972
2973 BT_DBG("%s num_hndl %d", hdev->name, ev->num_hndl);
2974
2975 for (i = 0; i < ev->num_hndl; i++) {
2976 struct hci_comp_pkts_info *info = &ev->handles[i];
2977 struct hci_conn *conn;
2978 __u16 handle, count;
2979
2980 handle = __le16_to_cpu(info->handle);
2981 count = __le16_to_cpu(info->count);
2982
2983 conn = hci_conn_hash_lookup_handle(hdev, handle);
2984 if (!conn)
2985 continue;
2986
2987 conn->sent -= count;
2988
2989 switch (conn->type) {
2990 case ACL_LINK:
2991 hdev->acl_cnt += count;
2992 if (hdev->acl_cnt > hdev->acl_pkts)
2993 hdev->acl_cnt = hdev->acl_pkts;
2994 break;
2995
2996 case LE_LINK:
2997 if (hdev->le_pkts) {
2998 hdev->le_cnt += count;
2999 if (hdev->le_cnt > hdev->le_pkts)
3000 hdev->le_cnt = hdev->le_pkts;
3001 } else {
3002 hdev->acl_cnt += count;
3003 if (hdev->acl_cnt > hdev->acl_pkts)
3004 hdev->acl_cnt = hdev->acl_pkts;
3005 }
3006 break;
3007
3008 case SCO_LINK:
3009 hdev->sco_cnt += count;
3010 if (hdev->sco_cnt > hdev->sco_pkts)
3011 hdev->sco_cnt = hdev->sco_pkts;
3012 break;
3013
3014 default:
3015 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3016 break;
3017 }
3018 }
3019
3020 queue_work(hdev->workqueue, &hdev->tx_work);
3021 }
3022
3023 static struct hci_conn *__hci_conn_lookup_handle(struct hci_dev *hdev,
3024 __u16 handle)
3025 {
3026 struct hci_chan *chan;
3027
3028 switch (hdev->dev_type) {
3029 case HCI_BREDR:
3030 return hci_conn_hash_lookup_handle(hdev, handle);
3031 case HCI_AMP:
3032 chan = hci_chan_lookup_handle(hdev, handle);
3033 if (chan)
3034 return chan->conn;
3035 break;
3036 default:
3037 BT_ERR("%s unknown dev_type %d", hdev->name, hdev->dev_type);
3038 break;
3039 }
3040
3041 return NULL;
3042 }
3043
3044 static void hci_num_comp_blocks_evt(struct hci_dev *hdev, struct sk_buff *skb)
3045 {
3046 struct hci_ev_num_comp_blocks *ev = (void *) skb->data;
3047 int i;
3048
3049 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_BLOCK_BASED) {
3050 BT_ERR("Wrong event for mode %d", hdev->flow_ctl_mode);
3051 return;
3052 }
3053
3054 if (skb->len < sizeof(*ev) || skb->len < sizeof(*ev) +
3055 ev->num_hndl * sizeof(struct hci_comp_blocks_info)) {
3056 BT_DBG("%s bad parameters", hdev->name);
3057 return;
3058 }
3059
3060 BT_DBG("%s num_blocks %d num_hndl %d", hdev->name, ev->num_blocks,
3061 ev->num_hndl);
3062
3063 for (i = 0; i < ev->num_hndl; i++) {
3064 struct hci_comp_blocks_info *info = &ev->handles[i];
3065 struct hci_conn *conn = NULL;
3066 __u16 handle, block_count;
3067
3068 handle = __le16_to_cpu(info->handle);
3069 block_count = __le16_to_cpu(info->blocks);
3070
3071 conn = __hci_conn_lookup_handle(hdev, handle);
3072 if (!conn)
3073 continue;
3074
3075 conn->sent -= block_count;
3076
3077 switch (conn->type) {
3078 case ACL_LINK:
3079 case AMP_LINK:
3080 hdev->block_cnt += block_count;
3081 if (hdev->block_cnt > hdev->num_blocks)
3082 hdev->block_cnt = hdev->num_blocks;
3083 break;
3084
3085 default:
3086 BT_ERR("Unknown type %d conn %p", conn->type, conn);
3087 break;
3088 }
3089 }
3090
3091 queue_work(hdev->workqueue, &hdev->tx_work);
3092 }
3093
3094 static void hci_mode_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3095 {
3096 struct hci_ev_mode_change *ev = (void *) skb->data;
3097 struct hci_conn *conn;
3098
3099 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3100
3101 hci_dev_lock(hdev);
3102
3103 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3104 if (conn) {
3105 conn->mode = ev->mode;
3106
3107 if (!test_and_clear_bit(HCI_CONN_MODE_CHANGE_PEND,
3108 &conn->flags)) {
3109 if (conn->mode == HCI_CM_ACTIVE)
3110 set_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3111 else
3112 clear_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3113 }
3114
3115 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
3116 hci_sco_setup(conn, ev->status);
3117 }
3118
3119 hci_dev_unlock(hdev);
3120 }
3121
3122 static void hci_pin_code_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3123 {
3124 struct hci_ev_pin_code_req *ev = (void *) skb->data;
3125 struct hci_conn *conn;
3126
3127 BT_DBG("%s", hdev->name);
3128
3129 hci_dev_lock(hdev);
3130
3131 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3132 if (!conn)
3133 goto unlock;
3134
3135 if (conn->state == BT_CONNECTED) {
3136 hci_conn_hold(conn);
3137 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
3138 hci_conn_drop(conn);
3139 }
3140
3141 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags) &&
3142 !test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags)) {
3143 hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3144 sizeof(ev->bdaddr), &ev->bdaddr);
3145 } else if (test_bit(HCI_MGMT, &hdev->dev_flags)) {
3146 u8 secure;
3147
3148 if (conn->pending_sec_level == BT_SECURITY_HIGH)
3149 secure = 1;
3150 else
3151 secure = 0;
3152
3153 mgmt_pin_code_request(hdev, &ev->bdaddr, secure);
3154 }
3155
3156 unlock:
3157 hci_dev_unlock(hdev);
3158 }
3159
3160 static void hci_link_key_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3161 {
3162 struct hci_ev_link_key_req *ev = (void *) skb->data;
3163 struct hci_cp_link_key_reply cp;
3164 struct hci_conn *conn;
3165 struct link_key *key;
3166
3167 BT_DBG("%s", hdev->name);
3168
3169 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3170 return;
3171
3172 hci_dev_lock(hdev);
3173
3174 key = hci_find_link_key(hdev, &ev->bdaddr);
3175 if (!key) {
3176 BT_DBG("%s link key not found for %pMR", hdev->name,
3177 &ev->bdaddr);
3178 goto not_found;
3179 }
3180
3181 BT_DBG("%s found key type %u for %pMR", hdev->name, key->type,
3182 &ev->bdaddr);
3183
3184 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3185 if (conn) {
3186 if ((key->type == HCI_LK_UNAUTH_COMBINATION_P192 ||
3187 key->type == HCI_LK_UNAUTH_COMBINATION_P256) &&
3188 conn->auth_type != 0xff && (conn->auth_type & 0x01)) {
3189 BT_DBG("%s ignoring unauthenticated key", hdev->name);
3190 goto not_found;
3191 }
3192
3193 if (key->type == HCI_LK_COMBINATION && key->pin_len < 16 &&
3194 (conn->pending_sec_level == BT_SECURITY_HIGH ||
3195 conn->pending_sec_level == BT_SECURITY_FIPS)) {
3196 BT_DBG("%s ignoring key unauthenticated for high security",
3197 hdev->name);
3198 goto not_found;
3199 }
3200
3201 conn->key_type = key->type;
3202 conn->pin_length = key->pin_len;
3203 }
3204
3205 bacpy(&cp.bdaddr, &ev->bdaddr);
3206 memcpy(cp.link_key, key->val, HCI_LINK_KEY_SIZE);
3207
3208 hci_send_cmd(hdev, HCI_OP_LINK_KEY_REPLY, sizeof(cp), &cp);
3209
3210 hci_dev_unlock(hdev);
3211
3212 return;
3213
3214 not_found:
3215 hci_send_cmd(hdev, HCI_OP_LINK_KEY_NEG_REPLY, 6, &ev->bdaddr);
3216 hci_dev_unlock(hdev);
3217 }
3218
3219 static void hci_link_key_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3220 {
3221 struct hci_ev_link_key_notify *ev = (void *) skb->data;
3222 struct hci_conn *conn;
3223 struct link_key *key;
3224 bool persistent;
3225 u8 pin_len = 0;
3226
3227 BT_DBG("%s", hdev->name);
3228
3229 hci_dev_lock(hdev);
3230
3231 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3232 if (conn) {
3233 hci_conn_hold(conn);
3234 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3235 pin_len = conn->pin_length;
3236
3237 if (ev->key_type != HCI_LK_CHANGED_COMBINATION)
3238 conn->key_type = ev->key_type;
3239
3240 hci_conn_drop(conn);
3241 }
3242
3243 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3244 goto unlock;
3245
3246 key = hci_add_link_key(hdev, conn, &ev->bdaddr, ev->link_key,
3247 ev->key_type, pin_len, &persistent);
3248 if (!key)
3249 goto unlock;
3250
3251 mgmt_new_link_key(hdev, key, persistent);
3252
3253 /* Keep debug keys around only if the HCI_KEEP_DEBUG_KEYS flag
3254 * is set. If it's not set simply remove the key from the kernel
3255 * list (we've still notified user space about it but with
3256 * store_hint being 0).
3257 */
3258 if (key->type == HCI_LK_DEBUG_COMBINATION &&
3259 !test_bit(HCI_KEEP_DEBUG_KEYS, &hdev->dev_flags)) {
3260 list_del(&key->list);
3261 kfree(key);
3262 } else if (conn) {
3263 if (persistent)
3264 clear_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3265 else
3266 set_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3267 }
3268
3269 unlock:
3270 hci_dev_unlock(hdev);
3271 }
3272
3273 static void hci_clock_offset_evt(struct hci_dev *hdev, struct sk_buff *skb)
3274 {
3275 struct hci_ev_clock_offset *ev = (void *) skb->data;
3276 struct hci_conn *conn;
3277
3278 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3279
3280 hci_dev_lock(hdev);
3281
3282 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3283 if (conn && !ev->status) {
3284 struct inquiry_entry *ie;
3285
3286 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3287 if (ie) {
3288 ie->data.clock_offset = ev->clock_offset;
3289 ie->timestamp = jiffies;
3290 }
3291 }
3292
3293 hci_dev_unlock(hdev);
3294 }
3295
3296 static void hci_pkt_type_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3297 {
3298 struct hci_ev_pkt_type_change *ev = (void *) skb->data;
3299 struct hci_conn *conn;
3300
3301 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3302
3303 hci_dev_lock(hdev);
3304
3305 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3306 if (conn && !ev->status)
3307 conn->pkt_type = __le16_to_cpu(ev->pkt_type);
3308
3309 hci_dev_unlock(hdev);
3310 }
3311
3312 static void hci_pscan_rep_mode_evt(struct hci_dev *hdev, struct sk_buff *skb)
3313 {
3314 struct hci_ev_pscan_rep_mode *ev = (void *) skb->data;
3315 struct inquiry_entry *ie;
3316
3317 BT_DBG("%s", hdev->name);
3318
3319 hci_dev_lock(hdev);
3320
3321 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3322 if (ie) {
3323 ie->data.pscan_rep_mode = ev->pscan_rep_mode;
3324 ie->timestamp = jiffies;
3325 }
3326
3327 hci_dev_unlock(hdev);
3328 }
3329
3330 static void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev,
3331 struct sk_buff *skb)
3332 {
3333 struct inquiry_data data;
3334 int num_rsp = *((__u8 *) skb->data);
3335
3336 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3337
3338 if (!num_rsp)
3339 return;
3340
3341 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3342 return;
3343
3344 hci_dev_lock(hdev);
3345
3346 if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) {
3347 struct inquiry_info_with_rssi_and_pscan_mode *info;
3348 info = (void *) (skb->data + 1);
3349
3350 for (; num_rsp; num_rsp--, info++) {
3351 u32 flags;
3352
3353 bacpy(&data.bdaddr, &info->bdaddr);
3354 data.pscan_rep_mode = info->pscan_rep_mode;
3355 data.pscan_period_mode = info->pscan_period_mode;
3356 data.pscan_mode = info->pscan_mode;
3357 memcpy(data.dev_class, info->dev_class, 3);
3358 data.clock_offset = info->clock_offset;
3359 data.rssi = info->rssi;
3360 data.ssp_mode = 0x00;
3361
3362 flags = hci_inquiry_cache_update(hdev, &data, false);
3363
3364 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3365 info->dev_class, info->rssi,
3366 flags, NULL, 0, NULL, 0);
3367 }
3368 } else {
3369 struct inquiry_info_with_rssi *info = (void *) (skb->data + 1);
3370
3371 for (; num_rsp; num_rsp--, info++) {
3372 u32 flags;
3373
3374 bacpy(&data.bdaddr, &info->bdaddr);
3375 data.pscan_rep_mode = info->pscan_rep_mode;
3376 data.pscan_period_mode = info->pscan_period_mode;
3377 data.pscan_mode = 0x00;
3378 memcpy(data.dev_class, info->dev_class, 3);
3379 data.clock_offset = info->clock_offset;
3380 data.rssi = info->rssi;
3381 data.ssp_mode = 0x00;
3382
3383 flags = hci_inquiry_cache_update(hdev, &data, false);
3384
3385 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3386 info->dev_class, info->rssi,
3387 flags, NULL, 0, NULL, 0);
3388 }
3389 }
3390
3391 hci_dev_unlock(hdev);
3392 }
3393
3394 static void hci_remote_ext_features_evt(struct hci_dev *hdev,
3395 struct sk_buff *skb)
3396 {
3397 struct hci_ev_remote_ext_features *ev = (void *) skb->data;
3398 struct hci_conn *conn;
3399
3400 BT_DBG("%s", hdev->name);
3401
3402 hci_dev_lock(hdev);
3403
3404 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3405 if (!conn)
3406 goto unlock;
3407
3408 if (ev->page < HCI_MAX_PAGES)
3409 memcpy(conn->features[ev->page], ev->features, 8);
3410
3411 if (!ev->status && ev->page == 0x01) {
3412 struct inquiry_entry *ie;
3413
3414 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3415 if (ie)
3416 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3417
3418 if (ev->features[0] & LMP_HOST_SSP) {
3419 set_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3420 } else {
3421 /* It is mandatory by the Bluetooth specification that
3422 * Extended Inquiry Results are only used when Secure
3423 * Simple Pairing is enabled, but some devices violate
3424 * this.
3425 *
3426 * To make these devices work, the internal SSP
3427 * enabled flag needs to be cleared if the remote host
3428 * features do not indicate SSP support */
3429 clear_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
3430 }
3431
3432 if (ev->features[0] & LMP_HOST_SC)
3433 set_bit(HCI_CONN_SC_ENABLED, &conn->flags);
3434 }
3435
3436 if (conn->state != BT_CONFIG)
3437 goto unlock;
3438
3439 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
3440 struct hci_cp_remote_name_req cp;
3441 memset(&cp, 0, sizeof(cp));
3442 bacpy(&cp.bdaddr, &conn->dst);
3443 cp.pscan_rep_mode = 0x02;
3444 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
3445 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
3446 mgmt_device_connected(hdev, conn, 0, NULL, 0);
3447
3448 if (!hci_outgoing_auth_needed(hdev, conn)) {
3449 conn->state = BT_CONNECTED;
3450 hci_proto_connect_cfm(conn, ev->status);
3451 hci_conn_drop(conn);
3452 }
3453
3454 unlock:
3455 hci_dev_unlock(hdev);
3456 }
3457
3458 static void hci_sync_conn_complete_evt(struct hci_dev *hdev,
3459 struct sk_buff *skb)
3460 {
3461 struct hci_ev_sync_conn_complete *ev = (void *) skb->data;
3462 struct hci_conn *conn;
3463
3464 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3465
3466 hci_dev_lock(hdev);
3467
3468 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
3469 if (!conn) {
3470 if (ev->link_type == ESCO_LINK)
3471 goto unlock;
3472
3473 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
3474 if (!conn)
3475 goto unlock;
3476
3477 conn->type = SCO_LINK;
3478 }
3479
3480 switch (ev->status) {
3481 case 0x00:
3482 conn->handle = __le16_to_cpu(ev->handle);
3483 conn->state = BT_CONNECTED;
3484
3485 hci_conn_add_sysfs(conn);
3486 break;
3487
3488 case 0x10: /* Connection Accept Timeout */
3489 case 0x0d: /* Connection Rejected due to Limited Resources */
3490 case 0x11: /* Unsupported Feature or Parameter Value */
3491 case 0x1c: /* SCO interval rejected */
3492 case 0x1a: /* Unsupported Remote Feature */
3493 case 0x1f: /* Unspecified error */
3494 case 0x20: /* Unsupported LMP Parameter value */
3495 if (conn->out) {
3496 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
3497 (hdev->esco_type & EDR_ESCO_MASK);
3498 if (hci_setup_sync(conn, conn->link->handle))
3499 goto unlock;
3500 }
3501 /* fall through */
3502
3503 default:
3504 conn->state = BT_CLOSED;
3505 break;
3506 }
3507
3508 hci_proto_connect_cfm(conn, ev->status);
3509 if (ev->status)
3510 hci_conn_del(conn);
3511
3512 unlock:
3513 hci_dev_unlock(hdev);
3514 }
3515
3516 static inline size_t eir_get_length(u8 *eir, size_t eir_len)
3517 {
3518 size_t parsed = 0;
3519
3520 while (parsed < eir_len) {
3521 u8 field_len = eir[0];
3522
3523 if (field_len == 0)
3524 return parsed;
3525
3526 parsed += field_len + 1;
3527 eir += field_len + 1;
3528 }
3529
3530 return eir_len;
3531 }
3532
3533 static void hci_extended_inquiry_result_evt(struct hci_dev *hdev,
3534 struct sk_buff *skb)
3535 {
3536 struct inquiry_data data;
3537 struct extended_inquiry_info *info = (void *) (skb->data + 1);
3538 int num_rsp = *((__u8 *) skb->data);
3539 size_t eir_len;
3540
3541 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3542
3543 if (!num_rsp)
3544 return;
3545
3546 if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags))
3547 return;
3548
3549 hci_dev_lock(hdev);
3550
3551 for (; num_rsp; num_rsp--, info++) {
3552 u32 flags;
3553 bool name_known;
3554
3555 bacpy(&data.bdaddr, &info->bdaddr);
3556 data.pscan_rep_mode = info->pscan_rep_mode;
3557 data.pscan_period_mode = info->pscan_period_mode;
3558 data.pscan_mode = 0x00;
3559 memcpy(data.dev_class, info->dev_class, 3);
3560 data.clock_offset = info->clock_offset;
3561 data.rssi = info->rssi;
3562 data.ssp_mode = 0x01;
3563
3564 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3565 name_known = eir_has_data_type(info->data,
3566 sizeof(info->data),
3567 EIR_NAME_COMPLETE);
3568 else
3569 name_known = true;
3570
3571 flags = hci_inquiry_cache_update(hdev, &data, name_known);
3572
3573 eir_len = eir_get_length(info->data, sizeof(info->data));
3574
3575 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
3576 info->dev_class, info->rssi,
3577 flags, info->data, eir_len, NULL, 0);
3578 }
3579
3580 hci_dev_unlock(hdev);
3581 }
3582
3583 static void hci_key_refresh_complete_evt(struct hci_dev *hdev,
3584 struct sk_buff *skb)
3585 {
3586 struct hci_ev_key_refresh_complete *ev = (void *) skb->data;
3587 struct hci_conn *conn;
3588
3589 BT_DBG("%s status 0x%2.2x handle 0x%4.4x", hdev->name, ev->status,
3590 __le16_to_cpu(ev->handle));
3591
3592 hci_dev_lock(hdev);
3593
3594 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3595 if (!conn)
3596 goto unlock;
3597
3598 /* For BR/EDR the necessary steps are taken through the
3599 * auth_complete event.
3600 */
3601 if (conn->type != LE_LINK)
3602 goto unlock;
3603
3604 if (!ev->status)
3605 conn->sec_level = conn->pending_sec_level;
3606
3607 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
3608
3609 if (ev->status && conn->state == BT_CONNECTED) {
3610 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
3611 hci_conn_drop(conn);
3612 goto unlock;
3613 }
3614
3615 if (conn->state == BT_CONFIG) {
3616 if (!ev->status)
3617 conn->state = BT_CONNECTED;
3618
3619 hci_proto_connect_cfm(conn, ev->status);
3620 hci_conn_drop(conn);
3621 } else {
3622 hci_auth_cfm(conn, ev->status);
3623
3624 hci_conn_hold(conn);
3625 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3626 hci_conn_drop(conn);
3627 }
3628
3629 unlock:
3630 hci_dev_unlock(hdev);
3631 }
3632
3633 static u8 hci_get_auth_req(struct hci_conn *conn)
3634 {
3635 /* If remote requests no-bonding follow that lead */
3636 if (conn->remote_auth == HCI_AT_NO_BONDING ||
3637 conn->remote_auth == HCI_AT_NO_BONDING_MITM)
3638 return conn->remote_auth | (conn->auth_type & 0x01);
3639
3640 /* If both remote and local have enough IO capabilities, require
3641 * MITM protection
3642 */
3643 if (conn->remote_cap != HCI_IO_NO_INPUT_OUTPUT &&
3644 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT)
3645 return conn->remote_auth | 0x01;
3646
3647 /* No MITM protection possible so ignore remote requirement */
3648 return (conn->remote_auth & ~0x01) | (conn->auth_type & 0x01);
3649 }
3650
3651 static void hci_io_capa_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3652 {
3653 struct hci_ev_io_capa_request *ev = (void *) skb->data;
3654 struct hci_conn *conn;
3655
3656 BT_DBG("%s", hdev->name);
3657
3658 hci_dev_lock(hdev);
3659
3660 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3661 if (!conn)
3662 goto unlock;
3663
3664 hci_conn_hold(conn);
3665
3666 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3667 goto unlock;
3668
3669 /* Allow pairing if we're pairable, the initiators of the
3670 * pairing or if the remote is not requesting bonding.
3671 */
3672 if (test_bit(HCI_BONDABLE, &hdev->dev_flags) ||
3673 test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags) ||
3674 (conn->remote_auth & ~0x01) == HCI_AT_NO_BONDING) {
3675 struct hci_cp_io_capability_reply cp;
3676
3677 bacpy(&cp.bdaddr, &ev->bdaddr);
3678 /* Change the IO capability from KeyboardDisplay
3679 * to DisplayYesNo as it is not supported by BT spec. */
3680 cp.capability = (conn->io_capability == 0x04) ?
3681 HCI_IO_DISPLAY_YESNO : conn->io_capability;
3682
3683 /* If we are initiators, there is no remote information yet */
3684 if (conn->remote_auth == 0xff) {
3685 /* Request MITM protection if our IO caps allow it
3686 * except for the no-bonding case.
3687 */
3688 if (conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3689 conn->auth_type != HCI_AT_NO_BONDING)
3690 conn->auth_type |= 0x01;
3691 } else {
3692 conn->auth_type = hci_get_auth_req(conn);
3693 }
3694
3695 /* If we're not bondable, force one of the non-bondable
3696 * authentication requirement values.
3697 */
3698 if (!test_bit(HCI_BONDABLE, &hdev->dev_flags))
3699 conn->auth_type &= HCI_AT_NO_BONDING_MITM;
3700
3701 cp.authentication = conn->auth_type;
3702
3703 if (hci_find_remote_oob_data(hdev, &conn->dst) &&
3704 (conn->out || test_bit(HCI_CONN_REMOTE_OOB, &conn->flags)))
3705 cp.oob_data = 0x01;
3706 else
3707 cp.oob_data = 0x00;
3708
3709 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_REPLY,
3710 sizeof(cp), &cp);
3711 } else {
3712 struct hci_cp_io_capability_neg_reply cp;
3713
3714 bacpy(&cp.bdaddr, &ev->bdaddr);
3715 cp.reason = HCI_ERROR_PAIRING_NOT_ALLOWED;
3716
3717 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_NEG_REPLY,
3718 sizeof(cp), &cp);
3719 }
3720
3721 unlock:
3722 hci_dev_unlock(hdev);
3723 }
3724
3725 static void hci_io_capa_reply_evt(struct hci_dev *hdev, struct sk_buff *skb)
3726 {
3727 struct hci_ev_io_capa_reply *ev = (void *) skb->data;
3728 struct hci_conn *conn;
3729
3730 BT_DBG("%s", hdev->name);
3731
3732 hci_dev_lock(hdev);
3733
3734 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3735 if (!conn)
3736 goto unlock;
3737
3738 conn->remote_cap = ev->capability;
3739 conn->remote_auth = ev->authentication;
3740 if (ev->oob_data)
3741 set_bit(HCI_CONN_REMOTE_OOB, &conn->flags);
3742
3743 unlock:
3744 hci_dev_unlock(hdev);
3745 }
3746
3747 static void hci_user_confirm_request_evt(struct hci_dev *hdev,
3748 struct sk_buff *skb)
3749 {
3750 struct hci_ev_user_confirm_req *ev = (void *) skb->data;
3751 int loc_mitm, rem_mitm, confirm_hint = 0;
3752 struct hci_conn *conn;
3753
3754 BT_DBG("%s", hdev->name);
3755
3756 hci_dev_lock(hdev);
3757
3758 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3759 goto unlock;
3760
3761 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3762 if (!conn)
3763 goto unlock;
3764
3765 loc_mitm = (conn->auth_type & 0x01);
3766 rem_mitm = (conn->remote_auth & 0x01);
3767
3768 /* If we require MITM but the remote device can't provide that
3769 * (it has NoInputNoOutput) then reject the confirmation
3770 * request. We check the security level here since it doesn't
3771 * necessarily match conn->auth_type.
3772 */
3773 if (conn->pending_sec_level > BT_SECURITY_MEDIUM &&
3774 conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) {
3775 BT_DBG("Rejecting request: remote device can't provide MITM");
3776 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_NEG_REPLY,
3777 sizeof(ev->bdaddr), &ev->bdaddr);
3778 goto unlock;
3779 }
3780
3781 /* If no side requires MITM protection; auto-accept */
3782 if ((!loc_mitm || conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) &&
3783 (!rem_mitm || conn->io_capability == HCI_IO_NO_INPUT_OUTPUT)) {
3784
3785 /* If we're not the initiators request authorization to
3786 * proceed from user space (mgmt_user_confirm with
3787 * confirm_hint set to 1). The exception is if neither
3788 * side had MITM or if the local IO capability is
3789 * NoInputNoOutput, in which case we do auto-accept
3790 */
3791 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) &&
3792 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
3793 (loc_mitm || rem_mitm)) {
3794 BT_DBG("Confirming auto-accept as acceptor");
3795 confirm_hint = 1;
3796 goto confirm;
3797 }
3798
3799 BT_DBG("Auto-accept of user confirmation with %ums delay",
3800 hdev->auto_accept_delay);
3801
3802 if (hdev->auto_accept_delay > 0) {
3803 int delay = msecs_to_jiffies(hdev->auto_accept_delay);
3804 queue_delayed_work(conn->hdev->workqueue,
3805 &conn->auto_accept_work, delay);
3806 goto unlock;
3807 }
3808
3809 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY,
3810 sizeof(ev->bdaddr), &ev->bdaddr);
3811 goto unlock;
3812 }
3813
3814 confirm:
3815 mgmt_user_confirm_request(hdev, &ev->bdaddr, ACL_LINK, 0,
3816 le32_to_cpu(ev->passkey), confirm_hint);
3817
3818 unlock:
3819 hci_dev_unlock(hdev);
3820 }
3821
3822 static void hci_user_passkey_request_evt(struct hci_dev *hdev,
3823 struct sk_buff *skb)
3824 {
3825 struct hci_ev_user_passkey_req *ev = (void *) skb->data;
3826
3827 BT_DBG("%s", hdev->name);
3828
3829 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3830 mgmt_user_passkey_request(hdev, &ev->bdaddr, ACL_LINK, 0);
3831 }
3832
3833 static void hci_user_passkey_notify_evt(struct hci_dev *hdev,
3834 struct sk_buff *skb)
3835 {
3836 struct hci_ev_user_passkey_notify *ev = (void *) skb->data;
3837 struct hci_conn *conn;
3838
3839 BT_DBG("%s", hdev->name);
3840
3841 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3842 if (!conn)
3843 return;
3844
3845 conn->passkey_notify = __le32_to_cpu(ev->passkey);
3846 conn->passkey_entered = 0;
3847
3848 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3849 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3850 conn->dst_type, conn->passkey_notify,
3851 conn->passkey_entered);
3852 }
3853
3854 static void hci_keypress_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3855 {
3856 struct hci_ev_keypress_notify *ev = (void *) skb->data;
3857 struct hci_conn *conn;
3858
3859 BT_DBG("%s", hdev->name);
3860
3861 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3862 if (!conn)
3863 return;
3864
3865 switch (ev->type) {
3866 case HCI_KEYPRESS_STARTED:
3867 conn->passkey_entered = 0;
3868 return;
3869
3870 case HCI_KEYPRESS_ENTERED:
3871 conn->passkey_entered++;
3872 break;
3873
3874 case HCI_KEYPRESS_ERASED:
3875 conn->passkey_entered--;
3876 break;
3877
3878 case HCI_KEYPRESS_CLEARED:
3879 conn->passkey_entered = 0;
3880 break;
3881
3882 case HCI_KEYPRESS_COMPLETED:
3883 return;
3884 }
3885
3886 if (test_bit(HCI_MGMT, &hdev->dev_flags))
3887 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
3888 conn->dst_type, conn->passkey_notify,
3889 conn->passkey_entered);
3890 }
3891
3892 static void hci_simple_pair_complete_evt(struct hci_dev *hdev,
3893 struct sk_buff *skb)
3894 {
3895 struct hci_ev_simple_pair_complete *ev = (void *) skb->data;
3896 struct hci_conn *conn;
3897
3898 BT_DBG("%s", hdev->name);
3899
3900 hci_dev_lock(hdev);
3901
3902 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3903 if (!conn)
3904 goto unlock;
3905
3906 /* Reset the authentication requirement to unknown */
3907 conn->remote_auth = 0xff;
3908
3909 /* To avoid duplicate auth_failed events to user space we check
3910 * the HCI_CONN_AUTH_PEND flag which will be set if we
3911 * initiated the authentication. A traditional auth_complete
3912 * event gets always produced as initiator and is also mapped to
3913 * the mgmt_auth_failed event */
3914 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) && ev->status)
3915 mgmt_auth_failed(conn, ev->status);
3916
3917 hci_conn_drop(conn);
3918
3919 unlock:
3920 hci_dev_unlock(hdev);
3921 }
3922
3923 static void hci_remote_host_features_evt(struct hci_dev *hdev,
3924 struct sk_buff *skb)
3925 {
3926 struct hci_ev_remote_host_features *ev = (void *) skb->data;
3927 struct inquiry_entry *ie;
3928 struct hci_conn *conn;
3929
3930 BT_DBG("%s", hdev->name);
3931
3932 hci_dev_lock(hdev);
3933
3934 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3935 if (conn)
3936 memcpy(conn->features[1], ev->features, 8);
3937
3938 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3939 if (ie)
3940 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
3941
3942 hci_dev_unlock(hdev);
3943 }
3944
3945 static void hci_remote_oob_data_request_evt(struct hci_dev *hdev,
3946 struct sk_buff *skb)
3947 {
3948 struct hci_ev_remote_oob_data_request *ev = (void *) skb->data;
3949 struct oob_data *data;
3950
3951 BT_DBG("%s", hdev->name);
3952
3953 hci_dev_lock(hdev);
3954
3955 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
3956 goto unlock;
3957
3958 data = hci_find_remote_oob_data(hdev, &ev->bdaddr);
3959 if (data) {
3960 if (test_bit(HCI_SC_ENABLED, &hdev->dev_flags)) {
3961 struct hci_cp_remote_oob_ext_data_reply cp;
3962
3963 bacpy(&cp.bdaddr, &ev->bdaddr);
3964 memcpy(cp.hash192, data->hash192, sizeof(cp.hash192));
3965 memcpy(cp.randomizer192, data->randomizer192,
3966 sizeof(cp.randomizer192));
3967 memcpy(cp.hash256, data->hash256, sizeof(cp.hash256));
3968 memcpy(cp.randomizer256, data->randomizer256,
3969 sizeof(cp.randomizer256));
3970
3971 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_EXT_DATA_REPLY,
3972 sizeof(cp), &cp);
3973 } else {
3974 struct hci_cp_remote_oob_data_reply cp;
3975
3976 bacpy(&cp.bdaddr, &ev->bdaddr);
3977 memcpy(cp.hash, data->hash192, sizeof(cp.hash));
3978 memcpy(cp.randomizer, data->randomizer192,
3979 sizeof(cp.randomizer));
3980
3981 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_REPLY,
3982 sizeof(cp), &cp);
3983 }
3984 } else {
3985 struct hci_cp_remote_oob_data_neg_reply cp;
3986
3987 bacpy(&cp.bdaddr, &ev->bdaddr);
3988 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_NEG_REPLY,
3989 sizeof(cp), &cp);
3990 }
3991
3992 unlock:
3993 hci_dev_unlock(hdev);
3994 }
3995
3996 static void hci_phy_link_complete_evt(struct hci_dev *hdev,
3997 struct sk_buff *skb)
3998 {
3999 struct hci_ev_phy_link_complete *ev = (void *) skb->data;
4000 struct hci_conn *hcon, *bredr_hcon;
4001
4002 BT_DBG("%s handle 0x%2.2x status 0x%2.2x", hdev->name, ev->phy_handle,
4003 ev->status);
4004
4005 hci_dev_lock(hdev);
4006
4007 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4008 if (!hcon) {
4009 hci_dev_unlock(hdev);
4010 return;
4011 }
4012
4013 if (ev->status) {
4014 hci_conn_del(hcon);
4015 hci_dev_unlock(hdev);
4016 return;
4017 }
4018
4019 bredr_hcon = hcon->amp_mgr->l2cap_conn->hcon;
4020
4021 hcon->state = BT_CONNECTED;
4022 bacpy(&hcon->dst, &bredr_hcon->dst);
4023
4024 hci_conn_hold(hcon);
4025 hcon->disc_timeout = HCI_DISCONN_TIMEOUT;
4026 hci_conn_drop(hcon);
4027
4028 hci_conn_add_sysfs(hcon);
4029
4030 amp_physical_cfm(bredr_hcon, hcon);
4031
4032 hci_dev_unlock(hdev);
4033 }
4034
4035 static void hci_loglink_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4036 {
4037 struct hci_ev_logical_link_complete *ev = (void *) skb->data;
4038 struct hci_conn *hcon;
4039 struct hci_chan *hchan;
4040 struct amp_mgr *mgr;
4041
4042 BT_DBG("%s log_handle 0x%4.4x phy_handle 0x%2.2x status 0x%2.2x",
4043 hdev->name, le16_to_cpu(ev->handle), ev->phy_handle,
4044 ev->status);
4045
4046 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4047 if (!hcon)
4048 return;
4049
4050 /* Create AMP hchan */
4051 hchan = hci_chan_create(hcon);
4052 if (!hchan)
4053 return;
4054
4055 hchan->handle = le16_to_cpu(ev->handle);
4056
4057 BT_DBG("hcon %p mgr %p hchan %p", hcon, hcon->amp_mgr, hchan);
4058
4059 mgr = hcon->amp_mgr;
4060 if (mgr && mgr->bredr_chan) {
4061 struct l2cap_chan *bredr_chan = mgr->bredr_chan;
4062
4063 l2cap_chan_lock(bredr_chan);
4064
4065 bredr_chan->conn->mtu = hdev->block_mtu;
4066 l2cap_logical_cfm(bredr_chan, hchan, 0);
4067 hci_conn_hold(hcon);
4068
4069 l2cap_chan_unlock(bredr_chan);
4070 }
4071 }
4072
4073 static void hci_disconn_loglink_complete_evt(struct hci_dev *hdev,
4074 struct sk_buff *skb)
4075 {
4076 struct hci_ev_disconn_logical_link_complete *ev = (void *) skb->data;
4077 struct hci_chan *hchan;
4078
4079 BT_DBG("%s log handle 0x%4.4x status 0x%2.2x", hdev->name,
4080 le16_to_cpu(ev->handle), ev->status);
4081
4082 if (ev->status)
4083 return;
4084
4085 hci_dev_lock(hdev);
4086
4087 hchan = hci_chan_lookup_handle(hdev, le16_to_cpu(ev->handle));
4088 if (!hchan)
4089 goto unlock;
4090
4091 amp_destroy_logical_link(hchan, ev->reason);
4092
4093 unlock:
4094 hci_dev_unlock(hdev);
4095 }
4096
4097 static void hci_disconn_phylink_complete_evt(struct hci_dev *hdev,
4098 struct sk_buff *skb)
4099 {
4100 struct hci_ev_disconn_phy_link_complete *ev = (void *) skb->data;
4101 struct hci_conn *hcon;
4102
4103 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4104
4105 if (ev->status)
4106 return;
4107
4108 hci_dev_lock(hdev);
4109
4110 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4111 if (hcon) {
4112 hcon->state = BT_CLOSED;
4113 hci_conn_del(hcon);
4114 }
4115
4116 hci_dev_unlock(hdev);
4117 }
4118
4119 static void hci_le_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4120 {
4121 struct hci_ev_le_conn_complete *ev = (void *) skb->data;
4122 struct hci_conn_params *params;
4123 struct hci_conn *conn;
4124 struct smp_irk *irk;
4125 u8 addr_type;
4126
4127 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4128
4129 hci_dev_lock(hdev);
4130
4131 /* All controllers implicitly stop advertising in the event of a
4132 * connection, so ensure that the state bit is cleared.
4133 */
4134 clear_bit(HCI_LE_ADV, &hdev->dev_flags);
4135
4136 conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
4137 if (!conn) {
4138 conn = hci_conn_add(hdev, LE_LINK, &ev->bdaddr, ev->role);
4139 if (!conn) {
4140 BT_ERR("No memory for new connection");
4141 goto unlock;
4142 }
4143
4144 conn->dst_type = ev->bdaddr_type;
4145
4146 /* If we didn't have a hci_conn object previously
4147 * but we're in master role this must be something
4148 * initiated using a white list. Since white list based
4149 * connections are not "first class citizens" we don't
4150 * have full tracking of them. Therefore, we go ahead
4151 * with a "best effort" approach of determining the
4152 * initiator address based on the HCI_PRIVACY flag.
4153 */
4154 if (conn->out) {
4155 conn->resp_addr_type = ev->bdaddr_type;
4156 bacpy(&conn->resp_addr, &ev->bdaddr);
4157 if (test_bit(HCI_PRIVACY, &hdev->dev_flags)) {
4158 conn->init_addr_type = ADDR_LE_DEV_RANDOM;
4159 bacpy(&conn->init_addr, &hdev->rpa);
4160 } else {
4161 hci_copy_identity_address(hdev,
4162 &conn->init_addr,
4163 &conn->init_addr_type);
4164 }
4165 }
4166 } else {
4167 cancel_delayed_work(&conn->le_conn_timeout);
4168 }
4169
4170 if (!conn->out) {
4171 /* Set the responder (our side) address type based on
4172 * the advertising address type.
4173 */
4174 conn->resp_addr_type = hdev->adv_addr_type;
4175 if (hdev->adv_addr_type == ADDR_LE_DEV_RANDOM)
4176 bacpy(&conn->resp_addr, &hdev->random_addr);
4177 else
4178 bacpy(&conn->resp_addr, &hdev->bdaddr);
4179
4180 conn->init_addr_type = ev->bdaddr_type;
4181 bacpy(&conn->init_addr, &ev->bdaddr);
4182
4183 /* For incoming connections, set the default minimum
4184 * and maximum connection interval. They will be used
4185 * to check if the parameters are in range and if not
4186 * trigger the connection update procedure.
4187 */
4188 conn->le_conn_min_interval = hdev->le_conn_min_interval;
4189 conn->le_conn_max_interval = hdev->le_conn_max_interval;
4190 }
4191
4192 /* Lookup the identity address from the stored connection
4193 * address and address type.
4194 *
4195 * When establishing connections to an identity address, the
4196 * connection procedure will store the resolvable random
4197 * address first. Now if it can be converted back into the
4198 * identity address, start using the identity address from
4199 * now on.
4200 */
4201 irk = hci_get_irk(hdev, &conn->dst, conn->dst_type);
4202 if (irk) {
4203 bacpy(&conn->dst, &irk->bdaddr);
4204 conn->dst_type = irk->addr_type;
4205 }
4206
4207 if (ev->status) {
4208 hci_le_conn_failed(conn, ev->status);
4209 goto unlock;
4210 }
4211
4212 if (conn->dst_type == ADDR_LE_DEV_PUBLIC)
4213 addr_type = BDADDR_LE_PUBLIC;
4214 else
4215 addr_type = BDADDR_LE_RANDOM;
4216
4217 /* Drop the connection if the device is blocked */
4218 if (hci_bdaddr_list_lookup(&hdev->blacklist, &conn->dst, addr_type)) {
4219 hci_conn_drop(conn);
4220 goto unlock;
4221 }
4222
4223 if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4224 mgmt_device_connected(hdev, conn, 0, NULL, 0);
4225
4226 conn->sec_level = BT_SECURITY_LOW;
4227 conn->handle = __le16_to_cpu(ev->handle);
4228 conn->state = BT_CONNECTED;
4229
4230 conn->le_conn_interval = le16_to_cpu(ev->interval);
4231 conn->le_conn_latency = le16_to_cpu(ev->latency);
4232 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4233
4234 hci_conn_add_sysfs(conn);
4235
4236 hci_proto_connect_cfm(conn, ev->status);
4237
4238 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, &conn->dst,
4239 conn->dst_type);
4240 if (params) {
4241 list_del_init(&params->action);
4242 if (params->conn) {
4243 hci_conn_drop(params->conn);
4244 hci_conn_put(params->conn);
4245 params->conn = NULL;
4246 }
4247 }
4248
4249 unlock:
4250 hci_update_background_scan(hdev);
4251 hci_dev_unlock(hdev);
4252 }
4253
4254 static void hci_le_conn_update_complete_evt(struct hci_dev *hdev,
4255 struct sk_buff *skb)
4256 {
4257 struct hci_ev_le_conn_update_complete *ev = (void *) skb->data;
4258 struct hci_conn *conn;
4259
4260 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4261
4262 if (ev->status)
4263 return;
4264
4265 hci_dev_lock(hdev);
4266
4267 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4268 if (conn) {
4269 conn->le_conn_interval = le16_to_cpu(ev->interval);
4270 conn->le_conn_latency = le16_to_cpu(ev->latency);
4271 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
4272 }
4273
4274 hci_dev_unlock(hdev);
4275 }
4276
4277 /* This function requires the caller holds hdev->lock */
4278 static struct hci_conn *check_pending_le_conn(struct hci_dev *hdev,
4279 bdaddr_t *addr,
4280 u8 addr_type, u8 adv_type)
4281 {
4282 struct hci_conn *conn;
4283 struct hci_conn_params *params;
4284
4285 /* If the event is not connectable don't proceed further */
4286 if (adv_type != LE_ADV_IND && adv_type != LE_ADV_DIRECT_IND)
4287 return NULL;
4288
4289 /* Ignore if the device is blocked */
4290 if (hci_bdaddr_list_lookup(&hdev->blacklist, addr, addr_type))
4291 return NULL;
4292
4293 /* Most controller will fail if we try to create new connections
4294 * while we have an existing one in slave role.
4295 */
4296 if (hdev->conn_hash.le_num_slave > 0)
4297 return NULL;
4298
4299 /* If we're not connectable only connect devices that we have in
4300 * our pend_le_conns list.
4301 */
4302 params = hci_pend_le_action_lookup(&hdev->pend_le_conns,
4303 addr, addr_type);
4304 if (!params)
4305 return NULL;
4306
4307 switch (params->auto_connect) {
4308 case HCI_AUTO_CONN_DIRECT:
4309 /* Only devices advertising with ADV_DIRECT_IND are
4310 * triggering a connection attempt. This is allowing
4311 * incoming connections from slave devices.
4312 */
4313 if (adv_type != LE_ADV_DIRECT_IND)
4314 return NULL;
4315 break;
4316 case HCI_AUTO_CONN_ALWAYS:
4317 /* Devices advertising with ADV_IND or ADV_DIRECT_IND
4318 * are triggering a connection attempt. This means
4319 * that incoming connectioms from slave device are
4320 * accepted and also outgoing connections to slave
4321 * devices are established when found.
4322 */
4323 break;
4324 default:
4325 return NULL;
4326 }
4327
4328 conn = hci_connect_le(hdev, addr, addr_type, BT_SECURITY_LOW,
4329 HCI_LE_AUTOCONN_TIMEOUT, HCI_ROLE_MASTER);
4330 if (!IS_ERR(conn)) {
4331 /* Store the pointer since we don't really have any
4332 * other owner of the object besides the params that
4333 * triggered it. This way we can abort the connection if
4334 * the parameters get removed and keep the reference
4335 * count consistent once the connection is established.
4336 */
4337 params->conn = hci_conn_get(conn);
4338 return conn;
4339 }
4340
4341 switch (PTR_ERR(conn)) {
4342 case -EBUSY:
4343 /* If hci_connect() returns -EBUSY it means there is already
4344 * an LE connection attempt going on. Since controllers don't
4345 * support more than one connection attempt at the time, we
4346 * don't consider this an error case.
4347 */
4348 break;
4349 default:
4350 BT_DBG("Failed to connect: err %ld", PTR_ERR(conn));
4351 return NULL;
4352 }
4353
4354 return NULL;
4355 }
4356
4357 static void process_adv_report(struct hci_dev *hdev, u8 type, bdaddr_t *bdaddr,
4358 u8 bdaddr_type, s8 rssi, u8 *data, u8 len)
4359 {
4360 struct discovery_state *d = &hdev->discovery;
4361 struct smp_irk *irk;
4362 struct hci_conn *conn;
4363 bool match;
4364 u32 flags;
4365
4366 /* Check if we need to convert to identity address */
4367 irk = hci_get_irk(hdev, bdaddr, bdaddr_type);
4368 if (irk) {
4369 bdaddr = &irk->bdaddr;
4370 bdaddr_type = irk->addr_type;
4371 }
4372
4373 /* Check if we have been requested to connect to this device */
4374 conn = check_pending_le_conn(hdev, bdaddr, bdaddr_type, type);
4375 if (conn && type == LE_ADV_IND) {
4376 /* Store report for later inclusion by
4377 * mgmt_device_connected
4378 */
4379 memcpy(conn->le_adv_data, data, len);
4380 conn->le_adv_data_len = len;
4381 }
4382
4383 /* Passive scanning shouldn't trigger any device found events,
4384 * except for devices marked as CONN_REPORT for which we do send
4385 * device found events.
4386 */
4387 if (hdev->le_scan_type == LE_SCAN_PASSIVE) {
4388 if (type == LE_ADV_DIRECT_IND)
4389 return;
4390
4391 if (!hci_pend_le_action_lookup(&hdev->pend_le_reports,
4392 bdaddr, bdaddr_type))
4393 return;
4394
4395 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND)
4396 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4397 else
4398 flags = 0;
4399 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4400 rssi, flags, data, len, NULL, 0);
4401 return;
4402 }
4403
4404 /* When receiving non-connectable or scannable undirected
4405 * advertising reports, this means that the remote device is
4406 * not connectable and then clearly indicate this in the
4407 * device found event.
4408 *
4409 * When receiving a scan response, then there is no way to
4410 * know if the remote device is connectable or not. However
4411 * since scan responses are merged with a previously seen
4412 * advertising report, the flags field from that report
4413 * will be used.
4414 *
4415 * In the really unlikely case that a controller get confused
4416 * and just sends a scan response event, then it is marked as
4417 * not connectable as well.
4418 */
4419 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND ||
4420 type == LE_ADV_SCAN_RSP)
4421 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
4422 else
4423 flags = 0;
4424
4425 /* If there's nothing pending either store the data from this
4426 * event or send an immediate device found event if the data
4427 * should not be stored for later.
4428 */
4429 if (!has_pending_adv_report(hdev)) {
4430 /* If the report will trigger a SCAN_REQ store it for
4431 * later merging.
4432 */
4433 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4434 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4435 rssi, flags, data, len);
4436 return;
4437 }
4438
4439 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4440 rssi, flags, data, len, NULL, 0);
4441 return;
4442 }
4443
4444 /* Check if the pending report is for the same device as the new one */
4445 match = (!bacmp(bdaddr, &d->last_adv_addr) &&
4446 bdaddr_type == d->last_adv_addr_type);
4447
4448 /* If the pending data doesn't match this report or this isn't a
4449 * scan response (e.g. we got a duplicate ADV_IND) then force
4450 * sending of the pending data.
4451 */
4452 if (type != LE_ADV_SCAN_RSP || !match) {
4453 /* Send out whatever is in the cache, but skip duplicates */
4454 if (!match)
4455 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4456 d->last_adv_addr_type, NULL,
4457 d->last_adv_rssi, d->last_adv_flags,
4458 d->last_adv_data,
4459 d->last_adv_data_len, NULL, 0);
4460
4461 /* If the new report will trigger a SCAN_REQ store it for
4462 * later merging.
4463 */
4464 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
4465 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
4466 rssi, flags, data, len);
4467 return;
4468 }
4469
4470 /* The advertising reports cannot be merged, so clear
4471 * the pending report and send out a device found event.
4472 */
4473 clear_pending_adv_report(hdev);
4474 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
4475 rssi, flags, data, len, NULL, 0);
4476 return;
4477 }
4478
4479 /* If we get here we've got a pending ADV_IND or ADV_SCAN_IND and
4480 * the new event is a SCAN_RSP. We can therefore proceed with
4481 * sending a merged device found event.
4482 */
4483 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
4484 d->last_adv_addr_type, NULL, rssi, d->last_adv_flags,
4485 d->last_adv_data, d->last_adv_data_len, data, len);
4486 clear_pending_adv_report(hdev);
4487 }
4488
4489 static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
4490 {
4491 u8 num_reports = skb->data[0];
4492 void *ptr = &skb->data[1];
4493
4494 hci_dev_lock(hdev);
4495
4496 while (num_reports--) {
4497 struct hci_ev_le_advertising_info *ev = ptr;
4498 s8 rssi;
4499
4500 rssi = ev->data[ev->length];
4501 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
4502 ev->bdaddr_type, rssi, ev->data, ev->length);
4503
4504 ptr += sizeof(*ev) + ev->length + 1;
4505 }
4506
4507 hci_dev_unlock(hdev);
4508 }
4509
4510 static void hci_le_ltk_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
4511 {
4512 struct hci_ev_le_ltk_req *ev = (void *) skb->data;
4513 struct hci_cp_le_ltk_reply cp;
4514 struct hci_cp_le_ltk_neg_reply neg;
4515 struct hci_conn *conn;
4516 struct smp_ltk *ltk;
4517
4518 BT_DBG("%s handle 0x%4.4x", hdev->name, __le16_to_cpu(ev->handle));
4519
4520 hci_dev_lock(hdev);
4521
4522 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4523 if (conn == NULL)
4524 goto not_found;
4525
4526 ltk = hci_find_ltk(hdev, ev->ediv, ev->rand, conn->role);
4527 if (ltk == NULL)
4528 goto not_found;
4529
4530 memcpy(cp.ltk, ltk->val, sizeof(ltk->val));
4531 cp.handle = cpu_to_le16(conn->handle);
4532
4533 conn->pending_sec_level = smp_ltk_sec_level(ltk);
4534
4535 conn->enc_key_size = ltk->enc_size;
4536
4537 hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp);
4538
4539 /* Ref. Bluetooth Core SPEC pages 1975 and 2004. STK is a
4540 * temporary key used to encrypt a connection following
4541 * pairing. It is used during the Encrypted Session Setup to
4542 * distribute the keys. Later, security can be re-established
4543 * using a distributed LTK.
4544 */
4545 if (ltk->type == SMP_STK) {
4546 set_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4547 list_del(&ltk->list);
4548 kfree(ltk);
4549 } else {
4550 clear_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
4551 }
4552
4553 hci_dev_unlock(hdev);
4554
4555 return;
4556
4557 not_found:
4558 neg.handle = ev->handle;
4559 hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(neg), &neg);
4560 hci_dev_unlock(hdev);
4561 }
4562
4563 static void send_conn_param_neg_reply(struct hci_dev *hdev, u16 handle,
4564 u8 reason)
4565 {
4566 struct hci_cp_le_conn_param_req_neg_reply cp;
4567
4568 cp.handle = cpu_to_le16(handle);
4569 cp.reason = reason;
4570
4571 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_NEG_REPLY, sizeof(cp),
4572 &cp);
4573 }
4574
4575 static void hci_le_remote_conn_param_req_evt(struct hci_dev *hdev,
4576 struct sk_buff *skb)
4577 {
4578 struct hci_ev_le_remote_conn_param_req *ev = (void *) skb->data;
4579 struct hci_cp_le_conn_param_req_reply cp;
4580 struct hci_conn *hcon;
4581 u16 handle, min, max, latency, timeout;
4582
4583 handle = le16_to_cpu(ev->handle);
4584 min = le16_to_cpu(ev->interval_min);
4585 max = le16_to_cpu(ev->interval_max);
4586 latency = le16_to_cpu(ev->latency);
4587 timeout = le16_to_cpu(ev->timeout);
4588
4589 hcon = hci_conn_hash_lookup_handle(hdev, handle);
4590 if (!hcon || hcon->state != BT_CONNECTED)
4591 return send_conn_param_neg_reply(hdev, handle,
4592 HCI_ERROR_UNKNOWN_CONN_ID);
4593
4594 if (hci_check_conn_params(min, max, latency, timeout))
4595 return send_conn_param_neg_reply(hdev, handle,
4596 HCI_ERROR_INVALID_LL_PARAMS);
4597
4598 if (hcon->role == HCI_ROLE_MASTER) {
4599 struct hci_conn_params *params;
4600 u8 store_hint;
4601
4602 hci_dev_lock(hdev);
4603
4604 params = hci_conn_params_lookup(hdev, &hcon->dst,
4605 hcon->dst_type);
4606 if (params) {
4607 params->conn_min_interval = min;
4608 params->conn_max_interval = max;
4609 params->conn_latency = latency;
4610 params->supervision_timeout = timeout;
4611 store_hint = 0x01;
4612 } else{
4613 store_hint = 0x00;
4614 }
4615
4616 hci_dev_unlock(hdev);
4617
4618 mgmt_new_conn_param(hdev, &hcon->dst, hcon->dst_type,
4619 store_hint, min, max, latency, timeout);
4620 }
4621
4622 cp.handle = ev->handle;
4623 cp.interval_min = ev->interval_min;
4624 cp.interval_max = ev->interval_max;
4625 cp.latency = ev->latency;
4626 cp.timeout = ev->timeout;
4627 cp.min_ce_len = 0;
4628 cp.max_ce_len = 0;
4629
4630 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_REPLY, sizeof(cp), &cp);
4631 }
4632
4633 static void hci_le_meta_evt(struct hci_dev *hdev, struct sk_buff *skb)
4634 {
4635 struct hci_ev_le_meta *le_ev = (void *) skb->data;
4636
4637 skb_pull(skb, sizeof(*le_ev));
4638
4639 switch (le_ev->subevent) {
4640 case HCI_EV_LE_CONN_COMPLETE:
4641 hci_le_conn_complete_evt(hdev, skb);
4642 break;
4643
4644 case HCI_EV_LE_CONN_UPDATE_COMPLETE:
4645 hci_le_conn_update_complete_evt(hdev, skb);
4646 break;
4647
4648 case HCI_EV_LE_ADVERTISING_REPORT:
4649 hci_le_adv_report_evt(hdev, skb);
4650 break;
4651
4652 case HCI_EV_LE_LTK_REQ:
4653 hci_le_ltk_request_evt(hdev, skb);
4654 break;
4655
4656 case HCI_EV_LE_REMOTE_CONN_PARAM_REQ:
4657 hci_le_remote_conn_param_req_evt(hdev, skb);
4658 break;
4659
4660 default:
4661 break;
4662 }
4663 }
4664
4665 static void hci_chan_selected_evt(struct hci_dev *hdev, struct sk_buff *skb)
4666 {
4667 struct hci_ev_channel_selected *ev = (void *) skb->data;
4668 struct hci_conn *hcon;
4669
4670 BT_DBG("%s handle 0x%2.2x", hdev->name, ev->phy_handle);
4671
4672 skb_pull(skb, sizeof(*ev));
4673
4674 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4675 if (!hcon)
4676 return;
4677
4678 amp_read_loc_assoc_final_data(hdev, hcon);
4679 }
4680
4681 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb)
4682 {
4683 struct hci_event_hdr *hdr = (void *) skb->data;
4684 __u8 event = hdr->evt;
4685
4686 hci_dev_lock(hdev);
4687
4688 /* Received events are (currently) only needed when a request is
4689 * ongoing so avoid unnecessary memory allocation.
4690 */
4691 if (hci_req_pending(hdev)) {
4692 kfree_skb(hdev->recv_evt);
4693 hdev->recv_evt = skb_clone(skb, GFP_KERNEL);
4694 }
4695
4696 hci_dev_unlock(hdev);
4697
4698 skb_pull(skb, HCI_EVENT_HDR_SIZE);
4699
4700 if (hdev->sent_cmd && bt_cb(hdev->sent_cmd)->req.event == event) {
4701 struct hci_command_hdr *cmd_hdr = (void *) hdev->sent_cmd->data;
4702 u16 opcode = __le16_to_cpu(cmd_hdr->opcode);
4703
4704 hci_req_cmd_complete(hdev, opcode, 0);
4705 }
4706
4707 switch (event) {
4708 case HCI_EV_INQUIRY_COMPLETE:
4709 hci_inquiry_complete_evt(hdev, skb);
4710 break;
4711
4712 case HCI_EV_INQUIRY_RESULT:
4713 hci_inquiry_result_evt(hdev, skb);
4714 break;
4715
4716 case HCI_EV_CONN_COMPLETE:
4717 hci_conn_complete_evt(hdev, skb);
4718 break;
4719
4720 case HCI_EV_CONN_REQUEST:
4721 hci_conn_request_evt(hdev, skb);
4722 break;
4723
4724 case HCI_EV_DISCONN_COMPLETE:
4725 hci_disconn_complete_evt(hdev, skb);
4726 break;
4727
4728 case HCI_EV_AUTH_COMPLETE:
4729 hci_auth_complete_evt(hdev, skb);
4730 break;
4731
4732 case HCI_EV_REMOTE_NAME:
4733 hci_remote_name_evt(hdev, skb);
4734 break;
4735
4736 case HCI_EV_ENCRYPT_CHANGE:
4737 hci_encrypt_change_evt(hdev, skb);
4738 break;
4739
4740 case HCI_EV_CHANGE_LINK_KEY_COMPLETE:
4741 hci_change_link_key_complete_evt(hdev, skb);
4742 break;
4743
4744 case HCI_EV_REMOTE_FEATURES:
4745 hci_remote_features_evt(hdev, skb);
4746 break;
4747
4748 case HCI_EV_CMD_COMPLETE:
4749 hci_cmd_complete_evt(hdev, skb);
4750 break;
4751
4752 case HCI_EV_CMD_STATUS:
4753 hci_cmd_status_evt(hdev, skb);
4754 break;
4755
4756 case HCI_EV_HARDWARE_ERROR:
4757 hci_hardware_error_evt(hdev, skb);
4758 break;
4759
4760 case HCI_EV_ROLE_CHANGE:
4761 hci_role_change_evt(hdev, skb);
4762 break;
4763
4764 case HCI_EV_NUM_COMP_PKTS:
4765 hci_num_comp_pkts_evt(hdev, skb);
4766 break;
4767
4768 case HCI_EV_MODE_CHANGE:
4769 hci_mode_change_evt(hdev, skb);
4770 break;
4771
4772 case HCI_EV_PIN_CODE_REQ:
4773 hci_pin_code_request_evt(hdev, skb);
4774 break;
4775
4776 case HCI_EV_LINK_KEY_REQ:
4777 hci_link_key_request_evt(hdev, skb);
4778 break;
4779
4780 case HCI_EV_LINK_KEY_NOTIFY:
4781 hci_link_key_notify_evt(hdev, skb);
4782 break;
4783
4784 case HCI_EV_CLOCK_OFFSET:
4785 hci_clock_offset_evt(hdev, skb);
4786 break;
4787
4788 case HCI_EV_PKT_TYPE_CHANGE:
4789 hci_pkt_type_change_evt(hdev, skb);
4790 break;
4791
4792 case HCI_EV_PSCAN_REP_MODE:
4793 hci_pscan_rep_mode_evt(hdev, skb);
4794 break;
4795
4796 case HCI_EV_INQUIRY_RESULT_WITH_RSSI:
4797 hci_inquiry_result_with_rssi_evt(hdev, skb);
4798 break;
4799
4800 case HCI_EV_REMOTE_EXT_FEATURES:
4801 hci_remote_ext_features_evt(hdev, skb);
4802 break;
4803
4804 case HCI_EV_SYNC_CONN_COMPLETE:
4805 hci_sync_conn_complete_evt(hdev, skb);
4806 break;
4807
4808 case HCI_EV_EXTENDED_INQUIRY_RESULT:
4809 hci_extended_inquiry_result_evt(hdev, skb);
4810 break;
4811
4812 case HCI_EV_KEY_REFRESH_COMPLETE:
4813 hci_key_refresh_complete_evt(hdev, skb);
4814 break;
4815
4816 case HCI_EV_IO_CAPA_REQUEST:
4817 hci_io_capa_request_evt(hdev, skb);
4818 break;
4819
4820 case HCI_EV_IO_CAPA_REPLY:
4821 hci_io_capa_reply_evt(hdev, skb);
4822 break;
4823
4824 case HCI_EV_USER_CONFIRM_REQUEST:
4825 hci_user_confirm_request_evt(hdev, skb);
4826 break;
4827
4828 case HCI_EV_USER_PASSKEY_REQUEST:
4829 hci_user_passkey_request_evt(hdev, skb);
4830 break;
4831
4832 case HCI_EV_USER_PASSKEY_NOTIFY:
4833 hci_user_passkey_notify_evt(hdev, skb);
4834 break;
4835
4836 case HCI_EV_KEYPRESS_NOTIFY:
4837 hci_keypress_notify_evt(hdev, skb);
4838 break;
4839
4840 case HCI_EV_SIMPLE_PAIR_COMPLETE:
4841 hci_simple_pair_complete_evt(hdev, skb);
4842 break;
4843
4844 case HCI_EV_REMOTE_HOST_FEATURES:
4845 hci_remote_host_features_evt(hdev, skb);
4846 break;
4847
4848 case HCI_EV_LE_META:
4849 hci_le_meta_evt(hdev, skb);
4850 break;
4851
4852 case HCI_EV_CHANNEL_SELECTED:
4853 hci_chan_selected_evt(hdev, skb);
4854 break;
4855
4856 case HCI_EV_REMOTE_OOB_DATA_REQUEST:
4857 hci_remote_oob_data_request_evt(hdev, skb);
4858 break;
4859
4860 case HCI_EV_PHY_LINK_COMPLETE:
4861 hci_phy_link_complete_evt(hdev, skb);
4862 break;
4863
4864 case HCI_EV_LOGICAL_LINK_COMPLETE:
4865 hci_loglink_complete_evt(hdev, skb);
4866 break;
4867
4868 case HCI_EV_DISCONN_LOGICAL_LINK_COMPLETE:
4869 hci_disconn_loglink_complete_evt(hdev, skb);
4870 break;
4871
4872 case HCI_EV_DISCONN_PHY_LINK_COMPLETE:
4873 hci_disconn_phylink_complete_evt(hdev, skb);
4874 break;
4875
4876 case HCI_EV_NUM_COMP_BLOCKS:
4877 hci_num_comp_blocks_evt(hdev, skb);
4878 break;
4879
4880 default:
4881 BT_DBG("%s event 0x%2.2x", hdev->name, event);
4882 break;
4883 }
4884
4885 kfree_skb(skb);
4886 hdev->stat.evt_rx++;
4887 }
This page took 0.122294 seconds and 4 git commands to generate.