2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
4 Copyright (C) 2011 ProFUSION Embedded Systems
6 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License version 2 as
10 published by the Free Software Foundation;
12 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
13 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
14 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
15 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
16 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
17 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
18 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
19 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
21 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
22 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
23 SOFTWARE IS DISCLAIMED.
26 /* Bluetooth HCI core. */
28 #include <linux/export.h>
29 #include <linux/idr.h>
31 #include <linux/rfkill.h>
33 #include <net/bluetooth/bluetooth.h>
34 #include <net/bluetooth/hci_core.h>
36 static void hci_rx_work(struct work_struct
*work
);
37 static void hci_cmd_work(struct work_struct
*work
);
38 static void hci_tx_work(struct work_struct
*work
);
41 LIST_HEAD(hci_dev_list
);
42 DEFINE_RWLOCK(hci_dev_list_lock
);
44 /* HCI callback list */
45 LIST_HEAD(hci_cb_list
);
46 DEFINE_RWLOCK(hci_cb_list_lock
);
48 /* HCI ID Numbering */
49 static DEFINE_IDA(hci_index_ida
);
51 /* ---- HCI notifications ---- */
53 static void hci_notify(struct hci_dev
*hdev
, int event
)
55 hci_sock_dev_event(hdev
, event
);
58 /* ---- HCI requests ---- */
60 static void hci_req_sync_complete(struct hci_dev
*hdev
, u8 result
)
62 BT_DBG("%s result 0x%2.2x", hdev
->name
, result
);
64 if (hdev
->req_status
== HCI_REQ_PEND
) {
65 hdev
->req_result
= result
;
66 hdev
->req_status
= HCI_REQ_DONE
;
67 wake_up_interruptible(&hdev
->req_wait_q
);
71 static void hci_req_cancel(struct hci_dev
*hdev
, int err
)
73 BT_DBG("%s err 0x%2.2x", hdev
->name
, err
);
75 if (hdev
->req_status
== HCI_REQ_PEND
) {
76 hdev
->req_result
= err
;
77 hdev
->req_status
= HCI_REQ_CANCELED
;
78 wake_up_interruptible(&hdev
->req_wait_q
);
82 static struct sk_buff
*hci_get_cmd_complete(struct hci_dev
*hdev
, u16 opcode
,
85 struct hci_ev_cmd_complete
*ev
;
86 struct hci_event_hdr
*hdr
;
92 hdev
->recv_evt
= NULL
;
97 return ERR_PTR(-ENODATA
);
99 if (skb
->len
< sizeof(*hdr
)) {
100 BT_ERR("Too short HCI event");
104 hdr
= (void *) skb
->data
;
105 skb_pull(skb
, HCI_EVENT_HDR_SIZE
);
108 if (hdr
->evt
!= event
)
113 if (hdr
->evt
!= HCI_EV_CMD_COMPLETE
) {
114 BT_DBG("Last event is not cmd complete (0x%2.2x)", hdr
->evt
);
118 if (skb
->len
< sizeof(*ev
)) {
119 BT_ERR("Too short cmd_complete event");
123 ev
= (void *) skb
->data
;
124 skb_pull(skb
, sizeof(*ev
));
126 if (opcode
== __le16_to_cpu(ev
->opcode
))
129 BT_DBG("opcode doesn't match (0x%2.2x != 0x%2.2x)", opcode
,
130 __le16_to_cpu(ev
->opcode
));
134 return ERR_PTR(-ENODATA
);
137 struct sk_buff
*__hci_cmd_sync_ev(struct hci_dev
*hdev
, u16 opcode
, u32 plen
,
138 const void *param
, u8 event
, u32 timeout
)
140 DECLARE_WAITQUEUE(wait
, current
);
141 struct hci_request req
;
144 BT_DBG("%s", hdev
->name
);
146 hci_req_init(&req
, hdev
);
148 hci_req_add_ev(&req
, opcode
, plen
, param
, event
);
150 hdev
->req_status
= HCI_REQ_PEND
;
152 err
= hci_req_run(&req
, hci_req_sync_complete
);
156 add_wait_queue(&hdev
->req_wait_q
, &wait
);
157 set_current_state(TASK_INTERRUPTIBLE
);
159 schedule_timeout(timeout
);
161 remove_wait_queue(&hdev
->req_wait_q
, &wait
);
163 if (signal_pending(current
))
164 return ERR_PTR(-EINTR
);
166 switch (hdev
->req_status
) {
168 err
= -bt_to_errno(hdev
->req_result
);
171 case HCI_REQ_CANCELED
:
172 err
= -hdev
->req_result
;
180 hdev
->req_status
= hdev
->req_result
= 0;
182 BT_DBG("%s end: err %d", hdev
->name
, err
);
187 return hci_get_cmd_complete(hdev
, opcode
, event
);
189 EXPORT_SYMBOL(__hci_cmd_sync_ev
);
191 struct sk_buff
*__hci_cmd_sync(struct hci_dev
*hdev
, u16 opcode
, u32 plen
,
192 const void *param
, u32 timeout
)
194 return __hci_cmd_sync_ev(hdev
, opcode
, plen
, param
, 0, timeout
);
196 EXPORT_SYMBOL(__hci_cmd_sync
);
198 /* Execute request and wait for completion. */
199 static int __hci_req_sync(struct hci_dev
*hdev
,
200 void (*func
)(struct hci_request
*req
,
202 unsigned long opt
, __u32 timeout
)
204 struct hci_request req
;
205 DECLARE_WAITQUEUE(wait
, current
);
208 BT_DBG("%s start", hdev
->name
);
210 hci_req_init(&req
, hdev
);
212 hdev
->req_status
= HCI_REQ_PEND
;
216 err
= hci_req_run(&req
, hci_req_sync_complete
);
218 hdev
->req_status
= 0;
220 /* ENODATA means the HCI request command queue is empty.
221 * This can happen when a request with conditionals doesn't
222 * trigger any commands to be sent. This is normal behavior
223 * and should not trigger an error return.
231 add_wait_queue(&hdev
->req_wait_q
, &wait
);
232 set_current_state(TASK_INTERRUPTIBLE
);
234 schedule_timeout(timeout
);
236 remove_wait_queue(&hdev
->req_wait_q
, &wait
);
238 if (signal_pending(current
))
241 switch (hdev
->req_status
) {
243 err
= -bt_to_errno(hdev
->req_result
);
246 case HCI_REQ_CANCELED
:
247 err
= -hdev
->req_result
;
255 hdev
->req_status
= hdev
->req_result
= 0;
257 BT_DBG("%s end: err %d", hdev
->name
, err
);
262 static int hci_req_sync(struct hci_dev
*hdev
,
263 void (*req
)(struct hci_request
*req
,
265 unsigned long opt
, __u32 timeout
)
269 if (!test_bit(HCI_UP
, &hdev
->flags
))
272 /* Serialize all requests */
274 ret
= __hci_req_sync(hdev
, req
, opt
, timeout
);
275 hci_req_unlock(hdev
);
280 static void hci_reset_req(struct hci_request
*req
, unsigned long opt
)
282 BT_DBG("%s %ld", req
->hdev
->name
, opt
);
285 set_bit(HCI_RESET
, &req
->hdev
->flags
);
286 hci_req_add(req
, HCI_OP_RESET
, 0, NULL
);
289 static void bredr_init(struct hci_request
*req
)
291 req
->hdev
->flow_ctl_mode
= HCI_FLOW_CTL_MODE_PACKET_BASED
;
293 /* Read Local Supported Features */
294 hci_req_add(req
, HCI_OP_READ_LOCAL_FEATURES
, 0, NULL
);
296 /* Read Local Version */
297 hci_req_add(req
, HCI_OP_READ_LOCAL_VERSION
, 0, NULL
);
299 /* Read BD Address */
300 hci_req_add(req
, HCI_OP_READ_BD_ADDR
, 0, NULL
);
303 static void amp_init(struct hci_request
*req
)
305 req
->hdev
->flow_ctl_mode
= HCI_FLOW_CTL_MODE_BLOCK_BASED
;
307 /* Read Local Version */
308 hci_req_add(req
, HCI_OP_READ_LOCAL_VERSION
, 0, NULL
);
310 /* Read Local AMP Info */
311 hci_req_add(req
, HCI_OP_READ_LOCAL_AMP_INFO
, 0, NULL
);
313 /* Read Data Blk size */
314 hci_req_add(req
, HCI_OP_READ_DATA_BLOCK_SIZE
, 0, NULL
);
317 static void hci_init1_req(struct hci_request
*req
, unsigned long opt
)
319 struct hci_dev
*hdev
= req
->hdev
;
321 BT_DBG("%s %ld", hdev
->name
, opt
);
324 if (!test_bit(HCI_QUIRK_RESET_ON_CLOSE
, &hdev
->quirks
))
325 hci_reset_req(req
, 0);
327 switch (hdev
->dev_type
) {
337 BT_ERR("Unknown device type %d", hdev
->dev_type
);
342 static void bredr_setup(struct hci_request
*req
)
347 /* Read Buffer Size (ACL mtu, max pkt, etc.) */
348 hci_req_add(req
, HCI_OP_READ_BUFFER_SIZE
, 0, NULL
);
350 /* Read Class of Device */
351 hci_req_add(req
, HCI_OP_READ_CLASS_OF_DEV
, 0, NULL
);
353 /* Read Local Name */
354 hci_req_add(req
, HCI_OP_READ_LOCAL_NAME
, 0, NULL
);
356 /* Read Voice Setting */
357 hci_req_add(req
, HCI_OP_READ_VOICE_SETTING
, 0, NULL
);
359 /* Clear Event Filters */
360 flt_type
= HCI_FLT_CLEAR_ALL
;
361 hci_req_add(req
, HCI_OP_SET_EVENT_FLT
, 1, &flt_type
);
363 /* Connection accept timeout ~20 secs */
364 param
= __constant_cpu_to_le16(0x7d00);
365 hci_req_add(req
, HCI_OP_WRITE_CA_TIMEOUT
, 2, ¶m
);
367 /* Read page scan parameters */
368 if (req
->hdev
->hci_ver
> BLUETOOTH_VER_1_1
) {
369 hci_req_add(req
, HCI_OP_READ_PAGE_SCAN_ACTIVITY
, 0, NULL
);
370 hci_req_add(req
, HCI_OP_READ_PAGE_SCAN_TYPE
, 0, NULL
);
374 static void le_setup(struct hci_request
*req
)
376 struct hci_dev
*hdev
= req
->hdev
;
378 /* Read LE Buffer Size */
379 hci_req_add(req
, HCI_OP_LE_READ_BUFFER_SIZE
, 0, NULL
);
381 /* Read LE Local Supported Features */
382 hci_req_add(req
, HCI_OP_LE_READ_LOCAL_FEATURES
, 0, NULL
);
384 /* Read LE Advertising Channel TX Power */
385 hci_req_add(req
, HCI_OP_LE_READ_ADV_TX_POWER
, 0, NULL
);
387 /* Read LE White List Size */
388 hci_req_add(req
, HCI_OP_LE_READ_WHITE_LIST_SIZE
, 0, NULL
);
390 /* Read LE Supported States */
391 hci_req_add(req
, HCI_OP_LE_READ_SUPPORTED_STATES
, 0, NULL
);
393 /* LE-only controllers have LE implicitly enabled */
394 if (!lmp_bredr_capable(hdev
))
395 set_bit(HCI_LE_ENABLED
, &hdev
->dev_flags
);
398 static u8
hci_get_inquiry_mode(struct hci_dev
*hdev
)
400 if (lmp_ext_inq_capable(hdev
))
403 if (lmp_inq_rssi_capable(hdev
))
406 if (hdev
->manufacturer
== 11 && hdev
->hci_rev
== 0x00 &&
407 hdev
->lmp_subver
== 0x0757)
410 if (hdev
->manufacturer
== 15) {
411 if (hdev
->hci_rev
== 0x03 && hdev
->lmp_subver
== 0x6963)
413 if (hdev
->hci_rev
== 0x09 && hdev
->lmp_subver
== 0x6963)
415 if (hdev
->hci_rev
== 0x00 && hdev
->lmp_subver
== 0x6965)
419 if (hdev
->manufacturer
== 31 && hdev
->hci_rev
== 0x2005 &&
420 hdev
->lmp_subver
== 0x1805)
426 static void hci_setup_inquiry_mode(struct hci_request
*req
)
430 mode
= hci_get_inquiry_mode(req
->hdev
);
432 hci_req_add(req
, HCI_OP_WRITE_INQUIRY_MODE
, 1, &mode
);
435 static void hci_setup_event_mask(struct hci_request
*req
)
437 struct hci_dev
*hdev
= req
->hdev
;
439 /* The second byte is 0xff instead of 0x9f (two reserved bits
440 * disabled) since a Broadcom 1.2 dongle doesn't respond to the
443 u8 events
[8] = { 0xff, 0xff, 0xfb, 0xff, 0x00, 0x00, 0x00, 0x00 };
445 /* CSR 1.1 dongles does not accept any bitfield so don't try to set
446 * any event mask for pre 1.2 devices.
448 if (hdev
->hci_ver
< BLUETOOTH_VER_1_2
)
451 if (lmp_bredr_capable(hdev
)) {
452 events
[4] |= 0x01; /* Flow Specification Complete */
453 events
[4] |= 0x02; /* Inquiry Result with RSSI */
454 events
[4] |= 0x04; /* Read Remote Extended Features Complete */
455 events
[5] |= 0x08; /* Synchronous Connection Complete */
456 events
[5] |= 0x10; /* Synchronous Connection Changed */
458 /* Use a different default for LE-only devices */
459 memset(events
, 0, sizeof(events
));
460 events
[0] |= 0x10; /* Disconnection Complete */
461 events
[0] |= 0x80; /* Encryption Change */
462 events
[1] |= 0x08; /* Read Remote Version Information Complete */
463 events
[1] |= 0x20; /* Command Complete */
464 events
[1] |= 0x40; /* Command Status */
465 events
[1] |= 0x80; /* Hardware Error */
466 events
[2] |= 0x04; /* Number of Completed Packets */
467 events
[3] |= 0x02; /* Data Buffer Overflow */
468 events
[5] |= 0x80; /* Encryption Key Refresh Complete */
471 if (lmp_inq_rssi_capable(hdev
))
472 events
[4] |= 0x02; /* Inquiry Result with RSSI */
474 if (lmp_sniffsubr_capable(hdev
))
475 events
[5] |= 0x20; /* Sniff Subrating */
477 if (lmp_pause_enc_capable(hdev
))
478 events
[5] |= 0x80; /* Encryption Key Refresh Complete */
480 if (lmp_ext_inq_capable(hdev
))
481 events
[5] |= 0x40; /* Extended Inquiry Result */
483 if (lmp_no_flush_capable(hdev
))
484 events
[7] |= 0x01; /* Enhanced Flush Complete */
486 if (lmp_lsto_capable(hdev
))
487 events
[6] |= 0x80; /* Link Supervision Timeout Changed */
489 if (lmp_ssp_capable(hdev
)) {
490 events
[6] |= 0x01; /* IO Capability Request */
491 events
[6] |= 0x02; /* IO Capability Response */
492 events
[6] |= 0x04; /* User Confirmation Request */
493 events
[6] |= 0x08; /* User Passkey Request */
494 events
[6] |= 0x10; /* Remote OOB Data Request */
495 events
[6] |= 0x20; /* Simple Pairing Complete */
496 events
[7] |= 0x04; /* User Passkey Notification */
497 events
[7] |= 0x08; /* Keypress Notification */
498 events
[7] |= 0x10; /* Remote Host Supported
499 * Features Notification
503 if (lmp_le_capable(hdev
))
504 events
[7] |= 0x20; /* LE Meta-Event */
506 hci_req_add(req
, HCI_OP_SET_EVENT_MASK
, sizeof(events
), events
);
508 if (lmp_le_capable(hdev
)) {
509 memset(events
, 0, sizeof(events
));
511 hci_req_add(req
, HCI_OP_LE_SET_EVENT_MASK
,
512 sizeof(events
), events
);
516 static void hci_init2_req(struct hci_request
*req
, unsigned long opt
)
518 struct hci_dev
*hdev
= req
->hdev
;
520 if (lmp_bredr_capable(hdev
))
523 if (lmp_le_capable(hdev
))
526 hci_setup_event_mask(req
);
528 /* AVM Berlin (31), aka "BlueFRITZ!", doesn't support the read
529 * local supported commands HCI command.
531 if (hdev
->manufacturer
!= 31 && hdev
->hci_ver
> BLUETOOTH_VER_1_1
)
532 hci_req_add(req
, HCI_OP_READ_LOCAL_COMMANDS
, 0, NULL
);
534 if (lmp_ssp_capable(hdev
)) {
535 if (test_bit(HCI_SSP_ENABLED
, &hdev
->dev_flags
)) {
537 hci_req_add(req
, HCI_OP_WRITE_SSP_MODE
,
538 sizeof(mode
), &mode
);
540 struct hci_cp_write_eir cp
;
542 memset(hdev
->eir
, 0, sizeof(hdev
->eir
));
543 memset(&cp
, 0, sizeof(cp
));
545 hci_req_add(req
, HCI_OP_WRITE_EIR
, sizeof(cp
), &cp
);
549 if (lmp_inq_rssi_capable(hdev
))
550 hci_setup_inquiry_mode(req
);
552 if (lmp_inq_tx_pwr_capable(hdev
))
553 hci_req_add(req
, HCI_OP_READ_INQ_RSP_TX_POWER
, 0, NULL
);
555 if (lmp_ext_feat_capable(hdev
)) {
556 struct hci_cp_read_local_ext_features cp
;
559 hci_req_add(req
, HCI_OP_READ_LOCAL_EXT_FEATURES
,
563 if (test_bit(HCI_LINK_SECURITY
, &hdev
->dev_flags
)) {
565 hci_req_add(req
, HCI_OP_WRITE_AUTH_ENABLE
, sizeof(enable
),
570 static void hci_setup_link_policy(struct hci_request
*req
)
572 struct hci_dev
*hdev
= req
->hdev
;
573 struct hci_cp_write_def_link_policy cp
;
576 if (lmp_rswitch_capable(hdev
))
577 link_policy
|= HCI_LP_RSWITCH
;
578 if (lmp_hold_capable(hdev
))
579 link_policy
|= HCI_LP_HOLD
;
580 if (lmp_sniff_capable(hdev
))
581 link_policy
|= HCI_LP_SNIFF
;
582 if (lmp_park_capable(hdev
))
583 link_policy
|= HCI_LP_PARK
;
585 cp
.policy
= cpu_to_le16(link_policy
);
586 hci_req_add(req
, HCI_OP_WRITE_DEF_LINK_POLICY
, sizeof(cp
), &cp
);
589 static void hci_set_le_support(struct hci_request
*req
)
591 struct hci_dev
*hdev
= req
->hdev
;
592 struct hci_cp_write_le_host_supported cp
;
594 /* LE-only devices do not support explicit enablement */
595 if (!lmp_bredr_capable(hdev
))
598 memset(&cp
, 0, sizeof(cp
));
600 if (test_bit(HCI_LE_ENABLED
, &hdev
->dev_flags
)) {
602 cp
.simul
= lmp_le_br_capable(hdev
);
605 if (cp
.le
!= lmp_host_le_capable(hdev
))
606 hci_req_add(req
, HCI_OP_WRITE_LE_HOST_SUPPORTED
, sizeof(cp
),
610 static void hci_init3_req(struct hci_request
*req
, unsigned long opt
)
612 struct hci_dev
*hdev
= req
->hdev
;
615 /* Some Broadcom based Bluetooth controllers do not support the
616 * Delete Stored Link Key command. They are clearly indicating its
617 * absence in the bit mask of supported commands.
619 * Check the supported commands and only if the the command is marked
620 * as supported send it. If not supported assume that the controller
621 * does not have actual support for stored link keys which makes this
622 * command redundant anyway.
624 if (hdev
->commands
[6] & 0x80) {
625 struct hci_cp_delete_stored_link_key cp
;
627 bacpy(&cp
.bdaddr
, BDADDR_ANY
);
628 cp
.delete_all
= 0x01;
629 hci_req_add(req
, HCI_OP_DELETE_STORED_LINK_KEY
,
633 if (hdev
->commands
[5] & 0x10)
634 hci_setup_link_policy(req
);
636 if (lmp_le_capable(hdev
)) {
637 hci_set_le_support(req
);
641 /* Read features beyond page 1 if available */
642 for (p
= 2; p
< HCI_MAX_PAGES
&& p
<= hdev
->max_page
; p
++) {
643 struct hci_cp_read_local_ext_features cp
;
646 hci_req_add(req
, HCI_OP_READ_LOCAL_EXT_FEATURES
,
651 static int __hci_init(struct hci_dev
*hdev
)
655 err
= __hci_req_sync(hdev
, hci_init1_req
, 0, HCI_INIT_TIMEOUT
);
659 /* HCI_BREDR covers both single-mode LE, BR/EDR and dual-mode
660 * BR/EDR/LE type controllers. AMP controllers only need the
663 if (hdev
->dev_type
!= HCI_BREDR
)
666 err
= __hci_req_sync(hdev
, hci_init2_req
, 0, HCI_INIT_TIMEOUT
);
670 return __hci_req_sync(hdev
, hci_init3_req
, 0, HCI_INIT_TIMEOUT
);
673 static void hci_scan_req(struct hci_request
*req
, unsigned long opt
)
677 BT_DBG("%s %x", req
->hdev
->name
, scan
);
679 /* Inquiry and Page scans */
680 hci_req_add(req
, HCI_OP_WRITE_SCAN_ENABLE
, 1, &scan
);
683 static void hci_auth_req(struct hci_request
*req
, unsigned long opt
)
687 BT_DBG("%s %x", req
->hdev
->name
, auth
);
690 hci_req_add(req
, HCI_OP_WRITE_AUTH_ENABLE
, 1, &auth
);
693 static void hci_encrypt_req(struct hci_request
*req
, unsigned long opt
)
697 BT_DBG("%s %x", req
->hdev
->name
, encrypt
);
700 hci_req_add(req
, HCI_OP_WRITE_ENCRYPT_MODE
, 1, &encrypt
);
703 static void hci_linkpol_req(struct hci_request
*req
, unsigned long opt
)
705 __le16 policy
= cpu_to_le16(opt
);
707 BT_DBG("%s %x", req
->hdev
->name
, policy
);
709 /* Default link policy */
710 hci_req_add(req
, HCI_OP_WRITE_DEF_LINK_POLICY
, 2, &policy
);
713 /* Get HCI device by index.
714 * Device is held on return. */
715 struct hci_dev
*hci_dev_get(int index
)
717 struct hci_dev
*hdev
= NULL
, *d
;
724 read_lock(&hci_dev_list_lock
);
725 list_for_each_entry(d
, &hci_dev_list
, list
) {
726 if (d
->id
== index
) {
727 hdev
= hci_dev_hold(d
);
731 read_unlock(&hci_dev_list_lock
);
735 /* ---- Inquiry support ---- */
737 bool hci_discovery_active(struct hci_dev
*hdev
)
739 struct discovery_state
*discov
= &hdev
->discovery
;
741 switch (discov
->state
) {
742 case DISCOVERY_FINDING
:
743 case DISCOVERY_RESOLVING
:
751 void hci_discovery_set_state(struct hci_dev
*hdev
, int state
)
753 BT_DBG("%s state %u -> %u", hdev
->name
, hdev
->discovery
.state
, state
);
755 if (hdev
->discovery
.state
== state
)
759 case DISCOVERY_STOPPED
:
760 if (hdev
->discovery
.state
!= DISCOVERY_STARTING
)
761 mgmt_discovering(hdev
, 0);
763 case DISCOVERY_STARTING
:
765 case DISCOVERY_FINDING
:
766 mgmt_discovering(hdev
, 1);
768 case DISCOVERY_RESOLVING
:
770 case DISCOVERY_STOPPING
:
774 hdev
->discovery
.state
= state
;
777 void hci_inquiry_cache_flush(struct hci_dev
*hdev
)
779 struct discovery_state
*cache
= &hdev
->discovery
;
780 struct inquiry_entry
*p
, *n
;
782 list_for_each_entry_safe(p
, n
, &cache
->all
, all
) {
787 INIT_LIST_HEAD(&cache
->unknown
);
788 INIT_LIST_HEAD(&cache
->resolve
);
791 struct inquiry_entry
*hci_inquiry_cache_lookup(struct hci_dev
*hdev
,
794 struct discovery_state
*cache
= &hdev
->discovery
;
795 struct inquiry_entry
*e
;
797 BT_DBG("cache %p, %pMR", cache
, bdaddr
);
799 list_for_each_entry(e
, &cache
->all
, all
) {
800 if (!bacmp(&e
->data
.bdaddr
, bdaddr
))
807 struct inquiry_entry
*hci_inquiry_cache_lookup_unknown(struct hci_dev
*hdev
,
810 struct discovery_state
*cache
= &hdev
->discovery
;
811 struct inquiry_entry
*e
;
813 BT_DBG("cache %p, %pMR", cache
, bdaddr
);
815 list_for_each_entry(e
, &cache
->unknown
, list
) {
816 if (!bacmp(&e
->data
.bdaddr
, bdaddr
))
823 struct inquiry_entry
*hci_inquiry_cache_lookup_resolve(struct hci_dev
*hdev
,
827 struct discovery_state
*cache
= &hdev
->discovery
;
828 struct inquiry_entry
*e
;
830 BT_DBG("cache %p bdaddr %pMR state %d", cache
, bdaddr
, state
);
832 list_for_each_entry(e
, &cache
->resolve
, list
) {
833 if (!bacmp(bdaddr
, BDADDR_ANY
) && e
->name_state
== state
)
835 if (!bacmp(&e
->data
.bdaddr
, bdaddr
))
842 void hci_inquiry_cache_update_resolve(struct hci_dev
*hdev
,
843 struct inquiry_entry
*ie
)
845 struct discovery_state
*cache
= &hdev
->discovery
;
846 struct list_head
*pos
= &cache
->resolve
;
847 struct inquiry_entry
*p
;
851 list_for_each_entry(p
, &cache
->resolve
, list
) {
852 if (p
->name_state
!= NAME_PENDING
&&
853 abs(p
->data
.rssi
) >= abs(ie
->data
.rssi
))
858 list_add(&ie
->list
, pos
);
861 bool hci_inquiry_cache_update(struct hci_dev
*hdev
, struct inquiry_data
*data
,
862 bool name_known
, bool *ssp
)
864 struct discovery_state
*cache
= &hdev
->discovery
;
865 struct inquiry_entry
*ie
;
867 BT_DBG("cache %p, %pMR", cache
, &data
->bdaddr
);
869 hci_remove_remote_oob_data(hdev
, &data
->bdaddr
);
872 *ssp
= data
->ssp_mode
;
874 ie
= hci_inquiry_cache_lookup(hdev
, &data
->bdaddr
);
876 if (ie
->data
.ssp_mode
&& ssp
)
879 if (ie
->name_state
== NAME_NEEDED
&&
880 data
->rssi
!= ie
->data
.rssi
) {
881 ie
->data
.rssi
= data
->rssi
;
882 hci_inquiry_cache_update_resolve(hdev
, ie
);
888 /* Entry not in the cache. Add new one. */
889 ie
= kzalloc(sizeof(struct inquiry_entry
), GFP_ATOMIC
);
893 list_add(&ie
->all
, &cache
->all
);
896 ie
->name_state
= NAME_KNOWN
;
898 ie
->name_state
= NAME_NOT_KNOWN
;
899 list_add(&ie
->list
, &cache
->unknown
);
903 if (name_known
&& ie
->name_state
!= NAME_KNOWN
&&
904 ie
->name_state
!= NAME_PENDING
) {
905 ie
->name_state
= NAME_KNOWN
;
909 memcpy(&ie
->data
, data
, sizeof(*data
));
910 ie
->timestamp
= jiffies
;
911 cache
->timestamp
= jiffies
;
913 if (ie
->name_state
== NAME_NOT_KNOWN
)
919 static int inquiry_cache_dump(struct hci_dev
*hdev
, int num
, __u8
*buf
)
921 struct discovery_state
*cache
= &hdev
->discovery
;
922 struct inquiry_info
*info
= (struct inquiry_info
*) buf
;
923 struct inquiry_entry
*e
;
926 list_for_each_entry(e
, &cache
->all
, all
) {
927 struct inquiry_data
*data
= &e
->data
;
932 bacpy(&info
->bdaddr
, &data
->bdaddr
);
933 info
->pscan_rep_mode
= data
->pscan_rep_mode
;
934 info
->pscan_period_mode
= data
->pscan_period_mode
;
935 info
->pscan_mode
= data
->pscan_mode
;
936 memcpy(info
->dev_class
, data
->dev_class
, 3);
937 info
->clock_offset
= data
->clock_offset
;
943 BT_DBG("cache %p, copied %d", cache
, copied
);
947 static void hci_inq_req(struct hci_request
*req
, unsigned long opt
)
949 struct hci_inquiry_req
*ir
= (struct hci_inquiry_req
*) opt
;
950 struct hci_dev
*hdev
= req
->hdev
;
951 struct hci_cp_inquiry cp
;
953 BT_DBG("%s", hdev
->name
);
955 if (test_bit(HCI_INQUIRY
, &hdev
->flags
))
959 memcpy(&cp
.lap
, &ir
->lap
, 3);
960 cp
.length
= ir
->length
;
961 cp
.num_rsp
= ir
->num_rsp
;
962 hci_req_add(req
, HCI_OP_INQUIRY
, sizeof(cp
), &cp
);
965 static int wait_inquiry(void *word
)
968 return signal_pending(current
);
971 int hci_inquiry(void __user
*arg
)
973 __u8 __user
*ptr
= arg
;
974 struct hci_inquiry_req ir
;
975 struct hci_dev
*hdev
;
976 int err
= 0, do_inquiry
= 0, max_rsp
;
980 if (copy_from_user(&ir
, ptr
, sizeof(ir
)))
983 hdev
= hci_dev_get(ir
.dev_id
);
988 if (inquiry_cache_age(hdev
) > INQUIRY_CACHE_AGE_MAX
||
989 inquiry_cache_empty(hdev
) || ir
.flags
& IREQ_CACHE_FLUSH
) {
990 hci_inquiry_cache_flush(hdev
);
993 hci_dev_unlock(hdev
);
995 timeo
= ir
.length
* msecs_to_jiffies(2000);
998 err
= hci_req_sync(hdev
, hci_inq_req
, (unsigned long) &ir
,
1003 /* Wait until Inquiry procedure finishes (HCI_INQUIRY flag is
1004 * cleared). If it is interrupted by a signal, return -EINTR.
1006 if (wait_on_bit(&hdev
->flags
, HCI_INQUIRY
, wait_inquiry
,
1007 TASK_INTERRUPTIBLE
))
1011 /* for unlimited number of responses we will use buffer with
1014 max_rsp
= (ir
.num_rsp
== 0) ? 255 : ir
.num_rsp
;
1016 /* cache_dump can't sleep. Therefore we allocate temp buffer and then
1017 * copy it to the user space.
1019 buf
= kmalloc(sizeof(struct inquiry_info
) * max_rsp
, GFP_KERNEL
);
1026 ir
.num_rsp
= inquiry_cache_dump(hdev
, max_rsp
, buf
);
1027 hci_dev_unlock(hdev
);
1029 BT_DBG("num_rsp %d", ir
.num_rsp
);
1031 if (!copy_to_user(ptr
, &ir
, sizeof(ir
))) {
1033 if (copy_to_user(ptr
, buf
, sizeof(struct inquiry_info
) *
1046 static u8
create_ad(struct hci_dev
*hdev
, u8
*ptr
)
1048 u8 ad_len
= 0, flags
= 0;
1051 if (test_bit(HCI_LE_PERIPHERAL
, &hdev
->dev_flags
))
1052 flags
|= LE_AD_GENERAL
;
1054 if (!lmp_bredr_capable(hdev
))
1055 flags
|= LE_AD_NO_BREDR
;
1057 if (lmp_le_br_capable(hdev
))
1058 flags
|= LE_AD_SIM_LE_BREDR_CTRL
;
1060 if (lmp_host_le_br_capable(hdev
))
1061 flags
|= LE_AD_SIM_LE_BREDR_HOST
;
1064 BT_DBG("adv flags 0x%02x", flags
);
1074 if (hdev
->adv_tx_power
!= HCI_TX_POWER_INVALID
) {
1076 ptr
[1] = EIR_TX_POWER
;
1077 ptr
[2] = (u8
) hdev
->adv_tx_power
;
1083 name_len
= strlen(hdev
->dev_name
);
1085 size_t max_len
= HCI_MAX_AD_LENGTH
- ad_len
- 2;
1087 if (name_len
> max_len
) {
1089 ptr
[1] = EIR_NAME_SHORT
;
1091 ptr
[1] = EIR_NAME_COMPLETE
;
1093 ptr
[0] = name_len
+ 1;
1095 memcpy(ptr
+ 2, hdev
->dev_name
, name_len
);
1097 ad_len
+= (name_len
+ 2);
1098 ptr
+= (name_len
+ 2);
1104 void hci_update_ad(struct hci_request
*req
)
1106 struct hci_dev
*hdev
= req
->hdev
;
1107 struct hci_cp_le_set_adv_data cp
;
1110 if (!lmp_le_capable(hdev
))
1113 memset(&cp
, 0, sizeof(cp
));
1115 len
= create_ad(hdev
, cp
.data
);
1117 if (hdev
->adv_data_len
== len
&&
1118 memcmp(cp
.data
, hdev
->adv_data
, len
) == 0)
1121 memcpy(hdev
->adv_data
, cp
.data
, sizeof(cp
.data
));
1122 hdev
->adv_data_len
= len
;
1126 hci_req_add(req
, HCI_OP_LE_SET_ADV_DATA
, sizeof(cp
), &cp
);
1129 /* ---- HCI ioctl helpers ---- */
1131 int hci_dev_open(__u16 dev
)
1133 struct hci_dev
*hdev
;
1136 hdev
= hci_dev_get(dev
);
1140 BT_DBG("%s %p", hdev
->name
, hdev
);
1144 if (test_bit(HCI_UNREGISTER
, &hdev
->dev_flags
)) {
1149 if (hdev
->rfkill
&& rfkill_blocked(hdev
->rfkill
)) {
1154 if (test_bit(HCI_UP
, &hdev
->flags
)) {
1159 if (hdev
->open(hdev
)) {
1164 atomic_set(&hdev
->cmd_cnt
, 1);
1165 set_bit(HCI_INIT
, &hdev
->flags
);
1167 if (hdev
->setup
&& test_bit(HCI_SETUP
, &hdev
->dev_flags
))
1168 ret
= hdev
->setup(hdev
);
1171 /* Treat all non BR/EDR controllers as raw devices if
1172 * enable_hs is not set.
1174 if (hdev
->dev_type
!= HCI_BREDR
&& !enable_hs
)
1175 set_bit(HCI_RAW
, &hdev
->flags
);
1177 if (test_bit(HCI_QUIRK_RAW_DEVICE
, &hdev
->quirks
))
1178 set_bit(HCI_RAW
, &hdev
->flags
);
1180 if (!test_bit(HCI_RAW
, &hdev
->flags
))
1181 ret
= __hci_init(hdev
);
1184 clear_bit(HCI_INIT
, &hdev
->flags
);
1188 set_bit(HCI_UP
, &hdev
->flags
);
1189 hci_notify(hdev
, HCI_DEV_UP
);
1190 if (!test_bit(HCI_SETUP
, &hdev
->dev_flags
) &&
1191 mgmt_valid_hdev(hdev
)) {
1193 mgmt_powered(hdev
, 1);
1194 hci_dev_unlock(hdev
);
1197 /* Init failed, cleanup */
1198 flush_work(&hdev
->tx_work
);
1199 flush_work(&hdev
->cmd_work
);
1200 flush_work(&hdev
->rx_work
);
1202 skb_queue_purge(&hdev
->cmd_q
);
1203 skb_queue_purge(&hdev
->rx_q
);
1208 if (hdev
->sent_cmd
) {
1209 kfree_skb(hdev
->sent_cmd
);
1210 hdev
->sent_cmd
= NULL
;
1218 hci_req_unlock(hdev
);
1223 static int hci_dev_do_close(struct hci_dev
*hdev
)
1225 BT_DBG("%s %p", hdev
->name
, hdev
);
1227 cancel_delayed_work(&hdev
->power_off
);
1229 hci_req_cancel(hdev
, ENODEV
);
1232 if (!test_and_clear_bit(HCI_UP
, &hdev
->flags
)) {
1233 del_timer_sync(&hdev
->cmd_timer
);
1234 hci_req_unlock(hdev
);
1238 /* Flush RX and TX works */
1239 flush_work(&hdev
->tx_work
);
1240 flush_work(&hdev
->rx_work
);
1242 if (hdev
->discov_timeout
> 0) {
1243 cancel_delayed_work(&hdev
->discov_off
);
1244 hdev
->discov_timeout
= 0;
1245 clear_bit(HCI_DISCOVERABLE
, &hdev
->dev_flags
);
1248 if (test_and_clear_bit(HCI_SERVICE_CACHE
, &hdev
->dev_flags
))
1249 cancel_delayed_work(&hdev
->service_cache
);
1251 cancel_delayed_work_sync(&hdev
->le_scan_disable
);
1254 hci_inquiry_cache_flush(hdev
);
1255 hci_conn_hash_flush(hdev
);
1256 hci_dev_unlock(hdev
);
1258 hci_notify(hdev
, HCI_DEV_DOWN
);
1264 skb_queue_purge(&hdev
->cmd_q
);
1265 atomic_set(&hdev
->cmd_cnt
, 1);
1266 if (!test_bit(HCI_RAW
, &hdev
->flags
) &&
1267 test_bit(HCI_QUIRK_RESET_ON_CLOSE
, &hdev
->quirks
)) {
1268 set_bit(HCI_INIT
, &hdev
->flags
);
1269 __hci_req_sync(hdev
, hci_reset_req
, 0, HCI_CMD_TIMEOUT
);
1270 clear_bit(HCI_INIT
, &hdev
->flags
);
1273 /* flush cmd work */
1274 flush_work(&hdev
->cmd_work
);
1277 skb_queue_purge(&hdev
->rx_q
);
1278 skb_queue_purge(&hdev
->cmd_q
);
1279 skb_queue_purge(&hdev
->raw_q
);
1281 /* Drop last sent command */
1282 if (hdev
->sent_cmd
) {
1283 del_timer_sync(&hdev
->cmd_timer
);
1284 kfree_skb(hdev
->sent_cmd
);
1285 hdev
->sent_cmd
= NULL
;
1288 kfree_skb(hdev
->recv_evt
);
1289 hdev
->recv_evt
= NULL
;
1291 /* After this point our queues are empty
1292 * and no tasks are scheduled. */
1297 hdev
->dev_flags
&= ~HCI_PERSISTENT_MASK
;
1299 if (!test_and_clear_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
) &&
1300 mgmt_valid_hdev(hdev
)) {
1302 mgmt_powered(hdev
, 0);
1303 hci_dev_unlock(hdev
);
1306 /* Controller radio is available but is currently powered down */
1307 hdev
->amp_status
= 0;
1309 memset(hdev
->eir
, 0, sizeof(hdev
->eir
));
1310 memset(hdev
->dev_class
, 0, sizeof(hdev
->dev_class
));
1312 hci_req_unlock(hdev
);
1318 int hci_dev_close(__u16 dev
)
1320 struct hci_dev
*hdev
;
1323 hdev
= hci_dev_get(dev
);
1327 if (test_and_clear_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
))
1328 cancel_delayed_work(&hdev
->power_off
);
1330 err
= hci_dev_do_close(hdev
);
1336 int hci_dev_reset(__u16 dev
)
1338 struct hci_dev
*hdev
;
1341 hdev
= hci_dev_get(dev
);
1347 if (!test_bit(HCI_UP
, &hdev
->flags
)) {
1353 skb_queue_purge(&hdev
->rx_q
);
1354 skb_queue_purge(&hdev
->cmd_q
);
1357 hci_inquiry_cache_flush(hdev
);
1358 hci_conn_hash_flush(hdev
);
1359 hci_dev_unlock(hdev
);
1364 atomic_set(&hdev
->cmd_cnt
, 1);
1365 hdev
->acl_cnt
= 0; hdev
->sco_cnt
= 0; hdev
->le_cnt
= 0;
1367 if (!test_bit(HCI_RAW
, &hdev
->flags
))
1368 ret
= __hci_req_sync(hdev
, hci_reset_req
, 0, HCI_INIT_TIMEOUT
);
1371 hci_req_unlock(hdev
);
1376 int hci_dev_reset_stat(__u16 dev
)
1378 struct hci_dev
*hdev
;
1381 hdev
= hci_dev_get(dev
);
1385 memset(&hdev
->stat
, 0, sizeof(struct hci_dev_stats
));
1392 int hci_dev_cmd(unsigned int cmd
, void __user
*arg
)
1394 struct hci_dev
*hdev
;
1395 struct hci_dev_req dr
;
1398 if (copy_from_user(&dr
, arg
, sizeof(dr
)))
1401 hdev
= hci_dev_get(dr
.dev_id
);
1407 err
= hci_req_sync(hdev
, hci_auth_req
, dr
.dev_opt
,
1412 if (!lmp_encrypt_capable(hdev
)) {
1417 if (!test_bit(HCI_AUTH
, &hdev
->flags
)) {
1418 /* Auth must be enabled first */
1419 err
= hci_req_sync(hdev
, hci_auth_req
, dr
.dev_opt
,
1425 err
= hci_req_sync(hdev
, hci_encrypt_req
, dr
.dev_opt
,
1430 err
= hci_req_sync(hdev
, hci_scan_req
, dr
.dev_opt
,
1435 err
= hci_req_sync(hdev
, hci_linkpol_req
, dr
.dev_opt
,
1439 case HCISETLINKMODE
:
1440 hdev
->link_mode
= ((__u16
) dr
.dev_opt
) &
1441 (HCI_LM_MASTER
| HCI_LM_ACCEPT
);
1445 hdev
->pkt_type
= (__u16
) dr
.dev_opt
;
1449 hdev
->acl_mtu
= *((__u16
*) &dr
.dev_opt
+ 1);
1450 hdev
->acl_pkts
= *((__u16
*) &dr
.dev_opt
+ 0);
1454 hdev
->sco_mtu
= *((__u16
*) &dr
.dev_opt
+ 1);
1455 hdev
->sco_pkts
= *((__u16
*) &dr
.dev_opt
+ 0);
1467 int hci_get_dev_list(void __user
*arg
)
1469 struct hci_dev
*hdev
;
1470 struct hci_dev_list_req
*dl
;
1471 struct hci_dev_req
*dr
;
1472 int n
= 0, size
, err
;
1475 if (get_user(dev_num
, (__u16 __user
*) arg
))
1478 if (!dev_num
|| dev_num
> (PAGE_SIZE
* 2) / sizeof(*dr
))
1481 size
= sizeof(*dl
) + dev_num
* sizeof(*dr
);
1483 dl
= kzalloc(size
, GFP_KERNEL
);
1489 read_lock(&hci_dev_list_lock
);
1490 list_for_each_entry(hdev
, &hci_dev_list
, list
) {
1491 if (test_and_clear_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
))
1492 cancel_delayed_work(&hdev
->power_off
);
1494 if (!test_bit(HCI_MGMT
, &hdev
->dev_flags
))
1495 set_bit(HCI_PAIRABLE
, &hdev
->dev_flags
);
1497 (dr
+ n
)->dev_id
= hdev
->id
;
1498 (dr
+ n
)->dev_opt
= hdev
->flags
;
1503 read_unlock(&hci_dev_list_lock
);
1506 size
= sizeof(*dl
) + n
* sizeof(*dr
);
1508 err
= copy_to_user(arg
, dl
, size
);
1511 return err
? -EFAULT
: 0;
1514 int hci_get_dev_info(void __user
*arg
)
1516 struct hci_dev
*hdev
;
1517 struct hci_dev_info di
;
1520 if (copy_from_user(&di
, arg
, sizeof(di
)))
1523 hdev
= hci_dev_get(di
.dev_id
);
1527 if (test_and_clear_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
))
1528 cancel_delayed_work_sync(&hdev
->power_off
);
1530 if (!test_bit(HCI_MGMT
, &hdev
->dev_flags
))
1531 set_bit(HCI_PAIRABLE
, &hdev
->dev_flags
);
1533 strcpy(di
.name
, hdev
->name
);
1534 di
.bdaddr
= hdev
->bdaddr
;
1535 di
.type
= (hdev
->bus
& 0x0f) | (hdev
->dev_type
<< 4);
1536 di
.flags
= hdev
->flags
;
1537 di
.pkt_type
= hdev
->pkt_type
;
1538 if (lmp_bredr_capable(hdev
)) {
1539 di
.acl_mtu
= hdev
->acl_mtu
;
1540 di
.acl_pkts
= hdev
->acl_pkts
;
1541 di
.sco_mtu
= hdev
->sco_mtu
;
1542 di
.sco_pkts
= hdev
->sco_pkts
;
1544 di
.acl_mtu
= hdev
->le_mtu
;
1545 di
.acl_pkts
= hdev
->le_pkts
;
1549 di
.link_policy
= hdev
->link_policy
;
1550 di
.link_mode
= hdev
->link_mode
;
1552 memcpy(&di
.stat
, &hdev
->stat
, sizeof(di
.stat
));
1553 memcpy(&di
.features
, &hdev
->features
, sizeof(di
.features
));
1555 if (copy_to_user(arg
, &di
, sizeof(di
)))
1563 /* ---- Interface to HCI drivers ---- */
1565 static int hci_rfkill_set_block(void *data
, bool blocked
)
1567 struct hci_dev
*hdev
= data
;
1569 BT_DBG("%p name %s blocked %d", hdev
, hdev
->name
, blocked
);
1574 hci_dev_do_close(hdev
);
1579 static const struct rfkill_ops hci_rfkill_ops
= {
1580 .set_block
= hci_rfkill_set_block
,
1583 static void hci_power_on(struct work_struct
*work
)
1585 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
, power_on
);
1588 BT_DBG("%s", hdev
->name
);
1590 err
= hci_dev_open(hdev
->id
);
1592 mgmt_set_powered_failed(hdev
, err
);
1596 if (test_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
))
1597 queue_delayed_work(hdev
->req_workqueue
, &hdev
->power_off
,
1598 HCI_AUTO_OFF_TIMEOUT
);
1600 if (test_and_clear_bit(HCI_SETUP
, &hdev
->dev_flags
))
1601 mgmt_index_added(hdev
);
1604 static void hci_power_off(struct work_struct
*work
)
1606 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
,
1609 BT_DBG("%s", hdev
->name
);
1611 hci_dev_do_close(hdev
);
1614 static void hci_discov_off(struct work_struct
*work
)
1616 struct hci_dev
*hdev
;
1617 u8 scan
= SCAN_PAGE
;
1619 hdev
= container_of(work
, struct hci_dev
, discov_off
.work
);
1621 BT_DBG("%s", hdev
->name
);
1625 hci_send_cmd(hdev
, HCI_OP_WRITE_SCAN_ENABLE
, sizeof(scan
), &scan
);
1627 hdev
->discov_timeout
= 0;
1629 hci_dev_unlock(hdev
);
1632 int hci_uuids_clear(struct hci_dev
*hdev
)
1634 struct bt_uuid
*uuid
, *tmp
;
1636 list_for_each_entry_safe(uuid
, tmp
, &hdev
->uuids
, list
) {
1637 list_del(&uuid
->list
);
1644 int hci_link_keys_clear(struct hci_dev
*hdev
)
1646 struct list_head
*p
, *n
;
1648 list_for_each_safe(p
, n
, &hdev
->link_keys
) {
1649 struct link_key
*key
;
1651 key
= list_entry(p
, struct link_key
, list
);
1660 int hci_smp_ltks_clear(struct hci_dev
*hdev
)
1662 struct smp_ltk
*k
, *tmp
;
1664 list_for_each_entry_safe(k
, tmp
, &hdev
->long_term_keys
, list
) {
1672 struct link_key
*hci_find_link_key(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
)
1676 list_for_each_entry(k
, &hdev
->link_keys
, list
)
1677 if (bacmp(bdaddr
, &k
->bdaddr
) == 0)
1683 static bool hci_persistent_key(struct hci_dev
*hdev
, struct hci_conn
*conn
,
1684 u8 key_type
, u8 old_key_type
)
1687 if (key_type
< 0x03)
1690 /* Debug keys are insecure so don't store them persistently */
1691 if (key_type
== HCI_LK_DEBUG_COMBINATION
)
1694 /* Changed combination key and there's no previous one */
1695 if (key_type
== HCI_LK_CHANGED_COMBINATION
&& old_key_type
== 0xff)
1698 /* Security mode 3 case */
1702 /* Neither local nor remote side had no-bonding as requirement */
1703 if (conn
->auth_type
> 0x01 && conn
->remote_auth
> 0x01)
1706 /* Local side had dedicated bonding as requirement */
1707 if (conn
->auth_type
== 0x02 || conn
->auth_type
== 0x03)
1710 /* Remote side had dedicated bonding as requirement */
1711 if (conn
->remote_auth
== 0x02 || conn
->remote_auth
== 0x03)
1714 /* If none of the above criteria match, then don't store the key
1719 struct smp_ltk
*hci_find_ltk(struct hci_dev
*hdev
, __le16 ediv
, u8 rand
[8])
1723 list_for_each_entry(k
, &hdev
->long_term_keys
, list
) {
1724 if (k
->ediv
!= ediv
||
1725 memcmp(rand
, k
->rand
, sizeof(k
->rand
)))
1734 struct smp_ltk
*hci_find_ltk_by_addr(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
,
1739 list_for_each_entry(k
, &hdev
->long_term_keys
, list
)
1740 if (addr_type
== k
->bdaddr_type
&&
1741 bacmp(bdaddr
, &k
->bdaddr
) == 0)
1747 int hci_add_link_key(struct hci_dev
*hdev
, struct hci_conn
*conn
, int new_key
,
1748 bdaddr_t
*bdaddr
, u8
*val
, u8 type
, u8 pin_len
)
1750 struct link_key
*key
, *old_key
;
1754 old_key
= hci_find_link_key(hdev
, bdaddr
);
1756 old_key_type
= old_key
->type
;
1759 old_key_type
= conn
? conn
->key_type
: 0xff;
1760 key
= kzalloc(sizeof(*key
), GFP_ATOMIC
);
1763 list_add(&key
->list
, &hdev
->link_keys
);
1766 BT_DBG("%s key for %pMR type %u", hdev
->name
, bdaddr
, type
);
1768 /* Some buggy controller combinations generate a changed
1769 * combination key for legacy pairing even when there's no
1771 if (type
== HCI_LK_CHANGED_COMBINATION
&&
1772 (!conn
|| conn
->remote_auth
== 0xff) && old_key_type
== 0xff) {
1773 type
= HCI_LK_COMBINATION
;
1775 conn
->key_type
= type
;
1778 bacpy(&key
->bdaddr
, bdaddr
);
1779 memcpy(key
->val
, val
, HCI_LINK_KEY_SIZE
);
1780 key
->pin_len
= pin_len
;
1782 if (type
== HCI_LK_CHANGED_COMBINATION
)
1783 key
->type
= old_key_type
;
1790 persistent
= hci_persistent_key(hdev
, conn
, type
, old_key_type
);
1792 mgmt_new_link_key(hdev
, key
, persistent
);
1795 conn
->flush_key
= !persistent
;
1800 int hci_add_ltk(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
, u8 addr_type
, u8 type
,
1801 int new_key
, u8 authenticated
, u8 tk
[16], u8 enc_size
, __le16
1804 struct smp_ltk
*key
, *old_key
;
1806 if (!(type
& HCI_SMP_STK
) && !(type
& HCI_SMP_LTK
))
1809 old_key
= hci_find_ltk_by_addr(hdev
, bdaddr
, addr_type
);
1813 key
= kzalloc(sizeof(*key
), GFP_ATOMIC
);
1816 list_add(&key
->list
, &hdev
->long_term_keys
);
1819 bacpy(&key
->bdaddr
, bdaddr
);
1820 key
->bdaddr_type
= addr_type
;
1821 memcpy(key
->val
, tk
, sizeof(key
->val
));
1822 key
->authenticated
= authenticated
;
1824 key
->enc_size
= enc_size
;
1826 memcpy(key
->rand
, rand
, sizeof(key
->rand
));
1831 if (type
& HCI_SMP_LTK
)
1832 mgmt_new_ltk(hdev
, key
, 1);
1837 int hci_remove_link_key(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
)
1839 struct link_key
*key
;
1841 key
= hci_find_link_key(hdev
, bdaddr
);
1845 BT_DBG("%s removing %pMR", hdev
->name
, bdaddr
);
1847 list_del(&key
->list
);
1853 int hci_remove_ltk(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
)
1855 struct smp_ltk
*k
, *tmp
;
1857 list_for_each_entry_safe(k
, tmp
, &hdev
->long_term_keys
, list
) {
1858 if (bacmp(bdaddr
, &k
->bdaddr
))
1861 BT_DBG("%s removing %pMR", hdev
->name
, bdaddr
);
1870 /* HCI command timer function */
1871 static void hci_cmd_timeout(unsigned long arg
)
1873 struct hci_dev
*hdev
= (void *) arg
;
1875 if (hdev
->sent_cmd
) {
1876 struct hci_command_hdr
*sent
= (void *) hdev
->sent_cmd
->data
;
1877 u16 opcode
= __le16_to_cpu(sent
->opcode
);
1879 BT_ERR("%s command 0x%4.4x tx timeout", hdev
->name
, opcode
);
1881 BT_ERR("%s command tx timeout", hdev
->name
);
1884 atomic_set(&hdev
->cmd_cnt
, 1);
1885 queue_work(hdev
->workqueue
, &hdev
->cmd_work
);
1888 struct oob_data
*hci_find_remote_oob_data(struct hci_dev
*hdev
,
1891 struct oob_data
*data
;
1893 list_for_each_entry(data
, &hdev
->remote_oob_data
, list
)
1894 if (bacmp(bdaddr
, &data
->bdaddr
) == 0)
1900 int hci_remove_remote_oob_data(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
)
1902 struct oob_data
*data
;
1904 data
= hci_find_remote_oob_data(hdev
, bdaddr
);
1908 BT_DBG("%s removing %pMR", hdev
->name
, bdaddr
);
1910 list_del(&data
->list
);
1916 int hci_remote_oob_data_clear(struct hci_dev
*hdev
)
1918 struct oob_data
*data
, *n
;
1920 list_for_each_entry_safe(data
, n
, &hdev
->remote_oob_data
, list
) {
1921 list_del(&data
->list
);
1928 int hci_add_remote_oob_data(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
, u8
*hash
,
1931 struct oob_data
*data
;
1933 data
= hci_find_remote_oob_data(hdev
, bdaddr
);
1936 data
= kmalloc(sizeof(*data
), GFP_ATOMIC
);
1940 bacpy(&data
->bdaddr
, bdaddr
);
1941 list_add(&data
->list
, &hdev
->remote_oob_data
);
1944 memcpy(data
->hash
, hash
, sizeof(data
->hash
));
1945 memcpy(data
->randomizer
, randomizer
, sizeof(data
->randomizer
));
1947 BT_DBG("%s for %pMR", hdev
->name
, bdaddr
);
1952 struct bdaddr_list
*hci_blacklist_lookup(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
)
1954 struct bdaddr_list
*b
;
1956 list_for_each_entry(b
, &hdev
->blacklist
, list
)
1957 if (bacmp(bdaddr
, &b
->bdaddr
) == 0)
1963 int hci_blacklist_clear(struct hci_dev
*hdev
)
1965 struct list_head
*p
, *n
;
1967 list_for_each_safe(p
, n
, &hdev
->blacklist
) {
1968 struct bdaddr_list
*b
;
1970 b
= list_entry(p
, struct bdaddr_list
, list
);
1979 int hci_blacklist_add(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
, u8 type
)
1981 struct bdaddr_list
*entry
;
1983 if (bacmp(bdaddr
, BDADDR_ANY
) == 0)
1986 if (hci_blacklist_lookup(hdev
, bdaddr
))
1989 entry
= kzalloc(sizeof(struct bdaddr_list
), GFP_KERNEL
);
1993 bacpy(&entry
->bdaddr
, bdaddr
);
1995 list_add(&entry
->list
, &hdev
->blacklist
);
1997 return mgmt_device_blocked(hdev
, bdaddr
, type
);
2000 int hci_blacklist_del(struct hci_dev
*hdev
, bdaddr_t
*bdaddr
, u8 type
)
2002 struct bdaddr_list
*entry
;
2004 if (bacmp(bdaddr
, BDADDR_ANY
) == 0)
2005 return hci_blacklist_clear(hdev
);
2007 entry
= hci_blacklist_lookup(hdev
, bdaddr
);
2011 list_del(&entry
->list
);
2014 return mgmt_device_unblocked(hdev
, bdaddr
, type
);
2017 static void inquiry_complete(struct hci_dev
*hdev
, u8 status
)
2020 BT_ERR("Failed to start inquiry: status %d", status
);
2023 hci_discovery_set_state(hdev
, DISCOVERY_STOPPED
);
2024 hci_dev_unlock(hdev
);
2029 static void le_scan_disable_work_complete(struct hci_dev
*hdev
, u8 status
)
2031 /* General inquiry access code (GIAC) */
2032 u8 lap
[3] = { 0x33, 0x8b, 0x9e };
2033 struct hci_request req
;
2034 struct hci_cp_inquiry cp
;
2038 BT_ERR("Failed to disable LE scanning: status %d", status
);
2042 switch (hdev
->discovery
.type
) {
2043 case DISCOV_TYPE_LE
:
2045 hci_discovery_set_state(hdev
, DISCOVERY_STOPPED
);
2046 hci_dev_unlock(hdev
);
2049 case DISCOV_TYPE_INTERLEAVED
:
2050 hci_req_init(&req
, hdev
);
2052 memset(&cp
, 0, sizeof(cp
));
2053 memcpy(&cp
.lap
, lap
, sizeof(cp
.lap
));
2054 cp
.length
= DISCOV_INTERLEAVED_INQUIRY_LEN
;
2055 hci_req_add(&req
, HCI_OP_INQUIRY
, sizeof(cp
), &cp
);
2059 hci_inquiry_cache_flush(hdev
);
2061 err
= hci_req_run(&req
, inquiry_complete
);
2063 BT_ERR("Inquiry request failed: err %d", err
);
2064 hci_discovery_set_state(hdev
, DISCOVERY_STOPPED
);
2067 hci_dev_unlock(hdev
);
2072 static void le_scan_disable_work(struct work_struct
*work
)
2074 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
,
2075 le_scan_disable
.work
);
2076 struct hci_cp_le_set_scan_enable cp
;
2077 struct hci_request req
;
2080 BT_DBG("%s", hdev
->name
);
2082 hci_req_init(&req
, hdev
);
2084 memset(&cp
, 0, sizeof(cp
));
2085 cp
.enable
= LE_SCAN_DISABLE
;
2086 hci_req_add(&req
, HCI_OP_LE_SET_SCAN_ENABLE
, sizeof(cp
), &cp
);
2088 err
= hci_req_run(&req
, le_scan_disable_work_complete
);
2090 BT_ERR("Disable LE scanning request failed: err %d", err
);
2093 /* Alloc HCI device */
2094 struct hci_dev
*hci_alloc_dev(void)
2096 struct hci_dev
*hdev
;
2098 hdev
= kzalloc(sizeof(struct hci_dev
), GFP_KERNEL
);
2102 hdev
->pkt_type
= (HCI_DM1
| HCI_DH1
| HCI_HV1
);
2103 hdev
->esco_type
= (ESCO_HV1
);
2104 hdev
->link_mode
= (HCI_LM_ACCEPT
);
2105 hdev
->io_capability
= 0x03; /* No Input No Output */
2106 hdev
->inq_tx_power
= HCI_TX_POWER_INVALID
;
2107 hdev
->adv_tx_power
= HCI_TX_POWER_INVALID
;
2109 hdev
->sniff_max_interval
= 800;
2110 hdev
->sniff_min_interval
= 80;
2112 mutex_init(&hdev
->lock
);
2113 mutex_init(&hdev
->req_lock
);
2115 INIT_LIST_HEAD(&hdev
->mgmt_pending
);
2116 INIT_LIST_HEAD(&hdev
->blacklist
);
2117 INIT_LIST_HEAD(&hdev
->uuids
);
2118 INIT_LIST_HEAD(&hdev
->link_keys
);
2119 INIT_LIST_HEAD(&hdev
->long_term_keys
);
2120 INIT_LIST_HEAD(&hdev
->remote_oob_data
);
2121 INIT_LIST_HEAD(&hdev
->conn_hash
.list
);
2123 INIT_WORK(&hdev
->rx_work
, hci_rx_work
);
2124 INIT_WORK(&hdev
->cmd_work
, hci_cmd_work
);
2125 INIT_WORK(&hdev
->tx_work
, hci_tx_work
);
2126 INIT_WORK(&hdev
->power_on
, hci_power_on
);
2128 INIT_DELAYED_WORK(&hdev
->power_off
, hci_power_off
);
2129 INIT_DELAYED_WORK(&hdev
->discov_off
, hci_discov_off
);
2130 INIT_DELAYED_WORK(&hdev
->le_scan_disable
, le_scan_disable_work
);
2132 skb_queue_head_init(&hdev
->rx_q
);
2133 skb_queue_head_init(&hdev
->cmd_q
);
2134 skb_queue_head_init(&hdev
->raw_q
);
2136 init_waitqueue_head(&hdev
->req_wait_q
);
2138 setup_timer(&hdev
->cmd_timer
, hci_cmd_timeout
, (unsigned long) hdev
);
2140 hci_init_sysfs(hdev
);
2141 discovery_init(hdev
);
2145 EXPORT_SYMBOL(hci_alloc_dev
);
2147 /* Free HCI device */
2148 void hci_free_dev(struct hci_dev
*hdev
)
2150 /* will free via device release */
2151 put_device(&hdev
->dev
);
2153 EXPORT_SYMBOL(hci_free_dev
);
2155 /* Register HCI device */
2156 int hci_register_dev(struct hci_dev
*hdev
)
2160 if (!hdev
->open
|| !hdev
->close
)
2163 /* Do not allow HCI_AMP devices to register at index 0,
2164 * so the index can be used as the AMP controller ID.
2166 switch (hdev
->dev_type
) {
2168 id
= ida_simple_get(&hci_index_ida
, 0, 0, GFP_KERNEL
);
2171 id
= ida_simple_get(&hci_index_ida
, 1, 0, GFP_KERNEL
);
2180 sprintf(hdev
->name
, "hci%d", id
);
2183 BT_DBG("%p name %s bus %d", hdev
, hdev
->name
, hdev
->bus
);
2185 hdev
->workqueue
= alloc_workqueue("%s", WQ_HIGHPRI
| WQ_UNBOUND
|
2186 WQ_MEM_RECLAIM
, 1, hdev
->name
);
2187 if (!hdev
->workqueue
) {
2192 hdev
->req_workqueue
= alloc_workqueue("%s", WQ_HIGHPRI
| WQ_UNBOUND
|
2193 WQ_MEM_RECLAIM
, 1, hdev
->name
);
2194 if (!hdev
->req_workqueue
) {
2195 destroy_workqueue(hdev
->workqueue
);
2200 error
= hci_add_sysfs(hdev
);
2204 hdev
->rfkill
= rfkill_alloc(hdev
->name
, &hdev
->dev
,
2205 RFKILL_TYPE_BLUETOOTH
, &hci_rfkill_ops
,
2208 if (rfkill_register(hdev
->rfkill
) < 0) {
2209 rfkill_destroy(hdev
->rfkill
);
2210 hdev
->rfkill
= NULL
;
2214 set_bit(HCI_SETUP
, &hdev
->dev_flags
);
2216 if (hdev
->dev_type
!= HCI_AMP
)
2217 set_bit(HCI_AUTO_OFF
, &hdev
->dev_flags
);
2219 write_lock(&hci_dev_list_lock
);
2220 list_add(&hdev
->list
, &hci_dev_list
);
2221 write_unlock(&hci_dev_list_lock
);
2223 hci_notify(hdev
, HCI_DEV_REG
);
2226 queue_work(hdev
->req_workqueue
, &hdev
->power_on
);
2231 destroy_workqueue(hdev
->workqueue
);
2232 destroy_workqueue(hdev
->req_workqueue
);
2234 ida_simple_remove(&hci_index_ida
, hdev
->id
);
2238 EXPORT_SYMBOL(hci_register_dev
);
2240 /* Unregister HCI device */
2241 void hci_unregister_dev(struct hci_dev
*hdev
)
2245 BT_DBG("%p name %s bus %d", hdev
, hdev
->name
, hdev
->bus
);
2247 set_bit(HCI_UNREGISTER
, &hdev
->dev_flags
);
2251 write_lock(&hci_dev_list_lock
);
2252 list_del(&hdev
->list
);
2253 write_unlock(&hci_dev_list_lock
);
2255 hci_dev_do_close(hdev
);
2257 for (i
= 0; i
< NUM_REASSEMBLY
; i
++)
2258 kfree_skb(hdev
->reassembly
[i
]);
2260 cancel_work_sync(&hdev
->power_on
);
2262 if (!test_bit(HCI_INIT
, &hdev
->flags
) &&
2263 !test_bit(HCI_SETUP
, &hdev
->dev_flags
)) {
2265 mgmt_index_removed(hdev
);
2266 hci_dev_unlock(hdev
);
2269 /* mgmt_index_removed should take care of emptying the
2271 BUG_ON(!list_empty(&hdev
->mgmt_pending
));
2273 hci_notify(hdev
, HCI_DEV_UNREG
);
2276 rfkill_unregister(hdev
->rfkill
);
2277 rfkill_destroy(hdev
->rfkill
);
2280 hci_del_sysfs(hdev
);
2282 destroy_workqueue(hdev
->workqueue
);
2283 destroy_workqueue(hdev
->req_workqueue
);
2286 hci_blacklist_clear(hdev
);
2287 hci_uuids_clear(hdev
);
2288 hci_link_keys_clear(hdev
);
2289 hci_smp_ltks_clear(hdev
);
2290 hci_remote_oob_data_clear(hdev
);
2291 hci_dev_unlock(hdev
);
2295 ida_simple_remove(&hci_index_ida
, id
);
2297 EXPORT_SYMBOL(hci_unregister_dev
);
2299 /* Suspend HCI device */
2300 int hci_suspend_dev(struct hci_dev
*hdev
)
2302 hci_notify(hdev
, HCI_DEV_SUSPEND
);
2305 EXPORT_SYMBOL(hci_suspend_dev
);
2307 /* Resume HCI device */
2308 int hci_resume_dev(struct hci_dev
*hdev
)
2310 hci_notify(hdev
, HCI_DEV_RESUME
);
2313 EXPORT_SYMBOL(hci_resume_dev
);
2315 /* Receive frame from HCI drivers */
2316 int hci_recv_frame(struct sk_buff
*skb
)
2318 struct hci_dev
*hdev
= (struct hci_dev
*) skb
->dev
;
2319 if (!hdev
|| (!test_bit(HCI_UP
, &hdev
->flags
)
2320 && !test_bit(HCI_INIT
, &hdev
->flags
))) {
2326 bt_cb(skb
)->incoming
= 1;
2329 __net_timestamp(skb
);
2331 skb_queue_tail(&hdev
->rx_q
, skb
);
2332 queue_work(hdev
->workqueue
, &hdev
->rx_work
);
2336 EXPORT_SYMBOL(hci_recv_frame
);
2338 static int hci_reassembly(struct hci_dev
*hdev
, int type
, void *data
,
2339 int count
, __u8 index
)
2344 struct sk_buff
*skb
;
2345 struct bt_skb_cb
*scb
;
2347 if ((type
< HCI_ACLDATA_PKT
|| type
> HCI_EVENT_PKT
) ||
2348 index
>= NUM_REASSEMBLY
)
2351 skb
= hdev
->reassembly
[index
];
2355 case HCI_ACLDATA_PKT
:
2356 len
= HCI_MAX_FRAME_SIZE
;
2357 hlen
= HCI_ACL_HDR_SIZE
;
2360 len
= HCI_MAX_EVENT_SIZE
;
2361 hlen
= HCI_EVENT_HDR_SIZE
;
2363 case HCI_SCODATA_PKT
:
2364 len
= HCI_MAX_SCO_SIZE
;
2365 hlen
= HCI_SCO_HDR_SIZE
;
2369 skb
= bt_skb_alloc(len
, GFP_ATOMIC
);
2373 scb
= (void *) skb
->cb
;
2375 scb
->pkt_type
= type
;
2377 skb
->dev
= (void *) hdev
;
2378 hdev
->reassembly
[index
] = skb
;
2382 scb
= (void *) skb
->cb
;
2383 len
= min_t(uint
, scb
->expect
, count
);
2385 memcpy(skb_put(skb
, len
), data
, len
);
2394 if (skb
->len
== HCI_EVENT_HDR_SIZE
) {
2395 struct hci_event_hdr
*h
= hci_event_hdr(skb
);
2396 scb
->expect
= h
->plen
;
2398 if (skb_tailroom(skb
) < scb
->expect
) {
2400 hdev
->reassembly
[index
] = NULL
;
2406 case HCI_ACLDATA_PKT
:
2407 if (skb
->len
== HCI_ACL_HDR_SIZE
) {
2408 struct hci_acl_hdr
*h
= hci_acl_hdr(skb
);
2409 scb
->expect
= __le16_to_cpu(h
->dlen
);
2411 if (skb_tailroom(skb
) < scb
->expect
) {
2413 hdev
->reassembly
[index
] = NULL
;
2419 case HCI_SCODATA_PKT
:
2420 if (skb
->len
== HCI_SCO_HDR_SIZE
) {
2421 struct hci_sco_hdr
*h
= hci_sco_hdr(skb
);
2422 scb
->expect
= h
->dlen
;
2424 if (skb_tailroom(skb
) < scb
->expect
) {
2426 hdev
->reassembly
[index
] = NULL
;
2433 if (scb
->expect
== 0) {
2434 /* Complete frame */
2436 bt_cb(skb
)->pkt_type
= type
;
2437 hci_recv_frame(skb
);
2439 hdev
->reassembly
[index
] = NULL
;
2447 int hci_recv_fragment(struct hci_dev
*hdev
, int type
, void *data
, int count
)
2451 if (type
< HCI_ACLDATA_PKT
|| type
> HCI_EVENT_PKT
)
2455 rem
= hci_reassembly(hdev
, type
, data
, count
, type
- 1);
2459 data
+= (count
- rem
);
2465 EXPORT_SYMBOL(hci_recv_fragment
);
2467 #define STREAM_REASSEMBLY 0
2469 int hci_recv_stream_fragment(struct hci_dev
*hdev
, void *data
, int count
)
2475 struct sk_buff
*skb
= hdev
->reassembly
[STREAM_REASSEMBLY
];
2478 struct { char type
; } *pkt
;
2480 /* Start of the frame */
2487 type
= bt_cb(skb
)->pkt_type
;
2489 rem
= hci_reassembly(hdev
, type
, data
, count
,
2494 data
+= (count
- rem
);
2500 EXPORT_SYMBOL(hci_recv_stream_fragment
);
2502 /* ---- Interface to upper protocols ---- */
2504 int hci_register_cb(struct hci_cb
*cb
)
2506 BT_DBG("%p name %s", cb
, cb
->name
);
2508 write_lock(&hci_cb_list_lock
);
2509 list_add(&cb
->list
, &hci_cb_list
);
2510 write_unlock(&hci_cb_list_lock
);
2514 EXPORT_SYMBOL(hci_register_cb
);
2516 int hci_unregister_cb(struct hci_cb
*cb
)
2518 BT_DBG("%p name %s", cb
, cb
->name
);
2520 write_lock(&hci_cb_list_lock
);
2521 list_del(&cb
->list
);
2522 write_unlock(&hci_cb_list_lock
);
2526 EXPORT_SYMBOL(hci_unregister_cb
);
2528 static int hci_send_frame(struct sk_buff
*skb
)
2530 struct hci_dev
*hdev
= (struct hci_dev
*) skb
->dev
;
2537 BT_DBG("%s type %d len %d", hdev
->name
, bt_cb(skb
)->pkt_type
, skb
->len
);
2540 __net_timestamp(skb
);
2542 /* Send copy to monitor */
2543 hci_send_to_monitor(hdev
, skb
);
2545 if (atomic_read(&hdev
->promisc
)) {
2546 /* Send copy to the sockets */
2547 hci_send_to_sock(hdev
, skb
);
2550 /* Get rid of skb owner, prior to sending to the driver. */
2553 return hdev
->send(skb
);
2556 void hci_req_init(struct hci_request
*req
, struct hci_dev
*hdev
)
2558 skb_queue_head_init(&req
->cmd_q
);
2563 int hci_req_run(struct hci_request
*req
, hci_req_complete_t complete
)
2565 struct hci_dev
*hdev
= req
->hdev
;
2566 struct sk_buff
*skb
;
2567 unsigned long flags
;
2569 BT_DBG("length %u", skb_queue_len(&req
->cmd_q
));
2571 /* If an error occured during request building, remove all HCI
2572 * commands queued on the HCI request queue.
2575 skb_queue_purge(&req
->cmd_q
);
2579 /* Do not allow empty requests */
2580 if (skb_queue_empty(&req
->cmd_q
))
2583 skb
= skb_peek_tail(&req
->cmd_q
);
2584 bt_cb(skb
)->req
.complete
= complete
;
2586 spin_lock_irqsave(&hdev
->cmd_q
.lock
, flags
);
2587 skb_queue_splice_tail(&req
->cmd_q
, &hdev
->cmd_q
);
2588 spin_unlock_irqrestore(&hdev
->cmd_q
.lock
, flags
);
2590 queue_work(hdev
->workqueue
, &hdev
->cmd_work
);
2595 static struct sk_buff
*hci_prepare_cmd(struct hci_dev
*hdev
, u16 opcode
,
2596 u32 plen
, const void *param
)
2598 int len
= HCI_COMMAND_HDR_SIZE
+ plen
;
2599 struct hci_command_hdr
*hdr
;
2600 struct sk_buff
*skb
;
2602 skb
= bt_skb_alloc(len
, GFP_ATOMIC
);
2606 hdr
= (struct hci_command_hdr
*) skb_put(skb
, HCI_COMMAND_HDR_SIZE
);
2607 hdr
->opcode
= cpu_to_le16(opcode
);
2611 memcpy(skb_put(skb
, plen
), param
, plen
);
2613 BT_DBG("skb len %d", skb
->len
);
2615 bt_cb(skb
)->pkt_type
= HCI_COMMAND_PKT
;
2616 skb
->dev
= (void *) hdev
;
2621 /* Send HCI command */
2622 int hci_send_cmd(struct hci_dev
*hdev
, __u16 opcode
, __u32 plen
,
2625 struct sk_buff
*skb
;
2627 BT_DBG("%s opcode 0x%4.4x plen %d", hdev
->name
, opcode
, plen
);
2629 skb
= hci_prepare_cmd(hdev
, opcode
, plen
, param
);
2631 BT_ERR("%s no memory for command", hdev
->name
);
2635 /* Stand-alone HCI commands must be flaged as
2636 * single-command requests.
2638 bt_cb(skb
)->req
.start
= true;
2640 skb_queue_tail(&hdev
->cmd_q
, skb
);
2641 queue_work(hdev
->workqueue
, &hdev
->cmd_work
);
2646 /* Queue a command to an asynchronous HCI request */
2647 void hci_req_add_ev(struct hci_request
*req
, u16 opcode
, u32 plen
,
2648 const void *param
, u8 event
)
2650 struct hci_dev
*hdev
= req
->hdev
;
2651 struct sk_buff
*skb
;
2653 BT_DBG("%s opcode 0x%4.4x plen %d", hdev
->name
, opcode
, plen
);
2655 /* If an error occured during request building, there is no point in
2656 * queueing the HCI command. We can simply return.
2661 skb
= hci_prepare_cmd(hdev
, opcode
, plen
, param
);
2663 BT_ERR("%s no memory for command (opcode 0x%4.4x)",
2664 hdev
->name
, opcode
);
2669 if (skb_queue_empty(&req
->cmd_q
))
2670 bt_cb(skb
)->req
.start
= true;
2672 bt_cb(skb
)->req
.event
= event
;
2674 skb_queue_tail(&req
->cmd_q
, skb
);
2677 void hci_req_add(struct hci_request
*req
, u16 opcode
, u32 plen
,
2680 hci_req_add_ev(req
, opcode
, plen
, param
, 0);
2683 /* Get data from the previously sent command */
2684 void *hci_sent_cmd_data(struct hci_dev
*hdev
, __u16 opcode
)
2686 struct hci_command_hdr
*hdr
;
2688 if (!hdev
->sent_cmd
)
2691 hdr
= (void *) hdev
->sent_cmd
->data
;
2693 if (hdr
->opcode
!= cpu_to_le16(opcode
))
2696 BT_DBG("%s opcode 0x%4.4x", hdev
->name
, opcode
);
2698 return hdev
->sent_cmd
->data
+ HCI_COMMAND_HDR_SIZE
;
2702 static void hci_add_acl_hdr(struct sk_buff
*skb
, __u16 handle
, __u16 flags
)
2704 struct hci_acl_hdr
*hdr
;
2707 skb_push(skb
, HCI_ACL_HDR_SIZE
);
2708 skb_reset_transport_header(skb
);
2709 hdr
= (struct hci_acl_hdr
*)skb_transport_header(skb
);
2710 hdr
->handle
= cpu_to_le16(hci_handle_pack(handle
, flags
));
2711 hdr
->dlen
= cpu_to_le16(len
);
2714 static void hci_queue_acl(struct hci_chan
*chan
, struct sk_buff_head
*queue
,
2715 struct sk_buff
*skb
, __u16 flags
)
2717 struct hci_conn
*conn
= chan
->conn
;
2718 struct hci_dev
*hdev
= conn
->hdev
;
2719 struct sk_buff
*list
;
2721 skb
->len
= skb_headlen(skb
);
2724 bt_cb(skb
)->pkt_type
= HCI_ACLDATA_PKT
;
2726 switch (hdev
->dev_type
) {
2728 hci_add_acl_hdr(skb
, conn
->handle
, flags
);
2731 hci_add_acl_hdr(skb
, chan
->handle
, flags
);
2734 BT_ERR("%s unknown dev_type %d", hdev
->name
, hdev
->dev_type
);
2738 list
= skb_shinfo(skb
)->frag_list
;
2740 /* Non fragmented */
2741 BT_DBG("%s nonfrag skb %p len %d", hdev
->name
, skb
, skb
->len
);
2743 skb_queue_tail(queue
, skb
);
2746 BT_DBG("%s frag %p len %d", hdev
->name
, skb
, skb
->len
);
2748 skb_shinfo(skb
)->frag_list
= NULL
;
2750 /* Queue all fragments atomically */
2751 spin_lock(&queue
->lock
);
2753 __skb_queue_tail(queue
, skb
);
2755 flags
&= ~ACL_START
;
2758 skb
= list
; list
= list
->next
;
2760 skb
->dev
= (void *) hdev
;
2761 bt_cb(skb
)->pkt_type
= HCI_ACLDATA_PKT
;
2762 hci_add_acl_hdr(skb
, conn
->handle
, flags
);
2764 BT_DBG("%s frag %p len %d", hdev
->name
, skb
, skb
->len
);
2766 __skb_queue_tail(queue
, skb
);
2769 spin_unlock(&queue
->lock
);
2773 void hci_send_acl(struct hci_chan
*chan
, struct sk_buff
*skb
, __u16 flags
)
2775 struct hci_dev
*hdev
= chan
->conn
->hdev
;
2777 BT_DBG("%s chan %p flags 0x%4.4x", hdev
->name
, chan
, flags
);
2779 skb
->dev
= (void *) hdev
;
2781 hci_queue_acl(chan
, &chan
->data_q
, skb
, flags
);
2783 queue_work(hdev
->workqueue
, &hdev
->tx_work
);
2787 void hci_send_sco(struct hci_conn
*conn
, struct sk_buff
*skb
)
2789 struct hci_dev
*hdev
= conn
->hdev
;
2790 struct hci_sco_hdr hdr
;
2792 BT_DBG("%s len %d", hdev
->name
, skb
->len
);
2794 hdr
.handle
= cpu_to_le16(conn
->handle
);
2795 hdr
.dlen
= skb
->len
;
2797 skb_push(skb
, HCI_SCO_HDR_SIZE
);
2798 skb_reset_transport_header(skb
);
2799 memcpy(skb_transport_header(skb
), &hdr
, HCI_SCO_HDR_SIZE
);
2801 skb
->dev
= (void *) hdev
;
2802 bt_cb(skb
)->pkt_type
= HCI_SCODATA_PKT
;
2804 skb_queue_tail(&conn
->data_q
, skb
);
2805 queue_work(hdev
->workqueue
, &hdev
->tx_work
);
2808 /* ---- HCI TX task (outgoing data) ---- */
2810 /* HCI Connection scheduler */
2811 static struct hci_conn
*hci_low_sent(struct hci_dev
*hdev
, __u8 type
,
2814 struct hci_conn_hash
*h
= &hdev
->conn_hash
;
2815 struct hci_conn
*conn
= NULL
, *c
;
2816 unsigned int num
= 0, min
= ~0;
2818 /* We don't have to lock device here. Connections are always
2819 * added and removed with TX task disabled. */
2823 list_for_each_entry_rcu(c
, &h
->list
, list
) {
2824 if (c
->type
!= type
|| skb_queue_empty(&c
->data_q
))
2827 if (c
->state
!= BT_CONNECTED
&& c
->state
!= BT_CONFIG
)
2832 if (c
->sent
< min
) {
2837 if (hci_conn_num(hdev
, type
) == num
)
2846 switch (conn
->type
) {
2848 cnt
= hdev
->acl_cnt
;
2852 cnt
= hdev
->sco_cnt
;
2855 cnt
= hdev
->le_mtu
? hdev
->le_cnt
: hdev
->acl_cnt
;
2859 BT_ERR("Unknown link type");
2867 BT_DBG("conn %p quote %d", conn
, *quote
);
2871 static void hci_link_tx_to(struct hci_dev
*hdev
, __u8 type
)
2873 struct hci_conn_hash
*h
= &hdev
->conn_hash
;
2876 BT_ERR("%s link tx timeout", hdev
->name
);
2880 /* Kill stalled connections */
2881 list_for_each_entry_rcu(c
, &h
->list
, list
) {
2882 if (c
->type
== type
&& c
->sent
) {
2883 BT_ERR("%s killing stalled connection %pMR",
2884 hdev
->name
, &c
->dst
);
2885 hci_disconnect(c
, HCI_ERROR_REMOTE_USER_TERM
);
2892 static struct hci_chan
*hci_chan_sent(struct hci_dev
*hdev
, __u8 type
,
2895 struct hci_conn_hash
*h
= &hdev
->conn_hash
;
2896 struct hci_chan
*chan
= NULL
;
2897 unsigned int num
= 0, min
= ~0, cur_prio
= 0;
2898 struct hci_conn
*conn
;
2899 int cnt
, q
, conn_num
= 0;
2901 BT_DBG("%s", hdev
->name
);
2905 list_for_each_entry_rcu(conn
, &h
->list
, list
) {
2906 struct hci_chan
*tmp
;
2908 if (conn
->type
!= type
)
2911 if (conn
->state
!= BT_CONNECTED
&& conn
->state
!= BT_CONFIG
)
2916 list_for_each_entry_rcu(tmp
, &conn
->chan_list
, list
) {
2917 struct sk_buff
*skb
;
2919 if (skb_queue_empty(&tmp
->data_q
))
2922 skb
= skb_peek(&tmp
->data_q
);
2923 if (skb
->priority
< cur_prio
)
2926 if (skb
->priority
> cur_prio
) {
2929 cur_prio
= skb
->priority
;
2934 if (conn
->sent
< min
) {
2940 if (hci_conn_num(hdev
, type
) == conn_num
)
2949 switch (chan
->conn
->type
) {
2951 cnt
= hdev
->acl_cnt
;
2954 cnt
= hdev
->block_cnt
;
2958 cnt
= hdev
->sco_cnt
;
2961 cnt
= hdev
->le_mtu
? hdev
->le_cnt
: hdev
->acl_cnt
;
2965 BT_ERR("Unknown link type");
2970 BT_DBG("chan %p quote %d", chan
, *quote
);
2974 static void hci_prio_recalculate(struct hci_dev
*hdev
, __u8 type
)
2976 struct hci_conn_hash
*h
= &hdev
->conn_hash
;
2977 struct hci_conn
*conn
;
2980 BT_DBG("%s", hdev
->name
);
2984 list_for_each_entry_rcu(conn
, &h
->list
, list
) {
2985 struct hci_chan
*chan
;
2987 if (conn
->type
!= type
)
2990 if (conn
->state
!= BT_CONNECTED
&& conn
->state
!= BT_CONFIG
)
2995 list_for_each_entry_rcu(chan
, &conn
->chan_list
, list
) {
2996 struct sk_buff
*skb
;
3003 if (skb_queue_empty(&chan
->data_q
))
3006 skb
= skb_peek(&chan
->data_q
);
3007 if (skb
->priority
>= HCI_PRIO_MAX
- 1)
3010 skb
->priority
= HCI_PRIO_MAX
- 1;
3012 BT_DBG("chan %p skb %p promoted to %d", chan
, skb
,
3016 if (hci_conn_num(hdev
, type
) == num
)
3024 static inline int __get_blocks(struct hci_dev
*hdev
, struct sk_buff
*skb
)
3026 /* Calculate count of blocks used by this packet */
3027 return DIV_ROUND_UP(skb
->len
- HCI_ACL_HDR_SIZE
, hdev
->block_len
);
3030 static void __check_timeout(struct hci_dev
*hdev
, unsigned int cnt
)
3032 if (!test_bit(HCI_RAW
, &hdev
->flags
)) {
3033 /* ACL tx timeout must be longer than maximum
3034 * link supervision timeout (40.9 seconds) */
3035 if (!cnt
&& time_after(jiffies
, hdev
->acl_last_tx
+
3036 HCI_ACL_TX_TIMEOUT
))
3037 hci_link_tx_to(hdev
, ACL_LINK
);
3041 static void hci_sched_acl_pkt(struct hci_dev
*hdev
)
3043 unsigned int cnt
= hdev
->acl_cnt
;
3044 struct hci_chan
*chan
;
3045 struct sk_buff
*skb
;
3048 __check_timeout(hdev
, cnt
);
3050 while (hdev
->acl_cnt
&&
3051 (chan
= hci_chan_sent(hdev
, ACL_LINK
, "e
))) {
3052 u32 priority
= (skb_peek(&chan
->data_q
))->priority
;
3053 while (quote
-- && (skb
= skb_peek(&chan
->data_q
))) {
3054 BT_DBG("chan %p skb %p len %d priority %u", chan
, skb
,
3055 skb
->len
, skb
->priority
);
3057 /* Stop if priority has changed */
3058 if (skb
->priority
< priority
)
3061 skb
= skb_dequeue(&chan
->data_q
);
3063 hci_conn_enter_active_mode(chan
->conn
,
3064 bt_cb(skb
)->force_active
);
3066 hci_send_frame(skb
);
3067 hdev
->acl_last_tx
= jiffies
;
3075 if (cnt
!= hdev
->acl_cnt
)
3076 hci_prio_recalculate(hdev
, ACL_LINK
);
3079 static void hci_sched_acl_blk(struct hci_dev
*hdev
)
3081 unsigned int cnt
= hdev
->block_cnt
;
3082 struct hci_chan
*chan
;
3083 struct sk_buff
*skb
;
3087 __check_timeout(hdev
, cnt
);
3089 BT_DBG("%s", hdev
->name
);
3091 if (hdev
->dev_type
== HCI_AMP
)
3096 while (hdev
->block_cnt
> 0 &&
3097 (chan
= hci_chan_sent(hdev
, type
, "e
))) {
3098 u32 priority
= (skb_peek(&chan
->data_q
))->priority
;
3099 while (quote
> 0 && (skb
= skb_peek(&chan
->data_q
))) {
3102 BT_DBG("chan %p skb %p len %d priority %u", chan
, skb
,
3103 skb
->len
, skb
->priority
);
3105 /* Stop if priority has changed */
3106 if (skb
->priority
< priority
)
3109 skb
= skb_dequeue(&chan
->data_q
);
3111 blocks
= __get_blocks(hdev
, skb
);
3112 if (blocks
> hdev
->block_cnt
)
3115 hci_conn_enter_active_mode(chan
->conn
,
3116 bt_cb(skb
)->force_active
);
3118 hci_send_frame(skb
);
3119 hdev
->acl_last_tx
= jiffies
;
3121 hdev
->block_cnt
-= blocks
;
3124 chan
->sent
+= blocks
;
3125 chan
->conn
->sent
+= blocks
;
3129 if (cnt
!= hdev
->block_cnt
)
3130 hci_prio_recalculate(hdev
, type
);
3133 static void hci_sched_acl(struct hci_dev
*hdev
)
3135 BT_DBG("%s", hdev
->name
);
3137 /* No ACL link over BR/EDR controller */
3138 if (!hci_conn_num(hdev
, ACL_LINK
) && hdev
->dev_type
== HCI_BREDR
)
3141 /* No AMP link over AMP controller */
3142 if (!hci_conn_num(hdev
, AMP_LINK
) && hdev
->dev_type
== HCI_AMP
)
3145 switch (hdev
->flow_ctl_mode
) {
3146 case HCI_FLOW_CTL_MODE_PACKET_BASED
:
3147 hci_sched_acl_pkt(hdev
);
3150 case HCI_FLOW_CTL_MODE_BLOCK_BASED
:
3151 hci_sched_acl_blk(hdev
);
3157 static void hci_sched_sco(struct hci_dev
*hdev
)
3159 struct hci_conn
*conn
;
3160 struct sk_buff
*skb
;
3163 BT_DBG("%s", hdev
->name
);
3165 if (!hci_conn_num(hdev
, SCO_LINK
))
3168 while (hdev
->sco_cnt
&& (conn
= hci_low_sent(hdev
, SCO_LINK
, "e
))) {
3169 while (quote
-- && (skb
= skb_dequeue(&conn
->data_q
))) {
3170 BT_DBG("skb %p len %d", skb
, skb
->len
);
3171 hci_send_frame(skb
);
3174 if (conn
->sent
== ~0)
3180 static void hci_sched_esco(struct hci_dev
*hdev
)
3182 struct hci_conn
*conn
;
3183 struct sk_buff
*skb
;
3186 BT_DBG("%s", hdev
->name
);
3188 if (!hci_conn_num(hdev
, ESCO_LINK
))
3191 while (hdev
->sco_cnt
&& (conn
= hci_low_sent(hdev
, ESCO_LINK
,
3193 while (quote
-- && (skb
= skb_dequeue(&conn
->data_q
))) {
3194 BT_DBG("skb %p len %d", skb
, skb
->len
);
3195 hci_send_frame(skb
);
3198 if (conn
->sent
== ~0)
3204 static void hci_sched_le(struct hci_dev
*hdev
)
3206 struct hci_chan
*chan
;
3207 struct sk_buff
*skb
;
3208 int quote
, cnt
, tmp
;
3210 BT_DBG("%s", hdev
->name
);
3212 if (!hci_conn_num(hdev
, LE_LINK
))
3215 if (!test_bit(HCI_RAW
, &hdev
->flags
)) {
3216 /* LE tx timeout must be longer than maximum
3217 * link supervision timeout (40.9 seconds) */
3218 if (!hdev
->le_cnt
&& hdev
->le_pkts
&&
3219 time_after(jiffies
, hdev
->le_last_tx
+ HZ
* 45))
3220 hci_link_tx_to(hdev
, LE_LINK
);
3223 cnt
= hdev
->le_pkts
? hdev
->le_cnt
: hdev
->acl_cnt
;
3225 while (cnt
&& (chan
= hci_chan_sent(hdev
, LE_LINK
, "e
))) {
3226 u32 priority
= (skb_peek(&chan
->data_q
))->priority
;
3227 while (quote
-- && (skb
= skb_peek(&chan
->data_q
))) {
3228 BT_DBG("chan %p skb %p len %d priority %u", chan
, skb
,
3229 skb
->len
, skb
->priority
);
3231 /* Stop if priority has changed */
3232 if (skb
->priority
< priority
)
3235 skb
= skb_dequeue(&chan
->data_q
);
3237 hci_send_frame(skb
);
3238 hdev
->le_last_tx
= jiffies
;
3249 hdev
->acl_cnt
= cnt
;
3252 hci_prio_recalculate(hdev
, LE_LINK
);
3255 static void hci_tx_work(struct work_struct
*work
)
3257 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
, tx_work
);
3258 struct sk_buff
*skb
;
3260 BT_DBG("%s acl %d sco %d le %d", hdev
->name
, hdev
->acl_cnt
,
3261 hdev
->sco_cnt
, hdev
->le_cnt
);
3263 /* Schedule queues and send stuff to HCI driver */
3265 hci_sched_acl(hdev
);
3267 hci_sched_sco(hdev
);
3269 hci_sched_esco(hdev
);
3273 /* Send next queued raw (unknown type) packet */
3274 while ((skb
= skb_dequeue(&hdev
->raw_q
)))
3275 hci_send_frame(skb
);
3278 /* ----- HCI RX task (incoming data processing) ----- */
3280 /* ACL data packet */
3281 static void hci_acldata_packet(struct hci_dev
*hdev
, struct sk_buff
*skb
)
3283 struct hci_acl_hdr
*hdr
= (void *) skb
->data
;
3284 struct hci_conn
*conn
;
3285 __u16 handle
, flags
;
3287 skb_pull(skb
, HCI_ACL_HDR_SIZE
);
3289 handle
= __le16_to_cpu(hdr
->handle
);
3290 flags
= hci_flags(handle
);
3291 handle
= hci_handle(handle
);
3293 BT_DBG("%s len %d handle 0x%4.4x flags 0x%4.4x", hdev
->name
, skb
->len
,
3296 hdev
->stat
.acl_rx
++;
3299 conn
= hci_conn_hash_lookup_handle(hdev
, handle
);
3300 hci_dev_unlock(hdev
);
3303 hci_conn_enter_active_mode(conn
, BT_POWER_FORCE_ACTIVE_OFF
);
3305 /* Send to upper protocol */
3306 l2cap_recv_acldata(conn
, skb
, flags
);
3309 BT_ERR("%s ACL packet for unknown connection handle %d",
3310 hdev
->name
, handle
);
3316 /* SCO data packet */
3317 static void hci_scodata_packet(struct hci_dev
*hdev
, struct sk_buff
*skb
)
3319 struct hci_sco_hdr
*hdr
= (void *) skb
->data
;
3320 struct hci_conn
*conn
;
3323 skb_pull(skb
, HCI_SCO_HDR_SIZE
);
3325 handle
= __le16_to_cpu(hdr
->handle
);
3327 BT_DBG("%s len %d handle 0x%4.4x", hdev
->name
, skb
->len
, handle
);
3329 hdev
->stat
.sco_rx
++;
3332 conn
= hci_conn_hash_lookup_handle(hdev
, handle
);
3333 hci_dev_unlock(hdev
);
3336 /* Send to upper protocol */
3337 sco_recv_scodata(conn
, skb
);
3340 BT_ERR("%s SCO packet for unknown connection handle %d",
3341 hdev
->name
, handle
);
3347 static bool hci_req_is_complete(struct hci_dev
*hdev
)
3349 struct sk_buff
*skb
;
3351 skb
= skb_peek(&hdev
->cmd_q
);
3355 return bt_cb(skb
)->req
.start
;
3358 static void hci_resend_last(struct hci_dev
*hdev
)
3360 struct hci_command_hdr
*sent
;
3361 struct sk_buff
*skb
;
3364 if (!hdev
->sent_cmd
)
3367 sent
= (void *) hdev
->sent_cmd
->data
;
3368 opcode
= __le16_to_cpu(sent
->opcode
);
3369 if (opcode
== HCI_OP_RESET
)
3372 skb
= skb_clone(hdev
->sent_cmd
, GFP_KERNEL
);
3376 skb_queue_head(&hdev
->cmd_q
, skb
);
3377 queue_work(hdev
->workqueue
, &hdev
->cmd_work
);
3380 void hci_req_cmd_complete(struct hci_dev
*hdev
, u16 opcode
, u8 status
)
3382 hci_req_complete_t req_complete
= NULL
;
3383 struct sk_buff
*skb
;
3384 unsigned long flags
;
3386 BT_DBG("opcode 0x%04x status 0x%02x", opcode
, status
);
3388 /* If the completed command doesn't match the last one that was
3389 * sent we need to do special handling of it.
3391 if (!hci_sent_cmd_data(hdev
, opcode
)) {
3392 /* Some CSR based controllers generate a spontaneous
3393 * reset complete event during init and any pending
3394 * command will never be completed. In such a case we
3395 * need to resend whatever was the last sent
3398 if (test_bit(HCI_INIT
, &hdev
->flags
) && opcode
== HCI_OP_RESET
)
3399 hci_resend_last(hdev
);
3404 /* If the command succeeded and there's still more commands in
3405 * this request the request is not yet complete.
3407 if (!status
&& !hci_req_is_complete(hdev
))
3410 /* If this was the last command in a request the complete
3411 * callback would be found in hdev->sent_cmd instead of the
3412 * command queue (hdev->cmd_q).
3414 if (hdev
->sent_cmd
) {
3415 req_complete
= bt_cb(hdev
->sent_cmd
)->req
.complete
;
3418 /* We must set the complete callback to NULL to
3419 * avoid calling the callback more than once if
3420 * this function gets called again.
3422 bt_cb(hdev
->sent_cmd
)->req
.complete
= NULL
;
3428 /* Remove all pending commands belonging to this request */
3429 spin_lock_irqsave(&hdev
->cmd_q
.lock
, flags
);
3430 while ((skb
= __skb_dequeue(&hdev
->cmd_q
))) {
3431 if (bt_cb(skb
)->req
.start
) {
3432 __skb_queue_head(&hdev
->cmd_q
, skb
);
3436 req_complete
= bt_cb(skb
)->req
.complete
;
3439 spin_unlock_irqrestore(&hdev
->cmd_q
.lock
, flags
);
3443 req_complete(hdev
, status
);
3446 static void hci_rx_work(struct work_struct
*work
)
3448 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
, rx_work
);
3449 struct sk_buff
*skb
;
3451 BT_DBG("%s", hdev
->name
);
3453 while ((skb
= skb_dequeue(&hdev
->rx_q
))) {
3454 /* Send copy to monitor */
3455 hci_send_to_monitor(hdev
, skb
);
3457 if (atomic_read(&hdev
->promisc
)) {
3458 /* Send copy to the sockets */
3459 hci_send_to_sock(hdev
, skb
);
3462 if (test_bit(HCI_RAW
, &hdev
->flags
)) {
3467 if (test_bit(HCI_INIT
, &hdev
->flags
)) {
3468 /* Don't process data packets in this states. */
3469 switch (bt_cb(skb
)->pkt_type
) {
3470 case HCI_ACLDATA_PKT
:
3471 case HCI_SCODATA_PKT
:
3478 switch (bt_cb(skb
)->pkt_type
) {
3480 BT_DBG("%s Event packet", hdev
->name
);
3481 hci_event_packet(hdev
, skb
);
3484 case HCI_ACLDATA_PKT
:
3485 BT_DBG("%s ACL data packet", hdev
->name
);
3486 hci_acldata_packet(hdev
, skb
);
3489 case HCI_SCODATA_PKT
:
3490 BT_DBG("%s SCO data packet", hdev
->name
);
3491 hci_scodata_packet(hdev
, skb
);
3501 static void hci_cmd_work(struct work_struct
*work
)
3503 struct hci_dev
*hdev
= container_of(work
, struct hci_dev
, cmd_work
);
3504 struct sk_buff
*skb
;
3506 BT_DBG("%s cmd_cnt %d cmd queued %d", hdev
->name
,
3507 atomic_read(&hdev
->cmd_cnt
), skb_queue_len(&hdev
->cmd_q
));
3509 /* Send queued commands */
3510 if (atomic_read(&hdev
->cmd_cnt
)) {
3511 skb
= skb_dequeue(&hdev
->cmd_q
);
3515 kfree_skb(hdev
->sent_cmd
);
3517 hdev
->sent_cmd
= skb_clone(skb
, GFP_ATOMIC
);
3518 if (hdev
->sent_cmd
) {
3519 atomic_dec(&hdev
->cmd_cnt
);
3520 hci_send_frame(skb
);
3521 if (test_bit(HCI_RESET
, &hdev
->flags
))
3522 del_timer(&hdev
->cmd_timer
);
3524 mod_timer(&hdev
->cmd_timer
,
3525 jiffies
+ HCI_CMD_TIMEOUT
);
3527 skb_queue_head(&hdev
->cmd_q
, skb
);
3528 queue_work(hdev
->workqueue
, &hdev
->cmd_work
);
3533 u8
bdaddr_to_le(u8 bdaddr_type
)
3535 switch (bdaddr_type
) {
3536 case BDADDR_LE_PUBLIC
:
3537 return ADDR_LE_DEV_PUBLIC
;
3540 /* Fallback to LE Random address type */
3541 return ADDR_LE_DEV_RANDOM
;