]> git.proxmox.com Git - mirror_ubuntu-eoan-kernel.git/blob - include/net/bluetooth/hci_core.h
Bluetooth: Turn hci_recv_frame into an exported function
[mirror_ubuntu-eoan-kernel.git] / include / net / bluetooth / hci_core.h
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 #ifndef __HCI_CORE_H
26 #define __HCI_CORE_H
27
28 #include <net/bluetooth/hci.h>
29
30 /* HCI upper protocols */
31 #define HCI_PROTO_L2CAP 0
32 #define HCI_PROTO_SCO 1
33
34 /* HCI Core structures */
35 struct inquiry_data {
36 bdaddr_t bdaddr;
37 __u8 pscan_rep_mode;
38 __u8 pscan_period_mode;
39 __u8 pscan_mode;
40 __u8 dev_class[3];
41 __le16 clock_offset;
42 __s8 rssi;
43 __u8 ssp_mode;
44 };
45
46 struct inquiry_entry {
47 struct inquiry_entry *next;
48 __u32 timestamp;
49 struct inquiry_data data;
50 };
51
52 struct inquiry_cache {
53 spinlock_t lock;
54 __u32 timestamp;
55 struct inquiry_entry *list;
56 };
57
58 struct hci_conn_hash {
59 struct list_head list;
60 spinlock_t lock;
61 unsigned int acl_num;
62 unsigned int sco_num;
63 };
64
65 struct hci_dev {
66 struct list_head list;
67 spinlock_t lock;
68 atomic_t refcnt;
69
70 char name[8];
71 unsigned long flags;
72 __u16 id;
73 __u8 type;
74 bdaddr_t bdaddr;
75 __u8 dev_name[248];
76 __u8 dev_class[3];
77 __u8 features[8];
78 __u8 commands[64];
79 __u8 ssp_mode;
80 __u8 hci_ver;
81 __u16 hci_rev;
82 __u16 manufacturer;
83 __u16 voice_setting;
84
85 __u16 pkt_type;
86 __u16 esco_type;
87 __u16 link_policy;
88 __u16 link_mode;
89
90 __u32 idle_timeout;
91 __u16 sniff_min_interval;
92 __u16 sniff_max_interval;
93
94 unsigned long quirks;
95
96 atomic_t cmd_cnt;
97 unsigned int acl_cnt;
98 unsigned int sco_cnt;
99
100 unsigned int acl_mtu;
101 unsigned int sco_mtu;
102 unsigned int acl_pkts;
103 unsigned int sco_pkts;
104
105 unsigned long cmd_last_tx;
106 unsigned long acl_last_tx;
107 unsigned long sco_last_tx;
108
109 struct tasklet_struct cmd_task;
110 struct tasklet_struct rx_task;
111 struct tasklet_struct tx_task;
112
113 struct sk_buff_head rx_q;
114 struct sk_buff_head raw_q;
115 struct sk_buff_head cmd_q;
116
117 struct sk_buff *sent_cmd;
118 struct sk_buff *reassembly[3];
119
120 struct mutex req_lock;
121 wait_queue_head_t req_wait_q;
122 __u32 req_status;
123 __u32 req_result;
124
125 struct inquiry_cache inq_cache;
126 struct hci_conn_hash conn_hash;
127
128 struct hci_dev_stats stat;
129
130 struct sk_buff_head driver_init;
131
132 void *driver_data;
133 void *core_data;
134
135 atomic_t promisc;
136
137 struct device *parent;
138 struct device dev;
139
140 struct rfkill *rfkill;
141
142 struct module *owner;
143
144 int (*open)(struct hci_dev *hdev);
145 int (*close)(struct hci_dev *hdev);
146 int (*flush)(struct hci_dev *hdev);
147 int (*send)(struct sk_buff *skb);
148 void (*destruct)(struct hci_dev *hdev);
149 void (*notify)(struct hci_dev *hdev, unsigned int evt);
150 int (*ioctl)(struct hci_dev *hdev, unsigned int cmd, unsigned long arg);
151 };
152
153 struct hci_conn {
154 struct list_head list;
155
156 atomic_t refcnt;
157 spinlock_t lock;
158
159 bdaddr_t dst;
160 __u16 handle;
161 __u16 state;
162 __u8 mode;
163 __u8 type;
164 __u8 out;
165 __u8 attempt;
166 __u8 dev_class[3];
167 __u8 features[8];
168 __u8 ssp_mode;
169 __u16 interval;
170 __u16 pkt_type;
171 __u16 link_policy;
172 __u32 link_mode;
173 __u8 auth_type;
174 __u8 sec_level;
175 __u8 power_save;
176 __u16 disc_timeout;
177 unsigned long pend;
178
179 unsigned int sent;
180
181 struct sk_buff_head data_q;
182
183 struct timer_list disc_timer;
184 struct timer_list idle_timer;
185
186 struct work_struct work_add;
187 struct work_struct work_del;
188
189 struct device dev;
190 atomic_t devref;
191
192 struct hci_dev *hdev;
193 void *l2cap_data;
194 void *sco_data;
195 void *priv;
196
197 struct hci_conn *link;
198 };
199
200 extern struct hci_proto *hci_proto[];
201 extern struct list_head hci_dev_list;
202 extern struct list_head hci_cb_list;
203 extern rwlock_t hci_dev_list_lock;
204 extern rwlock_t hci_cb_list_lock;
205
206 /* ----- Inquiry cache ----- */
207 #define INQUIRY_CACHE_AGE_MAX (HZ*30) // 30 seconds
208 #define INQUIRY_ENTRY_AGE_MAX (HZ*60) // 60 seconds
209
210 #define inquiry_cache_lock(c) spin_lock(&c->lock)
211 #define inquiry_cache_unlock(c) spin_unlock(&c->lock)
212 #define inquiry_cache_lock_bh(c) spin_lock_bh(&c->lock)
213 #define inquiry_cache_unlock_bh(c) spin_unlock_bh(&c->lock)
214
215 static inline void inquiry_cache_init(struct hci_dev *hdev)
216 {
217 struct inquiry_cache *c = &hdev->inq_cache;
218 spin_lock_init(&c->lock);
219 c->list = NULL;
220 }
221
222 static inline int inquiry_cache_empty(struct hci_dev *hdev)
223 {
224 struct inquiry_cache *c = &hdev->inq_cache;
225 return (c->list == NULL);
226 }
227
228 static inline long inquiry_cache_age(struct hci_dev *hdev)
229 {
230 struct inquiry_cache *c = &hdev->inq_cache;
231 return jiffies - c->timestamp;
232 }
233
234 static inline long inquiry_entry_age(struct inquiry_entry *e)
235 {
236 return jiffies - e->timestamp;
237 }
238
239 struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr);
240 void hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data);
241
242 /* ----- HCI Connections ----- */
243 enum {
244 HCI_CONN_AUTH_PEND,
245 HCI_CONN_ENCRYPT_PEND,
246 HCI_CONN_RSWITCH_PEND,
247 HCI_CONN_MODE_CHANGE_PEND,
248 };
249
250 static inline void hci_conn_hash_init(struct hci_dev *hdev)
251 {
252 struct hci_conn_hash *h = &hdev->conn_hash;
253 INIT_LIST_HEAD(&h->list);
254 spin_lock_init(&h->lock);
255 h->acl_num = 0;
256 h->sco_num = 0;
257 }
258
259 static inline void hci_conn_hash_add(struct hci_dev *hdev, struct hci_conn *c)
260 {
261 struct hci_conn_hash *h = &hdev->conn_hash;
262 list_add(&c->list, &h->list);
263 if (c->type == ACL_LINK)
264 h->acl_num++;
265 else
266 h->sco_num++;
267 }
268
269 static inline void hci_conn_hash_del(struct hci_dev *hdev, struct hci_conn *c)
270 {
271 struct hci_conn_hash *h = &hdev->conn_hash;
272 list_del(&c->list);
273 if (c->type == ACL_LINK)
274 h->acl_num--;
275 else
276 h->sco_num--;
277 }
278
279 static inline struct hci_conn *hci_conn_hash_lookup_handle(struct hci_dev *hdev,
280 __u16 handle)
281 {
282 struct hci_conn_hash *h = &hdev->conn_hash;
283 struct list_head *p;
284 struct hci_conn *c;
285
286 list_for_each(p, &h->list) {
287 c = list_entry(p, struct hci_conn, list);
288 if (c->handle == handle)
289 return c;
290 }
291 return NULL;
292 }
293
294 static inline struct hci_conn *hci_conn_hash_lookup_ba(struct hci_dev *hdev,
295 __u8 type, bdaddr_t *ba)
296 {
297 struct hci_conn_hash *h = &hdev->conn_hash;
298 struct list_head *p;
299 struct hci_conn *c;
300
301 list_for_each(p, &h->list) {
302 c = list_entry(p, struct hci_conn, list);
303 if (c->type == type && !bacmp(&c->dst, ba))
304 return c;
305 }
306 return NULL;
307 }
308
309 static inline struct hci_conn *hci_conn_hash_lookup_state(struct hci_dev *hdev,
310 __u8 type, __u16 state)
311 {
312 struct hci_conn_hash *h = &hdev->conn_hash;
313 struct list_head *p;
314 struct hci_conn *c;
315
316 list_for_each(p, &h->list) {
317 c = list_entry(p, struct hci_conn, list);
318 if (c->type == type && c->state == state)
319 return c;
320 }
321 return NULL;
322 }
323
324 void hci_acl_connect(struct hci_conn *conn);
325 void hci_acl_disconn(struct hci_conn *conn, __u8 reason);
326 void hci_add_sco(struct hci_conn *conn, __u16 handle);
327 void hci_setup_sync(struct hci_conn *conn, __u16 handle);
328
329 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst);
330 int hci_conn_del(struct hci_conn *conn);
331 void hci_conn_hash_flush(struct hci_dev *hdev);
332 void hci_conn_check_pending(struct hci_dev *hdev);
333
334 struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 sec_level, __u8 auth_type);
335 int hci_conn_check_link_mode(struct hci_conn *conn);
336 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type);
337 int hci_conn_change_link_key(struct hci_conn *conn);
338 int hci_conn_switch_role(struct hci_conn *conn, __u8 role);
339
340 void hci_conn_enter_active_mode(struct hci_conn *conn);
341 void hci_conn_enter_sniff_mode(struct hci_conn *conn);
342
343 void hci_conn_hold_device(struct hci_conn *conn);
344 void hci_conn_put_device(struct hci_conn *conn);
345
346 static inline void hci_conn_hold(struct hci_conn *conn)
347 {
348 atomic_inc(&conn->refcnt);
349 del_timer(&conn->disc_timer);
350 }
351
352 static inline void hci_conn_put(struct hci_conn *conn)
353 {
354 if (atomic_dec_and_test(&conn->refcnt)) {
355 unsigned long timeo;
356 if (conn->type == ACL_LINK) {
357 del_timer(&conn->idle_timer);
358 if (conn->state == BT_CONNECTED) {
359 timeo = msecs_to_jiffies(conn->disc_timeout);
360 if (!conn->out)
361 timeo *= 2;
362 } else
363 timeo = msecs_to_jiffies(10);
364 } else
365 timeo = msecs_to_jiffies(10);
366 mod_timer(&conn->disc_timer, jiffies + timeo);
367 }
368 }
369
370 /* ----- HCI tasks ----- */
371 static inline void hci_sched_cmd(struct hci_dev *hdev)
372 {
373 tasklet_schedule(&hdev->cmd_task);
374 }
375
376 static inline void hci_sched_rx(struct hci_dev *hdev)
377 {
378 tasklet_schedule(&hdev->rx_task);
379 }
380
381 static inline void hci_sched_tx(struct hci_dev *hdev)
382 {
383 tasklet_schedule(&hdev->tx_task);
384 }
385
386 /* ----- HCI Devices ----- */
387 static inline void __hci_dev_put(struct hci_dev *d)
388 {
389 if (atomic_dec_and_test(&d->refcnt))
390 d->destruct(d);
391 }
392
393 static inline void hci_dev_put(struct hci_dev *d)
394 {
395 __hci_dev_put(d);
396 module_put(d->owner);
397 }
398
399 static inline struct hci_dev *__hci_dev_hold(struct hci_dev *d)
400 {
401 atomic_inc(&d->refcnt);
402 return d;
403 }
404
405 static inline struct hci_dev *hci_dev_hold(struct hci_dev *d)
406 {
407 if (try_module_get(d->owner))
408 return __hci_dev_hold(d);
409 return NULL;
410 }
411
412 #define hci_dev_lock(d) spin_lock(&d->lock)
413 #define hci_dev_unlock(d) spin_unlock(&d->lock)
414 #define hci_dev_lock_bh(d) spin_lock_bh(&d->lock)
415 #define hci_dev_unlock_bh(d) spin_unlock_bh(&d->lock)
416
417 struct hci_dev *hci_dev_get(int index);
418 struct hci_dev *hci_get_route(bdaddr_t *src, bdaddr_t *dst);
419
420 struct hci_dev *hci_alloc_dev(void);
421 void hci_free_dev(struct hci_dev *hdev);
422 int hci_register_dev(struct hci_dev *hdev);
423 int hci_unregister_dev(struct hci_dev *hdev);
424 int hci_suspend_dev(struct hci_dev *hdev);
425 int hci_resume_dev(struct hci_dev *hdev);
426 int hci_dev_open(__u16 dev);
427 int hci_dev_close(__u16 dev);
428 int hci_dev_reset(__u16 dev);
429 int hci_dev_reset_stat(__u16 dev);
430 int hci_dev_cmd(unsigned int cmd, void __user *arg);
431 int hci_get_dev_list(void __user *arg);
432 int hci_get_dev_info(void __user *arg);
433 int hci_get_conn_list(void __user *arg);
434 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg);
435 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg);
436 int hci_inquiry(void __user *arg);
437
438 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb);
439
440 int hci_recv_frame(struct sk_buff *skb);
441 int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count);
442
443 int hci_register_sysfs(struct hci_dev *hdev);
444 void hci_unregister_sysfs(struct hci_dev *hdev);
445 void hci_conn_init_sysfs(struct hci_conn *conn);
446 void hci_conn_add_sysfs(struct hci_conn *conn);
447 void hci_conn_del_sysfs(struct hci_conn *conn);
448
449 #define SET_HCIDEV_DEV(hdev, pdev) ((hdev)->parent = (pdev))
450
451 /* ----- LMP capabilities ----- */
452 #define lmp_rswitch_capable(dev) ((dev)->features[0] & LMP_RSWITCH)
453 #define lmp_encrypt_capable(dev) ((dev)->features[0] & LMP_ENCRYPT)
454 #define lmp_sniff_capable(dev) ((dev)->features[0] & LMP_SNIFF)
455 #define lmp_sniffsubr_capable(dev) ((dev)->features[5] & LMP_SNIFF_SUBR)
456 #define lmp_esco_capable(dev) ((dev)->features[3] & LMP_ESCO)
457 #define lmp_ssp_capable(dev) ((dev)->features[6] & LMP_SIMPLE_PAIR)
458
459 /* ----- HCI protocols ----- */
460 struct hci_proto {
461 char *name;
462 unsigned int id;
463 unsigned long flags;
464
465 void *priv;
466
467 int (*connect_ind) (struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type);
468 int (*connect_cfm) (struct hci_conn *conn, __u8 status);
469 int (*disconn_ind) (struct hci_conn *conn);
470 int (*disconn_cfm) (struct hci_conn *conn, __u8 reason);
471 int (*recv_acldata) (struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
472 int (*recv_scodata) (struct hci_conn *conn, struct sk_buff *skb);
473 int (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
474 };
475
476 static inline int hci_proto_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type)
477 {
478 register struct hci_proto *hp;
479 int mask = 0;
480
481 hp = hci_proto[HCI_PROTO_L2CAP];
482 if (hp && hp->connect_ind)
483 mask |= hp->connect_ind(hdev, bdaddr, type);
484
485 hp = hci_proto[HCI_PROTO_SCO];
486 if (hp && hp->connect_ind)
487 mask |= hp->connect_ind(hdev, bdaddr, type);
488
489 return mask;
490 }
491
492 static inline void hci_proto_connect_cfm(struct hci_conn *conn, __u8 status)
493 {
494 register struct hci_proto *hp;
495
496 hp = hci_proto[HCI_PROTO_L2CAP];
497 if (hp && hp->connect_cfm)
498 hp->connect_cfm(conn, status);
499
500 hp = hci_proto[HCI_PROTO_SCO];
501 if (hp && hp->connect_cfm)
502 hp->connect_cfm(conn, status);
503 }
504
505 static inline int hci_proto_disconn_ind(struct hci_conn *conn)
506 {
507 register struct hci_proto *hp;
508 int reason = 0x13;
509
510 hp = hci_proto[HCI_PROTO_L2CAP];
511 if (hp && hp->disconn_ind)
512 reason = hp->disconn_ind(conn);
513
514 hp = hci_proto[HCI_PROTO_SCO];
515 if (hp && hp->disconn_ind)
516 reason = hp->disconn_ind(conn);
517
518 return reason;
519 }
520
521 static inline void hci_proto_disconn_cfm(struct hci_conn *conn, __u8 reason)
522 {
523 register struct hci_proto *hp;
524
525 hp = hci_proto[HCI_PROTO_L2CAP];
526 if (hp && hp->disconn_cfm)
527 hp->disconn_cfm(conn, reason);
528
529 hp = hci_proto[HCI_PROTO_SCO];
530 if (hp && hp->disconn_cfm)
531 hp->disconn_cfm(conn, reason);
532 }
533
534 static inline void hci_proto_auth_cfm(struct hci_conn *conn, __u8 status)
535 {
536 register struct hci_proto *hp;
537 __u8 encrypt;
538
539 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
540 return;
541
542 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
543
544 hp = hci_proto[HCI_PROTO_L2CAP];
545 if (hp && hp->security_cfm)
546 hp->security_cfm(conn, status, encrypt);
547
548 hp = hci_proto[HCI_PROTO_SCO];
549 if (hp && hp->security_cfm)
550 hp->security_cfm(conn, status, encrypt);
551 }
552
553 static inline void hci_proto_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
554 {
555 register struct hci_proto *hp;
556
557 hp = hci_proto[HCI_PROTO_L2CAP];
558 if (hp && hp->security_cfm)
559 hp->security_cfm(conn, status, encrypt);
560
561 hp = hci_proto[HCI_PROTO_SCO];
562 if (hp && hp->security_cfm)
563 hp->security_cfm(conn, status, encrypt);
564 }
565
566 int hci_register_proto(struct hci_proto *hproto);
567 int hci_unregister_proto(struct hci_proto *hproto);
568
569 /* ----- HCI callbacks ----- */
570 struct hci_cb {
571 struct list_head list;
572
573 char *name;
574
575 void (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
576 void (*key_change_cfm) (struct hci_conn *conn, __u8 status);
577 void (*role_switch_cfm) (struct hci_conn *conn, __u8 status, __u8 role);
578 };
579
580 static inline void hci_auth_cfm(struct hci_conn *conn, __u8 status)
581 {
582 struct list_head *p;
583 __u8 encrypt;
584
585 hci_proto_auth_cfm(conn, status);
586
587 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
588 return;
589
590 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
591
592 read_lock_bh(&hci_cb_list_lock);
593 list_for_each(p, &hci_cb_list) {
594 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
595 if (cb->security_cfm)
596 cb->security_cfm(conn, status, encrypt);
597 }
598 read_unlock_bh(&hci_cb_list_lock);
599 }
600
601 static inline void hci_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
602 {
603 struct list_head *p;
604
605 if (conn->sec_level == BT_SECURITY_SDP)
606 conn->sec_level = BT_SECURITY_LOW;
607
608 hci_proto_encrypt_cfm(conn, status, encrypt);
609
610 read_lock_bh(&hci_cb_list_lock);
611 list_for_each(p, &hci_cb_list) {
612 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
613 if (cb->security_cfm)
614 cb->security_cfm(conn, status, encrypt);
615 }
616 read_unlock_bh(&hci_cb_list_lock);
617 }
618
619 static inline void hci_key_change_cfm(struct hci_conn *conn, __u8 status)
620 {
621 struct list_head *p;
622
623 read_lock_bh(&hci_cb_list_lock);
624 list_for_each(p, &hci_cb_list) {
625 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
626 if (cb->key_change_cfm)
627 cb->key_change_cfm(conn, status);
628 }
629 read_unlock_bh(&hci_cb_list_lock);
630 }
631
632 static inline void hci_role_switch_cfm(struct hci_conn *conn, __u8 status, __u8 role)
633 {
634 struct list_head *p;
635
636 read_lock_bh(&hci_cb_list_lock);
637 list_for_each(p, &hci_cb_list) {
638 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
639 if (cb->role_switch_cfm)
640 cb->role_switch_cfm(conn, status, role);
641 }
642 read_unlock_bh(&hci_cb_list_lock);
643 }
644
645 int hci_register_cb(struct hci_cb *hcb);
646 int hci_unregister_cb(struct hci_cb *hcb);
647
648 int hci_register_notifier(struct notifier_block *nb);
649 int hci_unregister_notifier(struct notifier_block *nb);
650
651 int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param);
652 int hci_send_acl(struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
653 int hci_send_sco(struct hci_conn *conn, struct sk_buff *skb);
654
655 void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode);
656
657 void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data);
658
659 /* ----- HCI Sockets ----- */
660 void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb);
661
662 /* HCI info for socket */
663 #define hci_pi(sk) ((struct hci_pinfo *) sk)
664
665 struct hci_pinfo {
666 struct bt_sock bt;
667 struct hci_dev *hdev;
668 struct hci_filter filter;
669 __u32 cmsg_mask;
670 };
671
672 /* HCI security filter */
673 #define HCI_SFLT_MAX_OGF 5
674
675 struct hci_sec_filter {
676 __u32 type_mask;
677 __u32 event_mask[2];
678 __u32 ocf_mask[HCI_SFLT_MAX_OGF + 1][4];
679 };
680
681 /* ----- HCI requests ----- */
682 #define HCI_REQ_DONE 0
683 #define HCI_REQ_PEND 1
684 #define HCI_REQ_CANCELED 2
685
686 #define hci_req_lock(d) mutex_lock(&d->req_lock)
687 #define hci_req_unlock(d) mutex_unlock(&d->req_lock)
688
689 void hci_req_complete(struct hci_dev *hdev, int result);
690
691 #endif /* __HCI_CORE_H */