2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
28 #include <net/bluetooth/hci.h>
30 /* HCI upper protocols */
31 #define HCI_PROTO_L2CAP 0
32 #define HCI_PROTO_SCO 1
34 /* HCI Core structures */
38 __u8 pscan_period_mode;
46 struct inquiry_entry {
47 struct inquiry_entry *next;
49 struct inquiry_data data;
52 struct inquiry_cache {
55 struct inquiry_entry *list;
58 struct hci_conn_hash {
59 struct list_head list;
66 struct list_head list;
91 __u16 sniff_min_interval;
92 __u16 sniff_max_interval;
100 unsigned int acl_mtu;
101 unsigned int sco_mtu;
102 unsigned int acl_pkts;
103 unsigned int sco_pkts;
105 unsigned long cmd_last_tx;
106 unsigned long acl_last_tx;
107 unsigned long sco_last_tx;
109 struct tasklet_struct cmd_task;
110 struct tasklet_struct rx_task;
111 struct tasklet_struct tx_task;
113 struct sk_buff_head rx_q;
114 struct sk_buff_head raw_q;
115 struct sk_buff_head cmd_q;
117 struct sk_buff *sent_cmd;
118 struct sk_buff *reassembly[3];
120 struct semaphore req_lock;
121 wait_queue_head_t req_wait_q;
125 struct inquiry_cache inq_cache;
126 struct hci_conn_hash conn_hash;
128 struct hci_dev_stats stat;
130 struct sk_buff_head driver_init;
137 struct device *parent;
140 struct module *owner;
142 int (*open)(struct hci_dev *hdev);
143 int (*close)(struct hci_dev *hdev);
144 int (*flush)(struct hci_dev *hdev);
145 int (*send)(struct sk_buff *skb);
146 void (*destruct)(struct hci_dev *hdev);
147 void (*notify)(struct hci_dev *hdev, unsigned int evt);
148 int (*ioctl)(struct hci_dev *hdev, unsigned int cmd, unsigned long arg);
152 struct list_head list;
178 struct sk_buff_head data_q;
180 struct timer_list disc_timer;
181 struct timer_list idle_timer;
183 struct work_struct work;
187 struct hci_dev *hdev;
192 struct hci_conn *link;
195 extern struct hci_proto *hci_proto[];
196 extern struct list_head hci_dev_list;
197 extern struct list_head hci_cb_list;
198 extern rwlock_t hci_dev_list_lock;
199 extern rwlock_t hci_cb_list_lock;
201 /* ----- Inquiry cache ----- */
202 #define INQUIRY_CACHE_AGE_MAX (HZ*30) // 30 seconds
203 #define INQUIRY_ENTRY_AGE_MAX (HZ*60) // 60 seconds
205 #define inquiry_cache_lock(c) spin_lock(&c->lock)
206 #define inquiry_cache_unlock(c) spin_unlock(&c->lock)
207 #define inquiry_cache_lock_bh(c) spin_lock_bh(&c->lock)
208 #define inquiry_cache_unlock_bh(c) spin_unlock_bh(&c->lock)
210 static inline void inquiry_cache_init(struct hci_dev *hdev)
212 struct inquiry_cache *c = &hdev->inq_cache;
213 spin_lock_init(&c->lock);
217 static inline int inquiry_cache_empty(struct hci_dev *hdev)
219 struct inquiry_cache *c = &hdev->inq_cache;
220 return (c->list == NULL);
223 static inline long inquiry_cache_age(struct hci_dev *hdev)
225 struct inquiry_cache *c = &hdev->inq_cache;
226 return jiffies - c->timestamp;
229 static inline long inquiry_entry_age(struct inquiry_entry *e)
231 return jiffies - e->timestamp;
234 struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr);
235 void hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data);
237 /* ----- HCI Connections ----- */
240 HCI_CONN_ENCRYPT_PEND,
241 HCI_CONN_RSWITCH_PEND,
242 HCI_CONN_MODE_CHANGE_PEND,
245 static inline void hci_conn_hash_init(struct hci_dev *hdev)
247 struct hci_conn_hash *h = &hdev->conn_hash;
248 INIT_LIST_HEAD(&h->list);
249 spin_lock_init(&h->lock);
254 static inline void hci_conn_hash_add(struct hci_dev *hdev, struct hci_conn *c)
256 struct hci_conn_hash *h = &hdev->conn_hash;
257 list_add(&c->list, &h->list);
258 if (c->type == ACL_LINK)
264 static inline void hci_conn_hash_del(struct hci_dev *hdev, struct hci_conn *c)
266 struct hci_conn_hash *h = &hdev->conn_hash;
268 if (c->type == ACL_LINK)
274 static inline struct hci_conn *hci_conn_hash_lookup_handle(struct hci_dev *hdev,
277 struct hci_conn_hash *h = &hdev->conn_hash;
281 list_for_each(p, &h->list) {
282 c = list_entry(p, struct hci_conn, list);
283 if (c->handle == handle)
289 static inline struct hci_conn *hci_conn_hash_lookup_ba(struct hci_dev *hdev,
290 __u8 type, bdaddr_t *ba)
292 struct hci_conn_hash *h = &hdev->conn_hash;
296 list_for_each(p, &h->list) {
297 c = list_entry(p, struct hci_conn, list);
298 if (c->type == type && !bacmp(&c->dst, ba))
304 static inline struct hci_conn *hci_conn_hash_lookup_state(struct hci_dev *hdev,
305 __u8 type, __u16 state)
307 struct hci_conn_hash *h = &hdev->conn_hash;
311 list_for_each(p, &h->list) {
312 c = list_entry(p, struct hci_conn, list);
313 if (c->type == type && c->state == state)
319 void hci_acl_connect(struct hci_conn *conn);
320 void hci_acl_disconn(struct hci_conn *conn, __u8 reason);
321 void hci_add_sco(struct hci_conn *conn, __u16 handle);
322 void hci_setup_sync(struct hci_conn *conn, __u16 handle);
324 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst);
325 int hci_conn_del(struct hci_conn *conn);
326 void hci_conn_hash_flush(struct hci_dev *hdev);
327 void hci_conn_check_pending(struct hci_dev *hdev);
329 struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 sec_level, __u8 auth_type);
330 int hci_conn_check_link_mode(struct hci_conn *conn);
331 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type);
332 int hci_conn_change_link_key(struct hci_conn *conn);
333 int hci_conn_switch_role(struct hci_conn *conn, __u8 role);
335 void hci_conn_enter_active_mode(struct hci_conn *conn);
336 void hci_conn_enter_sniff_mode(struct hci_conn *conn);
338 static inline void hci_conn_hold(struct hci_conn *conn)
340 atomic_inc(&conn->refcnt);
341 del_timer(&conn->disc_timer);
344 static inline void hci_conn_put(struct hci_conn *conn)
346 if (atomic_dec_and_test(&conn->refcnt)) {
348 if (conn->type == ACL_LINK) {
349 del_timer(&conn->idle_timer);
350 if (conn->state == BT_CONNECTED) {
351 timeo = msecs_to_jiffies(HCI_DISCONN_TIMEOUT);
355 timeo = msecs_to_jiffies(10);
357 timeo = msecs_to_jiffies(10);
358 mod_timer(&conn->disc_timer, jiffies + timeo);
362 /* ----- HCI tasks ----- */
363 static inline void hci_sched_cmd(struct hci_dev *hdev)
365 tasklet_schedule(&hdev->cmd_task);
368 static inline void hci_sched_rx(struct hci_dev *hdev)
370 tasklet_schedule(&hdev->rx_task);
373 static inline void hci_sched_tx(struct hci_dev *hdev)
375 tasklet_schedule(&hdev->tx_task);
378 /* ----- HCI Devices ----- */
379 static inline void __hci_dev_put(struct hci_dev *d)
381 if (atomic_dec_and_test(&d->refcnt))
385 static inline void hci_dev_put(struct hci_dev *d)
388 module_put(d->owner);
391 static inline struct hci_dev *__hci_dev_hold(struct hci_dev *d)
393 atomic_inc(&d->refcnt);
397 static inline struct hci_dev *hci_dev_hold(struct hci_dev *d)
399 if (try_module_get(d->owner))
400 return __hci_dev_hold(d);
404 #define hci_dev_lock(d) spin_lock(&d->lock)
405 #define hci_dev_unlock(d) spin_unlock(&d->lock)
406 #define hci_dev_lock_bh(d) spin_lock_bh(&d->lock)
407 #define hci_dev_unlock_bh(d) spin_unlock_bh(&d->lock)
409 struct hci_dev *hci_dev_get(int index);
410 struct hci_dev *hci_get_route(bdaddr_t *src, bdaddr_t *dst);
412 struct hci_dev *hci_alloc_dev(void);
413 void hci_free_dev(struct hci_dev *hdev);
414 int hci_register_dev(struct hci_dev *hdev);
415 int hci_unregister_dev(struct hci_dev *hdev);
416 int hci_suspend_dev(struct hci_dev *hdev);
417 int hci_resume_dev(struct hci_dev *hdev);
418 int hci_dev_open(__u16 dev);
419 int hci_dev_close(__u16 dev);
420 int hci_dev_reset(__u16 dev);
421 int hci_dev_reset_stat(__u16 dev);
422 int hci_dev_cmd(unsigned int cmd, void __user *arg);
423 int hci_get_dev_list(void __user *arg);
424 int hci_get_dev_info(void __user *arg);
425 int hci_get_conn_list(void __user *arg);
426 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg);
427 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg);
428 int hci_inquiry(void __user *arg);
430 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb);
432 /* Receive frame from HCI drivers */
433 static inline int hci_recv_frame(struct sk_buff *skb)
435 struct hci_dev *hdev = (struct hci_dev *) skb->dev;
436 if (!hdev || (!test_bit(HCI_UP, &hdev->flags)
437 && !test_bit(HCI_INIT, &hdev->flags))) {
443 bt_cb(skb)->incoming = 1;
446 __net_timestamp(skb);
448 /* Queue frame for rx task */
449 skb_queue_tail(&hdev->rx_q, skb);
454 int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count);
456 int hci_register_sysfs(struct hci_dev *hdev);
457 void hci_unregister_sysfs(struct hci_dev *hdev);
458 void hci_conn_add_sysfs(struct hci_conn *conn);
459 void hci_conn_del_sysfs(struct hci_conn *conn);
461 #define SET_HCIDEV_DEV(hdev, pdev) ((hdev)->parent = (pdev))
463 /* ----- LMP capabilities ----- */
464 #define lmp_rswitch_capable(dev) ((dev)->features[0] & LMP_RSWITCH)
465 #define lmp_encrypt_capable(dev) ((dev)->features[0] & LMP_ENCRYPT)
466 #define lmp_sniff_capable(dev) ((dev)->features[0] & LMP_SNIFF)
467 #define lmp_sniffsubr_capable(dev) ((dev)->features[5] & LMP_SNIFF_SUBR)
468 #define lmp_esco_capable(dev) ((dev)->features[3] & LMP_ESCO)
469 #define lmp_ssp_capable(dev) ((dev)->features[6] & LMP_SIMPLE_PAIR)
471 /* ----- HCI protocols ----- */
479 int (*connect_ind) (struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type);
480 int (*connect_cfm) (struct hci_conn *conn, __u8 status);
481 int (*disconn_ind) (struct hci_conn *conn);
482 int (*disconn_cfm) (struct hci_conn *conn, __u8 reason);
483 int (*recv_acldata) (struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
484 int (*recv_scodata) (struct hci_conn *conn, struct sk_buff *skb);
485 int (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
488 static inline int hci_proto_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type)
490 register struct hci_proto *hp;
493 hp = hci_proto[HCI_PROTO_L2CAP];
494 if (hp && hp->connect_ind)
495 mask |= hp->connect_ind(hdev, bdaddr, type);
497 hp = hci_proto[HCI_PROTO_SCO];
498 if (hp && hp->connect_ind)
499 mask |= hp->connect_ind(hdev, bdaddr, type);
504 static inline void hci_proto_connect_cfm(struct hci_conn *conn, __u8 status)
506 register struct hci_proto *hp;
508 hp = hci_proto[HCI_PROTO_L2CAP];
509 if (hp && hp->connect_cfm)
510 hp->connect_cfm(conn, status);
512 hp = hci_proto[HCI_PROTO_SCO];
513 if (hp && hp->connect_cfm)
514 hp->connect_cfm(conn, status);
517 static inline int hci_proto_disconn_ind(struct hci_conn *conn)
519 register struct hci_proto *hp;
522 hp = hci_proto[HCI_PROTO_L2CAP];
523 if (hp && hp->disconn_ind)
524 reason = hp->disconn_ind(conn);
526 hp = hci_proto[HCI_PROTO_SCO];
527 if (hp && hp->disconn_ind)
528 reason = hp->disconn_ind(conn);
533 static inline void hci_proto_disconn_cfm(struct hci_conn *conn, __u8 reason)
535 register struct hci_proto *hp;
537 hp = hci_proto[HCI_PROTO_L2CAP];
538 if (hp && hp->disconn_cfm)
539 hp->disconn_cfm(conn, reason);
541 hp = hci_proto[HCI_PROTO_SCO];
542 if (hp && hp->disconn_cfm)
543 hp->disconn_cfm(conn, reason);
546 static inline void hci_proto_auth_cfm(struct hci_conn *conn, __u8 status)
548 register struct hci_proto *hp;
551 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
554 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
556 hp = hci_proto[HCI_PROTO_L2CAP];
557 if (hp && hp->security_cfm)
558 hp->security_cfm(conn, status, encrypt);
560 hp = hci_proto[HCI_PROTO_SCO];
561 if (hp && hp->security_cfm)
562 hp->security_cfm(conn, status, encrypt);
565 static inline void hci_proto_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
567 register struct hci_proto *hp;
569 hp = hci_proto[HCI_PROTO_L2CAP];
570 if (hp && hp->security_cfm)
571 hp->security_cfm(conn, status, encrypt);
573 hp = hci_proto[HCI_PROTO_SCO];
574 if (hp && hp->security_cfm)
575 hp->security_cfm(conn, status, encrypt);
578 int hci_register_proto(struct hci_proto *hproto);
579 int hci_unregister_proto(struct hci_proto *hproto);
581 /* ----- HCI callbacks ----- */
583 struct list_head list;
587 void (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
588 void (*key_change_cfm) (struct hci_conn *conn, __u8 status);
589 void (*role_switch_cfm) (struct hci_conn *conn, __u8 status, __u8 role);
592 static inline void hci_auth_cfm(struct hci_conn *conn, __u8 status)
597 hci_proto_auth_cfm(conn, status);
599 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
602 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
604 read_lock_bh(&hci_cb_list_lock);
605 list_for_each(p, &hci_cb_list) {
606 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
607 if (cb->security_cfm)
608 cb->security_cfm(conn, status, encrypt);
610 read_unlock_bh(&hci_cb_list_lock);
613 static inline void hci_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
617 if (conn->sec_level == BT_SECURITY_SDP)
618 conn->sec_level = BT_SECURITY_LOW;
620 hci_proto_encrypt_cfm(conn, status, encrypt);
622 read_lock_bh(&hci_cb_list_lock);
623 list_for_each(p, &hci_cb_list) {
624 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
625 if (cb->security_cfm)
626 cb->security_cfm(conn, status, encrypt);
628 read_unlock_bh(&hci_cb_list_lock);
631 static inline void hci_key_change_cfm(struct hci_conn *conn, __u8 status)
635 read_lock_bh(&hci_cb_list_lock);
636 list_for_each(p, &hci_cb_list) {
637 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
638 if (cb->key_change_cfm)
639 cb->key_change_cfm(conn, status);
641 read_unlock_bh(&hci_cb_list_lock);
644 static inline void hci_role_switch_cfm(struct hci_conn *conn, __u8 status, __u8 role)
648 read_lock_bh(&hci_cb_list_lock);
649 list_for_each(p, &hci_cb_list) {
650 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
651 if (cb->role_switch_cfm)
652 cb->role_switch_cfm(conn, status, role);
654 read_unlock_bh(&hci_cb_list_lock);
657 int hci_register_cb(struct hci_cb *hcb);
658 int hci_unregister_cb(struct hci_cb *hcb);
660 int hci_register_notifier(struct notifier_block *nb);
661 int hci_unregister_notifier(struct notifier_block *nb);
663 int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param);
664 int hci_send_acl(struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
665 int hci_send_sco(struct hci_conn *conn, struct sk_buff *skb);
667 void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode);
669 void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data);
671 /* ----- HCI Sockets ----- */
672 void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb);
674 /* HCI info for socket */
675 #define hci_pi(sk) ((struct hci_pinfo *) sk)
679 struct hci_dev *hdev;
680 struct hci_filter filter;
684 /* HCI security filter */
685 #define HCI_SFLT_MAX_OGF 5
687 struct hci_sec_filter {
690 __u32 ocf_mask[HCI_SFLT_MAX_OGF + 1][4];
693 /* ----- HCI requests ----- */
694 #define HCI_REQ_DONE 0
695 #define HCI_REQ_PEND 1
696 #define HCI_REQ_CANCELED 2
698 #define hci_req_lock(d) down(&d->req_lock)
699 #define hci_req_unlock(d) up(&d->req_lock)
701 void hci_req_complete(struct hci_dev *hdev, int result);
703 #endif /* __HCI_CORE_H */