ARM: 7409/1: Do not call flush_cache_user_range with mmap_sem held
[linux/fpc-iii.git] / net / bluetooth / hci_conn.c
blob4bb16b8decb135b532f4952df8a30436c0a03155
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
25 /* Bluetooth HCI connection handling. */
27 #include <linux/module.h>
29 #include <linux/types.h>
30 #include <linux/errno.h>
31 #include <linux/kernel.h>
32 #include <linux/slab.h>
33 #include <linux/poll.h>
34 #include <linux/fcntl.h>
35 #include <linux/init.h>
36 #include <linux/skbuff.h>
37 #include <linux/interrupt.h>
38 #include <linux/notifier.h>
39 #include <net/sock.h>
41 #include <asm/system.h>
42 #include <linux/uaccess.h>
43 #include <asm/unaligned.h>
45 #include <net/bluetooth/bluetooth.h>
46 #include <net/bluetooth/hci_core.h>
48 static void hci_le_connect(struct hci_conn *conn)
50 struct hci_dev *hdev = conn->hdev;
51 struct hci_cp_le_create_conn cp;
53 conn->state = BT_CONNECT;
54 conn->out = 1;
55 conn->link_mode |= HCI_LM_MASTER;
57 memset(&cp, 0, sizeof(cp));
58 cp.scan_interval = cpu_to_le16(0x0004);
59 cp.scan_window = cpu_to_le16(0x0004);
60 bacpy(&cp.peer_addr, &conn->dst);
61 cp.conn_interval_min = cpu_to_le16(0x0008);
62 cp.conn_interval_max = cpu_to_le16(0x0100);
63 cp.supervision_timeout = cpu_to_le16(0x0064);
64 cp.min_ce_len = cpu_to_le16(0x0001);
65 cp.max_ce_len = cpu_to_le16(0x0001);
67 hci_send_cmd(hdev, HCI_OP_LE_CREATE_CONN, sizeof(cp), &cp);
70 static void hci_le_connect_cancel(struct hci_conn *conn)
72 hci_send_cmd(conn->hdev, HCI_OP_LE_CREATE_CONN_CANCEL, 0, NULL);
75 void hci_acl_connect(struct hci_conn *conn)
77 struct hci_dev *hdev = conn->hdev;
78 struct inquiry_entry *ie;
79 struct hci_cp_create_conn cp;
81 BT_DBG("%p", conn);
83 conn->state = BT_CONNECT;
84 conn->out = 1;
86 conn->link_mode = HCI_LM_MASTER;
88 conn->attempt++;
90 conn->link_policy = hdev->link_policy;
92 memset(&cp, 0, sizeof(cp));
93 bacpy(&cp.bdaddr, &conn->dst);
94 cp.pscan_rep_mode = 0x02;
96 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
97 if (ie) {
98 if (inquiry_entry_age(ie) <= INQUIRY_ENTRY_AGE_MAX) {
99 cp.pscan_rep_mode = ie->data.pscan_rep_mode;
100 cp.pscan_mode = ie->data.pscan_mode;
101 cp.clock_offset = ie->data.clock_offset |
102 cpu_to_le16(0x8000);
105 memcpy(conn->dev_class, ie->data.dev_class, 3);
106 conn->ssp_mode = ie->data.ssp_mode;
109 cp.pkt_type = cpu_to_le16(conn->pkt_type);
110 if (lmp_rswitch_capable(hdev) && !(hdev->link_mode & HCI_LM_MASTER))
111 cp.role_switch = 0x01;
112 else
113 cp.role_switch = 0x00;
115 hci_send_cmd(hdev, HCI_OP_CREATE_CONN, sizeof(cp), &cp);
118 static void hci_acl_connect_cancel(struct hci_conn *conn)
120 struct hci_cp_create_conn_cancel cp;
122 BT_DBG("%p", conn);
124 if (conn->hdev->hci_ver < 2)
125 return;
127 bacpy(&cp.bdaddr, &conn->dst);
128 hci_send_cmd(conn->hdev, HCI_OP_CREATE_CONN_CANCEL, sizeof(cp), &cp);
131 void hci_acl_disconn(struct hci_conn *conn, __u8 reason)
133 struct hci_cp_disconnect cp;
135 BT_DBG("%p", conn);
137 conn->state = BT_DISCONN;
139 cp.handle = cpu_to_le16(conn->handle);
140 cp.reason = reason;
141 hci_send_cmd(conn->hdev, HCI_OP_DISCONNECT, sizeof(cp), &cp);
144 void hci_add_sco(struct hci_conn *conn, __u16 handle)
146 struct hci_dev *hdev = conn->hdev;
147 struct hci_cp_add_sco cp;
149 BT_DBG("%p", conn);
151 conn->state = BT_CONNECT;
152 conn->out = 1;
154 conn->attempt++;
156 cp.handle = cpu_to_le16(handle);
157 cp.pkt_type = cpu_to_le16(conn->pkt_type);
159 hci_send_cmd(hdev, HCI_OP_ADD_SCO, sizeof(cp), &cp);
162 void hci_setup_sync(struct hci_conn *conn, __u16 handle)
164 struct hci_dev *hdev = conn->hdev;
165 struct hci_cp_setup_sync_conn cp;
167 BT_DBG("%p", conn);
169 conn->state = BT_CONNECT;
170 conn->out = 1;
172 conn->attempt++;
174 cp.handle = cpu_to_le16(handle);
175 cp.pkt_type = cpu_to_le16(conn->pkt_type);
177 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
178 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
179 cp.max_latency = cpu_to_le16(0xffff);
180 cp.voice_setting = cpu_to_le16(hdev->voice_setting);
181 cp.retrans_effort = 0xff;
183 hci_send_cmd(hdev, HCI_OP_SETUP_SYNC_CONN, sizeof(cp), &cp);
186 void hci_le_conn_update(struct hci_conn *conn, u16 min, u16 max,
187 u16 latency, u16 to_multiplier)
189 struct hci_cp_le_conn_update cp;
190 struct hci_dev *hdev = conn->hdev;
192 memset(&cp, 0, sizeof(cp));
194 cp.handle = cpu_to_le16(conn->handle);
195 cp.conn_interval_min = cpu_to_le16(min);
196 cp.conn_interval_max = cpu_to_le16(max);
197 cp.conn_latency = cpu_to_le16(latency);
198 cp.supervision_timeout = cpu_to_le16(to_multiplier);
199 cp.min_ce_len = cpu_to_le16(0x0001);
200 cp.max_ce_len = cpu_to_le16(0x0001);
202 hci_send_cmd(hdev, HCI_OP_LE_CONN_UPDATE, sizeof(cp), &cp);
204 EXPORT_SYMBOL(hci_le_conn_update);
206 /* Device _must_ be locked */
207 void hci_sco_setup(struct hci_conn *conn, __u8 status)
209 struct hci_conn *sco = conn->link;
211 BT_DBG("%p", conn);
213 if (!sco)
214 return;
216 if (!status) {
217 if (lmp_esco_capable(conn->hdev))
218 hci_setup_sync(sco, conn->handle);
219 else
220 hci_add_sco(sco, conn->handle);
221 } else {
222 hci_proto_connect_cfm(sco, status);
223 hci_conn_del(sco);
227 static void hci_conn_timeout(unsigned long arg)
229 struct hci_conn *conn = (void *) arg;
230 struct hci_dev *hdev = conn->hdev;
231 __u8 reason;
233 BT_DBG("conn %p state %d", conn, conn->state);
235 if (atomic_read(&conn->refcnt))
236 return;
238 hci_dev_lock(hdev);
240 switch (conn->state) {
241 case BT_CONNECT:
242 case BT_CONNECT2:
243 if (conn->out) {
244 if (conn->type == ACL_LINK)
245 hci_acl_connect_cancel(conn);
246 else if (conn->type == LE_LINK)
247 hci_le_connect_cancel(conn);
249 break;
250 case BT_CONFIG:
251 case BT_CONNECTED:
252 reason = hci_proto_disconn_ind(conn);
253 hci_acl_disconn(conn, reason);
254 break;
255 default:
256 conn->state = BT_CLOSED;
257 break;
260 hci_dev_unlock(hdev);
263 static void hci_conn_idle(unsigned long arg)
265 struct hci_conn *conn = (void *) arg;
267 BT_DBG("conn %p mode %d", conn, conn->mode);
269 hci_conn_enter_sniff_mode(conn);
272 static void hci_conn_auto_accept(unsigned long arg)
274 struct hci_conn *conn = (void *) arg;
275 struct hci_dev *hdev = conn->hdev;
277 hci_dev_lock(hdev);
279 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY, sizeof(conn->dst),
280 &conn->dst);
282 hci_dev_unlock(hdev);
285 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst)
287 struct hci_conn *conn;
289 BT_DBG("%s dst %s", hdev->name, batostr(dst));
291 conn = kzalloc(sizeof(struct hci_conn), GFP_ATOMIC);
292 if (!conn)
293 return NULL;
295 bacpy(&conn->dst, dst);
296 conn->hdev = hdev;
297 conn->type = type;
298 conn->mode = HCI_CM_ACTIVE;
299 conn->state = BT_OPEN;
300 conn->auth_type = HCI_AT_GENERAL_BONDING;
301 conn->io_capability = hdev->io_capability;
302 conn->remote_auth = 0xff;
303 conn->key_type = 0xff;
305 conn->power_save = 1;
306 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
308 switch (type) {
309 case ACL_LINK:
310 conn->pkt_type = hdev->pkt_type & ACL_PTYPE_MASK;
311 break;
312 case SCO_LINK:
313 if (lmp_esco_capable(hdev))
314 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
315 (hdev->esco_type & EDR_ESCO_MASK);
316 else
317 conn->pkt_type = hdev->pkt_type & SCO_PTYPE_MASK;
318 break;
319 case ESCO_LINK:
320 conn->pkt_type = hdev->esco_type & ~EDR_ESCO_MASK;
321 break;
324 skb_queue_head_init(&conn->data_q);
326 setup_timer(&conn->disc_timer, hci_conn_timeout, (unsigned long)conn);
327 setup_timer(&conn->idle_timer, hci_conn_idle, (unsigned long)conn);
328 setup_timer(&conn->auto_accept_timer, hci_conn_auto_accept,
329 (unsigned long) conn);
331 atomic_set(&conn->refcnt, 0);
333 hci_dev_hold(hdev);
335 tasklet_disable(&hdev->tx_task);
337 hci_conn_hash_add(hdev, conn);
338 if (hdev->notify)
339 hdev->notify(hdev, HCI_NOTIFY_CONN_ADD);
341 atomic_set(&conn->devref, 0);
343 hci_conn_init_sysfs(conn);
345 tasklet_enable(&hdev->tx_task);
347 return conn;
350 int hci_conn_del(struct hci_conn *conn)
352 struct hci_dev *hdev = conn->hdev;
354 BT_DBG("%s conn %p handle %d", hdev->name, conn, conn->handle);
356 del_timer(&conn->idle_timer);
358 del_timer(&conn->disc_timer);
360 del_timer(&conn->auto_accept_timer);
362 if (conn->type == ACL_LINK) {
363 struct hci_conn *sco = conn->link;
364 if (sco)
365 sco->link = NULL;
367 /* Unacked frames */
368 hdev->acl_cnt += conn->sent;
369 } else if (conn->type == LE_LINK) {
370 if (hdev->le_pkts)
371 hdev->le_cnt += conn->sent;
372 else
373 hdev->acl_cnt += conn->sent;
374 } else {
375 struct hci_conn *acl = conn->link;
376 if (acl) {
377 acl->link = NULL;
378 hci_conn_put(acl);
382 tasklet_disable(&hdev->tx_task);
384 hci_conn_hash_del(hdev, conn);
385 if (hdev->notify)
386 hdev->notify(hdev, HCI_NOTIFY_CONN_DEL);
388 tasklet_enable(&hdev->tx_task);
390 skb_queue_purge(&conn->data_q);
392 hci_conn_put_device(conn);
394 hci_dev_put(hdev);
396 if (conn->handle == 0)
397 kfree(conn);
399 return 0;
402 struct hci_dev *hci_get_route(bdaddr_t *dst, bdaddr_t *src)
404 int use_src = bacmp(src, BDADDR_ANY);
405 struct hci_dev *hdev = NULL;
406 struct list_head *p;
408 BT_DBG("%s -> %s", batostr(src), batostr(dst));
410 read_lock_bh(&hci_dev_list_lock);
412 list_for_each(p, &hci_dev_list) {
413 struct hci_dev *d = list_entry(p, struct hci_dev, list);
415 if (!test_bit(HCI_UP, &d->flags) || test_bit(HCI_RAW, &d->flags))
416 continue;
418 /* Simple routing:
419 * No source address - find interface with bdaddr != dst
420 * Source address - find interface with bdaddr == src
423 if (use_src) {
424 if (!bacmp(&d->bdaddr, src)) {
425 hdev = d; break;
427 } else {
428 if (bacmp(&d->bdaddr, dst)) {
429 hdev = d; break;
434 if (hdev)
435 hdev = hci_dev_hold(hdev);
437 read_unlock_bh(&hci_dev_list_lock);
438 return hdev;
440 EXPORT_SYMBOL(hci_get_route);
442 /* Create SCO, ACL or LE connection.
443 * Device _must_ be locked */
444 struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 sec_level, __u8 auth_type)
446 struct hci_conn *acl;
447 struct hci_conn *sco;
448 struct hci_conn *le;
450 BT_DBG("%s dst %s", hdev->name, batostr(dst));
452 if (type == LE_LINK) {
453 le = hci_conn_hash_lookup_ba(hdev, LE_LINK, dst);
454 if (le)
455 return ERR_PTR(-EBUSY);
456 le = hci_conn_add(hdev, LE_LINK, dst);
457 if (!le)
458 return ERR_PTR(-ENOMEM);
459 if (le->state == BT_OPEN)
460 hci_le_connect(le);
462 hci_conn_hold(le);
464 return le;
467 acl = hci_conn_hash_lookup_ba(hdev, ACL_LINK, dst);
468 if (!acl) {
469 acl = hci_conn_add(hdev, ACL_LINK, dst);
470 if (!acl)
471 return NULL;
474 hci_conn_hold(acl);
476 if (acl->state == BT_OPEN || acl->state == BT_CLOSED) {
477 acl->sec_level = BT_SECURITY_LOW;
478 acl->pending_sec_level = sec_level;
479 acl->auth_type = auth_type;
480 hci_acl_connect(acl);
483 if (type == ACL_LINK)
484 return acl;
486 sco = hci_conn_hash_lookup_ba(hdev, type, dst);
487 if (!sco) {
488 sco = hci_conn_add(hdev, type, dst);
489 if (!sco) {
490 hci_conn_put(acl);
491 return NULL;
495 acl->link = sco;
496 sco->link = acl;
498 hci_conn_hold(sco);
500 if (acl->state == BT_CONNECTED &&
501 (sco->state == BT_OPEN || sco->state == BT_CLOSED)) {
502 acl->power_save = 1;
503 hci_conn_enter_active_mode(acl);
505 if (test_bit(HCI_CONN_MODE_CHANGE_PEND, &acl->pend)) {
506 /* defer SCO setup until mode change completed */
507 set_bit(HCI_CONN_SCO_SETUP_PEND, &acl->pend);
508 return sco;
511 hci_sco_setup(acl, 0x00);
514 return sco;
516 EXPORT_SYMBOL(hci_connect);
518 /* Check link security requirement */
519 int hci_conn_check_link_mode(struct hci_conn *conn)
521 BT_DBG("conn %p", conn);
523 if (conn->ssp_mode > 0 && conn->hdev->ssp_mode > 0 &&
524 !(conn->link_mode & HCI_LM_ENCRYPT))
525 return 0;
527 return 1;
529 EXPORT_SYMBOL(hci_conn_check_link_mode);
531 /* Authenticate remote device */
532 static int hci_conn_auth(struct hci_conn *conn, __u8 sec_level, __u8 auth_type)
534 BT_DBG("conn %p", conn);
536 if (conn->pending_sec_level > sec_level)
537 sec_level = conn->pending_sec_level;
539 if (sec_level > conn->sec_level)
540 conn->pending_sec_level = sec_level;
541 else if (conn->link_mode & HCI_LM_AUTH)
542 return 1;
544 /* Make sure we preserve an existing MITM requirement*/
545 auth_type |= (conn->auth_type & 0x01);
547 conn->auth_type = auth_type;
549 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->pend)) {
550 struct hci_cp_auth_requested cp;
552 /* encrypt must be pending if auth is also pending */
553 set_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend);
555 cp.handle = cpu_to_le16(conn->handle);
556 hci_send_cmd(conn->hdev, HCI_OP_AUTH_REQUESTED,
557 sizeof(cp), &cp);
560 return 0;
563 /* Encrypt the the link */
564 static void hci_conn_encrypt(struct hci_conn *conn)
566 BT_DBG("conn %p", conn);
568 if (!test_and_set_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend)) {
569 struct hci_cp_set_conn_encrypt cp;
570 cp.handle = cpu_to_le16(conn->handle);
571 cp.encrypt = 0x01;
572 hci_send_cmd(conn->hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
573 &cp);
577 /* Enable security */
578 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type)
580 BT_DBG("conn %p", conn);
582 /* For sdp we don't need the link key. */
583 if (sec_level == BT_SECURITY_SDP)
584 return 1;
586 /* For non 2.1 devices and low security level we don't need the link
587 key. */
588 if (sec_level == BT_SECURITY_LOW &&
589 (!conn->ssp_mode || !conn->hdev->ssp_mode))
590 return 1;
592 /* For other security levels we need the link key. */
593 if (!(conn->link_mode & HCI_LM_AUTH))
594 goto auth;
596 /* An authenticated combination key has sufficient security for any
597 security level. */
598 if (conn->key_type == HCI_LK_AUTH_COMBINATION)
599 goto encrypt;
601 /* An unauthenticated combination key has sufficient security for
602 security level 1 and 2. */
603 if (conn->key_type == HCI_LK_UNAUTH_COMBINATION &&
604 (sec_level == BT_SECURITY_MEDIUM ||
605 sec_level == BT_SECURITY_LOW))
606 goto encrypt;
608 /* A combination key has always sufficient security for the security
609 levels 1 or 2. High security level requires the combination key
610 is generated using maximum PIN code length (16).
611 For pre 2.1 units. */
612 if (conn->key_type == HCI_LK_COMBINATION &&
613 (sec_level != BT_SECURITY_HIGH ||
614 conn->pin_length == 16))
615 goto encrypt;
617 auth:
618 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
619 return 0;
621 if (!hci_conn_auth(conn, sec_level, auth_type))
622 return 0;
624 encrypt:
625 if (conn->link_mode & HCI_LM_ENCRYPT)
626 return 1;
628 hci_conn_encrypt(conn);
629 return 0;
631 EXPORT_SYMBOL(hci_conn_security);
633 /* Check secure link requirement */
634 int hci_conn_check_secure(struct hci_conn *conn, __u8 sec_level)
636 BT_DBG("conn %p", conn);
638 if (sec_level != BT_SECURITY_HIGH)
639 return 1; /* Accept if non-secure is required */
641 if (conn->key_type == HCI_LK_AUTH_COMBINATION ||
642 (conn->key_type == HCI_LK_COMBINATION &&
643 conn->pin_length == 16))
644 return 1;
646 return 0; /* Reject not secure link */
648 EXPORT_SYMBOL(hci_conn_check_secure);
650 /* Change link key */
651 int hci_conn_change_link_key(struct hci_conn *conn)
653 BT_DBG("conn %p", conn);
655 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->pend)) {
656 struct hci_cp_change_conn_link_key cp;
657 cp.handle = cpu_to_le16(conn->handle);
658 hci_send_cmd(conn->hdev, HCI_OP_CHANGE_CONN_LINK_KEY,
659 sizeof(cp), &cp);
662 return 0;
664 EXPORT_SYMBOL(hci_conn_change_link_key);
666 /* Switch role */
667 int hci_conn_switch_role(struct hci_conn *conn, __u8 role)
669 BT_DBG("conn %p", conn);
671 if (!role && conn->link_mode & HCI_LM_MASTER)
672 return 1;
674 if (!test_and_set_bit(HCI_CONN_RSWITCH_PEND, &conn->pend)) {
675 struct hci_cp_switch_role cp;
676 bacpy(&cp.bdaddr, &conn->dst);
677 cp.role = role;
678 hci_send_cmd(conn->hdev, HCI_OP_SWITCH_ROLE, sizeof(cp), &cp);
681 return 0;
683 EXPORT_SYMBOL(hci_conn_switch_role);
685 /* Enter active mode */
686 void hci_conn_enter_active_mode(struct hci_conn *conn)
688 struct hci_dev *hdev = conn->hdev;
690 BT_DBG("conn %p mode %d", conn, conn->mode);
692 if (test_bit(HCI_RAW, &hdev->flags))
693 return;
695 if (conn->mode != HCI_CM_SNIFF || !conn->power_save)
696 goto timer;
698 if (!test_and_set_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->pend)) {
699 struct hci_cp_exit_sniff_mode cp;
700 cp.handle = cpu_to_le16(conn->handle);
701 hci_send_cmd(hdev, HCI_OP_EXIT_SNIFF_MODE, sizeof(cp), &cp);
704 timer:
705 if (hdev->idle_timeout > 0)
706 mod_timer(&conn->idle_timer,
707 jiffies + msecs_to_jiffies(hdev->idle_timeout));
710 /* Enter sniff mode */
711 void hci_conn_enter_sniff_mode(struct hci_conn *conn)
713 struct hci_dev *hdev = conn->hdev;
715 BT_DBG("conn %p mode %d", conn, conn->mode);
717 if (test_bit(HCI_RAW, &hdev->flags))
718 return;
720 if (!lmp_sniff_capable(hdev) || !lmp_sniff_capable(conn))
721 return;
723 if (conn->mode != HCI_CM_ACTIVE || !(conn->link_policy & HCI_LP_SNIFF))
724 return;
726 if (lmp_sniffsubr_capable(hdev) && lmp_sniffsubr_capable(conn)) {
727 struct hci_cp_sniff_subrate cp;
728 cp.handle = cpu_to_le16(conn->handle);
729 cp.max_latency = cpu_to_le16(0);
730 cp.min_remote_timeout = cpu_to_le16(0);
731 cp.min_local_timeout = cpu_to_le16(0);
732 hci_send_cmd(hdev, HCI_OP_SNIFF_SUBRATE, sizeof(cp), &cp);
735 if (!test_and_set_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->pend)) {
736 struct hci_cp_sniff_mode cp;
737 cp.handle = cpu_to_le16(conn->handle);
738 cp.max_interval = cpu_to_le16(hdev->sniff_max_interval);
739 cp.min_interval = cpu_to_le16(hdev->sniff_min_interval);
740 cp.attempt = cpu_to_le16(4);
741 cp.timeout = cpu_to_le16(1);
742 hci_send_cmd(hdev, HCI_OP_SNIFF_MODE, sizeof(cp), &cp);
746 /* Drop all connection on the device */
747 void hci_conn_hash_flush(struct hci_dev *hdev)
749 struct hci_conn_hash *h = &hdev->conn_hash;
750 struct list_head *p;
752 BT_DBG("hdev %s", hdev->name);
754 p = h->list.next;
755 while (p != &h->list) {
756 struct hci_conn *c;
758 c = list_entry(p, struct hci_conn, list);
759 p = p->next;
761 c->state = BT_CLOSED;
763 hci_proto_disconn_cfm(c, 0x16);
764 hci_conn_del(c);
768 /* Check pending connect attempts */
769 void hci_conn_check_pending(struct hci_dev *hdev)
771 struct hci_conn *conn;
773 BT_DBG("hdev %s", hdev->name);
775 hci_dev_lock(hdev);
777 conn = hci_conn_hash_lookup_state(hdev, ACL_LINK, BT_CONNECT2);
778 if (conn)
779 hci_acl_connect(conn);
781 hci_dev_unlock(hdev);
784 void hci_conn_hold_device(struct hci_conn *conn)
786 atomic_inc(&conn->devref);
788 EXPORT_SYMBOL(hci_conn_hold_device);
790 void hci_conn_put_device(struct hci_conn *conn)
792 if (atomic_dec_and_test(&conn->devref))
793 hci_conn_del_sysfs(conn);
795 EXPORT_SYMBOL(hci_conn_put_device);
797 int hci_get_conn_list(void __user *arg)
799 struct hci_conn_list_req req, *cl;
800 struct hci_conn_info *ci;
801 struct hci_dev *hdev;
802 struct list_head *p;
803 int n = 0, size, err;
805 if (copy_from_user(&req, arg, sizeof(req)))
806 return -EFAULT;
808 if (!req.conn_num || req.conn_num > (PAGE_SIZE * 2) / sizeof(*ci))
809 return -EINVAL;
811 size = sizeof(req) + req.conn_num * sizeof(*ci);
813 cl = kmalloc(size, GFP_KERNEL);
814 if (!cl)
815 return -ENOMEM;
817 hdev = hci_dev_get(req.dev_id);
818 if (!hdev) {
819 kfree(cl);
820 return -ENODEV;
823 ci = cl->conn_info;
825 hci_dev_lock_bh(hdev);
826 list_for_each(p, &hdev->conn_hash.list) {
827 register struct hci_conn *c;
828 c = list_entry(p, struct hci_conn, list);
830 bacpy(&(ci + n)->bdaddr, &c->dst);
831 (ci + n)->handle = c->handle;
832 (ci + n)->type = c->type;
833 (ci + n)->out = c->out;
834 (ci + n)->state = c->state;
835 (ci + n)->link_mode = c->link_mode;
836 if (++n >= req.conn_num)
837 break;
839 hci_dev_unlock_bh(hdev);
841 cl->dev_id = hdev->id;
842 cl->conn_num = n;
843 size = sizeof(req) + n * sizeof(*ci);
845 hci_dev_put(hdev);
847 err = copy_to_user(arg, cl, size);
848 kfree(cl);
850 return err ? -EFAULT : 0;
853 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg)
855 struct hci_conn_info_req req;
856 struct hci_conn_info ci;
857 struct hci_conn *conn;
858 char __user *ptr = arg + sizeof(req);
860 if (copy_from_user(&req, arg, sizeof(req)))
861 return -EFAULT;
863 hci_dev_lock_bh(hdev);
864 conn = hci_conn_hash_lookup_ba(hdev, req.type, &req.bdaddr);
865 if (conn) {
866 bacpy(&ci.bdaddr, &conn->dst);
867 ci.handle = conn->handle;
868 ci.type = conn->type;
869 ci.out = conn->out;
870 ci.state = conn->state;
871 ci.link_mode = conn->link_mode;
873 hci_dev_unlock_bh(hdev);
875 if (!conn)
876 return -ENOENT;
878 return copy_to_user(ptr, &ci, sizeof(ci)) ? -EFAULT : 0;
881 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg)
883 struct hci_auth_info_req req;
884 struct hci_conn *conn;
886 if (copy_from_user(&req, arg, sizeof(req)))
887 return -EFAULT;
889 hci_dev_lock_bh(hdev);
890 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &req.bdaddr);
891 if (conn)
892 req.type = conn->auth_type;
893 hci_dev_unlock_bh(hdev);
895 if (!conn)
896 return -ENOENT;
898 return copy_to_user(arg, &req, sizeof(req)) ? -EFAULT : 0;