[PATCH] selinux: eliminate selinux_task_ctxid
[pv_ops_mirror.git] / net / bluetooth / l2cap.c
blobd56f60b392ac512b36a736f928844208ec88847a
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
25 /* Bluetooth L2CAP core and sockets. */
27 #include <linux/module.h>
29 #include <linux/types.h>
30 #include <linux/capability.h>
31 #include <linux/errno.h>
32 #include <linux/kernel.h>
33 #include <linux/sched.h>
34 #include <linux/slab.h>
35 #include <linux/poll.h>
36 #include <linux/fcntl.h>
37 #include <linux/init.h>
38 #include <linux/interrupt.h>
39 #include <linux/socket.h>
40 #include <linux/skbuff.h>
41 #include <linux/list.h>
42 #include <linux/device.h>
43 #include <net/sock.h>
45 #include <asm/system.h>
46 #include <asm/uaccess.h>
47 #include <asm/unaligned.h>
49 #include <net/bluetooth/bluetooth.h>
50 #include <net/bluetooth/hci_core.h>
51 #include <net/bluetooth/l2cap.h>
53 #ifndef CONFIG_BT_L2CAP_DEBUG
54 #undef BT_DBG
55 #define BT_DBG(D...)
56 #endif
58 #define VERSION "2.8"
60 static const struct proto_ops l2cap_sock_ops;
62 static struct bt_sock_list l2cap_sk_list = {
63 .lock = RW_LOCK_UNLOCKED
66 static void __l2cap_sock_close(struct sock *sk, int reason);
67 static void l2cap_sock_close(struct sock *sk);
68 static void l2cap_sock_kill(struct sock *sk);
70 static struct sk_buff *l2cap_build_cmd(struct l2cap_conn *conn,
71 u8 code, u8 ident, u16 dlen, void *data);
73 /* ---- L2CAP timers ---- */
74 static void l2cap_sock_timeout(unsigned long arg)
76 struct sock *sk = (struct sock *) arg;
78 BT_DBG("sock %p state %d", sk, sk->sk_state);
80 bh_lock_sock(sk);
81 __l2cap_sock_close(sk, ETIMEDOUT);
82 bh_unlock_sock(sk);
84 l2cap_sock_kill(sk);
85 sock_put(sk);
88 static void l2cap_sock_set_timer(struct sock *sk, long timeout)
90 BT_DBG("sk %p state %d timeout %ld", sk, sk->sk_state, timeout);
91 sk_reset_timer(sk, &sk->sk_timer, jiffies + timeout);
94 static void l2cap_sock_clear_timer(struct sock *sk)
96 BT_DBG("sock %p state %d", sk, sk->sk_state);
97 sk_stop_timer(sk, &sk->sk_timer);
100 static void l2cap_sock_init_timer(struct sock *sk)
102 init_timer(&sk->sk_timer);
103 sk->sk_timer.function = l2cap_sock_timeout;
104 sk->sk_timer.data = (unsigned long)sk;
107 /* ---- L2CAP channels ---- */
108 static struct sock *__l2cap_get_chan_by_dcid(struct l2cap_chan_list *l, u16 cid)
110 struct sock *s;
111 for (s = l->head; s; s = l2cap_pi(s)->next_c) {
112 if (l2cap_pi(s)->dcid == cid)
113 break;
115 return s;
118 static struct sock *__l2cap_get_chan_by_scid(struct l2cap_chan_list *l, u16 cid)
120 struct sock *s;
121 for (s = l->head; s; s = l2cap_pi(s)->next_c) {
122 if (l2cap_pi(s)->scid == cid)
123 break;
125 return s;
128 /* Find channel with given SCID.
129 * Returns locked socket */
130 static inline struct sock *l2cap_get_chan_by_scid(struct l2cap_chan_list *l, u16 cid)
132 struct sock *s;
133 read_lock(&l->lock);
134 s = __l2cap_get_chan_by_scid(l, cid);
135 if (s) bh_lock_sock(s);
136 read_unlock(&l->lock);
137 return s;
140 static struct sock *__l2cap_get_chan_by_ident(struct l2cap_chan_list *l, u8 ident)
142 struct sock *s;
143 for (s = l->head; s; s = l2cap_pi(s)->next_c) {
144 if (l2cap_pi(s)->ident == ident)
145 break;
147 return s;
150 static inline struct sock *l2cap_get_chan_by_ident(struct l2cap_chan_list *l, u8 ident)
152 struct sock *s;
153 read_lock(&l->lock);
154 s = __l2cap_get_chan_by_ident(l, ident);
155 if (s) bh_lock_sock(s);
156 read_unlock(&l->lock);
157 return s;
160 static u16 l2cap_alloc_cid(struct l2cap_chan_list *l)
162 u16 cid = 0x0040;
164 for (; cid < 0xffff; cid++) {
165 if(!__l2cap_get_chan_by_scid(l, cid))
166 return cid;
169 return 0;
172 static inline void __l2cap_chan_link(struct l2cap_chan_list *l, struct sock *sk)
174 sock_hold(sk);
176 if (l->head)
177 l2cap_pi(l->head)->prev_c = sk;
179 l2cap_pi(sk)->next_c = l->head;
180 l2cap_pi(sk)->prev_c = NULL;
181 l->head = sk;
184 static inline void l2cap_chan_unlink(struct l2cap_chan_list *l, struct sock *sk)
186 struct sock *next = l2cap_pi(sk)->next_c, *prev = l2cap_pi(sk)->prev_c;
188 write_lock_bh(&l->lock);
189 if (sk == l->head)
190 l->head = next;
192 if (next)
193 l2cap_pi(next)->prev_c = prev;
194 if (prev)
195 l2cap_pi(prev)->next_c = next;
196 write_unlock_bh(&l->lock);
198 __sock_put(sk);
201 static void __l2cap_chan_add(struct l2cap_conn *conn, struct sock *sk, struct sock *parent)
203 struct l2cap_chan_list *l = &conn->chan_list;
205 BT_DBG("conn %p, psm 0x%2.2x, dcid 0x%4.4x", conn, l2cap_pi(sk)->psm, l2cap_pi(sk)->dcid);
207 l2cap_pi(sk)->conn = conn;
209 if (sk->sk_type == SOCK_SEQPACKET) {
210 /* Alloc CID for connection-oriented socket */
211 l2cap_pi(sk)->scid = l2cap_alloc_cid(l);
212 } else if (sk->sk_type == SOCK_DGRAM) {
213 /* Connectionless socket */
214 l2cap_pi(sk)->scid = 0x0002;
215 l2cap_pi(sk)->dcid = 0x0002;
216 l2cap_pi(sk)->omtu = L2CAP_DEFAULT_MTU;
217 } else {
218 /* Raw socket can send/recv signalling messages only */
219 l2cap_pi(sk)->scid = 0x0001;
220 l2cap_pi(sk)->dcid = 0x0001;
221 l2cap_pi(sk)->omtu = L2CAP_DEFAULT_MTU;
224 __l2cap_chan_link(l, sk);
226 if (parent)
227 bt_accept_enqueue(parent, sk);
230 /* Delete channel.
231 * Must be called on the locked socket. */
232 static void l2cap_chan_del(struct sock *sk, int err)
234 struct l2cap_conn *conn = l2cap_pi(sk)->conn;
235 struct sock *parent = bt_sk(sk)->parent;
237 l2cap_sock_clear_timer(sk);
239 BT_DBG("sk %p, conn %p, err %d", sk, conn, err);
241 if (conn) {
242 /* Unlink from channel list */
243 l2cap_chan_unlink(&conn->chan_list, sk);
244 l2cap_pi(sk)->conn = NULL;
245 hci_conn_put(conn->hcon);
248 sk->sk_state = BT_CLOSED;
249 sock_set_flag(sk, SOCK_ZAPPED);
251 if (err)
252 sk->sk_err = err;
254 if (parent) {
255 bt_accept_unlink(sk);
256 parent->sk_data_ready(parent, 0);
257 } else
258 sk->sk_state_change(sk);
261 /* ---- L2CAP connections ---- */
262 static struct l2cap_conn *l2cap_conn_add(struct hci_conn *hcon, u8 status)
264 struct l2cap_conn *conn = hcon->l2cap_data;
266 if (conn || status)
267 return conn;
269 conn = kzalloc(sizeof(struct l2cap_conn), GFP_ATOMIC);
270 if (!conn)
271 return NULL;
273 hcon->l2cap_data = conn;
274 conn->hcon = hcon;
276 BT_DBG("hcon %p conn %p", hcon, conn);
278 conn->mtu = hcon->hdev->acl_mtu;
279 conn->src = &hcon->hdev->bdaddr;
280 conn->dst = &hcon->dst;
282 spin_lock_init(&conn->lock);
283 rwlock_init(&conn->chan_list.lock);
285 return conn;
288 static void l2cap_conn_del(struct hci_conn *hcon, int err)
290 struct l2cap_conn *conn = hcon->l2cap_data;
291 struct sock *sk;
293 if (!conn)
294 return;
296 BT_DBG("hcon %p conn %p, err %d", hcon, conn, err);
298 if (conn->rx_skb)
299 kfree_skb(conn->rx_skb);
301 /* Kill channels */
302 while ((sk = conn->chan_list.head)) {
303 bh_lock_sock(sk);
304 l2cap_chan_del(sk, err);
305 bh_unlock_sock(sk);
306 l2cap_sock_kill(sk);
309 hcon->l2cap_data = NULL;
310 kfree(conn);
313 static inline void l2cap_chan_add(struct l2cap_conn *conn, struct sock *sk, struct sock *parent)
315 struct l2cap_chan_list *l = &conn->chan_list;
316 write_lock_bh(&l->lock);
317 __l2cap_chan_add(conn, sk, parent);
318 write_unlock_bh(&l->lock);
321 static inline u8 l2cap_get_ident(struct l2cap_conn *conn)
323 u8 id;
325 /* Get next available identificator.
326 * 1 - 128 are used by kernel.
327 * 129 - 199 are reserved.
328 * 200 - 254 are used by utilities like l2ping, etc.
331 spin_lock_bh(&conn->lock);
333 if (++conn->tx_ident > 128)
334 conn->tx_ident = 1;
336 id = conn->tx_ident;
338 spin_unlock_bh(&conn->lock);
340 return id;
343 static inline int l2cap_send_cmd(struct l2cap_conn *conn, u8 ident, u8 code, u16 len, void *data)
345 struct sk_buff *skb = l2cap_build_cmd(conn, code, ident, len, data);
347 BT_DBG("code 0x%2.2x", code);
349 if (!skb)
350 return -ENOMEM;
352 return hci_send_acl(conn->hcon, skb, 0);
355 /* ---- Socket interface ---- */
356 static struct sock *__l2cap_get_sock_by_addr(u16 psm, bdaddr_t *src)
358 struct sock *sk;
359 struct hlist_node *node;
360 sk_for_each(sk, node, &l2cap_sk_list.head)
361 if (l2cap_pi(sk)->sport == psm && !bacmp(&bt_sk(sk)->src, src))
362 goto found;
363 sk = NULL;
364 found:
365 return sk;
368 /* Find socket with psm and source bdaddr.
369 * Returns closest match.
371 static struct sock *__l2cap_get_sock_by_psm(int state, u16 psm, bdaddr_t *src)
373 struct sock *sk = NULL, *sk1 = NULL;
374 struct hlist_node *node;
376 sk_for_each(sk, node, &l2cap_sk_list.head) {
377 if (state && sk->sk_state != state)
378 continue;
380 if (l2cap_pi(sk)->psm == psm) {
381 /* Exact match. */
382 if (!bacmp(&bt_sk(sk)->src, src))
383 break;
385 /* Closest match */
386 if (!bacmp(&bt_sk(sk)->src, BDADDR_ANY))
387 sk1 = sk;
390 return node ? sk : sk1;
393 /* Find socket with given address (psm, src).
394 * Returns locked socket */
395 static inline struct sock *l2cap_get_sock_by_psm(int state, u16 psm, bdaddr_t *src)
397 struct sock *s;
398 read_lock(&l2cap_sk_list.lock);
399 s = __l2cap_get_sock_by_psm(state, psm, src);
400 if (s) bh_lock_sock(s);
401 read_unlock(&l2cap_sk_list.lock);
402 return s;
405 static void l2cap_sock_destruct(struct sock *sk)
407 BT_DBG("sk %p", sk);
409 skb_queue_purge(&sk->sk_receive_queue);
410 skb_queue_purge(&sk->sk_write_queue);
413 static void l2cap_sock_cleanup_listen(struct sock *parent)
415 struct sock *sk;
417 BT_DBG("parent %p", parent);
419 /* Close not yet accepted channels */
420 while ((sk = bt_accept_dequeue(parent, NULL)))
421 l2cap_sock_close(sk);
423 parent->sk_state = BT_CLOSED;
424 sock_set_flag(parent, SOCK_ZAPPED);
427 /* Kill socket (only if zapped and orphan)
428 * Must be called on unlocked socket.
430 static void l2cap_sock_kill(struct sock *sk)
432 if (!sock_flag(sk, SOCK_ZAPPED) || sk->sk_socket)
433 return;
435 BT_DBG("sk %p state %d", sk, sk->sk_state);
437 /* Kill poor orphan */
438 bt_sock_unlink(&l2cap_sk_list, sk);
439 sock_set_flag(sk, SOCK_DEAD);
440 sock_put(sk);
443 static void __l2cap_sock_close(struct sock *sk, int reason)
445 BT_DBG("sk %p state %d socket %p", sk, sk->sk_state, sk->sk_socket);
447 switch (sk->sk_state) {
448 case BT_LISTEN:
449 l2cap_sock_cleanup_listen(sk);
450 break;
452 case BT_CONNECTED:
453 case BT_CONFIG:
454 case BT_CONNECT2:
455 if (sk->sk_type == SOCK_SEQPACKET) {
456 struct l2cap_conn *conn = l2cap_pi(sk)->conn;
457 struct l2cap_disconn_req req;
459 sk->sk_state = BT_DISCONN;
460 l2cap_sock_set_timer(sk, sk->sk_sndtimeo);
462 req.dcid = __cpu_to_le16(l2cap_pi(sk)->dcid);
463 req.scid = __cpu_to_le16(l2cap_pi(sk)->scid);
464 l2cap_send_cmd(conn, l2cap_get_ident(conn),
465 L2CAP_DISCONN_REQ, sizeof(req), &req);
466 } else {
467 l2cap_chan_del(sk, reason);
469 break;
471 case BT_CONNECT:
472 case BT_DISCONN:
473 l2cap_chan_del(sk, reason);
474 break;
476 default:
477 sock_set_flag(sk, SOCK_ZAPPED);
478 break;
482 /* Must be called on unlocked socket. */
483 static void l2cap_sock_close(struct sock *sk)
485 l2cap_sock_clear_timer(sk);
486 lock_sock(sk);
487 __l2cap_sock_close(sk, ECONNRESET);
488 release_sock(sk);
489 l2cap_sock_kill(sk);
492 static void l2cap_sock_init(struct sock *sk, struct sock *parent)
494 struct l2cap_pinfo *pi = l2cap_pi(sk);
496 BT_DBG("sk %p", sk);
498 if (parent) {
499 sk->sk_type = parent->sk_type;
500 pi->imtu = l2cap_pi(parent)->imtu;
501 pi->omtu = l2cap_pi(parent)->omtu;
502 pi->link_mode = l2cap_pi(parent)->link_mode;
503 } else {
504 pi->imtu = L2CAP_DEFAULT_MTU;
505 pi->omtu = 0;
506 pi->link_mode = 0;
509 /* Default config options */
510 pi->conf_mtu = L2CAP_DEFAULT_MTU;
511 pi->flush_to = L2CAP_DEFAULT_FLUSH_TO;
514 static struct proto l2cap_proto = {
515 .name = "L2CAP",
516 .owner = THIS_MODULE,
517 .obj_size = sizeof(struct l2cap_pinfo)
520 static struct sock *l2cap_sock_alloc(struct socket *sock, int proto, gfp_t prio)
522 struct sock *sk;
524 sk = sk_alloc(PF_BLUETOOTH, prio, &l2cap_proto, 1);
525 if (!sk)
526 return NULL;
528 sock_init_data(sock, sk);
529 INIT_LIST_HEAD(&bt_sk(sk)->accept_q);
531 sk->sk_destruct = l2cap_sock_destruct;
532 sk->sk_sndtimeo = L2CAP_CONN_TIMEOUT;
534 sock_reset_flag(sk, SOCK_ZAPPED);
536 sk->sk_protocol = proto;
537 sk->sk_state = BT_OPEN;
539 l2cap_sock_init_timer(sk);
541 bt_sock_link(&l2cap_sk_list, sk);
542 return sk;
545 static int l2cap_sock_create(struct socket *sock, int protocol)
547 struct sock *sk;
549 BT_DBG("sock %p", sock);
551 sock->state = SS_UNCONNECTED;
553 if (sock->type != SOCK_SEQPACKET &&
554 sock->type != SOCK_DGRAM && sock->type != SOCK_RAW)
555 return -ESOCKTNOSUPPORT;
557 if (sock->type == SOCK_RAW && !capable(CAP_NET_RAW))
558 return -EPERM;
560 sock->ops = &l2cap_sock_ops;
562 sk = l2cap_sock_alloc(sock, protocol, GFP_KERNEL);
563 if (!sk)
564 return -ENOMEM;
566 l2cap_sock_init(sk, NULL);
567 return 0;
570 static int l2cap_sock_bind(struct socket *sock, struct sockaddr *addr, int addr_len)
572 struct sockaddr_l2 *la = (struct sockaddr_l2 *) addr;
573 struct sock *sk = sock->sk;
574 int err = 0;
576 BT_DBG("sk %p, %s %d", sk, batostr(&la->l2_bdaddr), la->l2_psm);
578 if (!addr || addr->sa_family != AF_BLUETOOTH)
579 return -EINVAL;
581 lock_sock(sk);
583 if (sk->sk_state != BT_OPEN) {
584 err = -EBADFD;
585 goto done;
588 write_lock_bh(&l2cap_sk_list.lock);
590 if (la->l2_psm && __l2cap_get_sock_by_addr(la->l2_psm, &la->l2_bdaddr)) {
591 err = -EADDRINUSE;
592 } else {
593 /* Save source address */
594 bacpy(&bt_sk(sk)->src, &la->l2_bdaddr);
595 l2cap_pi(sk)->psm = la->l2_psm;
596 l2cap_pi(sk)->sport = la->l2_psm;
597 sk->sk_state = BT_BOUND;
600 write_unlock_bh(&l2cap_sk_list.lock);
602 done:
603 release_sock(sk);
604 return err;
607 static int l2cap_do_connect(struct sock *sk)
609 bdaddr_t *src = &bt_sk(sk)->src;
610 bdaddr_t *dst = &bt_sk(sk)->dst;
611 struct l2cap_conn *conn;
612 struct hci_conn *hcon;
613 struct hci_dev *hdev;
614 int err = 0;
616 BT_DBG("%s -> %s psm 0x%2.2x", batostr(src), batostr(dst), l2cap_pi(sk)->psm);
618 if (!(hdev = hci_get_route(dst, src)))
619 return -EHOSTUNREACH;
621 hci_dev_lock_bh(hdev);
623 err = -ENOMEM;
625 hcon = hci_connect(hdev, ACL_LINK, dst);
626 if (!hcon)
627 goto done;
629 conn = l2cap_conn_add(hcon, 0);
630 if (!conn) {
631 hci_conn_put(hcon);
632 goto done;
635 err = 0;
637 /* Update source addr of the socket */
638 bacpy(src, conn->src);
640 l2cap_chan_add(conn, sk, NULL);
642 sk->sk_state = BT_CONNECT;
643 l2cap_sock_set_timer(sk, sk->sk_sndtimeo);
645 if (hcon->state == BT_CONNECTED) {
646 if (sk->sk_type == SOCK_SEQPACKET) {
647 struct l2cap_conn_req req;
648 l2cap_pi(sk)->ident = l2cap_get_ident(conn);
649 req.scid = __cpu_to_le16(l2cap_pi(sk)->scid);
650 req.psm = l2cap_pi(sk)->psm;
651 l2cap_send_cmd(conn, l2cap_pi(sk)->ident,
652 L2CAP_CONN_REQ, sizeof(req), &req);
653 } else {
654 l2cap_sock_clear_timer(sk);
655 sk->sk_state = BT_CONNECTED;
659 done:
660 hci_dev_unlock_bh(hdev);
661 hci_dev_put(hdev);
662 return err;
665 static int l2cap_sock_connect(struct socket *sock, struct sockaddr *addr, int alen, int flags)
667 struct sockaddr_l2 *la = (struct sockaddr_l2 *) addr;
668 struct sock *sk = sock->sk;
669 int err = 0;
671 lock_sock(sk);
673 BT_DBG("sk %p", sk);
675 if (addr->sa_family != AF_BLUETOOTH || alen < sizeof(struct sockaddr_l2)) {
676 err = -EINVAL;
677 goto done;
680 if (sk->sk_type == SOCK_SEQPACKET && !la->l2_psm) {
681 err = -EINVAL;
682 goto done;
685 switch(sk->sk_state) {
686 case BT_CONNECT:
687 case BT_CONNECT2:
688 case BT_CONFIG:
689 /* Already connecting */
690 goto wait;
692 case BT_CONNECTED:
693 /* Already connected */
694 goto done;
696 case BT_OPEN:
697 case BT_BOUND:
698 /* Can connect */
699 break;
701 default:
702 err = -EBADFD;
703 goto done;
706 /* Set destination address and psm */
707 bacpy(&bt_sk(sk)->dst, &la->l2_bdaddr);
708 l2cap_pi(sk)->psm = la->l2_psm;
710 if ((err = l2cap_do_connect(sk)))
711 goto done;
713 wait:
714 err = bt_sock_wait_state(sk, BT_CONNECTED,
715 sock_sndtimeo(sk, flags & O_NONBLOCK));
716 done:
717 release_sock(sk);
718 return err;
721 static int l2cap_sock_listen(struct socket *sock, int backlog)
723 struct sock *sk = sock->sk;
724 int err = 0;
726 BT_DBG("sk %p backlog %d", sk, backlog);
728 lock_sock(sk);
730 if (sk->sk_state != BT_BOUND || sock->type != SOCK_SEQPACKET) {
731 err = -EBADFD;
732 goto done;
735 if (!l2cap_pi(sk)->psm) {
736 bdaddr_t *src = &bt_sk(sk)->src;
737 u16 psm;
739 err = -EINVAL;
741 write_lock_bh(&l2cap_sk_list.lock);
743 for (psm = 0x1001; psm < 0x1100; psm += 2)
744 if (!__l2cap_get_sock_by_addr(psm, src)) {
745 l2cap_pi(sk)->psm = htobs(psm);
746 l2cap_pi(sk)->sport = htobs(psm);
747 err = 0;
748 break;
751 write_unlock_bh(&l2cap_sk_list.lock);
753 if (err < 0)
754 goto done;
757 sk->sk_max_ack_backlog = backlog;
758 sk->sk_ack_backlog = 0;
759 sk->sk_state = BT_LISTEN;
761 done:
762 release_sock(sk);
763 return err;
766 static int l2cap_sock_accept(struct socket *sock, struct socket *newsock, int flags)
768 DECLARE_WAITQUEUE(wait, current);
769 struct sock *sk = sock->sk, *nsk;
770 long timeo;
771 int err = 0;
773 lock_sock(sk);
775 if (sk->sk_state != BT_LISTEN) {
776 err = -EBADFD;
777 goto done;
780 timeo = sock_rcvtimeo(sk, flags & O_NONBLOCK);
782 BT_DBG("sk %p timeo %ld", sk, timeo);
784 /* Wait for an incoming connection. (wake-one). */
785 add_wait_queue_exclusive(sk->sk_sleep, &wait);
786 while (!(nsk = bt_accept_dequeue(sk, newsock))) {
787 set_current_state(TASK_INTERRUPTIBLE);
788 if (!timeo) {
789 err = -EAGAIN;
790 break;
793 release_sock(sk);
794 timeo = schedule_timeout(timeo);
795 lock_sock(sk);
797 if (sk->sk_state != BT_LISTEN) {
798 err = -EBADFD;
799 break;
802 if (signal_pending(current)) {
803 err = sock_intr_errno(timeo);
804 break;
807 set_current_state(TASK_RUNNING);
808 remove_wait_queue(sk->sk_sleep, &wait);
810 if (err)
811 goto done;
813 newsock->state = SS_CONNECTED;
815 BT_DBG("new socket %p", nsk);
817 done:
818 release_sock(sk);
819 return err;
822 static int l2cap_sock_getname(struct socket *sock, struct sockaddr *addr, int *len, int peer)
824 struct sockaddr_l2 *la = (struct sockaddr_l2 *) addr;
825 struct sock *sk = sock->sk;
827 BT_DBG("sock %p, sk %p", sock, sk);
829 addr->sa_family = AF_BLUETOOTH;
830 *len = sizeof(struct sockaddr_l2);
832 if (peer)
833 bacpy(&la->l2_bdaddr, &bt_sk(sk)->dst);
834 else
835 bacpy(&la->l2_bdaddr, &bt_sk(sk)->src);
837 la->l2_psm = l2cap_pi(sk)->psm;
838 return 0;
841 static inline int l2cap_do_send(struct sock *sk, struct msghdr *msg, int len)
843 struct l2cap_conn *conn = l2cap_pi(sk)->conn;
844 struct sk_buff *skb, **frag;
845 int err, hlen, count, sent=0;
846 struct l2cap_hdr *lh;
848 BT_DBG("sk %p len %d", sk, len);
850 /* First fragment (with L2CAP header) */
851 if (sk->sk_type == SOCK_DGRAM)
852 hlen = L2CAP_HDR_SIZE + 2;
853 else
854 hlen = L2CAP_HDR_SIZE;
856 count = min_t(unsigned int, (conn->mtu - hlen), len);
858 skb = bt_skb_send_alloc(sk, hlen + count,
859 msg->msg_flags & MSG_DONTWAIT, &err);
860 if (!skb)
861 return err;
863 /* Create L2CAP header */
864 lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE);
865 lh->cid = __cpu_to_le16(l2cap_pi(sk)->dcid);
866 lh->len = __cpu_to_le16(len + (hlen - L2CAP_HDR_SIZE));
868 if (sk->sk_type == SOCK_DGRAM)
869 put_unaligned(l2cap_pi(sk)->psm, (u16 *) skb_put(skb, 2));
871 if (memcpy_fromiovec(skb_put(skb, count), msg->msg_iov, count)) {
872 err = -EFAULT;
873 goto fail;
876 sent += count;
877 len -= count;
879 /* Continuation fragments (no L2CAP header) */
880 frag = &skb_shinfo(skb)->frag_list;
881 while (len) {
882 count = min_t(unsigned int, conn->mtu, len);
884 *frag = bt_skb_send_alloc(sk, count, msg->msg_flags & MSG_DONTWAIT, &err);
885 if (!*frag)
886 goto fail;
888 if (memcpy_fromiovec(skb_put(*frag, count), msg->msg_iov, count)) {
889 err = -EFAULT;
890 goto fail;
893 sent += count;
894 len -= count;
896 frag = &(*frag)->next;
899 if ((err = hci_send_acl(conn->hcon, skb, 0)) < 0)
900 goto fail;
902 return sent;
904 fail:
905 kfree_skb(skb);
906 return err;
909 static int l2cap_sock_sendmsg(struct kiocb *iocb, struct socket *sock, struct msghdr *msg, size_t len)
911 struct sock *sk = sock->sk;
912 int err = 0;
914 BT_DBG("sock %p, sk %p", sock, sk);
916 err = sock_error(sk);
917 if (err)
918 return err;
920 if (msg->msg_flags & MSG_OOB)
921 return -EOPNOTSUPP;
923 /* Check outgoing MTU */
924 if (sk->sk_type != SOCK_RAW && len > l2cap_pi(sk)->omtu)
925 return -EINVAL;
927 lock_sock(sk);
929 if (sk->sk_state == BT_CONNECTED)
930 err = l2cap_do_send(sk, msg, len);
931 else
932 err = -ENOTCONN;
934 release_sock(sk);
935 return err;
938 static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, char __user *optval, int optlen)
940 struct sock *sk = sock->sk;
941 struct l2cap_options opts;
942 int err = 0, len;
943 u32 opt;
945 BT_DBG("sk %p", sk);
947 lock_sock(sk);
949 switch (optname) {
950 case L2CAP_OPTIONS:
951 len = min_t(unsigned int, sizeof(opts), optlen);
952 if (copy_from_user((char *) &opts, optval, len)) {
953 err = -EFAULT;
954 break;
956 l2cap_pi(sk)->imtu = opts.imtu;
957 l2cap_pi(sk)->omtu = opts.omtu;
958 break;
960 case L2CAP_LM:
961 if (get_user(opt, (u32 __user *) optval)) {
962 err = -EFAULT;
963 break;
966 l2cap_pi(sk)->link_mode = opt;
967 break;
969 default:
970 err = -ENOPROTOOPT;
971 break;
974 release_sock(sk);
975 return err;
978 static int l2cap_sock_getsockopt(struct socket *sock, int level, int optname, char __user *optval, int __user *optlen)
980 struct sock *sk = sock->sk;
981 struct l2cap_options opts;
982 struct l2cap_conninfo cinfo;
983 int len, err = 0;
985 BT_DBG("sk %p", sk);
987 if (get_user(len, optlen))
988 return -EFAULT;
990 lock_sock(sk);
992 switch (optname) {
993 case L2CAP_OPTIONS:
994 opts.imtu = l2cap_pi(sk)->imtu;
995 opts.omtu = l2cap_pi(sk)->omtu;
996 opts.flush_to = l2cap_pi(sk)->flush_to;
997 opts.mode = 0x00;
999 len = min_t(unsigned int, len, sizeof(opts));
1000 if (copy_to_user(optval, (char *) &opts, len))
1001 err = -EFAULT;
1003 break;
1005 case L2CAP_LM:
1006 if (put_user(l2cap_pi(sk)->link_mode, (u32 __user *) optval))
1007 err = -EFAULT;
1008 break;
1010 case L2CAP_CONNINFO:
1011 if (sk->sk_state != BT_CONNECTED) {
1012 err = -ENOTCONN;
1013 break;
1016 cinfo.hci_handle = l2cap_pi(sk)->conn->hcon->handle;
1017 memcpy(cinfo.dev_class, l2cap_pi(sk)->conn->hcon->dev_class, 3);
1019 len = min_t(unsigned int, len, sizeof(cinfo));
1020 if (copy_to_user(optval, (char *) &cinfo, len))
1021 err = -EFAULT;
1023 break;
1025 default:
1026 err = -ENOPROTOOPT;
1027 break;
1030 release_sock(sk);
1031 return err;
1034 static int l2cap_sock_shutdown(struct socket *sock, int how)
1036 struct sock *sk = sock->sk;
1037 int err = 0;
1039 BT_DBG("sock %p, sk %p", sock, sk);
1041 if (!sk)
1042 return 0;
1044 lock_sock(sk);
1045 if (!sk->sk_shutdown) {
1046 sk->sk_shutdown = SHUTDOWN_MASK;
1047 l2cap_sock_clear_timer(sk);
1048 __l2cap_sock_close(sk, 0);
1050 if (sock_flag(sk, SOCK_LINGER) && sk->sk_lingertime)
1051 err = bt_sock_wait_state(sk, BT_CLOSED, sk->sk_lingertime);
1053 release_sock(sk);
1054 return err;
1057 static int l2cap_sock_release(struct socket *sock)
1059 struct sock *sk = sock->sk;
1060 int err;
1062 BT_DBG("sock %p, sk %p", sock, sk);
1064 if (!sk)
1065 return 0;
1067 err = l2cap_sock_shutdown(sock, 2);
1069 sock_orphan(sk);
1070 l2cap_sock_kill(sk);
1071 return err;
1074 static void l2cap_conn_ready(struct l2cap_conn *conn)
1076 struct l2cap_chan_list *l = &conn->chan_list;
1077 struct sock *sk;
1079 BT_DBG("conn %p", conn);
1081 read_lock(&l->lock);
1083 for (sk = l->head; sk; sk = l2cap_pi(sk)->next_c) {
1084 bh_lock_sock(sk);
1086 if (sk->sk_type != SOCK_SEQPACKET) {
1087 l2cap_sock_clear_timer(sk);
1088 sk->sk_state = BT_CONNECTED;
1089 sk->sk_state_change(sk);
1090 } else if (sk->sk_state == BT_CONNECT) {
1091 struct l2cap_conn_req req;
1092 l2cap_pi(sk)->ident = l2cap_get_ident(conn);
1093 req.scid = __cpu_to_le16(l2cap_pi(sk)->scid);
1094 req.psm = l2cap_pi(sk)->psm;
1095 l2cap_send_cmd(conn, l2cap_pi(sk)->ident, L2CAP_CONN_REQ, sizeof(req), &req);
1098 bh_unlock_sock(sk);
1101 read_unlock(&l->lock);
1104 /* Notify sockets that we cannot guaranty reliability anymore */
1105 static void l2cap_conn_unreliable(struct l2cap_conn *conn, int err)
1107 struct l2cap_chan_list *l = &conn->chan_list;
1108 struct sock *sk;
1110 BT_DBG("conn %p", conn);
1112 read_lock(&l->lock);
1113 for (sk = l->head; sk; sk = l2cap_pi(sk)->next_c) {
1114 if (l2cap_pi(sk)->link_mode & L2CAP_LM_RELIABLE)
1115 sk->sk_err = err;
1117 read_unlock(&l->lock);
1120 static void l2cap_chan_ready(struct sock *sk)
1122 struct sock *parent = bt_sk(sk)->parent;
1124 BT_DBG("sk %p, parent %p", sk, parent);
1126 l2cap_pi(sk)->conf_state = 0;
1127 l2cap_sock_clear_timer(sk);
1129 if (!parent) {
1130 /* Outgoing channel.
1131 * Wake up socket sleeping on connect.
1133 sk->sk_state = BT_CONNECTED;
1134 sk->sk_state_change(sk);
1135 } else {
1136 /* Incoming channel.
1137 * Wake up socket sleeping on accept.
1139 parent->sk_data_ready(parent, 0);
1143 /* Copy frame to all raw sockets on that connection */
1144 static void l2cap_raw_recv(struct l2cap_conn *conn, struct sk_buff *skb)
1146 struct l2cap_chan_list *l = &conn->chan_list;
1147 struct sk_buff *nskb;
1148 struct sock * sk;
1150 BT_DBG("conn %p", conn);
1152 read_lock(&l->lock);
1153 for (sk = l->head; sk; sk = l2cap_pi(sk)->next_c) {
1154 if (sk->sk_type != SOCK_RAW)
1155 continue;
1157 /* Don't send frame to the socket it came from */
1158 if (skb->sk == sk)
1159 continue;
1161 if (!(nskb = skb_clone(skb, GFP_ATOMIC)))
1162 continue;
1164 if (sock_queue_rcv_skb(sk, nskb))
1165 kfree_skb(nskb);
1167 read_unlock(&l->lock);
1170 /* ---- L2CAP signalling commands ---- */
1171 static struct sk_buff *l2cap_build_cmd(struct l2cap_conn *conn,
1172 u8 code, u8 ident, u16 dlen, void *data)
1174 struct sk_buff *skb, **frag;
1175 struct l2cap_cmd_hdr *cmd;
1176 struct l2cap_hdr *lh;
1177 int len, count;
1179 BT_DBG("conn %p, code 0x%2.2x, ident 0x%2.2x, len %d", conn, code, ident, dlen);
1181 len = L2CAP_HDR_SIZE + L2CAP_CMD_HDR_SIZE + dlen;
1182 count = min_t(unsigned int, conn->mtu, len);
1184 skb = bt_skb_alloc(count, GFP_ATOMIC);
1185 if (!skb)
1186 return NULL;
1188 lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE);
1189 lh->len = __cpu_to_le16(L2CAP_CMD_HDR_SIZE + dlen);
1190 lh->cid = __cpu_to_le16(0x0001);
1192 cmd = (struct l2cap_cmd_hdr *) skb_put(skb, L2CAP_CMD_HDR_SIZE);
1193 cmd->code = code;
1194 cmd->ident = ident;
1195 cmd->len = __cpu_to_le16(dlen);
1197 if (dlen) {
1198 count -= L2CAP_HDR_SIZE + L2CAP_CMD_HDR_SIZE;
1199 memcpy(skb_put(skb, count), data, count);
1200 data += count;
1203 len -= skb->len;
1205 /* Continuation fragments (no L2CAP header) */
1206 frag = &skb_shinfo(skb)->frag_list;
1207 while (len) {
1208 count = min_t(unsigned int, conn->mtu, len);
1210 *frag = bt_skb_alloc(count, GFP_ATOMIC);
1211 if (!*frag)
1212 goto fail;
1214 memcpy(skb_put(*frag, count), data, count);
1216 len -= count;
1217 data += count;
1219 frag = &(*frag)->next;
1222 return skb;
1224 fail:
1225 kfree_skb(skb);
1226 return NULL;
1229 static inline int l2cap_get_conf_opt(void **ptr, int *type, int *olen, unsigned long *val)
1231 struct l2cap_conf_opt *opt = *ptr;
1232 int len;
1234 len = L2CAP_CONF_OPT_SIZE + opt->len;
1235 *ptr += len;
1237 *type = opt->type;
1238 *olen = opt->len;
1240 switch (opt->len) {
1241 case 1:
1242 *val = *((u8 *) opt->val);
1243 break;
1245 case 2:
1246 *val = __le16_to_cpu(*((u16 *)opt->val));
1247 break;
1249 case 4:
1250 *val = __le32_to_cpu(*((u32 *)opt->val));
1251 break;
1253 default:
1254 *val = (unsigned long) opt->val;
1255 break;
1258 BT_DBG("type 0x%2.2x len %d val 0x%lx", *type, opt->len, *val);
1259 return len;
1262 static inline void l2cap_parse_conf_req(struct sock *sk, void *data, int len)
1264 int type, hint, olen;
1265 unsigned long val;
1266 void *ptr = data;
1268 BT_DBG("sk %p len %d", sk, len);
1270 while (len >= L2CAP_CONF_OPT_SIZE) {
1271 len -= l2cap_get_conf_opt(&ptr, &type, &olen, &val);
1273 hint = type & 0x80;
1274 type &= 0x7f;
1276 switch (type) {
1277 case L2CAP_CONF_MTU:
1278 l2cap_pi(sk)->conf_mtu = val;
1279 break;
1281 case L2CAP_CONF_FLUSH_TO:
1282 l2cap_pi(sk)->flush_to = val;
1283 break;
1285 case L2CAP_CONF_QOS:
1286 break;
1288 default:
1289 if (hint)
1290 break;
1292 /* FIXME: Reject unknown option */
1293 break;
1298 static void l2cap_add_conf_opt(void **ptr, u8 type, u8 len, unsigned long val)
1300 struct l2cap_conf_opt *opt = *ptr;
1302 BT_DBG("type 0x%2.2x len %d val 0x%lx", type, len, val);
1304 opt->type = type;
1305 opt->len = len;
1307 switch (len) {
1308 case 1:
1309 *((u8 *) opt->val) = val;
1310 break;
1312 case 2:
1313 *((u16 *) opt->val) = __cpu_to_le16(val);
1314 break;
1316 case 4:
1317 *((u32 *) opt->val) = __cpu_to_le32(val);
1318 break;
1320 default:
1321 memcpy(opt->val, (void *) val, len);
1322 break;
1325 *ptr += L2CAP_CONF_OPT_SIZE + len;
1328 static int l2cap_build_conf_req(struct sock *sk, void *data)
1330 struct l2cap_pinfo *pi = l2cap_pi(sk);
1331 struct l2cap_conf_req *req = data;
1332 void *ptr = req->data;
1334 BT_DBG("sk %p", sk);
1336 if (pi->imtu != L2CAP_DEFAULT_MTU)
1337 l2cap_add_conf_opt(&ptr, L2CAP_CONF_MTU, 2, pi->imtu);
1339 /* FIXME: Need actual value of the flush timeout */
1340 //if (flush_to != L2CAP_DEFAULT_FLUSH_TO)
1341 // l2cap_add_conf_opt(&ptr, L2CAP_CONF_FLUSH_TO, 2, pi->flush_to);
1343 req->dcid = __cpu_to_le16(pi->dcid);
1344 req->flags = __cpu_to_le16(0);
1346 return ptr - data;
1349 static inline int l2cap_conf_output(struct sock *sk, void **ptr)
1351 struct l2cap_pinfo *pi = l2cap_pi(sk);
1352 int result = 0;
1354 /* Configure output options and let the other side know
1355 * which ones we don't like. */
1356 if (pi->conf_mtu < pi->omtu) {
1357 l2cap_add_conf_opt(ptr, L2CAP_CONF_MTU, 2, pi->omtu);
1358 result = L2CAP_CONF_UNACCEPT;
1359 } else {
1360 pi->omtu = pi->conf_mtu;
1363 BT_DBG("sk %p result %d", sk, result);
1364 return result;
1367 static int l2cap_build_conf_rsp(struct sock *sk, void *data, int *result)
1369 struct l2cap_conf_rsp *rsp = data;
1370 void *ptr = rsp->data;
1371 u16 flags = 0;
1373 BT_DBG("sk %p complete %d", sk, result ? 1 : 0);
1375 if (result)
1376 *result = l2cap_conf_output(sk, &ptr);
1377 else
1378 flags = 0x0001;
1380 rsp->scid = __cpu_to_le16(l2cap_pi(sk)->dcid);
1381 rsp->result = __cpu_to_le16(result ? *result : 0);
1382 rsp->flags = __cpu_to_le16(flags);
1384 return ptr - data;
1387 static inline int l2cap_connect_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1389 struct l2cap_chan_list *list = &conn->chan_list;
1390 struct l2cap_conn_req *req = (struct l2cap_conn_req *) data;
1391 struct l2cap_conn_rsp rsp;
1392 struct sock *sk, *parent;
1393 int result = 0, status = 0;
1395 u16 dcid = 0, scid = __le16_to_cpu(req->scid);
1396 u16 psm = req->psm;
1398 BT_DBG("psm 0x%2.2x scid 0x%4.4x", psm, scid);
1400 /* Check if we have socket listening on psm */
1401 parent = l2cap_get_sock_by_psm(BT_LISTEN, psm, conn->src);
1402 if (!parent) {
1403 result = L2CAP_CR_BAD_PSM;
1404 goto sendresp;
1407 result = L2CAP_CR_NO_MEM;
1409 /* Check for backlog size */
1410 if (sk_acceptq_is_full(parent)) {
1411 BT_DBG("backlog full %d", parent->sk_ack_backlog);
1412 goto response;
1415 sk = l2cap_sock_alloc(NULL, BTPROTO_L2CAP, GFP_ATOMIC);
1416 if (!sk)
1417 goto response;
1419 write_lock_bh(&list->lock);
1421 /* Check if we already have channel with that dcid */
1422 if (__l2cap_get_chan_by_dcid(list, scid)) {
1423 write_unlock_bh(&list->lock);
1424 sock_set_flag(sk, SOCK_ZAPPED);
1425 l2cap_sock_kill(sk);
1426 goto response;
1429 hci_conn_hold(conn->hcon);
1431 l2cap_sock_init(sk, parent);
1432 bacpy(&bt_sk(sk)->src, conn->src);
1433 bacpy(&bt_sk(sk)->dst, conn->dst);
1434 l2cap_pi(sk)->psm = psm;
1435 l2cap_pi(sk)->dcid = scid;
1437 __l2cap_chan_add(conn, sk, parent);
1438 dcid = l2cap_pi(sk)->scid;
1440 l2cap_sock_set_timer(sk, sk->sk_sndtimeo);
1442 /* Service level security */
1443 result = L2CAP_CR_PEND;
1444 status = L2CAP_CS_AUTHEN_PEND;
1445 sk->sk_state = BT_CONNECT2;
1446 l2cap_pi(sk)->ident = cmd->ident;
1448 if ((l2cap_pi(sk)->link_mode & L2CAP_LM_ENCRYPT) ||
1449 (l2cap_pi(sk)->link_mode & L2CAP_LM_SECURE)) {
1450 if (!hci_conn_encrypt(conn->hcon))
1451 goto done;
1452 } else if (l2cap_pi(sk)->link_mode & L2CAP_LM_AUTH) {
1453 if (!hci_conn_auth(conn->hcon))
1454 goto done;
1457 sk->sk_state = BT_CONFIG;
1458 result = status = 0;
1460 done:
1461 write_unlock_bh(&list->lock);
1463 response:
1464 bh_unlock_sock(parent);
1466 sendresp:
1467 rsp.scid = __cpu_to_le16(scid);
1468 rsp.dcid = __cpu_to_le16(dcid);
1469 rsp.result = __cpu_to_le16(result);
1470 rsp.status = __cpu_to_le16(status);
1471 l2cap_send_cmd(conn, cmd->ident, L2CAP_CONN_RSP, sizeof(rsp), &rsp);
1472 return 0;
1475 static inline int l2cap_connect_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1477 struct l2cap_conn_rsp *rsp = (struct l2cap_conn_rsp *) data;
1478 u16 scid, dcid, result, status;
1479 struct sock *sk;
1480 u8 req[128];
1482 scid = __le16_to_cpu(rsp->scid);
1483 dcid = __le16_to_cpu(rsp->dcid);
1484 result = __le16_to_cpu(rsp->result);
1485 status = __le16_to_cpu(rsp->status);
1487 BT_DBG("dcid 0x%4.4x scid 0x%4.4x result 0x%2.2x status 0x%2.2x", dcid, scid, result, status);
1489 if (scid) {
1490 if (!(sk = l2cap_get_chan_by_scid(&conn->chan_list, scid)))
1491 return 0;
1492 } else {
1493 if (!(sk = l2cap_get_chan_by_ident(&conn->chan_list, cmd->ident)))
1494 return 0;
1497 switch (result) {
1498 case L2CAP_CR_SUCCESS:
1499 sk->sk_state = BT_CONFIG;
1500 l2cap_pi(sk)->ident = 0;
1501 l2cap_pi(sk)->dcid = dcid;
1502 l2cap_pi(sk)->conf_state |= L2CAP_CONF_REQ_SENT;
1504 l2cap_send_cmd(conn, l2cap_get_ident(conn), L2CAP_CONF_REQ,
1505 l2cap_build_conf_req(sk, req), req);
1506 break;
1508 case L2CAP_CR_PEND:
1509 break;
1511 default:
1512 l2cap_chan_del(sk, ECONNREFUSED);
1513 break;
1516 bh_unlock_sock(sk);
1517 return 0;
1520 static inline int l2cap_config_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1522 struct l2cap_conf_req *req = (struct l2cap_conf_req *) data;
1523 u16 dcid, flags;
1524 u8 rsp[64];
1525 struct sock *sk;
1526 int result;
1528 dcid = __le16_to_cpu(req->dcid);
1529 flags = __le16_to_cpu(req->flags);
1531 BT_DBG("dcid 0x%4.4x flags 0x%2.2x", dcid, flags);
1533 if (!(sk = l2cap_get_chan_by_scid(&conn->chan_list, dcid)))
1534 return -ENOENT;
1536 l2cap_parse_conf_req(sk, req->data, cmd->len - sizeof(*req));
1538 if (flags & 0x0001) {
1539 /* Incomplete config. Send empty response. */
1540 l2cap_send_cmd(conn, cmd->ident, L2CAP_CONF_RSP,
1541 l2cap_build_conf_rsp(sk, rsp, NULL), rsp);
1542 goto unlock;
1545 /* Complete config. */
1546 l2cap_send_cmd(conn, cmd->ident, L2CAP_CONF_RSP,
1547 l2cap_build_conf_rsp(sk, rsp, &result), rsp);
1549 if (result)
1550 goto unlock;
1552 /* Output config done */
1553 l2cap_pi(sk)->conf_state |= L2CAP_CONF_OUTPUT_DONE;
1555 if (l2cap_pi(sk)->conf_state & L2CAP_CONF_INPUT_DONE) {
1556 sk->sk_state = BT_CONNECTED;
1557 l2cap_chan_ready(sk);
1558 } else if (!(l2cap_pi(sk)->conf_state & L2CAP_CONF_REQ_SENT)) {
1559 u8 req[64];
1560 l2cap_send_cmd(conn, l2cap_get_ident(conn), L2CAP_CONF_REQ,
1561 l2cap_build_conf_req(sk, req), req);
1564 unlock:
1565 bh_unlock_sock(sk);
1566 return 0;
1569 static inline int l2cap_config_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1571 struct l2cap_conf_rsp *rsp = (struct l2cap_conf_rsp *)data;
1572 u16 scid, flags, result;
1573 struct sock *sk;
1575 scid = __le16_to_cpu(rsp->scid);
1576 flags = __le16_to_cpu(rsp->flags);
1577 result = __le16_to_cpu(rsp->result);
1579 BT_DBG("scid 0x%4.4x flags 0x%2.2x result 0x%2.2x", scid, flags, result);
1581 if (!(sk = l2cap_get_chan_by_scid(&conn->chan_list, scid)))
1582 return 0;
1584 switch (result) {
1585 case L2CAP_CONF_SUCCESS:
1586 break;
1588 case L2CAP_CONF_UNACCEPT:
1589 if (++l2cap_pi(sk)->conf_retry < L2CAP_CONF_MAX_RETRIES) {
1590 char req[128];
1591 /* It does not make sense to adjust L2CAP parameters
1592 * that are currently defined in the spec. We simply
1593 * resend config request that we sent earlier. It is
1594 * stupid, but it helps qualification testing which
1595 * expects at least some response from us. */
1596 l2cap_send_cmd(conn, l2cap_get_ident(conn), L2CAP_CONF_REQ,
1597 l2cap_build_conf_req(sk, req), req);
1598 goto done;
1601 default:
1602 sk->sk_state = BT_DISCONN;
1603 sk->sk_err = ECONNRESET;
1604 l2cap_sock_set_timer(sk, HZ * 5);
1606 struct l2cap_disconn_req req;
1607 req.dcid = __cpu_to_le16(l2cap_pi(sk)->dcid);
1608 req.scid = __cpu_to_le16(l2cap_pi(sk)->scid);
1609 l2cap_send_cmd(conn, l2cap_get_ident(conn),
1610 L2CAP_DISCONN_REQ, sizeof(req), &req);
1612 goto done;
1615 if (flags & 0x01)
1616 goto done;
1618 /* Input config done */
1619 l2cap_pi(sk)->conf_state |= L2CAP_CONF_INPUT_DONE;
1621 if (l2cap_pi(sk)->conf_state & L2CAP_CONF_OUTPUT_DONE) {
1622 sk->sk_state = BT_CONNECTED;
1623 l2cap_chan_ready(sk);
1626 done:
1627 bh_unlock_sock(sk);
1628 return 0;
1631 static inline int l2cap_disconnect_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1633 struct l2cap_disconn_req *req = (struct l2cap_disconn_req *) data;
1634 struct l2cap_disconn_rsp rsp;
1635 u16 dcid, scid;
1636 struct sock *sk;
1638 scid = __le16_to_cpu(req->scid);
1639 dcid = __le16_to_cpu(req->dcid);
1641 BT_DBG("scid 0x%4.4x dcid 0x%4.4x", scid, dcid);
1643 if (!(sk = l2cap_get_chan_by_scid(&conn->chan_list, dcid)))
1644 return 0;
1646 rsp.dcid = __cpu_to_le16(l2cap_pi(sk)->scid);
1647 rsp.scid = __cpu_to_le16(l2cap_pi(sk)->dcid);
1648 l2cap_send_cmd(conn, cmd->ident, L2CAP_DISCONN_RSP, sizeof(rsp), &rsp);
1650 sk->sk_shutdown = SHUTDOWN_MASK;
1652 l2cap_chan_del(sk, ECONNRESET);
1653 bh_unlock_sock(sk);
1655 l2cap_sock_kill(sk);
1656 return 0;
1659 static inline int l2cap_disconnect_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1661 struct l2cap_disconn_rsp *rsp = (struct l2cap_disconn_rsp *) data;
1662 u16 dcid, scid;
1663 struct sock *sk;
1665 scid = __le16_to_cpu(rsp->scid);
1666 dcid = __le16_to_cpu(rsp->dcid);
1668 BT_DBG("dcid 0x%4.4x scid 0x%4.4x", dcid, scid);
1670 if (!(sk = l2cap_get_chan_by_scid(&conn->chan_list, scid)))
1671 return 0;
1673 l2cap_chan_del(sk, 0);
1674 bh_unlock_sock(sk);
1676 l2cap_sock_kill(sk);
1677 return 0;
1680 static inline int l2cap_information_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1682 struct l2cap_info_req *req = (struct l2cap_info_req *) data;
1683 struct l2cap_info_rsp rsp;
1684 u16 type;
1686 type = __le16_to_cpu(req->type);
1688 BT_DBG("type 0x%4.4x", type);
1690 rsp.type = __cpu_to_le16(type);
1691 rsp.result = __cpu_to_le16(L2CAP_IR_NOTSUPP);
1692 l2cap_send_cmd(conn, cmd->ident, L2CAP_INFO_RSP, sizeof(rsp), &rsp);
1694 return 0;
1697 static inline int l2cap_information_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data)
1699 struct l2cap_info_rsp *rsp = (struct l2cap_info_rsp *) data;
1700 u16 type, result;
1702 type = __le16_to_cpu(rsp->type);
1703 result = __le16_to_cpu(rsp->result);
1705 BT_DBG("type 0x%4.4x result 0x%2.2x", type, result);
1707 return 0;
1710 static inline void l2cap_sig_channel(struct l2cap_conn *conn, struct sk_buff *skb)
1712 u8 *data = skb->data;
1713 int len = skb->len;
1714 struct l2cap_cmd_hdr cmd;
1715 int err = 0;
1717 l2cap_raw_recv(conn, skb);
1719 while (len >= L2CAP_CMD_HDR_SIZE) {
1720 memcpy(&cmd, data, L2CAP_CMD_HDR_SIZE);
1721 data += L2CAP_CMD_HDR_SIZE;
1722 len -= L2CAP_CMD_HDR_SIZE;
1724 cmd.len = __le16_to_cpu(cmd.len);
1726 BT_DBG("code 0x%2.2x len %d id 0x%2.2x", cmd.code, cmd.len, cmd.ident);
1728 if (cmd.len > len || !cmd.ident) {
1729 BT_DBG("corrupted command");
1730 break;
1733 switch (cmd.code) {
1734 case L2CAP_COMMAND_REJ:
1735 /* FIXME: We should process this */
1736 break;
1738 case L2CAP_CONN_REQ:
1739 err = l2cap_connect_req(conn, &cmd, data);
1740 break;
1742 case L2CAP_CONN_RSP:
1743 err = l2cap_connect_rsp(conn, &cmd, data);
1744 break;
1746 case L2CAP_CONF_REQ:
1747 err = l2cap_config_req(conn, &cmd, data);
1748 break;
1750 case L2CAP_CONF_RSP:
1751 err = l2cap_config_rsp(conn, &cmd, data);
1752 break;
1754 case L2CAP_DISCONN_REQ:
1755 err = l2cap_disconnect_req(conn, &cmd, data);
1756 break;
1758 case L2CAP_DISCONN_RSP:
1759 err = l2cap_disconnect_rsp(conn, &cmd, data);
1760 break;
1762 case L2CAP_ECHO_REQ:
1763 l2cap_send_cmd(conn, cmd.ident, L2CAP_ECHO_RSP, cmd.len, data);
1764 break;
1766 case L2CAP_ECHO_RSP:
1767 break;
1769 case L2CAP_INFO_REQ:
1770 err = l2cap_information_req(conn, &cmd, data);
1771 break;
1773 case L2CAP_INFO_RSP:
1774 err = l2cap_information_rsp(conn, &cmd, data);
1775 break;
1777 default:
1778 BT_ERR("Unknown signaling command 0x%2.2x", cmd.code);
1779 err = -EINVAL;
1780 break;
1783 if (err) {
1784 struct l2cap_cmd_rej rej;
1785 BT_DBG("error %d", err);
1787 /* FIXME: Map err to a valid reason */
1788 rej.reason = __cpu_to_le16(0);
1789 l2cap_send_cmd(conn, cmd.ident, L2CAP_COMMAND_REJ, sizeof(rej), &rej);
1792 data += cmd.len;
1793 len -= cmd.len;
1796 kfree_skb(skb);
1799 static inline int l2cap_data_channel(struct l2cap_conn *conn, u16 cid, struct sk_buff *skb)
1801 struct sock *sk;
1803 sk = l2cap_get_chan_by_scid(&conn->chan_list, cid);
1804 if (!sk) {
1805 BT_DBG("unknown cid 0x%4.4x", cid);
1806 goto drop;
1809 BT_DBG("sk %p, len %d", sk, skb->len);
1811 if (sk->sk_state != BT_CONNECTED)
1812 goto drop;
1814 if (l2cap_pi(sk)->imtu < skb->len)
1815 goto drop;
1817 /* If socket recv buffers overflows we drop data here
1818 * which is *bad* because L2CAP has to be reliable.
1819 * But we don't have any other choice. L2CAP doesn't
1820 * provide flow control mechanism. */
1822 if (!sock_queue_rcv_skb(sk, skb))
1823 goto done;
1825 drop:
1826 kfree_skb(skb);
1828 done:
1829 if (sk)
1830 bh_unlock_sock(sk);
1832 return 0;
1835 static inline int l2cap_conless_channel(struct l2cap_conn *conn, u16 psm, struct sk_buff *skb)
1837 struct sock *sk;
1839 sk = l2cap_get_sock_by_psm(0, psm, conn->src);
1840 if (!sk)
1841 goto drop;
1843 BT_DBG("sk %p, len %d", sk, skb->len);
1845 if (sk->sk_state != BT_BOUND && sk->sk_state != BT_CONNECTED)
1846 goto drop;
1848 if (l2cap_pi(sk)->imtu < skb->len)
1849 goto drop;
1851 if (!sock_queue_rcv_skb(sk, skb))
1852 goto done;
1854 drop:
1855 kfree_skb(skb);
1857 done:
1858 if (sk) bh_unlock_sock(sk);
1859 return 0;
1862 static void l2cap_recv_frame(struct l2cap_conn *conn, struct sk_buff *skb)
1864 struct l2cap_hdr *lh = (void *) skb->data;
1865 u16 cid, psm, len;
1867 skb_pull(skb, L2CAP_HDR_SIZE);
1868 cid = __le16_to_cpu(lh->cid);
1869 len = __le16_to_cpu(lh->len);
1871 BT_DBG("len %d, cid 0x%4.4x", len, cid);
1873 switch (cid) {
1874 case 0x0001:
1875 l2cap_sig_channel(conn, skb);
1876 break;
1878 case 0x0002:
1879 psm = get_unaligned((u16 *) skb->data);
1880 skb_pull(skb, 2);
1881 l2cap_conless_channel(conn, psm, skb);
1882 break;
1884 default:
1885 l2cap_data_channel(conn, cid, skb);
1886 break;
1890 /* ---- L2CAP interface with lower layer (HCI) ---- */
1892 static int l2cap_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type)
1894 int exact = 0, lm1 = 0, lm2 = 0;
1895 register struct sock *sk;
1896 struct hlist_node *node;
1898 if (type != ACL_LINK)
1899 return 0;
1901 BT_DBG("hdev %s, bdaddr %s", hdev->name, batostr(bdaddr));
1903 /* Find listening sockets and check their link_mode */
1904 read_lock(&l2cap_sk_list.lock);
1905 sk_for_each(sk, node, &l2cap_sk_list.head) {
1906 if (sk->sk_state != BT_LISTEN)
1907 continue;
1909 if (!bacmp(&bt_sk(sk)->src, &hdev->bdaddr)) {
1910 lm1 |= (HCI_LM_ACCEPT | l2cap_pi(sk)->link_mode);
1911 exact++;
1912 } else if (!bacmp(&bt_sk(sk)->src, BDADDR_ANY))
1913 lm2 |= (HCI_LM_ACCEPT | l2cap_pi(sk)->link_mode);
1915 read_unlock(&l2cap_sk_list.lock);
1917 return exact ? lm1 : lm2;
1920 static int l2cap_connect_cfm(struct hci_conn *hcon, u8 status)
1922 struct l2cap_conn *conn;
1924 BT_DBG("hcon %p bdaddr %s status %d", hcon, batostr(&hcon->dst), status);
1926 if (hcon->type != ACL_LINK)
1927 return 0;
1929 if (!status) {
1930 conn = l2cap_conn_add(hcon, status);
1931 if (conn)
1932 l2cap_conn_ready(conn);
1933 } else
1934 l2cap_conn_del(hcon, bt_err(status));
1936 return 0;
1939 static int l2cap_disconn_ind(struct hci_conn *hcon, u8 reason)
1941 BT_DBG("hcon %p reason %d", hcon, reason);
1943 if (hcon->type != ACL_LINK)
1944 return 0;
1946 l2cap_conn_del(hcon, bt_err(reason));
1948 return 0;
1951 static int l2cap_auth_cfm(struct hci_conn *hcon, u8 status)
1953 struct l2cap_chan_list *l;
1954 struct l2cap_conn *conn = conn = hcon->l2cap_data;
1955 struct l2cap_conn_rsp rsp;
1956 struct sock *sk;
1957 int result;
1959 if (!conn)
1960 return 0;
1962 l = &conn->chan_list;
1964 BT_DBG("conn %p", conn);
1966 read_lock(&l->lock);
1968 for (sk = l->head; sk; sk = l2cap_pi(sk)->next_c) {
1969 bh_lock_sock(sk);
1971 if (sk->sk_state != BT_CONNECT2 ||
1972 (l2cap_pi(sk)->link_mode & L2CAP_LM_ENCRYPT) ||
1973 (l2cap_pi(sk)->link_mode & L2CAP_LM_SECURE)) {
1974 bh_unlock_sock(sk);
1975 continue;
1978 if (!status) {
1979 sk->sk_state = BT_CONFIG;
1980 result = 0;
1981 } else {
1982 sk->sk_state = BT_DISCONN;
1983 l2cap_sock_set_timer(sk, HZ/10);
1984 result = L2CAP_CR_SEC_BLOCK;
1987 rsp.scid = __cpu_to_le16(l2cap_pi(sk)->dcid);
1988 rsp.dcid = __cpu_to_le16(l2cap_pi(sk)->scid);
1989 rsp.result = __cpu_to_le16(result);
1990 rsp.status = __cpu_to_le16(0);
1991 l2cap_send_cmd(conn, l2cap_pi(sk)->ident,
1992 L2CAP_CONN_RSP, sizeof(rsp), &rsp);
1994 bh_unlock_sock(sk);
1997 read_unlock(&l->lock);
1998 return 0;
2001 static int l2cap_encrypt_cfm(struct hci_conn *hcon, u8 status)
2003 struct l2cap_chan_list *l;
2004 struct l2cap_conn *conn = hcon->l2cap_data;
2005 struct l2cap_conn_rsp rsp;
2006 struct sock *sk;
2007 int result;
2009 if (!conn)
2010 return 0;
2012 l = &conn->chan_list;
2014 BT_DBG("conn %p", conn);
2016 read_lock(&l->lock);
2018 for (sk = l->head; sk; sk = l2cap_pi(sk)->next_c) {
2019 bh_lock_sock(sk);
2021 if (sk->sk_state != BT_CONNECT2) {
2022 bh_unlock_sock(sk);
2023 continue;
2026 if (!status) {
2027 sk->sk_state = BT_CONFIG;
2028 result = 0;
2029 } else {
2030 sk->sk_state = BT_DISCONN;
2031 l2cap_sock_set_timer(sk, HZ/10);
2032 result = L2CAP_CR_SEC_BLOCK;
2035 rsp.scid = __cpu_to_le16(l2cap_pi(sk)->dcid);
2036 rsp.dcid = __cpu_to_le16(l2cap_pi(sk)->scid);
2037 rsp.result = __cpu_to_le16(result);
2038 rsp.status = __cpu_to_le16(0);
2039 l2cap_send_cmd(conn, l2cap_pi(sk)->ident,
2040 L2CAP_CONN_RSP, sizeof(rsp), &rsp);
2042 if (l2cap_pi(sk)->link_mode & L2CAP_LM_SECURE)
2043 hci_conn_change_link_key(hcon);
2045 bh_unlock_sock(sk);
2048 read_unlock(&l->lock);
2049 return 0;
2052 static int l2cap_recv_acldata(struct hci_conn *hcon, struct sk_buff *skb, u16 flags)
2054 struct l2cap_conn *conn = hcon->l2cap_data;
2056 if (!conn && !(conn = l2cap_conn_add(hcon, 0)))
2057 goto drop;
2059 BT_DBG("conn %p len %d flags 0x%x", conn, skb->len, flags);
2061 if (flags & ACL_START) {
2062 struct l2cap_hdr *hdr;
2063 int len;
2065 if (conn->rx_len) {
2066 BT_ERR("Unexpected start frame (len %d)", skb->len);
2067 kfree_skb(conn->rx_skb);
2068 conn->rx_skb = NULL;
2069 conn->rx_len = 0;
2070 l2cap_conn_unreliable(conn, ECOMM);
2073 if (skb->len < 2) {
2074 BT_ERR("Frame is too short (len %d)", skb->len);
2075 l2cap_conn_unreliable(conn, ECOMM);
2076 goto drop;
2079 hdr = (struct l2cap_hdr *) skb->data;
2080 len = __le16_to_cpu(hdr->len) + L2CAP_HDR_SIZE;
2082 if (len == skb->len) {
2083 /* Complete frame received */
2084 l2cap_recv_frame(conn, skb);
2085 return 0;
2088 BT_DBG("Start: total len %d, frag len %d", len, skb->len);
2090 if (skb->len > len) {
2091 BT_ERR("Frame is too long (len %d, expected len %d)",
2092 skb->len, len);
2093 l2cap_conn_unreliable(conn, ECOMM);
2094 goto drop;
2097 /* Allocate skb for the complete frame (with header) */
2098 if (!(conn->rx_skb = bt_skb_alloc(len, GFP_ATOMIC)))
2099 goto drop;
2101 memcpy(skb_put(conn->rx_skb, skb->len), skb->data, skb->len);
2102 conn->rx_len = len - skb->len;
2103 } else {
2104 BT_DBG("Cont: frag len %d (expecting %d)", skb->len, conn->rx_len);
2106 if (!conn->rx_len) {
2107 BT_ERR("Unexpected continuation frame (len %d)", skb->len);
2108 l2cap_conn_unreliable(conn, ECOMM);
2109 goto drop;
2112 if (skb->len > conn->rx_len) {
2113 BT_ERR("Fragment is too long (len %d, expected %d)",
2114 skb->len, conn->rx_len);
2115 kfree_skb(conn->rx_skb);
2116 conn->rx_skb = NULL;
2117 conn->rx_len = 0;
2118 l2cap_conn_unreliable(conn, ECOMM);
2119 goto drop;
2122 memcpy(skb_put(conn->rx_skb, skb->len), skb->data, skb->len);
2123 conn->rx_len -= skb->len;
2125 if (!conn->rx_len) {
2126 /* Complete frame received */
2127 l2cap_recv_frame(conn, conn->rx_skb);
2128 conn->rx_skb = NULL;
2132 drop:
2133 kfree_skb(skb);
2134 return 0;
2137 static ssize_t l2cap_sysfs_show(struct class *dev, char *buf)
2139 struct sock *sk;
2140 struct hlist_node *node;
2141 char *str = buf;
2143 read_lock_bh(&l2cap_sk_list.lock);
2145 sk_for_each(sk, node, &l2cap_sk_list.head) {
2146 struct l2cap_pinfo *pi = l2cap_pi(sk);
2148 str += sprintf(str, "%s %s %d %d 0x%4.4x 0x%4.4x %d %d 0x%x\n",
2149 batostr(&bt_sk(sk)->src), batostr(&bt_sk(sk)->dst),
2150 sk->sk_state, pi->psm, pi->scid, pi->dcid, pi->imtu,
2151 pi->omtu, pi->link_mode);
2154 read_unlock_bh(&l2cap_sk_list.lock);
2156 return (str - buf);
2159 static CLASS_ATTR(l2cap, S_IRUGO, l2cap_sysfs_show, NULL);
2161 static const struct proto_ops l2cap_sock_ops = {
2162 .family = PF_BLUETOOTH,
2163 .owner = THIS_MODULE,
2164 .release = l2cap_sock_release,
2165 .bind = l2cap_sock_bind,
2166 .connect = l2cap_sock_connect,
2167 .listen = l2cap_sock_listen,
2168 .accept = l2cap_sock_accept,
2169 .getname = l2cap_sock_getname,
2170 .sendmsg = l2cap_sock_sendmsg,
2171 .recvmsg = bt_sock_recvmsg,
2172 .poll = bt_sock_poll,
2173 .mmap = sock_no_mmap,
2174 .socketpair = sock_no_socketpair,
2175 .ioctl = sock_no_ioctl,
2176 .shutdown = l2cap_sock_shutdown,
2177 .setsockopt = l2cap_sock_setsockopt,
2178 .getsockopt = l2cap_sock_getsockopt
2181 static struct net_proto_family l2cap_sock_family_ops = {
2182 .family = PF_BLUETOOTH,
2183 .owner = THIS_MODULE,
2184 .create = l2cap_sock_create,
2187 static struct hci_proto l2cap_hci_proto = {
2188 .name = "L2CAP",
2189 .id = HCI_PROTO_L2CAP,
2190 .connect_ind = l2cap_connect_ind,
2191 .connect_cfm = l2cap_connect_cfm,
2192 .disconn_ind = l2cap_disconn_ind,
2193 .auth_cfm = l2cap_auth_cfm,
2194 .encrypt_cfm = l2cap_encrypt_cfm,
2195 .recv_acldata = l2cap_recv_acldata
2198 static int __init l2cap_init(void)
2200 int err;
2202 err = proto_register(&l2cap_proto, 0);
2203 if (err < 0)
2204 return err;
2206 err = bt_sock_register(BTPROTO_L2CAP, &l2cap_sock_family_ops);
2207 if (err < 0) {
2208 BT_ERR("L2CAP socket registration failed");
2209 goto error;
2212 err = hci_register_proto(&l2cap_hci_proto);
2213 if (err < 0) {
2214 BT_ERR("L2CAP protocol registration failed");
2215 bt_sock_unregister(BTPROTO_L2CAP);
2216 goto error;
2219 class_create_file(bt_class, &class_attr_l2cap);
2221 BT_INFO("L2CAP ver %s", VERSION);
2222 BT_INFO("L2CAP socket layer initialized");
2224 return 0;
2226 error:
2227 proto_unregister(&l2cap_proto);
2228 return err;
2231 static void __exit l2cap_exit(void)
2233 class_remove_file(bt_class, &class_attr_l2cap);
2235 if (bt_sock_unregister(BTPROTO_L2CAP) < 0)
2236 BT_ERR("L2CAP socket unregistration failed");
2238 if (hci_unregister_proto(&l2cap_hci_proto) < 0)
2239 BT_ERR("L2CAP protocol unregistration failed");
2241 proto_unregister(&l2cap_proto);
2244 void l2cap_load(void)
2246 /* Dummy function to trigger automatic L2CAP module loading by
2247 * other modules that use L2CAP sockets but don't use any other
2248 * symbols from it. */
2249 return;
2251 EXPORT_SYMBOL(l2cap_load);
2253 module_init(l2cap_init);
2254 module_exit(l2cap_exit);
2256 MODULE_AUTHOR("Maxim Krasnyansky <maxk@qualcomm.com>, Marcel Holtmann <marcel@holtmann.org>");
2257 MODULE_DESCRIPTION("Bluetooth L2CAP ver " VERSION);
2258 MODULE_VERSION(VERSION);
2259 MODULE_LICENSE("GPL");
2260 MODULE_ALIAS("bt-proto-0");