1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /* SCTP kernel implementation
3 * (C) Copyright IBM Corp. 2001, 2004
4 * Copyright (c) 1999-2000 Cisco, Inc.
5 * Copyright (c) 1999-2001 Motorola, Inc.
6 * Copyright (c) 2001 Intel Corp.
7 * Copyright (c) 2001 Nokia, Inc.
8 * Copyright (c) 2001 La Monte H.P. Yarroll
10 * This file is part of the SCTP kernel implementation
12 * Initialization/cleanup for SCTP protocol support.
14 * Please send any bug reports or fixes you make to the
16 * lksctp developers <linux-sctp@vger.kernel.org>
18 * Written or modified by:
19 * La Monte H.P. Yarroll <piggy@acm.org>
20 * Karl Knutson <karl@athena.chicago.il.us>
21 * Jon Grimm <jgrimm@us.ibm.com>
22 * Sridhar Samudrala <sri@us.ibm.com>
23 * Daisy Chang <daisyc@us.ibm.com>
24 * Ardelle Fan <ardelle.fan@intel.com>
27 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
29 #include <linux/module.h>
30 #include <linux/init.h>
31 #include <linux/netdevice.h>
32 #include <linux/inetdevice.h>
33 #include <linux/seq_file.h>
34 #include <linux/memblock.h>
35 #include <linux/highmem.h>
36 #include <linux/swap.h>
37 #include <linux/slab.h>
38 #include <net/net_namespace.h>
39 #include <net/protocol.h>
42 #include <net/route.h>
43 #include <net/sctp/sctp.h>
44 #include <net/addrconf.h>
45 #include <net/inet_common.h>
46 #include <net/inet_ecn.h>
48 #define MAX_SCTP_PORT_HASH_ENTRIES (64 * 1024)
50 /* Global data structures. */
51 struct sctp_globals sctp_globals __read_mostly
;
53 struct idr sctp_assocs_id
;
54 DEFINE_SPINLOCK(sctp_assocs_id_lock
);
56 static struct sctp_pf
*sctp_pf_inet6_specific
;
57 static struct sctp_pf
*sctp_pf_inet_specific
;
58 static struct sctp_af
*sctp_af_v4_specific
;
59 static struct sctp_af
*sctp_af_v6_specific
;
61 struct kmem_cache
*sctp_chunk_cachep __read_mostly
;
62 struct kmem_cache
*sctp_bucket_cachep __read_mostly
;
64 long sysctl_sctp_mem
[3];
65 int sysctl_sctp_rmem
[3];
66 int sysctl_sctp_wmem
[3];
68 /* Private helper to extract ipv4 address and stash them in
69 * the protocol structure.
71 static void sctp_v4_copy_addrlist(struct list_head
*addrlist
,
72 struct net_device
*dev
)
74 struct in_device
*in_dev
;
75 struct in_ifaddr
*ifa
;
76 struct sctp_sockaddr_entry
*addr
;
79 if ((in_dev
= __in_dev_get_rcu(dev
)) == NULL
) {
84 in_dev_for_each_ifa_rcu(ifa
, in_dev
) {
85 /* Add the address to the local list. */
86 addr
= kzalloc(sizeof(*addr
), GFP_ATOMIC
);
88 addr
->a
.v4
.sin_family
= AF_INET
;
89 addr
->a
.v4
.sin_addr
.s_addr
= ifa
->ifa_local
;
91 INIT_LIST_HEAD(&addr
->list
);
92 list_add_tail(&addr
->list
, addrlist
);
99 /* Extract our IP addresses from the system and stash them in the
100 * protocol structure.
102 static void sctp_get_local_addr_list(struct net
*net
)
104 struct net_device
*dev
;
105 struct list_head
*pos
;
109 for_each_netdev_rcu(net
, dev
) {
110 list_for_each(pos
, &sctp_address_families
) {
111 af
= list_entry(pos
, struct sctp_af
, list
);
112 af
->copy_addrlist(&net
->sctp
.local_addr_list
, dev
);
118 /* Free the existing local addresses. */
119 static void sctp_free_local_addr_list(struct net
*net
)
121 struct sctp_sockaddr_entry
*addr
;
122 struct list_head
*pos
, *temp
;
124 list_for_each_safe(pos
, temp
, &net
->sctp
.local_addr_list
) {
125 addr
= list_entry(pos
, struct sctp_sockaddr_entry
, list
);
131 /* Copy the local addresses which are valid for 'scope' into 'bp'. */
132 int sctp_copy_local_addr_list(struct net
*net
, struct sctp_bind_addr
*bp
,
133 enum sctp_scope scope
, gfp_t gfp
, int copy_flags
)
135 struct sctp_sockaddr_entry
*addr
;
136 union sctp_addr laddr
;
140 list_for_each_entry_rcu(addr
, &net
->sctp
.local_addr_list
, list
) {
143 if (!sctp_in_scope(net
, &addr
->a
, scope
))
146 /* Now that the address is in scope, check to see if
147 * the address type is really supported by the local
148 * sock as well as the remote peer.
150 if (addr
->a
.sa
.sa_family
== AF_INET
&&
151 (!(copy_flags
& SCTP_ADDR4_ALLOWED
) ||
152 !(copy_flags
& SCTP_ADDR4_PEERSUPP
)))
154 if (addr
->a
.sa
.sa_family
== AF_INET6
&&
155 (!(copy_flags
& SCTP_ADDR6_ALLOWED
) ||
156 !(copy_flags
& SCTP_ADDR6_PEERSUPP
)))
160 /* also works for setting ipv6 address port */
161 laddr
.v4
.sin_port
= htons(bp
->port
);
162 if (sctp_bind_addr_state(bp
, &laddr
) != -1)
165 error
= sctp_add_bind_addr(bp
, &addr
->a
, sizeof(addr
->a
),
166 SCTP_ADDR_SRC
, GFP_ATOMIC
);
175 /* Copy over any ip options */
176 static void sctp_v4_copy_ip_options(struct sock
*sk
, struct sock
*newsk
)
178 struct inet_sock
*newinet
, *inet
= inet_sk(sk
);
179 struct ip_options_rcu
*inet_opt
, *newopt
= NULL
;
181 newinet
= inet_sk(newsk
);
184 inet_opt
= rcu_dereference(inet
->inet_opt
);
186 newopt
= sock_kmalloc(newsk
, sizeof(*inet_opt
) +
187 inet_opt
->opt
.optlen
, GFP_ATOMIC
);
189 memcpy(newopt
, inet_opt
, sizeof(*inet_opt
) +
190 inet_opt
->opt
.optlen
);
192 pr_err("%s: Failed to copy ip options\n", __func__
);
194 RCU_INIT_POINTER(newinet
->inet_opt
, newopt
);
198 /* Account for the IP options */
199 static int sctp_v4_ip_options_len(struct sock
*sk
)
201 struct inet_sock
*inet
= inet_sk(sk
);
202 struct ip_options_rcu
*inet_opt
;
206 inet_opt
= rcu_dereference(inet
->inet_opt
);
208 len
= inet_opt
->opt
.optlen
;
214 /* Initialize a sctp_addr from in incoming skb. */
215 static void sctp_v4_from_skb(union sctp_addr
*addr
, struct sk_buff
*skb
,
218 /* Always called on head skb, so this is safe */
219 struct sctphdr
*sh
= sctp_hdr(skb
);
220 struct sockaddr_in
*sa
= &addr
->v4
;
222 addr
->v4
.sin_family
= AF_INET
;
225 sa
->sin_port
= sh
->source
;
226 sa
->sin_addr
.s_addr
= ip_hdr(skb
)->saddr
;
228 sa
->sin_port
= sh
->dest
;
229 sa
->sin_addr
.s_addr
= ip_hdr(skb
)->daddr
;
231 memset(sa
->sin_zero
, 0, sizeof(sa
->sin_zero
));
234 /* Initialize an sctp_addr from a socket. */
235 static void sctp_v4_from_sk(union sctp_addr
*addr
, struct sock
*sk
)
237 addr
->v4
.sin_family
= AF_INET
;
238 addr
->v4
.sin_port
= 0;
239 addr
->v4
.sin_addr
.s_addr
= inet_sk(sk
)->inet_rcv_saddr
;
240 memset(addr
->v4
.sin_zero
, 0, sizeof(addr
->v4
.sin_zero
));
243 /* Initialize sk->sk_rcv_saddr from sctp_addr. */
244 static void sctp_v4_to_sk_saddr(union sctp_addr
*addr
, struct sock
*sk
)
246 inet_sk(sk
)->inet_rcv_saddr
= addr
->v4
.sin_addr
.s_addr
;
249 /* Initialize sk->sk_daddr from sctp_addr. */
250 static void sctp_v4_to_sk_daddr(union sctp_addr
*addr
, struct sock
*sk
)
252 inet_sk(sk
)->inet_daddr
= addr
->v4
.sin_addr
.s_addr
;
255 /* Initialize a sctp_addr from an address parameter. */
256 static void sctp_v4_from_addr_param(union sctp_addr
*addr
,
257 union sctp_addr_param
*param
,
258 __be16 port
, int iif
)
260 addr
->v4
.sin_family
= AF_INET
;
261 addr
->v4
.sin_port
= port
;
262 addr
->v4
.sin_addr
.s_addr
= param
->v4
.addr
.s_addr
;
263 memset(addr
->v4
.sin_zero
, 0, sizeof(addr
->v4
.sin_zero
));
266 /* Initialize an address parameter from a sctp_addr and return the length
267 * of the address parameter.
269 static int sctp_v4_to_addr_param(const union sctp_addr
*addr
,
270 union sctp_addr_param
*param
)
272 int length
= sizeof(struct sctp_ipv4addr_param
);
274 param
->v4
.param_hdr
.type
= SCTP_PARAM_IPV4_ADDRESS
;
275 param
->v4
.param_hdr
.length
= htons(length
);
276 param
->v4
.addr
.s_addr
= addr
->v4
.sin_addr
.s_addr
;
281 /* Initialize a sctp_addr from a dst_entry. */
282 static void sctp_v4_dst_saddr(union sctp_addr
*saddr
, struct flowi4
*fl4
,
285 saddr
->v4
.sin_family
= AF_INET
;
286 saddr
->v4
.sin_port
= port
;
287 saddr
->v4
.sin_addr
.s_addr
= fl4
->saddr
;
288 memset(saddr
->v4
.sin_zero
, 0, sizeof(saddr
->v4
.sin_zero
));
291 /* Compare two addresses exactly. */
292 static int sctp_v4_cmp_addr(const union sctp_addr
*addr1
,
293 const union sctp_addr
*addr2
)
295 if (addr1
->sa
.sa_family
!= addr2
->sa
.sa_family
)
297 if (addr1
->v4
.sin_port
!= addr2
->v4
.sin_port
)
299 if (addr1
->v4
.sin_addr
.s_addr
!= addr2
->v4
.sin_addr
.s_addr
)
305 /* Initialize addr struct to INADDR_ANY. */
306 static void sctp_v4_inaddr_any(union sctp_addr
*addr
, __be16 port
)
308 addr
->v4
.sin_family
= AF_INET
;
309 addr
->v4
.sin_addr
.s_addr
= htonl(INADDR_ANY
);
310 addr
->v4
.sin_port
= port
;
311 memset(addr
->v4
.sin_zero
, 0, sizeof(addr
->v4
.sin_zero
));
314 /* Is this a wildcard address? */
315 static int sctp_v4_is_any(const union sctp_addr
*addr
)
317 return htonl(INADDR_ANY
) == addr
->v4
.sin_addr
.s_addr
;
320 /* This function checks if the address is a valid address to be used for
324 * Return 0 - If the address is a non-unicast or an illegal address.
325 * Return 1 - If the address is a unicast.
327 static int sctp_v4_addr_valid(union sctp_addr
*addr
,
328 struct sctp_sock
*sp
,
329 const struct sk_buff
*skb
)
331 /* IPv4 addresses not allowed */
332 if (sp
&& ipv6_only_sock(sctp_opt2sk(sp
)))
335 /* Is this a non-unicast address or a unusable SCTP address? */
336 if (IS_IPV4_UNUSABLE_ADDRESS(addr
->v4
.sin_addr
.s_addr
))
339 /* Is this a broadcast address? */
340 if (skb
&& skb_rtable(skb
)->rt_flags
& RTCF_BROADCAST
)
346 /* Should this be available for binding? */
347 static int sctp_v4_available(union sctp_addr
*addr
, struct sctp_sock
*sp
)
349 struct net
*net
= sock_net(&sp
->inet
.sk
);
350 int ret
= inet_addr_type(net
, addr
->v4
.sin_addr
.s_addr
);
353 if (addr
->v4
.sin_addr
.s_addr
!= htonl(INADDR_ANY
) &&
355 !sp
->inet
.freebind
&&
356 !net
->ipv4
.sysctl_ip_nonlocal_bind
)
359 if (ipv6_only_sock(sctp_opt2sk(sp
)))
365 /* Checking the loopback, private and other address scopes as defined in
366 * RFC 1918. The IPv4 scoping is based on the draft for SCTP IPv4
367 * scoping <draft-stewart-tsvwg-sctp-ipv4-00.txt>.
369 * Level 0 - unusable SCTP addresses
370 * Level 1 - loopback address
371 * Level 2 - link-local addresses
372 * Level 3 - private addresses.
373 * Level 4 - global addresses
374 * For INIT and INIT-ACK address list, let L be the level of
375 * of requested destination address, sender and receiver
376 * SHOULD include all of its addresses with level greater
377 * than or equal to L.
379 * IPv4 scoping can be controlled through sysctl option
380 * net.sctp.addr_scope_policy
382 static enum sctp_scope
sctp_v4_scope(union sctp_addr
*addr
)
384 enum sctp_scope retval
;
386 /* Check for unusable SCTP addresses. */
387 if (IS_IPV4_UNUSABLE_ADDRESS(addr
->v4
.sin_addr
.s_addr
)) {
388 retval
= SCTP_SCOPE_UNUSABLE
;
389 } else if (ipv4_is_loopback(addr
->v4
.sin_addr
.s_addr
)) {
390 retval
= SCTP_SCOPE_LOOPBACK
;
391 } else if (ipv4_is_linklocal_169(addr
->v4
.sin_addr
.s_addr
)) {
392 retval
= SCTP_SCOPE_LINK
;
393 } else if (ipv4_is_private_10(addr
->v4
.sin_addr
.s_addr
) ||
394 ipv4_is_private_172(addr
->v4
.sin_addr
.s_addr
) ||
395 ipv4_is_private_192(addr
->v4
.sin_addr
.s_addr
)) {
396 retval
= SCTP_SCOPE_PRIVATE
;
398 retval
= SCTP_SCOPE_GLOBAL
;
404 /* Returns a valid dst cache entry for the given source and destination ip
405 * addresses. If an association is passed, trys to get a dst entry with a
406 * source address that matches an address in the bind address list.
408 static void sctp_v4_get_dst(struct sctp_transport
*t
, union sctp_addr
*saddr
,
409 struct flowi
*fl
, struct sock
*sk
)
411 struct sctp_association
*asoc
= t
->asoc
;
414 struct flowi4
*fl4
= &_fl
.u
.ip4
;
415 struct sctp_bind_addr
*bp
;
416 struct sctp_sockaddr_entry
*laddr
;
417 struct dst_entry
*dst
= NULL
;
418 union sctp_addr
*daddr
= &t
->ipaddr
;
419 union sctp_addr dst_saddr
;
420 __u8 tos
= inet_sk(sk
)->tos
;
422 if (t
->dscp
& SCTP_DSCP_SET_MASK
)
423 tos
= t
->dscp
& SCTP_DSCP_VAL_MASK
;
424 memset(&_fl
, 0x0, sizeof(_fl
));
425 fl4
->daddr
= daddr
->v4
.sin_addr
.s_addr
;
426 fl4
->fl4_dport
= daddr
->v4
.sin_port
;
427 fl4
->flowi4_proto
= IPPROTO_SCTP
;
429 fl4
->flowi4_tos
= RT_CONN_FLAGS_TOS(asoc
->base
.sk
, tos
);
430 fl4
->flowi4_oif
= asoc
->base
.sk
->sk_bound_dev_if
;
431 fl4
->fl4_sport
= htons(asoc
->base
.bind_addr
.port
);
434 fl4
->saddr
= saddr
->v4
.sin_addr
.s_addr
;
436 fl4
->fl4_sport
= saddr
->v4
.sin_port
;
439 pr_debug("%s: dst:%pI4, src:%pI4 - ", __func__
, &fl4
->daddr
,
442 rt
= ip_route_output_key(sock_net(sk
), fl4
);
446 memcpy(fl
, &_fl
, sizeof(_fl
));
449 /* If there is no association or if a source address is passed, no
450 * more validation is required.
455 bp
= &asoc
->base
.bind_addr
;
458 /* Walk through the bind address list and look for a bind
459 * address that matches the source address of the returned dst.
461 sctp_v4_dst_saddr(&dst_saddr
, fl4
, htons(bp
->port
));
463 list_for_each_entry_rcu(laddr
, &bp
->address_list
, list
) {
464 if (!laddr
->valid
|| (laddr
->state
== SCTP_ADDR_DEL
) ||
465 (laddr
->state
!= SCTP_ADDR_SRC
&&
466 !asoc
->src_out_of_asoc_ok
))
468 if (sctp_v4_cmp_addr(&dst_saddr
, &laddr
->a
))
473 /* None of the bound addresses match the source address of the
474 * dst. So release it.
480 /* Walk through the bind address list and try to get a dst that
481 * matches a bind address as the source address.
484 list_for_each_entry_rcu(laddr
, &bp
->address_list
, list
) {
485 struct net_device
*odev
;
489 if (laddr
->state
!= SCTP_ADDR_SRC
||
490 AF_INET
!= laddr
->a
.sa
.sa_family
)
493 fl4
->fl4_sport
= laddr
->a
.v4
.sin_port
;
494 flowi4_update_output(fl4
,
495 asoc
->base
.sk
->sk_bound_dev_if
,
496 RT_CONN_FLAGS_TOS(asoc
->base
.sk
, tos
),
497 daddr
->v4
.sin_addr
.s_addr
,
498 laddr
->a
.v4
.sin_addr
.s_addr
);
500 rt
= ip_route_output_key(sock_net(sk
), fl4
);
504 /* Ensure the src address belongs to the output
507 odev
= __ip_dev_find(sock_net(sk
), laddr
->a
.v4
.sin_addr
.s_addr
,
509 if (!odev
|| odev
->ifindex
!= fl4
->flowi4_oif
) {
513 memcpy(fl
, &_fl
, sizeof(_fl
));
515 dst_release(&rt
->dst
);
523 memcpy(fl
, &_fl
, sizeof(_fl
));
531 pr_debug("rt_dst:%pI4, rt_src:%pI4\n",
532 &fl
->u
.ip4
.daddr
, &fl
->u
.ip4
.saddr
);
535 pr_debug("no route\n");
539 /* For v4, the source address is cached in the route entry(dst). So no need
540 * to cache it separately and hence this is an empty routine.
542 static void sctp_v4_get_saddr(struct sctp_sock
*sk
,
543 struct sctp_transport
*t
,
546 union sctp_addr
*saddr
= &t
->saddr
;
547 struct rtable
*rt
= (struct rtable
*)t
->dst
;
550 saddr
->v4
.sin_family
= AF_INET
;
551 saddr
->v4
.sin_addr
.s_addr
= fl
->u
.ip4
.saddr
;
555 /* What interface did this skb arrive on? */
556 static int sctp_v4_skb_iif(const struct sk_buff
*skb
)
558 return inet_iif(skb
);
561 /* Was this packet marked by Explicit Congestion Notification? */
562 static int sctp_v4_is_ce(const struct sk_buff
*skb
)
564 return INET_ECN_is_ce(ip_hdr(skb
)->tos
);
567 /* Create and initialize a new sk for the socket returned by accept(). */
568 static struct sock
*sctp_v4_create_accept_sk(struct sock
*sk
,
569 struct sctp_association
*asoc
,
572 struct sock
*newsk
= sk_alloc(sock_net(sk
), PF_INET
, GFP_KERNEL
,
574 struct inet_sock
*newinet
;
579 sock_init_data(NULL
, newsk
);
581 sctp_copy_sock(newsk
, sk
, asoc
);
582 sock_reset_flag(newsk
, SOCK_ZAPPED
);
584 sctp_v4_copy_ip_options(sk
, newsk
);
586 newinet
= inet_sk(newsk
);
588 newinet
->inet_daddr
= asoc
->peer
.primary_addr
.v4
.sin_addr
.s_addr
;
590 sk_refcnt_debug_inc(newsk
);
592 if (newsk
->sk_prot
->init(newsk
)) {
593 sk_common_release(newsk
);
601 static int sctp_v4_addr_to_user(struct sctp_sock
*sp
, union sctp_addr
*addr
)
603 /* No address mapping for V4 sockets */
604 memset(addr
->v4
.sin_zero
, 0, sizeof(addr
->v4
.sin_zero
));
605 return sizeof(struct sockaddr_in
);
608 /* Dump the v4 addr to the seq file. */
609 static void sctp_v4_seq_dump_addr(struct seq_file
*seq
, union sctp_addr
*addr
)
611 seq_printf(seq
, "%pI4 ", &addr
->v4
.sin_addr
);
614 static void sctp_v4_ecn_capable(struct sock
*sk
)
619 static void sctp_addr_wq_timeout_handler(struct timer_list
*t
)
621 struct net
*net
= from_timer(net
, t
, sctp
.addr_wq_timer
);
622 struct sctp_sockaddr_entry
*addrw
, *temp
;
623 struct sctp_sock
*sp
;
625 spin_lock_bh(&net
->sctp
.addr_wq_lock
);
627 list_for_each_entry_safe(addrw
, temp
, &net
->sctp
.addr_waitq
, list
) {
628 pr_debug("%s: the first ent in wq:%p is addr:%pISc for cmd:%d at "
629 "entry:%p\n", __func__
, &net
->sctp
.addr_waitq
, &addrw
->a
.sa
,
630 addrw
->state
, addrw
);
632 #if IS_ENABLED(CONFIG_IPV6)
633 /* Now we send an ASCONF for each association */
634 /* Note. we currently don't handle link local IPv6 addressees */
635 if (addrw
->a
.sa
.sa_family
== AF_INET6
) {
636 struct in6_addr
*in6
;
638 if (ipv6_addr_type(&addrw
->a
.v6
.sin6_addr
) &
642 in6
= (struct in6_addr
*)&addrw
->a
.v6
.sin6_addr
;
643 if (ipv6_chk_addr(net
, in6
, NULL
, 0) == 0 &&
644 addrw
->state
== SCTP_ADDR_NEW
) {
645 unsigned long timeo_val
;
647 pr_debug("%s: this is on DAD, trying %d sec "
649 SCTP_ADDRESS_TICK_DELAY
);
652 timeo_val
+= msecs_to_jiffies(SCTP_ADDRESS_TICK_DELAY
);
653 mod_timer(&net
->sctp
.addr_wq_timer
, timeo_val
);
658 list_for_each_entry(sp
, &net
->sctp
.auto_asconf_splist
, auto_asconf_list
) {
661 sk
= sctp_opt2sk(sp
);
662 /* ignore bound-specific endpoints */
663 if (!sctp_is_ep_boundall(sk
))
666 if (sctp_asconf_mgmt(sp
, addrw
) < 0)
667 pr_debug("%s: sctp_asconf_mgmt failed\n", __func__
);
670 #if IS_ENABLED(CONFIG_IPV6)
673 list_del(&addrw
->list
);
676 spin_unlock_bh(&net
->sctp
.addr_wq_lock
);
679 static void sctp_free_addr_wq(struct net
*net
)
681 struct sctp_sockaddr_entry
*addrw
;
682 struct sctp_sockaddr_entry
*temp
;
684 spin_lock_bh(&net
->sctp
.addr_wq_lock
);
685 del_timer(&net
->sctp
.addr_wq_timer
);
686 list_for_each_entry_safe(addrw
, temp
, &net
->sctp
.addr_waitq
, list
) {
687 list_del(&addrw
->list
);
690 spin_unlock_bh(&net
->sctp
.addr_wq_lock
);
693 /* lookup the entry for the same address in the addr_waitq
694 * sctp_addr_wq MUST be locked
696 static struct sctp_sockaddr_entry
*sctp_addr_wq_lookup(struct net
*net
,
697 struct sctp_sockaddr_entry
*addr
)
699 struct sctp_sockaddr_entry
*addrw
;
701 list_for_each_entry(addrw
, &net
->sctp
.addr_waitq
, list
) {
702 if (addrw
->a
.sa
.sa_family
!= addr
->a
.sa
.sa_family
)
704 if (addrw
->a
.sa
.sa_family
== AF_INET
) {
705 if (addrw
->a
.v4
.sin_addr
.s_addr
==
706 addr
->a
.v4
.sin_addr
.s_addr
)
708 } else if (addrw
->a
.sa
.sa_family
== AF_INET6
) {
709 if (ipv6_addr_equal(&addrw
->a
.v6
.sin6_addr
,
710 &addr
->a
.v6
.sin6_addr
))
717 void sctp_addr_wq_mgmt(struct net
*net
, struct sctp_sockaddr_entry
*addr
, int cmd
)
719 struct sctp_sockaddr_entry
*addrw
;
720 unsigned long timeo_val
;
722 /* first, we check if an opposite message already exist in the queue.
723 * If we found such message, it is removed.
724 * This operation is a bit stupid, but the DHCP client attaches the
725 * new address after a couple of addition and deletion of that address
728 spin_lock_bh(&net
->sctp
.addr_wq_lock
);
729 /* Offsets existing events in addr_wq */
730 addrw
= sctp_addr_wq_lookup(net
, addr
);
732 if (addrw
->state
!= cmd
) {
733 pr_debug("%s: offsets existing entry for %d, addr:%pISc "
734 "in wq:%p\n", __func__
, addrw
->state
, &addrw
->a
.sa
,
735 &net
->sctp
.addr_waitq
);
737 list_del(&addrw
->list
);
740 spin_unlock_bh(&net
->sctp
.addr_wq_lock
);
744 /* OK, we have to add the new address to the wait queue */
745 addrw
= kmemdup(addr
, sizeof(struct sctp_sockaddr_entry
), GFP_ATOMIC
);
747 spin_unlock_bh(&net
->sctp
.addr_wq_lock
);
751 list_add_tail(&addrw
->list
, &net
->sctp
.addr_waitq
);
753 pr_debug("%s: add new entry for cmd:%d, addr:%pISc in wq:%p\n",
754 __func__
, addrw
->state
, &addrw
->a
.sa
, &net
->sctp
.addr_waitq
);
756 if (!timer_pending(&net
->sctp
.addr_wq_timer
)) {
758 timeo_val
+= msecs_to_jiffies(SCTP_ADDRESS_TICK_DELAY
);
759 mod_timer(&net
->sctp
.addr_wq_timer
, timeo_val
);
761 spin_unlock_bh(&net
->sctp
.addr_wq_lock
);
764 /* Event handler for inet address addition/deletion events.
765 * The sctp_local_addr_list needs to be protocted by a spin lock since
766 * multiple notifiers (say IPv4 and IPv6) may be running at the same
767 * time and thus corrupt the list.
768 * The reader side is protected with RCU.
770 static int sctp_inetaddr_event(struct notifier_block
*this, unsigned long ev
,
773 struct in_ifaddr
*ifa
= (struct in_ifaddr
*)ptr
;
774 struct sctp_sockaddr_entry
*addr
= NULL
;
775 struct sctp_sockaddr_entry
*temp
;
776 struct net
*net
= dev_net(ifa
->ifa_dev
->dev
);
781 addr
= kzalloc(sizeof(*addr
), GFP_ATOMIC
);
783 addr
->a
.v4
.sin_family
= AF_INET
;
784 addr
->a
.v4
.sin_addr
.s_addr
= ifa
->ifa_local
;
786 spin_lock_bh(&net
->sctp
.local_addr_lock
);
787 list_add_tail_rcu(&addr
->list
, &net
->sctp
.local_addr_list
);
788 sctp_addr_wq_mgmt(net
, addr
, SCTP_ADDR_NEW
);
789 spin_unlock_bh(&net
->sctp
.local_addr_lock
);
793 spin_lock_bh(&net
->sctp
.local_addr_lock
);
794 list_for_each_entry_safe(addr
, temp
,
795 &net
->sctp
.local_addr_list
, list
) {
796 if (addr
->a
.sa
.sa_family
== AF_INET
&&
797 addr
->a
.v4
.sin_addr
.s_addr
==
799 sctp_addr_wq_mgmt(net
, addr
, SCTP_ADDR_DEL
);
802 list_del_rcu(&addr
->list
);
806 spin_unlock_bh(&net
->sctp
.local_addr_lock
);
808 kfree_rcu(addr
, rcu
);
816 * Initialize the control inode/socket with a control endpoint data
817 * structure. This endpoint is reserved exclusively for the OOTB processing.
819 static int sctp_ctl_sock_init(struct net
*net
)
822 sa_family_t family
= PF_INET
;
824 if (sctp_get_pf_specific(PF_INET6
))
827 err
= inet_ctl_sock_create(&net
->sctp
.ctl_sock
, family
,
828 SOCK_SEQPACKET
, IPPROTO_SCTP
, net
);
830 /* If IPv6 socket could not be created, try the IPv4 socket */
831 if (err
< 0 && family
== PF_INET6
)
832 err
= inet_ctl_sock_create(&net
->sctp
.ctl_sock
, AF_INET
,
833 SOCK_SEQPACKET
, IPPROTO_SCTP
,
837 pr_err("Failed to create the SCTP control socket\n");
843 /* Register address family specific functions. */
844 int sctp_register_af(struct sctp_af
*af
)
846 switch (af
->sa_family
) {
848 if (sctp_af_v4_specific
)
850 sctp_af_v4_specific
= af
;
853 if (sctp_af_v6_specific
)
855 sctp_af_v6_specific
= af
;
861 INIT_LIST_HEAD(&af
->list
);
862 list_add_tail(&af
->list
, &sctp_address_families
);
866 /* Get the table of functions for manipulating a particular address
869 struct sctp_af
*sctp_get_af_specific(sa_family_t family
)
873 return sctp_af_v4_specific
;
875 return sctp_af_v6_specific
;
881 /* Common code to initialize a AF_INET msg_name. */
882 static void sctp_inet_msgname(char *msgname
, int *addr_len
)
884 struct sockaddr_in
*sin
;
886 sin
= (struct sockaddr_in
*)msgname
;
887 *addr_len
= sizeof(struct sockaddr_in
);
888 sin
->sin_family
= AF_INET
;
889 memset(sin
->sin_zero
, 0, sizeof(sin
->sin_zero
));
892 /* Copy the primary address of the peer primary address as the msg_name. */
893 static void sctp_inet_event_msgname(struct sctp_ulpevent
*event
, char *msgname
,
896 struct sockaddr_in
*sin
, *sinfrom
;
899 struct sctp_association
*asoc
;
902 sctp_inet_msgname(msgname
, addr_len
);
903 sin
= (struct sockaddr_in
*)msgname
;
904 sinfrom
= &asoc
->peer
.primary_addr
.v4
;
905 sin
->sin_port
= htons(asoc
->peer
.port
);
906 sin
->sin_addr
.s_addr
= sinfrom
->sin_addr
.s_addr
;
910 /* Initialize and copy out a msgname from an inbound skb. */
911 static void sctp_inet_skb_msgname(struct sk_buff
*skb
, char *msgname
, int *len
)
914 struct sctphdr
*sh
= sctp_hdr(skb
);
915 struct sockaddr_in
*sin
= (struct sockaddr_in
*)msgname
;
917 sctp_inet_msgname(msgname
, len
);
918 sin
->sin_port
= sh
->source
;
919 sin
->sin_addr
.s_addr
= ip_hdr(skb
)->saddr
;
923 /* Do we support this AF? */
924 static int sctp_inet_af_supported(sa_family_t family
, struct sctp_sock
*sp
)
926 /* PF_INET only supports AF_INET addresses. */
927 return AF_INET
== family
;
930 /* Address matching with wildcards allowed. */
931 static int sctp_inet_cmp_addr(const union sctp_addr
*addr1
,
932 const union sctp_addr
*addr2
,
933 struct sctp_sock
*opt
)
935 /* PF_INET only supports AF_INET addresses. */
936 if (addr1
->sa
.sa_family
!= addr2
->sa
.sa_family
)
938 if (htonl(INADDR_ANY
) == addr1
->v4
.sin_addr
.s_addr
||
939 htonl(INADDR_ANY
) == addr2
->v4
.sin_addr
.s_addr
)
941 if (addr1
->v4
.sin_addr
.s_addr
== addr2
->v4
.sin_addr
.s_addr
)
947 /* Verify that provided sockaddr looks bindable. Common verification has
948 * already been taken care of.
950 static int sctp_inet_bind_verify(struct sctp_sock
*opt
, union sctp_addr
*addr
)
952 return sctp_v4_available(addr
, opt
);
955 /* Verify that sockaddr looks sendable. Common verification has already
956 * been taken care of.
958 static int sctp_inet_send_verify(struct sctp_sock
*opt
, union sctp_addr
*addr
)
963 /* Fill in Supported Address Type information for INIT and INIT-ACK
964 * chunks. Returns number of addresses supported.
966 static int sctp_inet_supported_addrs(const struct sctp_sock
*opt
,
969 types
[0] = SCTP_PARAM_IPV4_ADDRESS
;
973 /* Wrapper routine that calls the ip transmit routine. */
974 static inline int sctp_v4_xmit(struct sk_buff
*skb
,
975 struct sctp_transport
*transport
)
977 struct inet_sock
*inet
= inet_sk(skb
->sk
);
978 __u8 dscp
= inet
->tos
;
980 pr_debug("%s: skb:%p, len:%d, src:%pI4, dst:%pI4\n", __func__
, skb
,
981 skb
->len
, &transport
->fl
.u
.ip4
.saddr
,
982 &transport
->fl
.u
.ip4
.daddr
);
984 if (transport
->dscp
& SCTP_DSCP_SET_MASK
)
985 dscp
= transport
->dscp
& SCTP_DSCP_VAL_MASK
;
987 inet
->pmtudisc
= transport
->param_flags
& SPP_PMTUD_ENABLE
?
988 IP_PMTUDISC_DO
: IP_PMTUDISC_DONT
;
990 SCTP_INC_STATS(sock_net(&inet
->sk
), SCTP_MIB_OUTSCTPPACKS
);
992 return __ip_queue_xmit(&inet
->sk
, skb
, &transport
->fl
, dscp
);
995 static struct sctp_af sctp_af_inet
;
997 static struct sctp_pf sctp_pf_inet
= {
998 .event_msgname
= sctp_inet_event_msgname
,
999 .skb_msgname
= sctp_inet_skb_msgname
,
1000 .af_supported
= sctp_inet_af_supported
,
1001 .cmp_addr
= sctp_inet_cmp_addr
,
1002 .bind_verify
= sctp_inet_bind_verify
,
1003 .send_verify
= sctp_inet_send_verify
,
1004 .supported_addrs
= sctp_inet_supported_addrs
,
1005 .create_accept_sk
= sctp_v4_create_accept_sk
,
1006 .addr_to_user
= sctp_v4_addr_to_user
,
1007 .to_sk_saddr
= sctp_v4_to_sk_saddr
,
1008 .to_sk_daddr
= sctp_v4_to_sk_daddr
,
1009 .copy_ip_options
= sctp_v4_copy_ip_options
,
1013 /* Notifier for inetaddr addition/deletion events. */
1014 static struct notifier_block sctp_inetaddr_notifier
= {
1015 .notifier_call
= sctp_inetaddr_event
,
1018 /* Socket operations. */
1019 static const struct proto_ops inet_seqpacket_ops
= {
1021 .owner
= THIS_MODULE
,
1022 .release
= inet_release
, /* Needs to be wrapped... */
1024 .connect
= sctp_inet_connect
,
1025 .socketpair
= sock_no_socketpair
,
1026 .accept
= inet_accept
,
1027 .getname
= inet_getname
, /* Semantics are different. */
1029 .ioctl
= inet_ioctl
,
1030 .gettstamp
= sock_gettstamp
,
1031 .listen
= sctp_inet_listen
,
1032 .shutdown
= inet_shutdown
, /* Looks harmless. */
1033 .setsockopt
= sock_common_setsockopt
, /* IP_SOL IP_OPTION is a problem */
1034 .getsockopt
= sock_common_getsockopt
,
1035 .sendmsg
= inet_sendmsg
,
1036 .recvmsg
= inet_recvmsg
,
1037 .mmap
= sock_no_mmap
,
1038 .sendpage
= sock_no_sendpage
,
1039 #ifdef CONFIG_COMPAT
1040 .compat_setsockopt
= compat_sock_common_setsockopt
,
1041 .compat_getsockopt
= compat_sock_common_getsockopt
,
1045 /* Registration with AF_INET family. */
1046 static struct inet_protosw sctp_seqpacket_protosw
= {
1047 .type
= SOCK_SEQPACKET
,
1048 .protocol
= IPPROTO_SCTP
,
1050 .ops
= &inet_seqpacket_ops
,
1051 .flags
= SCTP_PROTOSW_FLAG
1053 static struct inet_protosw sctp_stream_protosw
= {
1054 .type
= SOCK_STREAM
,
1055 .protocol
= IPPROTO_SCTP
,
1057 .ops
= &inet_seqpacket_ops
,
1058 .flags
= SCTP_PROTOSW_FLAG
1061 /* Register with IP layer. */
1062 static const struct net_protocol sctp_protocol
= {
1063 .handler
= sctp_rcv
,
1064 .err_handler
= sctp_v4_err
,
1067 .icmp_strict_tag_validation
= 1,
1070 /* IPv4 address related functions. */
1071 static struct sctp_af sctp_af_inet
= {
1072 .sa_family
= AF_INET
,
1073 .sctp_xmit
= sctp_v4_xmit
,
1074 .setsockopt
= ip_setsockopt
,
1075 .getsockopt
= ip_getsockopt
,
1076 .get_dst
= sctp_v4_get_dst
,
1077 .get_saddr
= sctp_v4_get_saddr
,
1078 .copy_addrlist
= sctp_v4_copy_addrlist
,
1079 .from_skb
= sctp_v4_from_skb
,
1080 .from_sk
= sctp_v4_from_sk
,
1081 .from_addr_param
= sctp_v4_from_addr_param
,
1082 .to_addr_param
= sctp_v4_to_addr_param
,
1083 .cmp_addr
= sctp_v4_cmp_addr
,
1084 .addr_valid
= sctp_v4_addr_valid
,
1085 .inaddr_any
= sctp_v4_inaddr_any
,
1086 .is_any
= sctp_v4_is_any
,
1087 .available
= sctp_v4_available
,
1088 .scope
= sctp_v4_scope
,
1089 .skb_iif
= sctp_v4_skb_iif
,
1090 .is_ce
= sctp_v4_is_ce
,
1091 .seq_dump_addr
= sctp_v4_seq_dump_addr
,
1092 .ecn_capable
= sctp_v4_ecn_capable
,
1093 .net_header_len
= sizeof(struct iphdr
),
1094 .sockaddr_len
= sizeof(struct sockaddr_in
),
1095 .ip_options_len
= sctp_v4_ip_options_len
,
1096 #ifdef CONFIG_COMPAT
1097 .compat_setsockopt
= compat_ip_setsockopt
,
1098 .compat_getsockopt
= compat_ip_getsockopt
,
1102 struct sctp_pf
*sctp_get_pf_specific(sa_family_t family
)
1106 return sctp_pf_inet_specific
;
1108 return sctp_pf_inet6_specific
;
1114 /* Register the PF specific function table. */
1115 int sctp_register_pf(struct sctp_pf
*pf
, sa_family_t family
)
1119 if (sctp_pf_inet_specific
)
1121 sctp_pf_inet_specific
= pf
;
1124 if (sctp_pf_inet6_specific
)
1126 sctp_pf_inet6_specific
= pf
;
1134 static inline int init_sctp_mibs(struct net
*net
)
1136 net
->sctp
.sctp_statistics
= alloc_percpu(struct sctp_mib
);
1137 if (!net
->sctp
.sctp_statistics
)
1142 static inline void cleanup_sctp_mibs(struct net
*net
)
1144 free_percpu(net
->sctp
.sctp_statistics
);
1147 static void sctp_v4_pf_init(void)
1149 /* Initialize the SCTP specific PF functions. */
1150 sctp_register_pf(&sctp_pf_inet
, PF_INET
);
1151 sctp_register_af(&sctp_af_inet
);
1154 static void sctp_v4_pf_exit(void)
1156 list_del(&sctp_af_inet
.list
);
1159 static int sctp_v4_protosw_init(void)
1163 rc
= proto_register(&sctp_prot
, 1);
1167 /* Register SCTP(UDP and TCP style) with socket layer. */
1168 inet_register_protosw(&sctp_seqpacket_protosw
);
1169 inet_register_protosw(&sctp_stream_protosw
);
1174 static void sctp_v4_protosw_exit(void)
1176 inet_unregister_protosw(&sctp_stream_protosw
);
1177 inet_unregister_protosw(&sctp_seqpacket_protosw
);
1178 proto_unregister(&sctp_prot
);
1181 static int sctp_v4_add_protocol(void)
1183 /* Register notifier for inet address additions/deletions. */
1184 register_inetaddr_notifier(&sctp_inetaddr_notifier
);
1186 /* Register SCTP with inet layer. */
1187 if (inet_add_protocol(&sctp_protocol
, IPPROTO_SCTP
) < 0)
1193 static void sctp_v4_del_protocol(void)
1195 inet_del_protocol(&sctp_protocol
, IPPROTO_SCTP
);
1196 unregister_inetaddr_notifier(&sctp_inetaddr_notifier
);
1199 static int __net_init
sctp_defaults_init(struct net
*net
)
1204 * 14. Suggested SCTP Protocol Parameter Values
1206 /* The following protocol parameters are RECOMMENDED: */
1207 /* RTO.Initial - 3 seconds */
1208 net
->sctp
.rto_initial
= SCTP_RTO_INITIAL
;
1209 /* RTO.Min - 1 second */
1210 net
->sctp
.rto_min
= SCTP_RTO_MIN
;
1211 /* RTO.Max - 60 seconds */
1212 net
->sctp
.rto_max
= SCTP_RTO_MAX
;
1213 /* RTO.Alpha - 1/8 */
1214 net
->sctp
.rto_alpha
= SCTP_RTO_ALPHA
;
1215 /* RTO.Beta - 1/4 */
1216 net
->sctp
.rto_beta
= SCTP_RTO_BETA
;
1218 /* Valid.Cookie.Life - 60 seconds */
1219 net
->sctp
.valid_cookie_life
= SCTP_DEFAULT_COOKIE_LIFE
;
1221 /* Whether Cookie Preservative is enabled(1) or not(0) */
1222 net
->sctp
.cookie_preserve_enable
= 1;
1224 /* Default sctp sockets to use md5 as their hmac alg */
1225 #if defined (CONFIG_SCTP_DEFAULT_COOKIE_HMAC_MD5)
1226 net
->sctp
.sctp_hmac_alg
= "md5";
1227 #elif defined (CONFIG_SCTP_DEFAULT_COOKIE_HMAC_SHA1)
1228 net
->sctp
.sctp_hmac_alg
= "sha1";
1230 net
->sctp
.sctp_hmac_alg
= NULL
;
1234 net
->sctp
.max_burst
= SCTP_DEFAULT_MAX_BURST
;
1236 /* Disable of Primary Path Switchover by default */
1237 net
->sctp
.ps_retrans
= SCTP_PS_RETRANS_MAX
;
1239 /* Enable pf state by default */
1240 net
->sctp
.pf_enable
= 1;
1242 /* Ignore pf exposure feature by default */
1243 net
->sctp
.pf_expose
= SCTP_PF_EXPOSE_UNSET
;
1245 /* Association.Max.Retrans - 10 attempts
1246 * Path.Max.Retrans - 5 attempts (per destination address)
1247 * Max.Init.Retransmits - 8 attempts
1249 net
->sctp
.max_retrans_association
= 10;
1250 net
->sctp
.max_retrans_path
= 5;
1251 net
->sctp
.max_retrans_init
= 8;
1253 /* Sendbuffer growth - do per-socket accounting */
1254 net
->sctp
.sndbuf_policy
= 0;
1256 /* Rcvbuffer growth - do per-socket accounting */
1257 net
->sctp
.rcvbuf_policy
= 0;
1259 /* HB.interval - 30 seconds */
1260 net
->sctp
.hb_interval
= SCTP_DEFAULT_TIMEOUT_HEARTBEAT
;
1262 /* delayed SACK timeout */
1263 net
->sctp
.sack_timeout
= SCTP_DEFAULT_TIMEOUT_SACK
;
1265 /* Disable ADDIP by default. */
1266 net
->sctp
.addip_enable
= 0;
1267 net
->sctp
.addip_noauth
= 0;
1268 net
->sctp
.default_auto_asconf
= 0;
1270 /* Enable PR-SCTP by default. */
1271 net
->sctp
.prsctp_enable
= 1;
1273 /* Disable RECONF by default. */
1274 net
->sctp
.reconf_enable
= 0;
1276 /* Disable AUTH by default. */
1277 net
->sctp
.auth_enable
= 0;
1279 /* Enable ECN by default. */
1280 net
->sctp
.ecn_enable
= 1;
1282 /* Set SCOPE policy to enabled */
1283 net
->sctp
.scope_policy
= SCTP_SCOPE_POLICY_ENABLE
;
1285 /* Set the default rwnd update threshold */
1286 net
->sctp
.rwnd_upd_shift
= SCTP_DEFAULT_RWND_SHIFT
;
1288 /* Initialize maximum autoclose timeout. */
1289 net
->sctp
.max_autoclose
= INT_MAX
/ HZ
;
1291 status
= sctp_sysctl_net_register(net
);
1293 goto err_sysctl_register
;
1295 /* Allocate and initialise sctp mibs. */
1296 status
= init_sctp_mibs(net
);
1300 #ifdef CONFIG_PROC_FS
1301 /* Initialize proc fs directory. */
1302 status
= sctp_proc_init(net
);
1307 sctp_dbg_objcnt_init(net
);
1309 /* Initialize the local address list. */
1310 INIT_LIST_HEAD(&net
->sctp
.local_addr_list
);
1311 spin_lock_init(&net
->sctp
.local_addr_lock
);
1312 sctp_get_local_addr_list(net
);
1314 /* Initialize the address event list */
1315 INIT_LIST_HEAD(&net
->sctp
.addr_waitq
);
1316 INIT_LIST_HEAD(&net
->sctp
.auto_asconf_splist
);
1317 spin_lock_init(&net
->sctp
.addr_wq_lock
);
1318 net
->sctp
.addr_wq_timer
.expires
= 0;
1319 timer_setup(&net
->sctp
.addr_wq_timer
, sctp_addr_wq_timeout_handler
, 0);
1323 #ifdef CONFIG_PROC_FS
1325 cleanup_sctp_mibs(net
);
1328 sctp_sysctl_net_unregister(net
);
1329 err_sysctl_register
:
1333 static void __net_exit
sctp_defaults_exit(struct net
*net
)
1335 /* Free the local address list */
1336 sctp_free_addr_wq(net
);
1337 sctp_free_local_addr_list(net
);
1339 #ifdef CONFIG_PROC_FS
1340 remove_proc_subtree("sctp", net
->proc_net
);
1341 net
->sctp
.proc_net_sctp
= NULL
;
1343 cleanup_sctp_mibs(net
);
1344 sctp_sysctl_net_unregister(net
);
1347 static struct pernet_operations sctp_defaults_ops
= {
1348 .init
= sctp_defaults_init
,
1349 .exit
= sctp_defaults_exit
,
1352 static int __net_init
sctp_ctrlsock_init(struct net
*net
)
1356 /* Initialize the control inode/socket for handling OOTB packets. */
1357 status
= sctp_ctl_sock_init(net
);
1359 pr_err("Failed to initialize the SCTP control sock\n");
1364 static void __net_exit
sctp_ctrlsock_exit(struct net
*net
)
1366 /* Free the control endpoint. */
1367 inet_ctl_sock_destroy(net
->sctp
.ctl_sock
);
1370 static struct pernet_operations sctp_ctrlsock_ops
= {
1371 .init
= sctp_ctrlsock_init
,
1372 .exit
= sctp_ctrlsock_exit
,
1375 /* Initialize the universe into something sensible. */
1376 static __init
int sctp_init(void)
1379 int status
= -EINVAL
;
1381 unsigned long limit
;
1382 unsigned long nr_pages
= totalram_pages();
1386 int max_entry_order
;
1388 sock_skb_cb_check_size(sizeof(struct sctp_ulpevent
));
1390 /* Allocate bind_bucket and chunk caches. */
1392 sctp_bucket_cachep
= kmem_cache_create("sctp_bind_bucket",
1393 sizeof(struct sctp_bind_bucket
),
1394 0, SLAB_HWCACHE_ALIGN
,
1396 if (!sctp_bucket_cachep
)
1399 sctp_chunk_cachep
= kmem_cache_create("sctp_chunk",
1400 sizeof(struct sctp_chunk
),
1401 0, SLAB_HWCACHE_ALIGN
,
1403 if (!sctp_chunk_cachep
)
1404 goto err_chunk_cachep
;
1406 status
= percpu_counter_init(&sctp_sockets_allocated
, 0, GFP_KERNEL
);
1408 goto err_percpu_counter_init
;
1410 /* Implementation specific variables. */
1412 /* Initialize default stream count setup information. */
1413 sctp_max_instreams
= SCTP_DEFAULT_INSTREAMS
;
1414 sctp_max_outstreams
= SCTP_DEFAULT_OUTSTREAMS
;
1416 /* Initialize handle used for association ids. */
1417 idr_init(&sctp_assocs_id
);
1419 limit
= nr_free_buffer_pages() / 8;
1420 limit
= max(limit
, 128UL);
1421 sysctl_sctp_mem
[0] = limit
/ 4 * 3;
1422 sysctl_sctp_mem
[1] = limit
;
1423 sysctl_sctp_mem
[2] = sysctl_sctp_mem
[0] * 2;
1425 /* Set per-socket limits to no more than 1/128 the pressure threshold*/
1426 limit
= (sysctl_sctp_mem
[1]) << (PAGE_SHIFT
- 7);
1427 max_share
= min(4UL*1024*1024, limit
);
1429 sysctl_sctp_rmem
[0] = SK_MEM_QUANTUM
; /* give each asoc 1 page min */
1430 sysctl_sctp_rmem
[1] = 1500 * SKB_TRUESIZE(1);
1431 sysctl_sctp_rmem
[2] = max(sysctl_sctp_rmem
[1], max_share
);
1433 sysctl_sctp_wmem
[0] = SK_MEM_QUANTUM
;
1434 sysctl_sctp_wmem
[1] = 16*1024;
1435 sysctl_sctp_wmem
[2] = max(64*1024, max_share
);
1437 /* Size and allocate the association hash table.
1438 * The methodology is similar to that of the tcp hash tables.
1439 * Though not identical. Start by getting a goal size
1441 if (nr_pages
>= (128 * 1024))
1442 goal
= nr_pages
>> (22 - PAGE_SHIFT
);
1444 goal
= nr_pages
>> (24 - PAGE_SHIFT
);
1446 /* Then compute the page order for said goal */
1447 order
= get_order(goal
);
1449 /* Now compute the required page order for the maximum sized table we
1452 max_entry_order
= get_order(MAX_SCTP_PORT_HASH_ENTRIES
*
1453 sizeof(struct sctp_bind_hashbucket
));
1455 /* Limit the page order by that maximum hash table size */
1456 order
= min(order
, max_entry_order
);
1458 /* Allocate and initialize the endpoint hash table. */
1459 sctp_ep_hashsize
= 64;
1461 kmalloc_array(64, sizeof(struct sctp_hashbucket
), GFP_KERNEL
);
1462 if (!sctp_ep_hashtable
) {
1463 pr_err("Failed endpoint_hash alloc\n");
1465 goto err_ehash_alloc
;
1467 for (i
= 0; i
< sctp_ep_hashsize
; i
++) {
1468 rwlock_init(&sctp_ep_hashtable
[i
].lock
);
1469 INIT_HLIST_HEAD(&sctp_ep_hashtable
[i
].chain
);
1472 /* Allocate and initialize the SCTP port hash table.
1473 * Note that order is initalized to start at the max sized
1474 * table we want to support. If we can't get that many pages
1475 * reduce the order and try again
1478 sctp_port_hashtable
= (struct sctp_bind_hashbucket
*)
1479 __get_free_pages(GFP_KERNEL
| __GFP_NOWARN
, order
);
1480 } while (!sctp_port_hashtable
&& --order
> 0);
1482 if (!sctp_port_hashtable
) {
1483 pr_err("Failed bind hash alloc\n");
1485 goto err_bhash_alloc
;
1488 /* Now compute the number of entries that will fit in the
1489 * port hash space we allocated
1491 num_entries
= (1UL << order
) * PAGE_SIZE
/
1492 sizeof(struct sctp_bind_hashbucket
);
1494 /* And finish by rounding it down to the nearest power of two
1495 * this wastes some memory of course, but its needed because
1496 * the hash function operates based on the assumption that
1497 * that the number of entries is a power of two
1499 sctp_port_hashsize
= rounddown_pow_of_two(num_entries
);
1501 for (i
= 0; i
< sctp_port_hashsize
; i
++) {
1502 spin_lock_init(&sctp_port_hashtable
[i
].lock
);
1503 INIT_HLIST_HEAD(&sctp_port_hashtable
[i
].chain
);
1506 status
= sctp_transport_hashtable_init();
1508 goto err_thash_alloc
;
1510 pr_info("Hash tables configured (bind %d/%d)\n", sctp_port_hashsize
,
1513 sctp_sysctl_register();
1515 INIT_LIST_HEAD(&sctp_address_families
);
1518 sctp_sched_ops_init();
1520 status
= register_pernet_subsys(&sctp_defaults_ops
);
1522 goto err_register_defaults
;
1524 status
= sctp_v4_protosw_init();
1526 goto err_protosw_init
;
1528 status
= sctp_v6_protosw_init();
1530 goto err_v6_protosw_init
;
1532 status
= register_pernet_subsys(&sctp_ctrlsock_ops
);
1534 goto err_register_ctrlsock
;
1536 status
= sctp_v4_add_protocol();
1538 goto err_add_protocol
;
1540 /* Register SCTP with inet6 layer. */
1541 status
= sctp_v6_add_protocol();
1543 goto err_v6_add_protocol
;
1545 if (sctp_offload_init() < 0)
1546 pr_crit("%s: Cannot add SCTP protocol offload\n", __func__
);
1550 err_v6_add_protocol
:
1551 sctp_v4_del_protocol();
1553 unregister_pernet_subsys(&sctp_ctrlsock_ops
);
1554 err_register_ctrlsock
:
1555 sctp_v6_protosw_exit();
1556 err_v6_protosw_init
:
1557 sctp_v4_protosw_exit();
1559 unregister_pernet_subsys(&sctp_defaults_ops
);
1560 err_register_defaults
:
1563 sctp_sysctl_unregister();
1564 free_pages((unsigned long)sctp_port_hashtable
,
1565 get_order(sctp_port_hashsize
*
1566 sizeof(struct sctp_bind_hashbucket
)));
1568 sctp_transport_hashtable_destroy();
1570 kfree(sctp_ep_hashtable
);
1572 percpu_counter_destroy(&sctp_sockets_allocated
);
1573 err_percpu_counter_init
:
1574 kmem_cache_destroy(sctp_chunk_cachep
);
1576 kmem_cache_destroy(sctp_bucket_cachep
);
1580 /* Exit handler for the SCTP protocol. */
1581 static __exit
void sctp_exit(void)
1583 /* BUG. This should probably do something useful like clean
1584 * up all the remaining associations and all that memory.
1587 /* Unregister with inet6/inet layers. */
1588 sctp_v6_del_protocol();
1589 sctp_v4_del_protocol();
1591 unregister_pernet_subsys(&sctp_ctrlsock_ops
);
1593 /* Free protosw registrations */
1594 sctp_v6_protosw_exit();
1595 sctp_v4_protosw_exit();
1597 unregister_pernet_subsys(&sctp_defaults_ops
);
1599 /* Unregister with socket layer. */
1603 sctp_sysctl_unregister();
1605 free_pages((unsigned long)sctp_port_hashtable
,
1606 get_order(sctp_port_hashsize
*
1607 sizeof(struct sctp_bind_hashbucket
)));
1608 kfree(sctp_ep_hashtable
);
1609 sctp_transport_hashtable_destroy();
1611 percpu_counter_destroy(&sctp_sockets_allocated
);
1613 rcu_barrier(); /* Wait for completion of call_rcu()'s */
1615 kmem_cache_destroy(sctp_chunk_cachep
);
1616 kmem_cache_destroy(sctp_bucket_cachep
);
1619 module_init(sctp_init
);
1620 module_exit(sctp_exit
);
1623 * __stringify doesn't likes enums, so use IPPROTO_SCTP value (132) directly.
1625 MODULE_ALIAS("net-pf-" __stringify(PF_INET
) "-proto-132");
1626 MODULE_ALIAS("net-pf-" __stringify(PF_INET6
) "-proto-132");
1627 MODULE_AUTHOR("Linux Kernel SCTP developers <linux-sctp@vger.kernel.org>");
1628 MODULE_DESCRIPTION("Support for the SCTP protocol (RFC2960)");
1629 module_param_named(no_checksums
, sctp_checksum_disable
, bool, 0644);
1630 MODULE_PARM_DESC(no_checksums
, "Disable checksums computing and verification");
1631 MODULE_LICENSE("GPL");