2 * NET3 IP device support routines.
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
9 * Derived from the IP parts of dev.c 1.0.19
11 * Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
12 * Mark Evans, <evansmp@uhura.aston.ac.uk>
15 * Alan Cox, <gw4pts@gw4pts.ampr.org>
16 * Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
19 * Alexey Kuznetsov: pa_* fields are replaced with ifaddr
21 * Cyrus Durgin: updated for kmod
22 * Matthias Andree: in devinet_ioctl, compare label and
23 * address (4.4BSD alias style support),
24 * fall back to comparing just the label
29 #include <asm/uaccess.h>
30 #include <linux/bitops.h>
31 #include <linux/capability.h>
32 #include <linux/module.h>
33 #include <linux/types.h>
34 #include <linux/kernel.h>
35 #include <linux/string.h>
37 #include <linux/socket.h>
38 #include <linux/sockios.h>
40 #include <linux/errno.h>
41 #include <linux/interrupt.h>
42 #include <linux/if_addr.h>
43 #include <linux/if_ether.h>
44 #include <linux/inet.h>
45 #include <linux/netdevice.h>
46 #include <linux/etherdevice.h>
47 #include <linux/skbuff.h>
48 #include <linux/init.h>
49 #include <linux/notifier.h>
50 #include <linux/inetdevice.h>
51 #include <linux/igmp.h>
52 #include <linux/slab.h>
53 #include <linux/hash.h>
55 #include <linux/sysctl.h>
57 #include <linux/kmod.h>
58 #include <linux/netconf.h>
62 #include <net/route.h>
63 #include <net/ip_fib.h>
64 #include <net/rtnetlink.h>
65 #include <net/net_namespace.h>
66 #include <net/addrconf.h>
68 #include "fib_lookup.h"
70 static struct ipv4_devconf ipv4_devconf
= {
72 [IPV4_DEVCONF_ACCEPT_REDIRECTS
- 1] = 1,
73 [IPV4_DEVCONF_SEND_REDIRECTS
- 1] = 1,
74 [IPV4_DEVCONF_SECURE_REDIRECTS
- 1] = 1,
75 [IPV4_DEVCONF_SHARED_MEDIA
- 1] = 1,
76 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL
- 1] = 10000 /*ms*/,
77 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL
- 1] = 1000 /*ms*/,
81 static struct ipv4_devconf ipv4_devconf_dflt
= {
83 [IPV4_DEVCONF_ACCEPT_REDIRECTS
- 1] = 1,
84 [IPV4_DEVCONF_SEND_REDIRECTS
- 1] = 1,
85 [IPV4_DEVCONF_SECURE_REDIRECTS
- 1] = 1,
86 [IPV4_DEVCONF_SHARED_MEDIA
- 1] = 1,
87 [IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE
- 1] = 1,
88 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL
- 1] = 10000 /*ms*/,
89 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL
- 1] = 1000 /*ms*/,
93 #define IPV4_DEVCONF_DFLT(net, attr) \
94 IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr)
96 static const struct nla_policy ifa_ipv4_policy
[IFA_MAX
+1] = {
97 [IFA_LOCAL
] = { .type
= NLA_U32
},
98 [IFA_ADDRESS
] = { .type
= NLA_U32
},
99 [IFA_BROADCAST
] = { .type
= NLA_U32
},
100 [IFA_LABEL
] = { .type
= NLA_STRING
, .len
= IFNAMSIZ
- 1 },
101 [IFA_CACHEINFO
] = { .len
= sizeof(struct ifa_cacheinfo
) },
102 [IFA_FLAGS
] = { .type
= NLA_U32
},
105 #define IN4_ADDR_HSIZE_SHIFT 8
106 #define IN4_ADDR_HSIZE (1U << IN4_ADDR_HSIZE_SHIFT)
108 static struct hlist_head inet_addr_lst
[IN4_ADDR_HSIZE
];
110 static u32
inet_addr_hash(const struct net
*net
, __be32 addr
)
112 u32 val
= (__force u32
) addr
^ net_hash_mix(net
);
114 return hash_32(val
, IN4_ADDR_HSIZE_SHIFT
);
117 static void inet_hash_insert(struct net
*net
, struct in_ifaddr
*ifa
)
119 u32 hash
= inet_addr_hash(net
, ifa
->ifa_local
);
122 hlist_add_head_rcu(&ifa
->hash
, &inet_addr_lst
[hash
]);
125 static void inet_hash_remove(struct in_ifaddr
*ifa
)
128 hlist_del_init_rcu(&ifa
->hash
);
132 * __ip_dev_find - find the first device with a given source address.
133 * @net: the net namespace
134 * @addr: the source address
135 * @devref: if true, take a reference on the found device
137 * If a caller uses devref=false, it should be protected by RCU, or RTNL
139 struct net_device
*__ip_dev_find(struct net
*net
, __be32 addr
, bool devref
)
141 u32 hash
= inet_addr_hash(net
, addr
);
142 struct net_device
*result
= NULL
;
143 struct in_ifaddr
*ifa
;
146 hlist_for_each_entry_rcu(ifa
, &inet_addr_lst
[hash
], hash
) {
147 if (ifa
->ifa_local
== addr
) {
148 struct net_device
*dev
= ifa
->ifa_dev
->dev
;
150 if (!net_eq(dev_net(dev
), net
))
157 struct flowi4 fl4
= { .daddr
= addr
};
158 struct fib_result res
= { 0 };
159 struct fib_table
*local
;
161 /* Fallback to FIB local table so that communication
162 * over loopback subnets work.
164 local
= fib_get_table(net
, RT_TABLE_LOCAL
);
166 !fib_table_lookup(local
, &fl4
, &res
, FIB_LOOKUP_NOREF
) &&
167 res
.type
== RTN_LOCAL
)
168 result
= FIB_RES_DEV(res
);
170 if (result
&& devref
)
175 EXPORT_SYMBOL(__ip_dev_find
);
177 static void rtmsg_ifa(int event
, struct in_ifaddr
*, struct nlmsghdr
*, u32
);
179 static BLOCKING_NOTIFIER_HEAD(inetaddr_chain
);
180 static void inet_del_ifa(struct in_device
*in_dev
, struct in_ifaddr
**ifap
,
183 static int devinet_sysctl_register(struct in_device
*idev
);
184 static void devinet_sysctl_unregister(struct in_device
*idev
);
186 static int devinet_sysctl_register(struct in_device
*idev
)
190 static void devinet_sysctl_unregister(struct in_device
*idev
)
195 /* Locks all the inet devices. */
197 static struct in_ifaddr
*inet_alloc_ifa(void)
199 return kzalloc(sizeof(struct in_ifaddr
), GFP_KERNEL
);
202 static void inet_rcu_free_ifa(struct rcu_head
*head
)
204 struct in_ifaddr
*ifa
= container_of(head
, struct in_ifaddr
, rcu_head
);
206 in_dev_put(ifa
->ifa_dev
);
210 static void inet_free_ifa(struct in_ifaddr
*ifa
)
212 call_rcu(&ifa
->rcu_head
, inet_rcu_free_ifa
);
215 void in_dev_finish_destroy(struct in_device
*idev
)
217 struct net_device
*dev
= idev
->dev
;
219 WARN_ON(idev
->ifa_list
);
220 WARN_ON(idev
->mc_list
);
221 kfree(rcu_dereference_protected(idev
->mc_hash
, 1));
222 #ifdef NET_REFCNT_DEBUG
223 pr_debug("%s: %p=%s\n", __func__
, idev
, dev
? dev
->name
: "NIL");
227 pr_err("Freeing alive in_device %p\n", idev
);
231 EXPORT_SYMBOL(in_dev_finish_destroy
);
233 static struct in_device
*inetdev_init(struct net_device
*dev
)
235 struct in_device
*in_dev
;
240 in_dev
= kzalloc(sizeof(*in_dev
), GFP_KERNEL
);
243 memcpy(&in_dev
->cnf
, dev_net(dev
)->ipv4
.devconf_dflt
,
244 sizeof(in_dev
->cnf
));
245 in_dev
->cnf
.sysctl
= NULL
;
247 in_dev
->arp_parms
= neigh_parms_alloc(dev
, &arp_tbl
);
248 if (!in_dev
->arp_parms
)
250 if (IPV4_DEVCONF(in_dev
->cnf
, FORWARDING
))
251 dev_disable_lro(dev
);
252 /* Reference in_dev->dev */
254 /* Account for reference dev->ip_ptr (below) */
257 err
= devinet_sysctl_register(in_dev
);
264 ip_mc_init_dev(in_dev
);
265 if (dev
->flags
& IFF_UP
)
268 /* we can receive as soon as ip_ptr is set -- do this last */
269 rcu_assign_pointer(dev
->ip_ptr
, in_dev
);
271 return in_dev
?: ERR_PTR(err
);
278 static void in_dev_rcu_put(struct rcu_head
*head
)
280 struct in_device
*idev
= container_of(head
, struct in_device
, rcu_head
);
284 static void inetdev_destroy(struct in_device
*in_dev
)
286 struct in_ifaddr
*ifa
;
287 struct net_device
*dev
;
295 ip_mc_destroy_dev(in_dev
);
297 while ((ifa
= in_dev
->ifa_list
) != NULL
) {
298 inet_del_ifa(in_dev
, &in_dev
->ifa_list
, 0);
302 RCU_INIT_POINTER(dev
->ip_ptr
, NULL
);
304 devinet_sysctl_unregister(in_dev
);
305 neigh_parms_release(&arp_tbl
, in_dev
->arp_parms
);
308 call_rcu(&in_dev
->rcu_head
, in_dev_rcu_put
);
311 int inet_addr_onlink(struct in_device
*in_dev
, __be32 a
, __be32 b
)
314 for_primary_ifa(in_dev
) {
315 if (inet_ifa_match(a
, ifa
)) {
316 if (!b
|| inet_ifa_match(b
, ifa
)) {
321 } endfor_ifa(in_dev
);
326 static void __inet_del_ifa(struct in_device
*in_dev
, struct in_ifaddr
**ifap
,
327 int destroy
, struct nlmsghdr
*nlh
, u32 portid
)
329 struct in_ifaddr
*promote
= NULL
;
330 struct in_ifaddr
*ifa
, *ifa1
= *ifap
;
331 struct in_ifaddr
*last_prim
= in_dev
->ifa_list
;
332 struct in_ifaddr
*prev_prom
= NULL
;
333 int do_promote
= IN_DEV_PROMOTE_SECONDARIES(in_dev
);
337 /* 1. Deleting primary ifaddr forces deletion all secondaries
338 * unless alias promotion is set
341 if (!(ifa1
->ifa_flags
& IFA_F_SECONDARY
)) {
342 struct in_ifaddr
**ifap1
= &ifa1
->ifa_next
;
344 while ((ifa
= *ifap1
) != NULL
) {
345 if (!(ifa
->ifa_flags
& IFA_F_SECONDARY
) &&
346 ifa1
->ifa_scope
<= ifa
->ifa_scope
)
349 if (!(ifa
->ifa_flags
& IFA_F_SECONDARY
) ||
350 ifa1
->ifa_mask
!= ifa
->ifa_mask
||
351 !inet_ifa_match(ifa1
->ifa_address
, ifa
)) {
352 ifap1
= &ifa
->ifa_next
;
358 inet_hash_remove(ifa
);
359 *ifap1
= ifa
->ifa_next
;
361 rtmsg_ifa(RTM_DELADDR
, ifa
, nlh
, portid
);
362 blocking_notifier_call_chain(&inetaddr_chain
,
372 /* On promotion all secondaries from subnet are changing
373 * the primary IP, we must remove all their routes silently
374 * and later to add them back with new prefsrc. Do this
375 * while all addresses are on the device list.
377 for (ifa
= promote
; ifa
; ifa
= ifa
->ifa_next
) {
378 if (ifa1
->ifa_mask
== ifa
->ifa_mask
&&
379 inet_ifa_match(ifa1
->ifa_address
, ifa
))
380 fib_del_ifaddr(ifa
, ifa1
);
385 *ifap
= ifa1
->ifa_next
;
386 inet_hash_remove(ifa1
);
388 /* 3. Announce address deletion */
390 /* Send message first, then call notifier.
391 At first sight, FIB update triggered by notifier
392 will refer to already deleted ifaddr, that could confuse
393 netlink listeners. It is not true: look, gated sees
394 that route deleted and if it still thinks that ifaddr
395 is valid, it will try to restore deleted routes... Grr.
396 So that, this order is correct.
398 rtmsg_ifa(RTM_DELADDR
, ifa1
, nlh
, portid
);
399 blocking_notifier_call_chain(&inetaddr_chain
, NETDEV_DOWN
, ifa1
);
402 struct in_ifaddr
*next_sec
= promote
->ifa_next
;
405 prev_prom
->ifa_next
= promote
->ifa_next
;
406 promote
->ifa_next
= last_prim
->ifa_next
;
407 last_prim
->ifa_next
= promote
;
410 promote
->ifa_flags
&= ~IFA_F_SECONDARY
;
411 rtmsg_ifa(RTM_NEWADDR
, promote
, nlh
, portid
);
412 blocking_notifier_call_chain(&inetaddr_chain
,
414 for (ifa
= next_sec
; ifa
; ifa
= ifa
->ifa_next
) {
415 if (ifa1
->ifa_mask
!= ifa
->ifa_mask
||
416 !inet_ifa_match(ifa1
->ifa_address
, ifa
))
426 static void inet_del_ifa(struct in_device
*in_dev
, struct in_ifaddr
**ifap
,
429 __inet_del_ifa(in_dev
, ifap
, destroy
, NULL
, 0);
432 static void check_lifetime(struct work_struct
*work
);
434 static DECLARE_DELAYED_WORK(check_lifetime_work
, check_lifetime
);
436 static int __inet_insert_ifa(struct in_ifaddr
*ifa
, struct nlmsghdr
*nlh
,
439 struct in_device
*in_dev
= ifa
->ifa_dev
;
440 struct in_ifaddr
*ifa1
, **ifap
, **last_primary
;
444 if (!ifa
->ifa_local
) {
449 ifa
->ifa_flags
&= ~IFA_F_SECONDARY
;
450 last_primary
= &in_dev
->ifa_list
;
452 for (ifap
= &in_dev
->ifa_list
; (ifa1
= *ifap
) != NULL
;
453 ifap
= &ifa1
->ifa_next
) {
454 if (!(ifa1
->ifa_flags
& IFA_F_SECONDARY
) &&
455 ifa
->ifa_scope
<= ifa1
->ifa_scope
)
456 last_primary
= &ifa1
->ifa_next
;
457 if (ifa1
->ifa_mask
== ifa
->ifa_mask
&&
458 inet_ifa_match(ifa1
->ifa_address
, ifa
)) {
459 if (ifa1
->ifa_local
== ifa
->ifa_local
) {
463 if (ifa1
->ifa_scope
!= ifa
->ifa_scope
) {
467 ifa
->ifa_flags
|= IFA_F_SECONDARY
;
471 if (!(ifa
->ifa_flags
& IFA_F_SECONDARY
)) {
472 prandom_seed((__force u32
) ifa
->ifa_local
);
476 ifa
->ifa_next
= *ifap
;
479 inet_hash_insert(dev_net(in_dev
->dev
), ifa
);
481 cancel_delayed_work(&check_lifetime_work
);
482 queue_delayed_work(system_power_efficient_wq
, &check_lifetime_work
, 0);
484 /* Send message first, then call notifier.
485 Notifier will trigger FIB update, so that
486 listeners of netlink will know about new ifaddr */
487 rtmsg_ifa(RTM_NEWADDR
, ifa
, nlh
, portid
);
488 blocking_notifier_call_chain(&inetaddr_chain
, NETDEV_UP
, ifa
);
493 static int inet_insert_ifa(struct in_ifaddr
*ifa
)
495 return __inet_insert_ifa(ifa
, NULL
, 0);
498 static int inet_set_ifa(struct net_device
*dev
, struct in_ifaddr
*ifa
)
500 struct in_device
*in_dev
= __in_dev_get_rtnl(dev
);
508 ipv4_devconf_setall(in_dev
);
509 neigh_parms_data_state_setall(in_dev
->arp_parms
);
510 if (ifa
->ifa_dev
!= in_dev
) {
511 WARN_ON(ifa
->ifa_dev
);
513 ifa
->ifa_dev
= in_dev
;
515 if (ipv4_is_loopback(ifa
->ifa_local
))
516 ifa
->ifa_scope
= RT_SCOPE_HOST
;
517 return inet_insert_ifa(ifa
);
520 /* Caller must hold RCU or RTNL :
521 * We dont take a reference on found in_device
523 struct in_device
*inetdev_by_index(struct net
*net
, int ifindex
)
525 struct net_device
*dev
;
526 struct in_device
*in_dev
= NULL
;
529 dev
= dev_get_by_index_rcu(net
, ifindex
);
531 in_dev
= rcu_dereference_rtnl(dev
->ip_ptr
);
535 EXPORT_SYMBOL(inetdev_by_index
);
537 /* Called only from RTNL semaphored context. No locks. */
539 struct in_ifaddr
*inet_ifa_byprefix(struct in_device
*in_dev
, __be32 prefix
,
544 for_primary_ifa(in_dev
) {
545 if (ifa
->ifa_mask
== mask
&& inet_ifa_match(prefix
, ifa
))
547 } endfor_ifa(in_dev
);
551 static int ip_mc_config(struct sock
*sk
, bool join
, const struct in_ifaddr
*ifa
)
553 struct ip_mreqn mreq
= {
554 .imr_multiaddr
.s_addr
= ifa
->ifa_address
,
555 .imr_ifindex
= ifa
->ifa_dev
->dev
->ifindex
,
563 ret
= ip_mc_join_group(sk
, &mreq
);
565 ret
= ip_mc_leave_group(sk
, &mreq
);
571 static int inet_rtm_deladdr(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
573 struct net
*net
= sock_net(skb
->sk
);
574 struct nlattr
*tb
[IFA_MAX
+1];
575 struct in_device
*in_dev
;
576 struct ifaddrmsg
*ifm
;
577 struct in_ifaddr
*ifa
, **ifap
;
582 err
= nlmsg_parse(nlh
, sizeof(*ifm
), tb
, IFA_MAX
, ifa_ipv4_policy
);
586 ifm
= nlmsg_data(nlh
);
587 in_dev
= inetdev_by_index(net
, ifm
->ifa_index
);
593 for (ifap
= &in_dev
->ifa_list
; (ifa
= *ifap
) != NULL
;
594 ifap
= &ifa
->ifa_next
) {
596 ifa
->ifa_local
!= nla_get_in_addr(tb
[IFA_LOCAL
]))
599 if (tb
[IFA_LABEL
] && nla_strcmp(tb
[IFA_LABEL
], ifa
->ifa_label
))
602 if (tb
[IFA_ADDRESS
] &&
603 (ifm
->ifa_prefixlen
!= ifa
->ifa_prefixlen
||
604 !inet_ifa_match(nla_get_in_addr(tb
[IFA_ADDRESS
]), ifa
)))
607 if (ipv4_is_multicast(ifa
->ifa_address
))
608 ip_mc_config(net
->ipv4
.mc_autojoin_sk
, false, ifa
);
609 __inet_del_ifa(in_dev
, ifap
, 1, nlh
, NETLINK_CB(skb
).portid
);
613 err
= -EADDRNOTAVAIL
;
618 #define INFINITY_LIFE_TIME 0xFFFFFFFF
620 static void check_lifetime(struct work_struct
*work
)
622 unsigned long now
, next
, next_sec
, next_sched
;
623 struct in_ifaddr
*ifa
;
624 struct hlist_node
*n
;
628 next
= round_jiffies_up(now
+ ADDR_CHECK_FREQUENCY
);
630 for (i
= 0; i
< IN4_ADDR_HSIZE
; i
++) {
631 bool change_needed
= false;
634 hlist_for_each_entry_rcu(ifa
, &inet_addr_lst
[i
], hash
) {
637 if (ifa
->ifa_flags
& IFA_F_PERMANENT
)
640 /* We try to batch several events at once. */
641 age
= (now
- ifa
->ifa_tstamp
+
642 ADDRCONF_TIMER_FUZZ_MINUS
) / HZ
;
644 if (ifa
->ifa_valid_lft
!= INFINITY_LIFE_TIME
&&
645 age
>= ifa
->ifa_valid_lft
) {
646 change_needed
= true;
647 } else if (ifa
->ifa_preferred_lft
==
648 INFINITY_LIFE_TIME
) {
650 } else if (age
>= ifa
->ifa_preferred_lft
) {
651 if (time_before(ifa
->ifa_tstamp
+
652 ifa
->ifa_valid_lft
* HZ
, next
))
653 next
= ifa
->ifa_tstamp
+
654 ifa
->ifa_valid_lft
* HZ
;
656 if (!(ifa
->ifa_flags
& IFA_F_DEPRECATED
))
657 change_needed
= true;
658 } else if (time_before(ifa
->ifa_tstamp
+
659 ifa
->ifa_preferred_lft
* HZ
,
661 next
= ifa
->ifa_tstamp
+
662 ifa
->ifa_preferred_lft
* HZ
;
669 hlist_for_each_entry_safe(ifa
, n
, &inet_addr_lst
[i
], hash
) {
672 if (ifa
->ifa_flags
& IFA_F_PERMANENT
)
675 /* We try to batch several events at once. */
676 age
= (now
- ifa
->ifa_tstamp
+
677 ADDRCONF_TIMER_FUZZ_MINUS
) / HZ
;
679 if (ifa
->ifa_valid_lft
!= INFINITY_LIFE_TIME
&&
680 age
>= ifa
->ifa_valid_lft
) {
681 struct in_ifaddr
**ifap
;
683 for (ifap
= &ifa
->ifa_dev
->ifa_list
;
684 *ifap
!= NULL
; ifap
= &(*ifap
)->ifa_next
) {
686 inet_del_ifa(ifa
->ifa_dev
,
691 } else if (ifa
->ifa_preferred_lft
!=
692 INFINITY_LIFE_TIME
&&
693 age
>= ifa
->ifa_preferred_lft
&&
694 !(ifa
->ifa_flags
& IFA_F_DEPRECATED
)) {
695 ifa
->ifa_flags
|= IFA_F_DEPRECATED
;
696 rtmsg_ifa(RTM_NEWADDR
, ifa
, NULL
, 0);
702 next_sec
= round_jiffies_up(next
);
705 /* If rounded timeout is accurate enough, accept it. */
706 if (time_before(next_sec
, next
+ ADDRCONF_TIMER_FUZZ
))
707 next_sched
= next_sec
;
710 /* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */
711 if (time_before(next_sched
, now
+ ADDRCONF_TIMER_FUZZ_MAX
))
712 next_sched
= now
+ ADDRCONF_TIMER_FUZZ_MAX
;
714 queue_delayed_work(system_power_efficient_wq
, &check_lifetime_work
,
718 static void set_ifa_lifetime(struct in_ifaddr
*ifa
, __u32 valid_lft
,
721 unsigned long timeout
;
723 ifa
->ifa_flags
&= ~(IFA_F_PERMANENT
| IFA_F_DEPRECATED
);
725 timeout
= addrconf_timeout_fixup(valid_lft
, HZ
);
726 if (addrconf_finite_timeout(timeout
))
727 ifa
->ifa_valid_lft
= timeout
;
729 ifa
->ifa_flags
|= IFA_F_PERMANENT
;
731 timeout
= addrconf_timeout_fixup(prefered_lft
, HZ
);
732 if (addrconf_finite_timeout(timeout
)) {
734 ifa
->ifa_flags
|= IFA_F_DEPRECATED
;
735 ifa
->ifa_preferred_lft
= timeout
;
737 ifa
->ifa_tstamp
= jiffies
;
738 if (!ifa
->ifa_cstamp
)
739 ifa
->ifa_cstamp
= ifa
->ifa_tstamp
;
742 static struct in_ifaddr
*rtm_to_ifaddr(struct net
*net
, struct nlmsghdr
*nlh
,
743 __u32
*pvalid_lft
, __u32
*pprefered_lft
)
745 struct nlattr
*tb
[IFA_MAX
+1];
746 struct in_ifaddr
*ifa
;
747 struct ifaddrmsg
*ifm
;
748 struct net_device
*dev
;
749 struct in_device
*in_dev
;
752 err
= nlmsg_parse(nlh
, sizeof(*ifm
), tb
, IFA_MAX
, ifa_ipv4_policy
);
756 ifm
= nlmsg_data(nlh
);
758 if (ifm
->ifa_prefixlen
> 32 || !tb
[IFA_LOCAL
])
761 dev
= __dev_get_by_index(net
, ifm
->ifa_index
);
766 in_dev
= __in_dev_get_rtnl(dev
);
771 ifa
= inet_alloc_ifa();
774 * A potential indev allocation can be left alive, it stays
775 * assigned to its device and is destroy with it.
779 ipv4_devconf_setall(in_dev
);
780 neigh_parms_data_state_setall(in_dev
->arp_parms
);
783 if (!tb
[IFA_ADDRESS
])
784 tb
[IFA_ADDRESS
] = tb
[IFA_LOCAL
];
786 INIT_HLIST_NODE(&ifa
->hash
);
787 ifa
->ifa_prefixlen
= ifm
->ifa_prefixlen
;
788 ifa
->ifa_mask
= inet_make_mask(ifm
->ifa_prefixlen
);
789 ifa
->ifa_flags
= tb
[IFA_FLAGS
] ? nla_get_u32(tb
[IFA_FLAGS
]) :
791 ifa
->ifa_scope
= ifm
->ifa_scope
;
792 ifa
->ifa_dev
= in_dev
;
794 ifa
->ifa_local
= nla_get_in_addr(tb
[IFA_LOCAL
]);
795 ifa
->ifa_address
= nla_get_in_addr(tb
[IFA_ADDRESS
]);
797 if (tb
[IFA_BROADCAST
])
798 ifa
->ifa_broadcast
= nla_get_in_addr(tb
[IFA_BROADCAST
]);
801 nla_strlcpy(ifa
->ifa_label
, tb
[IFA_LABEL
], IFNAMSIZ
);
803 memcpy(ifa
->ifa_label
, dev
->name
, IFNAMSIZ
);
805 if (tb
[IFA_CACHEINFO
]) {
806 struct ifa_cacheinfo
*ci
;
808 ci
= nla_data(tb
[IFA_CACHEINFO
]);
809 if (!ci
->ifa_valid
|| ci
->ifa_prefered
> ci
->ifa_valid
) {
813 *pvalid_lft
= ci
->ifa_valid
;
814 *pprefered_lft
= ci
->ifa_prefered
;
825 static struct in_ifaddr
*find_matching_ifa(struct in_ifaddr
*ifa
)
827 struct in_device
*in_dev
= ifa
->ifa_dev
;
828 struct in_ifaddr
*ifa1
, **ifap
;
833 for (ifap
= &in_dev
->ifa_list
; (ifa1
= *ifap
) != NULL
;
834 ifap
= &ifa1
->ifa_next
) {
835 if (ifa1
->ifa_mask
== ifa
->ifa_mask
&&
836 inet_ifa_match(ifa1
->ifa_address
, ifa
) &&
837 ifa1
->ifa_local
== ifa
->ifa_local
)
843 static int inet_rtm_newaddr(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
845 struct net
*net
= sock_net(skb
->sk
);
846 struct in_ifaddr
*ifa
;
847 struct in_ifaddr
*ifa_existing
;
848 __u32 valid_lft
= INFINITY_LIFE_TIME
;
849 __u32 prefered_lft
= INFINITY_LIFE_TIME
;
853 ifa
= rtm_to_ifaddr(net
, nlh
, &valid_lft
, &prefered_lft
);
857 ifa_existing
= find_matching_ifa(ifa
);
859 /* It would be best to check for !NLM_F_CREATE here but
860 * userspace already relies on not having to provide this.
862 set_ifa_lifetime(ifa
, valid_lft
, prefered_lft
);
863 if (ifa
->ifa_flags
& IFA_F_MCAUTOJOIN
) {
864 int ret
= ip_mc_config(net
->ipv4
.mc_autojoin_sk
,
872 return __inet_insert_ifa(ifa
, nlh
, NETLINK_CB(skb
).portid
);
876 if (nlh
->nlmsg_flags
& NLM_F_EXCL
||
877 !(nlh
->nlmsg_flags
& NLM_F_REPLACE
))
880 set_ifa_lifetime(ifa
, valid_lft
, prefered_lft
);
881 cancel_delayed_work(&check_lifetime_work
);
882 queue_delayed_work(system_power_efficient_wq
,
883 &check_lifetime_work
, 0);
884 rtmsg_ifa(RTM_NEWADDR
, ifa
, nlh
, NETLINK_CB(skb
).portid
);
890 * Determine a default network mask, based on the IP address.
893 static int inet_abc_len(__be32 addr
)
895 int rc
= -1; /* Something else, probably a multicast. */
897 if (ipv4_is_zeronet(addr
))
900 __u32 haddr
= ntohl(addr
);
902 if (IN_CLASSA(haddr
))
904 else if (IN_CLASSB(haddr
))
906 else if (IN_CLASSC(haddr
))
914 int devinet_ioctl(struct net
*net
, unsigned int cmd
, void __user
*arg
)
917 struct sockaddr_in sin_orig
;
918 struct sockaddr_in
*sin
= (struct sockaddr_in
*)&ifr
.ifr_addr
;
919 struct in_device
*in_dev
;
920 struct in_ifaddr
**ifap
= NULL
;
921 struct in_ifaddr
*ifa
= NULL
;
922 struct net_device
*dev
;
925 int tryaddrmatch
= 0;
928 * Fetch the caller's info block into kernel space
931 if (copy_from_user(&ifr
, arg
, sizeof(struct ifreq
)))
933 ifr
.ifr_name
[IFNAMSIZ
- 1] = 0;
935 /* save original address for comparison */
936 memcpy(&sin_orig
, sin
, sizeof(*sin
));
938 colon
= strchr(ifr
.ifr_name
, ':');
942 dev_load(net
, ifr
.ifr_name
);
945 case SIOCGIFADDR
: /* Get interface address */
946 case SIOCGIFBRDADDR
: /* Get the broadcast address */
947 case SIOCGIFDSTADDR
: /* Get the destination address */
948 case SIOCGIFNETMASK
: /* Get the netmask for the interface */
949 /* Note that these ioctls will not sleep,
950 so that we do not impose a lock.
951 One day we will be forced to put shlock here (I mean SMP)
953 tryaddrmatch
= (sin_orig
.sin_family
== AF_INET
);
954 memset(sin
, 0, sizeof(*sin
));
955 sin
->sin_family
= AF_INET
;
960 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
963 case SIOCSIFADDR
: /* Set interface address (and family) */
964 case SIOCSIFBRDADDR
: /* Set the broadcast address */
965 case SIOCSIFDSTADDR
: /* Set the destination address */
966 case SIOCSIFNETMASK
: /* Set the netmask for the interface */
968 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
971 if (sin
->sin_family
!= AF_INET
)
982 dev
= __dev_get_by_name(net
, ifr
.ifr_name
);
989 in_dev
= __in_dev_get_rtnl(dev
);
992 /* Matthias Andree */
993 /* compare label and address (4.4BSD style) */
994 /* note: we only do this for a limited set of ioctls
995 and only if the original address family was AF_INET.
996 This is checked above. */
997 for (ifap
= &in_dev
->ifa_list
; (ifa
= *ifap
) != NULL
;
998 ifap
= &ifa
->ifa_next
) {
999 if (!strcmp(ifr
.ifr_name
, ifa
->ifa_label
) &&
1000 sin_orig
.sin_addr
.s_addr
==
1006 /* we didn't get a match, maybe the application is
1007 4.3BSD-style and passed in junk so we fall back to
1008 comparing just the label */
1010 for (ifap
= &in_dev
->ifa_list
; (ifa
= *ifap
) != NULL
;
1011 ifap
= &ifa
->ifa_next
)
1012 if (!strcmp(ifr
.ifr_name
, ifa
->ifa_label
))
1017 ret
= -EADDRNOTAVAIL
;
1018 if (!ifa
&& cmd
!= SIOCSIFADDR
&& cmd
!= SIOCSIFFLAGS
)
1022 case SIOCGIFADDR
: /* Get interface address */
1023 sin
->sin_addr
.s_addr
= ifa
->ifa_local
;
1026 case SIOCGIFBRDADDR
: /* Get the broadcast address */
1027 sin
->sin_addr
.s_addr
= ifa
->ifa_broadcast
;
1030 case SIOCGIFDSTADDR
: /* Get the destination address */
1031 sin
->sin_addr
.s_addr
= ifa
->ifa_address
;
1034 case SIOCGIFNETMASK
: /* Get the netmask for the interface */
1035 sin
->sin_addr
.s_addr
= ifa
->ifa_mask
;
1040 ret
= -EADDRNOTAVAIL
;
1044 if (!(ifr
.ifr_flags
& IFF_UP
))
1045 inet_del_ifa(in_dev
, ifap
, 1);
1048 ret
= dev_change_flags(dev
, ifr
.ifr_flags
);
1051 case SIOCSIFADDR
: /* Set interface address (and family) */
1053 if (inet_abc_len(sin
->sin_addr
.s_addr
) < 0)
1058 ifa
= inet_alloc_ifa();
1061 INIT_HLIST_NODE(&ifa
->hash
);
1063 memcpy(ifa
->ifa_label
, ifr
.ifr_name
, IFNAMSIZ
);
1065 memcpy(ifa
->ifa_label
, dev
->name
, IFNAMSIZ
);
1068 if (ifa
->ifa_local
== sin
->sin_addr
.s_addr
)
1070 inet_del_ifa(in_dev
, ifap
, 0);
1071 ifa
->ifa_broadcast
= 0;
1075 ifa
->ifa_address
= ifa
->ifa_local
= sin
->sin_addr
.s_addr
;
1077 if (!(dev
->flags
& IFF_POINTOPOINT
)) {
1078 ifa
->ifa_prefixlen
= inet_abc_len(ifa
->ifa_address
);
1079 ifa
->ifa_mask
= inet_make_mask(ifa
->ifa_prefixlen
);
1080 if ((dev
->flags
& IFF_BROADCAST
) &&
1081 ifa
->ifa_prefixlen
< 31)
1082 ifa
->ifa_broadcast
= ifa
->ifa_address
|
1085 ifa
->ifa_prefixlen
= 32;
1086 ifa
->ifa_mask
= inet_make_mask(32);
1088 set_ifa_lifetime(ifa
, INFINITY_LIFE_TIME
, INFINITY_LIFE_TIME
);
1089 ret
= inet_set_ifa(dev
, ifa
);
1092 case SIOCSIFBRDADDR
: /* Set the broadcast address */
1094 if (ifa
->ifa_broadcast
!= sin
->sin_addr
.s_addr
) {
1095 inet_del_ifa(in_dev
, ifap
, 0);
1096 ifa
->ifa_broadcast
= sin
->sin_addr
.s_addr
;
1097 inet_insert_ifa(ifa
);
1101 case SIOCSIFDSTADDR
: /* Set the destination address */
1103 if (ifa
->ifa_address
== sin
->sin_addr
.s_addr
)
1106 if (inet_abc_len(sin
->sin_addr
.s_addr
) < 0)
1109 inet_del_ifa(in_dev
, ifap
, 0);
1110 ifa
->ifa_address
= sin
->sin_addr
.s_addr
;
1111 inet_insert_ifa(ifa
);
1114 case SIOCSIFNETMASK
: /* Set the netmask for the interface */
1117 * The mask we set must be legal.
1120 if (bad_mask(sin
->sin_addr
.s_addr
, 0))
1123 if (ifa
->ifa_mask
!= sin
->sin_addr
.s_addr
) {
1124 __be32 old_mask
= ifa
->ifa_mask
;
1125 inet_del_ifa(in_dev
, ifap
, 0);
1126 ifa
->ifa_mask
= sin
->sin_addr
.s_addr
;
1127 ifa
->ifa_prefixlen
= inet_mask_len(ifa
->ifa_mask
);
1129 /* See if current broadcast address matches
1130 * with current netmask, then recalculate
1131 * the broadcast address. Otherwise it's a
1132 * funny address, so don't touch it since
1133 * the user seems to know what (s)he's doing...
1135 if ((dev
->flags
& IFF_BROADCAST
) &&
1136 (ifa
->ifa_prefixlen
< 31) &&
1137 (ifa
->ifa_broadcast
==
1138 (ifa
->ifa_local
|~old_mask
))) {
1139 ifa
->ifa_broadcast
= (ifa
->ifa_local
|
1140 ~sin
->sin_addr
.s_addr
);
1142 inet_insert_ifa(ifa
);
1152 ret
= copy_to_user(arg
, &ifr
, sizeof(struct ifreq
)) ? -EFAULT
: 0;
1156 static int inet_gifconf(struct net_device
*dev
, char __user
*buf
, int len
)
1158 struct in_device
*in_dev
= __in_dev_get_rtnl(dev
);
1159 struct in_ifaddr
*ifa
;
1166 for (ifa
= in_dev
->ifa_list
; ifa
; ifa
= ifa
->ifa_next
) {
1168 done
+= sizeof(ifr
);
1171 if (len
< (int) sizeof(ifr
))
1173 memset(&ifr
, 0, sizeof(struct ifreq
));
1174 strcpy(ifr
.ifr_name
, ifa
->ifa_label
);
1176 (*(struct sockaddr_in
*)&ifr
.ifr_addr
).sin_family
= AF_INET
;
1177 (*(struct sockaddr_in
*)&ifr
.ifr_addr
).sin_addr
.s_addr
=
1180 if (copy_to_user(buf
, &ifr
, sizeof(struct ifreq
))) {
1184 buf
+= sizeof(struct ifreq
);
1185 len
-= sizeof(struct ifreq
);
1186 done
+= sizeof(struct ifreq
);
1192 __be32
inet_select_addr(const struct net_device
*dev
, __be32 dst
, int scope
)
1195 struct in_device
*in_dev
;
1196 struct net
*net
= dev_net(dev
);
1199 in_dev
= __in_dev_get_rcu(dev
);
1203 for_primary_ifa(in_dev
) {
1204 if (ifa
->ifa_scope
> scope
)
1206 if (!dst
|| inet_ifa_match(dst
, ifa
)) {
1207 addr
= ifa
->ifa_local
;
1211 addr
= ifa
->ifa_local
;
1212 } endfor_ifa(in_dev
);
1218 /* Not loopback addresses on loopback should be preferred
1219 in this case. It is important that lo is the first interface
1222 for_each_netdev_rcu(net
, dev
) {
1223 in_dev
= __in_dev_get_rcu(dev
);
1227 for_primary_ifa(in_dev
) {
1228 if (ifa
->ifa_scope
!= RT_SCOPE_LINK
&&
1229 ifa
->ifa_scope
<= scope
) {
1230 addr
= ifa
->ifa_local
;
1233 } endfor_ifa(in_dev
);
1239 EXPORT_SYMBOL(inet_select_addr
);
1241 static __be32
confirm_addr_indev(struct in_device
*in_dev
, __be32 dst
,
1242 __be32 local
, int scope
)
1249 (local
== ifa
->ifa_local
|| !local
) &&
1250 ifa
->ifa_scope
<= scope
) {
1251 addr
= ifa
->ifa_local
;
1256 same
= (!local
|| inet_ifa_match(local
, ifa
)) &&
1257 (!dst
|| inet_ifa_match(dst
, ifa
));
1261 /* Is the selected addr into dst subnet? */
1262 if (inet_ifa_match(addr
, ifa
))
1264 /* No, then can we use new local src? */
1265 if (ifa
->ifa_scope
<= scope
) {
1266 addr
= ifa
->ifa_local
;
1269 /* search for large dst subnet for addr */
1273 } endfor_ifa(in_dev
);
1275 return same
? addr
: 0;
1279 * Confirm that local IP address exists using wildcards:
1280 * - net: netns to check, cannot be NULL
1281 * - in_dev: only on this interface, NULL=any interface
1282 * - dst: only in the same subnet as dst, 0=any dst
1283 * - local: address, 0=autoselect the local address
1284 * - scope: maximum allowed scope value for the local address
1286 __be32
inet_confirm_addr(struct net
*net
, struct in_device
*in_dev
,
1287 __be32 dst
, __be32 local
, int scope
)
1290 struct net_device
*dev
;
1293 return confirm_addr_indev(in_dev
, dst
, local
, scope
);
1296 for_each_netdev_rcu(net
, dev
) {
1297 in_dev
= __in_dev_get_rcu(dev
);
1299 addr
= confirm_addr_indev(in_dev
, dst
, local
, scope
);
1308 EXPORT_SYMBOL(inet_confirm_addr
);
1314 int register_inetaddr_notifier(struct notifier_block
*nb
)
1316 return blocking_notifier_chain_register(&inetaddr_chain
, nb
);
1318 EXPORT_SYMBOL(register_inetaddr_notifier
);
1320 int unregister_inetaddr_notifier(struct notifier_block
*nb
)
1322 return blocking_notifier_chain_unregister(&inetaddr_chain
, nb
);
1324 EXPORT_SYMBOL(unregister_inetaddr_notifier
);
1326 /* Rename ifa_labels for a device name change. Make some effort to preserve
1327 * existing alias numbering and to create unique labels if possible.
1329 static void inetdev_changename(struct net_device
*dev
, struct in_device
*in_dev
)
1331 struct in_ifaddr
*ifa
;
1334 for (ifa
= in_dev
->ifa_list
; ifa
; ifa
= ifa
->ifa_next
) {
1335 char old
[IFNAMSIZ
], *dot
;
1337 memcpy(old
, ifa
->ifa_label
, IFNAMSIZ
);
1338 memcpy(ifa
->ifa_label
, dev
->name
, IFNAMSIZ
);
1341 dot
= strchr(old
, ':');
1343 sprintf(old
, ":%d", named
);
1346 if (strlen(dot
) + strlen(dev
->name
) < IFNAMSIZ
)
1347 strcat(ifa
->ifa_label
, dot
);
1349 strcpy(ifa
->ifa_label
+ (IFNAMSIZ
- strlen(dot
) - 1), dot
);
1351 rtmsg_ifa(RTM_NEWADDR
, ifa
, NULL
, 0);
1355 static bool inetdev_valid_mtu(unsigned int mtu
)
1360 static void inetdev_send_gratuitous_arp(struct net_device
*dev
,
1361 struct in_device
*in_dev
)
1364 struct in_ifaddr
*ifa
;
1366 for (ifa
= in_dev
->ifa_list
; ifa
;
1367 ifa
= ifa
->ifa_next
) {
1368 arp_send(ARPOP_REQUEST
, ETH_P_ARP
,
1369 ifa
->ifa_local
, dev
,
1370 ifa
->ifa_local
, NULL
,
1371 dev
->dev_addr
, NULL
);
1375 /* Called only under RTNL semaphore */
1377 static int inetdev_event(struct notifier_block
*this, unsigned long event
,
1380 struct net_device
*dev
= netdev_notifier_info_to_dev(ptr
);
1381 struct in_device
*in_dev
= __in_dev_get_rtnl(dev
);
1386 if (event
== NETDEV_REGISTER
) {
1387 in_dev
= inetdev_init(dev
);
1389 return notifier_from_errno(PTR_ERR(in_dev
));
1390 if (dev
->flags
& IFF_LOOPBACK
) {
1391 IN_DEV_CONF_SET(in_dev
, NOXFRM
, 1);
1392 IN_DEV_CONF_SET(in_dev
, NOPOLICY
, 1);
1394 } else if (event
== NETDEV_CHANGEMTU
) {
1395 /* Re-enabling IP */
1396 if (inetdev_valid_mtu(dev
->mtu
))
1397 in_dev
= inetdev_init(dev
);
1403 case NETDEV_REGISTER
:
1404 pr_debug("%s: bug\n", __func__
);
1405 RCU_INIT_POINTER(dev
->ip_ptr
, NULL
);
1408 if (!inetdev_valid_mtu(dev
->mtu
))
1410 if (dev
->flags
& IFF_LOOPBACK
) {
1411 struct in_ifaddr
*ifa
= inet_alloc_ifa();
1414 INIT_HLIST_NODE(&ifa
->hash
);
1416 ifa
->ifa_address
= htonl(INADDR_LOOPBACK
);
1417 ifa
->ifa_prefixlen
= 8;
1418 ifa
->ifa_mask
= inet_make_mask(8);
1419 in_dev_hold(in_dev
);
1420 ifa
->ifa_dev
= in_dev
;
1421 ifa
->ifa_scope
= RT_SCOPE_HOST
;
1422 memcpy(ifa
->ifa_label
, dev
->name
, IFNAMSIZ
);
1423 set_ifa_lifetime(ifa
, INFINITY_LIFE_TIME
,
1424 INFINITY_LIFE_TIME
);
1425 ipv4_devconf_setall(in_dev
);
1426 neigh_parms_data_state_setall(in_dev
->arp_parms
);
1427 inet_insert_ifa(ifa
);
1432 case NETDEV_CHANGEADDR
:
1433 if (!IN_DEV_ARP_NOTIFY(in_dev
))
1436 case NETDEV_NOTIFY_PEERS
:
1437 /* Send gratuitous ARP to notify of link change */
1438 inetdev_send_gratuitous_arp(dev
, in_dev
);
1443 case NETDEV_PRE_TYPE_CHANGE
:
1444 ip_mc_unmap(in_dev
);
1446 case NETDEV_POST_TYPE_CHANGE
:
1447 ip_mc_remap(in_dev
);
1449 case NETDEV_CHANGEMTU
:
1450 if (inetdev_valid_mtu(dev
->mtu
))
1452 /* disable IP when MTU is not enough */
1453 case NETDEV_UNREGISTER
:
1454 inetdev_destroy(in_dev
);
1456 case NETDEV_CHANGENAME
:
1457 /* Do not notify about label change, this event is
1458 * not interesting to applications using netlink.
1460 inetdev_changename(dev
, in_dev
);
1462 devinet_sysctl_unregister(in_dev
);
1463 devinet_sysctl_register(in_dev
);
1470 static struct notifier_block ip_netdev_notifier
= {
1471 .notifier_call
= inetdev_event
,
1474 static size_t inet_nlmsg_size(void)
1476 return NLMSG_ALIGN(sizeof(struct ifaddrmsg
))
1477 + nla_total_size(4) /* IFA_ADDRESS */
1478 + nla_total_size(4) /* IFA_LOCAL */
1479 + nla_total_size(4) /* IFA_BROADCAST */
1480 + nla_total_size(IFNAMSIZ
) /* IFA_LABEL */
1481 + nla_total_size(4) /* IFA_FLAGS */
1482 + nla_total_size(sizeof(struct ifa_cacheinfo
)); /* IFA_CACHEINFO */
1485 static inline u32
cstamp_delta(unsigned long cstamp
)
1487 return (cstamp
- INITIAL_JIFFIES
) * 100UL / HZ
;
1490 static int put_cacheinfo(struct sk_buff
*skb
, unsigned long cstamp
,
1491 unsigned long tstamp
, u32 preferred
, u32 valid
)
1493 struct ifa_cacheinfo ci
;
1495 ci
.cstamp
= cstamp_delta(cstamp
);
1496 ci
.tstamp
= cstamp_delta(tstamp
);
1497 ci
.ifa_prefered
= preferred
;
1498 ci
.ifa_valid
= valid
;
1500 return nla_put(skb
, IFA_CACHEINFO
, sizeof(ci
), &ci
);
1503 static int inet_fill_ifaddr(struct sk_buff
*skb
, struct in_ifaddr
*ifa
,
1504 u32 portid
, u32 seq
, int event
, unsigned int flags
)
1506 struct ifaddrmsg
*ifm
;
1507 struct nlmsghdr
*nlh
;
1508 u32 preferred
, valid
;
1510 nlh
= nlmsg_put(skb
, portid
, seq
, event
, sizeof(*ifm
), flags
);
1514 ifm
= nlmsg_data(nlh
);
1515 ifm
->ifa_family
= AF_INET
;
1516 ifm
->ifa_prefixlen
= ifa
->ifa_prefixlen
;
1517 ifm
->ifa_flags
= ifa
->ifa_flags
;
1518 ifm
->ifa_scope
= ifa
->ifa_scope
;
1519 ifm
->ifa_index
= ifa
->ifa_dev
->dev
->ifindex
;
1521 if (!(ifm
->ifa_flags
& IFA_F_PERMANENT
)) {
1522 preferred
= ifa
->ifa_preferred_lft
;
1523 valid
= ifa
->ifa_valid_lft
;
1524 if (preferred
!= INFINITY_LIFE_TIME
) {
1525 long tval
= (jiffies
- ifa
->ifa_tstamp
) / HZ
;
1527 if (preferred
> tval
)
1531 if (valid
!= INFINITY_LIFE_TIME
) {
1539 preferred
= INFINITY_LIFE_TIME
;
1540 valid
= INFINITY_LIFE_TIME
;
1542 if ((ifa
->ifa_address
&&
1543 nla_put_in_addr(skb
, IFA_ADDRESS
, ifa
->ifa_address
)) ||
1545 nla_put_in_addr(skb
, IFA_LOCAL
, ifa
->ifa_local
)) ||
1546 (ifa
->ifa_broadcast
&&
1547 nla_put_in_addr(skb
, IFA_BROADCAST
, ifa
->ifa_broadcast
)) ||
1548 (ifa
->ifa_label
[0] &&
1549 nla_put_string(skb
, IFA_LABEL
, ifa
->ifa_label
)) ||
1550 nla_put_u32(skb
, IFA_FLAGS
, ifa
->ifa_flags
) ||
1551 put_cacheinfo(skb
, ifa
->ifa_cstamp
, ifa
->ifa_tstamp
,
1553 goto nla_put_failure
;
1555 nlmsg_end(skb
, nlh
);
1559 nlmsg_cancel(skb
, nlh
);
1563 static int inet_dump_ifaddr(struct sk_buff
*skb
, struct netlink_callback
*cb
)
1565 struct net
*net
= sock_net(skb
->sk
);
1568 int ip_idx
, s_ip_idx
;
1569 struct net_device
*dev
;
1570 struct in_device
*in_dev
;
1571 struct in_ifaddr
*ifa
;
1572 struct hlist_head
*head
;
1575 s_idx
= idx
= cb
->args
[1];
1576 s_ip_idx
= ip_idx
= cb
->args
[2];
1578 for (h
= s_h
; h
< NETDEV_HASHENTRIES
; h
++, s_idx
= 0) {
1580 head
= &net
->dev_index_head
[h
];
1582 cb
->seq
= atomic_read(&net
->ipv4
.dev_addr_genid
) ^
1584 hlist_for_each_entry_rcu(dev
, head
, index_hlist
) {
1587 if (h
> s_h
|| idx
> s_idx
)
1589 in_dev
= __in_dev_get_rcu(dev
);
1593 for (ifa
= in_dev
->ifa_list
, ip_idx
= 0; ifa
;
1594 ifa
= ifa
->ifa_next
, ip_idx
++) {
1595 if (ip_idx
< s_ip_idx
)
1597 if (inet_fill_ifaddr(skb
, ifa
,
1598 NETLINK_CB(cb
->skb
).portid
,
1600 RTM_NEWADDR
, NLM_F_MULTI
) < 0) {
1604 nl_dump_check_consistent(cb
, nlmsg_hdr(skb
));
1615 cb
->args
[2] = ip_idx
;
1620 static void rtmsg_ifa(int event
, struct in_ifaddr
*ifa
, struct nlmsghdr
*nlh
,
1623 struct sk_buff
*skb
;
1624 u32 seq
= nlh
? nlh
->nlmsg_seq
: 0;
1628 net
= dev_net(ifa
->ifa_dev
->dev
);
1629 skb
= nlmsg_new(inet_nlmsg_size(), GFP_KERNEL
);
1633 err
= inet_fill_ifaddr(skb
, ifa
, portid
, seq
, event
, 0);
1635 /* -EMSGSIZE implies BUG in inet_nlmsg_size() */
1636 WARN_ON(err
== -EMSGSIZE
);
1640 rtnl_notify(skb
, net
, portid
, RTNLGRP_IPV4_IFADDR
, nlh
, GFP_KERNEL
);
1644 rtnl_set_sk_err(net
, RTNLGRP_IPV4_IFADDR
, err
);
1647 static size_t inet_get_link_af_size(const struct net_device
*dev
,
1648 u32 ext_filter_mask
)
1650 struct in_device
*in_dev
= rcu_dereference_rtnl(dev
->ip_ptr
);
1655 return nla_total_size(IPV4_DEVCONF_MAX
* 4); /* IFLA_INET_CONF */
1658 static int inet_fill_link_af(struct sk_buff
*skb
, const struct net_device
*dev
,
1659 u32 ext_filter_mask
)
1661 struct in_device
*in_dev
= rcu_dereference_rtnl(dev
->ip_ptr
);
1668 nla
= nla_reserve(skb
, IFLA_INET_CONF
, IPV4_DEVCONF_MAX
* 4);
1672 for (i
= 0; i
< IPV4_DEVCONF_MAX
; i
++)
1673 ((u32
*) nla_data(nla
))[i
] = in_dev
->cnf
.data
[i
];
1678 static const struct nla_policy inet_af_policy
[IFLA_INET_MAX
+1] = {
1679 [IFLA_INET_CONF
] = { .type
= NLA_NESTED
},
1682 static int inet_validate_link_af(const struct net_device
*dev
,
1683 const struct nlattr
*nla
)
1685 struct nlattr
*a
, *tb
[IFLA_INET_MAX
+1];
1688 if (dev
&& !__in_dev_get_rtnl(dev
))
1689 return -EAFNOSUPPORT
;
1691 err
= nla_parse_nested(tb
, IFLA_INET_MAX
, nla
, inet_af_policy
);
1695 if (tb
[IFLA_INET_CONF
]) {
1696 nla_for_each_nested(a
, tb
[IFLA_INET_CONF
], rem
) {
1697 int cfgid
= nla_type(a
);
1702 if (cfgid
<= 0 || cfgid
> IPV4_DEVCONF_MAX
)
1710 static int inet_set_link_af(struct net_device
*dev
, const struct nlattr
*nla
)
1712 struct in_device
*in_dev
= __in_dev_get_rtnl(dev
);
1713 struct nlattr
*a
, *tb
[IFLA_INET_MAX
+1];
1717 return -EAFNOSUPPORT
;
1719 if (nla_parse_nested(tb
, IFLA_INET_MAX
, nla
, NULL
) < 0)
1722 if (tb
[IFLA_INET_CONF
]) {
1723 nla_for_each_nested(a
, tb
[IFLA_INET_CONF
], rem
)
1724 ipv4_devconf_set(in_dev
, nla_type(a
), nla_get_u32(a
));
1730 static int inet_netconf_msgsize_devconf(int type
)
1732 int size
= NLMSG_ALIGN(sizeof(struct netconfmsg
))
1733 + nla_total_size(4); /* NETCONFA_IFINDEX */
1735 /* type -1 is used for ALL */
1736 if (type
== -1 || type
== NETCONFA_FORWARDING
)
1737 size
+= nla_total_size(4);
1738 if (type
== -1 || type
== NETCONFA_RP_FILTER
)
1739 size
+= nla_total_size(4);
1740 if (type
== -1 || type
== NETCONFA_MC_FORWARDING
)
1741 size
+= nla_total_size(4);
1742 if (type
== -1 || type
== NETCONFA_PROXY_NEIGH
)
1743 size
+= nla_total_size(4);
1744 if (type
== -1 || type
== NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN
)
1745 size
+= nla_total_size(4);
1750 static int inet_netconf_fill_devconf(struct sk_buff
*skb
, int ifindex
,
1751 struct ipv4_devconf
*devconf
, u32 portid
,
1752 u32 seq
, int event
, unsigned int flags
,
1755 struct nlmsghdr
*nlh
;
1756 struct netconfmsg
*ncm
;
1758 nlh
= nlmsg_put(skb
, portid
, seq
, event
, sizeof(struct netconfmsg
),
1763 ncm
= nlmsg_data(nlh
);
1764 ncm
->ncm_family
= AF_INET
;
1766 if (nla_put_s32(skb
, NETCONFA_IFINDEX
, ifindex
) < 0)
1767 goto nla_put_failure
;
1769 /* type -1 is used for ALL */
1770 if ((type
== -1 || type
== NETCONFA_FORWARDING
) &&
1771 nla_put_s32(skb
, NETCONFA_FORWARDING
,
1772 IPV4_DEVCONF(*devconf
, FORWARDING
)) < 0)
1773 goto nla_put_failure
;
1774 if ((type
== -1 || type
== NETCONFA_RP_FILTER
) &&
1775 nla_put_s32(skb
, NETCONFA_RP_FILTER
,
1776 IPV4_DEVCONF(*devconf
, RP_FILTER
)) < 0)
1777 goto nla_put_failure
;
1778 if ((type
== -1 || type
== NETCONFA_MC_FORWARDING
) &&
1779 nla_put_s32(skb
, NETCONFA_MC_FORWARDING
,
1780 IPV4_DEVCONF(*devconf
, MC_FORWARDING
)) < 0)
1781 goto nla_put_failure
;
1782 if ((type
== -1 || type
== NETCONFA_PROXY_NEIGH
) &&
1783 nla_put_s32(skb
, NETCONFA_PROXY_NEIGH
,
1784 IPV4_DEVCONF(*devconf
, PROXY_ARP
)) < 0)
1785 goto nla_put_failure
;
1786 if ((type
== -1 || type
== NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN
) &&
1787 nla_put_s32(skb
, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN
,
1788 IPV4_DEVCONF(*devconf
, IGNORE_ROUTES_WITH_LINKDOWN
)) < 0)
1789 goto nla_put_failure
;
1791 nlmsg_end(skb
, nlh
);
1795 nlmsg_cancel(skb
, nlh
);
1799 void inet_netconf_notify_devconf(struct net
*net
, int type
, int ifindex
,
1800 struct ipv4_devconf
*devconf
)
1802 struct sk_buff
*skb
;
1805 skb
= nlmsg_new(inet_netconf_msgsize_devconf(type
), GFP_ATOMIC
);
1809 err
= inet_netconf_fill_devconf(skb
, ifindex
, devconf
, 0, 0,
1810 RTM_NEWNETCONF
, 0, type
);
1812 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1813 WARN_ON(err
== -EMSGSIZE
);
1817 rtnl_notify(skb
, net
, 0, RTNLGRP_IPV4_NETCONF
, NULL
, GFP_ATOMIC
);
1821 rtnl_set_sk_err(net
, RTNLGRP_IPV4_NETCONF
, err
);
1824 static const struct nla_policy devconf_ipv4_policy
[NETCONFA_MAX
+1] = {
1825 [NETCONFA_IFINDEX
] = { .len
= sizeof(int) },
1826 [NETCONFA_FORWARDING
] = { .len
= sizeof(int) },
1827 [NETCONFA_RP_FILTER
] = { .len
= sizeof(int) },
1828 [NETCONFA_PROXY_NEIGH
] = { .len
= sizeof(int) },
1829 [NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN
] = { .len
= sizeof(int) },
1832 static int inet_netconf_get_devconf(struct sk_buff
*in_skb
,
1833 struct nlmsghdr
*nlh
)
1835 struct net
*net
= sock_net(in_skb
->sk
);
1836 struct nlattr
*tb
[NETCONFA_MAX
+1];
1837 struct netconfmsg
*ncm
;
1838 struct sk_buff
*skb
;
1839 struct ipv4_devconf
*devconf
;
1840 struct in_device
*in_dev
;
1841 struct net_device
*dev
;
1845 err
= nlmsg_parse(nlh
, sizeof(*ncm
), tb
, NETCONFA_MAX
,
1846 devconf_ipv4_policy
);
1851 if (!tb
[NETCONFA_IFINDEX
])
1854 ifindex
= nla_get_s32(tb
[NETCONFA_IFINDEX
]);
1856 case NETCONFA_IFINDEX_ALL
:
1857 devconf
= net
->ipv4
.devconf_all
;
1859 case NETCONFA_IFINDEX_DEFAULT
:
1860 devconf
= net
->ipv4
.devconf_dflt
;
1863 dev
= __dev_get_by_index(net
, ifindex
);
1866 in_dev
= __in_dev_get_rtnl(dev
);
1869 devconf
= &in_dev
->cnf
;
1874 skb
= nlmsg_new(inet_netconf_msgsize_devconf(-1), GFP_ATOMIC
);
1878 err
= inet_netconf_fill_devconf(skb
, ifindex
, devconf
,
1879 NETLINK_CB(in_skb
).portid
,
1880 nlh
->nlmsg_seq
, RTM_NEWNETCONF
, 0,
1883 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1884 WARN_ON(err
== -EMSGSIZE
);
1888 err
= rtnl_unicast(skb
, net
, NETLINK_CB(in_skb
).portid
);
1893 static int inet_netconf_dump_devconf(struct sk_buff
*skb
,
1894 struct netlink_callback
*cb
)
1896 struct net
*net
= sock_net(skb
->sk
);
1899 struct net_device
*dev
;
1900 struct in_device
*in_dev
;
1901 struct hlist_head
*head
;
1904 s_idx
= idx
= cb
->args
[1];
1906 for (h
= s_h
; h
< NETDEV_HASHENTRIES
; h
++, s_idx
= 0) {
1908 head
= &net
->dev_index_head
[h
];
1910 cb
->seq
= atomic_read(&net
->ipv4
.dev_addr_genid
) ^
1912 hlist_for_each_entry_rcu(dev
, head
, index_hlist
) {
1915 in_dev
= __in_dev_get_rcu(dev
);
1919 if (inet_netconf_fill_devconf(skb
, dev
->ifindex
,
1921 NETLINK_CB(cb
->skb
).portid
,
1929 nl_dump_check_consistent(cb
, nlmsg_hdr(skb
));
1935 if (h
== NETDEV_HASHENTRIES
) {
1936 if (inet_netconf_fill_devconf(skb
, NETCONFA_IFINDEX_ALL
,
1937 net
->ipv4
.devconf_all
,
1938 NETLINK_CB(cb
->skb
).portid
,
1940 RTM_NEWNETCONF
, NLM_F_MULTI
,
1946 if (h
== NETDEV_HASHENTRIES
+ 1) {
1947 if (inet_netconf_fill_devconf(skb
, NETCONFA_IFINDEX_DEFAULT
,
1948 net
->ipv4
.devconf_dflt
,
1949 NETLINK_CB(cb
->skb
).portid
,
1951 RTM_NEWNETCONF
, NLM_F_MULTI
,
1964 #ifdef CONFIG_SYSCTL
1966 static void devinet_copy_dflt_conf(struct net
*net
, int i
)
1968 struct net_device
*dev
;
1971 for_each_netdev_rcu(net
, dev
) {
1972 struct in_device
*in_dev
;
1974 in_dev
= __in_dev_get_rcu(dev
);
1975 if (in_dev
&& !test_bit(i
, in_dev
->cnf
.state
))
1976 in_dev
->cnf
.data
[i
] = net
->ipv4
.devconf_dflt
->data
[i
];
1981 /* called with RTNL locked */
1982 static void inet_forward_change(struct net
*net
)
1984 struct net_device
*dev
;
1985 int on
= IPV4_DEVCONF_ALL(net
, FORWARDING
);
1987 IPV4_DEVCONF_ALL(net
, ACCEPT_REDIRECTS
) = !on
;
1988 IPV4_DEVCONF_DFLT(net
, FORWARDING
) = on
;
1989 inet_netconf_notify_devconf(net
, NETCONFA_FORWARDING
,
1990 NETCONFA_IFINDEX_ALL
,
1991 net
->ipv4
.devconf_all
);
1992 inet_netconf_notify_devconf(net
, NETCONFA_FORWARDING
,
1993 NETCONFA_IFINDEX_DEFAULT
,
1994 net
->ipv4
.devconf_dflt
);
1996 for_each_netdev(net
, dev
) {
1997 struct in_device
*in_dev
;
1999 dev_disable_lro(dev
);
2001 in_dev
= __in_dev_get_rcu(dev
);
2003 IN_DEV_CONF_SET(in_dev
, FORWARDING
, on
);
2004 inet_netconf_notify_devconf(net
, NETCONFA_FORWARDING
,
2005 dev
->ifindex
, &in_dev
->cnf
);
2011 static int devinet_conf_ifindex(struct net
*net
, struct ipv4_devconf
*cnf
)
2013 if (cnf
== net
->ipv4
.devconf_dflt
)
2014 return NETCONFA_IFINDEX_DEFAULT
;
2015 else if (cnf
== net
->ipv4
.devconf_all
)
2016 return NETCONFA_IFINDEX_ALL
;
2018 struct in_device
*idev
2019 = container_of(cnf
, struct in_device
, cnf
);
2020 return idev
->dev
->ifindex
;
2024 static int devinet_conf_proc(struct ctl_table
*ctl
, int write
,
2025 void __user
*buffer
,
2026 size_t *lenp
, loff_t
*ppos
)
2028 int old_value
= *(int *)ctl
->data
;
2029 int ret
= proc_dointvec(ctl
, write
, buffer
, lenp
, ppos
);
2030 int new_value
= *(int *)ctl
->data
;
2033 struct ipv4_devconf
*cnf
= ctl
->extra1
;
2034 struct net
*net
= ctl
->extra2
;
2035 int i
= (int *)ctl
->data
- cnf
->data
;
2038 set_bit(i
, cnf
->state
);
2040 if (cnf
== net
->ipv4
.devconf_dflt
)
2041 devinet_copy_dflt_conf(net
, i
);
2042 if (i
== IPV4_DEVCONF_ACCEPT_LOCAL
- 1 ||
2043 i
== IPV4_DEVCONF_ROUTE_LOCALNET
- 1)
2044 if ((new_value
== 0) && (old_value
!= 0))
2045 rt_cache_flush(net
);
2047 if (i
== IPV4_DEVCONF_RP_FILTER
- 1 &&
2048 new_value
!= old_value
) {
2049 ifindex
= devinet_conf_ifindex(net
, cnf
);
2050 inet_netconf_notify_devconf(net
, NETCONFA_RP_FILTER
,
2053 if (i
== IPV4_DEVCONF_PROXY_ARP
- 1 &&
2054 new_value
!= old_value
) {
2055 ifindex
= devinet_conf_ifindex(net
, cnf
);
2056 inet_netconf_notify_devconf(net
, NETCONFA_PROXY_NEIGH
,
2059 if (i
== IPV4_DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN
- 1 &&
2060 new_value
!= old_value
) {
2061 ifindex
= devinet_conf_ifindex(net
, cnf
);
2062 inet_netconf_notify_devconf(net
, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN
,
2070 static int devinet_sysctl_forward(struct ctl_table
*ctl
, int write
,
2071 void __user
*buffer
,
2072 size_t *lenp
, loff_t
*ppos
)
2074 int *valp
= ctl
->data
;
2077 int ret
= proc_dointvec(ctl
, write
, buffer
, lenp
, ppos
);
2079 if (write
&& *valp
!= val
) {
2080 struct net
*net
= ctl
->extra2
;
2082 if (valp
!= &IPV4_DEVCONF_DFLT(net
, FORWARDING
)) {
2083 if (!rtnl_trylock()) {
2084 /* Restore the original values before restarting */
2087 return restart_syscall();
2089 if (valp
== &IPV4_DEVCONF_ALL(net
, FORWARDING
)) {
2090 inet_forward_change(net
);
2092 struct ipv4_devconf
*cnf
= ctl
->extra1
;
2093 struct in_device
*idev
=
2094 container_of(cnf
, struct in_device
, cnf
);
2096 dev_disable_lro(idev
->dev
);
2097 inet_netconf_notify_devconf(net
,
2098 NETCONFA_FORWARDING
,
2103 rt_cache_flush(net
);
2105 inet_netconf_notify_devconf(net
, NETCONFA_FORWARDING
,
2106 NETCONFA_IFINDEX_DEFAULT
,
2107 net
->ipv4
.devconf_dflt
);
2113 static int ipv4_doint_and_flush(struct ctl_table
*ctl
, int write
,
2114 void __user
*buffer
,
2115 size_t *lenp
, loff_t
*ppos
)
2117 int *valp
= ctl
->data
;
2119 int ret
= proc_dointvec(ctl
, write
, buffer
, lenp
, ppos
);
2120 struct net
*net
= ctl
->extra2
;
2122 if (write
&& *valp
!= val
)
2123 rt_cache_flush(net
);
2128 #define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \
2131 .data = ipv4_devconf.data + \
2132 IPV4_DEVCONF_ ## attr - 1, \
2133 .maxlen = sizeof(int), \
2135 .proc_handler = proc, \
2136 .extra1 = &ipv4_devconf, \
2139 #define DEVINET_SYSCTL_RW_ENTRY(attr, name) \
2140 DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc)
2142 #define DEVINET_SYSCTL_RO_ENTRY(attr, name) \
2143 DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc)
2145 #define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \
2146 DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc)
2148 #define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \
2149 DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush)
2151 static struct devinet_sysctl_table
{
2152 struct ctl_table_header
*sysctl_header
;
2153 struct ctl_table devinet_vars
[__IPV4_DEVCONF_MAX
];
2154 } devinet_sysctl
= {
2156 DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING
, "forwarding",
2157 devinet_sysctl_forward
),
2158 DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING
, "mc_forwarding"),
2160 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS
, "accept_redirects"),
2161 DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS
, "secure_redirects"),
2162 DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA
, "shared_media"),
2163 DEVINET_SYSCTL_RW_ENTRY(RP_FILTER
, "rp_filter"),
2164 DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS
, "send_redirects"),
2165 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE
,
2166 "accept_source_route"),
2167 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL
, "accept_local"),
2168 DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK
, "src_valid_mark"),
2169 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP
, "proxy_arp"),
2170 DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID
, "medium_id"),
2171 DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY
, "bootp_relay"),
2172 DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS
, "log_martians"),
2173 DEVINET_SYSCTL_RW_ENTRY(TAG
, "tag"),
2174 DEVINET_SYSCTL_RW_ENTRY(ARPFILTER
, "arp_filter"),
2175 DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE
, "arp_announce"),
2176 DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE
, "arp_ignore"),
2177 DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT
, "arp_accept"),
2178 DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY
, "arp_notify"),
2179 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN
, "proxy_arp_pvlan"),
2180 DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION
,
2181 "force_igmp_version"),
2182 DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL
,
2183 "igmpv2_unsolicited_report_interval"),
2184 DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL
,
2185 "igmpv3_unsolicited_report_interval"),
2186 DEVINET_SYSCTL_RW_ENTRY(IGNORE_ROUTES_WITH_LINKDOWN
,
2187 "ignore_routes_with_linkdown"),
2189 DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM
, "disable_xfrm"),
2190 DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY
, "disable_policy"),
2191 DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES
,
2192 "promote_secondaries"),
2193 DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET
,
2198 static int __devinet_sysctl_register(struct net
*net
, char *dev_name
,
2199 struct ipv4_devconf
*p
)
2202 struct devinet_sysctl_table
*t
;
2203 char path
[sizeof("net/ipv4/conf/") + IFNAMSIZ
];
2205 t
= kmemdup(&devinet_sysctl
, sizeof(*t
), GFP_KERNEL
);
2209 for (i
= 0; i
< ARRAY_SIZE(t
->devinet_vars
) - 1; i
++) {
2210 t
->devinet_vars
[i
].data
+= (char *)p
- (char *)&ipv4_devconf
;
2211 t
->devinet_vars
[i
].extra1
= p
;
2212 t
->devinet_vars
[i
].extra2
= net
;
2215 snprintf(path
, sizeof(path
), "net/ipv4/conf/%s", dev_name
);
2217 t
->sysctl_header
= register_net_sysctl(net
, path
, t
->devinet_vars
);
2218 if (!t
->sysctl_header
)
2230 static void __devinet_sysctl_unregister(struct ipv4_devconf
*cnf
)
2232 struct devinet_sysctl_table
*t
= cnf
->sysctl
;
2238 unregister_net_sysctl_table(t
->sysctl_header
);
2242 static int devinet_sysctl_register(struct in_device
*idev
)
2246 if (!sysctl_dev_name_is_allowed(idev
->dev
->name
))
2249 err
= neigh_sysctl_register(idev
->dev
, idev
->arp_parms
, NULL
);
2252 err
= __devinet_sysctl_register(dev_net(idev
->dev
), idev
->dev
->name
,
2255 neigh_sysctl_unregister(idev
->arp_parms
);
2259 static void devinet_sysctl_unregister(struct in_device
*idev
)
2261 __devinet_sysctl_unregister(&idev
->cnf
);
2262 neigh_sysctl_unregister(idev
->arp_parms
);
2265 static struct ctl_table ctl_forward_entry
[] = {
2267 .procname
= "ip_forward",
2268 .data
= &ipv4_devconf
.data
[
2269 IPV4_DEVCONF_FORWARDING
- 1],
2270 .maxlen
= sizeof(int),
2272 .proc_handler
= devinet_sysctl_forward
,
2273 .extra1
= &ipv4_devconf
,
2274 .extra2
= &init_net
,
2280 static __net_init
int devinet_init_net(struct net
*net
)
2283 struct ipv4_devconf
*all
, *dflt
;
2284 #ifdef CONFIG_SYSCTL
2285 struct ctl_table
*tbl
= ctl_forward_entry
;
2286 struct ctl_table_header
*forw_hdr
;
2290 all
= &ipv4_devconf
;
2291 dflt
= &ipv4_devconf_dflt
;
2293 if (!net_eq(net
, &init_net
)) {
2294 all
= kmemdup(all
, sizeof(ipv4_devconf
), GFP_KERNEL
);
2298 dflt
= kmemdup(dflt
, sizeof(ipv4_devconf_dflt
), GFP_KERNEL
);
2300 goto err_alloc_dflt
;
2302 #ifdef CONFIG_SYSCTL
2303 tbl
= kmemdup(tbl
, sizeof(ctl_forward_entry
), GFP_KERNEL
);
2307 tbl
[0].data
= &all
->data
[IPV4_DEVCONF_FORWARDING
- 1];
2308 tbl
[0].extra1
= all
;
2309 tbl
[0].extra2
= net
;
2313 #ifdef CONFIG_SYSCTL
2314 err
= __devinet_sysctl_register(net
, "all", all
);
2318 err
= __devinet_sysctl_register(net
, "default", dflt
);
2323 forw_hdr
= register_net_sysctl(net
, "net/ipv4", tbl
);
2326 net
->ipv4
.forw_hdr
= forw_hdr
;
2329 net
->ipv4
.devconf_all
= all
;
2330 net
->ipv4
.devconf_dflt
= dflt
;
2333 #ifdef CONFIG_SYSCTL
2335 __devinet_sysctl_unregister(dflt
);
2337 __devinet_sysctl_unregister(all
);
2339 if (tbl
!= ctl_forward_entry
)
2343 if (dflt
!= &ipv4_devconf_dflt
)
2346 if (all
!= &ipv4_devconf
)
2352 static __net_exit
void devinet_exit_net(struct net
*net
)
2354 #ifdef CONFIG_SYSCTL
2355 struct ctl_table
*tbl
;
2357 tbl
= net
->ipv4
.forw_hdr
->ctl_table_arg
;
2358 unregister_net_sysctl_table(net
->ipv4
.forw_hdr
);
2359 __devinet_sysctl_unregister(net
->ipv4
.devconf_dflt
);
2360 __devinet_sysctl_unregister(net
->ipv4
.devconf_all
);
2363 kfree(net
->ipv4
.devconf_dflt
);
2364 kfree(net
->ipv4
.devconf_all
);
2367 static __net_initdata
struct pernet_operations devinet_ops
= {
2368 .init
= devinet_init_net
,
2369 .exit
= devinet_exit_net
,
2372 static struct rtnl_af_ops inet_af_ops __read_mostly
= {
2374 .fill_link_af
= inet_fill_link_af
,
2375 .get_link_af_size
= inet_get_link_af_size
,
2376 .validate_link_af
= inet_validate_link_af
,
2377 .set_link_af
= inet_set_link_af
,
2380 void __init
devinet_init(void)
2384 for (i
= 0; i
< IN4_ADDR_HSIZE
; i
++)
2385 INIT_HLIST_HEAD(&inet_addr_lst
[i
]);
2387 register_pernet_subsys(&devinet_ops
);
2389 register_gifconf(PF_INET
, inet_gifconf
);
2390 register_netdevice_notifier(&ip_netdev_notifier
);
2392 queue_delayed_work(system_power_efficient_wq
, &check_lifetime_work
, 0);
2394 rtnl_af_register(&inet_af_ops
);
2396 rtnl_register(PF_INET
, RTM_NEWADDR
, inet_rtm_newaddr
, NULL
, NULL
);
2397 rtnl_register(PF_INET
, RTM_DELADDR
, inet_rtm_deladdr
, NULL
, NULL
);
2398 rtnl_register(PF_INET
, RTM_GETADDR
, NULL
, inet_dump_ifaddr
, NULL
);
2399 rtnl_register(PF_INET
, RTM_GETNETCONF
, inet_netconf_get_devconf
,
2400 inet_netconf_dump_devconf
, NULL
);