ARM64: juno: disable NOR flash node by default
[linux/fpc-iii.git] / net / ipv4 / devinet.c
blob2d9cb1748f8191c785567632faf0ee14eaca628b
1 /*
2 * NET3 IP device support routines.
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
9 * Derived from the IP parts of dev.c 1.0.19
10 * Authors: Ross Biro
11 * Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
12 * Mark Evans, <evansmp@uhura.aston.ac.uk>
14 * Additional Authors:
15 * Alan Cox, <gw4pts@gw4pts.ampr.org>
16 * Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
18 * Changes:
19 * Alexey Kuznetsov: pa_* fields are replaced with ifaddr
20 * lists.
21 * Cyrus Durgin: updated for kmod
22 * Matthias Andree: in devinet_ioctl, compare label and
23 * address (4.4BSD alias style support),
24 * fall back to comparing just the label
25 * if no match found.
29 #include <asm/uaccess.h>
30 #include <linux/bitops.h>
31 #include <linux/capability.h>
32 #include <linux/module.h>
33 #include <linux/types.h>
34 #include <linux/kernel.h>
35 #include <linux/string.h>
36 #include <linux/mm.h>
37 #include <linux/socket.h>
38 #include <linux/sockios.h>
39 #include <linux/in.h>
40 #include <linux/errno.h>
41 #include <linux/interrupt.h>
42 #include <linux/if_addr.h>
43 #include <linux/if_ether.h>
44 #include <linux/inet.h>
45 #include <linux/netdevice.h>
46 #include <linux/etherdevice.h>
47 #include <linux/skbuff.h>
48 #include <linux/init.h>
49 #include <linux/notifier.h>
50 #include <linux/inetdevice.h>
51 #include <linux/igmp.h>
52 #include <linux/slab.h>
53 #include <linux/hash.h>
54 #ifdef CONFIG_SYSCTL
55 #include <linux/sysctl.h>
56 #endif
57 #include <linux/kmod.h>
58 #include <linux/netconf.h>
60 #include <net/arp.h>
61 #include <net/ip.h>
62 #include <net/route.h>
63 #include <net/ip_fib.h>
64 #include <net/rtnetlink.h>
65 #include <net/net_namespace.h>
66 #include <net/addrconf.h>
68 #include "fib_lookup.h"
70 static struct ipv4_devconf ipv4_devconf = {
71 .data = {
72 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
73 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
74 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
75 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
76 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
77 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/,
81 static struct ipv4_devconf ipv4_devconf_dflt = {
82 .data = {
83 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
84 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
85 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
86 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
87 [IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE - 1] = 1,
88 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
89 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/,
93 #define IPV4_DEVCONF_DFLT(net, attr) \
94 IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr)
96 static const struct nla_policy ifa_ipv4_policy[IFA_MAX+1] = {
97 [IFA_LOCAL] = { .type = NLA_U32 },
98 [IFA_ADDRESS] = { .type = NLA_U32 },
99 [IFA_BROADCAST] = { .type = NLA_U32 },
100 [IFA_LABEL] = { .type = NLA_STRING, .len = IFNAMSIZ - 1 },
101 [IFA_CACHEINFO] = { .len = sizeof(struct ifa_cacheinfo) },
102 [IFA_FLAGS] = { .type = NLA_U32 },
105 #define IN4_ADDR_HSIZE_SHIFT 8
106 #define IN4_ADDR_HSIZE (1U << IN4_ADDR_HSIZE_SHIFT)
108 static struct hlist_head inet_addr_lst[IN4_ADDR_HSIZE];
110 static u32 inet_addr_hash(const struct net *net, __be32 addr)
112 u32 val = (__force u32) addr ^ net_hash_mix(net);
114 return hash_32(val, IN4_ADDR_HSIZE_SHIFT);
117 static void inet_hash_insert(struct net *net, struct in_ifaddr *ifa)
119 u32 hash = inet_addr_hash(net, ifa->ifa_local);
121 ASSERT_RTNL();
122 hlist_add_head_rcu(&ifa->hash, &inet_addr_lst[hash]);
125 static void inet_hash_remove(struct in_ifaddr *ifa)
127 ASSERT_RTNL();
128 hlist_del_init_rcu(&ifa->hash);
132 * __ip_dev_find - find the first device with a given source address.
133 * @net: the net namespace
134 * @addr: the source address
135 * @devref: if true, take a reference on the found device
137 * If a caller uses devref=false, it should be protected by RCU, or RTNL
139 struct net_device *__ip_dev_find(struct net *net, __be32 addr, bool devref)
141 u32 hash = inet_addr_hash(net, addr);
142 struct net_device *result = NULL;
143 struct in_ifaddr *ifa;
145 rcu_read_lock();
146 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[hash], hash) {
147 if (ifa->ifa_local == addr) {
148 struct net_device *dev = ifa->ifa_dev->dev;
150 if (!net_eq(dev_net(dev), net))
151 continue;
152 result = dev;
153 break;
156 if (!result) {
157 struct flowi4 fl4 = { .daddr = addr };
158 struct fib_result res = { 0 };
159 struct fib_table *local;
161 /* Fallback to FIB local table so that communication
162 * over loopback subnets work.
164 local = fib_get_table(net, RT_TABLE_LOCAL);
165 if (local &&
166 !fib_table_lookup(local, &fl4, &res, FIB_LOOKUP_NOREF) &&
167 res.type == RTN_LOCAL)
168 result = FIB_RES_DEV(res);
170 if (result && devref)
171 dev_hold(result);
172 rcu_read_unlock();
173 return result;
175 EXPORT_SYMBOL(__ip_dev_find);
177 static void rtmsg_ifa(int event, struct in_ifaddr *, struct nlmsghdr *, u32);
179 static BLOCKING_NOTIFIER_HEAD(inetaddr_chain);
180 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
181 int destroy);
182 #ifdef CONFIG_SYSCTL
183 static int devinet_sysctl_register(struct in_device *idev);
184 static void devinet_sysctl_unregister(struct in_device *idev);
185 #else
186 static int devinet_sysctl_register(struct in_device *idev)
188 return 0;
190 static void devinet_sysctl_unregister(struct in_device *idev)
193 #endif
195 /* Locks all the inet devices. */
197 static struct in_ifaddr *inet_alloc_ifa(void)
199 return kzalloc(sizeof(struct in_ifaddr), GFP_KERNEL);
202 static void inet_rcu_free_ifa(struct rcu_head *head)
204 struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head);
205 if (ifa->ifa_dev)
206 in_dev_put(ifa->ifa_dev);
207 kfree(ifa);
210 static void inet_free_ifa(struct in_ifaddr *ifa)
212 call_rcu(&ifa->rcu_head, inet_rcu_free_ifa);
215 void in_dev_finish_destroy(struct in_device *idev)
217 struct net_device *dev = idev->dev;
219 WARN_ON(idev->ifa_list);
220 WARN_ON(idev->mc_list);
221 kfree(rcu_dereference_protected(idev->mc_hash, 1));
222 #ifdef NET_REFCNT_DEBUG
223 pr_debug("%s: %p=%s\n", __func__, idev, dev ? dev->name : "NIL");
224 #endif
225 dev_put(dev);
226 if (!idev->dead)
227 pr_err("Freeing alive in_device %p\n", idev);
228 else
229 kfree(idev);
231 EXPORT_SYMBOL(in_dev_finish_destroy);
233 static struct in_device *inetdev_init(struct net_device *dev)
235 struct in_device *in_dev;
236 int err = -ENOMEM;
238 ASSERT_RTNL();
240 in_dev = kzalloc(sizeof(*in_dev), GFP_KERNEL);
241 if (!in_dev)
242 goto out;
243 memcpy(&in_dev->cnf, dev_net(dev)->ipv4.devconf_dflt,
244 sizeof(in_dev->cnf));
245 in_dev->cnf.sysctl = NULL;
246 in_dev->dev = dev;
247 in_dev->arp_parms = neigh_parms_alloc(dev, &arp_tbl);
248 if (!in_dev->arp_parms)
249 goto out_kfree;
250 if (IPV4_DEVCONF(in_dev->cnf, FORWARDING))
251 dev_disable_lro(dev);
252 /* Reference in_dev->dev */
253 dev_hold(dev);
254 /* Account for reference dev->ip_ptr (below) */
255 in_dev_hold(in_dev);
257 err = devinet_sysctl_register(in_dev);
258 if (err) {
259 in_dev->dead = 1;
260 in_dev_put(in_dev);
261 in_dev = NULL;
262 goto out;
264 ip_mc_init_dev(in_dev);
265 if (dev->flags & IFF_UP)
266 ip_mc_up(in_dev);
268 /* we can receive as soon as ip_ptr is set -- do this last */
269 rcu_assign_pointer(dev->ip_ptr, in_dev);
270 out:
271 return in_dev ?: ERR_PTR(err);
272 out_kfree:
273 kfree(in_dev);
274 in_dev = NULL;
275 goto out;
278 static void in_dev_rcu_put(struct rcu_head *head)
280 struct in_device *idev = container_of(head, struct in_device, rcu_head);
281 in_dev_put(idev);
284 static void inetdev_destroy(struct in_device *in_dev)
286 struct in_ifaddr *ifa;
287 struct net_device *dev;
289 ASSERT_RTNL();
291 dev = in_dev->dev;
293 in_dev->dead = 1;
295 ip_mc_destroy_dev(in_dev);
297 while ((ifa = in_dev->ifa_list) != NULL) {
298 inet_del_ifa(in_dev, &in_dev->ifa_list, 0);
299 inet_free_ifa(ifa);
302 RCU_INIT_POINTER(dev->ip_ptr, NULL);
304 devinet_sysctl_unregister(in_dev);
305 neigh_parms_release(&arp_tbl, in_dev->arp_parms);
306 arp_ifdown(dev);
308 call_rcu(&in_dev->rcu_head, in_dev_rcu_put);
311 int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b)
313 rcu_read_lock();
314 for_primary_ifa(in_dev) {
315 if (inet_ifa_match(a, ifa)) {
316 if (!b || inet_ifa_match(b, ifa)) {
317 rcu_read_unlock();
318 return 1;
321 } endfor_ifa(in_dev);
322 rcu_read_unlock();
323 return 0;
326 static void __inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
327 int destroy, struct nlmsghdr *nlh, u32 portid)
329 struct in_ifaddr *promote = NULL;
330 struct in_ifaddr *ifa, *ifa1 = *ifap;
331 struct in_ifaddr *last_prim = in_dev->ifa_list;
332 struct in_ifaddr *prev_prom = NULL;
333 int do_promote = IN_DEV_PROMOTE_SECONDARIES(in_dev);
335 ASSERT_RTNL();
337 /* 1. Deleting primary ifaddr forces deletion all secondaries
338 * unless alias promotion is set
341 if (!(ifa1->ifa_flags & IFA_F_SECONDARY)) {
342 struct in_ifaddr **ifap1 = &ifa1->ifa_next;
344 while ((ifa = *ifap1) != NULL) {
345 if (!(ifa->ifa_flags & IFA_F_SECONDARY) &&
346 ifa1->ifa_scope <= ifa->ifa_scope)
347 last_prim = ifa;
349 if (!(ifa->ifa_flags & IFA_F_SECONDARY) ||
350 ifa1->ifa_mask != ifa->ifa_mask ||
351 !inet_ifa_match(ifa1->ifa_address, ifa)) {
352 ifap1 = &ifa->ifa_next;
353 prev_prom = ifa;
354 continue;
357 if (!do_promote) {
358 inet_hash_remove(ifa);
359 *ifap1 = ifa->ifa_next;
361 rtmsg_ifa(RTM_DELADDR, ifa, nlh, portid);
362 blocking_notifier_call_chain(&inetaddr_chain,
363 NETDEV_DOWN, ifa);
364 inet_free_ifa(ifa);
365 } else {
366 promote = ifa;
367 break;
372 /* On promotion all secondaries from subnet are changing
373 * the primary IP, we must remove all their routes silently
374 * and later to add them back with new prefsrc. Do this
375 * while all addresses are on the device list.
377 for (ifa = promote; ifa; ifa = ifa->ifa_next) {
378 if (ifa1->ifa_mask == ifa->ifa_mask &&
379 inet_ifa_match(ifa1->ifa_address, ifa))
380 fib_del_ifaddr(ifa, ifa1);
383 /* 2. Unlink it */
385 *ifap = ifa1->ifa_next;
386 inet_hash_remove(ifa1);
388 /* 3. Announce address deletion */
390 /* Send message first, then call notifier.
391 At first sight, FIB update triggered by notifier
392 will refer to already deleted ifaddr, that could confuse
393 netlink listeners. It is not true: look, gated sees
394 that route deleted and if it still thinks that ifaddr
395 is valid, it will try to restore deleted routes... Grr.
396 So that, this order is correct.
398 rtmsg_ifa(RTM_DELADDR, ifa1, nlh, portid);
399 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_DOWN, ifa1);
401 if (promote) {
402 struct in_ifaddr *next_sec = promote->ifa_next;
404 if (prev_prom) {
405 prev_prom->ifa_next = promote->ifa_next;
406 promote->ifa_next = last_prim->ifa_next;
407 last_prim->ifa_next = promote;
410 promote->ifa_flags &= ~IFA_F_SECONDARY;
411 rtmsg_ifa(RTM_NEWADDR, promote, nlh, portid);
412 blocking_notifier_call_chain(&inetaddr_chain,
413 NETDEV_UP, promote);
414 for (ifa = next_sec; ifa; ifa = ifa->ifa_next) {
415 if (ifa1->ifa_mask != ifa->ifa_mask ||
416 !inet_ifa_match(ifa1->ifa_address, ifa))
417 continue;
418 fib_add_ifaddr(ifa);
422 if (destroy)
423 inet_free_ifa(ifa1);
426 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
427 int destroy)
429 __inet_del_ifa(in_dev, ifap, destroy, NULL, 0);
432 static void check_lifetime(struct work_struct *work);
434 static DECLARE_DELAYED_WORK(check_lifetime_work, check_lifetime);
436 static int __inet_insert_ifa(struct in_ifaddr *ifa, struct nlmsghdr *nlh,
437 u32 portid)
439 struct in_device *in_dev = ifa->ifa_dev;
440 struct in_ifaddr *ifa1, **ifap, **last_primary;
442 ASSERT_RTNL();
444 if (!ifa->ifa_local) {
445 inet_free_ifa(ifa);
446 return 0;
449 ifa->ifa_flags &= ~IFA_F_SECONDARY;
450 last_primary = &in_dev->ifa_list;
452 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
453 ifap = &ifa1->ifa_next) {
454 if (!(ifa1->ifa_flags & IFA_F_SECONDARY) &&
455 ifa->ifa_scope <= ifa1->ifa_scope)
456 last_primary = &ifa1->ifa_next;
457 if (ifa1->ifa_mask == ifa->ifa_mask &&
458 inet_ifa_match(ifa1->ifa_address, ifa)) {
459 if (ifa1->ifa_local == ifa->ifa_local) {
460 inet_free_ifa(ifa);
461 return -EEXIST;
463 if (ifa1->ifa_scope != ifa->ifa_scope) {
464 inet_free_ifa(ifa);
465 return -EINVAL;
467 ifa->ifa_flags |= IFA_F_SECONDARY;
471 if (!(ifa->ifa_flags & IFA_F_SECONDARY)) {
472 prandom_seed((__force u32) ifa->ifa_local);
473 ifap = last_primary;
476 ifa->ifa_next = *ifap;
477 *ifap = ifa;
479 inet_hash_insert(dev_net(in_dev->dev), ifa);
481 cancel_delayed_work(&check_lifetime_work);
482 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
484 /* Send message first, then call notifier.
485 Notifier will trigger FIB update, so that
486 listeners of netlink will know about new ifaddr */
487 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, portid);
488 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa);
490 return 0;
493 static int inet_insert_ifa(struct in_ifaddr *ifa)
495 return __inet_insert_ifa(ifa, NULL, 0);
498 static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa)
500 struct in_device *in_dev = __in_dev_get_rtnl(dev);
502 ASSERT_RTNL();
504 if (!in_dev) {
505 inet_free_ifa(ifa);
506 return -ENOBUFS;
508 ipv4_devconf_setall(in_dev);
509 neigh_parms_data_state_setall(in_dev->arp_parms);
510 if (ifa->ifa_dev != in_dev) {
511 WARN_ON(ifa->ifa_dev);
512 in_dev_hold(in_dev);
513 ifa->ifa_dev = in_dev;
515 if (ipv4_is_loopback(ifa->ifa_local))
516 ifa->ifa_scope = RT_SCOPE_HOST;
517 return inet_insert_ifa(ifa);
520 /* Caller must hold RCU or RTNL :
521 * We dont take a reference on found in_device
523 struct in_device *inetdev_by_index(struct net *net, int ifindex)
525 struct net_device *dev;
526 struct in_device *in_dev = NULL;
528 rcu_read_lock();
529 dev = dev_get_by_index_rcu(net, ifindex);
530 if (dev)
531 in_dev = rcu_dereference_rtnl(dev->ip_ptr);
532 rcu_read_unlock();
533 return in_dev;
535 EXPORT_SYMBOL(inetdev_by_index);
537 /* Called only from RTNL semaphored context. No locks. */
539 struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix,
540 __be32 mask)
542 ASSERT_RTNL();
544 for_primary_ifa(in_dev) {
545 if (ifa->ifa_mask == mask && inet_ifa_match(prefix, ifa))
546 return ifa;
547 } endfor_ifa(in_dev);
548 return NULL;
551 static int ip_mc_config(struct sock *sk, bool join, const struct in_ifaddr *ifa)
553 struct ip_mreqn mreq = {
554 .imr_multiaddr.s_addr = ifa->ifa_address,
555 .imr_ifindex = ifa->ifa_dev->dev->ifindex,
557 int ret;
559 ASSERT_RTNL();
561 lock_sock(sk);
562 if (join)
563 ret = ip_mc_join_group(sk, &mreq);
564 else
565 ret = ip_mc_leave_group(sk, &mreq);
566 release_sock(sk);
568 return ret;
571 static int inet_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh)
573 struct net *net = sock_net(skb->sk);
574 struct nlattr *tb[IFA_MAX+1];
575 struct in_device *in_dev;
576 struct ifaddrmsg *ifm;
577 struct in_ifaddr *ifa, **ifap;
578 int err = -EINVAL;
580 ASSERT_RTNL();
582 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
583 if (err < 0)
584 goto errout;
586 ifm = nlmsg_data(nlh);
587 in_dev = inetdev_by_index(net, ifm->ifa_index);
588 if (!in_dev) {
589 err = -ENODEV;
590 goto errout;
593 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
594 ifap = &ifa->ifa_next) {
595 if (tb[IFA_LOCAL] &&
596 ifa->ifa_local != nla_get_in_addr(tb[IFA_LOCAL]))
597 continue;
599 if (tb[IFA_LABEL] && nla_strcmp(tb[IFA_LABEL], ifa->ifa_label))
600 continue;
602 if (tb[IFA_ADDRESS] &&
603 (ifm->ifa_prefixlen != ifa->ifa_prefixlen ||
604 !inet_ifa_match(nla_get_in_addr(tb[IFA_ADDRESS]), ifa)))
605 continue;
607 if (ipv4_is_multicast(ifa->ifa_address))
608 ip_mc_config(net->ipv4.mc_autojoin_sk, false, ifa);
609 __inet_del_ifa(in_dev, ifap, 1, nlh, NETLINK_CB(skb).portid);
610 return 0;
613 err = -EADDRNOTAVAIL;
614 errout:
615 return err;
618 #define INFINITY_LIFE_TIME 0xFFFFFFFF
620 static void check_lifetime(struct work_struct *work)
622 unsigned long now, next, next_sec, next_sched;
623 struct in_ifaddr *ifa;
624 struct hlist_node *n;
625 int i;
627 now = jiffies;
628 next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
630 for (i = 0; i < IN4_ADDR_HSIZE; i++) {
631 bool change_needed = false;
633 rcu_read_lock();
634 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[i], hash) {
635 unsigned long age;
637 if (ifa->ifa_flags & IFA_F_PERMANENT)
638 continue;
640 /* We try to batch several events at once. */
641 age = (now - ifa->ifa_tstamp +
642 ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
644 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
645 age >= ifa->ifa_valid_lft) {
646 change_needed = true;
647 } else if (ifa->ifa_preferred_lft ==
648 INFINITY_LIFE_TIME) {
649 continue;
650 } else if (age >= ifa->ifa_preferred_lft) {
651 if (time_before(ifa->ifa_tstamp +
652 ifa->ifa_valid_lft * HZ, next))
653 next = ifa->ifa_tstamp +
654 ifa->ifa_valid_lft * HZ;
656 if (!(ifa->ifa_flags & IFA_F_DEPRECATED))
657 change_needed = true;
658 } else if (time_before(ifa->ifa_tstamp +
659 ifa->ifa_preferred_lft * HZ,
660 next)) {
661 next = ifa->ifa_tstamp +
662 ifa->ifa_preferred_lft * HZ;
665 rcu_read_unlock();
666 if (!change_needed)
667 continue;
668 rtnl_lock();
669 hlist_for_each_entry_safe(ifa, n, &inet_addr_lst[i], hash) {
670 unsigned long age;
672 if (ifa->ifa_flags & IFA_F_PERMANENT)
673 continue;
675 /* We try to batch several events at once. */
676 age = (now - ifa->ifa_tstamp +
677 ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
679 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
680 age >= ifa->ifa_valid_lft) {
681 struct in_ifaddr **ifap;
683 for (ifap = &ifa->ifa_dev->ifa_list;
684 *ifap != NULL; ifap = &(*ifap)->ifa_next) {
685 if (*ifap == ifa) {
686 inet_del_ifa(ifa->ifa_dev,
687 ifap, 1);
688 break;
691 } else if (ifa->ifa_preferred_lft !=
692 INFINITY_LIFE_TIME &&
693 age >= ifa->ifa_preferred_lft &&
694 !(ifa->ifa_flags & IFA_F_DEPRECATED)) {
695 ifa->ifa_flags |= IFA_F_DEPRECATED;
696 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
699 rtnl_unlock();
702 next_sec = round_jiffies_up(next);
703 next_sched = next;
705 /* If rounded timeout is accurate enough, accept it. */
706 if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
707 next_sched = next_sec;
709 now = jiffies;
710 /* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */
711 if (time_before(next_sched, now + ADDRCONF_TIMER_FUZZ_MAX))
712 next_sched = now + ADDRCONF_TIMER_FUZZ_MAX;
714 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work,
715 next_sched - now);
718 static void set_ifa_lifetime(struct in_ifaddr *ifa, __u32 valid_lft,
719 __u32 prefered_lft)
721 unsigned long timeout;
723 ifa->ifa_flags &= ~(IFA_F_PERMANENT | IFA_F_DEPRECATED);
725 timeout = addrconf_timeout_fixup(valid_lft, HZ);
726 if (addrconf_finite_timeout(timeout))
727 ifa->ifa_valid_lft = timeout;
728 else
729 ifa->ifa_flags |= IFA_F_PERMANENT;
731 timeout = addrconf_timeout_fixup(prefered_lft, HZ);
732 if (addrconf_finite_timeout(timeout)) {
733 if (timeout == 0)
734 ifa->ifa_flags |= IFA_F_DEPRECATED;
735 ifa->ifa_preferred_lft = timeout;
737 ifa->ifa_tstamp = jiffies;
738 if (!ifa->ifa_cstamp)
739 ifa->ifa_cstamp = ifa->ifa_tstamp;
742 static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh,
743 __u32 *pvalid_lft, __u32 *pprefered_lft)
745 struct nlattr *tb[IFA_MAX+1];
746 struct in_ifaddr *ifa;
747 struct ifaddrmsg *ifm;
748 struct net_device *dev;
749 struct in_device *in_dev;
750 int err;
752 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
753 if (err < 0)
754 goto errout;
756 ifm = nlmsg_data(nlh);
757 err = -EINVAL;
758 if (ifm->ifa_prefixlen > 32 || !tb[IFA_LOCAL])
759 goto errout;
761 dev = __dev_get_by_index(net, ifm->ifa_index);
762 err = -ENODEV;
763 if (!dev)
764 goto errout;
766 in_dev = __in_dev_get_rtnl(dev);
767 err = -ENOBUFS;
768 if (!in_dev)
769 goto errout;
771 ifa = inet_alloc_ifa();
772 if (!ifa)
774 * A potential indev allocation can be left alive, it stays
775 * assigned to its device and is destroy with it.
777 goto errout;
779 ipv4_devconf_setall(in_dev);
780 neigh_parms_data_state_setall(in_dev->arp_parms);
781 in_dev_hold(in_dev);
783 if (!tb[IFA_ADDRESS])
784 tb[IFA_ADDRESS] = tb[IFA_LOCAL];
786 INIT_HLIST_NODE(&ifa->hash);
787 ifa->ifa_prefixlen = ifm->ifa_prefixlen;
788 ifa->ifa_mask = inet_make_mask(ifm->ifa_prefixlen);
789 ifa->ifa_flags = tb[IFA_FLAGS] ? nla_get_u32(tb[IFA_FLAGS]) :
790 ifm->ifa_flags;
791 ifa->ifa_scope = ifm->ifa_scope;
792 ifa->ifa_dev = in_dev;
794 ifa->ifa_local = nla_get_in_addr(tb[IFA_LOCAL]);
795 ifa->ifa_address = nla_get_in_addr(tb[IFA_ADDRESS]);
797 if (tb[IFA_BROADCAST])
798 ifa->ifa_broadcast = nla_get_in_addr(tb[IFA_BROADCAST]);
800 if (tb[IFA_LABEL])
801 nla_strlcpy(ifa->ifa_label, tb[IFA_LABEL], IFNAMSIZ);
802 else
803 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
805 if (tb[IFA_CACHEINFO]) {
806 struct ifa_cacheinfo *ci;
808 ci = nla_data(tb[IFA_CACHEINFO]);
809 if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) {
810 err = -EINVAL;
811 goto errout_free;
813 *pvalid_lft = ci->ifa_valid;
814 *pprefered_lft = ci->ifa_prefered;
817 return ifa;
819 errout_free:
820 inet_free_ifa(ifa);
821 errout:
822 return ERR_PTR(err);
825 static struct in_ifaddr *find_matching_ifa(struct in_ifaddr *ifa)
827 struct in_device *in_dev = ifa->ifa_dev;
828 struct in_ifaddr *ifa1, **ifap;
830 if (!ifa->ifa_local)
831 return NULL;
833 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
834 ifap = &ifa1->ifa_next) {
835 if (ifa1->ifa_mask == ifa->ifa_mask &&
836 inet_ifa_match(ifa1->ifa_address, ifa) &&
837 ifa1->ifa_local == ifa->ifa_local)
838 return ifa1;
840 return NULL;
843 static int inet_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh)
845 struct net *net = sock_net(skb->sk);
846 struct in_ifaddr *ifa;
847 struct in_ifaddr *ifa_existing;
848 __u32 valid_lft = INFINITY_LIFE_TIME;
849 __u32 prefered_lft = INFINITY_LIFE_TIME;
851 ASSERT_RTNL();
853 ifa = rtm_to_ifaddr(net, nlh, &valid_lft, &prefered_lft);
854 if (IS_ERR(ifa))
855 return PTR_ERR(ifa);
857 ifa_existing = find_matching_ifa(ifa);
858 if (!ifa_existing) {
859 /* It would be best to check for !NLM_F_CREATE here but
860 * userspace already relies on not having to provide this.
862 set_ifa_lifetime(ifa, valid_lft, prefered_lft);
863 if (ifa->ifa_flags & IFA_F_MCAUTOJOIN) {
864 int ret = ip_mc_config(net->ipv4.mc_autojoin_sk,
865 true, ifa);
867 if (ret < 0) {
868 inet_free_ifa(ifa);
869 return ret;
872 return __inet_insert_ifa(ifa, nlh, NETLINK_CB(skb).portid);
873 } else {
874 inet_free_ifa(ifa);
876 if (nlh->nlmsg_flags & NLM_F_EXCL ||
877 !(nlh->nlmsg_flags & NLM_F_REPLACE))
878 return -EEXIST;
879 ifa = ifa_existing;
880 set_ifa_lifetime(ifa, valid_lft, prefered_lft);
881 cancel_delayed_work(&check_lifetime_work);
882 queue_delayed_work(system_power_efficient_wq,
883 &check_lifetime_work, 0);
884 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, NETLINK_CB(skb).portid);
886 return 0;
890 * Determine a default network mask, based on the IP address.
893 static int inet_abc_len(__be32 addr)
895 int rc = -1; /* Something else, probably a multicast. */
897 if (ipv4_is_zeronet(addr))
898 rc = 0;
899 else {
900 __u32 haddr = ntohl(addr);
902 if (IN_CLASSA(haddr))
903 rc = 8;
904 else if (IN_CLASSB(haddr))
905 rc = 16;
906 else if (IN_CLASSC(haddr))
907 rc = 24;
910 return rc;
914 int devinet_ioctl(struct net *net, unsigned int cmd, void __user *arg)
916 struct ifreq ifr;
917 struct sockaddr_in sin_orig;
918 struct sockaddr_in *sin = (struct sockaddr_in *)&ifr.ifr_addr;
919 struct in_device *in_dev;
920 struct in_ifaddr **ifap = NULL;
921 struct in_ifaddr *ifa = NULL;
922 struct net_device *dev;
923 char *colon;
924 int ret = -EFAULT;
925 int tryaddrmatch = 0;
928 * Fetch the caller's info block into kernel space
931 if (copy_from_user(&ifr, arg, sizeof(struct ifreq)))
932 goto out;
933 ifr.ifr_name[IFNAMSIZ - 1] = 0;
935 /* save original address for comparison */
936 memcpy(&sin_orig, sin, sizeof(*sin));
938 colon = strchr(ifr.ifr_name, ':');
939 if (colon)
940 *colon = 0;
942 dev_load(net, ifr.ifr_name);
944 switch (cmd) {
945 case SIOCGIFADDR: /* Get interface address */
946 case SIOCGIFBRDADDR: /* Get the broadcast address */
947 case SIOCGIFDSTADDR: /* Get the destination address */
948 case SIOCGIFNETMASK: /* Get the netmask for the interface */
949 /* Note that these ioctls will not sleep,
950 so that we do not impose a lock.
951 One day we will be forced to put shlock here (I mean SMP)
953 tryaddrmatch = (sin_orig.sin_family == AF_INET);
954 memset(sin, 0, sizeof(*sin));
955 sin->sin_family = AF_INET;
956 break;
958 case SIOCSIFFLAGS:
959 ret = -EPERM;
960 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
961 goto out;
962 break;
963 case SIOCSIFADDR: /* Set interface address (and family) */
964 case SIOCSIFBRDADDR: /* Set the broadcast address */
965 case SIOCSIFDSTADDR: /* Set the destination address */
966 case SIOCSIFNETMASK: /* Set the netmask for the interface */
967 ret = -EPERM;
968 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
969 goto out;
970 ret = -EINVAL;
971 if (sin->sin_family != AF_INET)
972 goto out;
973 break;
974 default:
975 ret = -EINVAL;
976 goto out;
979 rtnl_lock();
981 ret = -ENODEV;
982 dev = __dev_get_by_name(net, ifr.ifr_name);
983 if (!dev)
984 goto done;
986 if (colon)
987 *colon = ':';
989 in_dev = __in_dev_get_rtnl(dev);
990 if (in_dev) {
991 if (tryaddrmatch) {
992 /* Matthias Andree */
993 /* compare label and address (4.4BSD style) */
994 /* note: we only do this for a limited set of ioctls
995 and only if the original address family was AF_INET.
996 This is checked above. */
997 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
998 ifap = &ifa->ifa_next) {
999 if (!strcmp(ifr.ifr_name, ifa->ifa_label) &&
1000 sin_orig.sin_addr.s_addr ==
1001 ifa->ifa_local) {
1002 break; /* found */
1006 /* we didn't get a match, maybe the application is
1007 4.3BSD-style and passed in junk so we fall back to
1008 comparing just the label */
1009 if (!ifa) {
1010 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
1011 ifap = &ifa->ifa_next)
1012 if (!strcmp(ifr.ifr_name, ifa->ifa_label))
1013 break;
1017 ret = -EADDRNOTAVAIL;
1018 if (!ifa && cmd != SIOCSIFADDR && cmd != SIOCSIFFLAGS)
1019 goto done;
1021 switch (cmd) {
1022 case SIOCGIFADDR: /* Get interface address */
1023 sin->sin_addr.s_addr = ifa->ifa_local;
1024 goto rarok;
1026 case SIOCGIFBRDADDR: /* Get the broadcast address */
1027 sin->sin_addr.s_addr = ifa->ifa_broadcast;
1028 goto rarok;
1030 case SIOCGIFDSTADDR: /* Get the destination address */
1031 sin->sin_addr.s_addr = ifa->ifa_address;
1032 goto rarok;
1034 case SIOCGIFNETMASK: /* Get the netmask for the interface */
1035 sin->sin_addr.s_addr = ifa->ifa_mask;
1036 goto rarok;
1038 case SIOCSIFFLAGS:
1039 if (colon) {
1040 ret = -EADDRNOTAVAIL;
1041 if (!ifa)
1042 break;
1043 ret = 0;
1044 if (!(ifr.ifr_flags & IFF_UP))
1045 inet_del_ifa(in_dev, ifap, 1);
1046 break;
1048 ret = dev_change_flags(dev, ifr.ifr_flags);
1049 break;
1051 case SIOCSIFADDR: /* Set interface address (and family) */
1052 ret = -EINVAL;
1053 if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1054 break;
1056 if (!ifa) {
1057 ret = -ENOBUFS;
1058 ifa = inet_alloc_ifa();
1059 if (!ifa)
1060 break;
1061 INIT_HLIST_NODE(&ifa->hash);
1062 if (colon)
1063 memcpy(ifa->ifa_label, ifr.ifr_name, IFNAMSIZ);
1064 else
1065 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1066 } else {
1067 ret = 0;
1068 if (ifa->ifa_local == sin->sin_addr.s_addr)
1069 break;
1070 inet_del_ifa(in_dev, ifap, 0);
1071 ifa->ifa_broadcast = 0;
1072 ifa->ifa_scope = 0;
1075 ifa->ifa_address = ifa->ifa_local = sin->sin_addr.s_addr;
1077 if (!(dev->flags & IFF_POINTOPOINT)) {
1078 ifa->ifa_prefixlen = inet_abc_len(ifa->ifa_address);
1079 ifa->ifa_mask = inet_make_mask(ifa->ifa_prefixlen);
1080 if ((dev->flags & IFF_BROADCAST) &&
1081 ifa->ifa_prefixlen < 31)
1082 ifa->ifa_broadcast = ifa->ifa_address |
1083 ~ifa->ifa_mask;
1084 } else {
1085 ifa->ifa_prefixlen = 32;
1086 ifa->ifa_mask = inet_make_mask(32);
1088 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, INFINITY_LIFE_TIME);
1089 ret = inet_set_ifa(dev, ifa);
1090 break;
1092 case SIOCSIFBRDADDR: /* Set the broadcast address */
1093 ret = 0;
1094 if (ifa->ifa_broadcast != sin->sin_addr.s_addr) {
1095 inet_del_ifa(in_dev, ifap, 0);
1096 ifa->ifa_broadcast = sin->sin_addr.s_addr;
1097 inet_insert_ifa(ifa);
1099 break;
1101 case SIOCSIFDSTADDR: /* Set the destination address */
1102 ret = 0;
1103 if (ifa->ifa_address == sin->sin_addr.s_addr)
1104 break;
1105 ret = -EINVAL;
1106 if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1107 break;
1108 ret = 0;
1109 inet_del_ifa(in_dev, ifap, 0);
1110 ifa->ifa_address = sin->sin_addr.s_addr;
1111 inet_insert_ifa(ifa);
1112 break;
1114 case SIOCSIFNETMASK: /* Set the netmask for the interface */
1117 * The mask we set must be legal.
1119 ret = -EINVAL;
1120 if (bad_mask(sin->sin_addr.s_addr, 0))
1121 break;
1122 ret = 0;
1123 if (ifa->ifa_mask != sin->sin_addr.s_addr) {
1124 __be32 old_mask = ifa->ifa_mask;
1125 inet_del_ifa(in_dev, ifap, 0);
1126 ifa->ifa_mask = sin->sin_addr.s_addr;
1127 ifa->ifa_prefixlen = inet_mask_len(ifa->ifa_mask);
1129 /* See if current broadcast address matches
1130 * with current netmask, then recalculate
1131 * the broadcast address. Otherwise it's a
1132 * funny address, so don't touch it since
1133 * the user seems to know what (s)he's doing...
1135 if ((dev->flags & IFF_BROADCAST) &&
1136 (ifa->ifa_prefixlen < 31) &&
1137 (ifa->ifa_broadcast ==
1138 (ifa->ifa_local|~old_mask))) {
1139 ifa->ifa_broadcast = (ifa->ifa_local |
1140 ~sin->sin_addr.s_addr);
1142 inet_insert_ifa(ifa);
1144 break;
1146 done:
1147 rtnl_unlock();
1148 out:
1149 return ret;
1150 rarok:
1151 rtnl_unlock();
1152 ret = copy_to_user(arg, &ifr, sizeof(struct ifreq)) ? -EFAULT : 0;
1153 goto out;
1156 static int inet_gifconf(struct net_device *dev, char __user *buf, int len)
1158 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1159 struct in_ifaddr *ifa;
1160 struct ifreq ifr;
1161 int done = 0;
1163 if (!in_dev)
1164 goto out;
1166 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1167 if (!buf) {
1168 done += sizeof(ifr);
1169 continue;
1171 if (len < (int) sizeof(ifr))
1172 break;
1173 memset(&ifr, 0, sizeof(struct ifreq));
1174 strcpy(ifr.ifr_name, ifa->ifa_label);
1176 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_family = AF_INET;
1177 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_addr.s_addr =
1178 ifa->ifa_local;
1180 if (copy_to_user(buf, &ifr, sizeof(struct ifreq))) {
1181 done = -EFAULT;
1182 break;
1184 buf += sizeof(struct ifreq);
1185 len -= sizeof(struct ifreq);
1186 done += sizeof(struct ifreq);
1188 out:
1189 return done;
1192 __be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope)
1194 __be32 addr = 0;
1195 struct in_device *in_dev;
1196 struct net *net = dev_net(dev);
1198 rcu_read_lock();
1199 in_dev = __in_dev_get_rcu(dev);
1200 if (!in_dev)
1201 goto no_in_dev;
1203 for_primary_ifa(in_dev) {
1204 if (ifa->ifa_scope > scope)
1205 continue;
1206 if (!dst || inet_ifa_match(dst, ifa)) {
1207 addr = ifa->ifa_local;
1208 break;
1210 if (!addr)
1211 addr = ifa->ifa_local;
1212 } endfor_ifa(in_dev);
1214 if (addr)
1215 goto out_unlock;
1216 no_in_dev:
1218 /* Not loopback addresses on loopback should be preferred
1219 in this case. It is important that lo is the first interface
1220 in dev_base list.
1222 for_each_netdev_rcu(net, dev) {
1223 in_dev = __in_dev_get_rcu(dev);
1224 if (!in_dev)
1225 continue;
1227 for_primary_ifa(in_dev) {
1228 if (ifa->ifa_scope != RT_SCOPE_LINK &&
1229 ifa->ifa_scope <= scope) {
1230 addr = ifa->ifa_local;
1231 goto out_unlock;
1233 } endfor_ifa(in_dev);
1235 out_unlock:
1236 rcu_read_unlock();
1237 return addr;
1239 EXPORT_SYMBOL(inet_select_addr);
1241 static __be32 confirm_addr_indev(struct in_device *in_dev, __be32 dst,
1242 __be32 local, int scope)
1244 int same = 0;
1245 __be32 addr = 0;
1247 for_ifa(in_dev) {
1248 if (!addr &&
1249 (local == ifa->ifa_local || !local) &&
1250 ifa->ifa_scope <= scope) {
1251 addr = ifa->ifa_local;
1252 if (same)
1253 break;
1255 if (!same) {
1256 same = (!local || inet_ifa_match(local, ifa)) &&
1257 (!dst || inet_ifa_match(dst, ifa));
1258 if (same && addr) {
1259 if (local || !dst)
1260 break;
1261 /* Is the selected addr into dst subnet? */
1262 if (inet_ifa_match(addr, ifa))
1263 break;
1264 /* No, then can we use new local src? */
1265 if (ifa->ifa_scope <= scope) {
1266 addr = ifa->ifa_local;
1267 break;
1269 /* search for large dst subnet for addr */
1270 same = 0;
1273 } endfor_ifa(in_dev);
1275 return same ? addr : 0;
1279 * Confirm that local IP address exists using wildcards:
1280 * - net: netns to check, cannot be NULL
1281 * - in_dev: only on this interface, NULL=any interface
1282 * - dst: only in the same subnet as dst, 0=any dst
1283 * - local: address, 0=autoselect the local address
1284 * - scope: maximum allowed scope value for the local address
1286 __be32 inet_confirm_addr(struct net *net, struct in_device *in_dev,
1287 __be32 dst, __be32 local, int scope)
1289 __be32 addr = 0;
1290 struct net_device *dev;
1292 if (in_dev)
1293 return confirm_addr_indev(in_dev, dst, local, scope);
1295 rcu_read_lock();
1296 for_each_netdev_rcu(net, dev) {
1297 in_dev = __in_dev_get_rcu(dev);
1298 if (in_dev) {
1299 addr = confirm_addr_indev(in_dev, dst, local, scope);
1300 if (addr)
1301 break;
1304 rcu_read_unlock();
1306 return addr;
1308 EXPORT_SYMBOL(inet_confirm_addr);
1311 * Device notifier
1314 int register_inetaddr_notifier(struct notifier_block *nb)
1316 return blocking_notifier_chain_register(&inetaddr_chain, nb);
1318 EXPORT_SYMBOL(register_inetaddr_notifier);
1320 int unregister_inetaddr_notifier(struct notifier_block *nb)
1322 return blocking_notifier_chain_unregister(&inetaddr_chain, nb);
1324 EXPORT_SYMBOL(unregister_inetaddr_notifier);
1326 /* Rename ifa_labels for a device name change. Make some effort to preserve
1327 * existing alias numbering and to create unique labels if possible.
1329 static void inetdev_changename(struct net_device *dev, struct in_device *in_dev)
1331 struct in_ifaddr *ifa;
1332 int named = 0;
1334 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1335 char old[IFNAMSIZ], *dot;
1337 memcpy(old, ifa->ifa_label, IFNAMSIZ);
1338 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1339 if (named++ == 0)
1340 goto skip;
1341 dot = strchr(old, ':');
1342 if (!dot) {
1343 sprintf(old, ":%d", named);
1344 dot = old;
1346 if (strlen(dot) + strlen(dev->name) < IFNAMSIZ)
1347 strcat(ifa->ifa_label, dot);
1348 else
1349 strcpy(ifa->ifa_label + (IFNAMSIZ - strlen(dot) - 1), dot);
1350 skip:
1351 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
1355 static bool inetdev_valid_mtu(unsigned int mtu)
1357 return mtu >= 68;
1360 static void inetdev_send_gratuitous_arp(struct net_device *dev,
1361 struct in_device *in_dev)
1364 struct in_ifaddr *ifa;
1366 for (ifa = in_dev->ifa_list; ifa;
1367 ifa = ifa->ifa_next) {
1368 arp_send(ARPOP_REQUEST, ETH_P_ARP,
1369 ifa->ifa_local, dev,
1370 ifa->ifa_local, NULL,
1371 dev->dev_addr, NULL);
1375 /* Called only under RTNL semaphore */
1377 static int inetdev_event(struct notifier_block *this, unsigned long event,
1378 void *ptr)
1380 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1381 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1383 ASSERT_RTNL();
1385 if (!in_dev) {
1386 if (event == NETDEV_REGISTER) {
1387 in_dev = inetdev_init(dev);
1388 if (IS_ERR(in_dev))
1389 return notifier_from_errno(PTR_ERR(in_dev));
1390 if (dev->flags & IFF_LOOPBACK) {
1391 IN_DEV_CONF_SET(in_dev, NOXFRM, 1);
1392 IN_DEV_CONF_SET(in_dev, NOPOLICY, 1);
1394 } else if (event == NETDEV_CHANGEMTU) {
1395 /* Re-enabling IP */
1396 if (inetdev_valid_mtu(dev->mtu))
1397 in_dev = inetdev_init(dev);
1399 goto out;
1402 switch (event) {
1403 case NETDEV_REGISTER:
1404 pr_debug("%s: bug\n", __func__);
1405 RCU_INIT_POINTER(dev->ip_ptr, NULL);
1406 break;
1407 case NETDEV_UP:
1408 if (!inetdev_valid_mtu(dev->mtu))
1409 break;
1410 if (dev->flags & IFF_LOOPBACK) {
1411 struct in_ifaddr *ifa = inet_alloc_ifa();
1413 if (ifa) {
1414 INIT_HLIST_NODE(&ifa->hash);
1415 ifa->ifa_local =
1416 ifa->ifa_address = htonl(INADDR_LOOPBACK);
1417 ifa->ifa_prefixlen = 8;
1418 ifa->ifa_mask = inet_make_mask(8);
1419 in_dev_hold(in_dev);
1420 ifa->ifa_dev = in_dev;
1421 ifa->ifa_scope = RT_SCOPE_HOST;
1422 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1423 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME,
1424 INFINITY_LIFE_TIME);
1425 ipv4_devconf_setall(in_dev);
1426 neigh_parms_data_state_setall(in_dev->arp_parms);
1427 inet_insert_ifa(ifa);
1430 ip_mc_up(in_dev);
1431 /* fall through */
1432 case NETDEV_CHANGEADDR:
1433 if (!IN_DEV_ARP_NOTIFY(in_dev))
1434 break;
1435 /* fall through */
1436 case NETDEV_NOTIFY_PEERS:
1437 /* Send gratuitous ARP to notify of link change */
1438 inetdev_send_gratuitous_arp(dev, in_dev);
1439 break;
1440 case NETDEV_DOWN:
1441 ip_mc_down(in_dev);
1442 break;
1443 case NETDEV_PRE_TYPE_CHANGE:
1444 ip_mc_unmap(in_dev);
1445 break;
1446 case NETDEV_POST_TYPE_CHANGE:
1447 ip_mc_remap(in_dev);
1448 break;
1449 case NETDEV_CHANGEMTU:
1450 if (inetdev_valid_mtu(dev->mtu))
1451 break;
1452 /* disable IP when MTU is not enough */
1453 case NETDEV_UNREGISTER:
1454 inetdev_destroy(in_dev);
1455 break;
1456 case NETDEV_CHANGENAME:
1457 /* Do not notify about label change, this event is
1458 * not interesting to applications using netlink.
1460 inetdev_changename(dev, in_dev);
1462 devinet_sysctl_unregister(in_dev);
1463 devinet_sysctl_register(in_dev);
1464 break;
1466 out:
1467 return NOTIFY_DONE;
1470 static struct notifier_block ip_netdev_notifier = {
1471 .notifier_call = inetdev_event,
1474 static size_t inet_nlmsg_size(void)
1476 return NLMSG_ALIGN(sizeof(struct ifaddrmsg))
1477 + nla_total_size(4) /* IFA_ADDRESS */
1478 + nla_total_size(4) /* IFA_LOCAL */
1479 + nla_total_size(4) /* IFA_BROADCAST */
1480 + nla_total_size(IFNAMSIZ) /* IFA_LABEL */
1481 + nla_total_size(4) /* IFA_FLAGS */
1482 + nla_total_size(sizeof(struct ifa_cacheinfo)); /* IFA_CACHEINFO */
1485 static inline u32 cstamp_delta(unsigned long cstamp)
1487 return (cstamp - INITIAL_JIFFIES) * 100UL / HZ;
1490 static int put_cacheinfo(struct sk_buff *skb, unsigned long cstamp,
1491 unsigned long tstamp, u32 preferred, u32 valid)
1493 struct ifa_cacheinfo ci;
1495 ci.cstamp = cstamp_delta(cstamp);
1496 ci.tstamp = cstamp_delta(tstamp);
1497 ci.ifa_prefered = preferred;
1498 ci.ifa_valid = valid;
1500 return nla_put(skb, IFA_CACHEINFO, sizeof(ci), &ci);
1503 static int inet_fill_ifaddr(struct sk_buff *skb, struct in_ifaddr *ifa,
1504 u32 portid, u32 seq, int event, unsigned int flags)
1506 struct ifaddrmsg *ifm;
1507 struct nlmsghdr *nlh;
1508 u32 preferred, valid;
1510 nlh = nlmsg_put(skb, portid, seq, event, sizeof(*ifm), flags);
1511 if (!nlh)
1512 return -EMSGSIZE;
1514 ifm = nlmsg_data(nlh);
1515 ifm->ifa_family = AF_INET;
1516 ifm->ifa_prefixlen = ifa->ifa_prefixlen;
1517 ifm->ifa_flags = ifa->ifa_flags;
1518 ifm->ifa_scope = ifa->ifa_scope;
1519 ifm->ifa_index = ifa->ifa_dev->dev->ifindex;
1521 if (!(ifm->ifa_flags & IFA_F_PERMANENT)) {
1522 preferred = ifa->ifa_preferred_lft;
1523 valid = ifa->ifa_valid_lft;
1524 if (preferred != INFINITY_LIFE_TIME) {
1525 long tval = (jiffies - ifa->ifa_tstamp) / HZ;
1527 if (preferred > tval)
1528 preferred -= tval;
1529 else
1530 preferred = 0;
1531 if (valid != INFINITY_LIFE_TIME) {
1532 if (valid > tval)
1533 valid -= tval;
1534 else
1535 valid = 0;
1538 } else {
1539 preferred = INFINITY_LIFE_TIME;
1540 valid = INFINITY_LIFE_TIME;
1542 if ((ifa->ifa_address &&
1543 nla_put_in_addr(skb, IFA_ADDRESS, ifa->ifa_address)) ||
1544 (ifa->ifa_local &&
1545 nla_put_in_addr(skb, IFA_LOCAL, ifa->ifa_local)) ||
1546 (ifa->ifa_broadcast &&
1547 nla_put_in_addr(skb, IFA_BROADCAST, ifa->ifa_broadcast)) ||
1548 (ifa->ifa_label[0] &&
1549 nla_put_string(skb, IFA_LABEL, ifa->ifa_label)) ||
1550 nla_put_u32(skb, IFA_FLAGS, ifa->ifa_flags) ||
1551 put_cacheinfo(skb, ifa->ifa_cstamp, ifa->ifa_tstamp,
1552 preferred, valid))
1553 goto nla_put_failure;
1555 nlmsg_end(skb, nlh);
1556 return 0;
1558 nla_put_failure:
1559 nlmsg_cancel(skb, nlh);
1560 return -EMSGSIZE;
1563 static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
1565 struct net *net = sock_net(skb->sk);
1566 int h, s_h;
1567 int idx, s_idx;
1568 int ip_idx, s_ip_idx;
1569 struct net_device *dev;
1570 struct in_device *in_dev;
1571 struct in_ifaddr *ifa;
1572 struct hlist_head *head;
1574 s_h = cb->args[0];
1575 s_idx = idx = cb->args[1];
1576 s_ip_idx = ip_idx = cb->args[2];
1578 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1579 idx = 0;
1580 head = &net->dev_index_head[h];
1581 rcu_read_lock();
1582 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1583 net->dev_base_seq;
1584 hlist_for_each_entry_rcu(dev, head, index_hlist) {
1585 if (idx < s_idx)
1586 goto cont;
1587 if (h > s_h || idx > s_idx)
1588 s_ip_idx = 0;
1589 in_dev = __in_dev_get_rcu(dev);
1590 if (!in_dev)
1591 goto cont;
1593 for (ifa = in_dev->ifa_list, ip_idx = 0; ifa;
1594 ifa = ifa->ifa_next, ip_idx++) {
1595 if (ip_idx < s_ip_idx)
1596 continue;
1597 if (inet_fill_ifaddr(skb, ifa,
1598 NETLINK_CB(cb->skb).portid,
1599 cb->nlh->nlmsg_seq,
1600 RTM_NEWADDR, NLM_F_MULTI) < 0) {
1601 rcu_read_unlock();
1602 goto done;
1604 nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1606 cont:
1607 idx++;
1609 rcu_read_unlock();
1612 done:
1613 cb->args[0] = h;
1614 cb->args[1] = idx;
1615 cb->args[2] = ip_idx;
1617 return skb->len;
1620 static void rtmsg_ifa(int event, struct in_ifaddr *ifa, struct nlmsghdr *nlh,
1621 u32 portid)
1623 struct sk_buff *skb;
1624 u32 seq = nlh ? nlh->nlmsg_seq : 0;
1625 int err = -ENOBUFS;
1626 struct net *net;
1628 net = dev_net(ifa->ifa_dev->dev);
1629 skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL);
1630 if (!skb)
1631 goto errout;
1633 err = inet_fill_ifaddr(skb, ifa, portid, seq, event, 0);
1634 if (err < 0) {
1635 /* -EMSGSIZE implies BUG in inet_nlmsg_size() */
1636 WARN_ON(err == -EMSGSIZE);
1637 kfree_skb(skb);
1638 goto errout;
1640 rtnl_notify(skb, net, portid, RTNLGRP_IPV4_IFADDR, nlh, GFP_KERNEL);
1641 return;
1642 errout:
1643 if (err < 0)
1644 rtnl_set_sk_err(net, RTNLGRP_IPV4_IFADDR, err);
1647 static size_t inet_get_link_af_size(const struct net_device *dev)
1649 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1651 if (!in_dev)
1652 return 0;
1654 return nla_total_size(IPV4_DEVCONF_MAX * 4); /* IFLA_INET_CONF */
1657 static int inet_fill_link_af(struct sk_buff *skb, const struct net_device *dev)
1659 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1660 struct nlattr *nla;
1661 int i;
1663 if (!in_dev)
1664 return -ENODATA;
1666 nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4);
1667 if (!nla)
1668 return -EMSGSIZE;
1670 for (i = 0; i < IPV4_DEVCONF_MAX; i++)
1671 ((u32 *) nla_data(nla))[i] = in_dev->cnf.data[i];
1673 return 0;
1676 static const struct nla_policy inet_af_policy[IFLA_INET_MAX+1] = {
1677 [IFLA_INET_CONF] = { .type = NLA_NESTED },
1680 static int inet_validate_link_af(const struct net_device *dev,
1681 const struct nlattr *nla)
1683 struct nlattr *a, *tb[IFLA_INET_MAX+1];
1684 int err, rem;
1686 if (dev && !__in_dev_get_rtnl(dev))
1687 return -EAFNOSUPPORT;
1689 err = nla_parse_nested(tb, IFLA_INET_MAX, nla, inet_af_policy);
1690 if (err < 0)
1691 return err;
1693 if (tb[IFLA_INET_CONF]) {
1694 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) {
1695 int cfgid = nla_type(a);
1697 if (nla_len(a) < 4)
1698 return -EINVAL;
1700 if (cfgid <= 0 || cfgid > IPV4_DEVCONF_MAX)
1701 return -EINVAL;
1705 return 0;
1708 static int inet_set_link_af(struct net_device *dev, const struct nlattr *nla)
1710 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1711 struct nlattr *a, *tb[IFLA_INET_MAX+1];
1712 int rem;
1714 if (!in_dev)
1715 return -EAFNOSUPPORT;
1717 if (nla_parse_nested(tb, IFLA_INET_MAX, nla, NULL) < 0)
1718 BUG();
1720 if (tb[IFLA_INET_CONF]) {
1721 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem)
1722 ipv4_devconf_set(in_dev, nla_type(a), nla_get_u32(a));
1725 return 0;
1728 static int inet_netconf_msgsize_devconf(int type)
1730 int size = NLMSG_ALIGN(sizeof(struct netconfmsg))
1731 + nla_total_size(4); /* NETCONFA_IFINDEX */
1733 /* type -1 is used for ALL */
1734 if (type == -1 || type == NETCONFA_FORWARDING)
1735 size += nla_total_size(4);
1736 if (type == -1 || type == NETCONFA_RP_FILTER)
1737 size += nla_total_size(4);
1738 if (type == -1 || type == NETCONFA_MC_FORWARDING)
1739 size += nla_total_size(4);
1740 if (type == -1 || type == NETCONFA_PROXY_NEIGH)
1741 size += nla_total_size(4);
1742 if (type == -1 || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN)
1743 size += nla_total_size(4);
1745 return size;
1748 static int inet_netconf_fill_devconf(struct sk_buff *skb, int ifindex,
1749 struct ipv4_devconf *devconf, u32 portid,
1750 u32 seq, int event, unsigned int flags,
1751 int type)
1753 struct nlmsghdr *nlh;
1754 struct netconfmsg *ncm;
1756 nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg),
1757 flags);
1758 if (!nlh)
1759 return -EMSGSIZE;
1761 ncm = nlmsg_data(nlh);
1762 ncm->ncm_family = AF_INET;
1764 if (nla_put_s32(skb, NETCONFA_IFINDEX, ifindex) < 0)
1765 goto nla_put_failure;
1767 /* type -1 is used for ALL */
1768 if ((type == -1 || type == NETCONFA_FORWARDING) &&
1769 nla_put_s32(skb, NETCONFA_FORWARDING,
1770 IPV4_DEVCONF(*devconf, FORWARDING)) < 0)
1771 goto nla_put_failure;
1772 if ((type == -1 || type == NETCONFA_RP_FILTER) &&
1773 nla_put_s32(skb, NETCONFA_RP_FILTER,
1774 IPV4_DEVCONF(*devconf, RP_FILTER)) < 0)
1775 goto nla_put_failure;
1776 if ((type == -1 || type == NETCONFA_MC_FORWARDING) &&
1777 nla_put_s32(skb, NETCONFA_MC_FORWARDING,
1778 IPV4_DEVCONF(*devconf, MC_FORWARDING)) < 0)
1779 goto nla_put_failure;
1780 if ((type == -1 || type == NETCONFA_PROXY_NEIGH) &&
1781 nla_put_s32(skb, NETCONFA_PROXY_NEIGH,
1782 IPV4_DEVCONF(*devconf, PROXY_ARP)) < 0)
1783 goto nla_put_failure;
1784 if ((type == -1 || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN) &&
1785 nla_put_s32(skb, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
1786 IPV4_DEVCONF(*devconf, IGNORE_ROUTES_WITH_LINKDOWN)) < 0)
1787 goto nla_put_failure;
1789 nlmsg_end(skb, nlh);
1790 return 0;
1792 nla_put_failure:
1793 nlmsg_cancel(skb, nlh);
1794 return -EMSGSIZE;
1797 void inet_netconf_notify_devconf(struct net *net, int type, int ifindex,
1798 struct ipv4_devconf *devconf)
1800 struct sk_buff *skb;
1801 int err = -ENOBUFS;
1803 skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_ATOMIC);
1804 if (!skb)
1805 goto errout;
1807 err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0,
1808 RTM_NEWNETCONF, 0, type);
1809 if (err < 0) {
1810 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1811 WARN_ON(err == -EMSGSIZE);
1812 kfree_skb(skb);
1813 goto errout;
1815 rtnl_notify(skb, net, 0, RTNLGRP_IPV4_NETCONF, NULL, GFP_ATOMIC);
1816 return;
1817 errout:
1818 if (err < 0)
1819 rtnl_set_sk_err(net, RTNLGRP_IPV4_NETCONF, err);
1822 static const struct nla_policy devconf_ipv4_policy[NETCONFA_MAX+1] = {
1823 [NETCONFA_IFINDEX] = { .len = sizeof(int) },
1824 [NETCONFA_FORWARDING] = { .len = sizeof(int) },
1825 [NETCONFA_RP_FILTER] = { .len = sizeof(int) },
1826 [NETCONFA_PROXY_NEIGH] = { .len = sizeof(int) },
1827 [NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN] = { .len = sizeof(int) },
1830 static int inet_netconf_get_devconf(struct sk_buff *in_skb,
1831 struct nlmsghdr *nlh)
1833 struct net *net = sock_net(in_skb->sk);
1834 struct nlattr *tb[NETCONFA_MAX+1];
1835 struct netconfmsg *ncm;
1836 struct sk_buff *skb;
1837 struct ipv4_devconf *devconf;
1838 struct in_device *in_dev;
1839 struct net_device *dev;
1840 int ifindex;
1841 int err;
1843 err = nlmsg_parse(nlh, sizeof(*ncm), tb, NETCONFA_MAX,
1844 devconf_ipv4_policy);
1845 if (err < 0)
1846 goto errout;
1848 err = EINVAL;
1849 if (!tb[NETCONFA_IFINDEX])
1850 goto errout;
1852 ifindex = nla_get_s32(tb[NETCONFA_IFINDEX]);
1853 switch (ifindex) {
1854 case NETCONFA_IFINDEX_ALL:
1855 devconf = net->ipv4.devconf_all;
1856 break;
1857 case NETCONFA_IFINDEX_DEFAULT:
1858 devconf = net->ipv4.devconf_dflt;
1859 break;
1860 default:
1861 dev = __dev_get_by_index(net, ifindex);
1862 if (!dev)
1863 goto errout;
1864 in_dev = __in_dev_get_rtnl(dev);
1865 if (!in_dev)
1866 goto errout;
1867 devconf = &in_dev->cnf;
1868 break;
1871 err = -ENOBUFS;
1872 skb = nlmsg_new(inet_netconf_msgsize_devconf(-1), GFP_ATOMIC);
1873 if (!skb)
1874 goto errout;
1876 err = inet_netconf_fill_devconf(skb, ifindex, devconf,
1877 NETLINK_CB(in_skb).portid,
1878 nlh->nlmsg_seq, RTM_NEWNETCONF, 0,
1879 -1);
1880 if (err < 0) {
1881 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1882 WARN_ON(err == -EMSGSIZE);
1883 kfree_skb(skb);
1884 goto errout;
1886 err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
1887 errout:
1888 return err;
1891 static int inet_netconf_dump_devconf(struct sk_buff *skb,
1892 struct netlink_callback *cb)
1894 struct net *net = sock_net(skb->sk);
1895 int h, s_h;
1896 int idx, s_idx;
1897 struct net_device *dev;
1898 struct in_device *in_dev;
1899 struct hlist_head *head;
1901 s_h = cb->args[0];
1902 s_idx = idx = cb->args[1];
1904 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1905 idx = 0;
1906 head = &net->dev_index_head[h];
1907 rcu_read_lock();
1908 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1909 net->dev_base_seq;
1910 hlist_for_each_entry_rcu(dev, head, index_hlist) {
1911 if (idx < s_idx)
1912 goto cont;
1913 in_dev = __in_dev_get_rcu(dev);
1914 if (!in_dev)
1915 goto cont;
1917 if (inet_netconf_fill_devconf(skb, dev->ifindex,
1918 &in_dev->cnf,
1919 NETLINK_CB(cb->skb).portid,
1920 cb->nlh->nlmsg_seq,
1921 RTM_NEWNETCONF,
1922 NLM_F_MULTI,
1923 -1) < 0) {
1924 rcu_read_unlock();
1925 goto done;
1927 nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1928 cont:
1929 idx++;
1931 rcu_read_unlock();
1933 if (h == NETDEV_HASHENTRIES) {
1934 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_ALL,
1935 net->ipv4.devconf_all,
1936 NETLINK_CB(cb->skb).portid,
1937 cb->nlh->nlmsg_seq,
1938 RTM_NEWNETCONF, NLM_F_MULTI,
1939 -1) < 0)
1940 goto done;
1941 else
1942 h++;
1944 if (h == NETDEV_HASHENTRIES + 1) {
1945 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_DEFAULT,
1946 net->ipv4.devconf_dflt,
1947 NETLINK_CB(cb->skb).portid,
1948 cb->nlh->nlmsg_seq,
1949 RTM_NEWNETCONF, NLM_F_MULTI,
1950 -1) < 0)
1951 goto done;
1952 else
1953 h++;
1955 done:
1956 cb->args[0] = h;
1957 cb->args[1] = idx;
1959 return skb->len;
1962 #ifdef CONFIG_SYSCTL
1964 static void devinet_copy_dflt_conf(struct net *net, int i)
1966 struct net_device *dev;
1968 rcu_read_lock();
1969 for_each_netdev_rcu(net, dev) {
1970 struct in_device *in_dev;
1972 in_dev = __in_dev_get_rcu(dev);
1973 if (in_dev && !test_bit(i, in_dev->cnf.state))
1974 in_dev->cnf.data[i] = net->ipv4.devconf_dflt->data[i];
1976 rcu_read_unlock();
1979 /* called with RTNL locked */
1980 static void inet_forward_change(struct net *net)
1982 struct net_device *dev;
1983 int on = IPV4_DEVCONF_ALL(net, FORWARDING);
1985 IPV4_DEVCONF_ALL(net, ACCEPT_REDIRECTS) = !on;
1986 IPV4_DEVCONF_DFLT(net, FORWARDING) = on;
1987 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
1988 NETCONFA_IFINDEX_ALL,
1989 net->ipv4.devconf_all);
1990 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
1991 NETCONFA_IFINDEX_DEFAULT,
1992 net->ipv4.devconf_dflt);
1994 for_each_netdev(net, dev) {
1995 struct in_device *in_dev;
1996 if (on)
1997 dev_disable_lro(dev);
1998 rcu_read_lock();
1999 in_dev = __in_dev_get_rcu(dev);
2000 if (in_dev) {
2001 IN_DEV_CONF_SET(in_dev, FORWARDING, on);
2002 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
2003 dev->ifindex, &in_dev->cnf);
2005 rcu_read_unlock();
2009 static int devinet_conf_ifindex(struct net *net, struct ipv4_devconf *cnf)
2011 if (cnf == net->ipv4.devconf_dflt)
2012 return NETCONFA_IFINDEX_DEFAULT;
2013 else if (cnf == net->ipv4.devconf_all)
2014 return NETCONFA_IFINDEX_ALL;
2015 else {
2016 struct in_device *idev
2017 = container_of(cnf, struct in_device, cnf);
2018 return idev->dev->ifindex;
2022 static int devinet_conf_proc(struct ctl_table *ctl, int write,
2023 void __user *buffer,
2024 size_t *lenp, loff_t *ppos)
2026 int old_value = *(int *)ctl->data;
2027 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2028 int new_value = *(int *)ctl->data;
2030 if (write) {
2031 struct ipv4_devconf *cnf = ctl->extra1;
2032 struct net *net = ctl->extra2;
2033 int i = (int *)ctl->data - cnf->data;
2034 int ifindex;
2036 set_bit(i, cnf->state);
2038 if (cnf == net->ipv4.devconf_dflt)
2039 devinet_copy_dflt_conf(net, i);
2040 if (i == IPV4_DEVCONF_ACCEPT_LOCAL - 1 ||
2041 i == IPV4_DEVCONF_ROUTE_LOCALNET - 1)
2042 if ((new_value == 0) && (old_value != 0))
2043 rt_cache_flush(net);
2045 if (i == IPV4_DEVCONF_RP_FILTER - 1 &&
2046 new_value != old_value) {
2047 ifindex = devinet_conf_ifindex(net, cnf);
2048 inet_netconf_notify_devconf(net, NETCONFA_RP_FILTER,
2049 ifindex, cnf);
2051 if (i == IPV4_DEVCONF_PROXY_ARP - 1 &&
2052 new_value != old_value) {
2053 ifindex = devinet_conf_ifindex(net, cnf);
2054 inet_netconf_notify_devconf(net, NETCONFA_PROXY_NEIGH,
2055 ifindex, cnf);
2057 if (i == IPV4_DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN - 1 &&
2058 new_value != old_value) {
2059 ifindex = devinet_conf_ifindex(net, cnf);
2060 inet_netconf_notify_devconf(net, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
2061 ifindex, cnf);
2065 return ret;
2068 static int devinet_sysctl_forward(struct ctl_table *ctl, int write,
2069 void __user *buffer,
2070 size_t *lenp, loff_t *ppos)
2072 int *valp = ctl->data;
2073 int val = *valp;
2074 loff_t pos = *ppos;
2075 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2077 if (write && *valp != val) {
2078 struct net *net = ctl->extra2;
2080 if (valp != &IPV4_DEVCONF_DFLT(net, FORWARDING)) {
2081 if (!rtnl_trylock()) {
2082 /* Restore the original values before restarting */
2083 *valp = val;
2084 *ppos = pos;
2085 return restart_syscall();
2087 if (valp == &IPV4_DEVCONF_ALL(net, FORWARDING)) {
2088 inet_forward_change(net);
2089 } else {
2090 struct ipv4_devconf *cnf = ctl->extra1;
2091 struct in_device *idev =
2092 container_of(cnf, struct in_device, cnf);
2093 if (*valp)
2094 dev_disable_lro(idev->dev);
2095 inet_netconf_notify_devconf(net,
2096 NETCONFA_FORWARDING,
2097 idev->dev->ifindex,
2098 cnf);
2100 rtnl_unlock();
2101 rt_cache_flush(net);
2102 } else
2103 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
2104 NETCONFA_IFINDEX_DEFAULT,
2105 net->ipv4.devconf_dflt);
2108 return ret;
2111 static int ipv4_doint_and_flush(struct ctl_table *ctl, int write,
2112 void __user *buffer,
2113 size_t *lenp, loff_t *ppos)
2115 int *valp = ctl->data;
2116 int val = *valp;
2117 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2118 struct net *net = ctl->extra2;
2120 if (write && *valp != val)
2121 rt_cache_flush(net);
2123 return ret;
2126 #define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \
2128 .procname = name, \
2129 .data = ipv4_devconf.data + \
2130 IPV4_DEVCONF_ ## attr - 1, \
2131 .maxlen = sizeof(int), \
2132 .mode = mval, \
2133 .proc_handler = proc, \
2134 .extra1 = &ipv4_devconf, \
2137 #define DEVINET_SYSCTL_RW_ENTRY(attr, name) \
2138 DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc)
2140 #define DEVINET_SYSCTL_RO_ENTRY(attr, name) \
2141 DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc)
2143 #define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \
2144 DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc)
2146 #define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \
2147 DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush)
2149 static struct devinet_sysctl_table {
2150 struct ctl_table_header *sysctl_header;
2151 struct ctl_table devinet_vars[__IPV4_DEVCONF_MAX];
2152 } devinet_sysctl = {
2153 .devinet_vars = {
2154 DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING, "forwarding",
2155 devinet_sysctl_forward),
2156 DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING, "mc_forwarding"),
2158 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS, "accept_redirects"),
2159 DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS, "secure_redirects"),
2160 DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA, "shared_media"),
2161 DEVINET_SYSCTL_RW_ENTRY(RP_FILTER, "rp_filter"),
2162 DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS, "send_redirects"),
2163 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE,
2164 "accept_source_route"),
2165 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL, "accept_local"),
2166 DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK, "src_valid_mark"),
2167 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP, "proxy_arp"),
2168 DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID, "medium_id"),
2169 DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY, "bootp_relay"),
2170 DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS, "log_martians"),
2171 DEVINET_SYSCTL_RW_ENTRY(TAG, "tag"),
2172 DEVINET_SYSCTL_RW_ENTRY(ARPFILTER, "arp_filter"),
2173 DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE, "arp_announce"),
2174 DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE, "arp_ignore"),
2175 DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT, "arp_accept"),
2176 DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY, "arp_notify"),
2177 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN, "proxy_arp_pvlan"),
2178 DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION,
2179 "force_igmp_version"),
2180 DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL,
2181 "igmpv2_unsolicited_report_interval"),
2182 DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL,
2183 "igmpv3_unsolicited_report_interval"),
2184 DEVINET_SYSCTL_RW_ENTRY(IGNORE_ROUTES_WITH_LINKDOWN,
2185 "ignore_routes_with_linkdown"),
2187 DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM, "disable_xfrm"),
2188 DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY, "disable_policy"),
2189 DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES,
2190 "promote_secondaries"),
2191 DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET,
2192 "route_localnet"),
2196 static int __devinet_sysctl_register(struct net *net, char *dev_name,
2197 struct ipv4_devconf *p)
2199 int i;
2200 struct devinet_sysctl_table *t;
2201 char path[sizeof("net/ipv4/conf/") + IFNAMSIZ];
2203 t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL);
2204 if (!t)
2205 goto out;
2207 for (i = 0; i < ARRAY_SIZE(t->devinet_vars) - 1; i++) {
2208 t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf;
2209 t->devinet_vars[i].extra1 = p;
2210 t->devinet_vars[i].extra2 = net;
2213 snprintf(path, sizeof(path), "net/ipv4/conf/%s", dev_name);
2215 t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars);
2216 if (!t->sysctl_header)
2217 goto free;
2219 p->sysctl = t;
2220 return 0;
2222 free:
2223 kfree(t);
2224 out:
2225 return -ENOBUFS;
2228 static void __devinet_sysctl_unregister(struct ipv4_devconf *cnf)
2230 struct devinet_sysctl_table *t = cnf->sysctl;
2232 if (!t)
2233 return;
2235 cnf->sysctl = NULL;
2236 unregister_net_sysctl_table(t->sysctl_header);
2237 kfree(t);
2240 static int devinet_sysctl_register(struct in_device *idev)
2242 int err;
2244 if (!sysctl_dev_name_is_allowed(idev->dev->name))
2245 return -EINVAL;
2247 err = neigh_sysctl_register(idev->dev, idev->arp_parms, NULL);
2248 if (err)
2249 return err;
2250 err = __devinet_sysctl_register(dev_net(idev->dev), idev->dev->name,
2251 &idev->cnf);
2252 if (err)
2253 neigh_sysctl_unregister(idev->arp_parms);
2254 return err;
2257 static void devinet_sysctl_unregister(struct in_device *idev)
2259 __devinet_sysctl_unregister(&idev->cnf);
2260 neigh_sysctl_unregister(idev->arp_parms);
2263 static struct ctl_table ctl_forward_entry[] = {
2265 .procname = "ip_forward",
2266 .data = &ipv4_devconf.data[
2267 IPV4_DEVCONF_FORWARDING - 1],
2268 .maxlen = sizeof(int),
2269 .mode = 0644,
2270 .proc_handler = devinet_sysctl_forward,
2271 .extra1 = &ipv4_devconf,
2272 .extra2 = &init_net,
2274 { },
2276 #endif
2278 static __net_init int devinet_init_net(struct net *net)
2280 int err;
2281 struct ipv4_devconf *all, *dflt;
2282 #ifdef CONFIG_SYSCTL
2283 struct ctl_table *tbl = ctl_forward_entry;
2284 struct ctl_table_header *forw_hdr;
2285 #endif
2287 err = -ENOMEM;
2288 all = &ipv4_devconf;
2289 dflt = &ipv4_devconf_dflt;
2291 if (!net_eq(net, &init_net)) {
2292 all = kmemdup(all, sizeof(ipv4_devconf), GFP_KERNEL);
2293 if (!all)
2294 goto err_alloc_all;
2296 dflt = kmemdup(dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL);
2297 if (!dflt)
2298 goto err_alloc_dflt;
2300 #ifdef CONFIG_SYSCTL
2301 tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL);
2302 if (!tbl)
2303 goto err_alloc_ctl;
2305 tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1];
2306 tbl[0].extra1 = all;
2307 tbl[0].extra2 = net;
2308 #endif
2311 #ifdef CONFIG_SYSCTL
2312 err = __devinet_sysctl_register(net, "all", all);
2313 if (err < 0)
2314 goto err_reg_all;
2316 err = __devinet_sysctl_register(net, "default", dflt);
2317 if (err < 0)
2318 goto err_reg_dflt;
2320 err = -ENOMEM;
2321 forw_hdr = register_net_sysctl(net, "net/ipv4", tbl);
2322 if (!forw_hdr)
2323 goto err_reg_ctl;
2324 net->ipv4.forw_hdr = forw_hdr;
2325 #endif
2327 net->ipv4.devconf_all = all;
2328 net->ipv4.devconf_dflt = dflt;
2329 return 0;
2331 #ifdef CONFIG_SYSCTL
2332 err_reg_ctl:
2333 __devinet_sysctl_unregister(dflt);
2334 err_reg_dflt:
2335 __devinet_sysctl_unregister(all);
2336 err_reg_all:
2337 if (tbl != ctl_forward_entry)
2338 kfree(tbl);
2339 err_alloc_ctl:
2340 #endif
2341 if (dflt != &ipv4_devconf_dflt)
2342 kfree(dflt);
2343 err_alloc_dflt:
2344 if (all != &ipv4_devconf)
2345 kfree(all);
2346 err_alloc_all:
2347 return err;
2350 static __net_exit void devinet_exit_net(struct net *net)
2352 #ifdef CONFIG_SYSCTL
2353 struct ctl_table *tbl;
2355 tbl = net->ipv4.forw_hdr->ctl_table_arg;
2356 unregister_net_sysctl_table(net->ipv4.forw_hdr);
2357 __devinet_sysctl_unregister(net->ipv4.devconf_dflt);
2358 __devinet_sysctl_unregister(net->ipv4.devconf_all);
2359 kfree(tbl);
2360 #endif
2361 kfree(net->ipv4.devconf_dflt);
2362 kfree(net->ipv4.devconf_all);
2365 static __net_initdata struct pernet_operations devinet_ops = {
2366 .init = devinet_init_net,
2367 .exit = devinet_exit_net,
2370 static struct rtnl_af_ops inet_af_ops __read_mostly = {
2371 .family = AF_INET,
2372 .fill_link_af = inet_fill_link_af,
2373 .get_link_af_size = inet_get_link_af_size,
2374 .validate_link_af = inet_validate_link_af,
2375 .set_link_af = inet_set_link_af,
2378 void __init devinet_init(void)
2380 int i;
2382 for (i = 0; i < IN4_ADDR_HSIZE; i++)
2383 INIT_HLIST_HEAD(&inet_addr_lst[i]);
2385 register_pernet_subsys(&devinet_ops);
2387 register_gifconf(PF_INET, inet_gifconf);
2388 register_netdevice_notifier(&ip_netdev_notifier);
2390 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
2392 rtnl_af_register(&inet_af_ops);
2394 rtnl_register(PF_INET, RTM_NEWADDR, inet_rtm_newaddr, NULL, NULL);
2395 rtnl_register(PF_INET, RTM_DELADDR, inet_rtm_deladdr, NULL, NULL);
2396 rtnl_register(PF_INET, RTM_GETADDR, NULL, inet_dump_ifaddr, NULL);
2397 rtnl_register(PF_INET, RTM_GETNETCONF, inet_netconf_get_devconf,
2398 inet_netconf_dump_devconf, NULL);