perf bpf: Move perf_event_output() from stdio.h to bpf.h
[linux/fpc-iii.git] / net / ipv4 / ipip.c
blobe65287c27e3d85f17a03d7958f197b54528b9d10
1 /*
2 * Linux NET3: IP/IP protocol decoder.
4 * Authors:
5 * Sam Lantinga (slouken@cs.ucdavis.edu) 02/01/95
7 * Fixes:
8 * Alan Cox : Merged and made usable non modular (its so tiny its silly as
9 * a module taking up 2 pages).
10 * Alan Cox : Fixed bug with 1.3.18 and IPIP not working (now needs to set skb->h.iph)
11 * to keep ip_forward happy.
12 * Alan Cox : More fixes for 1.3.21, and firewall fix. Maybe this will work soon 8).
13 * Kai Schulte : Fixed #defines for IP_FIREWALL->FIREWALL
14 * David Woodhouse : Perform some basic ICMP handling.
15 * IPIP Routing without decapsulation.
16 * Carlos Picoto : GRE over IP support
17 * Alexey Kuznetsov: Reworked. Really, now it is truncated version of ipv4/ip_gre.c.
18 * I do not want to merge them together.
20 * This program is free software; you can redistribute it and/or
21 * modify it under the terms of the GNU General Public License
22 * as published by the Free Software Foundation; either version
23 * 2 of the License, or (at your option) any later version.
27 /* tunnel.c: an IP tunnel driver
29 The purpose of this driver is to provide an IP tunnel through
30 which you can tunnel network traffic transparently across subnets.
32 This was written by looking at Nick Holloway's dummy driver
33 Thanks for the great code!
35 -Sam Lantinga (slouken@cs.ucdavis.edu) 02/01/95
37 Minor tweaks:
38 Cleaned up the code a little and added some pre-1.3.0 tweaks.
39 dev->hard_header/hard_header_len changed to use no headers.
40 Comments/bracketing tweaked.
41 Made the tunnels use dev->name not tunnel: when error reporting.
42 Added tx_dropped stat
44 -Alan Cox (alan@lxorguk.ukuu.org.uk) 21 March 95
46 Reworked:
47 Changed to tunnel to destination gateway in addition to the
48 tunnel's pointopoint address
49 Almost completely rewritten
50 Note: There is currently no firewall or ICMP handling done.
52 -Sam Lantinga (slouken@cs.ucdavis.edu) 02/13/96
56 /* Things I wish I had known when writing the tunnel driver:
58 When the tunnel_xmit() function is called, the skb contains the
59 packet to be sent (plus a great deal of extra info), and dev
60 contains the tunnel device that _we_ are.
62 When we are passed a packet, we are expected to fill in the
63 source address with our source IP address.
65 What is the proper way to allocate, copy and free a buffer?
66 After you allocate it, it is a "0 length" chunk of memory
67 starting at zero. If you want to add headers to the buffer
68 later, you'll have to call "skb_reserve(skb, amount)" with
69 the amount of memory you want reserved. Then, you call
70 "skb_put(skb, amount)" with the amount of space you want in
71 the buffer. skb_put() returns a pointer to the top (#0) of
72 that buffer. skb->len is set to the amount of space you have
73 "allocated" with skb_put(). You can then write up to skb->len
74 bytes to that buffer. If you need more, you can call skb_put()
75 again with the additional amount of space you need. You can
76 find out how much more space you can allocate by calling
77 "skb_tailroom(skb)".
78 Now, to add header space, call "skb_push(skb, header_len)".
79 This creates space at the beginning of the buffer and returns
80 a pointer to this new space. If later you need to strip a
81 header from a buffer, call "skb_pull(skb, header_len)".
82 skb_headroom() will return how much space is left at the top
83 of the buffer (before the main data). Remember, this headroom
84 space must be reserved before the skb_put() function is called.
88 This version of net/ipv4/ipip.c is cloned of net/ipv4/ip_gre.c
90 For comments look at net/ipv4/ip_gre.c --ANK
94 #include <linux/capability.h>
95 #include <linux/module.h>
96 #include <linux/types.h>
97 #include <linux/kernel.h>
98 #include <linux/slab.h>
99 #include <linux/uaccess.h>
100 #include <linux/skbuff.h>
101 #include <linux/netdevice.h>
102 #include <linux/in.h>
103 #include <linux/tcp.h>
104 #include <linux/udp.h>
105 #include <linux/if_arp.h>
106 #include <linux/init.h>
107 #include <linux/netfilter_ipv4.h>
108 #include <linux/if_ether.h>
110 #include <net/sock.h>
111 #include <net/ip.h>
112 #include <net/icmp.h>
113 #include <net/ip_tunnels.h>
114 #include <net/inet_ecn.h>
115 #include <net/xfrm.h>
116 #include <net/net_namespace.h>
117 #include <net/netns/generic.h>
118 #include <net/dst_metadata.h>
120 static bool log_ecn_error = true;
121 module_param(log_ecn_error, bool, 0644);
122 MODULE_PARM_DESC(log_ecn_error, "Log packets received with corrupted ECN");
124 static unsigned int ipip_net_id __read_mostly;
126 static int ipip_tunnel_init(struct net_device *dev);
127 static struct rtnl_link_ops ipip_link_ops __read_mostly;
129 static int ipip_err(struct sk_buff *skb, u32 info)
131 /* All the routers (except for Linux) return only
132 * 8 bytes of packet payload. It means, that precise relaying of
133 * ICMP in the real Internet is absolutely infeasible.
135 struct net *net = dev_net(skb->dev);
136 struct ip_tunnel_net *itn = net_generic(net, ipip_net_id);
137 const struct iphdr *iph = (const struct iphdr *)skb->data;
138 const int type = icmp_hdr(skb)->type;
139 const int code = icmp_hdr(skb)->code;
140 struct ip_tunnel *t;
141 int err = 0;
143 switch (type) {
144 case ICMP_DEST_UNREACH:
145 switch (code) {
146 case ICMP_SR_FAILED:
147 /* Impossible event. */
148 goto out;
149 default:
150 /* All others are translated to HOST_UNREACH.
151 * rfc2003 contains "deep thoughts" about NET_UNREACH,
152 * I believe they are just ether pollution. --ANK
154 break;
156 break;
158 case ICMP_TIME_EXCEEDED:
159 if (code != ICMP_EXC_TTL)
160 goto out;
161 break;
163 case ICMP_REDIRECT:
164 break;
166 default:
167 goto out;
170 t = ip_tunnel_lookup(itn, skb->dev->ifindex, TUNNEL_NO_KEY,
171 iph->daddr, iph->saddr, 0);
172 if (!t) {
173 err = -ENOENT;
174 goto out;
177 if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
178 ipv4_update_pmtu(skb, net, info, t->parms.link, iph->protocol);
179 goto out;
182 if (type == ICMP_REDIRECT) {
183 ipv4_redirect(skb, net, t->parms.link, iph->protocol);
184 goto out;
187 if (t->parms.iph.daddr == 0) {
188 err = -ENOENT;
189 goto out;
192 if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
193 goto out;
195 if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO))
196 t->err_count++;
197 else
198 t->err_count = 1;
199 t->err_time = jiffies;
201 out:
202 return err;
205 static const struct tnl_ptk_info ipip_tpi = {
206 /* no tunnel info required for ipip. */
207 .proto = htons(ETH_P_IP),
210 #if IS_ENABLED(CONFIG_MPLS)
211 static const struct tnl_ptk_info mplsip_tpi = {
212 /* no tunnel info required for mplsip. */
213 .proto = htons(ETH_P_MPLS_UC),
215 #endif
217 static int ipip_tunnel_rcv(struct sk_buff *skb, u8 ipproto)
219 struct net *net = dev_net(skb->dev);
220 struct ip_tunnel_net *itn = net_generic(net, ipip_net_id);
221 struct metadata_dst *tun_dst = NULL;
222 struct ip_tunnel *tunnel;
223 const struct iphdr *iph;
225 iph = ip_hdr(skb);
226 tunnel = ip_tunnel_lookup(itn, skb->dev->ifindex, TUNNEL_NO_KEY,
227 iph->saddr, iph->daddr, 0);
228 if (tunnel) {
229 const struct tnl_ptk_info *tpi;
231 if (tunnel->parms.iph.protocol != ipproto &&
232 tunnel->parms.iph.protocol != 0)
233 goto drop;
235 if (!xfrm4_policy_check(NULL, XFRM_POLICY_IN, skb))
236 goto drop;
237 #if IS_ENABLED(CONFIG_MPLS)
238 if (ipproto == IPPROTO_MPLS)
239 tpi = &mplsip_tpi;
240 else
241 #endif
242 tpi = &ipip_tpi;
243 if (iptunnel_pull_header(skb, 0, tpi->proto, false))
244 goto drop;
245 if (tunnel->collect_md) {
246 tun_dst = ip_tun_rx_dst(skb, 0, 0, 0);
247 if (!tun_dst)
248 return 0;
250 return ip_tunnel_rcv(tunnel, skb, tpi, tun_dst, log_ecn_error);
253 return -1;
255 drop:
256 kfree_skb(skb);
257 return 0;
260 static int ipip_rcv(struct sk_buff *skb)
262 return ipip_tunnel_rcv(skb, IPPROTO_IPIP);
265 #if IS_ENABLED(CONFIG_MPLS)
266 static int mplsip_rcv(struct sk_buff *skb)
268 return ipip_tunnel_rcv(skb, IPPROTO_MPLS);
270 #endif
273 * This function assumes it is being called from dev_queue_xmit()
274 * and that skb is filled properly by that function.
276 static netdev_tx_t ipip_tunnel_xmit(struct sk_buff *skb,
277 struct net_device *dev)
279 struct ip_tunnel *tunnel = netdev_priv(dev);
280 const struct iphdr *tiph = &tunnel->parms.iph;
281 u8 ipproto;
283 switch (skb->protocol) {
284 case htons(ETH_P_IP):
285 ipproto = IPPROTO_IPIP;
286 break;
287 #if IS_ENABLED(CONFIG_MPLS)
288 case htons(ETH_P_MPLS_UC):
289 ipproto = IPPROTO_MPLS;
290 break;
291 #endif
292 default:
293 goto tx_error;
296 if (tiph->protocol != ipproto && tiph->protocol != 0)
297 goto tx_error;
299 if (iptunnel_handle_offloads(skb, SKB_GSO_IPXIP4))
300 goto tx_error;
302 skb_set_inner_ipproto(skb, ipproto);
304 if (tunnel->collect_md)
305 ip_md_tunnel_xmit(skb, dev, ipproto);
306 else
307 ip_tunnel_xmit(skb, dev, tiph, ipproto);
308 return NETDEV_TX_OK;
310 tx_error:
311 kfree_skb(skb);
313 dev->stats.tx_errors++;
314 return NETDEV_TX_OK;
317 static bool ipip_tunnel_ioctl_verify_protocol(u8 ipproto)
319 switch (ipproto) {
320 case 0:
321 case IPPROTO_IPIP:
322 #if IS_ENABLED(CONFIG_MPLS)
323 case IPPROTO_MPLS:
324 #endif
325 return true;
328 return false;
331 static int
332 ipip_tunnel_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
334 int err = 0;
335 struct ip_tunnel_parm p;
337 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p)))
338 return -EFAULT;
340 if (cmd == SIOCADDTUNNEL || cmd == SIOCCHGTUNNEL) {
341 if (p.iph.version != 4 ||
342 !ipip_tunnel_ioctl_verify_protocol(p.iph.protocol) ||
343 p.iph.ihl != 5 || (p.iph.frag_off&htons(~IP_DF)))
344 return -EINVAL;
347 p.i_key = p.o_key = 0;
348 p.i_flags = p.o_flags = 0;
349 err = ip_tunnel_ioctl(dev, &p, cmd);
350 if (err)
351 return err;
353 if (copy_to_user(ifr->ifr_ifru.ifru_data, &p, sizeof(p)))
354 return -EFAULT;
356 return 0;
359 static const struct net_device_ops ipip_netdev_ops = {
360 .ndo_init = ipip_tunnel_init,
361 .ndo_uninit = ip_tunnel_uninit,
362 .ndo_start_xmit = ipip_tunnel_xmit,
363 .ndo_do_ioctl = ipip_tunnel_ioctl,
364 .ndo_change_mtu = ip_tunnel_change_mtu,
365 .ndo_get_stats64 = ip_tunnel_get_stats64,
366 .ndo_get_iflink = ip_tunnel_get_iflink,
369 #define IPIP_FEATURES (NETIF_F_SG | \
370 NETIF_F_FRAGLIST | \
371 NETIF_F_HIGHDMA | \
372 NETIF_F_GSO_SOFTWARE | \
373 NETIF_F_HW_CSUM)
375 static void ipip_tunnel_setup(struct net_device *dev)
377 dev->netdev_ops = &ipip_netdev_ops;
379 dev->type = ARPHRD_TUNNEL;
380 dev->flags = IFF_NOARP;
381 dev->addr_len = 4;
382 dev->features |= NETIF_F_LLTX;
383 netif_keep_dst(dev);
385 dev->features |= IPIP_FEATURES;
386 dev->hw_features |= IPIP_FEATURES;
387 ip_tunnel_setup(dev, ipip_net_id);
390 static int ipip_tunnel_init(struct net_device *dev)
392 struct ip_tunnel *tunnel = netdev_priv(dev);
394 memcpy(dev->dev_addr, &tunnel->parms.iph.saddr, 4);
395 memcpy(dev->broadcast, &tunnel->parms.iph.daddr, 4);
397 tunnel->tun_hlen = 0;
398 tunnel->hlen = tunnel->tun_hlen + tunnel->encap_hlen;
399 return ip_tunnel_init(dev);
402 static int ipip_tunnel_validate(struct nlattr *tb[], struct nlattr *data[],
403 struct netlink_ext_ack *extack)
405 u8 proto;
407 if (!data || !data[IFLA_IPTUN_PROTO])
408 return 0;
410 proto = nla_get_u8(data[IFLA_IPTUN_PROTO]);
411 if (proto != IPPROTO_IPIP && proto != IPPROTO_MPLS && proto != 0)
412 return -EINVAL;
414 return 0;
417 static void ipip_netlink_parms(struct nlattr *data[],
418 struct ip_tunnel_parm *parms, bool *collect_md,
419 __u32 *fwmark)
421 memset(parms, 0, sizeof(*parms));
423 parms->iph.version = 4;
424 parms->iph.protocol = IPPROTO_IPIP;
425 parms->iph.ihl = 5;
426 *collect_md = false;
428 if (!data)
429 return;
431 if (data[IFLA_IPTUN_LINK])
432 parms->link = nla_get_u32(data[IFLA_IPTUN_LINK]);
434 if (data[IFLA_IPTUN_LOCAL])
435 parms->iph.saddr = nla_get_in_addr(data[IFLA_IPTUN_LOCAL]);
437 if (data[IFLA_IPTUN_REMOTE])
438 parms->iph.daddr = nla_get_in_addr(data[IFLA_IPTUN_REMOTE]);
440 if (data[IFLA_IPTUN_TTL]) {
441 parms->iph.ttl = nla_get_u8(data[IFLA_IPTUN_TTL]);
442 if (parms->iph.ttl)
443 parms->iph.frag_off = htons(IP_DF);
446 if (data[IFLA_IPTUN_TOS])
447 parms->iph.tos = nla_get_u8(data[IFLA_IPTUN_TOS]);
449 if (data[IFLA_IPTUN_PROTO])
450 parms->iph.protocol = nla_get_u8(data[IFLA_IPTUN_PROTO]);
452 if (!data[IFLA_IPTUN_PMTUDISC] || nla_get_u8(data[IFLA_IPTUN_PMTUDISC]))
453 parms->iph.frag_off = htons(IP_DF);
455 if (data[IFLA_IPTUN_COLLECT_METADATA])
456 *collect_md = true;
458 if (data[IFLA_IPTUN_FWMARK])
459 *fwmark = nla_get_u32(data[IFLA_IPTUN_FWMARK]);
462 /* This function returns true when ENCAP attributes are present in the nl msg */
463 static bool ipip_netlink_encap_parms(struct nlattr *data[],
464 struct ip_tunnel_encap *ipencap)
466 bool ret = false;
468 memset(ipencap, 0, sizeof(*ipencap));
470 if (!data)
471 return ret;
473 if (data[IFLA_IPTUN_ENCAP_TYPE]) {
474 ret = true;
475 ipencap->type = nla_get_u16(data[IFLA_IPTUN_ENCAP_TYPE]);
478 if (data[IFLA_IPTUN_ENCAP_FLAGS]) {
479 ret = true;
480 ipencap->flags = nla_get_u16(data[IFLA_IPTUN_ENCAP_FLAGS]);
483 if (data[IFLA_IPTUN_ENCAP_SPORT]) {
484 ret = true;
485 ipencap->sport = nla_get_be16(data[IFLA_IPTUN_ENCAP_SPORT]);
488 if (data[IFLA_IPTUN_ENCAP_DPORT]) {
489 ret = true;
490 ipencap->dport = nla_get_be16(data[IFLA_IPTUN_ENCAP_DPORT]);
493 return ret;
496 static int ipip_newlink(struct net *src_net, struct net_device *dev,
497 struct nlattr *tb[], struct nlattr *data[],
498 struct netlink_ext_ack *extack)
500 struct ip_tunnel *t = netdev_priv(dev);
501 struct ip_tunnel_parm p;
502 struct ip_tunnel_encap ipencap;
503 __u32 fwmark = 0;
505 if (ipip_netlink_encap_parms(data, &ipencap)) {
506 int err = ip_tunnel_encap_setup(t, &ipencap);
508 if (err < 0)
509 return err;
512 ipip_netlink_parms(data, &p, &t->collect_md, &fwmark);
513 return ip_tunnel_newlink(dev, tb, &p, fwmark);
516 static int ipip_changelink(struct net_device *dev, struct nlattr *tb[],
517 struct nlattr *data[],
518 struct netlink_ext_ack *extack)
520 struct ip_tunnel *t = netdev_priv(dev);
521 struct ip_tunnel_parm p;
522 struct ip_tunnel_encap ipencap;
523 bool collect_md;
524 __u32 fwmark = t->fwmark;
526 if (ipip_netlink_encap_parms(data, &ipencap)) {
527 int err = ip_tunnel_encap_setup(t, &ipencap);
529 if (err < 0)
530 return err;
533 ipip_netlink_parms(data, &p, &collect_md, &fwmark);
534 if (collect_md)
535 return -EINVAL;
537 if (((dev->flags & IFF_POINTOPOINT) && !p.iph.daddr) ||
538 (!(dev->flags & IFF_POINTOPOINT) && p.iph.daddr))
539 return -EINVAL;
541 return ip_tunnel_changelink(dev, tb, &p, fwmark);
544 static size_t ipip_get_size(const struct net_device *dev)
546 return
547 /* IFLA_IPTUN_LINK */
548 nla_total_size(4) +
549 /* IFLA_IPTUN_LOCAL */
550 nla_total_size(4) +
551 /* IFLA_IPTUN_REMOTE */
552 nla_total_size(4) +
553 /* IFLA_IPTUN_TTL */
554 nla_total_size(1) +
555 /* IFLA_IPTUN_TOS */
556 nla_total_size(1) +
557 /* IFLA_IPTUN_PROTO */
558 nla_total_size(1) +
559 /* IFLA_IPTUN_PMTUDISC */
560 nla_total_size(1) +
561 /* IFLA_IPTUN_ENCAP_TYPE */
562 nla_total_size(2) +
563 /* IFLA_IPTUN_ENCAP_FLAGS */
564 nla_total_size(2) +
565 /* IFLA_IPTUN_ENCAP_SPORT */
566 nla_total_size(2) +
567 /* IFLA_IPTUN_ENCAP_DPORT */
568 nla_total_size(2) +
569 /* IFLA_IPTUN_COLLECT_METADATA */
570 nla_total_size(0) +
571 /* IFLA_IPTUN_FWMARK */
572 nla_total_size(4) +
576 static int ipip_fill_info(struct sk_buff *skb, const struct net_device *dev)
578 struct ip_tunnel *tunnel = netdev_priv(dev);
579 struct ip_tunnel_parm *parm = &tunnel->parms;
581 if (nla_put_u32(skb, IFLA_IPTUN_LINK, parm->link) ||
582 nla_put_in_addr(skb, IFLA_IPTUN_LOCAL, parm->iph.saddr) ||
583 nla_put_in_addr(skb, IFLA_IPTUN_REMOTE, parm->iph.daddr) ||
584 nla_put_u8(skb, IFLA_IPTUN_TTL, parm->iph.ttl) ||
585 nla_put_u8(skb, IFLA_IPTUN_TOS, parm->iph.tos) ||
586 nla_put_u8(skb, IFLA_IPTUN_PROTO, parm->iph.protocol) ||
587 nla_put_u8(skb, IFLA_IPTUN_PMTUDISC,
588 !!(parm->iph.frag_off & htons(IP_DF))) ||
589 nla_put_u32(skb, IFLA_IPTUN_FWMARK, tunnel->fwmark))
590 goto nla_put_failure;
592 if (nla_put_u16(skb, IFLA_IPTUN_ENCAP_TYPE,
593 tunnel->encap.type) ||
594 nla_put_be16(skb, IFLA_IPTUN_ENCAP_SPORT,
595 tunnel->encap.sport) ||
596 nla_put_be16(skb, IFLA_IPTUN_ENCAP_DPORT,
597 tunnel->encap.dport) ||
598 nla_put_u16(skb, IFLA_IPTUN_ENCAP_FLAGS,
599 tunnel->encap.flags))
600 goto nla_put_failure;
602 if (tunnel->collect_md)
603 if (nla_put_flag(skb, IFLA_IPTUN_COLLECT_METADATA))
604 goto nla_put_failure;
605 return 0;
607 nla_put_failure:
608 return -EMSGSIZE;
611 static const struct nla_policy ipip_policy[IFLA_IPTUN_MAX + 1] = {
612 [IFLA_IPTUN_LINK] = { .type = NLA_U32 },
613 [IFLA_IPTUN_LOCAL] = { .type = NLA_U32 },
614 [IFLA_IPTUN_REMOTE] = { .type = NLA_U32 },
615 [IFLA_IPTUN_TTL] = { .type = NLA_U8 },
616 [IFLA_IPTUN_TOS] = { .type = NLA_U8 },
617 [IFLA_IPTUN_PROTO] = { .type = NLA_U8 },
618 [IFLA_IPTUN_PMTUDISC] = { .type = NLA_U8 },
619 [IFLA_IPTUN_ENCAP_TYPE] = { .type = NLA_U16 },
620 [IFLA_IPTUN_ENCAP_FLAGS] = { .type = NLA_U16 },
621 [IFLA_IPTUN_ENCAP_SPORT] = { .type = NLA_U16 },
622 [IFLA_IPTUN_ENCAP_DPORT] = { .type = NLA_U16 },
623 [IFLA_IPTUN_COLLECT_METADATA] = { .type = NLA_FLAG },
624 [IFLA_IPTUN_FWMARK] = { .type = NLA_U32 },
627 static struct rtnl_link_ops ipip_link_ops __read_mostly = {
628 .kind = "ipip",
629 .maxtype = IFLA_IPTUN_MAX,
630 .policy = ipip_policy,
631 .priv_size = sizeof(struct ip_tunnel),
632 .setup = ipip_tunnel_setup,
633 .validate = ipip_tunnel_validate,
634 .newlink = ipip_newlink,
635 .changelink = ipip_changelink,
636 .dellink = ip_tunnel_dellink,
637 .get_size = ipip_get_size,
638 .fill_info = ipip_fill_info,
639 .get_link_net = ip_tunnel_get_link_net,
642 static struct xfrm_tunnel ipip_handler __read_mostly = {
643 .handler = ipip_rcv,
644 .err_handler = ipip_err,
645 .priority = 1,
648 #if IS_ENABLED(CONFIG_MPLS)
649 static struct xfrm_tunnel mplsip_handler __read_mostly = {
650 .handler = mplsip_rcv,
651 .err_handler = ipip_err,
652 .priority = 1,
654 #endif
656 static int __net_init ipip_init_net(struct net *net)
658 return ip_tunnel_init_net(net, ipip_net_id, &ipip_link_ops, "tunl0");
661 static void __net_exit ipip_exit_batch_net(struct list_head *list_net)
663 ip_tunnel_delete_nets(list_net, ipip_net_id, &ipip_link_ops);
666 static struct pernet_operations ipip_net_ops = {
667 .init = ipip_init_net,
668 .exit_batch = ipip_exit_batch_net,
669 .id = &ipip_net_id,
670 .size = sizeof(struct ip_tunnel_net),
673 static int __init ipip_init(void)
675 int err;
677 pr_info("ipip: IPv4 and MPLS over IPv4 tunneling driver\n");
679 err = register_pernet_device(&ipip_net_ops);
680 if (err < 0)
681 return err;
682 err = xfrm4_tunnel_register(&ipip_handler, AF_INET);
683 if (err < 0) {
684 pr_info("%s: can't register tunnel\n", __func__);
685 goto xfrm_tunnel_ipip_failed;
687 #if IS_ENABLED(CONFIG_MPLS)
688 err = xfrm4_tunnel_register(&mplsip_handler, AF_MPLS);
689 if (err < 0) {
690 pr_info("%s: can't register tunnel\n", __func__);
691 goto xfrm_tunnel_mplsip_failed;
693 #endif
694 err = rtnl_link_register(&ipip_link_ops);
695 if (err < 0)
696 goto rtnl_link_failed;
698 out:
699 return err;
701 rtnl_link_failed:
702 #if IS_ENABLED(CONFIG_MPLS)
703 xfrm4_tunnel_deregister(&mplsip_handler, AF_INET);
704 xfrm_tunnel_mplsip_failed:
706 #endif
707 xfrm4_tunnel_deregister(&ipip_handler, AF_INET);
708 xfrm_tunnel_ipip_failed:
709 unregister_pernet_device(&ipip_net_ops);
710 goto out;
713 static void __exit ipip_fini(void)
715 rtnl_link_unregister(&ipip_link_ops);
716 if (xfrm4_tunnel_deregister(&ipip_handler, AF_INET))
717 pr_info("%s: can't deregister tunnel\n", __func__);
718 #if IS_ENABLED(CONFIG_MPLS)
719 if (xfrm4_tunnel_deregister(&mplsip_handler, AF_MPLS))
720 pr_info("%s: can't deregister tunnel\n", __func__);
721 #endif
722 unregister_pernet_device(&ipip_net_ops);
725 module_init(ipip_init);
726 module_exit(ipip_fini);
727 MODULE_LICENSE("GPL");
728 MODULE_ALIAS_RTNL_LINK("ipip");
729 MODULE_ALIAS_NETDEV("tunl0");