1 /* SPDX-License-Identifier: GPL-2.0 */
2 #include <linux/kernel.h>
3 #include <linux/init.h>
4 #include <linux/module.h>
5 #include <linux/seqlock.h>
6 #include <linux/netlink.h>
7 #include <linux/netfilter.h>
8 #include <linux/netfilter/nf_tables.h>
9 #include <net/netfilter/nf_tables.h>
10 #include <net/dst_metadata.h>
11 #include <net/ip_tunnels.h>
12 #include <net/vxlan.h>
13 #include <net/erspan.h>
16 enum nft_tunnel_keys key
:8;
17 enum nft_registers dreg
:8;
18 enum nft_tunnel_mode mode
:8;
21 static void nft_tunnel_get_eval(const struct nft_expr
*expr
,
22 struct nft_regs
*regs
,
23 const struct nft_pktinfo
*pkt
)
25 const struct nft_tunnel
*priv
= nft_expr_priv(expr
);
26 u32
*dest
= ®s
->data
[priv
->dreg
];
27 struct ip_tunnel_info
*tun_info
;
29 tun_info
= skb_tunnel_info(pkt
->skb
);
34 nft_reg_store8(dest
, false);
37 if (priv
->mode
== NFT_TUNNEL_MODE_NONE
||
38 (priv
->mode
== NFT_TUNNEL_MODE_RX
&&
39 !(tun_info
->mode
& IP_TUNNEL_INFO_TX
)) ||
40 (priv
->mode
== NFT_TUNNEL_MODE_TX
&&
41 (tun_info
->mode
& IP_TUNNEL_INFO_TX
)))
42 nft_reg_store8(dest
, true);
44 nft_reg_store8(dest
, false);
48 regs
->verdict
.code
= NFT_BREAK
;
51 if (priv
->mode
== NFT_TUNNEL_MODE_NONE
||
52 (priv
->mode
== NFT_TUNNEL_MODE_RX
&&
53 !(tun_info
->mode
& IP_TUNNEL_INFO_TX
)) ||
54 (priv
->mode
== NFT_TUNNEL_MODE_TX
&&
55 (tun_info
->mode
& IP_TUNNEL_INFO_TX
)))
56 *dest
= ntohl(tunnel_id_to_key32(tun_info
->key
.tun_id
));
58 regs
->verdict
.code
= NFT_BREAK
;
62 regs
->verdict
.code
= NFT_BREAK
;
66 static const struct nla_policy nft_tunnel_policy
[NFTA_TUNNEL_MAX
+ 1] = {
67 [NFTA_TUNNEL_KEY
] = { .type
= NLA_U32
},
68 [NFTA_TUNNEL_DREG
] = { .type
= NLA_U32
},
69 [NFTA_TUNNEL_MODE
] = { .type
= NLA_U32
},
72 static int nft_tunnel_get_init(const struct nft_ctx
*ctx
,
73 const struct nft_expr
*expr
,
74 const struct nlattr
* const tb
[])
76 struct nft_tunnel
*priv
= nft_expr_priv(expr
);
79 if (!tb
[NFTA_TUNNEL_KEY
] ||
80 !tb
[NFTA_TUNNEL_DREG
])
83 priv
->key
= ntohl(nla_get_be32(tb
[NFTA_TUNNEL_KEY
]));
95 priv
->dreg
= nft_parse_register(tb
[NFTA_TUNNEL_DREG
]);
97 if (tb
[NFTA_TUNNEL_MODE
]) {
98 priv
->mode
= ntohl(nla_get_be32(tb
[NFTA_TUNNEL_MODE
]));
99 if (priv
->mode
> NFT_TUNNEL_MODE_MAX
)
102 priv
->mode
= NFT_TUNNEL_MODE_NONE
;
105 return nft_validate_register_store(ctx
, priv
->dreg
, NULL
,
106 NFT_DATA_VALUE
, len
);
109 static int nft_tunnel_get_dump(struct sk_buff
*skb
,
110 const struct nft_expr
*expr
)
112 const struct nft_tunnel
*priv
= nft_expr_priv(expr
);
114 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY
, htonl(priv
->key
)))
115 goto nla_put_failure
;
116 if (nft_dump_register(skb
, NFTA_TUNNEL_DREG
, priv
->dreg
))
117 goto nla_put_failure
;
118 if (nla_put_be32(skb
, NFTA_TUNNEL_MODE
, htonl(priv
->mode
)))
119 goto nla_put_failure
;
126 static struct nft_expr_type nft_tunnel_type
;
127 static const struct nft_expr_ops nft_tunnel_get_ops
= {
128 .type
= &nft_tunnel_type
,
129 .size
= NFT_EXPR_SIZE(sizeof(struct nft_tunnel
)),
130 .eval
= nft_tunnel_get_eval
,
131 .init
= nft_tunnel_get_init
,
132 .dump
= nft_tunnel_get_dump
,
135 static struct nft_expr_type nft_tunnel_type __read_mostly
= {
137 .ops
= &nft_tunnel_get_ops
,
138 .policy
= nft_tunnel_policy
,
139 .maxattr
= NFTA_TUNNEL_MAX
,
140 .owner
= THIS_MODULE
,
143 struct nft_tunnel_opts
{
145 struct vxlan_metadata vxlan
;
146 struct erspan_metadata erspan
;
152 struct nft_tunnel_obj
{
153 struct metadata_dst
*md
;
154 struct nft_tunnel_opts opts
;
157 static const struct nla_policy nft_tunnel_ip_policy
[NFTA_TUNNEL_KEY_IP_MAX
+ 1] = {
158 [NFTA_TUNNEL_KEY_IP_SRC
] = { .type
= NLA_U32
},
159 [NFTA_TUNNEL_KEY_IP_DST
] = { .type
= NLA_U32
},
162 static int nft_tunnel_obj_ip_init(const struct nft_ctx
*ctx
,
163 const struct nlattr
*attr
,
164 struct ip_tunnel_info
*info
)
166 struct nlattr
*tb
[NFTA_TUNNEL_KEY_IP_MAX
+ 1];
169 err
= nla_parse_nested_deprecated(tb
, NFTA_TUNNEL_KEY_IP_MAX
, attr
,
170 nft_tunnel_ip_policy
, NULL
);
174 if (!tb
[NFTA_TUNNEL_KEY_IP_DST
])
177 if (tb
[NFTA_TUNNEL_KEY_IP_SRC
])
178 info
->key
.u
.ipv4
.src
= nla_get_be32(tb
[NFTA_TUNNEL_KEY_IP_SRC
]);
179 if (tb
[NFTA_TUNNEL_KEY_IP_DST
])
180 info
->key
.u
.ipv4
.dst
= nla_get_be32(tb
[NFTA_TUNNEL_KEY_IP_DST
]);
185 static const struct nla_policy nft_tunnel_ip6_policy
[NFTA_TUNNEL_KEY_IP6_MAX
+ 1] = {
186 [NFTA_TUNNEL_KEY_IP6_SRC
] = { .len
= sizeof(struct in6_addr
), },
187 [NFTA_TUNNEL_KEY_IP6_DST
] = { .len
= sizeof(struct in6_addr
), },
188 [NFTA_TUNNEL_KEY_IP6_FLOWLABEL
] = { .type
= NLA_U32
, }
191 static int nft_tunnel_obj_ip6_init(const struct nft_ctx
*ctx
,
192 const struct nlattr
*attr
,
193 struct ip_tunnel_info
*info
)
195 struct nlattr
*tb
[NFTA_TUNNEL_KEY_IP6_MAX
+ 1];
198 err
= nla_parse_nested_deprecated(tb
, NFTA_TUNNEL_KEY_IP6_MAX
, attr
,
199 nft_tunnel_ip6_policy
, NULL
);
203 if (!tb
[NFTA_TUNNEL_KEY_IP6_DST
])
206 if (tb
[NFTA_TUNNEL_KEY_IP6_SRC
]) {
207 memcpy(&info
->key
.u
.ipv6
.src
,
208 nla_data(tb
[NFTA_TUNNEL_KEY_IP6_SRC
]),
209 sizeof(struct in6_addr
));
211 if (tb
[NFTA_TUNNEL_KEY_IP6_DST
]) {
212 memcpy(&info
->key
.u
.ipv6
.dst
,
213 nla_data(tb
[NFTA_TUNNEL_KEY_IP6_DST
]),
214 sizeof(struct in6_addr
));
216 if (tb
[NFTA_TUNNEL_KEY_IP6_FLOWLABEL
])
217 info
->key
.label
= nla_get_be32(tb
[NFTA_TUNNEL_KEY_IP6_FLOWLABEL
]);
219 info
->mode
|= IP_TUNNEL_INFO_IPV6
;
224 static const struct nla_policy nft_tunnel_opts_vxlan_policy
[NFTA_TUNNEL_KEY_VXLAN_MAX
+ 1] = {
225 [NFTA_TUNNEL_KEY_VXLAN_GBP
] = { .type
= NLA_U32
},
228 static int nft_tunnel_obj_vxlan_init(const struct nlattr
*attr
,
229 struct nft_tunnel_opts
*opts
)
231 struct nlattr
*tb
[NFTA_TUNNEL_KEY_VXLAN_MAX
+ 1];
234 err
= nla_parse_nested_deprecated(tb
, NFTA_TUNNEL_KEY_VXLAN_MAX
, attr
,
235 nft_tunnel_opts_vxlan_policy
, NULL
);
239 if (!tb
[NFTA_TUNNEL_KEY_VXLAN_GBP
])
242 opts
->u
.vxlan
.gbp
= ntohl(nla_get_be32(tb
[NFTA_TUNNEL_KEY_VXLAN_GBP
]));
244 opts
->len
= sizeof(struct vxlan_metadata
);
245 opts
->flags
= TUNNEL_VXLAN_OPT
;
250 static const struct nla_policy nft_tunnel_opts_erspan_policy
[NFTA_TUNNEL_KEY_ERSPAN_MAX
+ 1] = {
251 [NFTA_TUNNEL_KEY_ERSPAN_VERSION
] = { .type
= NLA_U32
},
252 [NFTA_TUNNEL_KEY_ERSPAN_V1_INDEX
] = { .type
= NLA_U32
},
253 [NFTA_TUNNEL_KEY_ERSPAN_V2_DIR
] = { .type
= NLA_U8
},
254 [NFTA_TUNNEL_KEY_ERSPAN_V2_HWID
] = { .type
= NLA_U8
},
257 static int nft_tunnel_obj_erspan_init(const struct nlattr
*attr
,
258 struct nft_tunnel_opts
*opts
)
260 struct nlattr
*tb
[NFTA_TUNNEL_KEY_ERSPAN_MAX
+ 1];
264 err
= nla_parse_nested_deprecated(tb
, NFTA_TUNNEL_KEY_ERSPAN_MAX
,
265 attr
, nft_tunnel_opts_erspan_policy
,
270 if (!tb
[NFTA_TUNNEL_KEY_ERSPAN_VERSION
])
273 version
= ntohl(nla_get_be32(tb
[NFTA_TUNNEL_KEY_ERSPAN_VERSION
]));
276 if (!tb
[NFTA_TUNNEL_KEY_ERSPAN_V1_INDEX
])
279 opts
->u
.erspan
.u
.index
=
280 nla_get_be32(tb
[NFTA_TUNNEL_KEY_ERSPAN_V1_INDEX
]);
282 case ERSPAN_VERSION2
:
283 if (!tb
[NFTA_TUNNEL_KEY_ERSPAN_V2_DIR
] ||
284 !tb
[NFTA_TUNNEL_KEY_ERSPAN_V2_HWID
])
287 hwid
= nla_get_u8(tb
[NFTA_TUNNEL_KEY_ERSPAN_V2_HWID
]);
288 dir
= nla_get_u8(tb
[NFTA_TUNNEL_KEY_ERSPAN_V2_DIR
]);
290 set_hwid(&opts
->u
.erspan
.u
.md2
, hwid
);
291 opts
->u
.erspan
.u
.md2
.dir
= dir
;
296 opts
->u
.erspan
.version
= version
;
298 opts
->len
= sizeof(struct erspan_metadata
);
299 opts
->flags
= TUNNEL_ERSPAN_OPT
;
304 static const struct nla_policy nft_tunnel_opts_policy
[NFTA_TUNNEL_KEY_OPTS_MAX
+ 1] = {
305 [NFTA_TUNNEL_KEY_OPTS_VXLAN
] = { .type
= NLA_NESTED
, },
306 [NFTA_TUNNEL_KEY_OPTS_ERSPAN
] = { .type
= NLA_NESTED
, },
309 static int nft_tunnel_obj_opts_init(const struct nft_ctx
*ctx
,
310 const struct nlattr
*attr
,
311 struct ip_tunnel_info
*info
,
312 struct nft_tunnel_opts
*opts
)
314 struct nlattr
*tb
[NFTA_TUNNEL_KEY_OPTS_MAX
+ 1];
317 err
= nla_parse_nested_deprecated(tb
, NFTA_TUNNEL_KEY_OPTS_MAX
, attr
,
318 nft_tunnel_opts_policy
, NULL
);
322 if (tb
[NFTA_TUNNEL_KEY_OPTS_VXLAN
]) {
323 err
= nft_tunnel_obj_vxlan_init(tb
[NFTA_TUNNEL_KEY_OPTS_VXLAN
],
325 } else if (tb
[NFTA_TUNNEL_KEY_OPTS_ERSPAN
]) {
326 err
= nft_tunnel_obj_erspan_init(tb
[NFTA_TUNNEL_KEY_OPTS_ERSPAN
],
335 static const struct nla_policy nft_tunnel_key_policy
[NFTA_TUNNEL_KEY_MAX
+ 1] = {
336 [NFTA_TUNNEL_KEY_IP
] = { .type
= NLA_NESTED
, },
337 [NFTA_TUNNEL_KEY_IP6
] = { .type
= NLA_NESTED
, },
338 [NFTA_TUNNEL_KEY_ID
] = { .type
= NLA_U32
, },
339 [NFTA_TUNNEL_KEY_FLAGS
] = { .type
= NLA_U32
, },
340 [NFTA_TUNNEL_KEY_TOS
] = { .type
= NLA_U8
, },
341 [NFTA_TUNNEL_KEY_TTL
] = { .type
= NLA_U8
, },
342 [NFTA_TUNNEL_KEY_OPTS
] = { .type
= NLA_NESTED
, },
345 static int nft_tunnel_obj_init(const struct nft_ctx
*ctx
,
346 const struct nlattr
* const tb
[],
347 struct nft_object
*obj
)
349 struct nft_tunnel_obj
*priv
= nft_obj_data(obj
);
350 struct ip_tunnel_info info
;
351 struct metadata_dst
*md
;
354 if (!tb
[NFTA_TUNNEL_KEY_ID
])
357 memset(&info
, 0, sizeof(info
));
358 info
.mode
= IP_TUNNEL_INFO_TX
;
359 info
.key
.tun_id
= key32_to_tunnel_id(nla_get_be32(tb
[NFTA_TUNNEL_KEY_ID
]));
360 info
.key
.tun_flags
= TUNNEL_KEY
| TUNNEL_CSUM
| TUNNEL_NOCACHE
;
362 if (tb
[NFTA_TUNNEL_KEY_IP
]) {
363 err
= nft_tunnel_obj_ip_init(ctx
, tb
[NFTA_TUNNEL_KEY_IP
], &info
);
366 } else if (tb
[NFTA_TUNNEL_KEY_IP6
]) {
367 err
= nft_tunnel_obj_ip6_init(ctx
, tb
[NFTA_TUNNEL_KEY_IP6
], &info
);
374 if (tb
[NFTA_TUNNEL_KEY_SPORT
]) {
375 info
.key
.tp_src
= nla_get_be16(tb
[NFTA_TUNNEL_KEY_SPORT
]);
377 if (tb
[NFTA_TUNNEL_KEY_DPORT
]) {
378 info
.key
.tp_dst
= nla_get_be16(tb
[NFTA_TUNNEL_KEY_DPORT
]);
381 if (tb
[NFTA_TUNNEL_KEY_FLAGS
]) {
384 tun_flags
= ntohl(nla_get_be32(tb
[NFTA_TUNNEL_KEY_FLAGS
]));
385 if (tun_flags
& ~NFT_TUNNEL_F_MASK
)
388 if (tun_flags
& NFT_TUNNEL_F_ZERO_CSUM_TX
)
389 info
.key
.tun_flags
&= ~TUNNEL_CSUM
;
390 if (tun_flags
& NFT_TUNNEL_F_DONT_FRAGMENT
)
391 info
.key
.tun_flags
|= TUNNEL_DONT_FRAGMENT
;
392 if (tun_flags
& NFT_TUNNEL_F_SEQ_NUMBER
)
393 info
.key
.tun_flags
|= TUNNEL_SEQ
;
395 if (tb
[NFTA_TUNNEL_KEY_TOS
])
396 info
.key
.tos
= nla_get_u8(tb
[NFTA_TUNNEL_KEY_TOS
]);
397 if (tb
[NFTA_TUNNEL_KEY_TTL
])
398 info
.key
.ttl
= nla_get_u8(tb
[NFTA_TUNNEL_KEY_TTL
]);
400 info
.key
.ttl
= U8_MAX
;
402 if (tb
[NFTA_TUNNEL_KEY_OPTS
]) {
403 err
= nft_tunnel_obj_opts_init(ctx
, tb
[NFTA_TUNNEL_KEY_OPTS
],
409 md
= metadata_dst_alloc(priv
->opts
.len
, METADATA_IP_TUNNEL
, GFP_KERNEL
);
413 memcpy(&md
->u
.tun_info
, &info
, sizeof(info
));
414 #ifdef CONFIG_DST_CACHE
415 err
= dst_cache_init(&md
->u
.tun_info
.dst_cache
, GFP_KERNEL
);
417 metadata_dst_free(md
);
421 ip_tunnel_info_opts_set(&md
->u
.tun_info
, &priv
->opts
.u
, priv
->opts
.len
,
428 static inline void nft_tunnel_obj_eval(struct nft_object
*obj
,
429 struct nft_regs
*regs
,
430 const struct nft_pktinfo
*pkt
)
432 struct nft_tunnel_obj
*priv
= nft_obj_data(obj
);
433 struct sk_buff
*skb
= pkt
->skb
;
436 dst_hold((struct dst_entry
*) priv
->md
);
437 skb_dst_set(skb
, (struct dst_entry
*) priv
->md
);
440 static int nft_tunnel_ip_dump(struct sk_buff
*skb
, struct ip_tunnel_info
*info
)
444 if (info
->mode
& IP_TUNNEL_INFO_IPV6
) {
445 nest
= nla_nest_start_noflag(skb
, NFTA_TUNNEL_KEY_IP6
);
449 if (nla_put_in6_addr(skb
, NFTA_TUNNEL_KEY_IP6_SRC
,
450 &info
->key
.u
.ipv6
.src
) < 0 ||
451 nla_put_in6_addr(skb
, NFTA_TUNNEL_KEY_IP6_DST
,
452 &info
->key
.u
.ipv6
.dst
) < 0 ||
453 nla_put_be32(skb
, NFTA_TUNNEL_KEY_IP6_FLOWLABEL
,
455 nla_nest_cancel(skb
, nest
);
459 nla_nest_end(skb
, nest
);
461 nest
= nla_nest_start_noflag(skb
, NFTA_TUNNEL_KEY_IP
);
465 if (nla_put_in_addr(skb
, NFTA_TUNNEL_KEY_IP_SRC
,
466 info
->key
.u
.ipv4
.src
) < 0 ||
467 nla_put_in_addr(skb
, NFTA_TUNNEL_KEY_IP_DST
,
468 info
->key
.u
.ipv4
.dst
) < 0) {
469 nla_nest_cancel(skb
, nest
);
473 nla_nest_end(skb
, nest
);
479 static int nft_tunnel_opts_dump(struct sk_buff
*skb
,
480 struct nft_tunnel_obj
*priv
)
482 struct nft_tunnel_opts
*opts
= &priv
->opts
;
483 struct nlattr
*nest
, *inner
;
485 nest
= nla_nest_start_noflag(skb
, NFTA_TUNNEL_KEY_OPTS
);
489 if (opts
->flags
& TUNNEL_VXLAN_OPT
) {
490 inner
= nla_nest_start_noflag(skb
, NFTA_TUNNEL_KEY_OPTS_VXLAN
);
493 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY_VXLAN_GBP
,
494 htonl(opts
->u
.vxlan
.gbp
)))
496 nla_nest_end(skb
, inner
);
497 } else if (opts
->flags
& TUNNEL_ERSPAN_OPT
) {
498 inner
= nla_nest_start_noflag(skb
, NFTA_TUNNEL_KEY_OPTS_ERSPAN
);
501 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY_ERSPAN_VERSION
,
502 htonl(opts
->u
.erspan
.version
)))
504 switch (opts
->u
.erspan
.version
) {
506 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY_ERSPAN_V1_INDEX
,
507 opts
->u
.erspan
.u
.index
))
510 case ERSPAN_VERSION2
:
511 if (nla_put_u8(skb
, NFTA_TUNNEL_KEY_ERSPAN_V2_HWID
,
512 get_hwid(&opts
->u
.erspan
.u
.md2
)) ||
513 nla_put_u8(skb
, NFTA_TUNNEL_KEY_ERSPAN_V2_DIR
,
514 opts
->u
.erspan
.u
.md2
.dir
))
518 nla_nest_end(skb
, inner
);
520 nla_nest_end(skb
, nest
);
524 nla_nest_cancel(skb
, inner
);
526 nla_nest_cancel(skb
, nest
);
530 static int nft_tunnel_ports_dump(struct sk_buff
*skb
,
531 struct ip_tunnel_info
*info
)
533 if (nla_put_be16(skb
, NFTA_TUNNEL_KEY_SPORT
, info
->key
.tp_src
) < 0 ||
534 nla_put_be16(skb
, NFTA_TUNNEL_KEY_DPORT
, info
->key
.tp_dst
) < 0)
540 static int nft_tunnel_flags_dump(struct sk_buff
*skb
,
541 struct ip_tunnel_info
*info
)
545 if (info
->key
.tun_flags
& TUNNEL_DONT_FRAGMENT
)
546 flags
|= NFT_TUNNEL_F_DONT_FRAGMENT
;
547 if (!(info
->key
.tun_flags
& TUNNEL_CSUM
))
548 flags
|= NFT_TUNNEL_F_ZERO_CSUM_TX
;
549 if (info
->key
.tun_flags
& TUNNEL_SEQ
)
550 flags
|= NFT_TUNNEL_F_SEQ_NUMBER
;
552 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY_FLAGS
, htonl(flags
)) < 0)
558 static int nft_tunnel_obj_dump(struct sk_buff
*skb
,
559 struct nft_object
*obj
, bool reset
)
561 struct nft_tunnel_obj
*priv
= nft_obj_data(obj
);
562 struct ip_tunnel_info
*info
= &priv
->md
->u
.tun_info
;
564 if (nla_put_be32(skb
, NFTA_TUNNEL_KEY_ID
,
565 tunnel_id_to_key32(info
->key
.tun_id
)) ||
566 nft_tunnel_ip_dump(skb
, info
) < 0 ||
567 nft_tunnel_ports_dump(skb
, info
) < 0 ||
568 nft_tunnel_flags_dump(skb
, info
) < 0 ||
569 nla_put_u8(skb
, NFTA_TUNNEL_KEY_TOS
, info
->key
.tos
) ||
570 nla_put_u8(skb
, NFTA_TUNNEL_KEY_TTL
, info
->key
.ttl
) ||
571 nft_tunnel_opts_dump(skb
, priv
) < 0)
572 goto nla_put_failure
;
580 static void nft_tunnel_obj_destroy(const struct nft_ctx
*ctx
,
581 struct nft_object
*obj
)
583 struct nft_tunnel_obj
*priv
= nft_obj_data(obj
);
585 metadata_dst_free(priv
->md
);
588 static struct nft_object_type nft_tunnel_obj_type
;
589 static const struct nft_object_ops nft_tunnel_obj_ops
= {
590 .type
= &nft_tunnel_obj_type
,
591 .size
= sizeof(struct nft_tunnel_obj
),
592 .eval
= nft_tunnel_obj_eval
,
593 .init
= nft_tunnel_obj_init
,
594 .destroy
= nft_tunnel_obj_destroy
,
595 .dump
= nft_tunnel_obj_dump
,
598 static struct nft_object_type nft_tunnel_obj_type __read_mostly
= {
599 .type
= NFT_OBJECT_TUNNEL
,
600 .ops
= &nft_tunnel_obj_ops
,
601 .maxattr
= NFTA_TUNNEL_KEY_MAX
,
602 .policy
= nft_tunnel_key_policy
,
603 .owner
= THIS_MODULE
,
606 static int __init
nft_tunnel_module_init(void)
610 err
= nft_register_expr(&nft_tunnel_type
);
614 err
= nft_register_obj(&nft_tunnel_obj_type
);
616 nft_unregister_expr(&nft_tunnel_type
);
621 static void __exit
nft_tunnel_module_exit(void)
623 nft_unregister_obj(&nft_tunnel_obj_type
);
624 nft_unregister_expr(&nft_tunnel_type
);
627 module_init(nft_tunnel_module_init
);
628 module_exit(nft_tunnel_module_exit
);
630 MODULE_LICENSE("GPL");
631 MODULE_AUTHOR("Pablo Neira Ayuso <pablo@netfilter.org>");
632 MODULE_ALIAS_NFT_EXPR("tunnel");
633 MODULE_ALIAS_NFT_OBJ(NFT_OBJECT_TUNNEL
);