2 * Extension Header handling for IPv6
3 * Linux INET6 implementation
6 * Pedro Roque <roque@di.fc.ul.pt>
7 * Andi Kleen <ak@muc.de>
8 * Alexey Kuznetsov <kuznet@ms2.inr.ac.ru>
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License
12 * as published by the Free Software Foundation; either version
13 * 2 of the License, or (at your option) any later version.
17 * yoshfuji : ensure not to overrun while parsing
19 * Mitsuru KANDA @USAGI and: Remove ipv6_parse_exthdrs().
20 * YOSHIFUJI Hideaki @USAGI Register inbound extension header
21 * handlers as inet6_protocol{}.
24 #include <linux/errno.h>
25 #include <linux/types.h>
26 #include <linux/socket.h>
27 #include <linux/sockios.h>
28 #include <linux/net.h>
29 #include <linux/netdevice.h>
30 #include <linux/in6.h>
31 #include <linux/icmpv6.h>
32 #include <linux/slab.h>
33 #include <linux/export.h>
40 #include <net/protocol.h>
41 #include <net/transp_v6.h>
42 #include <net/rawv6.h>
43 #include <net/ndisc.h>
44 #include <net/ip6_route.h>
45 #include <net/addrconf.h>
46 #include <net/calipso.h>
47 #if IS_ENABLED(CONFIG_IPV6_MIP6)
50 #include <linux/seg6.h>
52 #ifdef CONFIG_IPV6_SEG6_HMAC
53 #include <net/seg6_hmac.h>
56 #include <linux/uaccess.h>
59 * Parsing tlv encoded headers.
61 * Parsing function "func" returns true, if parsing succeed
62 * and false, if it failed.
63 * It MUST NOT touch skb->h.
68 bool (*func
)(struct sk_buff
*skb
, int offset
);
71 /*********************
73 *********************/
75 /* An unknown option is detected, decide what to do */
77 static bool ip6_tlvopt_unknown(struct sk_buff
*skb
, int optoff
,
78 bool disallow_unknowns
)
80 if (disallow_unknowns
) {
81 /* If unknown TLVs are disallowed by configuration
82 * then always silently drop packet. Note this also
83 * means no ICMP parameter problem is sent which
84 * could be a good property to mitigate a reflection DOS
91 switch ((skb_network_header(skb
)[optoff
] & 0xC0) >> 6) {
95 case 1: /* drop packet */
98 case 3: /* Send ICMP if not a multicast address and drop packet */
99 /* Actually, it is redundant check. icmp_send
100 will recheck in any case.
102 if (ipv6_addr_is_multicast(&ipv6_hdr(skb
)->daddr
))
105 case 2: /* send ICMP PARM PROB regardless and drop packet */
106 icmpv6_param_prob(skb
, ICMPV6_UNK_OPTION
, optoff
);
115 /* Parse tlv encoded option header (hop-by-hop or destination) */
117 static bool ip6_parse_tlv(const struct tlvtype_proc
*procs
,
121 int len
= (skb_transport_header(skb
)[1] + 1) << 3;
122 const unsigned char *nh
= skb_network_header(skb
);
123 int off
= skb_network_header_len(skb
);
124 const struct tlvtype_proc
*curr
;
125 bool disallow_unknowns
= false;
129 if (unlikely(max_count
< 0)) {
130 disallow_unknowns
= true;
131 max_count
= -max_count
;
134 if (skb_transport_offset(skb
) + len
> skb_headlen(skb
))
141 int optlen
= nh
[off
+ 1] + 2;
153 /* RFC 2460 states that the purpose of PadN is
154 * to align the containing header to multiples
155 * of 8. 7 is therefore the highest valid value.
156 * See also RFC 4942, Section 2.1.9.5.
161 /* RFC 4942 recommends receiving hosts to
162 * actively check PadN payload to contain
165 for (i
= 2; i
< optlen
; i
++) {
166 if (nh
[off
+ i
] != 0)
171 default: /* Other TLV code so scan list */
176 if (tlv_count
> max_count
)
179 for (curr
= procs
; curr
->type
>= 0; curr
++) {
180 if (curr
->type
== nh
[off
]) {
181 /* type specific length/alignment
182 checks will be performed in the
184 if (curr
->func(skb
, off
) == false)
189 if (curr
->type
< 0 &&
190 !ip6_tlvopt_unknown(skb
, off
, disallow_unknowns
))
207 /*****************************
208 Destination options header.
209 *****************************/
211 #if IS_ENABLED(CONFIG_IPV6_MIP6)
212 static bool ipv6_dest_hao(struct sk_buff
*skb
, int optoff
)
214 struct ipv6_destopt_hao
*hao
;
215 struct inet6_skb_parm
*opt
= IP6CB(skb
);
216 struct ipv6hdr
*ipv6h
= ipv6_hdr(skb
);
220 net_dbg_ratelimited("hao duplicated\n");
223 opt
->dsthao
= opt
->dst1
;
226 hao
= (struct ipv6_destopt_hao
*)(skb_network_header(skb
) + optoff
);
228 if (hao
->length
!= 16) {
229 net_dbg_ratelimited("hao invalid option length = %d\n",
234 if (!(ipv6_addr_type(&hao
->addr
) & IPV6_ADDR_UNICAST
)) {
235 net_dbg_ratelimited("hao is not an unicast addr: %pI6\n",
240 ret
= xfrm6_input_addr(skb
, (xfrm_address_t
*)&ipv6h
->daddr
,
241 (xfrm_address_t
*)&hao
->addr
, IPPROTO_DSTOPTS
);
242 if (unlikely(ret
< 0))
245 if (skb_cloned(skb
)) {
246 if (pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
))
249 /* update all variable using below by copied skbuff */
250 hao
= (struct ipv6_destopt_hao
*)(skb_network_header(skb
) +
252 ipv6h
= ipv6_hdr(skb
);
255 if (skb
->ip_summed
== CHECKSUM_COMPLETE
)
256 skb
->ip_summed
= CHECKSUM_NONE
;
258 swap(ipv6h
->saddr
, hao
->addr
);
260 if (skb
->tstamp
== 0)
261 __net_timestamp(skb
);
271 static const struct tlvtype_proc tlvprocdestopt_lst
[] = {
272 #if IS_ENABLED(CONFIG_IPV6_MIP6)
274 .type
= IPV6_TLV_HAO
,
275 .func
= ipv6_dest_hao
,
281 static int ipv6_destopt_rcv(struct sk_buff
*skb
)
283 struct inet6_skb_parm
*opt
= IP6CB(skb
);
284 #if IS_ENABLED(CONFIG_IPV6_MIP6)
287 struct dst_entry
*dst
= skb_dst(skb
);
288 struct net
*net
= dev_net(skb
->dev
);
291 if (!pskb_may_pull(skb
, skb_transport_offset(skb
) + 8) ||
292 !pskb_may_pull(skb
, (skb_transport_offset(skb
) +
293 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
294 __IP6_INC_STATS(dev_net(dst
->dev
), ip6_dst_idev(dst
),
295 IPSTATS_MIB_INHDRERRORS
);
301 extlen
= (skb_transport_header(skb
)[1] + 1) << 3;
302 if (extlen
> net
->ipv6
.sysctl
.max_dst_opts_len
)
305 opt
->lastopt
= opt
->dst1
= skb_network_header_len(skb
);
306 #if IS_ENABLED(CONFIG_IPV6_MIP6)
310 if (ip6_parse_tlv(tlvprocdestopt_lst
, skb
,
311 init_net
.ipv6
.sysctl
.max_dst_opts_cnt
)) {
312 skb
->transport_header
+= extlen
;
314 #if IS_ENABLED(CONFIG_IPV6_MIP6)
317 opt
->nhoff
= opt
->dst1
;
322 __IP6_INC_STATS(dev_net(dst
->dev
),
323 ip6_dst_idev(dst
), IPSTATS_MIB_INHDRERRORS
);
327 static void seg6_update_csum(struct sk_buff
*skb
)
329 struct ipv6_sr_hdr
*hdr
;
330 struct in6_addr
*addr
;
333 /* srh is at transport offset and seg_left is already decremented
334 * but daddr is not yet updated with next segment
337 hdr
= (struct ipv6_sr_hdr
*)skb_transport_header(skb
);
338 addr
= hdr
->segments
+ hdr
->segments_left
;
340 hdr
->segments_left
++;
341 from
= *(__be32
*)hdr
;
343 hdr
->segments_left
--;
346 /* update skb csum with diff resulting from seg_left decrement */
348 update_csum_diff4(skb
, from
, to
);
350 /* compute csum diff between current and next segment and update */
352 update_csum_diff16(skb
, (__be32
*)(&ipv6_hdr(skb
)->daddr
),
356 static int ipv6_srh_rcv(struct sk_buff
*skb
)
358 struct inet6_skb_parm
*opt
= IP6CB(skb
);
359 struct net
*net
= dev_net(skb
->dev
);
360 struct ipv6_sr_hdr
*hdr
;
361 struct inet6_dev
*idev
;
362 struct in6_addr
*addr
;
365 hdr
= (struct ipv6_sr_hdr
*)skb_transport_header(skb
);
367 idev
= __in6_dev_get(skb
->dev
);
369 accept_seg6
= net
->ipv6
.devconf_all
->seg6_enabled
;
370 if (accept_seg6
> idev
->cnf
.seg6_enabled
)
371 accept_seg6
= idev
->cnf
.seg6_enabled
;
378 #ifdef CONFIG_IPV6_SEG6_HMAC
379 if (!seg6_hmac_validate_skb(skb
)) {
386 if (hdr
->segments_left
== 0) {
387 if (hdr
->nexthdr
== NEXTHDR_IPV6
) {
388 int offset
= (hdr
->hdrlen
+ 1) << 3;
390 skb_postpull_rcsum(skb
, skb_network_header(skb
),
391 skb_network_header_len(skb
));
393 if (!pskb_pull(skb
, offset
)) {
397 skb_postpull_rcsum(skb
, skb_transport_header(skb
),
400 skb_reset_network_header(skb
);
401 skb_reset_transport_header(skb
);
402 skb
->encapsulation
= 0;
404 __skb_tunnel_rx(skb
, skb
->dev
, net
);
410 opt
->srcrt
= skb_network_header_len(skb
);
411 opt
->lastopt
= opt
->srcrt
;
412 skb
->transport_header
+= (hdr
->hdrlen
+ 1) << 3;
413 opt
->nhoff
= (&hdr
->nexthdr
) - skb_network_header(skb
);
418 if (hdr
->segments_left
>= (hdr
->hdrlen
>> 1)) {
419 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
420 IPSTATS_MIB_INHDRERRORS
);
421 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
,
422 ((&hdr
->segments_left
) -
423 skb_network_header(skb
)));
427 if (skb_cloned(skb
)) {
428 if (pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
)) {
429 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
430 IPSTATS_MIB_OUTDISCARDS
);
436 hdr
= (struct ipv6_sr_hdr
*)skb_transport_header(skb
);
438 hdr
->segments_left
--;
439 addr
= hdr
->segments
+ hdr
->segments_left
;
441 skb_push(skb
, sizeof(struct ipv6hdr
));
443 if (skb
->ip_summed
== CHECKSUM_COMPLETE
)
444 seg6_update_csum(skb
);
446 ipv6_hdr(skb
)->daddr
= *addr
;
450 ip6_route_input(skb
);
452 if (skb_dst(skb
)->error
) {
457 if (skb_dst(skb
)->dev
->flags
& IFF_LOOPBACK
) {
458 if (ipv6_hdr(skb
)->hop_limit
<= 1) {
459 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
460 IPSTATS_MIB_INHDRERRORS
);
461 icmpv6_send(skb
, ICMPV6_TIME_EXCEED
,
462 ICMPV6_EXC_HOPLIMIT
, 0);
466 ipv6_hdr(skb
)->hop_limit
--;
468 skb_pull(skb
, sizeof(struct ipv6hdr
));
477 /********************************
479 ********************************/
481 /* called with rcu_read_lock() */
482 static int ipv6_rthdr_rcv(struct sk_buff
*skb
)
484 struct inet6_skb_parm
*opt
= IP6CB(skb
);
485 struct in6_addr
*addr
= NULL
;
486 struct in6_addr daddr
;
487 struct inet6_dev
*idev
;
489 struct ipv6_rt_hdr
*hdr
;
490 struct rt0_hdr
*rthdr
;
491 struct net
*net
= dev_net(skb
->dev
);
492 int accept_source_route
= net
->ipv6
.devconf_all
->accept_source_route
;
494 idev
= __in6_dev_get(skb
->dev
);
495 if (idev
&& accept_source_route
> idev
->cnf
.accept_source_route
)
496 accept_source_route
= idev
->cnf
.accept_source_route
;
498 if (!pskb_may_pull(skb
, skb_transport_offset(skb
) + 8) ||
499 !pskb_may_pull(skb
, (skb_transport_offset(skb
) +
500 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
501 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
502 IPSTATS_MIB_INHDRERRORS
);
507 hdr
= (struct ipv6_rt_hdr
*)skb_transport_header(skb
);
509 if (ipv6_addr_is_multicast(&ipv6_hdr(skb
)->daddr
) ||
510 skb
->pkt_type
!= PACKET_HOST
) {
511 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
512 IPSTATS_MIB_INADDRERRORS
);
517 /* segment routing */
518 if (hdr
->type
== IPV6_SRCRT_TYPE_4
)
519 return ipv6_srh_rcv(skb
);
522 if (hdr
->segments_left
== 0) {
524 #if IS_ENABLED(CONFIG_IPV6_MIP6)
525 case IPV6_SRCRT_TYPE_2
:
526 /* Silently discard type 2 header unless it was
530 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
531 IPSTATS_MIB_INADDRERRORS
);
541 opt
->lastopt
= opt
->srcrt
= skb_network_header_len(skb
);
542 skb
->transport_header
+= (hdr
->hdrlen
+ 1) << 3;
543 opt
->dst0
= opt
->dst1
;
545 opt
->nhoff
= (&hdr
->nexthdr
) - skb_network_header(skb
);
550 #if IS_ENABLED(CONFIG_IPV6_MIP6)
551 case IPV6_SRCRT_TYPE_2
:
552 if (accept_source_route
< 0)
554 /* Silently discard invalid RTH type 2 */
555 if (hdr
->hdrlen
!= 2 || hdr
->segments_left
!= 1) {
556 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
557 IPSTATS_MIB_INHDRERRORS
);
568 * This is the routing header forwarding algorithm from
572 n
= hdr
->hdrlen
>> 1;
574 if (hdr
->segments_left
> n
) {
575 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
576 IPSTATS_MIB_INHDRERRORS
);
577 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
,
578 ((&hdr
->segments_left
) -
579 skb_network_header(skb
)));
583 /* We are about to mangle packet header. Be careful!
584 Do not damage packets queued somewhere.
586 if (skb_cloned(skb
)) {
587 /* the copy is a forwarded packet */
588 if (pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
)) {
589 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
590 IPSTATS_MIB_OUTDISCARDS
);
594 hdr
= (struct ipv6_rt_hdr
*)skb_transport_header(skb
);
597 if (skb
->ip_summed
== CHECKSUM_COMPLETE
)
598 skb
->ip_summed
= CHECKSUM_NONE
;
600 i
= n
- --hdr
->segments_left
;
602 rthdr
= (struct rt0_hdr
*) hdr
;
607 #if IS_ENABLED(CONFIG_IPV6_MIP6)
608 case IPV6_SRCRT_TYPE_2
:
609 if (xfrm6_input_addr(skb
, (xfrm_address_t
*)addr
,
610 (xfrm_address_t
*)&ipv6_hdr(skb
)->saddr
,
611 IPPROTO_ROUTING
) < 0) {
612 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
613 IPSTATS_MIB_INADDRERRORS
);
617 if (!ipv6_chk_home_addr(dev_net(skb_dst(skb
)->dev
), addr
)) {
618 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
619 IPSTATS_MIB_INADDRERRORS
);
629 if (ipv6_addr_is_multicast(addr
)) {
630 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
631 IPSTATS_MIB_INADDRERRORS
);
637 *addr
= ipv6_hdr(skb
)->daddr
;
638 ipv6_hdr(skb
)->daddr
= daddr
;
641 ip6_route_input(skb
);
642 if (skb_dst(skb
)->error
) {
643 skb_push(skb
, skb
->data
- skb_network_header(skb
));
648 if (skb_dst(skb
)->dev
->flags
&IFF_LOOPBACK
) {
649 if (ipv6_hdr(skb
)->hop_limit
<= 1) {
650 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)),
651 IPSTATS_MIB_INHDRERRORS
);
652 icmpv6_send(skb
, ICMPV6_TIME_EXCEED
, ICMPV6_EXC_HOPLIMIT
,
657 ipv6_hdr(skb
)->hop_limit
--;
661 skb_push(skb
, skb
->data
- skb_network_header(skb
));
666 __IP6_INC_STATS(net
, ip6_dst_idev(skb_dst(skb
)), IPSTATS_MIB_INHDRERRORS
);
667 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
,
668 (&hdr
->type
) - skb_network_header(skb
));
672 static const struct inet6_protocol rthdr_protocol
= {
673 .handler
= ipv6_rthdr_rcv
,
674 .flags
= INET6_PROTO_NOPOLICY
,
677 static const struct inet6_protocol destopt_protocol
= {
678 .handler
= ipv6_destopt_rcv
,
679 .flags
= INET6_PROTO_NOPOLICY
,
682 static const struct inet6_protocol nodata_protocol
= {
683 .handler
= dst_discard
,
684 .flags
= INET6_PROTO_NOPOLICY
,
687 int __init
ipv6_exthdrs_init(void)
691 ret
= inet6_add_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
695 ret
= inet6_add_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
699 ret
= inet6_add_protocol(&nodata_protocol
, IPPROTO_NONE
);
706 inet6_del_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
708 inet6_del_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
712 void ipv6_exthdrs_exit(void)
714 inet6_del_protocol(&nodata_protocol
, IPPROTO_NONE
);
715 inet6_del_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
716 inet6_del_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
719 /**********************************
721 **********************************/
724 * Note: we cannot rely on skb_dst(skb) before we assign it in ip6_route_input().
726 static inline struct inet6_dev
*ipv6_skb_idev(struct sk_buff
*skb
)
728 return skb_dst(skb
) ? ip6_dst_idev(skb_dst(skb
)) : __in6_dev_get(skb
->dev
);
731 static inline struct net
*ipv6_skb_net(struct sk_buff
*skb
)
733 return skb_dst(skb
) ? dev_net(skb_dst(skb
)->dev
) : dev_net(skb
->dev
);
736 /* Router Alert as of RFC 2711 */
738 static bool ipv6_hop_ra(struct sk_buff
*skb
, int optoff
)
740 const unsigned char *nh
= skb_network_header(skb
);
742 if (nh
[optoff
+ 1] == 2) {
743 IP6CB(skb
)->flags
|= IP6SKB_ROUTERALERT
;
744 memcpy(&IP6CB(skb
)->ra
, nh
+ optoff
+ 2, sizeof(IP6CB(skb
)->ra
));
747 net_dbg_ratelimited("ipv6_hop_ra: wrong RA length %d\n",
755 static bool ipv6_hop_jumbo(struct sk_buff
*skb
, int optoff
)
757 const unsigned char *nh
= skb_network_header(skb
);
758 struct net
*net
= ipv6_skb_net(skb
);
761 if (nh
[optoff
+ 1] != 4 || (optoff
& 3) != 2) {
762 net_dbg_ratelimited("ipv6_hop_jumbo: wrong jumbo opt length/alignment %d\n",
764 __IP6_INC_STATS(net
, ipv6_skb_idev(skb
),
765 IPSTATS_MIB_INHDRERRORS
);
769 pkt_len
= ntohl(*(__be32
*)(nh
+ optoff
+ 2));
770 if (pkt_len
<= IPV6_MAXPLEN
) {
771 __IP6_INC_STATS(net
, ipv6_skb_idev(skb
),
772 IPSTATS_MIB_INHDRERRORS
);
773 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
, optoff
+2);
776 if (ipv6_hdr(skb
)->payload_len
) {
777 __IP6_INC_STATS(net
, ipv6_skb_idev(skb
),
778 IPSTATS_MIB_INHDRERRORS
);
779 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
, optoff
);
783 if (pkt_len
> skb
->len
- sizeof(struct ipv6hdr
)) {
784 __IP6_INC_STATS(net
, ipv6_skb_idev(skb
),
785 IPSTATS_MIB_INTRUNCATEDPKTS
);
789 if (pskb_trim_rcsum(skb
, pkt_len
+ sizeof(struct ipv6hdr
)))
792 IP6CB(skb
)->flags
|= IP6SKB_JUMBOGRAM
;
800 /* CALIPSO RFC 5570 */
802 static bool ipv6_hop_calipso(struct sk_buff
*skb
, int optoff
)
804 const unsigned char *nh
= skb_network_header(skb
);
806 if (nh
[optoff
+ 1] < 8)
809 if (nh
[optoff
+ 6] * 4 + 8 > nh
[optoff
+ 1])
812 if (!calipso_validate(skb
, nh
+ optoff
))
822 static const struct tlvtype_proc tlvprochopopt_lst
[] = {
824 .type
= IPV6_TLV_ROUTERALERT
,
828 .type
= IPV6_TLV_JUMBO
,
829 .func
= ipv6_hop_jumbo
,
832 .type
= IPV6_TLV_CALIPSO
,
833 .func
= ipv6_hop_calipso
,
838 int ipv6_parse_hopopts(struct sk_buff
*skb
)
840 struct inet6_skb_parm
*opt
= IP6CB(skb
);
841 struct net
*net
= dev_net(skb
->dev
);
845 * skb_network_header(skb) is equal to skb->data, and
846 * skb_network_header_len(skb) is always equal to
847 * sizeof(struct ipv6hdr) by definition of
848 * hop-by-hop options.
850 if (!pskb_may_pull(skb
, sizeof(struct ipv6hdr
) + 8) ||
851 !pskb_may_pull(skb
, (sizeof(struct ipv6hdr
) +
852 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
858 extlen
= (skb_transport_header(skb
)[1] + 1) << 3;
859 if (extlen
> net
->ipv6
.sysctl
.max_hbh_opts_len
)
862 opt
->flags
|= IP6SKB_HOPBYHOP
;
863 if (ip6_parse_tlv(tlvprochopopt_lst
, skb
,
864 init_net
.ipv6
.sysctl
.max_hbh_opts_cnt
)) {
865 skb
->transport_header
+= extlen
;
867 opt
->nhoff
= sizeof(struct ipv6hdr
);
874 * Creating outbound headers.
876 * "build" functions work when skb is filled from head to tail (datagram)
877 * "push" functions work when headers are added from tail to head (tcp)
879 * In both cases we assume, that caller reserved enough room
883 static void ipv6_push_rthdr0(struct sk_buff
*skb
, u8
*proto
,
884 struct ipv6_rt_hdr
*opt
,
885 struct in6_addr
**addr_p
, struct in6_addr
*saddr
)
887 struct rt0_hdr
*phdr
, *ihdr
;
890 ihdr
= (struct rt0_hdr
*) opt
;
892 phdr
= skb_push(skb
, (ihdr
->rt_hdr
.hdrlen
+ 1) << 3);
893 memcpy(phdr
, ihdr
, sizeof(struct rt0_hdr
));
895 hops
= ihdr
->rt_hdr
.hdrlen
>> 1;
898 memcpy(phdr
->addr
, ihdr
->addr
+ 1,
899 (hops
- 1) * sizeof(struct in6_addr
));
901 phdr
->addr
[hops
- 1] = **addr_p
;
902 *addr_p
= ihdr
->addr
;
904 phdr
->rt_hdr
.nexthdr
= *proto
;
905 *proto
= NEXTHDR_ROUTING
;
908 static void ipv6_push_rthdr4(struct sk_buff
*skb
, u8
*proto
,
909 struct ipv6_rt_hdr
*opt
,
910 struct in6_addr
**addr_p
, struct in6_addr
*saddr
)
912 struct ipv6_sr_hdr
*sr_phdr
, *sr_ihdr
;
915 sr_ihdr
= (struct ipv6_sr_hdr
*)opt
;
916 plen
= (sr_ihdr
->hdrlen
+ 1) << 3;
918 sr_phdr
= skb_push(skb
, plen
);
919 memcpy(sr_phdr
, sr_ihdr
, sizeof(struct ipv6_sr_hdr
));
921 hops
= sr_ihdr
->first_segment
+ 1;
922 memcpy(sr_phdr
->segments
+ 1, sr_ihdr
->segments
+ 1,
923 (hops
- 1) * sizeof(struct in6_addr
));
925 sr_phdr
->segments
[0] = **addr_p
;
926 *addr_p
= &sr_ihdr
->segments
[sr_ihdr
->segments_left
];
928 if (sr_ihdr
->hdrlen
> hops
* 2) {
929 int tlvs_offset
, tlvs_length
;
931 tlvs_offset
= (1 + hops
* 2) << 3;
932 tlvs_length
= (sr_ihdr
->hdrlen
- hops
* 2) << 3;
933 memcpy((char *)sr_phdr
+ tlvs_offset
,
934 (char *)sr_ihdr
+ tlvs_offset
, tlvs_length
);
937 #ifdef CONFIG_IPV6_SEG6_HMAC
938 if (sr_has_hmac(sr_phdr
)) {
939 struct net
*net
= NULL
;
942 net
= dev_net(skb
->dev
);
944 net
= sock_net(skb
->sk
);
949 seg6_push_hmac(net
, saddr
, sr_phdr
);
953 sr_phdr
->nexthdr
= *proto
;
954 *proto
= NEXTHDR_ROUTING
;
957 static void ipv6_push_rthdr(struct sk_buff
*skb
, u8
*proto
,
958 struct ipv6_rt_hdr
*opt
,
959 struct in6_addr
**addr_p
, struct in6_addr
*saddr
)
962 case IPV6_SRCRT_TYPE_0
:
963 case IPV6_SRCRT_STRICT
:
964 case IPV6_SRCRT_TYPE_2
:
965 ipv6_push_rthdr0(skb
, proto
, opt
, addr_p
, saddr
);
967 case IPV6_SRCRT_TYPE_4
:
968 ipv6_push_rthdr4(skb
, proto
, opt
, addr_p
, saddr
);
975 static void ipv6_push_exthdr(struct sk_buff
*skb
, u8
*proto
, u8 type
, struct ipv6_opt_hdr
*opt
)
977 struct ipv6_opt_hdr
*h
= skb_push(skb
, ipv6_optlen(opt
));
979 memcpy(h
, opt
, ipv6_optlen(opt
));
984 void ipv6_push_nfrag_opts(struct sk_buff
*skb
, struct ipv6_txoptions
*opt
,
986 struct in6_addr
**daddr
, struct in6_addr
*saddr
)
989 ipv6_push_rthdr(skb
, proto
, opt
->srcrt
, daddr
, saddr
);
991 * IPV6_RTHDRDSTOPTS is ignored
992 * unless IPV6_RTHDR is set (RFC3542).
995 ipv6_push_exthdr(skb
, proto
, NEXTHDR_DEST
, opt
->dst0opt
);
998 ipv6_push_exthdr(skb
, proto
, NEXTHDR_HOP
, opt
->hopopt
);
1001 void ipv6_push_frag_opts(struct sk_buff
*skb
, struct ipv6_txoptions
*opt
, u8
*proto
)
1004 ipv6_push_exthdr(skb
, proto
, NEXTHDR_DEST
, opt
->dst1opt
);
1006 EXPORT_SYMBOL(ipv6_push_frag_opts
);
1008 struct ipv6_txoptions
*
1009 ipv6_dup_options(struct sock
*sk
, struct ipv6_txoptions
*opt
)
1011 struct ipv6_txoptions
*opt2
;
1013 opt2
= sock_kmalloc(sk
, opt
->tot_len
, GFP_ATOMIC
);
1015 long dif
= (char *)opt2
- (char *)opt
;
1016 memcpy(opt2
, opt
, opt
->tot_len
);
1018 *((char **)&opt2
->hopopt
) += dif
;
1020 *((char **)&opt2
->dst0opt
) += dif
;
1022 *((char **)&opt2
->dst1opt
) += dif
;
1024 *((char **)&opt2
->srcrt
) += dif
;
1025 refcount_set(&opt2
->refcnt
, 1);
1029 EXPORT_SYMBOL_GPL(ipv6_dup_options
);
1031 static int ipv6_renew_option(void *ohdr
,
1032 struct ipv6_opt_hdr __user
*newopt
, int newoptlen
,
1034 struct ipv6_opt_hdr
**hdr
,
1039 memcpy(*p
, ohdr
, ipv6_optlen((struct ipv6_opt_hdr
*)ohdr
));
1040 *hdr
= (struct ipv6_opt_hdr
*)*p
;
1041 *p
+= CMSG_ALIGN(ipv6_optlen(*hdr
));
1045 if (copy_from_user(*p
, newopt
, newoptlen
))
1047 *hdr
= (struct ipv6_opt_hdr
*)*p
;
1048 if (ipv6_optlen(*hdr
) > newoptlen
)
1050 *p
+= CMSG_ALIGN(newoptlen
);
1057 * ipv6_renew_options - replace a specific ext hdr with a new one.
1059 * @sk: sock from which to allocate memory
1060 * @opt: original options
1061 * @newtype: option type to replace in @opt
1062 * @newopt: new option of type @newtype to replace (user-mem)
1063 * @newoptlen: length of @newopt
1065 * Returns a new set of options which is a copy of @opt with the
1066 * option type @newtype replaced with @newopt.
1068 * @opt may be NULL, in which case a new set of options is returned
1069 * containing just @newopt.
1071 * @newopt may be NULL, in which case the specified option type is
1072 * not copied into the new set of options.
1074 * The new set of options is allocated from the socket option memory
1077 struct ipv6_txoptions
*
1078 ipv6_renew_options(struct sock
*sk
, struct ipv6_txoptions
*opt
,
1080 struct ipv6_opt_hdr __user
*newopt
, int newoptlen
)
1084 struct ipv6_txoptions
*opt2
;
1088 if (newtype
!= IPV6_HOPOPTS
&& opt
->hopopt
)
1089 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->hopopt
));
1090 if (newtype
!= IPV6_RTHDRDSTOPTS
&& opt
->dst0opt
)
1091 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->dst0opt
));
1092 if (newtype
!= IPV6_RTHDR
&& opt
->srcrt
)
1093 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->srcrt
));
1094 if (newtype
!= IPV6_DSTOPTS
&& opt
->dst1opt
)
1095 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->dst1opt
));
1098 if (newopt
&& newoptlen
)
1099 tot_len
+= CMSG_ALIGN(newoptlen
);
1104 tot_len
+= sizeof(*opt2
);
1105 opt2
= sock_kmalloc(sk
, tot_len
, GFP_ATOMIC
);
1107 return ERR_PTR(-ENOBUFS
);
1109 memset(opt2
, 0, tot_len
);
1110 refcount_set(&opt2
->refcnt
, 1);
1111 opt2
->tot_len
= tot_len
;
1112 p
= (char *)(opt2
+ 1);
1114 err
= ipv6_renew_option(opt
? opt
->hopopt
: NULL
, newopt
, newoptlen
,
1115 newtype
!= IPV6_HOPOPTS
,
1120 err
= ipv6_renew_option(opt
? opt
->dst0opt
: NULL
, newopt
, newoptlen
,
1121 newtype
!= IPV6_RTHDRDSTOPTS
,
1122 &opt2
->dst0opt
, &p
);
1126 err
= ipv6_renew_option(opt
? opt
->srcrt
: NULL
, newopt
, newoptlen
,
1127 newtype
!= IPV6_RTHDR
,
1128 (struct ipv6_opt_hdr
**)&opt2
->srcrt
, &p
);
1132 err
= ipv6_renew_option(opt
? opt
->dst1opt
: NULL
, newopt
, newoptlen
,
1133 newtype
!= IPV6_DSTOPTS
,
1134 &opt2
->dst1opt
, &p
);
1138 opt2
->opt_nflen
= (opt2
->hopopt
? ipv6_optlen(opt2
->hopopt
) : 0) +
1139 (opt2
->dst0opt
? ipv6_optlen(opt2
->dst0opt
) : 0) +
1140 (opt2
->srcrt
? ipv6_optlen(opt2
->srcrt
) : 0);
1141 opt2
->opt_flen
= (opt2
->dst1opt
? ipv6_optlen(opt2
->dst1opt
) : 0);
1145 sock_kfree_s(sk
, opt2
, opt2
->tot_len
);
1146 return ERR_PTR(err
);
1150 * ipv6_renew_options_kern - replace a specific ext hdr with a new one.
1152 * @sk: sock from which to allocate memory
1153 * @opt: original options
1154 * @newtype: option type to replace in @opt
1155 * @newopt: new option of type @newtype to replace (kernel-mem)
1156 * @newoptlen: length of @newopt
1158 * See ipv6_renew_options(). The difference is that @newopt is
1159 * kernel memory, rather than user memory.
1161 struct ipv6_txoptions
*
1162 ipv6_renew_options_kern(struct sock
*sk
, struct ipv6_txoptions
*opt
,
1163 int newtype
, struct ipv6_opt_hdr
*newopt
,
1166 struct ipv6_txoptions
*ret_val
;
1167 const mm_segment_t old_fs
= get_fs();
1170 ret_val
= ipv6_renew_options(sk
, opt
, newtype
,
1171 (struct ipv6_opt_hdr __user
*)newopt
,
1177 struct ipv6_txoptions
*ipv6_fixup_options(struct ipv6_txoptions
*opt_space
,
1178 struct ipv6_txoptions
*opt
)
1181 * ignore the dest before srcrt unless srcrt is being included.
1184 if (opt
&& opt
->dst0opt
&& !opt
->srcrt
) {
1185 if (opt_space
!= opt
) {
1186 memcpy(opt_space
, opt
, sizeof(*opt_space
));
1189 opt
->opt_nflen
-= ipv6_optlen(opt
->dst0opt
);
1190 opt
->dst0opt
= NULL
;
1195 EXPORT_SYMBOL_GPL(ipv6_fixup_options
);
1198 * fl6_update_dst - update flowi destination address with info given
1199 * by srcrt option, if any.
1201 * @fl6: flowi6 for which daddr is to be updated
1202 * @opt: struct ipv6_txoptions in which to look for srcrt opt
1203 * @orig: copy of original daddr address if modified
1205 * Returns NULL if no txoptions or no srcrt, otherwise returns orig
1206 * and initial value of fl6->daddr set in orig
1208 struct in6_addr
*fl6_update_dst(struct flowi6
*fl6
,
1209 const struct ipv6_txoptions
*opt
,
1210 struct in6_addr
*orig
)
1212 if (!opt
|| !opt
->srcrt
)
1217 switch (opt
->srcrt
->type
) {
1218 case IPV6_SRCRT_TYPE_0
:
1219 case IPV6_SRCRT_STRICT
:
1220 case IPV6_SRCRT_TYPE_2
:
1221 fl6
->daddr
= *((struct rt0_hdr
*)opt
->srcrt
)->addr
;
1223 case IPV6_SRCRT_TYPE_4
:
1225 struct ipv6_sr_hdr
*srh
= (struct ipv6_sr_hdr
*)opt
->srcrt
;
1227 fl6
->daddr
= srh
->segments
[srh
->segments_left
];
1236 EXPORT_SYMBOL_GPL(fl6_update_dst
);