2 * Copyright (C)2002 USAGI/WIDE Project
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2 of the License, or
7 * (at your option) any later version.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, see <http://www.gnu.org/licenses/>.
19 * Mitsuru KANDA @USAGI : IPv6 Support
20 * Kazunori MIYAZAWA @USAGI :
21 * Kunihiro Ishiguro <kunihiro@ipinfusion.com>
23 * This file is derived from net/ipv4/ah.c.
26 #define pr_fmt(fmt) "IPv6: " fmt
28 #include <crypto/hash.h>
29 #include <linux/module.h>
30 #include <linux/slab.h>
33 #include <linux/crypto.h>
34 #include <linux/pfkeyv2.h>
35 #include <linux/string.h>
36 #include <linux/scatterlist.h>
37 #include <net/ip6_route.h>
40 #include <net/protocol.h>
43 #define IPV6HDR_BASELEN 8
46 #if IS_ENABLED(CONFIG_IPV6_MIP6)
47 struct in6_addr saddr
;
49 struct in6_addr daddr
;
54 struct xfrm_skb_cb xfrm
;
58 #define AH_SKB_CB(__skb) ((struct ah_skb_cb *)&((__skb)->cb[0]))
60 static void *ah_alloc_tmp(struct crypto_ahash
*ahash
, int nfrags
,
65 len
= size
+ crypto_ahash_digestsize(ahash
) +
66 (crypto_ahash_alignmask(ahash
) &
67 ~(crypto_tfm_ctx_alignment() - 1));
69 len
= ALIGN(len
, crypto_tfm_ctx_alignment());
71 len
+= sizeof(struct ahash_request
) + crypto_ahash_reqsize(ahash
);
72 len
= ALIGN(len
, __alignof__(struct scatterlist
));
74 len
+= sizeof(struct scatterlist
) * nfrags
;
76 return kmalloc(len
, GFP_ATOMIC
);
79 static inline struct tmp_ext
*ah_tmp_ext(void *base
)
81 return base
+ IPV6HDR_BASELEN
;
84 static inline u8
*ah_tmp_auth(u8
*tmp
, unsigned int offset
)
89 static inline u8
*ah_tmp_icv(struct crypto_ahash
*ahash
, void *tmp
,
92 return PTR_ALIGN((u8
*)tmp
+ offset
, crypto_ahash_alignmask(ahash
) + 1);
95 static inline struct ahash_request
*ah_tmp_req(struct crypto_ahash
*ahash
,
98 struct ahash_request
*req
;
100 req
= (void *)PTR_ALIGN(icv
+ crypto_ahash_digestsize(ahash
),
101 crypto_tfm_ctx_alignment());
103 ahash_request_set_tfm(req
, ahash
);
108 static inline struct scatterlist
*ah_req_sg(struct crypto_ahash
*ahash
,
109 struct ahash_request
*req
)
111 return (void *)ALIGN((unsigned long)(req
+ 1) +
112 crypto_ahash_reqsize(ahash
),
113 __alignof__(struct scatterlist
));
116 static bool zero_out_mutable_opts(struct ipv6_opt_hdr
*opthdr
)
118 u8
*opt
= (u8
*)opthdr
;
119 int len
= ipv6_optlen(opthdr
);
136 optlen
= opt
[off
+1]+2;
140 memset(&opt
[off
+2], 0, opt
[off
+1]);
154 #if IS_ENABLED(CONFIG_IPV6_MIP6)
156 * ipv6_rearrange_destopt - rearrange IPv6 destination options header
158 * @destopt: destionation options header
160 static void ipv6_rearrange_destopt(struct ipv6hdr
*iph
, struct ipv6_opt_hdr
*destopt
)
162 u8
*opt
= (u8
*)destopt
;
163 int len
= ipv6_optlen(destopt
);
180 optlen
= opt
[off
+1]+2;
184 /* Rearrange the source address in @iph and the
185 * addresses in home address option for final source.
186 * See 11.3.2 of RFC 3775 for details.
188 if (opt
[off
] == IPV6_TLV_HAO
) {
189 struct in6_addr final_addr
;
190 struct ipv6_destopt_hao
*hao
;
192 hao
= (struct ipv6_destopt_hao
*)&opt
[off
];
193 if (hao
->length
!= sizeof(hao
->addr
)) {
194 net_warn_ratelimited("destopt hao: invalid header length: %u\n",
198 final_addr
= hao
->addr
;
199 hao
->addr
= iph
->saddr
;
200 iph
->saddr
= final_addr
;
208 /* Note: ok if len == 0 */
213 static void ipv6_rearrange_destopt(struct ipv6hdr
*iph
, struct ipv6_opt_hdr
*destopt
) {}
217 * ipv6_rearrange_rthdr - rearrange IPv6 routing header
219 * @rthdr: routing header
221 * Rearrange the destination address in @iph and the addresses in @rthdr
222 * so that they appear in the order they will at the final destination.
223 * See Appendix A2 of RFC 2402 for details.
225 static void ipv6_rearrange_rthdr(struct ipv6hdr
*iph
, struct ipv6_rt_hdr
*rthdr
)
227 int segments
, segments_left
;
228 struct in6_addr
*addrs
;
229 struct in6_addr final_addr
;
231 segments_left
= rthdr
->segments_left
;
232 if (segments_left
== 0)
234 rthdr
->segments_left
= 0;
236 /* The value of rthdr->hdrlen has been verified either by the system
237 * call if it is locally generated, or by ipv6_rthdr_rcv() for incoming
238 * packets. So we can assume that it is even and that segments is
239 * greater than or equal to segments_left.
241 * For the same reason we can assume that this option is of type 0.
243 segments
= rthdr
->hdrlen
>> 1;
245 addrs
= ((struct rt0_hdr
*)rthdr
)->addr
;
246 final_addr
= addrs
[segments
- 1];
248 addrs
+= segments
- segments_left
;
249 memmove(addrs
+ 1, addrs
, (segments_left
- 1) * sizeof(*addrs
));
251 addrs
[0] = iph
->daddr
;
252 iph
->daddr
= final_addr
;
255 static int ipv6_clear_mutable_options(struct ipv6hdr
*iph
, int len
, int dir
)
259 struct ipv6_opt_hdr
*opth
;
260 struct ipv6_rt_hdr
*rth
;
262 } exthdr
= { .iph
= iph
};
263 char *end
= exthdr
.raw
+ len
;
264 int nexthdr
= iph
->nexthdr
;
268 while (exthdr
.raw
< end
) {
271 if (dir
== XFRM_POLICY_OUT
)
272 ipv6_rearrange_destopt(iph
, exthdr
.opth
);
274 if (!zero_out_mutable_opts(exthdr
.opth
)) {
275 net_dbg_ratelimited("overrun %sopts\n",
276 nexthdr
== NEXTHDR_HOP
?
282 case NEXTHDR_ROUTING
:
283 ipv6_rearrange_rthdr(iph
, exthdr
.rth
);
290 nexthdr
= exthdr
.opth
->nexthdr
;
291 exthdr
.raw
+= ipv6_optlen(exthdr
.opth
);
297 static void ah6_output_done(struct crypto_async_request
*base
, int err
)
302 struct sk_buff
*skb
= base
->data
;
303 struct xfrm_state
*x
= skb_dst(skb
)->xfrm
;
304 struct ah_data
*ahp
= x
->data
;
305 struct ipv6hdr
*top_iph
= ipv6_hdr(skb
);
306 struct ip_auth_hdr
*ah
= ip_auth_hdr(skb
);
307 struct tmp_ext
*iph_ext
;
309 extlen
= skb_network_header_len(skb
) - sizeof(struct ipv6hdr
);
311 extlen
+= sizeof(*iph_ext
);
313 iph_base
= AH_SKB_CB(skb
)->tmp
;
314 iph_ext
= ah_tmp_ext(iph_base
);
315 icv
= ah_tmp_icv(ahp
->ahash
, iph_ext
, extlen
);
317 memcpy(ah
->auth_data
, icv
, ahp
->icv_trunc_len
);
318 memcpy(top_iph
, iph_base
, IPV6HDR_BASELEN
);
321 #if IS_ENABLED(CONFIG_IPV6_MIP6)
322 memcpy(&top_iph
->saddr
, iph_ext
, extlen
);
324 memcpy(&top_iph
->daddr
, iph_ext
, extlen
);
328 kfree(AH_SKB_CB(skb
)->tmp
);
329 xfrm_output_resume(skb
, err
);
332 static int ah6_output(struct xfrm_state
*x
, struct sk_buff
*skb
)
340 struct sk_buff
*trailer
;
341 struct crypto_ahash
*ahash
;
342 struct ahash_request
*req
;
343 struct scatterlist
*sg
;
344 struct ipv6hdr
*top_iph
;
345 struct ip_auth_hdr
*ah
;
347 struct tmp_ext
*iph_ext
;
351 struct scatterlist
*seqhisg
;
356 err
= skb_cow_data(skb
, 0, &trailer
);
361 skb_push(skb
, -skb_network_offset(skb
));
362 extlen
= skb_network_header_len(skb
) - sizeof(struct ipv6hdr
);
364 extlen
+= sizeof(*iph_ext
);
366 if (x
->props
.flags
& XFRM_STATE_ESN
) {
368 seqhi_len
= sizeof(*seqhi
);
371 iph_base
= ah_alloc_tmp(ahash
, nfrags
+ sglists
, IPV6HDR_BASELEN
+
376 iph_ext
= ah_tmp_ext(iph_base
);
377 seqhi
= (__be32
*)((char *)iph_ext
+ extlen
);
378 icv
= ah_tmp_icv(ahash
, seqhi
, seqhi_len
);
379 req
= ah_tmp_req(ahash
, icv
);
380 sg
= ah_req_sg(ahash
, req
);
381 seqhisg
= sg
+ nfrags
;
383 ah
= ip_auth_hdr(skb
);
384 memset(ah
->auth_data
, 0, ahp
->icv_trunc_len
);
386 top_iph
= ipv6_hdr(skb
);
387 top_iph
->payload_len
= htons(skb
->len
- sizeof(*top_iph
));
389 nexthdr
= *skb_mac_header(skb
);
390 *skb_mac_header(skb
) = IPPROTO_AH
;
392 /* When there are no extension headers, we only need to save the first
393 * 8 bytes of the base IP header.
395 memcpy(iph_base
, top_iph
, IPV6HDR_BASELEN
);
398 #if IS_ENABLED(CONFIG_IPV6_MIP6)
399 memcpy(iph_ext
, &top_iph
->saddr
, extlen
);
401 memcpy(iph_ext
, &top_iph
->daddr
, extlen
);
403 err
= ipv6_clear_mutable_options(top_iph
,
404 extlen
- sizeof(*iph_ext
) +
411 ah
->nexthdr
= nexthdr
;
413 top_iph
->priority
= 0;
414 top_iph
->flow_lbl
[0] = 0;
415 top_iph
->flow_lbl
[1] = 0;
416 top_iph
->flow_lbl
[2] = 0;
417 top_iph
->hop_limit
= 0;
419 ah
->hdrlen
= (XFRM_ALIGN8(sizeof(*ah
) + ahp
->icv_trunc_len
) >> 2) - 2;
423 ah
->seq_no
= htonl(XFRM_SKB_CB(skb
)->seq
.output
.low
);
425 sg_init_table(sg
, nfrags
+ sglists
);
426 skb_to_sgvec_nomark(skb
, sg
, 0, skb
->len
);
428 if (x
->props
.flags
& XFRM_STATE_ESN
) {
429 /* Attach seqhi sg right after packet payload */
430 *seqhi
= htonl(XFRM_SKB_CB(skb
)->seq
.output
.hi
);
431 sg_set_buf(seqhisg
, seqhi
, seqhi_len
);
433 ahash_request_set_crypt(req
, sg
, icv
, skb
->len
+ seqhi_len
);
434 ahash_request_set_callback(req
, 0, ah6_output_done
, skb
);
436 AH_SKB_CB(skb
)->tmp
= iph_base
;
438 err
= crypto_ahash_digest(req
);
440 if (err
== -EINPROGRESS
)
448 memcpy(ah
->auth_data
, icv
, ahp
->icv_trunc_len
);
449 memcpy(top_iph
, iph_base
, IPV6HDR_BASELEN
);
452 #if IS_ENABLED(CONFIG_IPV6_MIP6)
453 memcpy(&top_iph
->saddr
, iph_ext
, extlen
);
455 memcpy(&top_iph
->daddr
, iph_ext
, extlen
);
465 static void ah6_input_done(struct crypto_async_request
*base
, int err
)
470 struct sk_buff
*skb
= base
->data
;
471 struct xfrm_state
*x
= xfrm_input_state(skb
);
472 struct ah_data
*ahp
= x
->data
;
473 struct ip_auth_hdr
*ah
= ip_auth_hdr(skb
);
474 int hdr_len
= skb_network_header_len(skb
);
475 int ah_hlen
= (ah
->hdrlen
+ 2) << 2;
477 work_iph
= AH_SKB_CB(skb
)->tmp
;
478 auth_data
= ah_tmp_auth(work_iph
, hdr_len
);
479 icv
= ah_tmp_icv(ahp
->ahash
, auth_data
, ahp
->icv_trunc_len
);
481 err
= memcmp(icv
, auth_data
, ahp
->icv_trunc_len
) ? -EBADMSG
: 0;
487 skb
->network_header
+= ah_hlen
;
488 memcpy(skb_network_header(skb
), work_iph
, hdr_len
);
489 __skb_pull(skb
, ah_hlen
+ hdr_len
);
490 if (x
->props
.mode
== XFRM_MODE_TUNNEL
)
491 skb_reset_transport_header(skb
);
493 skb_set_transport_header(skb
, -hdr_len
);
495 kfree(AH_SKB_CB(skb
)->tmp
);
496 xfrm_input_resume(skb
, err
);
501 static int ah6_input(struct xfrm_state
*x
, struct sk_buff
*skb
)
505 * [IPv6][Ext1][Ext2][AH][Dest][Payload]
506 * |<-------------->| hdr_len
509 * Keeping copy of cleared headers. After AH processing,
510 * Moving the pointer of skb->network_header by using skb_pull as long
511 * as AH header length. Then copy back the copy as long as hdr_len
512 * If destination header following AH exists, copy it into after [Ext2].
514 * |<>|[IPv6][Ext1][Ext2][Dest][Payload]
515 * There is offset of AH before IPv6 header after the process.
521 struct sk_buff
*trailer
;
522 struct crypto_ahash
*ahash
;
523 struct ahash_request
*req
;
524 struct scatterlist
*sg
;
525 struct ip_auth_hdr
*ah
;
526 struct ipv6hdr
*ip6h
;
536 struct scatterlist
*seqhisg
;
538 if (!pskb_may_pull(skb
, sizeof(struct ip_auth_hdr
)))
541 /* We are going to _remove_ AH header to keep sockets happy,
542 * so... Later this can change. */
543 if (skb_unclone(skb
, GFP_ATOMIC
))
546 skb
->ip_summed
= CHECKSUM_NONE
;
548 hdr_len
= skb_network_header_len(skb
);
549 ah
= (struct ip_auth_hdr
*)skb
->data
;
553 nexthdr
= ah
->nexthdr
;
554 ah_hlen
= (ah
->hdrlen
+ 2) << 2;
556 if (ah_hlen
!= XFRM_ALIGN8(sizeof(*ah
) + ahp
->icv_full_len
) &&
557 ah_hlen
!= XFRM_ALIGN8(sizeof(*ah
) + ahp
->icv_trunc_len
))
560 if (!pskb_may_pull(skb
, ah_hlen
))
563 err
= skb_cow_data(skb
, 0, &trailer
);
568 ah
= (struct ip_auth_hdr
*)skb
->data
;
569 ip6h
= ipv6_hdr(skb
);
571 skb_push(skb
, hdr_len
);
573 if (x
->props
.flags
& XFRM_STATE_ESN
) {
575 seqhi_len
= sizeof(*seqhi
);
578 work_iph
= ah_alloc_tmp(ahash
, nfrags
+ sglists
, hdr_len
+
579 ahp
->icv_trunc_len
+ seqhi_len
);
583 auth_data
= ah_tmp_auth((u8
*)work_iph
, hdr_len
);
584 seqhi
= (__be32
*)(auth_data
+ ahp
->icv_trunc_len
);
585 icv
= ah_tmp_icv(ahash
, seqhi
, seqhi_len
);
586 req
= ah_tmp_req(ahash
, icv
);
587 sg
= ah_req_sg(ahash
, req
);
588 seqhisg
= sg
+ nfrags
;
590 memcpy(work_iph
, ip6h
, hdr_len
);
591 memcpy(auth_data
, ah
->auth_data
, ahp
->icv_trunc_len
);
592 memset(ah
->auth_data
, 0, ahp
->icv_trunc_len
);
594 if (ipv6_clear_mutable_options(ip6h
, hdr_len
, XFRM_POLICY_IN
))
598 ip6h
->flow_lbl
[0] = 0;
599 ip6h
->flow_lbl
[1] = 0;
600 ip6h
->flow_lbl
[2] = 0;
603 sg_init_table(sg
, nfrags
+ sglists
);
604 skb_to_sgvec_nomark(skb
, sg
, 0, skb
->len
);
606 if (x
->props
.flags
& XFRM_STATE_ESN
) {
607 /* Attach seqhi sg right after packet payload */
608 *seqhi
= XFRM_SKB_CB(skb
)->seq
.input
.hi
;
609 sg_set_buf(seqhisg
, seqhi
, seqhi_len
);
612 ahash_request_set_crypt(req
, sg
, icv
, skb
->len
+ seqhi_len
);
613 ahash_request_set_callback(req
, 0, ah6_input_done
, skb
);
615 AH_SKB_CB(skb
)->tmp
= work_iph
;
617 err
= crypto_ahash_digest(req
);
619 if (err
== -EINPROGRESS
)
625 err
= memcmp(icv
, auth_data
, ahp
->icv_trunc_len
) ? -EBADMSG
: 0;
629 skb
->network_header
+= ah_hlen
;
630 memcpy(skb_network_header(skb
), work_iph
, hdr_len
);
631 __skb_pull(skb
, ah_hlen
+ hdr_len
);
633 if (x
->props
.mode
== XFRM_MODE_TUNNEL
)
634 skb_reset_transport_header(skb
);
636 skb_set_transport_header(skb
, -hdr_len
);
646 static int ah6_err(struct sk_buff
*skb
, struct inet6_skb_parm
*opt
,
647 u8 type
, u8 code
, int offset
, __be32 info
)
649 struct net
*net
= dev_net(skb
->dev
);
650 struct ipv6hdr
*iph
= (struct ipv6hdr
*)skb
->data
;
651 struct ip_auth_hdr
*ah
= (struct ip_auth_hdr
*)(skb
->data
+offset
);
652 struct xfrm_state
*x
;
654 if (type
!= ICMPV6_PKT_TOOBIG
&&
655 type
!= NDISC_REDIRECT
)
658 x
= xfrm_state_lookup(net
, skb
->mark
, (xfrm_address_t
*)&iph
->daddr
, ah
->spi
, IPPROTO_AH
, AF_INET6
);
662 if (type
== NDISC_REDIRECT
)
663 ip6_redirect(skb
, net
, skb
->dev
->ifindex
, 0);
665 ip6_update_pmtu(skb
, net
, info
, 0, 0);
671 static int ah6_init_state(struct xfrm_state
*x
)
673 struct ah_data
*ahp
= NULL
;
674 struct xfrm_algo_desc
*aalg_desc
;
675 struct crypto_ahash
*ahash
;
683 ahp
= kzalloc(sizeof(*ahp
), GFP_KERNEL
);
687 ahash
= crypto_alloc_ahash(x
->aalg
->alg_name
, 0, 0);
692 if (crypto_ahash_setkey(ahash
, x
->aalg
->alg_key
,
693 (x
->aalg
->alg_key_len
+ 7) / 8))
697 * Lookup the algorithm description maintained by xfrm_algo,
698 * verify crypto transform properties, and store information
699 * we need for AH processing. This lookup cannot fail here
700 * after a successful crypto_alloc_hash().
702 aalg_desc
= xfrm_aalg_get_byname(x
->aalg
->alg_name
, 0);
705 if (aalg_desc
->uinfo
.auth
.icv_fullbits
/8 !=
706 crypto_ahash_digestsize(ahash
)) {
707 pr_info("AH: %s digestsize %u != %hu\n",
708 x
->aalg
->alg_name
, crypto_ahash_digestsize(ahash
),
709 aalg_desc
->uinfo
.auth
.icv_fullbits
/8);
713 ahp
->icv_full_len
= aalg_desc
->uinfo
.auth
.icv_fullbits
/8;
714 ahp
->icv_trunc_len
= x
->aalg
->alg_trunc_len
/8;
716 x
->props
.header_len
= XFRM_ALIGN8(sizeof(struct ip_auth_hdr
) +
718 switch (x
->props
.mode
) {
720 case XFRM_MODE_TRANSPORT
:
722 case XFRM_MODE_TUNNEL
:
723 x
->props
.header_len
+= sizeof(struct ipv6hdr
);
734 crypto_free_ahash(ahp
->ahash
);
740 static void ah6_destroy(struct xfrm_state
*x
)
742 struct ah_data
*ahp
= x
->data
;
747 crypto_free_ahash(ahp
->ahash
);
751 static int ah6_rcv_cb(struct sk_buff
*skb
, int err
)
756 static const struct xfrm_type ah6_type
= {
757 .description
= "AH6",
758 .owner
= THIS_MODULE
,
760 .flags
= XFRM_TYPE_REPLAY_PROT
,
761 .init_state
= ah6_init_state
,
762 .destructor
= ah6_destroy
,
764 .output
= ah6_output
,
765 .hdr_offset
= xfrm6_find_1stfragopt
,
768 static struct xfrm6_protocol ah6_protocol
= {
769 .handler
= xfrm6_rcv
,
770 .cb_handler
= ah6_rcv_cb
,
771 .err_handler
= ah6_err
,
775 static int __init
ah6_init(void)
777 if (xfrm_register_type(&ah6_type
, AF_INET6
) < 0) {
778 pr_info("%s: can't add xfrm type\n", __func__
);
782 if (xfrm6_protocol_register(&ah6_protocol
, IPPROTO_AH
) < 0) {
783 pr_info("%s: can't add protocol\n", __func__
);
784 xfrm_unregister_type(&ah6_type
, AF_INET6
);
791 static void __exit
ah6_fini(void)
793 if (xfrm6_protocol_deregister(&ah6_protocol
, IPPROTO_AH
) < 0)
794 pr_info("%s: can't remove protocol\n", __func__
);
796 if (xfrm_unregister_type(&ah6_type
, AF_INET6
) < 0)
797 pr_info("%s: can't remove xfrm type\n", __func__
);
801 module_init(ah6_init
);
802 module_exit(ah6_fini
);
804 MODULE_LICENSE("GPL");
805 MODULE_ALIAS_XFRM_TYPE(AF_INET6
, XFRM_PROTO_AH
);