3 * Ethernet-type device handling.
5 * Authors: Ben Greear <greearb@candelatech.com>
6 * Please send support related email to: netdev@vger.kernel.org
7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12 * Correct all the locking - David S. Miller <davem@redhat.com>;
13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * as published by the Free Software Foundation; either version
18 * 2 of the License, or (at your option) any later version.
21 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
23 #include <linux/capability.h>
24 #include <linux/module.h>
25 #include <linux/netdevice.h>
26 #include <linux/skbuff.h>
27 #include <linux/slab.h>
28 #include <linux/init.h>
29 #include <linux/rculist.h>
30 #include <net/p8022.h>
32 #include <linux/rtnetlink.h>
33 #include <linux/notifier.h>
34 #include <net/rtnetlink.h>
35 #include <net/net_namespace.h>
36 #include <net/netns/generic.h>
37 #include <linux/uaccess.h>
39 #include <linux/if_vlan.h>
43 #define DRV_VERSION "1.8"
45 /* Global VLAN variables */
47 unsigned int vlan_net_id __read_mostly
;
49 const char vlan_fullname
[] = "802.1Q VLAN Support";
50 const char vlan_version
[] = DRV_VERSION
;
52 /* End of global variables definitions. */
54 static int vlan_group_prealloc_vid(struct vlan_group
*vg
,
55 __be16 vlan_proto
, u16 vlan_id
)
57 struct net_device
**array
;
58 unsigned int pidx
, vidx
;
63 pidx
= vlan_proto_idx(vlan_proto
);
64 vidx
= vlan_id
/ VLAN_GROUP_ARRAY_PART_LEN
;
65 array
= vg
->vlan_devices_arrays
[pidx
][vidx
];
69 size
= sizeof(struct net_device
*) * VLAN_GROUP_ARRAY_PART_LEN
;
70 array
= kzalloc(size
, GFP_KERNEL
);
74 vg
->vlan_devices_arrays
[pidx
][vidx
] = array
;
78 void unregister_vlan_dev(struct net_device
*dev
, struct list_head
*head
)
80 struct vlan_dev_priv
*vlan
= vlan_dev_priv(dev
);
81 struct net_device
*real_dev
= vlan
->real_dev
;
82 struct vlan_info
*vlan_info
;
83 struct vlan_group
*grp
;
84 u16 vlan_id
= vlan
->vlan_id
;
88 vlan_info
= rtnl_dereference(real_dev
->vlan_info
);
91 grp
= &vlan_info
->grp
;
95 if (vlan
->flags
& VLAN_FLAG_MVRP
)
96 vlan_mvrp_request_leave(dev
);
97 if (vlan
->flags
& VLAN_FLAG_GVRP
)
98 vlan_gvrp_request_leave(dev
);
100 vlan_group_set_device(grp
, vlan
->vlan_proto
, vlan_id
, NULL
);
102 netdev_upper_dev_unlink(real_dev
, dev
);
103 /* Because unregister_netdevice_queue() makes sure at least one rcu
104 * grace period is respected before device freeing,
105 * we dont need to call synchronize_net() here.
107 unregister_netdevice_queue(dev
, head
);
109 if (grp
->nr_vlan_devs
== 0) {
110 vlan_mvrp_uninit_applicant(real_dev
);
111 vlan_gvrp_uninit_applicant(real_dev
);
114 /* Take it out of our own structures, but be sure to interlock with
115 * HW accelerating devices or SW vlan input packet processing if
116 * VLAN is not 0 (leave it there for 802.1p).
119 vlan_vid_del(real_dev
, vlan
->vlan_proto
, vlan_id
);
121 /* Get rid of the vlan's reference to real_dev */
125 int vlan_check_real_dev(struct net_device
*real_dev
,
126 __be16 protocol
, u16 vlan_id
)
128 const char *name
= real_dev
->name
;
130 if (real_dev
->features
& NETIF_F_VLAN_CHALLENGED
) {
131 pr_info("VLANs not supported on %s\n", name
);
135 if (vlan_find_dev(real_dev
, protocol
, vlan_id
) != NULL
)
141 int register_vlan_dev(struct net_device
*dev
)
143 struct vlan_dev_priv
*vlan
= vlan_dev_priv(dev
);
144 struct net_device
*real_dev
= vlan
->real_dev
;
145 u16 vlan_id
= vlan
->vlan_id
;
146 struct vlan_info
*vlan_info
;
147 struct vlan_group
*grp
;
150 err
= vlan_vid_add(real_dev
, vlan
->vlan_proto
, vlan_id
);
154 vlan_info
= rtnl_dereference(real_dev
->vlan_info
);
155 /* vlan_info should be there now. vlan_vid_add took care of it */
158 grp
= &vlan_info
->grp
;
159 if (grp
->nr_vlan_devs
== 0) {
160 err
= vlan_gvrp_init_applicant(real_dev
);
163 err
= vlan_mvrp_init_applicant(real_dev
);
165 goto out_uninit_gvrp
;
168 err
= vlan_group_prealloc_vid(grp
, vlan
->vlan_proto
, vlan_id
);
170 goto out_uninit_mvrp
;
172 vlan
->nest_level
= dev_get_nest_level(real_dev
) + 1;
173 err
= register_netdevice(dev
);
175 goto out_uninit_mvrp
;
177 err
= netdev_upper_dev_link(real_dev
, dev
);
179 goto out_unregister_netdev
;
181 /* Account for reference in struct vlan_dev_priv */
184 netif_stacked_transfer_operstate(real_dev
, dev
);
185 linkwatch_fire_event(dev
); /* _MUST_ call rfc2863_policy() */
187 /* So, got the sucker initialized, now lets place
188 * it into our local structure.
190 vlan_group_set_device(grp
, vlan
->vlan_proto
, vlan_id
, dev
);
195 out_unregister_netdev
:
196 unregister_netdevice(dev
);
198 if (grp
->nr_vlan_devs
== 0)
199 vlan_mvrp_uninit_applicant(real_dev
);
201 if (grp
->nr_vlan_devs
== 0)
202 vlan_gvrp_uninit_applicant(real_dev
);
204 vlan_vid_del(real_dev
, vlan
->vlan_proto
, vlan_id
);
208 /* Attach a VLAN device to a mac address (ie Ethernet Card).
209 * Returns 0 if the device was created or a negative error code otherwise.
211 static int register_vlan_device(struct net_device
*real_dev
, u16 vlan_id
)
213 struct net_device
*new_dev
;
214 struct vlan_dev_priv
*vlan
;
215 struct net
*net
= dev_net(real_dev
);
216 struct vlan_net
*vn
= net_generic(net
, vlan_net_id
);
220 if (vlan_id
>= VLAN_VID_MASK
)
223 err
= vlan_check_real_dev(real_dev
, htons(ETH_P_8021Q
), vlan_id
);
227 /* Gotta set up the fields for the device. */
228 switch (vn
->name_type
) {
229 case VLAN_NAME_TYPE_RAW_PLUS_VID
:
230 /* name will look like: eth1.0005 */
231 snprintf(name
, IFNAMSIZ
, "%s.%.4i", real_dev
->name
, vlan_id
);
233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD
:
234 /* Put our vlan.VID in the name.
235 * Name will look like: vlan5
237 snprintf(name
, IFNAMSIZ
, "vlan%i", vlan_id
);
239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD
:
240 /* Put our vlan.VID in the name.
241 * Name will look like: eth0.5
243 snprintf(name
, IFNAMSIZ
, "%s.%i", real_dev
->name
, vlan_id
);
245 case VLAN_NAME_TYPE_PLUS_VID
:
246 /* Put our vlan.VID in the name.
247 * Name will look like: vlan0005
250 snprintf(name
, IFNAMSIZ
, "vlan%.4i", vlan_id
);
253 new_dev
= alloc_netdev(sizeof(struct vlan_dev_priv
), name
,
254 NET_NAME_UNKNOWN
, vlan_setup
);
259 dev_net_set(new_dev
, net
);
260 /* need 4 bytes for extra VLAN header info,
261 * hope the underlying device can handle it.
263 new_dev
->mtu
= real_dev
->mtu
;
265 vlan
= vlan_dev_priv(new_dev
);
266 vlan
->vlan_proto
= htons(ETH_P_8021Q
);
267 vlan
->vlan_id
= vlan_id
;
268 vlan
->real_dev
= real_dev
;
270 vlan
->flags
= VLAN_FLAG_REORDER_HDR
;
272 new_dev
->rtnl_link_ops
= &vlan_link_ops
;
273 err
= register_vlan_dev(new_dev
);
275 goto out_free_newdev
;
280 if (new_dev
->reg_state
== NETREG_UNINITIALIZED
)
281 free_netdev(new_dev
);
285 static void vlan_sync_address(struct net_device
*dev
,
286 struct net_device
*vlandev
)
288 struct vlan_dev_priv
*vlan
= vlan_dev_priv(vlandev
);
290 /* May be called without an actual change */
291 if (ether_addr_equal(vlan
->real_dev_addr
, dev
->dev_addr
))
294 /* vlan continues to inherit address of lower device */
295 if (vlan_dev_inherit_address(vlandev
, dev
))
298 /* vlan address was different from the old address and is equal to
300 if (!ether_addr_equal(vlandev
->dev_addr
, vlan
->real_dev_addr
) &&
301 ether_addr_equal(vlandev
->dev_addr
, dev
->dev_addr
))
302 dev_uc_del(dev
, vlandev
->dev_addr
);
304 /* vlan address was equal to the old address and is different from
306 if (ether_addr_equal(vlandev
->dev_addr
, vlan
->real_dev_addr
) &&
307 !ether_addr_equal(vlandev
->dev_addr
, dev
->dev_addr
))
308 dev_uc_add(dev
, vlandev
->dev_addr
);
311 ether_addr_copy(vlan
->real_dev_addr
, dev
->dev_addr
);
314 static void vlan_transfer_features(struct net_device
*dev
,
315 struct net_device
*vlandev
)
317 struct vlan_dev_priv
*vlan
= vlan_dev_priv(vlandev
);
319 vlandev
->gso_max_size
= dev
->gso_max_size
;
320 vlandev
->gso_max_segs
= dev
->gso_max_segs
;
322 if (vlan_hw_offload_capable(dev
->features
, vlan
->vlan_proto
))
323 vlandev
->hard_header_len
= dev
->hard_header_len
;
325 vlandev
->hard_header_len
= dev
->hard_header_len
+ VLAN_HLEN
;
327 #if IS_ENABLED(CONFIG_FCOE)
328 vlandev
->fcoe_ddp_xid
= dev
->fcoe_ddp_xid
;
331 netdev_update_features(vlandev
);
334 static int __vlan_device_event(struct net_device
*dev
, unsigned long event
)
339 case NETDEV_CHANGENAME
:
340 vlan_proc_rem_dev(dev
);
341 err
= vlan_proc_add_dev(dev
);
343 case NETDEV_REGISTER
:
344 err
= vlan_proc_add_dev(dev
);
346 case NETDEV_UNREGISTER
:
347 vlan_proc_rem_dev(dev
);
354 static int vlan_device_event(struct notifier_block
*unused
, unsigned long event
,
357 struct net_device
*dev
= netdev_notifier_info_to_dev(ptr
);
358 struct vlan_group
*grp
;
359 struct vlan_info
*vlan_info
;
361 struct net_device
*vlandev
;
362 struct vlan_dev_priv
*vlan
;
366 if (is_vlan_dev(dev
)) {
367 int err
= __vlan_device_event(dev
, event
);
370 return notifier_from_errno(err
);
373 if ((event
== NETDEV_UP
) &&
374 (dev
->features
& NETIF_F_HW_VLAN_CTAG_FILTER
)) {
375 pr_info("adding VLAN 0 to HW filter on device %s\n",
377 vlan_vid_add(dev
, htons(ETH_P_8021Q
), 0);
380 vlan_info
= rtnl_dereference(dev
->vlan_info
);
383 grp
= &vlan_info
->grp
;
385 /* It is OK that we do not hold the group lock right now,
386 * as we run under the RTNL lock.
391 /* Propagate real device state to vlan devices */
392 vlan_group_for_each_dev(grp
, i
, vlandev
)
393 netif_stacked_transfer_operstate(dev
, vlandev
);
396 case NETDEV_CHANGEADDR
:
397 /* Adjust unicast filters on underlying device */
398 vlan_group_for_each_dev(grp
, i
, vlandev
) {
399 flgs
= vlandev
->flags
;
400 if (!(flgs
& IFF_UP
))
403 vlan_sync_address(dev
, vlandev
);
407 case NETDEV_CHANGEMTU
:
408 vlan_group_for_each_dev(grp
, i
, vlandev
) {
409 if (vlandev
->mtu
<= dev
->mtu
)
412 dev_set_mtu(vlandev
, dev
->mtu
);
416 case NETDEV_FEAT_CHANGE
:
417 /* Propagate device features to underlying device */
418 vlan_group_for_each_dev(grp
, i
, vlandev
)
419 vlan_transfer_features(dev
, vlandev
);
423 struct net_device
*tmp
;
424 LIST_HEAD(close_list
);
426 if (dev
->features
& NETIF_F_HW_VLAN_CTAG_FILTER
)
427 vlan_vid_del(dev
, htons(ETH_P_8021Q
), 0);
429 /* Put all VLANs for this dev in the down state too. */
430 vlan_group_for_each_dev(grp
, i
, vlandev
) {
431 flgs
= vlandev
->flags
;
432 if (!(flgs
& IFF_UP
))
435 vlan
= vlan_dev_priv(vlandev
);
436 if (!(vlan
->flags
& VLAN_FLAG_LOOSE_BINDING
))
437 list_add(&vlandev
->close_list
, &close_list
);
440 dev_close_many(&close_list
, false);
442 list_for_each_entry_safe(vlandev
, tmp
, &close_list
, close_list
) {
443 netif_stacked_transfer_operstate(dev
, vlandev
);
444 list_del_init(&vlandev
->close_list
);
446 list_del(&close_list
);
450 /* Put all VLANs for this dev in the up state too. */
451 vlan_group_for_each_dev(grp
, i
, vlandev
) {
452 flgs
= dev_get_flags(vlandev
);
456 vlan
= vlan_dev_priv(vlandev
);
457 if (!(vlan
->flags
& VLAN_FLAG_LOOSE_BINDING
))
458 dev_change_flags(vlandev
, flgs
| IFF_UP
);
459 netif_stacked_transfer_operstate(dev
, vlandev
);
463 case NETDEV_UNREGISTER
:
464 /* twiddle thumbs on netns device moves */
465 if (dev
->reg_state
!= NETREG_UNREGISTERING
)
468 vlan_group_for_each_dev(grp
, i
, vlandev
) {
469 /* removal of last vid destroys vlan_info, abort
471 if (vlan_info
->nr_vids
== 1)
474 unregister_vlan_dev(vlandev
, &list
);
478 unregister_netdevice_many(&list
);
481 case NETDEV_PRE_TYPE_CHANGE
:
482 /* Forbid underlaying device to change its type. */
483 if (vlan_uses_dev(dev
))
487 case NETDEV_NOTIFY_PEERS
:
488 case NETDEV_BONDING_FAILOVER
:
489 case NETDEV_RESEND_IGMP
:
490 /* Propagate to vlan devices */
491 vlan_group_for_each_dev(grp
, i
, vlandev
)
492 call_netdevice_notifiers(event
, vlandev
);
500 static struct notifier_block vlan_notifier_block __read_mostly
= {
501 .notifier_call
= vlan_device_event
,
505 * VLAN IOCTL handler.
506 * o execute requested action or pass command to the device driver
507 * arg is really a struct vlan_ioctl_args __user *.
509 static int vlan_ioctl_handler(struct net
*net
, void __user
*arg
)
512 struct vlan_ioctl_args args
;
513 struct net_device
*dev
= NULL
;
515 if (copy_from_user(&args
, arg
, sizeof(struct vlan_ioctl_args
)))
518 /* Null terminate this sucker, just in case. */
519 args
.device1
[sizeof(args
.device1
) - 1] = 0;
520 args
.u
.device2
[sizeof(args
.u
.device2
) - 1] = 0;
525 case SET_VLAN_INGRESS_PRIORITY_CMD
:
526 case SET_VLAN_EGRESS_PRIORITY_CMD
:
527 case SET_VLAN_FLAG_CMD
:
530 case GET_VLAN_REALDEV_NAME_CMD
:
531 case GET_VLAN_VID_CMD
:
533 dev
= __dev_get_by_name(net
, args
.device1
);
538 if (args
.cmd
!= ADD_VLAN_CMD
&& !is_vlan_dev(dev
))
543 case SET_VLAN_INGRESS_PRIORITY_CMD
:
545 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
547 vlan_dev_set_ingress_priority(dev
,
553 case SET_VLAN_EGRESS_PRIORITY_CMD
:
555 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
557 err
= vlan_dev_set_egress_priority(dev
,
562 case SET_VLAN_FLAG_CMD
:
564 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
566 err
= vlan_dev_change_flags(dev
,
567 args
.vlan_qos
? args
.u
.flag
: 0,
571 case SET_VLAN_NAME_TYPE_CMD
:
573 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
575 if (args
.u
.name_type
< VLAN_NAME_TYPE_HIGHEST
) {
578 vn
= net_generic(net
, vlan_net_id
);
579 vn
->name_type
= args
.u
.name_type
;
588 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
590 err
= register_vlan_device(dev
, args
.u
.VID
);
595 if (!ns_capable(net
->user_ns
, CAP_NET_ADMIN
))
597 unregister_vlan_dev(dev
, NULL
);
601 case GET_VLAN_REALDEV_NAME_CMD
:
603 vlan_dev_get_realdev_name(dev
, args
.u
.device2
);
604 if (copy_to_user(arg
, &args
,
605 sizeof(struct vlan_ioctl_args
)))
609 case GET_VLAN_VID_CMD
:
611 args
.u
.VID
= vlan_dev_vlan_id(dev
);
612 if (copy_to_user(arg
, &args
,
613 sizeof(struct vlan_ioctl_args
)))
626 static struct sk_buff
**vlan_gro_receive(struct sk_buff
**head
,
629 struct sk_buff
*p
, **pp
= NULL
;
630 struct vlan_hdr
*vhdr
;
631 unsigned int hlen
, off_vlan
;
632 const struct packet_offload
*ptype
;
636 off_vlan
= skb_gro_offset(skb
);
637 hlen
= off_vlan
+ sizeof(*vhdr
);
638 vhdr
= skb_gro_header_fast(skb
, off_vlan
);
639 if (skb_gro_header_hard(skb
, hlen
)) {
640 vhdr
= skb_gro_header_slow(skb
, hlen
, off_vlan
);
645 type
= vhdr
->h_vlan_encapsulated_proto
;
648 ptype
= gro_find_receive_by_type(type
);
654 for (p
= *head
; p
; p
= p
->next
) {
655 struct vlan_hdr
*vhdr2
;
657 if (!NAPI_GRO_CB(p
)->same_flow
)
660 vhdr2
= (struct vlan_hdr
*)(p
->data
+ off_vlan
);
661 if (compare_vlan_header(vhdr
, vhdr2
))
662 NAPI_GRO_CB(p
)->same_flow
= 0;
665 skb_gro_pull(skb
, sizeof(*vhdr
));
666 skb_gro_postpull_rcsum(skb
, vhdr
, sizeof(*vhdr
));
667 pp
= call_gro_receive(ptype
->callbacks
.gro_receive
, head
, skb
);
672 NAPI_GRO_CB(skb
)->flush
|= flush
;
677 static int vlan_gro_complete(struct sk_buff
*skb
, int nhoff
)
679 struct vlan_hdr
*vhdr
= (struct vlan_hdr
*)(skb
->data
+ nhoff
);
680 __be16 type
= vhdr
->h_vlan_encapsulated_proto
;
681 struct packet_offload
*ptype
;
685 ptype
= gro_find_complete_by_type(type
);
687 err
= ptype
->callbacks
.gro_complete(skb
, nhoff
+ sizeof(*vhdr
));
693 static struct packet_offload vlan_packet_offloads
[] __read_mostly
= {
695 .type
= cpu_to_be16(ETH_P_8021Q
),
698 .gro_receive
= vlan_gro_receive
,
699 .gro_complete
= vlan_gro_complete
,
703 .type
= cpu_to_be16(ETH_P_8021AD
),
706 .gro_receive
= vlan_gro_receive
,
707 .gro_complete
= vlan_gro_complete
,
712 static int __net_init
vlan_init_net(struct net
*net
)
714 struct vlan_net
*vn
= net_generic(net
, vlan_net_id
);
717 vn
->name_type
= VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD
;
719 err
= vlan_proc_init(net
);
724 static void __net_exit
vlan_exit_net(struct net
*net
)
726 vlan_proc_cleanup(net
);
729 static struct pernet_operations vlan_net_ops
= {
730 .init
= vlan_init_net
,
731 .exit
= vlan_exit_net
,
733 .size
= sizeof(struct vlan_net
),
736 static int __init
vlan_proto_init(void)
741 pr_info("%s v%s\n", vlan_fullname
, vlan_version
);
743 err
= register_pernet_subsys(&vlan_net_ops
);
747 err
= register_netdevice_notifier(&vlan_notifier_block
);
751 err
= vlan_gvrp_init();
755 err
= vlan_mvrp_init();
759 err
= vlan_netlink_init();
763 for (i
= 0; i
< ARRAY_SIZE(vlan_packet_offloads
); i
++)
764 dev_add_offload(&vlan_packet_offloads
[i
]);
766 vlan_ioctl_set(vlan_ioctl_handler
);
774 unregister_netdevice_notifier(&vlan_notifier_block
);
776 unregister_pernet_subsys(&vlan_net_ops
);
781 static void __exit
vlan_cleanup_module(void)
785 vlan_ioctl_set(NULL
);
787 for (i
= 0; i
< ARRAY_SIZE(vlan_packet_offloads
); i
++)
788 dev_remove_offload(&vlan_packet_offloads
[i
]);
792 unregister_netdevice_notifier(&vlan_notifier_block
);
794 unregister_pernet_subsys(&vlan_net_ops
);
795 rcu_barrier(); /* Wait for completion of call_rcu()'s */
801 module_init(vlan_proto_init
);
802 module_exit(vlan_cleanup_module
);
804 MODULE_LICENSE("GPL");
805 MODULE_VERSION(DRV_VERSION
);