1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/skbuff.h>
3 #include <linux/netdevice.h>
4 #include <linux/if_vlan.h>
5 #include <linux/netpoll.h>
6 #include <linux/export.h>
9 bool vlan_do_receive(struct sk_buff
**skbp
)
11 struct sk_buff
*skb
= *skbp
;
12 __be16 vlan_proto
= skb
->vlan_proto
;
13 u16 vlan_id
= skb_vlan_tag_get_id(skb
);
14 struct net_device
*vlan_dev
;
15 struct vlan_pcpu_stats
*rx_stats
;
17 vlan_dev
= vlan_find_dev(skb
->dev
, vlan_proto
, vlan_id
);
21 skb
= *skbp
= skb_share_check(skb
, GFP_ATOMIC
);
25 if (unlikely(!(vlan_dev
->flags
& IFF_UP
))) {
32 if (unlikely(skb
->pkt_type
== PACKET_OTHERHOST
)) {
33 /* Our lower layer thinks this is not local, let's make sure.
34 * This allows the VLAN to have a different MAC than the
35 * underlying device, and still route correctly. */
36 if (ether_addr_equal_64bits(eth_hdr(skb
)->h_dest
, vlan_dev
->dev_addr
))
37 skb
->pkt_type
= PACKET_HOST
;
40 if (!(vlan_dev_priv(vlan_dev
)->flags
& VLAN_FLAG_REORDER_HDR
) &&
41 !netif_is_macvlan_port(vlan_dev
) &&
42 !netif_is_bridge_port(vlan_dev
)) {
43 unsigned int offset
= skb
->data
- skb_mac_header(skb
);
46 * vlan_insert_tag expect skb->data pointing to mac header.
47 * So change skb->data before calling it and change back to
48 * original position later
50 skb_push(skb
, offset
);
51 skb
= *skbp
= vlan_insert_inner_tag(skb
, skb
->vlan_proto
,
52 skb
->vlan_tci
, skb
->mac_len
);
55 skb_pull(skb
, offset
+ VLAN_HLEN
);
56 skb_reset_mac_len(skb
);
59 skb
->priority
= vlan_get_ingress_priority(vlan_dev
, skb
->vlan_tci
);
62 rx_stats
= this_cpu_ptr(vlan_dev_priv(vlan_dev
)->vlan_pcpu_stats
);
64 u64_stats_update_begin(&rx_stats
->syncp
);
65 rx_stats
->rx_packets
++;
66 rx_stats
->rx_bytes
+= skb
->len
;
67 if (skb
->pkt_type
== PACKET_MULTICAST
)
68 rx_stats
->rx_multicast
++;
69 u64_stats_update_end(&rx_stats
->syncp
);
74 /* Must be invoked with rcu_read_lock. */
75 struct net_device
*__vlan_find_dev_deep_rcu(struct net_device
*dev
,
76 __be16 vlan_proto
, u16 vlan_id
)
78 struct vlan_info
*vlan_info
= rcu_dereference(dev
->vlan_info
);
81 return vlan_group_get_device(&vlan_info
->grp
,
85 * Lower devices of master uppers (bonding, team) do not have
86 * grp assigned to themselves. Grp is assigned to upper device
89 struct net_device
*upper_dev
;
91 upper_dev
= netdev_master_upper_dev_get_rcu(dev
);
93 return __vlan_find_dev_deep_rcu(upper_dev
,
99 EXPORT_SYMBOL(__vlan_find_dev_deep_rcu
);
101 struct net_device
*vlan_dev_real_dev(const struct net_device
*dev
)
103 struct net_device
*ret
= vlan_dev_priv(dev
)->real_dev
;
105 while (is_vlan_dev(ret
))
106 ret
= vlan_dev_priv(ret
)->real_dev
;
110 EXPORT_SYMBOL(vlan_dev_real_dev
);
112 u16
vlan_dev_vlan_id(const struct net_device
*dev
)
114 return vlan_dev_priv(dev
)->vlan_id
;
116 EXPORT_SYMBOL(vlan_dev_vlan_id
);
118 __be16
vlan_dev_vlan_proto(const struct net_device
*dev
)
120 return vlan_dev_priv(dev
)->vlan_proto
;
122 EXPORT_SYMBOL(vlan_dev_vlan_proto
);
125 * vlan info and vid list
128 static void vlan_group_free(struct vlan_group
*grp
)
132 for (i
= 0; i
< VLAN_PROTO_NUM
; i
++)
133 for (j
= 0; j
< VLAN_GROUP_ARRAY_SPLIT_PARTS
; j
++)
134 kfree(grp
->vlan_devices_arrays
[i
][j
]);
137 static void vlan_info_free(struct vlan_info
*vlan_info
)
139 vlan_group_free(&vlan_info
->grp
);
143 static void vlan_info_rcu_free(struct rcu_head
*rcu
)
145 vlan_info_free(container_of(rcu
, struct vlan_info
, rcu
));
148 static struct vlan_info
*vlan_info_alloc(struct net_device
*dev
)
150 struct vlan_info
*vlan_info
;
152 vlan_info
= kzalloc(sizeof(struct vlan_info
), GFP_KERNEL
);
156 vlan_info
->real_dev
= dev
;
157 INIT_LIST_HEAD(&vlan_info
->vid_list
);
161 struct vlan_vid_info
{
162 struct list_head list
;
168 static bool vlan_hw_filter_capable(const struct net_device
*dev
, __be16 proto
)
170 if (proto
== htons(ETH_P_8021Q
) &&
171 dev
->features
& NETIF_F_HW_VLAN_CTAG_FILTER
)
173 if (proto
== htons(ETH_P_8021AD
) &&
174 dev
->features
& NETIF_F_HW_VLAN_STAG_FILTER
)
179 static struct vlan_vid_info
*vlan_vid_info_get(struct vlan_info
*vlan_info
,
180 __be16 proto
, u16 vid
)
182 struct vlan_vid_info
*vid_info
;
184 list_for_each_entry(vid_info
, &vlan_info
->vid_list
, list
) {
185 if (vid_info
->proto
== proto
&& vid_info
->vid
== vid
)
191 static struct vlan_vid_info
*vlan_vid_info_alloc(__be16 proto
, u16 vid
)
193 struct vlan_vid_info
*vid_info
;
195 vid_info
= kzalloc(sizeof(struct vlan_vid_info
), GFP_KERNEL
);
198 vid_info
->proto
= proto
;
204 static int vlan_add_rx_filter_info(struct net_device
*dev
, __be16 proto
, u16 vid
)
206 if (!vlan_hw_filter_capable(dev
, proto
))
209 if (netif_device_present(dev
))
210 return dev
->netdev_ops
->ndo_vlan_rx_add_vid(dev
, proto
, vid
);
215 static int vlan_kill_rx_filter_info(struct net_device
*dev
, __be16 proto
, u16 vid
)
217 if (!vlan_hw_filter_capable(dev
, proto
))
220 if (netif_device_present(dev
))
221 return dev
->netdev_ops
->ndo_vlan_rx_kill_vid(dev
, proto
, vid
);
226 int vlan_filter_push_vids(struct vlan_info
*vlan_info
, __be16 proto
)
228 struct net_device
*real_dev
= vlan_info
->real_dev
;
229 struct vlan_vid_info
*vlan_vid_info
;
232 list_for_each_entry(vlan_vid_info
, &vlan_info
->vid_list
, list
) {
233 if (vlan_vid_info
->proto
== proto
) {
234 err
= vlan_add_rx_filter_info(real_dev
, proto
,
244 list_for_each_entry_continue_reverse(vlan_vid_info
,
245 &vlan_info
->vid_list
, list
) {
246 if (vlan_vid_info
->proto
== proto
)
247 vlan_kill_rx_filter_info(real_dev
, proto
,
253 EXPORT_SYMBOL(vlan_filter_push_vids
);
255 void vlan_filter_drop_vids(struct vlan_info
*vlan_info
, __be16 proto
)
257 struct vlan_vid_info
*vlan_vid_info
;
259 list_for_each_entry(vlan_vid_info
, &vlan_info
->vid_list
, list
)
260 if (vlan_vid_info
->proto
== proto
)
261 vlan_kill_rx_filter_info(vlan_info
->real_dev
,
262 vlan_vid_info
->proto
,
265 EXPORT_SYMBOL(vlan_filter_drop_vids
);
267 static int __vlan_vid_add(struct vlan_info
*vlan_info
, __be16 proto
, u16 vid
,
268 struct vlan_vid_info
**pvid_info
)
270 struct net_device
*dev
= vlan_info
->real_dev
;
271 struct vlan_vid_info
*vid_info
;
274 vid_info
= vlan_vid_info_alloc(proto
, vid
);
278 err
= vlan_add_rx_filter_info(dev
, proto
, vid
);
284 list_add(&vid_info
->list
, &vlan_info
->vid_list
);
285 vlan_info
->nr_vids
++;
286 *pvid_info
= vid_info
;
290 int vlan_vid_add(struct net_device
*dev
, __be16 proto
, u16 vid
)
292 struct vlan_info
*vlan_info
;
293 struct vlan_vid_info
*vid_info
;
294 bool vlan_info_created
= false;
299 vlan_info
= rtnl_dereference(dev
->vlan_info
);
301 vlan_info
= vlan_info_alloc(dev
);
304 vlan_info_created
= true;
306 vid_info
= vlan_vid_info_get(vlan_info
, proto
, vid
);
308 err
= __vlan_vid_add(vlan_info
, proto
, vid
, &vid_info
);
310 goto out_free_vlan_info
;
312 vid_info
->refcount
++;
314 if (vlan_info_created
)
315 rcu_assign_pointer(dev
->vlan_info
, vlan_info
);
320 if (vlan_info_created
)
324 EXPORT_SYMBOL(vlan_vid_add
);
326 static void __vlan_vid_del(struct vlan_info
*vlan_info
,
327 struct vlan_vid_info
*vid_info
)
329 struct net_device
*dev
= vlan_info
->real_dev
;
330 __be16 proto
= vid_info
->proto
;
331 u16 vid
= vid_info
->vid
;
334 err
= vlan_kill_rx_filter_info(dev
, proto
, vid
);
336 pr_warn("failed to kill vid %04x/%d for device %s\n",
337 proto
, vid
, dev
->name
);
339 list_del(&vid_info
->list
);
341 vlan_info
->nr_vids
--;
344 void vlan_vid_del(struct net_device
*dev
, __be16 proto
, u16 vid
)
346 struct vlan_info
*vlan_info
;
347 struct vlan_vid_info
*vid_info
;
351 vlan_info
= rtnl_dereference(dev
->vlan_info
);
355 vid_info
= vlan_vid_info_get(vlan_info
, proto
, vid
);
358 vid_info
->refcount
--;
359 if (vid_info
->refcount
== 0) {
360 __vlan_vid_del(vlan_info
, vid_info
);
361 if (vlan_info
->nr_vids
== 0) {
362 RCU_INIT_POINTER(dev
->vlan_info
, NULL
);
363 call_rcu(&vlan_info
->rcu
, vlan_info_rcu_free
);
367 EXPORT_SYMBOL(vlan_vid_del
);
369 int vlan_vids_add_by_dev(struct net_device
*dev
,
370 const struct net_device
*by_dev
)
372 struct vlan_vid_info
*vid_info
;
373 struct vlan_info
*vlan_info
;
378 vlan_info
= rtnl_dereference(by_dev
->vlan_info
);
382 list_for_each_entry(vid_info
, &vlan_info
->vid_list
, list
) {
383 err
= vlan_vid_add(dev
, vid_info
->proto
, vid_info
->vid
);
390 list_for_each_entry_continue_reverse(vid_info
,
391 &vlan_info
->vid_list
,
393 vlan_vid_del(dev
, vid_info
->proto
, vid_info
->vid
);
398 EXPORT_SYMBOL(vlan_vids_add_by_dev
);
400 void vlan_vids_del_by_dev(struct net_device
*dev
,
401 const struct net_device
*by_dev
)
403 struct vlan_vid_info
*vid_info
;
404 struct vlan_info
*vlan_info
;
408 vlan_info
= rtnl_dereference(by_dev
->vlan_info
);
412 list_for_each_entry(vid_info
, &vlan_info
->vid_list
, list
)
413 vlan_vid_del(dev
, vid_info
->proto
, vid_info
->vid
);
415 EXPORT_SYMBOL(vlan_vids_del_by_dev
);
417 bool vlan_uses_dev(const struct net_device
*dev
)
419 struct vlan_info
*vlan_info
;
423 vlan_info
= rtnl_dereference(dev
->vlan_info
);
426 return vlan_info
->grp
.nr_vlan_devs
? true : false;
428 EXPORT_SYMBOL(vlan_uses_dev
);