]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - net/8021q/vlan.c
UBUNTU: Ubuntu-4.13.0-45.50
[mirror_ubuntu-artful-kernel.git] / net / 8021q / vlan.c
1 /*
2 * INET 802.1Q VLAN
3 * Ethernet-type device handling.
4 *
5 * Authors: Ben Greear <greearb@candelatech.com>
6 * Please send support related email to: netdev@vger.kernel.org
7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
8 *
9 * Fixes:
10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12 * Correct all the locking - David S. Miller <davem@redhat.com>;
13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
14 *
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * as published by the Free Software Foundation; either version
18 * 2 of the License, or (at your option) any later version.
19 */
20
21 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22
23 #include <linux/capability.h>
24 #include <linux/module.h>
25 #include <linux/netdevice.h>
26 #include <linux/skbuff.h>
27 #include <linux/slab.h>
28 #include <linux/init.h>
29 #include <linux/rculist.h>
30 #include <net/p8022.h>
31 #include <net/arp.h>
32 #include <linux/rtnetlink.h>
33 #include <linux/notifier.h>
34 #include <net/rtnetlink.h>
35 #include <net/net_namespace.h>
36 #include <net/netns/generic.h>
37 #include <linux/uaccess.h>
38
39 #include <linux/if_vlan.h>
40 #include "vlan.h"
41 #include "vlanproc.h"
42
43 #define DRV_VERSION "1.8"
44
45 /* Global VLAN variables */
46
47 unsigned int vlan_net_id __read_mostly;
48
49 const char vlan_fullname[] = "802.1Q VLAN Support";
50 const char vlan_version[] = DRV_VERSION;
51
52 /* End of global variables definitions. */
53
54 static int vlan_group_prealloc_vid(struct vlan_group *vg,
55 __be16 vlan_proto, u16 vlan_id)
56 {
57 struct net_device **array;
58 unsigned int pidx, vidx;
59 unsigned int size;
60
61 ASSERT_RTNL();
62
63 pidx = vlan_proto_idx(vlan_proto);
64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 array = vg->vlan_devices_arrays[pidx][vidx];
66 if (array != NULL)
67 return 0;
68
69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 array = kzalloc(size, GFP_KERNEL);
71 if (array == NULL)
72 return -ENOBUFS;
73
74 vg->vlan_devices_arrays[pidx][vidx] = array;
75 return 0;
76 }
77
78 void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
79 {
80 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
81 struct net_device *real_dev = vlan->real_dev;
82 struct vlan_info *vlan_info;
83 struct vlan_group *grp;
84 u16 vlan_id = vlan->vlan_id;
85
86 ASSERT_RTNL();
87
88 vlan_info = rtnl_dereference(real_dev->vlan_info);
89 BUG_ON(!vlan_info);
90
91 grp = &vlan_info->grp;
92
93 grp->nr_vlan_devs--;
94
95 if (vlan->flags & VLAN_FLAG_MVRP)
96 vlan_mvrp_request_leave(dev);
97 if (vlan->flags & VLAN_FLAG_GVRP)
98 vlan_gvrp_request_leave(dev);
99
100 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
101
102 netdev_upper_dev_unlink(real_dev, dev);
103 /* Because unregister_netdevice_queue() makes sure at least one rcu
104 * grace period is respected before device freeing,
105 * we dont need to call synchronize_net() here.
106 */
107 unregister_netdevice_queue(dev, head);
108
109 if (grp->nr_vlan_devs == 0) {
110 vlan_mvrp_uninit_applicant(real_dev);
111 vlan_gvrp_uninit_applicant(real_dev);
112 }
113
114 /* Take it out of our own structures, but be sure to interlock with
115 * HW accelerating devices or SW vlan input packet processing if
116 * VLAN is not 0 (leave it there for 802.1p).
117 */
118 if (vlan_id)
119 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
120
121 /* Get rid of the vlan's reference to real_dev */
122 dev_put(real_dev);
123 }
124
125 int vlan_check_real_dev(struct net_device *real_dev,
126 __be16 protocol, u16 vlan_id)
127 {
128 const char *name = real_dev->name;
129
130 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
131 pr_info("VLANs not supported on %s\n", name);
132 return -EOPNOTSUPP;
133 }
134
135 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL)
136 return -EEXIST;
137
138 return 0;
139 }
140
141 int register_vlan_dev(struct net_device *dev)
142 {
143 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
144 struct net_device *real_dev = vlan->real_dev;
145 u16 vlan_id = vlan->vlan_id;
146 struct vlan_info *vlan_info;
147 struct vlan_group *grp;
148 int err;
149
150 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
151 if (err)
152 return err;
153
154 vlan_info = rtnl_dereference(real_dev->vlan_info);
155 /* vlan_info should be there now. vlan_vid_add took care of it */
156 BUG_ON(!vlan_info);
157
158 grp = &vlan_info->grp;
159 if (grp->nr_vlan_devs == 0) {
160 err = vlan_gvrp_init_applicant(real_dev);
161 if (err < 0)
162 goto out_vid_del;
163 err = vlan_mvrp_init_applicant(real_dev);
164 if (err < 0)
165 goto out_uninit_gvrp;
166 }
167
168 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
169 if (err < 0)
170 goto out_uninit_mvrp;
171
172 vlan->nest_level = dev_get_nest_level(real_dev) + 1;
173 err = register_netdevice(dev);
174 if (err < 0)
175 goto out_uninit_mvrp;
176
177 err = netdev_upper_dev_link(real_dev, dev);
178 if (err)
179 goto out_unregister_netdev;
180
181 /* Account for reference in struct vlan_dev_priv */
182 dev_hold(real_dev);
183
184 netif_stacked_transfer_operstate(real_dev, dev);
185 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
186
187 /* So, got the sucker initialized, now lets place
188 * it into our local structure.
189 */
190 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
191 grp->nr_vlan_devs++;
192
193 return 0;
194
195 out_unregister_netdev:
196 unregister_netdevice(dev);
197 out_uninit_mvrp:
198 if (grp->nr_vlan_devs == 0)
199 vlan_mvrp_uninit_applicant(real_dev);
200 out_uninit_gvrp:
201 if (grp->nr_vlan_devs == 0)
202 vlan_gvrp_uninit_applicant(real_dev);
203 out_vid_del:
204 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
205 return err;
206 }
207
208 /* Attach a VLAN device to a mac address (ie Ethernet Card).
209 * Returns 0 if the device was created or a negative error code otherwise.
210 */
211 static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
212 {
213 struct net_device *new_dev;
214 struct vlan_dev_priv *vlan;
215 struct net *net = dev_net(real_dev);
216 struct vlan_net *vn = net_generic(net, vlan_net_id);
217 char name[IFNAMSIZ];
218 int err;
219
220 if (vlan_id >= VLAN_VID_MASK)
221 return -ERANGE;
222
223 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id);
224 if (err < 0)
225 return err;
226
227 /* Gotta set up the fields for the device. */
228 switch (vn->name_type) {
229 case VLAN_NAME_TYPE_RAW_PLUS_VID:
230 /* name will look like: eth1.0005 */
231 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
232 break;
233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
234 /* Put our vlan.VID in the name.
235 * Name will look like: vlan5
236 */
237 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
238 break;
239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
240 /* Put our vlan.VID in the name.
241 * Name will look like: eth0.5
242 */
243 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
244 break;
245 case VLAN_NAME_TYPE_PLUS_VID:
246 /* Put our vlan.VID in the name.
247 * Name will look like: vlan0005
248 */
249 default:
250 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
251 }
252
253 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
254 NET_NAME_UNKNOWN, vlan_setup);
255
256 if (new_dev == NULL)
257 return -ENOBUFS;
258
259 dev_net_set(new_dev, net);
260 /* need 4 bytes for extra VLAN header info,
261 * hope the underlying device can handle it.
262 */
263 new_dev->mtu = real_dev->mtu;
264
265 vlan = vlan_dev_priv(new_dev);
266 vlan->vlan_proto = htons(ETH_P_8021Q);
267 vlan->vlan_id = vlan_id;
268 vlan->real_dev = real_dev;
269 vlan->dent = NULL;
270 vlan->flags = VLAN_FLAG_REORDER_HDR;
271
272 new_dev->rtnl_link_ops = &vlan_link_ops;
273 err = register_vlan_dev(new_dev);
274 if (err < 0)
275 goto out_free_newdev;
276
277 return 0;
278
279 out_free_newdev:
280 if (new_dev->reg_state == NETREG_UNINITIALIZED)
281 free_netdev(new_dev);
282 return err;
283 }
284
285 static void vlan_sync_address(struct net_device *dev,
286 struct net_device *vlandev)
287 {
288 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
289
290 /* May be called without an actual change */
291 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
292 return;
293
294 /* vlan continues to inherit address of lower device */
295 if (vlan_dev_inherit_address(vlandev, dev))
296 goto out;
297
298 /* vlan address was different from the old address and is equal to
299 * the new address */
300 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
301 ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
302 dev_uc_del(dev, vlandev->dev_addr);
303
304 /* vlan address was equal to the old address and is different from
305 * the new address */
306 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
307 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
308 dev_uc_add(dev, vlandev->dev_addr);
309
310 out:
311 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
312 }
313
314 static void vlan_transfer_features(struct net_device *dev,
315 struct net_device *vlandev)
316 {
317 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
318
319 vlandev->gso_max_size = dev->gso_max_size;
320 vlandev->gso_max_segs = dev->gso_max_segs;
321
322 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
323 vlandev->hard_header_len = dev->hard_header_len;
324 else
325 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
326
327 #if IS_ENABLED(CONFIG_FCOE)
328 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
329 #endif
330
331 netdev_update_features(vlandev);
332 }
333
334 static int __vlan_device_event(struct net_device *dev, unsigned long event)
335 {
336 int err = 0;
337
338 switch (event) {
339 case NETDEV_CHANGENAME:
340 vlan_proc_rem_dev(dev);
341 err = vlan_proc_add_dev(dev);
342 break;
343 case NETDEV_REGISTER:
344 err = vlan_proc_add_dev(dev);
345 break;
346 case NETDEV_UNREGISTER:
347 vlan_proc_rem_dev(dev);
348 break;
349 }
350
351 return err;
352 }
353
354 static int vlan_device_event(struct notifier_block *unused, unsigned long event,
355 void *ptr)
356 {
357 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
358 struct vlan_group *grp;
359 struct vlan_info *vlan_info;
360 int i, flgs;
361 struct net_device *vlandev;
362 struct vlan_dev_priv *vlan;
363 bool last = false;
364 LIST_HEAD(list);
365
366 if (is_vlan_dev(dev)) {
367 int err = __vlan_device_event(dev, event);
368
369 if (err)
370 return notifier_from_errno(err);
371 }
372
373 if ((event == NETDEV_UP) &&
374 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
375 pr_info("adding VLAN 0 to HW filter on device %s\n",
376 dev->name);
377 vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
378 }
379 if (event == NETDEV_DOWN &&
380 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER))
381 vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
382
383 vlan_info = rtnl_dereference(dev->vlan_info);
384 if (!vlan_info)
385 goto out;
386 grp = &vlan_info->grp;
387
388 /* It is OK that we do not hold the group lock right now,
389 * as we run under the RTNL lock.
390 */
391
392 switch (event) {
393 case NETDEV_CHANGE:
394 /* Propagate real device state to vlan devices */
395 vlan_group_for_each_dev(grp, i, vlandev)
396 netif_stacked_transfer_operstate(dev, vlandev);
397 break;
398
399 case NETDEV_CHANGEADDR:
400 /* Adjust unicast filters on underlying device */
401 vlan_group_for_each_dev(grp, i, vlandev) {
402 flgs = vlandev->flags;
403 if (!(flgs & IFF_UP))
404 continue;
405
406 vlan_sync_address(dev, vlandev);
407 }
408 break;
409
410 case NETDEV_CHANGEMTU:
411 vlan_group_for_each_dev(grp, i, vlandev) {
412 if (vlandev->mtu <= dev->mtu)
413 continue;
414
415 dev_set_mtu(vlandev, dev->mtu);
416 }
417 break;
418
419 case NETDEV_FEAT_CHANGE:
420 /* Propagate device features to underlying device */
421 vlan_group_for_each_dev(grp, i, vlandev)
422 vlan_transfer_features(dev, vlandev);
423 break;
424
425 case NETDEV_DOWN: {
426 struct net_device *tmp;
427 LIST_HEAD(close_list);
428
429 /* Put all VLANs for this dev in the down state too. */
430 vlan_group_for_each_dev(grp, i, vlandev) {
431 flgs = vlandev->flags;
432 if (!(flgs & IFF_UP))
433 continue;
434
435 vlan = vlan_dev_priv(vlandev);
436 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
437 list_add(&vlandev->close_list, &close_list);
438 }
439
440 dev_close_many(&close_list, false);
441
442 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
443 netif_stacked_transfer_operstate(dev, vlandev);
444 list_del_init(&vlandev->close_list);
445 }
446 list_del(&close_list);
447 break;
448 }
449 case NETDEV_UP:
450 /* Put all VLANs for this dev in the up state too. */
451 vlan_group_for_each_dev(grp, i, vlandev) {
452 flgs = dev_get_flags(vlandev);
453 if (flgs & IFF_UP)
454 continue;
455
456 vlan = vlan_dev_priv(vlandev);
457 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
458 dev_change_flags(vlandev, flgs | IFF_UP);
459 netif_stacked_transfer_operstate(dev, vlandev);
460 }
461 break;
462
463 case NETDEV_UNREGISTER:
464 /* twiddle thumbs on netns device moves */
465 if (dev->reg_state != NETREG_UNREGISTERING)
466 break;
467
468 vlan_group_for_each_dev(grp, i, vlandev) {
469 /* removal of last vid destroys vlan_info, abort
470 * afterwards */
471 if (vlan_info->nr_vids == 1)
472 last = true;
473
474 unregister_vlan_dev(vlandev, &list);
475 if (last)
476 break;
477 }
478 unregister_netdevice_many(&list);
479 break;
480
481 case NETDEV_PRE_TYPE_CHANGE:
482 /* Forbid underlaying device to change its type. */
483 if (vlan_uses_dev(dev))
484 return NOTIFY_BAD;
485 break;
486
487 case NETDEV_NOTIFY_PEERS:
488 case NETDEV_BONDING_FAILOVER:
489 case NETDEV_RESEND_IGMP:
490 /* Propagate to vlan devices */
491 vlan_group_for_each_dev(grp, i, vlandev)
492 call_netdevice_notifiers(event, vlandev);
493 break;
494 }
495
496 out:
497 return NOTIFY_DONE;
498 }
499
500 static struct notifier_block vlan_notifier_block __read_mostly = {
501 .notifier_call = vlan_device_event,
502 };
503
504 /*
505 * VLAN IOCTL handler.
506 * o execute requested action or pass command to the device driver
507 * arg is really a struct vlan_ioctl_args __user *.
508 */
509 static int vlan_ioctl_handler(struct net *net, void __user *arg)
510 {
511 int err;
512 struct vlan_ioctl_args args;
513 struct net_device *dev = NULL;
514
515 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
516 return -EFAULT;
517
518 /* Null terminate this sucker, just in case. */
519 args.device1[sizeof(args.device1) - 1] = 0;
520 args.u.device2[sizeof(args.u.device2) - 1] = 0;
521
522 rtnl_lock();
523
524 switch (args.cmd) {
525 case SET_VLAN_INGRESS_PRIORITY_CMD:
526 case SET_VLAN_EGRESS_PRIORITY_CMD:
527 case SET_VLAN_FLAG_CMD:
528 case ADD_VLAN_CMD:
529 case DEL_VLAN_CMD:
530 case GET_VLAN_REALDEV_NAME_CMD:
531 case GET_VLAN_VID_CMD:
532 err = -ENODEV;
533 dev = __dev_get_by_name(net, args.device1);
534 if (!dev)
535 goto out;
536
537 err = -EINVAL;
538 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
539 goto out;
540 }
541
542 switch (args.cmd) {
543 case SET_VLAN_INGRESS_PRIORITY_CMD:
544 err = -EPERM;
545 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
546 break;
547 vlan_dev_set_ingress_priority(dev,
548 args.u.skb_priority,
549 args.vlan_qos);
550 err = 0;
551 break;
552
553 case SET_VLAN_EGRESS_PRIORITY_CMD:
554 err = -EPERM;
555 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
556 break;
557 err = vlan_dev_set_egress_priority(dev,
558 args.u.skb_priority,
559 args.vlan_qos);
560 break;
561
562 case SET_VLAN_FLAG_CMD:
563 err = -EPERM;
564 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
565 break;
566 err = vlan_dev_change_flags(dev,
567 args.vlan_qos ? args.u.flag : 0,
568 args.u.flag);
569 break;
570
571 case SET_VLAN_NAME_TYPE_CMD:
572 err = -EPERM;
573 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
574 break;
575 if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) {
576 struct vlan_net *vn;
577
578 vn = net_generic(net, vlan_net_id);
579 vn->name_type = args.u.name_type;
580 err = 0;
581 } else {
582 err = -EINVAL;
583 }
584 break;
585
586 case ADD_VLAN_CMD:
587 err = -EPERM;
588 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
589 break;
590 err = register_vlan_device(dev, args.u.VID);
591 break;
592
593 case DEL_VLAN_CMD:
594 err = -EPERM;
595 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
596 break;
597 unregister_vlan_dev(dev, NULL);
598 err = 0;
599 break;
600
601 case GET_VLAN_REALDEV_NAME_CMD:
602 err = 0;
603 vlan_dev_get_realdev_name(dev, args.u.device2);
604 if (copy_to_user(arg, &args,
605 sizeof(struct vlan_ioctl_args)))
606 err = -EFAULT;
607 break;
608
609 case GET_VLAN_VID_CMD:
610 err = 0;
611 args.u.VID = vlan_dev_vlan_id(dev);
612 if (copy_to_user(arg, &args,
613 sizeof(struct vlan_ioctl_args)))
614 err = -EFAULT;
615 break;
616
617 default:
618 err = -EOPNOTSUPP;
619 break;
620 }
621 out:
622 rtnl_unlock();
623 return err;
624 }
625
626 static struct sk_buff **vlan_gro_receive(struct sk_buff **head,
627 struct sk_buff *skb)
628 {
629 struct sk_buff *p, **pp = NULL;
630 struct vlan_hdr *vhdr;
631 unsigned int hlen, off_vlan;
632 const struct packet_offload *ptype;
633 __be16 type;
634 int flush = 1;
635
636 off_vlan = skb_gro_offset(skb);
637 hlen = off_vlan + sizeof(*vhdr);
638 vhdr = skb_gro_header_fast(skb, off_vlan);
639 if (skb_gro_header_hard(skb, hlen)) {
640 vhdr = skb_gro_header_slow(skb, hlen, off_vlan);
641 if (unlikely(!vhdr))
642 goto out;
643 }
644
645 type = vhdr->h_vlan_encapsulated_proto;
646
647 rcu_read_lock();
648 ptype = gro_find_receive_by_type(type);
649 if (!ptype)
650 goto out_unlock;
651
652 flush = 0;
653
654 for (p = *head; p; p = p->next) {
655 struct vlan_hdr *vhdr2;
656
657 if (!NAPI_GRO_CB(p)->same_flow)
658 continue;
659
660 vhdr2 = (struct vlan_hdr *)(p->data + off_vlan);
661 if (compare_vlan_header(vhdr, vhdr2))
662 NAPI_GRO_CB(p)->same_flow = 0;
663 }
664
665 skb_gro_pull(skb, sizeof(*vhdr));
666 skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr));
667 pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
668
669 out_unlock:
670 rcu_read_unlock();
671 out:
672 NAPI_GRO_CB(skb)->flush |= flush;
673
674 return pp;
675 }
676
677 static int vlan_gro_complete(struct sk_buff *skb, int nhoff)
678 {
679 struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff);
680 __be16 type = vhdr->h_vlan_encapsulated_proto;
681 struct packet_offload *ptype;
682 int err = -ENOENT;
683
684 rcu_read_lock();
685 ptype = gro_find_complete_by_type(type);
686 if (ptype)
687 err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr));
688
689 rcu_read_unlock();
690 return err;
691 }
692
693 static struct packet_offload vlan_packet_offloads[] __read_mostly = {
694 {
695 .type = cpu_to_be16(ETH_P_8021Q),
696 .priority = 10,
697 .callbacks = {
698 .gro_receive = vlan_gro_receive,
699 .gro_complete = vlan_gro_complete,
700 },
701 },
702 {
703 .type = cpu_to_be16(ETH_P_8021AD),
704 .priority = 10,
705 .callbacks = {
706 .gro_receive = vlan_gro_receive,
707 .gro_complete = vlan_gro_complete,
708 },
709 },
710 };
711
712 static int __net_init vlan_init_net(struct net *net)
713 {
714 struct vlan_net *vn = net_generic(net, vlan_net_id);
715 int err;
716
717 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
718
719 err = vlan_proc_init(net);
720
721 return err;
722 }
723
724 static void __net_exit vlan_exit_net(struct net *net)
725 {
726 vlan_proc_cleanup(net);
727 }
728
729 static struct pernet_operations vlan_net_ops = {
730 .init = vlan_init_net,
731 .exit = vlan_exit_net,
732 .id = &vlan_net_id,
733 .size = sizeof(struct vlan_net),
734 };
735
736 static int __init vlan_proto_init(void)
737 {
738 int err;
739 unsigned int i;
740
741 pr_info("%s v%s\n", vlan_fullname, vlan_version);
742
743 err = register_pernet_subsys(&vlan_net_ops);
744 if (err < 0)
745 goto err0;
746
747 err = register_netdevice_notifier(&vlan_notifier_block);
748 if (err < 0)
749 goto err2;
750
751 err = vlan_gvrp_init();
752 if (err < 0)
753 goto err3;
754
755 err = vlan_mvrp_init();
756 if (err < 0)
757 goto err4;
758
759 err = vlan_netlink_init();
760 if (err < 0)
761 goto err5;
762
763 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
764 dev_add_offload(&vlan_packet_offloads[i]);
765
766 vlan_ioctl_set(vlan_ioctl_handler);
767 return 0;
768
769 err5:
770 vlan_mvrp_uninit();
771 err4:
772 vlan_gvrp_uninit();
773 err3:
774 unregister_netdevice_notifier(&vlan_notifier_block);
775 err2:
776 unregister_pernet_subsys(&vlan_net_ops);
777 err0:
778 return err;
779 }
780
781 static void __exit vlan_cleanup_module(void)
782 {
783 unsigned int i;
784
785 vlan_ioctl_set(NULL);
786
787 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
788 dev_remove_offload(&vlan_packet_offloads[i]);
789
790 vlan_netlink_fini();
791
792 unregister_netdevice_notifier(&vlan_notifier_block);
793
794 unregister_pernet_subsys(&vlan_net_ops);
795 rcu_barrier(); /* Wait for completion of call_rcu()'s */
796
797 vlan_mvrp_uninit();
798 vlan_gvrp_uninit();
799 }
800
801 module_init(vlan_proto_init);
802 module_exit(vlan_cleanup_module);
803
804 MODULE_LICENSE("GPL");
805 MODULE_VERSION(DRV_VERSION);