1/*
2 * INET		802.1Q VLAN
3 *		Ethernet-type device handling.
4 *
5 * Authors:	Ben Greear <greearb@candelatech.com>
6 *              Please send support related email to: netdev@vger.kernel.org
7 *              VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
8 *
9 * Fixes:
10 *              Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11 *		Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12 *		Correct all the locking - David S. Miller <davem@redhat.com>;
13 *		Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
14 *
15 *		This program is free software; you can redistribute it and/or
16 *		modify it under the terms of the GNU General Public License
17 *		as published by the Free Software Foundation; either version
18 *		2 of the License, or (at your option) any later version.
19 */
20
21#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22
23#include <linux/capability.h>
24#include <linux/module.h>
25#include <linux/netdevice.h>
26#include <linux/skbuff.h>
27#include <linux/slab.h>
28#include <linux/init.h>
29#include <linux/rculist.h>
30#include <net/p8022.h>
31#include <net/arp.h>
32#include <linux/rtnetlink.h>
33#include <linux/notifier.h>
34#include <net/rtnetlink.h>
35#include <net/net_namespace.h>
36#include <net/netns/generic.h>
37#include <asm/uaccess.h>
38
39#include <linux/if_vlan.h>
40#include "vlan.h"
41#include "vlanproc.h"
42
43#define DRV_VERSION "1.8"
44
45/* Global VLAN variables */
46
47int vlan_net_id __read_mostly;
48
49const char vlan_fullname[] = "802.1Q VLAN Support";
50const char vlan_version[] = DRV_VERSION;
51
52/* End of global variables definitions. */
53
54static int vlan_group_prealloc_vid(struct vlan_group *vg,
55				   __be16 vlan_proto, u16 vlan_id)
56{
57	struct net_device **array;
58	unsigned int pidx, vidx;
59	unsigned int size;
60
61	ASSERT_RTNL();
62
63	pidx  = vlan_proto_idx(vlan_proto);
64	vidx  = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65	array = vg->vlan_devices_arrays[pidx][vidx];
66	if (array != NULL)
67		return 0;
68
69	size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70	array = kzalloc(size, GFP_KERNEL);
71	if (array == NULL)
72		return -ENOBUFS;
73
74	vg->vlan_devices_arrays[pidx][vidx] = array;
75	return 0;
76}
77
78void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
79{
80	struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
81	struct net_device *real_dev = vlan->real_dev;
82	struct vlan_info *vlan_info;
83	struct vlan_group *grp;
84	u16 vlan_id = vlan->vlan_id;
85
86	ASSERT_RTNL();
87
88	vlan_info = rtnl_dereference(real_dev->vlan_info);
89	BUG_ON(!vlan_info);
90
91	grp = &vlan_info->grp;
92
93	grp->nr_vlan_devs--;
94
95	if (vlan->flags & VLAN_FLAG_MVRP)
96		vlan_mvrp_request_leave(dev);
97	if (vlan->flags & VLAN_FLAG_GVRP)
98		vlan_gvrp_request_leave(dev);
99
100	vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
101
102	netdev_upper_dev_unlink(real_dev, dev);
103	/* Because unregister_netdevice_queue() makes sure at least one rcu
104	 * grace period is respected before device freeing,
105	 * we dont need to call synchronize_net() here.
106	 */
107	unregister_netdevice_queue(dev, head);
108
109	if (grp->nr_vlan_devs == 0) {
110		vlan_mvrp_uninit_applicant(real_dev);
111		vlan_gvrp_uninit_applicant(real_dev);
112	}
113
114	/* Take it out of our own structures, but be sure to interlock with
115	 * HW accelerating devices or SW vlan input packet processing if
116	 * VLAN is not 0 (leave it there for 802.1p).
117	 */
118	if (vlan_id)
119		vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
120
121	/* Get rid of the vlan's reference to real_dev */
122	dev_put(real_dev);
123}
124
125int vlan_check_real_dev(struct net_device *real_dev,
126			__be16 protocol, u16 vlan_id)
127{
128	const char *name = real_dev->name;
129
130	if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
131		pr_info("VLANs not supported on %s\n", name);
132		return -EOPNOTSUPP;
133	}
134
135	if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL)
136		return -EEXIST;
137
138	return 0;
139}
140
141int register_vlan_dev(struct net_device *dev)
142{
143	struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
144	struct net_device *real_dev = vlan->real_dev;
145	u16 vlan_id = vlan->vlan_id;
146	struct vlan_info *vlan_info;
147	struct vlan_group *grp;
148	int err;
149
150	err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
151	if (err)
152		return err;
153
154	vlan_info = rtnl_dereference(real_dev->vlan_info);
155	/* vlan_info should be there now. vlan_vid_add took care of it */
156	BUG_ON(!vlan_info);
157
158	grp = &vlan_info->grp;
159	if (grp->nr_vlan_devs == 0) {
160		err = vlan_gvrp_init_applicant(real_dev);
161		if (err < 0)
162			goto out_vid_del;
163		err = vlan_mvrp_init_applicant(real_dev);
164		if (err < 0)
165			goto out_uninit_gvrp;
166	}
167
168	err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
169	if (err < 0)
170		goto out_uninit_mvrp;
171
172	vlan->nest_level = dev_get_nest_level(real_dev, is_vlan_dev) + 1;
173	err = register_netdevice(dev);
174	if (err < 0)
175		goto out_uninit_mvrp;
176
177	err = netdev_upper_dev_link(real_dev, dev);
178	if (err)
179		goto out_unregister_netdev;
180
181	/* Account for reference in struct vlan_dev_priv */
182	dev_hold(real_dev);
183
184	netif_stacked_transfer_operstate(real_dev, dev);
185	linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
186
187	/* So, got the sucker initialized, now lets place
188	 * it into our local structure.
189	 */
190	vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
191	grp->nr_vlan_devs++;
192
193	return 0;
194
195out_unregister_netdev:
196	unregister_netdevice(dev);
197out_uninit_mvrp:
198	if (grp->nr_vlan_devs == 0)
199		vlan_mvrp_uninit_applicant(real_dev);
200out_uninit_gvrp:
201	if (grp->nr_vlan_devs == 0)
202		vlan_gvrp_uninit_applicant(real_dev);
203out_vid_del:
204	vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
205	return err;
206}
207
208/*  Attach a VLAN device to a mac address (ie Ethernet Card).
209 *  Returns 0 if the device was created or a negative error code otherwise.
210 */
211static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
212{
213	struct net_device *new_dev;
214	struct vlan_dev_priv *vlan;
215	struct net *net = dev_net(real_dev);
216	struct vlan_net *vn = net_generic(net, vlan_net_id);
217	char name[IFNAMSIZ];
218	int err;
219
220	if (vlan_id >= VLAN_VID_MASK)
221		return -ERANGE;
222
223	err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id);
224	if (err < 0)
225		return err;
226
227	/* Gotta set up the fields for the device. */
228	switch (vn->name_type) {
229	case VLAN_NAME_TYPE_RAW_PLUS_VID:
230		/* name will look like:	 eth1.0005 */
231		snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
232		break;
233	case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
234		/* Put our vlan.VID in the name.
235		 * Name will look like:	 vlan5
236		 */
237		snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
238		break;
239	case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
240		/* Put our vlan.VID in the name.
241		 * Name will look like:	 eth0.5
242		 */
243		snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
244		break;
245	case VLAN_NAME_TYPE_PLUS_VID:
246		/* Put our vlan.VID in the name.
247		 * Name will look like:	 vlan0005
248		 */
249	default:
250		snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
251	}
252
253	new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
254			       NET_NAME_UNKNOWN, vlan_setup);
255
256	if (new_dev == NULL)
257		return -ENOBUFS;
258
259	dev_net_set(new_dev, net);
260	/* need 4 bytes for extra VLAN header info,
261	 * hope the underlying device can handle it.
262	 */
263	new_dev->mtu = real_dev->mtu;
264	new_dev->priv_flags |= (real_dev->priv_flags & IFF_UNICAST_FLT);
265
266	vlan = vlan_dev_priv(new_dev);
267	vlan->vlan_proto = htons(ETH_P_8021Q);
268	vlan->vlan_id = vlan_id;
269	vlan->real_dev = real_dev;
270	vlan->dent = NULL;
271	vlan->flags = VLAN_FLAG_REORDER_HDR;
272
273	new_dev->rtnl_link_ops = &vlan_link_ops;
274	err = register_vlan_dev(new_dev);
275	if (err < 0)
276		goto out_free_newdev;
277
278	return 0;
279
280out_free_newdev:
281	free_netdev(new_dev);
282	return err;
283}
284
285static void vlan_sync_address(struct net_device *dev,
286			      struct net_device *vlandev)
287{
288	struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
289
290	/* May be called without an actual change */
291	if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
292		return;
293
294	/* vlan address was different from the old address and is equal to
295	 * the new address */
296	if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
297	    ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
298		dev_uc_del(dev, vlandev->dev_addr);
299
300	/* vlan address was equal to the old address and is different from
301	 * the new address */
302	if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
303	    !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
304		dev_uc_add(dev, vlandev->dev_addr);
305
306	ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
307}
308
309static void vlan_transfer_features(struct net_device *dev,
310				   struct net_device *vlandev)
311{
312	struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
313
314	vlandev->gso_max_size = dev->gso_max_size;
315
316	if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
317		vlandev->hard_header_len = dev->hard_header_len;
318	else
319		vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
320
321#if IS_ENABLED(CONFIG_FCOE)
322	vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
323#endif
324
325	netdev_update_features(vlandev);
326}
327
328static int __vlan_device_event(struct net_device *dev, unsigned long event)
329{
330	int err = 0;
331
332	switch (event) {
333	case NETDEV_CHANGENAME:
334		vlan_proc_rem_dev(dev);
335		err = vlan_proc_add_dev(dev);
336		break;
337	case NETDEV_REGISTER:
338		err = vlan_proc_add_dev(dev);
339		break;
340	case NETDEV_UNREGISTER:
341		vlan_proc_rem_dev(dev);
342		break;
343	}
344
345	return err;
346}
347
348static int vlan_device_event(struct notifier_block *unused, unsigned long event,
349			     void *ptr)
350{
351	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
352	struct vlan_group *grp;
353	struct vlan_info *vlan_info;
354	int i, flgs;
355	struct net_device *vlandev;
356	struct vlan_dev_priv *vlan;
357	bool last = false;
358	LIST_HEAD(list);
359
360	if (is_vlan_dev(dev)) {
361		int err = __vlan_device_event(dev, event);
362
363		if (err)
364			return notifier_from_errno(err);
365	}
366
367	if ((event == NETDEV_UP) &&
368	    (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
369		pr_info("adding VLAN 0 to HW filter on device %s\n",
370			dev->name);
371		vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
372	}
373
374	vlan_info = rtnl_dereference(dev->vlan_info);
375	if (!vlan_info)
376		goto out;
377	grp = &vlan_info->grp;
378
379	/* It is OK that we do not hold the group lock right now,
380	 * as we run under the RTNL lock.
381	 */
382
383	switch (event) {
384	case NETDEV_CHANGE:
385		/* Propagate real device state to vlan devices */
386		vlan_group_for_each_dev(grp, i, vlandev)
387			netif_stacked_transfer_operstate(dev, vlandev);
388		break;
389
390	case NETDEV_CHANGEADDR:
391		/* Adjust unicast filters on underlying device */
392		vlan_group_for_each_dev(grp, i, vlandev) {
393			flgs = vlandev->flags;
394			if (!(flgs & IFF_UP))
395				continue;
396
397			vlan_sync_address(dev, vlandev);
398		}
399		break;
400
401	case NETDEV_CHANGEMTU:
402		vlan_group_for_each_dev(grp, i, vlandev) {
403			if (vlandev->mtu <= dev->mtu)
404				continue;
405
406			dev_set_mtu(vlandev, dev->mtu);
407		}
408		break;
409
410	case NETDEV_FEAT_CHANGE:
411		/* Propagate device features to underlying device */
412		vlan_group_for_each_dev(grp, i, vlandev)
413			vlan_transfer_features(dev, vlandev);
414		break;
415
416	case NETDEV_DOWN: {
417		struct net_device *tmp;
418		LIST_HEAD(close_list);
419
420		if (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)
421			vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
422
423		/* Put all VLANs for this dev in the down state too.  */
424		vlan_group_for_each_dev(grp, i, vlandev) {
425			flgs = vlandev->flags;
426			if (!(flgs & IFF_UP))
427				continue;
428
429			vlan = vlan_dev_priv(vlandev);
430			if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
431				list_add(&vlandev->close_list, &close_list);
432		}
433
434		dev_close_many(&close_list, false);
435
436		list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
437			netif_stacked_transfer_operstate(dev, vlandev);
438			list_del_init(&vlandev->close_list);
439		}
440		list_del(&close_list);
441		break;
442	}
443	case NETDEV_UP:
444		/* Put all VLANs for this dev in the up state too.  */
445		vlan_group_for_each_dev(grp, i, vlandev) {
446			flgs = dev_get_flags(vlandev);
447			if (flgs & IFF_UP)
448				continue;
449
450			vlan = vlan_dev_priv(vlandev);
451			if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
452				dev_change_flags(vlandev, flgs | IFF_UP);
453			netif_stacked_transfer_operstate(dev, vlandev);
454		}
455		break;
456
457	case NETDEV_UNREGISTER:
458		/* twiddle thumbs on netns device moves */
459		if (dev->reg_state != NETREG_UNREGISTERING)
460			break;
461
462		vlan_group_for_each_dev(grp, i, vlandev) {
463			/* removal of last vid destroys vlan_info, abort
464			 * afterwards */
465			if (vlan_info->nr_vids == 1)
466				last = true;
467
468			unregister_vlan_dev(vlandev, &list);
469			if (last)
470				break;
471		}
472		unregister_netdevice_many(&list);
473		break;
474
475	case NETDEV_PRE_TYPE_CHANGE:
476		/* Forbid underlaying device to change its type. */
477		if (vlan_uses_dev(dev))
478			return NOTIFY_BAD;
479		break;
480
481	case NETDEV_NOTIFY_PEERS:
482	case NETDEV_BONDING_FAILOVER:
483	case NETDEV_RESEND_IGMP:
484		/* Propagate to vlan devices */
485		vlan_group_for_each_dev(grp, i, vlandev)
486			call_netdevice_notifiers(event, vlandev);
487		break;
488	}
489
490out:
491	return NOTIFY_DONE;
492}
493
494static struct notifier_block vlan_notifier_block __read_mostly = {
495	.notifier_call = vlan_device_event,
496};
497
498/*
499 *	VLAN IOCTL handler.
500 *	o execute requested action or pass command to the device driver
501 *   arg is really a struct vlan_ioctl_args __user *.
502 */
503static int vlan_ioctl_handler(struct net *net, void __user *arg)
504{
505	int err;
506	struct vlan_ioctl_args args;
507	struct net_device *dev = NULL;
508
509	if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
510		return -EFAULT;
511
512	/* Null terminate this sucker, just in case. */
513	args.device1[23] = 0;
514	args.u.device2[23] = 0;
515
516	rtnl_lock();
517
518	switch (args.cmd) {
519	case SET_VLAN_INGRESS_PRIORITY_CMD:
520	case SET_VLAN_EGRESS_PRIORITY_CMD:
521	case SET_VLAN_FLAG_CMD:
522	case ADD_VLAN_CMD:
523	case DEL_VLAN_CMD:
524	case GET_VLAN_REALDEV_NAME_CMD:
525	case GET_VLAN_VID_CMD:
526		err = -ENODEV;
527		dev = __dev_get_by_name(net, args.device1);
528		if (!dev)
529			goto out;
530
531		err = -EINVAL;
532		if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
533			goto out;
534	}
535
536	switch (args.cmd) {
537	case SET_VLAN_INGRESS_PRIORITY_CMD:
538		err = -EPERM;
539		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
540			break;
541		vlan_dev_set_ingress_priority(dev,
542					      args.u.skb_priority,
543					      args.vlan_qos);
544		err = 0;
545		break;
546
547	case SET_VLAN_EGRESS_PRIORITY_CMD:
548		err = -EPERM;
549		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
550			break;
551		err = vlan_dev_set_egress_priority(dev,
552						   args.u.skb_priority,
553						   args.vlan_qos);
554		break;
555
556	case SET_VLAN_FLAG_CMD:
557		err = -EPERM;
558		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
559			break;
560		err = vlan_dev_change_flags(dev,
561					    args.vlan_qos ? args.u.flag : 0,
562					    args.u.flag);
563		break;
564
565	case SET_VLAN_NAME_TYPE_CMD:
566		err = -EPERM;
567		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
568			break;
569		if ((args.u.name_type >= 0) &&
570		    (args.u.name_type < VLAN_NAME_TYPE_HIGHEST)) {
571			struct vlan_net *vn;
572
573			vn = net_generic(net, vlan_net_id);
574			vn->name_type = args.u.name_type;
575			err = 0;
576		} else {
577			err = -EINVAL;
578		}
579		break;
580
581	case ADD_VLAN_CMD:
582		err = -EPERM;
583		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
584			break;
585		err = register_vlan_device(dev, args.u.VID);
586		break;
587
588	case DEL_VLAN_CMD:
589		err = -EPERM;
590		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
591			break;
592		unregister_vlan_dev(dev, NULL);
593		err = 0;
594		break;
595
596	case GET_VLAN_REALDEV_NAME_CMD:
597		err = 0;
598		vlan_dev_get_realdev_name(dev, args.u.device2);
599		if (copy_to_user(arg, &args,
600				 sizeof(struct vlan_ioctl_args)))
601			err = -EFAULT;
602		break;
603
604	case GET_VLAN_VID_CMD:
605		err = 0;
606		args.u.VID = vlan_dev_vlan_id(dev);
607		if (copy_to_user(arg, &args,
608				 sizeof(struct vlan_ioctl_args)))
609		      err = -EFAULT;
610		break;
611
612	default:
613		err = -EOPNOTSUPP;
614		break;
615	}
616out:
617	rtnl_unlock();
618	return err;
619}
620
621static struct sk_buff **vlan_gro_receive(struct sk_buff **head,
622					 struct sk_buff *skb)
623{
624	struct sk_buff *p, **pp = NULL;
625	struct vlan_hdr *vhdr;
626	unsigned int hlen, off_vlan;
627	const struct packet_offload *ptype;
628	__be16 type;
629	int flush = 1;
630
631	off_vlan = skb_gro_offset(skb);
632	hlen = off_vlan + sizeof(*vhdr);
633	vhdr = skb_gro_header_fast(skb, off_vlan);
634	if (skb_gro_header_hard(skb, hlen)) {
635		vhdr = skb_gro_header_slow(skb, hlen, off_vlan);
636		if (unlikely(!vhdr))
637			goto out;
638	}
639
640	type = vhdr->h_vlan_encapsulated_proto;
641
642	rcu_read_lock();
643	ptype = gro_find_receive_by_type(type);
644	if (!ptype)
645		goto out_unlock;
646
647	flush = 0;
648
649	for (p = *head; p; p = p->next) {
650		struct vlan_hdr *vhdr2;
651
652		if (!NAPI_GRO_CB(p)->same_flow)
653			continue;
654
655		vhdr2 = (struct vlan_hdr *)(p->data + off_vlan);
656		if (compare_vlan_header(vhdr, vhdr2))
657			NAPI_GRO_CB(p)->same_flow = 0;
658	}
659
660	skb_gro_pull(skb, sizeof(*vhdr));
661	skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr));
662	pp = ptype->callbacks.gro_receive(head, skb);
663
664out_unlock:
665	rcu_read_unlock();
666out:
667	NAPI_GRO_CB(skb)->flush |= flush;
668
669	return pp;
670}
671
672static int vlan_gro_complete(struct sk_buff *skb, int nhoff)
673{
674	struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff);
675	__be16 type = vhdr->h_vlan_encapsulated_proto;
676	struct packet_offload *ptype;
677	int err = -ENOENT;
678
679	rcu_read_lock();
680	ptype = gro_find_complete_by_type(type);
681	if (ptype)
682		err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr));
683
684	rcu_read_unlock();
685	return err;
686}
687
688static struct packet_offload vlan_packet_offloads[] __read_mostly = {
689	{
690		.type = cpu_to_be16(ETH_P_8021Q),
691		.priority = 10,
692		.callbacks = {
693			.gro_receive = vlan_gro_receive,
694			.gro_complete = vlan_gro_complete,
695		},
696	},
697	{
698		.type = cpu_to_be16(ETH_P_8021AD),
699		.priority = 10,
700		.callbacks = {
701			.gro_receive = vlan_gro_receive,
702			.gro_complete = vlan_gro_complete,
703		},
704	},
705};
706
707static int __net_init vlan_init_net(struct net *net)
708{
709	struct vlan_net *vn = net_generic(net, vlan_net_id);
710	int err;
711
712	vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
713
714	err = vlan_proc_init(net);
715
716	return err;
717}
718
719static void __net_exit vlan_exit_net(struct net *net)
720{
721	vlan_proc_cleanup(net);
722}
723
724static struct pernet_operations vlan_net_ops = {
725	.init = vlan_init_net,
726	.exit = vlan_exit_net,
727	.id   = &vlan_net_id,
728	.size = sizeof(struct vlan_net),
729};
730
731static int __init vlan_proto_init(void)
732{
733	int err;
734	unsigned int i;
735
736	pr_info("%s v%s\n", vlan_fullname, vlan_version);
737
738	err = register_pernet_subsys(&vlan_net_ops);
739	if (err < 0)
740		goto err0;
741
742	err = register_netdevice_notifier(&vlan_notifier_block);
743	if (err < 0)
744		goto err2;
745
746	err = vlan_gvrp_init();
747	if (err < 0)
748		goto err3;
749
750	err = vlan_mvrp_init();
751	if (err < 0)
752		goto err4;
753
754	err = vlan_netlink_init();
755	if (err < 0)
756		goto err5;
757
758	for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
759		dev_add_offload(&vlan_packet_offloads[i]);
760
761	vlan_ioctl_set(vlan_ioctl_handler);
762	return 0;
763
764err5:
765	vlan_mvrp_uninit();
766err4:
767	vlan_gvrp_uninit();
768err3:
769	unregister_netdevice_notifier(&vlan_notifier_block);
770err2:
771	unregister_pernet_subsys(&vlan_net_ops);
772err0:
773	return err;
774}
775
776static void __exit vlan_cleanup_module(void)
777{
778	unsigned int i;
779
780	vlan_ioctl_set(NULL);
781
782	for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
783		dev_remove_offload(&vlan_packet_offloads[i]);
784
785	vlan_netlink_fini();
786
787	unregister_netdevice_notifier(&vlan_notifier_block);
788
789	unregister_pernet_subsys(&vlan_net_ops);
790	rcu_barrier(); /* Wait for completion of call_rcu()'s */
791
792	vlan_mvrp_uninit();
793	vlan_gvrp_uninit();
794}
795
796module_init(vlan_proto_init);
797module_exit(vlan_cleanup_module);
798
799MODULE_LICENSE("GPL");
800MODULE_VERSION(DRV_VERSION);
801