]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* -*- linux-c -*- |
2 | * INET 802.1Q VLAN | |
3 | * Ethernet-type device handling. | |
4 | * | |
5 | * Authors: Ben Greear <greearb@candelatech.com> | |
6 | * Please send support related email to: vlan@scry.wanfear.com | |
7 | * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html | |
122952fc | 8 | * |
1da177e4 LT |
9 | * Fixes: Mar 22 2001: Martin Bokaemper <mbokaemper@unispherenetworks.com> |
10 | * - reset skb->pkt_type on incoming packets when MAC was changed | |
11 | * - see that changed MAC is saddr for outgoing packets | |
12 | * Oct 20, 2001: Ard van Breeman: | |
13 | * - Fix MC-list, finally. | |
14 | * - Flush MC-list on VLAN destroy. | |
122952fc | 15 | * |
1da177e4 LT |
16 | * |
17 | * This program is free software; you can redistribute it and/or | |
18 | * modify it under the terms of the GNU General Public License | |
19 | * as published by the Free Software Foundation; either version | |
20 | * 2 of the License, or (at your option) any later version. | |
21 | */ | |
22 | ||
23 | #include <linux/module.h> | |
24 | #include <linux/mm.h> | |
25 | #include <linux/in.h> | |
26 | #include <linux/init.h> | |
27 | #include <asm/uaccess.h> /* for copy_from_user */ | |
28 | #include <linux/skbuff.h> | |
29 | #include <linux/netdevice.h> | |
30 | #include <linux/etherdevice.h> | |
31 | #include <net/datalink.h> | |
32 | #include <net/p8022.h> | |
33 | #include <net/arp.h> | |
34 | ||
35 | #include "vlan.h" | |
36 | #include "vlanproc.h" | |
37 | #include <linux/if_vlan.h> | |
38 | #include <net/ip.h> | |
39 | ||
40 | /* | |
41 | * Rebuild the Ethernet MAC header. This is called after an ARP | |
42 | * (or in future other address resolution) has completed on this | |
43 | * sk_buff. We now let ARP fill in the other fields. | |
44 | * | |
45 | * This routine CANNOT use cached dst->neigh! | |
46 | * Really, it is used only when dst->neigh is wrong. | |
47 | * | |
48 | * TODO: This needs a checkup, I'm ignorant here. --BLG | |
49 | */ | |
ef3eb3e5 | 50 | static int vlan_dev_rebuild_header(struct sk_buff *skb) |
1da177e4 LT |
51 | { |
52 | struct net_device *dev = skb->dev; | |
53 | struct vlan_ethhdr *veth = (struct vlan_ethhdr *)(skb->data); | |
54 | ||
55 | switch (veth->h_vlan_encapsulated_proto) { | |
56 | #ifdef CONFIG_INET | |
57 | case __constant_htons(ETH_P_IP): | |
58 | ||
59 | /* TODO: Confirm this will work with VLAN headers... */ | |
60 | return arp_find(veth->h_dest, skb); | |
122952fc | 61 | #endif |
1da177e4 | 62 | default: |
40f98e1a PM |
63 | pr_debug("%s: unable to resolve type %X addresses.\n", |
64 | dev->name, ntohs(veth->h_vlan_encapsulated_proto)); | |
122952fc | 65 | |
1da177e4 LT |
66 | memcpy(veth->h_source, dev->dev_addr, ETH_ALEN); |
67 | break; | |
3ff50b79 | 68 | } |
1da177e4 LT |
69 | |
70 | return 0; | |
71 | } | |
72 | ||
73 | static inline struct sk_buff *vlan_check_reorder_header(struct sk_buff *skb) | |
74 | { | |
9dfebcc6 | 75 | if (vlan_dev_info(skb->dev)->flags & VLAN_FLAG_REORDER_HDR) { |
1da177e4 LT |
76 | if (skb_shared(skb) || skb_cloned(skb)) { |
77 | struct sk_buff *nskb = skb_copy(skb, GFP_ATOMIC); | |
78 | kfree_skb(skb); | |
79 | skb = nskb; | |
80 | } | |
81 | if (skb) { | |
82 | /* Lifted from Gleb's VLAN code... */ | |
83 | memmove(skb->data - ETH_HLEN, | |
84 | skb->data - VLAN_ETH_HLEN, 12); | |
b0e380b1 | 85 | skb->mac_header += VLAN_HLEN; |
1da177e4 LT |
86 | } |
87 | } | |
88 | ||
89 | return skb; | |
90 | } | |
91 | ||
92 | /* | |
122952fc | 93 | * Determine the packet's protocol ID. The rule here is that we |
1da177e4 LT |
94 | * assume 802.3 if the type field is short enough to be a length. |
95 | * This is normal practice and works for any 'now in use' protocol. | |
96 | * | |
97 | * Also, at this point we assume that we ARE dealing exclusively with | |
98 | * VLAN packets, or packets that should be made into VLAN packets based | |
99 | * on a default VLAN ID. | |
100 | * | |
101 | * NOTE: Should be similar to ethernet/eth.c. | |
102 | * | |
103 | * SANITY NOTE: This method is called when a packet is moving up the stack | |
104 | * towards userland. To get here, it would have already passed | |
105 | * through the ethernet/eth.c eth_type_trans() method. | |
106 | * SANITY NOTE 2: We are referencing to the VLAN_HDR frields, which MAY be | |
107 | * stored UNALIGNED in the memory. RISC systems don't like | |
108 | * such cases very much... | |
109 | * SANITY NOTE 2a: According to Dave Miller & Alexey, it will always be aligned, | |
110 | * so there doesn't need to be any of the unaligned stuff. It has | |
111 | * been commented out now... --Ben | |
112 | * | |
113 | */ | |
114 | int vlan_skb_recv(struct sk_buff *skb, struct net_device *dev, | |
122952fc | 115 | struct packet_type* ptype, struct net_device *orig_dev) |
1da177e4 LT |
116 | { |
117 | unsigned char *rawp = NULL; | |
e7c243c9 | 118 | struct vlan_hdr *vhdr; |
1da177e4 LT |
119 | unsigned short vid; |
120 | struct net_device_stats *stats; | |
121 | unsigned short vlan_TCI; | |
3c3f8f25 | 122 | __be16 proto; |
1da177e4 | 123 | |
e730c155 EB |
124 | if (dev->nd_net != &init_net) { |
125 | kfree_skb(skb); | |
126 | return -1; | |
127 | } | |
128 | ||
e7c243c9 EP |
129 | if ((skb = skb_share_check(skb, GFP_ATOMIC)) == NULL) |
130 | return -1; | |
131 | ||
132 | if (unlikely(!pskb_may_pull(skb, VLAN_HLEN))) { | |
133 | kfree_skb(skb); | |
134 | return -1; | |
135 | } | |
136 | ||
137 | vhdr = (struct vlan_hdr *)(skb->data); | |
138 | ||
1da177e4 LT |
139 | /* vlan_TCI = ntohs(get_unaligned(&vhdr->h_vlan_TCI)); */ |
140 | vlan_TCI = ntohs(vhdr->h_vlan_TCI); | |
141 | ||
142 | vid = (vlan_TCI & VLAN_VID_MASK); | |
143 | ||
1da177e4 LT |
144 | /* Ok, we will find the correct VLAN device, strip the header, |
145 | * and then go on as usual. | |
146 | */ | |
147 | ||
148 | /* We have 12 bits of vlan ID. | |
149 | * | |
150 | * We must not drop allow preempt until we hold a | |
151 | * reference to the device (netif_rx does that) or we | |
152 | * fail. | |
153 | */ | |
154 | ||
155 | rcu_read_lock(); | |
156 | skb->dev = __find_vlan_dev(dev, vid); | |
157 | if (!skb->dev) { | |
158 | rcu_read_unlock(); | |
40f98e1a PM |
159 | pr_debug("%s: ERROR: No net_device for VID: %u on dev: %s [%i]\n", |
160 | __FUNCTION__, (unsigned int)vid, dev->name, dev->ifindex); | |
1da177e4 LT |
161 | kfree_skb(skb); |
162 | return -1; | |
163 | } | |
164 | ||
165 | skb->dev->last_rx = jiffies; | |
166 | ||
167 | /* Bump the rx counters for the VLAN device. */ | |
7bd38d77 | 168 | stats = &skb->dev->stats; |
1da177e4 LT |
169 | stats->rx_packets++; |
170 | stats->rx_bytes += skb->len; | |
171 | ||
cbb042f9 HX |
172 | /* Take off the VLAN header (4 bytes currently) */ |
173 | skb_pull_rcsum(skb, VLAN_HLEN); | |
a388442c | 174 | |
1da177e4 LT |
175 | /* |
176 | * Deal with ingress priority mapping. | |
177 | */ | |
178 | skb->priority = vlan_get_ingress_priority(skb->dev, ntohs(vhdr->h_vlan_TCI)); | |
179 | ||
40f98e1a PM |
180 | pr_debug("%s: priority: %u for TCI: %hu\n", |
181 | __FUNCTION__, skb->priority, ntohs(vhdr->h_vlan_TCI)); | |
1da177e4 LT |
182 | |
183 | /* The ethernet driver already did the pkt_type calculations | |
184 | * for us... | |
185 | */ | |
186 | switch (skb->pkt_type) { | |
187 | case PACKET_BROADCAST: /* Yeah, stats collect these together.. */ | |
188 | // stats->broadcast ++; // no such counter :-( | |
189 | break; | |
190 | ||
191 | case PACKET_MULTICAST: | |
192 | stats->multicast++; | |
193 | break; | |
194 | ||
122952fc | 195 | case PACKET_OTHERHOST: |
1da177e4 LT |
196 | /* Our lower layer thinks this is not local, let's make sure. |
197 | * This allows the VLAN to have a different MAC than the underlying | |
198 | * device, and still route correctly. | |
199 | */ | |
d3f4a687 | 200 | if (!compare_ether_addr(eth_hdr(skb)->h_dest, skb->dev->dev_addr)) { |
1da177e4 LT |
201 | /* It is for our (changed) MAC-address! */ |
202 | skb->pkt_type = PACKET_HOST; | |
203 | } | |
204 | break; | |
205 | default: | |
206 | break; | |
3ff50b79 | 207 | } |
1da177e4 LT |
208 | |
209 | /* Was a VLAN packet, grab the encapsulated protocol, which the layer | |
210 | * three protocols care about. | |
211 | */ | |
212 | /* proto = get_unaligned(&vhdr->h_vlan_encapsulated_proto); */ | |
213 | proto = vhdr->h_vlan_encapsulated_proto; | |
214 | ||
215 | skb->protocol = proto; | |
216 | if (ntohs(proto) >= 1536) { | |
217 | /* place it back on the queue to be handled by | |
218 | * true layer 3 protocols. | |
219 | */ | |
220 | ||
221 | /* See if we are configured to re-write the VLAN header | |
222 | * to make it look like ethernet... | |
223 | */ | |
224 | skb = vlan_check_reorder_header(skb); | |
225 | ||
226 | /* Can be null if skb-clone fails when re-ordering */ | |
227 | if (skb) { | |
228 | netif_rx(skb); | |
229 | } else { | |
230 | /* TODO: Add a more specific counter here. */ | |
231 | stats->rx_errors++; | |
232 | } | |
233 | rcu_read_unlock(); | |
234 | return 0; | |
235 | } | |
236 | ||
237 | rawp = skb->data; | |
238 | ||
239 | /* | |
240 | * This is a magic hack to spot IPX packets. Older Novell breaks | |
241 | * the protocol design and runs IPX over 802.3 without an 802.2 LLC | |
242 | * layer. We look for FFFF which isn't a used 802.2 SSAP/DSAP. This | |
243 | * won't work for fault tolerant netware but does for the rest. | |
244 | */ | |
245 | if (*(unsigned short *)rawp == 0xFFFF) { | |
b93b7eeb | 246 | skb->protocol = htons(ETH_P_802_3); |
1da177e4 LT |
247 | /* place it back on the queue to be handled by true layer 3 protocols. |
248 | */ | |
249 | ||
250 | /* See if we are configured to re-write the VLAN header | |
251 | * to make it look like ethernet... | |
252 | */ | |
253 | skb = vlan_check_reorder_header(skb); | |
254 | ||
255 | /* Can be null if skb-clone fails when re-ordering */ | |
256 | if (skb) { | |
257 | netif_rx(skb); | |
258 | } else { | |
259 | /* TODO: Add a more specific counter here. */ | |
260 | stats->rx_errors++; | |
261 | } | |
262 | rcu_read_unlock(); | |
263 | return 0; | |
264 | } | |
265 | ||
266 | /* | |
267 | * Real 802.2 LLC | |
268 | */ | |
b93b7eeb | 269 | skb->protocol = htons(ETH_P_802_2); |
1da177e4 LT |
270 | /* place it back on the queue to be handled by upper layer protocols. |
271 | */ | |
272 | ||
273 | /* See if we are configured to re-write the VLAN header | |
274 | * to make it look like ethernet... | |
275 | */ | |
276 | skb = vlan_check_reorder_header(skb); | |
277 | ||
278 | /* Can be null if skb-clone fails when re-ordering */ | |
279 | if (skb) { | |
280 | netif_rx(skb); | |
281 | } else { | |
282 | /* TODO: Add a more specific counter here. */ | |
283 | stats->rx_errors++; | |
284 | } | |
285 | rcu_read_unlock(); | |
286 | return 0; | |
287 | } | |
288 | ||
289 | static inline unsigned short vlan_dev_get_egress_qos_mask(struct net_device* dev, | |
290 | struct sk_buff* skb) | |
291 | { | |
292 | struct vlan_priority_tci_mapping *mp = | |
9dfebcc6 | 293 | vlan_dev_info(dev)->egress_priority_map[(skb->priority & 0xF)]; |
1da177e4 LT |
294 | |
295 | while (mp) { | |
296 | if (mp->priority == skb->priority) { | |
297 | return mp->vlan_qos; /* This should already be shifted to mask | |
298 | * correctly with the VLAN's TCI | |
299 | */ | |
300 | } | |
301 | mp = mp->next; | |
302 | } | |
303 | return 0; | |
304 | } | |
305 | ||
306 | /* | |
122952fc | 307 | * Create the VLAN header for an arbitrary protocol layer |
1da177e4 LT |
308 | * |
309 | * saddr=NULL means use device source address | |
310 | * daddr=NULL means leave destination address (eg unresolved arp) | |
311 | * | |
312 | * This is called when the SKB is moving down the stack towards the | |
313 | * physical devices. | |
314 | */ | |
ef3eb3e5 PM |
315 | static int vlan_dev_hard_header(struct sk_buff *skb, struct net_device *dev, |
316 | unsigned short type, | |
317 | const void *daddr, const void *saddr, | |
318 | unsigned int len) | |
1da177e4 LT |
319 | { |
320 | struct vlan_hdr *vhdr; | |
321 | unsigned short veth_TCI = 0; | |
322 | int rc = 0; | |
323 | int build_vlan_header = 0; | |
324 | struct net_device *vdev = dev; /* save this for the bottom of the method */ | |
325 | ||
40f98e1a | 326 | pr_debug("%s: skb: %p type: %hx len: %u vlan_id: %hx, daddr: %p\n", |
9dfebcc6 | 327 | __FUNCTION__, skb, type, len, vlan_dev_info(dev)->vlan_id, daddr); |
1da177e4 LT |
328 | |
329 | /* build vlan header only if re_order_header flag is NOT set. This | |
330 | * fixes some programs that get confused when they see a VLAN device | |
331 | * sending a frame that is VLAN encoded (the consensus is that the VLAN | |
332 | * device should look completely like an Ethernet device when the | |
122952fc | 333 | * REORDER_HEADER flag is set) The drawback to this is some extra |
1da177e4 LT |
334 | * header shuffling in the hard_start_xmit. Users can turn off this |
335 | * REORDER behaviour with the vconfig tool. | |
336 | */ | |
9dfebcc6 | 337 | if (!(vlan_dev_info(dev)->flags & VLAN_FLAG_REORDER_HDR)) |
a4bf3af4 | 338 | build_vlan_header = 1; |
1da177e4 LT |
339 | |
340 | if (build_vlan_header) { | |
341 | vhdr = (struct vlan_hdr *) skb_push(skb, VLAN_HLEN); | |
342 | ||
343 | /* build the four bytes that make this a VLAN header. */ | |
344 | ||
345 | /* Now, construct the second two bytes. This field looks something | |
346 | * like: | |
347 | * usr_priority: 3 bits (high bits) | |
348 | * CFI 1 bit | |
349 | * VLAN ID 12 bits (low bits) | |
350 | * | |
351 | */ | |
9dfebcc6 | 352 | veth_TCI = vlan_dev_info(dev)->vlan_id; |
1da177e4 LT |
353 | veth_TCI |= vlan_dev_get_egress_qos_mask(dev, skb); |
354 | ||
355 | vhdr->h_vlan_TCI = htons(veth_TCI); | |
356 | ||
357 | /* | |
358 | * Set the protocol type. | |
359 | * For a packet of type ETH_P_802_3 we put the length in here instead. | |
360 | * It is up to the 802.2 layer to carry protocol information. | |
361 | */ | |
362 | ||
363 | if (type != ETH_P_802_3) { | |
364 | vhdr->h_vlan_encapsulated_proto = htons(type); | |
365 | } else { | |
366 | vhdr->h_vlan_encapsulated_proto = htons(len); | |
367 | } | |
279e172a JB |
368 | |
369 | skb->protocol = htons(ETH_P_8021Q); | |
be8bd863 | 370 | skb_reset_network_header(skb); |
1da177e4 LT |
371 | } |
372 | ||
373 | /* Before delegating work to the lower layer, enter our MAC-address */ | |
374 | if (saddr == NULL) | |
375 | saddr = dev->dev_addr; | |
376 | ||
9dfebcc6 | 377 | dev = vlan_dev_info(dev)->real_dev; |
1da177e4 LT |
378 | |
379 | /* MPLS can send us skbuffs w/out enough space. This check will grow the | |
380 | * skb if it doesn't have enough headroom. Not a beautiful solution, so | |
381 | * I'll tick a counter so that users can know it's happening... If they | |
382 | * care... | |
383 | */ | |
384 | ||
385 | /* NOTE: This may still break if the underlying device is not the final | |
386 | * device (and thus there are more headers to add...) It should work for | |
387 | * good-ole-ethernet though. | |
388 | */ | |
389 | if (skb_headroom(skb) < dev->hard_header_len) { | |
390 | struct sk_buff *sk_tmp = skb; | |
391 | skb = skb_realloc_headroom(sk_tmp, dev->hard_header_len); | |
392 | kfree_skb(sk_tmp); | |
393 | if (skb == NULL) { | |
7bd38d77 | 394 | struct net_device_stats *stats = &vdev->stats; |
1da177e4 LT |
395 | stats->tx_dropped++; |
396 | return -ENOMEM; | |
397 | } | |
9dfebcc6 | 398 | vlan_dev_info(vdev)->cnt_inc_headroom_on_tx++; |
40f98e1a | 399 | pr_debug("%s: %s: had to grow skb.\n", __FUNCTION__, vdev->name); |
1da177e4 LT |
400 | } |
401 | ||
402 | if (build_vlan_header) { | |
403 | /* Now make the underlying real hard header */ | |
0c4e8581 SH |
404 | rc = dev_hard_header(skb, dev, ETH_P_8021Q, daddr, saddr, |
405 | len + VLAN_HLEN); | |
406 | if (rc > 0) | |
1da177e4 | 407 | rc += VLAN_HLEN; |
0c4e8581 | 408 | else if (rc < 0) |
1da177e4 | 409 | rc -= VLAN_HLEN; |
0c4e8581 | 410 | } else |
1da177e4 LT |
411 | /* If here, then we'll just make a normal looking ethernet frame, |
412 | * but, the hard_start_xmit method will insert the tag (it has to | |
413 | * be able to do this for bridged and other skbs that don't come | |
414 | * down the protocol stack in an orderly manner. | |
415 | */ | |
0c4e8581 | 416 | rc = dev_hard_header(skb, dev, type, daddr, saddr, len); |
1da177e4 LT |
417 | |
418 | return rc; | |
419 | } | |
420 | ||
ef3eb3e5 | 421 | static int vlan_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) |
1da177e4 | 422 | { |
7bd38d77 | 423 | struct net_device_stats *stats = &dev->stats; |
1da177e4 LT |
424 | struct vlan_ethhdr *veth = (struct vlan_ethhdr *)(skb->data); |
425 | ||
426 | /* Handle non-VLAN frames if they are sent to us, for example by DHCP. | |
427 | * | |
428 | * NOTE: THIS ASSUMES DIX ETHERNET, SPECIFICALLY NOT SUPPORTING | |
429 | * OTHER THINGS LIKE FDDI/TokenRing/802.3 SNAPs... | |
430 | */ | |
431 | ||
6ab3b487 | 432 | if (veth->h_vlan_proto != htons(ETH_P_8021Q) || |
9dfebcc6 | 433 | vlan_dev_info(dev)->flags & VLAN_FLAG_REORDER_HDR) { |
1da177e4 LT |
434 | int orig_headroom = skb_headroom(skb); |
435 | unsigned short veth_TCI; | |
436 | ||
437 | /* This is not a VLAN frame...but we can fix that! */ | |
9dfebcc6 | 438 | vlan_dev_info(dev)->cnt_encap_on_xmit++; |
1da177e4 | 439 | |
40f98e1a PM |
440 | pr_debug("%s: proto to encap: 0x%hx\n", |
441 | __FUNCTION__, htons(veth->h_vlan_proto)); | |
1da177e4 LT |
442 | /* Construct the second two bytes. This field looks something |
443 | * like: | |
444 | * usr_priority: 3 bits (high bits) | |
445 | * CFI 1 bit | |
446 | * VLAN ID 12 bits (low bits) | |
447 | */ | |
9dfebcc6 | 448 | veth_TCI = vlan_dev_info(dev)->vlan_id; |
1da177e4 LT |
449 | veth_TCI |= vlan_dev_get_egress_qos_mask(dev, skb); |
450 | ||
451 | skb = __vlan_put_tag(skb, veth_TCI); | |
452 | if (!skb) { | |
453 | stats->tx_dropped++; | |
454 | return 0; | |
455 | } | |
456 | ||
457 | if (orig_headroom < VLAN_HLEN) { | |
9dfebcc6 | 458 | vlan_dev_info(dev)->cnt_inc_headroom_on_tx++; |
1da177e4 LT |
459 | } |
460 | } | |
461 | ||
40f98e1a | 462 | pr_debug("%s: about to send skb: %p to dev: %s\n", |
1da177e4 | 463 | __FUNCTION__, skb, skb->dev->name); |
40f98e1a PM |
464 | pr_debug(" " MAC_FMT " " MAC_FMT " %4hx %4hx %4hx\n", |
465 | veth->h_dest[0], veth->h_dest[1], veth->h_dest[2], | |
466 | veth->h_dest[3], veth->h_dest[4], veth->h_dest[5], | |
467 | veth->h_source[0], veth->h_source[1], veth->h_source[2], | |
468 | veth->h_source[3], veth->h_source[4], veth->h_source[5], | |
469 | veth->h_vlan_proto, veth->h_vlan_TCI, | |
470 | veth->h_vlan_encapsulated_proto); | |
1da177e4 LT |
471 | |
472 | stats->tx_packets++; /* for statics only */ | |
473 | stats->tx_bytes += skb->len; | |
474 | ||
9dfebcc6 | 475 | skb->dev = vlan_dev_info(dev)->real_dev; |
1da177e4 LT |
476 | dev_queue_xmit(skb); |
477 | ||
478 | return 0; | |
479 | } | |
480 | ||
ef3eb3e5 PM |
481 | static int vlan_dev_hwaccel_hard_start_xmit(struct sk_buff *skb, |
482 | struct net_device *dev) | |
1da177e4 | 483 | { |
7bd38d77 | 484 | struct net_device_stats *stats = &dev->stats; |
1da177e4 LT |
485 | unsigned short veth_TCI; |
486 | ||
487 | /* Construct the second two bytes. This field looks something | |
488 | * like: | |
489 | * usr_priority: 3 bits (high bits) | |
490 | * CFI 1 bit | |
491 | * VLAN ID 12 bits (low bits) | |
492 | */ | |
9dfebcc6 | 493 | veth_TCI = vlan_dev_info(dev)->vlan_id; |
1da177e4 LT |
494 | veth_TCI |= vlan_dev_get_egress_qos_mask(dev, skb); |
495 | skb = __vlan_hwaccel_put_tag(skb, veth_TCI); | |
496 | ||
497 | stats->tx_packets++; | |
498 | stats->tx_bytes += skb->len; | |
499 | ||
9dfebcc6 | 500 | skb->dev = vlan_dev_info(dev)->real_dev; |
1da177e4 LT |
501 | dev_queue_xmit(skb); |
502 | ||
503 | return 0; | |
504 | } | |
505 | ||
ef3eb3e5 | 506 | static int vlan_dev_change_mtu(struct net_device *dev, int new_mtu) |
1da177e4 LT |
507 | { |
508 | /* TODO: gotta make sure the underlying layer can handle it, | |
509 | * maybe an IFF_VLAN_CAPABLE flag for devices? | |
510 | */ | |
9dfebcc6 | 511 | if (vlan_dev_info(dev)->real_dev->mtu < new_mtu) |
1da177e4 LT |
512 | return -ERANGE; |
513 | ||
514 | dev->mtu = new_mtu; | |
515 | ||
516 | return 0; | |
517 | } | |
518 | ||
c17d8874 PM |
519 | void vlan_dev_set_ingress_priority(const struct net_device *dev, |
520 | u32 skb_prio, short vlan_prio) | |
1da177e4 | 521 | { |
9dfebcc6 | 522 | struct vlan_dev_info *vlan = vlan_dev_info(dev); |
b020cb48 PM |
523 | |
524 | if (vlan->ingress_priority_map[vlan_prio & 0x7] && !skb_prio) | |
525 | vlan->nr_ingress_mappings--; | |
526 | else if (!vlan->ingress_priority_map[vlan_prio & 0x7] && skb_prio) | |
527 | vlan->nr_ingress_mappings++; | |
528 | ||
529 | vlan->ingress_priority_map[vlan_prio & 0x7] = skb_prio; | |
1da177e4 LT |
530 | } |
531 | ||
c17d8874 PM |
532 | int vlan_dev_set_egress_priority(const struct net_device *dev, |
533 | u32 skb_prio, short vlan_prio) | |
1da177e4 | 534 | { |
9dfebcc6 | 535 | struct vlan_dev_info *vlan = vlan_dev_info(dev); |
1da177e4 LT |
536 | struct vlan_priority_tci_mapping *mp = NULL; |
537 | struct vlan_priority_tci_mapping *np; | |
b020cb48 | 538 | u32 vlan_qos = (vlan_prio << 13) & 0xE000; |
122952fc | 539 | |
c17d8874 | 540 | /* See if a priority mapping exists.. */ |
b020cb48 | 541 | mp = vlan->egress_priority_map[skb_prio & 0xF]; |
c17d8874 PM |
542 | while (mp) { |
543 | if (mp->priority == skb_prio) { | |
b020cb48 PM |
544 | if (mp->vlan_qos && !vlan_qos) |
545 | vlan->nr_egress_mappings--; | |
546 | else if (!mp->vlan_qos && vlan_qos) | |
547 | vlan->nr_egress_mappings++; | |
548 | mp->vlan_qos = vlan_qos; | |
c17d8874 | 549 | return 0; |
1da177e4 | 550 | } |
c17d8874 | 551 | mp = mp->next; |
1da177e4 | 552 | } |
c17d8874 PM |
553 | |
554 | /* Create a new mapping then. */ | |
b020cb48 | 555 | mp = vlan->egress_priority_map[skb_prio & 0xF]; |
c17d8874 PM |
556 | np = kmalloc(sizeof(struct vlan_priority_tci_mapping), GFP_KERNEL); |
557 | if (!np) | |
558 | return -ENOBUFS; | |
559 | ||
560 | np->next = mp; | |
561 | np->priority = skb_prio; | |
b020cb48 PM |
562 | np->vlan_qos = vlan_qos; |
563 | vlan->egress_priority_map[skb_prio & 0xF] = np; | |
564 | if (vlan_qos) | |
565 | vlan->nr_egress_mappings++; | |
c17d8874 | 566 | return 0; |
1da177e4 LT |
567 | } |
568 | ||
a4bf3af4 | 569 | /* Flags are defined in the vlan_flags enum in include/linux/if_vlan.h file. */ |
c17d8874 PM |
570 | int vlan_dev_set_vlan_flag(const struct net_device *dev, |
571 | u32 flag, short flag_val) | |
1da177e4 | 572 | { |
c17d8874 | 573 | /* verify flag is supported */ |
a4bf3af4 | 574 | if (flag == VLAN_FLAG_REORDER_HDR) { |
c17d8874 | 575 | if (flag_val) { |
9dfebcc6 | 576 | vlan_dev_info(dev)->flags |= VLAN_FLAG_REORDER_HDR; |
1da177e4 | 577 | } else { |
9dfebcc6 | 578 | vlan_dev_info(dev)->flags &= ~VLAN_FLAG_REORDER_HDR; |
1da177e4 | 579 | } |
c17d8874 | 580 | return 0; |
1da177e4 | 581 | } |
1da177e4 LT |
582 | return -EINVAL; |
583 | } | |
584 | ||
c17d8874 | 585 | void vlan_dev_get_realdev_name(const struct net_device *dev, char *result) |
1da177e4 | 586 | { |
9dfebcc6 | 587 | strncpy(result, vlan_dev_info(dev)->real_dev->name, 23); |
1da177e4 LT |
588 | } |
589 | ||
c17d8874 | 590 | void vlan_dev_get_vid(const struct net_device *dev, unsigned short *result) |
1da177e4 | 591 | { |
9dfebcc6 | 592 | *result = vlan_dev_info(dev)->vlan_id; |
1da177e4 LT |
593 | } |
594 | ||
ef3eb3e5 | 595 | static int vlan_dev_open(struct net_device *dev) |
1da177e4 | 596 | { |
9dfebcc6 | 597 | struct vlan_dev_info *vlan = vlan_dev_info(dev); |
8c979c26 PM |
598 | struct net_device *real_dev = vlan->real_dev; |
599 | int err; | |
600 | ||
601 | if (!(real_dev->flags & IFF_UP)) | |
1da177e4 LT |
602 | return -ENETDOWN; |
603 | ||
8c979c26 PM |
604 | if (compare_ether_addr(dev->dev_addr, real_dev->dev_addr)) { |
605 | err = dev_unicast_add(real_dev, dev->dev_addr, ETH_ALEN); | |
606 | if (err < 0) | |
607 | return err; | |
608 | } | |
609 | memcpy(vlan->real_dev_addr, real_dev->dev_addr, ETH_ALEN); | |
610 | ||
6c78dcbd PM |
611 | if (dev->flags & IFF_ALLMULTI) |
612 | dev_set_allmulti(real_dev, 1); | |
613 | if (dev->flags & IFF_PROMISC) | |
614 | dev_set_promiscuity(real_dev, 1); | |
615 | ||
1da177e4 LT |
616 | return 0; |
617 | } | |
618 | ||
ef3eb3e5 | 619 | static int vlan_dev_stop(struct net_device *dev) |
1da177e4 | 620 | { |
9dfebcc6 | 621 | struct net_device *real_dev = vlan_dev_info(dev)->real_dev; |
8c979c26 | 622 | |
56addd6e | 623 | dev_mc_unsync(real_dev, dev); |
6c78dcbd PM |
624 | if (dev->flags & IFF_ALLMULTI) |
625 | dev_set_allmulti(real_dev, -1); | |
626 | if (dev->flags & IFF_PROMISC) | |
627 | dev_set_promiscuity(real_dev, -1); | |
628 | ||
8c979c26 PM |
629 | if (compare_ether_addr(dev->dev_addr, real_dev->dev_addr)) |
630 | dev_unicast_delete(real_dev, dev->dev_addr, dev->addr_len); | |
631 | ||
1da177e4 LT |
632 | return 0; |
633 | } | |
634 | ||
ef3eb3e5 | 635 | static int vlan_dev_set_mac_address(struct net_device *dev, void *p) |
39aaac11 | 636 | { |
9dfebcc6 | 637 | struct net_device *real_dev = vlan_dev_info(dev)->real_dev; |
39aaac11 PM |
638 | struct sockaddr *addr = p; |
639 | int err; | |
640 | ||
641 | if (!is_valid_ether_addr(addr->sa_data)) | |
642 | return -EADDRNOTAVAIL; | |
643 | ||
644 | if (!(dev->flags & IFF_UP)) | |
645 | goto out; | |
646 | ||
647 | if (compare_ether_addr(addr->sa_data, real_dev->dev_addr)) { | |
648 | err = dev_unicast_add(real_dev, addr->sa_data, ETH_ALEN); | |
649 | if (err < 0) | |
650 | return err; | |
651 | } | |
652 | ||
653 | if (compare_ether_addr(dev->dev_addr, real_dev->dev_addr)) | |
654 | dev_unicast_delete(real_dev, dev->dev_addr, ETH_ALEN); | |
655 | ||
656 | out: | |
657 | memcpy(dev->dev_addr, addr->sa_data, ETH_ALEN); | |
658 | return 0; | |
659 | } | |
660 | ||
ef3eb3e5 | 661 | static int vlan_dev_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) |
1da177e4 | 662 | { |
9dfebcc6 | 663 | struct net_device *real_dev = vlan_dev_info(dev)->real_dev; |
1da177e4 LT |
664 | struct ifreq ifrr; |
665 | int err = -EOPNOTSUPP; | |
666 | ||
667 | strncpy(ifrr.ifr_name, real_dev->name, IFNAMSIZ); | |
668 | ifrr.ifr_ifru = ifr->ifr_ifru; | |
669 | ||
670 | switch(cmd) { | |
671 | case SIOCGMIIPHY: | |
672 | case SIOCGMIIREG: | |
673 | case SIOCSMIIREG: | |
122952fc | 674 | if (real_dev->do_ioctl && netif_device_present(real_dev)) |
1da177e4 LT |
675 | err = real_dev->do_ioctl(real_dev, &ifrr, cmd); |
676 | break; | |
1da177e4 LT |
677 | } |
678 | ||
122952fc | 679 | if (!err) |
1da177e4 LT |
680 | ifr->ifr_ifru = ifrr.ifr_ifru; |
681 | ||
682 | return err; | |
683 | } | |
684 | ||
ef3eb3e5 | 685 | static void vlan_dev_change_rx_flags(struct net_device *dev, int change) |
6c78dcbd | 686 | { |
9dfebcc6 | 687 | struct net_device *real_dev = vlan_dev_info(dev)->real_dev; |
6c78dcbd PM |
688 | |
689 | if (change & IFF_ALLMULTI) | |
690 | dev_set_allmulti(real_dev, dev->flags & IFF_ALLMULTI ? 1 : -1); | |
691 | if (change & IFF_PROMISC) | |
692 | dev_set_promiscuity(real_dev, dev->flags & IFF_PROMISC ? 1 : -1); | |
693 | } | |
694 | ||
ef3eb3e5 | 695 | static void vlan_dev_set_multicast_list(struct net_device *vlan_dev) |
1da177e4 | 696 | { |
9dfebcc6 | 697 | dev_mc_sync(vlan_dev_info(vlan_dev)->real_dev, vlan_dev); |
1da177e4 | 698 | } |
ef3eb3e5 PM |
699 | |
700 | /* | |
701 | * vlan network devices have devices nesting below it, and are a special | |
702 | * "super class" of normal network devices; split their locks off into a | |
703 | * separate class since they always nest. | |
704 | */ | |
705 | static struct lock_class_key vlan_netdev_xmit_lock_key; | |
706 | ||
707 | static const struct header_ops vlan_header_ops = { | |
708 | .create = vlan_dev_hard_header, | |
709 | .rebuild = vlan_dev_rebuild_header, | |
710 | .parse = eth_header_parse, | |
711 | }; | |
712 | ||
713 | static int vlan_dev_init(struct net_device *dev) | |
714 | { | |
9dfebcc6 | 715 | struct net_device *real_dev = vlan_dev_info(dev)->real_dev; |
ef3eb3e5 PM |
716 | int subclass = 0; |
717 | ||
718 | /* IFF_BROADCAST|IFF_MULTICAST; ??? */ | |
719 | dev->flags = real_dev->flags & ~IFF_UP; | |
720 | dev->iflink = real_dev->ifindex; | |
721 | dev->state = (real_dev->state & ((1<<__LINK_STATE_NOCARRIER) | | |
722 | (1<<__LINK_STATE_DORMANT))) | | |
723 | (1<<__LINK_STATE_PRESENT); | |
724 | ||
725 | /* ipv6 shared card related stuff */ | |
726 | dev->dev_id = real_dev->dev_id; | |
727 | ||
728 | if (is_zero_ether_addr(dev->dev_addr)) | |
729 | memcpy(dev->dev_addr, real_dev->dev_addr, dev->addr_len); | |
730 | if (is_zero_ether_addr(dev->broadcast)) | |
731 | memcpy(dev->broadcast, real_dev->broadcast, dev->addr_len); | |
732 | ||
733 | if (real_dev->features & NETIF_F_HW_VLAN_TX) { | |
734 | dev->header_ops = real_dev->header_ops; | |
735 | dev->hard_header_len = real_dev->hard_header_len; | |
736 | dev->hard_start_xmit = vlan_dev_hwaccel_hard_start_xmit; | |
737 | } else { | |
738 | dev->header_ops = &vlan_header_ops; | |
739 | dev->hard_header_len = real_dev->hard_header_len + VLAN_HLEN; | |
740 | dev->hard_start_xmit = vlan_dev_hard_start_xmit; | |
741 | } | |
742 | ||
743 | if (real_dev->priv_flags & IFF_802_1Q_VLAN) | |
744 | subclass = 1; | |
745 | ||
746 | lockdep_set_class_and_subclass(&dev->_xmit_lock, | |
747 | &vlan_netdev_xmit_lock_key, subclass); | |
748 | return 0; | |
749 | } | |
750 | ||
751 | void vlan_setup(struct net_device *dev) | |
752 | { | |
753 | ether_setup(dev); | |
754 | ||
755 | dev->priv_flags |= IFF_802_1Q_VLAN; | |
756 | dev->tx_queue_len = 0; | |
757 | ||
758 | dev->change_mtu = vlan_dev_change_mtu; | |
759 | dev->init = vlan_dev_init; | |
760 | dev->open = vlan_dev_open; | |
761 | dev->stop = vlan_dev_stop; | |
762 | dev->set_mac_address = vlan_dev_set_mac_address; | |
763 | dev->set_multicast_list = vlan_dev_set_multicast_list; | |
764 | dev->change_rx_flags = vlan_dev_change_rx_flags; | |
765 | dev->do_ioctl = vlan_dev_ioctl; | |
766 | dev->destructor = free_netdev; | |
767 | ||
768 | memset(dev->broadcast, 0, ETH_ALEN); | |
769 | } |