]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blame - arch/um/drivers/net_kern.c
Merge tag 'xfs-for-linus-v3.13-rc1-2' of git://oss.sgi.com/xfs/xfs
[mirror_ubuntu-artful-kernel.git] / arch / um / drivers / net_kern.c
CommitLineData
1da177e4 1/*
cd1ae0e4 2 * Copyright (C) 2001 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com)
f28169d2 3 * Copyright (C) 2001 Lennert Buytenhek (buytenh@gnu.org) and
1da177e4
LT
4 * James Leu (jleu@mindspring.net).
5 * Copyright (C) 2001 by various other people who didn't put their name here.
6 * Licensed under the GPL.
7 */
8
cd1ae0e4
JD
9#include <linux/bootmem.h>
10#include <linux/etherdevice.h>
11#include <linux/ethtool.h>
12#include <linux/inetdevice.h>
13#include <linux/init.h>
14#include <linux/list.h>
15#include <linux/netdevice.h>
16#include <linux/platform_device.h>
17#include <linux/rtnetlink.h>
18#include <linux/skbuff.h>
5a0e3ad6 19#include <linux/slab.h>
cd1ae0e4 20#include <linux/spinlock.h>
37185b33
AV
21#include <init.h>
22#include <irq_kern.h>
23#include <irq_user.h>
cd1ae0e4 24#include "mconsole_kern.h"
37185b33
AV
25#include <net_kern.h>
26#include <net_user.h>
1da177e4
LT
27
28#define DRIVER_NAME "uml-netdev"
29
30static DEFINE_SPINLOCK(opened_lock);
9010772c 31static LIST_HEAD(opened);
1da177e4 32
605c1e57
JD
33/*
34 * The drop_skb is used when we can't allocate an skb. The
35 * packet is read into drop_skb in order to get the data off the
36 * connection to the host.
37 * It is reallocated whenever a maximum packet size is seen which is
38 * larger than any seen before. update_drop_skb is called from
39 * eth_configure when a new interface is added.
40 */
41static DEFINE_SPINLOCK(drop_lock);
42static struct sk_buff *drop_skb;
43static int drop_max;
44
45static int update_drop_skb(int max)
46{
47 struct sk_buff *new;
48 unsigned long flags;
49 int err = 0;
50
51 spin_lock_irqsave(&drop_lock, flags);
52
53 if (max <= drop_max)
54 goto out;
55
56 err = -ENOMEM;
57 new = dev_alloc_skb(max);
58 if (new == NULL)
59 goto out;
60
61 skb_put(new, max);
62
63 kfree_skb(drop_skb);
64 drop_skb = new;
65 drop_max = max;
66 err = 0;
67out:
68 spin_unlock_irqrestore(&drop_lock, flags);
69
70 return err;
71}
72
1da177e4
LT
73static int uml_net_rx(struct net_device *dev)
74{
17c324fa 75 struct uml_net_private *lp = netdev_priv(dev);
1da177e4
LT
76 int pkt_len;
77 struct sk_buff *skb;
78
79 /* If we can't allocate memory, try again next round. */
b53f35a8 80 skb = dev_alloc_skb(lp->max_packet);
1da177e4 81 if (skb == NULL) {
605c1e57
JD
82 drop_skb->dev = dev;
83 /* Read a packet into drop_skb and don't do anything with it. */
84 (*lp->read)(lp->fd, drop_skb, lp);
cfa8707a 85 dev->stats.rx_dropped++;
1da177e4
LT
86 return 0;
87 }
88
89 skb->dev = dev;
b53f35a8 90 skb_put(skb, lp->max_packet);
459a98ed 91 skb_reset_mac_header(skb);
b53f35a8 92 pkt_len = (*lp->read)(lp->fd, skb, lp);
1da177e4
LT
93
94 if (pkt_len > 0) {
95 skb_trim(skb, pkt_len);
96 skb->protocol = (*lp->protocol)(skb);
1da177e4 97
cfa8707a
SH
98 dev->stats.rx_bytes += skb->len;
99 dev->stats.rx_packets++;
505a41d4 100 netif_rx(skb);
1da177e4
LT
101 return pkt_len;
102 }
103
104 kfree_skb(skb);
105 return pkt_len;
106}
107
eff3b634 108static void uml_dev_close(struct work_struct *work)
71c8d4c3 109{
eff3b634
PZ
110 struct uml_net_private *lp =
111 container_of(work, struct uml_net_private, work);
112 dev_close(lp->dev);
71c8d4c3
PBG
113}
114
074a0db8 115static irqreturn_t uml_net_interrupt(int irq, void *dev_id)
1da177e4
LT
116{
117 struct net_device *dev = dev_id;
17c324fa 118 struct uml_net_private *lp = netdev_priv(dev);
1da177e4
LT
119 int err;
120
cd1ae0e4
JD
121 if (!netif_running(dev))
122 return IRQ_NONE;
1da177e4
LT
123
124 spin_lock(&lp->lock);
cd1ae0e4
JD
125 while ((err = uml_net_rx(dev)) > 0) ;
126 if (err < 0) {
f28169d2
JD
127 printk(KERN_ERR
128 "Device '%s' read returned %d, shutting it down\n",
1da177e4 129 dev->name, err);
71c8d4c3
PBG
130 /* dev_close can't be called in interrupt context, and takes
131 * again lp->lock.
132 * And dev_close() can be safely called multiple times on the
133 * same device, since it tests for (dev->flags & IFF_UP). So
eff3b634
PZ
134 * there's no harm in delaying the device shutdown.
135 * Furthermore, the workqueue will not re-enqueue an already
136 * enqueued work item. */
137 schedule_work(&lp->work);
1da177e4
LT
138 goto out;
139 }
140 reactivate_fd(lp->fd, UM_ETH_IRQ);
141
71c8d4c3 142out:
1da177e4 143 spin_unlock(&lp->lock);
4ea21cd9 144 return IRQ_HANDLED;
1da177e4
LT
145}
146
147static int uml_net_open(struct net_device *dev)
148{
17c324fa 149 struct uml_net_private *lp = netdev_priv(dev);
1da177e4
LT
150 int err;
151
cd1ae0e4 152 if (lp->fd >= 0) {
1da177e4
LT
153 err = -ENXIO;
154 goto out;
155 }
156
1da177e4 157 lp->fd = (*lp->open)(&lp->user);
cd1ae0e4 158 if (lp->fd < 0) {
1da177e4
LT
159 err = lp->fd;
160 goto out;
161 }
162
163 err = um_request_irq(dev->irq, lp->fd, IRQ_READ, uml_net_interrupt,
c0b79a90 164 IRQF_SHARED, dev->name, dev);
cd1ae0e4 165 if (err != 0) {
1da177e4 166 printk(KERN_ERR "uml_net_open: failed to get irq(%d)\n", err);
1da177e4 167 err = -ENETUNREACH;
14d9ead0 168 goto out_close;
1da177e4
LT
169 }
170
171 lp->tl.data = (unsigned long) &lp->user;
172 netif_start_queue(dev);
173
174 /* clear buffer - it can happen that the host side of the interface
175 * is full when we get here. In this case, new data is never queued,
176 * SIGIOs never arrive, and the net never works.
177 */
cd1ae0e4 178 while ((err = uml_net_rx(dev)) > 0) ;
1da177e4 179
14d9ead0
JD
180 spin_lock(&opened_lock);
181 list_add(&lp->list, &opened);
182 spin_unlock(&opened_lock);
183
184 return 0;
185out_close:
cd1ae0e4 186 if (lp->close != NULL) (*lp->close)(lp->fd, &lp->user);
14d9ead0
JD
187 lp->fd = -1;
188out:
14d9ead0 189 return err;
1da177e4
LT
190}
191
192static int uml_net_close(struct net_device *dev)
193{
17c324fa 194 struct uml_net_private *lp = netdev_priv(dev);
f28169d2 195
1da177e4 196 netif_stop_queue(dev);
1da177e4 197
fa7a0449 198 um_free_irq(dev->irq, dev);
cd1ae0e4 199 if (lp->close != NULL)
1da177e4
LT
200 (*lp->close)(lp->fd, &lp->user);
201 lp->fd = -1;
202
14d9ead0
JD
203 spin_lock(&opened_lock);
204 list_del(&lp->list);
205 spin_unlock(&opened_lock);
206
1da177e4
LT
207 return 0;
208}
209
210static int uml_net_start_xmit(struct sk_buff *skb, struct net_device *dev)
211{
17c324fa 212 struct uml_net_private *lp = netdev_priv(dev);
1da177e4
LT
213 unsigned long flags;
214 int len;
215
216 netif_stop_queue(dev);
217
218 spin_lock_irqsave(&lp->lock, flags);
219
b53f35a8 220 len = (*lp->write)(lp->fd, skb, lp);
55ea1cfa 221 skb_tx_timestamp(skb);
1da177e4 222
cd1ae0e4 223 if (len == skb->len) {
cfa8707a
SH
224 dev->stats.tx_packets++;
225 dev->stats.tx_bytes += skb->len;
1da177e4
LT
226 dev->trans_start = jiffies;
227 netif_start_queue(dev);
228
229 /* this is normally done in the interrupt when tx finishes */
230 netif_wake_queue(dev);
f28169d2 231 }
cd1ae0e4 232 else if (len == 0) {
1da177e4 233 netif_start_queue(dev);
cfa8707a 234 dev->stats.tx_dropped++;
1da177e4
LT
235 }
236 else {
237 netif_start_queue(dev);
238 printk(KERN_ERR "uml_net_start_xmit: failed(%d)\n", len);
239 }
240
241 spin_unlock_irqrestore(&lp->lock, flags);
242
243 dev_kfree_skb(skb);
244
6ed10654 245 return NETDEV_TX_OK;
1da177e4
LT
246}
247
1da177e4
LT
248static void uml_net_set_multicast_list(struct net_device *dev)
249{
f70c1766 250 return;
1da177e4
LT
251}
252
253static void uml_net_tx_timeout(struct net_device *dev)
254{
255 dev->trans_start = jiffies;
256 netif_wake_queue(dev);
257}
258
1da177e4
LT
259static int uml_net_change_mtu(struct net_device *dev, int new_mtu)
260{
1da177e4
LT
261 dev->mtu = new_mtu;
262
b53f35a8 263 return 0;
1da177e4
LT
264}
265
dd71dc4c
RW
266#ifdef CONFIG_NET_POLL_CONTROLLER
267static void uml_net_poll_controller(struct net_device *dev)
268{
269 disable_irq(dev->irq);
270 uml_net_interrupt(dev->irq, dev);
271 enable_irq(dev->irq);
272}
273#endif
274
6d387484
CH
275static void uml_net_get_drvinfo(struct net_device *dev,
276 struct ethtool_drvinfo *info)
1da177e4 277{
7826d43f
JP
278 strlcpy(info->driver, DRIVER_NAME, sizeof(info->driver));
279 strlcpy(info->version, "42", sizeof(info->version));
1da177e4
LT
280}
281
0fc0b732 282static const struct ethtool_ops uml_net_ethtool_ops = {
6d387484
CH
283 .get_drvinfo = uml_net_get_drvinfo,
284 .get_link = ethtool_op_get_link,
55ea1cfa 285 .get_ts_info = ethtool_op_get_ts_info,
6d387484
CH
286};
287
074a0db8 288static void uml_net_user_timer_expire(unsigned long _conn)
1da177e4
LT
289{
290#ifdef undef
291 struct connection *conn = (struct connection *)_conn;
292
293 dprintk(KERN_INFO "uml_net_user_timer_expire [%p]\n", conn);
294 do_connect(conn);
295#endif
296}
297
646cbcda 298static void setup_etheraddr(struct net_device *dev, char *str)
b10aeeef 299{
646cbcda 300 unsigned char *addr = dev->dev_addr;
b10aeeef
JD
301 char *end;
302 int i;
303
cd1ae0e4 304 if (str == NULL)
b10aeeef
JD
305 goto random;
306
2278c5ac 307 for (i = 0; i < 6; i++) {
b10aeeef 308 addr[i] = simple_strtoul(str, &end, 16);
cd1ae0e4
JD
309 if ((end == str) ||
310 ((*end != ':') && (*end != ',') && (*end != '\0'))) {
b10aeeef
JD
311 printk(KERN_ERR
312 "setup_etheraddr: failed to parse '%s' "
313 "as an ethernet address\n", str);
314 goto random;
315 }
316 str = end + 1;
317 }
e024715f 318 if (is_multicast_ether_addr(addr)) {
b10aeeef 319 printk(KERN_ERR
e024715f 320 "Attempt to assign a multicast ethernet address to a "
b10aeeef
JD
321 "device disallowed\n");
322 goto random;
323 }
e024715f
PBG
324 if (!is_valid_ether_addr(addr)) {
325 printk(KERN_ERR
326 "Attempt to assign an invalid ethernet address to a "
327 "device disallowed\n");
328 goto random;
329 }
330 if (!is_local_ether_addr(addr)) {
331 printk(KERN_WARNING
2278c5ac 332 "Warning: Assigning a globally valid ethernet "
7d98230a 333 "address to a device\n");
2278c5ac
JD
334 printk(KERN_WARNING "You should set the 2nd rightmost bit in "
335 "the first byte of the MAC,\n");
7d98230a
JD
336 printk(KERN_WARNING "i.e. %02x:%02x:%02x:%02x:%02x:%02x\n",
337 addr[0] | 0x02, addr[1], addr[2], addr[3], addr[4],
338 addr[5]);
e024715f 339 }
646cbcda 340 return;
b10aeeef
JD
341
342random:
e024715f 343 printk(KERN_INFO
646cbcda
JP
344 "Choosing a random ethernet address for device %s\n", dev->name);
345 eth_hw_addr_random(dev);
b10aeeef
JD
346}
347
1da177e4 348static DEFINE_SPINLOCK(devices_lock);
9010772c 349static LIST_HEAD(devices);
1da177e4 350
3ae5eaec
RK
351static struct platform_driver uml_net_driver = {
352 .driver = {
353 .name = DRIVER_NAME,
354 },
1da177e4 355};
1da177e4 356
2e3f5251
JD
357static void net_device_release(struct device *dev)
358{
8691b97b 359 struct uml_net *device = dev_get_drvdata(dev);
2e3f5251 360 struct net_device *netdev = device->dev;
17c324fa 361 struct uml_net_private *lp = netdev_priv(netdev);
2e3f5251 362
cd1ae0e4 363 if (lp->remove != NULL)
2e3f5251
JD
364 (*lp->remove)(&lp->user);
365 list_del(&device->list);
366 kfree(device);
367 free_netdev(netdev);
368}
369
8bb95b39
SH
370static const struct net_device_ops uml_netdev_ops = {
371 .ndo_open = uml_net_open,
372 .ndo_stop = uml_net_close,
373 .ndo_start_xmit = uml_net_start_xmit,
afc4b13d 374 .ndo_set_rx_mode = uml_net_set_multicast_list,
8bb95b39 375 .ndo_tx_timeout = uml_net_tx_timeout,
9337057d 376 .ndo_set_mac_address = eth_mac_addr,
8bb95b39 377 .ndo_change_mtu = uml_net_change_mtu,
8bb95b39 378 .ndo_validate_addr = eth_validate_addr,
dd71dc4c
RW
379#ifdef CONFIG_NET_POLL_CONTROLLER
380 .ndo_poll_controller = uml_net_poll_controller,
381#endif
8bb95b39
SH
382};
383
80e39311
JD
384/*
385 * Ensures that platform_driver_register is called only once by
386 * eth_configure. Will be set in an initcall.
387 */
388static int driver_registered;
389
f34d9d2d
JD
390static void eth_configure(int n, void *init, char *mac,
391 struct transport *transport)
1da177e4
LT
392{
393 struct uml_net *device;
394 struct net_device *dev;
395 struct uml_net_private *lp;
c74c69b4 396 int err, size;
1da177e4 397
c74c69b4 398 size = transport->private_size + sizeof(struct uml_net_private);
1da177e4 399
0268bd0a 400 device = kzalloc(sizeof(*device), GFP_KERNEL);
1da177e4 401 if (device == NULL) {
8c840835
PBG
402 printk(KERN_ERR "eth_configure failed to allocate struct "
403 "uml_net\n");
f34d9d2d 404 return;
1da177e4
LT
405 }
406
8c840835
PBG
407 dev = alloc_etherdev(size);
408 if (dev == NULL) {
409 printk(KERN_ERR "eth_configure: failed to allocate struct "
410 "net_device for eth%d\n", n);
411 goto out_free_device;
412 }
413
1da177e4
LT
414 INIT_LIST_HEAD(&device->list);
415 device->index = n;
416
e024715f
PBG
417 /* If this name ends up conflicting with an existing registered
418 * netdevice, that is OK, register_netdev{,ice}() will notice this
419 * and fail.
420 */
8c840835 421 snprintf(dev->name, sizeof(dev->name), "eth%d", n);
e024715f 422
646cbcda 423 setup_etheraddr(dev, mac);
1da177e4 424
646cbcda 425 printk(KERN_INFO "Netdevice %d (%pM) : ", n, dev->dev_addr);
1da177e4 426
17c324fa 427 lp = netdev_priv(dev);
e56a7885
PBG
428 /* This points to the transport private data. It's still clear, but we
429 * must memset it to 0 *now*. Let's help the drivers. */
430 memset(lp, 0, size);
eff3b634 431 INIT_WORK(&lp->work, uml_dev_close);
e56a7885 432
1da177e4
LT
433 /* sysfs register */
434 if (!driver_registered) {
3ae5eaec 435 platform_driver_register(&uml_net_driver);
1da177e4
LT
436 driver_registered = 1;
437 }
438 device->pdev.id = n;
439 device->pdev.name = DRIVER_NAME;
2e3f5251 440 device->pdev.dev.release = net_device_release;
8691b97b 441 dev_set_drvdata(&device->pdev.dev, device);
cd1ae0e4 442 if (platform_device_register(&device->pdev))
f34d9d2d 443 goto out_free_netdev;
1da177e4
LT
444 SET_NETDEV_DEV(dev,&device->pdev.dev);
445
1da177e4
LT
446 device->dev = dev;
447
f34d9d2d
JD
448 /*
449 * These just fill in a data structure, so there's no failure
450 * to be worried about.
451 */
1da177e4
LT
452 (*transport->kern->init)(dev, init);
453
1da177e4
LT
454 *lp = ((struct uml_net_private)
455 { .list = LIST_HEAD_INIT(lp->list),
456 .dev = dev,
457 .fd = -1,
458 .mac = { 0xfe, 0xfd, 0x0, 0x0, 0x0, 0x0},
b53f35a8 459 .max_packet = transport->user->max_packet,
1da177e4
LT
460 .protocol = transport->kern->protocol,
461 .open = transport->user->open,
462 .close = transport->user->close,
463 .remove = transport->user->remove,
464 .read = transport->kern->read,
465 .write = transport->kern->write,
466 .add_address = transport->user->add_address,
b53f35a8 467 .delete_address = transport->user->delete_address });
1da177e4
LT
468
469 init_timer(&lp->tl);
470 spin_lock_init(&lp->lock);
471 lp->tl.function = uml_net_user_timer_expire;
646cbcda 472 memcpy(lp->mac, dev->dev_addr, sizeof(lp->mac));
1da177e4 473
f34d9d2d
JD
474 if ((transport->user->init != NULL) &&
475 ((*transport->user->init)(&lp->user, dev) != 0))
476 goto out_unregister;
1da177e4 477
b53f35a8 478 dev->mtu = transport->user->mtu;
8bb95b39 479 dev->netdev_ops = &uml_netdev_ops;
f34d9d2d
JD
480 dev->ethtool_ops = &uml_net_ethtool_ops;
481 dev->watchdog_timeo = (HZ >> 1);
482 dev->irq = UM_ETH_IRQ;
1da177e4 483
605c1e57
JD
484 err = update_drop_skb(lp->max_packet);
485 if (err)
486 goto out_undo_user_init;
487
f34d9d2d
JD
488 rtnl_lock();
489 err = register_netdevice(dev);
490 rtnl_unlock();
491 if (err)
492 goto out_undo_user_init;
493
494 spin_lock(&devices_lock);
495 list_add(&device->list, &devices);
496 spin_unlock(&devices_lock);
497
498 return;
499
500out_undo_user_init:
8c840835 501 if (transport->user->remove != NULL)
f34d9d2d
JD
502 (*transport->user->remove)(&lp->user);
503out_unregister:
504 platform_device_unregister(&device->pdev);
7d98230a 505 return; /* platform_device_unregister frees dev and device */
f34d9d2d
JD
506out_free_netdev:
507 free_netdev(dev);
8c840835 508out_free_device:
f34d9d2d 509 kfree(device);
1da177e4
LT
510}
511
512static struct uml_net *find_device(int n)
513{
514 struct uml_net *device;
515 struct list_head *ele;
516
517 spin_lock(&devices_lock);
cd1ae0e4 518 list_for_each(ele, &devices) {
1da177e4 519 device = list_entry(ele, struct uml_net, list);
cd1ae0e4 520 if (device->index == n)
1da177e4
LT
521 goto out;
522 }
523 device = NULL;
524 out:
525 spin_unlock(&devices_lock);
4ea21cd9 526 return device;
1da177e4
LT
527}
528
f28169d2
JD
529static int eth_parse(char *str, int *index_out, char **str_out,
530 char **error_out)
1da177e4
LT
531{
532 char *end;
a419aef8 533 int n, err = -EINVAL;
1da177e4
LT
534
535 n = simple_strtoul(str, &end, 0);
cd1ae0e4 536 if (end == str) {
f28169d2
JD
537 *error_out = "Bad device number";
538 return err;
1da177e4 539 }
f28169d2 540
1da177e4 541 str = end;
cd1ae0e4 542 if (*str != '=') {
f28169d2
JD
543 *error_out = "Expected '=' after device number";
544 return err;
1da177e4 545 }
f28169d2 546
1da177e4 547 str++;
cd1ae0e4 548 if (find_device(n)) {
f28169d2
JD
549 *error_out = "Device already configured";
550 return err;
1da177e4 551 }
f28169d2
JD
552
553 *index_out = n;
1da177e4 554 *str_out = str;
f28169d2 555 return 0;
1da177e4
LT
556}
557
558struct eth_init {
559 struct list_head list;
560 char *init;
561 int index;
562};
563
d3b7f69d
JD
564static DEFINE_SPINLOCK(transports_lock);
565static LIST_HEAD(transports);
1da177e4
LT
566
567/* Filled in during early boot */
c862fc32 568static LIST_HEAD(eth_cmd_line);
1da177e4
LT
569
570static int check_transport(struct transport *transport, char *eth, int n,
571 void **init_out, char **mac_out)
572{
573 int len;
574
575 len = strlen(transport->name);
cd1ae0e4 576 if (strncmp(eth, transport->name, len))
4ea21cd9 577 return 0;
1da177e4
LT
578
579 eth += len;
cd1ae0e4 580 if (*eth == ',')
1da177e4 581 eth++;
cd1ae0e4 582 else if (*eth != '\0')
4ea21cd9 583 return 0;
1da177e4
LT
584
585 *init_out = kmalloc(transport->setup_size, GFP_KERNEL);
cd1ae0e4 586 if (*init_out == NULL)
4ea21cd9 587 return 1;
1da177e4 588
cd1ae0e4 589 if (!transport->setup(eth, mac_out, *init_out)) {
1da177e4
LT
590 kfree(*init_out);
591 *init_out = NULL;
592 }
4ea21cd9 593 return 1;
1da177e4
LT
594}
595
596void register_transport(struct transport *new)
597{
598 struct list_head *ele, *next;
599 struct eth_init *eth;
600 void *init;
601 char *mac = NULL;
602 int match;
603
d3b7f69d
JD
604 spin_lock(&transports_lock);
605 BUG_ON(!list_empty(&new->list));
1da177e4 606 list_add(&new->list, &transports);
d3b7f69d 607 spin_unlock(&transports_lock);
1da177e4 608
cd1ae0e4 609 list_for_each_safe(ele, next, &eth_cmd_line) {
1da177e4
LT
610 eth = list_entry(ele, struct eth_init, list);
611 match = check_transport(new, eth->init, eth->index, &init,
612 &mac);
cd1ae0e4 613 if (!match)
1da177e4 614 continue;
cd1ae0e4 615 else if (init != NULL) {
1da177e4
LT
616 eth_configure(eth->index, init, mac, new);
617 kfree(init);
618 }
619 list_del(&eth->list);
620 }
621}
622
623static int eth_setup_common(char *str, int index)
624{
625 struct list_head *ele;
626 struct transport *transport;
627 void *init;
628 char *mac = NULL;
c862fc32 629 int found = 0;
1da177e4 630
c862fc32 631 spin_lock(&transports_lock);
cd1ae0e4 632 list_for_each(ele, &transports) {
1da177e4 633 transport = list_entry(ele, struct transport, list);
cd1ae0e4 634 if (!check_transport(transport, str, index, &init, &mac))
1da177e4 635 continue;
cd1ae0e4 636 if (init != NULL) {
1da177e4
LT
637 eth_configure(index, init, mac, transport);
638 kfree(init);
639 }
c862fc32
JD
640 found = 1;
641 break;
1da177e4 642 }
c862fc32
JD
643
644 spin_unlock(&transports_lock);
645 return found;
1da177e4
LT
646}
647
97a1fcbb 648static int __init eth_setup(char *str)
1da177e4
LT
649{
650 struct eth_init *new;
f28169d2 651 char *error;
1da177e4
LT
652 int n, err;
653
f28169d2 654 err = eth_parse(str, &n, &str, &error);
cd1ae0e4 655 if (err) {
f28169d2
JD
656 printk(KERN_ERR "eth_setup - Couldn't parse '%s' : %s\n",
657 str, error);
1183dc94 658 return 1;
f28169d2 659 }
1da177e4 660
1183dc94 661 new = alloc_bootmem(sizeof(*new));
cd1ae0e4
JD
662 if (new == NULL) {
663 printk(KERN_ERR "eth_init : alloc_bootmem failed\n");
1183dc94 664 return 1;
1da177e4
LT
665 }
666
667 INIT_LIST_HEAD(&new->list);
668 new->index = n;
669 new->init = str;
670
671 list_add_tail(&new->list, &eth_cmd_line);
1183dc94 672 return 1;
1da177e4
LT
673}
674
675__setup("eth", eth_setup);
676__uml_help(eth_setup,
677"eth[0-9]+=<transport>,<options>\n"
678" Configure a network device.\n\n"
679);
680
f28169d2 681static int net_config(char *str, char **error_out)
1da177e4
LT
682{
683 int n, err;
684
f28169d2 685 err = eth_parse(str, &n, &str, error_out);
cd1ae0e4 686 if (err)
f28169d2 687 return err;
1da177e4 688
f28169d2
JD
689 /* This string is broken up and the pieces used by the underlying
690 * driver. So, it is freed only if eth_setup_common fails.
691 */
970d6e3a 692 str = kstrdup(str, GFP_KERNEL);
cd1ae0e4 693 if (str == NULL) {
f28169d2
JD
694 *error_out = "net_config failed to strdup string";
695 return -ENOMEM;
1da177e4
LT
696 }
697 err = !eth_setup_common(str, n);
cd1ae0e4 698 if (err)
1da177e4 699 kfree(str);
cd1ae0e4 700 return err;
1da177e4
LT
701}
702
29d56cfe
JD
703static int net_id(char **str, int *start_out, int *end_out)
704{
cd1ae0e4
JD
705 char *end;
706 int n;
29d56cfe
JD
707
708 n = simple_strtoul(*str, &end, 0);
cd1ae0e4 709 if ((*end != '\0') || (end == *str))
29d56cfe
JD
710 return -1;
711
cd1ae0e4
JD
712 *start_out = n;
713 *end_out = n;
714 *str = end;
715 return n;
29d56cfe
JD
716}
717
f28169d2 718static int net_remove(int n, char **error_out)
1da177e4
LT
719{
720 struct uml_net *device;
721 struct net_device *dev;
722 struct uml_net_private *lp;
1da177e4
LT
723
724 device = find_device(n);
cd1ae0e4 725 if (device == NULL)
29d56cfe 726 return -ENODEV;
1da177e4
LT
727
728 dev = device->dev;
17c324fa 729 lp = netdev_priv(dev);
cd1ae0e4 730 if (lp->fd > 0)
4ea21cd9 731 return -EBUSY;
1da177e4
LT
732 unregister_netdev(dev);
733 platform_device_unregister(&device->pdev);
734
29d56cfe 735 return 0;
1da177e4
LT
736}
737
738static struct mc_device net_mc = {
84f48d4f 739 .list = LIST_HEAD_INIT(net_mc.list),
1da177e4
LT
740 .name = "eth",
741 .config = net_config,
742 .get_config = NULL,
4ea21cd9 743 .id = net_id,
1da177e4
LT
744 .remove = net_remove,
745};
746
32f862c3 747#ifdef CONFIG_INET
1da177e4
LT
748static int uml_inetaddr_event(struct notifier_block *this, unsigned long event,
749 void *ptr)
750{
751 struct in_ifaddr *ifa = ptr;
1da177e4
LT
752 struct net_device *dev = ifa->ifa_dev->dev;
753 struct uml_net_private *lp;
754 void (*proc)(unsigned char *, unsigned char *, void *);
755 unsigned char addr_buf[4], netmask_buf[4];
756
ebe28bb4 757 if (dev->netdev_ops->ndo_open != uml_net_open)
4ea21cd9 758 return NOTIFY_DONE;
1da177e4 759
17c324fa 760 lp = netdev_priv(dev);
1da177e4
LT
761
762 proc = NULL;
cd1ae0e4 763 switch (event) {
1da177e4
LT
764 case NETDEV_UP:
765 proc = lp->add_address;
766 break;
767 case NETDEV_DOWN:
768 proc = lp->delete_address;
769 break;
770 }
cd1ae0e4 771 if (proc != NULL) {
0e76422c
BS
772 memcpy(addr_buf, &ifa->ifa_address, sizeof(addr_buf));
773 memcpy(netmask_buf, &ifa->ifa_mask, sizeof(netmask_buf));
1da177e4
LT
774 (*proc)(addr_buf, netmask_buf, &lp->user);
775 }
4ea21cd9 776 return NOTIFY_DONE;
1da177e4
LT
777}
778
c862fc32 779/* uml_net_init shouldn't be called twice on two CPUs at the same time */
074a0db8 780static struct notifier_block uml_inetaddr_notifier = {
1da177e4
LT
781 .notifier_call = uml_inetaddr_event,
782};
783
32f862c3 784static void inet_register(void)
1da177e4
LT
785{
786 struct list_head *ele;
f28169d2 787 struct uml_net_private *lp;
1da177e4
LT
788 struct in_device *ip;
789 struct in_ifaddr *in;
790
1da177e4
LT
791 register_inetaddr_notifier(&uml_inetaddr_notifier);
792
793 /* Devices may have been opened already, so the uml_inetaddr_notifier
794 * didn't get a chance to run for them. This fakes it so that
795 * addresses which have already been set up get handled properly.
796 */
c862fc32 797 spin_lock(&opened_lock);
cd1ae0e4 798 list_for_each(ele, &opened) {
1da177e4
LT
799 lp = list_entry(ele, struct uml_net_private, list);
800 ip = lp->dev->ip_ptr;
cd1ae0e4 801 if (ip == NULL)
c862fc32 802 continue;
1da177e4 803 in = ip->ifa_list;
cd1ae0e4 804 while (in != NULL) {
1da177e4
LT
805 uml_inetaddr_event(NULL, NETDEV_UP, in);
806 in = in->ifa_next;
807 }
f28169d2 808 }
c862fc32 809 spin_unlock(&opened_lock);
32f862c3
JD
810}
811#else
812static inline void inet_register(void)
813{
814}
815#endif
1da177e4 816
32f862c3
JD
817static int uml_net_init(void)
818{
819 mconsole_register_dev(&net_mc);
820 inet_register();
c862fc32 821 return 0;
1da177e4
LT
822}
823
824__initcall(uml_net_init);
825
826static void close_devices(void)
827{
828 struct list_head *ele;
829 struct uml_net_private *lp;
830
c862fc32 831 spin_lock(&opened_lock);
cd1ae0e4 832 list_for_each(ele, &opened) {
1da177e4 833 lp = list_entry(ele, struct uml_net_private, list);
fa7a0449 834 um_free_irq(lp->dev->irq, lp->dev);
cd1ae0e4 835 if ((lp->close != NULL) && (lp->fd >= 0))
1da177e4 836 (*lp->close)(lp->fd, &lp->user);
cd1ae0e4 837 if (lp->remove != NULL)
c862fc32 838 (*lp->remove)(&lp->user);
1da177e4 839 }
c862fc32 840 spin_unlock(&opened_lock);
1da177e4
LT
841}
842
843__uml_exitcall(close_devices);
844
f28169d2
JD
845void iter_addresses(void *d, void (*cb)(unsigned char *, unsigned char *,
846 void *),
1da177e4
LT
847 void *arg)
848{
849 struct net_device *dev = d;
850 struct in_device *ip = dev->ip_ptr;
851 struct in_ifaddr *in;
852 unsigned char address[4], netmask[4];
853
cd1ae0e4 854 if (ip == NULL) return;
1da177e4 855 in = ip->ifa_list;
cd1ae0e4 856 while (in != NULL) {
0e76422c
BS
857 memcpy(address, &in->ifa_address, sizeof(address));
858 memcpy(netmask, &in->ifa_mask, sizeof(netmask));
1da177e4
LT
859 (*cb)(address, netmask, arg);
860 in = in->ifa_next;
861 }
862}
863
864int dev_netmask(void *d, void *m)
865{
866 struct net_device *dev = d;
867 struct in_device *ip = dev->ip_ptr;
868 struct in_ifaddr *in;
a144ea4b 869 __be32 *mask_out = m;
1da177e4 870
cd1ae0e4
JD
871 if (ip == NULL)
872 return 1;
1da177e4
LT
873
874 in = ip->ifa_list;
cd1ae0e4
JD
875 if (in == NULL)
876 return 1;
1da177e4
LT
877
878 *mask_out = in->ifa_mask;
cd1ae0e4 879 return 0;
1da177e4
LT
880}
881
882void *get_output_buffer(int *len_out)
883{
884 void *ret;
885
886 ret = (void *) __get_free_pages(GFP_KERNEL, 0);
cd1ae0e4 887 if (ret) *len_out = PAGE_SIZE;
1da177e4 888 else *len_out = 0;
4ea21cd9 889 return ret;
1da177e4
LT
890}
891
892void free_output_buffer(void *buffer)
893{
894 free_pages((unsigned long) buffer, 0);
895}
896
f28169d2 897int tap_setup_common(char *str, char *type, char **dev_name, char **mac_out,
1da177e4
LT
898 char **gate_addr)
899{
900 char *remain;
901
902 remain = split_if_spec(str, dev_name, mac_out, gate_addr, NULL);
cd1ae0e4
JD
903 if (remain != NULL) {
904 printk(KERN_ERR "tap_setup_common - Extra garbage on "
905 "specification : '%s'\n", remain);
906 return 1;
1da177e4
LT
907 }
908
cd1ae0e4 909 return 0;
1da177e4
LT
910}
911
912unsigned short eth_protocol(struct sk_buff *skb)
913{
cd1ae0e4 914 return eth_type_trans(skb, skb->dev);
1da177e4 915}