]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blame - drivers/net/ifb.c
net: lan78xx: fix rx handling before first packet is send
[mirror_ubuntu-bionic-kernel.git] / drivers / net / ifb.c
CommitLineData
6aa20a22 1/* drivers/net/ifb.c:
253af423
JHS
2
3 The purpose of this driver is to provide a device that allows
4 for sharing of resources:
5
6 1) qdiscs/policies that are per device as opposed to system wide.
7 ifb allows for a device which can be redirected to thus providing
8 an impression of sharing.
9
10 2) Allows for queueing incoming traffic for shaping instead of
6aa20a22
JG
11 dropping.
12
253af423
JHS
13 The original concept is based on what is known as the IMQ
14 driver initially written by Martin Devera, later rewritten
15 by Patrick McHardy and then maintained by Andre Correa.
16
17 You need the tc action mirror or redirect to feed this device
18 packets.
19
20 This program is free software; you can redistribute it and/or
21 modify it under the terms of the GNU General Public License
22 as published by the Free Software Foundation; either version
23 2 of the License, or (at your option) any later version.
6aa20a22 24
253af423 25 Authors: Jamal Hadi Salim (2005)
6aa20a22 26
253af423
JHS
27*/
28
29
253af423
JHS
30#include <linux/module.h>
31#include <linux/kernel.h>
32#include <linux/netdevice.h>
33#include <linux/etherdevice.h>
34#include <linux/init.h>
a6b7a407 35#include <linux/interrupt.h>
253af423 36#include <linux/moduleparam.h>
6aa20a22 37#include <net/pkt_sched.h>
881d966b 38#include <net/net_namespace.h>
253af423 39
253af423 40#define TX_Q_LIMIT 32
9e29e21a
ED
41struct ifb_q_private {
42 struct net_device *dev;
253af423 43 struct tasklet_struct ifb_tasklet;
9e29e21a
ED
44 int tasklet_pending;
45 int txqnum;
253af423 46 struct sk_buff_head rq;
9e29e21a
ED
47 u64 rx_packets;
48 u64 rx_bytes;
49 struct u64_stats_sync rsync;
3b0c9cbb 50
51 struct u64_stats_sync tsync;
9e29e21a
ED
52 u64 tx_packets;
53 u64 tx_bytes;
253af423 54 struct sk_buff_head tq;
9e29e21a 55} ____cacheline_aligned_in_smp;
253af423 56
9e29e21a
ED
57struct ifb_dev_private {
58 struct ifb_q_private *tx_private;
59};
253af423 60
424efe9c 61static netdev_tx_t ifb_xmit(struct sk_buff *skb, struct net_device *dev);
253af423
JHS
62static int ifb_open(struct net_device *dev);
63static int ifb_close(struct net_device *dev);
64
9e29e21a 65static void ifb_ri_tasklet(unsigned long _txp)
253af423 66{
9e29e21a 67 struct ifb_q_private *txp = (struct ifb_q_private *)_txp;
c3f26a26 68 struct netdev_queue *txq;
253af423
JHS
69 struct sk_buff *skb;
70
9e29e21a
ED
71 txq = netdev_get_tx_queue(txp->dev, txp->txqnum);
72 skb = skb_peek(&txp->tq);
73 if (!skb) {
74 if (!__netif_tx_trylock(txq))
253af423 75 goto resched;
9e29e21a
ED
76 skb_queue_splice_tail_init(&txp->rq, &txp->tq);
77 __netif_tx_unlock(txq);
253af423
JHS
78 }
79
9e29e21a 80 while ((skb = __skb_dequeue(&txp->tq)) != NULL) {
bc31c905 81 skb->tc_redirected = 0;
e7246e12 82 skb->tc_skip_classify = 1;
3b0c9cbb 83
9e29e21a
ED
84 u64_stats_update_begin(&txp->tsync);
85 txp->tx_packets++;
86 txp->tx_bytes += skb->len;
87 u64_stats_update_end(&txp->tsync);
c01003c2 88
05e8689c 89 rcu_read_lock();
9e29e21a 90 skb->dev = dev_get_by_index_rcu(dev_net(txp->dev), skb->skb_iif);
c01003c2 91 if (!skb->dev) {
05e8689c 92 rcu_read_unlock();
c01003c2 93 dev_kfree_skb(skb);
9e29e21a
ED
94 txp->dev->stats.tx_dropped++;
95 if (skb_queue_len(&txp->tq) != 0)
75c1c825 96 goto resched;
c01003c2
PM
97 break;
98 }
05e8689c 99 rcu_read_unlock();
9e29e21a 100 skb->skb_iif = txp->dev->ifindex;
c01003c2 101
bc31c905 102 if (!skb->tc_from_ingress) {
253af423 103 dev_queue_xmit(skb);
bc31c905 104 } else {
f40ae913 105 skb_pull(skb, skb->mac_len);
1a75972c 106 netif_receive_skb(skb);
bc31c905 107 }
253af423
JHS
108 }
109
c3f26a26 110 if (__netif_tx_trylock(txq)) {
9e29e21a
ED
111 skb = skb_peek(&txp->rq);
112 if (!skb) {
113 txp->tasklet_pending = 0;
114 if (netif_tx_queue_stopped(txq))
115 netif_tx_wake_queue(txq);
253af423 116 } else {
c3f26a26 117 __netif_tx_unlock(txq);
253af423
JHS
118 goto resched;
119 }
c3f26a26 120 __netif_tx_unlock(txq);
253af423
JHS
121 } else {
122resched:
9e29e21a
ED
123 txp->tasklet_pending = 1;
124 tasklet_schedule(&txp->ifb_tasklet);
253af423
JHS
125 }
126
127}
128
bc1f4470 129static void ifb_stats64(struct net_device *dev,
130 struct rtnl_link_stats64 *stats)
3b0c9cbb 131{
9e29e21a
ED
132 struct ifb_dev_private *dp = netdev_priv(dev);
133 struct ifb_q_private *txp = dp->tx_private;
3b0c9cbb 134 unsigned int start;
9e29e21a
ED
135 u64 packets, bytes;
136 int i;
137
138 for (i = 0; i < dev->num_tx_queues; i++,txp++) {
139 do {
140 start = u64_stats_fetch_begin_irq(&txp->rsync);
141 packets = txp->rx_packets;
142 bytes = txp->rx_bytes;
143 } while (u64_stats_fetch_retry_irq(&txp->rsync, start));
144 stats->rx_packets += packets;
145 stats->rx_bytes += bytes;
146
147 do {
148 start = u64_stats_fetch_begin_irq(&txp->tsync);
149 packets = txp->tx_packets;
150 bytes = txp->tx_bytes;
151 } while (u64_stats_fetch_retry_irq(&txp->tsync, start));
152 stats->tx_packets += packets;
153 stats->tx_bytes += bytes;
154 }
3b0c9cbb 155 stats->rx_dropped = dev->stats.rx_dropped;
156 stats->tx_dropped = dev->stats.tx_dropped;
3b0c9cbb 157}
158
9e29e21a
ED
159static int ifb_dev_init(struct net_device *dev)
160{
161 struct ifb_dev_private *dp = netdev_priv(dev);
162 struct ifb_q_private *txp;
163 int i;
164
165 txp = kcalloc(dev->num_tx_queues, sizeof(*txp), GFP_KERNEL);
166 if (!txp)
167 return -ENOMEM;
168 dp->tx_private = txp;
169 for (i = 0; i < dev->num_tx_queues; i++,txp++) {
170 txp->txqnum = i;
171 txp->dev = dev;
172 __skb_queue_head_init(&txp->rq);
173 __skb_queue_head_init(&txp->tq);
174 u64_stats_init(&txp->rsync);
175 u64_stats_init(&txp->tsync);
176 tasklet_init(&txp->ifb_tasklet, ifb_ri_tasklet,
177 (unsigned long)txp);
178 netif_tx_start_queue(netdev_get_tx_queue(dev, i));
179 }
180 return 0;
181}
3b0c9cbb 182
8dfcdf34 183static const struct net_device_ops ifb_netdev_ops = {
8dfcdf34
SH
184 .ndo_open = ifb_open,
185 .ndo_stop = ifb_close,
3b0c9cbb 186 .ndo_get_stats64 = ifb_stats64,
00829823
SH
187 .ndo_start_xmit = ifb_xmit,
188 .ndo_validate_addr = eth_validate_addr,
9e29e21a 189 .ndo_init = ifb_dev_init,
8dfcdf34
SH
190};
191
34324dc2 192#define IFB_FEATURES (NETIF_F_HW_CSUM | NETIF_F_SG | NETIF_F_FRAGLIST | \
39980292 193 NETIF_F_TSO_ECN | NETIF_F_TSO | NETIF_F_TSO6 | \
7d945796 194 NETIF_F_GSO_ENCAP_ALL | \
28d2b136
PM
195 NETIF_F_HIGHDMA | NETIF_F_HW_VLAN_CTAG_TX | \
196 NETIF_F_HW_VLAN_STAG_TX)
39980292 197
9e29e21a
ED
198static void ifb_dev_free(struct net_device *dev)
199{
200 struct ifb_dev_private *dp = netdev_priv(dev);
201 struct ifb_q_private *txp = dp->tx_private;
202 int i;
203
204 for (i = 0; i < dev->num_tx_queues; i++,txp++) {
205 tasklet_kill(&txp->ifb_tasklet);
206 __skb_queue_purge(&txp->rq);
207 __skb_queue_purge(&txp->tq);
208 }
209 kfree(dp->tx_private);
9e29e21a
ED
210}
211
9ba2cd65 212static void ifb_setup(struct net_device *dev)
253af423
JHS
213{
214 /* Initialize the device structure. */
8dfcdf34 215 dev->netdev_ops = &ifb_netdev_ops;
253af423
JHS
216
217 /* Fill in device structure with ethernet-generic values. */
218 ether_setup(dev);
219 dev->tx_queue_len = TX_Q_LIMIT;
8dfcdf34 220
39980292 221 dev->features |= IFB_FEATURES;
7d945796
ED
222 dev->hw_features |= dev->features;
223 dev->hw_enc_features |= dev->features;
8dd6e147
VY
224 dev->vlan_features |= IFB_FEATURES & ~(NETIF_F_HW_VLAN_CTAG_TX |
225 NETIF_F_HW_VLAN_STAG_TX);
39980292 226
253af423
JHS
227 dev->flags |= IFF_NOARP;
228 dev->flags &= ~IFF_MULTICAST;
02875878
ED
229 dev->priv_flags &= ~IFF_TX_SKB_SHARING;
230 netif_keep_dst(dev);
f2cedb63 231 eth_hw_addr_random(dev);
cf124db5
DM
232 dev->needs_free_netdev = true;
233 dev->priv_destructor = ifb_dev_free;
e94cd811
ZS
234
235 dev->min_mtu = 0;
236 dev->max_mtu = 0;
253af423
JHS
237}
238
424efe9c 239static netdev_tx_t ifb_xmit(struct sk_buff *skb, struct net_device *dev)
253af423 240{
9e29e21a 241 struct ifb_dev_private *dp = netdev_priv(dev);
9e29e21a 242 struct ifb_q_private *txp = dp->tx_private + skb_get_queue_mapping(skb);
253af423 243
9e29e21a
ED
244 u64_stats_update_begin(&txp->rsync);
245 txp->rx_packets++;
246 txp->rx_bytes += skb->len;
247 u64_stats_update_end(&txp->rsync);
253af423 248
bc31c905 249 if (!skb->tc_redirected || !skb->skb_iif) {
253af423 250 dev_kfree_skb(skb);
3b0c9cbb 251 dev->stats.rx_dropped++;
424efe9c 252 return NETDEV_TX_OK;
253af423
JHS
253 }
254
9e29e21a
ED
255 if (skb_queue_len(&txp->rq) >= dev->tx_queue_len)
256 netif_tx_stop_queue(netdev_get_tx_queue(dev, txp->txqnum));
253af423 257
9e29e21a
ED
258 __skb_queue_tail(&txp->rq, skb);
259 if (!txp->tasklet_pending) {
260 txp->tasklet_pending = 1;
261 tasklet_schedule(&txp->ifb_tasklet);
253af423
JHS
262 }
263
424efe9c 264 return NETDEV_TX_OK;
253af423
JHS
265}
266
253af423
JHS
267static int ifb_close(struct net_device *dev)
268{
9e29e21a 269 netif_tx_stop_all_queues(dev);
253af423
JHS
270 return 0;
271}
272
273static int ifb_open(struct net_device *dev)
274{
9e29e21a 275 netif_tx_start_all_queues(dev);
253af423
JHS
276 return 0;
277}
278
a8b8a889
MS
279static int ifb_validate(struct nlattr *tb[], struct nlattr *data[],
280 struct netlink_ext_ack *extack)
0e06877c
PM
281{
282 if (tb[IFLA_ADDRESS]) {
283 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
284 return -EINVAL;
285 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
286 return -EADDRNOTAVAIL;
287 }
288 return 0;
289}
290
9ba2cd65
PM
291static struct rtnl_link_ops ifb_link_ops __read_mostly = {
292 .kind = "ifb",
9e29e21a 293 .priv_size = sizeof(struct ifb_dev_private),
9ba2cd65 294 .setup = ifb_setup,
0e06877c 295 .validate = ifb_validate,
9ba2cd65
PM
296};
297
9e29e21a
ED
298/* Number of ifb devices to be set up by this module.
299 * Note that these legacy devices have one queue.
300 * Prefer something like : ip link add ifb10 numtxqueues 8 type ifb
301 */
302static int numifbs = 2;
2d85cba2
PM
303module_param(numifbs, int, 0);
304MODULE_PARM_DESC(numifbs, "Number of ifb devices");
305
253af423
JHS
306static int __init ifb_init_one(int index)
307{
308 struct net_device *dev_ifb;
309 int err;
310
9e29e21a 311 dev_ifb = alloc_netdev(sizeof(struct ifb_dev_private), "ifb%d",
c835a677 312 NET_NAME_UNKNOWN, ifb_setup);
253af423
JHS
313
314 if (!dev_ifb)
315 return -ENOMEM;
316
9ba2cd65
PM
317 dev_ifb->rtnl_link_ops = &ifb_link_ops;
318 err = register_netdevice(dev_ifb);
319 if (err < 0)
320 goto err;
94833dfb 321
9ba2cd65 322 return 0;
62b7ffca 323
9ba2cd65
PM
324err:
325 free_netdev(dev_ifb);
326 return err;
6aa20a22 327}
253af423
JHS
328
329static int __init ifb_init_module(void)
6aa20a22 330{
9ba2cd65
PM
331 int i, err;
332
333 rtnl_lock();
334 err = __rtnl_link_register(&ifb_link_ops);
f2966cd5 335 if (err < 0)
336 goto out;
62b7ffca 337
440d57bc 338 for (i = 0; i < numifbs && !err; i++) {
6aa20a22 339 err = ifb_init_one(i);
440d57bc 340 cond_resched();
341 }
2d85cba2 342 if (err)
9ba2cd65 343 __rtnl_link_unregister(&ifb_link_ops);
f2966cd5 344
345out:
9ba2cd65 346 rtnl_unlock();
253af423
JHS
347
348 return err;
6aa20a22 349}
253af423
JHS
350
351static void __exit ifb_cleanup_module(void)
352{
2d85cba2 353 rtnl_link_unregister(&ifb_link_ops);
253af423
JHS
354}
355
356module_init(ifb_init_module);
357module_exit(ifb_cleanup_module);
358MODULE_LICENSE("GPL");
359MODULE_AUTHOR("Jamal Hadi Salim");
9ba2cd65 360MODULE_ALIAS_RTNL_LINK("ifb");