]> git.proxmox.com Git - ceph.git/blob - ceph/src/spdk/dpdk/examples/kni/main.c
update sources to ceph Nautilus 14.2.1
[ceph.git] / ceph / src / spdk / dpdk / examples / kni / main.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2010-2014 Intel Corporation
3 */
4
5 #include <stdio.h>
6 #include <stdlib.h>
7 #include <stdint.h>
8 #include <inttypes.h>
9 #include <string.h>
10 #include <sys/queue.h>
11 #include <stdarg.h>
12 #include <errno.h>
13 #include <getopt.h>
14
15 #include <netinet/in.h>
16 #include <linux/if.h>
17 #include <linux/if_tun.h>
18 #include <fcntl.h>
19 #include <sys/ioctl.h>
20 #include <unistd.h>
21 #include <signal.h>
22
23 #include <rte_common.h>
24 #include <rte_log.h>
25 #include <rte_memory.h>
26 #include <rte_memcpy.h>
27 #include <rte_eal.h>
28 #include <rte_per_lcore.h>
29 #include <rte_launch.h>
30 #include <rte_atomic.h>
31 #include <rte_lcore.h>
32 #include <rte_branch_prediction.h>
33 #include <rte_interrupts.h>
34 #include <rte_bus_pci.h>
35 #include <rte_debug.h>
36 #include <rte_ether.h>
37 #include <rte_ethdev.h>
38 #include <rte_mempool.h>
39 #include <rte_mbuf.h>
40 #include <rte_string_fns.h>
41 #include <rte_cycles.h>
42 #include <rte_malloc.h>
43 #include <rte_kni.h>
44
45 /* Macros for printing using RTE_LOG */
46 #define RTE_LOGTYPE_APP RTE_LOGTYPE_USER1
47
48 /* Max size of a single packet */
49 #define MAX_PACKET_SZ 2048
50
51 /* Size of the data buffer in each mbuf */
52 #define MBUF_DATA_SZ (MAX_PACKET_SZ + RTE_PKTMBUF_HEADROOM)
53
54 /* Number of mbufs in mempool that is created */
55 #define NB_MBUF (8192 * 16)
56
57 /* How many packets to attempt to read from NIC in one go */
58 #define PKT_BURST_SZ 32
59
60 /* How many objects (mbufs) to keep in per-lcore mempool cache */
61 #define MEMPOOL_CACHE_SZ PKT_BURST_SZ
62
63 /* Number of RX ring descriptors */
64 #define NB_RXD 1024
65
66 /* Number of TX ring descriptors */
67 #define NB_TXD 1024
68
69 /* Total octets in ethernet header */
70 #define KNI_ENET_HEADER_SIZE 14
71
72 /* Total octets in the FCS */
73 #define KNI_ENET_FCS_SIZE 4
74
75 #define KNI_US_PER_SECOND 1000000
76 #define KNI_SECOND_PER_DAY 86400
77
78 #define KNI_MAX_KTHREAD 32
79 /*
80 * Structure of port parameters
81 */
82 struct kni_port_params {
83 uint16_t port_id;/* Port ID */
84 unsigned lcore_rx; /* lcore ID for RX */
85 unsigned lcore_tx; /* lcore ID for TX */
86 uint32_t nb_lcore_k; /* Number of lcores for KNI multi kernel threads */
87 uint32_t nb_kni; /* Number of KNI devices to be created */
88 unsigned lcore_k[KNI_MAX_KTHREAD]; /* lcore ID list for kthreads */
89 struct rte_kni *kni[KNI_MAX_KTHREAD]; /* KNI context pointers */
90 } __rte_cache_aligned;
91
92 static struct kni_port_params *kni_port_params_array[RTE_MAX_ETHPORTS];
93
94
95 /* Options for configuring ethernet port */
96 static struct rte_eth_conf port_conf = {
97 .rxmode = {
98 .offloads = DEV_RX_OFFLOAD_CRC_STRIP,
99 },
100 .txmode = {
101 .mq_mode = ETH_MQ_TX_NONE,
102 },
103 };
104
105 /* Mempool for mbufs */
106 static struct rte_mempool * pktmbuf_pool = NULL;
107
108 /* Mask of enabled ports */
109 static uint32_t ports_mask = 0;
110 /* Ports set in promiscuous mode off by default. */
111 static int promiscuous_on = 0;
112
113 /* Structure type for recording kni interface specific stats */
114 struct kni_interface_stats {
115 /* number of pkts received from NIC, and sent to KNI */
116 uint64_t rx_packets;
117
118 /* number of pkts received from NIC, but failed to send to KNI */
119 uint64_t rx_dropped;
120
121 /* number of pkts received from KNI, and sent to NIC */
122 uint64_t tx_packets;
123
124 /* number of pkts received from KNI, but failed to send to NIC */
125 uint64_t tx_dropped;
126 };
127
128 /* kni device statistics array */
129 static struct kni_interface_stats kni_stats[RTE_MAX_ETHPORTS];
130
131 static int kni_change_mtu(uint16_t port_id, unsigned int new_mtu);
132 static int kni_config_network_interface(uint16_t port_id, uint8_t if_up);
133 static int kni_config_mac_address(uint16_t port_id, uint8_t mac_addr[]);
134
135 static rte_atomic32_t kni_stop = RTE_ATOMIC32_INIT(0);
136
137 /* Print out statistics on packets handled */
138 static void
139 print_stats(void)
140 {
141 uint16_t i;
142
143 printf("\n**KNI example application statistics**\n"
144 "====== ============== ============ ============ ============ ============\n"
145 " Port Lcore(RX/TX) rx_packets rx_dropped tx_packets tx_dropped\n"
146 "------ -------------- ------------ ------------ ------------ ------------\n");
147 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
148 if (!kni_port_params_array[i])
149 continue;
150
151 printf("%7d %10u/%2u %13"PRIu64" %13"PRIu64" %13"PRIu64" "
152 "%13"PRIu64"\n", i,
153 kni_port_params_array[i]->lcore_rx,
154 kni_port_params_array[i]->lcore_tx,
155 kni_stats[i].rx_packets,
156 kni_stats[i].rx_dropped,
157 kni_stats[i].tx_packets,
158 kni_stats[i].tx_dropped);
159 }
160 printf("====== ============== ============ ============ ============ ============\n");
161 }
162
163 /* Custom handling of signals to handle stats and kni processing */
164 static void
165 signal_handler(int signum)
166 {
167 /* When we receive a USR1 signal, print stats */
168 if (signum == SIGUSR1) {
169 print_stats();
170 }
171
172 /* When we receive a USR2 signal, reset stats */
173 if (signum == SIGUSR2) {
174 memset(&kni_stats, 0, sizeof(kni_stats));
175 printf("\n**Statistics have been reset**\n");
176 return;
177 }
178
179 /* When we receive a RTMIN or SIGINT signal, stop kni processing */
180 if (signum == SIGRTMIN || signum == SIGINT){
181 printf("SIGRTMIN is received, and the KNI processing is "
182 "going to stop\n");
183 rte_atomic32_inc(&kni_stop);
184 return;
185 }
186 }
187
188 static void
189 kni_burst_free_mbufs(struct rte_mbuf **pkts, unsigned num)
190 {
191 unsigned i;
192
193 if (pkts == NULL)
194 return;
195
196 for (i = 0; i < num; i++) {
197 rte_pktmbuf_free(pkts[i]);
198 pkts[i] = NULL;
199 }
200 }
201
202 /**
203 * Interface to burst rx and enqueue mbufs into rx_q
204 */
205 static void
206 kni_ingress(struct kni_port_params *p)
207 {
208 uint8_t i;
209 uint16_t port_id;
210 unsigned nb_rx, num;
211 uint32_t nb_kni;
212 struct rte_mbuf *pkts_burst[PKT_BURST_SZ];
213
214 if (p == NULL)
215 return;
216
217 nb_kni = p->nb_kni;
218 port_id = p->port_id;
219 for (i = 0; i < nb_kni; i++) {
220 /* Burst rx from eth */
221 nb_rx = rte_eth_rx_burst(port_id, 0, pkts_burst, PKT_BURST_SZ);
222 if (unlikely(nb_rx > PKT_BURST_SZ)) {
223 RTE_LOG(ERR, APP, "Error receiving from eth\n");
224 return;
225 }
226 /* Burst tx to kni */
227 num = rte_kni_tx_burst(p->kni[i], pkts_burst, nb_rx);
228 kni_stats[port_id].rx_packets += num;
229
230 rte_kni_handle_request(p->kni[i]);
231 if (unlikely(num < nb_rx)) {
232 /* Free mbufs not tx to kni interface */
233 kni_burst_free_mbufs(&pkts_burst[num], nb_rx - num);
234 kni_stats[port_id].rx_dropped += nb_rx - num;
235 }
236 }
237 }
238
239 /**
240 * Interface to dequeue mbufs from tx_q and burst tx
241 */
242 static void
243 kni_egress(struct kni_port_params *p)
244 {
245 uint8_t i;
246 uint16_t port_id;
247 unsigned nb_tx, num;
248 uint32_t nb_kni;
249 struct rte_mbuf *pkts_burst[PKT_BURST_SZ];
250
251 if (p == NULL)
252 return;
253
254 nb_kni = p->nb_kni;
255 port_id = p->port_id;
256 for (i = 0; i < nb_kni; i++) {
257 /* Burst rx from kni */
258 num = rte_kni_rx_burst(p->kni[i], pkts_burst, PKT_BURST_SZ);
259 if (unlikely(num > PKT_BURST_SZ)) {
260 RTE_LOG(ERR, APP, "Error receiving from KNI\n");
261 return;
262 }
263 /* Burst tx to eth */
264 nb_tx = rte_eth_tx_burst(port_id, 0, pkts_burst, (uint16_t)num);
265 kni_stats[port_id].tx_packets += nb_tx;
266 if (unlikely(nb_tx < num)) {
267 /* Free mbufs not tx to NIC */
268 kni_burst_free_mbufs(&pkts_burst[nb_tx], num - nb_tx);
269 kni_stats[port_id].tx_dropped += num - nb_tx;
270 }
271 }
272 }
273
274 static int
275 main_loop(__rte_unused void *arg)
276 {
277 uint16_t i;
278 int32_t f_stop;
279 const unsigned lcore_id = rte_lcore_id();
280 enum lcore_rxtx {
281 LCORE_NONE,
282 LCORE_RX,
283 LCORE_TX,
284 LCORE_MAX
285 };
286 enum lcore_rxtx flag = LCORE_NONE;
287
288 RTE_ETH_FOREACH_DEV(i) {
289 if (!kni_port_params_array[i])
290 continue;
291 if (kni_port_params_array[i]->lcore_rx == (uint8_t)lcore_id) {
292 flag = LCORE_RX;
293 break;
294 } else if (kni_port_params_array[i]->lcore_tx ==
295 (uint8_t)lcore_id) {
296 flag = LCORE_TX;
297 break;
298 }
299 }
300
301 if (flag == LCORE_RX) {
302 RTE_LOG(INFO, APP, "Lcore %u is reading from port %d\n",
303 kni_port_params_array[i]->lcore_rx,
304 kni_port_params_array[i]->port_id);
305 while (1) {
306 f_stop = rte_atomic32_read(&kni_stop);
307 if (f_stop)
308 break;
309 kni_ingress(kni_port_params_array[i]);
310 }
311 } else if (flag == LCORE_TX) {
312 RTE_LOG(INFO, APP, "Lcore %u is writing to port %d\n",
313 kni_port_params_array[i]->lcore_tx,
314 kni_port_params_array[i]->port_id);
315 while (1) {
316 f_stop = rte_atomic32_read(&kni_stop);
317 if (f_stop)
318 break;
319 kni_egress(kni_port_params_array[i]);
320 }
321 } else
322 RTE_LOG(INFO, APP, "Lcore %u has nothing to do\n", lcore_id);
323
324 return 0;
325 }
326
327 /* Display usage instructions */
328 static void
329 print_usage(const char *prgname)
330 {
331 RTE_LOG(INFO, APP, "\nUsage: %s [EAL options] -- -p PORTMASK -P "
332 "[--config (port,lcore_rx,lcore_tx,lcore_kthread...)"
333 "[,(port,lcore_rx,lcore_tx,lcore_kthread...)]]\n"
334 " -p PORTMASK: hex bitmask of ports to use\n"
335 " -P : enable promiscuous mode\n"
336 " --config (port,lcore_rx,lcore_tx,lcore_kthread...): "
337 "port and lcore configurations\n",
338 prgname);
339 }
340
341 /* Convert string to unsigned number. 0 is returned if error occurs */
342 static uint32_t
343 parse_unsigned(const char *portmask)
344 {
345 char *end = NULL;
346 unsigned long num;
347
348 num = strtoul(portmask, &end, 16);
349 if ((portmask[0] == '\0') || (end == NULL) || (*end != '\0'))
350 return 0;
351
352 return (uint32_t)num;
353 }
354
355 static void
356 print_config(void)
357 {
358 uint32_t i, j;
359 struct kni_port_params **p = kni_port_params_array;
360
361 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
362 if (!p[i])
363 continue;
364 RTE_LOG(DEBUG, APP, "Port ID: %d\n", p[i]->port_id);
365 RTE_LOG(DEBUG, APP, "Rx lcore ID: %u, Tx lcore ID: %u\n",
366 p[i]->lcore_rx, p[i]->lcore_tx);
367 for (j = 0; j < p[i]->nb_lcore_k; j++)
368 RTE_LOG(DEBUG, APP, "Kernel thread lcore ID: %u\n",
369 p[i]->lcore_k[j]);
370 }
371 }
372
373 static int
374 parse_config(const char *arg)
375 {
376 const char *p, *p0 = arg;
377 char s[256], *end;
378 unsigned size;
379 enum fieldnames {
380 FLD_PORT = 0,
381 FLD_LCORE_RX,
382 FLD_LCORE_TX,
383 _NUM_FLD = KNI_MAX_KTHREAD + 3,
384 };
385 int i, j, nb_token;
386 char *str_fld[_NUM_FLD];
387 unsigned long int_fld[_NUM_FLD];
388 uint16_t port_id, nb_kni_port_params = 0;
389
390 memset(&kni_port_params_array, 0, sizeof(kni_port_params_array));
391 while (((p = strchr(p0, '(')) != NULL) &&
392 nb_kni_port_params < RTE_MAX_ETHPORTS) {
393 p++;
394 if ((p0 = strchr(p, ')')) == NULL)
395 goto fail;
396 size = p0 - p;
397 if (size >= sizeof(s)) {
398 printf("Invalid config parameters\n");
399 goto fail;
400 }
401 snprintf(s, sizeof(s), "%.*s", size, p);
402 nb_token = rte_strsplit(s, sizeof(s), str_fld, _NUM_FLD, ',');
403 if (nb_token <= FLD_LCORE_TX) {
404 printf("Invalid config parameters\n");
405 goto fail;
406 }
407 for (i = 0; i < nb_token; i++) {
408 errno = 0;
409 int_fld[i] = strtoul(str_fld[i], &end, 0);
410 if (errno != 0 || end == str_fld[i]) {
411 printf("Invalid config parameters\n");
412 goto fail;
413 }
414 }
415
416 i = 0;
417 port_id = int_fld[i++];
418 if (port_id >= RTE_MAX_ETHPORTS) {
419 printf("Port ID %d could not exceed the maximum %d\n",
420 port_id, RTE_MAX_ETHPORTS);
421 goto fail;
422 }
423 if (kni_port_params_array[port_id]) {
424 printf("Port %d has been configured\n", port_id);
425 goto fail;
426 }
427 kni_port_params_array[port_id] =
428 rte_zmalloc("KNI_port_params",
429 sizeof(struct kni_port_params), RTE_CACHE_LINE_SIZE);
430 kni_port_params_array[port_id]->port_id = port_id;
431 kni_port_params_array[port_id]->lcore_rx =
432 (uint8_t)int_fld[i++];
433 kni_port_params_array[port_id]->lcore_tx =
434 (uint8_t)int_fld[i++];
435 if (kni_port_params_array[port_id]->lcore_rx >= RTE_MAX_LCORE ||
436 kni_port_params_array[port_id]->lcore_tx >= RTE_MAX_LCORE) {
437 printf("lcore_rx %u or lcore_tx %u ID could not "
438 "exceed the maximum %u\n",
439 kni_port_params_array[port_id]->lcore_rx,
440 kni_port_params_array[port_id]->lcore_tx,
441 (unsigned)RTE_MAX_LCORE);
442 goto fail;
443 }
444 for (j = 0; i < nb_token && j < KNI_MAX_KTHREAD; i++, j++)
445 kni_port_params_array[port_id]->lcore_k[j] =
446 (uint8_t)int_fld[i];
447 kni_port_params_array[port_id]->nb_lcore_k = j;
448 }
449 print_config();
450
451 return 0;
452
453 fail:
454 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
455 if (kni_port_params_array[i]) {
456 rte_free(kni_port_params_array[i]);
457 kni_port_params_array[i] = NULL;
458 }
459 }
460
461 return -1;
462 }
463
464 static int
465 validate_parameters(uint32_t portmask)
466 {
467 uint32_t i;
468
469 if (!portmask) {
470 printf("No port configured in port mask\n");
471 return -1;
472 }
473
474 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
475 if (((portmask & (1 << i)) && !kni_port_params_array[i]) ||
476 (!(portmask & (1 << i)) && kni_port_params_array[i]))
477 rte_exit(EXIT_FAILURE, "portmask is not consistent "
478 "to port ids specified in --config\n");
479
480 if (kni_port_params_array[i] && !rte_lcore_is_enabled(\
481 (unsigned)(kni_port_params_array[i]->lcore_rx)))
482 rte_exit(EXIT_FAILURE, "lcore id %u for "
483 "port %d receiving not enabled\n",
484 kni_port_params_array[i]->lcore_rx,
485 kni_port_params_array[i]->port_id);
486
487 if (kni_port_params_array[i] && !rte_lcore_is_enabled(\
488 (unsigned)(kni_port_params_array[i]->lcore_tx)))
489 rte_exit(EXIT_FAILURE, "lcore id %u for "
490 "port %d transmitting not enabled\n",
491 kni_port_params_array[i]->lcore_tx,
492 kni_port_params_array[i]->port_id);
493
494 }
495
496 return 0;
497 }
498
499 #define CMDLINE_OPT_CONFIG "config"
500
501 /* Parse the arguments given in the command line of the application */
502 static int
503 parse_args(int argc, char **argv)
504 {
505 int opt, longindex, ret = 0;
506 const char *prgname = argv[0];
507 static struct option longopts[] = {
508 {CMDLINE_OPT_CONFIG, required_argument, NULL, 0},
509 {NULL, 0, NULL, 0}
510 };
511
512 /* Disable printing messages within getopt() */
513 opterr = 0;
514
515 /* Parse command line */
516 while ((opt = getopt_long(argc, argv, "p:P", longopts,
517 &longindex)) != EOF) {
518 switch (opt) {
519 case 'p':
520 ports_mask = parse_unsigned(optarg);
521 break;
522 case 'P':
523 promiscuous_on = 1;
524 break;
525 case 0:
526 if (!strncmp(longopts[longindex].name,
527 CMDLINE_OPT_CONFIG,
528 sizeof(CMDLINE_OPT_CONFIG))) {
529 ret = parse_config(optarg);
530 if (ret) {
531 printf("Invalid config\n");
532 print_usage(prgname);
533 return -1;
534 }
535 }
536 break;
537 default:
538 print_usage(prgname);
539 rte_exit(EXIT_FAILURE, "Invalid option specified\n");
540 }
541 }
542
543 /* Check that options were parsed ok */
544 if (validate_parameters(ports_mask) < 0) {
545 print_usage(prgname);
546 rte_exit(EXIT_FAILURE, "Invalid parameters\n");
547 }
548
549 return ret;
550 }
551
552 /* Initialize KNI subsystem */
553 static void
554 init_kni(void)
555 {
556 unsigned int num_of_kni_ports = 0, i;
557 struct kni_port_params **params = kni_port_params_array;
558
559 /* Calculate the maximum number of KNI interfaces that will be used */
560 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
561 if (kni_port_params_array[i]) {
562 num_of_kni_ports += (params[i]->nb_lcore_k ?
563 params[i]->nb_lcore_k : 1);
564 }
565 }
566
567 /* Invoke rte KNI init to preallocate the ports */
568 rte_kni_init(num_of_kni_ports);
569 }
570
571 /* Initialise a single port on an Ethernet device */
572 static void
573 init_port(uint16_t port)
574 {
575 int ret;
576 uint16_t nb_rxd = NB_RXD;
577 uint16_t nb_txd = NB_TXD;
578 struct rte_eth_dev_info dev_info;
579 struct rte_eth_rxconf rxq_conf;
580 struct rte_eth_txconf txq_conf;
581 struct rte_eth_conf local_port_conf = port_conf;
582
583 /* Initialise device and RX/TX queues */
584 RTE_LOG(INFO, APP, "Initialising port %u ...\n", (unsigned)port);
585 fflush(stdout);
586 rte_eth_dev_info_get(port, &dev_info);
587 if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)
588 local_port_conf.txmode.offloads |=
589 DEV_TX_OFFLOAD_MBUF_FAST_FREE;
590 ret = rte_eth_dev_configure(port, 1, 1, &local_port_conf);
591 if (ret < 0)
592 rte_exit(EXIT_FAILURE, "Could not configure port%u (%d)\n",
593 (unsigned)port, ret);
594
595 ret = rte_eth_dev_adjust_nb_rx_tx_desc(port, &nb_rxd, &nb_txd);
596 if (ret < 0)
597 rte_exit(EXIT_FAILURE, "Could not adjust number of descriptors "
598 "for port%u (%d)\n", (unsigned)port, ret);
599
600 rxq_conf = dev_info.default_rxconf;
601 rxq_conf.offloads = local_port_conf.rxmode.offloads;
602 ret = rte_eth_rx_queue_setup(port, 0, nb_rxd,
603 rte_eth_dev_socket_id(port), &rxq_conf, pktmbuf_pool);
604 if (ret < 0)
605 rte_exit(EXIT_FAILURE, "Could not setup up RX queue for "
606 "port%u (%d)\n", (unsigned)port, ret);
607
608 txq_conf = dev_info.default_txconf;
609 txq_conf.offloads = local_port_conf.txmode.offloads;
610 ret = rte_eth_tx_queue_setup(port, 0, nb_txd,
611 rte_eth_dev_socket_id(port), &txq_conf);
612 if (ret < 0)
613 rte_exit(EXIT_FAILURE, "Could not setup up TX queue for "
614 "port%u (%d)\n", (unsigned)port, ret);
615
616 ret = rte_eth_dev_start(port);
617 if (ret < 0)
618 rte_exit(EXIT_FAILURE, "Could not start port%u (%d)\n",
619 (unsigned)port, ret);
620
621 if (promiscuous_on)
622 rte_eth_promiscuous_enable(port);
623 }
624
625 /* Check the link status of all ports in up to 9s, and print them finally */
626 static void
627 check_all_ports_link_status(uint32_t port_mask)
628 {
629 #define CHECK_INTERVAL 100 /* 100ms */
630 #define MAX_CHECK_TIME 90 /* 9s (90 * 100ms) in total */
631 uint16_t portid;
632 uint8_t count, all_ports_up, print_flag = 0;
633 struct rte_eth_link link;
634
635 printf("\nChecking link status\n");
636 fflush(stdout);
637 for (count = 0; count <= MAX_CHECK_TIME; count++) {
638 all_ports_up = 1;
639 RTE_ETH_FOREACH_DEV(portid) {
640 if ((port_mask & (1 << portid)) == 0)
641 continue;
642 memset(&link, 0, sizeof(link));
643 rte_eth_link_get_nowait(portid, &link);
644 /* print link status if flag set */
645 if (print_flag == 1) {
646 if (link.link_status)
647 printf(
648 "Port%d Link Up - speed %uMbps - %s\n",
649 portid, link.link_speed,
650 (link.link_duplex == ETH_LINK_FULL_DUPLEX) ?
651 ("full-duplex") : ("half-duplex\n"));
652 else
653 printf("Port %d Link Down\n", portid);
654 continue;
655 }
656 /* clear all_ports_up flag if any link down */
657 if (link.link_status == ETH_LINK_DOWN) {
658 all_ports_up = 0;
659 break;
660 }
661 }
662 /* after finally printing all link status, get out */
663 if (print_flag == 1)
664 break;
665
666 if (all_ports_up == 0) {
667 printf(".");
668 fflush(stdout);
669 rte_delay_ms(CHECK_INTERVAL);
670 }
671
672 /* set the print_flag if all ports up or timeout */
673 if (all_ports_up == 1 || count == (MAX_CHECK_TIME - 1)) {
674 print_flag = 1;
675 printf("done\n");
676 }
677 }
678 }
679
680 /* Callback for request of changing MTU */
681 static int
682 kni_change_mtu(uint16_t port_id, unsigned int new_mtu)
683 {
684 int ret;
685 uint16_t nb_rxd = NB_RXD;
686 struct rte_eth_conf conf;
687 struct rte_eth_dev_info dev_info;
688 struct rte_eth_rxconf rxq_conf;
689
690 if (!rte_eth_dev_is_valid_port(port_id)) {
691 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
692 return -EINVAL;
693 }
694
695 RTE_LOG(INFO, APP, "Change MTU of port %d to %u\n", port_id, new_mtu);
696
697 /* Stop specific port */
698 rte_eth_dev_stop(port_id);
699
700 memcpy(&conf, &port_conf, sizeof(conf));
701 /* Set new MTU */
702 if (new_mtu > ETHER_MAX_LEN)
703 conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME;
704 else
705 conf.rxmode.offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME;
706
707 /* mtu + length of header + length of FCS = max pkt length */
708 conf.rxmode.max_rx_pkt_len = new_mtu + KNI_ENET_HEADER_SIZE +
709 KNI_ENET_FCS_SIZE;
710 ret = rte_eth_dev_configure(port_id, 1, 1, &conf);
711 if (ret < 0) {
712 RTE_LOG(ERR, APP, "Fail to reconfigure port %d\n", port_id);
713 return ret;
714 }
715
716 ret = rte_eth_dev_adjust_nb_rx_tx_desc(port_id, &nb_rxd, NULL);
717 if (ret < 0)
718 rte_exit(EXIT_FAILURE, "Could not adjust number of descriptors "
719 "for port%u (%d)\n", (unsigned int)port_id,
720 ret);
721
722 rte_eth_dev_info_get(port_id, &dev_info);
723 rxq_conf = dev_info.default_rxconf;
724 rxq_conf.offloads = conf.rxmode.offloads;
725 ret = rte_eth_rx_queue_setup(port_id, 0, nb_rxd,
726 rte_eth_dev_socket_id(port_id), &rxq_conf, pktmbuf_pool);
727 if (ret < 0) {
728 RTE_LOG(ERR, APP, "Fail to setup Rx queue of port %d\n",
729 port_id);
730 return ret;
731 }
732
733 /* Restart specific port */
734 ret = rte_eth_dev_start(port_id);
735 if (ret < 0) {
736 RTE_LOG(ERR, APP, "Fail to restart port %d\n", port_id);
737 return ret;
738 }
739
740 return 0;
741 }
742
743 /* Callback for request of configuring network interface up/down */
744 static int
745 kni_config_network_interface(uint16_t port_id, uint8_t if_up)
746 {
747 int ret = 0;
748
749 if (!rte_eth_dev_is_valid_port(port_id)) {
750 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
751 return -EINVAL;
752 }
753
754 RTE_LOG(INFO, APP, "Configure network interface of %d %s\n",
755 port_id, if_up ? "up" : "down");
756
757 if (if_up != 0) { /* Configure network interface up */
758 rte_eth_dev_stop(port_id);
759 ret = rte_eth_dev_start(port_id);
760 } else /* Configure network interface down */
761 rte_eth_dev_stop(port_id);
762
763 if (ret < 0)
764 RTE_LOG(ERR, APP, "Failed to start port %d\n", port_id);
765
766 return ret;
767 }
768
769 static void
770 print_ethaddr(const char *name, struct ether_addr *mac_addr)
771 {
772 char buf[ETHER_ADDR_FMT_SIZE];
773 ether_format_addr(buf, ETHER_ADDR_FMT_SIZE, mac_addr);
774 RTE_LOG(INFO, APP, "\t%s%s\n", name, buf);
775 }
776
777 /* Callback for request of configuring mac address */
778 static int
779 kni_config_mac_address(uint16_t port_id, uint8_t mac_addr[])
780 {
781 int ret = 0;
782
783 if (!rte_eth_dev_is_valid_port(port_id)) {
784 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
785 return -EINVAL;
786 }
787
788 RTE_LOG(INFO, APP, "Configure mac address of %d\n", port_id);
789 print_ethaddr("Address:", (struct ether_addr *)mac_addr);
790
791 ret = rte_eth_dev_default_mac_addr_set(port_id,
792 (struct ether_addr *)mac_addr);
793 if (ret < 0)
794 RTE_LOG(ERR, APP, "Failed to config mac_addr for port %d\n",
795 port_id);
796
797 return ret;
798 }
799
800 static int
801 kni_alloc(uint16_t port_id)
802 {
803 uint8_t i;
804 struct rte_kni *kni;
805 struct rte_kni_conf conf;
806 struct kni_port_params **params = kni_port_params_array;
807
808 if (port_id >= RTE_MAX_ETHPORTS || !params[port_id])
809 return -1;
810
811 params[port_id]->nb_kni = params[port_id]->nb_lcore_k ?
812 params[port_id]->nb_lcore_k : 1;
813
814 for (i = 0; i < params[port_id]->nb_kni; i++) {
815 /* Clear conf at first */
816 memset(&conf, 0, sizeof(conf));
817 if (params[port_id]->nb_lcore_k) {
818 snprintf(conf.name, RTE_KNI_NAMESIZE,
819 "vEth%u_%u", port_id, i);
820 conf.core_id = params[port_id]->lcore_k[i];
821 conf.force_bind = 1;
822 } else
823 snprintf(conf.name, RTE_KNI_NAMESIZE,
824 "vEth%u", port_id);
825 conf.group_id = port_id;
826 conf.mbuf_size = MAX_PACKET_SZ;
827 /*
828 * The first KNI device associated to a port
829 * is the master, for multiple kernel thread
830 * environment.
831 */
832 if (i == 0) {
833 struct rte_kni_ops ops;
834 struct rte_eth_dev_info dev_info;
835 const struct rte_pci_device *pci_dev;
836 const struct rte_bus *bus = NULL;
837
838 memset(&dev_info, 0, sizeof(dev_info));
839 rte_eth_dev_info_get(port_id, &dev_info);
840
841 if (dev_info.device)
842 bus = rte_bus_find_by_device(dev_info.device);
843 if (bus && !strcmp(bus->name, "pci")) {
844 pci_dev = RTE_DEV_TO_PCI(dev_info.device);
845 conf.addr = pci_dev->addr;
846 conf.id = pci_dev->id;
847 }
848 /* Get the interface default mac address */
849 rte_eth_macaddr_get(port_id,
850 (struct ether_addr *)&conf.mac_addr);
851
852 rte_eth_dev_get_mtu(port_id, &conf.mtu);
853
854 memset(&ops, 0, sizeof(ops));
855 ops.port_id = port_id;
856 ops.change_mtu = kni_change_mtu;
857 ops.config_network_if = kni_config_network_interface;
858 ops.config_mac_address = kni_config_mac_address;
859
860 kni = rte_kni_alloc(pktmbuf_pool, &conf, &ops);
861 } else
862 kni = rte_kni_alloc(pktmbuf_pool, &conf, NULL);
863
864 if (!kni)
865 rte_exit(EXIT_FAILURE, "Fail to create kni for "
866 "port: %d\n", port_id);
867 params[port_id]->kni[i] = kni;
868 }
869
870 return 0;
871 }
872
873 static int
874 kni_free_kni(uint16_t port_id)
875 {
876 uint8_t i;
877 struct kni_port_params **p = kni_port_params_array;
878
879 if (port_id >= RTE_MAX_ETHPORTS || !p[port_id])
880 return -1;
881
882 for (i = 0; i < p[port_id]->nb_kni; i++) {
883 if (rte_kni_release(p[port_id]->kni[i]))
884 printf("Fail to release kni\n");
885 p[port_id]->kni[i] = NULL;
886 }
887 rte_eth_dev_stop(port_id);
888
889 return 0;
890 }
891
892 /* Initialise ports/queues etc. and start main loop on each core */
893 int
894 main(int argc, char** argv)
895 {
896 int ret;
897 uint16_t nb_sys_ports, port;
898 unsigned i;
899
900 /* Associate signal_hanlder function with USR signals */
901 signal(SIGUSR1, signal_handler);
902 signal(SIGUSR2, signal_handler);
903 signal(SIGRTMIN, signal_handler);
904 signal(SIGINT, signal_handler);
905
906 /* Initialise EAL */
907 ret = rte_eal_init(argc, argv);
908 if (ret < 0)
909 rte_exit(EXIT_FAILURE, "Could not initialise EAL (%d)\n", ret);
910 argc -= ret;
911 argv += ret;
912
913 /* Parse application arguments (after the EAL ones) */
914 ret = parse_args(argc, argv);
915 if (ret < 0)
916 rte_exit(EXIT_FAILURE, "Could not parse input parameters\n");
917
918 /* Create the mbuf pool */
919 pktmbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF,
920 MEMPOOL_CACHE_SZ, 0, MBUF_DATA_SZ, rte_socket_id());
921 if (pktmbuf_pool == NULL) {
922 rte_exit(EXIT_FAILURE, "Could not initialise mbuf pool\n");
923 return -1;
924 }
925
926 /* Get number of ports found in scan */
927 nb_sys_ports = rte_eth_dev_count_avail();
928 if (nb_sys_ports == 0)
929 rte_exit(EXIT_FAILURE, "No supported Ethernet device found\n");
930
931 /* Check if the configured port ID is valid */
932 for (i = 0; i < RTE_MAX_ETHPORTS; i++)
933 if (kni_port_params_array[i] && !rte_eth_dev_is_valid_port(i))
934 rte_exit(EXIT_FAILURE, "Configured invalid "
935 "port ID %u\n", i);
936
937 /* Initialize KNI subsystem */
938 init_kni();
939
940 /* Initialise each port */
941 RTE_ETH_FOREACH_DEV(port) {
942 /* Skip ports that are not enabled */
943 if (!(ports_mask & (1 << port)))
944 continue;
945 init_port(port);
946
947 if (port >= RTE_MAX_ETHPORTS)
948 rte_exit(EXIT_FAILURE, "Can not use more than "
949 "%d ports for kni\n", RTE_MAX_ETHPORTS);
950
951 kni_alloc(port);
952 }
953 check_all_ports_link_status(ports_mask);
954
955 /* Launch per-lcore function on every lcore */
956 rte_eal_mp_remote_launch(main_loop, NULL, CALL_MASTER);
957 RTE_LCORE_FOREACH_SLAVE(i) {
958 if (rte_eal_wait_lcore(i) < 0)
959 return -1;
960 }
961
962 /* Release resources */
963 RTE_ETH_FOREACH_DEV(port) {
964 if (!(ports_mask & (1 << port)))
965 continue;
966 kni_free_kni(port);
967 }
968 for (i = 0; i < RTE_MAX_ETHPORTS; i++)
969 if (kni_port_params_array[i]) {
970 rte_free(kni_port_params_array[i]);
971 kni_port_params_array[i] = NULL;
972 }
973
974 return 0;
975 }