]> git.proxmox.com Git - ceph.git/blob - ceph/src/spdk/dpdk/examples/kni/main.c
update source to Ceph Pacific 16.2.2
[ceph.git] / ceph / src / spdk / dpdk / examples / kni / main.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2010-2014 Intel Corporation
3 */
4
5 #include <stdio.h>
6 #include <stdlib.h>
7 #include <stdint.h>
8 #include <inttypes.h>
9 #include <string.h>
10 #include <sys/queue.h>
11 #include <stdarg.h>
12 #include <errno.h>
13 #include <getopt.h>
14
15 #include <netinet/in.h>
16 #include <linux/if.h>
17 #include <linux/if_tun.h>
18 #include <fcntl.h>
19 #include <sys/ioctl.h>
20 #include <unistd.h>
21 #include <signal.h>
22
23 #include <rte_common.h>
24 #include <rte_log.h>
25 #include <rte_memory.h>
26 #include <rte_memcpy.h>
27 #include <rte_eal.h>
28 #include <rte_per_lcore.h>
29 #include <rte_launch.h>
30 #include <rte_atomic.h>
31 #include <rte_lcore.h>
32 #include <rte_branch_prediction.h>
33 #include <rte_interrupts.h>
34 #include <rte_bus_pci.h>
35 #include <rte_debug.h>
36 #include <rte_ether.h>
37 #include <rte_ethdev.h>
38 #include <rte_mempool.h>
39 #include <rte_mbuf.h>
40 #include <rte_string_fns.h>
41 #include <rte_cycles.h>
42 #include <rte_malloc.h>
43 #include <rte_kni.h>
44
45 /* Macros for printing using RTE_LOG */
46 #define RTE_LOGTYPE_APP RTE_LOGTYPE_USER1
47
48 /* Max size of a single packet */
49 #define MAX_PACKET_SZ 2048
50
51 /* Size of the data buffer in each mbuf */
52 #define MBUF_DATA_SZ (MAX_PACKET_SZ + RTE_PKTMBUF_HEADROOM)
53
54 /* Number of mbufs in mempool that is created */
55 #define NB_MBUF (8192 * 16)
56
57 /* How many packets to attempt to read from NIC in one go */
58 #define PKT_BURST_SZ 32
59
60 /* How many objects (mbufs) to keep in per-lcore mempool cache */
61 #define MEMPOOL_CACHE_SZ PKT_BURST_SZ
62
63 /* Number of RX ring descriptors */
64 #define NB_RXD 1024
65
66 /* Number of TX ring descriptors */
67 #define NB_TXD 1024
68
69 /* Total octets in ethernet header */
70 #define KNI_ENET_HEADER_SIZE 14
71
72 /* Total octets in the FCS */
73 #define KNI_ENET_FCS_SIZE 4
74
75 #define KNI_US_PER_SECOND 1000000
76 #define KNI_SECOND_PER_DAY 86400
77
78 #define KNI_MAX_KTHREAD 32
79 /*
80 * Structure of port parameters
81 */
82 struct kni_port_params {
83 uint16_t port_id;/* Port ID */
84 unsigned lcore_rx; /* lcore ID for RX */
85 unsigned lcore_tx; /* lcore ID for TX */
86 uint32_t nb_lcore_k; /* Number of lcores for KNI multi kernel threads */
87 uint32_t nb_kni; /* Number of KNI devices to be created */
88 unsigned lcore_k[KNI_MAX_KTHREAD]; /* lcore ID list for kthreads */
89 struct rte_kni *kni[KNI_MAX_KTHREAD]; /* KNI context pointers */
90 } __rte_cache_aligned;
91
92 static struct kni_port_params *kni_port_params_array[RTE_MAX_ETHPORTS];
93
94
95 /* Options for configuring ethernet port */
96 static struct rte_eth_conf port_conf = {
97 .txmode = {
98 .mq_mode = ETH_MQ_TX_NONE,
99 },
100 };
101
102 /* Mempool for mbufs */
103 static struct rte_mempool * pktmbuf_pool = NULL;
104
105 /* Mask of enabled ports */
106 static uint32_t ports_mask = 0;
107 /* Ports set in promiscuous mode off by default. */
108 static int promiscuous_on = 0;
109 /* Monitor link status continually. off by default. */
110 static int monitor_links;
111
112 /* Structure type for recording kni interface specific stats */
113 struct kni_interface_stats {
114 /* number of pkts received from NIC, and sent to KNI */
115 uint64_t rx_packets;
116
117 /* number of pkts received from NIC, but failed to send to KNI */
118 uint64_t rx_dropped;
119
120 /* number of pkts received from KNI, and sent to NIC */
121 uint64_t tx_packets;
122
123 /* number of pkts received from KNI, but failed to send to NIC */
124 uint64_t tx_dropped;
125 };
126
127 /* kni device statistics array */
128 static struct kni_interface_stats kni_stats[RTE_MAX_ETHPORTS];
129
130 static int kni_change_mtu(uint16_t port_id, unsigned int new_mtu);
131 static int kni_config_network_interface(uint16_t port_id, uint8_t if_up);
132 static int kni_config_mac_address(uint16_t port_id, uint8_t mac_addr[]);
133
134 static rte_atomic32_t kni_stop = RTE_ATOMIC32_INIT(0);
135 static rte_atomic32_t kni_pause = RTE_ATOMIC32_INIT(0);
136
137 /* Print out statistics on packets handled */
138 static void
139 print_stats(void)
140 {
141 uint16_t i;
142
143 printf("\n**KNI example application statistics**\n"
144 "====== ============== ============ ============ ============ ============\n"
145 " Port Lcore(RX/TX) rx_packets rx_dropped tx_packets tx_dropped\n"
146 "------ -------------- ------------ ------------ ------------ ------------\n");
147 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
148 if (!kni_port_params_array[i])
149 continue;
150
151 printf("%7d %10u/%2u %13"PRIu64" %13"PRIu64" %13"PRIu64" "
152 "%13"PRIu64"\n", i,
153 kni_port_params_array[i]->lcore_rx,
154 kni_port_params_array[i]->lcore_tx,
155 kni_stats[i].rx_packets,
156 kni_stats[i].rx_dropped,
157 kni_stats[i].tx_packets,
158 kni_stats[i].tx_dropped);
159 }
160 printf("====== ============== ============ ============ ============ ============\n");
161 }
162
163 /* Custom handling of signals to handle stats and kni processing */
164 static void
165 signal_handler(int signum)
166 {
167 /* When we receive a USR1 signal, print stats */
168 if (signum == SIGUSR1) {
169 print_stats();
170 }
171
172 /* When we receive a USR2 signal, reset stats */
173 if (signum == SIGUSR2) {
174 memset(&kni_stats, 0, sizeof(kni_stats));
175 printf("\n** Statistics have been reset **\n");
176 return;
177 }
178
179 /*
180 * When we receive a RTMIN or SIGINT or SIGTERM signal,
181 * stop kni processing
182 */
183 if (signum == SIGRTMIN || signum == SIGINT || signum == SIGTERM) {
184 printf("\nSIGRTMIN/SIGINT/SIGTERM received. "
185 "KNI processing stopping.\n");
186 rte_atomic32_inc(&kni_stop);
187 return;
188 }
189 }
190
191 static void
192 kni_burst_free_mbufs(struct rte_mbuf **pkts, unsigned num)
193 {
194 unsigned i;
195
196 if (pkts == NULL)
197 return;
198
199 for (i = 0; i < num; i++) {
200 rte_pktmbuf_free(pkts[i]);
201 pkts[i] = NULL;
202 }
203 }
204
205 /**
206 * Interface to burst rx and enqueue mbufs into rx_q
207 */
208 static void
209 kni_ingress(struct kni_port_params *p)
210 {
211 uint8_t i;
212 uint16_t port_id;
213 unsigned nb_rx, num;
214 uint32_t nb_kni;
215 struct rte_mbuf *pkts_burst[PKT_BURST_SZ];
216
217 if (p == NULL)
218 return;
219
220 nb_kni = p->nb_kni;
221 port_id = p->port_id;
222 for (i = 0; i < nb_kni; i++) {
223 /* Burst rx from eth */
224 nb_rx = rte_eth_rx_burst(port_id, 0, pkts_burst, PKT_BURST_SZ);
225 if (unlikely(nb_rx > PKT_BURST_SZ)) {
226 RTE_LOG(ERR, APP, "Error receiving from eth\n");
227 return;
228 }
229 /* Burst tx to kni */
230 num = rte_kni_tx_burst(p->kni[i], pkts_burst, nb_rx);
231 if (num)
232 kni_stats[port_id].rx_packets += num;
233
234 rte_kni_handle_request(p->kni[i]);
235 if (unlikely(num < nb_rx)) {
236 /* Free mbufs not tx to kni interface */
237 kni_burst_free_mbufs(&pkts_burst[num], nb_rx - num);
238 kni_stats[port_id].rx_dropped += nb_rx - num;
239 }
240 }
241 }
242
243 /**
244 * Interface to dequeue mbufs from tx_q and burst tx
245 */
246 static void
247 kni_egress(struct kni_port_params *p)
248 {
249 uint8_t i;
250 uint16_t port_id;
251 unsigned nb_tx, num;
252 uint32_t nb_kni;
253 struct rte_mbuf *pkts_burst[PKT_BURST_SZ];
254
255 if (p == NULL)
256 return;
257
258 nb_kni = p->nb_kni;
259 port_id = p->port_id;
260 for (i = 0; i < nb_kni; i++) {
261 /* Burst rx from kni */
262 num = rte_kni_rx_burst(p->kni[i], pkts_burst, PKT_BURST_SZ);
263 if (unlikely(num > PKT_BURST_SZ)) {
264 RTE_LOG(ERR, APP, "Error receiving from KNI\n");
265 return;
266 }
267 /* Burst tx to eth */
268 nb_tx = rte_eth_tx_burst(port_id, 0, pkts_burst, (uint16_t)num);
269 if (nb_tx)
270 kni_stats[port_id].tx_packets += nb_tx;
271 if (unlikely(nb_tx < num)) {
272 /* Free mbufs not tx to NIC */
273 kni_burst_free_mbufs(&pkts_burst[nb_tx], num - nb_tx);
274 kni_stats[port_id].tx_dropped += num - nb_tx;
275 }
276 }
277 }
278
279 static int
280 main_loop(__rte_unused void *arg)
281 {
282 uint16_t i;
283 int32_t f_stop;
284 int32_t f_pause;
285 const unsigned lcore_id = rte_lcore_id();
286 enum lcore_rxtx {
287 LCORE_NONE,
288 LCORE_RX,
289 LCORE_TX,
290 LCORE_MAX
291 };
292 enum lcore_rxtx flag = LCORE_NONE;
293
294 RTE_ETH_FOREACH_DEV(i) {
295 if (!kni_port_params_array[i])
296 continue;
297 if (kni_port_params_array[i]->lcore_rx == (uint8_t)lcore_id) {
298 flag = LCORE_RX;
299 break;
300 } else if (kni_port_params_array[i]->lcore_tx ==
301 (uint8_t)lcore_id) {
302 flag = LCORE_TX;
303 break;
304 }
305 }
306
307 if (flag == LCORE_RX) {
308 RTE_LOG(INFO, APP, "Lcore %u is reading from port %d\n",
309 kni_port_params_array[i]->lcore_rx,
310 kni_port_params_array[i]->port_id);
311 while (1) {
312 f_stop = rte_atomic32_read(&kni_stop);
313 f_pause = rte_atomic32_read(&kni_pause);
314 if (f_stop)
315 break;
316 if (f_pause)
317 continue;
318 kni_ingress(kni_port_params_array[i]);
319 }
320 } else if (flag == LCORE_TX) {
321 RTE_LOG(INFO, APP, "Lcore %u is writing to port %d\n",
322 kni_port_params_array[i]->lcore_tx,
323 kni_port_params_array[i]->port_id);
324 while (1) {
325 f_stop = rte_atomic32_read(&kni_stop);
326 f_pause = rte_atomic32_read(&kni_pause);
327 if (f_stop)
328 break;
329 if (f_pause)
330 continue;
331 kni_egress(kni_port_params_array[i]);
332 }
333 } else
334 RTE_LOG(INFO, APP, "Lcore %u has nothing to do\n", lcore_id);
335
336 return 0;
337 }
338
339 /* Display usage instructions */
340 static void
341 print_usage(const char *prgname)
342 {
343 RTE_LOG(INFO, APP, "\nUsage: %s [EAL options] -- -p PORTMASK -P -m "
344 "[--config (port,lcore_rx,lcore_tx,lcore_kthread...)"
345 "[,(port,lcore_rx,lcore_tx,lcore_kthread...)]]\n"
346 " -p PORTMASK: hex bitmask of ports to use\n"
347 " -P : enable promiscuous mode\n"
348 " -m : enable monitoring of port carrier state\n"
349 " --config (port,lcore_rx,lcore_tx,lcore_kthread...): "
350 "port and lcore configurations\n",
351 prgname);
352 }
353
354 /* Convert string to unsigned number. 0 is returned if error occurs */
355 static uint32_t
356 parse_unsigned(const char *portmask)
357 {
358 char *end = NULL;
359 unsigned long num;
360
361 num = strtoul(portmask, &end, 16);
362 if ((portmask[0] == '\0') || (end == NULL) || (*end != '\0'))
363 return 0;
364
365 return (uint32_t)num;
366 }
367
368 static void
369 print_config(void)
370 {
371 uint32_t i, j;
372 struct kni_port_params **p = kni_port_params_array;
373
374 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
375 if (!p[i])
376 continue;
377 RTE_LOG(DEBUG, APP, "Port ID: %d\n", p[i]->port_id);
378 RTE_LOG(DEBUG, APP, "Rx lcore ID: %u, Tx lcore ID: %u\n",
379 p[i]->lcore_rx, p[i]->lcore_tx);
380 for (j = 0; j < p[i]->nb_lcore_k; j++)
381 RTE_LOG(DEBUG, APP, "Kernel thread lcore ID: %u\n",
382 p[i]->lcore_k[j]);
383 }
384 }
385
386 static int
387 parse_config(const char *arg)
388 {
389 const char *p, *p0 = arg;
390 char s[256], *end;
391 unsigned size;
392 enum fieldnames {
393 FLD_PORT = 0,
394 FLD_LCORE_RX,
395 FLD_LCORE_TX,
396 _NUM_FLD = KNI_MAX_KTHREAD + 3,
397 };
398 int i, j, nb_token;
399 char *str_fld[_NUM_FLD];
400 unsigned long int_fld[_NUM_FLD];
401 uint16_t port_id, nb_kni_port_params = 0;
402
403 memset(&kni_port_params_array, 0, sizeof(kni_port_params_array));
404 while (((p = strchr(p0, '(')) != NULL) &&
405 nb_kni_port_params < RTE_MAX_ETHPORTS) {
406 p++;
407 if ((p0 = strchr(p, ')')) == NULL)
408 goto fail;
409 size = p0 - p;
410 if (size >= sizeof(s)) {
411 printf("Invalid config parameters\n");
412 goto fail;
413 }
414 snprintf(s, sizeof(s), "%.*s", size, p);
415 nb_token = rte_strsplit(s, sizeof(s), str_fld, _NUM_FLD, ',');
416 if (nb_token <= FLD_LCORE_TX) {
417 printf("Invalid config parameters\n");
418 goto fail;
419 }
420 for (i = 0; i < nb_token; i++) {
421 errno = 0;
422 int_fld[i] = strtoul(str_fld[i], &end, 0);
423 if (errno != 0 || end == str_fld[i]) {
424 printf("Invalid config parameters\n");
425 goto fail;
426 }
427 }
428
429 i = 0;
430 port_id = int_fld[i++];
431 if (port_id >= RTE_MAX_ETHPORTS) {
432 printf("Port ID %d could not exceed the maximum %d\n",
433 port_id, RTE_MAX_ETHPORTS);
434 goto fail;
435 }
436 if (kni_port_params_array[port_id]) {
437 printf("Port %d has been configured\n", port_id);
438 goto fail;
439 }
440 kni_port_params_array[port_id] =
441 rte_zmalloc("KNI_port_params",
442 sizeof(struct kni_port_params), RTE_CACHE_LINE_SIZE);
443 kni_port_params_array[port_id]->port_id = port_id;
444 kni_port_params_array[port_id]->lcore_rx =
445 (uint8_t)int_fld[i++];
446 kni_port_params_array[port_id]->lcore_tx =
447 (uint8_t)int_fld[i++];
448 if (kni_port_params_array[port_id]->lcore_rx >= RTE_MAX_LCORE ||
449 kni_port_params_array[port_id]->lcore_tx >= RTE_MAX_LCORE) {
450 printf("lcore_rx %u or lcore_tx %u ID could not "
451 "exceed the maximum %u\n",
452 kni_port_params_array[port_id]->lcore_rx,
453 kni_port_params_array[port_id]->lcore_tx,
454 (unsigned)RTE_MAX_LCORE);
455 goto fail;
456 }
457 for (j = 0; i < nb_token && j < KNI_MAX_KTHREAD; i++, j++)
458 kni_port_params_array[port_id]->lcore_k[j] =
459 (uint8_t)int_fld[i];
460 kni_port_params_array[port_id]->nb_lcore_k = j;
461 }
462 print_config();
463
464 return 0;
465
466 fail:
467 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
468 if (kni_port_params_array[i]) {
469 rte_free(kni_port_params_array[i]);
470 kni_port_params_array[i] = NULL;
471 }
472 }
473
474 return -1;
475 }
476
477 static int
478 validate_parameters(uint32_t portmask)
479 {
480 uint32_t i;
481
482 if (!portmask) {
483 printf("No port configured in port mask\n");
484 return -1;
485 }
486
487 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
488 if (((portmask & (1 << i)) && !kni_port_params_array[i]) ||
489 (!(portmask & (1 << i)) && kni_port_params_array[i]))
490 rte_exit(EXIT_FAILURE, "portmask is not consistent "
491 "to port ids specified in --config\n");
492
493 if (kni_port_params_array[i] && !rte_lcore_is_enabled(\
494 (unsigned)(kni_port_params_array[i]->lcore_rx)))
495 rte_exit(EXIT_FAILURE, "lcore id %u for "
496 "port %d receiving not enabled\n",
497 kni_port_params_array[i]->lcore_rx,
498 kni_port_params_array[i]->port_id);
499
500 if (kni_port_params_array[i] && !rte_lcore_is_enabled(\
501 (unsigned)(kni_port_params_array[i]->lcore_tx)))
502 rte_exit(EXIT_FAILURE, "lcore id %u for "
503 "port %d transmitting not enabled\n",
504 kni_port_params_array[i]->lcore_tx,
505 kni_port_params_array[i]->port_id);
506
507 }
508
509 return 0;
510 }
511
512 #define CMDLINE_OPT_CONFIG "config"
513
514 /* Parse the arguments given in the command line of the application */
515 static int
516 parse_args(int argc, char **argv)
517 {
518 int opt, longindex, ret = 0;
519 const char *prgname = argv[0];
520 static struct option longopts[] = {
521 {CMDLINE_OPT_CONFIG, required_argument, NULL, 0},
522 {NULL, 0, NULL, 0}
523 };
524
525 /* Disable printing messages within getopt() */
526 opterr = 0;
527
528 /* Parse command line */
529 while ((opt = getopt_long(argc, argv, "p:Pm", longopts,
530 &longindex)) != EOF) {
531 switch (opt) {
532 case 'p':
533 ports_mask = parse_unsigned(optarg);
534 break;
535 case 'P':
536 promiscuous_on = 1;
537 break;
538 case 'm':
539 monitor_links = 1;
540 break;
541 case 0:
542 if (!strncmp(longopts[longindex].name,
543 CMDLINE_OPT_CONFIG,
544 sizeof(CMDLINE_OPT_CONFIG))) {
545 ret = parse_config(optarg);
546 if (ret) {
547 printf("Invalid config\n");
548 print_usage(prgname);
549 return -1;
550 }
551 }
552 break;
553 default:
554 print_usage(prgname);
555 rte_exit(EXIT_FAILURE, "Invalid option specified\n");
556 }
557 }
558
559 /* Check that options were parsed ok */
560 if (validate_parameters(ports_mask) < 0) {
561 print_usage(prgname);
562 rte_exit(EXIT_FAILURE, "Invalid parameters\n");
563 }
564
565 return ret;
566 }
567
568 /* Initialize KNI subsystem */
569 static void
570 init_kni(void)
571 {
572 unsigned int num_of_kni_ports = 0, i;
573 struct kni_port_params **params = kni_port_params_array;
574
575 /* Calculate the maximum number of KNI interfaces that will be used */
576 for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
577 if (kni_port_params_array[i]) {
578 num_of_kni_ports += (params[i]->nb_lcore_k ?
579 params[i]->nb_lcore_k : 1);
580 }
581 }
582
583 /* Invoke rte KNI init to preallocate the ports */
584 rte_kni_init(num_of_kni_ports);
585 }
586
587 /* Initialise a single port on an Ethernet device */
588 static void
589 init_port(uint16_t port)
590 {
591 int ret;
592 uint16_t nb_rxd = NB_RXD;
593 uint16_t nb_txd = NB_TXD;
594 struct rte_eth_dev_info dev_info;
595 struct rte_eth_rxconf rxq_conf;
596 struct rte_eth_txconf txq_conf;
597 struct rte_eth_conf local_port_conf = port_conf;
598
599 /* Initialise device and RX/TX queues */
600 RTE_LOG(INFO, APP, "Initialising port %u ...\n", (unsigned)port);
601 fflush(stdout);
602
603 ret = rte_eth_dev_info_get(port, &dev_info);
604 if (ret != 0)
605 rte_exit(EXIT_FAILURE,
606 "Error during getting device (port %u) info: %s\n",
607 port, strerror(-ret));
608
609 if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)
610 local_port_conf.txmode.offloads |=
611 DEV_TX_OFFLOAD_MBUF_FAST_FREE;
612 ret = rte_eth_dev_configure(port, 1, 1, &local_port_conf);
613 if (ret < 0)
614 rte_exit(EXIT_FAILURE, "Could not configure port%u (%d)\n",
615 (unsigned)port, ret);
616
617 ret = rte_eth_dev_adjust_nb_rx_tx_desc(port, &nb_rxd, &nb_txd);
618 if (ret < 0)
619 rte_exit(EXIT_FAILURE, "Could not adjust number of descriptors "
620 "for port%u (%d)\n", (unsigned)port, ret);
621
622 rxq_conf = dev_info.default_rxconf;
623 rxq_conf.offloads = local_port_conf.rxmode.offloads;
624 ret = rte_eth_rx_queue_setup(port, 0, nb_rxd,
625 rte_eth_dev_socket_id(port), &rxq_conf, pktmbuf_pool);
626 if (ret < 0)
627 rte_exit(EXIT_FAILURE, "Could not setup up RX queue for "
628 "port%u (%d)\n", (unsigned)port, ret);
629
630 txq_conf = dev_info.default_txconf;
631 txq_conf.offloads = local_port_conf.txmode.offloads;
632 ret = rte_eth_tx_queue_setup(port, 0, nb_txd,
633 rte_eth_dev_socket_id(port), &txq_conf);
634 if (ret < 0)
635 rte_exit(EXIT_FAILURE, "Could not setup up TX queue for "
636 "port%u (%d)\n", (unsigned)port, ret);
637
638 ret = rte_eth_dev_start(port);
639 if (ret < 0)
640 rte_exit(EXIT_FAILURE, "Could not start port%u (%d)\n",
641 (unsigned)port, ret);
642
643 if (promiscuous_on) {
644 ret = rte_eth_promiscuous_enable(port);
645 if (ret != 0)
646 rte_exit(EXIT_FAILURE,
647 "Could not enable promiscuous mode for port%u: %s\n",
648 port, rte_strerror(-ret));
649 }
650 }
651
652 /* Check the link status of all ports in up to 9s, and print them finally */
653 static void
654 check_all_ports_link_status(uint32_t port_mask)
655 {
656 #define CHECK_INTERVAL 100 /* 100ms */
657 #define MAX_CHECK_TIME 90 /* 9s (90 * 100ms) in total */
658 uint16_t portid;
659 uint8_t count, all_ports_up, print_flag = 0;
660 struct rte_eth_link link;
661 int ret;
662
663 printf("\nChecking link status\n");
664 fflush(stdout);
665 for (count = 0; count <= MAX_CHECK_TIME; count++) {
666 all_ports_up = 1;
667 RTE_ETH_FOREACH_DEV(portid) {
668 if ((port_mask & (1 << portid)) == 0)
669 continue;
670 memset(&link, 0, sizeof(link));
671 ret = rte_eth_link_get_nowait(portid, &link);
672 if (ret < 0) {
673 all_ports_up = 0;
674 if (print_flag == 1)
675 printf("Port %u link get failed: %s\n",
676 portid, rte_strerror(-ret));
677 continue;
678 }
679 /* print link status if flag set */
680 if (print_flag == 1) {
681 if (link.link_status)
682 printf(
683 "Port%d Link Up - speed %uMbps - %s\n",
684 portid, link.link_speed,
685 (link.link_duplex == ETH_LINK_FULL_DUPLEX) ?
686 ("full-duplex") : ("half-duplex"));
687 else
688 printf("Port %d Link Down\n", portid);
689 continue;
690 }
691 /* clear all_ports_up flag if any link down */
692 if (link.link_status == ETH_LINK_DOWN) {
693 all_ports_up = 0;
694 break;
695 }
696 }
697 /* after finally printing all link status, get out */
698 if (print_flag == 1)
699 break;
700
701 if (all_ports_up == 0) {
702 printf(".");
703 fflush(stdout);
704 rte_delay_ms(CHECK_INTERVAL);
705 }
706
707 /* set the print_flag if all ports up or timeout */
708 if (all_ports_up == 1 || count == (MAX_CHECK_TIME - 1)) {
709 print_flag = 1;
710 printf("done\n");
711 }
712 }
713 }
714
715 static void
716 log_link_state(struct rte_kni *kni, int prev, struct rte_eth_link *link)
717 {
718 if (kni == NULL || link == NULL)
719 return;
720
721 if (prev == ETH_LINK_DOWN && link->link_status == ETH_LINK_UP) {
722 RTE_LOG(INFO, APP, "%s NIC Link is Up %d Mbps %s %s.\n",
723 rte_kni_get_name(kni),
724 link->link_speed,
725 link->link_autoneg ? "(AutoNeg)" : "(Fixed)",
726 link->link_duplex ? "Full Duplex" : "Half Duplex");
727 } else if (prev == ETH_LINK_UP && link->link_status == ETH_LINK_DOWN) {
728 RTE_LOG(INFO, APP, "%s NIC Link is Down.\n",
729 rte_kni_get_name(kni));
730 }
731 }
732
733 /*
734 * Monitor the link status of all ports and update the
735 * corresponding KNI interface(s)
736 */
737 static void *
738 monitor_all_ports_link_status(void *arg)
739 {
740 uint16_t portid;
741 struct rte_eth_link link;
742 unsigned int i;
743 struct kni_port_params **p = kni_port_params_array;
744 int prev;
745 (void) arg;
746 int ret;
747
748 while (monitor_links) {
749 rte_delay_ms(500);
750 RTE_ETH_FOREACH_DEV(portid) {
751 if ((ports_mask & (1 << portid)) == 0)
752 continue;
753 memset(&link, 0, sizeof(link));
754 ret = rte_eth_link_get_nowait(portid, &link);
755 if (ret < 0) {
756 RTE_LOG(ERR, APP,
757 "Get link failed (port %u): %s\n",
758 portid, rte_strerror(-ret));
759 continue;
760 }
761 for (i = 0; i < p[portid]->nb_kni; i++) {
762 prev = rte_kni_update_link(p[portid]->kni[i],
763 link.link_status);
764 log_link_state(p[portid]->kni[i], prev, &link);
765 }
766 }
767 }
768 return NULL;
769 }
770
771 static int
772 kni_change_mtu_(uint16_t port_id, unsigned int new_mtu)
773 {
774 int ret;
775 uint16_t nb_rxd = NB_RXD;
776 uint16_t nb_txd = NB_TXD;
777 struct rte_eth_conf conf;
778 struct rte_eth_dev_info dev_info;
779 struct rte_eth_rxconf rxq_conf;
780 struct rte_eth_txconf txq_conf;
781
782 if (!rte_eth_dev_is_valid_port(port_id)) {
783 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
784 return -EINVAL;
785 }
786
787 RTE_LOG(INFO, APP, "Change MTU of port %d to %u\n", port_id, new_mtu);
788
789 /* Stop specific port */
790 rte_eth_dev_stop(port_id);
791
792 memcpy(&conf, &port_conf, sizeof(conf));
793 /* Set new MTU */
794 if (new_mtu > RTE_ETHER_MAX_LEN)
795 conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME;
796 else
797 conf.rxmode.offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME;
798
799 /* mtu + length of header + length of FCS = max pkt length */
800 conf.rxmode.max_rx_pkt_len = new_mtu + KNI_ENET_HEADER_SIZE +
801 KNI_ENET_FCS_SIZE;
802 ret = rte_eth_dev_configure(port_id, 1, 1, &conf);
803 if (ret < 0) {
804 RTE_LOG(ERR, APP, "Fail to reconfigure port %d\n", port_id);
805 return ret;
806 }
807
808 ret = rte_eth_dev_adjust_nb_rx_tx_desc(port_id, &nb_rxd, &nb_txd);
809 if (ret < 0)
810 rte_exit(EXIT_FAILURE, "Could not adjust number of descriptors "
811 "for port%u (%d)\n", (unsigned int)port_id,
812 ret);
813
814 ret = rte_eth_dev_info_get(port_id, &dev_info);
815 if (ret != 0) {
816 RTE_LOG(ERR, APP,
817 "Error during getting device (port %u) info: %s\n",
818 port_id, strerror(-ret));
819
820 return ret;
821 }
822
823 rxq_conf = dev_info.default_rxconf;
824 rxq_conf.offloads = conf.rxmode.offloads;
825 ret = rte_eth_rx_queue_setup(port_id, 0, nb_rxd,
826 rte_eth_dev_socket_id(port_id), &rxq_conf, pktmbuf_pool);
827 if (ret < 0) {
828 RTE_LOG(ERR, APP, "Fail to setup Rx queue of port %d\n",
829 port_id);
830 return ret;
831 }
832
833 txq_conf = dev_info.default_txconf;
834 txq_conf.offloads = conf.txmode.offloads;
835 ret = rte_eth_tx_queue_setup(port_id, 0, nb_txd,
836 rte_eth_dev_socket_id(port_id), &txq_conf);
837 if (ret < 0) {
838 RTE_LOG(ERR, APP, "Fail to setup Tx queue of port %d\n",
839 port_id);
840 return ret;
841 }
842
843 /* Restart specific port */
844 ret = rte_eth_dev_start(port_id);
845 if (ret < 0) {
846 RTE_LOG(ERR, APP, "Fail to restart port %d\n", port_id);
847 return ret;
848 }
849
850 return 0;
851 }
852
853 /* Callback for request of changing MTU */
854 static int
855 kni_change_mtu(uint16_t port_id, unsigned int new_mtu)
856 {
857 int ret;
858
859 rte_atomic32_inc(&kni_pause);
860 ret = kni_change_mtu_(port_id, new_mtu);
861 rte_atomic32_dec(&kni_pause);
862
863 return ret;
864 }
865
866 /* Callback for request of configuring network interface up/down */
867 static int
868 kni_config_network_interface(uint16_t port_id, uint8_t if_up)
869 {
870 int ret = 0;
871
872 if (!rte_eth_dev_is_valid_port(port_id)) {
873 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
874 return -EINVAL;
875 }
876
877 RTE_LOG(INFO, APP, "Configure network interface of %d %s\n",
878 port_id, if_up ? "up" : "down");
879
880 rte_atomic32_inc(&kni_pause);
881
882 if (if_up != 0) { /* Configure network interface up */
883 rte_eth_dev_stop(port_id);
884 ret = rte_eth_dev_start(port_id);
885 } else /* Configure network interface down */
886 rte_eth_dev_stop(port_id);
887
888 rte_atomic32_dec(&kni_pause);
889
890 if (ret < 0)
891 RTE_LOG(ERR, APP, "Failed to start port %d\n", port_id);
892
893 return ret;
894 }
895
896 static void
897 print_ethaddr(const char *name, struct rte_ether_addr *mac_addr)
898 {
899 char buf[RTE_ETHER_ADDR_FMT_SIZE];
900 rte_ether_format_addr(buf, RTE_ETHER_ADDR_FMT_SIZE, mac_addr);
901 RTE_LOG(INFO, APP, "\t%s%s\n", name, buf);
902 }
903
904 /* Callback for request of configuring mac address */
905 static int
906 kni_config_mac_address(uint16_t port_id, uint8_t mac_addr[])
907 {
908 int ret = 0;
909
910 if (!rte_eth_dev_is_valid_port(port_id)) {
911 RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
912 return -EINVAL;
913 }
914
915 RTE_LOG(INFO, APP, "Configure mac address of %d\n", port_id);
916 print_ethaddr("Address:", (struct rte_ether_addr *)mac_addr);
917
918 ret = rte_eth_dev_default_mac_addr_set(port_id,
919 (struct rte_ether_addr *)mac_addr);
920 if (ret < 0)
921 RTE_LOG(ERR, APP, "Failed to config mac_addr for port %d\n",
922 port_id);
923
924 return ret;
925 }
926
927 static int
928 kni_alloc(uint16_t port_id)
929 {
930 uint8_t i;
931 struct rte_kni *kni;
932 struct rte_kni_conf conf;
933 struct kni_port_params **params = kni_port_params_array;
934 int ret;
935
936 if (port_id >= RTE_MAX_ETHPORTS || !params[port_id])
937 return -1;
938
939 params[port_id]->nb_kni = params[port_id]->nb_lcore_k ?
940 params[port_id]->nb_lcore_k : 1;
941
942 for (i = 0; i < params[port_id]->nb_kni; i++) {
943 /* Clear conf at first */
944 memset(&conf, 0, sizeof(conf));
945 if (params[port_id]->nb_lcore_k) {
946 snprintf(conf.name, RTE_KNI_NAMESIZE,
947 "vEth%u_%u", port_id, i);
948 conf.core_id = params[port_id]->lcore_k[i];
949 conf.force_bind = 1;
950 } else
951 snprintf(conf.name, RTE_KNI_NAMESIZE,
952 "vEth%u", port_id);
953 conf.group_id = port_id;
954 conf.mbuf_size = MAX_PACKET_SZ;
955 /*
956 * The first KNI device associated to a port
957 * is the master, for multiple kernel thread
958 * environment.
959 */
960 if (i == 0) {
961 struct rte_kni_ops ops;
962 struct rte_eth_dev_info dev_info;
963
964 ret = rte_eth_dev_info_get(port_id, &dev_info);
965 if (ret != 0)
966 rte_exit(EXIT_FAILURE,
967 "Error during getting device (port %u) info: %s\n",
968 port_id, strerror(-ret));
969
970 /* Get the interface default mac address */
971 ret = rte_eth_macaddr_get(port_id,
972 (struct rte_ether_addr *)&conf.mac_addr);
973 if (ret != 0)
974 rte_exit(EXIT_FAILURE,
975 "Failed to get MAC address (port %u): %s\n",
976 port_id, rte_strerror(-ret));
977
978 rte_eth_dev_get_mtu(port_id, &conf.mtu);
979
980 conf.min_mtu = dev_info.min_mtu;
981 conf.max_mtu = dev_info.max_mtu;
982
983 memset(&ops, 0, sizeof(ops));
984 ops.port_id = port_id;
985 ops.change_mtu = kni_change_mtu;
986 ops.config_network_if = kni_config_network_interface;
987 ops.config_mac_address = kni_config_mac_address;
988
989 kni = rte_kni_alloc(pktmbuf_pool, &conf, &ops);
990 } else
991 kni = rte_kni_alloc(pktmbuf_pool, &conf, NULL);
992
993 if (!kni)
994 rte_exit(EXIT_FAILURE, "Fail to create kni for "
995 "port: %d\n", port_id);
996 params[port_id]->kni[i] = kni;
997 }
998
999 return 0;
1000 }
1001
1002 static int
1003 kni_free_kni(uint16_t port_id)
1004 {
1005 uint8_t i;
1006 struct kni_port_params **p = kni_port_params_array;
1007
1008 if (port_id >= RTE_MAX_ETHPORTS || !p[port_id])
1009 return -1;
1010
1011 for (i = 0; i < p[port_id]->nb_kni; i++) {
1012 if (rte_kni_release(p[port_id]->kni[i]))
1013 printf("Fail to release kni\n");
1014 p[port_id]->kni[i] = NULL;
1015 }
1016 rte_eth_dev_stop(port_id);
1017
1018 return 0;
1019 }
1020
1021 /* Initialise ports/queues etc. and start main loop on each core */
1022 int
1023 main(int argc, char** argv)
1024 {
1025 int ret;
1026 uint16_t nb_sys_ports, port;
1027 unsigned i;
1028 void *retval;
1029 pthread_t kni_link_tid;
1030 int pid;
1031
1032 /* Associate signal_hanlder function with USR signals */
1033 signal(SIGUSR1, signal_handler);
1034 signal(SIGUSR2, signal_handler);
1035 signal(SIGRTMIN, signal_handler);
1036 signal(SIGINT, signal_handler);
1037 signal(SIGTERM, signal_handler);
1038
1039 /* Initialise EAL */
1040 ret = rte_eal_init(argc, argv);
1041 if (ret < 0)
1042 rte_exit(EXIT_FAILURE, "Could not initialise EAL (%d)\n", ret);
1043 argc -= ret;
1044 argv += ret;
1045
1046 /* Parse application arguments (after the EAL ones) */
1047 ret = parse_args(argc, argv);
1048 if (ret < 0)
1049 rte_exit(EXIT_FAILURE, "Could not parse input parameters\n");
1050
1051 /* Create the mbuf pool */
1052 pktmbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF,
1053 MEMPOOL_CACHE_SZ, 0, MBUF_DATA_SZ, rte_socket_id());
1054 if (pktmbuf_pool == NULL) {
1055 rte_exit(EXIT_FAILURE, "Could not initialise mbuf pool\n");
1056 return -1;
1057 }
1058
1059 /* Get number of ports found in scan */
1060 nb_sys_ports = rte_eth_dev_count_avail();
1061 if (nb_sys_ports == 0)
1062 rte_exit(EXIT_FAILURE, "No supported Ethernet device found\n");
1063
1064 /* Check if the configured port ID is valid */
1065 for (i = 0; i < RTE_MAX_ETHPORTS; i++)
1066 if (kni_port_params_array[i] && !rte_eth_dev_is_valid_port(i))
1067 rte_exit(EXIT_FAILURE, "Configured invalid "
1068 "port ID %u\n", i);
1069
1070 /* Initialize KNI subsystem */
1071 init_kni();
1072
1073 /* Initialise each port */
1074 RTE_ETH_FOREACH_DEV(port) {
1075 /* Skip ports that are not enabled */
1076 if (!(ports_mask & (1 << port)))
1077 continue;
1078 init_port(port);
1079
1080 if (port >= RTE_MAX_ETHPORTS)
1081 rte_exit(EXIT_FAILURE, "Can not use more than "
1082 "%d ports for kni\n", RTE_MAX_ETHPORTS);
1083
1084 kni_alloc(port);
1085 }
1086 check_all_ports_link_status(ports_mask);
1087
1088 pid = getpid();
1089 RTE_LOG(INFO, APP, "========================\n");
1090 RTE_LOG(INFO, APP, "KNI Running\n");
1091 RTE_LOG(INFO, APP, "kill -SIGUSR1 %d\n", pid);
1092 RTE_LOG(INFO, APP, " Show KNI Statistics.\n");
1093 RTE_LOG(INFO, APP, "kill -SIGUSR2 %d\n", pid);
1094 RTE_LOG(INFO, APP, " Zero KNI Statistics.\n");
1095 RTE_LOG(INFO, APP, "========================\n");
1096 fflush(stdout);
1097
1098 ret = rte_ctrl_thread_create(&kni_link_tid,
1099 "KNI link status check", NULL,
1100 monitor_all_ports_link_status, NULL);
1101 if (ret < 0)
1102 rte_exit(EXIT_FAILURE,
1103 "Could not create link status thread!\n");
1104
1105 /* Launch per-lcore function on every lcore */
1106 rte_eal_mp_remote_launch(main_loop, NULL, CALL_MASTER);
1107 RTE_LCORE_FOREACH_SLAVE(i) {
1108 if (rte_eal_wait_lcore(i) < 0)
1109 return -1;
1110 }
1111 monitor_links = 0;
1112 pthread_join(kni_link_tid, &retval);
1113
1114 /* Release resources */
1115 RTE_ETH_FOREACH_DEV(port) {
1116 if (!(ports_mask & (1 << port)))
1117 continue;
1118 kni_free_kni(port);
1119 }
1120 for (i = 0; i < RTE_MAX_ETHPORTS; i++)
1121 if (kni_port_params_array[i]) {
1122 rte_free(kni_port_params_array[i]);
1123 kni_port_params_array[i] = NULL;
1124 }
1125
1126 return 0;
1127 }