]> git.proxmox.com Git - ceph.git/blobdiff - ceph/src/dpdk/examples/load_balancer/init.c
update sources to ceph Nautilus 14.2.1
[ceph.git] / ceph / src / dpdk / examples / load_balancer / init.c
diff --git a/ceph/src/dpdk/examples/load_balancer/init.c b/ceph/src/dpdk/examples/load_balancer/init.c
deleted file mode 100644 (file)
index e07850b..0000000
+++ /dev/null
@@ -1,521 +0,0 @@
-/*-
- *   BSD LICENSE
- *
- *   Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
- *   All rights reserved.
- *
- *   Redistribution and use in source and binary forms, with or without
- *   modification, are permitted provided that the following conditions
- *   are met:
- *
- *     * Redistributions of source code must retain the above copyright
- *       notice, this list of conditions and the following disclaimer.
- *     * Redistributions in binary form must reproduce the above copyright
- *       notice, this list of conditions and the following disclaimer in
- *       the documentation and/or other materials provided with the
- *       distribution.
- *     * Neither the name of Intel Corporation nor the names of its
- *       contributors may be used to endorse or promote products derived
- *       from this software without specific prior written permission.
- *
- *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
- *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- */
-
-#include <stdio.h>
-#include <stdlib.h>
-#include <stdint.h>
-#include <inttypes.h>
-#include <sys/types.h>
-#include <string.h>
-#include <sys/queue.h>
-#include <stdarg.h>
-#include <errno.h>
-#include <getopt.h>
-
-#include <rte_common.h>
-#include <rte_byteorder.h>
-#include <rte_log.h>
-#include <rte_memory.h>
-#include <rte_memcpy.h>
-#include <rte_memzone.h>
-#include <rte_eal.h>
-#include <rte_per_lcore.h>
-#include <rte_launch.h>
-#include <rte_atomic.h>
-#include <rte_cycles.h>
-#include <rte_prefetch.h>
-#include <rte_lcore.h>
-#include <rte_per_lcore.h>
-#include <rte_branch_prediction.h>
-#include <rte_interrupts.h>
-#include <rte_pci.h>
-#include <rte_random.h>
-#include <rte_debug.h>
-#include <rte_ether.h>
-#include <rte_ethdev.h>
-#include <rte_ring.h>
-#include <rte_mempool.h>
-#include <rte_mbuf.h>
-#include <rte_string_fns.h>
-#include <rte_ip.h>
-#include <rte_tcp.h>
-#include <rte_lpm.h>
-
-#include "main.h"
-
-static struct rte_eth_conf port_conf = {
-       .rxmode = {
-               .mq_mode        = ETH_MQ_RX_RSS,
-               .split_hdr_size = 0,
-               .header_split   = 0, /**< Header Split disabled */
-               .hw_ip_checksum = 1, /**< IP checksum offload enabled */
-               .hw_vlan_filter = 0, /**< VLAN filtering disabled */
-               .jumbo_frame    = 0, /**< Jumbo Frame Support disabled */
-               .hw_strip_crc   = 0, /**< CRC stripped by hardware */
-       },
-       .rx_adv_conf = {
-               .rss_conf = {
-                       .rss_key = NULL,
-                       .rss_hf = ETH_RSS_IP,
-               },
-       },
-       .txmode = {
-               .mq_mode = ETH_MQ_TX_NONE,
-       },
-};
-
-static void
-app_assign_worker_ids(void)
-{
-       uint32_t lcore, worker_id;
-
-       /* Assign ID for each worker */
-       worker_id = 0;
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_worker *lp_worker = &app.lcore_params[lcore].worker;
-
-               if (app.lcore_params[lcore].type != e_APP_LCORE_WORKER) {
-                       continue;
-               }
-
-               lp_worker->worker_id = worker_id;
-               worker_id ++;
-       }
-}
-
-static void
-app_init_mbuf_pools(void)
-{
-       unsigned socket, lcore;
-
-       /* Init the buffer pools */
-       for (socket = 0; socket < APP_MAX_SOCKETS; socket ++) {
-               char name[32];
-               if (app_is_socket_used(socket) == 0) {
-                       continue;
-               }
-
-               snprintf(name, sizeof(name), "mbuf_pool_%u", socket);
-               printf("Creating the mbuf pool for socket %u ...\n", socket);
-               app.pools[socket] = rte_pktmbuf_pool_create(
-                       name, APP_DEFAULT_MEMPOOL_BUFFERS,
-                       APP_DEFAULT_MEMPOOL_CACHE_SIZE,
-                       0, APP_DEFAULT_MBUF_DATA_SIZE, socket);
-               if (app.pools[socket] == NULL) {
-                       rte_panic("Cannot create mbuf pool on socket %u\n", socket);
-               }
-       }
-
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               if (app.lcore_params[lcore].type == e_APP_LCORE_DISABLED) {
-                       continue;
-               }
-
-               socket = rte_lcore_to_socket_id(lcore);
-               app.lcore_params[lcore].pool = app.pools[socket];
-       }
-}
-
-static void
-app_init_lpm_tables(void)
-{
-       unsigned socket, lcore;
-
-       /* Init the LPM tables */
-       for (socket = 0; socket < APP_MAX_SOCKETS; socket ++) {
-               char name[32];
-               uint32_t rule;
-
-               if (app_is_socket_used(socket) == 0) {
-                       continue;
-               }
-
-               struct rte_lpm_config lpm_config;
-
-               lpm_config.max_rules = APP_MAX_LPM_RULES;
-               lpm_config.number_tbl8s = 256;
-               lpm_config.flags = 0;
-               snprintf(name, sizeof(name), "lpm_table_%u", socket);
-               printf("Creating the LPM table for socket %u ...\n", socket);
-               app.lpm_tables[socket] = rte_lpm_create(
-                       name,
-                       socket,
-                       &lpm_config);
-               if (app.lpm_tables[socket] == NULL) {
-                       rte_panic("Unable to create LPM table on socket %u\n", socket);
-               }
-
-               for (rule = 0; rule < app.n_lpm_rules; rule ++) {
-                       int ret;
-
-                       ret = rte_lpm_add(app.lpm_tables[socket],
-                               app.lpm_rules[rule].ip,
-                               app.lpm_rules[rule].depth,
-                               app.lpm_rules[rule].if_out);
-
-                       if (ret < 0) {
-                               rte_panic("Unable to add entry %u (%x/%u => %u) to the LPM table on socket %u (%d)\n",
-                                       (unsigned) rule,
-                                       (unsigned) app.lpm_rules[rule].ip,
-                                       (unsigned) app.lpm_rules[rule].depth,
-                                       (unsigned) app.lpm_rules[rule].if_out,
-                                       socket,
-                                       ret);
-                       }
-               }
-
-       }
-
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               if (app.lcore_params[lcore].type != e_APP_LCORE_WORKER) {
-                       continue;
-               }
-
-               socket = rte_lcore_to_socket_id(lcore);
-               app.lcore_params[lcore].worker.lpm_table = app.lpm_tables[socket];
-       }
-}
-
-static void
-app_init_rings_rx(void)
-{
-       unsigned lcore;
-
-       /* Initialize the rings for the RX side */
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_io *lp_io = &app.lcore_params[lcore].io;
-               unsigned socket_io, lcore_worker;
-
-               if ((app.lcore_params[lcore].type != e_APP_LCORE_IO) ||
-                   (lp_io->rx.n_nic_queues == 0)) {
-                       continue;
-               }
-
-               socket_io = rte_lcore_to_socket_id(lcore);
-
-               for (lcore_worker = 0; lcore_worker < APP_MAX_LCORES; lcore_worker ++) {
-                       char name[32];
-                       struct app_lcore_params_worker *lp_worker = &app.lcore_params[lcore_worker].worker;
-                       struct rte_ring *ring = NULL;
-
-                       if (app.lcore_params[lcore_worker].type != e_APP_LCORE_WORKER) {
-                               continue;
-                       }
-
-                       printf("Creating ring to connect I/O lcore %u (socket %u) with worker lcore %u ...\n",
-                               lcore,
-                               socket_io,
-                               lcore_worker);
-                       snprintf(name, sizeof(name), "app_ring_rx_s%u_io%u_w%u",
-                               socket_io,
-                               lcore,
-                               lcore_worker);
-                       ring = rte_ring_create(
-                               name,
-                               app.ring_rx_size,
-                               socket_io,
-                               RING_F_SP_ENQ | RING_F_SC_DEQ);
-                       if (ring == NULL) {
-                               rte_panic("Cannot create ring to connect I/O core %u with worker core %u\n",
-                                       lcore,
-                                       lcore_worker);
-                       }
-
-                       lp_io->rx.rings[lp_io->rx.n_rings] = ring;
-                       lp_io->rx.n_rings ++;
-
-                       lp_worker->rings_in[lp_worker->n_rings_in] = ring;
-                       lp_worker->n_rings_in ++;
-               }
-       }
-
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_io *lp_io = &app.lcore_params[lcore].io;
-
-               if ((app.lcore_params[lcore].type != e_APP_LCORE_IO) ||
-                   (lp_io->rx.n_nic_queues == 0)) {
-                       continue;
-               }
-
-               if (lp_io->rx.n_rings != app_get_lcores_worker()) {
-                       rte_panic("Algorithmic error (I/O RX rings)\n");
-               }
-       }
-
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_worker *lp_worker = &app.lcore_params[lcore].worker;
-
-               if (app.lcore_params[lcore].type != e_APP_LCORE_WORKER) {
-                       continue;
-               }
-
-               if (lp_worker->n_rings_in != app_get_lcores_io_rx()) {
-                       rte_panic("Algorithmic error (worker input rings)\n");
-               }
-       }
-}
-
-static void
-app_init_rings_tx(void)
-{
-       unsigned lcore;
-
-       /* Initialize the rings for the TX side */
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_worker *lp_worker = &app.lcore_params[lcore].worker;
-               unsigned port;
-
-               if (app.lcore_params[lcore].type != e_APP_LCORE_WORKER) {
-                       continue;
-               }
-
-               for (port = 0; port < APP_MAX_NIC_PORTS; port ++) {
-                       char name[32];
-                       struct app_lcore_params_io *lp_io = NULL;
-                       struct rte_ring *ring;
-                       uint32_t socket_io, lcore_io;
-
-                       if (app.nic_tx_port_mask[port] == 0) {
-                               continue;
-                       }
-
-                       if (app_get_lcore_for_nic_tx((uint8_t) port, &lcore_io) < 0) {
-                               rte_panic("Algorithmic error (no I/O core to handle TX of port %u)\n",
-                                       port);
-                       }
-
-                       lp_io = &app.lcore_params[lcore_io].io;
-                       socket_io = rte_lcore_to_socket_id(lcore_io);
-
-                       printf("Creating ring to connect worker lcore %u with TX port %u (through I/O lcore %u) (socket %u) ...\n",
-                               lcore, port, (unsigned)lcore_io, (unsigned)socket_io);
-                       snprintf(name, sizeof(name), "app_ring_tx_s%u_w%u_p%u", socket_io, lcore, port);
-                       ring = rte_ring_create(
-                               name,
-                               app.ring_tx_size,
-                               socket_io,
-                               RING_F_SP_ENQ | RING_F_SC_DEQ);
-                       if (ring == NULL) {
-                               rte_panic("Cannot create ring to connect worker core %u with TX port %u\n",
-                                       lcore,
-                                       port);
-                       }
-
-                       lp_worker->rings_out[port] = ring;
-                       lp_io->tx.rings[port][lp_worker->worker_id] = ring;
-               }
-       }
-
-       for (lcore = 0; lcore < APP_MAX_LCORES; lcore ++) {
-               struct app_lcore_params_io *lp_io = &app.lcore_params[lcore].io;
-               unsigned i;
-
-               if ((app.lcore_params[lcore].type != e_APP_LCORE_IO) ||
-                   (lp_io->tx.n_nic_ports == 0)) {
-                       continue;
-               }
-
-               for (i = 0; i < lp_io->tx.n_nic_ports; i ++){
-                       unsigned port, j;
-
-                       port = lp_io->tx.nic_ports[i];
-                       for (j = 0; j < app_get_lcores_worker(); j ++) {
-                               if (lp_io->tx.rings[port][j] == NULL) {
-                                       rte_panic("Algorithmic error (I/O TX rings)\n");
-                               }
-                       }
-               }
-       }
-}
-
-/* Check the link status of all ports in up to 9s, and print them finally */
-static void
-check_all_ports_link_status(uint8_t port_num, uint32_t port_mask)
-{
-#define CHECK_INTERVAL 100 /* 100ms */
-#define MAX_CHECK_TIME 90 /* 9s (90 * 100ms) in total */
-       uint8_t portid, count, all_ports_up, print_flag = 0;
-       struct rte_eth_link link;
-       uint32_t n_rx_queues, n_tx_queues;
-
-       printf("\nChecking link status");
-       fflush(stdout);
-       for (count = 0; count <= MAX_CHECK_TIME; count++) {
-               all_ports_up = 1;
-               for (portid = 0; portid < port_num; portid++) {
-                       if ((port_mask & (1 << portid)) == 0)
-                               continue;
-                       n_rx_queues = app_get_nic_rx_queues_per_port(portid);
-                       n_tx_queues = app.nic_tx_port_mask[portid];
-                       if ((n_rx_queues == 0) && (n_tx_queues == 0))
-                               continue;
-                       memset(&link, 0, sizeof(link));
-                       rte_eth_link_get_nowait(portid, &link);
-                       /* print link status if flag set */
-                       if (print_flag == 1) {
-                               if (link.link_status)
-                                       printf("Port %d Link Up - speed %u "
-                                               "Mbps - %s\n", (uint8_t)portid,
-                                               (unsigned)link.link_speed,
-                               (link.link_duplex == ETH_LINK_FULL_DUPLEX) ?
-                                       ("full-duplex") : ("half-duplex\n"));
-                               else
-                                       printf("Port %d Link Down\n",
-                                                       (uint8_t)portid);
-                               continue;
-                       }
-                       /* clear all_ports_up flag if any link down */
-                       if (link.link_status == ETH_LINK_DOWN) {
-                               all_ports_up = 0;
-                               break;
-                       }
-               }
-               /* after finally printing all link status, get out */
-               if (print_flag == 1)
-                       break;
-
-               if (all_ports_up == 0) {
-                       printf(".");
-                       fflush(stdout);
-                       rte_delay_ms(CHECK_INTERVAL);
-               }
-
-               /* set the print_flag if all ports up or timeout */
-               if (all_ports_up == 1 || count == (MAX_CHECK_TIME - 1)) {
-                       print_flag = 1;
-                       printf("done\n");
-               }
-       }
-}
-
-static void
-app_init_nics(void)
-{
-       unsigned socket;
-       uint32_t lcore;
-       uint8_t port, queue;
-       int ret;
-       uint32_t n_rx_queues, n_tx_queues;
-
-       /* Init NIC ports and queues, then start the ports */
-       for (port = 0; port < APP_MAX_NIC_PORTS; port ++) {
-               struct rte_mempool *pool;
-
-               n_rx_queues = app_get_nic_rx_queues_per_port(port);
-               n_tx_queues = app.nic_tx_port_mask[port];
-
-               if ((n_rx_queues == 0) && (n_tx_queues == 0)) {
-                       continue;
-               }
-
-               /* Init port */
-               printf("Initializing NIC port %u ...\n", (unsigned) port);
-               ret = rte_eth_dev_configure(
-                       port,
-                       (uint8_t) n_rx_queues,
-                       (uint8_t) n_tx_queues,
-                       &port_conf);
-               if (ret < 0) {
-                       rte_panic("Cannot init NIC port %u (%d)\n", (unsigned) port, ret);
-               }
-               rte_eth_promiscuous_enable(port);
-
-               /* Init RX queues */
-               for (queue = 0; queue < APP_MAX_RX_QUEUES_PER_NIC_PORT; queue ++) {
-                       if (app.nic_rx_queue_mask[port][queue] == 0) {
-                               continue;
-                       }
-
-                       app_get_lcore_for_nic_rx(port, queue, &lcore);
-                       socket = rte_lcore_to_socket_id(lcore);
-                       pool = app.lcore_params[lcore].pool;
-
-                       printf("Initializing NIC port %u RX queue %u ...\n",
-                               (unsigned) port,
-                               (unsigned) queue);
-                       ret = rte_eth_rx_queue_setup(
-                               port,
-                               queue,
-                               (uint16_t) app.nic_rx_ring_size,
-                               socket,
-                               NULL,
-                               pool);
-                       if (ret < 0) {
-                               rte_panic("Cannot init RX queue %u for port %u (%d)\n",
-                                       (unsigned) queue,
-                                       (unsigned) port,
-                                       ret);
-                       }
-               }
-
-               /* Init TX queues */
-               if (app.nic_tx_port_mask[port] == 1) {
-                       app_get_lcore_for_nic_tx(port, &lcore);
-                       socket = rte_lcore_to_socket_id(lcore);
-                       printf("Initializing NIC port %u TX queue 0 ...\n",
-                               (unsigned) port);
-                       ret = rte_eth_tx_queue_setup(
-                               port,
-                               0,
-                               (uint16_t) app.nic_tx_ring_size,
-                               socket,
-                               NULL);
-                       if (ret < 0) {
-                               rte_panic("Cannot init TX queue 0 for port %d (%d)\n",
-                                       port,
-                                       ret);
-                       }
-               }
-
-               /* Start port */
-               ret = rte_eth_dev_start(port);
-               if (ret < 0) {
-                       rte_panic("Cannot start port %d (%d)\n", port, ret);
-               }
-       }
-
-       check_all_ports_link_status(APP_MAX_NIC_PORTS, (~0x0));
-}
-
-void
-app_init(void)
-{
-       app_assign_worker_ids();
-       app_init_mbuf_pools();
-       app_init_lpm_tables();
-       app_init_rings_rx();
-       app_init_rings_tx();
-       app_init_nics();
-
-       printf("Initialization completed.\n");
-}