]> git.proxmox.com Git - ceph.git/blame - ceph/src/seastar/dpdk/examples/quota_watermark/qw/main.c
import 15.2.0 Octopus source
[ceph.git] / ceph / src / seastar / dpdk / examples / quota_watermark / qw / main.c
CommitLineData
9f95a23c
TL
1/* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2010-2014 Intel Corporation
11fdf7f2
TL
3 */
4
5#include <rte_eal.h>
6
7#include <rte_common.h>
8#include <rte_debug.h>
9#include <rte_errno.h>
10#include <rte_ethdev.h>
11#include <rte_launch.h>
12#include <rte_lcore.h>
13#include <rte_log.h>
14#include <rte_mbuf.h>
15#include <rte_ring.h>
16
17#include <rte_byteorder.h>
18
19#include "args.h"
20#include "main.h"
21#include "init.h"
22#include "../include/conf.h"
23
24
25#ifdef QW_SOFTWARE_FC
26#define SEND_PAUSE_FRAME(port_id, duration) send_pause_frame(port_id, duration)
27#else
28#define SEND_PAUSE_FRAME(port_id, duration) do { } while(0)
29#endif
30
31#define ETHER_TYPE_FLOW_CONTROL 0x8808
32
33struct ether_fc_frame {
34 uint16_t opcode;
35 uint16_t param;
36} __attribute__((__packed__));
37
38
39int *quota;
40unsigned int *low_watermark;
41unsigned int *high_watermark;
42
9f95a23c 43uint16_t port_pairs[RTE_MAX_ETHPORTS];
11fdf7f2
TL
44
45struct rte_ring *rings[RTE_MAX_LCORE][RTE_MAX_ETHPORTS];
46struct rte_mempool *mbuf_pool;
47
48
9f95a23c 49static void send_pause_frame(uint16_t port_id, uint16_t duration)
11fdf7f2
TL
50{
51 struct rte_mbuf *mbuf;
52 struct ether_fc_frame *pause_frame;
53 struct ether_hdr *hdr;
54 struct ether_addr mac_addr;
55
56 RTE_LOG_DP(DEBUG, USER1,
57 "Sending PAUSE frame (duration=%d) on port %d\n",
58 duration, port_id);
59
60 /* Get a mbuf from the pool */
61 mbuf = rte_pktmbuf_alloc(mbuf_pool);
62 if (unlikely(mbuf == NULL))
63 return;
64
65 /* Prepare a PAUSE frame */
66 hdr = rte_pktmbuf_mtod(mbuf, struct ether_hdr *);
67 pause_frame = (struct ether_fc_frame *) &hdr[1];
68
69 rte_eth_macaddr_get(port_id, &mac_addr);
70 ether_addr_copy(&mac_addr, &hdr->s_addr);
71
72 void *tmp = &hdr->d_addr.addr_bytes[0];
73 *((uint64_t *)tmp) = 0x010000C28001ULL;
74
75 hdr->ether_type = rte_cpu_to_be_16(ETHER_TYPE_FLOW_CONTROL);
76
77 pause_frame->opcode = rte_cpu_to_be_16(0x0001);
78 pause_frame->param = rte_cpu_to_be_16(duration);
79
80 mbuf->pkt_len = 60;
81 mbuf->data_len = 60;
82
83 rte_eth_tx_burst(port_id, 0, &mbuf, 1);
84}
85
86/**
87 * Get the previous enabled lcore ID
88 *
89 * @param lcore_id
90 * The current lcore ID.
91 * @return
92 * The previous enabled lcore_id or -1 if not found.
93 */
94static unsigned int
95get_previous_lcore_id(unsigned int lcore_id)
96{
97 int i;
98
99 for (i = lcore_id - 1; i >= 0; i--)
100 if (rte_lcore_is_enabled(i))
101 return i;
102
103 return -1;
104}
105
106/**
107 * Get the last enabled lcore ID
108 *
109 * @return
110 * The last enabled lcore_id.
111 */
112static unsigned int
113get_last_lcore_id(void)
114{
115 int i;
116
117 for (i = RTE_MAX_LCORE; i >= 0; i--)
118 if (rte_lcore_is_enabled(i))
119 return i;
120
121 return 0;
122}
123
124static void
125receive_stage(__attribute__((unused)) void *args)
126{
127 int i, ret;
128
9f95a23c 129 uint16_t port_id;
11fdf7f2
TL
130 uint16_t nb_rx_pkts;
131
132 unsigned int lcore_id;
133 unsigned int free;
134
135 struct rte_mbuf *pkts[MAX_PKT_QUOTA];
136 struct rte_ring *ring;
137 enum ring_state ring_state[RTE_MAX_ETHPORTS] = { RING_READY };
138
139 lcore_id = rte_lcore_id();
140
141 RTE_LOG(INFO, USER1,
142 "%s() started on core %u\n", __func__, lcore_id);
143
144 while (1) {
145
146 /* Process each port round robin style */
147 for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
148
149 if (!is_bit_set(port_id, portmask))
150 continue;
151
152 ring = rings[lcore_id][port_id];
153
154 if (ring_state[port_id] != RING_READY) {
155 if (rte_ring_count(ring) > *low_watermark)
156 continue;
157 else
158 ring_state[port_id] = RING_READY;
159 }
160
161 /* Enqueue received packets on the RX ring */
162 nb_rx_pkts = rte_eth_rx_burst(port_id, 0, pkts,
163 (uint16_t) *quota);
164 ret = rte_ring_enqueue_bulk(ring, (void *) pkts,
165 nb_rx_pkts, &free);
166 if (RING_SIZE - free > *high_watermark) {
167 ring_state[port_id] = RING_OVERLOADED;
168 send_pause_frame(port_id, 1337);
169 }
170
171 if (ret == 0) {
172
173 /*
174 * Return mbufs to the pool,
175 * effectively dropping packets
176 */
177 for (i = 0; i < nb_rx_pkts; i++)
178 rte_pktmbuf_free(pkts[i]);
179 }
180 }
181 }
182}
183
9f95a23c 184static int
11fdf7f2
TL
185pipeline_stage(__attribute__((unused)) void *args)
186{
187 int i, ret;
188 int nb_dq_pkts;
189
9f95a23c 190 uint16_t port_id;
11fdf7f2
TL
191
192 unsigned int lcore_id, previous_lcore_id;
193 unsigned int free;
194
195 void *pkts[MAX_PKT_QUOTA];
196 struct rte_ring *rx, *tx;
197 enum ring_state ring_state[RTE_MAX_ETHPORTS] = { RING_READY };
198
199 lcore_id = rte_lcore_id();
200 previous_lcore_id = get_previous_lcore_id(lcore_id);
201
202 RTE_LOG(INFO, USER1,
203 "%s() started on core %u - processing packets from core %u\n",
204 __func__, lcore_id, previous_lcore_id);
205
206 while (1) {
207
208 for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
209
210 if (!is_bit_set(port_id, portmask))
211 continue;
212
213 tx = rings[lcore_id][port_id];
214 rx = rings[previous_lcore_id][port_id];
215
216 if (ring_state[port_id] != RING_READY) {
217 if (rte_ring_count(tx) > *low_watermark)
218 continue;
219 else
220 ring_state[port_id] = RING_READY;
221 }
222
223 /* Dequeue up to quota mbuf from rx */
224 nb_dq_pkts = rte_ring_dequeue_burst(rx, pkts,
225 *quota, NULL);
226 if (unlikely(nb_dq_pkts < 0))
227 continue;
228
229 /* Enqueue them on tx */
230 ret = rte_ring_enqueue_bulk(tx, pkts,
231 nb_dq_pkts, &free);
232 if (RING_SIZE - free > *high_watermark)
233 ring_state[port_id] = RING_OVERLOADED;
234
235 if (ret == 0) {
236
237 /*
238 * Return mbufs to the pool,
239 * effectively dropping packets
240 */
241 for (i = 0; i < nb_dq_pkts; i++)
242 rte_pktmbuf_free(pkts[i]);
243 }
244 }
245 }
9f95a23c
TL
246
247 return 0;
11fdf7f2
TL
248}
249
9f95a23c 250static int
11fdf7f2
TL
251send_stage(__attribute__((unused)) void *args)
252{
253 uint16_t nb_dq_pkts;
254
9f95a23c
TL
255 uint16_t port_id;
256 uint16_t dest_port_id;
11fdf7f2
TL
257
258 unsigned int lcore_id, previous_lcore_id;
259
260 struct rte_ring *tx;
261 struct rte_mbuf *tx_pkts[MAX_PKT_QUOTA];
262
263 lcore_id = rte_lcore_id();
264 previous_lcore_id = get_previous_lcore_id(lcore_id);
265
266 RTE_LOG(INFO, USER1,
267 "%s() started on core %u - processing packets from core %u\n",
268 __func__, lcore_id, previous_lcore_id);
269
270 while (1) {
271
272 /* Process each ring round robin style */
273 for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
274
275 if (!is_bit_set(port_id, portmask))
276 continue;
277
278 dest_port_id = port_pairs[port_id];
279 tx = rings[previous_lcore_id][port_id];
280
281 if (rte_ring_empty(tx))
282 continue;
283
284 /* Dequeue packets from tx and send them */
285 nb_dq_pkts = (uint16_t) rte_ring_dequeue_burst(tx,
286 (void *) tx_pkts, *quota, NULL);
287 rte_eth_tx_burst(dest_port_id, 0, tx_pkts, nb_dq_pkts);
288
289 /* TODO: Check if nb_dq_pkts == nb_tx_pkts? */
290 }
291 }
9f95a23c
TL
292
293 return 0;
11fdf7f2
TL
294}
295
296int
297main(int argc, char **argv)
298{
299 int ret;
300 unsigned int lcore_id, master_lcore_id, last_lcore_id;
301
9f95a23c 302 uint16_t port_id;
11fdf7f2
TL
303
304 rte_log_set_global_level(RTE_LOG_INFO);
305
306 ret = rte_eal_init(argc, argv);
307 if (ret < 0)
308 rte_exit(EXIT_FAILURE, "Cannot initialize EAL\n");
309
310 argc -= ret;
311 argv += ret;
312
313 init_dpdk();
314 setup_shared_variables();
315
316 *quota = 32;
317 *low_watermark = 60 * RING_SIZE / 100;
318
319 last_lcore_id = get_last_lcore_id();
320 master_lcore_id = rte_get_master_lcore();
321
322 /* Parse the application's arguments */
323 ret = parse_qw_args(argc, argv);
324 if (ret < 0)
325 rte_exit(EXIT_FAILURE, "Invalid quota/watermark argument(s)\n");
326
327 /* Create a pool of mbuf to store packets */
328 mbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", MBUF_PER_POOL, 32, 0,
329 MBUF_DATA_SIZE, rte_socket_id());
330 if (mbuf_pool == NULL)
331 rte_panic("%s\n", rte_strerror(rte_errno));
332
333 for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++)
334 if (is_bit_set(port_id, portmask)) {
335 configure_eth_port(port_id);
336 init_ring(master_lcore_id, port_id);
337 }
338
339 pair_ports();
340
341 /*
342 * Start pipeline_connect() on all the available slave lcores
343 * but the last
344 */
345 for (lcore_id = 0 ; lcore_id < last_lcore_id; lcore_id++) {
346 if (rte_lcore_is_enabled(lcore_id) &&
347 lcore_id != master_lcore_id) {
348
349 for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++)
350 if (is_bit_set(port_id, portmask))
351 init_ring(lcore_id, port_id);
352
9f95a23c 353 rte_eal_remote_launch(pipeline_stage,
11fdf7f2
TL
354 NULL, lcore_id);
355 }
356 }
357
358 /* Start send_stage() on the last slave core */
9f95a23c 359 rte_eal_remote_launch(send_stage, NULL, last_lcore_id);
11fdf7f2
TL
360
361 /* Start receive_stage() on the master core */
362 receive_stage(NULL);
363
364 return 0;
365}