]> git.proxmox.com Git - ceph.git/blob - ceph/src/spdk/dpdk/drivers/event/octeontx2/otx2_evdev_adptr.c
update source to Ceph Pacific 16.2.2
[ceph.git] / ceph / src / spdk / dpdk / drivers / event / octeontx2 / otx2_evdev_adptr.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(C) 2019 Marvell International Ltd.
3 */
4
5 #include "otx2_evdev.h"
6
7 int
8 otx2_sso_rx_adapter_caps_get(const struct rte_eventdev *event_dev,
9 const struct rte_eth_dev *eth_dev, uint32_t *caps)
10 {
11 int rc;
12
13 RTE_SET_USED(event_dev);
14 rc = strncmp(eth_dev->device->driver->name, "net_octeontx2", 13);
15 if (rc)
16 *caps = RTE_EVENT_ETH_RX_ADAPTER_SW_CAP;
17 else
18 *caps = RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT |
19 RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ;
20
21 return 0;
22 }
23
24 static inline int
25 sso_rxq_enable(struct otx2_eth_dev *dev, uint16_t qid, uint8_t tt, uint8_t ggrp,
26 uint16_t eth_port_id)
27 {
28 struct otx2_mbox *mbox = dev->mbox;
29 struct nix_aq_enq_req *aq;
30 int rc;
31
32 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
33 aq->qidx = qid;
34 aq->ctype = NIX_AQ_CTYPE_CQ;
35 aq->op = NIX_AQ_INSTOP_WRITE;
36
37 aq->cq.ena = 0;
38 aq->cq.caching = 0;
39
40 otx2_mbox_memset(&aq->cq_mask, 0, sizeof(struct nix_cq_ctx_s));
41 aq->cq_mask.ena = ~(aq->cq_mask.ena);
42 aq->cq_mask.caching = ~(aq->cq_mask.caching);
43
44 rc = otx2_mbox_process(mbox);
45 if (rc < 0) {
46 otx2_err("Failed to disable cq context");
47 goto fail;
48 }
49
50 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
51 aq->qidx = qid;
52 aq->ctype = NIX_AQ_CTYPE_RQ;
53 aq->op = NIX_AQ_INSTOP_WRITE;
54
55 aq->rq.sso_ena = 1;
56 aq->rq.sso_tt = tt;
57 aq->rq.sso_grp = ggrp;
58 aq->rq.ena_wqwd = 1;
59 /* Mbuf Header generation :
60 * > FIRST_SKIP is a super set of WQE_SKIP, dont modify first skip as
61 * it already has data related to mbuf size, headroom, private area.
62 * > Using WQE_SKIP we can directly assign
63 * mbuf = wqe - sizeof(struct mbuf);
64 * so that mbuf header will not have unpredicted values while headroom
65 * and private data starts at the beginning of wqe_data.
66 */
67 aq->rq.wqe_skip = 1;
68 aq->rq.wqe_caching = 1;
69 aq->rq.spb_ena = 0;
70 aq->rq.flow_tagw = 20; /* 20-bits */
71
72 /* Flow Tag calculation :
73 *
74 * rq_tag <31:24> = good/bad_tag<8:0>;
75 * rq_tag <23:0> = [ltag]
76 *
77 * flow_tag_mask<31:0> = (1 << flow_tagw) - 1; <31:20>
78 * tag<31:0> = (~flow_tag_mask & rq_tag) | (flow_tag_mask & flow_tag);
79 *
80 * Setup :
81 * ltag<23:0> = (eth_port_id & 0xF) << 20;
82 * good/bad_tag<8:0> =
83 * ((eth_port_id >> 4) & 0xF) | (RTE_EVENT_TYPE_ETHDEV << 4);
84 *
85 * TAG<31:0> on getwork = <31:28>(RTE_EVENT_TYPE_ETHDEV) |
86 * <27:20> (eth_port_id) | <20:0> [TAG]
87 */
88
89 aq->rq.ltag = (eth_port_id & 0xF) << 20;
90 aq->rq.good_utag = ((eth_port_id >> 4) & 0xF) |
91 (RTE_EVENT_TYPE_ETHDEV << 4);
92 aq->rq.bad_utag = aq->rq.good_utag;
93
94 aq->rq.ena = 0; /* Don't enable RQ yet */
95 aq->rq.pb_caching = 0x2; /* First cache aligned block to LLC */
96 aq->rq.xqe_imm_size = 0; /* No pkt data copy to CQE */
97
98 otx2_mbox_memset(&aq->rq_mask, 0, sizeof(struct nix_rq_ctx_s));
99 /* mask the bits to write. */
100 aq->rq_mask.sso_ena = ~(aq->rq_mask.sso_ena);
101 aq->rq_mask.sso_tt = ~(aq->rq_mask.sso_tt);
102 aq->rq_mask.sso_grp = ~(aq->rq_mask.sso_grp);
103 aq->rq_mask.ena_wqwd = ~(aq->rq_mask.ena_wqwd);
104 aq->rq_mask.wqe_skip = ~(aq->rq_mask.wqe_skip);
105 aq->rq_mask.wqe_caching = ~(aq->rq_mask.wqe_caching);
106 aq->rq_mask.spb_ena = ~(aq->rq_mask.spb_ena);
107 aq->rq_mask.flow_tagw = ~(aq->rq_mask.flow_tagw);
108 aq->rq_mask.ltag = ~(aq->rq_mask.ltag);
109 aq->rq_mask.good_utag = ~(aq->rq_mask.good_utag);
110 aq->rq_mask.bad_utag = ~(aq->rq_mask.bad_utag);
111 aq->rq_mask.ena = ~(aq->rq_mask.ena);
112 aq->rq_mask.pb_caching = ~(aq->rq_mask.pb_caching);
113 aq->rq_mask.xqe_imm_size = ~(aq->rq_mask.xqe_imm_size);
114
115 rc = otx2_mbox_process(mbox);
116 if (rc < 0) {
117 otx2_err("Failed to init rx adapter context");
118 goto fail;
119 }
120
121 return 0;
122 fail:
123 return rc;
124 }
125
126 static inline int
127 sso_rxq_disable(struct otx2_eth_dev *dev, uint16_t qid)
128 {
129 struct otx2_mbox *mbox = dev->mbox;
130 struct nix_aq_enq_req *aq;
131 int rc;
132
133 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
134 aq->qidx = qid;
135 aq->ctype = NIX_AQ_CTYPE_CQ;
136 aq->op = NIX_AQ_INSTOP_WRITE;
137
138 aq->cq.ena = 1;
139 aq->cq.caching = 1;
140
141 otx2_mbox_memset(&aq->cq_mask, 0, sizeof(struct nix_cq_ctx_s));
142 aq->cq_mask.ena = ~(aq->cq_mask.ena);
143 aq->cq_mask.caching = ~(aq->cq_mask.caching);
144
145 rc = otx2_mbox_process(mbox);
146 if (rc < 0) {
147 otx2_err("Failed to enable cq context");
148 goto fail;
149 }
150
151 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
152 aq->qidx = qid;
153 aq->ctype = NIX_AQ_CTYPE_RQ;
154 aq->op = NIX_AQ_INSTOP_WRITE;
155
156 aq->rq.sso_ena = 0;
157 aq->rq.sso_tt = SSO_TT_UNTAGGED;
158 aq->rq.sso_grp = 0;
159 aq->rq.ena_wqwd = 0;
160 aq->rq.wqe_caching = 0;
161 aq->rq.wqe_skip = 0;
162 aq->rq.spb_ena = 0;
163 aq->rq.flow_tagw = 0x20;
164 aq->rq.ltag = 0;
165 aq->rq.good_utag = 0;
166 aq->rq.bad_utag = 0;
167 aq->rq.ena = 1;
168 aq->rq.pb_caching = 0x2; /* First cache aligned block to LLC */
169 aq->rq.xqe_imm_size = 0; /* No pkt data copy to CQE */
170
171 otx2_mbox_memset(&aq->rq_mask, 0, sizeof(struct nix_rq_ctx_s));
172 /* mask the bits to write. */
173 aq->rq_mask.sso_ena = ~(aq->rq_mask.sso_ena);
174 aq->rq_mask.sso_tt = ~(aq->rq_mask.sso_tt);
175 aq->rq_mask.sso_grp = ~(aq->rq_mask.sso_grp);
176 aq->rq_mask.ena_wqwd = ~(aq->rq_mask.ena_wqwd);
177 aq->rq_mask.wqe_caching = ~(aq->rq_mask.wqe_caching);
178 aq->rq_mask.wqe_skip = ~(aq->rq_mask.wqe_skip);
179 aq->rq_mask.spb_ena = ~(aq->rq_mask.spb_ena);
180 aq->rq_mask.flow_tagw = ~(aq->rq_mask.flow_tagw);
181 aq->rq_mask.ltag = ~(aq->rq_mask.ltag);
182 aq->rq_mask.good_utag = ~(aq->rq_mask.good_utag);
183 aq->rq_mask.bad_utag = ~(aq->rq_mask.bad_utag);
184 aq->rq_mask.ena = ~(aq->rq_mask.ena);
185 aq->rq_mask.pb_caching = ~(aq->rq_mask.pb_caching);
186 aq->rq_mask.xqe_imm_size = ~(aq->rq_mask.xqe_imm_size);
187
188 rc = otx2_mbox_process(mbox);
189 if (rc < 0) {
190 otx2_err("Failed to clear rx adapter context");
191 goto fail;
192 }
193
194 return 0;
195 fail:
196 return rc;
197 }
198
199 void
200 sso_updt_xae_cnt(struct otx2_sso_evdev *dev, void *data, uint32_t event_type)
201 {
202 int i;
203
204 switch (event_type) {
205 case RTE_EVENT_TYPE_ETHDEV:
206 {
207 struct otx2_eth_rxq *rxq = data;
208 uint64_t *old_ptr;
209
210 for (i = 0; i < dev->rx_adptr_pool_cnt; i++) {
211 if ((uint64_t)rxq->pool == dev->rx_adptr_pools[i])
212 return;
213 }
214
215 dev->rx_adptr_pool_cnt++;
216 old_ptr = dev->rx_adptr_pools;
217 dev->rx_adptr_pools = rte_realloc(dev->rx_adptr_pools,
218 sizeof(uint64_t) *
219 dev->rx_adptr_pool_cnt, 0);
220 if (dev->rx_adptr_pools == NULL) {
221 dev->adptr_xae_cnt += rxq->pool->size;
222 dev->rx_adptr_pools = old_ptr;
223 dev->rx_adptr_pool_cnt--;
224 return;
225 }
226 dev->rx_adptr_pools[dev->rx_adptr_pool_cnt - 1] =
227 (uint64_t)rxq->pool;
228
229 dev->adptr_xae_cnt += rxq->pool->size;
230 break;
231 }
232 case RTE_EVENT_TYPE_TIMER:
233 {
234 struct otx2_tim_ring *timr = data;
235 uint16_t *old_ring_ptr;
236 uint64_t *old_sz_ptr;
237
238 for (i = 0; i < dev->tim_adptr_ring_cnt; i++) {
239 if (timr->ring_id != dev->timer_adptr_rings[i])
240 continue;
241 if (timr->nb_timers == dev->timer_adptr_sz[i])
242 return;
243 dev->adptr_xae_cnt -= dev->timer_adptr_sz[i];
244 dev->adptr_xae_cnt += timr->nb_timers;
245 dev->timer_adptr_sz[i] = timr->nb_timers;
246
247 return;
248 }
249
250 dev->tim_adptr_ring_cnt++;
251 old_ring_ptr = dev->timer_adptr_rings;
252 old_sz_ptr = dev->timer_adptr_sz;
253
254 dev->timer_adptr_rings = rte_realloc(dev->timer_adptr_rings,
255 sizeof(uint16_t) *
256 dev->tim_adptr_ring_cnt,
257 0);
258 if (dev->timer_adptr_rings == NULL) {
259 dev->adptr_xae_cnt += timr->nb_timers;
260 dev->timer_adptr_rings = old_ring_ptr;
261 dev->tim_adptr_ring_cnt--;
262 return;
263 }
264
265 dev->timer_adptr_sz = rte_realloc(dev->timer_adptr_sz,
266 sizeof(uint64_t) *
267 dev->tim_adptr_ring_cnt,
268 0);
269
270 if (dev->timer_adptr_sz == NULL) {
271 dev->adptr_xae_cnt += timr->nb_timers;
272 dev->timer_adptr_sz = old_sz_ptr;
273 dev->tim_adptr_ring_cnt--;
274 return;
275 }
276
277 dev->timer_adptr_rings[dev->tim_adptr_ring_cnt - 1] =
278 timr->ring_id;
279 dev->timer_adptr_sz[dev->tim_adptr_ring_cnt - 1] =
280 timr->nb_timers;
281
282 dev->adptr_xae_cnt += timr->nb_timers;
283 break;
284 }
285 default:
286 break;
287 }
288 }
289
290 static inline void
291 sso_updt_lookup_mem(const struct rte_eventdev *event_dev, void *lookup_mem)
292 {
293 struct otx2_sso_evdev *dev = sso_pmd_priv(event_dev);
294 int i;
295
296 for (i = 0; i < dev->nb_event_ports; i++) {
297 if (dev->dual_ws) {
298 struct otx2_ssogws_dual *ws = event_dev->data->ports[i];
299
300 ws->lookup_mem = lookup_mem;
301 } else {
302 struct otx2_ssogws *ws = event_dev->data->ports[i];
303
304 ws->lookup_mem = lookup_mem;
305 }
306 }
307 }
308
309 int
310 otx2_sso_rx_adapter_queue_add(const struct rte_eventdev *event_dev,
311 const struct rte_eth_dev *eth_dev,
312 int32_t rx_queue_id,
313 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf)
314 {
315 struct otx2_eth_dev *otx2_eth_dev = eth_dev->data->dev_private;
316 struct otx2_sso_evdev *dev = sso_pmd_priv(event_dev);
317 uint16_t port = eth_dev->data->port_id;
318 struct otx2_eth_rxq *rxq;
319 int i, rc;
320
321 rc = strncmp(eth_dev->device->driver->name, "net_octeontx2", 13);
322 if (rc)
323 return -EINVAL;
324
325 if (rx_queue_id < 0) {
326 for (i = 0 ; i < eth_dev->data->nb_rx_queues; i++) {
327 rxq = eth_dev->data->rx_queues[i];
328 sso_updt_xae_cnt(dev, rxq, RTE_EVENT_TYPE_ETHDEV);
329 rc = sso_xae_reconfigure((struct rte_eventdev *)
330 (uintptr_t)event_dev);
331 rc |= sso_rxq_enable(otx2_eth_dev, i,
332 queue_conf->ev.sched_type,
333 queue_conf->ev.queue_id, port);
334 }
335 rxq = eth_dev->data->rx_queues[0];
336 sso_updt_lookup_mem(event_dev, rxq->lookup_mem);
337 } else {
338 rxq = eth_dev->data->rx_queues[rx_queue_id];
339 sso_updt_xae_cnt(dev, rxq, RTE_EVENT_TYPE_ETHDEV);
340 rc = sso_xae_reconfigure((struct rte_eventdev *)
341 (uintptr_t)event_dev);
342 rc |= sso_rxq_enable(otx2_eth_dev, (uint16_t)rx_queue_id,
343 queue_conf->ev.sched_type,
344 queue_conf->ev.queue_id, port);
345 sso_updt_lookup_mem(event_dev, rxq->lookup_mem);
346 }
347
348 if (rc < 0) {
349 otx2_err("Failed to configure Rx adapter port=%d, q=%d", port,
350 queue_conf->ev.queue_id);
351 return rc;
352 }
353
354 dev->rx_offloads |= otx2_eth_dev->rx_offload_flags;
355 dev->tstamp = &otx2_eth_dev->tstamp;
356 sso_fastpath_fns_set((struct rte_eventdev *)(uintptr_t)event_dev);
357
358 return 0;
359 }
360
361 int
362 otx2_sso_rx_adapter_queue_del(const struct rte_eventdev *event_dev,
363 const struct rte_eth_dev *eth_dev,
364 int32_t rx_queue_id)
365 {
366 struct otx2_eth_dev *dev = eth_dev->data->dev_private;
367 int i, rc;
368
369 RTE_SET_USED(event_dev);
370 rc = strncmp(eth_dev->device->driver->name, "net_octeontx2", 13);
371 if (rc)
372 return -EINVAL;
373
374 if (rx_queue_id < 0) {
375 for (i = 0 ; i < eth_dev->data->nb_rx_queues; i++)
376 rc = sso_rxq_disable(dev, i);
377 } else {
378 rc = sso_rxq_disable(dev, (uint16_t)rx_queue_id);
379 }
380
381 if (rc < 0)
382 otx2_err("Failed to clear Rx adapter config port=%d, q=%d",
383 eth_dev->data->port_id, rx_queue_id);
384
385 return rc;
386 }
387
388 int
389 otx2_sso_rx_adapter_start(const struct rte_eventdev *event_dev,
390 const struct rte_eth_dev *eth_dev)
391 {
392 RTE_SET_USED(event_dev);
393 RTE_SET_USED(eth_dev);
394
395 return 0;
396 }
397
398 int
399 otx2_sso_rx_adapter_stop(const struct rte_eventdev *event_dev,
400 const struct rte_eth_dev *eth_dev)
401 {
402 RTE_SET_USED(event_dev);
403 RTE_SET_USED(eth_dev);
404
405 return 0;
406 }
407
408 int
409 otx2_sso_tx_adapter_caps_get(const struct rte_eventdev *dev,
410 const struct rte_eth_dev *eth_dev, uint32_t *caps)
411 {
412 int ret;
413
414 RTE_SET_USED(dev);
415 ret = strncmp(eth_dev->device->driver->name, "net_octeontx2,", 13);
416 if (ret)
417 *caps = 0;
418 else
419 *caps = RTE_EVENT_ETH_TX_ADAPTER_CAP_INTERNAL_PORT;
420
421 return 0;
422 }
423
424 static int
425 sso_sqb_aura_limit_edit(struct rte_mempool *mp, uint16_t nb_sqb_bufs)
426 {
427 struct otx2_npa_lf *npa_lf = otx2_intra_dev_get_cfg()->npa_lf;
428 struct npa_aq_enq_req *aura_req;
429
430 aura_req = otx2_mbox_alloc_msg_npa_aq_enq(npa_lf->mbox);
431 aura_req->aura_id = npa_lf_aura_handle_to_aura(mp->pool_id);
432 aura_req->ctype = NPA_AQ_CTYPE_AURA;
433 aura_req->op = NPA_AQ_INSTOP_WRITE;
434
435 aura_req->aura.limit = nb_sqb_bufs;
436 aura_req->aura_mask.limit = ~(aura_req->aura_mask.limit);
437
438 return otx2_mbox_process(npa_lf->mbox);
439 }
440
441 int
442 otx2_sso_tx_adapter_queue_add(uint8_t id, const struct rte_eventdev *event_dev,
443 const struct rte_eth_dev *eth_dev,
444 int32_t tx_queue_id)
445 {
446 struct otx2_eth_dev *otx2_eth_dev = eth_dev->data->dev_private;
447 struct otx2_sso_evdev *dev = sso_pmd_priv(event_dev);
448 struct otx2_eth_txq *txq;
449 int i;
450
451 RTE_SET_USED(id);
452 if (tx_queue_id < 0) {
453 for (i = 0 ; i < eth_dev->data->nb_tx_queues; i++) {
454 txq = eth_dev->data->tx_queues[i];
455 sso_sqb_aura_limit_edit(txq->sqb_pool,
456 OTX2_SSO_SQB_LIMIT);
457 }
458 } else {
459 txq = eth_dev->data->tx_queues[tx_queue_id];
460 sso_sqb_aura_limit_edit(txq->sqb_pool, OTX2_SSO_SQB_LIMIT);
461 }
462
463 dev->tx_offloads |= otx2_eth_dev->tx_offload_flags;
464 sso_fastpath_fns_set((struct rte_eventdev *)(uintptr_t)event_dev);
465
466 return 0;
467 }
468
469 int
470 otx2_sso_tx_adapter_queue_del(uint8_t id, const struct rte_eventdev *event_dev,
471 const struct rte_eth_dev *eth_dev,
472 int32_t tx_queue_id)
473 {
474 struct otx2_eth_txq *txq;
475 int i;
476
477 RTE_SET_USED(id);
478 RTE_SET_USED(eth_dev);
479 RTE_SET_USED(event_dev);
480 if (tx_queue_id < 0) {
481 for (i = 0 ; i < eth_dev->data->nb_tx_queues; i++) {
482 txq = eth_dev->data->tx_queues[i];
483 sso_sqb_aura_limit_edit(txq->sqb_pool,
484 txq->nb_sqb_bufs);
485 }
486 } else {
487 txq = eth_dev->data->tx_queues[tx_queue_id];
488 sso_sqb_aura_limit_edit(txq->sqb_pool, txq->nb_sqb_bufs);
489 }
490
491 return 0;
492 }