]>
git.proxmox.com Git - ceph.git/blob - ceph/src/spdk/dpdk/drivers/net/thunderx/nicvf_ethdev.h
1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2016 Cavium, Inc
5 #ifndef __THUNDERX_NICVF_ETHDEV_H__
6 #define __THUNDERX_NICVF_ETHDEV_H__
8 #include <rte_ethdev_driver.h>
10 #define THUNDERX_NICVF_PMD_VERSION "2.0"
11 #define THUNDERX_REG_BYTES 8
13 #define NICVF_INTR_POLL_INTERVAL_MS 50
14 #define NICVF_HALF_DUPLEX 0x00
15 #define NICVF_FULL_DUPLEX 0x01
16 #define NICVF_UNKNOWN_DUPLEX 0xff
18 #define NICVF_RSS_OFFLOAD_PASS1 ( \
21 ETH_RSS_NONFRAG_IPV4_TCP | \
22 ETH_RSS_NONFRAG_IPV4_UDP | \
24 ETH_RSS_NONFRAG_IPV6_TCP | \
25 ETH_RSS_NONFRAG_IPV6_UDP)
27 #define NICVF_RSS_OFFLOAD_TUNNEL ( \
32 #define NICVF_TX_OFFLOAD_CAPA ( \
33 DEV_TX_OFFLOAD_IPV4_CKSUM | \
34 DEV_TX_OFFLOAD_UDP_CKSUM | \
35 DEV_TX_OFFLOAD_TCP_CKSUM | \
36 DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM | \
37 DEV_TX_OFFLOAD_MBUF_FAST_FREE | \
38 DEV_TX_OFFLOAD_MULTI_SEGS)
40 #define NICVF_RX_OFFLOAD_CAPA ( \
41 DEV_RX_OFFLOAD_CHECKSUM | \
42 DEV_RX_OFFLOAD_VLAN_STRIP | \
43 DEV_RX_OFFLOAD_JUMBO_FRAME | \
44 DEV_RX_OFFLOAD_SCATTER | \
45 DEV_RX_OFFLOAD_RSS_HASH)
47 #define NICVF_DEFAULT_RX_FREE_THRESH 224
48 #define NICVF_DEFAULT_TX_FREE_THRESH 224
49 #define NICVF_TX_FREE_MPOOL_THRESH 16
50 #define NICVF_MAX_RX_FREE_THRESH 1024
51 #define NICVF_MAX_TX_FREE_THRESH 1024
53 #define VLAN_TAG_SIZE 4 /* 802.3ac tag */
55 #define SKIP_DATA_BYTES "skip_data_bytes"
56 static inline struct nicvf
*
57 nicvf_pmd_priv(struct rte_eth_dev
*eth_dev
)
59 return eth_dev
->data
->dev_private
;
62 static inline uint64_t
63 nicvf_mempool_phy_offset(struct rte_mempool
*mp
)
65 struct rte_mempool_memhdr
*hdr
;
67 hdr
= STAILQ_FIRST(&mp
->mem_list
);
69 return (uint64_t)((uintptr_t)hdr
->addr
- hdr
->iova
);
72 static inline uint16_t
73 nicvf_mbuff_meta_length(struct rte_mbuf
*mbuf
)
75 return (uint16_t)((uintptr_t)mbuf
->buf_addr
- (uintptr_t)mbuf
);
78 static inline uint16_t
79 nicvf_netdev_qidx(struct nicvf
*nic
, uint8_t local_qidx
)
81 uint16_t global_qidx
= local_qidx
;
84 global_qidx
+= ((nic
->sqs_id
+ 1) * MAX_CMP_QUEUES_PER_QS
);
90 * Simple phy2virt functions assuming mbufs are in a single huge page
94 static inline uintptr_t
95 nicvf_mbuff_phy2virt(rte_iova_t phy
, uint64_t mbuf_phys_off
)
97 return (uintptr_t)(phy
+ mbuf_phys_off
);
100 static inline uintptr_t
101 nicvf_mbuff_virt2phy(uintptr_t virt
, uint64_t mbuf_phys_off
)
103 return (rte_iova_t
)(virt
- mbuf_phys_off
);
107 nicvf_tx_range(struct rte_eth_dev
*dev
, struct nicvf
*nic
, uint16_t *tx_start
,
112 *tx_start
= RTE_ALIGN_FLOOR(nicvf_netdev_qidx(nic
, 0),
113 MAX_SND_QUEUES_PER_QS
);
114 tmp
= RTE_ALIGN_CEIL(nicvf_netdev_qidx(nic
, 0) + 1,
115 MAX_SND_QUEUES_PER_QS
) - 1;
116 *tx_end
= dev
->data
->nb_tx_queues
?
117 RTE_MIN(tmp
, dev
->data
->nb_tx_queues
- 1) : 0;
121 nicvf_rx_range(struct rte_eth_dev
*dev
, struct nicvf
*nic
, uint16_t *rx_start
,
126 *rx_start
= RTE_ALIGN_FLOOR(nicvf_netdev_qidx(nic
, 0),
127 MAX_RCV_QUEUES_PER_QS
);
128 tmp
= RTE_ALIGN_CEIL(nicvf_netdev_qidx(nic
, 0) + 1,
129 MAX_RCV_QUEUES_PER_QS
) - 1;
130 *rx_end
= dev
->data
->nb_rx_queues
?
131 RTE_MIN(tmp
, dev
->data
->nb_rx_queues
- 1) : 0;
134 #endif /* __THUNDERX_NICVF_ETHDEV_H__ */