]> git.proxmox.com Git - mirror_ovs.git/blob - lib/dp-packet.h
flow: Support extra padding length.
[mirror_ovs.git] / lib / dp-packet.h
1 /*
2 * Copyright (c) 2008, 2009, 2010, 2011, 2012, 2013 Nicira, Inc.
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at:
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #ifndef DPBUF_H
18 #define DPBUF_H 1
19
20 #include <stddef.h>
21 #include <stdint.h>
22
23 #ifdef DPDK_NETDEV
24 #include <rte_config.h>
25 #include <rte_mbuf.h>
26 #endif
27
28 #include "netdev-afxdp.h"
29 #include "netdev-dpdk.h"
30 #include "openvswitch/list.h"
31 #include "packets.h"
32 #include "util.h"
33 #include "flow.h"
34
35 #ifdef __cplusplus
36 extern "C" {
37 #endif
38
39 enum OVS_PACKED_ENUM dp_packet_source {
40 DPBUF_MALLOC, /* Obtained via malloc(). */
41 DPBUF_STACK, /* Un-movable stack space or static buffer. */
42 DPBUF_STUB, /* Starts on stack, may expand into heap. */
43 DPBUF_DPDK, /* buffer data is from DPDK allocated memory.
44 * ref to dp_packet_init_dpdk() in dp-packet.c.
45 */
46 DPBUF_AFXDP, /* Buffer data from XDP frame. */
47 };
48
49 #define DP_PACKET_CONTEXT_SIZE 64
50
51 #ifdef DPDK_NETDEV
52 #define DEF_OL_FLAG(NAME, DPDK_DEF, GENERIC_DEF) NAME = DPDK_DEF
53 #else
54 #define DEF_OL_FLAG(NAME, DPDK_DEF, GENERIC_DEF) NAME = GENERIC_DEF
55 #endif
56
57 /* Bit masks for the 'ol_flags' member of the 'dp_packet' structure. */
58 enum dp_packet_offload_mask {
59 /* Value 0 is not used. */
60 /* Is the 'rss_hash' valid? */
61 DEF_OL_FLAG(DP_PACKET_OL_RSS_HASH, PKT_RX_RSS_HASH, 0x1),
62 /* Is the 'flow_mark' valid? */
63 DEF_OL_FLAG(DP_PACKET_OL_FLOW_MARK, PKT_RX_FDIR_ID, 0x2),
64 /* Bad L4 checksum in the packet. */
65 DEF_OL_FLAG(DP_PACKET_OL_RX_L4_CKSUM_BAD, PKT_RX_L4_CKSUM_BAD, 0x4),
66 /* Bad IP checksum in the packet. */
67 DEF_OL_FLAG(DP_PACKET_OL_RX_IP_CKSUM_BAD, PKT_RX_IP_CKSUM_BAD, 0x8),
68 /* Valid L4 checksum in the packet. */
69 DEF_OL_FLAG(DP_PACKET_OL_RX_L4_CKSUM_GOOD, PKT_RX_L4_CKSUM_GOOD, 0x10),
70 /* Valid IP checksum in the packet. */
71 DEF_OL_FLAG(DP_PACKET_OL_RX_IP_CKSUM_GOOD, PKT_RX_IP_CKSUM_GOOD, 0x20),
72 /* TCP Segmentation Offload. */
73 DEF_OL_FLAG(DP_PACKET_OL_TX_TCP_SEG, PKT_TX_TCP_SEG, 0x40),
74 /* Offloaded packet is IPv4. */
75 DEF_OL_FLAG(DP_PACKET_OL_TX_IPV4, PKT_TX_IPV4, 0x80),
76 /* Offloaded packet is IPv6. */
77 DEF_OL_FLAG(DP_PACKET_OL_TX_IPV6, PKT_TX_IPV6, 0x100),
78 /* Offload TCP checksum. */
79 DEF_OL_FLAG(DP_PACKET_OL_TX_TCP_CKSUM, PKT_TX_TCP_CKSUM, 0x200),
80 /* Offload UDP checksum. */
81 DEF_OL_FLAG(DP_PACKET_OL_TX_UDP_CKSUM, PKT_TX_UDP_CKSUM, 0x400),
82 /* Offload SCTP checksum. */
83 DEF_OL_FLAG(DP_PACKET_OL_TX_SCTP_CKSUM, PKT_TX_SCTP_CKSUM, 0x800),
84 /* Adding new field requires adding to DP_PACKET_OL_SUPPORTED_MASK. */
85 };
86
87 #define DP_PACKET_OL_SUPPORTED_MASK (DP_PACKET_OL_RSS_HASH | \
88 DP_PACKET_OL_FLOW_MARK | \
89 DP_PACKET_OL_RX_L4_CKSUM_BAD | \
90 DP_PACKET_OL_RX_IP_CKSUM_BAD | \
91 DP_PACKET_OL_RX_L4_CKSUM_GOOD | \
92 DP_PACKET_OL_RX_IP_CKSUM_GOOD | \
93 DP_PACKET_OL_TX_TCP_SEG | \
94 DP_PACKET_OL_TX_IPV4 | \
95 DP_PACKET_OL_TX_IPV6 | \
96 DP_PACKET_OL_TX_TCP_CKSUM | \
97 DP_PACKET_OL_TX_UDP_CKSUM | \
98 DP_PACKET_OL_TX_SCTP_CKSUM)
99
100 #define DP_PACKET_OL_TX_L4_MASK (DP_PACKET_OL_TX_TCP_CKSUM | \
101 DP_PACKET_OL_TX_UDP_CKSUM | \
102 DP_PACKET_OL_TX_SCTP_CKSUM)
103 #define DP_PACKET_OL_RX_IP_CKSUM_MASK (DP_PACKET_OL_RX_IP_CKSUM_GOOD | \
104 DP_PACKET_OL_RX_IP_CKSUM_BAD)
105 #define DP_PACKET_OL_RX_L4_CKSUM_MASK (DP_PACKET_OL_RX_L4_CKSUM_GOOD | \
106 DP_PACKET_OL_RX_L4_CKSUM_BAD)
107
108 /* Buffer for holding packet data. A dp_packet is automatically reallocated
109 * as necessary if it grows too large for the available memory.
110 * By default the packet type is set to Ethernet (PT_ETH).
111 */
112 struct dp_packet {
113 #ifdef DPDK_NETDEV
114 struct rte_mbuf mbuf; /* DPDK mbuf */
115 #else
116 void *base_; /* First byte of allocated space. */
117 uint16_t allocated_; /* Number of bytes allocated. */
118 uint16_t data_ofs; /* First byte actually in use. */
119 uint32_t size_; /* Number of bytes in use. */
120 uint32_t ol_flags; /* Offloading flags. */
121 uint32_t rss_hash; /* Packet hash. */
122 uint32_t flow_mark; /* Packet flow mark. */
123 #endif
124 enum dp_packet_source source; /* Source of memory allocated as 'base'. */
125
126 /* All the following elements of this struct are copied in a single call
127 * of memcpy in dp_packet_clone_with_headroom. */
128 uint16_t l2_pad_size; /* Detected l2 padding size.
129 * Padding is non-pullable. */
130 uint16_t l2_5_ofs; /* MPLS label stack offset, or UINT16_MAX */
131 uint16_t l3_ofs; /* Network-level header offset,
132 * or UINT16_MAX. */
133 uint16_t l4_ofs; /* Transport-level header offset,
134 or UINT16_MAX. */
135 uint32_t cutlen; /* length in bytes to cut from the end. */
136 ovs_be32 packet_type; /* Packet type as defined in OpenFlow */
137 union {
138 struct pkt_metadata md;
139 uint64_t data[DP_PACKET_CONTEXT_SIZE / 8];
140 };
141 };
142
143 #if HAVE_AF_XDP
144 struct dp_packet_afxdp {
145 struct umem_pool *mpool;
146 struct dp_packet packet;
147 };
148 #endif
149
150 static inline void *dp_packet_data(const struct dp_packet *);
151 static inline void dp_packet_set_data(struct dp_packet *, void *);
152 static inline void *dp_packet_base(const struct dp_packet *);
153 static inline void dp_packet_set_base(struct dp_packet *, void *);
154
155 static inline uint32_t dp_packet_size(const struct dp_packet *);
156 static inline void dp_packet_set_size(struct dp_packet *, uint32_t);
157
158 static inline uint16_t dp_packet_get_allocated(const struct dp_packet *);
159 static inline void dp_packet_set_allocated(struct dp_packet *, uint16_t);
160
161 void *dp_packet_resize_l2(struct dp_packet *, int increment);
162 void *dp_packet_resize_l2_5(struct dp_packet *, int increment);
163 static inline void *dp_packet_eth(const struct dp_packet *);
164 static inline void dp_packet_reset_offsets(struct dp_packet *);
165 static inline uint16_t dp_packet_l2_pad_size(const struct dp_packet *);
166 static inline void dp_packet_set_l2_pad_size(struct dp_packet *, uint16_t);
167 static inline void *dp_packet_l2_5(const struct dp_packet *);
168 static inline void dp_packet_set_l2_5(struct dp_packet *, void *);
169 static inline void *dp_packet_l3(const struct dp_packet *);
170 static inline void dp_packet_set_l3(struct dp_packet *, void *);
171 static inline void *dp_packet_l4(const struct dp_packet *);
172 static inline void dp_packet_set_l4(struct dp_packet *, void *);
173 static inline size_t dp_packet_l4_size(const struct dp_packet *);
174 static inline const void *dp_packet_get_tcp_payload(const struct dp_packet *);
175 static inline const void *dp_packet_get_udp_payload(const struct dp_packet *);
176 static inline const void *dp_packet_get_sctp_payload(const struct dp_packet *);
177 static inline const void *dp_packet_get_icmp_payload(const struct dp_packet *);
178 static inline const void *dp_packet_get_nd_payload(const struct dp_packet *);
179
180 void dp_packet_use(struct dp_packet *, void *, size_t);
181 void dp_packet_use_stub(struct dp_packet *, void *, size_t);
182 void dp_packet_use_const(struct dp_packet *, const void *, size_t);
183 #if HAVE_AF_XDP
184 void dp_packet_use_afxdp(struct dp_packet *, void *, size_t, size_t);
185 #endif
186 void dp_packet_init_dpdk(struct dp_packet *);
187
188 void dp_packet_init(struct dp_packet *, size_t);
189 void dp_packet_uninit(struct dp_packet *);
190
191 struct dp_packet *dp_packet_new(size_t);
192 struct dp_packet *dp_packet_new_with_headroom(size_t, size_t headroom);
193 struct dp_packet *dp_packet_clone(const struct dp_packet *);
194 struct dp_packet *dp_packet_clone_with_headroom(const struct dp_packet *,
195 size_t headroom);
196 struct dp_packet *dp_packet_clone_data(const void *, size_t);
197 struct dp_packet *dp_packet_clone_data_with_headroom(const void *, size_t,
198 size_t headroom);
199 void dp_packet_resize(struct dp_packet *b, size_t new_headroom,
200 size_t new_tailroom);
201 static inline void dp_packet_delete(struct dp_packet *);
202
203 static inline void *dp_packet_at(const struct dp_packet *, size_t offset,
204 size_t size);
205 static inline void *dp_packet_at_assert(const struct dp_packet *,
206 size_t offset, size_t size);
207 static inline void *dp_packet_tail(const struct dp_packet *);
208 static inline void *dp_packet_end(const struct dp_packet *);
209
210 void *dp_packet_put_uninit(struct dp_packet *, size_t);
211 void *dp_packet_put_zeros(struct dp_packet *, size_t);
212 void *dp_packet_put(struct dp_packet *, const void *, size_t);
213 char *dp_packet_put_hex(struct dp_packet *, const char *s, size_t *n);
214 void dp_packet_reserve(struct dp_packet *, size_t);
215 void dp_packet_reserve_with_tailroom(struct dp_packet *, size_t headroom,
216 size_t tailroom);
217 void *dp_packet_push_uninit(struct dp_packet *, size_t);
218 void *dp_packet_push_zeros(struct dp_packet *, size_t);
219 void *dp_packet_push(struct dp_packet *, const void *, size_t);
220
221 static inline size_t dp_packet_headroom(const struct dp_packet *);
222 static inline size_t dp_packet_tailroom(const struct dp_packet *);
223 void dp_packet_prealloc_headroom(struct dp_packet *, size_t);
224 void dp_packet_prealloc_tailroom(struct dp_packet *, size_t);
225 void dp_packet_shift(struct dp_packet *, int);
226
227 static inline void dp_packet_clear(struct dp_packet *);
228 static inline void *dp_packet_pull(struct dp_packet *, size_t);
229 static inline void *dp_packet_try_pull(struct dp_packet *, size_t);
230
231 void *dp_packet_steal_data(struct dp_packet *);
232
233 static inline bool dp_packet_equal(const struct dp_packet *,
234 const struct dp_packet *);
235
236 \f
237 /* Frees memory that 'b' points to, as well as 'b' itself. */
238 static inline void
239 dp_packet_delete(struct dp_packet *b)
240 {
241 if (b) {
242 if (b->source == DPBUF_DPDK) {
243 /* If this dp_packet was allocated by DPDK it must have been
244 * created as a dp_packet */
245 free_dpdk_buf((struct dp_packet*) b);
246 return;
247 }
248
249 if (b->source == DPBUF_AFXDP) {
250 free_afxdp_buf(b);
251 return;
252 }
253
254 dp_packet_uninit(b);
255 free(b);
256 }
257 }
258
259 /* If 'b' contains at least 'offset + size' bytes of data, returns a pointer to
260 * byte 'offset'. Otherwise, returns a null pointer. */
261 static inline void *
262 dp_packet_at(const struct dp_packet *b, size_t offset, size_t size)
263 {
264 return offset + size <= dp_packet_size(b)
265 ? (char *) dp_packet_data(b) + offset
266 : NULL;
267 }
268
269 /* Returns a pointer to byte 'offset' in 'b', which must contain at least
270 * 'offset + size' bytes of data. */
271 static inline void *
272 dp_packet_at_assert(const struct dp_packet *b, size_t offset, size_t size)
273 {
274 ovs_assert(offset + size <= dp_packet_size(b));
275 return ((char *) dp_packet_data(b)) + offset;
276 }
277
278 /* Returns a pointer to byte following the last byte of data in use in 'b'. */
279 static inline void *
280 dp_packet_tail(const struct dp_packet *b)
281 {
282 return (char *) dp_packet_data(b) + dp_packet_size(b);
283 }
284
285 /* Returns a pointer to byte following the last byte allocated for use (but
286 * not necessarily in use) in 'b'. */
287 static inline void *
288 dp_packet_end(const struct dp_packet *b)
289 {
290 return (char *) dp_packet_base(b) + dp_packet_get_allocated(b);
291 }
292
293 /* Returns the number of bytes of headroom in 'b', that is, the number of bytes
294 * of unused space in dp_packet 'b' before the data that is in use. (Most
295 * commonly, the data in a dp_packet is at its beginning, and thus the
296 * dp_packet's headroom is 0.) */
297 static inline size_t
298 dp_packet_headroom(const struct dp_packet *b)
299 {
300 return (char *) dp_packet_data(b) - (char *) dp_packet_base(b);
301 }
302
303 /* Returns the number of bytes that may be appended to the tail end of
304 * dp_packet 'b' before the dp_packet must be reallocated. */
305 static inline size_t
306 dp_packet_tailroom(const struct dp_packet *b)
307 {
308 return (char *) dp_packet_end(b) - (char *) dp_packet_tail(b);
309 }
310
311 /* Clears any data from 'b'. */
312 static inline void
313 dp_packet_clear(struct dp_packet *b)
314 {
315 dp_packet_set_data(b, dp_packet_base(b));
316 dp_packet_set_size(b, 0);
317 }
318
319 /* Removes 'size' bytes from the head end of 'b', which must contain at least
320 * 'size' bytes of data. Returns the first byte of data removed. */
321 static inline void *
322 dp_packet_pull(struct dp_packet *b, size_t size)
323 {
324 void *data = dp_packet_data(b);
325 ovs_assert(dp_packet_size(b) - dp_packet_l2_pad_size(b) >= size);
326 dp_packet_set_data(b, (char *) dp_packet_data(b) + size);
327 dp_packet_set_size(b, dp_packet_size(b) - size);
328 return data;
329 }
330
331 /* If 'b' has at least 'size' bytes of data, removes that many bytes from the
332 * head end of 'b' and returns the first byte removed. Otherwise, returns a
333 * null pointer without modifying 'b'. */
334 static inline void *
335 dp_packet_try_pull(struct dp_packet *b, size_t size)
336 {
337 return dp_packet_size(b) - dp_packet_l2_pad_size(b) >= size
338 ? dp_packet_pull(b, size) : NULL;
339 }
340
341 static inline bool
342 dp_packet_equal(const struct dp_packet *a, const struct dp_packet *b)
343 {
344 return dp_packet_size(a) == dp_packet_size(b) &&
345 !memcmp(dp_packet_data(a), dp_packet_data(b), dp_packet_size(a));
346 }
347
348 static inline bool
349 dp_packet_is_eth(const struct dp_packet *b)
350 {
351 return b->packet_type == htonl(PT_ETH);
352 }
353
354 /* Get the start of the Ethernet frame. 'l3_ofs' marks the end of the l2
355 * headers, so return NULL if it is not set. */
356 static inline void *
357 dp_packet_eth(const struct dp_packet *b)
358 {
359 return (dp_packet_is_eth(b) && b->l3_ofs != UINT16_MAX)
360 ? dp_packet_data(b) : NULL;
361 }
362
363 /* Resets all layer offsets. 'l3' offset must be set before 'l2' can be
364 * retrieved. */
365 static inline void
366 dp_packet_reset_offsets(struct dp_packet *b)
367 {
368 b->l2_pad_size = 0;
369 b->l2_5_ofs = UINT16_MAX;
370 b->l3_ofs = UINT16_MAX;
371 b->l4_ofs = UINT16_MAX;
372 }
373
374 static inline uint16_t
375 dp_packet_l2_pad_size(const struct dp_packet *b)
376 {
377 return b->l2_pad_size;
378 }
379
380 static inline void
381 dp_packet_set_l2_pad_size(struct dp_packet *b, uint16_t pad_size)
382 {
383 ovs_assert(pad_size <= dp_packet_size(b));
384 b->l2_pad_size = pad_size;
385 }
386
387 static inline void *
388 dp_packet_l2_5(const struct dp_packet *b)
389 {
390 return b->l2_5_ofs != UINT16_MAX
391 ? (char *) dp_packet_data(b) + b->l2_5_ofs
392 : NULL;
393 }
394
395 static inline void
396 dp_packet_set_l2_5(struct dp_packet *b, void *l2_5)
397 {
398 b->l2_5_ofs = l2_5
399 ? (char *) l2_5 - (char *) dp_packet_data(b)
400 : UINT16_MAX;
401 }
402
403 static inline void *
404 dp_packet_l3(const struct dp_packet *b)
405 {
406 return b->l3_ofs != UINT16_MAX
407 ? (char *) dp_packet_data(b) + b->l3_ofs
408 : NULL;
409 }
410
411 static inline void
412 dp_packet_set_l3(struct dp_packet *b, void *l3)
413 {
414 b->l3_ofs = l3 ? (char *) l3 - (char *) dp_packet_data(b) : UINT16_MAX;
415 }
416
417 static inline void *
418 dp_packet_l4(const struct dp_packet *b)
419 {
420 return b->l4_ofs != UINT16_MAX
421 ? (char *) dp_packet_data(b) + b->l4_ofs
422 : NULL;
423 }
424
425 static inline void
426 dp_packet_set_l4(struct dp_packet *b, void *l4)
427 {
428 b->l4_ofs = l4 ? (char *) l4 - (char *) dp_packet_data(b) : UINT16_MAX;
429 }
430
431 /* Returns the size of the packet from the beginning of the L3 header to the
432 * end of the L3 payload. Hence L2 padding is not included. */
433 static inline size_t
434 dp_packet_l3_size(const struct dp_packet *b)
435 {
436 return OVS_LIKELY(b->l3_ofs != UINT16_MAX)
437 ? (const char *)dp_packet_tail(b) - (const char *)dp_packet_l3(b)
438 - dp_packet_l2_pad_size(b)
439 : 0;
440 }
441
442 /* Returns the size of the packet from the beginning of the L4 header to the
443 * end of the L4 payload. Hence L2 padding is not included. */
444 static inline size_t
445 dp_packet_l4_size(const struct dp_packet *b)
446 {
447 return OVS_LIKELY(b->l4_ofs != UINT16_MAX)
448 ? (const char *)dp_packet_tail(b) - (const char *)dp_packet_l4(b)
449 - dp_packet_l2_pad_size(b)
450 : 0;
451 }
452
453 static inline const void *
454 dp_packet_get_tcp_payload(const struct dp_packet *b)
455 {
456 size_t l4_size = dp_packet_l4_size(b);
457
458 if (OVS_LIKELY(l4_size >= TCP_HEADER_LEN)) {
459 struct tcp_header *tcp = dp_packet_l4(b);
460 int tcp_len = TCP_OFFSET(tcp->tcp_ctl) * 4;
461
462 if (OVS_LIKELY(tcp_len >= TCP_HEADER_LEN && tcp_len <= l4_size)) {
463 return (const char *)tcp + tcp_len;
464 }
465 }
466 return NULL;
467 }
468
469 static inline uint32_t
470 dp_packet_get_tcp_payload_length(const struct dp_packet *pkt)
471 {
472 const char *tcp_payload = dp_packet_get_tcp_payload(pkt);
473 if (tcp_payload) {
474 return ((char *) dp_packet_tail(pkt) - dp_packet_l2_pad_size(pkt)
475 - tcp_payload);
476 } else {
477 return 0;
478 }
479 }
480
481 static inline const void *
482 dp_packet_get_udp_payload(const struct dp_packet *b)
483 {
484 return OVS_LIKELY(dp_packet_l4_size(b) >= UDP_HEADER_LEN)
485 ? (const char *)dp_packet_l4(b) + UDP_HEADER_LEN : NULL;
486 }
487
488 static inline const void *
489 dp_packet_get_sctp_payload(const struct dp_packet *b)
490 {
491 return OVS_LIKELY(dp_packet_l4_size(b) >= SCTP_HEADER_LEN)
492 ? (const char *)dp_packet_l4(b) + SCTP_HEADER_LEN : NULL;
493 }
494
495 static inline const void *
496 dp_packet_get_icmp_payload(const struct dp_packet *b)
497 {
498 return OVS_LIKELY(dp_packet_l4_size(b) >= ICMP_HEADER_LEN)
499 ? (const char *)dp_packet_l4(b) + ICMP_HEADER_LEN : NULL;
500 }
501
502 static inline const void *
503 dp_packet_get_nd_payload(const struct dp_packet *b)
504 {
505 return OVS_LIKELY(dp_packet_l4_size(b) >= ND_MSG_LEN)
506 ? (const char *)dp_packet_l4(b) + ND_MSG_LEN : NULL;
507 }
508
509 #ifdef DPDK_NETDEV
510 static inline uint64_t *
511 dp_packet_ol_flags_ptr(const struct dp_packet *b)
512 {
513 return CONST_CAST(uint64_t *, &b->mbuf.ol_flags);
514 }
515
516 static inline uint32_t *
517 dp_packet_rss_ptr(const struct dp_packet *b)
518 {
519 return CONST_CAST(uint32_t *, &b->mbuf.hash.rss);
520 }
521
522 static inline uint32_t *
523 dp_packet_flow_mark_ptr(const struct dp_packet *b)
524 {
525 return CONST_CAST(uint32_t *, &b->mbuf.hash.fdir.hi);
526 }
527
528 #else
529 static inline uint32_t *
530 dp_packet_ol_flags_ptr(const struct dp_packet *b)
531 {
532 return CONST_CAST(uint32_t *, &b->ol_flags);
533 }
534
535 static inline uint32_t *
536 dp_packet_rss_ptr(const struct dp_packet *b)
537 {
538 return CONST_CAST(uint32_t *, &b->rss_hash);
539 }
540
541 static inline uint32_t *
542 dp_packet_flow_mark_ptr(const struct dp_packet *b)
543 {
544 return CONST_CAST(uint32_t *, &b->flow_mark);
545 }
546 #endif
547
548 #ifdef DPDK_NETDEV
549 BUILD_ASSERT_DECL(offsetof(struct dp_packet, mbuf) == 0);
550
551 static inline void
552 dp_packet_init_specific(struct dp_packet *p)
553 {
554 /* This initialization is needed for packets that do not come from DPDK
555 * interfaces, when vswitchd is built with --with-dpdk. */
556 p->mbuf.ol_flags = p->mbuf.tx_offload = p->mbuf.packet_type = 0;
557 p->mbuf.nb_segs = 1;
558 p->mbuf.next = NULL;
559 }
560
561 static inline void *
562 dp_packet_base(const struct dp_packet *b)
563 {
564 return b->mbuf.buf_addr;
565 }
566
567 static inline void
568 dp_packet_set_base(struct dp_packet *b, void *d)
569 {
570 b->mbuf.buf_addr = d;
571 }
572
573 static inline uint32_t
574 dp_packet_size(const struct dp_packet *b)
575 {
576 return b->mbuf.pkt_len;
577 }
578
579 static inline void
580 dp_packet_set_size(struct dp_packet *b, uint32_t v)
581 {
582 /* netdev-dpdk does not currently support segmentation; consequently, for
583 * all intents and purposes, 'data_len' (16 bit) and 'pkt_len' (32 bit) may
584 * be used interchangably.
585 *
586 * On the datapath, it is expected that the size of packets
587 * (and thus 'v') will always be <= UINT16_MAX; this means that there is no
588 * loss of accuracy in assigning 'v' to 'data_len'.
589 */
590 b->mbuf.data_len = (uint16_t)v; /* Current seg length. */
591 b->mbuf.pkt_len = v; /* Total length of all segments linked to
592 * this segment. */
593 }
594
595 static inline uint16_t
596 __packet_data(const struct dp_packet *b)
597 {
598 return b->mbuf.data_off;
599 }
600
601 static inline void
602 __packet_set_data(struct dp_packet *b, uint16_t v)
603 {
604 b->mbuf.data_off = v;
605 }
606
607 static inline uint16_t
608 dp_packet_get_allocated(const struct dp_packet *b)
609 {
610 return b->mbuf.buf_len;
611 }
612
613 static inline void
614 dp_packet_set_allocated(struct dp_packet *b, uint16_t s)
615 {
616 b->mbuf.buf_len = s;
617 }
618
619 #else /* DPDK_NETDEV */
620
621 static inline void
622 dp_packet_init_specific(struct dp_packet *p OVS_UNUSED)
623 {
624 /* There are no implementation-specific fields for initialization. */
625 }
626
627 static inline void *
628 dp_packet_base(const struct dp_packet *b)
629 {
630 return b->base_;
631 }
632
633 static inline void
634 dp_packet_set_base(struct dp_packet *b, void *d)
635 {
636 b->base_ = d;
637 }
638
639 static inline uint32_t
640 dp_packet_size(const struct dp_packet *b)
641 {
642 return b->size_;
643 }
644
645 static inline void
646 dp_packet_set_size(struct dp_packet *b, uint32_t v)
647 {
648 b->size_ = v;
649 }
650
651 static inline uint16_t
652 __packet_data(const struct dp_packet *b)
653 {
654 return b->data_ofs;
655 }
656
657 static inline void
658 __packet_set_data(struct dp_packet *b, uint16_t v)
659 {
660 b->data_ofs = v;
661 }
662
663 static inline uint16_t
664 dp_packet_get_allocated(const struct dp_packet *b)
665 {
666 return b->allocated_;
667 }
668
669 static inline void
670 dp_packet_set_allocated(struct dp_packet *b, uint16_t s)
671 {
672 b->allocated_ = s;
673 }
674
675 #endif /* DPDK_NETDEV */
676
677 static inline void
678 dp_packet_reset_cutlen(struct dp_packet *b)
679 {
680 b->cutlen = 0;
681 }
682
683 static inline uint32_t
684 dp_packet_set_cutlen(struct dp_packet *b, uint32_t max_len)
685 {
686 if (max_len < ETH_HEADER_LEN) {
687 max_len = ETH_HEADER_LEN;
688 }
689
690 if (max_len >= dp_packet_size(b)) {
691 b->cutlen = 0;
692 } else {
693 b->cutlen = dp_packet_size(b) - max_len;
694 }
695 return b->cutlen;
696 }
697
698 static inline uint32_t
699 dp_packet_get_cutlen(const struct dp_packet *b)
700 {
701 /* Always in valid range if user uses dp_packet_set_cutlen. */
702 return b->cutlen;
703 }
704
705 static inline uint32_t
706 dp_packet_get_send_len(const struct dp_packet *b)
707 {
708 return dp_packet_size(b) - dp_packet_get_cutlen(b);
709 }
710
711 static inline void *
712 dp_packet_data(const struct dp_packet *b)
713 {
714 return __packet_data(b) != UINT16_MAX
715 ? (char *) dp_packet_base(b) + __packet_data(b) : NULL;
716 }
717
718 static inline void
719 dp_packet_set_data(struct dp_packet *b, void *data)
720 {
721 if (data) {
722 __packet_set_data(b, (char *) data - (char *) dp_packet_base(b));
723 } else {
724 __packet_set_data(b, UINT16_MAX);
725 }
726 }
727
728 static inline void
729 dp_packet_reset_packet(struct dp_packet *b, int off)
730 {
731 dp_packet_set_size(b, dp_packet_size(b) - off);
732 dp_packet_set_data(b, ((unsigned char *) dp_packet_data(b) + off));
733 dp_packet_reset_offsets(b);
734 }
735
736 enum { NETDEV_MAX_BURST = 32 }; /* Maximum number packets in a batch. */
737
738 struct dp_packet_batch {
739 size_t count;
740 bool trunc; /* true if the batch needs truncate. */
741 bool do_not_steal; /* Indicate that the packets should not be stolen. */
742 struct dp_packet *packets[NETDEV_MAX_BURST];
743 };
744
745 static inline void
746 dp_packet_batch_init(struct dp_packet_batch *batch)
747 {
748 batch->count = 0;
749 batch->trunc = false;
750 batch->do_not_steal = false;
751 }
752
753 static inline void
754 dp_packet_batch_add__(struct dp_packet_batch *batch,
755 struct dp_packet *packet, size_t limit)
756 {
757 if (batch->count < limit) {
758 batch->packets[batch->count++] = packet;
759 } else {
760 dp_packet_delete(packet);
761 }
762 }
763
764 /* When the batch is full, 'packet' will be dropped and freed. */
765 static inline void
766 dp_packet_batch_add(struct dp_packet_batch *batch, struct dp_packet *packet)
767 {
768 dp_packet_batch_add__(batch, packet, NETDEV_MAX_BURST);
769 }
770
771 static inline size_t
772 dp_packet_batch_size(const struct dp_packet_batch *batch)
773 {
774 return batch->count;
775 }
776
777 /* Clear 'batch' for refill. Use dp_packet_batch_refill() to add
778 * packets back into the 'batch'. */
779 static inline void
780 dp_packet_batch_refill_init(struct dp_packet_batch *batch)
781 {
782 batch->count = 0;
783 };
784
785 static inline void
786 dp_packet_batch_refill(struct dp_packet_batch *batch,
787 struct dp_packet *packet, size_t idx)
788 {
789 dp_packet_batch_add__(batch, packet, MIN(NETDEV_MAX_BURST, idx + 1));
790 }
791
792 static inline void
793 dp_packet_batch_init_packet(struct dp_packet_batch *batch, struct dp_packet *p)
794 {
795 dp_packet_batch_init(batch);
796 batch->count = 1;
797 batch->packets[0] = p;
798 }
799
800 static inline bool
801 dp_packet_batch_is_empty(const struct dp_packet_batch *batch)
802 {
803 return !dp_packet_batch_size(batch);
804 }
805
806 static inline bool
807 dp_packet_batch_is_full(const struct dp_packet_batch *batch)
808 {
809 return dp_packet_batch_size(batch) == NETDEV_MAX_BURST;
810 }
811
812 #define DP_PACKET_BATCH_FOR_EACH(IDX, PACKET, BATCH) \
813 for (size_t IDX = 0; IDX < dp_packet_batch_size(BATCH); IDX++) \
814 if (PACKET = BATCH->packets[IDX], true)
815
816 /* Use this macro for cases where some packets in the 'BATCH' may be
817 * dropped after going through each packet in the 'BATCH'.
818 *
819 * For packets to stay in the 'BATCH', they need to be refilled back
820 * into the 'BATCH' by calling dp_packet_batch_refill(). Caller owns
821 * the packets that are not refilled.
822 *
823 * Caller needs to supply 'SIZE', that stores the current number of
824 * packets in 'BATCH'. It is best to declare this variable with
825 * the 'const' modifier since it should not be modified by
826 * the iterator. */
827 #define DP_PACKET_BATCH_REFILL_FOR_EACH(IDX, SIZE, PACKET, BATCH) \
828 for (dp_packet_batch_refill_init(BATCH), IDX=0; IDX < SIZE; IDX++) \
829 if (PACKET = BATCH->packets[IDX], true)
830
831 static inline void
832 dp_packet_batch_clone(struct dp_packet_batch *dst,
833 struct dp_packet_batch *src)
834 {
835 struct dp_packet *packet;
836
837 dp_packet_batch_init(dst);
838 DP_PACKET_BATCH_FOR_EACH (i, packet, src) {
839 if (i + 1 < dp_packet_batch_size(src)) {
840 OVS_PREFETCH(src->packets[i + 1]);
841 }
842
843 uint32_t headroom = dp_packet_headroom(packet);
844 struct dp_packet *pkt_clone;
845
846 pkt_clone = dp_packet_clone_with_headroom(packet, headroom);
847 dp_packet_batch_add(dst, pkt_clone);
848 }
849 dst->trunc = src->trunc;
850 }
851
852 static inline void
853 dp_packet_delete_batch(struct dp_packet_batch *batch, bool should_steal)
854 {
855 if (should_steal) {
856 struct dp_packet *packet;
857
858 DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
859 dp_packet_delete(packet);
860 }
861 dp_packet_batch_init(batch);
862 }
863 }
864
865 static inline void
866 dp_packet_batch_init_packet_fields(struct dp_packet_batch *batch)
867 {
868 struct dp_packet *packet;
869
870 DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
871 dp_packet_reset_cutlen(packet);
872 packet->packet_type = htonl(PT_ETH);
873 }
874 }
875
876 static inline void
877 dp_packet_batch_apply_cutlen(struct dp_packet_batch *batch)
878 {
879 if (batch->trunc) {
880 struct dp_packet *packet;
881
882 DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
883 dp_packet_set_size(packet, dp_packet_get_send_len(packet));
884 dp_packet_reset_cutlen(packet);
885 }
886 batch->trunc = false;
887 }
888 }
889
890 static inline void
891 dp_packet_batch_reset_cutlen(struct dp_packet_batch *batch)
892 {
893 if (batch->trunc) {
894 struct dp_packet *packet;
895
896 DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
897 dp_packet_reset_cutlen(packet);
898 }
899 batch->trunc = false;
900 }
901 }
902
903 /* Returns the RSS hash of the packet 'p'. Note that the returned value is
904 * correct only if 'dp_packet_rss_valid(p)' returns 'true'. */
905 static inline uint32_t
906 dp_packet_get_rss_hash(const struct dp_packet *p)
907 {
908 return *dp_packet_rss_ptr(p);
909 }
910
911 static inline void
912 dp_packet_set_rss_hash(struct dp_packet *p, uint32_t hash)
913 {
914 *dp_packet_rss_ptr(p) = hash;
915 *dp_packet_ol_flags_ptr(p) |= DP_PACKET_OL_RSS_HASH;
916 }
917
918 static inline bool
919 dp_packet_rss_valid(const struct dp_packet *p)
920 {
921 return *dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_RSS_HASH;
922 }
923
924 static inline void
925 dp_packet_reset_offload(struct dp_packet *p)
926 {
927 *dp_packet_ol_flags_ptr(p) &= ~DP_PACKET_OL_SUPPORTED_MASK;
928 }
929
930 static inline bool
931 dp_packet_has_flow_mark(const struct dp_packet *p, uint32_t *mark)
932 {
933 if (*dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_FLOW_MARK) {
934 *mark = *dp_packet_flow_mark_ptr(p);
935 return true;
936 }
937
938 return false;
939 }
940
941 static inline void
942 dp_packet_set_flow_mark(struct dp_packet *p, uint32_t mark)
943 {
944 *dp_packet_flow_mark_ptr(p) = mark;
945 *dp_packet_ol_flags_ptr(p) |= DP_PACKET_OL_FLOW_MARK;
946 }
947
948 /* Returns the L4 cksum offload bitmask. */
949 static inline uint64_t
950 dp_packet_hwol_l4_mask(const struct dp_packet *b)
951 {
952 return *dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_L4_MASK;
953 }
954
955 /* Return true if the packet 'b' requested L4 checksum offload. */
956 static inline bool
957 dp_packet_hwol_tx_l4_checksum(const struct dp_packet *b)
958 {
959 return !!dp_packet_hwol_l4_mask(b);
960 }
961
962 /* Returns 'true' if packet 'b' is marked for TCP segmentation offloading. */
963 static inline bool
964 dp_packet_hwol_is_tso(const struct dp_packet *b)
965 {
966 return !!(*dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_TCP_SEG);
967 }
968
969 /* Returns 'true' if packet 'b' is marked for IPv4 checksum offloading. */
970 static inline bool
971 dp_packet_hwol_is_ipv4(const struct dp_packet *b)
972 {
973 return !!(*dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_IPV4);
974 }
975
976 /* Returns 'true' if packet 'b' is marked for TCP checksum offloading. */
977 static inline bool
978 dp_packet_hwol_l4_is_tcp(const struct dp_packet *b)
979 {
980 return (*dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_L4_MASK) ==
981 DP_PACKET_OL_TX_TCP_CKSUM;
982 }
983
984 /* Returns 'true' if packet 'b' is marked for UDP checksum offloading. */
985 static inline bool
986 dp_packet_hwol_l4_is_udp(struct dp_packet *b)
987 {
988 return (*dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_L4_MASK) ==
989 DP_PACKET_OL_TX_UDP_CKSUM;
990 }
991
992 /* Returns 'true' if packet 'b' is marked for SCTP checksum offloading. */
993 static inline bool
994 dp_packet_hwol_l4_is_sctp(struct dp_packet *b)
995 {
996 return (*dp_packet_ol_flags_ptr(b) & DP_PACKET_OL_TX_L4_MASK) ==
997 DP_PACKET_OL_TX_SCTP_CKSUM;
998 }
999
1000 /* Mark packet 'b' for IPv4 checksum offloading. */
1001 static inline void
1002 dp_packet_hwol_set_tx_ipv4(struct dp_packet *b)
1003 {
1004 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_IPV4;
1005 }
1006
1007 /* Mark packet 'b' for IPv6 checksum offloading. */
1008 static inline void
1009 dp_packet_hwol_set_tx_ipv6(struct dp_packet *b)
1010 {
1011 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_IPV6;
1012 }
1013
1014 /* Mark packet 'b' for TCP checksum offloading. It implies that either
1015 * the packet 'b' is marked for IPv4 or IPv6 checksum offloading. */
1016 static inline void
1017 dp_packet_hwol_set_csum_tcp(struct dp_packet *b)
1018 {
1019 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_TCP_CKSUM;
1020 }
1021
1022 /* Mark packet 'b' for UDP checksum offloading. It implies that either
1023 * the packet 'b' is marked for IPv4 or IPv6 checksum offloading. */
1024 static inline void
1025 dp_packet_hwol_set_csum_udp(struct dp_packet *b)
1026 {
1027 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_UDP_CKSUM;
1028 }
1029
1030 /* Mark packet 'b' for SCTP checksum offloading. It implies that either
1031 * the packet 'b' is marked for IPv4 or IPv6 checksum offloading. */
1032 static inline void
1033 dp_packet_hwol_set_csum_sctp(struct dp_packet *b)
1034 {
1035 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_SCTP_CKSUM;
1036 }
1037
1038 /* Mark packet 'b' for TCP segmentation offloading. It implies that
1039 * either the packet 'b' is marked for IPv4 or IPv6 checksum offloading
1040 * and also for TCP checksum offloading. */
1041 static inline void
1042 dp_packet_hwol_set_tcp_seg(struct dp_packet *b)
1043 {
1044 *dp_packet_ol_flags_ptr(b) |= DP_PACKET_OL_TX_TCP_SEG;
1045 }
1046
1047 static inline bool
1048 dp_packet_ip_checksum_valid(const struct dp_packet *p)
1049 {
1050 return (*dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_RX_IP_CKSUM_MASK) ==
1051 DP_PACKET_OL_RX_IP_CKSUM_GOOD;
1052 }
1053
1054 static inline bool
1055 dp_packet_ip_checksum_bad(const struct dp_packet *p)
1056 {
1057 return (*dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_RX_IP_CKSUM_MASK) ==
1058 DP_PACKET_OL_RX_IP_CKSUM_BAD;
1059 }
1060
1061 static inline bool
1062 dp_packet_l4_checksum_valid(const struct dp_packet *p)
1063 {
1064 return (*dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_RX_L4_CKSUM_MASK) ==
1065 DP_PACKET_OL_RX_L4_CKSUM_GOOD;
1066 }
1067
1068 static inline bool
1069 dp_packet_l4_checksum_bad(const struct dp_packet *p)
1070 {
1071 return (*dp_packet_ol_flags_ptr(p) & DP_PACKET_OL_RX_L4_CKSUM_MASK) ==
1072 DP_PACKET_OL_RX_L4_CKSUM_BAD;
1073 }
1074
1075 #ifdef __cplusplus
1076 }
1077 #endif
1078
1079 #endif /* dp-packet.h */