2 * Copyright (c) 2008, 2009, 2010, 2011, 2012 Nicira, Inc.
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at:
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
17 #include <sys/types.h>
22 #include <netinet/in.h>
23 #include <netinet/icmp6.h>
24 #include <netinet/ip6.h>
28 #include "byte-order.h"
31 #include "dynamic-string.h"
36 #include "openflow/openflow.h"
38 #include "unaligned.h"
41 VLOG_DEFINE_THIS_MODULE(flow
);
43 COVERAGE_DEFINE(flow_extract
);
44 COVERAGE_DEFINE(miniflow_malloc
);
46 static struct arp_eth_header
*
47 pull_arp(struct ofpbuf
*packet
)
49 return ofpbuf_try_pull(packet
, ARP_ETH_HEADER_LEN
);
52 static struct ip_header
*
53 pull_ip(struct ofpbuf
*packet
)
55 if (packet
->size
>= IP_HEADER_LEN
) {
56 struct ip_header
*ip
= packet
->data
;
57 int ip_len
= IP_IHL(ip
->ip_ihl_ver
) * 4;
58 if (ip_len
>= IP_HEADER_LEN
&& packet
->size
>= ip_len
) {
59 return ofpbuf_pull(packet
, ip_len
);
65 static struct tcp_header
*
66 pull_tcp(struct ofpbuf
*packet
)
68 if (packet
->size
>= TCP_HEADER_LEN
) {
69 struct tcp_header
*tcp
= packet
->data
;
70 int tcp_len
= TCP_OFFSET(tcp
->tcp_ctl
) * 4;
71 if (tcp_len
>= TCP_HEADER_LEN
&& packet
->size
>= tcp_len
) {
72 return ofpbuf_pull(packet
, tcp_len
);
78 static struct udp_header
*
79 pull_udp(struct ofpbuf
*packet
)
81 return ofpbuf_try_pull(packet
, UDP_HEADER_LEN
);
84 static struct icmp_header
*
85 pull_icmp(struct ofpbuf
*packet
)
87 return ofpbuf_try_pull(packet
, ICMP_HEADER_LEN
);
90 static struct icmp6_hdr
*
91 pull_icmpv6(struct ofpbuf
*packet
)
93 return ofpbuf_try_pull(packet
, sizeof(struct icmp6_hdr
));
97 parse_mpls(struct ofpbuf
*b
, struct flow
*flow
)
101 while ((mh
= ofpbuf_try_pull(b
, sizeof *mh
))) {
102 if (flow
->mpls_depth
++ == 0) {
103 flow
->mpls_lse
= mh
->mpls_lse
;
105 if (mh
->mpls_lse
& htonl(MPLS_BOS_MASK
)) {
112 parse_vlan(struct ofpbuf
*b
, struct flow
*flow
)
115 ovs_be16 eth_type
; /* ETH_TYPE_VLAN */
119 if (b
->size
>= sizeof(struct qtag_prefix
) + sizeof(ovs_be16
)) {
120 struct qtag_prefix
*qp
= ofpbuf_pull(b
, sizeof *qp
);
121 flow
->vlan_tci
= qp
->tci
| htons(VLAN_CFI
);
126 parse_ethertype(struct ofpbuf
*b
)
128 struct llc_snap_header
*llc
;
131 proto
= *(ovs_be16
*) ofpbuf_pull(b
, sizeof proto
);
132 if (ntohs(proto
) >= ETH_TYPE_MIN
) {
136 if (b
->size
< sizeof *llc
) {
137 return htons(FLOW_DL_TYPE_NONE
);
141 if (llc
->llc
.llc_dsap
!= LLC_DSAP_SNAP
142 || llc
->llc
.llc_ssap
!= LLC_SSAP_SNAP
143 || llc
->llc
.llc_cntl
!= LLC_CNTL_SNAP
144 || memcmp(llc
->snap
.snap_org
, SNAP_ORG_ETHERNET
,
145 sizeof llc
->snap
.snap_org
)) {
146 return htons(FLOW_DL_TYPE_NONE
);
149 ofpbuf_pull(b
, sizeof *llc
);
151 if (ntohs(llc
->snap
.snap_type
) >= ETH_TYPE_MIN
) {
152 return llc
->snap
.snap_type
;
155 return htons(FLOW_DL_TYPE_NONE
);
159 parse_ipv6(struct ofpbuf
*packet
, struct flow
*flow
)
161 const struct ip6_hdr
*nh
;
165 nh
= ofpbuf_try_pull(packet
, sizeof *nh
);
170 nexthdr
= nh
->ip6_nxt
;
172 flow
->ipv6_src
= nh
->ip6_src
;
173 flow
->ipv6_dst
= nh
->ip6_dst
;
175 tc_flow
= get_unaligned_be32(&nh
->ip6_flow
);
176 flow
->nw_tos
= ntohl(tc_flow
) >> 20;
177 flow
->ipv6_label
= tc_flow
& htonl(IPV6_LABEL_MASK
);
178 flow
->nw_ttl
= nh
->ip6_hlim
;
179 flow
->nw_proto
= IPPROTO_NONE
;
182 if ((nexthdr
!= IPPROTO_HOPOPTS
)
183 && (nexthdr
!= IPPROTO_ROUTING
)
184 && (nexthdr
!= IPPROTO_DSTOPTS
)
185 && (nexthdr
!= IPPROTO_AH
)
186 && (nexthdr
!= IPPROTO_FRAGMENT
)) {
187 /* It's either a terminal header (e.g., TCP, UDP) or one we
188 * don't understand. In either case, we're done with the
189 * packet, so use it to fill in 'nw_proto'. */
193 /* We only verify that at least 8 bytes of the next header are
194 * available, but many of these headers are longer. Ensure that
195 * accesses within the extension header are within those first 8
196 * bytes. All extension headers are required to be at least 8
198 if (packet
->size
< 8) {
202 if ((nexthdr
== IPPROTO_HOPOPTS
)
203 || (nexthdr
== IPPROTO_ROUTING
)
204 || (nexthdr
== IPPROTO_DSTOPTS
)) {
205 /* These headers, while different, have the fields we care about
206 * in the same location and with the same interpretation. */
207 const struct ip6_ext
*ext_hdr
= packet
->data
;
208 nexthdr
= ext_hdr
->ip6e_nxt
;
209 if (!ofpbuf_try_pull(packet
, (ext_hdr
->ip6e_len
+ 1) * 8)) {
212 } else if (nexthdr
== IPPROTO_AH
) {
213 /* A standard AH definition isn't available, but the fields
214 * we care about are in the same location as the generic
215 * option header--only the header length is calculated
217 const struct ip6_ext
*ext_hdr
= packet
->data
;
218 nexthdr
= ext_hdr
->ip6e_nxt
;
219 if (!ofpbuf_try_pull(packet
, (ext_hdr
->ip6e_len
+ 2) * 4)) {
222 } else if (nexthdr
== IPPROTO_FRAGMENT
) {
223 const struct ip6_frag
*frag_hdr
= packet
->data
;
225 nexthdr
= frag_hdr
->ip6f_nxt
;
226 if (!ofpbuf_try_pull(packet
, sizeof *frag_hdr
)) {
230 /* We only process the first fragment. */
231 if (frag_hdr
->ip6f_offlg
!= htons(0)) {
232 flow
->nw_frag
= FLOW_NW_FRAG_ANY
;
233 if ((frag_hdr
->ip6f_offlg
& IP6F_OFF_MASK
) != htons(0)) {
234 flow
->nw_frag
|= FLOW_NW_FRAG_LATER
;
235 nexthdr
= IPPROTO_FRAGMENT
;
242 flow
->nw_proto
= nexthdr
;
247 parse_tcp(struct ofpbuf
*packet
, struct ofpbuf
*b
, struct flow
*flow
)
249 const struct tcp_header
*tcp
= pull_tcp(b
);
251 flow
->tp_src
= tcp
->tcp_src
;
252 flow
->tp_dst
= tcp
->tcp_dst
;
253 packet
->l7
= b
->data
;
258 parse_udp(struct ofpbuf
*packet
, struct ofpbuf
*b
, struct flow
*flow
)
260 const struct udp_header
*udp
= pull_udp(b
);
262 flow
->tp_src
= udp
->udp_src
;
263 flow
->tp_dst
= udp
->udp_dst
;
264 packet
->l7
= b
->data
;
269 parse_icmpv6(struct ofpbuf
*b
, struct flow
*flow
)
271 const struct icmp6_hdr
*icmp
= pull_icmpv6(b
);
277 /* The ICMPv6 type and code fields use the 16-bit transport port
278 * fields, so we need to store them in 16-bit network byte order. */
279 flow
->tp_src
= htons(icmp
->icmp6_type
);
280 flow
->tp_dst
= htons(icmp
->icmp6_code
);
282 if (icmp
->icmp6_code
== 0 &&
283 (icmp
->icmp6_type
== ND_NEIGHBOR_SOLICIT
||
284 icmp
->icmp6_type
== ND_NEIGHBOR_ADVERT
)) {
285 const struct in6_addr
*nd_target
;
287 nd_target
= ofpbuf_try_pull(b
, sizeof *nd_target
);
291 flow
->nd_target
= *nd_target
;
293 while (b
->size
>= 8) {
294 /* The minimum size of an option is 8 bytes, which also is
295 * the size of Ethernet link-layer options. */
296 const struct nd_opt_hdr
*nd_opt
= b
->data
;
297 int opt_len
= nd_opt
->nd_opt_len
* 8;
299 if (!opt_len
|| opt_len
> b
->size
) {
303 /* Store the link layer address if the appropriate option is
304 * provided. It is considered an error if the same link
305 * layer option is specified twice. */
306 if (nd_opt
->nd_opt_type
== ND_OPT_SOURCE_LINKADDR
308 if (eth_addr_is_zero(flow
->arp_sha
)) {
309 memcpy(flow
->arp_sha
, nd_opt
+ 1, ETH_ADDR_LEN
);
313 } else if (nd_opt
->nd_opt_type
== ND_OPT_TARGET_LINKADDR
315 if (eth_addr_is_zero(flow
->arp_tha
)) {
316 memcpy(flow
->arp_tha
, nd_opt
+ 1, ETH_ADDR_LEN
);
322 if (!ofpbuf_try_pull(b
, opt_len
)) {
331 memset(&flow
->nd_target
, 0, sizeof(flow
->nd_target
));
332 memset(flow
->arp_sha
, 0, sizeof(flow
->arp_sha
));
333 memset(flow
->arp_tha
, 0, sizeof(flow
->arp_tha
));
339 /* Initializes 'flow' members from 'packet', 'skb_priority', 'tnl', and
342 * Initializes 'packet' header pointers as follows:
344 * - packet->l2 to the start of the Ethernet header.
346 * - packet->l2_5 to the start of the MPLS shim header.
348 * - packet->l3 to just past the Ethernet header, or just past the
349 * vlan_header if one is present, to the first byte of the payload of the
352 * - packet->l4 to just past the IPv4 header, if one is present and has a
353 * correct length, and otherwise NULL.
355 * - packet->l7 to just past the TCP or UDP or ICMP header, if one is
356 * present and has a correct length, and otherwise NULL.
359 flow_extract(struct ofpbuf
*packet
, uint32_t skb_priority
, uint32_t skb_mark
,
360 const struct flow_tnl
*tnl
, uint16_t ofp_in_port
,
363 struct ofpbuf b
= *packet
;
364 struct eth_header
*eth
;
366 COVERAGE_INC(flow_extract
);
368 memset(flow
, 0, sizeof *flow
);
371 ovs_assert(tnl
!= &flow
->tunnel
);
374 flow
->in_port
= ofp_in_port
;
375 flow
->skb_priority
= skb_priority
;
376 flow
->skb_mark
= skb_mark
;
384 if (b
.size
< sizeof *eth
) {
390 memcpy(flow
->dl_src
, eth
->eth_src
, ETH_ADDR_LEN
);
391 memcpy(flow
->dl_dst
, eth
->eth_dst
, ETH_ADDR_LEN
);
393 /* dl_type, vlan_tci. */
394 ofpbuf_pull(&b
, ETH_ADDR_LEN
* 2);
395 if (eth
->eth_type
== htons(ETH_TYPE_VLAN
)) {
396 parse_vlan(&b
, flow
);
398 flow
->dl_type
= parse_ethertype(&b
);
400 /* Parse mpls, copy l3 ttl. */
401 if (eth_type_mpls(flow
->dl_type
)) {
402 packet
->l2_5
= b
.data
;
403 parse_mpls(&b
, flow
);
408 if (flow
->dl_type
== htons(ETH_TYPE_IP
)) {
409 const struct ip_header
*nh
= pull_ip(&b
);
413 flow
->nw_src
= get_unaligned_be32(&nh
->ip_src
);
414 flow
->nw_dst
= get_unaligned_be32(&nh
->ip_dst
);
415 flow
->nw_proto
= nh
->ip_proto
;
417 flow
->nw_tos
= nh
->ip_tos
;
418 if (IP_IS_FRAGMENT(nh
->ip_frag_off
)) {
419 flow
->nw_frag
= FLOW_NW_FRAG_ANY
;
420 if (nh
->ip_frag_off
& htons(IP_FRAG_OFF_MASK
)) {
421 flow
->nw_frag
|= FLOW_NW_FRAG_LATER
;
424 flow
->nw_ttl
= nh
->ip_ttl
;
426 if (!(nh
->ip_frag_off
& htons(IP_FRAG_OFF_MASK
))) {
427 if (flow
->nw_proto
== IPPROTO_TCP
) {
428 parse_tcp(packet
, &b
, flow
);
429 } else if (flow
->nw_proto
== IPPROTO_UDP
) {
430 parse_udp(packet
, &b
, flow
);
431 } else if (flow
->nw_proto
== IPPROTO_ICMP
) {
432 const struct icmp_header
*icmp
= pull_icmp(&b
);
434 flow
->tp_src
= htons(icmp
->icmp_type
);
435 flow
->tp_dst
= htons(icmp
->icmp_code
);
441 } else if (flow
->dl_type
== htons(ETH_TYPE_IPV6
)) {
442 if (parse_ipv6(&b
, flow
)) {
447 if (flow
->nw_proto
== IPPROTO_TCP
) {
448 parse_tcp(packet
, &b
, flow
);
449 } else if (flow
->nw_proto
== IPPROTO_UDP
) {
450 parse_udp(packet
, &b
, flow
);
451 } else if (flow
->nw_proto
== IPPROTO_ICMPV6
) {
452 if (parse_icmpv6(&b
, flow
)) {
456 } else if (flow
->dl_type
== htons(ETH_TYPE_ARP
) ||
457 flow
->dl_type
== htons(ETH_TYPE_RARP
)) {
458 const struct arp_eth_header
*arp
= pull_arp(&b
);
459 if (arp
&& arp
->ar_hrd
== htons(1)
460 && arp
->ar_pro
== htons(ETH_TYPE_IP
)
461 && arp
->ar_hln
== ETH_ADDR_LEN
462 && arp
->ar_pln
== 4) {
463 /* We only match on the lower 8 bits of the opcode. */
464 if (ntohs(arp
->ar_op
) <= 0xff) {
465 flow
->nw_proto
= ntohs(arp
->ar_op
);
468 flow
->nw_src
= arp
->ar_spa
;
469 flow
->nw_dst
= arp
->ar_tpa
;
470 memcpy(flow
->arp_sha
, arp
->ar_sha
, ETH_ADDR_LEN
);
471 memcpy(flow
->arp_tha
, arp
->ar_tha
, ETH_ADDR_LEN
);
476 /* For every bit of a field that is wildcarded in 'wildcards', sets the
477 * corresponding bit in 'flow' to zero. */
479 flow_zero_wildcards(struct flow
*flow
, const struct flow_wildcards
*wildcards
)
481 uint32_t *flow_u32
= (uint32_t *) flow
;
482 const uint32_t *wc_u32
= (const uint32_t *) &wildcards
->masks
;
485 for (i
= 0; i
< FLOW_U32S
; i
++) {
486 flow_u32
[i
] &= wc_u32
[i
];
490 /* Initializes 'fmd' with the metadata found in 'flow'. */
492 flow_get_metadata(const struct flow
*flow
, struct flow_metadata
*fmd
)
494 BUILD_ASSERT_DECL(FLOW_WC_SEQ
== 20);
496 fmd
->tun_id
= flow
->tunnel
.tun_id
;
497 fmd
->tun_src
= flow
->tunnel
.ip_src
;
498 fmd
->tun_dst
= flow
->tunnel
.ip_dst
;
499 fmd
->metadata
= flow
->metadata
;
500 memcpy(fmd
->regs
, flow
->regs
, sizeof fmd
->regs
);
501 fmd
->in_port
= flow
->in_port
;
505 flow_to_string(const struct flow
*flow
)
507 struct ds ds
= DS_EMPTY_INITIALIZER
;
508 flow_format(&ds
, flow
);
513 flow_tun_flag_to_string(uint32_t flags
)
516 case FLOW_TNL_F_DONT_FRAGMENT
:
518 case FLOW_TNL_F_CSUM
:
528 format_flags(struct ds
*ds
, const char *(*bit_to_string
)(uint32_t),
529 uint32_t flags
, char del
)
537 uint32_t bit
= rightmost_1bit(flags
);
540 s
= bit_to_string(bit
);
542 ds_put_format(ds
, "%s%c", s
, del
);
551 ds_put_format(ds
, "0x%"PRIx32
"%c", bad
, del
);
557 flow_format(struct ds
*ds
, const struct flow
*flow
)
561 match_wc_init(&match
, flow
);
562 match_format(&match
, ds
, OFP_DEFAULT_PRIORITY
);
566 flow_print(FILE *stream
, const struct flow
*flow
)
568 char *s
= flow_to_string(flow
);
573 /* flow_wildcards functions. */
575 /* Initializes 'wc' as a set of wildcards that matches every packet. */
577 flow_wildcards_init_catchall(struct flow_wildcards
*wc
)
579 memset(&wc
->masks
, 0, sizeof wc
->masks
);
582 /* Initializes 'wc' as an exact-match set of wildcards; that is, 'wc' does not
583 * wildcard any bits or fields. */
585 flow_wildcards_init_exact(struct flow_wildcards
*wc
)
587 memset(&wc
->masks
, 0xff, sizeof wc
->masks
);
588 memset(wc
->masks
.zeros
, 0, sizeof wc
->masks
.zeros
);
591 /* Returns true if 'wc' matches every packet, false if 'wc' fixes any bits or
594 flow_wildcards_is_catchall(const struct flow_wildcards
*wc
)
596 const uint32_t *wc_u32
= (const uint32_t *) &wc
->masks
;
599 for (i
= 0; i
< FLOW_U32S
; i
++) {
607 /* Sets 'dst' as the bitwise AND of wildcards in 'src1' and 'src2'.
608 * That is, a bit or a field is wildcarded in 'dst' if it is wildcarded
609 * in 'src1' or 'src2' or both. */
611 flow_wildcards_and(struct flow_wildcards
*dst
,
612 const struct flow_wildcards
*src1
,
613 const struct flow_wildcards
*src2
)
615 uint32_t *dst_u32
= (uint32_t *) &dst
->masks
;
616 const uint32_t *src1_u32
= (const uint32_t *) &src1
->masks
;
617 const uint32_t *src2_u32
= (const uint32_t *) &src2
->masks
;
620 for (i
= 0; i
< FLOW_U32S
; i
++) {
621 dst_u32
[i
] = src1_u32
[i
] & src2_u32
[i
];
625 /* Sets 'dst' as the bitwise OR of wildcards in 'src1' and 'src2'. That
626 * is, a bit or a field is wildcarded in 'dst' if it is neither
627 * wildcarded in 'src1' nor 'src2'. */
629 flow_wildcards_or(struct flow_wildcards
*dst
,
630 const struct flow_wildcards
*src1
,
631 const struct flow_wildcards
*src2
)
633 uint32_t *dst_u32
= (uint32_t *) &dst
->masks
;
634 const uint32_t *src1_u32
= (const uint32_t *) &src1
->masks
;
635 const uint32_t *src2_u32
= (const uint32_t *) &src2
->masks
;
638 for (i
= 0; i
< FLOW_U32S
; i
++) {
639 dst_u32
[i
] = src1_u32
[i
] | src2_u32
[i
];
643 /* Perform a bitwise OR of miniflow 'src' flow data with the equivalent
644 * fields in 'dst', storing the result in 'dst'. */
646 flow_union_with_miniflow(struct flow
*dst
, const struct miniflow
*src
)
648 uint32_t *dst_u32
= (uint32_t *) dst
;
653 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
656 for (map
= src
->map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
657 dst_u32
[raw_ctz(map
) + i
* 32] |= src
->values
[ofs
++];
662 /* Fold minimask 'mask''s wildcard mask into 'wc's wildcard mask. */
664 flow_wildcards_fold_minimask(struct flow_wildcards
*wc
,
665 const struct minimask
*mask
)
667 flow_union_with_miniflow(&wc
->masks
, &mask
->masks
);
670 /* Returns a hash of the wildcards in 'wc'. */
672 flow_wildcards_hash(const struct flow_wildcards
*wc
, uint32_t basis
)
674 return flow_hash(&wc
->masks
, basis
);
677 /* Returns true if 'a' and 'b' represent the same wildcards, false if they are
680 flow_wildcards_equal(const struct flow_wildcards
*a
,
681 const struct flow_wildcards
*b
)
683 return flow_equal(&a
->masks
, &b
->masks
);
686 /* Returns true if at least one bit or field is wildcarded in 'a' but not in
687 * 'b', false otherwise. */
689 flow_wildcards_has_extra(const struct flow_wildcards
*a
,
690 const struct flow_wildcards
*b
)
692 const uint32_t *a_u32
= (const uint32_t *) &a
->masks
;
693 const uint32_t *b_u32
= (const uint32_t *) &b
->masks
;
696 for (i
= 0; i
< FLOW_U32S
; i
++) {
697 if ((a_u32
[i
] & b_u32
[i
]) != b_u32
[i
]) {
704 /* Returns true if 'a' and 'b' are equal, except that 0-bits (wildcarded bits)
705 * in 'wc' do not need to be equal in 'a' and 'b'. */
707 flow_equal_except(const struct flow
*a
, const struct flow
*b
,
708 const struct flow_wildcards
*wc
)
710 const uint32_t *a_u32
= (const uint32_t *) a
;
711 const uint32_t *b_u32
= (const uint32_t *) b
;
712 const uint32_t *wc_u32
= (const uint32_t *) &wc
->masks
;
715 for (i
= 0; i
< FLOW_U32S
; i
++) {
716 if ((a_u32
[i
] ^ b_u32
[i
]) & wc_u32
[i
]) {
723 /* Sets the wildcard mask for register 'idx' in 'wc' to 'mask'.
724 * (A 0-bit indicates a wildcard bit.) */
726 flow_wildcards_set_reg_mask(struct flow_wildcards
*wc
, int idx
, uint32_t mask
)
728 wc
->masks
.regs
[idx
] = mask
;
731 /* Hashes 'flow' based on its L2 through L4 protocol information. */
733 flow_hash_symmetric_l4(const struct flow
*flow
, uint32_t basis
)
738 struct in6_addr ipv6_addr
;
743 uint8_t eth_addr
[ETH_ADDR_LEN
];
749 memset(&fields
, 0, sizeof fields
);
750 for (i
= 0; i
< ETH_ADDR_LEN
; i
++) {
751 fields
.eth_addr
[i
] = flow
->dl_src
[i
] ^ flow
->dl_dst
[i
];
753 fields
.vlan_tci
= flow
->vlan_tci
& htons(VLAN_VID_MASK
);
754 fields
.eth_type
= flow
->dl_type
;
756 /* UDP source and destination port are not taken into account because they
757 * will not necessarily be symmetric in a bidirectional flow. */
758 if (fields
.eth_type
== htons(ETH_TYPE_IP
)) {
759 fields
.ipv4_addr
= flow
->nw_src
^ flow
->nw_dst
;
760 fields
.ip_proto
= flow
->nw_proto
;
761 if (fields
.ip_proto
== IPPROTO_TCP
) {
762 fields
.tp_port
= flow
->tp_src
^ flow
->tp_dst
;
764 } else if (fields
.eth_type
== htons(ETH_TYPE_IPV6
)) {
765 const uint8_t *a
= &flow
->ipv6_src
.s6_addr
[0];
766 const uint8_t *b
= &flow
->ipv6_dst
.s6_addr
[0];
767 uint8_t *ipv6_addr
= &fields
.ipv6_addr
.s6_addr
[0];
769 for (i
=0; i
<16; i
++) {
770 ipv6_addr
[i
] = a
[i
] ^ b
[i
];
772 fields
.ip_proto
= flow
->nw_proto
;
773 if (fields
.ip_proto
== IPPROTO_TCP
) {
774 fields
.tp_port
= flow
->tp_src
^ flow
->tp_dst
;
777 return jhash_bytes(&fields
, sizeof fields
, basis
);
780 /* Masks the fields in 'wc' that are used by the flow hash 'fields'. */
782 flow_mask_hash_fields(struct flow_wildcards
*wc
, enum nx_hash_fields fields
)
785 case NX_HASH_FIELDS_ETH_SRC
:
786 memset(&wc
->masks
.dl_src
, 0xff, sizeof wc
->masks
.dl_src
);
789 case NX_HASH_FIELDS_SYMMETRIC_L4
:
790 memset(&wc
->masks
.dl_src
, 0xff, sizeof wc
->masks
.dl_src
);
791 memset(&wc
->masks
.dl_dst
, 0xff, sizeof wc
->masks
.dl_dst
);
792 memset(&wc
->masks
.dl_type
, 0xff, sizeof wc
->masks
.dl_type
);
793 memset(&wc
->masks
.vlan_tci
, 0xff, sizeof wc
->masks
.vlan_tci
);
794 memset(&wc
->masks
.nw_proto
, 0xff, sizeof wc
->masks
.nw_proto
);
795 memset(&wc
->masks
.nw_src
, 0xff, sizeof wc
->masks
.nw_src
);
796 memset(&wc
->masks
.nw_dst
, 0xff, sizeof wc
->masks
.nw_dst
);
797 memset(&wc
->masks
.tp_src
, 0xff, sizeof wc
->masks
.tp_src
);
798 memset(&wc
->masks
.tp_dst
, 0xff, sizeof wc
->masks
.tp_dst
);
806 /* Hashes the portions of 'flow' designated by 'fields'. */
808 flow_hash_fields(const struct flow
*flow
, enum nx_hash_fields fields
,
813 case NX_HASH_FIELDS_ETH_SRC
:
814 return jhash_bytes(flow
->dl_src
, sizeof flow
->dl_src
, basis
);
816 case NX_HASH_FIELDS_SYMMETRIC_L4
:
817 return flow_hash_symmetric_l4(flow
, basis
);
823 /* Returns a string representation of 'fields'. */
825 flow_hash_fields_to_str(enum nx_hash_fields fields
)
828 case NX_HASH_FIELDS_ETH_SRC
: return "eth_src";
829 case NX_HASH_FIELDS_SYMMETRIC_L4
: return "symmetric_l4";
830 default: return "<unknown>";
834 /* Returns true if the value of 'fields' is supported. Otherwise false. */
836 flow_hash_fields_valid(enum nx_hash_fields fields
)
838 return fields
== NX_HASH_FIELDS_ETH_SRC
839 || fields
== NX_HASH_FIELDS_SYMMETRIC_L4
;
842 /* Returns a hash value for the bits of 'flow' that are active based on
843 * 'wc', given 'basis'. */
845 flow_hash_in_wildcards(const struct flow
*flow
,
846 const struct flow_wildcards
*wc
, uint32_t basis
)
848 const uint32_t *wc_u32
= (const uint32_t *) &wc
->masks
;
849 const uint32_t *flow_u32
= (const uint32_t *) flow
;
854 for (i
= 0; i
< FLOW_U32S
; i
++) {
855 hash
= mhash_add(hash
, flow_u32
[i
] & wc_u32
[i
]);
857 return mhash_finish(hash
, 4 * FLOW_U32S
);
860 /* Sets the VLAN VID that 'flow' matches to 'vid', which is interpreted as an
861 * OpenFlow 1.0 "dl_vlan" value:
863 * - If it is in the range 0...4095, 'flow->vlan_tci' is set to match
864 * that VLAN. Any existing PCP match is unchanged (it becomes 0 if
865 * 'flow' previously matched packets without a VLAN header).
867 * - If it is OFP_VLAN_NONE, 'flow->vlan_tci' is set to match a packet
868 * without a VLAN tag.
870 * - Other values of 'vid' should not be used. */
872 flow_set_dl_vlan(struct flow
*flow
, ovs_be16 vid
)
874 if (vid
== htons(OFP10_VLAN_NONE
)) {
875 flow
->vlan_tci
= htons(0);
877 vid
&= htons(VLAN_VID_MASK
);
878 flow
->vlan_tci
&= ~htons(VLAN_VID_MASK
);
879 flow
->vlan_tci
|= htons(VLAN_CFI
) | vid
;
883 /* Sets the VLAN VID that 'flow' matches to 'vid', which is interpreted as an
884 * OpenFlow 1.2 "vlan_vid" value, that is, the low 13 bits of 'vlan_tci' (VID
887 flow_set_vlan_vid(struct flow
*flow
, ovs_be16 vid
)
889 ovs_be16 mask
= htons(VLAN_VID_MASK
| VLAN_CFI
);
890 flow
->vlan_tci
&= ~mask
;
891 flow
->vlan_tci
|= vid
& mask
;
894 /* Sets the VLAN PCP that 'flow' matches to 'pcp', which should be in the
897 * This function has no effect on the VLAN ID that 'flow' matches.
899 * After calling this function, 'flow' will not match packets without a VLAN
902 flow_set_vlan_pcp(struct flow
*flow
, uint8_t pcp
)
905 flow
->vlan_tci
&= ~htons(VLAN_PCP_MASK
);
906 flow
->vlan_tci
|= htons((pcp
<< VLAN_PCP_SHIFT
) | VLAN_CFI
);
909 /* Sets the MPLS Label that 'flow' matches to 'label', which is interpreted
910 * as an OpenFlow 1.1 "mpls_label" value. */
912 flow_set_mpls_label(struct flow
*flow
, ovs_be32 label
)
914 set_mpls_lse_label(&flow
->mpls_lse
, label
);
917 /* Sets the MPLS TTL that 'flow' matches to 'ttl', which should be in the
920 flow_set_mpls_ttl(struct flow
*flow
, uint8_t ttl
)
922 set_mpls_lse_ttl(&flow
->mpls_lse
, ttl
);
925 /* Sets the MPLS TC that 'flow' matches to 'tc', which should be in the
928 flow_set_mpls_tc(struct flow
*flow
, uint8_t tc
)
930 set_mpls_lse_tc(&flow
->mpls_lse
, tc
);
933 /* Sets the MPLS BOS bit that 'flow' matches to which should be 0 or 1. */
935 flow_set_mpls_bos(struct flow
*flow
, uint8_t bos
)
937 set_mpls_lse_bos(&flow
->mpls_lse
, bos
);
940 /* Puts into 'b' a packet that flow_extract() would parse as having the given
943 * (This is useful only for testing, obviously, and the packet isn't really
944 * valid. It hasn't got some checksums filled in, for one, and lots of fields
945 * are just zeroed.) */
947 flow_compose(struct ofpbuf
*b
, const struct flow
*flow
)
949 eth_compose(b
, flow
->dl_dst
, flow
->dl_src
, ntohs(flow
->dl_type
), 0);
950 if (flow
->dl_type
== htons(FLOW_DL_TYPE_NONE
)) {
951 struct eth_header
*eth
= b
->l2
;
952 eth
->eth_type
= htons(b
->size
);
956 if (flow
->vlan_tci
& htons(VLAN_CFI
)) {
957 eth_push_vlan(b
, flow
->vlan_tci
);
960 if (flow
->dl_type
== htons(ETH_TYPE_IP
)) {
961 struct ip_header
*ip
;
963 b
->l3
= ip
= ofpbuf_put_zeros(b
, sizeof *ip
);
964 ip
->ip_ihl_ver
= IP_IHL_VER(5, 4);
965 ip
->ip_tos
= flow
->nw_tos
;
966 ip
->ip_ttl
= flow
->nw_ttl
;
967 ip
->ip_proto
= flow
->nw_proto
;
968 ip
->ip_src
= flow
->nw_src
;
969 ip
->ip_dst
= flow
->nw_dst
;
971 if (flow
->nw_frag
& FLOW_NW_FRAG_ANY
) {
972 ip
->ip_frag_off
|= htons(IP_MORE_FRAGMENTS
);
973 if (flow
->nw_frag
& FLOW_NW_FRAG_LATER
) {
974 ip
->ip_frag_off
|= htons(100);
977 if (!(flow
->nw_frag
& FLOW_NW_FRAG_ANY
)
978 || !(flow
->nw_frag
& FLOW_NW_FRAG_LATER
)) {
979 if (flow
->nw_proto
== IPPROTO_TCP
) {
980 struct tcp_header
*tcp
;
982 b
->l4
= tcp
= ofpbuf_put_zeros(b
, sizeof *tcp
);
983 tcp
->tcp_src
= flow
->tp_src
;
984 tcp
->tcp_dst
= flow
->tp_dst
;
985 tcp
->tcp_ctl
= TCP_CTL(0, 5);
986 } else if (flow
->nw_proto
== IPPROTO_UDP
) {
987 struct udp_header
*udp
;
989 b
->l4
= udp
= ofpbuf_put_zeros(b
, sizeof *udp
);
990 udp
->udp_src
= flow
->tp_src
;
991 udp
->udp_dst
= flow
->tp_dst
;
992 } else if (flow
->nw_proto
== IPPROTO_ICMP
) {
993 struct icmp_header
*icmp
;
995 b
->l4
= icmp
= ofpbuf_put_zeros(b
, sizeof *icmp
);
996 icmp
->icmp_type
= ntohs(flow
->tp_src
);
997 icmp
->icmp_code
= ntohs(flow
->tp_dst
);
998 icmp
->icmp_csum
= csum(icmp
, ICMP_HEADER_LEN
);
1003 ip
->ip_tot_len
= htons((uint8_t *) b
->data
+ b
->size
1004 - (uint8_t *) b
->l3
);
1005 ip
->ip_csum
= csum(ip
, sizeof *ip
);
1006 } else if (flow
->dl_type
== htons(ETH_TYPE_IPV6
)) {
1008 } else if (flow
->dl_type
== htons(ETH_TYPE_ARP
) ||
1009 flow
->dl_type
== htons(ETH_TYPE_RARP
)) {
1010 struct arp_eth_header
*arp
;
1012 b
->l3
= arp
= ofpbuf_put_zeros(b
, sizeof *arp
);
1013 arp
->ar_hrd
= htons(1);
1014 arp
->ar_pro
= htons(ETH_TYPE_IP
);
1015 arp
->ar_hln
= ETH_ADDR_LEN
;
1017 arp
->ar_op
= htons(flow
->nw_proto
);
1019 if (flow
->nw_proto
== ARP_OP_REQUEST
||
1020 flow
->nw_proto
== ARP_OP_REPLY
) {
1021 arp
->ar_spa
= flow
->nw_src
;
1022 arp
->ar_tpa
= flow
->nw_dst
;
1023 memcpy(arp
->ar_sha
, flow
->arp_sha
, ETH_ADDR_LEN
);
1024 memcpy(arp
->ar_tha
, flow
->arp_tha
, ETH_ADDR_LEN
);
1028 if (eth_type_mpls(flow
->dl_type
)) {
1030 push_mpls(b
, flow
->dl_type
, flow
->mpls_lse
);
1034 /* Compressed flow. */
1037 miniflow_n_values(const struct miniflow
*flow
)
1042 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1043 n
+= popcount(flow
->map
[i
]);
1049 miniflow_alloc_values(struct miniflow
*flow
, int n
)
1051 if (n
<= MINI_N_INLINE
) {
1052 return flow
->inline_values
;
1054 COVERAGE_INC(miniflow_malloc
);
1055 return xmalloc(n
* sizeof *flow
->values
);
1059 /* Initializes 'dst' as a copy of 'src'. The caller must eventually free 'dst'
1060 * with miniflow_destroy(). */
1062 miniflow_init(struct miniflow
*dst
, const struct flow
*src
)
1064 const uint32_t *src_u32
= (const uint32_t *) src
;
1069 /* Initialize dst->map, counting the number of nonzero elements. */
1071 memset(dst
->map
, 0, sizeof dst
->map
);
1072 for (i
= 0; i
< FLOW_U32S
; i
++) {
1074 dst
->map
[i
/ 32] |= 1u << (i
% 32);
1079 /* Initialize dst->values. */
1080 dst
->values
= miniflow_alloc_values(dst
, n
);
1082 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1085 for (map
= dst
->map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
1086 dst
->values
[ofs
++] = src_u32
[raw_ctz(map
) + i
* 32];
1091 /* Initializes 'dst' as a copy of 'src'. The caller must eventually free 'dst'
1092 * with miniflow_destroy(). */
1094 miniflow_clone(struct miniflow
*dst
, const struct miniflow
*src
)
1096 int n
= miniflow_n_values(src
);
1097 memcpy(dst
->map
, src
->map
, sizeof dst
->map
);
1098 dst
->values
= miniflow_alloc_values(dst
, n
);
1099 memcpy(dst
->values
, src
->values
, n
* sizeof *dst
->values
);
1102 /* Frees any memory owned by 'flow'. Does not free the storage in which 'flow'
1103 * itself resides; the caller is responsible for that. */
1105 miniflow_destroy(struct miniflow
*flow
)
1107 if (flow
->values
!= flow
->inline_values
) {
1112 /* Initializes 'dst' as a copy of 'src'. */
1114 miniflow_expand(const struct miniflow
*src
, struct flow
*dst
)
1116 memset(dst
, 0, sizeof *dst
);
1117 flow_union_with_miniflow(dst
, src
);
1120 static const uint32_t *
1121 miniflow_get__(const struct miniflow
*flow
, unsigned int u32_ofs
)
1123 if (!(flow
->map
[u32_ofs
/ 32] & (1u << (u32_ofs
% 32)))) {
1124 static const uint32_t zero
= 0;
1127 const uint32_t *p
= flow
->values
;
1129 BUILD_ASSERT(MINI_N_MAPS
== 2);
1131 p
+= popcount(flow
->map
[0] & ((1u << u32_ofs
) - 1));
1133 p
+= popcount(flow
->map
[0]);
1134 p
+= popcount(flow
->map
[1] & ((1u << (u32_ofs
- 32)) - 1));
1140 /* Returns the uint32_t that would be at byte offset '4 * u32_ofs' if 'flow'
1141 * were expanded into a "struct flow". */
1143 miniflow_get(const struct miniflow
*flow
, unsigned int u32_ofs
)
1145 return *miniflow_get__(flow
, u32_ofs
);
1148 /* Returns the ovs_be16 that would be at byte offset 'u8_ofs' if 'flow' were
1149 * expanded into a "struct flow". */
1151 miniflow_get_be16(const struct miniflow
*flow
, unsigned int u8_ofs
)
1153 const uint32_t *u32p
= miniflow_get__(flow
, u8_ofs
/ 4);
1154 const ovs_be16
*be16p
= (const ovs_be16
*) u32p
;
1155 return be16p
[u8_ofs
% 4 != 0];
1158 /* Returns the VID within the vlan_tci member of the "struct flow" represented
1161 miniflow_get_vid(const struct miniflow
*flow
)
1163 ovs_be16 tci
= miniflow_get_be16(flow
, offsetof(struct flow
, vlan_tci
));
1164 return vlan_tci_to_vid(tci
);
1167 /* Returns true if 'a' and 'b' are the same flow, false otherwise. */
1169 miniflow_equal(const struct miniflow
*a
, const struct miniflow
*b
)
1173 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1174 if (a
->map
[i
] != b
->map
[i
]) {
1179 return !memcmp(a
->values
, b
->values
,
1180 miniflow_n_values(a
) * sizeof *a
->values
);
1183 /* Returns true if 'a' and 'b' are equal at the places where there are 1-bits
1184 * in 'mask', false if they differ. */
1186 miniflow_equal_in_minimask(const struct miniflow
*a
, const struct miniflow
*b
,
1187 const struct minimask
*mask
)
1192 p
= mask
->masks
.values
;
1193 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1196 for (map
= mask
->masks
.map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
1197 int ofs
= raw_ctz(map
) + i
* 32;
1199 if ((miniflow_get(a
, ofs
) ^ miniflow_get(b
, ofs
)) & *p
) {
1209 /* Returns true if 'a' and 'b' are equal at the places where there are 1-bits
1210 * in 'mask', false if they differ. */
1212 miniflow_equal_flow_in_minimask(const struct miniflow
*a
, const struct flow
*b
,
1213 const struct minimask
*mask
)
1215 const uint32_t *b_u32
= (const uint32_t *) b
;
1219 p
= mask
->masks
.values
;
1220 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1223 for (map
= mask
->masks
.map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
1224 int ofs
= raw_ctz(map
) + i
* 32;
1226 if ((miniflow_get(a
, ofs
) ^ b_u32
[ofs
]) & *p
) {
1236 /* Returns a hash value for 'flow', given 'basis'. */
1238 miniflow_hash(const struct miniflow
*flow
, uint32_t basis
)
1240 BUILD_ASSERT_DECL(MINI_N_MAPS
== 2);
1241 return hash_3words(flow
->map
[0], flow
->map
[1],
1242 hash_words(flow
->values
, miniflow_n_values(flow
),
1246 /* Returns a hash value for the bits of 'flow' where there are 1-bits in
1247 * 'mask', given 'basis'.
1249 * The hash values returned by this function are the same as those returned by
1250 * flow_hash_in_minimask(), only the form of the arguments differ. */
1252 miniflow_hash_in_minimask(const struct miniflow
*flow
,
1253 const struct minimask
*mask
, uint32_t basis
)
1255 const uint32_t *p
= mask
->masks
.values
;
1260 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1263 for (map
= mask
->masks
.map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
1264 int ofs
= raw_ctz(map
) + i
* 32;
1266 hash
= mhash_add(hash
, miniflow_get(flow
, ofs
) & *p
);
1271 return mhash_finish(hash
, (p
- mask
->masks
.values
) * 4);
1274 /* Returns a hash value for the bits of 'flow' where there are 1-bits in
1275 * 'mask', given 'basis'.
1277 * The hash values returned by this function are the same as those returned by
1278 * miniflow_hash_in_minimask(), only the form of the arguments differ. */
1280 flow_hash_in_minimask(const struct flow
*flow
, const struct minimask
*mask
,
1283 const uint32_t *flow_u32
= (const uint32_t *) flow
;
1284 const uint32_t *p
= mask
->masks
.values
;
1289 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1292 for (map
= mask
->masks
.map
[i
]; map
; map
= zero_rightmost_1bit(map
)) {
1293 int ofs
= raw_ctz(map
) + i
* 32;
1295 hash
= mhash_add(hash
, flow_u32
[ofs
] & *p
);
1300 return mhash_finish(hash
, (p
- mask
->masks
.values
) * 4);
1303 /* Initializes 'dst' as a copy of 'src'. The caller must eventually free 'dst'
1304 * with minimask_destroy(). */
1306 minimask_init(struct minimask
*mask
, const struct flow_wildcards
*wc
)
1308 miniflow_init(&mask
->masks
, &wc
->masks
);
1311 /* Initializes 'dst' as a copy of 'src'. The caller must eventually free 'dst'
1312 * with minimask_destroy(). */
1314 minimask_clone(struct minimask
*dst
, const struct minimask
*src
)
1316 miniflow_clone(&dst
->masks
, &src
->masks
);
1319 /* Initializes 'dst_' as the bit-wise "and" of 'a_' and 'b_'.
1321 * The caller must provide room for FLOW_U32S "uint32_t"s in 'storage', for use
1322 * by 'dst_'. The caller must *not* free 'dst_' with minimask_destroy(). */
1324 minimask_combine(struct minimask
*dst_
,
1325 const struct minimask
*a_
, const struct minimask
*b_
,
1326 uint32_t storage
[FLOW_U32S
])
1328 struct miniflow
*dst
= &dst_
->masks
;
1329 const struct miniflow
*a
= &a_
->masks
;
1330 const struct miniflow
*b
= &b_
->masks
;
1334 dst
->values
= storage
;
1335 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1339 for (map
= a
->map
[i
] & b
->map
[i
]; map
;
1340 map
= zero_rightmost_1bit(map
)) {
1341 int ofs
= raw_ctz(map
) + i
* 32;
1342 uint32_t mask
= miniflow_get(a
, ofs
) & miniflow_get(b
, ofs
);
1345 dst
->map
[i
] |= rightmost_1bit(map
);
1346 dst
->values
[n
++] = mask
;
1352 /* Frees any memory owned by 'mask'. Does not free the storage in which 'mask'
1353 * itself resides; the caller is responsible for that. */
1355 minimask_destroy(struct minimask
*mask
)
1357 miniflow_destroy(&mask
->masks
);
1360 /* Initializes 'dst' as a copy of 'src'. */
1362 minimask_expand(const struct minimask
*mask
, struct flow_wildcards
*wc
)
1364 miniflow_expand(&mask
->masks
, &wc
->masks
);
1367 /* Returns the uint32_t that would be at byte offset '4 * u32_ofs' if 'mask'
1368 * were expanded into a "struct flow_wildcards". */
1370 minimask_get(const struct minimask
*mask
, unsigned int u32_ofs
)
1372 return miniflow_get(&mask
->masks
, u32_ofs
);
1375 /* Returns the VID mask within the vlan_tci member of the "struct
1376 * flow_wildcards" represented by 'mask'. */
1378 minimask_get_vid_mask(const struct minimask
*mask
)
1380 return miniflow_get_vid(&mask
->masks
);
1383 /* Returns true if 'a' and 'b' are the same flow mask, false otherwise. */
1385 minimask_equal(const struct minimask
*a
, const struct minimask
*b
)
1387 return miniflow_equal(&a
->masks
, &b
->masks
);
1390 /* Returns a hash value for 'mask', given 'basis'. */
1392 minimask_hash(const struct minimask
*mask
, uint32_t basis
)
1394 return miniflow_hash(&mask
->masks
, basis
);
1397 /* Returns true if at least one bit is wildcarded in 'a_' but not in 'b_',
1398 * false otherwise. */
1400 minimask_has_extra(const struct minimask
*a_
, const struct minimask
*b_
)
1402 const struct miniflow
*a
= &a_
->masks
;
1403 const struct miniflow
*b
= &b_
->masks
;
1406 for (i
= 0; i
< MINI_N_MAPS
; i
++) {
1409 for (map
= a
->map
[i
] | b
->map
[i
]; map
;
1410 map
= zero_rightmost_1bit(map
)) {
1411 int ofs
= raw_ctz(map
) + i
* 32;
1412 uint32_t a_u32
= miniflow_get(a
, ofs
);
1413 uint32_t b_u32
= miniflow_get(b
, ofs
);
1415 if ((a_u32
& b_u32
) != b_u32
) {
1424 /* Returns true if 'mask' matches every packet, false if 'mask' fixes any bits
1427 minimask_is_catchall(const struct minimask
*mask_
)
1429 const struct miniflow
*mask
= &mask_
->masks
;
1431 BUILD_ASSERT(MINI_N_MAPS
== 2);
1432 return !(mask
->map
[0] | mask
->map
[1]);