2 * Copyright (c) 2010, 2011, 2012 Nicira, Inc.
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at:
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
21 #include <netinet/icmp6.h>
23 #include "classifier.h"
24 #include "dynamic-string.h"
25 #include "meta-flow.h"
26 #include "ofp-actions.h"
27 #include "ofp-errors.h"
30 #include "openflow/nicira-ext.h"
32 #include "unaligned.h"
36 VLOG_DEFINE_THIS_MODULE(nx_match
);
38 /* Rate limit for nx_match parse errors. These always indicate a bug in the
39 * peer and so there's not much point in showing a lot of them. */
40 static struct vlog_rate_limit rl
= VLOG_RATE_LIMIT_INIT(1, 5);
42 /* Returns the width of the data for a field with the given 'header', in
45 nxm_field_bytes(uint32_t header
)
47 unsigned int length
= NXM_LENGTH(header
);
48 return NXM_HASMASK(header
) ? length
/ 2 : length
;
51 /* Returns the width of the data for a field with the given 'header', in
54 nxm_field_bits(uint32_t header
)
56 return nxm_field_bytes(header
) * 8;
59 /* nx_pull_match() and helpers. */
62 nx_entry_ok(const void *p
, unsigned int match_len
)
64 unsigned int payload_len
;
70 VLOG_DBG_RL(&rl
, "nx_match ends with partial (%u-byte) nxm_header",
75 memcpy(&header_be
, p
, 4);
76 header
= ntohl(header_be
);
78 payload_len
= NXM_LENGTH(header
);
80 VLOG_DBG_RL(&rl
, "nxm_entry %08"PRIx32
" has invalid payload "
84 if (match_len
< payload_len
+ 4) {
85 VLOG_DBG_RL(&rl
, "%"PRIu32
"-byte nxm_entry but only "
86 "%u bytes left in nx_match", payload_len
+ 4, match_len
);
94 nx_pull_match__(struct ofpbuf
*b
, unsigned int match_len
, bool strict
,
95 uint16_t priority
, struct cls_rule
*rule
,
96 ovs_be64
*cookie
, ovs_be64
*cookie_mask
)
101 assert((cookie
!= NULL
) == (cookie_mask
!= NULL
));
103 cls_rule_init_catchall(rule
, priority
);
105 *cookie
= *cookie_mask
= htonll(0);
111 p
= ofpbuf_try_pull(b
, ROUND_UP(match_len
, 8));
113 VLOG_DBG_RL(&rl
, "nx_match length %u, rounded up to a "
114 "multiple of 8, is longer than space in message (max "
115 "length %zu)", match_len
, b
->size
);
116 return OFPERR_OFPBMC_BAD_LEN
;
120 (header
= nx_entry_ok(p
, match_len
)) != 0;
121 p
+= 4 + NXM_LENGTH(header
), match_len
-= 4 + NXM_LENGTH(header
)) {
122 const struct mf_field
*mf
;
125 mf
= mf_from_nxm_header(header
);
128 error
= OFPERR_OFPBMC_BAD_FIELD
;
132 } else if (!mf_are_prereqs_ok(mf
, &rule
->flow
)) {
133 error
= OFPERR_OFPBMC_BAD_PREREQ
;
134 } else if (!mf_is_all_wild(mf
, &rule
->wc
)) {
135 error
= OFPERR_OFPBMC_DUP_FIELD
;
136 } else if (header
!= OXM_OF_IN_PORT
) {
137 unsigned int width
= mf
->n_bytes
;
138 union mf_value value
;
140 memcpy(&value
, p
+ 4, width
);
141 if (!mf_is_value_valid(mf
, &value
)) {
142 error
= OFPERR_OFPBMC_BAD_VALUE
;
143 } else if (!NXM_HASMASK(header
)) {
145 mf_set_value(mf
, &value
, rule
);
149 memcpy(&mask
, p
+ 4 + width
, width
);
150 if (!mf_is_mask_valid(mf
, &mask
)) {
151 error
= OFPERR_OFPBMC_BAD_MASK
;
154 mf_set(mf
, &value
, &mask
, rule
);
158 /* Special case for 32bit ports when using OXM,
159 * ports are 16 bits wide otherwise. */
163 memcpy(&port_of11
, p
+ 4, sizeof port_of11
);
164 error
= ofputil_port_from_ofp11(port_of11
, &port
);
166 cls_rule_set_in_port(rule
, port
);
170 /* Check if the match is for a cookie rather than a classifier rule. */
171 if ((header
== NXM_NX_COOKIE
|| header
== NXM_NX_COOKIE_W
) && cookie
) {
173 error
= OFPERR_OFPBMC_DUP_FIELD
;
175 unsigned int width
= sizeof *cookie
;
177 memcpy(cookie
, p
+ 4, width
);
178 if (NXM_HASMASK(header
)) {
179 memcpy(cookie_mask
, p
+ 4 + width
, width
);
181 *cookie_mask
= htonll(UINT64_MAX
);
188 VLOG_DBG_RL(&rl
, "bad nxm_entry %#08"PRIx32
" (vendor=%"PRIu32
", "
189 "field=%"PRIu32
", hasmask=%"PRIu32
", len=%"PRIu32
"), "
191 NXM_VENDOR(header
), NXM_FIELD(header
),
192 NXM_HASMASK(header
), NXM_LENGTH(header
),
193 ofperr_to_string(error
));
198 return match_len
? OFPERR_OFPBMC_BAD_LEN
: 0;
201 /* Parses the nx_match formatted match description in 'b' with length
202 * 'match_len'. The results are stored in 'rule', which is initialized with
203 * 'priority'. If 'cookie' and 'cookie_mask' contain valid pointers, then the
204 * cookie and mask will be stored in them if a "NXM_NX_COOKIE*" match is
205 * defined. Otherwise, 0 is stored in both.
207 * Fails with an error when encountering unknown NXM headers.
209 * Returns 0 if successful, otherwise an OpenFlow error code. */
211 nx_pull_match(struct ofpbuf
*b
, unsigned int match_len
,
212 uint16_t priority
, struct cls_rule
*rule
,
213 ovs_be64
*cookie
, ovs_be64
*cookie_mask
)
215 return nx_pull_match__(b
, match_len
, true, priority
, rule
, cookie
,
219 /* Behaves the same as nx_pull_match() with one exception. Skips over unknown
220 * NXM headers instead of failing with an error when they are encountered. */
222 nx_pull_match_loose(struct ofpbuf
*b
, unsigned int match_len
,
223 uint16_t priority
, struct cls_rule
*rule
,
224 ovs_be64
*cookie
, ovs_be64
*cookie_mask
)
226 return nx_pull_match__(b
, match_len
, false, priority
, rule
, cookie
,
230 /* nx_put_match() and helpers.
232 * 'put' functions whose names end in 'w' add a wildcarded field.
233 * 'put' functions whose names end in 'm' add a field that might be wildcarded.
234 * Other 'put' functions add exact-match fields.
238 nxm_put_header(struct ofpbuf
*b
, uint32_t header
)
240 ovs_be32 n_header
= htonl(header
);
241 ofpbuf_put(b
, &n_header
, sizeof n_header
);
245 nxm_put_8(struct ofpbuf
*b
, uint32_t header
, uint8_t value
)
247 nxm_put_header(b
, header
);
248 ofpbuf_put(b
, &value
, sizeof value
);
252 nxm_put_8m(struct ofpbuf
*b
, uint32_t header
, uint8_t value
, uint8_t mask
)
259 nxm_put_8(b
, header
, value
);
263 nxm_put_header(b
, NXM_MAKE_WILD_HEADER(header
));
264 ofpbuf_put(b
, &value
, sizeof value
);
265 ofpbuf_put(b
, &mask
, sizeof mask
);
270 nxm_put_16(struct ofpbuf
*b
, uint32_t header
, ovs_be16 value
)
272 nxm_put_header(b
, header
);
273 ofpbuf_put(b
, &value
, sizeof value
);
277 nxm_put_16w(struct ofpbuf
*b
, uint32_t header
, ovs_be16 value
, ovs_be16 mask
)
279 nxm_put_header(b
, header
);
280 ofpbuf_put(b
, &value
, sizeof value
);
281 ofpbuf_put(b
, &mask
, sizeof mask
);
285 nxm_put_16m(struct ofpbuf
*b
, uint32_t header
, ovs_be16 value
, ovs_be16 mask
)
291 case CONSTANT_HTONS(UINT16_MAX
):
292 nxm_put_16(b
, header
, value
);
296 nxm_put_16w(b
, NXM_MAKE_WILD_HEADER(header
), value
, mask
);
302 nxm_put_32(struct ofpbuf
*b
, uint32_t header
, ovs_be32 value
)
304 nxm_put_header(b
, header
);
305 ofpbuf_put(b
, &value
, sizeof value
);
309 nxm_put_32w(struct ofpbuf
*b
, uint32_t header
, ovs_be32 value
, ovs_be32 mask
)
311 nxm_put_header(b
, header
);
312 ofpbuf_put(b
, &value
, sizeof value
);
313 ofpbuf_put(b
, &mask
, sizeof mask
);
317 nxm_put_32m(struct ofpbuf
*b
, uint32_t header
, ovs_be32 value
, ovs_be32 mask
)
323 case CONSTANT_HTONL(UINT32_MAX
):
324 nxm_put_32(b
, header
, value
);
328 nxm_put_32w(b
, NXM_MAKE_WILD_HEADER(header
), value
, mask
);
334 nxm_put_64(struct ofpbuf
*b
, uint32_t header
, ovs_be64 value
)
336 nxm_put_header(b
, header
);
337 ofpbuf_put(b
, &value
, sizeof value
);
341 nxm_put_64w(struct ofpbuf
*b
, uint32_t header
, ovs_be64 value
, ovs_be64 mask
)
343 nxm_put_header(b
, header
);
344 ofpbuf_put(b
, &value
, sizeof value
);
345 ofpbuf_put(b
, &mask
, sizeof mask
);
349 nxm_put_64m(struct ofpbuf
*b
, uint32_t header
, ovs_be64 value
, ovs_be64 mask
)
355 case CONSTANT_HTONLL(UINT64_MAX
):
356 nxm_put_64(b
, header
, value
);
360 nxm_put_64w(b
, NXM_MAKE_WILD_HEADER(header
), value
, mask
);
366 nxm_put_eth(struct ofpbuf
*b
, uint32_t header
,
367 const uint8_t value
[ETH_ADDR_LEN
])
369 nxm_put_header(b
, header
);
370 ofpbuf_put(b
, value
, ETH_ADDR_LEN
);
374 nxm_put_eth_masked(struct ofpbuf
*b
, uint32_t header
,
375 const uint8_t value
[ETH_ADDR_LEN
],
376 const uint8_t mask
[ETH_ADDR_LEN
])
378 if (!eth_addr_is_zero(mask
)) {
379 if (eth_mask_is_exact(mask
)) {
380 nxm_put_eth(b
, header
, value
);
382 nxm_put_header(b
, NXM_MAKE_WILD_HEADER(header
));
383 ofpbuf_put(b
, value
, ETH_ADDR_LEN
);
384 ofpbuf_put(b
, mask
, ETH_ADDR_LEN
);
390 nxm_put_ipv6(struct ofpbuf
*b
, uint32_t header
,
391 const struct in6_addr
*value
, const struct in6_addr
*mask
)
393 if (ipv6_mask_is_any(mask
)) {
395 } else if (ipv6_mask_is_exact(mask
)) {
396 nxm_put_header(b
, header
);
397 ofpbuf_put(b
, value
, sizeof *value
);
399 nxm_put_header(b
, NXM_MAKE_WILD_HEADER(header
));
400 ofpbuf_put(b
, value
, sizeof *value
);
401 ofpbuf_put(b
, mask
, sizeof *mask
);
406 nxm_put_frag(struct ofpbuf
*b
, const struct cls_rule
*cr
)
408 uint8_t nw_frag
= cr
->flow
.nw_frag
;
409 uint8_t nw_frag_mask
= cr
->wc
.nw_frag_mask
;
411 switch (nw_frag_mask
) {
415 case FLOW_NW_FRAG_MASK
:
416 nxm_put_8(b
, NXM_NX_IP_FRAG
, nw_frag
);
420 nxm_put_8m(b
, NXM_NX_IP_FRAG
, nw_frag
,
421 nw_frag_mask
& FLOW_NW_FRAG_MASK
);
427 nxm_put_ip(struct ofpbuf
*b
, const struct cls_rule
*cr
,
428 uint8_t icmp_proto
, uint32_t icmp_type
, uint32_t icmp_code
,
431 const flow_wildcards_t wc
= cr
->wc
.wildcards
;
432 const struct flow
*flow
= &cr
->flow
;
436 if (!(wc
& FWW_NW_DSCP
)) {
437 nxm_put_8(b
, oxm
? OXM_OF_IP_DSCP
: NXM_OF_IP_TOS
,
438 flow
->nw_tos
& IP_DSCP_MASK
);
441 if (!(wc
& FWW_NW_ECN
)) {
442 nxm_put_8(b
, oxm
? OXM_OF_IP_ECN
: NXM_NX_IP_ECN
,
443 flow
->nw_tos
& IP_ECN_MASK
);
446 if (!oxm
&& !(wc
& FWW_NW_TTL
)) {
447 nxm_put_8(b
, NXM_NX_IP_TTL
, flow
->nw_ttl
);
450 if (!(wc
& FWW_NW_PROTO
)) {
451 nxm_put_8(b
, oxm
? OXM_OF_IP_PROTO
: NXM_OF_IP_PROTO
, flow
->nw_proto
);
453 if (flow
->nw_proto
== IPPROTO_TCP
) {
454 nxm_put_16m(b
, oxm
? OXM_OF_TCP_SRC
: NXM_OF_TCP_SRC
,
455 flow
->tp_src
, cr
->wc
.tp_src_mask
);
456 nxm_put_16m(b
, oxm
? OXM_OF_TCP_DST
: NXM_OF_TCP_DST
,
457 flow
->tp_dst
, cr
->wc
.tp_dst_mask
);
458 } else if (flow
->nw_proto
== IPPROTO_UDP
) {
459 nxm_put_16m(b
, oxm
? OXM_OF_UDP_SRC
: NXM_OF_UDP_SRC
,
460 flow
->tp_src
, cr
->wc
.tp_src_mask
);
461 nxm_put_16m(b
, oxm
? OXM_OF_UDP_DST
: NXM_OF_UDP_DST
,
462 flow
->tp_dst
, cr
->wc
.tp_dst_mask
);
463 } else if (flow
->nw_proto
== icmp_proto
) {
464 if (cr
->wc
.tp_src_mask
) {
465 nxm_put_8(b
, icmp_type
, ntohs(flow
->tp_src
));
467 if (cr
->wc
.tp_dst_mask
) {
468 nxm_put_8(b
, icmp_code
, ntohs(flow
->tp_dst
));
474 /* Appends to 'b' the nx_match format that expresses 'cr' (except for
475 * 'cr->priority', because priority is not part of nx_match), plus enough
476 * zero bytes to pad the nx_match out to a multiple of 8. For Flow Mod
477 * and Flow Stats Requests messages, a 'cookie' and 'cookie_mask' may be
478 * supplied. Otherwise, 'cookie_mask' should be zero.
480 * This function can cause 'b''s data to be reallocated.
482 * Returns the number of bytes appended to 'b', excluding padding.
484 * If 'cr' is a catch-all rule that matches every packet, then this function
485 * appends nothing to 'b' and returns 0. */
487 nx_put_match(struct ofpbuf
*b
, bool oxm
, const struct cls_rule
*cr
,
488 ovs_be64 cookie
, ovs_be64 cookie_mask
)
490 const flow_wildcards_t wc
= cr
->wc
.wildcards
;
491 const struct flow
*flow
= &cr
->flow
;
492 const size_t start_len
= b
->size
;
496 BUILD_ASSERT_DECL(FLOW_WC_SEQ
== 12);
499 if (!(wc
& FWW_IN_PORT
)) {
500 uint16_t in_port
= flow
->in_port
;
502 nxm_put_32(b
, OXM_OF_IN_PORT
, ofputil_port_to_ofp11(in_port
));
504 nxm_put_16(b
, NXM_OF_IN_PORT
, htons(in_port
));
509 nxm_put_eth_masked(b
, oxm
? OXM_OF_ETH_SRC
: NXM_OF_ETH_SRC
,
510 flow
->dl_src
, cr
->wc
.dl_src_mask
);
511 nxm_put_eth_masked(b
, oxm
? OXM_OF_ETH_DST
: NXM_OF_ETH_DST
,
512 flow
->dl_dst
, cr
->wc
.dl_dst_mask
);
513 if (!(wc
& FWW_DL_TYPE
)) {
514 nxm_put_16(b
, oxm
? OXM_OF_ETH_TYPE
: NXM_OF_ETH_TYPE
,
515 ofputil_dl_type_to_openflow(flow
->dl_type
));
520 * XXX missing OXM support */
521 nxm_put_16m(b
, NXM_OF_VLAN_TCI
, flow
->vlan_tci
, cr
->wc
.vlan_tci_mask
);
524 if (!(wc
& FWW_DL_TYPE
) && flow
->dl_type
== htons(ETH_TYPE_IP
)) {
526 nxm_put_32m(b
, oxm
? OXM_OF_IPV4_SRC
: NXM_OF_IP_SRC
,
527 flow
->nw_src
, cr
->wc
.nw_src_mask
);
528 nxm_put_32m(b
, oxm
? OXM_OF_IPV4_DST
: NXM_OF_IP_DST
,
529 flow
->nw_dst
, cr
->wc
.nw_dst_mask
);
530 nxm_put_ip(b
, cr
, IPPROTO_ICMP
,
531 oxm
? OXM_OF_ICMPV4_TYPE
: NXM_OF_ICMP_TYPE
,
532 oxm
? OXM_OF_ICMPV4_CODE
: NXM_OF_ICMP_CODE
, oxm
);
533 } else if (!(wc
& FWW_DL_TYPE
) && flow
->dl_type
== htons(ETH_TYPE_IPV6
)) {
535 nxm_put_ipv6(b
, oxm
? OXM_OF_IPV6_SRC
: NXM_NX_IPV6_SRC
,
536 &flow
->ipv6_src
, &cr
->wc
.ipv6_src_mask
);
537 nxm_put_ipv6(b
, oxm
? OXM_OF_IPV6_DST
: NXM_NX_IPV6_DST
,
538 &flow
->ipv6_dst
, &cr
->wc
.ipv6_dst_mask
);
539 nxm_put_ip(b
, cr
, IPPROTO_ICMPV6
,
540 oxm
? OXM_OF_ICMPV6_TYPE
: NXM_NX_ICMPV6_TYPE
,
541 oxm
? OXM_OF_ICMPV6_CODE
: NXM_NX_ICMPV6_CODE
, oxm
);
543 if (!(wc
& FWW_IPV6_LABEL
)) {
544 nxm_put_32(b
, oxm
? OXM_OF_IPV6_FLABEL
: NXM_NX_IPV6_LABEL
,
548 if (flow
->nw_proto
== IPPROTO_ICMPV6
549 && (flow
->tp_src
== htons(ND_NEIGHBOR_SOLICIT
) ||
550 flow
->tp_src
== htons(ND_NEIGHBOR_ADVERT
))) {
551 nxm_put_ipv6(b
, oxm
? OXM_OF_IPV6_ND_TARGET
: NXM_NX_ND_TARGET
,
552 &flow
->nd_target
, &cr
->wc
.nd_target_mask
);
553 if (!(wc
& FWW_ARP_SHA
)
554 && flow
->tp_src
== htons(ND_NEIGHBOR_SOLICIT
)) {
555 nxm_put_eth(b
, oxm
? OXM_OF_IPV6_ND_SLL
: NXM_NX_ND_SLL
,
558 if (!(wc
& FWW_ARP_THA
)
559 && flow
->tp_src
== htons(ND_NEIGHBOR_ADVERT
)) {
560 nxm_put_eth(b
, oxm
? OXM_OF_IPV6_ND_TLL
: NXM_NX_ND_TLL
,
564 } else if (!(wc
& FWW_DL_TYPE
) && flow
->dl_type
== htons(ETH_TYPE_ARP
)) {
566 if (!(wc
& FWW_NW_PROTO
)) {
567 nxm_put_16(b
, oxm
? OXM_OF_ARP_OP
: NXM_OF_ARP_OP
,
568 htons(flow
->nw_proto
));
570 nxm_put_32m(b
, oxm
? OXM_OF_ARP_SPA
: NXM_OF_ARP_SPA
,
571 flow
->nw_src
, cr
->wc
.nw_src_mask
);
572 nxm_put_32m(b
, oxm
? OXM_OF_ARP_TPA
: NXM_OF_ARP_TPA
,
573 flow
->nw_dst
, cr
->wc
.nw_dst_mask
);
574 if (!(wc
& FWW_ARP_SHA
)) {
575 nxm_put_eth(b
, oxm
? OXM_OF_ARP_SHA
: NXM_NX_ARP_SHA
,
578 if (!(wc
& FWW_ARP_THA
)) {
579 nxm_put_eth(b
, oxm
? OXM_OF_ARP_THA
: NXM_NX_ARP_THA
,
585 nxm_put_64m(b
, NXM_NX_TUN_ID
, flow
->tun_id
, cr
->wc
.tun_id_mask
);
588 for (i
= 0; i
< FLOW_N_REGS
; i
++) {
589 nxm_put_32m(b
, NXM_NX_REG(i
),
590 htonl(flow
->regs
[i
]), htonl(cr
->wc
.reg_masks
[i
]));
593 /* OpenFlow 1.1+ Metadata. */
594 nxm_put_64m(b
, OXM_OF_METADATA
, flow
->metadata
, cr
->wc
.metadata_mask
);
597 nxm_put_64m(b
, NXM_NX_COOKIE
, cookie
, cookie_mask
);
599 match_len
= b
->size
- start_len
;
600 ofpbuf_put_zeros(b
, ROUND_UP(match_len
, 8) - match_len
);
604 /* nx_match_to_string() and helpers. */
606 static void format_nxm_field_name(struct ds
*, uint32_t header
);
609 nx_match_to_string(const uint8_t *p
, unsigned int match_len
)
615 return xstrdup("<any>");
619 while ((header
= nx_entry_ok(p
, match_len
)) != 0) {
620 unsigned int length
= NXM_LENGTH(header
);
621 unsigned int value_len
= nxm_field_bytes(header
);
622 const uint8_t *value
= p
+ 4;
623 const uint8_t *mask
= value
+ value_len
;
627 ds_put_cstr(&s
, ", ");
630 format_nxm_field_name(&s
, header
);
631 ds_put_char(&s
, '(');
633 for (i
= 0; i
< value_len
; i
++) {
634 ds_put_format(&s
, "%02x", value
[i
]);
636 if (NXM_HASMASK(header
)) {
637 ds_put_char(&s
, '/');
638 for (i
= 0; i
< value_len
; i
++) {
639 ds_put_format(&s
, "%02x", mask
[i
]);
642 ds_put_char(&s
, ')');
645 match_len
-= 4 + length
;
650 ds_put_cstr(&s
, ", ");
653 ds_put_format(&s
, "<%u invalid bytes>", match_len
);
656 return ds_steal_cstr(&s
);
660 format_nxm_field_name(struct ds
*s
, uint32_t header
)
662 const struct mf_field
*mf
= mf_from_nxm_header(header
);
664 ds_put_cstr(s
, IS_OXM_HEADER(header
) ? mf
->oxm_name
: mf
->nxm_name
);
665 if (NXM_HASMASK(header
)) {
666 ds_put_cstr(s
, "_W");
668 } else if (header
== NXM_NX_COOKIE
) {
669 ds_put_cstr(s
, "NXM_NX_COOKIE");
670 } else if (header
== NXM_NX_COOKIE_W
) {
671 ds_put_cstr(s
, "NXM_NX_COOKIE_W");
673 ds_put_format(s
, "%d:%d", NXM_VENDOR(header
), NXM_FIELD(header
));
678 parse_nxm_field_name(const char *name
, int name_len
)
683 /* Check whether it's a field name. */
684 wild
= name_len
> 2 && !memcmp(&name
[name_len
- 2], "_W", 2);
689 for (i
= 0; i
< MFF_N_IDS
; i
++) {
690 const struct mf_field
*mf
= mf_from_id(i
);
694 !strncmp(mf
->nxm_name
, name
, name_len
) &&
695 mf
->nxm_name
[name_len
] == '\0') {
696 header
= mf
->nxm_header
;
697 } else if (mf
->oxm_name
&&
698 !strncmp(mf
->oxm_name
, name
, name_len
) &&
699 mf
->oxm_name
[name_len
] == '\0') {
700 header
= mf
->oxm_header
;
707 } else if (mf
->maskable
!= MFM_NONE
) {
708 return NXM_MAKE_WILD_HEADER(header
);
712 if (!strncmp("NXM_NX_COOKIE", name
, name_len
) &&
713 (name_len
== strlen("NXM_NX_COOKIE"))) {
715 return NXM_NX_COOKIE
;
717 return NXM_NX_COOKIE_W
;
721 /* Check whether it's a 32-bit field header value as hex.
722 * (This isn't ordinarily useful except for testing error behavior.) */
724 uint32_t header
= hexits_value(name
, name_len
, NULL
);
725 if (header
!= UINT_MAX
) {
733 /* nx_match_from_string(). */
736 nx_match_from_string(const char *s
, struct ofpbuf
*b
)
738 const char *full_s
= s
;
739 const size_t start_len
= b
->size
;
742 if (!strcmp(s
, "<any>")) {
743 /* Ensure that 'b->data' isn't actually null. */
744 ofpbuf_prealloc_tailroom(b
, 1);
748 for (s
+= strspn(s
, ", "); *s
; s
+= strspn(s
, ", ")) {
755 name_len
= strcspn(s
, "(");
756 if (s
[name_len
] != '(') {
757 ovs_fatal(0, "%s: missing ( at end of nx_match", full_s
);
760 header
= parse_nxm_field_name(name
, name_len
);
762 ovs_fatal(0, "%s: unknown field `%.*s'", full_s
, name_len
, s
);
767 nxm_put_header(b
, header
);
768 s
= ofpbuf_put_hex(b
, s
, &n
);
769 if (n
!= nxm_field_bytes(header
)) {
770 ovs_fatal(0, "%.2s: hex digits expected", s
);
772 if (NXM_HASMASK(header
)) {
775 ovs_fatal(0, "%s: missing / in masked field %.*s",
776 full_s
, name_len
, name
);
778 s
= ofpbuf_put_hex(b
, s
+ 1, &n
);
779 if (n
!= nxm_field_bytes(header
)) {
780 ovs_fatal(0, "%.2s: hex digits expected", s
);
786 ovs_fatal(0, "%s: missing ) following field %.*s",
787 full_s
, name_len
, name
);
792 match_len
= b
->size
- start_len
;
793 ofpbuf_put_zeros(b
, ROUND_UP(match_len
, 8) - match_len
);
798 nxm_parse_reg_move(struct ofpact_reg_move
*move
, const char *s
)
800 const char *full_s
= s
;
802 s
= mf_parse_subfield(&move
->src
, s
);
803 if (strncmp(s
, "->", 2)) {
804 ovs_fatal(0, "%s: missing `->' following source", full_s
);
807 s
= mf_parse_subfield(&move
->dst
, s
);
809 ovs_fatal(0, "%s: trailing garbage following destination", full_s
);
812 if (move
->src
.n_bits
!= move
->dst
.n_bits
) {
813 ovs_fatal(0, "%s: source field is %d bits wide but destination is "
814 "%d bits wide", full_s
,
815 move
->src
.n_bits
, move
->dst
.n_bits
);
820 nxm_parse_reg_load(struct ofpact_reg_load
*load
, const char *s
)
822 const char *full_s
= s
;
824 load
->value
= strtoull(s
, (char **) &s
, 0);
825 if (strncmp(s
, "->", 2)) {
826 ovs_fatal(0, "%s: missing `->' following value", full_s
);
829 s
= mf_parse_subfield(&load
->dst
, s
);
831 ovs_fatal(0, "%s: trailing garbage following destination", full_s
);
834 if (load
->dst
.n_bits
< 64 && (load
->value
>> load
->dst
.n_bits
) != 0) {
835 ovs_fatal(0, "%s: value %"PRIu64
" does not fit into %d bits",
836 full_s
, load
->value
, load
->dst
.n_bits
);
840 /* nxm_format_reg_move(), nxm_format_reg_load(). */
843 nxm_format_reg_move(const struct ofpact_reg_move
*move
, struct ds
*s
)
845 ds_put_format(s
, "move:");
846 mf_format_subfield(&move
->src
, s
);
847 ds_put_cstr(s
, "->");
848 mf_format_subfield(&move
->dst
, s
);
852 nxm_format_reg_load(const struct ofpact_reg_load
*load
, struct ds
*s
)
854 ds_put_format(s
, "load:%#"PRIx64
"->", load
->value
);
855 mf_format_subfield(&load
->dst
, s
);
859 nxm_reg_move_from_openflow(const struct nx_action_reg_move
*narm
,
860 struct ofpbuf
*ofpacts
)
862 struct ofpact_reg_move
*move
;
864 move
= ofpact_put_REG_MOVE(ofpacts
);
865 move
->src
.field
= mf_from_nxm_header(ntohl(narm
->src
));
866 move
->src
.ofs
= ntohs(narm
->src_ofs
);
867 move
->src
.n_bits
= ntohs(narm
->n_bits
);
868 move
->dst
.field
= mf_from_nxm_header(ntohl(narm
->dst
));
869 move
->dst
.ofs
= ntohs(narm
->dst_ofs
);
870 move
->dst
.n_bits
= ntohs(narm
->n_bits
);
872 return nxm_reg_move_check(move
, NULL
);
876 nxm_reg_load_from_openflow(const struct nx_action_reg_load
*narl
,
877 struct ofpbuf
*ofpacts
)
879 struct ofpact_reg_load
*load
;
881 load
= ofpact_put_REG_LOAD(ofpacts
);
882 load
->dst
.field
= mf_from_nxm_header(ntohl(narl
->dst
));
883 load
->dst
.ofs
= nxm_decode_ofs(narl
->ofs_nbits
);
884 load
->dst
.n_bits
= nxm_decode_n_bits(narl
->ofs_nbits
);
885 load
->value
= ntohll(narl
->value
);
887 /* Reject 'narl' if a bit numbered 'n_bits' or higher is set to 1 in
889 if (load
->dst
.n_bits
< 64 && load
->value
>> load
->dst
.n_bits
) {
890 return OFPERR_OFPBAC_BAD_ARGUMENT
;
893 return nxm_reg_load_check(load
, NULL
);
897 nxm_reg_move_check(const struct ofpact_reg_move
*move
, const struct flow
*flow
)
901 error
= mf_check_src(&move
->src
, flow
);
906 return mf_check_dst(&move
->dst
, NULL
);
910 nxm_reg_load_check(const struct ofpact_reg_load
*load
, const struct flow
*flow
)
912 return mf_check_dst(&load
->dst
, flow
);
916 nxm_reg_move_to_nxast(const struct ofpact_reg_move
*move
,
917 struct ofpbuf
*openflow
)
919 struct nx_action_reg_move
*narm
;
921 narm
= ofputil_put_NXAST_REG_MOVE(openflow
);
922 narm
->n_bits
= htons(move
->dst
.n_bits
);
923 narm
->src_ofs
= htons(move
->src
.ofs
);
924 narm
->dst_ofs
= htons(move
->dst
.ofs
);
925 narm
->src
= htonl(move
->src
.field
->nxm_header
);
926 narm
->dst
= htonl(move
->dst
.field
->nxm_header
);
930 nxm_reg_load_to_nxast(const struct ofpact_reg_load
*load
,
931 struct ofpbuf
*openflow
)
933 struct nx_action_reg_load
*narl
;
935 narl
= ofputil_put_NXAST_REG_LOAD(openflow
);
936 narl
->ofs_nbits
= nxm_encode_ofs_nbits(load
->dst
.ofs
, load
->dst
.n_bits
);
937 narl
->dst
= htonl(load
->dst
.field
->nxm_header
);
938 narl
->value
= htonll(load
->value
);
941 /* nxm_execute_reg_move(), nxm_execute_reg_load(). */
944 nxm_execute_reg_move(const struct ofpact_reg_move
*move
,
947 union mf_value src_value
;
948 union mf_value dst_value
;
950 mf_get_value(move
->dst
.field
, flow
, &dst_value
);
951 mf_get_value(move
->src
.field
, flow
, &src_value
);
952 bitwise_copy(&src_value
, move
->src
.field
->n_bytes
, move
->src
.ofs
,
953 &dst_value
, move
->dst
.field
->n_bytes
, move
->dst
.ofs
,
955 mf_set_flow_value(move
->dst
.field
, &dst_value
, flow
);
959 nxm_execute_reg_load(const struct ofpact_reg_load
*load
, struct flow
*flow
)
961 nxm_reg_load(&load
->dst
, load
->value
, flow
);
965 nxm_reg_load(const struct mf_subfield
*dst
, uint64_t src_data
,
968 union mf_value dst_value
;
969 union mf_value src_value
;
971 mf_get_value(dst
->field
, flow
, &dst_value
);
972 src_value
.be64
= htonll(src_data
);
973 bitwise_copy(&src_value
, sizeof src_value
.be64
, 0,
974 &dst_value
, dst
->field
->n_bytes
, dst
->ofs
,
976 mf_set_flow_value(dst
->field
, &dst_value
, flow
);