2 * Copyright (c) 2009, 2010, 2011, 2012, 2013, 2014, 2015 Nicira, Inc.
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at:
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
17 /* Tests for classifier, written with knowledge of and to advantage of the
18 * classifier's internal structure.
20 * With very few exceptions, these tests obtain complete coverage of every
21 * basic block and every branch in the classifier implementation, e.g. a clean
22 * report from "gcov -b". (Covering the exceptions would require finding
23 * collisions in the hash function used for flow data, etc.)
25 * This test should receive a clean report from "valgrind --leak-check=full":
26 * it frees every heap block that it allocates.
31 #include "classifier.h"
35 #include "byte-order.h"
36 #include "classifier-private.h"
37 #include "command-line.h"
38 #include "fatal-signal.h"
41 #include "ovs-atomic.h"
42 #include "ovs-thread.h"
46 #include "unaligned.h"
49 static bool versioned
= false;
51 /* Fields in a rule. */
53 /* struct flow all-caps */ \
54 /* member name name */ \
55 /* ----------- -------- */ \
56 CLS_FIELD(tunnel.tun_id, TUN_ID) \
57 CLS_FIELD(metadata, METADATA) \
58 CLS_FIELD(nw_src, NW_SRC) \
59 CLS_FIELD(nw_dst, NW_DST) \
60 CLS_FIELD(in_port.ofp_port, IN_PORT) \
61 CLS_FIELD(vlans[0].tci, VLAN_TCI) \
62 CLS_FIELD(dl_type, DL_TYPE) \
63 CLS_FIELD(tp_src, TP_SRC) \
64 CLS_FIELD(tp_dst, TP_DST) \
65 CLS_FIELD(dl_src, DL_SRC) \
66 CLS_FIELD(dl_dst, DL_DST) \
67 CLS_FIELD(nw_proto, NW_PROTO) \
68 CLS_FIELD(nw_tos, NW_DSCP)
72 * (These are also indexed into struct classifier's 'tables' array.) */
74 #define CLS_FIELD(MEMBER, NAME) CLS_F_IDX_##NAME,
80 /* Field information. */
82 int ofs
; /* Offset in struct flow. */
83 int len
; /* Length in bytes. */
84 const char *name
; /* Name (for debugging). */
87 static const struct cls_field cls_fields
[CLS_N_FIELDS
] = {
88 #define CLS_FIELD(MEMBER, NAME) \
89 { offsetof(struct flow, MEMBER), \
90 sizeof ((struct flow *)0)->MEMBER, \
97 struct ovs_list list_node
;
98 int aux
; /* Auxiliary data. */
99 struct cls_rule cls_rule
; /* Classifier rule data. */
102 static struct test_rule
*
103 test_rule_from_cls_rule(const struct cls_rule
*rule
)
105 return rule
? CONTAINER_OF(rule
, struct test_rule
, cls_rule
) : NULL
;
109 test_rule_destroy(struct test_rule
*rule
)
112 cls_rule_destroy(&rule
->cls_rule
);
117 static struct test_rule
*make_rule(int wc_fields
, int priority
, int value_pat
);
118 static void free_rule(struct test_rule
*);
119 static struct test_rule
*clone_rule(const struct test_rule
*);
121 /* Trivial (linear) classifier. */
124 size_t allocated_rules
;
125 struct test_rule
**rules
;
129 tcls_init(struct tcls
*tcls
)
132 tcls
->allocated_rules
= 0;
137 tcls_destroy(struct tcls
*tcls
)
142 for (i
= 0; i
< tcls
->n_rules
; i
++) {
143 test_rule_destroy(tcls
->rules
[i
]);
150 tcls_is_empty(const struct tcls
*tcls
)
152 return tcls
->n_rules
== 0;
155 static struct test_rule
*
156 tcls_insert(struct tcls
*tcls
, const struct test_rule
*rule
)
160 for (i
= 0; i
< tcls
->n_rules
; i
++) {
161 const struct cls_rule
*pos
= &tcls
->rules
[i
]->cls_rule
;
162 if (cls_rule_equal(pos
, &rule
->cls_rule
)) {
164 ovsrcu_postpone(free_rule
, tcls
->rules
[i
]);
165 tcls
->rules
[i
] = clone_rule(rule
);
166 return tcls
->rules
[i
];
167 } else if (pos
->priority
< rule
->cls_rule
.priority
) {
172 if (tcls
->n_rules
>= tcls
->allocated_rules
) {
173 tcls
->rules
= x2nrealloc(tcls
->rules
, &tcls
->allocated_rules
,
174 sizeof *tcls
->rules
);
176 if (i
!= tcls
->n_rules
) {
177 memmove(&tcls
->rules
[i
+ 1], &tcls
->rules
[i
],
178 sizeof *tcls
->rules
* (tcls
->n_rules
- i
));
180 tcls
->rules
[i
] = clone_rule(rule
);
182 return tcls
->rules
[i
];
186 tcls_remove(struct tcls
*cls
, const struct test_rule
*rule
)
190 for (i
= 0; i
< cls
->n_rules
; i
++) {
191 struct test_rule
*pos
= cls
->rules
[i
];
193 test_rule_destroy(pos
);
195 memmove(&cls
->rules
[i
], &cls
->rules
[i
+ 1],
196 sizeof *cls
->rules
* (cls
->n_rules
- i
- 1));
206 match(const struct cls_rule
*wild_
, const struct flow
*fixed
)
211 minimatch_expand(&wild_
->match
, &wild
);
212 for (f_idx
= 0; f_idx
< CLS_N_FIELDS
; f_idx
++) {
215 if (f_idx
== CLS_F_IDX_NW_SRC
) {
216 eq
= !((fixed
->nw_src
^ wild
.flow
.nw_src
)
217 & wild
.wc
.masks
.nw_src
);
218 } else if (f_idx
== CLS_F_IDX_NW_DST
) {
219 eq
= !((fixed
->nw_dst
^ wild
.flow
.nw_dst
)
220 & wild
.wc
.masks
.nw_dst
);
221 } else if (f_idx
== CLS_F_IDX_TP_SRC
) {
222 eq
= !((fixed
->tp_src
^ wild
.flow
.tp_src
)
223 & wild
.wc
.masks
.tp_src
);
224 } else if (f_idx
== CLS_F_IDX_TP_DST
) {
225 eq
= !((fixed
->tp_dst
^ wild
.flow
.tp_dst
)
226 & wild
.wc
.masks
.tp_dst
);
227 } else if (f_idx
== CLS_F_IDX_DL_SRC
) {
228 eq
= eth_addr_equal_except(fixed
->dl_src
, wild
.flow
.dl_src
,
229 wild
.wc
.masks
.dl_src
);
230 } else if (f_idx
== CLS_F_IDX_DL_DST
) {
231 eq
= eth_addr_equal_except(fixed
->dl_dst
, wild
.flow
.dl_dst
,
232 wild
.wc
.masks
.dl_dst
);
233 } else if (f_idx
== CLS_F_IDX_VLAN_TCI
) {
234 eq
= !((fixed
->vlans
[0].tci
^ wild
.flow
.vlans
[0].tci
)
235 & wild
.wc
.masks
.vlans
[0].tci
);
236 } else if (f_idx
== CLS_F_IDX_TUN_ID
) {
237 eq
= !((fixed
->tunnel
.tun_id
^ wild
.flow
.tunnel
.tun_id
)
238 & wild
.wc
.masks
.tunnel
.tun_id
);
239 } else if (f_idx
== CLS_F_IDX_METADATA
) {
240 eq
= !((fixed
->metadata
^ wild
.flow
.metadata
)
241 & wild
.wc
.masks
.metadata
);
242 } else if (f_idx
== CLS_F_IDX_NW_DSCP
) {
243 eq
= !((fixed
->nw_tos
^ wild
.flow
.nw_tos
) &
244 (wild
.wc
.masks
.nw_tos
& IP_DSCP_MASK
));
245 } else if (f_idx
== CLS_F_IDX_NW_PROTO
) {
246 eq
= !((fixed
->nw_proto
^ wild
.flow
.nw_proto
)
247 & wild
.wc
.masks
.nw_proto
);
248 } else if (f_idx
== CLS_F_IDX_DL_TYPE
) {
249 eq
= !((fixed
->dl_type
^ wild
.flow
.dl_type
)
250 & wild
.wc
.masks
.dl_type
);
251 } else if (f_idx
== CLS_F_IDX_IN_PORT
) {
252 eq
= !((fixed
->in_port
.ofp_port
253 ^ wild
.flow
.in_port
.ofp_port
)
254 & wild
.wc
.masks
.in_port
.ofp_port
);
266 static struct cls_rule
*
267 tcls_lookup(const struct tcls
*cls
, const struct flow
*flow
)
271 for (i
= 0; i
< cls
->n_rules
; i
++) {
272 struct test_rule
*pos
= cls
->rules
[i
];
273 if (match(&pos
->cls_rule
, flow
)) {
274 return &pos
->cls_rule
;
281 tcls_delete_matches(struct tcls
*cls
, const struct cls_rule
*target
)
285 for (i
= 0; i
< cls
->n_rules
; ) {
286 struct test_rule
*pos
= cls
->rules
[i
];
287 if (!minimask_has_extra(pos
->cls_rule
.match
.mask
,
288 target
->match
.mask
)) {
291 miniflow_expand(pos
->cls_rule
.match
.flow
, &flow
);
292 if (match(target
, &flow
)) {
293 tcls_remove(cls
, pos
);
301 static ovs_be32 nw_src_values
[] = { CONSTANT_HTONL(0xc0a80001),
302 CONSTANT_HTONL(0xc0a04455) };
303 static ovs_be32 nw_dst_values
[] = { CONSTANT_HTONL(0xc0a80002),
304 CONSTANT_HTONL(0xc0a04455) };
305 static ovs_be64 tun_id_values
[] = {
307 CONSTANT_HTONLL(UINT64_C(0xfedcba9876543210)) };
308 static ovs_be64 metadata_values
[] = {
310 CONSTANT_HTONLL(UINT64_C(0xfedcba9876543210)) };
311 static ofp_port_t in_port_values
[] = { OFP_PORT_C(1), OFPP_LOCAL
};
312 static ovs_be16 vlan_tci_values
[] = { CONSTANT_HTONS(101), CONSTANT_HTONS(0) };
313 static ovs_be16 dl_type_values
[]
314 = { CONSTANT_HTONS(ETH_TYPE_IP
), CONSTANT_HTONS(ETH_TYPE_ARP
) };
315 static ovs_be16 tp_src_values
[] = { CONSTANT_HTONS(49362),
316 CONSTANT_HTONS(80) };
317 static ovs_be16 tp_dst_values
[] = { CONSTANT_HTONS(6667), CONSTANT_HTONS(22) };
318 static struct eth_addr dl_src_values
[] = {
319 ETH_ADDR_C(00,02,e3
,0f
,80,a4
),
320 ETH_ADDR_C(5e
,33,7f
,5f
,1e
,99)
322 static struct eth_addr dl_dst_values
[] = {
323 ETH_ADDR_C(4a
,27,71,ae
,64,c1
),
324 ETH_ADDR_C(ff
,ff
,ff
,ff
,ff
,ff
)
326 static uint8_t nw_proto_values
[] = { IPPROTO_TCP
, IPPROTO_ICMP
};
327 static uint8_t nw_dscp_values
[] = { 48, 0 };
329 static void *values
[CLS_N_FIELDS
][2];
334 values
[CLS_F_IDX_TUN_ID
][0] = &tun_id_values
[0];
335 values
[CLS_F_IDX_TUN_ID
][1] = &tun_id_values
[1];
337 values
[CLS_F_IDX_METADATA
][0] = &metadata_values
[0];
338 values
[CLS_F_IDX_METADATA
][1] = &metadata_values
[1];
340 values
[CLS_F_IDX_IN_PORT
][0] = &in_port_values
[0];
341 values
[CLS_F_IDX_IN_PORT
][1] = &in_port_values
[1];
343 values
[CLS_F_IDX_VLAN_TCI
][0] = &vlan_tci_values
[0];
344 values
[CLS_F_IDX_VLAN_TCI
][1] = &vlan_tci_values
[1];
346 values
[CLS_F_IDX_DL_SRC
][0] = &dl_src_values
[0];
347 values
[CLS_F_IDX_DL_SRC
][1] = &dl_src_values
[1];
349 values
[CLS_F_IDX_DL_DST
][0] = &dl_dst_values
[0];
350 values
[CLS_F_IDX_DL_DST
][1] = &dl_dst_values
[1];
352 values
[CLS_F_IDX_DL_TYPE
][0] = &dl_type_values
[0];
353 values
[CLS_F_IDX_DL_TYPE
][1] = &dl_type_values
[1];
355 values
[CLS_F_IDX_NW_SRC
][0] = &nw_src_values
[0];
356 values
[CLS_F_IDX_NW_SRC
][1] = &nw_src_values
[1];
358 values
[CLS_F_IDX_NW_DST
][0] = &nw_dst_values
[0];
359 values
[CLS_F_IDX_NW_DST
][1] = &nw_dst_values
[1];
361 values
[CLS_F_IDX_NW_PROTO
][0] = &nw_proto_values
[0];
362 values
[CLS_F_IDX_NW_PROTO
][1] = &nw_proto_values
[1];
364 values
[CLS_F_IDX_NW_DSCP
][0] = &nw_dscp_values
[0];
365 values
[CLS_F_IDX_NW_DSCP
][1] = &nw_dscp_values
[1];
367 values
[CLS_F_IDX_TP_SRC
][0] = &tp_src_values
[0];
368 values
[CLS_F_IDX_TP_SRC
][1] = &tp_src_values
[1];
370 values
[CLS_F_IDX_TP_DST
][0] = &tp_dst_values
[0];
371 values
[CLS_F_IDX_TP_DST
][1] = &tp_dst_values
[1];
374 #define N_NW_SRC_VALUES ARRAY_SIZE(nw_src_values)
375 #define N_NW_DST_VALUES ARRAY_SIZE(nw_dst_values)
376 #define N_TUN_ID_VALUES ARRAY_SIZE(tun_id_values)
377 #define N_METADATA_VALUES ARRAY_SIZE(metadata_values)
378 #define N_IN_PORT_VALUES ARRAY_SIZE(in_port_values)
379 #define N_VLAN_TCI_VALUES ARRAY_SIZE(vlan_tci_values)
380 #define N_DL_TYPE_VALUES ARRAY_SIZE(dl_type_values)
381 #define N_TP_SRC_VALUES ARRAY_SIZE(tp_src_values)
382 #define N_TP_DST_VALUES ARRAY_SIZE(tp_dst_values)
383 #define N_DL_SRC_VALUES ARRAY_SIZE(dl_src_values)
384 #define N_DL_DST_VALUES ARRAY_SIZE(dl_dst_values)
385 #define N_NW_PROTO_VALUES ARRAY_SIZE(nw_proto_values)
386 #define N_NW_DSCP_VALUES ARRAY_SIZE(nw_dscp_values)
388 #define N_FLOW_VALUES (N_NW_SRC_VALUES * \
392 N_VLAN_TCI_VALUES * \
398 N_NW_PROTO_VALUES * \
402 get_value(unsigned int *x
, unsigned n_values
)
404 unsigned int rem
= *x
% n_values
;
410 compare_classifiers(struct classifier
*cls
, size_t n_invisible_rules
,
411 ovs_version_t version
, struct tcls
*tcls
)
413 static const int confidence
= 500;
416 assert(classifier_count(cls
) == tcls
->n_rules
+ n_invisible_rules
);
417 for (i
= 0; i
< confidence
; i
++) {
418 const struct cls_rule
*cr0
, *cr1
, *cr2
;
420 struct flow_wildcards wc
;
423 flow_wildcards_init_catchall(&wc
);
424 x
= random_range(N_FLOW_VALUES
);
425 memset(&flow
, 0, sizeof flow
);
426 flow
.nw_src
= nw_src_values
[get_value(&x
, N_NW_SRC_VALUES
)];
427 flow
.nw_dst
= nw_dst_values
[get_value(&x
, N_NW_DST_VALUES
)];
428 flow
.tunnel
.tun_id
= tun_id_values
[get_value(&x
, N_TUN_ID_VALUES
)];
429 flow
.metadata
= metadata_values
[get_value(&x
, N_METADATA_VALUES
)];
430 flow
.in_port
.ofp_port
= in_port_values
[get_value(&x
,
432 flow
.vlans
[0].tci
= vlan_tci_values
[get_value(&x
, N_VLAN_TCI_VALUES
)];
433 flow
.dl_type
= dl_type_values
[get_value(&x
, N_DL_TYPE_VALUES
)];
434 flow
.tp_src
= tp_src_values
[get_value(&x
, N_TP_SRC_VALUES
)];
435 flow
.tp_dst
= tp_dst_values
[get_value(&x
, N_TP_DST_VALUES
)];
436 flow
.dl_src
= dl_src_values
[get_value(&x
, N_DL_SRC_VALUES
)];
437 flow
.dl_dst
= dl_dst_values
[get_value(&x
, N_DL_DST_VALUES
)];
438 flow
.nw_proto
= nw_proto_values
[get_value(&x
, N_NW_PROTO_VALUES
)];
439 flow
.nw_tos
= nw_dscp_values
[get_value(&x
, N_NW_DSCP_VALUES
)];
441 /* This assertion is here to suppress a GCC 4.9 array-bounds warning */
442 ovs_assert(cls
->n_tries
<= CLS_MAX_TRIES
);
444 cr0
= classifier_lookup(cls
, version
, &flow
, &wc
);
445 cr1
= tcls_lookup(tcls
, &flow
);
446 assert((cr0
== NULL
) == (cr1
== NULL
));
448 const struct test_rule
*tr0
= test_rule_from_cls_rule(cr0
);
449 const struct test_rule
*tr1
= test_rule_from_cls_rule(cr1
);
451 assert(cls_rule_equal(cr0
, cr1
));
452 assert(tr0
->aux
== tr1
->aux
);
454 /* Make sure the rule should have been visible. */
455 assert(cls_rule_visible_in_version(cr0
, version
));
457 cr2
= classifier_lookup(cls
, version
, &flow
, NULL
);
463 destroy_classifier(struct classifier
*cls
)
465 struct test_rule
*rule
;
467 classifier_defer(cls
);
468 CLS_FOR_EACH (rule
, cls_rule
, cls
) {
469 classifier_remove_assert(cls
, &rule
->cls_rule
);
470 ovsrcu_postpone(free_rule
, rule
);
472 classifier_destroy(cls
);
476 pvector_verify(const struct pvector
*pvec
)
478 void *ptr OVS_UNUSED
;
479 int prev_priority
= INT_MAX
;
481 PVECTOR_FOR_EACH (ptr
, pvec
) {
482 int priority
= cursor__
.vector
[cursor__
.entry_idx
].priority
;
483 if (priority
> prev_priority
) {
484 ovs_abort(0, "Priority vector is out of order (%u > %u)",
485 priority
, prev_priority
);
487 prev_priority
= priority
;
492 trie_verify(const rcu_trie_ptr
*trie
, unsigned int ofs
, unsigned int n_bits
)
494 const struct trie_node
*node
= ovsrcu_get(struct trie_node
*, trie
);
497 assert(node
->n_rules
== 0 || node
->n_bits
> 0);
499 assert((ofs
> 0 || (ofs
== 0 && node
->n_bits
== 0)) && ofs
<= n_bits
);
502 + trie_verify(&node
->edges
[0], ofs
, n_bits
)
503 + trie_verify(&node
->edges
[1], ofs
, n_bits
);
509 verify_tries(struct classifier
*cls
)
510 OVS_NO_THREAD_SAFETY_ANALYSIS
512 unsigned int n_rules
= 0;
515 for (i
= 0; i
< cls
->n_tries
; i
++) {
516 const struct mf_field
* cls_field
517 = ovsrcu_get(struct mf_field
*, &cls
->tries
[i
].field
);
518 n_rules
+= trie_verify(&cls
->tries
[i
].root
, 0, cls_field
->n_bits
);
520 assert(n_rules
<= cls
->n_rules
);
524 check_tables(const struct classifier
*cls
, int n_tables
, int n_rules
,
525 int n_dups
, int n_invisible
, ovs_version_t version
)
526 OVS_NO_THREAD_SAFETY_ANALYSIS
528 const struct cls_subtable
*table
;
529 struct test_rule
*test_rule
;
530 int found_tables
= 0;
531 int found_tables_with_visible_rules
= 0;
534 int found_invisible
= 0;
535 int found_visible_but_removable
= 0;
536 int found_rules2
= 0;
538 pvector_verify(&cls
->subtables
);
539 CMAP_FOR_EACH (table
, cmap_node
, &cls
->subtables_map
) {
540 const struct cls_match
*head
;
541 int max_priority
= INT_MIN
;
542 unsigned int max_count
= 0;
544 bool found_visible_rules
= false;
545 const struct cls_subtable
*iter
;
547 /* Locate the subtable from 'subtables'. */
548 PVECTOR_FOR_EACH (iter
, &cls
->subtables
) {
551 ovs_abort(0, "Subtable %p duplicated in 'subtables'.",
558 ovs_abort(0, "Subtable %p not found from 'subtables'.", table
);
561 assert(!cmap_is_empty(&table
->rules
));
562 assert(trie_verify(&table
->ports_trie
, 0, table
->ports_mask_len
)
563 == (table
->ports_mask_len
? cmap_count(&table
->rules
) : 0));
567 CMAP_FOR_EACH (head
, cmap_node
, &table
->rules
) {
568 int prev_priority
= INT_MAX
;
569 ovs_version_t prev_version
= 0;
570 const struct cls_match
*rule
, *prev
;
571 bool found_visible_rules_in_list
= false;
573 assert(head
->priority
<= table
->max_priority
);
575 if (head
->priority
> max_priority
) {
576 max_priority
= head
->priority
;
580 FOR_EACH_RULE_IN_LIST_PROTECTED(rule
, prev
, head
) {
581 ovs_version_t rule_version
;
582 const struct cls_rule
*found_rule
;
584 /* Priority may not increase. */
585 assert(rule
->priority
<= prev_priority
);
587 if (rule
->priority
== max_priority
) {
591 /* Count invisible rules and visible duplicates. */
592 if (!cls_match_visible_in_version(rule
, version
)) {
595 if (cls_match_is_eventually_invisible(rule
)) {
596 found_visible_but_removable
++;
598 if (found_visible_rules_in_list
) {
601 found_visible_rules_in_list
= true;
602 found_visible_rules
= true;
605 /* Rule must be visible in the version it was inserted. */
606 rule_version
= rule
->versions
.add_version
;
607 assert(cls_match_visible_in_version(rule
, rule_version
));
609 /* We should always find the latest version of the rule,
610 * unless all rules have been marked for removal.
611 * Later versions must always be later in the list. */
612 found_rule
= classifier_find_rule_exactly(cls
, rule
->cls_rule
,
614 if (found_rule
&& found_rule
!= rule
->cls_rule
) {
615 struct cls_match
*cls_match
;
616 cls_match
= get_cls_match_protected(found_rule
);
618 assert(found_rule
->priority
== rule
->priority
);
620 /* Found rule may not have a lower version. */
621 assert(cls_match
->versions
.add_version
>= rule_version
);
623 /* This rule must not be visible in the found rule's
625 assert(!cls_match_visible_in_version(
626 rule
, cls_match
->versions
.add_version
));
629 if (rule
->priority
== prev_priority
) {
630 /* Exact duplicate rule may not have a lower version. */
631 assert(rule_version
>= prev_version
);
633 /* Previous rule must not be visible in rule's version. */
634 assert(!cls_match_visible_in_version(prev
, rule_version
));
637 prev_priority
= rule
->priority
;
638 prev_version
= rule_version
;
643 if (found_visible_rules
) {
644 found_tables_with_visible_rules
++;
647 assert(table
->max_priority
== max_priority
);
648 assert(table
->max_count
== max_count
);
651 assert(found_tables
== cmap_count(&cls
->subtables_map
));
652 assert(found_tables
== pvector_count(&cls
->subtables
));
653 assert(n_tables
== -1 || n_tables
== found_tables_with_visible_rules
);
654 assert(n_rules
== -1 || found_rules
== n_rules
+ found_invisible
);
655 assert(n_dups
== -1 || found_dups
== n_dups
);
656 assert(found_invisible
== n_invisible
);
658 CLS_FOR_EACH (test_rule
, cls_rule
, cls
) {
661 /* Iteration does not see removable rules. */
663 == found_rules2
+ found_visible_but_removable
+ found_invisible
);
666 static struct test_rule
*
667 make_rule(int wc_fields
, int priority
, int value_pat
)
669 const struct cls_field
*f
;
670 struct test_rule
*rule
;
673 match_init_catchall(&match
);
674 for (f
= &cls_fields
[0]; f
< &cls_fields
[CLS_N_FIELDS
]; f
++) {
675 int f_idx
= f
- cls_fields
;
676 int value_idx
= (value_pat
& (1u << f_idx
)) != 0;
677 memcpy((char *) &match
.flow
+ f
->ofs
,
678 values
[f_idx
][value_idx
], f
->len
);
680 if (f_idx
== CLS_F_IDX_NW_SRC
) {
681 match
.wc
.masks
.nw_src
= OVS_BE32_MAX
;
682 } else if (f_idx
== CLS_F_IDX_NW_DST
) {
683 match
.wc
.masks
.nw_dst
= OVS_BE32_MAX
;
684 } else if (f_idx
== CLS_F_IDX_TP_SRC
) {
685 match
.wc
.masks
.tp_src
= OVS_BE16_MAX
;
686 } else if (f_idx
== CLS_F_IDX_TP_DST
) {
687 match
.wc
.masks
.tp_dst
= OVS_BE16_MAX
;
688 } else if (f_idx
== CLS_F_IDX_DL_SRC
) {
689 WC_MASK_FIELD(&match
.wc
, dl_src
);
690 } else if (f_idx
== CLS_F_IDX_DL_DST
) {
691 WC_MASK_FIELD(&match
.wc
, dl_dst
);
692 } else if (f_idx
== CLS_F_IDX_VLAN_TCI
) {
693 match
.wc
.masks
.vlans
[0].tci
= OVS_BE16_MAX
;
694 } else if (f_idx
== CLS_F_IDX_TUN_ID
) {
695 match
.wc
.masks
.tunnel
.tun_id
= OVS_BE64_MAX
;
696 } else if (f_idx
== CLS_F_IDX_METADATA
) {
697 match
.wc
.masks
.metadata
= OVS_BE64_MAX
;
698 } else if (f_idx
== CLS_F_IDX_NW_DSCP
) {
699 match
.wc
.masks
.nw_tos
|= IP_DSCP_MASK
;
700 } else if (f_idx
== CLS_F_IDX_NW_PROTO
) {
701 match
.wc
.masks
.nw_proto
= UINT8_MAX
;
702 } else if (f_idx
== CLS_F_IDX_DL_TYPE
) {
703 match
.wc
.masks
.dl_type
= OVS_BE16_MAX
;
704 } else if (f_idx
== CLS_F_IDX_IN_PORT
) {
705 match
.wc
.masks
.in_port
.ofp_port
= u16_to_ofp(UINT16_MAX
);
711 rule
= xzalloc(sizeof *rule
);
712 cls_rule_init(&rule
->cls_rule
, &match
, wc_fields
713 ? (priority
== INT_MIN
? priority
+ 1 :
714 priority
== INT_MAX
? priority
- 1 : priority
)
719 static struct test_rule
*
720 clone_rule(const struct test_rule
*src
)
722 struct test_rule
*dst
;
724 dst
= xmalloc(sizeof *dst
);
726 cls_rule_clone(&dst
->cls_rule
, &src
->cls_rule
);
731 free_rule(struct test_rule
*rule
)
733 cls_rule_destroy(&rule
->cls_rule
);
738 shuffle(int *p
, size_t n
)
740 for (; n
> 1; n
--, p
++) {
741 int *q
= &p
[random_range(n
)];
749 shuffle_u32s(uint32_t *p
, size_t n
)
751 for (; n
> 1; n
--, p
++) {
752 uint32_t *q
= &p
[random_range(n
)];
759 /* Classifier tests. */
761 static enum mf_field_id trie_fields
[2] = {
762 MFF_IPV4_DST
, MFF_IPV4_SRC
766 set_prefix_fields(struct classifier
*cls
)
769 classifier_set_prefix_fields(cls
, trie_fields
, ARRAY_SIZE(trie_fields
));
773 /* Tests an empty classifier. */
775 test_empty(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
777 struct classifier cls
;
780 classifier_init(&cls
, flow_segment_u64s
);
781 set_prefix_fields(&cls
);
783 assert(classifier_is_empty(&cls
));
784 assert(tcls_is_empty(&tcls
));
785 compare_classifiers(&cls
, 0, OVS_VERSION_MIN
, &tcls
);
786 classifier_destroy(&cls
);
790 /* Destroys a null classifier. */
792 test_destroy_null(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
794 classifier_destroy(NULL
);
797 /* Tests classification with one rule at a time. */
799 test_single_rule(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
801 unsigned int wc_fields
; /* Hilarious. */
803 for (wc_fields
= 0; wc_fields
< (1u << CLS_N_FIELDS
); wc_fields
++) {
804 struct classifier cls
;
805 struct test_rule
*rule
, *tcls_rule
;
808 rule
= make_rule(wc_fields
,
809 hash_bytes(&wc_fields
, sizeof wc_fields
, 0), 0);
810 classifier_init(&cls
, flow_segment_u64s
);
811 set_prefix_fields(&cls
);
813 tcls_rule
= tcls_insert(&tcls
, rule
);
815 classifier_insert(&cls
, &rule
->cls_rule
, OVS_VERSION_MIN
, NULL
, 0);
816 compare_classifiers(&cls
, 0, OVS_VERSION_MIN
, &tcls
);
817 check_tables(&cls
, 1, 1, 0, 0, OVS_VERSION_MIN
);
819 classifier_remove_assert(&cls
, &rule
->cls_rule
);
820 tcls_remove(&tcls
, tcls_rule
);
821 assert(classifier_is_empty(&cls
));
822 assert(tcls_is_empty(&tcls
));
823 compare_classifiers(&cls
, 0, OVS_VERSION_MIN
, &tcls
);
825 ovsrcu_postpone(free_rule
, rule
);
826 classifier_destroy(&cls
);
831 /* Tests replacing one rule by another. */
833 test_rule_replacement(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
835 unsigned int wc_fields
;
837 for (wc_fields
= 0; wc_fields
< (1u << CLS_N_FIELDS
); wc_fields
++) {
838 struct classifier cls
;
839 struct test_rule
*rule1
;
840 struct test_rule
*rule2
;
843 rule1
= make_rule(wc_fields
, OFP_DEFAULT_PRIORITY
, UINT_MAX
);
844 rule2
= make_rule(wc_fields
, OFP_DEFAULT_PRIORITY
, UINT_MAX
);
848 classifier_init(&cls
, flow_segment_u64s
);
849 set_prefix_fields(&cls
);
851 tcls_insert(&tcls
, rule1
);
852 classifier_insert(&cls
, &rule1
->cls_rule
, OVS_VERSION_MIN
, NULL
, 0);
853 compare_classifiers(&cls
, 0, OVS_VERSION_MIN
, &tcls
);
854 check_tables(&cls
, 1, 1, 0, 0, OVS_VERSION_MIN
);
858 tcls_insert(&tcls
, rule2
);
860 assert(test_rule_from_cls_rule(
861 classifier_replace(&cls
, &rule2
->cls_rule
, OVS_VERSION_MIN
,
863 ovsrcu_postpone(free_rule
, rule1
);
864 compare_classifiers(&cls
, 0, OVS_VERSION_MIN
, &tcls
);
865 check_tables(&cls
, 1, 1, 0, 0, OVS_VERSION_MIN
);
866 classifier_defer(&cls
);
867 classifier_remove_assert(&cls
, &rule2
->cls_rule
);
870 destroy_classifier(&cls
);
875 factorial(int n_items
)
880 for (i
= 2; i
<= n_items
; i
++) {
895 reverse(int *a
, int n
)
899 for (i
= 0; i
< n
/ 2; i
++) {
906 next_permutation(int *a
, int n
)
910 for (k
= n
- 2; k
>= 0; k
--) {
911 if (a
[k
] < a
[k
+ 1]) {
914 for (l
= n
- 1; ; l
--) {
917 reverse(a
+ (k
+ 1), n
- (k
+ 1));
926 /* Tests classification with rules that have the same matching criteria. */
928 test_many_rules_in_one_list (struct ovs_cmdl_context
*ctx OVS_UNUSED
)
930 enum { N_RULES
= 3 };
933 for (n_pris
= N_RULES
; n_pris
>= 1; n_pris
--) {
934 int ops
[N_RULES
* 2];
940 for (i
= 1; i
< N_RULES
; i
++) {
941 pris
[i
] = pris
[i
- 1] + (n_pris
> i
);
944 for (i
= 0; i
< N_RULES
* 2; i
++) {
950 struct test_rule
*rules
[N_RULES
];
951 struct test_rule
*tcls_rules
[N_RULES
];
952 int pri_rules
[N_RULES
];
953 struct classifier cls
;
955 ovs_version_t version
= OVS_VERSION_MIN
;
956 size_t n_invisible_rules
= 0;
960 for (i
= 0; i
< N_RULES
; i
++) {
961 rules
[i
] = make_rule(456, pris
[i
], 0);
962 tcls_rules
[i
] = NULL
;
966 classifier_init(&cls
, flow_segment_u64s
);
967 set_prefix_fields(&cls
);
970 for (i
= 0; i
< ARRAY_SIZE(ops
); i
++) {
971 struct test_rule
*displaced_rule
= NULL
;
972 struct cls_rule
*removable_rule
= NULL
;
976 if (!tcls_rules
[j
]) {
977 tcls_rules
[j
] = tcls_insert(&tcls
, rules
[j
]);
979 /* Insert the new rule in the next version. */
982 displaced_rule
= test_rule_from_cls_rule(
983 classifier_find_rule_exactly(&cls
,
986 if (displaced_rule
) {
987 /* Mark the old rule for removal after the current
989 cls_rule_make_invisible_in_version(
990 &displaced_rule
->cls_rule
, version
);
992 removable_rule
= &displaced_rule
->cls_rule
;
994 classifier_insert(&cls
, &rules
[j
]->cls_rule
, version
,
997 displaced_rule
= test_rule_from_cls_rule(
998 classifier_replace(&cls
, &rules
[j
]->cls_rule
,
1001 if (pri_rules
[pris
[j
]] >= 0) {
1002 int k
= pri_rules
[pris
[j
]];
1003 assert(displaced_rule
!= NULL
);
1004 assert(displaced_rule
!= rules
[j
]);
1005 assert(pris
[j
] == displaced_rule
->cls_rule
.priority
);
1006 tcls_rules
[k
] = NULL
;
1008 assert(displaced_rule
== NULL
);
1010 pri_rules
[pris
[j
]] = j
;
1013 /* Mark the rule for removal after the current
1016 cls_rule_make_invisible_in_version(
1017 &rules
[j
]->cls_rule
, version
);
1018 n_invisible_rules
++;
1019 removable_rule
= &rules
[j
]->cls_rule
;
1021 classifier_remove_assert(&cls
, &rules
[j
]->cls_rule
);
1023 tcls_remove(&tcls
, tcls_rules
[j
]);
1024 tcls_rules
[j
] = NULL
;
1025 pri_rules
[pris
[j
]] = -1;
1027 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1029 for (m
= 0; m
< N_RULES
; m
++) {
1030 n
+= tcls_rules
[m
] != NULL
;
1032 check_tables(&cls
, n
> 0, n
, n
- 1, n_invisible_rules
,
1035 if (versioned
&& removable_rule
) {
1036 struct cls_match
*cls_match
=
1037 get_cls_match_protected(removable_rule
);
1039 /* Removable rule is no longer visible. */
1041 assert(!cls_match_visible_in_version(cls_match
, version
));
1042 classifier_remove_assert(&cls
, removable_rule
);
1043 n_invisible_rules
--;
1047 classifier_defer(&cls
);
1048 for (i
= 0; i
< N_RULES
; i
++) {
1049 if (classifier_remove(&cls
, &rules
[i
]->cls_rule
)) {
1050 ovsrcu_postpone(free_rule
, rules
[i
]);
1053 classifier_destroy(&cls
);
1054 tcls_destroy(&tcls
);
1055 } while (next_permutation(ops
, ARRAY_SIZE(ops
)));
1056 assert(n_permutations
== (factorial(N_RULES
* 2) >> N_RULES
));
1061 count_ones(unsigned long int x
)
1066 x
= zero_rightmost_1bit(x
);
1074 array_contains(int *array
, int n
, int value
)
1078 for (i
= 0; i
< n
; i
++) {
1079 if (array
[i
] == value
) {
1087 /* Tests classification with two rules at a time that fall into the same
1088 * table but different lists. */
1090 test_many_rules_in_one_table(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1094 for (iteration
= 0; iteration
< 50; iteration
++) {
1095 enum { N_RULES
= 20 };
1096 struct test_rule
*rules
[N_RULES
];
1097 struct test_rule
*tcls_rules
[N_RULES
];
1098 struct classifier cls
;
1100 ovs_version_t version
= OVS_VERSION_MIN
;
1101 size_t n_invisible_rules
= 0;
1102 int value_pats
[N_RULES
];
1108 wcf
= random_uint32() & ((1u << CLS_N_FIELDS
) - 1);
1109 value_mask
= ~wcf
& ((1u << CLS_N_FIELDS
) - 1);
1110 } while ((1 << count_ones(value_mask
)) < N_RULES
);
1112 classifier_init(&cls
, flow_segment_u64s
);
1113 set_prefix_fields(&cls
);
1116 for (i
= 0; i
< N_RULES
; i
++) {
1117 int priority
= random_range(INT_MAX
);
1120 value_pats
[i
] = random_uint32() & value_mask
;
1121 } while (array_contains(value_pats
, i
, value_pats
[i
]));
1124 rules
[i
] = make_rule(wcf
, priority
, value_pats
[i
]);
1125 tcls_rules
[i
] = tcls_insert(&tcls
, rules
[i
]);
1127 classifier_insert(&cls
, &rules
[i
]->cls_rule
, version
, NULL
, 0);
1128 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1130 check_tables(&cls
, 1, i
+ 1, 0, n_invisible_rules
, version
);
1133 for (i
= 0; i
< N_RULES
; i
++) {
1134 tcls_remove(&tcls
, tcls_rules
[i
]);
1136 /* Mark the rule for removal after the current version. */
1138 cls_rule_make_invisible_in_version(&rules
[i
]->cls_rule
,
1140 n_invisible_rules
++;
1142 classifier_remove_assert(&cls
, &rules
[i
]->cls_rule
);
1144 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1145 check_tables(&cls
, i
< N_RULES
- 1, N_RULES
- (i
+ 1), 0,
1146 n_invisible_rules
, version
);
1148 ovsrcu_postpone(free_rule
, rules
[i
]);
1153 for (i
= 0; i
< N_RULES
; i
++) {
1154 classifier_remove_assert(&cls
, &rules
[i
]->cls_rule
);
1155 n_invisible_rules
--;
1157 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1158 check_tables(&cls
, 0, 0, 0, n_invisible_rules
, version
);
1159 ovsrcu_postpone(free_rule
, rules
[i
]);
1163 classifier_destroy(&cls
);
1164 tcls_destroy(&tcls
);
1168 /* Tests classification with many rules at a time that fall into random lists
1171 test_many_rules_in_n_tables(int n_tables
)
1173 enum { MAX_RULES
= 50 };
1178 assert(n_tables
< 10);
1179 for (i
= 0; i
< n_tables
; i
++) {
1181 wcfs
[i
] = random_uint32() & ((1u << CLS_N_FIELDS
) - 1);
1182 } while (array_contains(wcfs
, i
, wcfs
[i
]));
1185 for (iteration
= 0; iteration
< 30; iteration
++) {
1186 int priorities
[MAX_RULES
];
1187 struct classifier cls
;
1189 ovs_version_t version
= OVS_VERSION_MIN
;
1190 size_t n_invisible_rules
= 0;
1191 struct ovs_list list
= OVS_LIST_INITIALIZER(&list
);
1193 random_set_seed(iteration
+ 1);
1194 for (i
= 0; i
< MAX_RULES
; i
++) {
1195 priorities
[i
] = (i
* 129) & INT_MAX
;
1197 shuffle(priorities
, ARRAY_SIZE(priorities
));
1199 classifier_init(&cls
, flow_segment_u64s
);
1200 set_prefix_fields(&cls
);
1203 for (i
= 0; i
< MAX_RULES
; i
++) {
1204 struct test_rule
*rule
;
1205 int priority
= priorities
[i
];
1206 int wcf
= wcfs
[random_range(n_tables
)];
1207 int value_pat
= random_uint32() & ((1u << CLS_N_FIELDS
) - 1);
1208 rule
= make_rule(wcf
, priority
, value_pat
);
1209 tcls_insert(&tcls
, rule
);
1210 classifier_insert(&cls
, &rule
->cls_rule
, version
, NULL
, 0);
1211 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1212 check_tables(&cls
, -1, i
+ 1, -1, n_invisible_rules
, version
);
1215 while (classifier_count(&cls
) - n_invisible_rules
> 0) {
1216 struct test_rule
*target
;
1217 struct test_rule
*rule
;
1218 size_t n_removable_rules
= 0;
1220 target
= clone_rule(tcls
.rules
[random_range(tcls
.n_rules
)]);
1222 CLS_FOR_EACH_TARGET (rule
, cls_rule
, &cls
, &target
->cls_rule
,
1225 /* Mark the rule for removal after the current version. */
1226 cls_rule_make_invisible_in_version(&rule
->cls_rule
,
1228 n_removable_rules
++;
1229 compare_classifiers(&cls
, n_invisible_rules
, version
,
1231 check_tables(&cls
, -1, -1, -1, n_invisible_rules
, version
);
1233 ovs_list_push_back(&list
, &rule
->list_node
);
1234 } else if (classifier_remove(&cls
, &rule
->cls_rule
)) {
1235 ovsrcu_postpone(free_rule
, rule
);
1240 n_invisible_rules
+= n_removable_rules
;
1242 tcls_delete_matches(&tcls
, &target
->cls_rule
);
1245 compare_classifiers(&cls
, n_invisible_rules
, version
, &tcls
);
1246 check_tables(&cls
, -1, -1, -1, n_invisible_rules
, version
);
1249 struct test_rule
*rule
;
1251 /* Remove rules that are no longer visible. */
1252 LIST_FOR_EACH_POP (rule
, list_node
, &list
) {
1253 classifier_remove_assert(&cls
, &rule
->cls_rule
);
1254 n_invisible_rules
--;
1256 compare_classifiers(&cls
, n_invisible_rules
, version
,
1258 check_tables(&cls
, -1, -1, -1, n_invisible_rules
, version
);
1262 destroy_classifier(&cls
);
1263 tcls_destroy(&tcls
);
1268 test_many_rules_in_two_tables(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1270 test_many_rules_in_n_tables(2);
1274 test_many_rules_in_five_tables(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1276 test_many_rules_in_n_tables(5);
1279 /* Classifier benchmarks. */
1281 static int n_rules
; /* Number of rules to insert. */
1282 static int n_priorities
; /* Number of priorities to use. */
1283 static int n_tables
; /* Number of subtables. */
1284 static int n_threads
; /* Number of threads to search and mutate. */
1285 static int n_lookups
; /* Number of lookups each thread performs. */
1287 static void benchmark(bool use_wc
);
1290 elapsed(const struct timeval
*start
)
1294 xgettimeofday(&end
);
1295 return timeval_to_msec(&end
) - timeval_to_msec(start
);
1299 run_benchmarks(struct ovs_cmdl_context
*ctx
)
1302 || (ctx
->argc
> 1 && !strcmp(ctx
->argv
[1], "--help"))) {
1304 "usage: ovstest %s benchmark <n_rules> <n_priorities> <n_subtables> <n_threads> <n_lookups>\n"
1308 "<n_rules> - The number of rules to install for lookups. More rules\n"
1309 " makes misses less likely.\n"
1310 "<n_priorities> - How many different priorities to use. Using only 1\n"
1311 " priority will force lookups to continue through all\n"
1313 "<n_subtables> - Number of subtables to use. Normally a classifier has\n"
1314 " rules with different kinds of masks, resulting in\n"
1315 " multiple subtables (one per mask). However, in some\n"
1316 " special cases a table may consist of only one kind of\n"
1317 " rules, so there will be only one subtable.\n"
1318 "<n_threads> - How many lookup threads to use. Using one thread should\n"
1319 " give less variance accross runs, but classifier\n"
1320 " scaling can be tested with multiple threads.\n"
1321 "<n_lookups> - How many lookups each thread should perform.\n"
1322 "\n", program_name
);
1326 n_rules
= strtol(ctx
->argv
[1], NULL
, 10);
1327 n_priorities
= strtol(ctx
->argv
[2], NULL
, 10);
1328 n_tables
= strtol(ctx
->argv
[3], NULL
, 10);
1329 n_threads
= strtol(ctx
->argv
[4], NULL
, 10);
1330 n_lookups
= strtol(ctx
->argv
[5], NULL
, 10);
1332 printf("\nBenchmarking with:\n"
1333 "%d rules with %d priorities in %d tables, "
1334 "%d threads doing %d lookups each\n",
1335 n_rules
, n_priorities
, n_tables
, n_threads
, n_lookups
);
1337 puts("\nWithout wildcards: \n");
1339 puts("\nWith wildcards: \n");
1344 const struct classifier
*cls
;
1345 size_t n_lookup_flows
;
1346 struct flow
*lookup_flows
;
1353 lookup_classifier(void *aux_
)
1355 struct cls_aux
*aux
= aux_
;
1356 ovs_version_t version
= OVS_VERSION_MIN
;
1357 int hits
= 0, old_hits
;
1358 int misses
= 0, old_misses
;
1363 for (i
= 0; i
< n_lookups
; i
++) {
1364 const struct cls_rule
*cr
;
1365 struct flow_wildcards wc
;
1368 x
= random_range(aux
->n_lookup_flows
);
1371 flow_wildcards_init_catchall(&wc
);
1372 cr
= classifier_lookup(aux
->cls
, version
, &aux
->lookup_flows
[x
],
1375 cr
= classifier_lookup(aux
->cls
, version
, &aux
->lookup_flows
[x
],
1384 atomic_add(&aux
->hits
, hits
, &old_hits
);
1385 atomic_add(&aux
->misses
, misses
, &old_misses
);
1389 /* Benchmark classification. */
1391 benchmark(bool use_wc
)
1393 struct classifier cls
;
1394 ovs_version_t version
= OVS_VERSION_MIN
;
1396 int *wcfs
= xmalloc(n_tables
* sizeof *wcfs
);
1397 int *priorities
= xmalloc(n_priorities
* sizeof *priorities
);
1398 struct timeval start
;
1402 fatal_signal_init();
1406 for (i
= 0; i
< n_tables
; i
++) {
1408 wcfs
[i
] = random_uint32() & ((1u << CLS_N_FIELDS
) - 1);
1409 } while (array_contains(wcfs
, i
, wcfs
[i
]));
1412 for (i
= 0; i
< n_priorities
; i
++) {
1413 priorities
[i
] = (i
* 129) & INT_MAX
;
1415 shuffle(priorities
, n_priorities
);
1417 classifier_init(&cls
, flow_segment_u64s
);
1418 set_prefix_fields(&cls
);
1420 /* Create lookup flows. */
1421 aux
.use_wc
= use_wc
;
1423 aux
.n_lookup_flows
= 2 * N_FLOW_VALUES
;
1424 aux
.lookup_flows
= xzalloc(aux
.n_lookup_flows
* sizeof *aux
.lookup_flows
);
1425 for (i
= 0; i
< aux
.n_lookup_flows
; i
++) {
1426 struct flow
*flow
= &aux
.lookup_flows
[i
];
1429 x
= random_range(N_FLOW_VALUES
);
1430 flow
->nw_src
= nw_src_values
[get_value(&x
, N_NW_SRC_VALUES
)];
1431 flow
->nw_dst
= nw_dst_values
[get_value(&x
, N_NW_DST_VALUES
)];
1432 flow
->tunnel
.tun_id
= tun_id_values
[get_value(&x
, N_TUN_ID_VALUES
)];
1433 flow
->metadata
= metadata_values
[get_value(&x
, N_METADATA_VALUES
)];
1434 flow
->in_port
.ofp_port
= in_port_values
[get_value(&x
,
1436 flow
->vlans
[0].tci
= vlan_tci_values
[get_value(&x
, N_VLAN_TCI_VALUES
)];
1437 flow
->dl_type
= dl_type_values
[get_value(&x
, N_DL_TYPE_VALUES
)];
1438 flow
->tp_src
= tp_src_values
[get_value(&x
, N_TP_SRC_VALUES
)];
1439 flow
->tp_dst
= tp_dst_values
[get_value(&x
, N_TP_DST_VALUES
)];
1440 flow
->dl_src
= dl_src_values
[get_value(&x
, N_DL_SRC_VALUES
)];
1441 flow
->dl_dst
= dl_dst_values
[get_value(&x
, N_DL_DST_VALUES
)];
1442 flow
->nw_proto
= nw_proto_values
[get_value(&x
, N_NW_PROTO_VALUES
)];
1443 flow
->nw_tos
= nw_dscp_values
[get_value(&x
, N_NW_DSCP_VALUES
)];
1445 atomic_init(&aux
.hits
, 0);
1446 atomic_init(&aux
.misses
, 0);
1448 /* Rule insertion. */
1449 for (i
= 0; i
< n_rules
; i
++) {
1450 struct test_rule
*rule
;
1451 const struct cls_rule
*old_cr
;
1453 int priority
= priorities
[random_range(n_priorities
)];
1454 int wcf
= wcfs
[random_range(n_tables
)];
1455 int value_pat
= random_uint32() & ((1u << CLS_N_FIELDS
) - 1);
1457 rule
= make_rule(wcf
, priority
, value_pat
);
1458 old_cr
= classifier_find_rule_exactly(&cls
, &rule
->cls_rule
, version
);
1460 classifier_insert(&cls
, &rule
->cls_rule
, version
, NULL
, 0);
1467 xgettimeofday(&start
);
1468 threads
= xmalloc(n_threads
* sizeof *threads
);
1469 for (i
= 0; i
< n_threads
; i
++) {
1470 threads
[i
] = ovs_thread_create("lookups", lookup_classifier
, &aux
);
1472 for (i
= 0; i
< n_threads
; i
++) {
1473 xpthread_join(threads
[i
], NULL
);
1476 int elapsed_msec
= elapsed(&start
);
1481 atomic_read(&aux
.hits
, &hits
);
1482 atomic_read(&aux
.misses
, &misses
);
1483 printf("hits: %d, misses: %d\n", hits
, misses
);
1485 printf("classifier lookups: %5d ms, %"PRId64
" lookups/sec\n",
1487 (((uint64_t)hits
+ misses
) * 1000) / elapsed_msec
);
1489 destroy_classifier(&cls
);
1490 free(aux
.lookup_flows
);
1495 /* Miniflow tests. */
1500 static const uint32_t values_
[] =
1501 { 0xffffffff, 0xaaaaaaaa, 0x55555555, 0x80000000,
1502 0x00000001, 0xface0000, 0x00d00d1e, 0xdeadbeef };
1504 return values_
[random_range(ARRAY_SIZE(values_
))];
1508 choose(unsigned int n
, unsigned int *idxp
)
1518 #define FLOW_U32S (FLOW_U64S * 2)
1521 init_consecutive_values(int n_consecutive
, struct flow
*flow
,
1524 uint32_t *flow_u32
= (uint32_t *) flow
;
1526 if (choose(FLOW_U32S
- n_consecutive
+ 1, idxp
)) {
1529 for (i
= 0; i
< n_consecutive
; i
++) {
1530 flow_u32
[*idxp
+ i
] = random_value();
1539 next_random_flow(struct flow
*flow
, unsigned int idx
)
1541 uint32_t *flow_u32
= (uint32_t *) flow
;
1543 memset(flow
, 0, sizeof *flow
);
1546 if (choose(1, &idx
)) {
1550 /* All flows with a small number of consecutive nonzero values. */
1551 for (int i
= 1; i
<= 4; i
++) {
1552 if (init_consecutive_values(i
, flow
, &idx
)) {
1557 /* All flows with a large number of consecutive nonzero values. */
1558 for (int i
= FLOW_U32S
- 4; i
<= FLOW_U32S
; i
++) {
1559 if (init_consecutive_values(i
, flow
, &idx
)) {
1564 /* All flows with exactly two nonconsecutive nonzero values. */
1565 if (choose((FLOW_U32S
- 1) * (FLOW_U32S
- 2) / 2, &idx
)) {
1568 for (ofs1
= 0; ofs1
< FLOW_U32S
- 2; ofs1
++) {
1571 for (ofs2
= ofs1
+ 2; ofs2
< FLOW_U32S
; ofs2
++) {
1572 if (choose(1, &idx
)) {
1573 flow_u32
[ofs1
] = random_value();
1574 flow_u32
[ofs2
] = random_value();
1582 /* 16 randomly chosen flows with N >= 3 nonzero values. */
1583 if (choose(16 * (FLOW_U32S
- 4), &idx
)) {
1584 int n
= idx
/ 16 + 3;
1586 for (int i
= 0; i
< n
; i
++) {
1587 flow_u32
[i
] = random_value();
1589 shuffle_u32s(flow_u32
, FLOW_U32S
);
1598 any_random_flow(struct flow
*flow
)
1600 static unsigned int max
;
1602 while (next_random_flow(flow
, max
)) {
1607 next_random_flow(flow
, random_range(max
));
1611 toggle_masked_flow_bits(struct flow
*flow
, const struct flow_wildcards
*mask
)
1613 const uint32_t *mask_u32
= (const uint32_t *) &mask
->masks
;
1614 uint32_t *flow_u32
= (uint32_t *) flow
;
1617 for (i
= 0; i
< FLOW_U32S
; i
++) {
1618 if (mask_u32
[i
] != 0) {
1622 bit
= 1u << random_range(32);
1623 } while (!(bit
& mask_u32
[i
]));
1630 wildcard_extra_bits(struct flow_wildcards
*mask
)
1632 uint32_t *mask_u32
= (uint32_t *) &mask
->masks
;
1635 for (i
= 0; i
< FLOW_U32S
; i
++) {
1636 if (mask_u32
[i
] != 0) {
1640 bit
= 1u << random_range(32);
1641 } while (!(bit
& mask_u32
[i
]));
1642 mask_u32
[i
] &= ~bit
;
1647 /* Returns a copy of 'src'. The caller must eventually free the returned
1648 * miniflow with free(). */
1649 static struct miniflow
*
1650 miniflow_clone__(const struct miniflow
*src
)
1652 struct miniflow
*dst
;
1655 data_size
= miniflow_alloc(&dst
, 1, src
);
1656 miniflow_clone(dst
, src
, data_size
/ sizeof(uint64_t));
1660 /* Returns a hash value for 'flow', given 'basis'. */
1661 static inline uint32_t
1662 miniflow_hash__(const struct miniflow
*flow
, uint32_t basis
)
1664 const uint64_t *p
= miniflow_get_values(flow
);
1665 size_t n_values
= miniflow_n_values(flow
);
1666 struct flowmap hash_map
= FLOWMAP_EMPTY_INITIALIZER
;
1667 uint32_t hash
= basis
;
1670 FLOWMAP_FOR_EACH_INDEX(idx
, flow
->map
) {
1671 uint64_t value
= *p
++;
1674 hash
= hash_add64(hash
, value
);
1675 flowmap_set(&hash_map
, idx
, 1);
1679 FLOWMAP_FOR_EACH_MAP (map
, hash_map
) {
1680 hash
= hash_add64(hash
, map
);
1683 return hash_finish(hash
, n_values
);
1687 test_miniflow(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1692 random_set_seed(0xb3faca38);
1693 for (idx
= 0; next_random_flow(&flow
, idx
); idx
++) {
1694 const uint64_t *flow_u64
= (const uint64_t *) &flow
;
1695 struct miniflow
*miniflow
, *miniflow2
, *miniflow3
;
1696 struct flow flow2
, flow3
;
1697 struct flow_wildcards mask
;
1698 struct minimask
*minimask
;
1701 /* Convert flow to miniflow. */
1702 miniflow
= miniflow_create(&flow
);
1704 /* Check that the flow equals its miniflow. */
1705 for (i
= 0; i
< FLOW_MAX_VLAN_HEADERS
; i
++) {
1706 assert(miniflow_get_vid(miniflow
, i
) ==
1707 vlan_tci_to_vid(flow
.vlans
[i
].tci
));
1709 for (i
= 0; i
< FLOW_U64S
; i
++) {
1710 assert(miniflow_get(miniflow
, i
) == flow_u64
[i
]);
1713 /* Check that the miniflow equals itself. */
1714 assert(miniflow_equal(miniflow
, miniflow
));
1716 /* Convert miniflow back to flow and verify that it's the same. */
1717 miniflow_expand(miniflow
, &flow2
);
1718 assert(flow_equal(&flow
, &flow2
));
1720 /* Check that copying a miniflow works properly. */
1721 miniflow2
= miniflow_clone__(miniflow
);
1722 assert(miniflow_equal(miniflow
, miniflow2
));
1723 assert(miniflow_hash__(miniflow
, 0) == miniflow_hash__(miniflow2
, 0));
1724 miniflow_expand(miniflow2
, &flow3
);
1725 assert(flow_equal(&flow
, &flow3
));
1727 /* Check that masked matches work as expected for identical flows and
1730 next_random_flow(&mask
.masks
, 1);
1731 } while (flow_wildcards_is_catchall(&mask
));
1732 minimask
= minimask_create(&mask
);
1733 assert(minimask_is_catchall(minimask
)
1734 == flow_wildcards_is_catchall(&mask
));
1735 assert(miniflow_equal_in_minimask(miniflow
, miniflow2
, minimask
));
1736 assert(miniflow_equal_flow_in_minimask(miniflow
, &flow2
, minimask
));
1737 assert(miniflow_hash_in_minimask(miniflow
, minimask
, 0x12345678) ==
1738 flow_hash_in_minimask(&flow
, minimask
, 0x12345678));
1739 assert(minimask_hash(minimask
, 0) ==
1740 miniflow_hash__(&minimask
->masks
, 0));
1742 /* Check that masked matches work as expected for differing flows and
1744 toggle_masked_flow_bits(&flow2
, &mask
);
1745 assert(!miniflow_equal_flow_in_minimask(miniflow
, &flow2
, minimask
));
1746 miniflow3
= miniflow_create(&flow2
);
1747 assert(!miniflow_equal_in_minimask(miniflow
, miniflow3
, minimask
));
1758 test_minimask_has_extra(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1760 struct flow_wildcards catchall
;
1761 struct minimask
*minicatchall
;
1765 flow_wildcards_init_catchall(&catchall
);
1766 minicatchall
= minimask_create(&catchall
);
1767 assert(minimask_is_catchall(minicatchall
));
1769 random_set_seed(0x2ec7905b);
1770 for (idx
= 0; next_random_flow(&flow
, idx
); idx
++) {
1771 struct flow_wildcards mask
;
1772 struct minimask
*minimask
;
1775 minimask
= minimask_create(&mask
);
1776 assert(!minimask_has_extra(minimask
, minimask
));
1777 assert(minimask_has_extra(minicatchall
, minimask
)
1778 == !minimask_is_catchall(minimask
));
1779 if (!minimask_is_catchall(minimask
)) {
1780 struct minimask
*minimask2
;
1782 wildcard_extra_bits(&mask
);
1783 minimask2
= minimask_create(&mask
);
1784 assert(minimask_has_extra(minimask2
, minimask
));
1785 assert(!minimask_has_extra(minimask
, minimask2
));
1796 test_minimask_combine(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1798 struct flow_wildcards catchall
;
1799 struct minimask
*minicatchall
;
1803 flow_wildcards_init_catchall(&catchall
);
1804 minicatchall
= minimask_create(&catchall
);
1805 assert(minimask_is_catchall(minicatchall
));
1807 random_set_seed(0x181bf0cd);
1808 for (idx
= 0; next_random_flow(&flow
, idx
); idx
++) {
1809 struct minimask
*minimask
, *minimask2
;
1810 struct flow_wildcards mask
, mask2
, combined
, combined2
;
1812 struct minimask minicombined
;
1813 uint64_t storage
[FLOW_U64S
];
1818 minimask
= minimask_create(&mask
);
1820 minimask_combine(&m
.minicombined
, minimask
, minicatchall
, m
.storage
);
1821 assert(minimask_is_catchall(&m
.minicombined
));
1823 any_random_flow(&flow2
);
1824 mask2
.masks
= flow2
;
1825 minimask2
= minimask_create(&mask2
);
1827 minimask_combine(&m
.minicombined
, minimask
, minimask2
, m
.storage
);
1828 flow_wildcards_and(&combined
, &mask
, &mask2
);
1829 minimask_expand(&m
.minicombined
, &combined2
);
1830 assert(flow_wildcards_equal(&combined
, &combined2
));
1840 static void help(struct ovs_cmdl_context
*ctx
);
1842 static const struct ovs_cmdl_command commands
[] = {
1843 /* Classifier tests. */
1844 {"empty", NULL
, 0, 0, test_empty
, OVS_RO
},
1845 {"destroy-null", NULL
, 0, 0, test_destroy_null
, OVS_RO
},
1846 {"single-rule", NULL
, 0, 0, test_single_rule
, OVS_RO
},
1847 {"rule-replacement", NULL
, 0, 0, test_rule_replacement
, OVS_RO
},
1848 {"many-rules-in-one-list", NULL
, 0, 1, test_many_rules_in_one_list
, OVS_RO
},
1849 {"many-rules-in-one-table", NULL
, 0, 1, test_many_rules_in_one_table
, OVS_RO
},
1850 {"many-rules-in-two-tables", NULL
, 0, 0, test_many_rules_in_two_tables
, OVS_RO
},
1851 {"many-rules-in-five-tables", NULL
, 0, 0, test_many_rules_in_five_tables
, OVS_RO
},
1852 {"benchmark", NULL
, 0, 5, run_benchmarks
, OVS_RO
},
1854 /* Miniflow and minimask tests. */
1855 {"miniflow", NULL
, 0, 0, test_miniflow
, OVS_RO
},
1856 {"minimask_has_extra", NULL
, 0, 0, test_minimask_has_extra
, OVS_RO
},
1857 {"minimask_combine", NULL
, 0, 0, test_minimask_combine
, OVS_RO
},
1859 {"--help", NULL
, 0, 0, help
, OVS_RO
},
1860 {NULL
, NULL
, 0, 0, NULL
, OVS_RO
},
1864 help(struct ovs_cmdl_context
*ctx OVS_UNUSED
)
1866 const struct ovs_cmdl_command
*p
;
1867 struct ds test_names
= DS_EMPTY_INITIALIZER
;
1868 const int linesize
= 80;
1870 printf("usage: ovstest %s TEST [TESTARGS]\n"
1871 "where TEST is one of the following:\n\n",
1874 for (p
= commands
; p
->name
!= NULL
; p
++) {
1875 if (*p
->name
!= '-') { /* Skip internal commands */
1876 if (test_names
.length
> 1
1877 && test_names
.length
+ strlen(p
->name
) + 1 >= linesize
) {
1878 test_names
.length
-= 1;
1879 printf ("%s\n", ds_cstr(&test_names
));
1880 ds_clear(&test_names
);
1882 ds_put_format(&test_names
, "%s, ", p
->name
);
1885 if (test_names
.length
> 2) {
1886 test_names
.length
-= 2;
1887 printf("%s\n", ds_cstr(&test_names
));
1889 ds_destroy(&test_names
);
1893 test_classifier_main(int argc
, char *argv
[])
1895 struct ovs_cmdl_context ctx
= {
1899 set_program_name(argv
[0]);
1901 if (argc
> 1 && !strcmp(argv
[1], "--versioned")) {
1908 ovs_cmdl_run_command(&ctx
, commands
);
1911 OVSTEST_REGISTER("test-classifier", test_classifier_main
);