2 * Copyright (c) 2014, 2015 Nicira, Inc.
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at:
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
19 #include "tnl-ports.h"
25 #include "classifier.h"
26 #include "openvswitch/dynamic-string.h"
28 #include "openvswitch/list.h"
31 #include "ovs-thread.h"
33 #include "ovs-thread.h"
37 static struct ovs_mutex mutex
= OVS_MUTEX_INITIALIZER
;
38 static struct classifier cls
; /* Tunnel ports. */
43 struct in6_addr
*addr
;
47 char dev_name
[IFNAMSIZ
];
50 static struct ovs_list addr_list
;
55 char dev_name
[IFNAMSIZ
];
59 static struct ovs_list port_list
;
64 struct ovs_refcount ref_cnt
;
65 char dev_name
[IFNAMSIZ
];
68 static struct tnl_port_in
*
69 tnl_port_cast(const struct cls_rule
*cr
)
71 BUILD_ASSERT_DECL(offsetof(struct tnl_port_in
, cr
) == 0);
73 return CONTAINER_OF(cr
, struct tnl_port_in
, cr
);
77 tnl_port_free(struct tnl_port_in
*p
)
79 cls_rule_destroy(&p
->cr
);
84 tnl_port_init_flow(struct flow
*flow
, struct eth_addr mac
,
85 struct in6_addr
*addr
, ovs_be16 udp_port
)
87 memset(flow
, 0, sizeof *flow
);
90 if (IN6_IS_ADDR_V4MAPPED(addr
)) {
91 flow
->dl_type
= htons(ETH_TYPE_IP
);
92 flow
->nw_dst
= in6_addr_get_mapped_ipv4(addr
);
94 flow
->dl_type
= htons(ETH_TYPE_IPV6
);
95 flow
->ipv6_dst
= *addr
;
99 flow
->nw_proto
= IPPROTO_UDP
;
101 flow
->nw_proto
= IPPROTO_GRE
;
103 flow
->tp_dst
= udp_port
;
107 map_insert(odp_port_t port
, struct eth_addr mac
, struct in6_addr
*addr
,
108 ovs_be16 udp_port
, const char dev_name
[])
110 const struct cls_rule
*cr
;
111 struct tnl_port_in
*p
;
114 memset(&match
, 0, sizeof match
);
115 tnl_port_init_flow(&match
.flow
, mac
, addr
, udp_port
);
118 cr
= classifier_lookup(&cls
, CLS_MAX_VERSION
, &match
.flow
, NULL
);
119 p
= tnl_port_cast(cr
);
120 /* Try again if the rule was released before we get the reference. */
121 } while (p
&& !ovs_refcount_try_ref_rcu(&p
->ref_cnt
));
124 p
= xzalloc(sizeof *p
);
127 match
.wc
.masks
.dl_type
= OVS_BE16_MAX
;
128 match
.wc
.masks
.nw_proto
= 0xff;
129 /* XXX: No fragments support. */
130 match
.wc
.masks
.nw_frag
= FLOW_NW_FRAG_MASK
;
132 /* 'udp_port' is zero for non-UDP tunnels (e.g. GRE). In this case it
133 * doesn't make sense to match on UDP port numbers. */
135 match
.wc
.masks
.tp_dst
= OVS_BE16_MAX
;
137 if (IN6_IS_ADDR_V4MAPPED(addr
)) {
138 match
.wc
.masks
.nw_dst
= OVS_BE32_MAX
;
140 match
.wc
.masks
.ipv6_dst
= in6addr_exact
;
142 match
.wc
.masks
.vlan_tci
= OVS_BE16_MAX
;
143 memset(&match
.wc
.masks
.dl_dst
, 0xff, sizeof (struct eth_addr
));
145 cls_rule_init(&p
->cr
, &match
, 0); /* Priority == 0. */
146 ovs_refcount_init(&p
->ref_cnt
);
147 ovs_strlcpy(p
->dev_name
, dev_name
, sizeof p
->dev_name
);
149 classifier_insert(&cls
, &p
->cr
, CLS_MIN_VERSION
, NULL
, 0);
154 map_insert_ipdev__(struct ip_device
*ip_dev
, char dev_name
[],
155 odp_port_t port
, ovs_be16 udp_port
)
157 if (ip_dev
->n_addr
) {
160 for (i
= 0; i
< ip_dev
->n_addr
; i
++) {
161 map_insert(port
, ip_dev
->mac
, &ip_dev
->addr
[i
],
168 tnl_port_map_insert(odp_port_t port
,
169 ovs_be16 udp_port
, const char dev_name
[])
172 struct ip_device
*ip_dev
;
174 ovs_mutex_lock(&mutex
);
175 LIST_FOR_EACH(p
, node
, &port_list
) {
176 if (udp_port
== p
->udp_port
) {
181 p
= xzalloc(sizeof *p
);
183 p
->udp_port
= udp_port
;
184 ovs_strlcpy(p
->dev_name
, dev_name
, sizeof p
->dev_name
);
185 list_insert(&port_list
, &p
->node
);
187 LIST_FOR_EACH(ip_dev
, node
, &addr_list
) {
188 map_insert_ipdev__(ip_dev
, p
->dev_name
, p
->port
, p
->udp_port
);
192 ovs_mutex_unlock(&mutex
);
196 tnl_port_unref(const struct cls_rule
*cr
)
198 struct tnl_port_in
*p
= tnl_port_cast(cr
);
200 if (cr
&& ovs_refcount_unref_relaxed(&p
->ref_cnt
) == 1) {
201 if (classifier_remove(&cls
, cr
)) {
202 ovsrcu_postpone(tnl_port_free
, p
);
208 map_delete(struct eth_addr mac
, struct in6_addr
*addr
, ovs_be16 udp_port
)
210 const struct cls_rule
*cr
;
213 tnl_port_init_flow(&flow
, mac
, addr
, udp_port
);
215 cr
= classifier_lookup(&cls
, CLS_MAX_VERSION
, &flow
, NULL
);
220 ipdev_map_delete(struct ip_device
*ip_dev
, ovs_be16 udp_port
)
222 if (ip_dev
->n_addr
) {
225 for (i
= 0; i
< ip_dev
->n_addr
; i
++) {
226 map_delete(ip_dev
->mac
, &ip_dev
->addr
[i
], udp_port
);
232 tnl_port_map_delete(ovs_be16 udp_port
)
234 struct tnl_port
*p
, *next
;
235 struct ip_device
*ip_dev
;
238 ovs_mutex_lock(&mutex
);
239 LIST_FOR_EACH_SAFE(p
, next
, node
, &port_list
) {
240 if (p
->udp_port
== udp_port
) {
241 list_remove(&p
->node
);
250 LIST_FOR_EACH(ip_dev
, node
, &addr_list
) {
251 ipdev_map_delete(ip_dev
, p
->udp_port
);
256 ovs_mutex_unlock(&mutex
);
259 /* 'flow' is non-const to allow for temporary modifications during the lookup.
260 * Any changes are restored before returning. */
262 tnl_port_map_lookup(struct flow
*flow
, struct flow_wildcards
*wc
)
264 const struct cls_rule
*cr
= classifier_lookup(&cls
, CLS_MAX_VERSION
, flow
,
267 return (cr
) ? tnl_port_cast(cr
)->portno
: ODPP_NONE
;
271 tnl_port_show_v(struct ds
*ds
)
273 const struct tnl_port_in
*p
;
275 CLS_FOR_EACH(p
, cr
, &cls
) {
276 struct odputil_keybuf keybuf
;
277 struct odputil_keybuf maskbuf
;
279 const struct nlattr
*key
, *mask
;
280 size_t key_len
, mask_len
;
281 struct flow_wildcards wc
;
283 struct odp_flow_key_parms odp_parms
= {
288 ds_put_format(ds
, "%s (%"PRIu32
") : ", p
->dev_name
, p
->portno
);
289 minimask_expand(p
->cr
.match
.mask
, &wc
);
290 miniflow_expand(p
->cr
.match
.flow
, &flow
);
293 odp_parms
.odp_in_port
= flow
.in_port
.odp_port
;
294 odp_parms
.support
.recirc
= true;
295 ofpbuf_use_stack(&buf
, &keybuf
, sizeof keybuf
);
296 odp_flow_key_from_flow(&odp_parms
, &buf
);
301 odp_parms
.odp_in_port
= wc
.masks
.in_port
.odp_port
;
302 odp_parms
.support
.recirc
= false;
303 ofpbuf_use_stack(&buf
, &maskbuf
, sizeof maskbuf
);
304 odp_flow_key_from_mask(&odp_parms
, &buf
);
309 odp_flow_format(key
, key_len
, mask
, mask_len
, NULL
, ds
, false);
310 ds_put_format(ds
, "\n");
315 tnl_port_show(struct unixctl_conn
*conn
, int argc OVS_UNUSED
,
316 const char *argv
[] OVS_UNUSED
, void *aux OVS_UNUSED
)
318 struct ds ds
= DS_EMPTY_INITIALIZER
;
321 ds_put_format(&ds
, "Listening ports:\n");
322 ovs_mutex_lock(&mutex
);
324 if (!strcasecmp(argv
[1], "-v")) {
325 tnl_port_show_v(&ds
);
330 LIST_FOR_EACH(p
, node
, &port_list
) {
331 ds_put_format(&ds
, "%s (%"PRIu32
")\n", p
->dev_name
, p
->port
);
335 ovs_mutex_unlock(&mutex
);
336 unixctl_command_reply(conn
, ds_cstr(&ds
));
341 map_insert_ipdev(struct ip_device
*ip_dev
)
345 LIST_FOR_EACH(p
, node
, &port_list
) {
346 map_insert_ipdev__(ip_dev
, p
->dev_name
, p
->port
, p
->udp_port
);
351 insert_ipdev__(struct netdev
*dev
,
352 struct in6_addr
*addr
, int n_addr
)
354 struct ip_device
*ip_dev
;
355 enum netdev_flags flags
;
358 error
= netdev_get_flags(dev
, &flags
);
359 if (error
|| (flags
& NETDEV_LOOPBACK
)) {
363 ip_dev
= xzalloc(sizeof *ip_dev
);
364 ip_dev
->dev
= netdev_ref(dev
);
365 ip_dev
->change_seq
= netdev_get_change_seq(dev
);
366 error
= netdev_get_etheraddr(ip_dev
->dev
, &ip_dev
->mac
);
371 ip_dev
->n_addr
= n_addr
;
372 ovs_strlcpy(ip_dev
->dev_name
, netdev_get_name(dev
), sizeof ip_dev
->dev_name
);
373 list_insert(&addr_list
, &ip_dev
->node
);
374 map_insert_ipdev(ip_dev
);
378 netdev_close(ip_dev
->dev
);
385 insert_ipdev(const char dev_name
[])
387 struct in6_addr
*addr
, *mask
;
391 error
= netdev_open(dev_name
, NULL
, &dev
);
396 error
= netdev_get_addr_list(dev
, &addr
, &mask
, &n_in6
);
402 insert_ipdev__(dev
, addr
, n_in6
);
407 delete_ipdev(struct ip_device
*ip_dev
)
411 LIST_FOR_EACH(p
, node
, &port_list
) {
412 ipdev_map_delete(ip_dev
, p
->udp_port
);
415 list_remove(&ip_dev
->node
);
416 netdev_close(ip_dev
->dev
);
422 tnl_port_map_insert_ipdev(const char dev_name
[])
424 struct ip_device
*ip_dev
, *next
;
426 ovs_mutex_lock(&mutex
);
428 LIST_FOR_EACH_SAFE(ip_dev
, next
, node
, &addr_list
) {
429 if (!strcmp(netdev_get_name(ip_dev
->dev
), dev_name
)) {
430 if (ip_dev
->change_seq
== netdev_get_change_seq(ip_dev
->dev
)) {
433 /* Address changed. */
434 delete_ipdev(ip_dev
);
437 insert_ipdev(dev_name
);
440 ovs_mutex_unlock(&mutex
);
444 tnl_port_map_delete_ipdev(const char dev_name
[])
446 struct ip_device
*ip_dev
, *next
;
448 ovs_mutex_lock(&mutex
);
449 LIST_FOR_EACH_SAFE(ip_dev
, next
, node
, &addr_list
) {
450 if (!strcmp(netdev_get_name(ip_dev
->dev
), dev_name
)) {
451 delete_ipdev(ip_dev
);
454 ovs_mutex_unlock(&mutex
);
458 tnl_port_map_run(void)
460 struct ip_device
*ip_dev
, *next
;
462 ovs_mutex_lock(&mutex
);
463 LIST_FOR_EACH_SAFE(ip_dev
, next
, node
, &addr_list
) {
464 char dev_name
[IFNAMSIZ
];
466 if (ip_dev
->change_seq
== netdev_get_change_seq(ip_dev
->dev
)) {
470 /* Address changed. */
471 ovs_strlcpy(dev_name
, ip_dev
->dev_name
, sizeof dev_name
);
472 delete_ipdev(ip_dev
);
473 insert_ipdev(dev_name
);
475 ovs_mutex_unlock(&mutex
);
479 tnl_port_map_init(void)
481 classifier_init(&cls
, flow_segment_u64s
);
482 list_init(&addr_list
);
483 list_init(&port_list
);
484 unixctl_command_register("tnl/ports/show", "-v", 0, 1, tnl_port_show
, NULL
);