2 * net/sched/cls_flower.c Flower classifier
4 * Copyright (c) 2015 Jiri Pirko <jiri@resnulli.us>
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
12 #include <linux/kernel.h>
13 #include <linux/init.h>
14 #include <linux/module.h>
15 #include <linux/rhashtable.h>
16 #include <linux/workqueue.h>
18 #include <linux/if_ether.h>
19 #include <linux/in6.h>
21 #include <linux/mpls.h>
23 #include <net/sch_generic.h>
24 #include <net/pkt_cls.h>
26 #include <net/flow_dissector.h>
29 #include <net/dst_metadata.h>
33 struct flow_dissector_key_control control
;
34 struct flow_dissector_key_control enc_control
;
35 struct flow_dissector_key_basic basic
;
36 struct flow_dissector_key_eth_addrs eth
;
37 struct flow_dissector_key_vlan vlan
;
39 struct flow_dissector_key_ipv4_addrs ipv4
;
40 struct flow_dissector_key_ipv6_addrs ipv6
;
42 struct flow_dissector_key_ports tp
;
43 struct flow_dissector_key_icmp icmp
;
44 struct flow_dissector_key_arp arp
;
45 struct flow_dissector_key_keyid enc_key_id
;
47 struct flow_dissector_key_ipv4_addrs enc_ipv4
;
48 struct flow_dissector_key_ipv6_addrs enc_ipv6
;
50 struct flow_dissector_key_ports enc_tp
;
51 struct flow_dissector_key_mpls mpls
;
52 struct flow_dissector_key_tcp tcp
;
53 struct flow_dissector_key_ip ip
;
54 } __aligned(BITS_PER_LONG
/ 8); /* Ensure that we can do comparisons as longs. */
56 struct fl_flow_mask_range
{
57 unsigned short int start
;
58 unsigned short int end
;
62 struct fl_flow_key key
;
63 struct fl_flow_mask_range range
;
69 struct fl_flow_mask mask
;
70 struct flow_dissector dissector
;
72 struct list_head filters
;
73 struct rhashtable_params ht_params
;
74 struct rcu_work rwork
;
75 struct idr handle_idr
;
78 struct cls_fl_filter
{
79 struct rhash_head ht_node
;
80 struct fl_flow_key mkey
;
82 struct tcf_result res
;
83 struct fl_flow_key key
;
84 struct list_head list
;
87 struct rcu_work rwork
;
88 struct net_device
*hw_dev
;
91 static unsigned short int fl_mask_range(const struct fl_flow_mask
*mask
)
93 return mask
->range
.end
- mask
->range
.start
;
96 static void fl_mask_update_range(struct fl_flow_mask
*mask
)
98 const u8
*bytes
= (const u8
*) &mask
->key
;
99 size_t size
= sizeof(mask
->key
);
100 size_t i
, first
= 0, last
= size
- 1;
102 for (i
= 0; i
< sizeof(mask
->key
); i
++) {
109 mask
->range
.start
= rounddown(first
, sizeof(long));
110 mask
->range
.end
= roundup(last
+ 1, sizeof(long));
113 static void *fl_key_get_start(struct fl_flow_key
*key
,
114 const struct fl_flow_mask
*mask
)
116 return (u8
*) key
+ mask
->range
.start
;
119 static void fl_set_masked_key(struct fl_flow_key
*mkey
, struct fl_flow_key
*key
,
120 struct fl_flow_mask
*mask
)
122 const long *lkey
= fl_key_get_start(key
, mask
);
123 const long *lmask
= fl_key_get_start(&mask
->key
, mask
);
124 long *lmkey
= fl_key_get_start(mkey
, mask
);
127 for (i
= 0; i
< fl_mask_range(mask
); i
+= sizeof(long))
128 *lmkey
++ = *lkey
++ & *lmask
++;
131 static void fl_clear_masked_range(struct fl_flow_key
*key
,
132 struct fl_flow_mask
*mask
)
134 memset(fl_key_get_start(key
, mask
), 0, fl_mask_range(mask
));
137 static struct cls_fl_filter
*fl_lookup(struct cls_fl_head
*head
,
138 struct fl_flow_key
*mkey
)
140 return rhashtable_lookup_fast(&head
->ht
,
141 fl_key_get_start(mkey
, &head
->mask
),
145 static int fl_classify(struct sk_buff
*skb
, const struct tcf_proto
*tp
,
146 struct tcf_result
*res
)
148 struct cls_fl_head
*head
= rcu_dereference_bh(tp
->root
);
149 struct cls_fl_filter
*f
;
150 struct fl_flow_key skb_key
;
151 struct fl_flow_key skb_mkey
;
153 if (!atomic_read(&head
->ht
.nelems
))
156 fl_clear_masked_range(&skb_key
, &head
->mask
);
158 skb_key
.indev_ifindex
= skb
->skb_iif
;
159 /* skb_flow_dissect() does not set n_proto in case an unknown protocol,
160 * so do it rather here.
162 skb_key
.basic
.n_proto
= skb
->protocol
;
163 skb_flow_dissect(skb
, &head
->dissector
, &skb_key
, 0);
165 fl_set_masked_key(&skb_mkey
, &skb_key
, &head
->mask
);
167 f
= fl_lookup(head
, &skb_mkey
);
168 if (f
&& !tc_skip_sw(f
->flags
)) {
170 return tcf_exts_exec(skb
, &f
->exts
, res
);
175 static int fl_init(struct tcf_proto
*tp
)
177 struct cls_fl_head
*head
;
179 head
= kzalloc(sizeof(*head
), GFP_KERNEL
);
183 INIT_LIST_HEAD_RCU(&head
->filters
);
184 rcu_assign_pointer(tp
->root
, head
);
185 idr_init(&head
->handle_idr
);
190 static void __fl_destroy_filter(struct cls_fl_filter
*f
)
192 tcf_exts_destroy(&f
->exts
);
193 tcf_exts_put_net(&f
->exts
);
197 static void fl_destroy_filter_work(struct work_struct
*work
)
199 struct cls_fl_filter
*f
= container_of(to_rcu_work(work
),
200 struct cls_fl_filter
, rwork
);
203 __fl_destroy_filter(f
);
207 static void fl_hw_destroy_filter(struct tcf_proto
*tp
, struct cls_fl_filter
*f
)
209 struct tc_cls_flower_offload cls_flower
= {};
210 struct tcf_block
*block
= tp
->chain
->block
;
212 tc_cls_common_offload_init(&cls_flower
.common
, tp
);
213 cls_flower
.command
= TC_CLSFLOWER_DESTROY
;
214 cls_flower
.cookie
= (unsigned long) f
;
216 tc_setup_cb_call(block
, &f
->exts
, TC_SETUP_CLSFLOWER
,
220 static int fl_hw_replace_filter(struct tcf_proto
*tp
,
221 struct flow_dissector
*dissector
,
222 struct fl_flow_key
*mask
,
223 struct cls_fl_filter
*f
)
225 struct tc_cls_flower_offload cls_flower
= {};
226 struct tcf_block
*block
= tp
->chain
->block
;
227 bool skip_sw
= tc_skip_sw(f
->flags
);
230 tc_cls_common_offload_init(&cls_flower
.common
, tp
);
231 cls_flower
.command
= TC_CLSFLOWER_REPLACE
;
232 cls_flower
.cookie
= (unsigned long) f
;
233 cls_flower
.dissector
= dissector
;
234 cls_flower
.mask
= mask
;
235 cls_flower
.key
= &f
->mkey
;
236 cls_flower
.exts
= &f
->exts
;
237 cls_flower
.classid
= f
->res
.classid
;
239 err
= tc_setup_cb_call(block
, &f
->exts
, TC_SETUP_CLSFLOWER
,
240 &cls_flower
, skip_sw
);
242 fl_hw_destroy_filter(tp
, f
);
244 } else if (err
> 0) {
245 f
->flags
|= TCA_CLS_FLAGS_IN_HW
;
248 if (skip_sw
&& !(f
->flags
& TCA_CLS_FLAGS_IN_HW
))
254 static void fl_hw_update_stats(struct tcf_proto
*tp
, struct cls_fl_filter
*f
)
256 struct tc_cls_flower_offload cls_flower
= {};
257 struct tcf_block
*block
= tp
->chain
->block
;
259 tc_cls_common_offload_init(&cls_flower
.common
, tp
);
260 cls_flower
.command
= TC_CLSFLOWER_STATS
;
261 cls_flower
.cookie
= (unsigned long) f
;
262 cls_flower
.exts
= &f
->exts
;
263 cls_flower
.classid
= f
->res
.classid
;
265 tc_setup_cb_call(block
, &f
->exts
, TC_SETUP_CLSFLOWER
,
269 static void __fl_delete(struct tcf_proto
*tp
, struct cls_fl_filter
*f
)
271 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
273 idr_remove_ext(&head
->handle_idr
, f
->handle
);
274 list_del_rcu(&f
->list
);
275 if (!tc_skip_hw(f
->flags
))
276 fl_hw_destroy_filter(tp
, f
);
277 tcf_unbind_filter(tp
, &f
->res
);
278 if (tcf_exts_get_net(&f
->exts
))
279 tcf_queue_work(&f
->rwork
, fl_destroy_filter_work
);
281 __fl_destroy_filter(f
);
284 static void fl_destroy_sleepable(struct work_struct
*work
)
286 struct cls_fl_head
*head
= container_of(to_rcu_work(work
),
289 if (head
->mask_assigned
)
290 rhashtable_destroy(&head
->ht
);
292 module_put(THIS_MODULE
);
295 static void fl_destroy(struct tcf_proto
*tp
)
297 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
298 struct cls_fl_filter
*f
, *next
;
300 list_for_each_entry_safe(f
, next
, &head
->filters
, list
)
302 idr_destroy(&head
->handle_idr
);
304 __module_get(THIS_MODULE
);
305 tcf_queue_work(&head
->rwork
, fl_destroy_sleepable
);
308 static void *fl_get(struct tcf_proto
*tp
, u32 handle
)
310 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
312 return idr_find_ext(&head
->handle_idr
, handle
);
315 static const struct nla_policy fl_policy
[TCA_FLOWER_MAX
+ 1] = {
316 [TCA_FLOWER_UNSPEC
] = { .type
= NLA_UNSPEC
},
317 [TCA_FLOWER_CLASSID
] = { .type
= NLA_U32
},
318 [TCA_FLOWER_INDEV
] = { .type
= NLA_STRING
,
320 [TCA_FLOWER_KEY_ETH_DST
] = { .len
= ETH_ALEN
},
321 [TCA_FLOWER_KEY_ETH_DST_MASK
] = { .len
= ETH_ALEN
},
322 [TCA_FLOWER_KEY_ETH_SRC
] = { .len
= ETH_ALEN
},
323 [TCA_FLOWER_KEY_ETH_SRC_MASK
] = { .len
= ETH_ALEN
},
324 [TCA_FLOWER_KEY_ETH_TYPE
] = { .type
= NLA_U16
},
325 [TCA_FLOWER_KEY_IP_PROTO
] = { .type
= NLA_U8
},
326 [TCA_FLOWER_KEY_IPV4_SRC
] = { .type
= NLA_U32
},
327 [TCA_FLOWER_KEY_IPV4_SRC_MASK
] = { .type
= NLA_U32
},
328 [TCA_FLOWER_KEY_IPV4_DST
] = { .type
= NLA_U32
},
329 [TCA_FLOWER_KEY_IPV4_DST_MASK
] = { .type
= NLA_U32
},
330 [TCA_FLOWER_KEY_IPV6_SRC
] = { .len
= sizeof(struct in6_addr
) },
331 [TCA_FLOWER_KEY_IPV6_SRC_MASK
] = { .len
= sizeof(struct in6_addr
) },
332 [TCA_FLOWER_KEY_IPV6_DST
] = { .len
= sizeof(struct in6_addr
) },
333 [TCA_FLOWER_KEY_IPV6_DST_MASK
] = { .len
= sizeof(struct in6_addr
) },
334 [TCA_FLOWER_KEY_TCP_SRC
] = { .type
= NLA_U16
},
335 [TCA_FLOWER_KEY_TCP_DST
] = { .type
= NLA_U16
},
336 [TCA_FLOWER_KEY_UDP_SRC
] = { .type
= NLA_U16
},
337 [TCA_FLOWER_KEY_UDP_DST
] = { .type
= NLA_U16
},
338 [TCA_FLOWER_KEY_VLAN_ID
] = { .type
= NLA_U16
},
339 [TCA_FLOWER_KEY_VLAN_PRIO
] = { .type
= NLA_U8
},
340 [TCA_FLOWER_KEY_VLAN_ETH_TYPE
] = { .type
= NLA_U16
},
341 [TCA_FLOWER_KEY_ENC_KEY_ID
] = { .type
= NLA_U32
},
342 [TCA_FLOWER_KEY_ENC_IPV4_SRC
] = { .type
= NLA_U32
},
343 [TCA_FLOWER_KEY_ENC_IPV4_SRC_MASK
] = { .type
= NLA_U32
},
344 [TCA_FLOWER_KEY_ENC_IPV4_DST
] = { .type
= NLA_U32
},
345 [TCA_FLOWER_KEY_ENC_IPV4_DST_MASK
] = { .type
= NLA_U32
},
346 [TCA_FLOWER_KEY_ENC_IPV6_SRC
] = { .len
= sizeof(struct in6_addr
) },
347 [TCA_FLOWER_KEY_ENC_IPV6_SRC_MASK
] = { .len
= sizeof(struct in6_addr
) },
348 [TCA_FLOWER_KEY_ENC_IPV6_DST
] = { .len
= sizeof(struct in6_addr
) },
349 [TCA_FLOWER_KEY_ENC_IPV6_DST_MASK
] = { .len
= sizeof(struct in6_addr
) },
350 [TCA_FLOWER_KEY_TCP_SRC_MASK
] = { .type
= NLA_U16
},
351 [TCA_FLOWER_KEY_TCP_DST_MASK
] = { .type
= NLA_U16
},
352 [TCA_FLOWER_KEY_UDP_SRC_MASK
] = { .type
= NLA_U16
},
353 [TCA_FLOWER_KEY_UDP_DST_MASK
] = { .type
= NLA_U16
},
354 [TCA_FLOWER_KEY_SCTP_SRC_MASK
] = { .type
= NLA_U16
},
355 [TCA_FLOWER_KEY_SCTP_DST_MASK
] = { .type
= NLA_U16
},
356 [TCA_FLOWER_KEY_SCTP_SRC
] = { .type
= NLA_U16
},
357 [TCA_FLOWER_KEY_SCTP_DST
] = { .type
= NLA_U16
},
358 [TCA_FLOWER_KEY_ENC_UDP_SRC_PORT
] = { .type
= NLA_U16
},
359 [TCA_FLOWER_KEY_ENC_UDP_SRC_PORT_MASK
] = { .type
= NLA_U16
},
360 [TCA_FLOWER_KEY_ENC_UDP_DST_PORT
] = { .type
= NLA_U16
},
361 [TCA_FLOWER_KEY_ENC_UDP_DST_PORT_MASK
] = { .type
= NLA_U16
},
362 [TCA_FLOWER_KEY_FLAGS
] = { .type
= NLA_U32
},
363 [TCA_FLOWER_KEY_FLAGS_MASK
] = { .type
= NLA_U32
},
364 [TCA_FLOWER_KEY_ICMPV4_TYPE
] = { .type
= NLA_U8
},
365 [TCA_FLOWER_KEY_ICMPV4_TYPE_MASK
] = { .type
= NLA_U8
},
366 [TCA_FLOWER_KEY_ICMPV4_CODE
] = { .type
= NLA_U8
},
367 [TCA_FLOWER_KEY_ICMPV4_CODE_MASK
] = { .type
= NLA_U8
},
368 [TCA_FLOWER_KEY_ICMPV6_TYPE
] = { .type
= NLA_U8
},
369 [TCA_FLOWER_KEY_ICMPV6_TYPE_MASK
] = { .type
= NLA_U8
},
370 [TCA_FLOWER_KEY_ICMPV6_CODE
] = { .type
= NLA_U8
},
371 [TCA_FLOWER_KEY_ICMPV6_CODE_MASK
] = { .type
= NLA_U8
},
372 [TCA_FLOWER_KEY_ARP_SIP
] = { .type
= NLA_U32
},
373 [TCA_FLOWER_KEY_ARP_SIP_MASK
] = { .type
= NLA_U32
},
374 [TCA_FLOWER_KEY_ARP_TIP
] = { .type
= NLA_U32
},
375 [TCA_FLOWER_KEY_ARP_TIP_MASK
] = { .type
= NLA_U32
},
376 [TCA_FLOWER_KEY_ARP_OP
] = { .type
= NLA_U8
},
377 [TCA_FLOWER_KEY_ARP_OP_MASK
] = { .type
= NLA_U8
},
378 [TCA_FLOWER_KEY_ARP_SHA
] = { .len
= ETH_ALEN
},
379 [TCA_FLOWER_KEY_ARP_SHA_MASK
] = { .len
= ETH_ALEN
},
380 [TCA_FLOWER_KEY_ARP_THA
] = { .len
= ETH_ALEN
},
381 [TCA_FLOWER_KEY_ARP_THA_MASK
] = { .len
= ETH_ALEN
},
382 [TCA_FLOWER_KEY_MPLS_TTL
] = { .type
= NLA_U8
},
383 [TCA_FLOWER_KEY_MPLS_BOS
] = { .type
= NLA_U8
},
384 [TCA_FLOWER_KEY_MPLS_TC
] = { .type
= NLA_U8
},
385 [TCA_FLOWER_KEY_MPLS_LABEL
] = { .type
= NLA_U32
},
386 [TCA_FLOWER_KEY_TCP_FLAGS
] = { .type
= NLA_U16
},
387 [TCA_FLOWER_KEY_TCP_FLAGS_MASK
] = { .type
= NLA_U16
},
388 [TCA_FLOWER_KEY_IP_TOS
] = { .type
= NLA_U8
},
389 [TCA_FLOWER_KEY_IP_TOS_MASK
] = { .type
= NLA_U8
},
390 [TCA_FLOWER_KEY_IP_TTL
] = { .type
= NLA_U8
},
391 [TCA_FLOWER_KEY_IP_TTL_MASK
] = { .type
= NLA_U8
},
394 static void fl_set_key_val(struct nlattr
**tb
,
395 void *val
, int val_type
,
396 void *mask
, int mask_type
, int len
)
400 memcpy(val
, nla_data(tb
[val_type
]), len
);
401 if (mask_type
== TCA_FLOWER_UNSPEC
|| !tb
[mask_type
])
402 memset(mask
, 0xff, len
);
404 memcpy(mask
, nla_data(tb
[mask_type
]), len
);
407 static int fl_set_key_mpls(struct nlattr
**tb
,
408 struct flow_dissector_key_mpls
*key_val
,
409 struct flow_dissector_key_mpls
*key_mask
)
411 if (tb
[TCA_FLOWER_KEY_MPLS_TTL
]) {
412 key_val
->mpls_ttl
= nla_get_u8(tb
[TCA_FLOWER_KEY_MPLS_TTL
]);
413 key_mask
->mpls_ttl
= MPLS_TTL_MASK
;
415 if (tb
[TCA_FLOWER_KEY_MPLS_BOS
]) {
416 u8 bos
= nla_get_u8(tb
[TCA_FLOWER_KEY_MPLS_BOS
]);
418 if (bos
& ~MPLS_BOS_MASK
)
420 key_val
->mpls_bos
= bos
;
421 key_mask
->mpls_bos
= MPLS_BOS_MASK
;
423 if (tb
[TCA_FLOWER_KEY_MPLS_TC
]) {
424 u8 tc
= nla_get_u8(tb
[TCA_FLOWER_KEY_MPLS_TC
]);
426 if (tc
& ~MPLS_TC_MASK
)
428 key_val
->mpls_tc
= tc
;
429 key_mask
->mpls_tc
= MPLS_TC_MASK
;
431 if (tb
[TCA_FLOWER_KEY_MPLS_LABEL
]) {
432 u32 label
= nla_get_u32(tb
[TCA_FLOWER_KEY_MPLS_LABEL
]);
434 if (label
& ~MPLS_LABEL_MASK
)
436 key_val
->mpls_label
= label
;
437 key_mask
->mpls_label
= MPLS_LABEL_MASK
;
442 static void fl_set_key_vlan(struct nlattr
**tb
,
443 struct flow_dissector_key_vlan
*key_val
,
444 struct flow_dissector_key_vlan
*key_mask
)
446 #define VLAN_PRIORITY_MASK 0x7
448 if (tb
[TCA_FLOWER_KEY_VLAN_ID
]) {
450 nla_get_u16(tb
[TCA_FLOWER_KEY_VLAN_ID
]) & VLAN_VID_MASK
;
451 key_mask
->vlan_id
= VLAN_VID_MASK
;
453 if (tb
[TCA_FLOWER_KEY_VLAN_PRIO
]) {
454 key_val
->vlan_priority
=
455 nla_get_u8(tb
[TCA_FLOWER_KEY_VLAN_PRIO
]) &
457 key_mask
->vlan_priority
= VLAN_PRIORITY_MASK
;
461 static void fl_set_key_flag(u32 flower_key
, u32 flower_mask
,
462 u32
*dissector_key
, u32
*dissector_mask
,
463 u32 flower_flag_bit
, u32 dissector_flag_bit
)
465 if (flower_mask
& flower_flag_bit
) {
466 *dissector_mask
|= dissector_flag_bit
;
467 if (flower_key
& flower_flag_bit
)
468 *dissector_key
|= dissector_flag_bit
;
472 static int fl_set_key_flags(struct nlattr
**tb
,
473 u32
*flags_key
, u32
*flags_mask
)
477 /* mask is mandatory for flags */
478 if (!tb
[TCA_FLOWER_KEY_FLAGS_MASK
])
481 key
= be32_to_cpu(nla_get_u32(tb
[TCA_FLOWER_KEY_FLAGS
]));
482 mask
= be32_to_cpu(nla_get_u32(tb
[TCA_FLOWER_KEY_FLAGS_MASK
]));
487 fl_set_key_flag(key
, mask
, flags_key
, flags_mask
,
488 TCA_FLOWER_KEY_FLAGS_IS_FRAGMENT
, FLOW_DIS_IS_FRAGMENT
);
493 static void fl_set_key_ip(struct nlattr
**tb
,
494 struct flow_dissector_key_ip
*key
,
495 struct flow_dissector_key_ip
*mask
)
497 fl_set_key_val(tb
, &key
->tos
, TCA_FLOWER_KEY_IP_TOS
,
498 &mask
->tos
, TCA_FLOWER_KEY_IP_TOS_MASK
,
501 fl_set_key_val(tb
, &key
->ttl
, TCA_FLOWER_KEY_IP_TTL
,
502 &mask
->ttl
, TCA_FLOWER_KEY_IP_TTL_MASK
,
506 static int fl_set_key(struct net
*net
, struct nlattr
**tb
,
507 struct fl_flow_key
*key
, struct fl_flow_key
*mask
)
511 #ifdef CONFIG_NET_CLS_IND
512 if (tb
[TCA_FLOWER_INDEV
]) {
513 int err
= tcf_change_indev(net
, tb
[TCA_FLOWER_INDEV
]);
516 key
->indev_ifindex
= err
;
517 mask
->indev_ifindex
= 0xffffffff;
521 fl_set_key_val(tb
, key
->eth
.dst
, TCA_FLOWER_KEY_ETH_DST
,
522 mask
->eth
.dst
, TCA_FLOWER_KEY_ETH_DST_MASK
,
523 sizeof(key
->eth
.dst
));
524 fl_set_key_val(tb
, key
->eth
.src
, TCA_FLOWER_KEY_ETH_SRC
,
525 mask
->eth
.src
, TCA_FLOWER_KEY_ETH_SRC_MASK
,
526 sizeof(key
->eth
.src
));
528 if (tb
[TCA_FLOWER_KEY_ETH_TYPE
]) {
529 ethertype
= nla_get_be16(tb
[TCA_FLOWER_KEY_ETH_TYPE
]);
531 if (ethertype
== htons(ETH_P_8021Q
)) {
532 fl_set_key_vlan(tb
, &key
->vlan
, &mask
->vlan
);
533 fl_set_key_val(tb
, &key
->basic
.n_proto
,
534 TCA_FLOWER_KEY_VLAN_ETH_TYPE
,
535 &mask
->basic
.n_proto
, TCA_FLOWER_UNSPEC
,
536 sizeof(key
->basic
.n_proto
));
538 key
->basic
.n_proto
= ethertype
;
539 mask
->basic
.n_proto
= cpu_to_be16(~0);
543 if (key
->basic
.n_proto
== htons(ETH_P_IP
) ||
544 key
->basic
.n_proto
== htons(ETH_P_IPV6
)) {
545 fl_set_key_val(tb
, &key
->basic
.ip_proto
, TCA_FLOWER_KEY_IP_PROTO
,
546 &mask
->basic
.ip_proto
, TCA_FLOWER_UNSPEC
,
547 sizeof(key
->basic
.ip_proto
));
548 fl_set_key_ip(tb
, &key
->ip
, &mask
->ip
);
551 if (tb
[TCA_FLOWER_KEY_IPV4_SRC
] || tb
[TCA_FLOWER_KEY_IPV4_DST
]) {
552 key
->control
.addr_type
= FLOW_DISSECTOR_KEY_IPV4_ADDRS
;
553 mask
->control
.addr_type
= ~0;
554 fl_set_key_val(tb
, &key
->ipv4
.src
, TCA_FLOWER_KEY_IPV4_SRC
,
555 &mask
->ipv4
.src
, TCA_FLOWER_KEY_IPV4_SRC_MASK
,
556 sizeof(key
->ipv4
.src
));
557 fl_set_key_val(tb
, &key
->ipv4
.dst
, TCA_FLOWER_KEY_IPV4_DST
,
558 &mask
->ipv4
.dst
, TCA_FLOWER_KEY_IPV4_DST_MASK
,
559 sizeof(key
->ipv4
.dst
));
560 } else if (tb
[TCA_FLOWER_KEY_IPV6_SRC
] || tb
[TCA_FLOWER_KEY_IPV6_DST
]) {
561 key
->control
.addr_type
= FLOW_DISSECTOR_KEY_IPV6_ADDRS
;
562 mask
->control
.addr_type
= ~0;
563 fl_set_key_val(tb
, &key
->ipv6
.src
, TCA_FLOWER_KEY_IPV6_SRC
,
564 &mask
->ipv6
.src
, TCA_FLOWER_KEY_IPV6_SRC_MASK
,
565 sizeof(key
->ipv6
.src
));
566 fl_set_key_val(tb
, &key
->ipv6
.dst
, TCA_FLOWER_KEY_IPV6_DST
,
567 &mask
->ipv6
.dst
, TCA_FLOWER_KEY_IPV6_DST_MASK
,
568 sizeof(key
->ipv6
.dst
));
571 if (key
->basic
.ip_proto
== IPPROTO_TCP
) {
572 fl_set_key_val(tb
, &key
->tp
.src
, TCA_FLOWER_KEY_TCP_SRC
,
573 &mask
->tp
.src
, TCA_FLOWER_KEY_TCP_SRC_MASK
,
574 sizeof(key
->tp
.src
));
575 fl_set_key_val(tb
, &key
->tp
.dst
, TCA_FLOWER_KEY_TCP_DST
,
576 &mask
->tp
.dst
, TCA_FLOWER_KEY_TCP_DST_MASK
,
577 sizeof(key
->tp
.dst
));
578 fl_set_key_val(tb
, &key
->tcp
.flags
, TCA_FLOWER_KEY_TCP_FLAGS
,
579 &mask
->tcp
.flags
, TCA_FLOWER_KEY_TCP_FLAGS_MASK
,
580 sizeof(key
->tcp
.flags
));
581 } else if (key
->basic
.ip_proto
== IPPROTO_UDP
) {
582 fl_set_key_val(tb
, &key
->tp
.src
, TCA_FLOWER_KEY_UDP_SRC
,
583 &mask
->tp
.src
, TCA_FLOWER_KEY_UDP_SRC_MASK
,
584 sizeof(key
->tp
.src
));
585 fl_set_key_val(tb
, &key
->tp
.dst
, TCA_FLOWER_KEY_UDP_DST
,
586 &mask
->tp
.dst
, TCA_FLOWER_KEY_UDP_DST_MASK
,
587 sizeof(key
->tp
.dst
));
588 } else if (key
->basic
.ip_proto
== IPPROTO_SCTP
) {
589 fl_set_key_val(tb
, &key
->tp
.src
, TCA_FLOWER_KEY_SCTP_SRC
,
590 &mask
->tp
.src
, TCA_FLOWER_KEY_SCTP_SRC_MASK
,
591 sizeof(key
->tp
.src
));
592 fl_set_key_val(tb
, &key
->tp
.dst
, TCA_FLOWER_KEY_SCTP_DST
,
593 &mask
->tp
.dst
, TCA_FLOWER_KEY_SCTP_DST_MASK
,
594 sizeof(key
->tp
.dst
));
595 } else if (key
->basic
.n_proto
== htons(ETH_P_IP
) &&
596 key
->basic
.ip_proto
== IPPROTO_ICMP
) {
597 fl_set_key_val(tb
, &key
->icmp
.type
, TCA_FLOWER_KEY_ICMPV4_TYPE
,
599 TCA_FLOWER_KEY_ICMPV4_TYPE_MASK
,
600 sizeof(key
->icmp
.type
));
601 fl_set_key_val(tb
, &key
->icmp
.code
, TCA_FLOWER_KEY_ICMPV4_CODE
,
603 TCA_FLOWER_KEY_ICMPV4_CODE_MASK
,
604 sizeof(key
->icmp
.code
));
605 } else if (key
->basic
.n_proto
== htons(ETH_P_IPV6
) &&
606 key
->basic
.ip_proto
== IPPROTO_ICMPV6
) {
607 fl_set_key_val(tb
, &key
->icmp
.type
, TCA_FLOWER_KEY_ICMPV6_TYPE
,
609 TCA_FLOWER_KEY_ICMPV6_TYPE_MASK
,
610 sizeof(key
->icmp
.type
));
611 fl_set_key_val(tb
, &key
->icmp
.code
, TCA_FLOWER_KEY_ICMPV6_CODE
,
613 TCA_FLOWER_KEY_ICMPV6_CODE_MASK
,
614 sizeof(key
->icmp
.code
));
615 } else if (key
->basic
.n_proto
== htons(ETH_P_MPLS_UC
) ||
616 key
->basic
.n_proto
== htons(ETH_P_MPLS_MC
)) {
617 ret
= fl_set_key_mpls(tb
, &key
->mpls
, &mask
->mpls
);
620 } else if (key
->basic
.n_proto
== htons(ETH_P_ARP
) ||
621 key
->basic
.n_proto
== htons(ETH_P_RARP
)) {
622 fl_set_key_val(tb
, &key
->arp
.sip
, TCA_FLOWER_KEY_ARP_SIP
,
623 &mask
->arp
.sip
, TCA_FLOWER_KEY_ARP_SIP_MASK
,
624 sizeof(key
->arp
.sip
));
625 fl_set_key_val(tb
, &key
->arp
.tip
, TCA_FLOWER_KEY_ARP_TIP
,
626 &mask
->arp
.tip
, TCA_FLOWER_KEY_ARP_TIP_MASK
,
627 sizeof(key
->arp
.tip
));
628 fl_set_key_val(tb
, &key
->arp
.op
, TCA_FLOWER_KEY_ARP_OP
,
629 &mask
->arp
.op
, TCA_FLOWER_KEY_ARP_OP_MASK
,
630 sizeof(key
->arp
.op
));
631 fl_set_key_val(tb
, key
->arp
.sha
, TCA_FLOWER_KEY_ARP_SHA
,
632 mask
->arp
.sha
, TCA_FLOWER_KEY_ARP_SHA_MASK
,
633 sizeof(key
->arp
.sha
));
634 fl_set_key_val(tb
, key
->arp
.tha
, TCA_FLOWER_KEY_ARP_THA
,
635 mask
->arp
.tha
, TCA_FLOWER_KEY_ARP_THA_MASK
,
636 sizeof(key
->arp
.tha
));
639 if (tb
[TCA_FLOWER_KEY_ENC_IPV4_SRC
] ||
640 tb
[TCA_FLOWER_KEY_ENC_IPV4_DST
]) {
641 key
->enc_control
.addr_type
= FLOW_DISSECTOR_KEY_IPV4_ADDRS
;
642 mask
->enc_control
.addr_type
= ~0;
643 fl_set_key_val(tb
, &key
->enc_ipv4
.src
,
644 TCA_FLOWER_KEY_ENC_IPV4_SRC
,
646 TCA_FLOWER_KEY_ENC_IPV4_SRC_MASK
,
647 sizeof(key
->enc_ipv4
.src
));
648 fl_set_key_val(tb
, &key
->enc_ipv4
.dst
,
649 TCA_FLOWER_KEY_ENC_IPV4_DST
,
651 TCA_FLOWER_KEY_ENC_IPV4_DST_MASK
,
652 sizeof(key
->enc_ipv4
.dst
));
655 if (tb
[TCA_FLOWER_KEY_ENC_IPV6_SRC
] ||
656 tb
[TCA_FLOWER_KEY_ENC_IPV6_DST
]) {
657 key
->enc_control
.addr_type
= FLOW_DISSECTOR_KEY_IPV6_ADDRS
;
658 mask
->enc_control
.addr_type
= ~0;
659 fl_set_key_val(tb
, &key
->enc_ipv6
.src
,
660 TCA_FLOWER_KEY_ENC_IPV6_SRC
,
662 TCA_FLOWER_KEY_ENC_IPV6_SRC_MASK
,
663 sizeof(key
->enc_ipv6
.src
));
664 fl_set_key_val(tb
, &key
->enc_ipv6
.dst
,
665 TCA_FLOWER_KEY_ENC_IPV6_DST
,
667 TCA_FLOWER_KEY_ENC_IPV6_DST_MASK
,
668 sizeof(key
->enc_ipv6
.dst
));
671 fl_set_key_val(tb
, &key
->enc_key_id
.keyid
, TCA_FLOWER_KEY_ENC_KEY_ID
,
672 &mask
->enc_key_id
.keyid
, TCA_FLOWER_UNSPEC
,
673 sizeof(key
->enc_key_id
.keyid
));
675 fl_set_key_val(tb
, &key
->enc_tp
.src
, TCA_FLOWER_KEY_ENC_UDP_SRC_PORT
,
676 &mask
->enc_tp
.src
, TCA_FLOWER_KEY_ENC_UDP_SRC_PORT_MASK
,
677 sizeof(key
->enc_tp
.src
));
679 fl_set_key_val(tb
, &key
->enc_tp
.dst
, TCA_FLOWER_KEY_ENC_UDP_DST_PORT
,
680 &mask
->enc_tp
.dst
, TCA_FLOWER_KEY_ENC_UDP_DST_PORT_MASK
,
681 sizeof(key
->enc_tp
.dst
));
683 if (tb
[TCA_FLOWER_KEY_FLAGS
])
684 ret
= fl_set_key_flags(tb
, &key
->control
.flags
, &mask
->control
.flags
);
689 static bool fl_mask_eq(struct fl_flow_mask
*mask1
,
690 struct fl_flow_mask
*mask2
)
692 const long *lmask1
= fl_key_get_start(&mask1
->key
, mask1
);
693 const long *lmask2
= fl_key_get_start(&mask2
->key
, mask2
);
695 return !memcmp(&mask1
->range
, &mask2
->range
, sizeof(mask1
->range
)) &&
696 !memcmp(lmask1
, lmask2
, fl_mask_range(mask1
));
699 static const struct rhashtable_params fl_ht_params
= {
700 .key_offset
= offsetof(struct cls_fl_filter
, mkey
), /* base offset */
701 .head_offset
= offsetof(struct cls_fl_filter
, ht_node
),
702 .automatic_shrinking
= true,
705 static int fl_init_hashtable(struct cls_fl_head
*head
,
706 struct fl_flow_mask
*mask
)
708 head
->ht_params
= fl_ht_params
;
709 head
->ht_params
.key_len
= fl_mask_range(mask
);
710 head
->ht_params
.key_offset
+= mask
->range
.start
;
712 return rhashtable_init(&head
->ht
, &head
->ht_params
);
715 #define FL_KEY_MEMBER_OFFSET(member) offsetof(struct fl_flow_key, member)
716 #define FL_KEY_MEMBER_SIZE(member) (sizeof(((struct fl_flow_key *) 0)->member))
718 #define FL_KEY_IS_MASKED(mask, member) \
719 memchr_inv(((char *)mask) + FL_KEY_MEMBER_OFFSET(member), \
720 0, FL_KEY_MEMBER_SIZE(member)) \
722 #define FL_KEY_SET(keys, cnt, id, member) \
724 keys[cnt].key_id = id; \
725 keys[cnt].offset = FL_KEY_MEMBER_OFFSET(member); \
729 #define FL_KEY_SET_IF_MASKED(mask, keys, cnt, id, member) \
731 if (FL_KEY_IS_MASKED(mask, member)) \
732 FL_KEY_SET(keys, cnt, id, member); \
735 static void fl_init_dissector(struct cls_fl_head
*head
,
736 struct fl_flow_mask
*mask
)
738 struct flow_dissector_key keys
[FLOW_DISSECTOR_KEY_MAX
];
741 FL_KEY_SET(keys
, cnt
, FLOW_DISSECTOR_KEY_CONTROL
, control
);
742 FL_KEY_SET(keys
, cnt
, FLOW_DISSECTOR_KEY_BASIC
, basic
);
743 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
744 FLOW_DISSECTOR_KEY_ETH_ADDRS
, eth
);
745 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
746 FLOW_DISSECTOR_KEY_IPV4_ADDRS
, ipv4
);
747 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
748 FLOW_DISSECTOR_KEY_IPV6_ADDRS
, ipv6
);
749 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
750 FLOW_DISSECTOR_KEY_PORTS
, tp
);
751 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
752 FLOW_DISSECTOR_KEY_IP
, ip
);
753 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
754 FLOW_DISSECTOR_KEY_TCP
, tcp
);
755 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
756 FLOW_DISSECTOR_KEY_ICMP
, icmp
);
757 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
758 FLOW_DISSECTOR_KEY_ARP
, arp
);
759 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
760 FLOW_DISSECTOR_KEY_MPLS
, mpls
);
761 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
762 FLOW_DISSECTOR_KEY_VLAN
, vlan
);
763 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
764 FLOW_DISSECTOR_KEY_ENC_KEYID
, enc_key_id
);
765 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
766 FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS
, enc_ipv4
);
767 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
768 FLOW_DISSECTOR_KEY_ENC_IPV6_ADDRS
, enc_ipv6
);
769 if (FL_KEY_IS_MASKED(&mask
->key
, enc_ipv4
) ||
770 FL_KEY_IS_MASKED(&mask
->key
, enc_ipv6
))
771 FL_KEY_SET(keys
, cnt
, FLOW_DISSECTOR_KEY_ENC_CONTROL
,
773 FL_KEY_SET_IF_MASKED(&mask
->key
, keys
, cnt
,
774 FLOW_DISSECTOR_KEY_ENC_PORTS
, enc_tp
);
776 skb_flow_dissector_init(&head
->dissector
, keys
, cnt
);
779 static int fl_check_assign_mask(struct cls_fl_head
*head
,
780 struct fl_flow_mask
*mask
)
784 if (head
->mask_assigned
) {
785 if (!fl_mask_eq(&head
->mask
, mask
))
791 /* Mask is not assigned yet. So assign it and init hashtable
794 err
= fl_init_hashtable(head
, mask
);
797 memcpy(&head
->mask
, mask
, sizeof(head
->mask
));
798 head
->mask_assigned
= true;
800 fl_init_dissector(head
, mask
);
805 static int fl_set_parms(struct net
*net
, struct tcf_proto
*tp
,
806 struct cls_fl_filter
*f
, struct fl_flow_mask
*mask
,
807 unsigned long base
, struct nlattr
**tb
,
808 struct nlattr
*est
, bool ovr
)
812 err
= tcf_exts_validate(net
, tp
, tb
, est
, &f
->exts
, ovr
);
816 if (tb
[TCA_FLOWER_CLASSID
]) {
817 f
->res
.classid
= nla_get_u32(tb
[TCA_FLOWER_CLASSID
]);
818 tcf_bind_filter(tp
, &f
->res
, base
);
821 err
= fl_set_key(net
, tb
, &f
->key
, &mask
->key
);
825 fl_mask_update_range(mask
);
826 fl_set_masked_key(&f
->mkey
, &f
->key
, mask
);
831 static int fl_change(struct net
*net
, struct sk_buff
*in_skb
,
832 struct tcf_proto
*tp
, unsigned long base
,
833 u32 handle
, struct nlattr
**tca
,
834 void **arg
, bool ovr
)
836 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
837 struct cls_fl_filter
*fold
= *arg
;
838 struct cls_fl_filter
*fnew
;
840 struct fl_flow_mask mask
= {};
841 unsigned long idr_index
;
844 if (!tca
[TCA_OPTIONS
])
847 tb
= kcalloc(TCA_FLOWER_MAX
+ 1, sizeof(struct nlattr
*), GFP_KERNEL
);
851 err
= nla_parse_nested(tb
, TCA_FLOWER_MAX
, tca
[TCA_OPTIONS
],
856 if (fold
&& handle
&& fold
->handle
!= handle
) {
861 fnew
= kzalloc(sizeof(*fnew
), GFP_KERNEL
);
867 err
= tcf_exts_init(&fnew
->exts
, TCA_FLOWER_ACT
, 0);
872 err
= idr_alloc_ext(&head
->handle_idr
, fnew
, &idr_index
,
873 1, 0x80000000, GFP_KERNEL
);
876 fnew
->handle
= idr_index
;
879 /* user specifies a handle and it doesn't exist */
880 if (handle
&& !fold
) {
881 err
= idr_alloc_ext(&head
->handle_idr
, fnew
, &idr_index
,
882 handle
, handle
+ 1, GFP_KERNEL
);
885 fnew
->handle
= idr_index
;
888 if (tb
[TCA_FLOWER_FLAGS
]) {
889 fnew
->flags
= nla_get_u32(tb
[TCA_FLOWER_FLAGS
]);
891 if (!tc_flags_valid(fnew
->flags
)) {
897 err
= fl_set_parms(net
, tp
, fnew
, &mask
, base
, tb
, tca
[TCA_RATE
], ovr
);
901 err
= fl_check_assign_mask(head
, &mask
);
905 if (!tc_skip_sw(fnew
->flags
)) {
906 if (!fold
&& fl_lookup(head
, &fnew
->mkey
)) {
911 err
= rhashtable_insert_fast(&head
->ht
, &fnew
->ht_node
,
917 if (!tc_skip_hw(fnew
->flags
)) {
918 err
= fl_hw_replace_filter(tp
,
926 if (!tc_in_hw(fnew
->flags
))
927 fnew
->flags
|= TCA_CLS_FLAGS_NOT_IN_HW
;
930 if (!tc_skip_sw(fold
->flags
))
931 rhashtable_remove_fast(&head
->ht
, &fold
->ht_node
,
933 if (!tc_skip_hw(fold
->flags
))
934 fl_hw_destroy_filter(tp
, fold
);
940 fnew
->handle
= handle
;
941 idr_replace_ext(&head
->handle_idr
, fnew
, fnew
->handle
);
942 list_replace_rcu(&fold
->list
, &fnew
->list
);
943 tcf_unbind_filter(tp
, &fold
->res
);
944 tcf_exts_get_net(&fold
->exts
);
945 tcf_queue_work(&fold
->rwork
, fl_destroy_filter_work
);
947 list_add_tail_rcu(&fnew
->list
, &head
->filters
);
955 idr_remove_ext(&head
->handle_idr
, fnew
->handle
);
957 tcf_exts_destroy(&fnew
->exts
);
964 static int fl_delete(struct tcf_proto
*tp
, void *arg
, bool *last
)
966 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
967 struct cls_fl_filter
*f
= arg
;
969 if (!tc_skip_sw(f
->flags
))
970 rhashtable_remove_fast(&head
->ht
, &f
->ht_node
,
973 *last
= list_empty(&head
->filters
);
977 static void fl_walk(struct tcf_proto
*tp
, struct tcf_walker
*arg
)
979 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
980 struct cls_fl_filter
*f
;
982 list_for_each_entry_rcu(f
, &head
->filters
, list
) {
983 if (arg
->count
< arg
->skip
)
985 if (arg
->fn(tp
, f
, arg
) < 0) {
994 static int fl_dump_key_val(struct sk_buff
*skb
,
995 void *val
, int val_type
,
996 void *mask
, int mask_type
, int len
)
1000 if (!memchr_inv(mask
, 0, len
))
1002 err
= nla_put(skb
, val_type
, len
, val
);
1005 if (mask_type
!= TCA_FLOWER_UNSPEC
) {
1006 err
= nla_put(skb
, mask_type
, len
, mask
);
1013 static int fl_dump_key_mpls(struct sk_buff
*skb
,
1014 struct flow_dissector_key_mpls
*mpls_key
,
1015 struct flow_dissector_key_mpls
*mpls_mask
)
1019 if (!memchr_inv(mpls_mask
, 0, sizeof(*mpls_mask
)))
1021 if (mpls_mask
->mpls_ttl
) {
1022 err
= nla_put_u8(skb
, TCA_FLOWER_KEY_MPLS_TTL
,
1023 mpls_key
->mpls_ttl
);
1027 if (mpls_mask
->mpls_tc
) {
1028 err
= nla_put_u8(skb
, TCA_FLOWER_KEY_MPLS_TC
,
1033 if (mpls_mask
->mpls_label
) {
1034 err
= nla_put_u32(skb
, TCA_FLOWER_KEY_MPLS_LABEL
,
1035 mpls_key
->mpls_label
);
1039 if (mpls_mask
->mpls_bos
) {
1040 err
= nla_put_u8(skb
, TCA_FLOWER_KEY_MPLS_BOS
,
1041 mpls_key
->mpls_bos
);
1048 static int fl_dump_key_ip(struct sk_buff
*skb
,
1049 struct flow_dissector_key_ip
*key
,
1050 struct flow_dissector_key_ip
*mask
)
1052 if (fl_dump_key_val(skb
, &key
->tos
, TCA_FLOWER_KEY_IP_TOS
, &mask
->tos
,
1053 TCA_FLOWER_KEY_IP_TOS_MASK
, sizeof(key
->tos
)) ||
1054 fl_dump_key_val(skb
, &key
->ttl
, TCA_FLOWER_KEY_IP_TTL
, &mask
->ttl
,
1055 TCA_FLOWER_KEY_IP_TTL_MASK
, sizeof(key
->ttl
)))
1061 static int fl_dump_key_vlan(struct sk_buff
*skb
,
1062 struct flow_dissector_key_vlan
*vlan_key
,
1063 struct flow_dissector_key_vlan
*vlan_mask
)
1067 if (!memchr_inv(vlan_mask
, 0, sizeof(*vlan_mask
)))
1069 if (vlan_mask
->vlan_id
) {
1070 err
= nla_put_u16(skb
, TCA_FLOWER_KEY_VLAN_ID
,
1075 if (vlan_mask
->vlan_priority
) {
1076 err
= nla_put_u8(skb
, TCA_FLOWER_KEY_VLAN_PRIO
,
1077 vlan_key
->vlan_priority
);
1084 static void fl_get_key_flag(u32 dissector_key
, u32 dissector_mask
,
1085 u32
*flower_key
, u32
*flower_mask
,
1086 u32 flower_flag_bit
, u32 dissector_flag_bit
)
1088 if (dissector_mask
& dissector_flag_bit
) {
1089 *flower_mask
|= flower_flag_bit
;
1090 if (dissector_key
& dissector_flag_bit
)
1091 *flower_key
|= flower_flag_bit
;
1095 static int fl_dump_key_flags(struct sk_buff
*skb
, u32 flags_key
, u32 flags_mask
)
1101 if (!memchr_inv(&flags_mask
, 0, sizeof(flags_mask
)))
1107 fl_get_key_flag(flags_key
, flags_mask
, &key
, &mask
,
1108 TCA_FLOWER_KEY_FLAGS_IS_FRAGMENT
, FLOW_DIS_IS_FRAGMENT
);
1110 _key
= cpu_to_be32(key
);
1111 _mask
= cpu_to_be32(mask
);
1113 err
= nla_put(skb
, TCA_FLOWER_KEY_FLAGS
, 4, &_key
);
1117 return nla_put(skb
, TCA_FLOWER_KEY_FLAGS_MASK
, 4, &_mask
);
1120 static int fl_dump(struct net
*net
, struct tcf_proto
*tp
, void *fh
,
1121 struct sk_buff
*skb
, struct tcmsg
*t
)
1123 struct cls_fl_head
*head
= rtnl_dereference(tp
->root
);
1124 struct cls_fl_filter
*f
= fh
;
1125 struct nlattr
*nest
;
1126 struct fl_flow_key
*key
, *mask
;
1131 t
->tcm_handle
= f
->handle
;
1133 nest
= nla_nest_start(skb
, TCA_OPTIONS
);
1135 goto nla_put_failure
;
1137 if (f
->res
.classid
&&
1138 nla_put_u32(skb
, TCA_FLOWER_CLASSID
, f
->res
.classid
))
1139 goto nla_put_failure
;
1142 mask
= &head
->mask
.key
;
1144 if (mask
->indev_ifindex
) {
1145 struct net_device
*dev
;
1147 dev
= __dev_get_by_index(net
, key
->indev_ifindex
);
1148 if (dev
&& nla_put_string(skb
, TCA_FLOWER_INDEV
, dev
->name
))
1149 goto nla_put_failure
;
1152 if (!tc_skip_hw(f
->flags
))
1153 fl_hw_update_stats(tp
, f
);
1155 if (fl_dump_key_val(skb
, key
->eth
.dst
, TCA_FLOWER_KEY_ETH_DST
,
1156 mask
->eth
.dst
, TCA_FLOWER_KEY_ETH_DST_MASK
,
1157 sizeof(key
->eth
.dst
)) ||
1158 fl_dump_key_val(skb
, key
->eth
.src
, TCA_FLOWER_KEY_ETH_SRC
,
1159 mask
->eth
.src
, TCA_FLOWER_KEY_ETH_SRC_MASK
,
1160 sizeof(key
->eth
.src
)) ||
1161 fl_dump_key_val(skb
, &key
->basic
.n_proto
, TCA_FLOWER_KEY_ETH_TYPE
,
1162 &mask
->basic
.n_proto
, TCA_FLOWER_UNSPEC
,
1163 sizeof(key
->basic
.n_proto
)))
1164 goto nla_put_failure
;
1166 if (fl_dump_key_mpls(skb
, &key
->mpls
, &mask
->mpls
))
1167 goto nla_put_failure
;
1169 if (fl_dump_key_vlan(skb
, &key
->vlan
, &mask
->vlan
))
1170 goto nla_put_failure
;
1172 if ((key
->basic
.n_proto
== htons(ETH_P_IP
) ||
1173 key
->basic
.n_proto
== htons(ETH_P_IPV6
)) &&
1174 (fl_dump_key_val(skb
, &key
->basic
.ip_proto
, TCA_FLOWER_KEY_IP_PROTO
,
1175 &mask
->basic
.ip_proto
, TCA_FLOWER_UNSPEC
,
1176 sizeof(key
->basic
.ip_proto
)) ||
1177 fl_dump_key_ip(skb
, &key
->ip
, &mask
->ip
)))
1178 goto nla_put_failure
;
1180 if (key
->control
.addr_type
== FLOW_DISSECTOR_KEY_IPV4_ADDRS
&&
1181 (fl_dump_key_val(skb
, &key
->ipv4
.src
, TCA_FLOWER_KEY_IPV4_SRC
,
1182 &mask
->ipv4
.src
, TCA_FLOWER_KEY_IPV4_SRC_MASK
,
1183 sizeof(key
->ipv4
.src
)) ||
1184 fl_dump_key_val(skb
, &key
->ipv4
.dst
, TCA_FLOWER_KEY_IPV4_DST
,
1185 &mask
->ipv4
.dst
, TCA_FLOWER_KEY_IPV4_DST_MASK
,
1186 sizeof(key
->ipv4
.dst
))))
1187 goto nla_put_failure
;
1188 else if (key
->control
.addr_type
== FLOW_DISSECTOR_KEY_IPV6_ADDRS
&&
1189 (fl_dump_key_val(skb
, &key
->ipv6
.src
, TCA_FLOWER_KEY_IPV6_SRC
,
1190 &mask
->ipv6
.src
, TCA_FLOWER_KEY_IPV6_SRC_MASK
,
1191 sizeof(key
->ipv6
.src
)) ||
1192 fl_dump_key_val(skb
, &key
->ipv6
.dst
, TCA_FLOWER_KEY_IPV6_DST
,
1193 &mask
->ipv6
.dst
, TCA_FLOWER_KEY_IPV6_DST_MASK
,
1194 sizeof(key
->ipv6
.dst
))))
1195 goto nla_put_failure
;
1197 if (key
->basic
.ip_proto
== IPPROTO_TCP
&&
1198 (fl_dump_key_val(skb
, &key
->tp
.src
, TCA_FLOWER_KEY_TCP_SRC
,
1199 &mask
->tp
.src
, TCA_FLOWER_KEY_TCP_SRC_MASK
,
1200 sizeof(key
->tp
.src
)) ||
1201 fl_dump_key_val(skb
, &key
->tp
.dst
, TCA_FLOWER_KEY_TCP_DST
,
1202 &mask
->tp
.dst
, TCA_FLOWER_KEY_TCP_DST_MASK
,
1203 sizeof(key
->tp
.dst
)) ||
1204 fl_dump_key_val(skb
, &key
->tcp
.flags
, TCA_FLOWER_KEY_TCP_FLAGS
,
1205 &mask
->tcp
.flags
, TCA_FLOWER_KEY_TCP_FLAGS_MASK
,
1206 sizeof(key
->tcp
.flags
))))
1207 goto nla_put_failure
;
1208 else if (key
->basic
.ip_proto
== IPPROTO_UDP
&&
1209 (fl_dump_key_val(skb
, &key
->tp
.src
, TCA_FLOWER_KEY_UDP_SRC
,
1210 &mask
->tp
.src
, TCA_FLOWER_KEY_UDP_SRC_MASK
,
1211 sizeof(key
->tp
.src
)) ||
1212 fl_dump_key_val(skb
, &key
->tp
.dst
, TCA_FLOWER_KEY_UDP_DST
,
1213 &mask
->tp
.dst
, TCA_FLOWER_KEY_UDP_DST_MASK
,
1214 sizeof(key
->tp
.dst
))))
1215 goto nla_put_failure
;
1216 else if (key
->basic
.ip_proto
== IPPROTO_SCTP
&&
1217 (fl_dump_key_val(skb
, &key
->tp
.src
, TCA_FLOWER_KEY_SCTP_SRC
,
1218 &mask
->tp
.src
, TCA_FLOWER_KEY_SCTP_SRC_MASK
,
1219 sizeof(key
->tp
.src
)) ||
1220 fl_dump_key_val(skb
, &key
->tp
.dst
, TCA_FLOWER_KEY_SCTP_DST
,
1221 &mask
->tp
.dst
, TCA_FLOWER_KEY_SCTP_DST_MASK
,
1222 sizeof(key
->tp
.dst
))))
1223 goto nla_put_failure
;
1224 else if (key
->basic
.n_proto
== htons(ETH_P_IP
) &&
1225 key
->basic
.ip_proto
== IPPROTO_ICMP
&&
1226 (fl_dump_key_val(skb
, &key
->icmp
.type
,
1227 TCA_FLOWER_KEY_ICMPV4_TYPE
, &mask
->icmp
.type
,
1228 TCA_FLOWER_KEY_ICMPV4_TYPE_MASK
,
1229 sizeof(key
->icmp
.type
)) ||
1230 fl_dump_key_val(skb
, &key
->icmp
.code
,
1231 TCA_FLOWER_KEY_ICMPV4_CODE
, &mask
->icmp
.code
,
1232 TCA_FLOWER_KEY_ICMPV4_CODE_MASK
,
1233 sizeof(key
->icmp
.code
))))
1234 goto nla_put_failure
;
1235 else if (key
->basic
.n_proto
== htons(ETH_P_IPV6
) &&
1236 key
->basic
.ip_proto
== IPPROTO_ICMPV6
&&
1237 (fl_dump_key_val(skb
, &key
->icmp
.type
,
1238 TCA_FLOWER_KEY_ICMPV6_TYPE
, &mask
->icmp
.type
,
1239 TCA_FLOWER_KEY_ICMPV6_TYPE_MASK
,
1240 sizeof(key
->icmp
.type
)) ||
1241 fl_dump_key_val(skb
, &key
->icmp
.code
,
1242 TCA_FLOWER_KEY_ICMPV6_CODE
, &mask
->icmp
.code
,
1243 TCA_FLOWER_KEY_ICMPV6_CODE_MASK
,
1244 sizeof(key
->icmp
.code
))))
1245 goto nla_put_failure
;
1246 else if ((key
->basic
.n_proto
== htons(ETH_P_ARP
) ||
1247 key
->basic
.n_proto
== htons(ETH_P_RARP
)) &&
1248 (fl_dump_key_val(skb
, &key
->arp
.sip
,
1249 TCA_FLOWER_KEY_ARP_SIP
, &mask
->arp
.sip
,
1250 TCA_FLOWER_KEY_ARP_SIP_MASK
,
1251 sizeof(key
->arp
.sip
)) ||
1252 fl_dump_key_val(skb
, &key
->arp
.tip
,
1253 TCA_FLOWER_KEY_ARP_TIP
, &mask
->arp
.tip
,
1254 TCA_FLOWER_KEY_ARP_TIP_MASK
,
1255 sizeof(key
->arp
.tip
)) ||
1256 fl_dump_key_val(skb
, &key
->arp
.op
,
1257 TCA_FLOWER_KEY_ARP_OP
, &mask
->arp
.op
,
1258 TCA_FLOWER_KEY_ARP_OP_MASK
,
1259 sizeof(key
->arp
.op
)) ||
1260 fl_dump_key_val(skb
, key
->arp
.sha
, TCA_FLOWER_KEY_ARP_SHA
,
1261 mask
->arp
.sha
, TCA_FLOWER_KEY_ARP_SHA_MASK
,
1262 sizeof(key
->arp
.sha
)) ||
1263 fl_dump_key_val(skb
, key
->arp
.tha
, TCA_FLOWER_KEY_ARP_THA
,
1264 mask
->arp
.tha
, TCA_FLOWER_KEY_ARP_THA_MASK
,
1265 sizeof(key
->arp
.tha
))))
1266 goto nla_put_failure
;
1268 if (key
->enc_control
.addr_type
== FLOW_DISSECTOR_KEY_IPV4_ADDRS
&&
1269 (fl_dump_key_val(skb
, &key
->enc_ipv4
.src
,
1270 TCA_FLOWER_KEY_ENC_IPV4_SRC
, &mask
->enc_ipv4
.src
,
1271 TCA_FLOWER_KEY_ENC_IPV4_SRC_MASK
,
1272 sizeof(key
->enc_ipv4
.src
)) ||
1273 fl_dump_key_val(skb
, &key
->enc_ipv4
.dst
,
1274 TCA_FLOWER_KEY_ENC_IPV4_DST
, &mask
->enc_ipv4
.dst
,
1275 TCA_FLOWER_KEY_ENC_IPV4_DST_MASK
,
1276 sizeof(key
->enc_ipv4
.dst
))))
1277 goto nla_put_failure
;
1278 else if (key
->enc_control
.addr_type
== FLOW_DISSECTOR_KEY_IPV6_ADDRS
&&
1279 (fl_dump_key_val(skb
, &key
->enc_ipv6
.src
,
1280 TCA_FLOWER_KEY_ENC_IPV6_SRC
, &mask
->enc_ipv6
.src
,
1281 TCA_FLOWER_KEY_ENC_IPV6_SRC_MASK
,
1282 sizeof(key
->enc_ipv6
.src
)) ||
1283 fl_dump_key_val(skb
, &key
->enc_ipv6
.dst
,
1284 TCA_FLOWER_KEY_ENC_IPV6_DST
,
1285 &mask
->enc_ipv6
.dst
,
1286 TCA_FLOWER_KEY_ENC_IPV6_DST_MASK
,
1287 sizeof(key
->enc_ipv6
.dst
))))
1288 goto nla_put_failure
;
1290 if (fl_dump_key_val(skb
, &key
->enc_key_id
, TCA_FLOWER_KEY_ENC_KEY_ID
,
1291 &mask
->enc_key_id
, TCA_FLOWER_UNSPEC
,
1292 sizeof(key
->enc_key_id
)) ||
1293 fl_dump_key_val(skb
, &key
->enc_tp
.src
,
1294 TCA_FLOWER_KEY_ENC_UDP_SRC_PORT
,
1296 TCA_FLOWER_KEY_ENC_UDP_SRC_PORT_MASK
,
1297 sizeof(key
->enc_tp
.src
)) ||
1298 fl_dump_key_val(skb
, &key
->enc_tp
.dst
,
1299 TCA_FLOWER_KEY_ENC_UDP_DST_PORT
,
1301 TCA_FLOWER_KEY_ENC_UDP_DST_PORT_MASK
,
1302 sizeof(key
->enc_tp
.dst
)))
1303 goto nla_put_failure
;
1305 if (fl_dump_key_flags(skb
, key
->control
.flags
, mask
->control
.flags
))
1306 goto nla_put_failure
;
1308 if (f
->flags
&& nla_put_u32(skb
, TCA_FLOWER_FLAGS
, f
->flags
))
1309 goto nla_put_failure
;
1311 if (tcf_exts_dump(skb
, &f
->exts
))
1312 goto nla_put_failure
;
1314 nla_nest_end(skb
, nest
);
1316 if (tcf_exts_dump_stats(skb
, &f
->exts
) < 0)
1317 goto nla_put_failure
;
1322 nla_nest_cancel(skb
, nest
);
1326 static void fl_bind_class(void *fh
, u32 classid
, unsigned long cl
)
1328 struct cls_fl_filter
*f
= fh
;
1330 if (f
&& f
->res
.classid
== classid
)
1334 static struct tcf_proto_ops cls_fl_ops __read_mostly
= {
1336 .classify
= fl_classify
,
1338 .destroy
= fl_destroy
,
1340 .change
= fl_change
,
1341 .delete = fl_delete
,
1344 .bind_class
= fl_bind_class
,
1345 .owner
= THIS_MODULE
,
1348 static int __init
cls_fl_init(void)
1350 return register_tcf_proto_ops(&cls_fl_ops
);
1353 static void __exit
cls_fl_exit(void)
1355 unregister_tcf_proto_ops(&cls_fl_ops
);
1358 module_init(cls_fl_init
);
1359 module_exit(cls_fl_exit
);
1361 MODULE_AUTHOR("Jiri Pirko <jiri@resnulli.us>");
1362 MODULE_DESCRIPTION("Flower classifier");
1363 MODULE_LICENSE("GPL v2");