]>
Commit | Line | Data |
---|---|---|
e8f887ac AV |
1 | /* |
2 | * Copyright (c) 2016, Mellanox Technologies. All rights reserved. | |
3 | * | |
4 | * This software is available to you under a choice of one of two | |
5 | * licenses. You may choose to be licensed under the terms of the GNU | |
6 | * General Public License (GPL) Version 2, available from the file | |
7 | * COPYING in the main directory of this source tree, or the | |
8 | * OpenIB.org BSD license below: | |
9 | * | |
10 | * Redistribution and use in source and binary forms, with or | |
11 | * without modification, are permitted provided that the following | |
12 | * conditions are met: | |
13 | * | |
14 | * - Redistributions of source code must retain the above | |
15 | * copyright notice, this list of conditions and the following | |
16 | * disclaimer. | |
17 | * | |
18 | * - Redistributions in binary form must reproduce the above | |
19 | * copyright notice, this list of conditions and the following | |
20 | * disclaimer in the documentation and/or other materials | |
21 | * provided with the distribution. | |
22 | * | |
23 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
24 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
25 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
26 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
27 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
28 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
29 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
30 | * SOFTWARE. | |
31 | */ | |
32 | ||
e3a2b7ed AV |
33 | #include <net/flow_dissector.h> |
34 | #include <net/pkt_cls.h> | |
35 | #include <net/tc_act/tc_gact.h> | |
12185a9f | 36 | #include <net/tc_act/tc_skbedit.h> |
e8f887ac AV |
37 | #include <linux/mlx5/fs.h> |
38 | #include <linux/mlx5/device.h> | |
39 | #include <linux/rhashtable.h> | |
03a9d11e OG |
40 | #include <net/switchdev.h> |
41 | #include <net/tc_act/tc_mirred.h> | |
776b12b6 | 42 | #include <net/tc_act/tc_vlan.h> |
bbd00f7e | 43 | #include <net/tc_act/tc_tunnel_key.h> |
a54e20b4 | 44 | #include <net/vxlan.h> |
e8f887ac AV |
45 | #include "en.h" |
46 | #include "en_tc.h" | |
03a9d11e | 47 | #include "eswitch.h" |
bbd00f7e | 48 | #include "vxlan.h" |
e8f887ac AV |
49 | |
50 | struct mlx5e_tc_flow { | |
51 | struct rhash_head node; | |
52 | u64 cookie; | |
74491de9 | 53 | struct mlx5_flow_handle *rule; |
a54e20b4 | 54 | struct list_head encap; /* flows sharing the same encap */ |
776b12b6 | 55 | struct mlx5_esw_flow_attr *attr; |
e8f887ac AV |
56 | }; |
57 | ||
a54e20b4 HHZ |
58 | enum { |
59 | MLX5_HEADER_TYPE_VXLAN = 0x0, | |
60 | MLX5_HEADER_TYPE_NVGRE = 0x1, | |
61 | }; | |
62 | ||
acff797c MG |
63 | #define MLX5E_TC_TABLE_NUM_ENTRIES 1024 |
64 | #define MLX5E_TC_TABLE_NUM_GROUPS 4 | |
e8f887ac | 65 | |
74491de9 MB |
66 | static struct mlx5_flow_handle * |
67 | mlx5e_tc_add_nic_flow(struct mlx5e_priv *priv, | |
68 | struct mlx5_flow_spec *spec, | |
69 | u32 action, u32 flow_tag) | |
e8f887ac | 70 | { |
aad7e08d AV |
71 | struct mlx5_core_dev *dev = priv->mdev; |
72 | struct mlx5_flow_destination dest = { 0 }; | |
66958ed9 HHZ |
73 | struct mlx5_flow_act flow_act = { |
74 | .action = action, | |
75 | .flow_tag = flow_tag, | |
76 | .encap_id = 0, | |
77 | }; | |
aad7e08d | 78 | struct mlx5_fc *counter = NULL; |
74491de9 | 79 | struct mlx5_flow_handle *rule; |
e8f887ac AV |
80 | bool table_created = false; |
81 | ||
aad7e08d AV |
82 | if (action & MLX5_FLOW_CONTEXT_ACTION_FWD_DEST) { |
83 | dest.type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE; | |
84 | dest.ft = priv->fs.vlan.ft.t; | |
55130287 | 85 | } else if (action & MLX5_FLOW_CONTEXT_ACTION_COUNT) { |
aad7e08d AV |
86 | counter = mlx5_fc_create(dev, true); |
87 | if (IS_ERR(counter)) | |
88 | return ERR_CAST(counter); | |
89 | ||
90 | dest.type = MLX5_FLOW_DESTINATION_TYPE_COUNTER; | |
91 | dest.counter = counter; | |
92 | } | |
93 | ||
acff797c MG |
94 | if (IS_ERR_OR_NULL(priv->fs.tc.t)) { |
95 | priv->fs.tc.t = | |
96 | mlx5_create_auto_grouped_flow_table(priv->fs.ns, | |
97 | MLX5E_TC_PRIO, | |
98 | MLX5E_TC_TABLE_NUM_ENTRIES, | |
99 | MLX5E_TC_TABLE_NUM_GROUPS, | |
c9f1b073 | 100 | 0, 0); |
acff797c | 101 | if (IS_ERR(priv->fs.tc.t)) { |
e8f887ac AV |
102 | netdev_err(priv->netdev, |
103 | "Failed to create tc offload table\n"); | |
aad7e08d AV |
104 | rule = ERR_CAST(priv->fs.tc.t); |
105 | goto err_create_ft; | |
e8f887ac AV |
106 | } |
107 | ||
108 | table_created = true; | |
109 | } | |
110 | ||
c5bb1730 | 111 | spec->match_criteria_enable = MLX5_MATCH_OUTER_HEADERS; |
66958ed9 | 112 | rule = mlx5_add_flow_rules(priv->fs.tc.t, spec, &flow_act, &dest, 1); |
aad7e08d AV |
113 | |
114 | if (IS_ERR(rule)) | |
115 | goto err_add_rule; | |
116 | ||
117 | return rule; | |
e8f887ac | 118 | |
aad7e08d AV |
119 | err_add_rule: |
120 | if (table_created) { | |
acff797c MG |
121 | mlx5_destroy_flow_table(priv->fs.tc.t); |
122 | priv->fs.tc.t = NULL; | |
e8f887ac | 123 | } |
aad7e08d AV |
124 | err_create_ft: |
125 | mlx5_fc_destroy(dev, counter); | |
e8f887ac AV |
126 | |
127 | return rule; | |
128 | } | |
129 | ||
74491de9 MB |
130 | static struct mlx5_flow_handle * |
131 | mlx5e_tc_add_fdb_flow(struct mlx5e_priv *priv, | |
132 | struct mlx5_flow_spec *spec, | |
133 | struct mlx5_esw_flow_attr *attr) | |
adb4c123 OG |
134 | { |
135 | struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; | |
8b32580d OG |
136 | int err; |
137 | ||
138 | err = mlx5_eswitch_add_vlan_action(esw, attr); | |
139 | if (err) | |
140 | return ERR_PTR(err); | |
adb4c123 | 141 | |
776b12b6 | 142 | return mlx5_eswitch_add_offloaded_rule(esw, spec, attr); |
adb4c123 OG |
143 | } |
144 | ||
e8f887ac | 145 | static void mlx5e_tc_del_flow(struct mlx5e_priv *priv, |
74491de9 | 146 | struct mlx5_flow_handle *rule, |
8b32580d | 147 | struct mlx5_esw_flow_attr *attr) |
e8f887ac | 148 | { |
8b32580d | 149 | struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; |
aad7e08d AV |
150 | struct mlx5_fc *counter = NULL; |
151 | ||
152 | counter = mlx5_flow_rule_counter(rule); | |
153 | ||
8b32580d OG |
154 | if (esw && esw->mode == SRIOV_OFFLOADS) |
155 | mlx5_eswitch_del_vlan_action(esw, attr); | |
156 | ||
74491de9 | 157 | mlx5_del_flow_rules(rule); |
e8f887ac | 158 | |
aad7e08d AV |
159 | mlx5_fc_destroy(priv->mdev, counter); |
160 | ||
5c40348c | 161 | if (!mlx5e_tc_num_filters(priv) && (priv->fs.tc.t)) { |
acff797c MG |
162 | mlx5_destroy_flow_table(priv->fs.tc.t); |
163 | priv->fs.tc.t = NULL; | |
e8f887ac AV |
164 | } |
165 | } | |
166 | ||
bbd00f7e HHZ |
167 | static void parse_vxlan_attr(struct mlx5_flow_spec *spec, |
168 | struct tc_cls_flower_offload *f) | |
169 | { | |
170 | void *headers_c = MLX5_ADDR_OF(fte_match_param, spec->match_criteria, | |
171 | outer_headers); | |
172 | void *headers_v = MLX5_ADDR_OF(fte_match_param, spec->match_value, | |
173 | outer_headers); | |
174 | void *misc_c = MLX5_ADDR_OF(fte_match_param, spec->match_criteria, | |
175 | misc_parameters); | |
176 | void *misc_v = MLX5_ADDR_OF(fte_match_param, spec->match_value, | |
177 | misc_parameters); | |
178 | ||
179 | MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, headers_c, ip_protocol); | |
180 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_protocol, IPPROTO_UDP); | |
181 | ||
182 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_KEYID)) { | |
183 | struct flow_dissector_key_keyid *key = | |
184 | skb_flow_dissector_target(f->dissector, | |
185 | FLOW_DISSECTOR_KEY_ENC_KEYID, | |
186 | f->key); | |
187 | struct flow_dissector_key_keyid *mask = | |
188 | skb_flow_dissector_target(f->dissector, | |
189 | FLOW_DISSECTOR_KEY_ENC_KEYID, | |
190 | f->mask); | |
191 | MLX5_SET(fte_match_set_misc, misc_c, vxlan_vni, | |
192 | be32_to_cpu(mask->keyid)); | |
193 | MLX5_SET(fte_match_set_misc, misc_v, vxlan_vni, | |
194 | be32_to_cpu(key->keyid)); | |
195 | } | |
196 | } | |
197 | ||
198 | static int parse_tunnel_attr(struct mlx5e_priv *priv, | |
199 | struct mlx5_flow_spec *spec, | |
200 | struct tc_cls_flower_offload *f) | |
201 | { | |
202 | void *headers_c = MLX5_ADDR_OF(fte_match_param, spec->match_criteria, | |
203 | outer_headers); | |
204 | void *headers_v = MLX5_ADDR_OF(fte_match_param, spec->match_value, | |
205 | outer_headers); | |
206 | ||
207 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_PORTS)) { | |
208 | struct flow_dissector_key_ports *key = | |
209 | skb_flow_dissector_target(f->dissector, | |
210 | FLOW_DISSECTOR_KEY_ENC_PORTS, | |
211 | f->key); | |
212 | struct flow_dissector_key_ports *mask = | |
213 | skb_flow_dissector_target(f->dissector, | |
214 | FLOW_DISSECTOR_KEY_ENC_PORTS, | |
215 | f->mask); | |
216 | ||
217 | /* Full udp dst port must be given */ | |
218 | if (memchr_inv(&mask->dst, 0xff, sizeof(mask->dst))) | |
219 | return -EOPNOTSUPP; | |
220 | ||
221 | /* udp src port isn't supported */ | |
222 | if (memchr_inv(&mask->src, 0, sizeof(mask->src))) | |
223 | return -EOPNOTSUPP; | |
224 | ||
225 | if (mlx5e_vxlan_lookup_port(priv, be16_to_cpu(key->dst)) && | |
226 | MLX5_CAP_ESW(priv->mdev, vxlan_encap_decap)) | |
227 | parse_vxlan_attr(spec, f); | |
228 | else | |
229 | return -EOPNOTSUPP; | |
230 | ||
231 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
232 | udp_dport, ntohs(mask->dst)); | |
233 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
234 | udp_dport, ntohs(key->dst)); | |
235 | ||
236 | } else { /* udp dst port must be given */ | |
237 | return -EOPNOTSUPP; | |
238 | } | |
239 | ||
240 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS)) { | |
241 | struct flow_dissector_key_ipv4_addrs *key = | |
242 | skb_flow_dissector_target(f->dissector, | |
243 | FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS, | |
244 | f->key); | |
245 | struct flow_dissector_key_ipv4_addrs *mask = | |
246 | skb_flow_dissector_target(f->dissector, | |
247 | FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS, | |
248 | f->mask); | |
249 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
250 | src_ipv4_src_ipv6.ipv4_layout.ipv4, | |
251 | ntohl(mask->src)); | |
252 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
253 | src_ipv4_src_ipv6.ipv4_layout.ipv4, | |
254 | ntohl(key->src)); | |
255 | ||
256 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
257 | dst_ipv4_dst_ipv6.ipv4_layout.ipv4, | |
258 | ntohl(mask->dst)); | |
259 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
260 | dst_ipv4_dst_ipv6.ipv4_layout.ipv4, | |
261 | ntohl(key->dst)); | |
262 | } | |
263 | ||
264 | MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, headers_c, ethertype); | |
265 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype, ETH_P_IP); | |
266 | ||
267 | /* Enforce DMAC when offloading incoming tunneled flows. | |
268 | * Flow counters require a match on the DMAC. | |
269 | */ | |
270 | MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, headers_c, dmac_47_16); | |
271 | MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, headers_c, dmac_15_0); | |
272 | ether_addr_copy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
273 | dmac_47_16), priv->netdev->dev_addr); | |
274 | ||
275 | /* let software handle IP fragments */ | |
276 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, frag, 1); | |
277 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, frag, 0); | |
278 | ||
279 | return 0; | |
280 | } | |
281 | ||
c5bb1730 | 282 | static int parse_cls_flower(struct mlx5e_priv *priv, struct mlx5_flow_spec *spec, |
e3a2b7ed AV |
283 | struct tc_cls_flower_offload *f) |
284 | { | |
c5bb1730 MG |
285 | void *headers_c = MLX5_ADDR_OF(fte_match_param, spec->match_criteria, |
286 | outer_headers); | |
287 | void *headers_v = MLX5_ADDR_OF(fte_match_param, spec->match_value, | |
288 | outer_headers); | |
e3a2b7ed AV |
289 | u16 addr_type = 0; |
290 | u8 ip_proto = 0; | |
291 | ||
292 | if (f->dissector->used_keys & | |
293 | ~(BIT(FLOW_DISSECTOR_KEY_CONTROL) | | |
294 | BIT(FLOW_DISSECTOR_KEY_BASIC) | | |
295 | BIT(FLOW_DISSECTOR_KEY_ETH_ADDRS) | | |
095b6cfd | 296 | BIT(FLOW_DISSECTOR_KEY_VLAN) | |
e3a2b7ed AV |
297 | BIT(FLOW_DISSECTOR_KEY_IPV4_ADDRS) | |
298 | BIT(FLOW_DISSECTOR_KEY_IPV6_ADDRS) | | |
bbd00f7e HHZ |
299 | BIT(FLOW_DISSECTOR_KEY_PORTS) | |
300 | BIT(FLOW_DISSECTOR_KEY_ENC_KEYID) | | |
301 | BIT(FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS) | | |
302 | BIT(FLOW_DISSECTOR_KEY_ENC_IPV6_ADDRS) | | |
303 | BIT(FLOW_DISSECTOR_KEY_ENC_PORTS) | | |
304 | BIT(FLOW_DISSECTOR_KEY_ENC_CONTROL))) { | |
e3a2b7ed AV |
305 | netdev_warn(priv->netdev, "Unsupported key used: 0x%x\n", |
306 | f->dissector->used_keys); | |
307 | return -EOPNOTSUPP; | |
308 | } | |
309 | ||
bbd00f7e HHZ |
310 | if ((dissector_uses_key(f->dissector, |
311 | FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS) || | |
312 | dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_KEYID) || | |
313 | dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_PORTS)) && | |
314 | dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ENC_CONTROL)) { | |
315 | struct flow_dissector_key_control *key = | |
316 | skb_flow_dissector_target(f->dissector, | |
317 | FLOW_DISSECTOR_KEY_ENC_CONTROL, | |
318 | f->key); | |
319 | switch (key->addr_type) { | |
320 | case FLOW_DISSECTOR_KEY_IPV4_ADDRS: | |
321 | if (parse_tunnel_attr(priv, spec, f)) | |
322 | return -EOPNOTSUPP; | |
323 | break; | |
324 | default: | |
325 | return -EOPNOTSUPP; | |
326 | } | |
327 | ||
328 | /* In decap flow, header pointers should point to the inner | |
329 | * headers, outer header were already set by parse_tunnel_attr | |
330 | */ | |
331 | headers_c = MLX5_ADDR_OF(fte_match_param, spec->match_criteria, | |
332 | inner_headers); | |
333 | headers_v = MLX5_ADDR_OF(fte_match_param, spec->match_value, | |
334 | inner_headers); | |
335 | } | |
336 | ||
e3a2b7ed AV |
337 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_CONTROL)) { |
338 | struct flow_dissector_key_control *key = | |
339 | skb_flow_dissector_target(f->dissector, | |
1dbd0d37 | 340 | FLOW_DISSECTOR_KEY_CONTROL, |
e3a2b7ed AV |
341 | f->key); |
342 | addr_type = key->addr_type; | |
343 | } | |
344 | ||
345 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_BASIC)) { | |
346 | struct flow_dissector_key_basic *key = | |
347 | skb_flow_dissector_target(f->dissector, | |
348 | FLOW_DISSECTOR_KEY_BASIC, | |
349 | f->key); | |
350 | struct flow_dissector_key_basic *mask = | |
351 | skb_flow_dissector_target(f->dissector, | |
352 | FLOW_DISSECTOR_KEY_BASIC, | |
353 | f->mask); | |
354 | ip_proto = key->ip_proto; | |
355 | ||
356 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, ethertype, | |
357 | ntohs(mask->n_proto)); | |
358 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype, | |
359 | ntohs(key->n_proto)); | |
360 | ||
361 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, ip_protocol, | |
362 | mask->ip_proto); | |
363 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_protocol, | |
364 | key->ip_proto); | |
365 | } | |
366 | ||
367 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_ETH_ADDRS)) { | |
368 | struct flow_dissector_key_eth_addrs *key = | |
369 | skb_flow_dissector_target(f->dissector, | |
370 | FLOW_DISSECTOR_KEY_ETH_ADDRS, | |
371 | f->key); | |
372 | struct flow_dissector_key_eth_addrs *mask = | |
373 | skb_flow_dissector_target(f->dissector, | |
374 | FLOW_DISSECTOR_KEY_ETH_ADDRS, | |
375 | f->mask); | |
376 | ||
377 | ether_addr_copy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
378 | dmac_47_16), | |
379 | mask->dst); | |
380 | ether_addr_copy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
381 | dmac_47_16), | |
382 | key->dst); | |
383 | ||
384 | ether_addr_copy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
385 | smac_47_16), | |
386 | mask->src); | |
387 | ether_addr_copy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
388 | smac_47_16), | |
389 | key->src); | |
390 | } | |
391 | ||
095b6cfd OG |
392 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_VLAN)) { |
393 | struct flow_dissector_key_vlan *key = | |
394 | skb_flow_dissector_target(f->dissector, | |
395 | FLOW_DISSECTOR_KEY_VLAN, | |
396 | f->key); | |
397 | struct flow_dissector_key_vlan *mask = | |
398 | skb_flow_dissector_target(f->dissector, | |
399 | FLOW_DISSECTOR_KEY_VLAN, | |
400 | f->mask); | |
358d79a4 | 401 | if (mask->vlan_id || mask->vlan_priority) { |
095b6cfd OG |
402 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, vlan_tag, 1); |
403 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, vlan_tag, 1); | |
404 | ||
405 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, first_vid, mask->vlan_id); | |
406 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_vid, key->vlan_id); | |
358d79a4 OG |
407 | |
408 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, first_prio, mask->vlan_priority); | |
409 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_prio, key->vlan_priority); | |
095b6cfd OG |
410 | } |
411 | } | |
412 | ||
e3a2b7ed AV |
413 | if (addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) { |
414 | struct flow_dissector_key_ipv4_addrs *key = | |
415 | skb_flow_dissector_target(f->dissector, | |
416 | FLOW_DISSECTOR_KEY_IPV4_ADDRS, | |
417 | f->key); | |
418 | struct flow_dissector_key_ipv4_addrs *mask = | |
419 | skb_flow_dissector_target(f->dissector, | |
420 | FLOW_DISSECTOR_KEY_IPV4_ADDRS, | |
421 | f->mask); | |
422 | ||
423 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
424 | src_ipv4_src_ipv6.ipv4_layout.ipv4), | |
425 | &mask->src, sizeof(mask->src)); | |
426 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
427 | src_ipv4_src_ipv6.ipv4_layout.ipv4), | |
428 | &key->src, sizeof(key->src)); | |
429 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
430 | dst_ipv4_dst_ipv6.ipv4_layout.ipv4), | |
431 | &mask->dst, sizeof(mask->dst)); | |
432 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
433 | dst_ipv4_dst_ipv6.ipv4_layout.ipv4), | |
434 | &key->dst, sizeof(key->dst)); | |
435 | } | |
436 | ||
437 | if (addr_type == FLOW_DISSECTOR_KEY_IPV6_ADDRS) { | |
438 | struct flow_dissector_key_ipv6_addrs *key = | |
439 | skb_flow_dissector_target(f->dissector, | |
440 | FLOW_DISSECTOR_KEY_IPV6_ADDRS, | |
441 | f->key); | |
442 | struct flow_dissector_key_ipv6_addrs *mask = | |
443 | skb_flow_dissector_target(f->dissector, | |
444 | FLOW_DISSECTOR_KEY_IPV6_ADDRS, | |
445 | f->mask); | |
446 | ||
447 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
448 | src_ipv4_src_ipv6.ipv6_layout.ipv6), | |
449 | &mask->src, sizeof(mask->src)); | |
450 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
451 | src_ipv4_src_ipv6.ipv6_layout.ipv6), | |
452 | &key->src, sizeof(key->src)); | |
453 | ||
454 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_c, | |
455 | dst_ipv4_dst_ipv6.ipv6_layout.ipv6), | |
456 | &mask->dst, sizeof(mask->dst)); | |
457 | memcpy(MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, | |
458 | dst_ipv4_dst_ipv6.ipv6_layout.ipv6), | |
459 | &key->dst, sizeof(key->dst)); | |
460 | } | |
461 | ||
462 | if (dissector_uses_key(f->dissector, FLOW_DISSECTOR_KEY_PORTS)) { | |
463 | struct flow_dissector_key_ports *key = | |
464 | skb_flow_dissector_target(f->dissector, | |
465 | FLOW_DISSECTOR_KEY_PORTS, | |
466 | f->key); | |
467 | struct flow_dissector_key_ports *mask = | |
468 | skb_flow_dissector_target(f->dissector, | |
469 | FLOW_DISSECTOR_KEY_PORTS, | |
470 | f->mask); | |
471 | switch (ip_proto) { | |
472 | case IPPROTO_TCP: | |
473 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
474 | tcp_sport, ntohs(mask->src)); | |
475 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
476 | tcp_sport, ntohs(key->src)); | |
477 | ||
478 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
479 | tcp_dport, ntohs(mask->dst)); | |
480 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
481 | tcp_dport, ntohs(key->dst)); | |
482 | break; | |
483 | ||
484 | case IPPROTO_UDP: | |
485 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
486 | udp_sport, ntohs(mask->src)); | |
487 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
488 | udp_sport, ntohs(key->src)); | |
489 | ||
490 | MLX5_SET(fte_match_set_lyr_2_4, headers_c, | |
491 | udp_dport, ntohs(mask->dst)); | |
492 | MLX5_SET(fte_match_set_lyr_2_4, headers_v, | |
493 | udp_dport, ntohs(key->dst)); | |
494 | break; | |
495 | default: | |
496 | netdev_err(priv->netdev, | |
497 | "Only UDP and TCP transport are supported\n"); | |
498 | return -EINVAL; | |
499 | } | |
500 | } | |
501 | ||
502 | return 0; | |
503 | } | |
504 | ||
5c40348c OG |
505 | static int parse_tc_nic_actions(struct mlx5e_priv *priv, struct tcf_exts *exts, |
506 | u32 *action, u32 *flow_tag) | |
e3a2b7ed AV |
507 | { |
508 | const struct tc_action *a; | |
22dc13c8 | 509 | LIST_HEAD(actions); |
e3a2b7ed AV |
510 | |
511 | if (tc_no_actions(exts)) | |
512 | return -EINVAL; | |
513 | ||
514 | *flow_tag = MLX5_FS_DEFAULT_FLOW_TAG; | |
515 | *action = 0; | |
516 | ||
22dc13c8 WC |
517 | tcf_exts_to_list(exts, &actions); |
518 | list_for_each_entry(a, &actions, list) { | |
e3a2b7ed AV |
519 | /* Only support a single action per rule */ |
520 | if (*action) | |
521 | return -EINVAL; | |
522 | ||
523 | if (is_tcf_gact_shot(a)) { | |
524 | *action |= MLX5_FLOW_CONTEXT_ACTION_DROP; | |
aad7e08d AV |
525 | if (MLX5_CAP_FLOWTABLE(priv->mdev, |
526 | flow_table_properties_nic_receive.flow_counter)) | |
527 | *action |= MLX5_FLOW_CONTEXT_ACTION_COUNT; | |
e3a2b7ed AV |
528 | continue; |
529 | } | |
530 | ||
531 | if (is_tcf_skbedit_mark(a)) { | |
532 | u32 mark = tcf_skbedit_mark(a); | |
533 | ||
534 | if (mark & ~MLX5E_TC_FLOW_ID_MASK) { | |
535 | netdev_warn(priv->netdev, "Bad flow mark - only 16 bit is supported: 0x%x\n", | |
536 | mark); | |
537 | return -EINVAL; | |
538 | } | |
539 | ||
540 | *flow_tag = mark; | |
541 | *action |= MLX5_FLOW_CONTEXT_ACTION_FWD_DEST; | |
542 | continue; | |
543 | } | |
544 | ||
545 | return -EINVAL; | |
546 | } | |
547 | ||
548 | return 0; | |
549 | } | |
550 | ||
a54e20b4 HHZ |
551 | static inline int cmp_encap_info(struct mlx5_encap_info *a, |
552 | struct mlx5_encap_info *b) | |
553 | { | |
554 | return memcmp(a, b, sizeof(*a)); | |
555 | } | |
556 | ||
557 | static inline int hash_encap_info(struct mlx5_encap_info *info) | |
558 | { | |
559 | return jhash(info, sizeof(*info), 0); | |
560 | } | |
561 | ||
562 | static int mlx5e_route_lookup_ipv4(struct mlx5e_priv *priv, | |
563 | struct net_device *mirred_dev, | |
564 | struct net_device **out_dev, | |
565 | struct flowi4 *fl4, | |
566 | struct neighbour **out_n, | |
567 | __be32 *saddr, | |
568 | int *out_ttl) | |
569 | { | |
570 | struct rtable *rt; | |
571 | struct neighbour *n = NULL; | |
572 | int ttl; | |
573 | ||
574 | #if IS_ENABLED(CONFIG_INET) | |
575 | rt = ip_route_output_key(dev_net(mirred_dev), fl4); | |
576 | if (IS_ERR(rt)) { | |
577 | pr_warn("%s: no route to %pI4\n", __func__, &fl4->daddr); | |
578 | return -EOPNOTSUPP; | |
579 | } | |
580 | #else | |
581 | return -EOPNOTSUPP; | |
582 | #endif | |
583 | ||
584 | if (!switchdev_port_same_parent_id(priv->netdev, rt->dst.dev)) { | |
585 | pr_warn("%s: Can't offload the flow, netdevices aren't on the same HW e-switch\n", | |
586 | __func__); | |
587 | ip_rt_put(rt); | |
588 | return -EOPNOTSUPP; | |
589 | } | |
590 | ||
591 | ttl = ip4_dst_hoplimit(&rt->dst); | |
592 | n = dst_neigh_lookup(&rt->dst, &fl4->daddr); | |
593 | ip_rt_put(rt); | |
594 | if (!n) | |
595 | return -ENOMEM; | |
596 | ||
597 | *out_n = n; | |
598 | *saddr = fl4->saddr; | |
599 | *out_ttl = ttl; | |
600 | *out_dev = rt->dst.dev; | |
601 | ||
602 | return 0; | |
603 | } | |
604 | ||
605 | static int gen_vxlan_header_ipv4(struct net_device *out_dev, | |
606 | char buf[], | |
607 | unsigned char h_dest[ETH_ALEN], | |
608 | int ttl, | |
609 | __be32 daddr, | |
610 | __be32 saddr, | |
611 | __be16 udp_dst_port, | |
612 | __be32 vx_vni) | |
613 | { | |
614 | int encap_size = VXLAN_HLEN + sizeof(struct iphdr) + ETH_HLEN; | |
615 | struct ethhdr *eth = (struct ethhdr *)buf; | |
616 | struct iphdr *ip = (struct iphdr *)((char *)eth + sizeof(struct ethhdr)); | |
617 | struct udphdr *udp = (struct udphdr *)((char *)ip + sizeof(struct iphdr)); | |
618 | struct vxlanhdr *vxh = (struct vxlanhdr *)((char *)udp + sizeof(struct udphdr)); | |
619 | ||
620 | memset(buf, 0, encap_size); | |
621 | ||
622 | ether_addr_copy(eth->h_dest, h_dest); | |
623 | ether_addr_copy(eth->h_source, out_dev->dev_addr); | |
624 | eth->h_proto = htons(ETH_P_IP); | |
625 | ||
626 | ip->daddr = daddr; | |
627 | ip->saddr = saddr; | |
628 | ||
629 | ip->ttl = ttl; | |
630 | ip->protocol = IPPROTO_UDP; | |
631 | ip->version = 0x4; | |
632 | ip->ihl = 0x5; | |
633 | ||
634 | udp->dest = udp_dst_port; | |
635 | vxh->vx_flags = VXLAN_HF_VNI; | |
636 | vxh->vx_vni = vxlan_vni_field(vx_vni); | |
637 | ||
638 | return encap_size; | |
639 | } | |
640 | ||
641 | static int mlx5e_create_encap_header_ipv4(struct mlx5e_priv *priv, | |
642 | struct net_device *mirred_dev, | |
643 | struct mlx5_encap_entry *e, | |
644 | struct net_device **out_dev) | |
645 | { | |
646 | int max_encap_size = MLX5_CAP_ESW(priv->mdev, max_encap_header_size); | |
647 | struct flowi4 fl4 = {}; | |
648 | struct neighbour *n; | |
649 | char *encap_header; | |
650 | int encap_size; | |
651 | __be32 saddr; | |
652 | int ttl; | |
653 | int err; | |
654 | ||
655 | encap_header = kzalloc(max_encap_size, GFP_KERNEL); | |
656 | if (!encap_header) | |
657 | return -ENOMEM; | |
658 | ||
659 | switch (e->tunnel_type) { | |
660 | case MLX5_HEADER_TYPE_VXLAN: | |
661 | fl4.flowi4_proto = IPPROTO_UDP; | |
662 | fl4.fl4_dport = e->tun_info.tp_dst; | |
663 | break; | |
664 | default: | |
665 | err = -EOPNOTSUPP; | |
666 | goto out; | |
667 | } | |
668 | fl4.daddr = e->tun_info.daddr; | |
669 | ||
670 | err = mlx5e_route_lookup_ipv4(priv, mirred_dev, out_dev, | |
671 | &fl4, &n, &saddr, &ttl); | |
672 | if (err) | |
673 | goto out; | |
674 | ||
675 | e->n = n; | |
676 | e->out_dev = *out_dev; | |
677 | ||
678 | if (!(n->nud_state & NUD_VALID)) { | |
679 | err = -ENOTSUPP; | |
680 | goto out; | |
681 | } | |
682 | ||
683 | neigh_ha_snapshot(e->h_dest, n, *out_dev); | |
684 | ||
685 | switch (e->tunnel_type) { | |
686 | case MLX5_HEADER_TYPE_VXLAN: | |
687 | encap_size = gen_vxlan_header_ipv4(*out_dev, encap_header, | |
688 | e->h_dest, ttl, | |
689 | e->tun_info.daddr, | |
690 | saddr, e->tun_info.tp_dst, | |
691 | e->tun_info.tun_id); | |
692 | break; | |
693 | default: | |
694 | err = -EOPNOTSUPP; | |
695 | goto out; | |
696 | } | |
697 | ||
698 | err = mlx5_encap_alloc(priv->mdev, e->tunnel_type, | |
699 | encap_size, encap_header, &e->encap_id); | |
700 | out: | |
701 | kfree(encap_header); | |
702 | return err; | |
703 | } | |
704 | ||
705 | static int mlx5e_attach_encap(struct mlx5e_priv *priv, | |
706 | struct ip_tunnel_info *tun_info, | |
707 | struct net_device *mirred_dev, | |
708 | struct mlx5_esw_flow_attr *attr) | |
709 | { | |
710 | struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; | |
711 | unsigned short family = ip_tunnel_info_af(tun_info); | |
712 | struct ip_tunnel_key *key = &tun_info->key; | |
713 | struct mlx5_encap_info info; | |
714 | struct mlx5_encap_entry *e; | |
715 | struct net_device *out_dev; | |
716 | uintptr_t hash_key; | |
717 | bool found = false; | |
718 | int tunnel_type; | |
719 | int err; | |
720 | ||
721 | /* udp dst port must be given */ | |
722 | if (!memchr_inv(&key->tp_dst, 0, sizeof(key->tp_dst))) | |
723 | return -EOPNOTSUPP; | |
724 | ||
725 | if (mlx5e_vxlan_lookup_port(priv, be16_to_cpu(key->tp_dst)) && | |
726 | MLX5_CAP_ESW(priv->mdev, vxlan_encap_decap)) { | |
727 | info.tp_dst = key->tp_dst; | |
728 | info.tun_id = tunnel_id_to_key32(key->tun_id); | |
729 | tunnel_type = MLX5_HEADER_TYPE_VXLAN; | |
730 | } else { | |
731 | return -EOPNOTSUPP; | |
732 | } | |
733 | ||
734 | switch (family) { | |
735 | case AF_INET: | |
736 | info.daddr = key->u.ipv4.dst; | |
737 | break; | |
738 | default: | |
739 | return -EOPNOTSUPP; | |
740 | } | |
741 | ||
742 | hash_key = hash_encap_info(&info); | |
743 | ||
744 | hash_for_each_possible_rcu(esw->offloads.encap_tbl, e, | |
745 | encap_hlist, hash_key) { | |
746 | if (!cmp_encap_info(&e->tun_info, &info)) { | |
747 | found = true; | |
748 | break; | |
749 | } | |
750 | } | |
751 | ||
752 | if (found) { | |
753 | attr->encap = e; | |
754 | return 0; | |
755 | } | |
756 | ||
757 | e = kzalloc(sizeof(*e), GFP_KERNEL); | |
758 | if (!e) | |
759 | return -ENOMEM; | |
760 | ||
761 | e->tun_info = info; | |
762 | e->tunnel_type = tunnel_type; | |
763 | INIT_LIST_HEAD(&e->flows); | |
764 | ||
765 | err = mlx5e_create_encap_header_ipv4(priv, mirred_dev, e, &out_dev); | |
766 | if (err) | |
767 | goto out_err; | |
768 | ||
769 | attr->encap = e; | |
770 | hash_add_rcu(esw->offloads.encap_tbl, &e->encap_hlist, hash_key); | |
771 | ||
772 | return err; | |
773 | ||
774 | out_err: | |
775 | kfree(e); | |
776 | return err; | |
777 | } | |
778 | ||
03a9d11e | 779 | static int parse_tc_fdb_actions(struct mlx5e_priv *priv, struct tcf_exts *exts, |
a54e20b4 | 780 | struct mlx5e_tc_flow *flow) |
03a9d11e | 781 | { |
a54e20b4 HHZ |
782 | struct mlx5_esw_flow_attr *attr = flow->attr; |
783 | struct ip_tunnel_info *info = NULL; | |
03a9d11e | 784 | const struct tc_action *a; |
22dc13c8 | 785 | LIST_HEAD(actions); |
a54e20b4 HHZ |
786 | bool encap = false; |
787 | int err; | |
03a9d11e OG |
788 | |
789 | if (tc_no_actions(exts)) | |
790 | return -EINVAL; | |
791 | ||
776b12b6 OG |
792 | memset(attr, 0, sizeof(*attr)); |
793 | attr->in_rep = priv->ppriv; | |
03a9d11e | 794 | |
22dc13c8 WC |
795 | tcf_exts_to_list(exts, &actions); |
796 | list_for_each_entry(a, &actions, list) { | |
03a9d11e | 797 | if (is_tcf_gact_shot(a)) { |
8b32580d OG |
798 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_DROP | |
799 | MLX5_FLOW_CONTEXT_ACTION_COUNT; | |
03a9d11e OG |
800 | continue; |
801 | } | |
802 | ||
5724b8b5 | 803 | if (is_tcf_mirred_egress_redirect(a)) { |
03a9d11e OG |
804 | int ifindex = tcf_mirred_ifindex(a); |
805 | struct net_device *out_dev; | |
806 | struct mlx5e_priv *out_priv; | |
03a9d11e OG |
807 | |
808 | out_dev = __dev_get_by_index(dev_net(priv->netdev), ifindex); | |
809 | ||
a54e20b4 HHZ |
810 | if (switchdev_port_same_parent_id(priv->netdev, |
811 | out_dev)) { | |
812 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_FWD_DEST | | |
813 | MLX5_FLOW_CONTEXT_ACTION_COUNT; | |
814 | out_priv = netdev_priv(out_dev); | |
815 | attr->out_rep = out_priv->ppriv; | |
816 | } else if (encap) { | |
817 | err = mlx5e_attach_encap(priv, info, | |
818 | out_dev, attr); | |
819 | if (err) | |
820 | return err; | |
821 | list_add(&flow->encap, &attr->encap->flows); | |
822 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_ENCAP | | |
823 | MLX5_FLOW_CONTEXT_ACTION_FWD_DEST | | |
824 | MLX5_FLOW_CONTEXT_ACTION_COUNT; | |
825 | out_priv = netdev_priv(attr->encap->out_dev); | |
826 | attr->out_rep = out_priv->ppriv; | |
827 | } else { | |
03a9d11e OG |
828 | pr_err("devices %s %s not on same switch HW, can't offload forwarding\n", |
829 | priv->netdev->name, out_dev->name); | |
830 | return -EINVAL; | |
831 | } | |
a54e20b4 HHZ |
832 | continue; |
833 | } | |
03a9d11e | 834 | |
a54e20b4 HHZ |
835 | if (is_tcf_tunnel_set(a)) { |
836 | info = tcf_tunnel_info(a); | |
837 | if (info) | |
838 | encap = true; | |
839 | else | |
840 | return -EOPNOTSUPP; | |
03a9d11e OG |
841 | continue; |
842 | } | |
843 | ||
8b32580d OG |
844 | if (is_tcf_vlan(a)) { |
845 | if (tcf_vlan_action(a) == VLAN_F_POP) { | |
846 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_POP; | |
847 | } else if (tcf_vlan_action(a) == VLAN_F_PUSH) { | |
848 | if (tcf_vlan_push_proto(a) != htons(ETH_P_8021Q)) | |
849 | return -EOPNOTSUPP; | |
850 | ||
851 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH; | |
852 | attr->vlan = tcf_vlan_push_vid(a); | |
853 | } | |
854 | continue; | |
855 | } | |
856 | ||
bbd00f7e HHZ |
857 | if (is_tcf_tunnel_release(a)) { |
858 | attr->action |= MLX5_FLOW_CONTEXT_ACTION_DECAP; | |
859 | continue; | |
860 | } | |
861 | ||
03a9d11e OG |
862 | return -EINVAL; |
863 | } | |
864 | return 0; | |
865 | } | |
866 | ||
e3a2b7ed AV |
867 | int mlx5e_configure_flower(struct mlx5e_priv *priv, __be16 protocol, |
868 | struct tc_cls_flower_offload *f) | |
869 | { | |
acff797c | 870 | struct mlx5e_tc_table *tc = &priv->fs.tc; |
e3a2b7ed | 871 | int err = 0; |
776b12b6 OG |
872 | bool fdb_flow = false; |
873 | u32 flow_tag, action; | |
e3a2b7ed | 874 | struct mlx5e_tc_flow *flow; |
c5bb1730 | 875 | struct mlx5_flow_spec *spec; |
74491de9 | 876 | struct mlx5_flow_handle *old = NULL; |
d0debb76 | 877 | struct mlx5_esw_flow_attr *old_attr = NULL; |
adb4c123 | 878 | struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; |
e3a2b7ed | 879 | |
776b12b6 OG |
880 | if (esw && esw->mode == SRIOV_OFFLOADS) |
881 | fdb_flow = true; | |
882 | ||
e3a2b7ed AV |
883 | flow = rhashtable_lookup_fast(&tc->ht, &f->cookie, |
884 | tc->ht_params); | |
776b12b6 | 885 | if (flow) { |
e3a2b7ed | 886 | old = flow->rule; |
8b32580d | 887 | old_attr = flow->attr; |
776b12b6 OG |
888 | } else { |
889 | if (fdb_flow) | |
890 | flow = kzalloc(sizeof(*flow) + sizeof(struct mlx5_esw_flow_attr), | |
891 | GFP_KERNEL); | |
892 | else | |
893 | flow = kzalloc(sizeof(*flow), GFP_KERNEL); | |
894 | } | |
e3a2b7ed | 895 | |
c5bb1730 MG |
896 | spec = mlx5_vzalloc(sizeof(*spec)); |
897 | if (!spec || !flow) { | |
e3a2b7ed AV |
898 | err = -ENOMEM; |
899 | goto err_free; | |
900 | } | |
901 | ||
902 | flow->cookie = f->cookie; | |
903 | ||
c5bb1730 | 904 | err = parse_cls_flower(priv, spec, f); |
e3a2b7ed AV |
905 | if (err < 0) |
906 | goto err_free; | |
907 | ||
776b12b6 OG |
908 | if (fdb_flow) { |
909 | flow->attr = (struct mlx5_esw_flow_attr *)(flow + 1); | |
a54e20b4 | 910 | err = parse_tc_fdb_actions(priv, f->exts, flow); |
adb4c123 OG |
911 | if (err < 0) |
912 | goto err_free; | |
776b12b6 | 913 | flow->rule = mlx5e_tc_add_fdb_flow(priv, spec, flow->attr); |
adb4c123 OG |
914 | } else { |
915 | err = parse_tc_nic_actions(priv, f->exts, &action, &flow_tag); | |
916 | if (err < 0) | |
917 | goto err_free; | |
918 | flow->rule = mlx5e_tc_add_nic_flow(priv, spec, action, flow_tag); | |
919 | } | |
e3a2b7ed | 920 | |
e3a2b7ed AV |
921 | if (IS_ERR(flow->rule)) { |
922 | err = PTR_ERR(flow->rule); | |
5c40348c | 923 | goto err_free; |
e3a2b7ed AV |
924 | } |
925 | ||
5c40348c OG |
926 | err = rhashtable_insert_fast(&tc->ht, &flow->node, |
927 | tc->ht_params); | |
928 | if (err) | |
929 | goto err_del_rule; | |
930 | ||
e3a2b7ed | 931 | if (old) |
8b32580d | 932 | mlx5e_tc_del_flow(priv, old, old_attr); |
e3a2b7ed AV |
933 | |
934 | goto out; | |
935 | ||
5c40348c | 936 | err_del_rule: |
74491de9 | 937 | mlx5_del_flow_rules(flow->rule); |
e3a2b7ed AV |
938 | |
939 | err_free: | |
940 | if (!old) | |
941 | kfree(flow); | |
942 | out: | |
c5bb1730 | 943 | kvfree(spec); |
e3a2b7ed AV |
944 | return err; |
945 | } | |
946 | ||
a54e20b4 HHZ |
947 | static void mlx5e_detach_encap(struct mlx5e_priv *priv, |
948 | struct mlx5e_tc_flow *flow) { | |
949 | struct list_head *next = flow->encap.next; | |
950 | ||
951 | list_del(&flow->encap); | |
952 | if (list_empty(next)) { | |
953 | struct mlx5_encap_entry *e; | |
954 | ||
955 | e = list_entry(next, struct mlx5_encap_entry, flows); | |
956 | if (e->n) { | |
957 | mlx5_encap_dealloc(priv->mdev, e->encap_id); | |
958 | neigh_release(e->n); | |
959 | } | |
960 | hlist_del_rcu(&e->encap_hlist); | |
961 | kfree(e); | |
962 | } | |
963 | } | |
964 | ||
e3a2b7ed AV |
965 | int mlx5e_delete_flower(struct mlx5e_priv *priv, |
966 | struct tc_cls_flower_offload *f) | |
967 | { | |
968 | struct mlx5e_tc_flow *flow; | |
acff797c | 969 | struct mlx5e_tc_table *tc = &priv->fs.tc; |
e3a2b7ed AV |
970 | |
971 | flow = rhashtable_lookup_fast(&tc->ht, &f->cookie, | |
972 | tc->ht_params); | |
973 | if (!flow) | |
974 | return -EINVAL; | |
975 | ||
976 | rhashtable_remove_fast(&tc->ht, &flow->node, tc->ht_params); | |
977 | ||
8b32580d | 978 | mlx5e_tc_del_flow(priv, flow->rule, flow->attr); |
e3a2b7ed | 979 | |
a54e20b4 HHZ |
980 | if (flow->attr->action & MLX5_FLOW_CONTEXT_ACTION_ENCAP) |
981 | mlx5e_detach_encap(priv, flow); | |
982 | ||
e3a2b7ed AV |
983 | kfree(flow); |
984 | ||
985 | return 0; | |
986 | } | |
987 | ||
aad7e08d AV |
988 | int mlx5e_stats_flower(struct mlx5e_priv *priv, |
989 | struct tc_cls_flower_offload *f) | |
990 | { | |
991 | struct mlx5e_tc_table *tc = &priv->fs.tc; | |
992 | struct mlx5e_tc_flow *flow; | |
993 | struct tc_action *a; | |
994 | struct mlx5_fc *counter; | |
22dc13c8 | 995 | LIST_HEAD(actions); |
aad7e08d AV |
996 | u64 bytes; |
997 | u64 packets; | |
998 | u64 lastuse; | |
999 | ||
1000 | flow = rhashtable_lookup_fast(&tc->ht, &f->cookie, | |
1001 | tc->ht_params); | |
1002 | if (!flow) | |
1003 | return -EINVAL; | |
1004 | ||
1005 | counter = mlx5_flow_rule_counter(flow->rule); | |
1006 | if (!counter) | |
1007 | return 0; | |
1008 | ||
1009 | mlx5_fc_query_cached(counter, &bytes, &packets, &lastuse); | |
1010 | ||
22dc13c8 WC |
1011 | tcf_exts_to_list(f->exts, &actions); |
1012 | list_for_each_entry(a, &actions, list) | |
aad7e08d AV |
1013 | tcf_action_stats_update(a, bytes, packets, lastuse); |
1014 | ||
1015 | return 0; | |
1016 | } | |
1017 | ||
e8f887ac AV |
1018 | static const struct rhashtable_params mlx5e_tc_flow_ht_params = { |
1019 | .head_offset = offsetof(struct mlx5e_tc_flow, node), | |
1020 | .key_offset = offsetof(struct mlx5e_tc_flow, cookie), | |
1021 | .key_len = sizeof(((struct mlx5e_tc_flow *)0)->cookie), | |
1022 | .automatic_shrinking = true, | |
1023 | }; | |
1024 | ||
1025 | int mlx5e_tc_init(struct mlx5e_priv *priv) | |
1026 | { | |
acff797c | 1027 | struct mlx5e_tc_table *tc = &priv->fs.tc; |
e8f887ac AV |
1028 | |
1029 | tc->ht_params = mlx5e_tc_flow_ht_params; | |
1030 | return rhashtable_init(&tc->ht, &tc->ht_params); | |
1031 | } | |
1032 | ||
1033 | static void _mlx5e_tc_del_flow(void *ptr, void *arg) | |
1034 | { | |
1035 | struct mlx5e_tc_flow *flow = ptr; | |
1036 | struct mlx5e_priv *priv = arg; | |
1037 | ||
8b32580d | 1038 | mlx5e_tc_del_flow(priv, flow->rule, flow->attr); |
e8f887ac AV |
1039 | kfree(flow); |
1040 | } | |
1041 | ||
1042 | void mlx5e_tc_cleanup(struct mlx5e_priv *priv) | |
1043 | { | |
acff797c | 1044 | struct mlx5e_tc_table *tc = &priv->fs.tc; |
e8f887ac AV |
1045 | |
1046 | rhashtable_free_and_destroy(&tc->ht, _mlx5e_tc_del_flow, priv); | |
1047 | ||
acff797c MG |
1048 | if (!IS_ERR_OR_NULL(tc->t)) { |
1049 | mlx5_destroy_flow_table(tc->t); | |
1050 | tc->t = NULL; | |
e8f887ac AV |
1051 | } |
1052 | } |