]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/blob - net/dsa/switch.c
tcp: Don't acquire inet_listen_hashbucket::lock with disabled BH.
[mirror_ubuntu-jammy-kernel.git] / net / dsa / switch.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Handling of a single switch chip, part of a switch fabric
4 *
5 * Copyright (c) 2017 Savoir-faire Linux Inc.
6 * Vivien Didelot <vivien.didelot@savoirfairelinux.com>
7 */
8
9 #include <linux/if_bridge.h>
10 #include <linux/netdevice.h>
11 #include <linux/notifier.h>
12 #include <linux/if_vlan.h>
13 #include <net/switchdev.h>
14
15 #include "dsa_priv.h"
16
17 static unsigned int dsa_switch_fastest_ageing_time(struct dsa_switch *ds,
18 unsigned int ageing_time)
19 {
20 int i;
21
22 for (i = 0; i < ds->num_ports; ++i) {
23 struct dsa_port *dp = dsa_to_port(ds, i);
24
25 if (dp->ageing_time && dp->ageing_time < ageing_time)
26 ageing_time = dp->ageing_time;
27 }
28
29 return ageing_time;
30 }
31
32 static int dsa_switch_ageing_time(struct dsa_switch *ds,
33 struct dsa_notifier_ageing_time_info *info)
34 {
35 unsigned int ageing_time = info->ageing_time;
36
37 if (ds->ageing_time_min && ageing_time < ds->ageing_time_min)
38 return -ERANGE;
39
40 if (ds->ageing_time_max && ageing_time > ds->ageing_time_max)
41 return -ERANGE;
42
43 /* Program the fastest ageing time in case of multiple bridges */
44 ageing_time = dsa_switch_fastest_ageing_time(ds, ageing_time);
45
46 if (ds->ops->set_ageing_time)
47 return ds->ops->set_ageing_time(ds, ageing_time);
48
49 return 0;
50 }
51
52 static bool dsa_switch_mtu_match(struct dsa_switch *ds, int port,
53 struct dsa_notifier_mtu_info *info)
54 {
55 if (ds->index == info->sw_index && port == info->port)
56 return true;
57
58 /* Do not propagate to other switches in the tree if the notifier was
59 * targeted for a single switch.
60 */
61 if (info->targeted_match)
62 return false;
63
64 if (dsa_is_dsa_port(ds, port) || dsa_is_cpu_port(ds, port))
65 return true;
66
67 return false;
68 }
69
70 static int dsa_switch_mtu(struct dsa_switch *ds,
71 struct dsa_notifier_mtu_info *info)
72 {
73 int port, ret;
74
75 if (!ds->ops->port_change_mtu)
76 return -EOPNOTSUPP;
77
78 for (port = 0; port < ds->num_ports; port++) {
79 if (dsa_switch_mtu_match(ds, port, info)) {
80 ret = ds->ops->port_change_mtu(ds, port, info->mtu);
81 if (ret)
82 return ret;
83 }
84 }
85
86 return 0;
87 }
88
89 static int dsa_switch_bridge_join(struct dsa_switch *ds,
90 struct dsa_notifier_bridge_info *info)
91 {
92 struct dsa_switch_tree *dst = ds->dst;
93 int err;
94
95 if (dst->index == info->tree_index && ds->index == info->sw_index) {
96 if (!ds->ops->port_bridge_join)
97 return -EOPNOTSUPP;
98
99 err = ds->ops->port_bridge_join(ds, info->port, info->br);
100 if (err)
101 return err;
102 }
103
104 if ((dst->index != info->tree_index || ds->index != info->sw_index) &&
105 ds->ops->crosschip_bridge_join) {
106 err = ds->ops->crosschip_bridge_join(ds, info->tree_index,
107 info->sw_index,
108 info->port, info->br);
109 if (err)
110 return err;
111 }
112
113 return dsa_tag_8021q_bridge_join(ds, info);
114 }
115
116 static int dsa_switch_bridge_leave(struct dsa_switch *ds,
117 struct dsa_notifier_bridge_info *info)
118 {
119 struct dsa_switch_tree *dst = ds->dst;
120 struct netlink_ext_ack extack = {0};
121 bool change_vlan_filtering = false;
122 bool vlan_filtering;
123 int err, port;
124
125 if (dst->index == info->tree_index && ds->index == info->sw_index &&
126 ds->ops->port_bridge_leave)
127 ds->ops->port_bridge_leave(ds, info->port, info->br);
128
129 if ((dst->index != info->tree_index || ds->index != info->sw_index) &&
130 ds->ops->crosschip_bridge_leave)
131 ds->ops->crosschip_bridge_leave(ds, info->tree_index,
132 info->sw_index, info->port,
133 info->br);
134
135 if (ds->needs_standalone_vlan_filtering && !br_vlan_enabled(info->br)) {
136 change_vlan_filtering = true;
137 vlan_filtering = true;
138 } else if (!ds->needs_standalone_vlan_filtering &&
139 br_vlan_enabled(info->br)) {
140 change_vlan_filtering = true;
141 vlan_filtering = false;
142 }
143
144 /* If the bridge was vlan_filtering, the bridge core doesn't trigger an
145 * event for changing vlan_filtering setting upon slave ports leaving
146 * it. That is a good thing, because that lets us handle it and also
147 * handle the case where the switch's vlan_filtering setting is global
148 * (not per port). When that happens, the correct moment to trigger the
149 * vlan_filtering callback is only when the last port leaves the last
150 * VLAN-aware bridge.
151 */
152 if (change_vlan_filtering && ds->vlan_filtering_is_global) {
153 for (port = 0; port < ds->num_ports; port++) {
154 struct net_device *bridge_dev;
155
156 bridge_dev = dsa_to_port(ds, port)->bridge_dev;
157
158 if (bridge_dev && br_vlan_enabled(bridge_dev)) {
159 change_vlan_filtering = false;
160 break;
161 }
162 }
163 }
164
165 if (change_vlan_filtering) {
166 err = dsa_port_vlan_filtering(dsa_to_port(ds, info->port),
167 vlan_filtering, &extack);
168 if (extack._msg)
169 dev_err(ds->dev, "port %d: %s\n", info->port,
170 extack._msg);
171 if (err && err != -EOPNOTSUPP)
172 return err;
173 }
174
175 return dsa_tag_8021q_bridge_leave(ds, info);
176 }
177
178 /* Matches for all upstream-facing ports (the CPU port and all upstream-facing
179 * DSA links) that sit between the targeted port on which the notifier was
180 * emitted and its dedicated CPU port.
181 */
182 static bool dsa_switch_host_address_match(struct dsa_switch *ds, int port,
183 int info_sw_index, int info_port)
184 {
185 struct dsa_port *targeted_dp, *cpu_dp;
186 struct dsa_switch *targeted_ds;
187
188 targeted_ds = dsa_switch_find(ds->dst->index, info_sw_index);
189 targeted_dp = dsa_to_port(targeted_ds, info_port);
190 cpu_dp = targeted_dp->cpu_dp;
191
192 if (dsa_switch_is_upstream_of(ds, targeted_ds))
193 return port == dsa_towards_port(ds, cpu_dp->ds->index,
194 cpu_dp->index);
195
196 return false;
197 }
198
199 static struct dsa_mac_addr *dsa_mac_addr_find(struct list_head *addr_list,
200 const unsigned char *addr,
201 u16 vid)
202 {
203 struct dsa_mac_addr *a;
204
205 list_for_each_entry(a, addr_list, list)
206 if (ether_addr_equal(a->addr, addr) && a->vid == vid)
207 return a;
208
209 return NULL;
210 }
211
212 static int dsa_switch_do_mdb_add(struct dsa_switch *ds, int port,
213 const struct switchdev_obj_port_mdb *mdb)
214 {
215 struct dsa_port *dp = dsa_to_port(ds, port);
216 struct dsa_mac_addr *a;
217 int err;
218
219 /* No need to bother with refcounting for user ports */
220 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
221 return ds->ops->port_mdb_add(ds, port, mdb);
222
223 a = dsa_mac_addr_find(&dp->mdbs, mdb->addr, mdb->vid);
224 if (a) {
225 refcount_inc(&a->refcount);
226 return 0;
227 }
228
229 a = kzalloc(sizeof(*a), GFP_KERNEL);
230 if (!a)
231 return -ENOMEM;
232
233 err = ds->ops->port_mdb_add(ds, port, mdb);
234 if (err) {
235 kfree(a);
236 return err;
237 }
238
239 ether_addr_copy(a->addr, mdb->addr);
240 a->vid = mdb->vid;
241 refcount_set(&a->refcount, 1);
242 list_add_tail(&a->list, &dp->mdbs);
243
244 return 0;
245 }
246
247 static int dsa_switch_do_mdb_del(struct dsa_switch *ds, int port,
248 const struct switchdev_obj_port_mdb *mdb)
249 {
250 struct dsa_port *dp = dsa_to_port(ds, port);
251 struct dsa_mac_addr *a;
252 int err;
253
254 /* No need to bother with refcounting for user ports */
255 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
256 return ds->ops->port_mdb_del(ds, port, mdb);
257
258 a = dsa_mac_addr_find(&dp->mdbs, mdb->addr, mdb->vid);
259 if (!a)
260 return -ENOENT;
261
262 if (!refcount_dec_and_test(&a->refcount))
263 return 0;
264
265 err = ds->ops->port_mdb_del(ds, port, mdb);
266 if (err) {
267 refcount_set(&a->refcount, 1);
268 return err;
269 }
270
271 list_del(&a->list);
272 kfree(a);
273
274 return 0;
275 }
276
277 static int dsa_switch_do_fdb_add(struct dsa_switch *ds, int port,
278 const unsigned char *addr, u16 vid)
279 {
280 struct dsa_port *dp = dsa_to_port(ds, port);
281 struct dsa_mac_addr *a;
282 int err;
283
284 /* No need to bother with refcounting for user ports */
285 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
286 return ds->ops->port_fdb_add(ds, port, addr, vid);
287
288 a = dsa_mac_addr_find(&dp->fdbs, addr, vid);
289 if (a) {
290 refcount_inc(&a->refcount);
291 return 0;
292 }
293
294 a = kzalloc(sizeof(*a), GFP_KERNEL);
295 if (!a)
296 return -ENOMEM;
297
298 err = ds->ops->port_fdb_add(ds, port, addr, vid);
299 if (err) {
300 kfree(a);
301 return err;
302 }
303
304 ether_addr_copy(a->addr, addr);
305 a->vid = vid;
306 refcount_set(&a->refcount, 1);
307 list_add_tail(&a->list, &dp->fdbs);
308
309 return 0;
310 }
311
312 static int dsa_switch_do_fdb_del(struct dsa_switch *ds, int port,
313 const unsigned char *addr, u16 vid)
314 {
315 struct dsa_port *dp = dsa_to_port(ds, port);
316 struct dsa_mac_addr *a;
317 int err;
318
319 /* No need to bother with refcounting for user ports */
320 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
321 return ds->ops->port_fdb_del(ds, port, addr, vid);
322
323 a = dsa_mac_addr_find(&dp->fdbs, addr, vid);
324 if (!a)
325 return -ENOENT;
326
327 if (!refcount_dec_and_test(&a->refcount))
328 return 0;
329
330 err = ds->ops->port_fdb_del(ds, port, addr, vid);
331 if (err) {
332 refcount_set(&a->refcount, 1);
333 return err;
334 }
335
336 list_del(&a->list);
337 kfree(a);
338
339 return 0;
340 }
341
342 static int dsa_switch_host_fdb_add(struct dsa_switch *ds,
343 struct dsa_notifier_fdb_info *info)
344 {
345 int err = 0;
346 int port;
347
348 if (!ds->ops->port_fdb_add)
349 return -EOPNOTSUPP;
350
351 for (port = 0; port < ds->num_ports; port++) {
352 if (dsa_switch_host_address_match(ds, port, info->sw_index,
353 info->port)) {
354 err = dsa_switch_do_fdb_add(ds, port, info->addr,
355 info->vid);
356 if (err)
357 break;
358 }
359 }
360
361 return err;
362 }
363
364 static int dsa_switch_host_fdb_del(struct dsa_switch *ds,
365 struct dsa_notifier_fdb_info *info)
366 {
367 int err = 0;
368 int port;
369
370 if (!ds->ops->port_fdb_del)
371 return -EOPNOTSUPP;
372
373 for (port = 0; port < ds->num_ports; port++) {
374 if (dsa_switch_host_address_match(ds, port, info->sw_index,
375 info->port)) {
376 err = dsa_switch_do_fdb_del(ds, port, info->addr,
377 info->vid);
378 if (err)
379 break;
380 }
381 }
382
383 return err;
384 }
385
386 static int dsa_switch_fdb_add(struct dsa_switch *ds,
387 struct dsa_notifier_fdb_info *info)
388 {
389 int port = dsa_towards_port(ds, info->sw_index, info->port);
390
391 if (!ds->ops->port_fdb_add)
392 return -EOPNOTSUPP;
393
394 return dsa_switch_do_fdb_add(ds, port, info->addr, info->vid);
395 }
396
397 static int dsa_switch_fdb_del(struct dsa_switch *ds,
398 struct dsa_notifier_fdb_info *info)
399 {
400 int port = dsa_towards_port(ds, info->sw_index, info->port);
401
402 if (!ds->ops->port_fdb_del)
403 return -EOPNOTSUPP;
404
405 return dsa_switch_do_fdb_del(ds, port, info->addr, info->vid);
406 }
407
408 static int dsa_switch_hsr_join(struct dsa_switch *ds,
409 struct dsa_notifier_hsr_info *info)
410 {
411 if (ds->index == info->sw_index && ds->ops->port_hsr_join)
412 return ds->ops->port_hsr_join(ds, info->port, info->hsr);
413
414 return -EOPNOTSUPP;
415 }
416
417 static int dsa_switch_hsr_leave(struct dsa_switch *ds,
418 struct dsa_notifier_hsr_info *info)
419 {
420 if (ds->index == info->sw_index && ds->ops->port_hsr_leave)
421 return ds->ops->port_hsr_leave(ds, info->port, info->hsr);
422
423 return -EOPNOTSUPP;
424 }
425
426 static int dsa_switch_lag_change(struct dsa_switch *ds,
427 struct dsa_notifier_lag_info *info)
428 {
429 if (ds->index == info->sw_index && ds->ops->port_lag_change)
430 return ds->ops->port_lag_change(ds, info->port);
431
432 if (ds->index != info->sw_index && ds->ops->crosschip_lag_change)
433 return ds->ops->crosschip_lag_change(ds, info->sw_index,
434 info->port);
435
436 return 0;
437 }
438
439 static int dsa_switch_lag_join(struct dsa_switch *ds,
440 struct dsa_notifier_lag_info *info)
441 {
442 if (ds->index == info->sw_index && ds->ops->port_lag_join)
443 return ds->ops->port_lag_join(ds, info->port, info->lag,
444 info->info);
445
446 if (ds->index != info->sw_index && ds->ops->crosschip_lag_join)
447 return ds->ops->crosschip_lag_join(ds, info->sw_index,
448 info->port, info->lag,
449 info->info);
450
451 return -EOPNOTSUPP;
452 }
453
454 static int dsa_switch_lag_leave(struct dsa_switch *ds,
455 struct dsa_notifier_lag_info *info)
456 {
457 if (ds->index == info->sw_index && ds->ops->port_lag_leave)
458 return ds->ops->port_lag_leave(ds, info->port, info->lag);
459
460 if (ds->index != info->sw_index && ds->ops->crosschip_lag_leave)
461 return ds->ops->crosschip_lag_leave(ds, info->sw_index,
462 info->port, info->lag);
463
464 return -EOPNOTSUPP;
465 }
466
467 static int dsa_switch_mdb_add(struct dsa_switch *ds,
468 struct dsa_notifier_mdb_info *info)
469 {
470 int port = dsa_towards_port(ds, info->sw_index, info->port);
471
472 if (!ds->ops->port_mdb_add)
473 return -EOPNOTSUPP;
474
475 return dsa_switch_do_mdb_add(ds, port, info->mdb);
476 }
477
478 static int dsa_switch_mdb_del(struct dsa_switch *ds,
479 struct dsa_notifier_mdb_info *info)
480 {
481 int port = dsa_towards_port(ds, info->sw_index, info->port);
482
483 if (!ds->ops->port_mdb_del)
484 return -EOPNOTSUPP;
485
486 return dsa_switch_do_mdb_del(ds, port, info->mdb);
487 }
488
489 static int dsa_switch_host_mdb_add(struct dsa_switch *ds,
490 struct dsa_notifier_mdb_info *info)
491 {
492 int err = 0;
493 int port;
494
495 if (!ds->ops->port_mdb_add)
496 return -EOPNOTSUPP;
497
498 for (port = 0; port < ds->num_ports; port++) {
499 if (dsa_switch_host_address_match(ds, port, info->sw_index,
500 info->port)) {
501 err = dsa_switch_do_mdb_add(ds, port, info->mdb);
502 if (err)
503 break;
504 }
505 }
506
507 return err;
508 }
509
510 static int dsa_switch_host_mdb_del(struct dsa_switch *ds,
511 struct dsa_notifier_mdb_info *info)
512 {
513 int err = 0;
514 int port;
515
516 if (!ds->ops->port_mdb_del)
517 return -EOPNOTSUPP;
518
519 for (port = 0; port < ds->num_ports; port++) {
520 if (dsa_switch_host_address_match(ds, port, info->sw_index,
521 info->port)) {
522 err = dsa_switch_do_mdb_del(ds, port, info->mdb);
523 if (err)
524 break;
525 }
526 }
527
528 return err;
529 }
530
531 static bool dsa_switch_vlan_match(struct dsa_switch *ds, int port,
532 struct dsa_notifier_vlan_info *info)
533 {
534 if (ds->index == info->sw_index && port == info->port)
535 return true;
536
537 if (dsa_is_dsa_port(ds, port))
538 return true;
539
540 return false;
541 }
542
543 static int dsa_switch_vlan_add(struct dsa_switch *ds,
544 struct dsa_notifier_vlan_info *info)
545 {
546 int port, err;
547
548 if (!ds->ops->port_vlan_add)
549 return -EOPNOTSUPP;
550
551 for (port = 0; port < ds->num_ports; port++) {
552 if (dsa_switch_vlan_match(ds, port, info)) {
553 err = ds->ops->port_vlan_add(ds, port, info->vlan,
554 info->extack);
555 if (err)
556 return err;
557 }
558 }
559
560 return 0;
561 }
562
563 static int dsa_switch_vlan_del(struct dsa_switch *ds,
564 struct dsa_notifier_vlan_info *info)
565 {
566 if (!ds->ops->port_vlan_del)
567 return -EOPNOTSUPP;
568
569 if (ds->index == info->sw_index)
570 return ds->ops->port_vlan_del(ds, info->port, info->vlan);
571
572 /* Do not deprogram the DSA links as they may be used as conduit
573 * for other VLAN members in the fabric.
574 */
575 return 0;
576 }
577
578 static int dsa_switch_change_tag_proto(struct dsa_switch *ds,
579 struct dsa_notifier_tag_proto_info *info)
580 {
581 const struct dsa_device_ops *tag_ops = info->tag_ops;
582 int port, err;
583
584 if (!ds->ops->change_tag_protocol)
585 return -EOPNOTSUPP;
586
587 ASSERT_RTNL();
588
589 for (port = 0; port < ds->num_ports; port++) {
590 if (!dsa_is_cpu_port(ds, port))
591 continue;
592
593 err = ds->ops->change_tag_protocol(ds, port, tag_ops->proto);
594 if (err)
595 return err;
596
597 dsa_port_set_tag_protocol(dsa_to_port(ds, port), tag_ops);
598 }
599
600 /* Now that changing the tag protocol can no longer fail, let's update
601 * the remaining bits which are "duplicated for faster access", and the
602 * bits that depend on the tagger, such as the MTU.
603 */
604 for (port = 0; port < ds->num_ports; port++) {
605 if (dsa_is_user_port(ds, port)) {
606 struct net_device *slave;
607
608 slave = dsa_to_port(ds, port)->slave;
609 dsa_slave_setup_tagger(slave);
610
611 /* rtnl_mutex is held in dsa_tree_change_tag_proto */
612 dsa_slave_change_mtu(slave, slave->mtu);
613 }
614 }
615
616 return 0;
617 }
618
619 static int dsa_switch_mrp_add(struct dsa_switch *ds,
620 struct dsa_notifier_mrp_info *info)
621 {
622 if (!ds->ops->port_mrp_add)
623 return -EOPNOTSUPP;
624
625 if (ds->index == info->sw_index)
626 return ds->ops->port_mrp_add(ds, info->port, info->mrp);
627
628 return 0;
629 }
630
631 static int dsa_switch_mrp_del(struct dsa_switch *ds,
632 struct dsa_notifier_mrp_info *info)
633 {
634 if (!ds->ops->port_mrp_del)
635 return -EOPNOTSUPP;
636
637 if (ds->index == info->sw_index)
638 return ds->ops->port_mrp_del(ds, info->port, info->mrp);
639
640 return 0;
641 }
642
643 static int
644 dsa_switch_mrp_add_ring_role(struct dsa_switch *ds,
645 struct dsa_notifier_mrp_ring_role_info *info)
646 {
647 if (!ds->ops->port_mrp_add_ring_role)
648 return -EOPNOTSUPP;
649
650 if (ds->index == info->sw_index)
651 return ds->ops->port_mrp_add_ring_role(ds, info->port,
652 info->mrp);
653
654 return 0;
655 }
656
657 static int
658 dsa_switch_mrp_del_ring_role(struct dsa_switch *ds,
659 struct dsa_notifier_mrp_ring_role_info *info)
660 {
661 if (!ds->ops->port_mrp_del_ring_role)
662 return -EOPNOTSUPP;
663
664 if (ds->index == info->sw_index)
665 return ds->ops->port_mrp_del_ring_role(ds, info->port,
666 info->mrp);
667
668 return 0;
669 }
670
671 static int dsa_switch_event(struct notifier_block *nb,
672 unsigned long event, void *info)
673 {
674 struct dsa_switch *ds = container_of(nb, struct dsa_switch, nb);
675 int err;
676
677 switch (event) {
678 case DSA_NOTIFIER_AGEING_TIME:
679 err = dsa_switch_ageing_time(ds, info);
680 break;
681 case DSA_NOTIFIER_BRIDGE_JOIN:
682 err = dsa_switch_bridge_join(ds, info);
683 break;
684 case DSA_NOTIFIER_BRIDGE_LEAVE:
685 err = dsa_switch_bridge_leave(ds, info);
686 break;
687 case DSA_NOTIFIER_FDB_ADD:
688 err = dsa_switch_fdb_add(ds, info);
689 break;
690 case DSA_NOTIFIER_FDB_DEL:
691 err = dsa_switch_fdb_del(ds, info);
692 break;
693 case DSA_NOTIFIER_HOST_FDB_ADD:
694 err = dsa_switch_host_fdb_add(ds, info);
695 break;
696 case DSA_NOTIFIER_HOST_FDB_DEL:
697 err = dsa_switch_host_fdb_del(ds, info);
698 break;
699 case DSA_NOTIFIER_HSR_JOIN:
700 err = dsa_switch_hsr_join(ds, info);
701 break;
702 case DSA_NOTIFIER_HSR_LEAVE:
703 err = dsa_switch_hsr_leave(ds, info);
704 break;
705 case DSA_NOTIFIER_LAG_CHANGE:
706 err = dsa_switch_lag_change(ds, info);
707 break;
708 case DSA_NOTIFIER_LAG_JOIN:
709 err = dsa_switch_lag_join(ds, info);
710 break;
711 case DSA_NOTIFIER_LAG_LEAVE:
712 err = dsa_switch_lag_leave(ds, info);
713 break;
714 case DSA_NOTIFIER_MDB_ADD:
715 err = dsa_switch_mdb_add(ds, info);
716 break;
717 case DSA_NOTIFIER_MDB_DEL:
718 err = dsa_switch_mdb_del(ds, info);
719 break;
720 case DSA_NOTIFIER_HOST_MDB_ADD:
721 err = dsa_switch_host_mdb_add(ds, info);
722 break;
723 case DSA_NOTIFIER_HOST_MDB_DEL:
724 err = dsa_switch_host_mdb_del(ds, info);
725 break;
726 case DSA_NOTIFIER_VLAN_ADD:
727 err = dsa_switch_vlan_add(ds, info);
728 break;
729 case DSA_NOTIFIER_VLAN_DEL:
730 err = dsa_switch_vlan_del(ds, info);
731 break;
732 case DSA_NOTIFIER_MTU:
733 err = dsa_switch_mtu(ds, info);
734 break;
735 case DSA_NOTIFIER_TAG_PROTO:
736 err = dsa_switch_change_tag_proto(ds, info);
737 break;
738 case DSA_NOTIFIER_MRP_ADD:
739 err = dsa_switch_mrp_add(ds, info);
740 break;
741 case DSA_NOTIFIER_MRP_DEL:
742 err = dsa_switch_mrp_del(ds, info);
743 break;
744 case DSA_NOTIFIER_MRP_ADD_RING_ROLE:
745 err = dsa_switch_mrp_add_ring_role(ds, info);
746 break;
747 case DSA_NOTIFIER_MRP_DEL_RING_ROLE:
748 err = dsa_switch_mrp_del_ring_role(ds, info);
749 break;
750 case DSA_NOTIFIER_TAG_8021Q_VLAN_ADD:
751 err = dsa_switch_tag_8021q_vlan_add(ds, info);
752 break;
753 case DSA_NOTIFIER_TAG_8021Q_VLAN_DEL:
754 err = dsa_switch_tag_8021q_vlan_del(ds, info);
755 break;
756 default:
757 err = -EOPNOTSUPP;
758 break;
759 }
760
761 if (err)
762 dev_dbg(ds->dev, "breaking chain for DSA event %lu (%d)\n",
763 event, err);
764
765 return notifier_from_errno(err);
766 }
767
768 int dsa_switch_register_notifier(struct dsa_switch *ds)
769 {
770 ds->nb.notifier_call = dsa_switch_event;
771
772 return raw_notifier_chain_register(&ds->dst->nh, &ds->nb);
773 }
774
775 void dsa_switch_unregister_notifier(struct dsa_switch *ds)
776 {
777 int err;
778
779 err = raw_notifier_chain_unregister(&ds->dst->nh, &ds->nb);
780 if (err)
781 dev_err(ds->dev, "failed to unregister notifier (%d)\n", err);
782 }