blob: 0f99297b2fb3517942bf74fee08ed3e61d65a4f0 [file] [log] [blame]
Andrew Lunn83c0afa2016-06-04 21:17:07 +02001/*
2 * net/dsa/dsa2.c - Hardware switch handling, binding version 2
3 * Copyright (c) 2008-2009 Marvell Semiconductor
4 * Copyright (c) 2013 Florian Fainelli <florian@openwrt.org>
5 * Copyright (c) 2016 Andrew Lunn <andrew@lunn.ch>
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License as published by
9 * the Free Software Foundation; either version 2 of the License, or
10 * (at your option) any later version.
11 */
12
13#include <linux/device.h>
14#include <linux/err.h>
15#include <linux/list.h>
16#include <linux/slab.h>
17#include <linux/rtnetlink.h>
18#include <net/dsa.h>
19#include <linux/of.h>
20#include <linux/of_net.h>
21#include "dsa_priv.h"
22
23static LIST_HEAD(dsa_switch_trees);
24static DEFINE_MUTEX(dsa2_mutex);
25
26static struct dsa_switch_tree *dsa_get_dst(u32 tree)
27{
28 struct dsa_switch_tree *dst;
29
30 list_for_each_entry(dst, &dsa_switch_trees, list)
Nikita Yushchenko7a99cd62016-11-28 09:48:48 +030031 if (dst->tree == tree) {
32 kref_get(&dst->refcount);
Andrew Lunn83c0afa2016-06-04 21:17:07 +020033 return dst;
Nikita Yushchenko7a99cd62016-11-28 09:48:48 +030034 }
Andrew Lunn83c0afa2016-06-04 21:17:07 +020035 return NULL;
36}
37
38static void dsa_free_dst(struct kref *ref)
39{
40 struct dsa_switch_tree *dst = container_of(ref, struct dsa_switch_tree,
41 refcount);
42
43 list_del(&dst->list);
44 kfree(dst);
45}
46
47static void dsa_put_dst(struct dsa_switch_tree *dst)
48{
49 kref_put(&dst->refcount, dsa_free_dst);
50}
51
52static struct dsa_switch_tree *dsa_add_dst(u32 tree)
53{
54 struct dsa_switch_tree *dst;
55
56 dst = kzalloc(sizeof(*dst), GFP_KERNEL);
57 if (!dst)
58 return NULL;
59 dst->tree = tree;
60 dst->cpu_switch = -1;
61 INIT_LIST_HEAD(&dst->list);
62 list_add_tail(&dsa_switch_trees, &dst->list);
63 kref_init(&dst->refcount);
64
65 return dst;
66}
67
68static void dsa_dst_add_ds(struct dsa_switch_tree *dst,
69 struct dsa_switch *ds, u32 index)
70{
71 kref_get(&dst->refcount);
72 dst->ds[index] = ds;
73}
74
75static void dsa_dst_del_ds(struct dsa_switch_tree *dst,
76 struct dsa_switch *ds, u32 index)
77{
78 dst->ds[index] = NULL;
79 kref_put(&dst->refcount, dsa_free_dst);
80}
81
82static bool dsa_port_is_dsa(struct device_node *port)
83{
84 const char *name;
85
86 name = of_get_property(port, "label", NULL);
87 if (!name)
88 return false;
89
90 if (!strcmp(name, "dsa"))
91 return true;
92
93 return false;
94}
95
96static bool dsa_port_is_cpu(struct device_node *port)
97{
98 const char *name;
99
100 name = of_get_property(port, "label", NULL);
101 if (!name)
102 return false;
103
104 if (!strcmp(name, "cpu"))
105 return true;
106
107 return false;
108}
109
110static bool dsa_ds_find_port(struct dsa_switch *ds,
111 struct device_node *port)
112{
113 u32 index;
114
115 for (index = 0; index < DSA_MAX_PORTS; index++)
116 if (ds->ports[index].dn == port)
117 return true;
118 return false;
119}
120
121static struct dsa_switch *dsa_dst_find_port(struct dsa_switch_tree *dst,
122 struct device_node *port)
123{
124 struct dsa_switch *ds;
125 u32 index;
126
127 for (index = 0; index < DSA_MAX_SWITCHES; index++) {
128 ds = dst->ds[index];
129 if (!ds)
130 continue;
131
132 if (dsa_ds_find_port(ds, port))
133 return ds;
134 }
135
136 return NULL;
137}
138
139static int dsa_port_complete(struct dsa_switch_tree *dst,
140 struct dsa_switch *src_ds,
141 struct device_node *port,
142 u32 src_port)
143{
144 struct device_node *link;
145 int index;
146 struct dsa_switch *dst_ds;
147
148 for (index = 0;; index++) {
149 link = of_parse_phandle(port, "link", index);
150 if (!link)
151 break;
152
153 dst_ds = dsa_dst_find_port(dst, link);
154 of_node_put(link);
155
156 if (!dst_ds)
157 return 1;
158
159 src_ds->rtable[dst_ds->index] = src_port;
160 }
161
162 return 0;
163}
164
165/* A switch is complete if all the DSA ports phandles point to ports
166 * known in the tree. A return value of 1 means the tree is not
167 * complete. This is not an error condition. A value of 0 is
168 * success.
169 */
170static int dsa_ds_complete(struct dsa_switch_tree *dst, struct dsa_switch *ds)
171{
172 struct device_node *port;
173 u32 index;
174 int err;
175
176 for (index = 0; index < DSA_MAX_PORTS; index++) {
177 port = ds->ports[index].dn;
178 if (!port)
179 continue;
180
181 if (!dsa_port_is_dsa(port))
182 continue;
183
184 err = dsa_port_complete(dst, ds, port, index);
185 if (err != 0)
186 return err;
187
188 ds->dsa_port_mask |= BIT(index);
189 }
190
191 return 0;
192}
193
194/* A tree is complete if all the DSA ports phandles point to ports
195 * known in the tree. A return value of 1 means the tree is not
196 * complete. This is not an error condition. A value of 0 is
197 * success.
198 */
199static int dsa_dst_complete(struct dsa_switch_tree *dst)
200{
201 struct dsa_switch *ds;
202 u32 index;
203 int err;
204
205 for (index = 0; index < DSA_MAX_SWITCHES; index++) {
206 ds = dst->ds[index];
207 if (!ds)
208 continue;
209
210 err = dsa_ds_complete(dst, ds);
211 if (err != 0)
212 return err;
213 }
214
215 return 0;
216}
217
218static int dsa_dsa_port_apply(struct device_node *port, u32 index,
219 struct dsa_switch *ds)
220{
221 int err;
222
223 err = dsa_cpu_dsa_setup(ds, ds->dev, port, index);
224 if (err) {
225 dev_warn(ds->dev, "Failed to setup dsa port %d: %d\n",
226 index, err);
227 return err;
228 }
229
230 return 0;
231}
232
233static void dsa_dsa_port_unapply(struct device_node *port, u32 index,
234 struct dsa_switch *ds)
235{
236 dsa_cpu_dsa_destroy(port);
237}
238
239static int dsa_cpu_port_apply(struct device_node *port, u32 index,
240 struct dsa_switch *ds)
241{
242 int err;
243
244 err = dsa_cpu_dsa_setup(ds, ds->dev, port, index);
245 if (err) {
246 dev_warn(ds->dev, "Failed to setup cpu port %d: %d\n",
247 index, err);
248 return err;
249 }
250
251 ds->cpu_port_mask |= BIT(index);
252
253 return 0;
254}
255
256static void dsa_cpu_port_unapply(struct device_node *port, u32 index,
257 struct dsa_switch *ds)
258{
259 dsa_cpu_dsa_destroy(port);
260 ds->cpu_port_mask &= ~BIT(index);
261
262}
263
264static int dsa_user_port_apply(struct device_node *port, u32 index,
265 struct dsa_switch *ds)
266{
267 const char *name;
268 int err;
269
270 name = of_get_property(port, "label", NULL);
271
272 err = dsa_slave_create(ds, ds->dev, index, name);
273 if (err) {
274 dev_warn(ds->dev, "Failed to create slave %d: %d\n",
275 index, err);
Florian Fainelli12758a22017-02-07 23:10:13 -0800276 ds->ports[index].netdev = NULL;
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200277 return err;
278 }
279
280 return 0;
281}
282
283static void dsa_user_port_unapply(struct device_node *port, u32 index,
284 struct dsa_switch *ds)
285{
286 if (ds->ports[index].netdev) {
287 dsa_slave_destroy(ds->ports[index].netdev);
288 ds->ports[index].netdev = NULL;
Florian Fainelli6e830d82016-06-07 16:32:39 -0700289 ds->enabled_port_mask &= ~(1 << index);
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200290 }
291}
292
293static int dsa_ds_apply(struct dsa_switch_tree *dst, struct dsa_switch *ds)
294{
295 struct device_node *port;
296 u32 index;
297 int err;
298
Florian Fainelli6e830d82016-06-07 16:32:39 -0700299 /* Initialize ds->phys_mii_mask before registering the slave MDIO bus
Vivien Didelot9d490b42016-08-23 12:38:56 -0400300 * driver and before ops->setup() has run, since the switch drivers and
Florian Fainelli6e830d82016-06-07 16:32:39 -0700301 * the slave MDIO bus driver rely on these values for probing PHY
302 * devices or not
303 */
304 ds->phys_mii_mask = ds->enabled_port_mask;
305
Vivien Didelot9d490b42016-08-23 12:38:56 -0400306 err = ds->ops->setup(ds);
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200307 if (err < 0)
308 return err;
309
John Crispin092183d2016-09-19 15:28:01 +0200310 if (ds->ops->set_addr) {
311 err = ds->ops->set_addr(ds, dst->master_netdev->dev_addr);
312 if (err < 0)
313 return err;
314 }
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200315
Vivien Didelot9d490b42016-08-23 12:38:56 -0400316 if (!ds->slave_mii_bus && ds->ops->phy_read) {
Florian Fainelli1eb59442016-06-07 16:32:40 -0700317 ds->slave_mii_bus = devm_mdiobus_alloc(ds->dev);
318 if (!ds->slave_mii_bus)
319 return -ENOMEM;
320
321 dsa_slave_mii_bus_init(ds);
322
323 err = mdiobus_register(ds->slave_mii_bus);
324 if (err < 0)
325 return err;
326 }
327
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200328 for (index = 0; index < DSA_MAX_PORTS; index++) {
329 port = ds->ports[index].dn;
330 if (!port)
331 continue;
332
333 if (dsa_port_is_dsa(port)) {
334 err = dsa_dsa_port_apply(port, index, ds);
335 if (err)
336 return err;
337 continue;
338 }
339
340 if (dsa_port_is_cpu(port)) {
341 err = dsa_cpu_port_apply(port, index, ds);
342 if (err)
343 return err;
344 continue;
345 }
346
347 err = dsa_user_port_apply(port, index, ds);
348 if (err)
349 continue;
350 }
351
352 return 0;
353}
354
355static void dsa_ds_unapply(struct dsa_switch_tree *dst, struct dsa_switch *ds)
356{
357 struct device_node *port;
358 u32 index;
359
360 for (index = 0; index < DSA_MAX_PORTS; index++) {
361 port = ds->ports[index].dn;
362 if (!port)
363 continue;
364
365 if (dsa_port_is_dsa(port)) {
366 dsa_dsa_port_unapply(port, index, ds);
367 continue;
368 }
369
370 if (dsa_port_is_cpu(port)) {
371 dsa_cpu_port_unapply(port, index, ds);
372 continue;
373 }
374
375 dsa_user_port_unapply(port, index, ds);
376 }
Florian Fainelli1eb59442016-06-07 16:32:40 -0700377
Vivien Didelot9d490b42016-08-23 12:38:56 -0400378 if (ds->slave_mii_bus && ds->ops->phy_read)
Florian Fainelli1eb59442016-06-07 16:32:40 -0700379 mdiobus_unregister(ds->slave_mii_bus);
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200380}
381
382static int dsa_dst_apply(struct dsa_switch_tree *dst)
383{
384 struct dsa_switch *ds;
385 u32 index;
386 int err;
387
388 for (index = 0; index < DSA_MAX_SWITCHES; index++) {
389 ds = dst->ds[index];
390 if (!ds)
391 continue;
392
393 err = dsa_ds_apply(dst, ds);
394 if (err)
395 return err;
396 }
397
Florian Fainelli2ff4a022017-01-09 11:58:34 -0800398 if (dst->ds[0]) {
399 err = dsa_cpu_port_ethtool_setup(dst->ds[0]);
400 if (err)
401 return err;
402 }
Florian Fainelli0c73c522016-06-07 16:32:42 -0700403
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200404 /* If we use a tagging format that doesn't have an ethertype
405 * field, make sure that all packets from this point on get
406 * sent to the tag format's receive function.
407 */
408 wmb();
409 dst->master_netdev->dsa_ptr = (void *)dst;
410 dst->applied = true;
411
412 return 0;
413}
414
415static void dsa_dst_unapply(struct dsa_switch_tree *dst)
416{
417 struct dsa_switch *ds;
418 u32 index;
419
420 if (!dst->applied)
421 return;
422
423 dst->master_netdev->dsa_ptr = NULL;
424
425 /* If we used a tagging format that doesn't have an ethertype
426 * field, make sure that all packets from this point get sent
427 * without the tag and go through the regular receive path.
428 */
429 wmb();
430
431 for (index = 0; index < DSA_MAX_SWITCHES; index++) {
432 ds = dst->ds[index];
433 if (!ds)
434 continue;
435
436 dsa_ds_unapply(dst, ds);
437 }
438
Florian Fainelli2ff4a022017-01-09 11:58:34 -0800439 if (dst->ds[0])
440 dsa_cpu_port_ethtool_restore(dst->ds[0]);
Florian Fainelli0c73c522016-06-07 16:32:42 -0700441
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200442 pr_info("DSA: tree %d unapplied\n", dst->tree);
443 dst->applied = false;
444}
445
446static int dsa_cpu_parse(struct device_node *port, u32 index,
447 struct dsa_switch_tree *dst,
448 struct dsa_switch *ds)
449{
Andrew Lunn7b314362016-08-22 16:01:01 +0200450 enum dsa_tag_protocol tag_protocol;
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200451 struct net_device *ethernet_dev;
452 struct device_node *ethernet;
453
454 ethernet = of_parse_phandle(port, "ethernet", 0);
455 if (!ethernet)
456 return -EINVAL;
457
458 ethernet_dev = of_find_net_device_by_node(ethernet);
459 if (!ethernet_dev)
460 return -EPROBE_DEFER;
461
462 if (!ds->master_netdev)
463 ds->master_netdev = ethernet_dev;
464
465 if (!dst->master_netdev)
466 dst->master_netdev = ethernet_dev;
467
468 if (dst->cpu_switch == -1) {
469 dst->cpu_switch = ds->index;
470 dst->cpu_port = index;
471 }
472
Vivien Didelot9d490b42016-08-23 12:38:56 -0400473 tag_protocol = ds->ops->get_tag_protocol(ds);
Andrew Lunn7b314362016-08-22 16:01:01 +0200474 dst->tag_ops = dsa_resolve_tag_protocol(tag_protocol);
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200475 if (IS_ERR(dst->tag_ops)) {
476 dev_warn(ds->dev, "No tagger for this switch\n");
477 return PTR_ERR(dst->tag_ops);
478 }
479
480 dst->rcv = dst->tag_ops->rcv;
481
482 return 0;
483}
484
485static int dsa_ds_parse(struct dsa_switch_tree *dst, struct dsa_switch *ds)
486{
487 struct device_node *port;
488 u32 index;
489 int err;
490
491 for (index = 0; index < DSA_MAX_PORTS; index++) {
492 port = ds->ports[index].dn;
493 if (!port)
494 continue;
495
496 if (dsa_port_is_cpu(port)) {
497 err = dsa_cpu_parse(port, index, dst, ds);
498 if (err)
499 return err;
500 }
501 }
502
503 pr_info("DSA: switch %d %d parsed\n", dst->tree, ds->index);
504
505 return 0;
506}
507
508static int dsa_dst_parse(struct dsa_switch_tree *dst)
509{
510 struct dsa_switch *ds;
511 u32 index;
512 int err;
513
514 for (index = 0; index < DSA_MAX_SWITCHES; index++) {
515 ds = dst->ds[index];
516 if (!ds)
517 continue;
518
519 err = dsa_ds_parse(dst, ds);
520 if (err)
521 return err;
522 }
523
524 if (!dst->master_netdev) {
525 pr_warn("Tree has no master device\n");
526 return -EINVAL;
527 }
528
529 pr_info("DSA: tree %d parsed\n", dst->tree);
530
531 return 0;
532}
533
534static int dsa_parse_ports_dn(struct device_node *ports, struct dsa_switch *ds)
535{
536 struct device_node *port;
537 int err;
538 u32 reg;
539
540 for_each_available_child_of_node(ports, port) {
541 err = of_property_read_u32(port, "reg", &reg);
542 if (err)
543 return err;
544
545 if (reg >= DSA_MAX_PORTS)
546 return -EINVAL;
547
548 ds->ports[reg].dn = port;
Florian Fainelli6e830d82016-06-07 16:32:39 -0700549
Vivien Didelot9d490b42016-08-23 12:38:56 -0400550 /* Initialize enabled_port_mask now for ops->setup()
Florian Fainelli6e830d82016-06-07 16:32:39 -0700551 * to have access to a correct value, just like what
552 * net/dsa/dsa.c::dsa_switch_setup_one does.
553 */
554 if (!dsa_port_is_cpu(port))
555 ds->enabled_port_mask |= 1 << reg;
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200556 }
557
558 return 0;
559}
560
561static int dsa_parse_member(struct device_node *np, u32 *tree, u32 *index)
562{
563 int err;
564
565 *tree = *index = 0;
566
567 err = of_property_read_u32_index(np, "dsa,member", 0, tree);
568 if (err) {
569 /* Does not exist, but it is optional */
570 if (err == -EINVAL)
571 return 0;
572 return err;
573 }
574
575 err = of_property_read_u32_index(np, "dsa,member", 1, index);
576 if (err)
577 return err;
578
579 if (*index >= DSA_MAX_SWITCHES)
580 return -EINVAL;
581
582 return 0;
583}
584
585static struct device_node *dsa_get_ports(struct dsa_switch *ds,
586 struct device_node *np)
587{
588 struct device_node *ports;
589
590 ports = of_get_child_by_name(np, "ports");
591 if (!ports) {
592 dev_err(ds->dev, "no ports child node found\n");
593 return ERR_PTR(-EINVAL);
594 }
595
596 return ports;
597}
598
599static int _dsa_register_switch(struct dsa_switch *ds, struct device_node *np)
600{
601 struct device_node *ports = dsa_get_ports(ds, np);
602 struct dsa_switch_tree *dst;
603 u32 tree, index;
Vivien Didelotd3902382016-07-06 20:03:54 -0400604 int i, err;
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200605
606 err = dsa_parse_member(np, &tree, &index);
607 if (err)
608 return err;
609
610 if (IS_ERR(ports))
611 return PTR_ERR(ports);
612
613 err = dsa_parse_ports_dn(ports, ds);
614 if (err)
615 return err;
616
617 dst = dsa_get_dst(tree);
618 if (!dst) {
619 dst = dsa_add_dst(tree);
620 if (!dst)
621 return -ENOMEM;
622 }
623
624 if (dst->ds[index]) {
625 err = -EBUSY;
626 goto out;
627 }
628
629 ds->dst = dst;
630 ds->index = index;
Vivien Didelotd3902382016-07-06 20:03:54 -0400631
632 /* Initialize the routing table */
633 for (i = 0; i < DSA_MAX_SWITCHES; ++i)
634 ds->rtable[i] = DSA_RTABLE_NONE;
635
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200636 dsa_dst_add_ds(dst, ds, index);
637
638 err = dsa_dst_complete(dst);
639 if (err < 0)
640 goto out_del_dst;
641
642 if (err == 1) {
643 /* Not all switches registered yet */
644 err = 0;
645 goto out;
646 }
647
648 if (dst->applied) {
649 pr_info("DSA: Disjoint trees?\n");
650 return -EINVAL;
651 }
652
653 err = dsa_dst_parse(dst);
654 if (err)
655 goto out_del_dst;
656
657 err = dsa_dst_apply(dst);
658 if (err) {
659 dsa_dst_unapply(dst);
660 goto out_del_dst;
661 }
662
663 dsa_put_dst(dst);
664 return 0;
665
666out_del_dst:
667 dsa_dst_del_ds(dst, ds, ds->index);
668out:
669 dsa_put_dst(dst);
670
671 return err;
672}
673
674int dsa_register_switch(struct dsa_switch *ds, struct device_node *np)
675{
676 int err;
677
678 mutex_lock(&dsa2_mutex);
679 err = _dsa_register_switch(ds, np);
680 mutex_unlock(&dsa2_mutex);
681
682 return err;
683}
684EXPORT_SYMBOL_GPL(dsa_register_switch);
685
Wei Yongjun85c22ba2016-07-12 15:24:10 +0000686static void _dsa_unregister_switch(struct dsa_switch *ds)
Andrew Lunn83c0afa2016-06-04 21:17:07 +0200687{
688 struct dsa_switch_tree *dst = ds->dst;
689
690 dsa_dst_unapply(dst);
691
692 dsa_dst_del_ds(dst, ds, ds->index);
693}
694
695void dsa_unregister_switch(struct dsa_switch *ds)
696{
697 mutex_lock(&dsa2_mutex);
698 _dsa_unregister_switch(ds);
699 mutex_unlock(&dsa2_mutex);
700}
701EXPORT_SYMBOL_GPL(dsa_unregister_switch);