blob: 9af16cb68f76b92bd929d4d941a5e8c9a04d779e [file] [log] [blame]
Arvid Brodin70ebe4a2014-07-04 23:34:38 +02001/* Copyright 2011-2014 Autronica Fire and Security AS
Arvid Brodinf4214362013-10-30 21:10:47 +01002 *
3 * This program is free software; you can redistribute it and/or modify it
4 * under the terms of the GNU General Public License as published by the Free
5 * Software Foundation; either version 2 of the License, or (at your option)
6 * any later version.
7 *
8 * Author(s):
Arvid Brodin70ebe4a2014-07-04 23:34:38 +02009 * 2011-2014 Arvid Brodin, arvid.brodin@alten.se
Arvid Brodinf4214362013-10-30 21:10:47 +010010 *
11 * The HSR spec says never to forward the same frame twice on the same
12 * interface. A frame is identified by its source MAC address and its HSR
13 * sequence number. This code keeps track of senders and their sequence numbers
14 * to allow filtering of duplicate frames, and to detect HSR ring errors.
15 */
16
17#include <linux/if_ether.h>
18#include <linux/etherdevice.h>
19#include <linux/slab.h>
20#include <linux/rculist.h>
21#include "hsr_main.h"
22#include "hsr_framereg.h"
23#include "hsr_netlink.h"
24
25
Arvid Brodin70ebe4a2014-07-04 23:34:38 +020026struct hsr_node {
27 struct list_head mac_list;
28 unsigned char MacAddressA[ETH_ALEN];
29 unsigned char MacAddressB[ETH_ALEN];
Arvid Brodinc5a75912014-07-04 23:38:05 +020030 /* Local slave through which AddrB frames are received from this node */
31 enum hsr_port_type AddrB_port;
32 unsigned long time_in[HSR_PT_PORTS];
33 bool time_in_stale[HSR_PT_PORTS];
34 u16 seq_out[HSR_PT_PORTS];
Arvid Brodin70ebe4a2014-07-04 23:34:38 +020035 struct rcu_head rcu_head;
Arvid Brodinf4214362013-10-30 21:10:47 +010036};
37
Arvid Brodinf266a682014-07-04 23:41:03 +020038
Arvid Brodinf4214362013-10-30 21:10:47 +010039/* TODO: use hash lists for mac addresses (linux/jhash.h)? */
40
41
Arvid Brodinf266a682014-07-04 23:41:03 +020042/* seq_nr_after(a, b) - return true if a is after (higher in sequence than) b,
43 * false otherwise.
44 */
45static bool seq_nr_after(u16 a, u16 b)
46{
47 /* Remove inconsistency where
48 * seq_nr_after(a, b) == seq_nr_before(a, b)
49 */
50 if ((int) b - a == 32768)
51 return false;
52
53 return (((s16) (b - a)) < 0);
54}
55#define seq_nr_before(a, b) seq_nr_after((b), (a))
56#define seq_nr_after_or_eq(a, b) (!seq_nr_before((a), (b)))
57#define seq_nr_before_or_eq(a, b) (!seq_nr_after((a), (b)))
58
59
60bool hsr_addr_is_self(struct hsr_priv *hsr, unsigned char *addr)
61{
62 struct hsr_node *node;
63
64 node = list_first_or_null_rcu(&hsr->self_node_db, struct hsr_node,
65 mac_list);
66 if (!node) {
67 WARN_ONCE(1, "HSR: No self node\n");
68 return false;
69 }
70
71 if (ether_addr_equal(addr, node->MacAddressA))
72 return true;
73 if (ether_addr_equal(addr, node->MacAddressB))
74 return true;
75
76 return false;
77}
Arvid Brodinf4214362013-10-30 21:10:47 +010078
79/* Search for mac entry. Caller must hold rcu read lock.
80 */
Arvid Brodin70ebe4a2014-07-04 23:34:38 +020081static struct hsr_node *find_node_by_AddrA(struct list_head *node_db,
82 const unsigned char addr[ETH_ALEN])
Arvid Brodinf4214362013-10-30 21:10:47 +010083{
Arvid Brodin70ebe4a2014-07-04 23:34:38 +020084 struct hsr_node *node;
Arvid Brodinf4214362013-10-30 21:10:47 +010085
86 list_for_each_entry_rcu(node, node_db, mac_list) {
87 if (ether_addr_equal(node->MacAddressA, addr))
88 return node;
89 }
90
91 return NULL;
92}
93
94
Arvid Brodinf4214362013-10-30 21:10:47 +010095/* Helper for device init; the self_node_db is used in hsr_rcv() to recognize
96 * frames from self that's been looped over the HSR ring.
97 */
98int hsr_create_self_node(struct list_head *self_node_db,
99 unsigned char addr_a[ETH_ALEN],
100 unsigned char addr_b[ETH_ALEN])
101{
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200102 struct hsr_node *node, *oldnode;
Arvid Brodinf4214362013-10-30 21:10:47 +0100103
104 node = kmalloc(sizeof(*node), GFP_KERNEL);
105 if (!node)
106 return -ENOMEM;
107
Joe Perchese83abe32014-02-18 10:37:20 -0800108 ether_addr_copy(node->MacAddressA, addr_a);
109 ether_addr_copy(node->MacAddressB, addr_b);
Arvid Brodinf4214362013-10-30 21:10:47 +0100110
111 rcu_read_lock();
112 oldnode = list_first_or_null_rcu(self_node_db,
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200113 struct hsr_node, mac_list);
Arvid Brodinf4214362013-10-30 21:10:47 +0100114 if (oldnode) {
115 list_replace_rcu(&oldnode->mac_list, &node->mac_list);
116 rcu_read_unlock();
117 synchronize_rcu();
118 kfree(oldnode);
119 } else {
120 rcu_read_unlock();
121 list_add_tail_rcu(&node->mac_list, self_node_db);
122 }
123
124 return 0;
125}
126
Mao Wenan996ee1a2019-03-06 22:45:01 +0800127void hsr_del_node(struct list_head *self_node_db)
128{
129 struct hsr_node *node;
130
131 rcu_read_lock();
132 node = list_first_or_null_rcu(self_node_db, struct hsr_node, mac_list);
133 rcu_read_unlock();
134 if (node) {
135 list_del_rcu(&node->mac_list);
136 kfree(node);
137 }
138}
Arvid Brodinf4214362013-10-30 21:10:47 +0100139
Arvid Brodinf266a682014-07-04 23:41:03 +0200140/* Allocate an hsr_node and add it to node_db. 'addr' is the node's AddressA;
141 * seq_out is used to initialize filtering of outgoing duplicate frames
142 * originating from the newly added node.
Arvid Brodinf4214362013-10-30 21:10:47 +0100143 */
Arvid Brodinf266a682014-07-04 23:41:03 +0200144struct hsr_node *hsr_add_node(struct list_head *node_db, unsigned char addr[],
145 u16 seq_out)
Arvid Brodinf4214362013-10-30 21:10:47 +0100146{
Arvid Brodinf266a682014-07-04 23:41:03 +0200147 struct hsr_node *node;
Arvid Brodinf4214362013-10-30 21:10:47 +0100148 unsigned long now;
Arvid Brodinf266a682014-07-04 23:41:03 +0200149 int i;
Arvid Brodinf4214362013-10-30 21:10:47 +0100150
151 node = kzalloc(sizeof(*node), GFP_ATOMIC);
152 if (!node)
153 return NULL;
154
Arvid Brodinf266a682014-07-04 23:41:03 +0200155 ether_addr_copy(node->MacAddressA, addr);
Arvid Brodinf4214362013-10-30 21:10:47 +0100156
157 /* We are only interested in time diffs here, so use current jiffies
158 * as initialization. (0 could trigger an spurious ring error warning).
159 */
160 now = jiffies;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200161 for (i = 0; i < HSR_PT_PORTS; i++)
Arvid Brodinf4214362013-10-30 21:10:47 +0100162 node->time_in[i] = now;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200163 for (i = 0; i < HSR_PT_PORTS; i++)
Arvid Brodinf266a682014-07-04 23:41:03 +0200164 node->seq_out[i] = seq_out;
Arvid Brodinf4214362013-10-30 21:10:47 +0100165
Arvid Brodinf266a682014-07-04 23:41:03 +0200166 list_add_tail_rcu(&node->mac_list, node_db);
Arvid Brodinf4214362013-10-30 21:10:47 +0100167
168 return node;
169}
170
Arvid Brodinf266a682014-07-04 23:41:03 +0200171/* Get the hsr_node from which 'skb' was sent.
172 */
Karicheri, Muralidharan675c8da2017-06-12 15:06:26 -0400173struct hsr_node *hsr_get_node(struct hsr_port *port, struct sk_buff *skb,
Arvid Brodinf266a682014-07-04 23:41:03 +0200174 bool is_sup)
175{
Karicheri, Muralidharan675c8da2017-06-12 15:06:26 -0400176 struct list_head *node_db = &port->hsr->node_db;
Arvid Brodinf266a682014-07-04 23:41:03 +0200177 struct hsr_node *node;
178 struct ethhdr *ethhdr;
179 u16 seq_out;
180
181 if (!skb_mac_header_was_set(skb))
182 return NULL;
183
184 ethhdr = (struct ethhdr *) skb_mac_header(skb);
185
186 list_for_each_entry_rcu(node, node_db, mac_list) {
187 if (ether_addr_equal(node->MacAddressA, ethhdr->h_source))
188 return node;
189 if (ether_addr_equal(node->MacAddressB, ethhdr->h_source))
190 return node;
191 }
192
Peter Heiseee1c2792016-04-13 13:52:22 +0200193 /* Everyone may create a node entry, connected node to a HSR device. */
Arvid Brodinf266a682014-07-04 23:41:03 +0200194
Peter Heiseee1c2792016-04-13 13:52:22 +0200195 if (ethhdr->h_proto == htons(ETH_P_PRP)
196 || ethhdr->h_proto == htons(ETH_P_HSR)) {
Arvid Brodinf266a682014-07-04 23:41:03 +0200197 /* Use the existing sequence_nr from the tag as starting point
198 * for filtering duplicate frames.
199 */
200 seq_out = hsr_get_skb_sequence_nr(skb) - 1;
201 } else {
Karicheri, Muralidharan675c8da2017-06-12 15:06:26 -0400202 /* this is called also for frames from master port and
203 * so warn only for non master ports
204 */
205 if (port->type != HSR_PT_MASTER)
206 WARN_ONCE(1, "%s: Non-HSR frame\n", __func__);
Peter Heiseee1c2792016-04-13 13:52:22 +0200207 seq_out = HSR_SEQNR_START;
Arvid Brodinf266a682014-07-04 23:41:03 +0200208 }
209
210 return hsr_add_node(node_db, ethhdr->h_source, seq_out);
211}
212
213/* Use the Supervision frame's info about an eventual MacAddressB for merging
214 * nodes that has previously had their MacAddressB registered as a separate
215 * node.
216 */
217void hsr_handle_sup_frame(struct sk_buff *skb, struct hsr_node *node_curr,
218 struct hsr_port *port_rcv)
219{
Peter Heiseee1c2792016-04-13 13:52:22 +0200220 struct ethhdr *ethhdr;
Arvid Brodinf266a682014-07-04 23:41:03 +0200221 struct hsr_node *node_real;
222 struct hsr_sup_payload *hsr_sp;
223 struct list_head *node_db;
224 int i;
225
Peter Heiseee1c2792016-04-13 13:52:22 +0200226 ethhdr = (struct ethhdr *) skb_mac_header(skb);
Arvid Brodinf266a682014-07-04 23:41:03 +0200227
Peter Heiseee1c2792016-04-13 13:52:22 +0200228 /* Leave the ethernet header. */
229 skb_pull(skb, sizeof(struct ethhdr));
230
231 /* And leave the HSR tag. */
232 if (ethhdr->h_proto == htons(ETH_P_HSR))
233 skb_pull(skb, sizeof(struct hsr_tag));
234
235 /* And leave the HSR sup tag. */
236 skb_pull(skb, sizeof(struct hsr_sup_tag));
237
238 hsr_sp = (struct hsr_sup_payload *) skb->data;
Arvid Brodinf266a682014-07-04 23:41:03 +0200239
240 /* Merge node_curr (registered on MacAddressB) into node_real */
241 node_db = &port_rcv->hsr->node_db;
242 node_real = find_node_by_AddrA(node_db, hsr_sp->MacAddressA);
243 if (!node_real)
244 /* No frame received from AddrA of this node yet */
245 node_real = hsr_add_node(node_db, hsr_sp->MacAddressA,
246 HSR_SEQNR_START - 1);
247 if (!node_real)
248 goto done; /* No mem */
249 if (node_real == node_curr)
250 /* Node has already been merged */
251 goto done;
252
Peter Heiseee1c2792016-04-13 13:52:22 +0200253 ether_addr_copy(node_real->MacAddressB, ethhdr->h_source);
Arvid Brodinf266a682014-07-04 23:41:03 +0200254 for (i = 0; i < HSR_PT_PORTS; i++) {
255 if (!node_curr->time_in_stale[i] &&
256 time_after(node_curr->time_in[i], node_real->time_in[i])) {
257 node_real->time_in[i] = node_curr->time_in[i];
258 node_real->time_in_stale[i] = node_curr->time_in_stale[i];
259 }
260 if (seq_nr_after(node_curr->seq_out[i], node_real->seq_out[i]))
261 node_real->seq_out[i] = node_curr->seq_out[i];
262 }
263 node_real->AddrB_port = port_rcv->type;
264
265 list_del_rcu(&node_curr->mac_list);
266 kfree_rcu(node_curr, rcu_head);
267
268done:
Peter Heiseee1c2792016-04-13 13:52:22 +0200269 skb_push(skb, sizeof(struct hsrv1_ethhdr_sp));
Arvid Brodinf266a682014-07-04 23:41:03 +0200270}
271
Arvid Brodinf4214362013-10-30 21:10:47 +0100272
273/* 'skb' is a frame meant for this host, that is to be passed to upper layers.
274 *
Arvid Brodinf266a682014-07-04 23:41:03 +0200275 * If the frame was sent by a node's B interface, replace the source
Arvid Brodinf4214362013-10-30 21:10:47 +0100276 * address with that node's "official" address (MacAddressA) so that upper
277 * layers recognize where it came from.
278 */
Arvid Brodinf266a682014-07-04 23:41:03 +0200279void hsr_addr_subst_source(struct hsr_node *node, struct sk_buff *skb)
Arvid Brodinf4214362013-10-30 21:10:47 +0100280{
Arvid Brodinf4214362013-10-30 21:10:47 +0100281 if (!skb_mac_header_was_set(skb)) {
282 WARN_ONCE(1, "%s: Mac header not set\n", __func__);
283 return;
284 }
Arvid Brodinf4214362013-10-30 21:10:47 +0100285
Arvid Brodinf266a682014-07-04 23:41:03 +0200286 memcpy(&eth_hdr(skb)->h_source, node->MacAddressA, ETH_ALEN);
Arvid Brodinf4214362013-10-30 21:10:47 +0100287}
288
Arvid Brodinf4214362013-10-30 21:10:47 +0100289/* 'skb' is a frame meant for another host.
Arvid Brodinf266a682014-07-04 23:41:03 +0200290 * 'port' is the outgoing interface
Arvid Brodinf4214362013-10-30 21:10:47 +0100291 *
292 * Substitute the target (dest) MAC address if necessary, so the it matches the
293 * recipient interface MAC address, regardless of whether that is the
294 * recipient's A or B interface.
295 * This is needed to keep the packets flowing through switches that learn on
296 * which "side" the different interfaces are.
297 */
Arvid Brodinf266a682014-07-04 23:41:03 +0200298void hsr_addr_subst_dest(struct hsr_node *node_src, struct sk_buff *skb,
Arvid Brodinc5a75912014-07-04 23:38:05 +0200299 struct hsr_port *port)
Arvid Brodinf4214362013-10-30 21:10:47 +0100300{
Arvid Brodinf266a682014-07-04 23:41:03 +0200301 struct hsr_node *node_dst;
Arvid Brodinf4214362013-10-30 21:10:47 +0100302
Arvid Brodinf266a682014-07-04 23:41:03 +0200303 if (!skb_mac_header_was_set(skb)) {
304 WARN_ONCE(1, "%s: Mac header not set\n", __func__);
305 return;
306 }
307
308 if (!is_unicast_ether_addr(eth_hdr(skb)->h_dest))
309 return;
310
311 node_dst = find_node_by_AddrA(&port->hsr->node_db, eth_hdr(skb)->h_dest);
312 if (!node_dst) {
313 WARN_ONCE(1, "%s: Unknown node\n", __func__);
314 return;
315 }
316 if (port->type != node_dst->AddrB_port)
317 return;
Arvid Brodinf266a682014-07-04 23:41:03 +0200318
319 ether_addr_copy(eth_hdr(skb)->h_dest, node_dst->MacAddressB);
Arvid Brodinf4214362013-10-30 21:10:47 +0100320}
321
322
Arvid Brodinf266a682014-07-04 23:41:03 +0200323void hsr_register_frame_in(struct hsr_node *node, struct hsr_port *port,
324 u16 sequence_nr)
Arvid Brodinf4214362013-10-30 21:10:47 +0100325{
Arvid Brodinf266a682014-07-04 23:41:03 +0200326 /* Don't register incoming frames without a valid sequence number. This
327 * ensures entries of restarted nodes gets pruned so that they can
328 * re-register and resume communications.
Arvid Brodin213e3bc2013-11-29 23:37:07 +0100329 */
Arvid Brodinf266a682014-07-04 23:41:03 +0200330 if (seq_nr_before(sequence_nr, node->seq_out[port->type]))
331 return;
Arvid Brodinf4214362013-10-30 21:10:47 +0100332
Arvid Brodinc5a75912014-07-04 23:38:05 +0200333 node->time_in[port->type] = jiffies;
334 node->time_in_stale[port->type] = false;
Arvid Brodinf4214362013-10-30 21:10:47 +0100335}
336
Arvid Brodinf4214362013-10-30 21:10:47 +0100337/* 'skb' is a HSR Ethernet frame (with a HSR tag inserted), with a valid
338 * ethhdr->h_source address and skb->mac_header set.
339 *
340 * Return:
341 * 1 if frame can be shown to have been sent recently on this interface,
342 * 0 otherwise, or
343 * negative error code on error
344 */
Arvid Brodinf266a682014-07-04 23:41:03 +0200345int hsr_register_frame_out(struct hsr_port *port, struct hsr_node *node,
346 u16 sequence_nr)
Arvid Brodinf4214362013-10-30 21:10:47 +0100347{
Arvid Brodinc5a75912014-07-04 23:38:05 +0200348 if (seq_nr_before_or_eq(sequence_nr, node->seq_out[port->type]))
Arvid Brodinf4214362013-10-30 21:10:47 +0100349 return 1;
350
Arvid Brodinc5a75912014-07-04 23:38:05 +0200351 node->seq_out[port->type] = sequence_nr;
Arvid Brodinf4214362013-10-30 21:10:47 +0100352 return 0;
353}
354
355
Arvid Brodinc5a75912014-07-04 23:38:05 +0200356static struct hsr_port *get_late_port(struct hsr_priv *hsr,
357 struct hsr_node *node)
Arvid Brodinf4214362013-10-30 21:10:47 +0100358{
Arvid Brodinc5a75912014-07-04 23:38:05 +0200359 if (node->time_in_stale[HSR_PT_SLAVE_A])
360 return hsr_port_get_hsr(hsr, HSR_PT_SLAVE_A);
361 if (node->time_in_stale[HSR_PT_SLAVE_B])
362 return hsr_port_get_hsr(hsr, HSR_PT_SLAVE_B);
Arvid Brodinf4214362013-10-30 21:10:47 +0100363
Arvid Brodinc5a75912014-07-04 23:38:05 +0200364 if (time_after(node->time_in[HSR_PT_SLAVE_B],
365 node->time_in[HSR_PT_SLAVE_A] +
366 msecs_to_jiffies(MAX_SLAVE_DIFF)))
367 return hsr_port_get_hsr(hsr, HSR_PT_SLAVE_A);
368 if (time_after(node->time_in[HSR_PT_SLAVE_A],
369 node->time_in[HSR_PT_SLAVE_B] +
370 msecs_to_jiffies(MAX_SLAVE_DIFF)))
371 return hsr_port_get_hsr(hsr, HSR_PT_SLAVE_B);
Arvid Brodinf4214362013-10-30 21:10:47 +0100372
Arvid Brodinc5a75912014-07-04 23:38:05 +0200373 return NULL;
Arvid Brodinf4214362013-10-30 21:10:47 +0100374}
375
376
377/* Remove stale sequence_nr records. Called by timer every
378 * HSR_LIFE_CHECK_INTERVAL (two seconds or so).
379 */
Kees Cookdda436b2017-10-24 01:46:16 -0700380void hsr_prune_nodes(struct timer_list *t)
Arvid Brodinf4214362013-10-30 21:10:47 +0100381{
Kees Cookdda436b2017-10-24 01:46:16 -0700382 struct hsr_priv *hsr = from_timer(hsr, t, prune_timer);
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200383 struct hsr_node *node;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200384 struct hsr_port *port;
Arvid Brodinf4214362013-10-30 21:10:47 +0100385 unsigned long timestamp;
386 unsigned long time_a, time_b;
387
388 rcu_read_lock();
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200389 list_for_each_entry_rcu(node, &hsr->node_db, mac_list) {
Arvid Brodinf4214362013-10-30 21:10:47 +0100390 /* Shorthand */
Arvid Brodinc5a75912014-07-04 23:38:05 +0200391 time_a = node->time_in[HSR_PT_SLAVE_A];
392 time_b = node->time_in[HSR_PT_SLAVE_B];
Arvid Brodinf4214362013-10-30 21:10:47 +0100393
394 /* Check for timestamps old enough to risk wrap-around */
395 if (time_after(jiffies, time_a + MAX_JIFFY_OFFSET/2))
Arvid Brodinc5a75912014-07-04 23:38:05 +0200396 node->time_in_stale[HSR_PT_SLAVE_A] = true;
Arvid Brodinf4214362013-10-30 21:10:47 +0100397 if (time_after(jiffies, time_b + MAX_JIFFY_OFFSET/2))
Arvid Brodinc5a75912014-07-04 23:38:05 +0200398 node->time_in_stale[HSR_PT_SLAVE_B] = true;
Arvid Brodinf4214362013-10-30 21:10:47 +0100399
400 /* Get age of newest frame from node.
401 * At least one time_in is OK here; nodes get pruned long
402 * before both time_ins can get stale
403 */
404 timestamp = time_a;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200405 if (node->time_in_stale[HSR_PT_SLAVE_A] ||
406 (!node->time_in_stale[HSR_PT_SLAVE_B] &&
Arvid Brodinf4214362013-10-30 21:10:47 +0100407 time_after(time_b, time_a)))
408 timestamp = time_b;
409
410 /* Warn of ring error only as long as we get frames at all */
411 if (time_is_after_jiffies(timestamp +
412 msecs_to_jiffies(1.5*MAX_SLAVE_DIFF))) {
Arvid Brodinc5a75912014-07-04 23:38:05 +0200413 rcu_read_lock();
414 port = get_late_port(hsr, node);
415 if (port != NULL)
416 hsr_nl_ringerror(hsr, node->MacAddressA, port);
417 rcu_read_unlock();
Arvid Brodinf4214362013-10-30 21:10:47 +0100418 }
419
420 /* Prune old entries */
421 if (time_is_before_jiffies(timestamp +
422 msecs_to_jiffies(HSR_NODE_FORGET_TIME))) {
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200423 hsr_nl_nodedown(hsr, node->MacAddressA);
Arvid Brodinf4214362013-10-30 21:10:47 +0100424 list_del_rcu(&node->mac_list);
425 /* Note that we need to free this entry later: */
Wei Yongjun1aee6cc2013-12-16 14:05:50 +0800426 kfree_rcu(node, rcu_head);
Arvid Brodinf4214362013-10-30 21:10:47 +0100427 }
428 }
429 rcu_read_unlock();
430}
431
432
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200433void *hsr_get_next_node(struct hsr_priv *hsr, void *_pos,
Arvid Brodinf4214362013-10-30 21:10:47 +0100434 unsigned char addr[ETH_ALEN])
435{
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200436 struct hsr_node *node;
Arvid Brodinf4214362013-10-30 21:10:47 +0100437
438 if (!_pos) {
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200439 node = list_first_or_null_rcu(&hsr->node_db,
440 struct hsr_node, mac_list);
Arvid Brodinf4214362013-10-30 21:10:47 +0100441 if (node)
Joe Perchese83abe32014-02-18 10:37:20 -0800442 ether_addr_copy(addr, node->MacAddressA);
Arvid Brodinf4214362013-10-30 21:10:47 +0100443 return node;
444 }
445
446 node = _pos;
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200447 list_for_each_entry_continue_rcu(node, &hsr->node_db, mac_list) {
Joe Perchese83abe32014-02-18 10:37:20 -0800448 ether_addr_copy(addr, node->MacAddressA);
Arvid Brodinf4214362013-10-30 21:10:47 +0100449 return node;
450 }
451
452 return NULL;
453}
454
455
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200456int hsr_get_node_data(struct hsr_priv *hsr,
Arvid Brodinf4214362013-10-30 21:10:47 +0100457 const unsigned char *addr,
458 unsigned char addr_b[ETH_ALEN],
459 unsigned int *addr_b_ifindex,
460 int *if1_age,
461 u16 *if1_seq,
462 int *if2_age,
463 u16 *if2_seq)
464{
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200465 struct hsr_node *node;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200466 struct hsr_port *port;
Arvid Brodinf4214362013-10-30 21:10:47 +0100467 unsigned long tdiff;
468
469
470 rcu_read_lock();
Arvid Brodin70ebe4a2014-07-04 23:34:38 +0200471 node = find_node_by_AddrA(&hsr->node_db, addr);
Arvid Brodinf4214362013-10-30 21:10:47 +0100472 if (!node) {
473 rcu_read_unlock();
474 return -ENOENT; /* No such entry */
475 }
476
Joe Perchese83abe32014-02-18 10:37:20 -0800477 ether_addr_copy(addr_b, node->MacAddressB);
Arvid Brodinf4214362013-10-30 21:10:47 +0100478
Arvid Brodinc5a75912014-07-04 23:38:05 +0200479 tdiff = jiffies - node->time_in[HSR_PT_SLAVE_A];
480 if (node->time_in_stale[HSR_PT_SLAVE_A])
Arvid Brodinf4214362013-10-30 21:10:47 +0100481 *if1_age = INT_MAX;
482#if HZ <= MSEC_PER_SEC
483 else if (tdiff > msecs_to_jiffies(INT_MAX))
484 *if1_age = INT_MAX;
485#endif
486 else
487 *if1_age = jiffies_to_msecs(tdiff);
488
Arvid Brodinc5a75912014-07-04 23:38:05 +0200489 tdiff = jiffies - node->time_in[HSR_PT_SLAVE_B];
490 if (node->time_in_stale[HSR_PT_SLAVE_B])
Arvid Brodinf4214362013-10-30 21:10:47 +0100491 *if2_age = INT_MAX;
492#if HZ <= MSEC_PER_SEC
493 else if (tdiff > msecs_to_jiffies(INT_MAX))
494 *if2_age = INT_MAX;
495#endif
496 else
497 *if2_age = jiffies_to_msecs(tdiff);
498
499 /* Present sequence numbers as if they were incoming on interface */
Arvid Brodinc5a75912014-07-04 23:38:05 +0200500 *if1_seq = node->seq_out[HSR_PT_SLAVE_B];
501 *if2_seq = node->seq_out[HSR_PT_SLAVE_A];
Arvid Brodinf4214362013-10-30 21:10:47 +0100502
Arvid Brodinc5a75912014-07-04 23:38:05 +0200503 if (node->AddrB_port != HSR_PT_NONE) {
504 port = hsr_port_get_hsr(hsr, node->AddrB_port);
505 *addr_b_ifindex = port->dev->ifindex;
506 } else {
Arvid Brodinf4214362013-10-30 21:10:47 +0100507 *addr_b_ifindex = -1;
Arvid Brodinc5a75912014-07-04 23:38:05 +0200508 }
Arvid Brodinf4214362013-10-30 21:10:47 +0100509
510 rcu_read_unlock();
511
512 return 0;
513}