blob: 58e98c89a68f94ac0cb7c6660761fec70fe84e31 [file] [log] [blame]
Simon Wunderliche19f9752014-01-04 18:04:25 +01001/* Copyright (C) 2007-2014 B.A.T.M.A.N. contributors:
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +00002 *
3 * Marek Lindner, Simon Wunderlich
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of version 2 of the GNU General Public
7 * License as published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but
10 * WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
12 * General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
Antonio Quartulliebf38fb2013-11-03 20:40:48 +010015 * along with this program; if not, see <http://www.gnu.org/licenses/>.
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000016 */
17
Sven Eckelmann95a066d2012-10-17 21:10:39 +020018#include <linux/crc32c.h>
19#include <linux/highmem.h>
Simon Wunderlichc54f38c92013-07-29 17:56:44 +020020#include <linux/if_vlan.h>
21#include <net/ip.h>
22#include <net/ipv6.h>
23#include <net/dsfield.h>
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000024#include "main.h"
Sven Eckelmannb706b132012-06-10 23:58:51 +020025#include "sysfs.h"
26#include "debugfs.h"
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000027#include "routing.h"
28#include "send.h"
29#include "originator.h"
30#include "soft-interface.h"
31#include "icmp_socket.h"
32#include "translation-table.h"
33#include "hard-interface.h"
34#include "gateway_client.h"
Simon Wunderlich23721382012-01-22 20:00:19 +010035#include "bridge_loop_avoidance.h"
Antonio Quartulli2f1dfbe2012-06-30 20:01:19 +020036#include "distributed-arp-table.h"
Linus Lüssingc5caf4e2014-02-15 17:47:49 +010037#include "multicast.h"
Marek Lindner414254e2013-04-23 21:39:58 +080038#include "gateway_common.h"
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000039#include "hash.h"
Marek Lindner1c280472011-11-28 17:40:17 +080040#include "bat_algo.h"
Martin Hundebølld353d8d2013-01-25 11:12:38 +010041#include "network-coding.h"
Martin Hundebøll610bfc6bc2013-05-23 16:53:02 +020042#include "fragmentation.h"
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000043
Sven Eckelmannc3caf512011-05-03 11:51:38 +020044
45/* List manipulations on hardif_list have to be rtnl_lock()'ed,
Sven Eckelmann9cfc7bd2012-05-12 02:09:43 +020046 * list traversals just rcu-locked
47 */
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020048struct list_head batadv_hardif_list;
Sven Eckelmannee11ad62012-05-16 20:23:19 +020049static int (*batadv_rx_handler[256])(struct sk_buff *,
Sven Eckelmann56303d32012-06-05 22:31:31 +020050 struct batadv_hard_iface *);
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020051char batadv_routing_algo[20] = "BATMAN_IV";
Sven Eckelmannee11ad62012-05-16 20:23:19 +020052static struct hlist_head batadv_algo_list;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000053
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020054unsigned char batadv_broadcast_addr[] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff};
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000055
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020056struct workqueue_struct *batadv_event_workqueue;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000057
Sven Eckelmannee11ad62012-05-16 20:23:19 +020058static void batadv_recv_handler_init(void);
Marek Lindnerffa995e2012-03-01 15:35:17 +080059
Sven Eckelmannee11ad62012-05-16 20:23:19 +020060static int __init batadv_init(void)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000061{
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020062 INIT_LIST_HEAD(&batadv_hardif_list);
Sven Eckelmannee11ad62012-05-16 20:23:19 +020063 INIT_HLIST_HEAD(&batadv_algo_list);
Marek Lindner1c280472011-11-28 17:40:17 +080064
Sven Eckelmannee11ad62012-05-16 20:23:19 +020065 batadv_recv_handler_init();
Marek Lindnerffa995e2012-03-01 15:35:17 +080066
Sven Eckelmann81c524f2012-05-12 02:09:22 +020067 batadv_iv_init();
Matthias Schiffer6c519ba2013-09-27 18:03:39 +020068 batadv_nc_init();
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000069
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020070 batadv_event_workqueue = create_singlethread_workqueue("bat_events");
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000071
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020072 if (!batadv_event_workqueue)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000073 return -ENOMEM;
74
Sven Eckelmann9039dc72012-05-12 02:09:33 +020075 batadv_socket_init();
Sven Eckelmann40a072d2012-05-12 02:09:23 +020076 batadv_debugfs_init();
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000077
Sven Eckelmann95638772012-05-12 02:09:31 +020078 register_netdevice_notifier(&batadv_hard_if_notifier);
Sven Eckelmanna4ac28c2013-02-11 17:10:26 +080079 rtnl_link_register(&batadv_link_ops);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000080
Sven Eckelmann86ceb362012-03-07 09:07:45 +010081 pr_info("B.A.T.M.A.N. advanced %s (compatibility version %i) loaded\n",
Sven Eckelmann42d0b042012-06-03 22:19:17 +020082 BATADV_SOURCE_VERSION, BATADV_COMPAT_VERSION);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000083
84 return 0;
85}
86
Sven Eckelmannee11ad62012-05-16 20:23:19 +020087static void __exit batadv_exit(void)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000088{
Sven Eckelmann40a072d2012-05-12 02:09:23 +020089 batadv_debugfs_destroy();
Sven Eckelmanna4ac28c2013-02-11 17:10:26 +080090 rtnl_link_unregister(&batadv_link_ops);
Sven Eckelmann95638772012-05-12 02:09:31 +020091 unregister_netdevice_notifier(&batadv_hard_if_notifier);
92 batadv_hardif_remove_interfaces();
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000093
Sven Eckelmann3193e8f2012-05-12 02:09:42 +020094 flush_workqueue(batadv_event_workqueue);
95 destroy_workqueue(batadv_event_workqueue);
96 batadv_event_workqueue = NULL;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +000097
98 rcu_barrier();
99}
100
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200101int batadv_mesh_init(struct net_device *soft_iface)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000102{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200103 struct batadv_priv *bat_priv = netdev_priv(soft_iface);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200104 int ret;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000105
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000106 spin_lock_init(&bat_priv->forw_bat_list_lock);
107 spin_lock_init(&bat_priv->forw_bcast_list_lock);
Sven Eckelmann807736f2012-07-15 22:26:51 +0200108 spin_lock_init(&bat_priv->tt.changes_list_lock);
109 spin_lock_init(&bat_priv->tt.req_list_lock);
110 spin_lock_init(&bat_priv->tt.roam_list_lock);
111 spin_lock_init(&bat_priv->tt.last_changeset_lock);
Antonio Quartullia70a9aa2013-07-30 22:16:24 +0200112 spin_lock_init(&bat_priv->tt.commit_lock);
Sven Eckelmann807736f2012-07-15 22:26:51 +0200113 spin_lock_init(&bat_priv->gw.list_lock);
Marek Lindneref261572013-04-23 21:39:57 +0800114 spin_lock_init(&bat_priv->tvlv.container_list_lock);
115 spin_lock_init(&bat_priv->tvlv.handler_list_lock);
Antonio Quartulli5d2c05b2013-07-02 11:04:34 +0200116 spin_lock_init(&bat_priv->softif_vlan_list_lock);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000117
118 INIT_HLIST_HEAD(&bat_priv->forw_bat_list);
119 INIT_HLIST_HEAD(&bat_priv->forw_bcast_list);
Sven Eckelmann807736f2012-07-15 22:26:51 +0200120 INIT_HLIST_HEAD(&bat_priv->gw.list);
121 INIT_LIST_HEAD(&bat_priv->tt.changes_list);
122 INIT_LIST_HEAD(&bat_priv->tt.req_list);
123 INIT_LIST_HEAD(&bat_priv->tt.roam_list);
Linus Lüssingc5caf4e2014-02-15 17:47:49 +0100124#ifdef CONFIG_BATMAN_ADV_MCAST
125 INIT_HLIST_HEAD(&bat_priv->mcast.mla_list);
126#endif
Marek Lindneref261572013-04-23 21:39:57 +0800127 INIT_HLIST_HEAD(&bat_priv->tvlv.container_list);
128 INIT_HLIST_HEAD(&bat_priv->tvlv.handler_list);
Antonio Quartulli5d2c05b2013-07-02 11:04:34 +0200129 INIT_HLIST_HEAD(&bat_priv->softif_vlan_list);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000130
Sven Eckelmann7d211ef2012-05-12 02:09:34 +0200131 ret = batadv_originator_init(bat_priv);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200132 if (ret < 0)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000133 goto err;
134
Sven Eckelmann08c36d32012-05-12 02:09:39 +0200135 ret = batadv_tt_init(bat_priv);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200136 if (ret < 0)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000137 goto err;
138
Sven Eckelmann08adf152012-05-12 13:38:47 +0200139 ret = batadv_bla_init(bat_priv);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200140 if (ret < 0)
Simon Wunderlich23721382012-01-22 20:00:19 +0100141 goto err;
142
Antonio Quartulli2f1dfbe2012-06-30 20:01:19 +0200143 ret = batadv_dat_init(bat_priv);
144 if (ret < 0)
145 goto err;
146
Matthias Schiffer6c519ba2013-09-27 18:03:39 +0200147 ret = batadv_nc_mesh_init(bat_priv);
Martin Hundebølld353d8d2013-01-25 11:12:38 +0100148 if (ret < 0)
149 goto err;
150
Marek Lindner414254e2013-04-23 21:39:58 +0800151 batadv_gw_init(bat_priv);
152
Sven Eckelmann807736f2012-07-15 22:26:51 +0200153 atomic_set(&bat_priv->gw.reselect, 0);
Sven Eckelmann39c75a52012-06-03 22:19:22 +0200154 atomic_set(&bat_priv->mesh_state, BATADV_MESH_ACTIVE);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200155
156 return 0;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000157
158err:
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200159 batadv_mesh_free(soft_iface);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200160 return ret;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000161}
162
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200163void batadv_mesh_free(struct net_device *soft_iface)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000164{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200165 struct batadv_priv *bat_priv = netdev_priv(soft_iface);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000166
Sven Eckelmann39c75a52012-06-03 22:19:22 +0200167 atomic_set(&bat_priv->mesh_state, BATADV_MESH_DEACTIVATING);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000168
Sven Eckelmann9455e342012-05-12 02:09:37 +0200169 batadv_purge_outstanding_packets(bat_priv, NULL);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000170
Sven Eckelmann7cf06bc2012-05-12 02:09:29 +0200171 batadv_gw_node_purge(bat_priv);
Matthias Schiffer6c519ba2013-09-27 18:03:39 +0200172 batadv_nc_mesh_free(bat_priv);
Antonio Quartullia4361862013-05-07 01:06:18 +0200173 batadv_dat_free(bat_priv);
Sven Eckelmann08adf152012-05-12 13:38:47 +0200174 batadv_bla_free(bat_priv);
Simon Wunderlich23721382012-01-22 20:00:19 +0100175
Linus Lüssingc5caf4e2014-02-15 17:47:49 +0100176 batadv_mcast_free(bat_priv);
177
Antonio Quartullia4361862013-05-07 01:06:18 +0200178 /* Free the TT and the originator tables only after having terminated
179 * all the other depending components which may use these structures for
180 * their purposes.
181 */
182 batadv_tt_free(bat_priv);
183
184 /* Since the originator table clean up routine is accessing the TT
185 * tables as well, it has to be invoked after the TT tables have been
186 * freed and marked as empty. This ensures that no cleanup RCU callbacks
187 * accessing the TT data are scheduled for later execution.
188 */
189 batadv_originator_free(bat_priv);
Antonio Quartulli2f1dfbe2012-06-30 20:01:19 +0200190
Marek Lindner414254e2013-04-23 21:39:58 +0800191 batadv_gw_free(bat_priv);
192
Martin Hundebøllf8214862012-04-20 17:02:45 +0200193 free_percpu(bat_priv->bat_counters);
Martin Hundebøllf69ae772013-04-17 21:13:16 +0200194 bat_priv->bat_counters = NULL;
Martin Hundebøllf8214862012-04-20 17:02:45 +0200195
Sven Eckelmann39c75a52012-06-03 22:19:22 +0200196 atomic_set(&bat_priv->mesh_state, BATADV_MESH_INACTIVE);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000197}
198
David S. Miller6e0895c2013-04-22 20:32:51 -0400199/**
200 * batadv_is_my_mac - check if the given mac address belongs to any of the real
201 * interfaces in the current mesh
202 * @bat_priv: the bat priv with all the soft interface information
203 * @addr: the address to check
204 */
Antonio Quartullife8a93b2013-04-03 19:10:26 +0200205int batadv_is_my_mac(struct batadv_priv *bat_priv, const uint8_t *addr)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000206{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200207 const struct batadv_hard_iface *hard_iface;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000208
209 rcu_read_lock();
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200210 list_for_each_entry_rcu(hard_iface, &batadv_hardif_list, list) {
Sven Eckelmanne9a4f292012-06-03 22:19:19 +0200211 if (hard_iface->if_status != BATADV_IF_ACTIVE)
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000212 continue;
213
Antonio Quartullife8a93b2013-04-03 19:10:26 +0200214 if (hard_iface->soft_iface != bat_priv->soft_iface)
215 continue;
216
Sven Eckelmann1eda58b2012-05-12 13:48:58 +0200217 if (batadv_compare_eth(hard_iface->net_dev->dev_addr, addr)) {
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000218 rcu_read_unlock();
219 return 1;
220 }
221 }
222 rcu_read_unlock();
223 return 0;
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +0000224}
225
Marek Lindner30da63a2012-08-03 17:15:46 +0200226/**
227 * batadv_seq_print_text_primary_if_get - called from debugfs table printing
228 * function that requires the primary interface
229 * @seq: debugfs table seq_file struct
230 *
231 * Returns primary interface if found or NULL otherwise.
232 */
233struct batadv_hard_iface *
234batadv_seq_print_text_primary_if_get(struct seq_file *seq)
235{
236 struct net_device *net_dev = (struct net_device *)seq->private;
237 struct batadv_priv *bat_priv = netdev_priv(net_dev);
238 struct batadv_hard_iface *primary_if;
239
240 primary_if = batadv_primary_if_get_selected(bat_priv);
241
242 if (!primary_if) {
243 seq_printf(seq,
244 "BATMAN mesh %s disabled - please specify interfaces to enable it\n",
245 net_dev->name);
246 goto out;
247 }
248
249 if (primary_if->if_status == BATADV_IF_ACTIVE)
250 goto out;
251
252 seq_printf(seq,
253 "BATMAN mesh %s disabled - primary interface not active\n",
254 net_dev->name);
255 batadv_hardif_free_ref(primary_if);
256 primary_if = NULL;
257
258out:
259 return primary_if;
260}
261
Simon Wunderlichc54f38c92013-07-29 17:56:44 +0200262/**
Marek Lindner411d6ed2013-05-08 13:31:59 +0800263 * batadv_max_header_len - calculate maximum encapsulation overhead for a
264 * payload packet
265 *
266 * Return the maximum encapsulation overhead in bytes.
267 */
268int batadv_max_header_len(void)
269{
270 int header_len = 0;
271
272 header_len = max_t(int, header_len,
273 sizeof(struct batadv_unicast_packet));
274 header_len = max_t(int, header_len,
275 sizeof(struct batadv_unicast_4addr_packet));
276 header_len = max_t(int, header_len,
277 sizeof(struct batadv_bcast_packet));
278
279#ifdef CONFIG_BATMAN_ADV_NC
280 header_len = max_t(int, header_len,
281 sizeof(struct batadv_coded_packet));
282#endif
283
Marek Lindner1df0cbd2014-01-15 20:31:18 +0800284 return header_len + ETH_HLEN;
Marek Lindner411d6ed2013-05-08 13:31:59 +0800285}
286
287/**
Simon Wunderlichc54f38c92013-07-29 17:56:44 +0200288 * batadv_skb_set_priority - sets skb priority according to packet content
289 * @skb: the packet to be sent
290 * @offset: offset to the packet content
291 *
292 * This function sets a value between 256 and 263 (802.1d priority), which
293 * can be interpreted by the cfg80211 or other drivers.
294 */
295void batadv_skb_set_priority(struct sk_buff *skb, int offset)
296{
297 struct iphdr ip_hdr_tmp, *ip_hdr;
298 struct ipv6hdr ip6_hdr_tmp, *ip6_hdr;
299 struct ethhdr ethhdr_tmp, *ethhdr;
300 struct vlan_ethhdr *vhdr, vhdr_tmp;
301 u32 prio;
302
303 /* already set, do nothing */
304 if (skb->priority >= 256 && skb->priority <= 263)
305 return;
306
307 ethhdr = skb_header_pointer(skb, offset, sizeof(*ethhdr), &ethhdr_tmp);
308 if (!ethhdr)
309 return;
310
311 switch (ethhdr->h_proto) {
312 case htons(ETH_P_8021Q):
313 vhdr = skb_header_pointer(skb, offset + sizeof(*vhdr),
314 sizeof(*vhdr), &vhdr_tmp);
315 if (!vhdr)
316 return;
317 prio = ntohs(vhdr->h_vlan_TCI) & VLAN_PRIO_MASK;
318 prio = prio >> VLAN_PRIO_SHIFT;
319 break;
320 case htons(ETH_P_IP):
321 ip_hdr = skb_header_pointer(skb, offset + sizeof(*ethhdr),
322 sizeof(*ip_hdr), &ip_hdr_tmp);
323 if (!ip_hdr)
324 return;
325 prio = (ipv4_get_dsfield(ip_hdr) & 0xfc) >> 5;
326 break;
327 case htons(ETH_P_IPV6):
328 ip6_hdr = skb_header_pointer(skb, offset + sizeof(*ethhdr),
329 sizeof(*ip6_hdr), &ip6_hdr_tmp);
330 if (!ip6_hdr)
331 return;
332 prio = (ipv6_get_dsfield(ip6_hdr) & 0xfc) >> 5;
333 break;
334 default:
335 return;
336 }
337
338 skb->priority = prio + 256;
339}
340
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200341static int batadv_recv_unhandled_packet(struct sk_buff *skb,
Sven Eckelmann56303d32012-06-05 22:31:31 +0200342 struct batadv_hard_iface *recv_if)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800343{
344 return NET_RX_DROP;
345}
346
347/* incoming packets with the batman ethertype received on any active hard
348 * interface
349 */
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200350int batadv_batman_skb_recv(struct sk_buff *skb, struct net_device *dev,
351 struct packet_type *ptype,
352 struct net_device *orig_dev)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800353{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200354 struct batadv_priv *bat_priv;
Sven Eckelmann96412692012-06-05 22:31:30 +0200355 struct batadv_ogm_packet *batadv_ogm_packet;
Sven Eckelmann56303d32012-06-05 22:31:31 +0200356 struct batadv_hard_iface *hard_iface;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800357 uint8_t idx;
358 int ret;
359
Sven Eckelmann56303d32012-06-05 22:31:31 +0200360 hard_iface = container_of(ptype, struct batadv_hard_iface,
361 batman_adv_ptype);
Marek Lindnerffa995e2012-03-01 15:35:17 +0800362 skb = skb_share_check(skb, GFP_ATOMIC);
363
364 /* skb was released by skb_share_check() */
365 if (!skb)
366 goto err_out;
367
368 /* packet should hold at least type and version */
369 if (unlikely(!pskb_may_pull(skb, 2)))
370 goto err_free;
371
372 /* expect a valid ethernet header here. */
373 if (unlikely(skb->mac_len != ETH_HLEN || !skb_mac_header(skb)))
374 goto err_free;
375
376 if (!hard_iface->soft_iface)
377 goto err_free;
378
379 bat_priv = netdev_priv(hard_iface->soft_iface);
380
Sven Eckelmann39c75a52012-06-03 22:19:22 +0200381 if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800382 goto err_free;
383
384 /* discard frames on not active interfaces */
Sven Eckelmanne9a4f292012-06-03 22:19:19 +0200385 if (hard_iface->if_status != BATADV_IF_ACTIVE)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800386 goto err_free;
387
Sven Eckelmann96412692012-06-05 22:31:30 +0200388 batadv_ogm_packet = (struct batadv_ogm_packet *)skb->data;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800389
Simon Wunderlicha40d9b02013-12-02 20:38:31 +0100390 if (batadv_ogm_packet->version != BATADV_COMPAT_VERSION) {
Sven Eckelmann39c75a52012-06-03 22:19:22 +0200391 batadv_dbg(BATADV_DBG_BATMAN, bat_priv,
Sven Eckelmann1eda58b2012-05-12 13:48:58 +0200392 "Drop packet: incompatible batman version (%i)\n",
Simon Wunderlicha40d9b02013-12-02 20:38:31 +0100393 batadv_ogm_packet->version);
Marek Lindnerffa995e2012-03-01 15:35:17 +0800394 goto err_free;
395 }
396
397 /* all receive handlers return whether they received or reused
398 * the supplied skb. if not, we have to free the skb.
399 */
Simon Wunderlicha40d9b02013-12-02 20:38:31 +0100400 idx = batadv_ogm_packet->packet_type;
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200401 ret = (*batadv_rx_handler[idx])(skb, hard_iface);
Marek Lindnerffa995e2012-03-01 15:35:17 +0800402
403 if (ret == NET_RX_DROP)
404 kfree_skb(skb);
405
406 /* return NET_RX_SUCCESS in any case as we
407 * most probably dropped the packet for
408 * routing-logical reasons.
409 */
410 return NET_RX_SUCCESS;
411
412err_free:
413 kfree_skb(skb);
414err_out:
415 return NET_RX_DROP;
416}
417
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200418static void batadv_recv_handler_init(void)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800419{
420 int i;
421
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200422 for (i = 0; i < ARRAY_SIZE(batadv_rx_handler); i++)
423 batadv_rx_handler[i] = batadv_recv_unhandled_packet;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800424
Simon Wunderlicha1f1ac52013-04-25 10:37:23 +0200425 for (i = BATADV_UNICAST_MIN; i <= BATADV_UNICAST_MAX; i++)
426 batadv_rx_handler[i] = batadv_recv_unhandled_unicast_packet;
427
Simon Wunderlich031ace82013-12-17 19:12:12 +0100428 /* compile time checks for sizes */
429 BUILD_BUG_ON(sizeof(struct batadv_bla_claim_dst) != 6);
430 BUILD_BUG_ON(sizeof(struct batadv_ogm_packet) != 24);
431 BUILD_BUG_ON(sizeof(struct batadv_icmp_header) != 20);
432 BUILD_BUG_ON(sizeof(struct batadv_icmp_packet) != 20);
433 BUILD_BUG_ON(sizeof(struct batadv_icmp_packet_rr) != 116);
434 BUILD_BUG_ON(sizeof(struct batadv_unicast_packet) != 10);
435 BUILD_BUG_ON(sizeof(struct batadv_unicast_4addr_packet) != 18);
436 BUILD_BUG_ON(sizeof(struct batadv_frag_packet) != 20);
437 BUILD_BUG_ON(sizeof(struct batadv_bcast_packet) != 14);
438 BUILD_BUG_ON(sizeof(struct batadv_coded_packet) != 46);
439 BUILD_BUG_ON(sizeof(struct batadv_unicast_tvlv_packet) != 20);
440 BUILD_BUG_ON(sizeof(struct batadv_tvlv_hdr) != 4);
441 BUILD_BUG_ON(sizeof(struct batadv_tvlv_gateway_data) != 8);
442 BUILD_BUG_ON(sizeof(struct batadv_tvlv_tt_vlan_data) != 8);
443 BUILD_BUG_ON(sizeof(struct batadv_tvlv_tt_change) != 12);
444 BUILD_BUG_ON(sizeof(struct batadv_tvlv_roam_adv) != 8);
Simon Wunderlich80067c82013-04-25 10:37:22 +0200445
Simon Wunderlicha1f1ac52013-04-25 10:37:23 +0200446 /* broadcast packet */
447 batadv_rx_handler[BATADV_BCAST] = batadv_recv_bcast_packet;
448
449 /* unicast packets ... */
Antonio Quartulli7cdcf6d2012-10-01 09:57:35 +0200450 /* unicast with 4 addresses packet */
451 batadv_rx_handler[BATADV_UNICAST_4ADDR] = batadv_recv_unicast_packet;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800452 /* unicast packet */
Sven Eckelmannacd34af2012-06-03 22:19:21 +0200453 batadv_rx_handler[BATADV_UNICAST] = batadv_recv_unicast_packet;
Marek Lindneref261572013-04-23 21:39:57 +0800454 /* unicast tvlv packet */
455 batadv_rx_handler[BATADV_UNICAST_TVLV] = batadv_recv_unicast_tvlv;
Simon Wunderlicha1f1ac52013-04-25 10:37:23 +0200456 /* batman icmp packet */
457 batadv_rx_handler[BATADV_ICMP] = batadv_recv_icmp_packet;
Martin Hundebøll610bfc6bc2013-05-23 16:53:02 +0200458 /* Fragmented packets */
459 batadv_rx_handler[BATADV_UNICAST_FRAG] = batadv_recv_frag_packet;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800460}
461
Sven Eckelmann56303d32012-06-05 22:31:31 +0200462int
463batadv_recv_handler_register(uint8_t packet_type,
464 int (*recv_handler)(struct sk_buff *,
465 struct batadv_hard_iface *))
Marek Lindnerffa995e2012-03-01 15:35:17 +0800466{
Simon Wunderlicha1f1ac52013-04-25 10:37:23 +0200467 int (*curr)(struct sk_buff *,
468 struct batadv_hard_iface *);
469 curr = batadv_rx_handler[packet_type];
470
471 if ((curr != batadv_recv_unhandled_packet) &&
472 (curr != batadv_recv_unhandled_unicast_packet))
Marek Lindnerffa995e2012-03-01 15:35:17 +0800473 return -EBUSY;
474
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200475 batadv_rx_handler[packet_type] = recv_handler;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800476 return 0;
477}
478
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200479void batadv_recv_handler_unregister(uint8_t packet_type)
Marek Lindnerffa995e2012-03-01 15:35:17 +0800480{
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200481 batadv_rx_handler[packet_type] = batadv_recv_unhandled_packet;
Marek Lindnerffa995e2012-03-01 15:35:17 +0800482}
483
Sven Eckelmann56303d32012-06-05 22:31:31 +0200484static struct batadv_algo_ops *batadv_algo_get(char *name)
Marek Lindner1c280472011-11-28 17:40:17 +0800485{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200486 struct batadv_algo_ops *bat_algo_ops = NULL, *bat_algo_ops_tmp;
Marek Lindner1c280472011-11-28 17:40:17 +0800487
Sasha Levinb67bfe02013-02-27 17:06:00 -0800488 hlist_for_each_entry(bat_algo_ops_tmp, &batadv_algo_list, list) {
Marek Lindner1c280472011-11-28 17:40:17 +0800489 if (strcmp(bat_algo_ops_tmp->name, name) != 0)
490 continue;
491
492 bat_algo_ops = bat_algo_ops_tmp;
493 break;
494 }
495
496 return bat_algo_ops;
497}
498
Sven Eckelmann56303d32012-06-05 22:31:31 +0200499int batadv_algo_register(struct batadv_algo_ops *bat_algo_ops)
Marek Lindner1c280472011-11-28 17:40:17 +0800500{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200501 struct batadv_algo_ops *bat_algo_ops_tmp;
Sven Eckelmann5346c352012-05-05 13:27:28 +0200502 int ret;
Marek Lindner1c280472011-11-28 17:40:17 +0800503
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200504 bat_algo_ops_tmp = batadv_algo_get(bat_algo_ops->name);
Marek Lindner1c280472011-11-28 17:40:17 +0800505 if (bat_algo_ops_tmp) {
Sven Eckelmann86ceb362012-03-07 09:07:45 +0100506 pr_info("Trying to register already registered routing algorithm: %s\n",
507 bat_algo_ops->name);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200508 ret = -EEXIST;
Marek Lindner1c280472011-11-28 17:40:17 +0800509 goto out;
510 }
511
Marek Lindner01c42242011-11-28 21:31:55 +0800512 /* all algorithms must implement all ops (for now) */
Marek Lindnerc2aca022012-02-07 17:20:45 +0800513 if (!bat_algo_ops->bat_iface_enable ||
Marek Lindner00a50072012-02-07 17:20:47 +0800514 !bat_algo_ops->bat_iface_disable ||
Marek Lindnerc3229392012-03-11 06:17:50 +0800515 !bat_algo_ops->bat_iface_update_mac ||
Marek Lindnercd8b78e2012-02-07 17:20:49 +0800516 !bat_algo_ops->bat_primary_iface_set ||
Marek Lindner01c42242011-11-28 21:31:55 +0800517 !bat_algo_ops->bat_ogm_schedule ||
Antonio Quartullia3285a82013-09-02 12:15:04 +0200518 !bat_algo_ops->bat_ogm_emit ||
Antonio Quartullic43c9812013-09-02 12:15:05 +0200519 !bat_algo_ops->bat_neigh_cmp ||
520 !bat_algo_ops->bat_neigh_is_equiv_or_better) {
Marek Lindner01c42242011-11-28 21:31:55 +0800521 pr_info("Routing algo '%s' does not implement required ops\n",
522 bat_algo_ops->name);
Sven Eckelmann5346c352012-05-05 13:27:28 +0200523 ret = -EINVAL;
Marek Lindner01c42242011-11-28 21:31:55 +0800524 goto out;
525 }
526
Marek Lindner1c280472011-11-28 17:40:17 +0800527 INIT_HLIST_NODE(&bat_algo_ops->list);
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200528 hlist_add_head(&bat_algo_ops->list, &batadv_algo_list);
Marek Lindner1c280472011-11-28 17:40:17 +0800529 ret = 0;
530
531out:
532 return ret;
533}
534
Sven Eckelmann56303d32012-06-05 22:31:31 +0200535int batadv_algo_select(struct batadv_priv *bat_priv, char *name)
Marek Lindner1c280472011-11-28 17:40:17 +0800536{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200537 struct batadv_algo_ops *bat_algo_ops;
Sven Eckelmann5346c352012-05-05 13:27:28 +0200538 int ret = -EINVAL;
Marek Lindner1c280472011-11-28 17:40:17 +0800539
Sven Eckelmannee11ad62012-05-16 20:23:19 +0200540 bat_algo_ops = batadv_algo_get(name);
Marek Lindner1c280472011-11-28 17:40:17 +0800541 if (!bat_algo_ops)
542 goto out;
543
544 bat_priv->bat_algo_ops = bat_algo_ops;
545 ret = 0;
546
547out:
548 return ret;
549}
550
Sven Eckelmann3193e8f2012-05-12 02:09:42 +0200551int batadv_algo_seq_print_text(struct seq_file *seq, void *offset)
Marek Lindner1c280472011-11-28 17:40:17 +0800552{
Sven Eckelmann56303d32012-06-05 22:31:31 +0200553 struct batadv_algo_ops *bat_algo_ops;
Marek Lindner1c280472011-11-28 17:40:17 +0800554
Antonio Quartulli0c814652013-03-21 09:23:29 +0100555 seq_puts(seq, "Available routing algorithms:\n");
Marek Lindner1c280472011-11-28 17:40:17 +0800556
Sasha Levinb67bfe02013-02-27 17:06:00 -0800557 hlist_for_each_entry(bat_algo_ops, &batadv_algo_list, list) {
Marek Lindner1c280472011-11-28 17:40:17 +0800558 seq_printf(seq, "%s\n", bat_algo_ops->name);
559 }
560
561 return 0;
562}
563
Sven Eckelmann95a066d2012-10-17 21:10:39 +0200564/**
565 * batadv_skb_crc32 - calculate CRC32 of the whole packet and skip bytes in
566 * the header
567 * @skb: skb pointing to fragmented socket buffers
568 * @payload_ptr: Pointer to position inside the head buffer of the skb
569 * marking the start of the data to be CRC'ed
570 *
571 * payload_ptr must always point to an address in the skb head buffer and not to
572 * a fragment.
573 */
574__be32 batadv_skb_crc32(struct sk_buff *skb, u8 *payload_ptr)
575{
576 u32 crc = 0;
577 unsigned int from;
578 unsigned int to = skb->len;
579 struct skb_seq_state st;
580 const u8 *data;
581 unsigned int len;
582 unsigned int consumed = 0;
583
584 from = (unsigned int)(payload_ptr - skb->data);
585
586 skb_prepare_seq_read(skb, from, to, &st);
587 while ((len = skb_seq_read(consumed, &data, &st)) != 0) {
588 crc = crc32c(crc, data, len);
589 consumed += len;
590 }
Sven Eckelmann95a066d2012-10-17 21:10:39 +0200591
592 return htonl(crc);
593}
594
Marek Lindneref261572013-04-23 21:39:57 +0800595/**
596 * batadv_tvlv_handler_free_ref - decrement the tvlv handler refcounter and
597 * possibly free it
598 * @tvlv_handler: the tvlv handler to free
599 */
600static void
601batadv_tvlv_handler_free_ref(struct batadv_tvlv_handler *tvlv_handler)
602{
603 if (atomic_dec_and_test(&tvlv_handler->refcount))
604 kfree_rcu(tvlv_handler, rcu);
605}
606
607/**
608 * batadv_tvlv_handler_get - retrieve tvlv handler from the tvlv handler list
609 * based on the provided type and version (both need to match)
610 * @bat_priv: the bat priv with all the soft interface information
611 * @type: tvlv handler type to look for
612 * @version: tvlv handler version to look for
613 *
614 * Returns tvlv handler if found or NULL otherwise.
615 */
616static struct batadv_tvlv_handler
617*batadv_tvlv_handler_get(struct batadv_priv *bat_priv,
618 uint8_t type, uint8_t version)
619{
620 struct batadv_tvlv_handler *tvlv_handler_tmp, *tvlv_handler = NULL;
621
622 rcu_read_lock();
623 hlist_for_each_entry_rcu(tvlv_handler_tmp,
624 &bat_priv->tvlv.handler_list, list) {
625 if (tvlv_handler_tmp->type != type)
626 continue;
627
628 if (tvlv_handler_tmp->version != version)
629 continue;
630
631 if (!atomic_inc_not_zero(&tvlv_handler_tmp->refcount))
632 continue;
633
634 tvlv_handler = tvlv_handler_tmp;
635 break;
636 }
637 rcu_read_unlock();
638
639 return tvlv_handler;
640}
641
642/**
643 * batadv_tvlv_container_free_ref - decrement the tvlv container refcounter and
644 * possibly free it
645 * @tvlv_handler: the tvlv container to free
646 */
647static void batadv_tvlv_container_free_ref(struct batadv_tvlv_container *tvlv)
648{
649 if (atomic_dec_and_test(&tvlv->refcount))
650 kfree(tvlv);
651}
652
653/**
654 * batadv_tvlv_container_get - retrieve tvlv container from the tvlv container
655 * list based on the provided type and version (both need to match)
656 * @bat_priv: the bat priv with all the soft interface information
657 * @type: tvlv container type to look for
658 * @version: tvlv container version to look for
659 *
660 * Has to be called with the appropriate locks being acquired
661 * (tvlv.container_list_lock).
662 *
663 * Returns tvlv container if found or NULL otherwise.
664 */
665static struct batadv_tvlv_container
666*batadv_tvlv_container_get(struct batadv_priv *bat_priv,
667 uint8_t type, uint8_t version)
668{
669 struct batadv_tvlv_container *tvlv_tmp, *tvlv = NULL;
670
671 hlist_for_each_entry(tvlv_tmp, &bat_priv->tvlv.container_list, list) {
672 if (tvlv_tmp->tvlv_hdr.type != type)
673 continue;
674
675 if (tvlv_tmp->tvlv_hdr.version != version)
676 continue;
677
678 if (!atomic_inc_not_zero(&tvlv_tmp->refcount))
679 continue;
680
681 tvlv = tvlv_tmp;
682 break;
683 }
684
685 return tvlv;
686}
687
688/**
689 * batadv_tvlv_container_list_size - calculate the size of the tvlv container
690 * list entries
691 * @bat_priv: the bat priv with all the soft interface information
692 *
693 * Has to be called with the appropriate locks being acquired
694 * (tvlv.container_list_lock).
695 *
696 * Returns size of all currently registered tvlv containers in bytes.
697 */
698static uint16_t batadv_tvlv_container_list_size(struct batadv_priv *bat_priv)
699{
700 struct batadv_tvlv_container *tvlv;
701 uint16_t tvlv_len = 0;
702
703 hlist_for_each_entry(tvlv, &bat_priv->tvlv.container_list, list) {
704 tvlv_len += sizeof(struct batadv_tvlv_hdr);
705 tvlv_len += ntohs(tvlv->tvlv_hdr.len);
706 }
707
708 return tvlv_len;
709}
710
711/**
712 * batadv_tvlv_container_remove - remove tvlv container from the tvlv container
713 * list
714 * @tvlv: the to be removed tvlv container
715 *
716 * Has to be called with the appropriate locks being acquired
717 * (tvlv.container_list_lock).
718 */
719static void batadv_tvlv_container_remove(struct batadv_tvlv_container *tvlv)
720{
721 if (!tvlv)
722 return;
723
724 hlist_del(&tvlv->list);
725
726 /* first call to decrement the counter, second call to free */
727 batadv_tvlv_container_free_ref(tvlv);
728 batadv_tvlv_container_free_ref(tvlv);
729}
730
731/**
732 * batadv_tvlv_container_unregister - unregister tvlv container based on the
733 * provided type and version (both need to match)
734 * @bat_priv: the bat priv with all the soft interface information
735 * @type: tvlv container type to unregister
736 * @version: tvlv container type to unregister
737 */
738void batadv_tvlv_container_unregister(struct batadv_priv *bat_priv,
739 uint8_t type, uint8_t version)
740{
741 struct batadv_tvlv_container *tvlv;
742
743 spin_lock_bh(&bat_priv->tvlv.container_list_lock);
744 tvlv = batadv_tvlv_container_get(bat_priv, type, version);
745 batadv_tvlv_container_remove(tvlv);
746 spin_unlock_bh(&bat_priv->tvlv.container_list_lock);
747}
748
749/**
750 * batadv_tvlv_container_register - register tvlv type, version and content
751 * to be propagated with each (primary interface) OGM
752 * @bat_priv: the bat priv with all the soft interface information
753 * @type: tvlv container type
754 * @version: tvlv container version
755 * @tvlv_value: tvlv container content
756 * @tvlv_value_len: tvlv container content length
757 *
758 * If a container of the same type and version was already registered the new
759 * content is going to replace the old one.
760 */
761void batadv_tvlv_container_register(struct batadv_priv *bat_priv,
762 uint8_t type, uint8_t version,
763 void *tvlv_value, uint16_t tvlv_value_len)
764{
765 struct batadv_tvlv_container *tvlv_old, *tvlv_new;
766
767 if (!tvlv_value)
768 tvlv_value_len = 0;
769
770 tvlv_new = kzalloc(sizeof(*tvlv_new) + tvlv_value_len, GFP_ATOMIC);
771 if (!tvlv_new)
772 return;
773
774 tvlv_new->tvlv_hdr.version = version;
775 tvlv_new->tvlv_hdr.type = type;
776 tvlv_new->tvlv_hdr.len = htons(tvlv_value_len);
777
778 memcpy(tvlv_new + 1, tvlv_value, ntohs(tvlv_new->tvlv_hdr.len));
779 INIT_HLIST_NODE(&tvlv_new->list);
780 atomic_set(&tvlv_new->refcount, 1);
781
782 spin_lock_bh(&bat_priv->tvlv.container_list_lock);
783 tvlv_old = batadv_tvlv_container_get(bat_priv, type, version);
784 batadv_tvlv_container_remove(tvlv_old);
785 hlist_add_head(&tvlv_new->list, &bat_priv->tvlv.container_list);
786 spin_unlock_bh(&bat_priv->tvlv.container_list_lock);
787}
788
789/**
790 * batadv_tvlv_realloc_packet_buff - reallocate packet buffer to accomodate
791 * requested packet size
792 * @packet_buff: packet buffer
793 * @packet_buff_len: packet buffer size
794 * @packet_min_len: requested packet minimum size
795 * @additional_packet_len: requested additional packet size on top of minimum
796 * size
797 *
798 * Returns true of the packet buffer could be changed to the requested size,
799 * false otherwise.
800 */
801static bool batadv_tvlv_realloc_packet_buff(unsigned char **packet_buff,
802 int *packet_buff_len,
803 int min_packet_len,
804 int additional_packet_len)
805{
806 unsigned char *new_buff;
807
808 new_buff = kmalloc(min_packet_len + additional_packet_len, GFP_ATOMIC);
809
810 /* keep old buffer if kmalloc should fail */
811 if (new_buff) {
812 memcpy(new_buff, *packet_buff, min_packet_len);
813 kfree(*packet_buff);
814 *packet_buff = new_buff;
815 *packet_buff_len = min_packet_len + additional_packet_len;
816 return true;
817 }
818
819 return false;
820}
821
822/**
823 * batadv_tvlv_container_ogm_append - append tvlv container content to given
824 * OGM packet buffer
825 * @bat_priv: the bat priv with all the soft interface information
826 * @packet_buff: ogm packet buffer
827 * @packet_buff_len: ogm packet buffer size including ogm header and tvlv
828 * content
829 * @packet_min_len: ogm header size to be preserved for the OGM itself
830 *
831 * The ogm packet might be enlarged or shrunk depending on the current size
832 * and the size of the to-be-appended tvlv containers.
833 *
834 * Returns size of all appended tvlv containers in bytes.
835 */
836uint16_t batadv_tvlv_container_ogm_append(struct batadv_priv *bat_priv,
837 unsigned char **packet_buff,
838 int *packet_buff_len,
839 int packet_min_len)
840{
841 struct batadv_tvlv_container *tvlv;
842 struct batadv_tvlv_hdr *tvlv_hdr;
843 uint16_t tvlv_value_len;
844 void *tvlv_value;
845 bool ret;
846
847 spin_lock_bh(&bat_priv->tvlv.container_list_lock);
848 tvlv_value_len = batadv_tvlv_container_list_size(bat_priv);
849
850 ret = batadv_tvlv_realloc_packet_buff(packet_buff, packet_buff_len,
851 packet_min_len, tvlv_value_len);
852
853 if (!ret)
854 goto end;
855
856 if (!tvlv_value_len)
857 goto end;
858
859 tvlv_value = (*packet_buff) + packet_min_len;
860
861 hlist_for_each_entry(tvlv, &bat_priv->tvlv.container_list, list) {
862 tvlv_hdr = tvlv_value;
863 tvlv_hdr->type = tvlv->tvlv_hdr.type;
864 tvlv_hdr->version = tvlv->tvlv_hdr.version;
865 tvlv_hdr->len = tvlv->tvlv_hdr.len;
866 tvlv_value = tvlv_hdr + 1;
867 memcpy(tvlv_value, tvlv + 1, ntohs(tvlv->tvlv_hdr.len));
868 tvlv_value = (uint8_t *)tvlv_value + ntohs(tvlv->tvlv_hdr.len);
869 }
870
871end:
872 spin_unlock_bh(&bat_priv->tvlv.container_list_lock);
873 return tvlv_value_len;
874}
875
876/**
877 * batadv_tvlv_call_handler - parse the given tvlv buffer to call the
878 * appropriate handlers
879 * @bat_priv: the bat priv with all the soft interface information
880 * @tvlv_handler: tvlv callback function handling the tvlv content
881 * @ogm_source: flag indicating wether the tvlv is an ogm or a unicast packet
882 * @orig_node: orig node emitting the ogm packet
883 * @src: source mac address of the unicast packet
884 * @dst: destination mac address of the unicast packet
885 * @tvlv_value: tvlv content
886 * @tvlv_value_len: tvlv content length
887 *
888 * Returns success if handler was not found or the return value of the handler
889 * callback.
890 */
891static int batadv_tvlv_call_handler(struct batadv_priv *bat_priv,
892 struct batadv_tvlv_handler *tvlv_handler,
893 bool ogm_source,
894 struct batadv_orig_node *orig_node,
895 uint8_t *src, uint8_t *dst,
896 void *tvlv_value, uint16_t tvlv_value_len)
897{
898 if (!tvlv_handler)
899 return NET_RX_SUCCESS;
900
901 if (ogm_source) {
902 if (!tvlv_handler->ogm_handler)
903 return NET_RX_SUCCESS;
904
905 if (!orig_node)
906 return NET_RX_SUCCESS;
907
908 tvlv_handler->ogm_handler(bat_priv, orig_node,
909 BATADV_NO_FLAGS,
910 tvlv_value, tvlv_value_len);
911 tvlv_handler->flags |= BATADV_TVLV_HANDLER_OGM_CALLED;
912 } else {
913 if (!src)
914 return NET_RX_SUCCESS;
915
916 if (!dst)
917 return NET_RX_SUCCESS;
918
919 if (!tvlv_handler->unicast_handler)
920 return NET_RX_SUCCESS;
921
922 return tvlv_handler->unicast_handler(bat_priv, src,
923 dst, tvlv_value,
924 tvlv_value_len);
925 }
926
927 return NET_RX_SUCCESS;
928}
929
930/**
931 * batadv_tvlv_containers_process - parse the given tvlv buffer to call the
932 * appropriate handlers
933 * @bat_priv: the bat priv with all the soft interface information
934 * @ogm_source: flag indicating wether the tvlv is an ogm or a unicast packet
935 * @orig_node: orig node emitting the ogm packet
936 * @src: source mac address of the unicast packet
937 * @dst: destination mac address of the unicast packet
938 * @tvlv_value: tvlv content
939 * @tvlv_value_len: tvlv content length
940 *
941 * Returns success when processing an OGM or the return value of all called
942 * handler callbacks.
943 */
944int batadv_tvlv_containers_process(struct batadv_priv *bat_priv,
945 bool ogm_source,
946 struct batadv_orig_node *orig_node,
947 uint8_t *src, uint8_t *dst,
948 void *tvlv_value, uint16_t tvlv_value_len)
949{
950 struct batadv_tvlv_handler *tvlv_handler;
951 struct batadv_tvlv_hdr *tvlv_hdr;
952 uint16_t tvlv_value_cont_len;
953 uint8_t cifnotfound = BATADV_TVLV_HANDLER_OGM_CIFNOTFND;
954 int ret = NET_RX_SUCCESS;
955
956 while (tvlv_value_len >= sizeof(*tvlv_hdr)) {
957 tvlv_hdr = tvlv_value;
958 tvlv_value_cont_len = ntohs(tvlv_hdr->len);
959 tvlv_value = tvlv_hdr + 1;
960 tvlv_value_len -= sizeof(*tvlv_hdr);
961
962 if (tvlv_value_cont_len > tvlv_value_len)
963 break;
964
965 tvlv_handler = batadv_tvlv_handler_get(bat_priv,
966 tvlv_hdr->type,
967 tvlv_hdr->version);
968
969 ret |= batadv_tvlv_call_handler(bat_priv, tvlv_handler,
970 ogm_source, orig_node,
971 src, dst, tvlv_value,
972 tvlv_value_cont_len);
973 if (tvlv_handler)
974 batadv_tvlv_handler_free_ref(tvlv_handler);
975 tvlv_value = (uint8_t *)tvlv_value + tvlv_value_cont_len;
976 tvlv_value_len -= tvlv_value_cont_len;
977 }
978
979 if (!ogm_source)
980 return ret;
981
982 rcu_read_lock();
983 hlist_for_each_entry_rcu(tvlv_handler,
984 &bat_priv->tvlv.handler_list, list) {
985 if ((tvlv_handler->flags & BATADV_TVLV_HANDLER_OGM_CIFNOTFND) &&
986 !(tvlv_handler->flags & BATADV_TVLV_HANDLER_OGM_CALLED))
987 tvlv_handler->ogm_handler(bat_priv, orig_node,
988 cifnotfound, NULL, 0);
989
990 tvlv_handler->flags &= ~BATADV_TVLV_HANDLER_OGM_CALLED;
991 }
992 rcu_read_unlock();
993
994 return NET_RX_SUCCESS;
995}
996
997/**
998 * batadv_tvlv_ogm_receive - process an incoming ogm and call the appropriate
999 * handlers
1000 * @bat_priv: the bat priv with all the soft interface information
1001 * @batadv_ogm_packet: ogm packet containing the tvlv containers
1002 * @orig_node: orig node emitting the ogm packet
1003 */
1004void batadv_tvlv_ogm_receive(struct batadv_priv *bat_priv,
1005 struct batadv_ogm_packet *batadv_ogm_packet,
1006 struct batadv_orig_node *orig_node)
1007{
1008 void *tvlv_value;
1009 uint16_t tvlv_value_len;
1010
1011 if (!batadv_ogm_packet)
1012 return;
1013
1014 tvlv_value_len = ntohs(batadv_ogm_packet->tvlv_len);
1015 if (!tvlv_value_len)
1016 return;
1017
1018 tvlv_value = batadv_ogm_packet + 1;
1019
1020 batadv_tvlv_containers_process(bat_priv, true, orig_node, NULL, NULL,
1021 tvlv_value, tvlv_value_len);
1022}
1023
1024/**
1025 * batadv_tvlv_handler_register - register tvlv handler based on the provided
1026 * type and version (both need to match) for ogm tvlv payload and/or unicast
1027 * payload
1028 * @bat_priv: the bat priv with all the soft interface information
1029 * @optr: ogm tvlv handler callback function. This function receives the orig
1030 * node, flags and the tvlv content as argument to process.
1031 * @uptr: unicast tvlv handler callback function. This function receives the
1032 * source & destination of the unicast packet as well as the tvlv content
1033 * to process.
1034 * @type: tvlv handler type to be registered
1035 * @version: tvlv handler version to be registered
1036 * @flags: flags to enable or disable TVLV API behavior
1037 */
1038void batadv_tvlv_handler_register(struct batadv_priv *bat_priv,
1039 void (*optr)(struct batadv_priv *bat_priv,
1040 struct batadv_orig_node *orig,
1041 uint8_t flags,
1042 void *tvlv_value,
1043 uint16_t tvlv_value_len),
1044 int (*uptr)(struct batadv_priv *bat_priv,
1045 uint8_t *src, uint8_t *dst,
1046 void *tvlv_value,
1047 uint16_t tvlv_value_len),
1048 uint8_t type, uint8_t version, uint8_t flags)
1049{
1050 struct batadv_tvlv_handler *tvlv_handler;
1051
1052 tvlv_handler = batadv_tvlv_handler_get(bat_priv, type, version);
1053 if (tvlv_handler) {
1054 batadv_tvlv_handler_free_ref(tvlv_handler);
1055 return;
1056 }
1057
1058 tvlv_handler = kzalloc(sizeof(*tvlv_handler), GFP_ATOMIC);
1059 if (!tvlv_handler)
1060 return;
1061
1062 tvlv_handler->ogm_handler = optr;
1063 tvlv_handler->unicast_handler = uptr;
1064 tvlv_handler->type = type;
1065 tvlv_handler->version = version;
1066 tvlv_handler->flags = flags;
1067 atomic_set(&tvlv_handler->refcount, 1);
1068 INIT_HLIST_NODE(&tvlv_handler->list);
1069
1070 spin_lock_bh(&bat_priv->tvlv.handler_list_lock);
1071 hlist_add_head_rcu(&tvlv_handler->list, &bat_priv->tvlv.handler_list);
1072 spin_unlock_bh(&bat_priv->tvlv.handler_list_lock);
1073}
1074
1075/**
1076 * batadv_tvlv_handler_unregister - unregister tvlv handler based on the
1077 * provided type and version (both need to match)
1078 * @bat_priv: the bat priv with all the soft interface information
1079 * @type: tvlv handler type to be unregistered
1080 * @version: tvlv handler version to be unregistered
1081 */
1082void batadv_tvlv_handler_unregister(struct batadv_priv *bat_priv,
1083 uint8_t type, uint8_t version)
1084{
1085 struct batadv_tvlv_handler *tvlv_handler;
1086
1087 tvlv_handler = batadv_tvlv_handler_get(bat_priv, type, version);
1088 if (!tvlv_handler)
1089 return;
1090
1091 batadv_tvlv_handler_free_ref(tvlv_handler);
1092 spin_lock_bh(&bat_priv->tvlv.handler_list_lock);
1093 hlist_del_rcu(&tvlv_handler->list);
1094 spin_unlock_bh(&bat_priv->tvlv.handler_list_lock);
1095 batadv_tvlv_handler_free_ref(tvlv_handler);
1096}
1097
1098/**
1099 * batadv_tvlv_unicast_send - send a unicast packet with tvlv payload to the
1100 * specified host
1101 * @bat_priv: the bat priv with all the soft interface information
1102 * @src: source mac address of the unicast packet
1103 * @dst: destination mac address of the unicast packet
1104 * @type: tvlv type
1105 * @version: tvlv version
1106 * @tvlv_value: tvlv content
1107 * @tvlv_value_len: tvlv content length
1108 */
1109void batadv_tvlv_unicast_send(struct batadv_priv *bat_priv, uint8_t *src,
1110 uint8_t *dst, uint8_t type, uint8_t version,
1111 void *tvlv_value, uint16_t tvlv_value_len)
1112{
1113 struct batadv_unicast_tvlv_packet *unicast_tvlv_packet;
1114 struct batadv_tvlv_hdr *tvlv_hdr;
1115 struct batadv_orig_node *orig_node;
1116 struct sk_buff *skb = NULL;
1117 unsigned char *tvlv_buff;
1118 unsigned int tvlv_len;
1119 ssize_t hdr_len = sizeof(*unicast_tvlv_packet);
1120 bool ret = false;
1121
1122 orig_node = batadv_orig_hash_find(bat_priv, dst);
1123 if (!orig_node)
1124 goto out;
1125
1126 tvlv_len = sizeof(*tvlv_hdr) + tvlv_value_len;
1127
1128 skb = netdev_alloc_skb_ip_align(NULL, ETH_HLEN + hdr_len + tvlv_len);
1129 if (!skb)
1130 goto out;
1131
1132 skb->priority = TC_PRIO_CONTROL;
1133 skb_reserve(skb, ETH_HLEN);
1134 tvlv_buff = skb_put(skb, sizeof(*unicast_tvlv_packet) + tvlv_len);
1135 unicast_tvlv_packet = (struct batadv_unicast_tvlv_packet *)tvlv_buff;
Simon Wunderlicha40d9b02013-12-02 20:38:31 +01001136 unicast_tvlv_packet->packet_type = BATADV_UNICAST_TVLV;
1137 unicast_tvlv_packet->version = BATADV_COMPAT_VERSION;
1138 unicast_tvlv_packet->ttl = BATADV_TTL;
Marek Lindneref261572013-04-23 21:39:57 +08001139 unicast_tvlv_packet->reserved = 0;
1140 unicast_tvlv_packet->tvlv_len = htons(tvlv_len);
1141 unicast_tvlv_packet->align = 0;
Antonio Quartulli8fdd0152014-01-22 00:42:11 +01001142 ether_addr_copy(unicast_tvlv_packet->src, src);
1143 ether_addr_copy(unicast_tvlv_packet->dst, dst);
Marek Lindneref261572013-04-23 21:39:57 +08001144
1145 tvlv_buff = (unsigned char *)(unicast_tvlv_packet + 1);
1146 tvlv_hdr = (struct batadv_tvlv_hdr *)tvlv_buff;
1147 tvlv_hdr->version = version;
1148 tvlv_hdr->type = type;
1149 tvlv_hdr->len = htons(tvlv_value_len);
1150 tvlv_buff += sizeof(*tvlv_hdr);
1151 memcpy(tvlv_buff, tvlv_value, tvlv_value_len);
1152
1153 if (batadv_send_skb_to_orig(skb, orig_node, NULL) != NET_XMIT_DROP)
1154 ret = true;
1155
1156out:
1157 if (skb && !ret)
1158 kfree_skb(skb);
1159 if (orig_node)
1160 batadv_orig_node_free_ref(orig_node);
1161}
1162
Antonio Quartullic018ad32013-06-04 12:11:39 +02001163/**
1164 * batadv_get_vid - extract the VLAN identifier from skb if any
1165 * @skb: the buffer containing the packet
1166 * @header_len: length of the batman header preceding the ethernet header
1167 *
1168 * If the packet embedded in the skb is vlan tagged this function returns the
1169 * VID with the BATADV_VLAN_HAS_TAG flag. Otherwise BATADV_NO_FLAGS is returned.
1170 */
1171unsigned short batadv_get_vid(struct sk_buff *skb, size_t header_len)
1172{
1173 struct ethhdr *ethhdr = (struct ethhdr *)(skb->data + header_len);
1174 struct vlan_ethhdr *vhdr;
1175 unsigned short vid;
1176
1177 if (ethhdr->h_proto != htons(ETH_P_8021Q))
1178 return BATADV_NO_FLAGS;
1179
1180 if (!pskb_may_pull(skb, header_len + VLAN_ETH_HLEN))
1181 return BATADV_NO_FLAGS;
1182
1183 vhdr = (struct vlan_ethhdr *)(skb->data + header_len);
1184 vid = ntohs(vhdr->h_vlan_TCI) & VLAN_VID_MASK;
1185 vid |= BATADV_VLAN_HAS_TAG;
1186
1187 return vid;
1188}
1189
Antonio Quartullieceb22a2013-11-16 12:03:51 +01001190/**
1191 * batadv_vlan_ap_isola_get - return the AP isolation status for the given vlan
1192 * @bat_priv: the bat priv with all the soft interface information
1193 * @vid: the VLAN identifier for which the AP isolation attributed as to be
1194 * looked up
1195 *
1196 * Returns true if AP isolation is on for the VLAN idenfied by vid, false
1197 * otherwise
1198 */
1199bool batadv_vlan_ap_isola_get(struct batadv_priv *bat_priv, unsigned short vid)
1200{
1201 bool ap_isolation_enabled = false;
1202 struct batadv_softif_vlan *vlan;
1203
1204 /* if the AP isolation is requested on a VLAN, then check for its
1205 * setting in the proper VLAN private data structure
1206 */
1207 vlan = batadv_softif_vlan_get(bat_priv, vid);
1208 if (vlan) {
1209 ap_isolation_enabled = atomic_read(&vlan->ap_isolation);
1210 batadv_softif_vlan_free_ref(vlan);
1211 }
1212
1213 return ap_isolation_enabled;
1214}
1215
Sven Eckelmannee11ad62012-05-16 20:23:19 +02001216static int batadv_param_set_ra(const char *val, const struct kernel_param *kp)
Marek Lindnerd419be12011-12-10 19:45:53 +08001217{
Sven Eckelmann56303d32012-06-05 22:31:31 +02001218 struct batadv_algo_ops *bat_algo_ops;
Marek Lindnerd8cb54862012-04-18 17:16:39 +08001219 char *algo_name = (char *)val;
1220 size_t name_len = strlen(algo_name);
Marek Lindnerd419be12011-12-10 19:45:53 +08001221
Marek Lindner293c9c12013-04-27 16:22:28 +08001222 if (name_len > 0 && algo_name[name_len - 1] == '\n')
Marek Lindnerd8cb54862012-04-18 17:16:39 +08001223 algo_name[name_len - 1] = '\0';
1224
Sven Eckelmannee11ad62012-05-16 20:23:19 +02001225 bat_algo_ops = batadv_algo_get(algo_name);
Marek Lindnerd419be12011-12-10 19:45:53 +08001226 if (!bat_algo_ops) {
Marek Lindnerd8cb54862012-04-18 17:16:39 +08001227 pr_err("Routing algorithm '%s' is not supported\n", algo_name);
Marek Lindnerd419be12011-12-10 19:45:53 +08001228 return -EINVAL;
1229 }
1230
Marek Lindnerd8cb54862012-04-18 17:16:39 +08001231 return param_set_copystring(algo_name, kp);
Marek Lindnerd419be12011-12-10 19:45:53 +08001232}
1233
Sven Eckelmannee11ad62012-05-16 20:23:19 +02001234static const struct kernel_param_ops batadv_param_ops_ra = {
1235 .set = batadv_param_set_ra,
Marek Lindnerd419be12011-12-10 19:45:53 +08001236 .get = param_get_string,
1237};
1238
Sven Eckelmannee11ad62012-05-16 20:23:19 +02001239static struct kparam_string batadv_param_string_ra = {
Sven Eckelmann3193e8f2012-05-12 02:09:42 +02001240 .maxlen = sizeof(batadv_routing_algo),
1241 .string = batadv_routing_algo,
Marek Lindnerd419be12011-12-10 19:45:53 +08001242};
1243
Sven Eckelmannee11ad62012-05-16 20:23:19 +02001244module_param_cb(routing_algo, &batadv_param_ops_ra, &batadv_param_string_ra,
1245 0644);
1246module_init(batadv_init);
1247module_exit(batadv_exit);
Sven Eckelmannc6c8fea2010-12-13 11:19:28 +00001248
1249MODULE_LICENSE("GPL");
1250
Sven Eckelmann42d0b042012-06-03 22:19:17 +02001251MODULE_AUTHOR(BATADV_DRIVER_AUTHOR);
1252MODULE_DESCRIPTION(BATADV_DRIVER_DESC);
1253MODULE_SUPPORTED_DEVICE(BATADV_DRIVER_DEVICE);
1254MODULE_VERSION(BATADV_SOURCE_VERSION);