Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 1 | /* |
Sven Eckelmann | 64afe35 | 2011-01-27 10:38:15 +0100 | [diff] [blame] | 2 | * Copyright (C) 2007-2011 B.A.T.M.A.N. contributors: |
Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 3 | * |
| 4 | * Marek Lindner, Simon Wunderlich |
| 5 | * |
| 6 | * This program is free software; you can redistribute it and/or |
| 7 | * modify it under the terms of version 2 of the GNU General Public |
| 8 | * License as published by the Free Software Foundation. |
| 9 | * |
| 10 | * This program is distributed in the hope that it will be useful, but |
| 11 | * WITHOUT ANY WARRANTY; without even the implied warranty of |
| 12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 13 | * General Public License for more details. |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License |
| 16 | * along with this program; if not, write to the Free Software |
| 17 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA |
| 18 | * 02110-1301, USA |
| 19 | * |
| 20 | */ |
| 21 | |
| 22 | #include "main.h" |
| 23 | #include "aggregation.h" |
| 24 | #include "send.h" |
| 25 | #include "routing.h" |
| 26 | |
| 27 | /* calculate the size of the hna information for a given packet */ |
| 28 | static int hna_len(struct batman_packet *batman_packet) |
| 29 | { |
| 30 | return batman_packet->num_hna * ETH_ALEN; |
| 31 | } |
| 32 | |
| 33 | /* return true if new_packet can be aggregated with forw_packet */ |
| 34 | static bool can_aggregate_with(struct batman_packet *new_batman_packet, |
| 35 | int packet_len, |
| 36 | unsigned long send_time, |
| 37 | bool directlink, |
Marek Lindner | e6c10f4 | 2011-02-18 12:33:20 +0000 | [diff] [blame] | 38 | struct hard_iface *if_incoming, |
Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 39 | struct forw_packet *forw_packet) |
| 40 | { |
| 41 | struct batman_packet *batman_packet = |
| 42 | (struct batman_packet *)forw_packet->skb->data; |
| 43 | int aggregated_bytes = forw_packet->packet_len + packet_len; |
| 44 | |
| 45 | /** |
| 46 | * we can aggregate the current packet to this aggregated packet |
| 47 | * if: |
| 48 | * |
| 49 | * - the send time is within our MAX_AGGREGATION_MS time |
| 50 | * - the resulting packet wont be bigger than |
| 51 | * MAX_AGGREGATION_BYTES |
| 52 | */ |
| 53 | |
| 54 | if (time_before(send_time, forw_packet->send_time) && |
| 55 | time_after_eq(send_time + msecs_to_jiffies(MAX_AGGREGATION_MS), |
| 56 | forw_packet->send_time) && |
| 57 | (aggregated_bytes <= MAX_AGGREGATION_BYTES)) { |
| 58 | |
| 59 | /** |
| 60 | * check aggregation compatibility |
| 61 | * -> direct link packets are broadcasted on |
| 62 | * their interface only |
| 63 | * -> aggregate packet if the current packet is |
| 64 | * a "global" packet as well as the base |
| 65 | * packet |
| 66 | */ |
| 67 | |
| 68 | /* packets without direct link flag and high TTL |
| 69 | * are flooded through the net */ |
| 70 | if ((!directlink) && |
| 71 | (!(batman_packet->flags & DIRECTLINK)) && |
| 72 | (batman_packet->ttl != 1) && |
| 73 | |
| 74 | /* own packets originating non-primary |
| 75 | * interfaces leave only that interface */ |
| 76 | ((!forw_packet->own) || |
| 77 | (forw_packet->if_incoming->if_num == 0))) |
| 78 | return true; |
| 79 | |
| 80 | /* if the incoming packet is sent via this one |
| 81 | * interface only - we still can aggregate */ |
| 82 | if ((directlink) && |
| 83 | (new_batman_packet->ttl == 1) && |
| 84 | (forw_packet->if_incoming == if_incoming) && |
| 85 | |
| 86 | /* packets from direct neighbors or |
| 87 | * own secondary interface packets |
| 88 | * (= secondary interface packets in general) */ |
| 89 | (batman_packet->flags & DIRECTLINK || |
| 90 | (forw_packet->own && |
| 91 | forw_packet->if_incoming->if_num != 0))) |
| 92 | return true; |
| 93 | } |
| 94 | |
| 95 | return false; |
| 96 | } |
| 97 | |
| 98 | #define atomic_dec_not_zero(v) atomic_add_unless((v), -1, 0) |
| 99 | /* create a new aggregated packet and add this packet to it */ |
| 100 | static void new_aggregated_packet(unsigned char *packet_buff, int packet_len, |
| 101 | unsigned long send_time, bool direct_link, |
Marek Lindner | e6c10f4 | 2011-02-18 12:33:20 +0000 | [diff] [blame] | 102 | struct hard_iface *if_incoming, |
Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 103 | int own_packet) |
| 104 | { |
| 105 | struct bat_priv *bat_priv = netdev_priv(if_incoming->soft_iface); |
| 106 | struct forw_packet *forw_packet_aggr; |
| 107 | unsigned char *skb_buff; |
| 108 | |
| 109 | /* own packet should always be scheduled */ |
| 110 | if (!own_packet) { |
| 111 | if (!atomic_dec_not_zero(&bat_priv->batman_queue_left)) { |
| 112 | bat_dbg(DBG_BATMAN, bat_priv, |
| 113 | "batman packet queue full\n"); |
| 114 | return; |
| 115 | } |
| 116 | } |
| 117 | |
| 118 | forw_packet_aggr = kmalloc(sizeof(struct forw_packet), GFP_ATOMIC); |
| 119 | if (!forw_packet_aggr) { |
| 120 | if (!own_packet) |
| 121 | atomic_inc(&bat_priv->batman_queue_left); |
| 122 | return; |
| 123 | } |
| 124 | |
| 125 | if ((atomic_read(&bat_priv->aggregated_ogms)) && |
| 126 | (packet_len < MAX_AGGREGATION_BYTES)) |
| 127 | forw_packet_aggr->skb = dev_alloc_skb(MAX_AGGREGATION_BYTES + |
| 128 | sizeof(struct ethhdr)); |
| 129 | else |
| 130 | forw_packet_aggr->skb = dev_alloc_skb(packet_len + |
| 131 | sizeof(struct ethhdr)); |
| 132 | |
| 133 | if (!forw_packet_aggr->skb) { |
| 134 | if (!own_packet) |
| 135 | atomic_inc(&bat_priv->batman_queue_left); |
| 136 | kfree(forw_packet_aggr); |
| 137 | return; |
| 138 | } |
| 139 | skb_reserve(forw_packet_aggr->skb, sizeof(struct ethhdr)); |
| 140 | |
| 141 | INIT_HLIST_NODE(&forw_packet_aggr->list); |
| 142 | |
| 143 | skb_buff = skb_put(forw_packet_aggr->skb, packet_len); |
| 144 | forw_packet_aggr->packet_len = packet_len; |
| 145 | memcpy(skb_buff, packet_buff, packet_len); |
| 146 | |
| 147 | forw_packet_aggr->own = own_packet; |
| 148 | forw_packet_aggr->if_incoming = if_incoming; |
| 149 | forw_packet_aggr->num_packets = 0; |
| 150 | forw_packet_aggr->direct_link_flags = 0; |
| 151 | forw_packet_aggr->send_time = send_time; |
| 152 | |
| 153 | /* save packet direct link flag status */ |
| 154 | if (direct_link) |
| 155 | forw_packet_aggr->direct_link_flags |= 1; |
| 156 | |
| 157 | /* add new packet to packet list */ |
| 158 | spin_lock_bh(&bat_priv->forw_bat_list_lock); |
| 159 | hlist_add_head(&forw_packet_aggr->list, &bat_priv->forw_bat_list); |
| 160 | spin_unlock_bh(&bat_priv->forw_bat_list_lock); |
| 161 | |
| 162 | /* start timer for this packet */ |
| 163 | INIT_DELAYED_WORK(&forw_packet_aggr->delayed_work, |
| 164 | send_outstanding_bat_packet); |
| 165 | queue_delayed_work(bat_event_workqueue, |
| 166 | &forw_packet_aggr->delayed_work, |
| 167 | send_time - jiffies); |
| 168 | } |
| 169 | |
| 170 | /* aggregate a new packet into the existing aggregation */ |
| 171 | static void aggregate(struct forw_packet *forw_packet_aggr, |
| 172 | unsigned char *packet_buff, |
| 173 | int packet_len, |
| 174 | bool direct_link) |
| 175 | { |
| 176 | unsigned char *skb_buff; |
| 177 | |
| 178 | skb_buff = skb_put(forw_packet_aggr->skb, packet_len); |
| 179 | memcpy(skb_buff, packet_buff, packet_len); |
| 180 | forw_packet_aggr->packet_len += packet_len; |
| 181 | forw_packet_aggr->num_packets++; |
| 182 | |
| 183 | /* save packet direct link flag status */ |
| 184 | if (direct_link) |
| 185 | forw_packet_aggr->direct_link_flags |= |
| 186 | (1 << forw_packet_aggr->num_packets); |
| 187 | } |
| 188 | |
| 189 | void add_bat_packet_to_list(struct bat_priv *bat_priv, |
| 190 | unsigned char *packet_buff, int packet_len, |
Marek Lindner | e6c10f4 | 2011-02-18 12:33:20 +0000 | [diff] [blame] | 191 | struct hard_iface *if_incoming, char own_packet, |
Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 192 | unsigned long send_time) |
| 193 | { |
| 194 | /** |
| 195 | * _aggr -> pointer to the packet we want to aggregate with |
| 196 | * _pos -> pointer to the position in the queue |
| 197 | */ |
| 198 | struct forw_packet *forw_packet_aggr = NULL, *forw_packet_pos = NULL; |
| 199 | struct hlist_node *tmp_node; |
| 200 | struct batman_packet *batman_packet = |
| 201 | (struct batman_packet *)packet_buff; |
| 202 | bool direct_link = batman_packet->flags & DIRECTLINK ? 1 : 0; |
| 203 | |
| 204 | /* find position for the packet in the forward queue */ |
| 205 | spin_lock_bh(&bat_priv->forw_bat_list_lock); |
| 206 | /* own packets are not to be aggregated */ |
| 207 | if ((atomic_read(&bat_priv->aggregated_ogms)) && (!own_packet)) { |
| 208 | hlist_for_each_entry(forw_packet_pos, tmp_node, |
| 209 | &bat_priv->forw_bat_list, list) { |
| 210 | if (can_aggregate_with(batman_packet, |
| 211 | packet_len, |
| 212 | send_time, |
| 213 | direct_link, |
| 214 | if_incoming, |
| 215 | forw_packet_pos)) { |
| 216 | forw_packet_aggr = forw_packet_pos; |
| 217 | break; |
| 218 | } |
| 219 | } |
| 220 | } |
| 221 | |
| 222 | /* nothing to aggregate with - either aggregation disabled or no |
| 223 | * suitable aggregation packet found */ |
| 224 | if (!forw_packet_aggr) { |
| 225 | /* the following section can run without the lock */ |
| 226 | spin_unlock_bh(&bat_priv->forw_bat_list_lock); |
| 227 | |
| 228 | /** |
| 229 | * if we could not aggregate this packet with one of the others |
| 230 | * we hold it back for a while, so that it might be aggregated |
| 231 | * later on |
| 232 | */ |
| 233 | if ((!own_packet) && |
| 234 | (atomic_read(&bat_priv->aggregated_ogms))) |
| 235 | send_time += msecs_to_jiffies(MAX_AGGREGATION_MS); |
| 236 | |
| 237 | new_aggregated_packet(packet_buff, packet_len, |
| 238 | send_time, direct_link, |
| 239 | if_incoming, own_packet); |
| 240 | } else { |
| 241 | aggregate(forw_packet_aggr, |
| 242 | packet_buff, packet_len, |
| 243 | direct_link); |
| 244 | spin_unlock_bh(&bat_priv->forw_bat_list_lock); |
| 245 | } |
| 246 | } |
| 247 | |
| 248 | /* unpack the aggregated packets and process them one by one */ |
| 249 | void receive_aggr_bat_packet(struct ethhdr *ethhdr, unsigned char *packet_buff, |
Marek Lindner | e6c10f4 | 2011-02-18 12:33:20 +0000 | [diff] [blame] | 250 | int packet_len, struct hard_iface *if_incoming) |
Sven Eckelmann | c6c8fea | 2010-12-13 11:19:28 +0000 | [diff] [blame] | 251 | { |
| 252 | struct batman_packet *batman_packet; |
| 253 | int buff_pos = 0; |
| 254 | unsigned char *hna_buff; |
| 255 | |
| 256 | batman_packet = (struct batman_packet *)packet_buff; |
| 257 | |
| 258 | do { |
| 259 | /* network to host order for our 32bit seqno, and the |
| 260 | orig_interval. */ |
| 261 | batman_packet->seqno = ntohl(batman_packet->seqno); |
| 262 | |
| 263 | hna_buff = packet_buff + buff_pos + BAT_PACKET_LEN; |
| 264 | receive_bat_packet(ethhdr, batman_packet, |
| 265 | hna_buff, hna_len(batman_packet), |
| 266 | if_incoming); |
| 267 | |
| 268 | buff_pos += BAT_PACKET_LEN + hna_len(batman_packet); |
| 269 | batman_packet = (struct batman_packet *) |
| 270 | (packet_buff + buff_pos); |
| 271 | } while (aggregated_packet(buff_pos, packet_len, |
| 272 | batman_packet->num_hna)); |
| 273 | } |