blob: a018e42d117eb49b9b89716e70910f299de2ef68 [file] [log] [blame]
Avinash Patil838e4f42012-08-03 18:06:08 -07001/*
2 * Marvell Wireless LAN device driver: AP TX and RX data handling
3 *
4 * Copyright (C) 2012, Marvell International Ltd.
5 *
6 * This software file (the "File") is distributed by Marvell International
7 * Ltd. under the terms of the GNU General Public License Version 2, June 1991
8 * (the "License"). You may use, redistribute and/or modify this File in
9 * accordance with the terms and conditions of the License, a copy of which
10 * is available by writing to the Free Software Foundation, Inc.,
11 * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA or on the
12 * worldwide web at http://www.gnu.org/licenses/old-licenses/gpl-2.0.txt.
13 *
14 * THE FILE IS DISTRIBUTED AS-IS, WITHOUT WARRANTY OF ANY KIND, AND THE
15 * IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE
16 * ARE EXPRESSLY DISCLAIMED. The License provides additional details about
17 * this warranty disclaimer.
18 */
19
20#include "decl.h"
21#include "ioctl.h"
22#include "main.h"
23#include "wmm.h"
Avinash Patild1cf3b92012-08-03 18:06:09 -070024#include "11n_aggr.h"
25#include "11n_rxreorder.h"
Avinash Patil838e4f42012-08-03 18:06:08 -070026
27static void mwifiex_uap_queue_bridged_pkt(struct mwifiex_private *priv,
28 struct sk_buff *skb)
29{
30 struct mwifiex_adapter *adapter = priv->adapter;
31 struct uap_rxpd *uap_rx_pd;
32 struct rx_packet_hdr *rx_pkt_hdr;
33 struct sk_buff *new_skb;
34 struct mwifiex_txinfo *tx_info;
35 int hdr_chop;
36 struct timeval tv;
37 u8 rfc1042_eth_hdr[ETH_ALEN] = { 0xaa, 0xaa, 0x03, 0x00, 0x00, 0x00 };
38
39 uap_rx_pd = (struct uap_rxpd *)(skb->data);
40 rx_pkt_hdr = (void *)uap_rx_pd + le16_to_cpu(uap_rx_pd->rx_pkt_offset);
41
42 if ((atomic_read(&adapter->pending_bridged_pkts) >=
43 MWIFIEX_BRIDGED_PKTS_THRESHOLD)) {
44 dev_err(priv->adapter->dev,
45 "Tx: Bridge packet limit reached. Drop packet!\n");
46 kfree_skb(skb);
47 return;
48 }
49
50 if (!memcmp(&rx_pkt_hdr->rfc1042_hdr,
51 rfc1042_eth_hdr, sizeof(rfc1042_eth_hdr)))
52 /* Chop off the rxpd + the excess memory from
53 * 802.2/llc/snap header that was removed.
54 */
55 hdr_chop = (u8 *)eth_hdr - (u8 *)uap_rx_pd;
56 else
57 /* Chop off the rxpd */
58 hdr_chop = (u8 *)&rx_pkt_hdr->eth803_hdr - (u8 *)uap_rx_pd;
59
60 /* Chop off the leading header bytes so the it points
61 * to the start of either the reconstructed EthII frame
62 * or the 802.2/llc/snap frame.
63 */
64 skb_pull(skb, hdr_chop);
65
66 if (skb_headroom(skb) < MWIFIEX_MIN_DATA_HEADER_LEN) {
67 dev_dbg(priv->adapter->dev,
68 "data: Tx: insufficient skb headroom %d\n",
69 skb_headroom(skb));
70 /* Insufficient skb headroom - allocate a new skb */
71 new_skb =
72 skb_realloc_headroom(skb, MWIFIEX_MIN_DATA_HEADER_LEN);
73 if (unlikely(!new_skb)) {
74 dev_err(priv->adapter->dev,
75 "Tx: cannot allocate new_skb\n");
76 kfree_skb(skb);
77 priv->stats.tx_dropped++;
78 return;
79 }
80
81 kfree_skb(skb);
82 skb = new_skb;
83 dev_dbg(priv->adapter->dev, "info: new skb headroom %d\n",
84 skb_headroom(skb));
85 }
86
87 tx_info = MWIFIEX_SKB_TXCB(skb);
88 tx_info->bss_num = priv->bss_num;
89 tx_info->bss_type = priv->bss_type;
90 tx_info->flags |= MWIFIEX_BUF_FLAG_BRIDGED_PKT;
91
92 do_gettimeofday(&tv);
93 skb->tstamp = timeval_to_ktime(tv);
94 mwifiex_wmm_add_buf_txqueue(priv, skb);
95 atomic_inc(&adapter->tx_pending);
96 atomic_inc(&adapter->pending_bridged_pkts);
97
98 if ((atomic_read(&adapter->tx_pending) >= MAX_TX_PENDING)) {
99 mwifiex_set_trans_start(priv->netdev);
100 mwifiex_stop_net_dev_queue(priv->netdev, priv->adapter);
101 }
102 return;
103}
104
105/*
106 * This function contains logic for AP packet forwarding.
107 *
108 * If a packet is multicast/broadcast, it is sent to kernel/upper layer
109 * as well as queued back to AP TX queue so that it can be sent to other
110 * associated stations.
111 * If a packet is unicast and RA is present in associated station list,
112 * it is again requeued into AP TX queue.
113 * If a packet is unicast and RA is not in associated station list,
114 * packet is forwarded to kernel to handle routing logic.
115 */
116int mwifiex_handle_uap_rx_forward(struct mwifiex_private *priv,
117 struct sk_buff *skb)
118{
119 struct mwifiex_adapter *adapter = priv->adapter;
120 struct uap_rxpd *uap_rx_pd;
121 struct rx_packet_hdr *rx_pkt_hdr;
122 u8 ra[ETH_ALEN];
123 struct sk_buff *skb_uap;
124
125 uap_rx_pd = (struct uap_rxpd *)(skb->data);
126 rx_pkt_hdr = (void *)uap_rx_pd + le16_to_cpu(uap_rx_pd->rx_pkt_offset);
127
128 /* don't do packet forwarding in disconnected state */
129 if (!priv->media_connected) {
130 dev_err(adapter->dev, "drop packet in disconnected state.\n");
131 dev_kfree_skb_any(skb);
132 return 0;
133 }
134
135 memcpy(ra, rx_pkt_hdr->eth803_hdr.h_dest, ETH_ALEN);
136
137 if (is_multicast_ether_addr(ra)) {
138 skb_uap = skb_copy(skb, GFP_ATOMIC);
139 mwifiex_uap_queue_bridged_pkt(priv, skb_uap);
140 } else {
141 if (mwifiex_get_sta_entry(priv, ra)) {
142 /* Requeue Intra-BSS packet */
143 mwifiex_uap_queue_bridged_pkt(priv, skb);
144 return 0;
145 }
146 }
147
148 /* Forward unicat/Inter-BSS packets to kernel. */
Avinash Patilf3b369e2012-10-19 19:19:21 -0700149 return mwifiex_process_rx_packet(priv, skb);
Avinash Patil838e4f42012-08-03 18:06:08 -0700150}
151
152/*
153 * This function processes the packet received on AP interface.
154 *
155 * The function looks into the RxPD and performs sanity tests on the
156 * received buffer to ensure its a valid packet before processing it
157 * further. If the packet is determined to be aggregated, it is
158 * de-aggregated accordingly. Then skb is passed to AP packet forwarding logic.
159 *
160 * The completion callback is called after processing is complete.
161 */
Avinash Patilf3b369e2012-10-19 19:19:21 -0700162int mwifiex_process_uap_rx_packet(struct mwifiex_private *priv,
Avinash Patil838e4f42012-08-03 18:06:08 -0700163 struct sk_buff *skb)
164{
Avinash Patilf3b369e2012-10-19 19:19:21 -0700165 struct mwifiex_adapter *adapter = priv->adapter;
Avinash Patil838e4f42012-08-03 18:06:08 -0700166 int ret;
167 struct uap_rxpd *uap_rx_pd;
Avinash Patil838e4f42012-08-03 18:06:08 -0700168 struct rx_packet_hdr *rx_pkt_hdr;
169 u16 rx_pkt_type;
Avinash Patild1cf3b92012-08-03 18:06:09 -0700170 u8 ta[ETH_ALEN], pkt_type;
171 struct mwifiex_sta_node *node;
172
Avinash Patil838e4f42012-08-03 18:06:08 -0700173 uap_rx_pd = (struct uap_rxpd *)(skb->data);
174 rx_pkt_type = le16_to_cpu(uap_rx_pd->rx_pkt_type);
175 rx_pkt_hdr = (void *)uap_rx_pd + le16_to_cpu(uap_rx_pd->rx_pkt_offset);
176
177 if ((le16_to_cpu(uap_rx_pd->rx_pkt_offset) +
178 le16_to_cpu(uap_rx_pd->rx_pkt_length)) > (u16) skb->len) {
179 dev_err(adapter->dev,
180 "wrong rx packet: len=%d, offset=%d, length=%d\n",
181 skb->len, le16_to_cpu(uap_rx_pd->rx_pkt_offset),
182 le16_to_cpu(uap_rx_pd->rx_pkt_length));
183 priv->stats.rx_dropped++;
184
185 if (adapter->if_ops.data_complete)
186 adapter->if_ops.data_complete(adapter, skb);
187 else
188 dev_kfree_skb_any(skb);
189
190 return 0;
191 }
Avinash Patil838e4f42012-08-03 18:06:08 -0700192
Avinash Patild1cf3b92012-08-03 18:06:09 -0700193 if (le16_to_cpu(uap_rx_pd->rx_pkt_type) == PKT_TYPE_AMSDU) {
194 struct sk_buff_head list;
195 struct sk_buff *rx_skb;
196
197 __skb_queue_head_init(&list);
198 skb_pull(skb, le16_to_cpu(uap_rx_pd->rx_pkt_offset));
199 skb_trim(skb, le16_to_cpu(uap_rx_pd->rx_pkt_length));
200
201 ieee80211_amsdu_to_8023s(skb, &list, priv->curr_addr,
202 priv->wdev->iftype, 0, false);
203
204 while (!skb_queue_empty(&list)) {
205 rx_skb = __skb_dequeue(&list);
Avinash Patilf3b369e2012-10-19 19:19:21 -0700206 ret = mwifiex_recv_packet(priv, rx_skb);
Avinash Patild1cf3b92012-08-03 18:06:09 -0700207 if (ret)
208 dev_err(adapter->dev,
209 "AP:Rx A-MSDU failed");
210 }
211
212 return 0;
Stone Piao2dbaf752012-09-25 20:23:35 -0700213 } else if (rx_pkt_type == PKT_TYPE_MGMT) {
Avinash Patilf3b369e2012-10-19 19:19:21 -0700214 ret = mwifiex_process_mgmt_packet(priv, skb);
Stone Piao2dbaf752012-09-25 20:23:35 -0700215 if (ret)
216 dev_err(adapter->dev, "Rx of mgmt packet failed");
217 dev_kfree_skb_any(skb);
218 return ret;
Avinash Patild1cf3b92012-08-03 18:06:09 -0700219 }
220
221 memcpy(ta, rx_pkt_hdr->eth803_hdr.h_source, ETH_ALEN);
222
223 if (rx_pkt_type != PKT_TYPE_BAR && uap_rx_pd->priority < MAX_NUM_TID) {
224 node = mwifiex_get_sta_entry(priv, ta);
225 if (node)
226 node->rx_seq[uap_rx_pd->priority] =
227 le16_to_cpu(uap_rx_pd->seq_num);
228 }
229
230 if (!priv->ap_11n_enabled ||
231 (!mwifiex_11n_get_rx_reorder_tbl(priv, uap_rx_pd->priority, ta) &&
232 (le16_to_cpu(uap_rx_pd->rx_pkt_type) != PKT_TYPE_AMSDU))) {
233 ret = mwifiex_handle_uap_rx_forward(priv, skb);
234 return ret;
235 }
236
237 /* Reorder and send to kernel */
238 pkt_type = (u8)le16_to_cpu(uap_rx_pd->rx_pkt_type);
239 ret = mwifiex_11n_rx_reorder_pkt(priv, le16_to_cpu(uap_rx_pd->seq_num),
240 uap_rx_pd->priority, ta, pkt_type,
241 skb);
242
243 if (ret || (rx_pkt_type == PKT_TYPE_BAR)) {
Avinash Patil838e4f42012-08-03 18:06:08 -0700244 if (adapter->if_ops.data_complete)
245 adapter->if_ops.data_complete(adapter, skb);
246 else
247 dev_kfree_skb_any(skb);
248 }
249
Avinash Patild1cf3b92012-08-03 18:06:09 -0700250 if (ret)
251 priv->stats.rx_dropped++;
252
Avinash Patil838e4f42012-08-03 18:06:08 -0700253 return ret;
254}
Avinash Patil4ac87642012-09-10 18:30:49 -0700255
256/*
257 * This function fills the TxPD for AP tx packets.
258 *
259 * The Tx buffer received by this function should already have the
260 * header space allocated for TxPD.
261 *
262 * This function inserts the TxPD in between interface header and actual
263 * data and adjusts the buffer pointers accordingly.
264 *
265 * The following TxPD fields are set by this function, as required -
266 * - BSS number
267 * - Tx packet length and offset
268 * - Priority
269 * - Packet delay
270 * - Priority specific Tx control
271 * - Flags
272 */
273void *mwifiex_process_uap_txpd(struct mwifiex_private *priv,
274 struct sk_buff *skb)
275{
276 struct mwifiex_adapter *adapter = priv->adapter;
277 struct uap_txpd *txpd;
278 struct mwifiex_txinfo *tx_info = MWIFIEX_SKB_TXCB(skb);
279 int pad, len;
Stone Piao32152152012-09-25 20:23:44 -0700280 u16 pkt_type;
Avinash Patil4ac87642012-09-10 18:30:49 -0700281
282 if (!skb->len) {
283 dev_err(adapter->dev, "Tx: bad packet length: %d\n", skb->len);
284 tx_info->status_code = -1;
285 return skb->data;
286 }
287
Stone Piao32152152012-09-25 20:23:44 -0700288 pkt_type = mwifiex_is_skb_mgmt_frame(skb) ? PKT_TYPE_MGMT : 0;
289
Avinash Patil4ac87642012-09-10 18:30:49 -0700290 /* If skb->data is not aligned, add padding */
291 pad = (4 - (((void *)skb->data - NULL) & 0x3)) % 4;
292
293 len = sizeof(*txpd) + pad;
294
295 BUG_ON(skb_headroom(skb) < len + INTF_HEADER_LEN);
296
297 skb_push(skb, len);
298
299 txpd = (struct uap_txpd *)skb->data;
300 memset(txpd, 0, sizeof(*txpd));
301 txpd->bss_num = priv->bss_num;
302 txpd->bss_type = priv->bss_type;
303 txpd->tx_pkt_length = cpu_to_le16((u16)(skb->len - len));
304
305 txpd->priority = (u8)skb->priority;
306 txpd->pkt_delay_2ms = mwifiex_wmm_compute_drv_pkt_delay(priv, skb);
307
308 if (txpd->priority < ARRAY_SIZE(priv->wmm.user_pri_pkt_tx_ctrl))
309 /*
310 * Set the priority specific tx_control field, setting of 0 will
311 * cause the default value to be used later in this function.
312 */
313 txpd->tx_control =
314 cpu_to_le32(priv->wmm.user_pri_pkt_tx_ctrl[txpd->priority]);
315
316 /* Offset of actual data */
Stone Piao32152152012-09-25 20:23:44 -0700317 if (pkt_type == PKT_TYPE_MGMT) {
318 /* Set the packet type and add header for management frame */
319 txpd->tx_pkt_type = cpu_to_le16(pkt_type);
320 len += MWIFIEX_MGMT_FRAME_HEADER_SIZE;
321 }
322
Avinash Patil4ac87642012-09-10 18:30:49 -0700323 txpd->tx_pkt_offset = cpu_to_le16(len);
324
325 /* make space for INTF_HEADER_LEN */
326 skb_push(skb, INTF_HEADER_LEN);
327
328 if (!txpd->tx_control)
329 /* TxCtrl set by user or default */
330 txpd->tx_control = cpu_to_le32(priv->pkt_tx_ctrl);
331
332 return skb->data;
333}