blob: a056b12fbd1c2df694fc6c723d602d6dc3322e1a [file] [log] [blame]
Ivo van Doorn95ea3622007-09-25 17:57:13 -07001/*
Ivo van Doorn811aa9c2008-02-03 15:42:53 +01002 Copyright (C) 2004 - 2008 rt2x00 SourceForge Project
Ivo van Doorn95ea3622007-09-25 17:57:13 -07003 <http://rt2x00.serialmonkey.com>
4
5 This program is free software; you can redistribute it and/or modify
6 it under the terms of the GNU General Public License as published by
7 the Free Software Foundation; either version 2 of the License, or
8 (at your option) any later version.
9
10 This program is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 GNU General Public License for more details.
14
15 You should have received a copy of the GNU General Public License
16 along with this program; if not, write to the
17 Free Software Foundation, Inc.,
18 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
19 */
20
21/*
22 Module: rt2x00pci
23 Abstract: rt2x00 generic pci device routines.
24 */
25
Ivo van Doorn95ea3622007-09-25 17:57:13 -070026#include <linux/dma-mapping.h>
27#include <linux/kernel.h>
28#include <linux/module.h>
29#include <linux/pci.h>
30
31#include "rt2x00.h"
32#include "rt2x00pci.h"
33
34/*
Ivo van Doorn95ea3622007-09-25 17:57:13 -070035 * TX data handlers.
36 */
37int rt2x00pci_write_tx_data(struct rt2x00_dev *rt2x00dev,
Ivo van Doorn181d6902008-02-05 16:42:23 -050038 struct data_queue *queue, struct sk_buff *skb,
Ivo van Doorn95ea3622007-09-25 17:57:13 -070039 struct ieee80211_tx_control *control)
40{
Ivo van Doorn181d6902008-02-05 16:42:23 -050041 struct queue_entry *entry = rt2x00queue_get_entry(queue, Q_INDEX);
42 struct queue_entry_priv_pci_tx *priv_tx = entry->priv_data;
43 struct skb_frame_desc *skbdesc;
Ivo van Doorn7050ec82008-05-10 13:46:13 +020044 struct txentry_desc txdesc;
Ivo van Doorn95ea3622007-09-25 17:57:13 -070045 u32 word;
46
Ivo van Doorn181d6902008-02-05 16:42:23 -050047 if (rt2x00queue_full(queue))
Ivo van Doorn95ea3622007-09-25 17:57:13 -070048 return -EINVAL;
Ivo van Doorn95ea3622007-09-25 17:57:13 -070049
Ivo van Doorn181d6902008-02-05 16:42:23 -050050 rt2x00_desc_read(priv_tx->desc, 0, &word);
Ivo van Doorn95ea3622007-09-25 17:57:13 -070051
52 if (rt2x00_get_field32(word, TXD_ENTRY_OWNER_NIC) ||
53 rt2x00_get_field32(word, TXD_ENTRY_VALID)) {
54 ERROR(rt2x00dev,
55 "Arrived at non-free entry in the non-full queue %d.\n"
56 "Please file bug report to %s.\n",
Ivo van Doorne58c6ac2008-04-21 19:00:47 +020057 entry->queue->qid, DRV_PROJECT);
Ivo van Doorn95ea3622007-09-25 17:57:13 -070058 return -EINVAL;
59 }
60
Ivo van Doorn08992f72008-01-24 01:56:25 -080061 /*
Ivo van Doorn7050ec82008-05-10 13:46:13 +020062 * Copy all TX descriptor information into txdesc,
63 * after that we are free to use the skb->cb array
64 * for our information.
65 */
66 entry->skb = skb;
67 rt2x00queue_create_tx_descriptor(entry, &txdesc, control);
68
69 /*
Ivo van Doorn08992f72008-01-24 01:56:25 -080070 * Fill in skb descriptor
71 */
Ivo van Doorn181d6902008-02-05 16:42:23 -050072 skbdesc = get_skb_frame_desc(skb);
Ivo van Doorn181d6902008-02-05 16:42:23 -050073 skbdesc->data = skb->data;
Ivo van Doorn647d0ca2008-02-10 22:51:21 +010074 skbdesc->data_len = skb->len;
Ivo van Doorn181d6902008-02-05 16:42:23 -050075 skbdesc->desc = priv_tx->desc;
76 skbdesc->desc_len = queue->desc_size;
77 skbdesc->entry = entry;
Ivo van Doorn08992f72008-01-24 01:56:25 -080078
Mattias Nissler92f5ac62008-03-09 22:38:54 +010079 memcpy(&priv_tx->control, control, sizeof(priv_tx->control));
Ivo van Doorn181d6902008-02-05 16:42:23 -050080 memcpy(priv_tx->data, skb->data, skb->len);
Ivo van Doorn95ea3622007-09-25 17:57:13 -070081
Ivo van Doorn7050ec82008-05-10 13:46:13 +020082 rt2x00queue_write_tx_descriptor(entry, &txdesc);
Ivo van Doorn181d6902008-02-05 16:42:23 -050083 rt2x00queue_index_inc(queue, Q_INDEX);
Ivo van Doorn95ea3622007-09-25 17:57:13 -070084
Ivo van Doorn95ea3622007-09-25 17:57:13 -070085 return 0;
86}
87EXPORT_SYMBOL_GPL(rt2x00pci_write_tx_data);
88
89/*
Ivo van Doorn3957ccb2007-11-12 15:02:40 +010090 * TX/RX data handlers.
Ivo van Doorn95ea3622007-09-25 17:57:13 -070091 */
92void rt2x00pci_rxdone(struct rt2x00_dev *rt2x00dev)
93{
Ivo van Doorn181d6902008-02-05 16:42:23 -050094 struct data_queue *queue = rt2x00dev->rx;
95 struct queue_entry *entry;
96 struct queue_entry_priv_pci_rx *priv_rx;
Ivo van Doornc5d0dc52008-01-06 23:40:27 +010097 struct ieee80211_hdr *hdr;
Ivo van Doorn181d6902008-02-05 16:42:23 -050098 struct skb_frame_desc *skbdesc;
99 struct rxdone_entry_desc rxdesc;
Ivo van Doornc5d0dc52008-01-06 23:40:27 +0100100 int header_size;
101 int align;
Johannes Berg4150c572007-09-17 01:29:23 -0400102 u32 word;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700103
104 while (1) {
Ivo van Doorn181d6902008-02-05 16:42:23 -0500105 entry = rt2x00queue_get_entry(queue, Q_INDEX);
106 priv_rx = entry->priv_data;
107 rt2x00_desc_read(priv_rx->desc, 0, &word);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700108
Johannes Berg4150c572007-09-17 01:29:23 -0400109 if (rt2x00_get_field32(word, RXD_ENTRY_OWNER_NIC))
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700110 break;
111
Ivo van Doorn181d6902008-02-05 16:42:23 -0500112 memset(&rxdesc, 0, sizeof(rxdesc));
113 rt2x00dev->ops->lib->fill_rxdone(entry, &rxdesc);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700114
Ivo van Doorn181d6902008-02-05 16:42:23 -0500115 hdr = (struct ieee80211_hdr *)priv_rx->data;
Ivo van Doornc5d0dc52008-01-06 23:40:27 +0100116 header_size =
117 ieee80211_get_hdrlen(le16_to_cpu(hdr->frame_control));
118
119 /*
120 * The data behind the ieee80211 header must be
121 * aligned on a 4 byte boundary.
122 */
Ivo van Doornd101f642008-01-11 20:53:07 +0100123 align = header_size % 4;
Ivo van Doornc5d0dc52008-01-06 23:40:27 +0100124
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700125 /*
126 * Allocate the sk_buffer, initialize it and copy
127 * all data into it.
128 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500129 entry->skb = dev_alloc_skb(rxdesc.size + align);
130 if (!entry->skb)
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700131 return;
132
Ivo van Doorn181d6902008-02-05 16:42:23 -0500133 skb_reserve(entry->skb, align);
134 memcpy(skb_put(entry->skb, rxdesc.size),
135 priv_rx->data, rxdesc.size);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700136
137 /*
Ivo van Doorn08992f72008-01-24 01:56:25 -0800138 * Fill in skb descriptor
139 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500140 skbdesc = get_skb_frame_desc(entry->skb);
141 memset(skbdesc, 0, sizeof(*skbdesc));
142 skbdesc->data = entry->skb->data;
Ivo van Doorn647d0ca2008-02-10 22:51:21 +0100143 skbdesc->data_len = entry->skb->len;
Ivo van Doorn181d6902008-02-05 16:42:23 -0500144 skbdesc->desc = priv_rx->desc;
145 skbdesc->desc_len = queue->desc_size;
Ivo van Doorn08992f72008-01-24 01:56:25 -0800146 skbdesc->entry = entry;
147
148 /*
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700149 * Send the frame to rt2x00lib for further processing.
150 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500151 rt2x00lib_rxdone(entry, &rxdesc);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700152
Ivo van Doorn181d6902008-02-05 16:42:23 -0500153 if (test_bit(DEVICE_ENABLED_RADIO, &queue->rt2x00dev->flags)) {
Johannes Berg4150c572007-09-17 01:29:23 -0400154 rt2x00_set_field32(&word, RXD_ENTRY_OWNER_NIC, 1);
Ivo van Doorn181d6902008-02-05 16:42:23 -0500155 rt2x00_desc_write(priv_rx->desc, 0, word);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700156 }
157
Ivo van Doorn181d6902008-02-05 16:42:23 -0500158 rt2x00queue_index_inc(queue, Q_INDEX);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700159 }
160}
161EXPORT_SYMBOL_GPL(rt2x00pci_rxdone);
162
Ivo van Doorn181d6902008-02-05 16:42:23 -0500163void rt2x00pci_txdone(struct rt2x00_dev *rt2x00dev, struct queue_entry *entry,
164 struct txdone_entry_desc *txdesc)
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100165{
Ivo van Doorn181d6902008-02-05 16:42:23 -0500166 struct queue_entry_priv_pci_tx *priv_tx = entry->priv_data;
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100167 u32 word;
168
Ivo van Doorn181d6902008-02-05 16:42:23 -0500169 txdesc->control = &priv_tx->control;
170 rt2x00lib_txdone(entry, txdesc);
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100171
172 /*
173 * Make this entry available for reuse.
174 */
175 entry->flags = 0;
176
Ivo van Doorn181d6902008-02-05 16:42:23 -0500177 rt2x00_desc_read(priv_tx->desc, 0, &word);
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100178 rt2x00_set_field32(&word, TXD_ENTRY_OWNER_NIC, 0);
179 rt2x00_set_field32(&word, TXD_ENTRY_VALID, 0);
Ivo van Doorn181d6902008-02-05 16:42:23 -0500180 rt2x00_desc_write(priv_tx->desc, 0, word);
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100181
Ivo van Doorn181d6902008-02-05 16:42:23 -0500182 rt2x00queue_index_inc(entry->queue, Q_INDEX_DONE);
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100183
184 /*
Ivo van Doorn181d6902008-02-05 16:42:23 -0500185 * If the data queue was full before the txdone handler
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100186 * we must make sure the packet queue in the mac80211 stack
187 * is reenabled when the txdone handler has finished.
188 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500189 if (!rt2x00queue_full(entry->queue))
190 ieee80211_wake_queue(rt2x00dev->hw, priv_tx->control.queue);
Ivo van Doorn3957ccb2007-11-12 15:02:40 +0100191
192}
193EXPORT_SYMBOL_GPL(rt2x00pci_txdone);
194
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700195/*
196 * Device initialization handlers.
197 */
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100198#define desc_size(__queue) \
199({ \
200 ((__queue)->limit * (__queue)->desc_size);\
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700201})
202
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100203#define data_size(__queue) \
204({ \
205 ((__queue)->limit * (__queue)->data_size);\
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700206})
207
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100208#define dma_size(__queue) \
209({ \
210 data_size(__queue) + desc_size(__queue);\
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700211})
212
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100213#define desc_offset(__queue, __base, __i) \
214({ \
215 (__base) + data_size(__queue) + \
216 ((__i) * (__queue)->desc_size); \
217})
218
219#define data_offset(__queue, __base, __i) \
220({ \
221 (__base) + \
222 ((__i) * (__queue)->data_size); \
Ivo van Doorn181d6902008-02-05 16:42:23 -0500223})
224
225static int rt2x00pci_alloc_queue_dma(struct rt2x00_dev *rt2x00dev,
226 struct data_queue *queue)
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700227{
Ivo van Doorn181d6902008-02-05 16:42:23 -0500228 struct pci_dev *pci_dev = rt2x00dev_pci(rt2x00dev);
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100229 struct queue_entry_priv_pci_rx *priv_rx;
Ivo van Doorn181d6902008-02-05 16:42:23 -0500230 struct queue_entry_priv_pci_tx *priv_tx;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100231 void *addr;
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100232 dma_addr_t dma;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100233 void *desc_addr;
234 dma_addr_t desc_dma;
235 void *data_addr;
236 dma_addr_t data_dma;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700237 unsigned int i;
238
239 /*
240 * Allocate DMA memory for descriptor and buffer.
241 */
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100242 addr = pci_alloc_consistent(pci_dev, dma_size(queue), &dma);
243 if (!addr)
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700244 return -ENOMEM;
245
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100246 memset(addr, 0, dma_size(queue));
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100247
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700248 /*
Ivo van Doorn181d6902008-02-05 16:42:23 -0500249 * Initialize all queue entries to contain valid addresses.
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700250 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500251 for (i = 0; i < queue->limit; i++) {
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100252 desc_addr = desc_offset(queue, addr, i);
253 desc_dma = desc_offset(queue, dma, i);
254 data_addr = data_offset(queue, addr, i);
255 data_dma = data_offset(queue, dma, i);
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100256
257 if (queue->qid == QID_RX) {
258 priv_rx = queue->entries[i].priv_data;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100259 priv_rx->desc = desc_addr;
260 priv_rx->desc_dma = desc_dma;
261 priv_rx->data = data_addr;
262 priv_rx->data_dma = data_dma;
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100263 } else {
264 priv_tx = queue->entries[i].priv_data;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100265 priv_tx->desc = desc_addr;
266 priv_tx->desc_dma = desc_dma;
267 priv_tx->data = data_addr;
268 priv_tx->data_dma = data_dma;
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100269 }
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700270 }
271
272 return 0;
273}
274
Ivo van Doorn181d6902008-02-05 16:42:23 -0500275static void rt2x00pci_free_queue_dma(struct rt2x00_dev *rt2x00dev,
276 struct data_queue *queue)
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700277{
Ivo van Doorn181d6902008-02-05 16:42:23 -0500278 struct pci_dev *pci_dev = rt2x00dev_pci(rt2x00dev);
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100279 struct queue_entry_priv_pci_rx *priv_rx;
280 struct queue_entry_priv_pci_tx *priv_tx;
281 void *data_addr;
282 dma_addr_t data_dma;
Ivo van Doorn181d6902008-02-05 16:42:23 -0500283
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100284 if (queue->qid == QID_RX) {
285 priv_rx = queue->entries[0].priv_data;
286 data_addr = priv_rx->data;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100287 data_dma = priv_rx->data_dma;
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100288
289 priv_rx->data = NULL;
290 } else {
291 priv_tx = queue->entries[0].priv_data;
292 data_addr = priv_tx->data;
Ivo van Doorn30b3a232008-02-17 17:33:24 +0100293 data_dma = priv_tx->data_dma;
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100294
295 priv_tx->data = NULL;
296 }
297
298 if (data_addr)
Ivo van Doorn181d6902008-02-05 16:42:23 -0500299 pci_free_consistent(pci_dev, dma_size(queue),
Ivo van Doorn9c9dd2c2008-02-10 22:46:52 +0100300 data_addr, data_dma);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700301}
302
303int rt2x00pci_initialize(struct rt2x00_dev *rt2x00dev)
304{
305 struct pci_dev *pci_dev = rt2x00dev_pci(rt2x00dev);
Ivo van Doorn181d6902008-02-05 16:42:23 -0500306 struct data_queue *queue;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700307 int status;
308
309 /*
310 * Allocate DMA
311 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500312 queue_for_each(rt2x00dev, queue) {
313 status = rt2x00pci_alloc_queue_dma(rt2x00dev, queue);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700314 if (status)
315 goto exit;
316 }
317
318 /*
319 * Register interrupt handler.
320 */
321 status = request_irq(pci_dev->irq, rt2x00dev->ops->lib->irq_handler,
322 IRQF_SHARED, pci_name(pci_dev), rt2x00dev);
323 if (status) {
324 ERROR(rt2x00dev, "IRQ %d allocation failed (error %d).\n",
325 pci_dev->irq, status);
Ivo van Doornb30cdfc2008-05-05 17:24:03 +0200326 goto exit;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700327 }
328
329 return 0;
330
331exit:
Ivo van Doornb30cdfc2008-05-05 17:24:03 +0200332 queue_for_each(rt2x00dev, queue)
333 rt2x00pci_free_queue_dma(rt2x00dev, queue);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700334
335 return status;
336}
337EXPORT_SYMBOL_GPL(rt2x00pci_initialize);
338
339void rt2x00pci_uninitialize(struct rt2x00_dev *rt2x00dev)
340{
Ivo van Doorn181d6902008-02-05 16:42:23 -0500341 struct data_queue *queue;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700342
343 /*
344 * Free irq line.
345 */
346 free_irq(rt2x00dev_pci(rt2x00dev)->irq, rt2x00dev);
347
348 /*
349 * Free DMA
350 */
Ivo van Doorn181d6902008-02-05 16:42:23 -0500351 queue_for_each(rt2x00dev, queue)
352 rt2x00pci_free_queue_dma(rt2x00dev, queue);
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700353}
354EXPORT_SYMBOL_GPL(rt2x00pci_uninitialize);
355
356/*
357 * PCI driver handlers.
358 */
359static void rt2x00pci_free_reg(struct rt2x00_dev *rt2x00dev)
360{
361 kfree(rt2x00dev->rf);
362 rt2x00dev->rf = NULL;
363
364 kfree(rt2x00dev->eeprom);
365 rt2x00dev->eeprom = NULL;
366
Ivo van Doorn21795092008-02-10 22:49:13 +0100367 if (rt2x00dev->csr.base) {
368 iounmap(rt2x00dev->csr.base);
369 rt2x00dev->csr.base = NULL;
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700370 }
371}
372
373static int rt2x00pci_alloc_reg(struct rt2x00_dev *rt2x00dev)
374{
375 struct pci_dev *pci_dev = rt2x00dev_pci(rt2x00dev);
376
Ivo van Doorn21795092008-02-10 22:49:13 +0100377 rt2x00dev->csr.base = ioremap(pci_resource_start(pci_dev, 0),
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700378 pci_resource_len(pci_dev, 0));
Ivo van Doorn21795092008-02-10 22:49:13 +0100379 if (!rt2x00dev->csr.base)
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700380 goto exit;
381
382 rt2x00dev->eeprom = kzalloc(rt2x00dev->ops->eeprom_size, GFP_KERNEL);
383 if (!rt2x00dev->eeprom)
384 goto exit;
385
386 rt2x00dev->rf = kzalloc(rt2x00dev->ops->rf_size, GFP_KERNEL);
387 if (!rt2x00dev->rf)
388 goto exit;
389
390 return 0;
391
392exit:
393 ERROR_PROBE("Failed to allocate registers.\n");
394
395 rt2x00pci_free_reg(rt2x00dev);
396
397 return -ENOMEM;
398}
399
400int rt2x00pci_probe(struct pci_dev *pci_dev, const struct pci_device_id *id)
401{
402 struct rt2x00_ops *ops = (struct rt2x00_ops *)id->driver_data;
403 struct ieee80211_hw *hw;
404 struct rt2x00_dev *rt2x00dev;
405 int retval;
406
407 retval = pci_request_regions(pci_dev, pci_name(pci_dev));
408 if (retval) {
409 ERROR_PROBE("PCI request regions failed.\n");
410 return retval;
411 }
412
413 retval = pci_enable_device(pci_dev);
414 if (retval) {
415 ERROR_PROBE("Enable device failed.\n");
416 goto exit_release_regions;
417 }
418
419 pci_set_master(pci_dev);
420
421 if (pci_set_mwi(pci_dev))
422 ERROR_PROBE("MWI not available.\n");
423
424 if (pci_set_dma_mask(pci_dev, DMA_64BIT_MASK) &&
425 pci_set_dma_mask(pci_dev, DMA_32BIT_MASK)) {
426 ERROR_PROBE("PCI DMA not supported.\n");
427 retval = -EIO;
428 goto exit_disable_device;
429 }
430
431 hw = ieee80211_alloc_hw(sizeof(struct rt2x00_dev), ops->hw);
432 if (!hw) {
433 ERROR_PROBE("Failed to allocate hardware.\n");
434 retval = -ENOMEM;
435 goto exit_disable_device;
436 }
437
438 pci_set_drvdata(pci_dev, hw);
439
440 rt2x00dev = hw->priv;
441 rt2x00dev->dev = pci_dev;
442 rt2x00dev->ops = ops;
443 rt2x00dev->hw = hw;
444
445 retval = rt2x00pci_alloc_reg(rt2x00dev);
446 if (retval)
447 goto exit_free_device;
448
449 retval = rt2x00lib_probe_dev(rt2x00dev);
450 if (retval)
451 goto exit_free_reg;
452
453 return 0;
454
455exit_free_reg:
456 rt2x00pci_free_reg(rt2x00dev);
457
458exit_free_device:
459 ieee80211_free_hw(hw);
460
461exit_disable_device:
462 if (retval != -EBUSY)
463 pci_disable_device(pci_dev);
464
465exit_release_regions:
466 pci_release_regions(pci_dev);
467
468 pci_set_drvdata(pci_dev, NULL);
469
470 return retval;
471}
472EXPORT_SYMBOL_GPL(rt2x00pci_probe);
473
474void rt2x00pci_remove(struct pci_dev *pci_dev)
475{
476 struct ieee80211_hw *hw = pci_get_drvdata(pci_dev);
477 struct rt2x00_dev *rt2x00dev = hw->priv;
478
479 /*
480 * Free all allocated data.
481 */
482 rt2x00lib_remove_dev(rt2x00dev);
483 rt2x00pci_free_reg(rt2x00dev);
484 ieee80211_free_hw(hw);
485
486 /*
487 * Free the PCI device data.
488 */
489 pci_set_drvdata(pci_dev, NULL);
490 pci_disable_device(pci_dev);
491 pci_release_regions(pci_dev);
492}
493EXPORT_SYMBOL_GPL(rt2x00pci_remove);
494
495#ifdef CONFIG_PM
496int rt2x00pci_suspend(struct pci_dev *pci_dev, pm_message_t state)
497{
498 struct ieee80211_hw *hw = pci_get_drvdata(pci_dev);
499 struct rt2x00_dev *rt2x00dev = hw->priv;
500 int retval;
501
502 retval = rt2x00lib_suspend(rt2x00dev, state);
503 if (retval)
504 return retval;
505
506 rt2x00pci_free_reg(rt2x00dev);
507
508 pci_save_state(pci_dev);
509 pci_disable_device(pci_dev);
510 return pci_set_power_state(pci_dev, pci_choose_state(pci_dev, state));
511}
512EXPORT_SYMBOL_GPL(rt2x00pci_suspend);
513
514int rt2x00pci_resume(struct pci_dev *pci_dev)
515{
516 struct ieee80211_hw *hw = pci_get_drvdata(pci_dev);
517 struct rt2x00_dev *rt2x00dev = hw->priv;
518 int retval;
519
520 if (pci_set_power_state(pci_dev, PCI_D0) ||
521 pci_enable_device(pci_dev) ||
522 pci_restore_state(pci_dev)) {
523 ERROR(rt2x00dev, "Failed to resume device.\n");
524 return -EIO;
525 }
526
527 retval = rt2x00pci_alloc_reg(rt2x00dev);
528 if (retval)
529 return retval;
530
531 retval = rt2x00lib_resume(rt2x00dev);
532 if (retval)
533 goto exit_free_reg;
534
535 return 0;
536
537exit_free_reg:
538 rt2x00pci_free_reg(rt2x00dev);
539
540 return retval;
541}
542EXPORT_SYMBOL_GPL(rt2x00pci_resume);
543#endif /* CONFIG_PM */
544
545/*
546 * rt2x00pci module information.
547 */
548MODULE_AUTHOR(DRV_PROJECT);
549MODULE_VERSION(DRV_VERSION);
Ivo van Doorn181d6902008-02-05 16:42:23 -0500550MODULE_DESCRIPTION("rt2x00 pci library");
Ivo van Doorn95ea3622007-09-25 17:57:13 -0700551MODULE_LICENSE("GPL");