blob: 9a32f2d174b0bc549ce79a33d4ed962f19c71094 [file] [log] [blame]
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +08001/*
2 * Copyright 2012 Marvell International Ltd.
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
7 */
Thierry Reding73312052013-01-21 11:09:00 +01008#include <linux/err.h>
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +08009#include <linux/module.h>
10#include <linux/init.h>
11#include <linux/types.h>
12#include <linux/interrupt.h>
13#include <linux/dma-mapping.h>
14#include <linux/slab.h>
15#include <linux/dmaengine.h>
16#include <linux/platform_device.h>
17#include <linux/device.h>
18#include <linux/platform_data/mmp_dma.h>
19#include <linux/dmapool.h>
20#include <linux/of_device.h>
21#include <linux/of.h>
22
23#include "dmaengine.h"
24
25#define DCSR 0x0000
26#define DALGN 0x00a0
27#define DINT 0x00f0
28#define DDADR 0x0200
29#define DSADR 0x0204
30#define DTADR 0x0208
31#define DCMD 0x020c
32
33#define DCSR_RUN (1 << 31) /* Run Bit (read / write) */
34#define DCSR_NODESC (1 << 30) /* No-Descriptor Fetch (read / write) */
35#define DCSR_STOPIRQEN (1 << 29) /* Stop Interrupt Enable (read / write) */
36#define DCSR_REQPEND (1 << 8) /* Request Pending (read-only) */
37#define DCSR_STOPSTATE (1 << 3) /* Stop State (read-only) */
38#define DCSR_ENDINTR (1 << 2) /* End Interrupt (read / write) */
39#define DCSR_STARTINTR (1 << 1) /* Start Interrupt (read / write) */
40#define DCSR_BUSERR (1 << 0) /* Bus Error Interrupt (read / write) */
41
42#define DCSR_EORIRQEN (1 << 28) /* End of Receive Interrupt Enable (R/W) */
43#define DCSR_EORJMPEN (1 << 27) /* Jump to next descriptor on EOR */
44#define DCSR_EORSTOPEN (1 << 26) /* STOP on an EOR */
45#define DCSR_SETCMPST (1 << 25) /* Set Descriptor Compare Status */
46#define DCSR_CLRCMPST (1 << 24) /* Clear Descriptor Compare Status */
47#define DCSR_CMPST (1 << 10) /* The Descriptor Compare Status */
48#define DCSR_EORINTR (1 << 9) /* The end of Receive */
49
50#define DRCMR_MAPVLD (1 << 7) /* Map Valid (read / write) */
51#define DRCMR_CHLNUM 0x1f /* mask for Channel Number (read / write) */
52
53#define DDADR_DESCADDR 0xfffffff0 /* Address of next descriptor (mask) */
54#define DDADR_STOP (1 << 0) /* Stop (read / write) */
55
56#define DCMD_INCSRCADDR (1 << 31) /* Source Address Increment Setting. */
57#define DCMD_INCTRGADDR (1 << 30) /* Target Address Increment Setting. */
58#define DCMD_FLOWSRC (1 << 29) /* Flow Control by the source. */
59#define DCMD_FLOWTRG (1 << 28) /* Flow Control by the target. */
60#define DCMD_STARTIRQEN (1 << 22) /* Start Interrupt Enable */
61#define DCMD_ENDIRQEN (1 << 21) /* End Interrupt Enable */
62#define DCMD_ENDIAN (1 << 18) /* Device Endian-ness. */
63#define DCMD_BURST8 (1 << 16) /* 8 byte burst */
64#define DCMD_BURST16 (2 << 16) /* 16 byte burst */
65#define DCMD_BURST32 (3 << 16) /* 32 byte burst */
66#define DCMD_WIDTH1 (1 << 14) /* 1 byte width */
67#define DCMD_WIDTH2 (2 << 14) /* 2 byte width (HalfWord) */
68#define DCMD_WIDTH4 (3 << 14) /* 4 byte width (Word) */
69#define DCMD_LENGTH 0x01fff /* length mask (max = 8K - 1) */
70
71#define PDMA_ALIGNMENT 3
72#define PDMA_MAX_DESC_BYTES 0x1000
73
74struct mmp_pdma_desc_hw {
75 u32 ddadr; /* Points to the next descriptor + flags */
76 u32 dsadr; /* DSADR value for the current transfer */
77 u32 dtadr; /* DTADR value for the current transfer */
78 u32 dcmd; /* DCMD value for the current transfer */
79} __aligned(32);
80
81struct mmp_pdma_desc_sw {
82 struct mmp_pdma_desc_hw desc;
83 struct list_head node;
84 struct list_head tx_list;
85 struct dma_async_tx_descriptor async_tx;
86};
87
88struct mmp_pdma_phy;
89
90struct mmp_pdma_chan {
91 struct device *dev;
92 struct dma_chan chan;
93 struct dma_async_tx_descriptor desc;
94 struct mmp_pdma_phy *phy;
95 enum dma_transfer_direction dir;
96
97 /* channel's basic info */
98 struct tasklet_struct tasklet;
99 u32 dcmd;
100 u32 drcmr;
101 u32 dev_addr;
102
103 /* list for desc */
104 spinlock_t desc_lock; /* Descriptor list lock */
105 struct list_head chain_pending; /* Link descriptors queue for pending */
106 struct list_head chain_running; /* Link descriptors queue for running */
107 bool idle; /* channel statue machine */
108
109 struct dma_pool *desc_pool; /* Descriptors pool */
110};
111
112struct mmp_pdma_phy {
113 int idx;
114 void __iomem *base;
115 struct mmp_pdma_chan *vchan;
116};
117
118struct mmp_pdma_device {
119 int dma_channels;
120 void __iomem *base;
121 struct device *dev;
122 struct dma_device device;
123 struct mmp_pdma_phy *phy;
Xiang Wang027f28b2013-06-18 14:55:58 +0800124 spinlock_t phy_lock; /* protect alloc/free phy channels */
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800125};
126
127#define tx_to_mmp_pdma_desc(tx) container_of(tx, struct mmp_pdma_desc_sw, async_tx)
128#define to_mmp_pdma_desc(lh) container_of(lh, struct mmp_pdma_desc_sw, node)
129#define to_mmp_pdma_chan(dchan) container_of(dchan, struct mmp_pdma_chan, chan)
130#define to_mmp_pdma_dev(dmadev) container_of(dmadev, struct mmp_pdma_device, device)
131
132static void set_desc(struct mmp_pdma_phy *phy, dma_addr_t addr)
133{
134 u32 reg = (phy->idx << 4) + DDADR;
135
136 writel(addr, phy->base + reg);
137}
138
139static void enable_chan(struct mmp_pdma_phy *phy)
140{
141 u32 reg;
142
143 if (!phy->vchan)
144 return;
145
146 reg = phy->vchan->drcmr;
147 reg = (((reg) < 64) ? 0x0100 : 0x1100) + (((reg) & 0x3f) << 2);
148 writel(DRCMR_MAPVLD | phy->idx, phy->base + reg);
149
150 reg = (phy->idx << 2) + DCSR;
151 writel(readl(phy->base + reg) | DCSR_RUN,
152 phy->base + reg);
153}
154
155static void disable_chan(struct mmp_pdma_phy *phy)
156{
157 u32 reg;
158
159 if (phy) {
160 reg = (phy->idx << 2) + DCSR;
161 writel(readl(phy->base + reg) & ~DCSR_RUN,
162 phy->base + reg);
163 }
164}
165
166static int clear_chan_irq(struct mmp_pdma_phy *phy)
167{
168 u32 dcsr;
169 u32 dint = readl(phy->base + DINT);
170 u32 reg = (phy->idx << 2) + DCSR;
171
172 if (dint & BIT(phy->idx)) {
173 /* clear irq */
174 dcsr = readl(phy->base + reg);
175 writel(dcsr, phy->base + reg);
176 if ((dcsr & DCSR_BUSERR) && (phy->vchan))
177 dev_warn(phy->vchan->dev, "DCSR_BUSERR\n");
178 return 0;
179 }
180 return -EAGAIN;
181}
182
183static irqreturn_t mmp_pdma_chan_handler(int irq, void *dev_id)
184{
185 struct mmp_pdma_phy *phy = dev_id;
186
187 if (clear_chan_irq(phy) == 0) {
188 tasklet_schedule(&phy->vchan->tasklet);
189 return IRQ_HANDLED;
190 } else
191 return IRQ_NONE;
192}
193
194static irqreturn_t mmp_pdma_int_handler(int irq, void *dev_id)
195{
196 struct mmp_pdma_device *pdev = dev_id;
197 struct mmp_pdma_phy *phy;
198 u32 dint = readl(pdev->base + DINT);
199 int i, ret;
200 int irq_num = 0;
201
202 while (dint) {
203 i = __ffs(dint);
204 dint &= (dint - 1);
205 phy = &pdev->phy[i];
206 ret = mmp_pdma_chan_handler(irq, phy);
207 if (ret == IRQ_HANDLED)
208 irq_num++;
209 }
210
211 if (irq_num)
212 return IRQ_HANDLED;
213 else
214 return IRQ_NONE;
215}
216
217/* lookup free phy channel as descending priority */
218static struct mmp_pdma_phy *lookup_phy(struct mmp_pdma_chan *pchan)
219{
220 int prio, i;
221 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device);
222 struct mmp_pdma_phy *phy;
Xiang Wang027f28b2013-06-18 14:55:58 +0800223 unsigned long flags;
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800224
225 /*
226 * dma channel priorities
227 * ch 0 - 3, 16 - 19 <--> (0)
228 * ch 4 - 7, 20 - 23 <--> (1)
229 * ch 8 - 11, 24 - 27 <--> (2)
230 * ch 12 - 15, 28 - 31 <--> (3)
231 */
Xiang Wang027f28b2013-06-18 14:55:58 +0800232
233 spin_lock_irqsave(&pdev->phy_lock, flags);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800234 for (prio = 0; prio <= (((pdev->dma_channels - 1) & 0xf) >> 2); prio++) {
235 for (i = 0; i < pdev->dma_channels; i++) {
236 if (prio != ((i & 0xf) >> 2))
237 continue;
238 phy = &pdev->phy[i];
239 if (!phy->vchan) {
240 phy->vchan = pchan;
Xiang Wang027f28b2013-06-18 14:55:58 +0800241 spin_unlock_irqrestore(&pdev->phy_lock, flags);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800242 return phy;
243 }
244 }
245 }
246
Xiang Wang027f28b2013-06-18 14:55:58 +0800247 spin_unlock_irqrestore(&pdev->phy_lock, flags);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800248 return NULL;
249}
250
Xiang Wang027f28b2013-06-18 14:55:58 +0800251static void mmp_pdma_free_phy(struct mmp_pdma_chan *pchan)
252{
253 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device);
254 unsigned long flags;
Xiang Wang26a2dfd2013-06-18 14:55:59 +0800255 u32 reg;
Xiang Wang027f28b2013-06-18 14:55:58 +0800256
257 if (!pchan->phy)
258 return;
259
Xiang Wang26a2dfd2013-06-18 14:55:59 +0800260 /* clear the channel mapping in DRCMR */
261 reg = pchan->phy->vchan->drcmr;
262 reg = ((reg < 64) ? 0x0100 : 0x1100) + ((reg & 0x3f) << 2);
263 writel(0, pchan->phy->base + reg);
264
Xiang Wang027f28b2013-06-18 14:55:58 +0800265 spin_lock_irqsave(&pdev->phy_lock, flags);
266 pchan->phy->vchan = NULL;
267 pchan->phy = NULL;
268 spin_unlock_irqrestore(&pdev->phy_lock, flags);
269}
270
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800271/* desc->tx_list ==> pending list */
272static void append_pending_queue(struct mmp_pdma_chan *chan,
273 struct mmp_pdma_desc_sw *desc)
274{
275 struct mmp_pdma_desc_sw *tail =
276 to_mmp_pdma_desc(chan->chain_pending.prev);
277
278 if (list_empty(&chan->chain_pending))
279 goto out_splice;
280
281 /* one irq per queue, even appended */
282 tail->desc.ddadr = desc->async_tx.phys;
283 tail->desc.dcmd &= ~DCMD_ENDIRQEN;
284
285 /* softly link to pending list */
286out_splice:
287 list_splice_tail_init(&desc->tx_list, &chan->chain_pending);
288}
289
290/**
291 * start_pending_queue - transfer any pending transactions
292 * pending list ==> running list
293 */
294static void start_pending_queue(struct mmp_pdma_chan *chan)
295{
296 struct mmp_pdma_desc_sw *desc;
297
298 /* still in running, irq will start the pending list */
299 if (!chan->idle) {
300 dev_dbg(chan->dev, "DMA controller still busy\n");
301 return;
302 }
303
304 if (list_empty(&chan->chain_pending)) {
305 /* chance to re-fetch phy channel with higher prio */
Xiang Wang027f28b2013-06-18 14:55:58 +0800306 mmp_pdma_free_phy(chan);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800307 dev_dbg(chan->dev, "no pending list\n");
308 return;
309 }
310
311 if (!chan->phy) {
312 chan->phy = lookup_phy(chan);
313 if (!chan->phy) {
314 dev_dbg(chan->dev, "no free dma channel\n");
315 return;
316 }
317 }
318
319 /*
320 * pending -> running
321 * reintilize pending list
322 */
323 desc = list_first_entry(&chan->chain_pending,
324 struct mmp_pdma_desc_sw, node);
325 list_splice_tail_init(&chan->chain_pending, &chan->chain_running);
326
327 /*
328 * Program the descriptor's address into the DMA controller,
329 * then start the DMA transaction
330 */
331 set_desc(chan->phy, desc->async_tx.phys);
332 enable_chan(chan->phy);
333 chan->idle = false;
334}
335
336
337/* desc->tx_list ==> pending list */
338static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx)
339{
340 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan);
341 struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx);
342 struct mmp_pdma_desc_sw *child;
343 unsigned long flags;
344 dma_cookie_t cookie = -EBUSY;
345
346 spin_lock_irqsave(&chan->desc_lock, flags);
347
348 list_for_each_entry(child, &desc->tx_list, node) {
349 cookie = dma_cookie_assign(&child->async_tx);
350 }
351
352 append_pending_queue(chan, desc);
353
354 spin_unlock_irqrestore(&chan->desc_lock, flags);
355
356 return cookie;
357}
358
359struct mmp_pdma_desc_sw *mmp_pdma_alloc_descriptor(struct mmp_pdma_chan *chan)
360{
361 struct mmp_pdma_desc_sw *desc;
362 dma_addr_t pdesc;
363
364 desc = dma_pool_alloc(chan->desc_pool, GFP_ATOMIC, &pdesc);
365 if (!desc) {
366 dev_err(chan->dev, "out of memory for link descriptor\n");
367 return NULL;
368 }
369
370 memset(desc, 0, sizeof(*desc));
371 INIT_LIST_HEAD(&desc->tx_list);
372 dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan);
373 /* each desc has submit */
374 desc->async_tx.tx_submit = mmp_pdma_tx_submit;
375 desc->async_tx.phys = pdesc;
376
377 return desc;
378}
379
380/**
381 * mmp_pdma_alloc_chan_resources - Allocate resources for DMA channel.
382 *
383 * This function will create a dma pool for descriptor allocation.
384 * Request irq only when channel is requested
385 * Return - The number of allocated descriptors.
386 */
387
388static int mmp_pdma_alloc_chan_resources(struct dma_chan *dchan)
389{
390 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan);
391
392 if (chan->desc_pool)
393 return 1;
394
395 chan->desc_pool =
396 dma_pool_create(dev_name(&dchan->dev->device), chan->dev,
397 sizeof(struct mmp_pdma_desc_sw),
398 __alignof__(struct mmp_pdma_desc_sw), 0);
399 if (!chan->desc_pool) {
400 dev_err(chan->dev, "unable to allocate descriptor pool\n");
401 return -ENOMEM;
402 }
Xiang Wang027f28b2013-06-18 14:55:58 +0800403 mmp_pdma_free_phy(chan);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800404 chan->idle = true;
405 chan->dev_addr = 0;
406 return 1;
407}
408
409static void mmp_pdma_free_desc_list(struct mmp_pdma_chan *chan,
410 struct list_head *list)
411{
412 struct mmp_pdma_desc_sw *desc, *_desc;
413
414 list_for_each_entry_safe(desc, _desc, list, node) {
415 list_del(&desc->node);
416 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys);
417 }
418}
419
420static void mmp_pdma_free_chan_resources(struct dma_chan *dchan)
421{
422 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan);
423 unsigned long flags;
424
425 spin_lock_irqsave(&chan->desc_lock, flags);
426 mmp_pdma_free_desc_list(chan, &chan->chain_pending);
427 mmp_pdma_free_desc_list(chan, &chan->chain_running);
428 spin_unlock_irqrestore(&chan->desc_lock, flags);
429
430 dma_pool_destroy(chan->desc_pool);
431 chan->desc_pool = NULL;
432 chan->idle = true;
433 chan->dev_addr = 0;
Xiang Wang027f28b2013-06-18 14:55:58 +0800434 mmp_pdma_free_phy(chan);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800435 return;
436}
437
438static struct dma_async_tx_descriptor *
439mmp_pdma_prep_memcpy(struct dma_chan *dchan,
440 dma_addr_t dma_dst, dma_addr_t dma_src,
441 size_t len, unsigned long flags)
442{
443 struct mmp_pdma_chan *chan;
444 struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new;
445 size_t copy = 0;
446
447 if (!dchan)
448 return NULL;
449
450 if (!len)
451 return NULL;
452
453 chan = to_mmp_pdma_chan(dchan);
454
455 if (!chan->dir) {
456 chan->dir = DMA_MEM_TO_MEM;
457 chan->dcmd = DCMD_INCTRGADDR | DCMD_INCSRCADDR;
458 chan->dcmd |= DCMD_BURST32;
459 }
460
461 do {
462 /* Allocate the link descriptor from DMA pool */
463 new = mmp_pdma_alloc_descriptor(chan);
464 if (!new) {
465 dev_err(chan->dev, "no memory for desc\n");
466 goto fail;
467 }
468
469 copy = min_t(size_t, len, PDMA_MAX_DESC_BYTES);
470
471 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & copy);
472 new->desc.dsadr = dma_src;
473 new->desc.dtadr = dma_dst;
474
475 if (!first)
476 first = new;
477 else
478 prev->desc.ddadr = new->async_tx.phys;
479
480 new->async_tx.cookie = 0;
481 async_tx_ack(&new->async_tx);
482
483 prev = new;
484 len -= copy;
485
486 if (chan->dir == DMA_MEM_TO_DEV) {
487 dma_src += copy;
488 } else if (chan->dir == DMA_DEV_TO_MEM) {
489 dma_dst += copy;
490 } else if (chan->dir == DMA_MEM_TO_MEM) {
491 dma_src += copy;
492 dma_dst += copy;
493 }
494
495 /* Insert the link descriptor to the LD ring */
496 list_add_tail(&new->node, &first->tx_list);
497 } while (len);
498
499 first->async_tx.flags = flags; /* client is in control of this ack */
500 first->async_tx.cookie = -EBUSY;
501
502 /* last desc and fire IRQ */
503 new->desc.ddadr = DDADR_STOP;
504 new->desc.dcmd |= DCMD_ENDIRQEN;
505
506 return &first->async_tx;
507
508fail:
509 if (first)
510 mmp_pdma_free_desc_list(chan, &first->tx_list);
511 return NULL;
512}
513
514static struct dma_async_tx_descriptor *
515mmp_pdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl,
516 unsigned int sg_len, enum dma_transfer_direction dir,
517 unsigned long flags, void *context)
518{
519 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan);
520 struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new = NULL;
521 size_t len, avail;
522 struct scatterlist *sg;
523 dma_addr_t addr;
524 int i;
525
526 if ((sgl == NULL) || (sg_len == 0))
527 return NULL;
528
529 for_each_sg(sgl, sg, sg_len, i) {
530 addr = sg_dma_address(sg);
531 avail = sg_dma_len(sgl);
532
533 do {
534 len = min_t(size_t, avail, PDMA_MAX_DESC_BYTES);
535
536 /* allocate and populate the descriptor */
537 new = mmp_pdma_alloc_descriptor(chan);
538 if (!new) {
539 dev_err(chan->dev, "no memory for desc\n");
540 goto fail;
541 }
542
543 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & len);
544 if (dir == DMA_MEM_TO_DEV) {
545 new->desc.dsadr = addr;
546 new->desc.dtadr = chan->dev_addr;
547 } else {
548 new->desc.dsadr = chan->dev_addr;
549 new->desc.dtadr = addr;
550 }
551
552 if (!first)
553 first = new;
554 else
555 prev->desc.ddadr = new->async_tx.phys;
556
557 new->async_tx.cookie = 0;
558 async_tx_ack(&new->async_tx);
559 prev = new;
560
561 /* Insert the link descriptor to the LD ring */
562 list_add_tail(&new->node, &first->tx_list);
563
564 /* update metadata */
565 addr += len;
566 avail -= len;
567 } while (avail);
568 }
569
570 first->async_tx.cookie = -EBUSY;
571 first->async_tx.flags = flags;
572
573 /* last desc and fire IRQ */
574 new->desc.ddadr = DDADR_STOP;
575 new->desc.dcmd |= DCMD_ENDIRQEN;
576
577 return &first->async_tx;
578
579fail:
580 if (first)
581 mmp_pdma_free_desc_list(chan, &first->tx_list);
582 return NULL;
583}
584
585static int mmp_pdma_control(struct dma_chan *dchan, enum dma_ctrl_cmd cmd,
586 unsigned long arg)
587{
588 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan);
589 struct dma_slave_config *cfg = (void *)arg;
590 unsigned long flags;
591 int ret = 0;
592 u32 maxburst = 0, addr = 0;
593 enum dma_slave_buswidth width = DMA_SLAVE_BUSWIDTH_UNDEFINED;
594
595 if (!dchan)
596 return -EINVAL;
597
598 switch (cmd) {
599 case DMA_TERMINATE_ALL:
600 disable_chan(chan->phy);
Xiang Wang027f28b2013-06-18 14:55:58 +0800601 mmp_pdma_free_phy(chan);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800602 spin_lock_irqsave(&chan->desc_lock, flags);
603 mmp_pdma_free_desc_list(chan, &chan->chain_pending);
604 mmp_pdma_free_desc_list(chan, &chan->chain_running);
605 spin_unlock_irqrestore(&chan->desc_lock, flags);
606 chan->idle = true;
607 break;
608 case DMA_SLAVE_CONFIG:
609 if (cfg->direction == DMA_DEV_TO_MEM) {
610 chan->dcmd = DCMD_INCTRGADDR | DCMD_FLOWSRC;
611 maxburst = cfg->src_maxburst;
612 width = cfg->src_addr_width;
613 addr = cfg->src_addr;
614 } else if (cfg->direction == DMA_MEM_TO_DEV) {
615 chan->dcmd = DCMD_INCSRCADDR | DCMD_FLOWTRG;
616 maxburst = cfg->dst_maxburst;
617 width = cfg->dst_addr_width;
618 addr = cfg->dst_addr;
619 }
620
621 if (width == DMA_SLAVE_BUSWIDTH_1_BYTE)
622 chan->dcmd |= DCMD_WIDTH1;
623 else if (width == DMA_SLAVE_BUSWIDTH_2_BYTES)
624 chan->dcmd |= DCMD_WIDTH2;
625 else if (width == DMA_SLAVE_BUSWIDTH_4_BYTES)
626 chan->dcmd |= DCMD_WIDTH4;
627
628 if (maxburst == 8)
629 chan->dcmd |= DCMD_BURST8;
630 else if (maxburst == 16)
631 chan->dcmd |= DCMD_BURST16;
632 else if (maxburst == 32)
633 chan->dcmd |= DCMD_BURST32;
634
Cong Dinged30933e2013-01-15 01:19:48 +0100635 chan->dir = cfg->direction;
636 chan->drcmr = cfg->slave_id;
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800637 chan->dev_addr = addr;
638 break;
639 default:
640 return -ENOSYS;
641 }
642
643 return ret;
644}
645
646static enum dma_status mmp_pdma_tx_status(struct dma_chan *dchan,
647 dma_cookie_t cookie, struct dma_tx_state *txstate)
648{
Andy Shevchenko4aa9fe02013-05-27 15:14:36 +0300649 return dma_cookie_status(dchan, cookie, txstate);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800650}
651
652/**
653 * mmp_pdma_issue_pending - Issue the DMA start command
654 * pending list ==> running list
655 */
656static void mmp_pdma_issue_pending(struct dma_chan *dchan)
657{
658 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan);
659 unsigned long flags;
660
661 spin_lock_irqsave(&chan->desc_lock, flags);
662 start_pending_queue(chan);
663 spin_unlock_irqrestore(&chan->desc_lock, flags);
664}
665
666/*
667 * dma_do_tasklet
668 * Do call back
669 * Start pending list
670 */
671static void dma_do_tasklet(unsigned long data)
672{
673 struct mmp_pdma_chan *chan = (struct mmp_pdma_chan *)data;
674 struct mmp_pdma_desc_sw *desc, *_desc;
675 LIST_HEAD(chain_cleanup);
676 unsigned long flags;
677
678 /* submit pending list; callback for each desc; free desc */
679
680 spin_lock_irqsave(&chan->desc_lock, flags);
681
682 /* update the cookie if we have some descriptors to cleanup */
683 if (!list_empty(&chan->chain_running)) {
684 dma_cookie_t cookie;
685
686 desc = to_mmp_pdma_desc(chan->chain_running.prev);
687 cookie = desc->async_tx.cookie;
688 dma_cookie_complete(&desc->async_tx);
689
690 dev_dbg(chan->dev, "completed_cookie=%d\n", cookie);
691 }
692
693 /*
694 * move the descriptors to a temporary list so we can drop the lock
695 * during the entire cleanup operation
696 */
697 list_splice_tail_init(&chan->chain_running, &chain_cleanup);
698
699 /* the hardware is now idle and ready for more */
700 chan->idle = true;
701
702 /* Start any pending transactions automatically */
703 start_pending_queue(chan);
704 spin_unlock_irqrestore(&chan->desc_lock, flags);
705
706 /* Run the callback for each descriptor, in order */
707 list_for_each_entry_safe(desc, _desc, &chain_cleanup, node) {
708 struct dma_async_tx_descriptor *txd = &desc->async_tx;
709
710 /* Remove from the list of transactions */
711 list_del(&desc->node);
712 /* Run the link descriptor callback function */
713 if (txd->callback)
714 txd->callback(txd->callback_param);
715
716 dma_pool_free(chan->desc_pool, desc, txd->phys);
717 }
718}
719
Greg Kroah-Hartman4bf27b82012-12-21 15:09:59 -0800720static int mmp_pdma_remove(struct platform_device *op)
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800721{
722 struct mmp_pdma_device *pdev = platform_get_drvdata(op);
723
724 dma_async_device_unregister(&pdev->device);
725 return 0;
726}
727
Bill Pemberton463a1f82012-11-19 13:22:55 -0500728static int mmp_pdma_chan_init(struct mmp_pdma_device *pdev,
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800729 int idx, int irq)
730{
731 struct mmp_pdma_phy *phy = &pdev->phy[idx];
732 struct mmp_pdma_chan *chan;
733 int ret;
734
735 chan = devm_kzalloc(pdev->dev,
736 sizeof(struct mmp_pdma_chan), GFP_KERNEL);
737 if (chan == NULL)
738 return -ENOMEM;
739
740 phy->idx = idx;
741 phy->base = pdev->base;
742
743 if (irq) {
744 ret = devm_request_irq(pdev->dev, irq,
745 mmp_pdma_chan_handler, IRQF_DISABLED, "pdma", phy);
746 if (ret) {
747 dev_err(pdev->dev, "channel request irq fail!\n");
748 return ret;
749 }
750 }
751
752 spin_lock_init(&chan->desc_lock);
753 chan->dev = pdev->dev;
754 chan->chan.device = &pdev->device;
755 tasklet_init(&chan->tasklet, dma_do_tasklet, (unsigned long)chan);
756 INIT_LIST_HEAD(&chan->chain_pending);
757 INIT_LIST_HEAD(&chan->chain_running);
758
759 /* register virt channel to dma engine */
760 list_add_tail(&chan->chan.device_node,
761 &pdev->device.channels);
762
763 return 0;
764}
765
766static struct of_device_id mmp_pdma_dt_ids[] = {
767 { .compatible = "marvell,pdma-1.0", },
768 {}
769};
770MODULE_DEVICE_TABLE(of, mmp_pdma_dt_ids);
771
Bill Pemberton463a1f82012-11-19 13:22:55 -0500772static int mmp_pdma_probe(struct platform_device *op)
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800773{
774 struct mmp_pdma_device *pdev;
775 const struct of_device_id *of_id;
776 struct mmp_dma_platdata *pdata = dev_get_platdata(&op->dev);
777 struct resource *iores;
778 int i, ret, irq = 0;
779 int dma_channels = 0, irq_num = 0;
780
781 pdev = devm_kzalloc(&op->dev, sizeof(*pdev), GFP_KERNEL);
782 if (!pdev)
783 return -ENOMEM;
784 pdev->dev = &op->dev;
785
Xiang Wang027f28b2013-06-18 14:55:58 +0800786 spin_lock_init(&pdev->phy_lock);
787
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800788 iores = platform_get_resource(op, IORESOURCE_MEM, 0);
789 if (!iores)
790 return -EINVAL;
791
Thierry Reding73312052013-01-21 11:09:00 +0100792 pdev->base = devm_ioremap_resource(pdev->dev, iores);
793 if (IS_ERR(pdev->base))
794 return PTR_ERR(pdev->base);
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800795
796 of_id = of_match_device(mmp_pdma_dt_ids, pdev->dev);
797 if (of_id)
798 of_property_read_u32(pdev->dev->of_node,
799 "#dma-channels", &dma_channels);
800 else if (pdata && pdata->dma_channels)
801 dma_channels = pdata->dma_channels;
802 else
803 dma_channels = 32; /* default 32 channel */
804 pdev->dma_channels = dma_channels;
805
806 for (i = 0; i < dma_channels; i++) {
807 if (platform_get_irq(op, i) > 0)
808 irq_num++;
809 }
810
811 pdev->phy = devm_kzalloc(pdev->dev,
812 dma_channels * sizeof(struct mmp_pdma_chan), GFP_KERNEL);
813 if (pdev->phy == NULL)
814 return -ENOMEM;
815
816 INIT_LIST_HEAD(&pdev->device.channels);
817
818 if (irq_num != dma_channels) {
819 /* all chan share one irq, demux inside */
820 irq = platform_get_irq(op, 0);
821 ret = devm_request_irq(pdev->dev, irq,
822 mmp_pdma_int_handler, IRQF_DISABLED, "pdma", pdev);
823 if (ret)
824 return ret;
825 }
826
827 for (i = 0; i < dma_channels; i++) {
828 irq = (irq_num != dma_channels) ? 0 : platform_get_irq(op, i);
829 ret = mmp_pdma_chan_init(pdev, i, irq);
830 if (ret)
831 return ret;
832 }
833
834 dma_cap_set(DMA_SLAVE, pdev->device.cap_mask);
835 dma_cap_set(DMA_MEMCPY, pdev->device.cap_mask);
836 dma_cap_set(DMA_SLAVE, pdev->device.cap_mask);
837 pdev->device.dev = &op->dev;
838 pdev->device.device_alloc_chan_resources = mmp_pdma_alloc_chan_resources;
839 pdev->device.device_free_chan_resources = mmp_pdma_free_chan_resources;
840 pdev->device.device_tx_status = mmp_pdma_tx_status;
841 pdev->device.device_prep_dma_memcpy = mmp_pdma_prep_memcpy;
842 pdev->device.device_prep_slave_sg = mmp_pdma_prep_slave_sg;
843 pdev->device.device_issue_pending = mmp_pdma_issue_pending;
844 pdev->device.device_control = mmp_pdma_control;
845 pdev->device.copy_align = PDMA_ALIGNMENT;
846
847 if (pdev->dev->coherent_dma_mask)
848 dma_set_mask(pdev->dev, pdev->dev->coherent_dma_mask);
849 else
850 dma_set_mask(pdev->dev, DMA_BIT_MASK(64));
851
852 ret = dma_async_device_register(&pdev->device);
853 if (ret) {
854 dev_err(pdev->device.dev, "unable to register\n");
855 return ret;
856 }
857
858 dev_info(pdev->device.dev, "initialized\n");
859 return 0;
860}
861
862static const struct platform_device_id mmp_pdma_id_table[] = {
863 { "mmp-pdma", },
864 { },
865};
866
867static struct platform_driver mmp_pdma_driver = {
868 .driver = {
869 .name = "mmp-pdma",
870 .owner = THIS_MODULE,
871 .of_match_table = mmp_pdma_dt_ids,
872 },
873 .id_table = mmp_pdma_id_table,
874 .probe = mmp_pdma_probe,
Bill Pembertona7d6e3e2012-11-19 13:20:04 -0500875 .remove = mmp_pdma_remove,
Zhangfei Gaoc8acd6a2012-09-03 11:03:45 +0800876};
877
878module_platform_driver(mmp_pdma_driver);
879
880MODULE_DESCRIPTION("MARVELL MMP Periphera DMA Driver");
881MODULE_AUTHOR("Marvell International Ltd.");
882MODULE_LICENSE("GPL v2");