blob: 93e5a062df880ee991c332edca71a532eb2341cb [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/* $Id: dma.h,v 1.21 2001/12/13 04:16:52 davem Exp $
2 * include/asm-sparc64/dma.h
3 *
4 * Copyright 1996 (C) David S. Miller (davem@caip.rutgers.edu)
5 */
6
7#ifndef _ASM_SPARC64_DMA_H
8#define _ASM_SPARC64_DMA_H
9
Linus Torvalds1da177e2005-04-16 15:20:36 -070010#include <linux/kernel.h>
11#include <linux/types.h>
12#include <linux/spinlock.h>
13
14#include <asm/sbus.h>
15#include <asm/delay.h>
16#include <asm/oplib.h>
17
18extern spinlock_t dma_spin_lock;
19
20#define claim_dma_lock() \
21({ unsigned long flags; \
22 spin_lock_irqsave(&dma_spin_lock, flags); \
23 flags; \
24})
25
26#define release_dma_lock(__flags) \
27 spin_unlock_irqrestore(&dma_spin_lock, __flags);
28
29/* These are irrelevant for Sparc DMA, but we leave it in so that
30 * things can compile.
31 */
32#define MAX_DMA_CHANNELS 8
33#define DMA_MODE_READ 1
34#define DMA_MODE_WRITE 2
35#define MAX_DMA_ADDRESS (~0UL)
36
37/* Useful constants */
38#define SIZE_16MB (16*1024*1024)
39#define SIZE_64K (64*1024)
40
41/* SBUS DMA controller reg offsets */
42#define DMA_CSR 0x00UL /* rw DMA control/status register 0x00 */
43#define DMA_ADDR 0x04UL /* rw DMA transfer address register 0x04 */
44#define DMA_COUNT 0x08UL /* rw DMA transfer count register 0x08 */
45#define DMA_TEST 0x0cUL /* rw DMA test/debug register 0x0c */
46
47/* DVMA chip revisions */
48enum dvma_rev {
49 dvmarev0,
50 dvmaesc1,
51 dvmarev1,
52 dvmarev2,
53 dvmarev3,
54 dvmarevplus,
55 dvmahme
56};
57
58#define DMA_HASCOUNT(rev) ((rev)==dvmaesc1)
59
60/* Linux DMA information structure, filled during probe. */
61struct sbus_dma {
62 struct sbus_dma *next;
63 struct sbus_dev *sdev;
64 void __iomem *regs;
65
66 /* Status, misc info */
67 int node; /* Prom node for this DMA device */
68 int running; /* Are we doing DMA now? */
69 int allocated; /* Are we "owned" by anyone yet? */
70
71 /* Transfer information. */
72 u32 addr; /* Start address of current transfer */
73 int nbytes; /* Size of current transfer */
74 int realbytes; /* For splitting up large transfers, etc. */
75
76 /* DMA revision */
77 enum dvma_rev revision;
78};
79
80extern struct sbus_dma *dma_chain;
81
82/* Broken hardware... */
83#define DMA_ISBROKEN(dma) ((dma)->revision == dvmarev1)
84#define DMA_ISESC1(dma) ((dma)->revision == dvmaesc1)
85
86/* Main routines in dma.c */
87extern void dvma_init(struct sbus_bus *);
88
89/* Fields in the cond_reg register */
90/* First, the version identification bits */
91#define DMA_DEVICE_ID 0xf0000000 /* Device identification bits */
92#define DMA_VERS0 0x00000000 /* Sunray DMA version */
93#define DMA_ESCV1 0x40000000 /* DMA ESC Version 1 */
94#define DMA_VERS1 0x80000000 /* DMA rev 1 */
95#define DMA_VERS2 0xa0000000 /* DMA rev 2 */
96#define DMA_VERHME 0xb0000000 /* DMA hme gate array */
97#define DMA_VERSPLUS 0x90000000 /* DMA rev 1 PLUS */
98
99#define DMA_HNDL_INTR 0x00000001 /* An IRQ needs to be handled */
100#define DMA_HNDL_ERROR 0x00000002 /* We need to take an error */
101#define DMA_FIFO_ISDRAIN 0x0000000c /* The DMA FIFO is draining */
102#define DMA_INT_ENAB 0x00000010 /* Turn on interrupts */
103#define DMA_FIFO_INV 0x00000020 /* Invalidate the FIFO */
104#define DMA_ACC_SZ_ERR 0x00000040 /* The access size was bad */
105#define DMA_FIFO_STDRAIN 0x00000040 /* DMA_VERS1 Drain the FIFO */
106#define DMA_RST_SCSI 0x00000080 /* Reset the SCSI controller */
107#define DMA_RST_ENET DMA_RST_SCSI /* Reset the ENET controller */
108#define DMA_ST_WRITE 0x00000100 /* write from device to memory */
109#define DMA_ENABLE 0x00000200 /* Fire up DMA, handle requests */
110#define DMA_PEND_READ 0x00000400 /* DMA_VERS1/0/PLUS Pending Read */
111#define DMA_ESC_BURST 0x00000800 /* 1=16byte 0=32byte */
112#define DMA_READ_AHEAD 0x00001800 /* DMA read ahead partial longword */
113#define DMA_DSBL_RD_DRN 0x00001000 /* No EC drain on slave reads */
114#define DMA_BCNT_ENAB 0x00002000 /* If on, use the byte counter */
115#define DMA_TERM_CNTR 0x00004000 /* Terminal counter */
116#define DMA_SCSI_SBUS64 0x00008000 /* HME: Enable 64-bit SBUS mode. */
117#define DMA_CSR_DISAB 0x00010000 /* No FIFO drains during csr */
118#define DMA_SCSI_DISAB 0x00020000 /* No FIFO drains during reg */
119#define DMA_DSBL_WR_INV 0x00020000 /* No EC inval. on slave writes */
120#define DMA_ADD_ENABLE 0x00040000 /* Special ESC DVMA optimization */
121#define DMA_E_BURSTS 0x000c0000 /* ENET: SBUS r/w burst mask */
122#define DMA_E_BURST32 0x00040000 /* ENET: SBUS 32 byte r/w burst */
123#define DMA_E_BURST16 0x00000000 /* ENET: SBUS 16 byte r/w burst */
124#define DMA_BRST_SZ 0x000c0000 /* SCSI: SBUS r/w burst size */
125#define DMA_BRST64 0x000c0000 /* SCSI: 64byte bursts (HME on UltraSparc only) */
126#define DMA_BRST32 0x00040000 /* SCSI: 32byte bursts */
127#define DMA_BRST16 0x00000000 /* SCSI: 16byte bursts */
128#define DMA_BRST0 0x00080000 /* SCSI: no bursts (non-HME gate arrays) */
129#define DMA_ADDR_DISAB 0x00100000 /* No FIFO drains during addr */
130#define DMA_2CLKS 0x00200000 /* Each transfer = 2 clock ticks */
131#define DMA_3CLKS 0x00400000 /* Each transfer = 3 clock ticks */
132#define DMA_EN_ENETAUI DMA_3CLKS /* Put lance into AUI-cable mode */
133#define DMA_CNTR_DISAB 0x00800000 /* No IRQ when DMA_TERM_CNTR set */
134#define DMA_AUTO_NADDR 0x01000000 /* Use "auto nxt addr" feature */
135#define DMA_SCSI_ON 0x02000000 /* Enable SCSI dma */
136#define DMA_PARITY_OFF 0x02000000 /* HME: disable parity checking */
137#define DMA_LOADED_ADDR 0x04000000 /* Address has been loaded */
138#define DMA_LOADED_NADDR 0x08000000 /* Next address has been loaded */
139#define DMA_RESET_FAS366 0x08000000 /* HME: Assert RESET to FAS366 */
140
141/* Values describing the burst-size property from the PROM */
142#define DMA_BURST1 0x01
143#define DMA_BURST2 0x02
144#define DMA_BURST4 0x04
145#define DMA_BURST8 0x08
146#define DMA_BURST16 0x10
147#define DMA_BURST32 0x20
148#define DMA_BURST64 0x40
149#define DMA_BURSTBITS 0x7f
150
151/* Determine highest possible final transfer address given a base */
152#define DMA_MAXEND(addr) (0x01000000UL-(((unsigned long)(addr))&0x00ffffffUL))
153
154/* Yes, I hack a lot of elisp in my spare time... */
Mariusz Kozlowski5ff42452006-12-01 20:19:59 -0800155#define DMA_ERROR_P(regs) ((sbus_readl((regs) + DMA_CSR) & DMA_HNDL_ERROR))
156#define DMA_IRQ_P(regs) ((sbus_readl((regs) + DMA_CSR)) & (DMA_HNDL_INTR | DMA_HNDL_ERROR))
157#define DMA_WRITE_P(regs) ((sbus_readl((regs) + DMA_CSR) & DMA_ST_WRITE))
Linus Torvalds1da177e2005-04-16 15:20:36 -0700158#define DMA_OFF(__regs) \
159do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
160 tmp &= ~DMA_ENABLE; \
161 sbus_writel(tmp, (__regs) + DMA_CSR); \
162} while(0)
163#define DMA_INTSOFF(__regs) \
164do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
165 tmp &= ~DMA_INT_ENAB; \
166 sbus_writel(tmp, (__regs) + DMA_CSR); \
167} while(0)
168#define DMA_INTSON(__regs) \
169do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
170 tmp |= DMA_INT_ENAB; \
171 sbus_writel(tmp, (__regs) + DMA_CSR); \
172} while(0)
173#define DMA_PUNTFIFO(__regs) \
174do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
175 tmp |= DMA_FIFO_INV; \
176 sbus_writel(tmp, (__regs) + DMA_CSR); \
177} while(0)
178#define DMA_SETSTART(__regs, __addr) \
179 sbus_writel((u32)(__addr), (__regs) + DMA_ADDR);
180#define DMA_BEGINDMA_W(__regs) \
181do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
182 tmp |= (DMA_ST_WRITE|DMA_ENABLE|DMA_INT_ENAB); \
183 sbus_writel(tmp, (__regs) + DMA_CSR); \
184} while(0)
185#define DMA_BEGINDMA_R(__regs) \
186do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
187 tmp |= (DMA_ENABLE|DMA_INT_ENAB); \
188 tmp &= ~DMA_ST_WRITE; \
189 sbus_writel(tmp, (__regs) + DMA_CSR); \
190} while(0)
191
192/* For certain DMA chips, we need to disable ints upon irq entry
193 * and turn them back on when we are done. So in any ESP interrupt
194 * handler you *must* call DMA_IRQ_ENTRY upon entry and DMA_IRQ_EXIT
195 * when leaving the handler. You have been warned...
196 */
197#define DMA_IRQ_ENTRY(dma, dregs) do { \
198 if(DMA_ISBROKEN(dma)) DMA_INTSOFF(dregs); \
199 } while (0)
200
201#define DMA_IRQ_EXIT(dma, dregs) do { \
202 if(DMA_ISBROKEN(dma)) DMA_INTSON(dregs); \
203 } while(0)
204
205#define for_each_dvma(dma) \
206 for((dma) = dma_chain; (dma); (dma) = (dma)->next)
207
208extern int get_dma_list(char *);
209extern int request_dma(unsigned int, __const__ char *);
210extern void free_dma(unsigned int);
211
212/* From PCI */
213
214#ifdef CONFIG_PCI
215extern int isa_dma_bridge_buggy;
216#else
217#define isa_dma_bridge_buggy (0)
218#endif
219
220#endif /* !(_ASM_SPARC64_DMA_H) */