blob: ca5ea994d68872998302e20f6a2566e143b3e1cc [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * Machine vector for IA-64.
3 *
4 * Copyright (C) 1999 Silicon Graphics, Inc.
5 * Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com>
6 * Copyright (C) Vijay Chander <vijay@engr.sgi.com>
7 * Copyright (C) 1999-2001, 2003-2004 Hewlett-Packard Co.
8 * David Mosberger-Tang <davidm@hpl.hp.com>
9 */
10#ifndef _ASM_IA64_MACHVEC_H
11#define _ASM_IA64_MACHVEC_H
12
13#include <linux/config.h>
14#include <linux/types.h>
15
16/* forward declarations: */
17struct device;
18struct pt_regs;
19struct scatterlist;
20struct page;
21struct mm_struct;
22struct pci_bus;
23
24typedef void ia64_mv_setup_t (char **);
25typedef void ia64_mv_cpu_init_t (void);
26typedef void ia64_mv_irq_init_t (void);
27typedef void ia64_mv_send_ipi_t (int, int, int, int);
28typedef void ia64_mv_timer_interrupt_t (int, void *, struct pt_regs *);
Dean Roec1902aa2005-10-27 15:41:04 -050029typedef void ia64_mv_global_tlb_purge_t (struct mm_struct *, unsigned long, unsigned long, unsigned long);
Linus Torvalds1da177e2005-04-16 15:20:36 -070030typedef void ia64_mv_tlb_migrate_finish_t (struct mm_struct *);
31typedef unsigned int ia64_mv_local_vector_to_irq (u8);
32typedef char *ia64_mv_pci_get_legacy_mem_t (struct pci_bus *);
33typedef int ia64_mv_pci_legacy_read_t (struct pci_bus *, u16 port, u32 *val,
34 u8 size);
35typedef int ia64_mv_pci_legacy_write_t (struct pci_bus *, u16 port, u32 val,
36 u8 size);
37
38/* DMA-mapping interface: */
39typedef void ia64_mv_dma_init (void);
Al Viro06a54492005-10-21 03:21:03 -040040typedef void *ia64_mv_dma_alloc_coherent (struct device *, size_t, dma_addr_t *, gfp_t);
Linus Torvalds1da177e2005-04-16 15:20:36 -070041typedef void ia64_mv_dma_free_coherent (struct device *, size_t, void *, dma_addr_t);
42typedef dma_addr_t ia64_mv_dma_map_single (struct device *, void *, size_t, int);
43typedef void ia64_mv_dma_unmap_single (struct device *, dma_addr_t, size_t, int);
44typedef int ia64_mv_dma_map_sg (struct device *, struct scatterlist *, int, int);
45typedef void ia64_mv_dma_unmap_sg (struct device *, struct scatterlist *, int, int);
46typedef void ia64_mv_dma_sync_single_for_cpu (struct device *, dma_addr_t, size_t, int);
47typedef void ia64_mv_dma_sync_sg_for_cpu (struct device *, struct scatterlist *, int, int);
48typedef void ia64_mv_dma_sync_single_for_device (struct device *, dma_addr_t, size_t, int);
49typedef void ia64_mv_dma_sync_sg_for_device (struct device *, struct scatterlist *, int, int);
50typedef int ia64_mv_dma_mapping_error (dma_addr_t dma_addr);
51typedef int ia64_mv_dma_supported (struct device *, u64);
52
53/*
54 * WARNING: The legacy I/O space is _architected_. Platforms are
55 * expected to follow this architected model (see Section 10.7 in the
56 * IA-64 Architecture Software Developer's Manual). Unfortunately,
57 * some broken machines do not follow that model, which is why we have
58 * to make the inX/outX operations part of the machine vector.
59 * Platform designers should follow the architected model whenever
60 * possible.
61 */
62typedef unsigned int ia64_mv_inb_t (unsigned long);
63typedef unsigned int ia64_mv_inw_t (unsigned long);
64typedef unsigned int ia64_mv_inl_t (unsigned long);
65typedef void ia64_mv_outb_t (unsigned char, unsigned long);
66typedef void ia64_mv_outw_t (unsigned short, unsigned long);
67typedef void ia64_mv_outl_t (unsigned int, unsigned long);
68typedef void ia64_mv_mmiowb_t (void);
69typedef unsigned char ia64_mv_readb_t (const volatile void __iomem *);
70typedef unsigned short ia64_mv_readw_t (const volatile void __iomem *);
71typedef unsigned int ia64_mv_readl_t (const volatile void __iomem *);
72typedef unsigned long ia64_mv_readq_t (const volatile void __iomem *);
73typedef unsigned char ia64_mv_readb_relaxed_t (const volatile void __iomem *);
74typedef unsigned short ia64_mv_readw_relaxed_t (const volatile void __iomem *);
75typedef unsigned int ia64_mv_readl_relaxed_t (const volatile void __iomem *);
76typedef unsigned long ia64_mv_readq_relaxed_t (const volatile void __iomem *);
77
78static inline void
79machvec_noop (void)
80{
81}
82
83static inline void
84machvec_noop_mm (struct mm_struct *mm)
85{
86}
87
88extern void machvec_setup (char **);
89extern void machvec_timer_interrupt (int, void *, struct pt_regs *);
90extern void machvec_dma_sync_single (struct device *, dma_addr_t, size_t, int);
91extern void machvec_dma_sync_sg (struct device *, struct scatterlist *, int, int);
92extern void machvec_tlb_migrate_finish (struct mm_struct *);
93
94# if defined (CONFIG_IA64_HP_SIM)
95# include <asm/machvec_hpsim.h>
96# elif defined (CONFIG_IA64_DIG)
97# include <asm/machvec_dig.h>
98# elif defined (CONFIG_IA64_HP_ZX1)
99# include <asm/machvec_hpzx1.h>
100# elif defined (CONFIG_IA64_HP_ZX1_SWIOTLB)
101# include <asm/machvec_hpzx1_swiotlb.h>
102# elif defined (CONFIG_IA64_SGI_SN2)
103# include <asm/machvec_sn2.h>
104# elif defined (CONFIG_IA64_GENERIC)
105
106# ifdef MACHVEC_PLATFORM_HEADER
107# include MACHVEC_PLATFORM_HEADER
108# else
109# define platform_name ia64_mv.name
110# define platform_setup ia64_mv.setup
111# define platform_cpu_init ia64_mv.cpu_init
112# define platform_irq_init ia64_mv.irq_init
113# define platform_send_ipi ia64_mv.send_ipi
114# define platform_timer_interrupt ia64_mv.timer_interrupt
115# define platform_global_tlb_purge ia64_mv.global_tlb_purge
116# define platform_tlb_migrate_finish ia64_mv.tlb_migrate_finish
117# define platform_dma_init ia64_mv.dma_init
118# define platform_dma_alloc_coherent ia64_mv.dma_alloc_coherent
119# define platform_dma_free_coherent ia64_mv.dma_free_coherent
120# define platform_dma_map_single ia64_mv.dma_map_single
121# define platform_dma_unmap_single ia64_mv.dma_unmap_single
122# define platform_dma_map_sg ia64_mv.dma_map_sg
123# define platform_dma_unmap_sg ia64_mv.dma_unmap_sg
124# define platform_dma_sync_single_for_cpu ia64_mv.dma_sync_single_for_cpu
125# define platform_dma_sync_sg_for_cpu ia64_mv.dma_sync_sg_for_cpu
126# define platform_dma_sync_single_for_device ia64_mv.dma_sync_single_for_device
127# define platform_dma_sync_sg_for_device ia64_mv.dma_sync_sg_for_device
128# define platform_dma_mapping_error ia64_mv.dma_mapping_error
129# define platform_dma_supported ia64_mv.dma_supported
130# define platform_local_vector_to_irq ia64_mv.local_vector_to_irq
131# define platform_pci_get_legacy_mem ia64_mv.pci_get_legacy_mem
132# define platform_pci_legacy_read ia64_mv.pci_legacy_read
133# define platform_pci_legacy_write ia64_mv.pci_legacy_write
134# define platform_inb ia64_mv.inb
135# define platform_inw ia64_mv.inw
136# define platform_inl ia64_mv.inl
137# define platform_outb ia64_mv.outb
138# define platform_outw ia64_mv.outw
139# define platform_outl ia64_mv.outl
140# define platform_mmiowb ia64_mv.mmiowb
141# define platform_readb ia64_mv.readb
142# define platform_readw ia64_mv.readw
143# define platform_readl ia64_mv.readl
144# define platform_readq ia64_mv.readq
145# define platform_readb_relaxed ia64_mv.readb_relaxed
146# define platform_readw_relaxed ia64_mv.readw_relaxed
147# define platform_readl_relaxed ia64_mv.readl_relaxed
148# define platform_readq_relaxed ia64_mv.readq_relaxed
149# endif
150
151/* __attribute__((__aligned__(16))) is required to make size of the
152 * structure multiple of 16 bytes.
153 * This will fillup the holes created because of section 3.3.1 in
154 * Software Conventions guide.
155 */
156struct ia64_machine_vector {
157 const char *name;
158 ia64_mv_setup_t *setup;
159 ia64_mv_cpu_init_t *cpu_init;
160 ia64_mv_irq_init_t *irq_init;
161 ia64_mv_send_ipi_t *send_ipi;
162 ia64_mv_timer_interrupt_t *timer_interrupt;
163 ia64_mv_global_tlb_purge_t *global_tlb_purge;
164 ia64_mv_tlb_migrate_finish_t *tlb_migrate_finish;
165 ia64_mv_dma_init *dma_init;
166 ia64_mv_dma_alloc_coherent *dma_alloc_coherent;
167 ia64_mv_dma_free_coherent *dma_free_coherent;
168 ia64_mv_dma_map_single *dma_map_single;
169 ia64_mv_dma_unmap_single *dma_unmap_single;
170 ia64_mv_dma_map_sg *dma_map_sg;
171 ia64_mv_dma_unmap_sg *dma_unmap_sg;
172 ia64_mv_dma_sync_single_for_cpu *dma_sync_single_for_cpu;
173 ia64_mv_dma_sync_sg_for_cpu *dma_sync_sg_for_cpu;
174 ia64_mv_dma_sync_single_for_device *dma_sync_single_for_device;
175 ia64_mv_dma_sync_sg_for_device *dma_sync_sg_for_device;
176 ia64_mv_dma_mapping_error *dma_mapping_error;
177 ia64_mv_dma_supported *dma_supported;
178 ia64_mv_local_vector_to_irq *local_vector_to_irq;
179 ia64_mv_pci_get_legacy_mem_t *pci_get_legacy_mem;
180 ia64_mv_pci_legacy_read_t *pci_legacy_read;
181 ia64_mv_pci_legacy_write_t *pci_legacy_write;
182 ia64_mv_inb_t *inb;
183 ia64_mv_inw_t *inw;
184 ia64_mv_inl_t *inl;
185 ia64_mv_outb_t *outb;
186 ia64_mv_outw_t *outw;
187 ia64_mv_outl_t *outl;
188 ia64_mv_mmiowb_t *mmiowb;
189 ia64_mv_readb_t *readb;
190 ia64_mv_readw_t *readw;
191 ia64_mv_readl_t *readl;
192 ia64_mv_readq_t *readq;
193 ia64_mv_readb_relaxed_t *readb_relaxed;
194 ia64_mv_readw_relaxed_t *readw_relaxed;
195 ia64_mv_readl_relaxed_t *readl_relaxed;
196 ia64_mv_readq_relaxed_t *readq_relaxed;
197} __attribute__((__aligned__(16))); /* align attrib? see above comment */
198
199#define MACHVEC_INIT(name) \
200{ \
201 #name, \
202 platform_setup, \
203 platform_cpu_init, \
204 platform_irq_init, \
205 platform_send_ipi, \
206 platform_timer_interrupt, \
207 platform_global_tlb_purge, \
208 platform_tlb_migrate_finish, \
209 platform_dma_init, \
210 platform_dma_alloc_coherent, \
211 platform_dma_free_coherent, \
212 platform_dma_map_single, \
213 platform_dma_unmap_single, \
214 platform_dma_map_sg, \
215 platform_dma_unmap_sg, \
216 platform_dma_sync_single_for_cpu, \
217 platform_dma_sync_sg_for_cpu, \
218 platform_dma_sync_single_for_device, \
219 platform_dma_sync_sg_for_device, \
220 platform_dma_mapping_error, \
221 platform_dma_supported, \
222 platform_local_vector_to_irq, \
223 platform_pci_get_legacy_mem, \
224 platform_pci_legacy_read, \
225 platform_pci_legacy_write, \
226 platform_inb, \
227 platform_inw, \
228 platform_inl, \
229 platform_outb, \
230 platform_outw, \
231 platform_outl, \
232 platform_mmiowb, \
233 platform_readb, \
234 platform_readw, \
235 platform_readl, \
236 platform_readq, \
237 platform_readb_relaxed, \
238 platform_readw_relaxed, \
239 platform_readl_relaxed, \
240 platform_readq_relaxed, \
241}
242
243extern struct ia64_machine_vector ia64_mv;
244extern void machvec_init (const char *name);
245
246# else
247# error Unknown configuration. Update asm-ia64/machvec.h.
248# endif /* CONFIG_IA64_GENERIC */
249
250/*
251 * Declare default routines which aren't declared anywhere else:
252 */
253extern ia64_mv_dma_init swiotlb_init;
254extern ia64_mv_dma_alloc_coherent swiotlb_alloc_coherent;
255extern ia64_mv_dma_free_coherent swiotlb_free_coherent;
256extern ia64_mv_dma_map_single swiotlb_map_single;
257extern ia64_mv_dma_unmap_single swiotlb_unmap_single;
258extern ia64_mv_dma_map_sg swiotlb_map_sg;
259extern ia64_mv_dma_unmap_sg swiotlb_unmap_sg;
260extern ia64_mv_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu;
261extern ia64_mv_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu;
262extern ia64_mv_dma_sync_single_for_device swiotlb_sync_single_for_device;
263extern ia64_mv_dma_sync_sg_for_device swiotlb_sync_sg_for_device;
264extern ia64_mv_dma_mapping_error swiotlb_dma_mapping_error;
265extern ia64_mv_dma_supported swiotlb_dma_supported;
266
267/*
268 * Define default versions so we can extend machvec for new platforms without having
269 * to update the machvec files for all existing platforms.
270 */
271#ifndef platform_setup
272# define platform_setup machvec_setup
273#endif
274#ifndef platform_cpu_init
275# define platform_cpu_init machvec_noop
276#endif
277#ifndef platform_irq_init
278# define platform_irq_init machvec_noop
279#endif
280
281#ifndef platform_send_ipi
282# define platform_send_ipi ia64_send_ipi /* default to architected version */
283#endif
284#ifndef platform_timer_interrupt
285# define platform_timer_interrupt machvec_timer_interrupt
286#endif
287#ifndef platform_global_tlb_purge
288# define platform_global_tlb_purge ia64_global_tlb_purge /* default to architected version */
289#endif
290#ifndef platform_tlb_migrate_finish
291# define platform_tlb_migrate_finish machvec_noop_mm
292#endif
293#ifndef platform_dma_init
294# define platform_dma_init swiotlb_init
295#endif
296#ifndef platform_dma_alloc_coherent
297# define platform_dma_alloc_coherent swiotlb_alloc_coherent
298#endif
299#ifndef platform_dma_free_coherent
300# define platform_dma_free_coherent swiotlb_free_coherent
301#endif
302#ifndef platform_dma_map_single
303# define platform_dma_map_single swiotlb_map_single
304#endif
305#ifndef platform_dma_unmap_single
306# define platform_dma_unmap_single swiotlb_unmap_single
307#endif
308#ifndef platform_dma_map_sg
309# define platform_dma_map_sg swiotlb_map_sg
310#endif
311#ifndef platform_dma_unmap_sg
312# define platform_dma_unmap_sg swiotlb_unmap_sg
313#endif
314#ifndef platform_dma_sync_single_for_cpu
315# define platform_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu
316#endif
317#ifndef platform_dma_sync_sg_for_cpu
318# define platform_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu
319#endif
320#ifndef platform_dma_sync_single_for_device
321# define platform_dma_sync_single_for_device swiotlb_sync_single_for_device
322#endif
323#ifndef platform_dma_sync_sg_for_device
324# define platform_dma_sync_sg_for_device swiotlb_sync_sg_for_device
325#endif
326#ifndef platform_dma_mapping_error
327# define platform_dma_mapping_error swiotlb_dma_mapping_error
328#endif
329#ifndef platform_dma_supported
330# define platform_dma_supported swiotlb_dma_supported
331#endif
332#ifndef platform_local_vector_to_irq
333# define platform_local_vector_to_irq __ia64_local_vector_to_irq
334#endif
335#ifndef platform_pci_get_legacy_mem
336# define platform_pci_get_legacy_mem ia64_pci_get_legacy_mem
337#endif
338#ifndef platform_pci_legacy_read
339# define platform_pci_legacy_read ia64_pci_legacy_read
340#endif
341#ifndef platform_pci_legacy_write
342# define platform_pci_legacy_write ia64_pci_legacy_write
343#endif
344#ifndef platform_inb
345# define platform_inb __ia64_inb
346#endif
347#ifndef platform_inw
348# define platform_inw __ia64_inw
349#endif
350#ifndef platform_inl
351# define platform_inl __ia64_inl
352#endif
353#ifndef platform_outb
354# define platform_outb __ia64_outb
355#endif
356#ifndef platform_outw
357# define platform_outw __ia64_outw
358#endif
359#ifndef platform_outl
360# define platform_outl __ia64_outl
361#endif
362#ifndef platform_mmiowb
363# define platform_mmiowb __ia64_mmiowb
364#endif
365#ifndef platform_readb
366# define platform_readb __ia64_readb
367#endif
368#ifndef platform_readw
369# define platform_readw __ia64_readw
370#endif
371#ifndef platform_readl
372# define platform_readl __ia64_readl
373#endif
374#ifndef platform_readq
375# define platform_readq __ia64_readq
376#endif
377#ifndef platform_readb_relaxed
378# define platform_readb_relaxed __ia64_readb_relaxed
379#endif
380#ifndef platform_readw_relaxed
381# define platform_readw_relaxed __ia64_readw_relaxed
382#endif
383#ifndef platform_readl_relaxed
384# define platform_readl_relaxed __ia64_readl_relaxed
385#endif
386#ifndef platform_readq_relaxed
387# define platform_readq_relaxed __ia64_readq_relaxed
388#endif
389
390#endif /* _ASM_IA64_MACHVEC_H */