| /* |
| * Machine vector for IA-64. |
| * |
| * Copyright (C) 1999 Silicon Graphics, Inc. |
| * Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com> |
| * Copyright (C) Vijay Chander <vijay@engr.sgi.com> |
| * Copyright (C) 1999-2001, 2003-2004 Hewlett-Packard Co. |
| * David Mosberger-Tang <davidm@hpl.hp.com> |
| */ |
| #ifndef _ASM_IA64_MACHVEC_H |
| #define _ASM_IA64_MACHVEC_H |
| |
| #include <linux/config.h> |
| #include <linux/types.h> |
| |
| /* forward declarations: */ |
| struct device; |
| struct pt_regs; |
| struct scatterlist; |
| struct page; |
| struct mm_struct; |
| struct pci_bus; |
| struct task_struct; |
| |
| typedef void ia64_mv_setup_t (char **); |
| typedef void ia64_mv_cpu_init_t (void); |
| typedef void ia64_mv_irq_init_t (void); |
| typedef void ia64_mv_send_ipi_t (int, int, int, int); |
| typedef void ia64_mv_timer_interrupt_t (int, void *, struct pt_regs *); |
| typedef void ia64_mv_global_tlb_purge_t (struct mm_struct *, unsigned long, unsigned long, unsigned long); |
| typedef void ia64_mv_tlb_migrate_finish_t (struct mm_struct *); |
| typedef unsigned int ia64_mv_local_vector_to_irq (u8); |
| typedef char *ia64_mv_pci_get_legacy_mem_t (struct pci_bus *); |
| typedef int ia64_mv_pci_legacy_read_t (struct pci_bus *, u16 port, u32 *val, |
| u8 size); |
| typedef int ia64_mv_pci_legacy_write_t (struct pci_bus *, u16 port, u32 val, |
| u8 size); |
| typedef void ia64_mv_migrate_t(struct task_struct * task); |
| |
| /* DMA-mapping interface: */ |
| typedef void ia64_mv_dma_init (void); |
| typedef void *ia64_mv_dma_alloc_coherent (struct device *, size_t, dma_addr_t *, gfp_t); |
| typedef void ia64_mv_dma_free_coherent (struct device *, size_t, void *, dma_addr_t); |
| typedef dma_addr_t ia64_mv_dma_map_single (struct device *, void *, size_t, int); |
| typedef void ia64_mv_dma_unmap_single (struct device *, dma_addr_t, size_t, int); |
| typedef int ia64_mv_dma_map_sg (struct device *, struct scatterlist *, int, int); |
| typedef void ia64_mv_dma_unmap_sg (struct device *, struct scatterlist *, int, int); |
| typedef void ia64_mv_dma_sync_single_for_cpu (struct device *, dma_addr_t, size_t, int); |
| typedef void ia64_mv_dma_sync_sg_for_cpu (struct device *, struct scatterlist *, int, int); |
| typedef void ia64_mv_dma_sync_single_for_device (struct device *, dma_addr_t, size_t, int); |
| typedef void ia64_mv_dma_sync_sg_for_device (struct device *, struct scatterlist *, int, int); |
| typedef int ia64_mv_dma_mapping_error (dma_addr_t dma_addr); |
| typedef int ia64_mv_dma_supported (struct device *, u64); |
| |
| /* |
| * WARNING: The legacy I/O space is _architected_. Platforms are |
| * expected to follow this architected model (see Section 10.7 in the |
| * IA-64 Architecture Software Developer's Manual). Unfortunately, |
| * some broken machines do not follow that model, which is why we have |
| * to make the inX/outX operations part of the machine vector. |
| * Platform designers should follow the architected model whenever |
| * possible. |
| */ |
| typedef unsigned int ia64_mv_inb_t (unsigned long); |
| typedef unsigned int ia64_mv_inw_t (unsigned long); |
| typedef unsigned int ia64_mv_inl_t (unsigned long); |
| typedef void ia64_mv_outb_t (unsigned char, unsigned long); |
| typedef void ia64_mv_outw_t (unsigned short, unsigned long); |
| typedef void ia64_mv_outl_t (unsigned int, unsigned long); |
| typedef void ia64_mv_mmiowb_t (void); |
| typedef unsigned char ia64_mv_readb_t (const volatile void __iomem *); |
| typedef unsigned short ia64_mv_readw_t (const volatile void __iomem *); |
| typedef unsigned int ia64_mv_readl_t (const volatile void __iomem *); |
| typedef unsigned long ia64_mv_readq_t (const volatile void __iomem *); |
| typedef unsigned char ia64_mv_readb_relaxed_t (const volatile void __iomem *); |
| typedef unsigned short ia64_mv_readw_relaxed_t (const volatile void __iomem *); |
| typedef unsigned int ia64_mv_readl_relaxed_t (const volatile void __iomem *); |
| typedef unsigned long ia64_mv_readq_relaxed_t (const volatile void __iomem *); |
| |
| static inline void |
| machvec_noop (void) |
| { |
| } |
| |
| static inline void |
| machvec_noop_mm (struct mm_struct *mm) |
| { |
| } |
| |
| static inline void |
| machvec_noop_task (struct task_struct *task) |
| { |
| } |
| |
| extern void machvec_setup (char **); |
| extern void machvec_timer_interrupt (int, void *, struct pt_regs *); |
| extern void machvec_dma_sync_single (struct device *, dma_addr_t, size_t, int); |
| extern void machvec_dma_sync_sg (struct device *, struct scatterlist *, int, int); |
| extern void machvec_tlb_migrate_finish (struct mm_struct *); |
| |
| # if defined (CONFIG_IA64_HP_SIM) |
| # include <asm/machvec_hpsim.h> |
| # elif defined (CONFIG_IA64_DIG) |
| # include <asm/machvec_dig.h> |
| # elif defined (CONFIG_IA64_HP_ZX1) |
| # include <asm/machvec_hpzx1.h> |
| # elif defined (CONFIG_IA64_HP_ZX1_SWIOTLB) |
| # include <asm/machvec_hpzx1_swiotlb.h> |
| # elif defined (CONFIG_IA64_SGI_SN2) |
| # include <asm/machvec_sn2.h> |
| # elif defined (CONFIG_IA64_GENERIC) |
| |
| # ifdef MACHVEC_PLATFORM_HEADER |
| # include MACHVEC_PLATFORM_HEADER |
| # else |
| # define platform_name ia64_mv.name |
| # define platform_setup ia64_mv.setup |
| # define platform_cpu_init ia64_mv.cpu_init |
| # define platform_irq_init ia64_mv.irq_init |
| # define platform_send_ipi ia64_mv.send_ipi |
| # define platform_timer_interrupt ia64_mv.timer_interrupt |
| # define platform_global_tlb_purge ia64_mv.global_tlb_purge |
| # define platform_tlb_migrate_finish ia64_mv.tlb_migrate_finish |
| # define platform_dma_init ia64_mv.dma_init |
| # define platform_dma_alloc_coherent ia64_mv.dma_alloc_coherent |
| # define platform_dma_free_coherent ia64_mv.dma_free_coherent |
| # define platform_dma_map_single ia64_mv.dma_map_single |
| # define platform_dma_unmap_single ia64_mv.dma_unmap_single |
| # define platform_dma_map_sg ia64_mv.dma_map_sg |
| # define platform_dma_unmap_sg ia64_mv.dma_unmap_sg |
| # define platform_dma_sync_single_for_cpu ia64_mv.dma_sync_single_for_cpu |
| # define platform_dma_sync_sg_for_cpu ia64_mv.dma_sync_sg_for_cpu |
| # define platform_dma_sync_single_for_device ia64_mv.dma_sync_single_for_device |
| # define platform_dma_sync_sg_for_device ia64_mv.dma_sync_sg_for_device |
| # define platform_dma_mapping_error ia64_mv.dma_mapping_error |
| # define platform_dma_supported ia64_mv.dma_supported |
| # define platform_local_vector_to_irq ia64_mv.local_vector_to_irq |
| # define platform_pci_get_legacy_mem ia64_mv.pci_get_legacy_mem |
| # define platform_pci_legacy_read ia64_mv.pci_legacy_read |
| # define platform_pci_legacy_write ia64_mv.pci_legacy_write |
| # define platform_inb ia64_mv.inb |
| # define platform_inw ia64_mv.inw |
| # define platform_inl ia64_mv.inl |
| # define platform_outb ia64_mv.outb |
| # define platform_outw ia64_mv.outw |
| # define platform_outl ia64_mv.outl |
| # define platform_mmiowb ia64_mv.mmiowb |
| # define platform_readb ia64_mv.readb |
| # define platform_readw ia64_mv.readw |
| # define platform_readl ia64_mv.readl |
| # define platform_readq ia64_mv.readq |
| # define platform_readb_relaxed ia64_mv.readb_relaxed |
| # define platform_readw_relaxed ia64_mv.readw_relaxed |
| # define platform_readl_relaxed ia64_mv.readl_relaxed |
| # define platform_readq_relaxed ia64_mv.readq_relaxed |
| # define platform_migrate ia64_mv.migrate |
| # endif |
| |
| /* __attribute__((__aligned__(16))) is required to make size of the |
| * structure multiple of 16 bytes. |
| * This will fillup the holes created because of section 3.3.1 in |
| * Software Conventions guide. |
| */ |
| struct ia64_machine_vector { |
| const char *name; |
| ia64_mv_setup_t *setup; |
| ia64_mv_cpu_init_t *cpu_init; |
| ia64_mv_irq_init_t *irq_init; |
| ia64_mv_send_ipi_t *send_ipi; |
| ia64_mv_timer_interrupt_t *timer_interrupt; |
| ia64_mv_global_tlb_purge_t *global_tlb_purge; |
| ia64_mv_tlb_migrate_finish_t *tlb_migrate_finish; |
| ia64_mv_dma_init *dma_init; |
| ia64_mv_dma_alloc_coherent *dma_alloc_coherent; |
| ia64_mv_dma_free_coherent *dma_free_coherent; |
| ia64_mv_dma_map_single *dma_map_single; |
| ia64_mv_dma_unmap_single *dma_unmap_single; |
| ia64_mv_dma_map_sg *dma_map_sg; |
| ia64_mv_dma_unmap_sg *dma_unmap_sg; |
| ia64_mv_dma_sync_single_for_cpu *dma_sync_single_for_cpu; |
| ia64_mv_dma_sync_sg_for_cpu *dma_sync_sg_for_cpu; |
| ia64_mv_dma_sync_single_for_device *dma_sync_single_for_device; |
| ia64_mv_dma_sync_sg_for_device *dma_sync_sg_for_device; |
| ia64_mv_dma_mapping_error *dma_mapping_error; |
| ia64_mv_dma_supported *dma_supported; |
| ia64_mv_local_vector_to_irq *local_vector_to_irq; |
| ia64_mv_pci_get_legacy_mem_t *pci_get_legacy_mem; |
| ia64_mv_pci_legacy_read_t *pci_legacy_read; |
| ia64_mv_pci_legacy_write_t *pci_legacy_write; |
| ia64_mv_inb_t *inb; |
| ia64_mv_inw_t *inw; |
| ia64_mv_inl_t *inl; |
| ia64_mv_outb_t *outb; |
| ia64_mv_outw_t *outw; |
| ia64_mv_outl_t *outl; |
| ia64_mv_mmiowb_t *mmiowb; |
| ia64_mv_readb_t *readb; |
| ia64_mv_readw_t *readw; |
| ia64_mv_readl_t *readl; |
| ia64_mv_readq_t *readq; |
| ia64_mv_readb_relaxed_t *readb_relaxed; |
| ia64_mv_readw_relaxed_t *readw_relaxed; |
| ia64_mv_readl_relaxed_t *readl_relaxed; |
| ia64_mv_readq_relaxed_t *readq_relaxed; |
| ia64_mv_migrate_t *migrate; |
| } __attribute__((__aligned__(16))); /* align attrib? see above comment */ |
| |
| #define MACHVEC_INIT(name) \ |
| { \ |
| #name, \ |
| platform_setup, \ |
| platform_cpu_init, \ |
| platform_irq_init, \ |
| platform_send_ipi, \ |
| platform_timer_interrupt, \ |
| platform_global_tlb_purge, \ |
| platform_tlb_migrate_finish, \ |
| platform_dma_init, \ |
| platform_dma_alloc_coherent, \ |
| platform_dma_free_coherent, \ |
| platform_dma_map_single, \ |
| platform_dma_unmap_single, \ |
| platform_dma_map_sg, \ |
| platform_dma_unmap_sg, \ |
| platform_dma_sync_single_for_cpu, \ |
| platform_dma_sync_sg_for_cpu, \ |
| platform_dma_sync_single_for_device, \ |
| platform_dma_sync_sg_for_device, \ |
| platform_dma_mapping_error, \ |
| platform_dma_supported, \ |
| platform_local_vector_to_irq, \ |
| platform_pci_get_legacy_mem, \ |
| platform_pci_legacy_read, \ |
| platform_pci_legacy_write, \ |
| platform_inb, \ |
| platform_inw, \ |
| platform_inl, \ |
| platform_outb, \ |
| platform_outw, \ |
| platform_outl, \ |
| platform_mmiowb, \ |
| platform_readb, \ |
| platform_readw, \ |
| platform_readl, \ |
| platform_readq, \ |
| platform_readb_relaxed, \ |
| platform_readw_relaxed, \ |
| platform_readl_relaxed, \ |
| platform_readq_relaxed, \ |
| platform_migrate, \ |
| } |
| |
| extern struct ia64_machine_vector ia64_mv; |
| extern void machvec_init (const char *name); |
| |
| # else |
| # error Unknown configuration. Update asm-ia64/machvec.h. |
| # endif /* CONFIG_IA64_GENERIC */ |
| |
| /* |
| * Declare default routines which aren't declared anywhere else: |
| */ |
| extern ia64_mv_dma_init swiotlb_init; |
| extern ia64_mv_dma_alloc_coherent swiotlb_alloc_coherent; |
| extern ia64_mv_dma_free_coherent swiotlb_free_coherent; |
| extern ia64_mv_dma_map_single swiotlb_map_single; |
| extern ia64_mv_dma_unmap_single swiotlb_unmap_single; |
| extern ia64_mv_dma_map_sg swiotlb_map_sg; |
| extern ia64_mv_dma_unmap_sg swiotlb_unmap_sg; |
| extern ia64_mv_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu; |
| extern ia64_mv_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu; |
| extern ia64_mv_dma_sync_single_for_device swiotlb_sync_single_for_device; |
| extern ia64_mv_dma_sync_sg_for_device swiotlb_sync_sg_for_device; |
| extern ia64_mv_dma_mapping_error swiotlb_dma_mapping_error; |
| extern ia64_mv_dma_supported swiotlb_dma_supported; |
| |
| /* |
| * Define default versions so we can extend machvec for new platforms without having |
| * to update the machvec files for all existing platforms. |
| */ |
| #ifndef platform_setup |
| # define platform_setup machvec_setup |
| #endif |
| #ifndef platform_cpu_init |
| # define platform_cpu_init machvec_noop |
| #endif |
| #ifndef platform_irq_init |
| # define platform_irq_init machvec_noop |
| #endif |
| |
| #ifndef platform_send_ipi |
| # define platform_send_ipi ia64_send_ipi /* default to architected version */ |
| #endif |
| #ifndef platform_timer_interrupt |
| # define platform_timer_interrupt machvec_timer_interrupt |
| #endif |
| #ifndef platform_global_tlb_purge |
| # define platform_global_tlb_purge ia64_global_tlb_purge /* default to architected version */ |
| #endif |
| #ifndef platform_tlb_migrate_finish |
| # define platform_tlb_migrate_finish machvec_noop_mm |
| #endif |
| #ifndef platform_dma_init |
| # define platform_dma_init swiotlb_init |
| #endif |
| #ifndef platform_dma_alloc_coherent |
| # define platform_dma_alloc_coherent swiotlb_alloc_coherent |
| #endif |
| #ifndef platform_dma_free_coherent |
| # define platform_dma_free_coherent swiotlb_free_coherent |
| #endif |
| #ifndef platform_dma_map_single |
| # define platform_dma_map_single swiotlb_map_single |
| #endif |
| #ifndef platform_dma_unmap_single |
| # define platform_dma_unmap_single swiotlb_unmap_single |
| #endif |
| #ifndef platform_dma_map_sg |
| # define platform_dma_map_sg swiotlb_map_sg |
| #endif |
| #ifndef platform_dma_unmap_sg |
| # define platform_dma_unmap_sg swiotlb_unmap_sg |
| #endif |
| #ifndef platform_dma_sync_single_for_cpu |
| # define platform_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu |
| #endif |
| #ifndef platform_dma_sync_sg_for_cpu |
| # define platform_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu |
| #endif |
| #ifndef platform_dma_sync_single_for_device |
| # define platform_dma_sync_single_for_device swiotlb_sync_single_for_device |
| #endif |
| #ifndef platform_dma_sync_sg_for_device |
| # define platform_dma_sync_sg_for_device swiotlb_sync_sg_for_device |
| #endif |
| #ifndef platform_dma_mapping_error |
| # define platform_dma_mapping_error swiotlb_dma_mapping_error |
| #endif |
| #ifndef platform_dma_supported |
| # define platform_dma_supported swiotlb_dma_supported |
| #endif |
| #ifndef platform_local_vector_to_irq |
| # define platform_local_vector_to_irq __ia64_local_vector_to_irq |
| #endif |
| #ifndef platform_pci_get_legacy_mem |
| # define platform_pci_get_legacy_mem ia64_pci_get_legacy_mem |
| #endif |
| #ifndef platform_pci_legacy_read |
| # define platform_pci_legacy_read ia64_pci_legacy_read |
| extern int ia64_pci_legacy_read(struct pci_bus *bus, u16 port, u32 *val, u8 size); |
| #endif |
| #ifndef platform_pci_legacy_write |
| # define platform_pci_legacy_write ia64_pci_legacy_write |
| extern int ia64_pci_legacy_write(struct pci_bus *bus, u16 port, u32 val, u8 size); |
| #endif |
| #ifndef platform_inb |
| # define platform_inb __ia64_inb |
| #endif |
| #ifndef platform_inw |
| # define platform_inw __ia64_inw |
| #endif |
| #ifndef platform_inl |
| # define platform_inl __ia64_inl |
| #endif |
| #ifndef platform_outb |
| # define platform_outb __ia64_outb |
| #endif |
| #ifndef platform_outw |
| # define platform_outw __ia64_outw |
| #endif |
| #ifndef platform_outl |
| # define platform_outl __ia64_outl |
| #endif |
| #ifndef platform_mmiowb |
| # define platform_mmiowb __ia64_mmiowb |
| #endif |
| #ifndef platform_readb |
| # define platform_readb __ia64_readb |
| #endif |
| #ifndef platform_readw |
| # define platform_readw __ia64_readw |
| #endif |
| #ifndef platform_readl |
| # define platform_readl __ia64_readl |
| #endif |
| #ifndef platform_readq |
| # define platform_readq __ia64_readq |
| #endif |
| #ifndef platform_readb_relaxed |
| # define platform_readb_relaxed __ia64_readb_relaxed |
| #endif |
| #ifndef platform_readw_relaxed |
| # define platform_readw_relaxed __ia64_readw_relaxed |
| #endif |
| #ifndef platform_readl_relaxed |
| # define platform_readl_relaxed __ia64_readl_relaxed |
| #endif |
| #ifndef platform_readq_relaxed |
| # define platform_readq_relaxed __ia64_readq_relaxed |
| #endif |
| #ifndef platform_migrate |
| # define platform_migrate machvec_noop_task |
| #endif |
| |
| #endif /* _ASM_IA64_MACHVEC_H */ |