62c4f0a2d5
Signed-off-by: David Woodhouse <dwmw2@infradead.org>
403 lines
14 KiB
C
403 lines
14 KiB
C
/*
|
|
* Machine vector for IA-64.
|
|
*
|
|
* Copyright (C) 1999 Silicon Graphics, Inc.
|
|
* Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com>
|
|
* Copyright (C) Vijay Chander <vijay@engr.sgi.com>
|
|
* Copyright (C) 1999-2001, 2003-2004 Hewlett-Packard Co.
|
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
|
*/
|
|
#ifndef _ASM_IA64_MACHVEC_H
|
|
#define _ASM_IA64_MACHVEC_H
|
|
|
|
#include <linux/types.h>
|
|
|
|
/* forward declarations: */
|
|
struct device;
|
|
struct pt_regs;
|
|
struct scatterlist;
|
|
struct page;
|
|
struct mm_struct;
|
|
struct pci_bus;
|
|
struct task_struct;
|
|
|
|
typedef void ia64_mv_setup_t (char **);
|
|
typedef void ia64_mv_cpu_init_t (void);
|
|
typedef void ia64_mv_irq_init_t (void);
|
|
typedef void ia64_mv_send_ipi_t (int, int, int, int);
|
|
typedef void ia64_mv_timer_interrupt_t (int, void *, struct pt_regs *);
|
|
typedef void ia64_mv_global_tlb_purge_t (struct mm_struct *, unsigned long, unsigned long, unsigned long);
|
|
typedef void ia64_mv_tlb_migrate_finish_t (struct mm_struct *);
|
|
typedef unsigned int ia64_mv_local_vector_to_irq (u8);
|
|
typedef char *ia64_mv_pci_get_legacy_mem_t (struct pci_bus *);
|
|
typedef int ia64_mv_pci_legacy_read_t (struct pci_bus *, u16 port, u32 *val,
|
|
u8 size);
|
|
typedef int ia64_mv_pci_legacy_write_t (struct pci_bus *, u16 port, u32 val,
|
|
u8 size);
|
|
typedef void ia64_mv_migrate_t(struct task_struct * task);
|
|
|
|
/* DMA-mapping interface: */
|
|
typedef void ia64_mv_dma_init (void);
|
|
typedef void *ia64_mv_dma_alloc_coherent (struct device *, size_t, dma_addr_t *, gfp_t);
|
|
typedef void ia64_mv_dma_free_coherent (struct device *, size_t, void *, dma_addr_t);
|
|
typedef dma_addr_t ia64_mv_dma_map_single (struct device *, void *, size_t, int);
|
|
typedef void ia64_mv_dma_unmap_single (struct device *, dma_addr_t, size_t, int);
|
|
typedef int ia64_mv_dma_map_sg (struct device *, struct scatterlist *, int, int);
|
|
typedef void ia64_mv_dma_unmap_sg (struct device *, struct scatterlist *, int, int);
|
|
typedef void ia64_mv_dma_sync_single_for_cpu (struct device *, dma_addr_t, size_t, int);
|
|
typedef void ia64_mv_dma_sync_sg_for_cpu (struct device *, struct scatterlist *, int, int);
|
|
typedef void ia64_mv_dma_sync_single_for_device (struct device *, dma_addr_t, size_t, int);
|
|
typedef void ia64_mv_dma_sync_sg_for_device (struct device *, struct scatterlist *, int, int);
|
|
typedef int ia64_mv_dma_mapping_error (dma_addr_t dma_addr);
|
|
typedef int ia64_mv_dma_supported (struct device *, u64);
|
|
|
|
/*
|
|
* WARNING: The legacy I/O space is _architected_. Platforms are
|
|
* expected to follow this architected model (see Section 10.7 in the
|
|
* IA-64 Architecture Software Developer's Manual). Unfortunately,
|
|
* some broken machines do not follow that model, which is why we have
|
|
* to make the inX/outX operations part of the machine vector.
|
|
* Platform designers should follow the architected model whenever
|
|
* possible.
|
|
*/
|
|
typedef unsigned int ia64_mv_inb_t (unsigned long);
|
|
typedef unsigned int ia64_mv_inw_t (unsigned long);
|
|
typedef unsigned int ia64_mv_inl_t (unsigned long);
|
|
typedef void ia64_mv_outb_t (unsigned char, unsigned long);
|
|
typedef void ia64_mv_outw_t (unsigned short, unsigned long);
|
|
typedef void ia64_mv_outl_t (unsigned int, unsigned long);
|
|
typedef void ia64_mv_mmiowb_t (void);
|
|
typedef unsigned char ia64_mv_readb_t (const volatile void __iomem *);
|
|
typedef unsigned short ia64_mv_readw_t (const volatile void __iomem *);
|
|
typedef unsigned int ia64_mv_readl_t (const volatile void __iomem *);
|
|
typedef unsigned long ia64_mv_readq_t (const volatile void __iomem *);
|
|
typedef unsigned char ia64_mv_readb_relaxed_t (const volatile void __iomem *);
|
|
typedef unsigned short ia64_mv_readw_relaxed_t (const volatile void __iomem *);
|
|
typedef unsigned int ia64_mv_readl_relaxed_t (const volatile void __iomem *);
|
|
typedef unsigned long ia64_mv_readq_relaxed_t (const volatile void __iomem *);
|
|
|
|
static inline void
|
|
machvec_noop (void)
|
|
{
|
|
}
|
|
|
|
static inline void
|
|
machvec_noop_mm (struct mm_struct *mm)
|
|
{
|
|
}
|
|
|
|
static inline void
|
|
machvec_noop_task (struct task_struct *task)
|
|
{
|
|
}
|
|
|
|
extern void machvec_setup (char **);
|
|
extern void machvec_timer_interrupt (int, void *, struct pt_regs *);
|
|
extern void machvec_dma_sync_single (struct device *, dma_addr_t, size_t, int);
|
|
extern void machvec_dma_sync_sg (struct device *, struct scatterlist *, int, int);
|
|
extern void machvec_tlb_migrate_finish (struct mm_struct *);
|
|
|
|
# if defined (CONFIG_IA64_HP_SIM)
|
|
# include <asm/machvec_hpsim.h>
|
|
# elif defined (CONFIG_IA64_DIG)
|
|
# include <asm/machvec_dig.h>
|
|
# elif defined (CONFIG_IA64_HP_ZX1)
|
|
# include <asm/machvec_hpzx1.h>
|
|
# elif defined (CONFIG_IA64_HP_ZX1_SWIOTLB)
|
|
# include <asm/machvec_hpzx1_swiotlb.h>
|
|
# elif defined (CONFIG_IA64_SGI_SN2)
|
|
# include <asm/machvec_sn2.h>
|
|
# elif defined (CONFIG_IA64_GENERIC)
|
|
|
|
# ifdef MACHVEC_PLATFORM_HEADER
|
|
# include MACHVEC_PLATFORM_HEADER
|
|
# else
|
|
# define platform_name ia64_mv.name
|
|
# define platform_setup ia64_mv.setup
|
|
# define platform_cpu_init ia64_mv.cpu_init
|
|
# define platform_irq_init ia64_mv.irq_init
|
|
# define platform_send_ipi ia64_mv.send_ipi
|
|
# define platform_timer_interrupt ia64_mv.timer_interrupt
|
|
# define platform_global_tlb_purge ia64_mv.global_tlb_purge
|
|
# define platform_tlb_migrate_finish ia64_mv.tlb_migrate_finish
|
|
# define platform_dma_init ia64_mv.dma_init
|
|
# define platform_dma_alloc_coherent ia64_mv.dma_alloc_coherent
|
|
# define platform_dma_free_coherent ia64_mv.dma_free_coherent
|
|
# define platform_dma_map_single ia64_mv.dma_map_single
|
|
# define platform_dma_unmap_single ia64_mv.dma_unmap_single
|
|
# define platform_dma_map_sg ia64_mv.dma_map_sg
|
|
# define platform_dma_unmap_sg ia64_mv.dma_unmap_sg
|
|
# define platform_dma_sync_single_for_cpu ia64_mv.dma_sync_single_for_cpu
|
|
# define platform_dma_sync_sg_for_cpu ia64_mv.dma_sync_sg_for_cpu
|
|
# define platform_dma_sync_single_for_device ia64_mv.dma_sync_single_for_device
|
|
# define platform_dma_sync_sg_for_device ia64_mv.dma_sync_sg_for_device
|
|
# define platform_dma_mapping_error ia64_mv.dma_mapping_error
|
|
# define platform_dma_supported ia64_mv.dma_supported
|
|
# define platform_local_vector_to_irq ia64_mv.local_vector_to_irq
|
|
# define platform_pci_get_legacy_mem ia64_mv.pci_get_legacy_mem
|
|
# define platform_pci_legacy_read ia64_mv.pci_legacy_read
|
|
# define platform_pci_legacy_write ia64_mv.pci_legacy_write
|
|
# define platform_inb ia64_mv.inb
|
|
# define platform_inw ia64_mv.inw
|
|
# define platform_inl ia64_mv.inl
|
|
# define platform_outb ia64_mv.outb
|
|
# define platform_outw ia64_mv.outw
|
|
# define platform_outl ia64_mv.outl
|
|
# define platform_mmiowb ia64_mv.mmiowb
|
|
# define platform_readb ia64_mv.readb
|
|
# define platform_readw ia64_mv.readw
|
|
# define platform_readl ia64_mv.readl
|
|
# define platform_readq ia64_mv.readq
|
|
# define platform_readb_relaxed ia64_mv.readb_relaxed
|
|
# define platform_readw_relaxed ia64_mv.readw_relaxed
|
|
# define platform_readl_relaxed ia64_mv.readl_relaxed
|
|
# define platform_readq_relaxed ia64_mv.readq_relaxed
|
|
# define platform_migrate ia64_mv.migrate
|
|
# endif
|
|
|
|
/* __attribute__((__aligned__(16))) is required to make size of the
|
|
* structure multiple of 16 bytes.
|
|
* This will fillup the holes created because of section 3.3.1 in
|
|
* Software Conventions guide.
|
|
*/
|
|
struct ia64_machine_vector {
|
|
const char *name;
|
|
ia64_mv_setup_t *setup;
|
|
ia64_mv_cpu_init_t *cpu_init;
|
|
ia64_mv_irq_init_t *irq_init;
|
|
ia64_mv_send_ipi_t *send_ipi;
|
|
ia64_mv_timer_interrupt_t *timer_interrupt;
|
|
ia64_mv_global_tlb_purge_t *global_tlb_purge;
|
|
ia64_mv_tlb_migrate_finish_t *tlb_migrate_finish;
|
|
ia64_mv_dma_init *dma_init;
|
|
ia64_mv_dma_alloc_coherent *dma_alloc_coherent;
|
|
ia64_mv_dma_free_coherent *dma_free_coherent;
|
|
ia64_mv_dma_map_single *dma_map_single;
|
|
ia64_mv_dma_unmap_single *dma_unmap_single;
|
|
ia64_mv_dma_map_sg *dma_map_sg;
|
|
ia64_mv_dma_unmap_sg *dma_unmap_sg;
|
|
ia64_mv_dma_sync_single_for_cpu *dma_sync_single_for_cpu;
|
|
ia64_mv_dma_sync_sg_for_cpu *dma_sync_sg_for_cpu;
|
|
ia64_mv_dma_sync_single_for_device *dma_sync_single_for_device;
|
|
ia64_mv_dma_sync_sg_for_device *dma_sync_sg_for_device;
|
|
ia64_mv_dma_mapping_error *dma_mapping_error;
|
|
ia64_mv_dma_supported *dma_supported;
|
|
ia64_mv_local_vector_to_irq *local_vector_to_irq;
|
|
ia64_mv_pci_get_legacy_mem_t *pci_get_legacy_mem;
|
|
ia64_mv_pci_legacy_read_t *pci_legacy_read;
|
|
ia64_mv_pci_legacy_write_t *pci_legacy_write;
|
|
ia64_mv_inb_t *inb;
|
|
ia64_mv_inw_t *inw;
|
|
ia64_mv_inl_t *inl;
|
|
ia64_mv_outb_t *outb;
|
|
ia64_mv_outw_t *outw;
|
|
ia64_mv_outl_t *outl;
|
|
ia64_mv_mmiowb_t *mmiowb;
|
|
ia64_mv_readb_t *readb;
|
|
ia64_mv_readw_t *readw;
|
|
ia64_mv_readl_t *readl;
|
|
ia64_mv_readq_t *readq;
|
|
ia64_mv_readb_relaxed_t *readb_relaxed;
|
|
ia64_mv_readw_relaxed_t *readw_relaxed;
|
|
ia64_mv_readl_relaxed_t *readl_relaxed;
|
|
ia64_mv_readq_relaxed_t *readq_relaxed;
|
|
ia64_mv_migrate_t *migrate;
|
|
} __attribute__((__aligned__(16))); /* align attrib? see above comment */
|
|
|
|
#define MACHVEC_INIT(name) \
|
|
{ \
|
|
#name, \
|
|
platform_setup, \
|
|
platform_cpu_init, \
|
|
platform_irq_init, \
|
|
platform_send_ipi, \
|
|
platform_timer_interrupt, \
|
|
platform_global_tlb_purge, \
|
|
platform_tlb_migrate_finish, \
|
|
platform_dma_init, \
|
|
platform_dma_alloc_coherent, \
|
|
platform_dma_free_coherent, \
|
|
platform_dma_map_single, \
|
|
platform_dma_unmap_single, \
|
|
platform_dma_map_sg, \
|
|
platform_dma_unmap_sg, \
|
|
platform_dma_sync_single_for_cpu, \
|
|
platform_dma_sync_sg_for_cpu, \
|
|
platform_dma_sync_single_for_device, \
|
|
platform_dma_sync_sg_for_device, \
|
|
platform_dma_mapping_error, \
|
|
platform_dma_supported, \
|
|
platform_local_vector_to_irq, \
|
|
platform_pci_get_legacy_mem, \
|
|
platform_pci_legacy_read, \
|
|
platform_pci_legacy_write, \
|
|
platform_inb, \
|
|
platform_inw, \
|
|
platform_inl, \
|
|
platform_outb, \
|
|
platform_outw, \
|
|
platform_outl, \
|
|
platform_mmiowb, \
|
|
platform_readb, \
|
|
platform_readw, \
|
|
platform_readl, \
|
|
platform_readq, \
|
|
platform_readb_relaxed, \
|
|
platform_readw_relaxed, \
|
|
platform_readl_relaxed, \
|
|
platform_readq_relaxed, \
|
|
platform_migrate, \
|
|
}
|
|
|
|
extern struct ia64_machine_vector ia64_mv;
|
|
extern void machvec_init (const char *name);
|
|
|
|
# else
|
|
# error Unknown configuration. Update asm-ia64/machvec.h.
|
|
# endif /* CONFIG_IA64_GENERIC */
|
|
|
|
/*
|
|
* Declare default routines which aren't declared anywhere else:
|
|
*/
|
|
extern ia64_mv_dma_init swiotlb_init;
|
|
extern ia64_mv_dma_alloc_coherent swiotlb_alloc_coherent;
|
|
extern ia64_mv_dma_free_coherent swiotlb_free_coherent;
|
|
extern ia64_mv_dma_map_single swiotlb_map_single;
|
|
extern ia64_mv_dma_unmap_single swiotlb_unmap_single;
|
|
extern ia64_mv_dma_map_sg swiotlb_map_sg;
|
|
extern ia64_mv_dma_unmap_sg swiotlb_unmap_sg;
|
|
extern ia64_mv_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu;
|
|
extern ia64_mv_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu;
|
|
extern ia64_mv_dma_sync_single_for_device swiotlb_sync_single_for_device;
|
|
extern ia64_mv_dma_sync_sg_for_device swiotlb_sync_sg_for_device;
|
|
extern ia64_mv_dma_mapping_error swiotlb_dma_mapping_error;
|
|
extern ia64_mv_dma_supported swiotlb_dma_supported;
|
|
|
|
/*
|
|
* Define default versions so we can extend machvec for new platforms without having
|
|
* to update the machvec files for all existing platforms.
|
|
*/
|
|
#ifndef platform_setup
|
|
# define platform_setup machvec_setup
|
|
#endif
|
|
#ifndef platform_cpu_init
|
|
# define platform_cpu_init machvec_noop
|
|
#endif
|
|
#ifndef platform_irq_init
|
|
# define platform_irq_init machvec_noop
|
|
#endif
|
|
|
|
#ifndef platform_send_ipi
|
|
# define platform_send_ipi ia64_send_ipi /* default to architected version */
|
|
#endif
|
|
#ifndef platform_timer_interrupt
|
|
# define platform_timer_interrupt machvec_timer_interrupt
|
|
#endif
|
|
#ifndef platform_global_tlb_purge
|
|
# define platform_global_tlb_purge ia64_global_tlb_purge /* default to architected version */
|
|
#endif
|
|
#ifndef platform_tlb_migrate_finish
|
|
# define platform_tlb_migrate_finish machvec_noop_mm
|
|
#endif
|
|
#ifndef platform_dma_init
|
|
# define platform_dma_init swiotlb_init
|
|
#endif
|
|
#ifndef platform_dma_alloc_coherent
|
|
# define platform_dma_alloc_coherent swiotlb_alloc_coherent
|
|
#endif
|
|
#ifndef platform_dma_free_coherent
|
|
# define platform_dma_free_coherent swiotlb_free_coherent
|
|
#endif
|
|
#ifndef platform_dma_map_single
|
|
# define platform_dma_map_single swiotlb_map_single
|
|
#endif
|
|
#ifndef platform_dma_unmap_single
|
|
# define platform_dma_unmap_single swiotlb_unmap_single
|
|
#endif
|
|
#ifndef platform_dma_map_sg
|
|
# define platform_dma_map_sg swiotlb_map_sg
|
|
#endif
|
|
#ifndef platform_dma_unmap_sg
|
|
# define platform_dma_unmap_sg swiotlb_unmap_sg
|
|
#endif
|
|
#ifndef platform_dma_sync_single_for_cpu
|
|
# define platform_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu
|
|
#endif
|
|
#ifndef platform_dma_sync_sg_for_cpu
|
|
# define platform_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu
|
|
#endif
|
|
#ifndef platform_dma_sync_single_for_device
|
|
# define platform_dma_sync_single_for_device swiotlb_sync_single_for_device
|
|
#endif
|
|
#ifndef platform_dma_sync_sg_for_device
|
|
# define platform_dma_sync_sg_for_device swiotlb_sync_sg_for_device
|
|
#endif
|
|
#ifndef platform_dma_mapping_error
|
|
# define platform_dma_mapping_error swiotlb_dma_mapping_error
|
|
#endif
|
|
#ifndef platform_dma_supported
|
|
# define platform_dma_supported swiotlb_dma_supported
|
|
#endif
|
|
#ifndef platform_local_vector_to_irq
|
|
# define platform_local_vector_to_irq __ia64_local_vector_to_irq
|
|
#endif
|
|
#ifndef platform_pci_get_legacy_mem
|
|
# define platform_pci_get_legacy_mem ia64_pci_get_legacy_mem
|
|
#endif
|
|
#ifndef platform_pci_legacy_read
|
|
# define platform_pci_legacy_read ia64_pci_legacy_read
|
|
#endif
|
|
#ifndef platform_pci_legacy_write
|
|
# define platform_pci_legacy_write ia64_pci_legacy_write
|
|
#endif
|
|
#ifndef platform_inb
|
|
# define platform_inb __ia64_inb
|
|
#endif
|
|
#ifndef platform_inw
|
|
# define platform_inw __ia64_inw
|
|
#endif
|
|
#ifndef platform_inl
|
|
# define platform_inl __ia64_inl
|
|
#endif
|
|
#ifndef platform_outb
|
|
# define platform_outb __ia64_outb
|
|
#endif
|
|
#ifndef platform_outw
|
|
# define platform_outw __ia64_outw
|
|
#endif
|
|
#ifndef platform_outl
|
|
# define platform_outl __ia64_outl
|
|
#endif
|
|
#ifndef platform_mmiowb
|
|
# define platform_mmiowb __ia64_mmiowb
|
|
#endif
|
|
#ifndef platform_readb
|
|
# define platform_readb __ia64_readb
|
|
#endif
|
|
#ifndef platform_readw
|
|
# define platform_readw __ia64_readw
|
|
#endif
|
|
#ifndef platform_readl
|
|
# define platform_readl __ia64_readl
|
|
#endif
|
|
#ifndef platform_readq
|
|
# define platform_readq __ia64_readq
|
|
#endif
|
|
#ifndef platform_readb_relaxed
|
|
# define platform_readb_relaxed __ia64_readb_relaxed
|
|
#endif
|
|
#ifndef platform_readw_relaxed
|
|
# define platform_readw_relaxed __ia64_readw_relaxed
|
|
#endif
|
|
#ifndef platform_readl_relaxed
|
|
# define platform_readl_relaxed __ia64_readl_relaxed
|
|
#endif
|
|
#ifndef platform_readq_relaxed
|
|
# define platform_readq_relaxed __ia64_readq_relaxed
|
|
#endif
|
|
#ifndef platform_migrate
|
|
# define platform_migrate machvec_noop_task
|
|
#endif
|
|
|
|
#endif /* _ASM_IA64_MACHVEC_H */
|