568 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			568 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * arch/arm/mach-ixp4xx/include/mach/io.h
 | |
|  *
 | |
|  * Author: Deepak Saxena <dsaxena@plexity.net>
 | |
|  *
 | |
|  * Copyright (C) 2002-2005  MontaVista Software, Inc.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License version 2 as
 | |
|  * published by the Free Software Foundation.
 | |
|  */
 | |
| 
 | |
| #ifndef __ASM_ARM_ARCH_IO_H
 | |
| #define __ASM_ARM_ARCH_IO_H
 | |
| 
 | |
| #include <linux/bitops.h>
 | |
| 
 | |
| #include <mach/hardware.h>
 | |
| 
 | |
| #define IO_SPACE_LIMIT 0x0000ffff
 | |
| 
 | |
| extern int (*ixp4xx_pci_read)(u32 addr, u32 cmd, u32* data);
 | |
| extern int ixp4xx_pci_write(u32 addr, u32 cmd, u32 data);
 | |
| 
 | |
| 
 | |
| /*
 | |
|  * IXP4xx provides two methods of accessing PCI memory space:
 | |
|  *
 | |
|  * 1) A direct mapped window from 0x48000000 to 0x4bffffff (64MB).
 | |
|  *    To access PCI via this space, we simply ioremap() the BAR
 | |
|  *    into the kernel and we can use the standard read[bwl]/write[bwl]
 | |
|  *    macros. This is the preffered method due to speed but it
 | |
|  *    limits the system to just 64MB of PCI memory. This can be 
 | |
|  *    problamatic if using video cards and other memory-heavy
 | |
|  *    targets.
 | |
|  *
 | |
|  * 2) If > 64MB of memory space is required, the IXP4xx can be configured
 | |
|  *    to use indirect registers to access PCI (as we do below for I/O
 | |
|  *    transactions). This allows for up to 128MB (0x48000000 to 0x4fffffff)
 | |
|  *    of memory on the bus. The disadvantage of this is that every 
 | |
|  *    PCI access requires three local register accesses plus a spinlock,
 | |
|  *    but in some cases the performance hit is acceptable. In addition,
 | |
|  *    you cannot mmap() PCI devices in this case.
 | |
|  *
 | |
|  */
 | |
| #ifndef	CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 
 | |
| #define __mem_pci(a)		(a)
 | |
| 
 | |
| #else
 | |
| 
 | |
| /*
 | |
|  * In the case of using indirect PCI, we simply return the actual PCI
 | |
|  * address and our read/write implementation use that to drive the 
 | |
|  * access registers. If something outside of PCI is ioremap'd, we
 | |
|  * fallback to the default.
 | |
|  */
 | |
| static inline void __iomem *
 | |
| __ixp4xx_ioremap(unsigned long addr, size_t size, unsigned int mtype)
 | |
| {
 | |
| 	if((addr < PCIBIOS_MIN_MEM) || (addr > 0x4fffffff))
 | |
| 		return __arm_ioremap(addr, size, mtype);
 | |
| 
 | |
| 	return (void __iomem *)addr;
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iounmap(void __iomem *addr)
 | |
| {
 | |
| 	if ((__force u32)addr >= VMALLOC_START)
 | |
| 		__iounmap(addr);
 | |
| }
 | |
| 
 | |
| #define __arch_ioremap(a, s, f)		__ixp4xx_ioremap(a, s, f)
 | |
| #define	__arch_iounmap(a)		__ixp4xx_iounmap(a)
 | |
| 
 | |
| #define	writeb(v, p)			__ixp4xx_writeb(v, p)
 | |
| #define	writew(v, p)			__ixp4xx_writew(v, p)
 | |
| #define	writel(v, p)			__ixp4xx_writel(v, p)
 | |
| 
 | |
| #define	writesb(p, v, l)		__ixp4xx_writesb(p, v, l)
 | |
| #define	writesw(p, v, l)		__ixp4xx_writesw(p, v, l)
 | |
| #define	writesl(p, v, l)		__ixp4xx_writesl(p, v, l)
 | |
| 	
 | |
| #define	readb(p)			__ixp4xx_readb(p)
 | |
| #define	readw(p)			__ixp4xx_readw(p)
 | |
| #define	readl(p)			__ixp4xx_readl(p)
 | |
| 	
 | |
| #define	readsb(p, v, l)			__ixp4xx_readsb(p, v, l)
 | |
| #define	readsw(p, v, l)			__ixp4xx_readsw(p, v, l)
 | |
| #define	readsl(p, v, l)			__ixp4xx_readsl(p, v, l)
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_writeb(u8 value, volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (u32)p;
 | |
| 	u32 n, byte_enables, data;
 | |
| 
 | |
| 	if (addr >= VMALLOC_START) {
 | |
| 		__raw_writeb(value, addr);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~BIT(n)) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	data = value << (8*n);
 | |
| 	ixp4xx_pci_write(addr, byte_enables | NP_CMD_MEMWRITE, data);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_writesb(volatile void __iomem *bus_addr, const u8 *vaddr, int count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		writeb(*vaddr++, bus_addr);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_writew(u16 value, volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (u32)p;
 | |
| 	u32 n, byte_enables, data;
 | |
| 
 | |
| 	if (addr >= VMALLOC_START) {
 | |
| 		__raw_writew(value, addr);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~(BIT(n) | BIT(n+1))) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	data = value << (8*n);
 | |
| 	ixp4xx_pci_write(addr, byte_enables | NP_CMD_MEMWRITE, data);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_writesw(volatile void __iomem *bus_addr, const u16 *vaddr, int count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		writew(*vaddr++, bus_addr);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_writel(u32 value, volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (__force u32)p;
 | |
| 	if (addr >= VMALLOC_START) {
 | |
| 		__raw_writel(value, p);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	ixp4xx_pci_write(addr, NP_CMD_MEMWRITE, value);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_writesl(volatile void __iomem *bus_addr, const u32 *vaddr, int count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		writel(*vaddr++, bus_addr);
 | |
| }
 | |
| 
 | |
| static inline unsigned char 
 | |
| __ixp4xx_readb(const volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (u32)p;
 | |
| 	u32 n, byte_enables, data;
 | |
| 
 | |
| 	if (addr >= VMALLOC_START)
 | |
| 		return __raw_readb(addr);
 | |
| 
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~BIT(n)) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	if (ixp4xx_pci_read(addr, byte_enables | NP_CMD_MEMREAD, &data))
 | |
| 		return 0xff;
 | |
| 
 | |
| 	return data >> (8*n);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_readsb(const volatile void __iomem *bus_addr, u8 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = readb(bus_addr);
 | |
| }
 | |
| 
 | |
| static inline unsigned short 
 | |
| __ixp4xx_readw(const volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (u32)p;
 | |
| 	u32 n, byte_enables, data;
 | |
| 
 | |
| 	if (addr >= VMALLOC_START)
 | |
| 		return __raw_readw(addr);
 | |
| 
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~(BIT(n) | BIT(n+1))) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	if (ixp4xx_pci_read(addr, byte_enables | NP_CMD_MEMREAD, &data))
 | |
| 		return 0xffff;
 | |
| 
 | |
| 	return data>>(8*n);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_readsw(const volatile void __iomem *bus_addr, u16 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = readw(bus_addr);
 | |
| }
 | |
| 
 | |
| static inline unsigned long 
 | |
| __ixp4xx_readl(const volatile void __iomem *p)
 | |
| {
 | |
| 	u32 addr = (__force u32)p;
 | |
| 	u32 data;
 | |
| 
 | |
| 	if (addr >= VMALLOC_START)
 | |
| 		return __raw_readl(p);
 | |
| 
 | |
| 	if (ixp4xx_pci_read(addr, NP_CMD_MEMREAD, &data))
 | |
| 		return 0xffffffff;
 | |
| 
 | |
| 	return data;
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_readsl(const volatile void __iomem *bus_addr, u32 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = readl(bus_addr);
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|  * We can use the built-in functions b/c they end up calling writeb/readb
 | |
|  */
 | |
| #define memset_io(c,v,l)		_memset_io((c),(v),(l))
 | |
| #define memcpy_fromio(a,c,l)		_memcpy_fromio((a),(c),(l))
 | |
| #define memcpy_toio(c,a,l)		_memcpy_toio((c),(a),(l))
 | |
| 
 | |
| #endif
 | |
| 
 | |
| #ifndef CONFIG_PCI
 | |
| 
 | |
| #define	__io(v)		__typesafe_io(v)
 | |
| 
 | |
| #else
 | |
| 
 | |
| /*
 | |
|  * IXP4xx does not have a transparent cpu -> PCI I/O translation
 | |
|  * window.  Instead, it has a set of registers that must be tweaked
 | |
|  * with the proper byte lanes, command types, and address for the
 | |
|  * transaction.  This means that we need to override the default
 | |
|  * I/O functions.
 | |
|  */
 | |
| #define	outb(p, v)			__ixp4xx_outb(p, v)
 | |
| #define	outw(p, v)			__ixp4xx_outw(p, v)
 | |
| #define	outl(p, v)			__ixp4xx_outl(p, v)
 | |
| 	
 | |
| #define	outsb(p, v, l)			__ixp4xx_outsb(p, v, l)
 | |
| #define	outsw(p, v, l)			__ixp4xx_outsw(p, v, l)
 | |
| #define	outsl(p, v, l)			__ixp4xx_outsl(p, v, l)
 | |
| 
 | |
| #define	inb(p)				__ixp4xx_inb(p)
 | |
| #define	inw(p)				__ixp4xx_inw(p)
 | |
| #define	inl(p)				__ixp4xx_inl(p)
 | |
| 
 | |
| #define	insb(p, v, l)			__ixp4xx_insb(p, v, l)
 | |
| #define	insw(p, v, l)			__ixp4xx_insw(p, v, l)
 | |
| #define	insl(p, v, l)			__ixp4xx_insl(p, v, l)
 | |
| 
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outb(u8 value, u32 addr)
 | |
| {
 | |
| 	u32 n, byte_enables, data;
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~BIT(n)) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	data = value << (8*n);
 | |
| 	ixp4xx_pci_write(addr, byte_enables | NP_CMD_IOWRITE, data);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outsb(u32 io_addr, const u8 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		outb(*vaddr++, io_addr);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outw(u16 value, u32 addr)
 | |
| {
 | |
| 	u32 n, byte_enables, data;
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~(BIT(n) | BIT(n+1))) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	data = value << (8*n);
 | |
| 	ixp4xx_pci_write(addr, byte_enables | NP_CMD_IOWRITE, data);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outsw(u32 io_addr, const u16 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		outw(cpu_to_le16(*vaddr++), io_addr);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outl(u32 value, u32 addr)
 | |
| {
 | |
| 	ixp4xx_pci_write(addr, NP_CMD_IOWRITE, value);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_outsl(u32 io_addr, const u32 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		outl(*vaddr++, io_addr);
 | |
| }
 | |
| 
 | |
| static inline u8 
 | |
| __ixp4xx_inb(u32 addr)
 | |
| {
 | |
| 	u32 n, byte_enables, data;
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~BIT(n)) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	if (ixp4xx_pci_read(addr, byte_enables | NP_CMD_IOREAD, &data))
 | |
| 		return 0xff;
 | |
| 
 | |
| 	return data >> (8*n);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_insb(u32 io_addr, u8 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = inb(io_addr);
 | |
| }
 | |
| 
 | |
| static inline u16 
 | |
| __ixp4xx_inw(u32 addr)
 | |
| {
 | |
| 	u32 n, byte_enables, data;
 | |
| 	n = addr % 4;
 | |
| 	byte_enables = (0xf & ~(BIT(n) | BIT(n+1))) << IXP4XX_PCI_NP_CBE_BESL;
 | |
| 	if (ixp4xx_pci_read(addr, byte_enables | NP_CMD_IOREAD, &data))
 | |
| 		return 0xffff;
 | |
| 
 | |
| 	return data>>(8*n);
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_insw(u32 io_addr, u16 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = le16_to_cpu(inw(io_addr));
 | |
| }
 | |
| 
 | |
| static inline u32 
 | |
| __ixp4xx_inl(u32 addr)
 | |
| {
 | |
| 	u32 data;
 | |
| 	if (ixp4xx_pci_read(addr, NP_CMD_IOREAD, &data))
 | |
| 		return 0xffffffff;
 | |
| 
 | |
| 	return data;
 | |
| }
 | |
| 
 | |
| static inline void 
 | |
| __ixp4xx_insl(u32 io_addr, u32 *vaddr, u32 count)
 | |
| {
 | |
| 	while (count--)
 | |
| 		*vaddr++ = inl(io_addr);
 | |
| }
 | |
| 
 | |
| #define PIO_OFFSET      0x10000UL
 | |
| #define PIO_MASK        0x0ffffUL
 | |
| 
 | |
| #define	__is_io_address(p)	(((unsigned long)p >= PIO_OFFSET) && \
 | |
| 					((unsigned long)p <= (PIO_MASK + PIO_OFFSET)))
 | |
| static inline unsigned int
 | |
| __ixp4xx_ioread8(const void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		return	(unsigned int)__ixp4xx_inb(port & PIO_MASK);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		return (unsigned int)__raw_readb(port);
 | |
| #else
 | |
| 		return (unsigned int)__ixp4xx_readb(addr);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_ioread8_rep(const void __iomem *addr, void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_insb(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef	CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_readsb(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_readsb(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| __ixp4xx_ioread16(const void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		return	(unsigned int)__ixp4xx_inw(port & PIO_MASK);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		return le16_to_cpu(__raw_readw((u32)port));
 | |
| #else
 | |
| 		return (unsigned int)__ixp4xx_readw(addr);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_ioread16_rep(const void __iomem *addr, void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_insw(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef	CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_readsw(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_readsw(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| __ixp4xx_ioread32(const void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		return	(unsigned int)__ixp4xx_inl(port & PIO_MASK);
 | |
| 	else {
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		return le32_to_cpu((__force __le32)__raw_readl(addr));
 | |
| #else
 | |
| 		return (unsigned int)__ixp4xx_readl(addr);
 | |
| #endif
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_ioread32_rep(const void __iomem *addr, void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_insl(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef	CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_readsl(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_readsl(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite8(u8 value, void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outb(value, port & PIO_MASK);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writeb(value, port);
 | |
| #else
 | |
| 		__ixp4xx_writeb(value, addr);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite8_rep(void __iomem *addr, const void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outsb(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writesb(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_writesb(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite16(u16 value, void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outw(value, port & PIO_MASK);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writew(cpu_to_le16(value), addr);
 | |
| #else
 | |
| 		__ixp4xx_writew(value, addr);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite16_rep(void __iomem *addr, const void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outsw(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writesw(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_writesw(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite32(u32 value, void __iomem *addr)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outl(value, port & PIO_MASK);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writel((u32 __force)cpu_to_le32(value), addr);
 | |
| #else
 | |
| 		__ixp4xx_writel(value, addr);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| __ixp4xx_iowrite32_rep(void __iomem *addr, const void *vaddr, u32 count)
 | |
| {
 | |
| 	unsigned long port = (unsigned long __force)addr;
 | |
| 	if (__is_io_address(port))
 | |
| 		__ixp4xx_outsl(port & PIO_MASK, vaddr, count);
 | |
| 	else
 | |
| #ifndef CONFIG_IXP4XX_INDIRECT_PCI
 | |
| 		__raw_writesl(addr, vaddr, count);
 | |
| #else
 | |
| 		__ixp4xx_writesl(addr, vaddr, count);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| #define	ioread8(p)			__ixp4xx_ioread8(p)
 | |
| #define	ioread16(p)			__ixp4xx_ioread16(p)
 | |
| #define	ioread32(p)			__ixp4xx_ioread32(p)
 | |
| 
 | |
| #define	ioread8_rep(p, v, c)		__ixp4xx_ioread8_rep(p, v, c)
 | |
| #define	ioread16_rep(p, v, c)		__ixp4xx_ioread16_rep(p, v, c)
 | |
| #define	ioread32_rep(p, v, c)		__ixp4xx_ioread32_rep(p, v, c)
 | |
| 
 | |
| #define	iowrite8(v,p)			__ixp4xx_iowrite8(v,p)
 | |
| #define	iowrite16(v,p)			__ixp4xx_iowrite16(v,p)
 | |
| #define	iowrite32(v,p)			__ixp4xx_iowrite32(v,p)
 | |
| 
 | |
| #define	iowrite8_rep(p, v, c)		__ixp4xx_iowrite8_rep(p, v, c)
 | |
| #define	iowrite16_rep(p, v, c)		__ixp4xx_iowrite16_rep(p, v, c)
 | |
| #define	iowrite32_rep(p, v, c)		__ixp4xx_iowrite32_rep(p, v, c)
 | |
| 
 | |
| #define	ioport_map(port, nr)		((void __iomem*)(port + PIO_OFFSET))
 | |
| #define	ioport_unmap(addr)
 | |
| #endif	// !CONFIG_PCI
 | |
| 
 | |
| #endif	//  __ASM_ARM_ARCH_IO_H
 | |
| 
 |