From c13d383c2be23f646e0de054eb8ea7d4bce115f6 Mon Sep 17 00:00:00 2001 From: Felix Fietkau Date: Thu, 15 Feb 2007 21:52:13 +0000 Subject: [PATCH] some ar531x cleanup SVN-Revision: 6302 --- .../files/arch/mips/atheros/ar5312.c | 24 +- .../files/arch/mips/atheros/ar5312.h | 11 + .../files/arch/mips/atheros/ar5315.c | 16 +- .../files/arch/mips/atheros/ar5315.h | 44 +- .../asm-mips/mach-atheros/dma-coherence.h | 41 + .../patches/150-mips_cache_cleanup.patch | 2028 +++++++++++++++++ 6 files changed, 2121 insertions(+), 43 deletions(-) create mode 100644 target/linux/atheros-2.6/files/include/asm-mips/mach-atheros/dma-coherence.h create mode 100644 target/linux/atheros-2.6/patches/150-mips_cache_cleanup.patch diff --git a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.c b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.c index 28b9d3f958..4e1beaae38 100644 --- a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.c +++ b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.c @@ -30,12 +30,6 @@ #include "ar531x.h" -#define AR531X_IRQ_WLAN0_INTRS MIPS_CPU_IRQ_BASE+2 /* C0_CAUSE: 0x0400 */ -#define AR531X_IRQ_ENET0_INTRS MIPS_CPU_IRQ_BASE+3 /* C0_CAUSE: 0x0800 */ -#define AR531X_IRQ_ENET1_INTRS MIPS_CPU_IRQ_BASE+4 /* C0_CAUSE: 0x1000 */ -#define AR531X_IRQ_WLAN1_INTRS MIPS_CPU_IRQ_BASE+5 /* C0_CAUSE: 0x2000 */ -#define AR531X_IRQ_MISC_INTRS MIPS_CPU_IRQ_BASE+6 /* C0_CAUSE: 0x4000 */ - static struct platform_device *ar5312_devs[5]; @@ -49,8 +43,8 @@ static struct resource ar5312_eth0_res[] = { { .name = "eth_irq", .flags = IORESOURCE_IRQ, - .start = AR531X_IRQ_ENET0_INTRS, - .end = AR531X_IRQ_ENET0_INTRS, + .start = AR5312_IRQ_ENET0_INTRS, + .end = AR5312_IRQ_ENET0_INTRS, }, }; @@ -65,8 +59,8 @@ static struct resource ar5312_eth1_res[] = { { .name = "eth_irq", .flags = IORESOURCE_IRQ, - .start = AR531X_IRQ_ENET1_INTRS, - .end = AR531X_IRQ_ENET1_INTRS, + .start = AR5312_IRQ_ENET1_INTRS, + .end = AR5312_IRQ_ENET1_INTRS, }, }; @@ -221,13 +215,13 @@ asmlinkage void ar5312_irq_dispatch(void) int pending = read_c0_status() & read_c0_cause(); if (pending & CAUSEF_IP2) - do_IRQ(AR531X_IRQ_WLAN0_INTRS); + do_IRQ(AR5312_IRQ_WLAN0_INTRS); else if (pending & CAUSEF_IP3) - do_IRQ(AR531X_IRQ_ENET0_INTRS); + do_IRQ(AR5312_IRQ_ENET0_INTRS); else if (pending & CAUSEF_IP4) - do_IRQ(AR531X_IRQ_ENET1_INTRS); + do_IRQ(AR5312_IRQ_ENET1_INTRS); else if (pending & CAUSEF_IP5) - do_IRQ(AR531X_IRQ_WLAN1_INTRS); + do_IRQ(AR5312_IRQ_WLAN1_INTRS); else if (pending & CAUSEF_IP6) { unsigned int ar531x_misc_intrs = sysRegRead(AR531X_ISR) & sysRegRead(AR531X_IMR); @@ -448,7 +442,7 @@ void __init ar5312_misc_intr_init(int irq_base) irq_desc[i].chip = &ar5312_misc_intr_controller; } setup_irq(AR531X_MISC_IRQ_AHB_PROC, &ar5312_ahb_proc_interrupt); - setup_irq(AR531X_IRQ_MISC_INTRS, &cascade); + setup_irq(AR5312_IRQ_MISC_INTRS, &cascade); } diff --git a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.h b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.h index b6e71b8901..62d5797933 100644 --- a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.h +++ b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5312.h @@ -13,6 +13,17 @@ #include +/* + * IRQs + */ + +#define AR5312_IRQ_WLAN0_INTRS MIPS_CPU_IRQ_BASE+2 /* C0_CAUSE: 0x0400 */ +#define AR5312_IRQ_ENET0_INTRS MIPS_CPU_IRQ_BASE+3 /* C0_CAUSE: 0x0800 */ +#define AR5312_IRQ_ENET1_INTRS MIPS_CPU_IRQ_BASE+4 /* C0_CAUSE: 0x1000 */ +#define AR5312_IRQ_WLAN1_INTRS MIPS_CPU_IRQ_BASE+5 /* C0_CAUSE: 0x2000 */ +#define AR5312_IRQ_MISC_INTRS MIPS_CPU_IRQ_BASE+6 /* C0_CAUSE: 0x4000 */ + + /* Address Map */ #define AR531X_WLAN0 0x18000000 #define AR531X_WLAN1 0x18500000 diff --git a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.c b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.c index 78b2835914..b0db7c1d02 100644 --- a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.c +++ b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.c @@ -28,12 +28,6 @@ #include #include "ar531x.h" -#define AR531X_IRQ_MISC_INTRS MIPS_CPU_IRQ_BASE+2 /* C0_CAUSE: 0x0400 */ -#define AR531X_IRQ_WLAN0_INTRS MIPS_CPU_IRQ_BASE+3 /* C0_CAUSE: 0x0800 */ -#define AR531X_IRQ_ENET0_INTRS MIPS_CPU_IRQ_BASE+4 /* C0_CAUSE: 0x1000 */ -#define AR531X_IRQ_LCBUS_PCI MIPS_CPU_IRQ_BASE+5 /* C0_CAUSE: 0x2000 */ -#define AR531X_IRQ_WLAN0_POLL MIPS_CPU_IRQ_BASE+6 /* C0_CAUSE: 0x4000 */ - static struct resource ar5315_eth_res[] = { { .name = "eth_membase", @@ -44,8 +38,8 @@ static struct resource ar5315_eth_res[] = { { .name = "eth_irq", .flags = IORESOURCE_IRQ, - .start = AR531X_IRQ_ENET0_INTRS, - .end = AR531X_IRQ_ENET0_INTRS, + .start = AR5315_IRQ_ENET0_INTRS, + .end = AR5315_IRQ_ENET0_INTRS, }, }; @@ -218,9 +212,9 @@ asmlinkage void ar5315_irq_dispatch(void) int pending = read_c0_status() & read_c0_cause(); if (pending & CAUSEF_IP3) - do_IRQ(AR531X_IRQ_WLAN0_INTRS); + do_IRQ(AR5315_IRQ_WLAN0_INTRS); else if (pending & CAUSEF_IP4) - do_IRQ(AR531X_IRQ_ENET0_INTRS); + do_IRQ(AR5315_IRQ_ENET0_INTRS); else if (pending & CAUSEF_IP2) { unsigned int ar531x_misc_intrs = sysRegRead(AR5315_ISR) & sysRegRead(AR5315_IMR); @@ -504,7 +498,7 @@ void ar5315_misc_intr_init(int irq_base) irq_desc[i].chip = &ar5315_misc_intr_controller; } setup_irq(AR531X_MISC_IRQ_AHB_PROC, &ar5315_ahb_proc_interrupt); - setup_irq(AR531X_IRQ_MISC_INTRS, &cascade); + setup_irq(AR5315_IRQ_MISC_INTRS, &cascade); } void __init ar5315_plat_setup(void) diff --git a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.h b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.h index b523f80405..ef2df8778d 100644 --- a/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.h +++ b/target/linux/atheros-2.6/files/arch/mips/atheros/ar5315.h @@ -12,6 +12,16 @@ #ifndef AR5315_H #define AR5315_H +/* + * IRQs + */ +#define AR5315_IRQ_MISC_INTRS MIPS_CPU_IRQ_BASE+2 /* C0_CAUSE: 0x0400 */ +#define AR5315_IRQ_WLAN0_INTRS MIPS_CPU_IRQ_BASE+3 /* C0_CAUSE: 0x0800 */ +#define AR5315_IRQ_ENET0_INTRS MIPS_CPU_IRQ_BASE+4 /* C0_CAUSE: 0x1000 */ +#define AR5315_IRQ_LCBUS_PCI MIPS_CPU_IRQ_BASE+5 /* C0_CAUSE: 0x2000 */ +#define AR5315_IRQ_WLAN0_POLL MIPS_CPU_IRQ_BASE+6 /* C0_CAUSE: 0x4000 */ + + /* * Address map */ @@ -80,23 +90,23 @@ */ #define AR5315_ENDIAN_CTL (AR5315_DSLBASE + 0x000c) -#define CONFIG_AHB 0x00000001 /* EC - AHB bridge endianess */ -#define CONFIG_WLAN 0x00000002 /* WLAN byteswap */ -#define CONFIG_MPEGTS_RSVD 0x00000004 /* MPEG-TS byteswap */ -#define CONFIG_PCI 0x00000008 /* PCI byteswap */ -#define CONFIG_MEMCTL 0x00000010 /* Memory controller endianess */ -#define CONFIG_LOCAL 0x00000020 /* Local bus byteswap */ -#define CONFIG_ETHERNET 0x00000040 /* Ethernet byteswap */ - -#define CONFIG_MERGE 0x00000200 /* CPU write buffer merge */ -#define CONFIG_CPU 0x00000400 /* CPU big endian */ -#define CONFIG_PCIAHB 0x00000800 -#define CONFIG_PCIAHB_BRIDGE 0x00001000 -#define CONFIG_SPI 0x00008000 /* SPI byteswap */ -#define CONFIG_CPU_DRAM 0x00010000 -#define CONFIG_CPU_PCI 0x00020000 -#define CONFIG_CPU_MMR 0x00040000 -#define CONFIG_BIG 0x00000400 +#define AR5315_CONFIG_AHB 0x00000001 /* EC - AHB bridge endianess */ +#define AR5315_CONFIG_WLAN 0x00000002 /* WLAN byteswap */ +#define AR5315_CONFIG_MPEGTS_RSVD 0x00000004 /* MPEG-TS byteswap */ +#define AR5315_CONFIG_PCI 0x00000008 /* PCI byteswap */ +#define AR5315_CONFIG_MEMCTL 0x00000010 /* Memory controller endianess */ +#define AR5315_CONFIG_LOCAL 0x00000020 /* Local bus byteswap */ +#define AR5315_CONFIG_ETHERNET 0x00000040 /* Ethernet byteswap */ + +#define AR5315_CONFIG_MERGE 0x00000200 /* CPU write buffer merge */ +#define AR5315_CONFIG_CPU 0x00000400 /* CPU big endian */ +#define AR5315_CONFIG_PCIAHB 0x00000800 +#define AR5315_CONFIG_PCIAHB_BRIDGE 0x00001000 +#define AR5315_CONFIG_SPI 0x00008000 /* SPI byteswap */ +#define AR5315_CONFIG_CPU_DRAM 0x00010000 +#define AR5315_CONFIG_CPU_PCI 0x00020000 +#define AR5315_CONFIG_CPU_MMR 0x00040000 +#define AR5315_CONFIG_BIG 0x00000400 /* diff --git a/target/linux/atheros-2.6/files/include/asm-mips/mach-atheros/dma-coherence.h b/target/linux/atheros-2.6/files/include/asm-mips/mach-atheros/dma-coherence.h new file mode 100644 index 0000000000..fe53556881 --- /dev/null +++ b/target/linux/atheros-2.6/files/include/asm-mips/mach-atheros/dma-coherence.h @@ -0,0 +1,41 @@ +/* + * This file is subject to the terms and conditions of the GNU General Public + * License. See the file "COPYING" in the main directory of this archive + * for more details. + * + * Copyright (C) 2006 Ralf Baechle + * Copyright (C) 2007 Felix Fietkau + * + */ +#ifndef __ASM_MACH_GENERIC_DMA_COHERENCE_H +#define __ASM_MACH_GENERIC_DMA_COHERENCE_H + +#define PCI_DMA_OFFSET 0x20000000 + +struct device; + +static dma_addr_t plat_map_dma_mem(struct device *dev, void *addr, size_t size) +{ + return virt_to_phys(addr) + (dev != NULL ? PCI_DMA_OFFSET : 0); +} + +static dma_addr_t plat_map_dma_mem_page(struct device *dev, struct page *page) +{ + return page_to_phys(page) + (dev != NULL ? PCI_DMA_OFFSET : 0); +} + +static unsigned long plat_dma_addr_to_phys(dma_addr_t dma_addr) +{ + return (dma_addr > PCI_DMA_OFFSET ? dma_addr - PCI_DMA_OFFSET : dma_addr); +} + +static void plat_unmap_dma_mem(dma_addr_t dma_addr) +{ +} + +static inline int plat_device_is_coherent(struct device *dev) +{ + return 0; +} + +#endif /* __ASM_MACH_GENERIC_DMA_COHERENCE_H */ diff --git a/target/linux/atheros-2.6/patches/150-mips_cache_cleanup.patch b/target/linux/atheros-2.6/patches/150-mips_cache_cleanup.patch new file mode 100644 index 0000000000..e39cb82473 --- /dev/null +++ b/target/linux/atheros-2.6/patches/150-mips_cache_cleanup.patch @@ -0,0 +1,2028 @@ +Platforms will now have to supply a function dma_device_is_coherent which +returns if a particular device participates in the coherence domain. For +most platforms this function will always return 0 or 1. + +Signed-off-by: Ralf Baechle +Signed-off-by: Felix Fietkau + +diff -urN linux.old/arch/mips/Kconfig linux.dev/arch/mips/Kconfig +--- linux.old/arch/mips/Kconfig 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/Kconfig 2007-02-09 20:26:45.367388152 +0100 +@@ -571,8 +571,6 @@ + select ARC + select ARC32 + select BOOT_ELF32 +- select OWN_DMA +- select DMA_IP32 + select DMA_NONCOHERENT + select HW_HAS_PCI + select R5000_CPU_SCACHE +@@ -835,9 +833,6 @@ + config DMA_NEED_PCI_MAP_STATE + bool + +-config OWN_DMA +- bool +- + config EARLY_PRINTK + bool + +diff -urN linux.old/arch/mips/mm/dma-coherent.c linux.dev/arch/mips/mm/dma-coherent.c +--- linux.old/arch/mips/mm/dma-coherent.c 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/mm/dma-coherent.c 1970-01-01 01:00:00.000000000 +0100 +@@ -1,254 +0,0 @@ +-/* +- * This file is subject to the terms and conditions of the GNU General Public +- * License. See the file "COPYING" in the main directory of this archive +- * for more details. +- * +- * Copyright (C) 2000 Ani Joshi +- * Copyright (C) 2000, 2001 Ralf Baechle +- * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. +- */ +-#include +-#include +-#include +-#include +-#include +- +-#include +-#include +- +-void *dma_alloc_noncoherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- /* ignore region specifiers */ +- gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); +- +- if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) +- gfp |= GFP_DMA; +- ret = (void *) __get_free_pages(gfp, get_order(size)); +- +- if (ret != NULL) { +- memset(ret, 0, size); +- *dma_handle = virt_to_phys(ret); +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_noncoherent); +- +-void *dma_alloc_coherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +- __attribute__((alias("dma_alloc_noncoherent"))); +- +-EXPORT_SYMBOL(dma_alloc_coherent); +- +-void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- unsigned long addr = (unsigned long) vaddr; +- +- free_pages(addr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_noncoherent); +- +-void dma_free_coherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) __attribute__((alias("dma_free_noncoherent"))); +- +-EXPORT_SYMBOL(dma_free_coherent); +- +-dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- return __pa(ptr); +-} +- +-EXPORT_SYMBOL(dma_map_single); +- +-void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_single); +- +-int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- for (i = 0; i < nents; i++, sg++) { +- sg->dma_address = (dma_addr_t)page_to_phys(sg->page) + sg->offset; +- } +- +- return nents; +-} +- +-EXPORT_SYMBOL(dma_map_sg); +- +-dma_addr_t dma_map_page(struct device *dev, struct page *page, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- return page_to_phys(page) + offset; +-} +- +-EXPORT_SYMBOL(dma_map_page); +- +-void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_page); +- +-void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_sg); +- +-void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_cpu); +- +-void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_device); +- +-void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_cpu); +- +-void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_device); +- +-void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_cpu); +- +-void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_device); +- +-int dma_mapping_error(dma_addr_t dma_addr) +-{ +- return 0; +-} +- +-EXPORT_SYMBOL(dma_mapping_error); +- +-int dma_supported(struct device *dev, u64 mask) +-{ +- /* +- * we fall back to GFP_DMA when the mask isn't all 1s, +- * so we can't guarantee allocations that must be +- * within a tighter range than GFP_DMA.. +- */ +- if (mask < 0x00ffffff) +- return 0; +- +- return 1; +-} +- +-EXPORT_SYMBOL(dma_supported); +- +-int dma_is_consistent(dma_addr_t dma_addr) +-{ +- return 1; +-} +- +-EXPORT_SYMBOL(dma_is_consistent); +- +-void dma_cache_sync(void *vaddr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_cache_sync); +- +-/* The DAC routines are a PCIism.. */ +- +-#ifdef CONFIG_PCI +- +-#include +- +-dma64_addr_t pci_dac_page_to_dma(struct pci_dev *pdev, +- struct page *page, unsigned long offset, int direction) +-{ +- return (dma64_addr_t)page_to_phys(page) + offset; +-} +- +-EXPORT_SYMBOL(pci_dac_page_to_dma); +- +-struct page *pci_dac_dma_to_page(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- return mem_map + (dma_addr >> PAGE_SHIFT); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_page); +- +-unsigned long pci_dac_dma_to_offset(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- return dma_addr & ~PAGE_MASK; +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_offset); +- +-void pci_dac_dma_sync_single_for_cpu(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_cpu); +- +-void pci_dac_dma_sync_single_for_device(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_device); +- +-#endif /* CONFIG_PCI */ +diff -urN linux.old/arch/mips/mm/dma-default.c linux.dev/arch/mips/mm/dma-default.c +--- linux.old/arch/mips/mm/dma-default.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/arch/mips/mm/dma-default.c 2007-02-09 20:26:48.671885792 +0100 +@@ -0,0 +1,363 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2000 Ani Joshi ++ * Copyright (C) 2000, 2001, 06 Ralf Baechle ++ * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. ++ */ ++ ++#include ++#include ++#include ++#include ++#include ++ ++#include ++#include ++ ++#include ++ ++/* ++ * Warning on the terminology - Linux calls an uncached area coherent; ++ * MIPS terminology calls memory areas with hardware maintained coherency ++ * coherent. ++ */ ++ ++static inline int cpu_is_noncoherent_r10000(struct device *dev) ++{ ++ return !plat_device_is_coherent(dev) && ++ (current_cpu_data.cputype == CPU_R10000 && ++ current_cpu_data.cputype == CPU_R12000); ++} ++ ++void *dma_alloc_noncoherent(struct device *dev, size_t size, ++ dma_addr_t * dma_handle, gfp_t gfp) ++{ ++ void *ret; ++ ++ /* ignore region specifiers */ ++ gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); ++ ++ if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) ++ gfp |= GFP_DMA; ++ ret = (void *) __get_free_pages(gfp, get_order(size)); ++ ++ if (ret != NULL) { ++ memset(ret, 0, size); ++ *dma_handle = plat_map_dma_mem(dev, ret, size); ++ } ++ ++ return ret; ++} ++ ++EXPORT_SYMBOL(dma_alloc_noncoherent); ++ ++void *dma_alloc_coherent(struct device *dev, size_t size, ++ dma_addr_t * dma_handle, gfp_t gfp) ++{ ++ void *ret; ++ ++ /* ignore region specifiers */ ++ gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); ++ ++ if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) ++ gfp |= GFP_DMA; ++ ret = (void *) __get_free_pages(gfp, get_order(size)); ++ ++ if (ret) { ++ memset(ret, 0, size); ++ *dma_handle = plat_map_dma_mem(dev, ret, size); ++ ++ if (!plat_device_is_coherent(dev)) { ++ dma_cache_wback_inv((unsigned long) ret, size); ++ ret = UNCAC_ADDR(ret); ++ } ++ } ++ ++ return ret; ++} ++ ++EXPORT_SYMBOL(dma_alloc_coherent); ++ ++void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, ++ dma_addr_t dma_handle) ++{ ++ free_pages((unsigned long) vaddr, get_order(size)); ++} ++ ++EXPORT_SYMBOL(dma_free_noncoherent); ++ ++void dma_free_coherent(struct device *dev, size_t size, void *vaddr, ++ dma_addr_t dma_handle) ++{ ++ unsigned long addr = (unsigned long) vaddr; ++ ++ if (!plat_device_is_coherent(dev)) ++ addr = CAC_ADDR(addr); ++ ++ free_pages(addr, get_order(size)); ++} ++ ++EXPORT_SYMBOL(dma_free_coherent); ++ ++static inline void __dma_sync(unsigned long addr, size_t size, ++ enum dma_data_direction direction) ++{ ++ switch (direction) { ++ case DMA_TO_DEVICE: ++ dma_cache_wback(addr, size); ++ break; ++ ++ case DMA_FROM_DEVICE: ++ dma_cache_inv(addr, size); ++ break; ++ ++ case DMA_BIDIRECTIONAL: ++ dma_cache_wback_inv(addr, size); ++ break; ++ ++ default: ++ BUG(); ++ } ++} ++ ++dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, ++ enum dma_data_direction direction) ++{ ++ unsigned long addr = (unsigned long) ptr; ++ ++ if (!plat_device_is_coherent(dev)) ++ __dma_sync(addr, size, direction); ++ ++ return plat_map_dma_mem(dev, ptr, size); ++} ++ ++EXPORT_SYMBOL(dma_map_single); ++ ++void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, ++ enum dma_data_direction direction) ++{ ++ if (cpu_is_noncoherent_r10000(dev)) ++ __dma_sync(plat_dma_addr_to_phys(dma_addr) + PAGE_OFFSET, size, ++ direction); ++ ++ plat_unmap_dma_mem(dma_addr); ++} ++ ++EXPORT_SYMBOL(dma_unmap_single); ++ ++int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, ++ enum dma_data_direction direction) ++{ ++ int i; ++ ++ BUG_ON(direction == DMA_NONE); ++ ++ for (i = 0; i < nents; i++, sg++) { ++ unsigned long addr; ++ ++ addr = (unsigned long) page_address(sg->page); ++ if (!plat_device_is_coherent(dev) && addr) ++ __dma_sync(addr + sg->offset, sg->length, direction); ++ sg->dma_address = plat_map_dma_mem_page(dev, sg->page) + ++ sg->offset; ++ } ++ ++ return nents; ++} ++ ++EXPORT_SYMBOL(dma_map_sg); ++ ++dma_addr_t dma_map_page(struct device *dev, struct page *page, ++ unsigned long offset, size_t size, enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (!plat_device_is_coherent(dev)) { ++ unsigned long addr; ++ ++ addr = (unsigned long) page_address(page) + offset; ++ dma_cache_wback_inv(addr, size); ++ } ++ ++ return plat_map_dma_mem_page(dev, page) + offset; ++} ++ ++EXPORT_SYMBOL(dma_map_page); ++ ++void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, ++ enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (!plat_device_is_coherent(dev) && direction != DMA_TO_DEVICE) { ++ unsigned long addr; ++ ++ addr = plat_dma_addr_to_phys(dma_address); ++ dma_cache_wback_inv(addr, size); ++ } ++ ++ plat_unmap_dma_mem(dma_address); ++} ++ ++EXPORT_SYMBOL(dma_unmap_page); ++ ++void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, ++ enum dma_data_direction direction) ++{ ++ unsigned long addr; ++ int i; ++ ++ BUG_ON(direction == DMA_NONE); ++ ++ for (i = 0; i < nhwentries; i++, sg++) { ++ if (!plat_device_is_coherent(dev) && ++ direction != DMA_TO_DEVICE) { ++ addr = (unsigned long) page_address(sg->page); ++ if (addr) ++ __dma_sync(addr + sg->offset, sg->length, ++ direction); ++ } ++ plat_unmap_dma_mem(sg->dma_address); ++ } ++} ++ ++EXPORT_SYMBOL(dma_unmap_sg); ++ ++void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, ++ size_t size, enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (cpu_is_noncoherent_r10000(dev)) { ++ unsigned long addr; ++ ++ addr = PAGE_OFFSET + plat_dma_addr_to_phys(dma_handle); ++ __dma_sync(addr, size, direction); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_single_for_cpu); ++ ++void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, ++ size_t size, enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (cpu_is_noncoherent_r10000(dev)) { ++ unsigned long addr; ++ ++ addr = plat_dma_addr_to_phys(dma_handle); ++ __dma_sync(addr, size, direction); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_single_for_device); ++ ++void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, ++ unsigned long offset, size_t size, enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (cpu_is_noncoherent_r10000(dev)) { ++ unsigned long addr; ++ ++ addr = PAGE_OFFSET + plat_dma_addr_to_phys(dma_handle); ++ __dma_sync(addr + offset, size, direction); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_single_range_for_cpu); ++ ++void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, ++ unsigned long offset, size_t size, enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (cpu_is_noncoherent_r10000(dev)) { ++ unsigned long addr; ++ ++ addr = PAGE_OFFSET + plat_dma_addr_to_phys(dma_handle); ++ __dma_sync(addr + offset, size, direction); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_single_range_for_device); ++ ++void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, ++ enum dma_data_direction direction) ++{ ++ int i; ++ ++ BUG_ON(direction == DMA_NONE); ++ ++ /* Make sure that gcc doesn't leave the empty loop body. */ ++ for (i = 0; i < nelems; i++, sg++) { ++ if (!plat_device_is_coherent(dev)) ++ __dma_sync((unsigned long)page_address(sg->page), ++ sg->length, direction); ++ plat_unmap_dma_mem(sg->dma_address); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_sg_for_cpu); ++ ++void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, ++ enum dma_data_direction direction) ++{ ++ int i; ++ ++ BUG_ON(direction == DMA_NONE); ++ ++ /* Make sure that gcc doesn't leave the empty loop body. */ ++ for (i = 0; i < nelems; i++, sg++) { ++ if (!plat_device_is_coherent(dev)) ++ __dma_sync((unsigned long)page_address(sg->page), ++ sg->length, direction); ++ plat_unmap_dma_mem(sg->dma_address); ++ } ++} ++ ++EXPORT_SYMBOL(dma_sync_sg_for_device); ++ ++int dma_mapping_error(dma_addr_t dma_addr) ++{ ++ return 0; ++} ++ ++EXPORT_SYMBOL(dma_mapping_error); ++ ++int dma_supported(struct device *dev, u64 mask) ++{ ++ /* ++ * we fall back to GFP_DMA when the mask isn't all 1s, ++ * so we can't guarantee allocations that must be ++ * within a tighter range than GFP_DMA.. ++ */ ++ if (mask < 0x00ffffff) ++ return 0; ++ ++ return 1; ++} ++ ++EXPORT_SYMBOL(dma_supported); ++ ++int dma_is_consistent(dma_addr_t dma_addr) ++{ ++ return plat_device_is_coherent(NULL); ++} ++ ++EXPORT_SYMBOL(dma_is_consistent); ++ ++void dma_cache_sync(void *vaddr, size_t size, ++ enum dma_data_direction direction) ++{ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (!plat_device_is_coherent(NULL)) ++ dma_cache_wback_inv((unsigned long)vaddr, size); ++} ++ ++EXPORT_SYMBOL(dma_cache_sync); +diff -urN linux.old/arch/mips/mm/dma-ip27.c linux.dev/arch/mips/mm/dma-ip27.c +--- linux.old/arch/mips/mm/dma-ip27.c 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/mm/dma-ip27.c 1970-01-01 01:00:00.000000000 +0100 +@@ -1,257 +0,0 @@ +-/* +- * This file is subject to the terms and conditions of the GNU General Public +- * License. See the file "COPYING" in the main directory of this archive +- * for more details. +- * +- * Copyright (C) 2000 Ani Joshi +- * Copyright (C) 2000, 2001 Ralf Baechle +- * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. +- */ +-#include +-#include +-#include +-#include +-#include +- +-#include +-#include +- +-#define pdev_to_baddr(pdev, addr) \ +- (BRIDGE_CONTROLLER(pdev->bus)->baddr + (addr)) +-#define dev_to_baddr(dev, addr) \ +- pdev_to_baddr(to_pci_dev(dev), (addr)) +- +-void *dma_alloc_noncoherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- +- /* ignore region specifiers */ +- gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); +- +- if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) +- gfp |= GFP_DMA; +- ret = (void *) __get_free_pages(gfp, get_order(size)); +- +- if (ret != NULL) { +- memset(ret, 0, size); +- *dma_handle = dev_to_baddr(dev, virt_to_phys(ret)); +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_noncoherent); +- +-void *dma_alloc_coherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +- __attribute__((alias("dma_alloc_noncoherent"))); +- +-EXPORT_SYMBOL(dma_alloc_coherent); +- +-void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- unsigned long addr = (unsigned long) vaddr; +- +- free_pages(addr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_noncoherent); +- +-void dma_free_coherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) __attribute__((alias("dma_free_noncoherent"))); +- +-EXPORT_SYMBOL(dma_free_coherent); +- +-dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- return dev_to_baddr(dev, __pa(ptr)); +-} +- +-EXPORT_SYMBOL(dma_map_single); +- +-void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_single); +- +-int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- for (i = 0; i < nents; i++, sg++) { +- sg->dma_address = (dma_addr_t) dev_to_baddr(dev, +- page_to_phys(sg->page) + sg->offset); +- } +- +- return nents; +-} +- +-EXPORT_SYMBOL(dma_map_sg); +- +-dma_addr_t dma_map_page(struct device *dev, struct page *page, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- return dev_to_baddr(dev, page_to_phys(page) + offset); +-} +- +-EXPORT_SYMBOL(dma_map_page); +- +-void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_page); +- +-void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_unmap_sg); +- +-void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_cpu); +- +-void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_device); +- +-void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_cpu); +- +-void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_device); +- +-void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_cpu); +- +-void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_device); +- +-int dma_mapping_error(dma_addr_t dma_addr) +-{ +- return 0; +-} +- +-EXPORT_SYMBOL(dma_mapping_error); +- +-int dma_supported(struct device *dev, u64 mask) +-{ +- /* +- * we fall back to GFP_DMA when the mask isn't all 1s, +- * so we can't guarantee allocations that must be +- * within a tighter range than GFP_DMA.. +- */ +- if (mask < 0x00ffffff) +- return 0; +- +- return 1; +-} +- +-EXPORT_SYMBOL(dma_supported); +- +-int dma_is_consistent(dma_addr_t dma_addr) +-{ +- return 1; +-} +- +-EXPORT_SYMBOL(dma_is_consistent); +- +-void dma_cache_sync(void *vaddr, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +-} +- +-EXPORT_SYMBOL(dma_cache_sync); +- +-dma64_addr_t pci_dac_page_to_dma(struct pci_dev *pdev, +- struct page *page, unsigned long offset, int direction) +-{ +- dma64_addr_t addr = page_to_phys(page) + offset; +- +- return (dma64_addr_t) pdev_to_baddr(pdev, addr); +-} +- +-EXPORT_SYMBOL(pci_dac_page_to_dma); +- +-struct page *pci_dac_dma_to_page(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- struct bridge_controller *bc = BRIDGE_CONTROLLER(pdev->bus); +- +- return pfn_to_page((dma_addr - bc->baddr) >> PAGE_SHIFT); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_page); +- +-unsigned long pci_dac_dma_to_offset(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- return dma_addr & ~PAGE_MASK; +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_offset); +- +-void pci_dac_dma_sync_single_for_cpu(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_cpu); +- +-void pci_dac_dma_sync_single_for_device(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_device); +diff -urN linux.old/arch/mips/mm/dma-ip32.c linux.dev/arch/mips/mm/dma-ip32.c +--- linux.old/arch/mips/mm/dma-ip32.c 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/mm/dma-ip32.c 1970-01-01 01:00:00.000000000 +0100 +@@ -1,382 +0,0 @@ +-/* +- * This file is subject to the terms and conditions of the GNU General Public +- * License. See the file "COPYING" in the main directory of this archive +- * for more details. +- * +- * Copyright (C) 2000 Ani Joshi +- * Copyright (C) 2000, 2001 Ralf Baechle +- * Copyright (C) 2005 Ilya A. Volynets-Evenbakh +- * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. +- * IP32 changes by Ilya. +- */ +-#include +-#include +-#include +-#include +-#include +- +-#include +-#include +-#include +- +-/* +- * Warning on the terminology - Linux calls an uncached area coherent; +- * MIPS terminology calls memory areas with hardware maintained coherency +- * coherent. +- */ +- +-/* +- * Few notes. +- * 1. CPU sees memory as two chunks: 0-256M@0x0, and the rest @0x40000000+256M +- * 2. PCI sees memory as one big chunk @0x0 (or we could use 0x40000000 for native-endian) +- * 3. All other devices see memory as one big chunk at 0x40000000 +- * 4. Non-PCI devices will pass NULL as struct device* +- * Thus we translate differently, depending on device. +- */ +- +-#define RAM_OFFSET_MASK 0x3fffffff +- +-void *dma_alloc_noncoherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- /* ignore region specifiers */ +- gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); +- +- if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) +- gfp |= GFP_DMA; +- ret = (void *) __get_free_pages(gfp, get_order(size)); +- +- if (ret != NULL) { +- unsigned long addr = virt_to_phys(ret)&RAM_OFFSET_MASK; +- memset(ret, 0, size); +- if(dev==NULL) +- addr+= CRIME_HI_MEM_BASE; +- *dma_handle = addr; +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_noncoherent); +- +-void *dma_alloc_coherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- +- ret = dma_alloc_noncoherent(dev, size, dma_handle, gfp); +- if (ret) { +- dma_cache_wback_inv((unsigned long) ret, size); +- ret = UNCAC_ADDR(ret); +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_coherent); +- +-void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- free_pages((unsigned long) vaddr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_noncoherent); +- +-void dma_free_coherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- unsigned long addr = (unsigned long) vaddr; +- +- addr = CAC_ADDR(addr); +- free_pages(addr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_coherent); +- +-static inline void __dma_sync(unsigned long addr, size_t size, +- enum dma_data_direction direction) +-{ +- switch (direction) { +- case DMA_TO_DEVICE: +- dma_cache_wback(addr, size); +- break; +- +- case DMA_FROM_DEVICE: +- dma_cache_inv(addr, size); +- break; +- +- case DMA_BIDIRECTIONAL: +- dma_cache_wback_inv(addr, size); +- break; +- +- default: +- BUG(); +- } +-} +- +-dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, +- enum dma_data_direction direction) +-{ +- unsigned long addr = (unsigned long) ptr; +- +- switch (direction) { +- case DMA_TO_DEVICE: +- dma_cache_wback(addr, size); +- break; +- +- case DMA_FROM_DEVICE: +- dma_cache_inv(addr, size); +- break; +- +- case DMA_BIDIRECTIONAL: +- dma_cache_wback_inv(addr, size); +- break; +- +- default: +- BUG(); +- } +- +- addr = virt_to_phys(ptr)&RAM_OFFSET_MASK; +- if(dev == NULL) +- addr+=CRIME_HI_MEM_BASE; +- return (dma_addr_t)addr; +-} +- +-EXPORT_SYMBOL(dma_map_single); +- +-void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, +- enum dma_data_direction direction) +-{ +- switch (direction) { +- case DMA_TO_DEVICE: +- break; +- +- case DMA_FROM_DEVICE: +- break; +- +- case DMA_BIDIRECTIONAL: +- break; +- +- default: +- BUG(); +- } +-} +- +-EXPORT_SYMBOL(dma_unmap_single); +- +-int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- for (i = 0; i < nents; i++, sg++) { +- unsigned long addr; +- +- addr = (unsigned long) page_address(sg->page)+sg->offset; +- if (addr) +- __dma_sync(addr, sg->length, direction); +- addr = __pa(addr)&RAM_OFFSET_MASK; +- if(dev == NULL) +- addr += CRIME_HI_MEM_BASE; +- sg->dma_address = (dma_addr_t)addr; +- } +- +- return nents; +-} +- +-EXPORT_SYMBOL(dma_map_sg); +- +-dma_addr_t dma_map_page(struct device *dev, struct page *page, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = (unsigned long) page_address(page) + offset; +- dma_cache_wback_inv(addr, size); +- addr = __pa(addr)&RAM_OFFSET_MASK; +- if(dev == NULL) +- addr += CRIME_HI_MEM_BASE; +- +- return (dma_addr_t)addr; +-} +- +-EXPORT_SYMBOL(dma_map_page); +- +-void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- if (direction != DMA_TO_DEVICE) { +- unsigned long addr; +- +- dma_address&=RAM_OFFSET_MASK; +- addr = dma_address + PAGE_OFFSET; +- if(dma_address>=256*1024*1024) +- addr+=CRIME_HI_MEM_BASE; +- dma_cache_wback_inv(addr, size); +- } +-} +- +-EXPORT_SYMBOL(dma_unmap_page); +- +-void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, +- enum dma_data_direction direction) +-{ +- unsigned long addr; +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- if (direction == DMA_TO_DEVICE) +- return; +- +- for (i = 0; i < nhwentries; i++, sg++) { +- addr = (unsigned long) page_address(sg->page); +- if (!addr) +- continue; +- dma_cache_wback_inv(addr + sg->offset, sg->length); +- } +-} +- +-EXPORT_SYMBOL(dma_unmap_sg); +- +-void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- dma_handle&=RAM_OFFSET_MASK; +- addr = dma_handle + PAGE_OFFSET; +- if(dma_handle>=256*1024*1024) +- addr+=CRIME_HI_MEM_BASE; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_cpu); +- +-void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- dma_handle&=RAM_OFFSET_MASK; +- addr = dma_handle + PAGE_OFFSET; +- if(dma_handle>=256*1024*1024) +- addr+=CRIME_HI_MEM_BASE; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_device); +- +-void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- dma_handle&=RAM_OFFSET_MASK; +- addr = dma_handle + offset + PAGE_OFFSET; +- if(dma_handle>=256*1024*1024) +- addr+=CRIME_HI_MEM_BASE; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_cpu); +- +-void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- dma_handle&=RAM_OFFSET_MASK; +- addr = dma_handle + offset + PAGE_OFFSET; +- if(dma_handle>=256*1024*1024) +- addr+=CRIME_HI_MEM_BASE; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_device); +- +-void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- /* Make sure that gcc doesn't leave the empty loop body. */ +- for (i = 0; i < nelems; i++, sg++) +- __dma_sync((unsigned long)page_address(sg->page), +- sg->length, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_cpu); +- +-void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- /* Make sure that gcc doesn't leave the empty loop body. */ +- for (i = 0; i < nelems; i++, sg++) +- __dma_sync((unsigned long)page_address(sg->page), +- sg->length, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_device); +- +-int dma_mapping_error(dma_addr_t dma_addr) +-{ +- return 0; +-} +- +-EXPORT_SYMBOL(dma_mapping_error); +- +-int dma_supported(struct device *dev, u64 mask) +-{ +- /* +- * we fall back to GFP_DMA when the mask isn't all 1s, +- * so we can't guarantee allocations that must be +- * within a tighter range than GFP_DMA.. +- */ +- if (mask < 0x00ffffff) +- return 0; +- +- return 1; +-} +- +-EXPORT_SYMBOL(dma_supported); +- +-int dma_is_consistent(dma_addr_t dma_addr) +-{ +- return 1; +-} +- +-EXPORT_SYMBOL(dma_is_consistent); +- +-void dma_cache_sync(void *vaddr, size_t size, enum dma_data_direction direction) +-{ +- if (direction == DMA_NONE) +- return; +- +- dma_cache_wback_inv((unsigned long)vaddr, size); +-} +- +-EXPORT_SYMBOL(dma_cache_sync); +- +diff -urN linux.old/arch/mips/mm/dma-noncoherent.c linux.dev/arch/mips/mm/dma-noncoherent.c +--- linux.old/arch/mips/mm/dma-noncoherent.c 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/mm/dma-noncoherent.c 1970-01-01 01:00:00.000000000 +0100 +@@ -1,369 +0,0 @@ +-/* +- * This file is subject to the terms and conditions of the GNU General Public +- * License. See the file "COPYING" in the main directory of this archive +- * for more details. +- * +- * Copyright (C) 2000 Ani Joshi +- * Copyright (C) 2000, 2001 Ralf Baechle +- * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. +- */ +-#include +-#include +-#include +-#include +-#include +- +-#include +-#include +- +-/* +- * Warning on the terminology - Linux calls an uncached area coherent; +- * MIPS terminology calls memory areas with hardware maintained coherency +- * coherent. +- */ +- +-void *dma_alloc_noncoherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- /* ignore region specifiers */ +- gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); +- +- if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff)) +- gfp |= GFP_DMA; +- ret = (void *) __get_free_pages(gfp, get_order(size)); +- +- if (ret != NULL) { +- memset(ret, 0, size); +- *dma_handle = virt_to_phys(ret); +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_noncoherent); +- +-void *dma_alloc_coherent(struct device *dev, size_t size, +- dma_addr_t * dma_handle, gfp_t gfp) +-{ +- void *ret; +- +- ret = dma_alloc_noncoherent(dev, size, dma_handle, gfp); +- if (ret) { +- dma_cache_wback_inv((unsigned long) ret, size); +- ret = UNCAC_ADDR(ret); +- } +- +- return ret; +-} +- +-EXPORT_SYMBOL(dma_alloc_coherent); +- +-void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- free_pages((unsigned long) vaddr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_noncoherent); +- +-void dma_free_coherent(struct device *dev, size_t size, void *vaddr, +- dma_addr_t dma_handle) +-{ +- unsigned long addr = (unsigned long) vaddr; +- +- addr = CAC_ADDR(addr); +- free_pages(addr, get_order(size)); +-} +- +-EXPORT_SYMBOL(dma_free_coherent); +- +-static inline void __dma_sync(unsigned long addr, size_t size, +- enum dma_data_direction direction) +-{ +- switch (direction) { +- case DMA_TO_DEVICE: +- dma_cache_wback(addr, size); +- break; +- +- case DMA_FROM_DEVICE: +- dma_cache_inv(addr, size); +- break; +- +- case DMA_BIDIRECTIONAL: +- dma_cache_wback_inv(addr, size); +- break; +- +- default: +- BUG(); +- } +-} +- +-dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, +- enum dma_data_direction direction) +-{ +- unsigned long addr = (unsigned long) ptr; +- +- __dma_sync(addr, size, direction); +- +- return virt_to_phys(ptr); +-} +- +-EXPORT_SYMBOL(dma_map_single); +- +-void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, +- enum dma_data_direction direction) +-{ +- unsigned long addr; +- addr = dma_addr + PAGE_OFFSET; +- +- //__dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_unmap_single); +- +-int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- for (i = 0; i < nents; i++, sg++) { +- unsigned long addr; +- +- addr = (unsigned long) page_address(sg->page); +- if (addr) { +- __dma_sync(addr + sg->offset, sg->length, direction); +- sg->dma_address = (dma_addr_t)page_to_phys(sg->page) +- + sg->offset; +- } +- } +- +- return nents; +-} +- +-EXPORT_SYMBOL(dma_map_sg); +- +-dma_addr_t dma_map_page(struct device *dev, struct page *page, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = (unsigned long) page_address(page) + offset; +- dma_cache_wback_inv(addr, size); +- +- return page_to_phys(page) + offset; +-} +- +-EXPORT_SYMBOL(dma_map_page); +- +-void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, +- enum dma_data_direction direction) +-{ +- BUG_ON(direction == DMA_NONE); +- +- if (direction != DMA_TO_DEVICE) { +- unsigned long addr; +- +- addr = dma_address + PAGE_OFFSET; +- dma_cache_wback_inv(addr, size); +- } +-} +- +-EXPORT_SYMBOL(dma_unmap_page); +- +-void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, +- enum dma_data_direction direction) +-{ +- unsigned long addr; +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- if (direction == DMA_TO_DEVICE) +- return; +- +- for (i = 0; i < nhwentries; i++, sg++) { +- addr = (unsigned long) page_address(sg->page); +- if (addr) +- __dma_sync(addr + sg->offset, sg->length, direction); +- } +-} +- +-EXPORT_SYMBOL(dma_unmap_sg); +- +-void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = dma_handle + PAGE_OFFSET; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_cpu); +- +-void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, +- size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = dma_handle + PAGE_OFFSET; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_for_device); +- +-void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = dma_handle + offset + PAGE_OFFSET; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_cpu); +- +-void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, +- unsigned long offset, size_t size, enum dma_data_direction direction) +-{ +- unsigned long addr; +- +- BUG_ON(direction == DMA_NONE); +- +- addr = dma_handle + offset + PAGE_OFFSET; +- __dma_sync(addr, size, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_single_range_for_device); +- +-void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- /* Make sure that gcc doesn't leave the empty loop body. */ +- for (i = 0; i < nelems; i++, sg++) +- __dma_sync((unsigned long)page_address(sg->page), +- sg->length, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_cpu); +- +-void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, +- enum dma_data_direction direction) +-{ +- int i; +- +- BUG_ON(direction == DMA_NONE); +- +- /* Make sure that gcc doesn't leave the empty loop body. */ +- for (i = 0; i < nelems; i++, sg++) +- __dma_sync((unsigned long)page_address(sg->page), +- sg->length, direction); +-} +- +-EXPORT_SYMBOL(dma_sync_sg_for_device); +- +-int dma_mapping_error(dma_addr_t dma_addr) +-{ +- return 0; +-} +- +-EXPORT_SYMBOL(dma_mapping_error); +- +-int dma_supported(struct device *dev, u64 mask) +-{ +- /* +- * we fall back to GFP_DMA when the mask isn't all 1s, +- * so we can't guarantee allocations that must be +- * within a tighter range than GFP_DMA.. +- */ +- if (mask < 0x00ffffff) +- return 0; +- +- return 1; +-} +- +-EXPORT_SYMBOL(dma_supported); +- +-int dma_is_consistent(dma_addr_t dma_addr) +-{ +- return 1; +-} +- +-EXPORT_SYMBOL(dma_is_consistent); +- +-void dma_cache_sync(void *vaddr, size_t size, enum dma_data_direction direction) +-{ +- if (direction == DMA_NONE) +- return; +- +- dma_cache_wback_inv((unsigned long)vaddr, size); +-} +- +-EXPORT_SYMBOL(dma_cache_sync); +- +-/* The DAC routines are a PCIism.. */ +- +-#ifdef CONFIG_PCI +- +-#include +- +-dma64_addr_t pci_dac_page_to_dma(struct pci_dev *pdev, +- struct page *page, unsigned long offset, int direction) +-{ +- return (dma64_addr_t)page_to_phys(page) + offset; +-} +- +-EXPORT_SYMBOL(pci_dac_page_to_dma); +- +-struct page *pci_dac_dma_to_page(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- return mem_map + (dma_addr >> PAGE_SHIFT); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_page); +- +-unsigned long pci_dac_dma_to_offset(struct pci_dev *pdev, +- dma64_addr_t dma_addr) +-{ +- return dma_addr & ~PAGE_MASK; +-} +- +-EXPORT_SYMBOL(pci_dac_dma_to_offset); +- +-void pci_dac_dma_sync_single_for_cpu(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +- +- dma_cache_wback_inv(dma_addr + PAGE_OFFSET, len); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_cpu); +- +-void pci_dac_dma_sync_single_for_device(struct pci_dev *pdev, +- dma64_addr_t dma_addr, size_t len, int direction) +-{ +- BUG_ON(direction == PCI_DMA_NONE); +- +- dma_cache_wback_inv(dma_addr + PAGE_OFFSET, len); +-} +- +-EXPORT_SYMBOL(pci_dac_dma_sync_single_for_device); +- +-#endif /* CONFIG_PCI */ +diff -urN linux.old/arch/mips/mm/Makefile linux.dev/arch/mips/mm/Makefile +--- linux.old/arch/mips/mm/Makefile 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/mm/Makefile 2007-02-09 20:26:45.376386784 +0100 +@@ -2,8 +2,8 @@ + # Makefile for the Linux/MIPS-specific parts of the memory manager. + # + +-obj-y += cache.o extable.o fault.o init.o pgtable.o \ +- tlbex.o tlbex-fault.o ++obj-y += cache.o dma-default.o extable.o fault.o \ ++ init.o pgtable.o tlbex.o tlbex-fault.o + + obj-$(CONFIG_32BIT) += ioremap.o pgtable-32.o + obj-$(CONFIG_64BIT) += pgtable-64.o +@@ -32,14 +32,4 @@ + obj-$(CONFIG_RM7000_CPU_SCACHE) += sc-rm7k.o + obj-$(CONFIG_MIPS_CPU_SCACHE) += sc-mips.o + +-# +-# Choose one DMA coherency model +-# +-ifndef CONFIG_OWN_DMA +-obj-$(CONFIG_DMA_COHERENT) += dma-coherent.o +-obj-$(CONFIG_DMA_NONCOHERENT) += dma-noncoherent.o +-endif +-obj-$(CONFIG_DMA_IP27) += dma-ip27.o +-obj-$(CONFIG_DMA_IP32) += dma-ip32.o +- + EXTRA_AFLAGS := $(CFLAGS) +diff -urN linux.old/arch/mips/pci/Makefile linux.dev/arch/mips/pci/Makefile +--- linux.old/arch/mips/pci/Makefile 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/arch/mips/pci/Makefile 2007-02-09 20:26:50.961537712 +0100 +@@ -2,7 +2,7 @@ + # Makefile for the PCI specific kernel interface routines under Linux. + # + +-obj-y += pci.o ++obj-y += pci.o pci-dac.o + + # + # PCI bus host bridge specific code +diff -urN linux.old/arch/mips/pci/pci-dac.c linux.dev/arch/mips/pci/pci-dac.c +--- linux.old/arch/mips/pci/pci-dac.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/arch/mips/pci/pci-dac.c 2007-02-09 20:26:50.961537712 +0100 +@@ -0,0 +1,79 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2000 Ani Joshi ++ * Copyright (C) 2000, 2001, 06 Ralf Baechle ++ * swiped from i386, and cloned for MIPS by Geert, polished by Ralf. ++ */ ++ ++#include ++#include ++#include ++#include ++#include ++ ++#include ++#include ++ ++#include ++ ++#include ++ ++dma64_addr_t pci_dac_page_to_dma(struct pci_dev *pdev, ++ struct page *page, unsigned long offset, int direction) ++{ ++ struct device *dev = &pdev->dev; ++ ++ BUG_ON(direction == DMA_NONE); ++ ++ if (!plat_device_is_coherent(dev)) { ++ unsigned long addr; ++ ++ addr = (unsigned long) page_address(page) + offset; ++ dma_cache_wback_inv(addr, PAGE_SIZE); ++ } ++ ++ return plat_map_dma_mem_page(dev, page) + offset; ++} ++ ++EXPORT_SYMBOL(pci_dac_page_to_dma); ++ ++struct page *pci_dac_dma_to_page(struct pci_dev *pdev, ++ dma64_addr_t dma_addr) ++{ ++ return pfn_to_page(plat_dma_addr_to_phys(dma_addr) >> PAGE_SHIFT); ++} ++ ++EXPORT_SYMBOL(pci_dac_dma_to_page); ++ ++unsigned long pci_dac_dma_to_offset(struct pci_dev *pdev, ++ dma64_addr_t dma_addr) ++{ ++ return dma_addr & ~PAGE_MASK; ++} ++ ++EXPORT_SYMBOL(pci_dac_dma_to_offset); ++ ++void pci_dac_dma_sync_single_for_cpu(struct pci_dev *pdev, ++ dma64_addr_t dma_addr, size_t len, int direction) ++{ ++ BUG_ON(direction == PCI_DMA_NONE); ++ ++ if (!plat_device_is_coherent(&pdev->dev)) ++ dma_cache_wback_inv(dma_addr + PAGE_OFFSET, len); ++} ++ ++EXPORT_SYMBOL(pci_dac_dma_sync_single_for_cpu); ++ ++void pci_dac_dma_sync_single_for_device(struct pci_dev *pdev, ++ dma64_addr_t dma_addr, size_t len, int direction) ++{ ++ BUG_ON(direction == PCI_DMA_NONE); ++ ++ if (!plat_device_is_coherent(&pdev->dev)) ++ dma_cache_wback_inv(dma_addr + PAGE_OFFSET, len); ++} ++ ++EXPORT_SYMBOL(pci_dac_dma_sync_single_for_device); +diff -urN linux.old/include/asm-mips/mach-generic/dma-coherence.h linux.dev/include/asm-mips/mach-generic/dma-coherence.h +--- linux.old/include/asm-mips/mach-generic/dma-coherence.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/include/asm-mips/mach-generic/dma-coherence.h 2007-02-09 20:26:50.962537560 +0100 +@@ -0,0 +1,43 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2006 Ralf Baechle ++ * ++ */ ++#ifndef __ASM_MACH_GENERIC_DMA_COHERENCE_H ++#define __ASM_MACH_GENERIC_DMA_COHERENCE_H ++ ++struct device; ++ ++static dma_addr_t plat_map_dma_mem(struct device *dev, void *addr, size_t size) ++{ ++ return virt_to_phys(addr); ++} ++ ++static dma_addr_t plat_map_dma_mem_page(struct device *dev, struct page *page) ++{ ++ return page_to_phys(page); ++} ++ ++static unsigned long plat_dma_addr_to_phys(dma_addr_t dma_addr) ++{ ++ return dma_addr; ++} ++ ++static void plat_unmap_dma_mem(dma_addr_t dma_addr) ++{ ++} ++ ++static inline int plat_device_is_coherent(struct device *dev) ++{ ++#ifdef CONFIG_DMA_COHERENT ++ return 1; ++#endif ++#ifdef CONFIG_DMA_NONCOHERENT ++ return 0; ++#endif ++} ++ ++#endif /* __ASM_MACH_GENERIC_DMA_COHERENCE_H */ +diff -urN linux.old/include/asm-mips/mach-generic/kmalloc.h linux.dev/include/asm-mips/mach-generic/kmalloc.h +--- linux.old/include/asm-mips/mach-generic/kmalloc.h 2007-01-10 20:10:37.000000000 +0100 ++++ linux.dev/include/asm-mips/mach-generic/kmalloc.h 2007-02-09 20:26:50.962537560 +0100 +@@ -5,6 +5,7 @@ + #ifndef CONFIG_DMA_COHERENT + /* + * Total overkill for most systems but need as a safe default. ++ * Set this one if any device in the system might do non-coherent DMA. + */ + #define ARCH_KMALLOC_MINALIGN 128 + #endif +diff -urN linux.old/include/asm-mips/mach-ip27/dma-coherence.h linux.dev/include/asm-mips/mach-ip27/dma-coherence.h +--- linux.old/include/asm-mips/mach-ip27/dma-coherence.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/include/asm-mips/mach-ip27/dma-coherence.h 2007-02-09 20:26:50.962537560 +0100 +@@ -0,0 +1,49 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2006 Ralf Baechle ++ * ++ */ ++#ifndef __ASM_MACH_IP27_DMA_COHERENCE_H ++#define __ASM_MACH_IP27_DMA_COHERENCE_H ++ ++#include ++ ++#define pdev_to_baddr(pdev, addr) \ ++ (BRIDGE_CONTROLLER(pdev->bus)->baddr + (addr)) ++#define dev_to_baddr(dev, addr) \ ++ pdev_to_baddr(to_pci_dev(dev), (addr)) ++ ++struct device; ++ ++static dma_addr_t plat_map_dma_mem(struct device *dev, void *addr, size_t size) ++{ ++ dma_addr_t pa = dev_to_baddr(dev, virt_to_phys(addr)); ++ ++ return pa; ++} ++ ++static dma_addr_t plat_map_dma_mem_page(struct device *dev, struct page *page) ++{ ++ dma_addr_t pa = dev_to_baddr(dev, page_to_phys(page)); ++ ++ return pa; ++} ++ ++static unsigned long plat_dma_addr_to_phys(dma_addr_t dma_addr) ++{ ++ return dma_addr & (0xffUL << 56); ++} ++ ++static void plat_unmap_dma_mem(dma_addr_t dma_addr) ++{ ++} ++ ++static inline int plat_device_is_coherent(struct device *dev) ++{ ++ return 1; /* IP27 non-cohernet mode is unsupported */ ++} ++ ++#endif /* __ASM_MACH_IP27_DMA_COHERENCE_H */ +diff -urN linux.old/include/asm-mips/mach-ip32/dma-coherence.h linux.dev/include/asm-mips/mach-ip32/dma-coherence.h +--- linux.old/include/asm-mips/mach-ip32/dma-coherence.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/include/asm-mips/mach-ip32/dma-coherence.h 2007-02-09 20:26:50.963537408 +0100 +@@ -0,0 +1,71 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2006 Ralf Baechle ++ * ++ */ ++#ifndef __ASM_MACH_IP35_DMA_COHERENCE_H ++#define __ASM_MACH_IP35_DMA_COHERENCE_H ++ ++#include ++ ++struct device; ++ ++/* ++ * Few notes. ++ * 1. CPU sees memory as two chunks: 0-256M@0x0, and the rest @0x40000000+256M ++ * 2. PCI sees memory as one big chunk @0x0 (or we could use 0x40000000 for ++ * native-endian) ++ * 3. All other devices see memory as one big chunk at 0x40000000 ++ * 4. Non-PCI devices will pass NULL as struct device* ++ * ++ * Thus we translate differently, depending on device. ++ */ ++ ++#define RAM_OFFSET_MASK 0x3fffffffUL ++ ++static dma_addr_t plat_map_dma_mem(struct device *dev, void *addr, size_t size) ++{ ++ dma_addr_t pa = virt_to_phys(addr) & RAM_OFFSET_MASK; ++ ++ if (dev == NULL) ++ pa += CRIME_HI_MEM_BASE; ++ ++ return pa; ++} ++ ++static dma_addr_t plat_map_dma_mem_page(struct device *dev, struct page *page) ++{ ++ dma_addr_t pa; ++ ++ pa = page_to_phys(page) & RAM_OFFSET_MASK; ++ ++ if (dev == NULL) ++ pa += CRIME_HI_MEM_BASE; ++ ++ return pa; ++} ++ ++/* This is almost certainly wrong but it's what dma-ip32.c used to use */ ++static unsigned long plat_dma_addr_to_phys(dma_addr_t dma_addr) ++{ ++ unsigned long addr = dma_addr & RAM_OFFSET_MASK; ++ ++ if (dma_addr >= 256*1024*1024) ++ addr += CRIME_HI_MEM_BASE; ++ ++ return addr; ++} ++ ++static void plat_unmap_dma_mem(dma_addr_t dma_addr) ++{ ++} ++ ++static inline int plat_device_is_coherent(struct device *dev) ++{ ++ return 0; /* IP32 is non-cohernet */ ++} ++ ++#endif /* __ASM_MACH_IP35_DMA_COHERENCE_H */ +diff -urN linux.old/include/asm-mips/mach-jazz/dma-coherence.h linux.dev/include/asm-mips/mach-jazz/dma-coherence.h +--- linux.old/include/asm-mips/mach-jazz/dma-coherence.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux.dev/include/asm-mips/mach-jazz/dma-coherence.h 2007-02-09 20:26:50.963537408 +0100 +@@ -0,0 +1,40 @@ ++/* ++ * This file is subject to the terms and conditions of the GNU General Public ++ * License. See the file "COPYING" in the main directory of this archive ++ * for more details. ++ * ++ * Copyright (C) 2006 Ralf Baechle ++ */ ++#ifndef __ASM_MACH_JAZZ_DMA_COHERENCE_H ++#define __ASM_MACH_JAZZ_DMA_COHERENCE_H ++ ++#include ++ ++struct device; ++ ++static dma_addr_t plat_map_dma_mem(struct device *dev, void *addr, size_t size) ++{ ++ return vdma_alloc(virt_to_phys(addr), size); ++} ++ ++static dma_addr_t plat_map_dma_mem_page(struct device *dev, struct page *page) ++{ ++ return vdma_alloc(page_to_phys(page), PAGE_SIZE); ++} ++ ++static unsigned long plat_dma_addr_to_phys(dma_addr_t dma_addr) ++{ ++ return vdma_log2phys(dma_addr); ++} ++ ++static void plat_unmap_dma_mem(dma_addr_t dma_addr) ++{ ++ vdma_free(dma_addr); ++} ++ ++static inline int plat_device_is_coherent(struct device *dev) ++{ ++ return 0; ++} ++ ++#endif /* __ASM_MACH_JAZZ_DMA_COHERENCE_H */ -- 2.25.1