Merge branch 'core/generic-dma-coherent' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip into for-linus
This commit is contained in:
Коммит
756f7bc668
|
@ -17,6 +17,7 @@ config ARM
|
||||||
select HAVE_KRETPROBES if (HAVE_KPROBES)
|
select HAVE_KRETPROBES if (HAVE_KPROBES)
|
||||||
select HAVE_FTRACE if (!XIP_KERNEL)
|
select HAVE_FTRACE if (!XIP_KERNEL)
|
||||||
select HAVE_DYNAMIC_FTRACE if (HAVE_FTRACE)
|
select HAVE_DYNAMIC_FTRACE if (HAVE_FTRACE)
|
||||||
|
select HAVE_GENERIC_DMA_COHERENT
|
||||||
help
|
help
|
||||||
The ARM series is a line of low-power-consumption RISC chip designs
|
The ARM series is a line of low-power-consumption RISC chip designs
|
||||||
licensed by ARM Ltd and targeted at embedded applications and
|
licensed by ARM Ltd and targeted at embedded applications and
|
||||||
|
|
|
@ -274,6 +274,11 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
|
||||||
void *
|
void *
|
||||||
dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp)
|
dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp)
|
||||||
{
|
{
|
||||||
|
void *memory;
|
||||||
|
|
||||||
|
if (dma_alloc_from_coherent(dev, size, handle, &memory))
|
||||||
|
return memory;
|
||||||
|
|
||||||
if (arch_is_coherent()) {
|
if (arch_is_coherent()) {
|
||||||
void *virt;
|
void *virt;
|
||||||
|
|
||||||
|
@ -362,6 +367,9 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr
|
||||||
|
|
||||||
WARN_ON(irqs_disabled());
|
WARN_ON(irqs_disabled());
|
||||||
|
|
||||||
|
if (dma_release_from_coherent(dev, get_order(size), cpu_addr))
|
||||||
|
return;
|
||||||
|
|
||||||
if (arch_is_coherent()) {
|
if (arch_is_coherent()) {
|
||||||
kfree(cpu_addr);
|
kfree(cpu_addr);
|
||||||
return;
|
return;
|
||||||
|
|
|
@ -641,6 +641,7 @@ config PCI
|
||||||
bool
|
bool
|
||||||
depends on ETRAX_CARDBUS
|
depends on ETRAX_CARDBUS
|
||||||
default y
|
default y
|
||||||
|
select HAVE_GENERIC_DMA_COHERENT
|
||||||
|
|
||||||
config ETRAX_IOP_FW_LOAD
|
config ETRAX_IOP_FW_LOAD
|
||||||
tristate "IO-processor hotplug firmware loading support"
|
tristate "IO-processor hotplug firmware loading support"
|
||||||
|
|
|
@ -15,35 +15,16 @@
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <asm/io.h>
|
#include <asm/io.h>
|
||||||
|
|
||||||
struct dma_coherent_mem {
|
|
||||||
void *virt_base;
|
|
||||||
u32 device_base;
|
|
||||||
int size;
|
|
||||||
int flags;
|
|
||||||
unsigned long *bitmap;
|
|
||||||
};
|
|
||||||
|
|
||||||
void *dma_alloc_coherent(struct device *dev, size_t size,
|
void *dma_alloc_coherent(struct device *dev, size_t size,
|
||||||
dma_addr_t *dma_handle, gfp_t gfp)
|
dma_addr_t *dma_handle, gfp_t gfp)
|
||||||
{
|
{
|
||||||
void *ret;
|
void *ret;
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
|
||||||
int order = get_order(size);
|
int order = get_order(size);
|
||||||
/* ignore region specifiers */
|
/* ignore region specifiers */
|
||||||
gfp &= ~(__GFP_DMA | __GFP_HIGHMEM);
|
gfp &= ~(__GFP_DMA | __GFP_HIGHMEM);
|
||||||
|
|
||||||
if (mem) {
|
if (dma_alloc_from_coherent(dev, size, dma_handle, &ret))
|
||||||
int page = bitmap_find_free_region(mem->bitmap, mem->size,
|
return ret;
|
||||||
order);
|
|
||||||
if (page >= 0) {
|
|
||||||
*dma_handle = mem->device_base + (page << PAGE_SHIFT);
|
|
||||||
ret = mem->virt_base + (page << PAGE_SHIFT);
|
|
||||||
memset(ret, 0, size);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
if (mem->flags & DMA_MEMORY_EXCLUSIVE)
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff))
|
if (dev == NULL || (dev->coherent_dma_mask < 0xffffffff))
|
||||||
gfp |= GFP_DMA;
|
gfp |= GFP_DMA;
|
||||||
|
@ -60,90 +41,9 @@ void *dma_alloc_coherent(struct device *dev, size_t size,
|
||||||
void dma_free_coherent(struct device *dev, size_t size,
|
void dma_free_coherent(struct device *dev, size_t size,
|
||||||
void *vaddr, dma_addr_t dma_handle)
|
void *vaddr, dma_addr_t dma_handle)
|
||||||
{
|
{
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
|
||||||
int order = get_order(size);
|
int order = get_order(size);
|
||||||
|
|
||||||
if (mem && vaddr >= mem->virt_base && vaddr < (mem->virt_base + (mem->size << PAGE_SHIFT))) {
|
if (!dma_release_from_coherent(dev, order, vaddr))
|
||||||
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
|
|
||||||
|
|
||||||
bitmap_release_region(mem->bitmap, page, order);
|
|
||||||
} else
|
|
||||||
free_pages((unsigned long)vaddr, order);
|
free_pages((unsigned long)vaddr, order);
|
||||||
}
|
}
|
||||||
|
|
||||||
int dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
|
||||||
dma_addr_t device_addr, size_t size, int flags)
|
|
||||||
{
|
|
||||||
void __iomem *mem_base;
|
|
||||||
int pages = size >> PAGE_SHIFT;
|
|
||||||
int bitmap_size = BITS_TO_LONGS(pages) * sizeof(long);
|
|
||||||
|
|
||||||
if ((flags & (DMA_MEMORY_MAP | DMA_MEMORY_IO)) == 0)
|
|
||||||
goto out;
|
|
||||||
if (!size)
|
|
||||||
goto out;
|
|
||||||
if (dev->dma_mem)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
/* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
|
|
||||||
|
|
||||||
mem_base = ioremap(bus_addr, size);
|
|
||||||
if (!mem_base)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
dev->dma_mem = kzalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem)
|
|
||||||
goto iounmap_out;
|
|
||||||
dev->dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem->bitmap)
|
|
||||||
goto free1_out;
|
|
||||||
|
|
||||||
dev->dma_mem->virt_base = mem_base;
|
|
||||||
dev->dma_mem->device_base = device_addr;
|
|
||||||
dev->dma_mem->size = pages;
|
|
||||||
dev->dma_mem->flags = flags;
|
|
||||||
|
|
||||||
if (flags & DMA_MEMORY_MAP)
|
|
||||||
return DMA_MEMORY_MAP;
|
|
||||||
|
|
||||||
return DMA_MEMORY_IO;
|
|
||||||
|
|
||||||
free1_out:
|
|
||||||
kfree(dev->dma_mem);
|
|
||||||
iounmap_out:
|
|
||||||
iounmap(mem_base);
|
|
||||||
out:
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_declare_coherent_memory);
|
|
||||||
|
|
||||||
void dma_release_declared_memory(struct device *dev)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
|
|
||||||
if(!mem)
|
|
||||||
return;
|
|
||||||
dev->dma_mem = NULL;
|
|
||||||
iounmap(mem->virt_base);
|
|
||||||
kfree(mem->bitmap);
|
|
||||||
kfree(mem);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_release_declared_memory);
|
|
||||||
|
|
||||||
void *dma_mark_declared_memory_occupied(struct device *dev,
|
|
||||||
dma_addr_t device_addr, size_t size)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
int pages = (size + (device_addr & ~PAGE_MASK) + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
||||||
int pos, err;
|
|
||||||
|
|
||||||
if (!mem)
|
|
||||||
return ERR_PTR(-EINVAL);
|
|
||||||
|
|
||||||
pos = (device_addr - mem->device_base) >> PAGE_SHIFT;
|
|
||||||
err = bitmap_allocate_region(mem->bitmap, pos, get_order(pages));
|
|
||||||
if (err != 0)
|
|
||||||
return ERR_PTR(err);
|
|
||||||
return mem->virt_base + (pos << PAGE_SHIFT);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_mark_declared_memory_occupied);
|
|
||||||
|
|
|
@ -11,6 +11,7 @@ config SUPERH
|
||||||
select HAVE_CLK
|
select HAVE_CLK
|
||||||
select HAVE_IDE
|
select HAVE_IDE
|
||||||
select HAVE_OPROFILE
|
select HAVE_OPROFILE
|
||||||
|
select HAVE_GENERIC_DMA_COHERENT
|
||||||
help
|
help
|
||||||
The SuperH is a RISC processor targeted for use in embedded systems
|
The SuperH is a RISC processor targeted for use in embedded systems
|
||||||
and consumer electronics; it was also used in the Sega Dreamcast
|
and consumer electronics; it was also used in the Sega Dreamcast
|
||||||
|
|
|
@ -28,21 +28,10 @@ void *dma_alloc_coherent(struct device *dev, size_t size,
|
||||||
dma_addr_t *dma_handle, gfp_t gfp)
|
dma_addr_t *dma_handle, gfp_t gfp)
|
||||||
{
|
{
|
||||||
void *ret, *ret_nocache;
|
void *ret, *ret_nocache;
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
|
||||||
int order = get_order(size);
|
int order = get_order(size);
|
||||||
|
|
||||||
if (mem) {
|
if (dma_alloc_from_coherent(dev, size, dma_handle, &ret))
|
||||||
int page = bitmap_find_free_region(mem->bitmap, mem->size,
|
return ret;
|
||||||
order);
|
|
||||||
if (page >= 0) {
|
|
||||||
*dma_handle = mem->device_base + (page << PAGE_SHIFT);
|
|
||||||
ret = mem->virt_base + (page << PAGE_SHIFT);
|
|
||||||
memset(ret, 0, size);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
if (mem->flags & DMA_MEMORY_EXCLUSIVE)
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret = (void *)__get_free_pages(gfp, order);
|
ret = (void *)__get_free_pages(gfp, order);
|
||||||
if (!ret)
|
if (!ret)
|
||||||
|
@ -72,11 +61,7 @@ void dma_free_coherent(struct device *dev, size_t size,
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
||||||
int order = get_order(size);
|
int order = get_order(size);
|
||||||
|
|
||||||
if (mem && vaddr >= mem->virt_base && vaddr < (mem->virt_base + (mem->size << PAGE_SHIFT))) {
|
if (!dma_release_from_coherent(dev, order, vaddr)) {
|
||||||
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
|
|
||||||
|
|
||||||
bitmap_release_region(mem->bitmap, page, order);
|
|
||||||
} else {
|
|
||||||
WARN_ON(irqs_disabled()); /* for portability */
|
WARN_ON(irqs_disabled()); /* for portability */
|
||||||
BUG_ON(mem && mem->flags & DMA_MEMORY_EXCLUSIVE);
|
BUG_ON(mem && mem->flags & DMA_MEMORY_EXCLUSIVE);
|
||||||
free_pages((unsigned long)phys_to_virt(dma_handle), order);
|
free_pages((unsigned long)phys_to_virt(dma_handle), order);
|
||||||
|
@ -85,83 +70,6 @@ void dma_free_coherent(struct device *dev, size_t size,
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(dma_free_coherent);
|
EXPORT_SYMBOL(dma_free_coherent);
|
||||||
|
|
||||||
int dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
|
||||||
dma_addr_t device_addr, size_t size, int flags)
|
|
||||||
{
|
|
||||||
void __iomem *mem_base = NULL;
|
|
||||||
int pages = size >> PAGE_SHIFT;
|
|
||||||
int bitmap_size = BITS_TO_LONGS(pages) * sizeof(long);
|
|
||||||
|
|
||||||
if ((flags & (DMA_MEMORY_MAP | DMA_MEMORY_IO)) == 0)
|
|
||||||
goto out;
|
|
||||||
if (!size)
|
|
||||||
goto out;
|
|
||||||
if (dev->dma_mem)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
/* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
|
|
||||||
|
|
||||||
mem_base = ioremap_nocache(bus_addr, size);
|
|
||||||
if (!mem_base)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
dev->dma_mem = kmalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem)
|
|
||||||
goto out;
|
|
||||||
dev->dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem->bitmap)
|
|
||||||
goto free1_out;
|
|
||||||
|
|
||||||
dev->dma_mem->virt_base = mem_base;
|
|
||||||
dev->dma_mem->device_base = device_addr;
|
|
||||||
dev->dma_mem->size = pages;
|
|
||||||
dev->dma_mem->flags = flags;
|
|
||||||
|
|
||||||
if (flags & DMA_MEMORY_MAP)
|
|
||||||
return DMA_MEMORY_MAP;
|
|
||||||
|
|
||||||
return DMA_MEMORY_IO;
|
|
||||||
|
|
||||||
free1_out:
|
|
||||||
kfree(dev->dma_mem);
|
|
||||||
out:
|
|
||||||
if (mem_base)
|
|
||||||
iounmap(mem_base);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_declare_coherent_memory);
|
|
||||||
|
|
||||||
void dma_release_declared_memory(struct device *dev)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
|
|
||||||
if (!mem)
|
|
||||||
return;
|
|
||||||
dev->dma_mem = NULL;
|
|
||||||
iounmap(mem->virt_base);
|
|
||||||
kfree(mem->bitmap);
|
|
||||||
kfree(mem);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_release_declared_memory);
|
|
||||||
|
|
||||||
void *dma_mark_declared_memory_occupied(struct device *dev,
|
|
||||||
dma_addr_t device_addr, size_t size)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
int pages = (size + (device_addr & ~PAGE_MASK) + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
||||||
int pos, err;
|
|
||||||
|
|
||||||
if (!mem)
|
|
||||||
return ERR_PTR(-EINVAL);
|
|
||||||
|
|
||||||
pos = (device_addr - mem->device_base) >> PAGE_SHIFT;
|
|
||||||
err = bitmap_allocate_region(mem->bitmap, pos, get_order(pages));
|
|
||||||
if (err != 0)
|
|
||||||
return ERR_PTR(err);
|
|
||||||
return mem->virt_base + (pos << PAGE_SHIFT);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_mark_declared_memory_occupied);
|
|
||||||
|
|
||||||
void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
|
void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
|
||||||
enum dma_data_direction direction)
|
enum dma_data_direction direction)
|
||||||
{
|
{
|
||||||
|
|
|
@ -30,6 +30,7 @@ config X86
|
||||||
select HAVE_FTRACE
|
select HAVE_FTRACE
|
||||||
select HAVE_KVM if ((X86_32 && !X86_VOYAGER && !X86_VISWS && !X86_NUMAQ) || X86_64)
|
select HAVE_KVM if ((X86_32 && !X86_VOYAGER && !X86_VISWS && !X86_NUMAQ) || X86_64)
|
||||||
select HAVE_ARCH_KGDB if !X86_VOYAGER
|
select HAVE_ARCH_KGDB if !X86_VOYAGER
|
||||||
|
select HAVE_GENERIC_DMA_COHERENT if X86_32
|
||||||
select HAVE_EFFICIENT_UNALIGNED_ACCESS
|
select HAVE_EFFICIENT_UNALIGNED_ACCESS
|
||||||
|
|
||||||
config ARCH_DEFCONFIG
|
config ARCH_DEFCONFIG
|
||||||
|
|
|
@ -192,124 +192,6 @@ static __init int iommu_setup(char *p)
|
||||||
}
|
}
|
||||||
early_param("iommu", iommu_setup);
|
early_param("iommu", iommu_setup);
|
||||||
|
|
||||||
#ifdef CONFIG_X86_32
|
|
||||||
int dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
|
||||||
dma_addr_t device_addr, size_t size, int flags)
|
|
||||||
{
|
|
||||||
void __iomem *mem_base = NULL;
|
|
||||||
int pages = size >> PAGE_SHIFT;
|
|
||||||
int bitmap_size = BITS_TO_LONGS(pages) * sizeof(long);
|
|
||||||
|
|
||||||
if ((flags & (DMA_MEMORY_MAP | DMA_MEMORY_IO)) == 0)
|
|
||||||
goto out;
|
|
||||||
if (!size)
|
|
||||||
goto out;
|
|
||||||
if (dev->dma_mem)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
/* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
|
|
||||||
|
|
||||||
mem_base = ioremap(bus_addr, size);
|
|
||||||
if (!mem_base)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
dev->dma_mem = kzalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem)
|
|
||||||
goto out;
|
|
||||||
dev->dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
|
|
||||||
if (!dev->dma_mem->bitmap)
|
|
||||||
goto free1_out;
|
|
||||||
|
|
||||||
dev->dma_mem->virt_base = mem_base;
|
|
||||||
dev->dma_mem->device_base = device_addr;
|
|
||||||
dev->dma_mem->size = pages;
|
|
||||||
dev->dma_mem->flags = flags;
|
|
||||||
|
|
||||||
if (flags & DMA_MEMORY_MAP)
|
|
||||||
return DMA_MEMORY_MAP;
|
|
||||||
|
|
||||||
return DMA_MEMORY_IO;
|
|
||||||
|
|
||||||
free1_out:
|
|
||||||
kfree(dev->dma_mem);
|
|
||||||
out:
|
|
||||||
if (mem_base)
|
|
||||||
iounmap(mem_base);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_declare_coherent_memory);
|
|
||||||
|
|
||||||
void dma_release_declared_memory(struct device *dev)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
|
|
||||||
if (!mem)
|
|
||||||
return;
|
|
||||||
dev->dma_mem = NULL;
|
|
||||||
iounmap(mem->virt_base);
|
|
||||||
kfree(mem->bitmap);
|
|
||||||
kfree(mem);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_release_declared_memory);
|
|
||||||
|
|
||||||
void *dma_mark_declared_memory_occupied(struct device *dev,
|
|
||||||
dma_addr_t device_addr, size_t size)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev->dma_mem;
|
|
||||||
int pos, err;
|
|
||||||
int pages = (size + (device_addr & ~PAGE_MASK) + PAGE_SIZE - 1);
|
|
||||||
|
|
||||||
pages >>= PAGE_SHIFT;
|
|
||||||
|
|
||||||
if (!mem)
|
|
||||||
return ERR_PTR(-EINVAL);
|
|
||||||
|
|
||||||
pos = (device_addr - mem->device_base) >> PAGE_SHIFT;
|
|
||||||
err = bitmap_allocate_region(mem->bitmap, pos, get_order(pages));
|
|
||||||
if (err != 0)
|
|
||||||
return ERR_PTR(err);
|
|
||||||
return mem->virt_base + (pos << PAGE_SHIFT);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(dma_mark_declared_memory_occupied);
|
|
||||||
|
|
||||||
static int dma_alloc_from_coherent_mem(struct device *dev, ssize_t size,
|
|
||||||
dma_addr_t *dma_handle, void **ret)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
|
||||||
int order = get_order(size);
|
|
||||||
|
|
||||||
if (mem) {
|
|
||||||
int page = bitmap_find_free_region(mem->bitmap, mem->size,
|
|
||||||
order);
|
|
||||||
if (page >= 0) {
|
|
||||||
*dma_handle = mem->device_base + (page << PAGE_SHIFT);
|
|
||||||
*ret = mem->virt_base + (page << PAGE_SHIFT);
|
|
||||||
memset(*ret, 0, size);
|
|
||||||
}
|
|
||||||
if (mem->flags & DMA_MEMORY_EXCLUSIVE)
|
|
||||||
*ret = NULL;
|
|
||||||
}
|
|
||||||
return (mem != NULL);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int dma_release_coherent(struct device *dev, int order, void *vaddr)
|
|
||||||
{
|
|
||||||
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
|
||||||
|
|
||||||
if (mem && vaddr >= mem->virt_base && vaddr <
|
|
||||||
(mem->virt_base + (mem->size << PAGE_SHIFT))) {
|
|
||||||
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
|
|
||||||
|
|
||||||
bitmap_release_region(mem->bitmap, page, order);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
#define dma_alloc_from_coherent_mem(dev, size, handle, ret) (0)
|
|
||||||
#define dma_release_coherent(dev, order, vaddr) (0)
|
|
||||||
#endif /* CONFIG_X86_32 */
|
|
||||||
|
|
||||||
int dma_supported(struct device *dev, u64 mask)
|
int dma_supported(struct device *dev, u64 mask)
|
||||||
{
|
{
|
||||||
struct dma_mapping_ops *ops = get_dma_ops(dev);
|
struct dma_mapping_ops *ops = get_dma_ops(dev);
|
||||||
|
@ -379,7 +261,7 @@ dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle,
|
||||||
/* ignore region specifiers */
|
/* ignore region specifiers */
|
||||||
gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32);
|
gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32);
|
||||||
|
|
||||||
if (dma_alloc_from_coherent_mem(dev, size, dma_handle, &memory))
|
if (dma_alloc_from_coherent(dev, size, dma_handle, &memory))
|
||||||
return memory;
|
return memory;
|
||||||
|
|
||||||
if (!dev) {
|
if (!dev) {
|
||||||
|
@ -484,7 +366,7 @@ void dma_free_coherent(struct device *dev, size_t size,
|
||||||
|
|
||||||
int order = get_order(size);
|
int order = get_order(size);
|
||||||
WARN_ON(irqs_disabled()); /* for portability */
|
WARN_ON(irqs_disabled()); /* for portability */
|
||||||
if (dma_release_coherent(dev, order, vaddr))
|
if (dma_release_from_coherent(dev, order, vaddr))
|
||||||
return;
|
return;
|
||||||
if (ops->unmap_single)
|
if (ops->unmap_single)
|
||||||
ops->unmap_single(dev, bus, size, 0);
|
ops->unmap_single(dev, bus, size, 0);
|
||||||
|
|
|
@ -7,6 +7,8 @@
|
||||||
|
|
||||||
#include <linux/scatterlist.h>
|
#include <linux/scatterlist.h>
|
||||||
|
|
||||||
|
#include <asm-generic/dma-coherent.h>
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* DMA-consistent mapping functions. These allocate/free a region of
|
* DMA-consistent mapping functions. These allocate/free a region of
|
||||||
* uncached, unwrite-buffered mapped memory space for use with DMA
|
* uncached, unwrite-buffered mapped memory space for use with DMA
|
||||||
|
|
|
@ -14,6 +14,8 @@
|
||||||
#define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h)
|
#define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h)
|
||||||
|
|
||||||
#ifdef CONFIG_PCI
|
#ifdef CONFIG_PCI
|
||||||
|
#include <asm-generic/dma-coherent.h>
|
||||||
|
|
||||||
void *dma_alloc_coherent(struct device *dev, size_t size,
|
void *dma_alloc_coherent(struct device *dev, size_t size,
|
||||||
dma_addr_t *dma_handle, gfp_t flag);
|
dma_addr_t *dma_handle, gfp_t flag);
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,32 @@
|
||||||
|
#ifndef DMA_COHERENT_H
|
||||||
|
#define DMA_COHERENT_H
|
||||||
|
|
||||||
|
#ifdef CONFIG_HAVE_GENERIC_DMA_COHERENT
|
||||||
|
/*
|
||||||
|
* These two functions are only for dma allocator.
|
||||||
|
* Don't use them in device drivers.
|
||||||
|
*/
|
||||||
|
int dma_alloc_from_coherent(struct device *dev, ssize_t size,
|
||||||
|
dma_addr_t *dma_handle, void **ret);
|
||||||
|
int dma_release_from_coherent(struct device *dev, int order, void *vaddr);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Standard interface
|
||||||
|
*/
|
||||||
|
#define ARCH_HAS_DMA_DECLARE_COHERENT_MEMORY
|
||||||
|
extern int
|
||||||
|
dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
||||||
|
dma_addr_t device_addr, size_t size, int flags);
|
||||||
|
|
||||||
|
extern void
|
||||||
|
dma_release_declared_memory(struct device *dev);
|
||||||
|
|
||||||
|
extern void *
|
||||||
|
dma_mark_declared_memory_occupied(struct device *dev,
|
||||||
|
dma_addr_t device_addr, size_t size);
|
||||||
|
#else
|
||||||
|
#define dma_alloc_from_coherent(dev, size, handle, ret) (0)
|
||||||
|
#define dma_release_from_coherent(dev, order, vaddr) (0)
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif
|
|
@ -5,6 +5,7 @@
|
||||||
#include <linux/scatterlist.h>
|
#include <linux/scatterlist.h>
|
||||||
#include <asm/cacheflush.h>
|
#include <asm/cacheflush.h>
|
||||||
#include <asm/io.h>
|
#include <asm/io.h>
|
||||||
|
#include <asm-generic/dma-coherent.h>
|
||||||
|
|
||||||
extern struct bus_type pci_bus_type;
|
extern struct bus_type pci_bus_type;
|
||||||
|
|
||||||
|
|
|
@ -249,25 +249,5 @@ static inline int dma_get_cache_alignment(void)
|
||||||
|
|
||||||
#define dma_is_consistent(d, h) (1)
|
#define dma_is_consistent(d, h) (1)
|
||||||
|
|
||||||
#ifdef CONFIG_X86_32
|
#include <asm-generic/dma-coherent.h>
|
||||||
# define ARCH_HAS_DMA_DECLARE_COHERENT_MEMORY
|
|
||||||
struct dma_coherent_mem {
|
|
||||||
void *virt_base;
|
|
||||||
u32 device_base;
|
|
||||||
int size;
|
|
||||||
int flags;
|
|
||||||
unsigned long *bitmap;
|
|
||||||
};
|
|
||||||
|
|
||||||
extern int
|
|
||||||
dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
|
||||||
dma_addr_t device_addr, size_t size, int flags);
|
|
||||||
|
|
||||||
extern void
|
|
||||||
dma_release_declared_memory(struct device *dev);
|
|
||||||
|
|
||||||
extern void *
|
|
||||||
dma_mark_declared_memory_occupied(struct device *dev,
|
|
||||||
dma_addr_t device_addr, size_t size);
|
|
||||||
#endif /* CONFIG_X86_32 */
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -802,6 +802,10 @@ config PROC_PAGE_MONITOR
|
||||||
|
|
||||||
endmenu # General setup
|
endmenu # General setup
|
||||||
|
|
||||||
|
config HAVE_GENERIC_DMA_COHERENT
|
||||||
|
bool
|
||||||
|
default n
|
||||||
|
|
||||||
config SLABINFO
|
config SLABINFO
|
||||||
bool
|
bool
|
||||||
depends on PROC_FS
|
depends on PROC_FS
|
||||||
|
|
|
@ -84,6 +84,7 @@ obj-$(CONFIG_TASK_DELAY_ACCT) += delayacct.o
|
||||||
obj-$(CONFIG_TASKSTATS) += taskstats.o tsacct.o
|
obj-$(CONFIG_TASKSTATS) += taskstats.o tsacct.o
|
||||||
obj-$(CONFIG_MARKERS) += marker.o
|
obj-$(CONFIG_MARKERS) += marker.o
|
||||||
obj-$(CONFIG_LATENCYTOP) += latencytop.o
|
obj-$(CONFIG_LATENCYTOP) += latencytop.o
|
||||||
|
obj-$(CONFIG_HAVE_GENERIC_DMA_COHERENT) += dma-coherent.o
|
||||||
obj-$(CONFIG_FTRACE) += trace/
|
obj-$(CONFIG_FTRACE) += trace/
|
||||||
obj-$(CONFIG_TRACING) += trace/
|
obj-$(CONFIG_TRACING) += trace/
|
||||||
obj-$(CONFIG_SMP) += sched_cpupri.o
|
obj-$(CONFIG_SMP) += sched_cpupri.o
|
||||||
|
|
|
@ -0,0 +1,154 @@
|
||||||
|
/*
|
||||||
|
* Coherent per-device memory handling.
|
||||||
|
* Borrowed from i386
|
||||||
|
*/
|
||||||
|
#include <linux/kernel.h>
|
||||||
|
#include <linux/dma-mapping.h>
|
||||||
|
|
||||||
|
struct dma_coherent_mem {
|
||||||
|
void *virt_base;
|
||||||
|
u32 device_base;
|
||||||
|
int size;
|
||||||
|
int flags;
|
||||||
|
unsigned long *bitmap;
|
||||||
|
};
|
||||||
|
|
||||||
|
int dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
|
||||||
|
dma_addr_t device_addr, size_t size, int flags)
|
||||||
|
{
|
||||||
|
void __iomem *mem_base = NULL;
|
||||||
|
int pages = size >> PAGE_SHIFT;
|
||||||
|
int bitmap_size = BITS_TO_LONGS(pages) * sizeof(long);
|
||||||
|
|
||||||
|
if ((flags & (DMA_MEMORY_MAP | DMA_MEMORY_IO)) == 0)
|
||||||
|
goto out;
|
||||||
|
if (!size)
|
||||||
|
goto out;
|
||||||
|
if (dev->dma_mem)
|
||||||
|
goto out;
|
||||||
|
|
||||||
|
/* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
|
||||||
|
|
||||||
|
mem_base = ioremap(bus_addr, size);
|
||||||
|
if (!mem_base)
|
||||||
|
goto out;
|
||||||
|
|
||||||
|
dev->dma_mem = kzalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL);
|
||||||
|
if (!dev->dma_mem)
|
||||||
|
goto out;
|
||||||
|
dev->dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
|
||||||
|
if (!dev->dma_mem->bitmap)
|
||||||
|
goto free1_out;
|
||||||
|
|
||||||
|
dev->dma_mem->virt_base = mem_base;
|
||||||
|
dev->dma_mem->device_base = device_addr;
|
||||||
|
dev->dma_mem->size = pages;
|
||||||
|
dev->dma_mem->flags = flags;
|
||||||
|
|
||||||
|
if (flags & DMA_MEMORY_MAP)
|
||||||
|
return DMA_MEMORY_MAP;
|
||||||
|
|
||||||
|
return DMA_MEMORY_IO;
|
||||||
|
|
||||||
|
free1_out:
|
||||||
|
kfree(dev->dma_mem);
|
||||||
|
out:
|
||||||
|
if (mem_base)
|
||||||
|
iounmap(mem_base);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(dma_declare_coherent_memory);
|
||||||
|
|
||||||
|
void dma_release_declared_memory(struct device *dev)
|
||||||
|
{
|
||||||
|
struct dma_coherent_mem *mem = dev->dma_mem;
|
||||||
|
|
||||||
|
if (!mem)
|
||||||
|
return;
|
||||||
|
dev->dma_mem = NULL;
|
||||||
|
iounmap(mem->virt_base);
|
||||||
|
kfree(mem->bitmap);
|
||||||
|
kfree(mem);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(dma_release_declared_memory);
|
||||||
|
|
||||||
|
void *dma_mark_declared_memory_occupied(struct device *dev,
|
||||||
|
dma_addr_t device_addr, size_t size)
|
||||||
|
{
|
||||||
|
struct dma_coherent_mem *mem = dev->dma_mem;
|
||||||
|
int pos, err;
|
||||||
|
int pages = (size + (device_addr & ~PAGE_MASK) + PAGE_SIZE - 1);
|
||||||
|
|
||||||
|
pages >>= PAGE_SHIFT;
|
||||||
|
|
||||||
|
if (!mem)
|
||||||
|
return ERR_PTR(-EINVAL);
|
||||||
|
|
||||||
|
pos = (device_addr - mem->device_base) >> PAGE_SHIFT;
|
||||||
|
err = bitmap_allocate_region(mem->bitmap, pos, get_order(pages));
|
||||||
|
if (err != 0)
|
||||||
|
return ERR_PTR(err);
|
||||||
|
return mem->virt_base + (pos << PAGE_SHIFT);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(dma_mark_declared_memory_occupied);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Try to allocate memory from the per-device coherent area.
|
||||||
|
*
|
||||||
|
* @dev: device from which we allocate memory
|
||||||
|
* @size: size of requested memory area
|
||||||
|
* @dma_handle: This will be filled with the correct dma handle
|
||||||
|
* @ret: This pointer will be filled with the virtual address
|
||||||
|
* to allocated area.
|
||||||
|
*
|
||||||
|
* This function should be only called from per-arch %dma_alloc_coherent()
|
||||||
|
* to support allocation from per-device coherent memory pools.
|
||||||
|
*
|
||||||
|
* Returns 0 if dma_alloc_coherent should continue with allocating from
|
||||||
|
* generic memory areas, or !0 if dma_alloc_coherent should return %ret.
|
||||||
|
*/
|
||||||
|
int dma_alloc_from_coherent(struct device *dev, ssize_t size,
|
||||||
|
dma_addr_t *dma_handle, void **ret)
|
||||||
|
{
|
||||||
|
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
||||||
|
int order = get_order(size);
|
||||||
|
|
||||||
|
if (mem) {
|
||||||
|
int page = bitmap_find_free_region(mem->bitmap, mem->size,
|
||||||
|
order);
|
||||||
|
if (page >= 0) {
|
||||||
|
*dma_handle = mem->device_base + (page << PAGE_SHIFT);
|
||||||
|
*ret = mem->virt_base + (page << PAGE_SHIFT);
|
||||||
|
memset(*ret, 0, size);
|
||||||
|
} else if (mem->flags & DMA_MEMORY_EXCLUSIVE)
|
||||||
|
*ret = NULL;
|
||||||
|
}
|
||||||
|
return (mem != NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Try to free the memory allocated from per-device coherent memory pool.
|
||||||
|
* @dev: device from which the memory was allocated
|
||||||
|
* @order: the order of pages allocated
|
||||||
|
* @vaddr: virtual address of allocated pages
|
||||||
|
*
|
||||||
|
* This checks whether the memory was allocated from the per-device
|
||||||
|
* coherent memory pool and if so, releases that memory.
|
||||||
|
*
|
||||||
|
* Returns 1 if we correctly released the memory, or 0 if
|
||||||
|
* %dma_release_coherent() should proceed with releasing memory from
|
||||||
|
* generic pools.
|
||||||
|
*/
|
||||||
|
int dma_release_from_coherent(struct device *dev, int order, void *vaddr)
|
||||||
|
{
|
||||||
|
struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
|
||||||
|
|
||||||
|
if (mem && vaddr >= mem->virt_base && vaddr <
|
||||||
|
(mem->virt_base + (mem->size << PAGE_SHIFT))) {
|
||||||
|
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
|
||||||
|
|
||||||
|
bitmap_release_region(mem->bitmap, page, order);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
Загрузка…
Ссылка в новой задаче