2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-23 04:34:11 +08:00

sparc: remove the sparc32_dma_ops indirection

There is no good reason to have a double indirection for the sparc32
dma ops, so remove the sparc32_dma_ops and define separate dma_map_ops
instance for the different IOMMU types.

Signed-off-by: Christoph Hellwig <hch@lst.de>
Acked-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
Christoph Hellwig 2018-12-03 14:04:32 +01:00
parent 53b7670e57
commit ce65d36f3e
4 changed files with 138 additions and 236 deletions

View File

@ -91,54 +91,10 @@ extern int isa_dma_bridge_buggy;
#endif #endif
#ifdef CONFIG_SPARC32 #ifdef CONFIG_SPARC32
/* Routines for data transfer buffers. */
struct device; struct device;
struct scatterlist;
struct sparc32_dma_ops {
__u32 (*get_scsi_one)(struct device *, char *, unsigned long);
void (*get_scsi_sgl)(struct device *, struct scatterlist *, int);
void (*release_scsi_one)(struct device *, __u32, unsigned long);
void (*release_scsi_sgl)(struct device *, struct scatterlist *,int);
#ifdef CONFIG_SBUS
int (*map_dma_area)(struct device *, dma_addr_t *, unsigned long, unsigned long, int);
void (*unmap_dma_area)(struct device *, unsigned long, int);
#endif
};
extern const struct sparc32_dma_ops *sparc32_dma_ops;
#define mmu_get_scsi_one(dev,vaddr,len) \
sparc32_dma_ops->get_scsi_one(dev, vaddr, len)
#define mmu_get_scsi_sgl(dev,sg,sz) \
sparc32_dma_ops->get_scsi_sgl(dev, sg, sz)
#define mmu_release_scsi_one(dev,vaddr,len) \
sparc32_dma_ops->release_scsi_one(dev, vaddr,len)
#define mmu_release_scsi_sgl(dev,sg,sz) \
sparc32_dma_ops->release_scsi_sgl(dev, sg, sz)
#ifdef CONFIG_SBUS
/*
* mmu_map/unmap are provided by iommu/iounit; Invalid to call on IIep.
*
* The mmu_map_dma_area establishes two mappings in one go.
* These mappings point to pages normally mapped at 'va' (linear address).
* First mapping is for CPU visible address at 'a', uncached.
* This is an alias, but it works because it is an uncached mapping.
* Second mapping is for device visible address, or "bus" address.
* The bus address is returned at '*pba'.
*
* These functions seem distinct, but are hard to split.
* On sun4m, page attributes depend on the CPU type, so we have to
* know if we are mapping RAM or I/O, so it has to be an additional argument
* to a separate mapping function for CPU visible mappings.
*/
#define sbus_map_dma_area(dev,pba,va,a,len) \
sparc32_dma_ops->map_dma_area(dev, pba, va, a, len)
#define sbus_unmap_dma_area(dev,ba,len) \
sparc32_dma_ops->unmap_dma_area(dev, ba, len)
#endif /* CONFIG_SBUS */
unsigned long sparc_dma_alloc_resource(struct device *dev, size_t len);
bool sparc_dma_free_resource(void *cpu_addr, size_t size);
#endif #endif
#endif /* !(_ASM_SPARC_DMA_H) */ #endif /* !(_ASM_SPARC_DMA_H) */

View File

@ -52,8 +52,6 @@
#include <asm/io-unit.h> #include <asm/io-unit.h>
#include <asm/leon.h> #include <asm/leon.h>
const struct sparc32_dma_ops *sparc32_dma_ops;
/* This function must make sure that caches and memory are coherent after DMA /* This function must make sure that caches and memory are coherent after DMA
* On LEON systems without cache snooping it flushes the entire D-CACHE. * On LEON systems without cache snooping it flushes the entire D-CACHE.
*/ */
@ -247,7 +245,7 @@ static void _sparc_free_io(struct resource *res)
release_resource(res); release_resource(res);
} }
static unsigned long sparc_dma_alloc_resource(struct device *dev, size_t len) unsigned long sparc_dma_alloc_resource(struct device *dev, size_t len)
{ {
struct resource *res; struct resource *res;
@ -266,7 +264,7 @@ static unsigned long sparc_dma_alloc_resource(struct device *dev, size_t len)
return res->start; return res->start;
} }
static bool sparc_dma_free_resource(void *cpu_addr, size_t size) bool sparc_dma_free_resource(void *cpu_addr, size_t size)
{ {
unsigned long addr = (unsigned long)cpu_addr; unsigned long addr = (unsigned long)cpu_addr;
struct resource *res; struct resource *res;
@ -302,122 +300,6 @@ void sbus_set_sbus64(struct device *dev, int x)
} }
EXPORT_SYMBOL(sbus_set_sbus64); EXPORT_SYMBOL(sbus_set_sbus64);
/*
* Allocate a chunk of memory suitable for DMA.
* Typically devices use them for control blocks.
* CPU may access them without any explicit flushing.
*/
static void *sbus_alloc_coherent(struct device *dev, size_t len,
dma_addr_t *dma_addrp, gfp_t gfp,
unsigned long attrs)
{
unsigned long len_total = PAGE_ALIGN(len);
unsigned long va, addr;
int order;
/* XXX why are some lengths signed, others unsigned? */
if (len <= 0) {
return NULL;
}
/* XXX So what is maxphys for us and how do drivers know it? */
if (len > 256*1024) { /* __get_free_pages() limit */
return NULL;
}
order = get_order(len_total);
va = __get_free_pages(gfp, order);
if (va == 0)
goto err_nopages;
addr = sparc_dma_alloc_resource(dev, len_total);
if (!addr)
goto err_nomem;
// XXX The sbus_map_dma_area does this for us below, see comments.
// srmmu_mapiorange(0, virt_to_phys(va), res->start, len_total);
/*
* XXX That's where sdev would be used. Currently we load
* all iommu tables with the same translations.
*/
if (sbus_map_dma_area(dev, dma_addrp, va, addr, len_total) != 0)
goto err_noiommu;
return (void *)addr;
err_noiommu:
sparc_dma_free_resource((void *)addr, len_total);
err_nomem:
free_pages(va, order);
err_nopages:
return NULL;
}
static void sbus_free_coherent(struct device *dev, size_t n, void *p,
dma_addr_t ba, unsigned long attrs)
{
struct page *pgv;
n = PAGE_ALIGN(n);
if (!sparc_dma_free_resource(p, n))
return;
pgv = virt_to_page(p);
sbus_unmap_dma_area(dev, ba, n);
__free_pages(pgv, get_order(n));
}
/*
* Map a chunk of memory so that devices can see it.
* CPU view of this memory may be inconsistent with
* a device view and explicit flushing is necessary.
*/
static dma_addr_t sbus_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t len,
enum dma_data_direction dir,
unsigned long attrs)
{
void *va = page_address(page) + offset;
/* XXX why are some lengths signed, others unsigned? */
if (len <= 0) {
return 0;
}
/* XXX So what is maxphys for us and how do drivers know it? */
if (len > 256*1024) { /* __get_free_pages() limit */
return 0;
}
return mmu_get_scsi_one(dev, va, len);
}
static void sbus_unmap_page(struct device *dev, dma_addr_t ba, size_t n,
enum dma_data_direction dir, unsigned long attrs)
{
mmu_release_scsi_one(dev, ba, n);
}
static int sbus_map_sg(struct device *dev, struct scatterlist *sg, int n,
enum dma_data_direction dir, unsigned long attrs)
{
mmu_get_scsi_sgl(dev, sg, n);
return n;
}
static void sbus_unmap_sg(struct device *dev, struct scatterlist *sg, int n,
enum dma_data_direction dir, unsigned long attrs)
{
mmu_release_scsi_sgl(dev, sg, n);
}
static const struct dma_map_ops sbus_dma_ops = {
.alloc = sbus_alloc_coherent,
.free = sbus_free_coherent,
.map_page = sbus_map_page,
.unmap_page = sbus_unmap_page,
.map_sg = sbus_map_sg,
.unmap_sg = sbus_unmap_sg,
};
static int __init sparc_register_ioport(void) static int __init sparc_register_ioport(void)
{ {
register_proc_sparc_ioport(); register_proc_sparc_ioport();
@ -491,7 +373,7 @@ void arch_sync_dma_for_cpu(struct device *dev, phys_addr_t paddr,
dma_make_coherent(paddr, PAGE_ALIGN(size)); dma_make_coherent(paddr, PAGE_ALIGN(size));
} }
const struct dma_map_ops *dma_ops = &sbus_dma_ops; const struct dma_map_ops *dma_ops;
EXPORT_SYMBOL(dma_ops); EXPORT_SYMBOL(dma_ops);
#ifdef CONFIG_PROC_FS #ifdef CONFIG_PROC_FS

View File

@ -12,7 +12,7 @@
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/highmem.h> /* pte_offset_map => kmap_atomic */ #include <linux/highmem.h> /* pte_offset_map => kmap_atomic */
#include <linux/bitops.h> #include <linux/bitops.h>
#include <linux/scatterlist.h> #include <linux/dma-mapping.h>
#include <linux/of.h> #include <linux/of.h>
#include <linux/of_device.h> #include <linux/of_device.h>
@ -140,18 +140,26 @@ nexti: scan = find_next_zero_bit(iounit->bmap, limit, scan);
return vaddr; return vaddr;
} }
static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len) static dma_addr_t iounit_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t len, enum dma_data_direction dir,
unsigned long attrs)
{ {
void *vaddr = page_address(page) + offset;
struct iounit_struct *iounit = dev->archdata.iommu; struct iounit_struct *iounit = dev->archdata.iommu;
unsigned long ret, flags; unsigned long ret, flags;
/* XXX So what is maxphys for us and how do drivers know it? */
if (!len || len > 256 * 1024)
return DMA_MAPPING_ERROR;
spin_lock_irqsave(&iounit->lock, flags); spin_lock_irqsave(&iounit->lock, flags);
ret = iounit_get_area(iounit, (unsigned long)vaddr, len); ret = iounit_get_area(iounit, (unsigned long)vaddr, len);
spin_unlock_irqrestore(&iounit->lock, flags); spin_unlock_irqrestore(&iounit->lock, flags);
return ret; return ret;
} }
static void iounit_get_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz) static int iounit_map_sg(struct device *dev, struct scatterlist *sg, int sz,
enum dma_data_direction dir, unsigned long attrs)
{ {
struct iounit_struct *iounit = dev->archdata.iommu; struct iounit_struct *iounit = dev->archdata.iommu;
unsigned long flags; unsigned long flags;
@ -165,9 +173,11 @@ static void iounit_get_scsi_sgl(struct device *dev, struct scatterlist *sg, int
sg = sg_next(sg); sg = sg_next(sg);
} }
spin_unlock_irqrestore(&iounit->lock, flags); spin_unlock_irqrestore(&iounit->lock, flags);
return sz;
} }
static void iounit_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len) static void iounit_unmap_page(struct device *dev, dma_addr_t vaddr, size_t len,
enum dma_data_direction dir, unsigned long attrs)
{ {
struct iounit_struct *iounit = dev->archdata.iommu; struct iounit_struct *iounit = dev->archdata.iommu;
unsigned long flags; unsigned long flags;
@ -181,7 +191,8 @@ static void iounit_release_scsi_one(struct device *dev, __u32 vaddr, unsigned lo
spin_unlock_irqrestore(&iounit->lock, flags); spin_unlock_irqrestore(&iounit->lock, flags);
} }
static void iounit_release_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz) static void iounit_unmap_sg(struct device *dev, struct scatterlist *sg, int sz,
enum dma_data_direction dir, unsigned long attrs)
{ {
struct iounit_struct *iounit = dev->archdata.iommu; struct iounit_struct *iounit = dev->archdata.iommu;
unsigned long flags; unsigned long flags;
@ -201,14 +212,27 @@ static void iounit_release_scsi_sgl(struct device *dev, struct scatterlist *sg,
} }
#ifdef CONFIG_SBUS #ifdef CONFIG_SBUS
static int iounit_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned long va, unsigned long addr, int len) static void *iounit_alloc(struct device *dev, size_t len,
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
{ {
struct iounit_struct *iounit = dev->archdata.iommu; struct iounit_struct *iounit = dev->archdata.iommu;
unsigned long page, end; unsigned long va, addr, page, end, ret;
pgprot_t dvma_prot; pgprot_t dvma_prot;
iopte_t __iomem *iopte; iopte_t __iomem *iopte;
*pba = addr; /* XXX So what is maxphys for us and how do drivers know it? */
if (!len || len > 256 * 1024)
return NULL;
len = PAGE_ALIGN(len);
va = __get_free_pages(gfp, get_order(len));
if (!va)
return NULL;
addr = ret = sparc_dma_alloc_resource(dev, len);
if (!addr)
goto out_free_pages;
*dma_handle = addr;
dvma_prot = __pgprot(SRMMU_CACHE | SRMMU_ET_PTE | SRMMU_PRIV); dvma_prot = __pgprot(SRMMU_CACHE | SRMMU_ET_PTE | SRMMU_PRIV);
end = PAGE_ALIGN((addr + len)); end = PAGE_ALIGN((addr + len));
@ -237,27 +261,32 @@ static int iounit_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned lon
flush_cache_all(); flush_cache_all();
flush_tlb_all(); flush_tlb_all();
return 0; return (void *)ret;
out_free_pages:
free_pages(va, get_order(len));
return NULL;
} }
static void iounit_unmap_dma_area(struct device *dev, unsigned long addr, int len) static void iounit_free(struct device *dev, size_t size, void *cpu_addr,
dma_addr_t dma_addr, unsigned long attrs)
{ {
/* XXX Somebody please fill this in */ /* XXX Somebody please fill this in */
} }
#endif #endif
static const struct sparc32_dma_ops iounit_dma_ops = { static const struct dma_map_ops iounit_dma_ops = {
.get_scsi_one = iounit_get_scsi_one,
.get_scsi_sgl = iounit_get_scsi_sgl,
.release_scsi_one = iounit_release_scsi_one,
.release_scsi_sgl = iounit_release_scsi_sgl,
#ifdef CONFIG_SBUS #ifdef CONFIG_SBUS
.map_dma_area = iounit_map_dma_area, .alloc = iounit_alloc,
.unmap_dma_area = iounit_unmap_dma_area, .free = iounit_free,
#endif #endif
.map_page = iounit_map_page,
.unmap_page = iounit_unmap_page,
.map_sg = iounit_map_sg,
.unmap_sg = iounit_unmap_sg,
}; };
void __init ld_mmu_iounit(void) void __init ld_mmu_iounit(void)
{ {
sparc32_dma_ops = &iounit_dma_ops; dma_ops = &iounit_dma_ops;
} }

View File

@ -13,7 +13,7 @@
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/highmem.h> /* pte_offset_map => kmap_atomic */ #include <linux/highmem.h> /* pte_offset_map => kmap_atomic */
#include <linux/scatterlist.h> #include <linux/dma-mapping.h>
#include <linux/of.h> #include <linux/of.h>
#include <linux/of_device.h> #include <linux/of_device.h>
@ -205,38 +205,44 @@ static u32 iommu_get_one(struct device *dev, struct page *page, int npages)
return busa0; return busa0;
} }
static u32 iommu_get_scsi_one(struct device *dev, char *vaddr, unsigned int len) static dma_addr_t __sbus_iommu_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t len)
{ {
unsigned long off; void *vaddr = page_address(page) + offset;
int npages; unsigned long off = (unsigned long)vaddr & ~PAGE_MASK;
struct page *page; unsigned long npages = (off + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
u32 busa;
off = (unsigned long)vaddr & ~PAGE_MASK; /* XXX So what is maxphys for us and how do drivers know it? */
npages = (off + len + PAGE_SIZE-1) >> PAGE_SHIFT; if (!len || len > 256 * 1024)
page = virt_to_page((unsigned long)vaddr & PAGE_MASK); return DMA_MAPPING_ERROR;
busa = iommu_get_one(dev, page, npages); return iommu_get_one(dev, virt_to_page(vaddr), npages) + off;
return busa + off;
} }
static __u32 iommu_get_scsi_one_gflush(struct device *dev, char *vaddr, unsigned long len) static dma_addr_t sbus_iommu_map_page_gflush(struct device *dev,
struct page *page, unsigned long offset, size_t len,
enum dma_data_direction dir, unsigned long attrs)
{ {
flush_page_for_dma(0); flush_page_for_dma(0);
return iommu_get_scsi_one(dev, vaddr, len); return __sbus_iommu_map_page(dev, page, offset, len);
} }
static __u32 iommu_get_scsi_one_pflush(struct device *dev, char *vaddr, unsigned long len) static dma_addr_t sbus_iommu_map_page_pflush(struct device *dev,
struct page *page, unsigned long offset, size_t len,
enum dma_data_direction dir, unsigned long attrs)
{ {
unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; void *vaddr = page_address(page) + offset;
unsigned long p = ((unsigned long)vaddr) & PAGE_MASK;
while(page < ((unsigned long)(vaddr + len))) { while (p < (unsigned long)vaddr + len) {
flush_page_for_dma(page); flush_page_for_dma(p);
page += PAGE_SIZE; p += PAGE_SIZE;
} }
return iommu_get_scsi_one(dev, vaddr, len);
return __sbus_iommu_map_page(dev, page, offset, len);
} }
static void iommu_get_scsi_sgl_gflush(struct device *dev, struct scatterlist *sg, int sz) static int sbus_iommu_map_sg_gflush(struct device *dev, struct scatterlist *sg,
int sz, enum dma_data_direction dir, unsigned long attrs)
{ {
int n; int n;
@ -248,9 +254,12 @@ static void iommu_get_scsi_sgl_gflush(struct device *dev, struct scatterlist *sg
sg->dma_length = sg->length; sg->dma_length = sg->length;
sg = sg_next(sg); sg = sg_next(sg);
} }
return sz;
} }
static void iommu_get_scsi_sgl_pflush(struct device *dev, struct scatterlist *sg, int sz) static int sbus_iommu_map_sg_pflush(struct device *dev, struct scatterlist *sg,
int sz, enum dma_data_direction dir, unsigned long attrs)
{ {
unsigned long page, oldpage = 0; unsigned long page, oldpage = 0;
int n, i; int n, i;
@ -279,6 +288,8 @@ static void iommu_get_scsi_sgl_pflush(struct device *dev, struct scatterlist *sg
sg->dma_length = sg->length; sg->dma_length = sg->length;
sg = sg_next(sg); sg = sg_next(sg);
} }
return sz;
} }
static void iommu_release_one(struct device *dev, u32 busa, int npages) static void iommu_release_one(struct device *dev, u32 busa, int npages)
@ -297,23 +308,23 @@ static void iommu_release_one(struct device *dev, u32 busa, int npages)
bit_map_clear(&iommu->usemap, ioptex, npages); bit_map_clear(&iommu->usemap, ioptex, npages);
} }
static void iommu_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len) static void sbus_iommu_unmap_page(struct device *dev, dma_addr_t dma_addr,
size_t len, enum dma_data_direction dir, unsigned long attrs)
{ {
unsigned long off; unsigned long off = dma_addr & ~PAGE_MASK;
int npages; int npages;
off = vaddr & ~PAGE_MASK;
npages = (off + len + PAGE_SIZE-1) >> PAGE_SHIFT; npages = (off + len + PAGE_SIZE-1) >> PAGE_SHIFT;
iommu_release_one(dev, vaddr & PAGE_MASK, npages); iommu_release_one(dev, dma_addr & PAGE_MASK, npages);
} }
static void iommu_release_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz) static void sbus_iommu_unmap_sg(struct device *dev, struct scatterlist *sg,
int sz, enum dma_data_direction dir, unsigned long attrs)
{ {
int n; int n;
while(sz != 0) { while(sz != 0) {
--sz; --sz;
n = (sg->length + sg->offset + PAGE_SIZE-1) >> PAGE_SHIFT; n = (sg->length + sg->offset + PAGE_SIZE-1) >> PAGE_SHIFT;
iommu_release_one(dev, sg->dma_address & PAGE_MASK, n); iommu_release_one(dev, sg->dma_address & PAGE_MASK, n);
sg->dma_address = 0x21212121; sg->dma_address = 0x21212121;
@ -322,15 +333,28 @@ static void iommu_release_scsi_sgl(struct device *dev, struct scatterlist *sg, i
} }
#ifdef CONFIG_SBUS #ifdef CONFIG_SBUS
static int iommu_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned long va, static void *sbus_iommu_alloc(struct device *dev, size_t len,
unsigned long addr, int len) dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
{ {
struct iommu_struct *iommu = dev->archdata.iommu; struct iommu_struct *iommu = dev->archdata.iommu;
unsigned long page, end; unsigned long va, addr, page, end, ret;
iopte_t *iopte = iommu->page_table; iopte_t *iopte = iommu->page_table;
iopte_t *first; iopte_t *first;
int ioptex; int ioptex;
/* XXX So what is maxphys for us and how do drivers know it? */
if (!len || len > 256 * 1024)
return NULL;
len = PAGE_ALIGN(len);
va = __get_free_pages(gfp, get_order(len));
if (va == 0)
return NULL;
addr = ret = sparc_dma_alloc_resource(dev, len);
if (!addr)
goto out_free_pages;
BUG_ON((va & ~PAGE_MASK) != 0); BUG_ON((va & ~PAGE_MASK) != 0);
BUG_ON((addr & ~PAGE_MASK) != 0); BUG_ON((addr & ~PAGE_MASK) != 0);
BUG_ON((len & ~PAGE_MASK) != 0); BUG_ON((len & ~PAGE_MASK) != 0);
@ -385,16 +409,25 @@ static int iommu_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned long
flush_tlb_all(); flush_tlb_all();
iommu_invalidate(iommu->regs); iommu_invalidate(iommu->regs);
*pba = iommu->start + (ioptex << PAGE_SHIFT); *dma_handle = iommu->start + (ioptex << PAGE_SHIFT);
return 0; return (void *)ret;
out_free_pages:
free_pages(va, get_order(len));
return NULL;
} }
static void iommu_unmap_dma_area(struct device *dev, unsigned long busa, int len) static void sbus_iommu_free(struct device *dev, size_t len, void *cpu_addr,
dma_addr_t busa, unsigned long attrs)
{ {
struct iommu_struct *iommu = dev->archdata.iommu; struct iommu_struct *iommu = dev->archdata.iommu;
iopte_t *iopte = iommu->page_table; iopte_t *iopte = iommu->page_table;
unsigned long end; struct page *page = virt_to_page(cpu_addr);
int ioptex = (busa - iommu->start) >> PAGE_SHIFT; int ioptex = (busa - iommu->start) >> PAGE_SHIFT;
unsigned long end;
if (!sparc_dma_free_resource(cpu_addr, len))
return;
BUG_ON((busa & ~PAGE_MASK) != 0); BUG_ON((busa & ~PAGE_MASK) != 0);
BUG_ON((len & ~PAGE_MASK) != 0); BUG_ON((len & ~PAGE_MASK) != 0);
@ -408,38 +441,40 @@ static void iommu_unmap_dma_area(struct device *dev, unsigned long busa, int len
flush_tlb_all(); flush_tlb_all();
iommu_invalidate(iommu->regs); iommu_invalidate(iommu->regs);
bit_map_clear(&iommu->usemap, ioptex, len >> PAGE_SHIFT); bit_map_clear(&iommu->usemap, ioptex, len >> PAGE_SHIFT);
__free_pages(page, get_order(len));
} }
#endif #endif
static const struct sparc32_dma_ops iommu_dma_gflush_ops = { static const struct dma_map_ops sbus_iommu_dma_gflush_ops = {
.get_scsi_one = iommu_get_scsi_one_gflush,
.get_scsi_sgl = iommu_get_scsi_sgl_gflush,
.release_scsi_one = iommu_release_scsi_one,
.release_scsi_sgl = iommu_release_scsi_sgl,
#ifdef CONFIG_SBUS #ifdef CONFIG_SBUS
.map_dma_area = iommu_map_dma_area, .alloc = sbus_iommu_alloc,
.unmap_dma_area = iommu_unmap_dma_area, .free = sbus_iommu_free,
#endif #endif
.map_page = sbus_iommu_map_page_gflush,
.unmap_page = sbus_iommu_unmap_page,
.map_sg = sbus_iommu_map_sg_gflush,
.unmap_sg = sbus_iommu_unmap_sg,
}; };
static const struct sparc32_dma_ops iommu_dma_pflush_ops = { static const struct dma_map_ops sbus_iommu_dma_pflush_ops = {
.get_scsi_one = iommu_get_scsi_one_pflush,
.get_scsi_sgl = iommu_get_scsi_sgl_pflush,
.release_scsi_one = iommu_release_scsi_one,
.release_scsi_sgl = iommu_release_scsi_sgl,
#ifdef CONFIG_SBUS #ifdef CONFIG_SBUS
.map_dma_area = iommu_map_dma_area, .alloc = sbus_iommu_alloc,
.unmap_dma_area = iommu_unmap_dma_area, .free = sbus_iommu_free,
#endif #endif
.map_page = sbus_iommu_map_page_pflush,
.unmap_page = sbus_iommu_unmap_page,
.map_sg = sbus_iommu_map_sg_pflush,
.unmap_sg = sbus_iommu_unmap_sg,
}; };
void __init ld_mmu_iommu(void) void __init ld_mmu_iommu(void)
{ {
if (flush_page_for_dma_global) { if (flush_page_for_dma_global) {
/* flush_page_for_dma flushes everything, no matter of what page is it */ /* flush_page_for_dma flushes everything, no matter of what page is it */
sparc32_dma_ops = &iommu_dma_gflush_ops; dma_ops = &sbus_iommu_dma_gflush_ops;
} else { } else {
sparc32_dma_ops = &iommu_dma_pflush_ops; dma_ops = &sbus_iommu_dma_pflush_ops;
} }
if (viking_mxcc_present || srmmu_modtype == HyperSparc) { if (viking_mxcc_present || srmmu_modtype == HyperSparc) {