mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-04 01:24:12 +08:00
26ff6c11ef
Cleanup of page table allocators, using generic folded PMD and PUD helpers. TLB flushing operations are moved to a more sensible spot. The page fault handler is also optimized slightly, we no longer waste cycles on IRQ disabling for flushing of the page from the ITLB, since we're already under CLI protection by the initial exception handler. Signed-off-by: Paul Mundt <lethal@linux-sh.org>
87 lines
1.8 KiB
C
87 lines
1.8 KiB
C
/*
|
|
* arch/sh/mm/consistent.c
|
|
*
|
|
* Copyright (C) 2004 Paul Mundt
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*/
|
|
#include <linux/mm.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/addrspace.h>
|
|
#include <asm/io.h>
|
|
|
|
void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *handle)
|
|
{
|
|
struct page *page, *end, *free;
|
|
void *ret;
|
|
int order;
|
|
|
|
size = PAGE_ALIGN(size);
|
|
order = get_order(size);
|
|
|
|
page = alloc_pages(gfp, order);
|
|
if (!page)
|
|
return NULL;
|
|
split_page(page, order);
|
|
|
|
ret = page_address(page);
|
|
*handle = virt_to_phys(ret);
|
|
|
|
/*
|
|
* We must flush the cache before we pass it on to the device
|
|
*/
|
|
dma_cache_wback_inv(ret, size);
|
|
|
|
page = virt_to_page(ret);
|
|
free = page + (size >> PAGE_SHIFT);
|
|
end = page + (1 << order);
|
|
|
|
while (++page < end) {
|
|
/* Free any unused pages */
|
|
if (page >= free) {
|
|
__free_page(page);
|
|
}
|
|
}
|
|
|
|
return P2SEGADDR(ret);
|
|
}
|
|
|
|
void consistent_free(void *vaddr, size_t size)
|
|
{
|
|
unsigned long addr = P1SEGADDR((unsigned long)vaddr);
|
|
struct page *page=virt_to_page(addr);
|
|
int num_pages=(size+PAGE_SIZE-1) >> PAGE_SHIFT;
|
|
int i;
|
|
|
|
for(i=0;i<num_pages;i++) {
|
|
__free_page((page+i));
|
|
}
|
|
}
|
|
|
|
void consistent_sync(void *vaddr, size_t size, int direction)
|
|
{
|
|
void * p1addr = (void*) P1SEGADDR((unsigned long)vaddr);
|
|
|
|
switch (direction) {
|
|
case DMA_FROM_DEVICE: /* invalidate only */
|
|
dma_cache_inv(p1addr, size);
|
|
break;
|
|
case DMA_TO_DEVICE: /* writeback only */
|
|
dma_cache_wback(p1addr, size);
|
|
break;
|
|
case DMA_BIDIRECTIONAL: /* writeback and invalidate */
|
|
dma_cache_wback_inv(p1addr, size);
|
|
break;
|
|
default:
|
|
BUG();
|
|
}
|
|
}
|
|
|
|
EXPORT_SYMBOL(consistent_alloc);
|
|
EXPORT_SYMBOL(consistent_free);
|
|
EXPORT_SYMBOL(consistent_sync);
|
|
|