263 lines
7.2 KiB
C
263 lines
7.2 KiB
C
/*
|
|
* arch/m68k/include/asm/cf_548x_cacheflush.h - Coldfire 547x/548x Cache
|
|
*
|
|
* Copyright 2007-2009 Freescale Semiconductor, Inc. All Rights Reserved.
|
|
* Kurt Mahan kmahan@freescale.com
|
|
* Shrek Wu b16972@freescale.com
|
|
*
|
|
* Based on include/asm-m68k/cacheflush.h
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2 of the License, or (at your
|
|
* option) any later version.
|
|
*/
|
|
#ifndef M68K_CF_548x_CACHEFLUSH_H
|
|
#define M68K_CF_548x_CACHEFLUSH_H
|
|
|
|
#include <asm/cfcache.h>
|
|
/*
|
|
* Cache handling functions
|
|
*/
|
|
|
|
#define flush_icache() \
|
|
({ \
|
|
unsigned long set; \
|
|
unsigned long start_set; \
|
|
unsigned long end_set; \
|
|
\
|
|
start_set = 0; \
|
|
end_set = (unsigned long)LAST_DCACHE_ADDR; \
|
|
\
|
|
for (set = start_set; set <= end_set; set += (0x10 - 3)) {\
|
|
asm volatile("cpushl %%ic,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%ic,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%ic,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set)); \
|
|
} \
|
|
})
|
|
|
|
#define flush_dcache() \
|
|
({ \
|
|
unsigned long set; \
|
|
unsigned long start_set; \
|
|
unsigned long end_set; \
|
|
\
|
|
start_set = 0; \
|
|
end_set = (unsigned long)LAST_DCACHE_ADDR; \
|
|
\
|
|
for (set = start_set; set <= end_set; set += (0x10 - 3)) { \
|
|
asm volatile("cpushl %%dc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%dc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%dc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%dc,(%0)" : "=a" (set) : "a" (set)); \
|
|
} \
|
|
})
|
|
|
|
#define flush_bcache() \
|
|
({ \
|
|
unsigned long set; \
|
|
unsigned long start_set; \
|
|
unsigned long end_set; \
|
|
\
|
|
start_set = 0; \
|
|
end_set = (unsigned long)LAST_DCACHE_ADDR; \
|
|
\
|
|
for (set = start_set; set <= end_set; set += (0x10 - 3)) { \
|
|
asm volatile("cpushl %%bc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%bc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%bc,(%0)\n" \
|
|
"\taddq%.l #1,%0\n" \
|
|
"\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set)); \
|
|
} \
|
|
})
|
|
|
|
/*
|
|
* invalidate the cache for the specified memory range.
|
|
* It starts at the physical address specified for
|
|
* the given number of bytes.
|
|
*/
|
|
extern void cache_clear(unsigned long paddr, int len);
|
|
/*
|
|
* push any dirty cache in the specified memory range.
|
|
* It starts at the physical address specified for
|
|
* the given number of bytes.
|
|
*/
|
|
extern void cache_push(unsigned long paddr, int len);
|
|
|
|
/*
|
|
* push and invalidate pages in the specified user virtual
|
|
* memory range.
|
|
*/
|
|
extern void cache_push_v(unsigned long vaddr, int len);
|
|
|
|
/* This is needed whenever the virtual mapping of the current
|
|
process changes. */
|
|
|
|
/**
|
|
* flush_cache_mm - Flush an mm_struct
|
|
* @mm: mm_struct to flush
|
|
*/
|
|
static inline void flush_cache_mm(struct mm_struct *mm)
|
|
{
|
|
if (mm == current->mm)
|
|
flush_bcache();
|
|
}
|
|
|
|
#define flush_cache_dup_mm(mm) flush_cache_mm(mm)
|
|
|
|
#define flush_cache_all() flush_bcache()
|
|
|
|
/**
|
|
* flush_cache_range - Flush a cache range
|
|
* @vma: vma struct
|
|
* @start: Starting address
|
|
* @end: Ending address
|
|
*
|
|
* flush_cache_range must be a macro to avoid a dependency on
|
|
* linux/mm.h which includes this file.
|
|
*/
|
|
static inline void flush_cache_range(struct vm_area_struct *vma,
|
|
unsigned long start, unsigned long end)
|
|
{
|
|
if (vma->vm_mm == current->mm)
|
|
flush_bcache();
|
|
/*cf_cache_flush_range(start, end);*/
|
|
}
|
|
|
|
/**
|
|
* flush_cache_page - Flush a page of the cache
|
|
* @vma: vma struct
|
|
* @vmaddr:
|
|
* @pfn: page numer
|
|
*
|
|
* flush_cache_page must be a macro to avoid a dependency on
|
|
* linux/mm.h which includes this file.
|
|
*/
|
|
static inline void flush_cache_page(struct vm_area_struct *vma,
|
|
unsigned long vmaddr, unsigned long pfn)
|
|
{
|
|
if (vma->vm_mm == current->mm)
|
|
flush_bcache();
|
|
/*cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);*/
|
|
}
|
|
|
|
/* Push the page at kernel virtual address and clear the icache */
|
|
/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
|
|
#define flush_page_to_ram(page) __flush_page_to_ram((void *) page_address(page))
|
|
extern inline void __flush_page_to_ram(void *address)
|
|
{
|
|
unsigned long set;
|
|
unsigned long start_set;
|
|
unsigned long end_set;
|
|
unsigned long addr = (unsigned long) address;
|
|
|
|
addr &= ~(PAGE_SIZE - 1);
|
|
/* round down to page start address */
|
|
|
|
start_set = addr & _ICACHE_SET_MASK;
|
|
end_set = (addr + PAGE_SIZE-1) & _ICACHE_SET_MASK;
|
|
|
|
if (start_set > end_set) {
|
|
/* from the begining to the lowest address */
|
|
for (set = 0; set <= end_set; set += (0x10 - 3)) {
|
|
asm volatile("cpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
|
|
}
|
|
/* next loop will finish the cache ie pass the hole */
|
|
end_set = LAST_ICACHE_ADDR;
|
|
}
|
|
|
|
for (set = start_set; set <= end_set; set += (0x10 - 3)) {
|
|
asm volatile("cpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
|
|
}
|
|
}
|
|
|
|
/* Use __flush_page_to_ram() for flush_dcache_page all values are same - MW */
|
|
#define flush_dcache_page(page) \
|
|
__flush_page_to_ram((void *) page_address(page))
|
|
#define flush_icache_page(vma, pg) \
|
|
__flush_page_to_ram((void *) page_address(pg))
|
|
#define flush_icache_user_range(adr, len) \
|
|
do { } while (0)
|
|
/* NL */
|
|
#define flush_icache_user_page(vma, page, addr, len) \
|
|
do { } while (0)
|
|
|
|
/* Push n pages at kernel virtual address and clear the icache */
|
|
/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
|
|
extern inline void flush_icache_range(unsigned long address,
|
|
unsigned long endaddr)
|
|
{
|
|
unsigned long set;
|
|
unsigned long start_set;
|
|
unsigned long end_set;
|
|
|
|
start_set = address & _ICACHE_SET_MASK;
|
|
end_set = endaddr & _ICACHE_SET_MASK;
|
|
|
|
if (start_set > end_set) {
|
|
/* from the begining to the lowest address */
|
|
for (set = 0; set <= end_set; set += (0x10 - 3)) {
|
|
asm volatile("cpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
|
|
}
|
|
/* next loop will finish the cache ie pass the hole */
|
|
end_set = LAST_ICACHE_ADDR;
|
|
}
|
|
for (set = start_set; set <= end_set; set += (0x10 - 3)) {
|
|
asm volatile("cpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)\n"
|
|
"\taddq%.l #1,%0\n"
|
|
"\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
|
|
}
|
|
}
|
|
|
|
static inline void copy_to_user_page(struct vm_area_struct *vma,
|
|
struct page *page, unsigned long vaddr,
|
|
void *dst, void *src, int len)
|
|
{
|
|
memcpy(dst, src, len);
|
|
flush_icache_user_page(vma, page, vaddr, len);
|
|
}
|
|
static inline void copy_from_user_page(struct vm_area_struct *vma,
|
|
struct page *page, unsigned long vaddr,
|
|
void *dst, void *src, int len)
|
|
{
|
|
memcpy(dst, src, len);
|
|
}
|
|
|
|
#define flush_cache_vmap(start, end) flush_cache_all()
|
|
#define flush_cache_vunmap(start, end) flush_cache_all()
|
|
#define flush_dcache_mmap_lock(mapping) do { } while (0)
|
|
#define flush_dcache_mmap_unlock(mapping) do { } while (0)
|
|
|
|
#endif /* M68K_CF_548x_CACHEFLUSH_H */
|