.. | .. |
---|
254 | 254 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) |
---|
255 | 255 | #define flush_icache_page(vma, page) __flush_page_to_ram(page_address(page)) |
---|
256 | 256 | |
---|
257 | | -extern void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, |
---|
| 257 | +extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page, |
---|
258 | 258 | unsigned long addr, int len); |
---|
259 | 259 | extern void flush_icache_range(unsigned long address, unsigned long endaddr); |
---|
| 260 | +extern void flush_icache_user_range(unsigned long address, |
---|
| 261 | + unsigned long endaddr); |
---|
260 | 262 | |
---|
261 | 263 | static inline void copy_to_user_page(struct vm_area_struct *vma, |
---|
262 | 264 | struct page *page, unsigned long vaddr, |
---|
.. | .. |
---|
264 | 266 | { |
---|
265 | 267 | flush_cache_page(vma, vaddr, page_to_pfn(page)); |
---|
266 | 268 | memcpy(dst, src, len); |
---|
267 | | - flush_icache_user_range(vma, page, vaddr, len); |
---|
| 269 | + flush_icache_user_page(vma, page, vaddr, len); |
---|
268 | 270 | } |
---|
269 | 271 | static inline void copy_from_user_page(struct vm_area_struct *vma, |
---|
270 | 272 | struct page *page, unsigned long vaddr, |
---|