.. | .. |
---|
14 | 14 | #include <linux/slab.h> |
---|
15 | 15 | #include <linux/vmalloc.h> |
---|
16 | 16 | #include <linux/mm_types.h> |
---|
| 17 | +#include <linux/io.h> |
---|
17 | 18 | #include <asm/cacheflush.h> |
---|
18 | | -#include <asm/io.h> |
---|
19 | 19 | #include <asm/tlbflush.h> |
---|
| 20 | +#include <ioremap.h> |
---|
20 | 21 | |
---|
21 | | -static inline void remap_area_pte(pte_t * pte, unsigned long address, |
---|
22 | | - phys_addr_t size, phys_addr_t phys_addr, unsigned long flags) |
---|
23 | | -{ |
---|
24 | | - phys_addr_t end; |
---|
25 | | - unsigned long pfn; |
---|
26 | | - pgprot_t pgprot = __pgprot(_PAGE_GLOBAL | _PAGE_PRESENT | __READABLE |
---|
27 | | - | __WRITEABLE | flags); |
---|
28 | | - |
---|
29 | | - address &= ~PMD_MASK; |
---|
30 | | - end = address + size; |
---|
31 | | - if (end > PMD_SIZE) |
---|
32 | | - end = PMD_SIZE; |
---|
33 | | - BUG_ON(address >= end); |
---|
34 | | - pfn = phys_addr >> PAGE_SHIFT; |
---|
35 | | - do { |
---|
36 | | - if (!pte_none(*pte)) { |
---|
37 | | - printk("remap_area_pte: page already exists\n"); |
---|
38 | | - BUG(); |
---|
39 | | - } |
---|
40 | | - set_pte(pte, pfn_pte(pfn, pgprot)); |
---|
41 | | - address += PAGE_SIZE; |
---|
42 | | - pfn++; |
---|
43 | | - pte++; |
---|
44 | | - } while (address && (address < end)); |
---|
45 | | -} |
---|
46 | | - |
---|
47 | | -static inline int remap_area_pmd(pmd_t * pmd, unsigned long address, |
---|
48 | | - phys_addr_t size, phys_addr_t phys_addr, unsigned long flags) |
---|
49 | | -{ |
---|
50 | | - phys_addr_t end; |
---|
51 | | - |
---|
52 | | - address &= ~PGDIR_MASK; |
---|
53 | | - end = address + size; |
---|
54 | | - if (end > PGDIR_SIZE) |
---|
55 | | - end = PGDIR_SIZE; |
---|
56 | | - phys_addr -= address; |
---|
57 | | - BUG_ON(address >= end); |
---|
58 | | - do { |
---|
59 | | - pte_t * pte = pte_alloc_kernel(pmd, address); |
---|
60 | | - if (!pte) |
---|
61 | | - return -ENOMEM; |
---|
62 | | - remap_area_pte(pte, address, end - address, address + phys_addr, flags); |
---|
63 | | - address = (address + PMD_SIZE) & PMD_MASK; |
---|
64 | | - pmd++; |
---|
65 | | - } while (address && (address < end)); |
---|
66 | | - return 0; |
---|
67 | | -} |
---|
68 | | - |
---|
69 | | -static int remap_area_pages(unsigned long address, phys_addr_t phys_addr, |
---|
70 | | - phys_addr_t size, unsigned long flags) |
---|
71 | | -{ |
---|
72 | | - int error; |
---|
73 | | - pgd_t * dir; |
---|
74 | | - unsigned long end = address + size; |
---|
75 | | - |
---|
76 | | - phys_addr -= address; |
---|
77 | | - dir = pgd_offset(&init_mm, address); |
---|
78 | | - flush_cache_all(); |
---|
79 | | - BUG_ON(address >= end); |
---|
80 | | - do { |
---|
81 | | - pud_t *pud; |
---|
82 | | - pmd_t *pmd; |
---|
83 | | - |
---|
84 | | - error = -ENOMEM; |
---|
85 | | - pud = pud_alloc(&init_mm, dir, address); |
---|
86 | | - if (!pud) |
---|
87 | | - break; |
---|
88 | | - pmd = pmd_alloc(&init_mm, pud, address); |
---|
89 | | - if (!pmd) |
---|
90 | | - break; |
---|
91 | | - if (remap_area_pmd(pmd, address, end - address, |
---|
92 | | - phys_addr + address, flags)) |
---|
93 | | - break; |
---|
94 | | - error = 0; |
---|
95 | | - address = (address + PGDIR_SIZE) & PGDIR_MASK; |
---|
96 | | - dir++; |
---|
97 | | - } while (address && (address < end)); |
---|
98 | | - flush_tlb_all(); |
---|
99 | | - return error; |
---|
100 | | -} |
---|
| 22 | +#define IS_LOW512(addr) (!((phys_addr_t)(addr) & (phys_addr_t) ~0x1fffffffULL)) |
---|
| 23 | +#define IS_KSEG1(addr) (((unsigned long)(addr) & ~0x1fffffffUL) == CKSEG1) |
---|
101 | 24 | |
---|
102 | 25 | static int __ioremap_check_ram(unsigned long start_pfn, unsigned long nr_pages, |
---|
103 | 26 | void *arg) |
---|
.. | .. |
---|
114 | 37 | } |
---|
115 | 38 | |
---|
116 | 39 | /* |
---|
117 | | - * Generic mapping function (not visible outside): |
---|
118 | | - */ |
---|
119 | | - |
---|
120 | | -/* |
---|
121 | | - * Remap an arbitrary physical address space into the kernel virtual |
---|
122 | | - * address space. Needed when the kernel wants to access high addresses |
---|
123 | | - * directly. |
---|
| 40 | + * ioremap_prot - map bus memory into CPU space |
---|
| 41 | + * @phys_addr: bus address of the memory |
---|
| 42 | + * @size: size of the resource to map |
---|
124 | 43 | * |
---|
125 | | - * NOTE! We need to allow non-page-aligned mappings too: we will obviously |
---|
126 | | - * have to convert them into an offset in a page-aligned mapping, but the |
---|
127 | | - * caller shouldn't need to know that small detail. |
---|
| 44 | + * ioremap_prot gives the caller control over cache coherency attributes (CCA) |
---|
128 | 45 | */ |
---|
129 | | - |
---|
130 | | -#define IS_LOW512(addr) (!((phys_addr_t)(addr) & (phys_addr_t) ~0x1fffffffULL)) |
---|
131 | | - |
---|
132 | | -void __iomem * __ioremap(phys_addr_t phys_addr, phys_addr_t size, unsigned long flags) |
---|
| 46 | +void __iomem *ioremap_prot(phys_addr_t phys_addr, unsigned long size, |
---|
| 47 | + unsigned long prot_val) |
---|
133 | 48 | { |
---|
| 49 | + unsigned long flags = prot_val & _CACHE_MASK; |
---|
134 | 50 | unsigned long offset, pfn, last_pfn; |
---|
135 | | - struct vm_struct * area; |
---|
| 51 | + struct vm_struct *area; |
---|
136 | 52 | phys_addr_t last_addr; |
---|
137 | | - void * addr; |
---|
| 53 | + unsigned long vaddr; |
---|
| 54 | + void __iomem *cpu_addr; |
---|
| 55 | + |
---|
| 56 | + cpu_addr = plat_ioremap(phys_addr, size, flags); |
---|
| 57 | + if (cpu_addr) |
---|
| 58 | + return cpu_addr; |
---|
138 | 59 | |
---|
139 | 60 | phys_addr = fixup_bigphys_addr(phys_addr, size); |
---|
140 | 61 | |
---|
.. | .. |
---|
177 | 98 | area = get_vm_area(size, VM_IOREMAP); |
---|
178 | 99 | if (!area) |
---|
179 | 100 | return NULL; |
---|
180 | | - addr = area->addr; |
---|
181 | | - if (remap_area_pages((unsigned long) addr, phys_addr, size, flags)) { |
---|
182 | | - vunmap(addr); |
---|
| 101 | + vaddr = (unsigned long)area->addr; |
---|
| 102 | + |
---|
| 103 | + flags |= _PAGE_GLOBAL | _PAGE_PRESENT | __READABLE | __WRITEABLE; |
---|
| 104 | + if (ioremap_page_range(vaddr, vaddr + size, phys_addr, |
---|
| 105 | + __pgprot(flags))) { |
---|
| 106 | + free_vm_area(area); |
---|
183 | 107 | return NULL; |
---|
184 | 108 | } |
---|
185 | 109 | |
---|
186 | | - return (void __iomem *) (offset + (char *)addr); |
---|
| 110 | + return (void __iomem *)(vaddr + offset); |
---|
187 | 111 | } |
---|
| 112 | +EXPORT_SYMBOL(ioremap_prot); |
---|
188 | 113 | |
---|
189 | | -#define IS_KSEG1(addr) (((unsigned long)(addr) & ~0x1fffffffUL) == CKSEG1) |
---|
190 | | - |
---|
191 | | -void __iounmap(const volatile void __iomem *addr) |
---|
| 114 | +void iounmap(const volatile void __iomem *addr) |
---|
192 | 115 | { |
---|
193 | | - struct vm_struct *p; |
---|
194 | | - |
---|
195 | | - if (IS_KSEG1(addr)) |
---|
196 | | - return; |
---|
197 | | - |
---|
198 | | - p = remove_vm_area((void *) (PAGE_MASK & (unsigned long __force) addr)); |
---|
199 | | - if (!p) |
---|
200 | | - printk(KERN_ERR "iounmap: bad address %p\n", addr); |
---|
201 | | - |
---|
202 | | - kfree(p); |
---|
| 116 | + if (!plat_iounmap(addr) && !IS_KSEG1(addr)) |
---|
| 117 | + vunmap((void *)((unsigned long)addr & PAGE_MASK)); |
---|
203 | 118 | } |
---|
204 | | - |
---|
205 | | -EXPORT_SYMBOL(__ioremap); |
---|
206 | | -EXPORT_SYMBOL(__iounmap); |
---|
| 119 | +EXPORT_SYMBOL(iounmap); |
---|