.. | .. |
---|
12 | 12 | #include <linux/highmem.h> |
---|
13 | 13 | #include <asm/tlbflush.h> |
---|
14 | 14 | |
---|
15 | | -static pte_t *kmap_pte; |
---|
16 | | - |
---|
17 | 15 | #if DCACHE_WAY_SIZE > PAGE_SIZE |
---|
18 | 16 | unsigned int last_pkmap_nr_arr[DCACHE_N_COLORS]; |
---|
19 | 17 | wait_queue_head_t pkmap_map_wait_arr[DCACHE_N_COLORS]; |
---|
.. | .. |
---|
33 | 31 | |
---|
34 | 32 | static inline enum fixed_addresses kmap_idx(int type, unsigned long color) |
---|
35 | 33 | { |
---|
36 | | - return (type + KM_TYPE_NR * smp_processor_id()) * DCACHE_N_COLORS + |
---|
| 34 | + return (type + KM_MAX_IDX * smp_processor_id()) * DCACHE_N_COLORS + |
---|
37 | 35 | color; |
---|
38 | 36 | } |
---|
39 | 37 | |
---|
40 | | -void *kmap_atomic(struct page *page) |
---|
| 38 | +enum fixed_addresses kmap_local_map_idx(int type, unsigned long pfn) |
---|
41 | 39 | { |
---|
42 | | - enum fixed_addresses idx; |
---|
43 | | - unsigned long vaddr; |
---|
44 | | - |
---|
45 | | - preempt_disable(); |
---|
46 | | - pagefault_disable(); |
---|
47 | | - if (!PageHighMem(page)) |
---|
48 | | - return page_address(page); |
---|
49 | | - |
---|
50 | | - idx = kmap_idx(kmap_atomic_idx_push(), |
---|
51 | | - DCACHE_ALIAS(page_to_phys(page))); |
---|
52 | | - vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); |
---|
53 | | -#ifdef CONFIG_DEBUG_HIGHMEM |
---|
54 | | - BUG_ON(!pte_none(*(kmap_pte + idx))); |
---|
55 | | -#endif |
---|
56 | | - set_pte(kmap_pte + idx, mk_pte(page, PAGE_KERNEL_EXEC)); |
---|
57 | | - |
---|
58 | | - return (void *)vaddr; |
---|
| 40 | + return kmap_idx(type, DCACHE_ALIAS(pfn << PAGE_SHIFT)); |
---|
59 | 41 | } |
---|
60 | | -EXPORT_SYMBOL(kmap_atomic); |
---|
61 | 42 | |
---|
62 | | -void __kunmap_atomic(void *kvaddr) |
---|
| 43 | +enum fixed_addresses kmap_local_unmap_idx(int type, unsigned long addr) |
---|
63 | 44 | { |
---|
64 | | - if (kvaddr >= (void *)FIXADDR_START && |
---|
65 | | - kvaddr < (void *)FIXADDR_TOP) { |
---|
66 | | - int idx = kmap_idx(kmap_atomic_idx(), |
---|
67 | | - DCACHE_ALIAS((unsigned long)kvaddr)); |
---|
68 | | - |
---|
69 | | - /* |
---|
70 | | - * Force other mappings to Oops if they'll try to access this |
---|
71 | | - * pte without first remap it. Keeping stale mappings around |
---|
72 | | - * is a bad idea also, in case the page changes cacheability |
---|
73 | | - * attributes or becomes a protected page in a hypervisor. |
---|
74 | | - */ |
---|
75 | | - pte_clear(&init_mm, kvaddr, kmap_pte + idx); |
---|
76 | | - local_flush_tlb_kernel_range((unsigned long)kvaddr, |
---|
77 | | - (unsigned long)kvaddr + PAGE_SIZE); |
---|
78 | | - |
---|
79 | | - kmap_atomic_idx_pop(); |
---|
80 | | - } |
---|
81 | | - |
---|
82 | | - pagefault_enable(); |
---|
83 | | - preempt_enable(); |
---|
| 45 | + return kmap_idx(type, DCACHE_ALIAS(addr)); |
---|
84 | 46 | } |
---|
85 | | -EXPORT_SYMBOL(__kunmap_atomic); |
---|
86 | 47 | |
---|
87 | 48 | void __init kmap_init(void) |
---|
88 | 49 | { |
---|
89 | | - unsigned long kmap_vstart; |
---|
90 | | - |
---|
91 | | - /* cache the first kmap pte */ |
---|
92 | | - kmap_vstart = __fix_to_virt(FIX_KMAP_BEGIN); |
---|
93 | | - kmap_pte = kmap_get_fixmap_pte(kmap_vstart); |
---|
| 50 | + /* Check if this memory layout is broken because PKMAP overlaps |
---|
| 51 | + * page table. |
---|
| 52 | + */ |
---|
| 53 | + BUILD_BUG_ON(PKMAP_BASE < TLBTEMP_BASE_1 + TLBTEMP_SIZE); |
---|
94 | 54 | kmap_waitqueues_init(); |
---|
95 | 55 | } |
---|