| .. | .. |
|---|
| 12 | 12 | |
|---|
| 13 | 13 | unsigned long highstart_pfn, highend_pfn; |
|---|
| 14 | 14 | |
|---|
| 15 | | -void *kmap(struct page *page) |
|---|
| 15 | +void kmap_flush_tlb(unsigned long addr) |
|---|
| 16 | 16 | { |
|---|
| 17 | | - void *addr; |
|---|
| 18 | | - |
|---|
| 19 | | - might_sleep(); |
|---|
| 20 | | - if (!PageHighMem(page)) |
|---|
| 21 | | - return page_address(page); |
|---|
| 22 | | - addr = kmap_high(page); |
|---|
| 23 | | - flush_tlb_one((unsigned long)addr); |
|---|
| 24 | | - |
|---|
| 25 | | - return addr; |
|---|
| 17 | + flush_tlb_one(addr); |
|---|
| 26 | 18 | } |
|---|
| 27 | | -EXPORT_SYMBOL(kmap); |
|---|
| 19 | +EXPORT_SYMBOL(kmap_flush_tlb); |
|---|
| 28 | 20 | |
|---|
| 29 | | -void kunmap(struct page *page) |
|---|
| 30 | | -{ |
|---|
| 31 | | - BUG_ON(in_interrupt()); |
|---|
| 32 | | - if (!PageHighMem(page)) |
|---|
| 33 | | - return; |
|---|
| 34 | | - kunmap_high(page); |
|---|
| 35 | | -} |
|---|
| 36 | | -EXPORT_SYMBOL(kunmap); |
|---|
| 37 | | - |
|---|
| 38 | | -/* |
|---|
| 39 | | - * kmap_atomic/kunmap_atomic is significantly faster than kmap/kunmap because |
|---|
| 40 | | - * no global lock is needed and because the kmap code must perform a global TLB |
|---|
| 41 | | - * invalidation when the kmap pool wraps. |
|---|
| 42 | | - * |
|---|
| 43 | | - * However when holding an atomic kmap is is not legal to sleep, so atomic |
|---|
| 44 | | - * kmaps are appropriate for short, tight code paths only. |
|---|
| 45 | | - */ |
|---|
| 46 | | - |
|---|
| 47 | | -void *kmap_atomic(struct page *page) |
|---|
| 21 | +void *kmap_atomic_high_prot(struct page *page, pgprot_t prot) |
|---|
| 48 | 22 | { |
|---|
| 49 | 23 | unsigned long vaddr; |
|---|
| 50 | 24 | int idx, type; |
|---|
| 51 | | - |
|---|
| 52 | | - preempt_disable(); |
|---|
| 53 | | - pagefault_disable(); |
|---|
| 54 | | - if (!PageHighMem(page)) |
|---|
| 55 | | - return page_address(page); |
|---|
| 56 | 25 | |
|---|
| 57 | 26 | type = kmap_atomic_idx_push(); |
|---|
| 58 | 27 | idx = type + KM_TYPE_NR*smp_processor_id(); |
|---|
| .. | .. |
|---|
| 60 | 29 | #ifdef CONFIG_DEBUG_HIGHMEM |
|---|
| 61 | 30 | BUG_ON(!pte_none(*(kmap_pte - idx))); |
|---|
| 62 | 31 | #endif |
|---|
| 63 | | - set_pte(kmap_pte-idx, mk_pte(page, PAGE_KERNEL)); |
|---|
| 32 | + set_pte(kmap_pte-idx, mk_pte(page, prot)); |
|---|
| 64 | 33 | local_flush_tlb_one((unsigned long)vaddr); |
|---|
| 65 | 34 | |
|---|
| 66 | 35 | return (void*) vaddr; |
|---|
| 67 | 36 | } |
|---|
| 68 | | -EXPORT_SYMBOL(kmap_atomic); |
|---|
| 37 | +EXPORT_SYMBOL(kmap_atomic_high_prot); |
|---|
| 69 | 38 | |
|---|
| 70 | | -void __kunmap_atomic(void *kvaddr) |
|---|
| 39 | +void kunmap_atomic_high(void *kvaddr) |
|---|
| 71 | 40 | { |
|---|
| 72 | 41 | unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; |
|---|
| 73 | 42 | int type __maybe_unused; |
|---|
| 74 | 43 | |
|---|
| 75 | | - if (vaddr < FIXADDR_START) { // FIXME |
|---|
| 76 | | - pagefault_enable(); |
|---|
| 77 | | - preempt_enable(); |
|---|
| 44 | + if (vaddr < FIXADDR_START) |
|---|
| 78 | 45 | return; |
|---|
| 79 | | - } |
|---|
| 80 | 46 | |
|---|
| 81 | 47 | type = kmap_atomic_idx(); |
|---|
| 82 | 48 | #ifdef CONFIG_DEBUG_HIGHMEM |
|---|
| .. | .. |
|---|
| 94 | 60 | } |
|---|
| 95 | 61 | #endif |
|---|
| 96 | 62 | kmap_atomic_idx_pop(); |
|---|
| 97 | | - pagefault_enable(); |
|---|
| 98 | | - preempt_enable(); |
|---|
| 99 | 63 | } |
|---|
| 100 | | -EXPORT_SYMBOL(__kunmap_atomic); |
|---|
| 64 | +EXPORT_SYMBOL(kunmap_atomic_high); |
|---|
| 101 | 65 | |
|---|
| 102 | 66 | /* |
|---|
| 103 | 67 | * This is the same as kmap_atomic() but can map memory that doesn't |
|---|
| .. | .. |
|---|
| 126 | 90 | |
|---|
| 127 | 91 | /* cache the first kmap pte */ |
|---|
| 128 | 92 | kmap_vstart = __fix_to_virt(FIX_KMAP_BEGIN); |
|---|
| 129 | | - kmap_pte = kmap_get_fixmap_pte(kmap_vstart); |
|---|
| 93 | + kmap_pte = virt_to_kpte(kmap_vstart); |
|---|
| 130 | 94 | } |
|---|