.. | .. |
---|
32 | 32 | */ |
---|
33 | 33 | #include <asm/tlbflush.h> |
---|
34 | 34 | |
---|
35 | | -void *kmap_atomic_prot(struct page *page, pgprot_t prot) |
---|
| 35 | +void *kmap_atomic_high_prot(struct page *page, pgprot_t prot) |
---|
36 | 36 | { |
---|
37 | 37 | |
---|
38 | 38 | unsigned long vaddr; |
---|
39 | 39 | int idx, type; |
---|
40 | | - |
---|
41 | | - preempt_disable(); |
---|
42 | | - pagefault_disable(); |
---|
43 | | - if (!PageHighMem(page)) |
---|
44 | | - return page_address(page); |
---|
45 | | - |
---|
46 | 40 | |
---|
47 | 41 | type = kmap_atomic_idx_push(); |
---|
48 | 42 | idx = type + KM_TYPE_NR*smp_processor_id(); |
---|
.. | .. |
---|
55 | 49 | |
---|
56 | 50 | return (void *) vaddr; |
---|
57 | 51 | } |
---|
58 | | -EXPORT_SYMBOL(kmap_atomic_prot); |
---|
| 52 | +EXPORT_SYMBOL(kmap_atomic_high_prot); |
---|
59 | 53 | |
---|
60 | | -void __kunmap_atomic(void *kvaddr) |
---|
| 54 | +void kunmap_atomic_high(void *kvaddr) |
---|
61 | 55 | { |
---|
62 | 56 | unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; |
---|
63 | 57 | int type; |
---|
64 | 58 | unsigned int idx; |
---|
65 | 59 | |
---|
66 | | - if (vaddr < __fix_to_virt(FIX_KMAP_END)) { |
---|
67 | | - pagefault_enable(); |
---|
68 | | - preempt_enable(); |
---|
| 60 | + if (vaddr < __fix_to_virt(FIX_KMAP_END)) |
---|
69 | 61 | return; |
---|
70 | | - } |
---|
71 | 62 | |
---|
72 | 63 | type = kmap_atomic_idx(); |
---|
73 | 64 | |
---|
.. | .. |
---|
83 | 74 | local_flush_tlb_page(NULL, vaddr); |
---|
84 | 75 | |
---|
85 | 76 | kmap_atomic_idx_pop(); |
---|
86 | | - pagefault_enable(); |
---|
87 | | - preempt_enable(); |
---|
88 | 77 | } |
---|
89 | | -EXPORT_SYMBOL(__kunmap_atomic); |
---|
| 78 | +EXPORT_SYMBOL(kunmap_atomic_high); |
---|