| .. | .. |
|---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-only |
|---|
| 1 | 2 | #include <linux/highmem.h> |
|---|
| 2 | 3 | #include <linux/export.h> |
|---|
| 3 | 4 | #include <linux/swap.h> /* for totalram_pages */ |
|---|
| 4 | | -#include <linux/bootmem.h> |
|---|
| 5 | +#include <linux/memblock.h> |
|---|
| 5 | 6 | |
|---|
| 6 | | -void *kmap(struct page *page) |
|---|
| 7 | +void *kmap_atomic_high_prot(struct page *page, pgprot_t prot) |
|---|
| 7 | 8 | { |
|---|
| 8 | | - might_sleep(); |
|---|
| 9 | | - if (!PageHighMem(page)) |
|---|
| 10 | | - return page_address(page); |
|---|
| 11 | | - return kmap_high(page); |
|---|
| 12 | | -} |
|---|
| 13 | | -EXPORT_SYMBOL(kmap); |
|---|
| 14 | | - |
|---|
| 15 | | -void kunmap(struct page *page) |
|---|
| 16 | | -{ |
|---|
| 17 | | - if (in_interrupt()) |
|---|
| 18 | | - BUG(); |
|---|
| 19 | | - if (!PageHighMem(page)) |
|---|
| 20 | | - return; |
|---|
| 21 | | - kunmap_high(page); |
|---|
| 22 | | -} |
|---|
| 23 | | -EXPORT_SYMBOL(kunmap); |
|---|
| 24 | | - |
|---|
| 25 | | -/* |
|---|
| 26 | | - * kmap_atomic/kunmap_atomic is significantly faster than kmap/kunmap because |
|---|
| 27 | | - * no global lock is needed and because the kmap code must perform a global TLB |
|---|
| 28 | | - * invalidation when the kmap pool wraps. |
|---|
| 29 | | - * |
|---|
| 30 | | - * However when holding an atomic kmap it is not legal to sleep, so atomic |
|---|
| 31 | | - * kmaps are appropriate for short, tight code paths only. |
|---|
| 32 | | - */ |
|---|
| 33 | | -void *kmap_atomic_prot(struct page *page, pgprot_t prot) |
|---|
| 34 | | -{ |
|---|
| 35 | | - pte_t pte = mk_pte(page, prot); |
|---|
| 36 | 9 | unsigned long vaddr; |
|---|
| 37 | 10 | int idx, type; |
|---|
| 38 | | - |
|---|
| 39 | | - preempt_disable_nort(); |
|---|
| 40 | | - pagefault_disable(); |
|---|
| 41 | | - |
|---|
| 42 | | - if (!PageHighMem(page)) |
|---|
| 43 | | - return page_address(page); |
|---|
| 44 | 11 | |
|---|
| 45 | 12 | type = kmap_atomic_idx_push(); |
|---|
| 46 | 13 | idx = type + KM_TYPE_NR*smp_processor_id(); |
|---|
| 47 | 14 | vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); |
|---|
| 48 | 15 | BUG_ON(!pte_none(*(kmap_pte-idx))); |
|---|
| 49 | | -#ifdef CONFIG_PREEMPT_RT_FULL |
|---|
| 50 | | - current->kmap_pte[type] = pte; |
|---|
| 51 | | -#endif |
|---|
| 52 | | - set_pte(kmap_pte-idx, pte); |
|---|
| 16 | + set_pte(kmap_pte-idx, mk_pte(page, prot)); |
|---|
| 53 | 17 | arch_flush_lazy_mmu_mode(); |
|---|
| 54 | 18 | |
|---|
| 55 | 19 | return (void *)vaddr; |
|---|
| 56 | 20 | } |
|---|
| 57 | | -EXPORT_SYMBOL(kmap_atomic_prot); |
|---|
| 58 | | - |
|---|
| 59 | | -void *kmap_atomic(struct page *page) |
|---|
| 60 | | -{ |
|---|
| 61 | | - return kmap_atomic_prot(page, kmap_prot); |
|---|
| 62 | | -} |
|---|
| 63 | | -EXPORT_SYMBOL(kmap_atomic); |
|---|
| 21 | +EXPORT_SYMBOL(kmap_atomic_high_prot); |
|---|
| 64 | 22 | |
|---|
| 65 | 23 | /* |
|---|
| 66 | 24 | * This is the same as kmap_atomic() but can map memory that doesn't |
|---|
| .. | .. |
|---|
| 72 | 30 | } |
|---|
| 73 | 31 | EXPORT_SYMBOL_GPL(kmap_atomic_pfn); |
|---|
| 74 | 32 | |
|---|
| 75 | | -void __kunmap_atomic(void *kvaddr) |
|---|
| 33 | +void kunmap_atomic_high(void *kvaddr) |
|---|
| 76 | 34 | { |
|---|
| 77 | 35 | unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; |
|---|
| 78 | 36 | |
|---|
| .. | .. |
|---|
| 92 | 50 | * is a bad idea also, in case the page changes cacheability |
|---|
| 93 | 51 | * attributes or becomes a protected page in a hypervisor. |
|---|
| 94 | 52 | */ |
|---|
| 95 | | -#ifdef CONFIG_PREEMPT_RT_FULL |
|---|
| 96 | | - current->kmap_pte[type] = __pte(0); |
|---|
| 97 | | -#endif |
|---|
| 98 | 53 | kpte_clear_flush(kmap_pte-idx, vaddr); |
|---|
| 99 | 54 | kmap_atomic_idx_pop(); |
|---|
| 100 | 55 | arch_flush_lazy_mmu_mode(); |
|---|
| .. | .. |
|---|
| 105 | 60 | BUG_ON(vaddr >= (unsigned long)high_memory); |
|---|
| 106 | 61 | } |
|---|
| 107 | 62 | #endif |
|---|
| 108 | | - |
|---|
| 109 | | - pagefault_enable(); |
|---|
| 110 | | - preempt_enable_nort(); |
|---|
| 111 | 63 | } |
|---|
| 112 | | -EXPORT_SYMBOL(__kunmap_atomic); |
|---|
| 64 | +EXPORT_SYMBOL(kunmap_atomic_high); |
|---|
| 113 | 65 | |
|---|
| 114 | 66 | void __init set_highmem_pages_init(void) |
|---|
| 115 | 67 | { |
|---|
| .. | .. |
|---|
| 118 | 70 | |
|---|
| 119 | 71 | /* |
|---|
| 120 | 72 | * Explicitly reset zone->managed_pages because set_highmem_pages_init() |
|---|
| 121 | | - * is invoked before free_all_bootmem() |
|---|
| 73 | + * is invoked before memblock_free_all() |
|---|
| 122 | 74 | */ |
|---|
| 123 | 75 | reset_all_zones_managed_pages(); |
|---|
| 124 | 76 | for_each_zone(zone) { |
|---|