.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0 |
---|
1 | 2 | /* |
---|
2 | 3 | * Copyright (C) 2000 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com) |
---|
3 | | - * Licensed under the GPL |
---|
4 | 4 | */ |
---|
5 | 5 | |
---|
6 | 6 | #include <linux/stddef.h> |
---|
7 | 7 | #include <linux/module.h> |
---|
8 | | -#include <linux/bootmem.h> |
---|
| 8 | +#include <linux/memblock.h> |
---|
9 | 9 | #include <linux/highmem.h> |
---|
10 | 10 | #include <linux/mm.h> |
---|
11 | 11 | #include <linux/swap.h> |
---|
.. | .. |
---|
31 | 31 | |
---|
32 | 32 | /* Initialized at boot time, and readonly after that */ |
---|
33 | 33 | unsigned long long highmem; |
---|
| 34 | +EXPORT_SYMBOL(highmem); |
---|
34 | 35 | int kmalloc_ok = 0; |
---|
35 | 36 | |
---|
36 | 37 | /* Used during early boot */ |
---|
.. | .. |
---|
46 | 47 | */ |
---|
47 | 48 | brk_end = (unsigned long) UML_ROUND_UP(sbrk(0)); |
---|
48 | 49 | map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0); |
---|
49 | | - free_bootmem(__pa(brk_end), uml_reserved - brk_end); |
---|
| 50 | + memblock_free(__pa(brk_end), uml_reserved - brk_end); |
---|
50 | 51 | uml_reserved = brk_end; |
---|
51 | 52 | |
---|
52 | 53 | /* this will put all low memory onto the freelists */ |
---|
53 | | - free_all_bootmem(); |
---|
54 | | - max_low_pfn = totalram_pages; |
---|
55 | | - max_pfn = totalram_pages; |
---|
| 54 | + memblock_free_all(); |
---|
| 55 | + max_low_pfn = totalram_pages(); |
---|
| 56 | + max_pfn = max_low_pfn; |
---|
56 | 57 | mem_init_print_info(NULL); |
---|
57 | 58 | kmalloc_ok = 1; |
---|
58 | 59 | } |
---|
.. | .. |
---|
64 | 65 | static void __init one_page_table_init(pmd_t *pmd) |
---|
65 | 66 | { |
---|
66 | 67 | if (pmd_none(*pmd)) { |
---|
67 | | - pte_t *pte = (pte_t *) alloc_bootmem_low_pages(PAGE_SIZE); |
---|
| 68 | + pte_t *pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, |
---|
| 69 | + PAGE_SIZE); |
---|
| 70 | + if (!pte) |
---|
| 71 | + panic("%s: Failed to allocate %lu bytes align=%lx\n", |
---|
| 72 | + __func__, PAGE_SIZE, PAGE_SIZE); |
---|
| 73 | + |
---|
68 | 74 | set_pmd(pmd, __pmd(_KERNPG_TABLE + |
---|
69 | 75 | (unsigned long) __pa(pte))); |
---|
70 | 76 | if (pte != pte_offset_kernel(pmd, 0)) |
---|
.. | .. |
---|
75 | 81 | static void __init one_md_table_init(pud_t *pud) |
---|
76 | 82 | { |
---|
77 | 83 | #ifdef CONFIG_3_LEVEL_PGTABLES |
---|
78 | | - pmd_t *pmd_table = (pmd_t *) alloc_bootmem_low_pages(PAGE_SIZE); |
---|
| 84 | + pmd_t *pmd_table = (pmd_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); |
---|
| 85 | + if (!pmd_table) |
---|
| 86 | + panic("%s: Failed to allocate %lu bytes align=%lx\n", |
---|
| 87 | + __func__, PAGE_SIZE, PAGE_SIZE); |
---|
| 88 | + |
---|
79 | 89 | set_pud(pud, __pud(_KERNPG_TABLE + (unsigned long) __pa(pmd_table))); |
---|
80 | 90 | if (pmd_table != pmd_offset(pud, 0)) |
---|
81 | 91 | BUG(); |
---|
.. | .. |
---|
86 | 96 | pgd_t *pgd_base) |
---|
87 | 97 | { |
---|
88 | 98 | pgd_t *pgd; |
---|
| 99 | + p4d_t *p4d; |
---|
89 | 100 | pud_t *pud; |
---|
90 | 101 | pmd_t *pmd; |
---|
91 | 102 | int i, j; |
---|
.. | .. |
---|
97 | 108 | pgd = pgd_base + i; |
---|
98 | 109 | |
---|
99 | 110 | for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { |
---|
100 | | - pud = pud_offset(pgd, vaddr); |
---|
| 111 | + p4d = p4d_offset(pgd, vaddr); |
---|
| 112 | + pud = pud_offset(p4d, vaddr); |
---|
101 | 113 | if (pud_none(*pud)) |
---|
102 | 114 | one_md_table_init(pud); |
---|
103 | 115 | pmd = pmd_offset(pud, vaddr); |
---|
.. | .. |
---|
113 | 125 | { |
---|
114 | 126 | #ifdef CONFIG_ARCH_REUSE_HOST_VSYSCALL_AREA |
---|
115 | 127 | long size = FIXADDR_USER_END - FIXADDR_USER_START; |
---|
116 | | - pgd_t *pgd; |
---|
117 | | - pud_t *pud; |
---|
118 | | - pmd_t *pmd; |
---|
119 | 128 | pte_t *pte; |
---|
120 | 129 | phys_t p; |
---|
121 | 130 | unsigned long v, vaddr = FIXADDR_USER_START; |
---|
.. | .. |
---|
124 | 133 | return; |
---|
125 | 134 | |
---|
126 | 135 | fixrange_init( FIXADDR_USER_START, FIXADDR_USER_END, swapper_pg_dir); |
---|
127 | | - v = (unsigned long) alloc_bootmem_low_pages(size); |
---|
| 136 | + v = (unsigned long) memblock_alloc_low(size, PAGE_SIZE); |
---|
| 137 | + if (!v) |
---|
| 138 | + panic("%s: Failed to allocate %lu bytes align=%lx\n", |
---|
| 139 | + __func__, size, PAGE_SIZE); |
---|
| 140 | + |
---|
128 | 141 | memcpy((void *) v , (void *) FIXADDR_USER_START, size); |
---|
129 | 142 | p = __pa(v); |
---|
130 | 143 | for ( ; size > 0; size -= PAGE_SIZE, vaddr += PAGE_SIZE, |
---|
131 | 144 | p += PAGE_SIZE) { |
---|
132 | | - pgd = swapper_pg_dir + pgd_index(vaddr); |
---|
133 | | - pud = pud_offset(pgd, vaddr); |
---|
134 | | - pmd = pmd_offset(pud, vaddr); |
---|
135 | | - pte = pte_offset_kernel(pmd, vaddr); |
---|
| 145 | + pte = virt_to_kpte(vaddr); |
---|
136 | 146 | pte_set_val(*pte, p, PAGE_READONLY); |
---|
137 | 147 | } |
---|
138 | 148 | #endif |
---|
.. | .. |
---|
140 | 150 | |
---|
141 | 151 | void __init paging_init(void) |
---|
142 | 152 | { |
---|
143 | | - unsigned long zones_size[MAX_NR_ZONES], vaddr; |
---|
144 | | - int i; |
---|
| 153 | + unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 }; |
---|
| 154 | + unsigned long vaddr; |
---|
145 | 155 | |
---|
146 | | - empty_zero_page = (unsigned long *) alloc_bootmem_low_pages(PAGE_SIZE); |
---|
147 | | - for (i = 0; i < ARRAY_SIZE(zones_size); i++) |
---|
148 | | - zones_size[i] = 0; |
---|
| 156 | + empty_zero_page = (unsigned long *) memblock_alloc_low(PAGE_SIZE, |
---|
| 157 | + PAGE_SIZE); |
---|
| 158 | + if (!empty_zero_page) |
---|
| 159 | + panic("%s: Failed to allocate %lu bytes align=%lx\n", |
---|
| 160 | + __func__, PAGE_SIZE, PAGE_SIZE); |
---|
149 | 161 | |
---|
150 | | - zones_size[ZONE_NORMAL] = (end_iomem >> PAGE_SHIFT) - |
---|
151 | | - (uml_physmem >> PAGE_SHIFT); |
---|
152 | | - free_area_init(zones_size); |
---|
| 162 | + max_zone_pfn[ZONE_NORMAL] = end_iomem >> PAGE_SHIFT; |
---|
| 163 | + free_area_init(max_zone_pfn); |
---|
153 | 164 | |
---|
154 | 165 | /* |
---|
155 | 166 | * Fixed mappings, only the page table structure has to be |
---|
.. | .. |
---|
170 | 181 | { |
---|
171 | 182 | } |
---|
172 | 183 | |
---|
173 | | -#ifdef CONFIG_BLK_DEV_INITRD |
---|
174 | | -void free_initrd_mem(unsigned long start, unsigned long end) |
---|
175 | | -{ |
---|
176 | | - free_reserved_area((void *)start, (void *)end, -1, "initrd"); |
---|
177 | | -} |
---|
178 | | -#endif |
---|
179 | | - |
---|
180 | 184 | /* Allocate and free page tables. */ |
---|
181 | 185 | |
---|
182 | 186 | pgd_t *pgd_alloc(struct mm_struct *mm) |
---|
.. | .. |
---|
191 | 195 | } |
---|
192 | 196 | return pgd; |
---|
193 | 197 | } |
---|
194 | | - |
---|
195 | | -void pgd_free(struct mm_struct *mm, pgd_t *pgd) |
---|
196 | | -{ |
---|
197 | | - free_page((unsigned long) pgd); |
---|
198 | | -} |
---|
199 | | - |
---|
200 | | -pte_t *pte_alloc_one_kernel(struct mm_struct *mm, unsigned long address) |
---|
201 | | -{ |
---|
202 | | - pte_t *pte; |
---|
203 | | - |
---|
204 | | - pte = (pte_t *)__get_free_page(GFP_KERNEL|__GFP_ZERO); |
---|
205 | | - return pte; |
---|
206 | | -} |
---|
207 | | - |
---|
208 | | -pgtable_t pte_alloc_one(struct mm_struct *mm, unsigned long address) |
---|
209 | | -{ |
---|
210 | | - struct page *pte; |
---|
211 | | - |
---|
212 | | - pte = alloc_page(GFP_KERNEL|__GFP_ZERO); |
---|
213 | | - if (!pte) |
---|
214 | | - return NULL; |
---|
215 | | - if (!pgtable_page_ctor(pte)) { |
---|
216 | | - __free_page(pte); |
---|
217 | | - return NULL; |
---|
218 | | - } |
---|
219 | | - return pte; |
---|
220 | | -} |
---|
221 | | - |
---|
222 | | -#ifdef CONFIG_3_LEVEL_PGTABLES |
---|
223 | | -pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) |
---|
224 | | -{ |
---|
225 | | - pmd_t *pmd = (pmd_t *) __get_free_page(GFP_KERNEL); |
---|
226 | | - |
---|
227 | | - if (pmd) |
---|
228 | | - memset(pmd, 0, PAGE_SIZE); |
---|
229 | | - |
---|
230 | | - return pmd; |
---|
231 | | -} |
---|
232 | | -#endif |
---|
233 | 198 | |
---|
234 | 199 | void *uml_kmalloc(int size, int flags) |
---|
235 | 200 | { |
---|