.. | .. |
---|
19 | 19 | #include <linux/mm.h> |
---|
20 | 20 | #include <linux/nmi.h> |
---|
21 | 21 | #include <linux/swap.h> |
---|
22 | | -#include <linux/bootmem.h> |
---|
23 | 22 | #include <linux/memblock.h> |
---|
24 | 23 | #include <linux/acpi.h> |
---|
25 | 24 | #include <linux/efi.h> |
---|
26 | 25 | #include <linux/nodemask.h> |
---|
27 | 26 | #include <linux/slab.h> |
---|
28 | | -#include <asm/pgalloc.h> |
---|
29 | 27 | #include <asm/tlb.h> |
---|
30 | 28 | #include <asm/meminit.h> |
---|
31 | 29 | #include <asm/numa.h> |
---|
.. | .. |
---|
181 | 179 | void __init setup_per_cpu_areas(void) |
---|
182 | 180 | { |
---|
183 | 181 | struct pcpu_alloc_info *ai; |
---|
184 | | - struct pcpu_group_info *uninitialized_var(gi); |
---|
| 182 | + struct pcpu_group_info *gi; |
---|
185 | 183 | unsigned int *cpu_map; |
---|
186 | 184 | void *base; |
---|
187 | 185 | unsigned long base_offset; |
---|
188 | 186 | unsigned int cpu; |
---|
189 | 187 | ssize_t static_size, reserved_size, dyn_size; |
---|
190 | | - int node, prev_node, unit, nr_units, rc; |
---|
| 188 | + int node, prev_node, unit, nr_units; |
---|
191 | 189 | |
---|
192 | 190 | ai = pcpu_alloc_alloc_info(MAX_NUMNODES, nr_cpu_ids); |
---|
193 | 191 | if (!ai) |
---|
.. | .. |
---|
228 | 226 | * CPUs are put into groups according to node. Walk cpu_map |
---|
229 | 227 | * and create new groups at node boundaries. |
---|
230 | 228 | */ |
---|
231 | | - prev_node = -1; |
---|
| 229 | + prev_node = NUMA_NO_NODE; |
---|
232 | 230 | ai->nr_groups = 0; |
---|
233 | 231 | for (unit = 0; unit < nr_units; unit++) { |
---|
234 | 232 | cpu = cpu_map[unit]; |
---|
.. | .. |
---|
246 | 244 | gi->cpu_map = &cpu_map[unit]; |
---|
247 | 245 | } |
---|
248 | 246 | |
---|
249 | | - rc = pcpu_setup_first_chunk(ai, base); |
---|
250 | | - if (rc) |
---|
251 | | - panic("failed to setup percpu area (err=%d)", rc); |
---|
252 | | - |
---|
| 247 | + pcpu_setup_first_chunk(ai, base); |
---|
253 | 248 | pcpu_free_alloc_info(ai); |
---|
254 | 249 | } |
---|
255 | 250 | #endif |
---|
.. | .. |
---|
397 | 392 | * |
---|
398 | 393 | * Each node's per-node area has a copy of the global pg_data_t list, so |
---|
399 | 394 | * we copy that to each node here, as well as setting the per-cpu pointer |
---|
400 | | - * to the local node data structure. The active_cpus field of the per-node |
---|
401 | | - * structure gets setup by the platform_cpu_init() function later. |
---|
| 395 | + * to the local node data structure. |
---|
402 | 396 | */ |
---|
403 | 397 | static void __init initialize_pernode_data(void) |
---|
404 | 398 | { |
---|
.. | .. |
---|
436 | 430 | { |
---|
437 | 431 | void *ptr = NULL; |
---|
438 | 432 | u8 best = 0xff; |
---|
439 | | - int bestnode = -1, node, anynode = 0; |
---|
| 433 | + int bestnode = NUMA_NO_NODE, node, anynode = 0; |
---|
440 | 434 | |
---|
441 | 435 | for_each_online_node(node) { |
---|
442 | 436 | if (node_isset(node, memory_less_mask)) |
---|
.. | .. |
---|
448 | 442 | anynode = node; |
---|
449 | 443 | } |
---|
450 | 444 | |
---|
451 | | - if (bestnode == -1) |
---|
| 445 | + if (bestnode == NUMA_NO_NODE) |
---|
452 | 446 | bestnode = anynode; |
---|
453 | 447 | |
---|
454 | | - ptr = __alloc_bootmem_node(pgdat_list[bestnode], pernodesize, |
---|
455 | | - PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); |
---|
| 448 | + ptr = memblock_alloc_try_nid(pernodesize, PERCPU_PAGE_SIZE, |
---|
| 449 | + __pa(MAX_DMA_ADDRESS), |
---|
| 450 | + MEMBLOCK_ALLOC_ACCESSIBLE, |
---|
| 451 | + bestnode); |
---|
| 452 | + if (!ptr) |
---|
| 453 | + panic("%s: Failed to allocate %lu bytes align=0x%lx nid=%d from=%lx\n", |
---|
| 454 | + __func__, pernodesize, PERCPU_PAGE_SIZE, bestnode, |
---|
| 455 | + __pa(MAX_DMA_ADDRESS)); |
---|
456 | 456 | |
---|
457 | 457 | return ptr; |
---|
458 | 458 | } |
---|
.. | .. |
---|
600 | 600 | |
---|
601 | 601 | max_dma = virt_to_phys((void *) MAX_DMA_ADDRESS) >> PAGE_SHIFT; |
---|
602 | 602 | |
---|
603 | | - sparse_memory_present_with_active_regions(MAX_NUMNODES); |
---|
604 | 603 | sparse_init(); |
---|
605 | 604 | |
---|
606 | 605 | #ifdef CONFIG_VIRTUAL_MEM_MAP |
---|
.. | .. |
---|
626 | 625 | max_zone_pfns[ZONE_DMA32] = max_dma; |
---|
627 | 626 | #endif |
---|
628 | 627 | max_zone_pfns[ZONE_NORMAL] = max_pfn; |
---|
629 | | - free_area_init_nodes(max_zone_pfns); |
---|
| 628 | + free_area_init(max_zone_pfns); |
---|
630 | 629 | |
---|
631 | 630 | zero_page_memmap_ptr = virt_to_page(ia64_imva(empty_zero_page)); |
---|
632 | 631 | } |
---|
.. | .. |
---|
655 | 654 | int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, |
---|
656 | 655 | struct vmem_altmap *altmap) |
---|
657 | 656 | { |
---|
658 | | - return vmemmap_populate_basepages(start, end, node); |
---|
| 657 | + return vmemmap_populate_basepages(start, end, node, NULL); |
---|
659 | 658 | } |
---|
660 | 659 | |
---|
661 | 660 | void vmemmap_free(unsigned long start, unsigned long end, |
---|