| .. | .. |
|---|
| 19 | 19 | #include <linux/mm.h> |
|---|
| 20 | 20 | #include <linux/nmi.h> |
|---|
| 21 | 21 | #include <linux/swap.h> |
|---|
| 22 | | -#include <linux/bootmem.h> |
|---|
| 23 | 22 | #include <linux/memblock.h> |
|---|
| 24 | 23 | #include <linux/acpi.h> |
|---|
| 25 | 24 | #include <linux/efi.h> |
|---|
| 26 | 25 | #include <linux/nodemask.h> |
|---|
| 27 | 26 | #include <linux/slab.h> |
|---|
| 28 | | -#include <asm/pgalloc.h> |
|---|
| 29 | 27 | #include <asm/tlb.h> |
|---|
| 30 | 28 | #include <asm/meminit.h> |
|---|
| 31 | 29 | #include <asm/numa.h> |
|---|
| .. | .. |
|---|
| 181 | 179 | void __init setup_per_cpu_areas(void) |
|---|
| 182 | 180 | { |
|---|
| 183 | 181 | struct pcpu_alloc_info *ai; |
|---|
| 184 | | - struct pcpu_group_info *uninitialized_var(gi); |
|---|
| 182 | + struct pcpu_group_info *gi; |
|---|
| 185 | 183 | unsigned int *cpu_map; |
|---|
| 186 | 184 | void *base; |
|---|
| 187 | 185 | unsigned long base_offset; |
|---|
| 188 | 186 | unsigned int cpu; |
|---|
| 189 | 187 | ssize_t static_size, reserved_size, dyn_size; |
|---|
| 190 | | - int node, prev_node, unit, nr_units, rc; |
|---|
| 188 | + int node, prev_node, unit, nr_units; |
|---|
| 191 | 189 | |
|---|
| 192 | 190 | ai = pcpu_alloc_alloc_info(MAX_NUMNODES, nr_cpu_ids); |
|---|
| 193 | 191 | if (!ai) |
|---|
| .. | .. |
|---|
| 228 | 226 | * CPUs are put into groups according to node. Walk cpu_map |
|---|
| 229 | 227 | * and create new groups at node boundaries. |
|---|
| 230 | 228 | */ |
|---|
| 231 | | - prev_node = -1; |
|---|
| 229 | + prev_node = NUMA_NO_NODE; |
|---|
| 232 | 230 | ai->nr_groups = 0; |
|---|
| 233 | 231 | for (unit = 0; unit < nr_units; unit++) { |
|---|
| 234 | 232 | cpu = cpu_map[unit]; |
|---|
| .. | .. |
|---|
| 246 | 244 | gi->cpu_map = &cpu_map[unit]; |
|---|
| 247 | 245 | } |
|---|
| 248 | 246 | |
|---|
| 249 | | - rc = pcpu_setup_first_chunk(ai, base); |
|---|
| 250 | | - if (rc) |
|---|
| 251 | | - panic("failed to setup percpu area (err=%d)", rc); |
|---|
| 252 | | - |
|---|
| 247 | + pcpu_setup_first_chunk(ai, base); |
|---|
| 253 | 248 | pcpu_free_alloc_info(ai); |
|---|
| 254 | 249 | } |
|---|
| 255 | 250 | #endif |
|---|
| .. | .. |
|---|
| 397 | 392 | * |
|---|
| 398 | 393 | * Each node's per-node area has a copy of the global pg_data_t list, so |
|---|
| 399 | 394 | * we copy that to each node here, as well as setting the per-cpu pointer |
|---|
| 400 | | - * to the local node data structure. The active_cpus field of the per-node |
|---|
| 401 | | - * structure gets setup by the platform_cpu_init() function later. |
|---|
| 395 | + * to the local node data structure. |
|---|
| 402 | 396 | */ |
|---|
| 403 | 397 | static void __init initialize_pernode_data(void) |
|---|
| 404 | 398 | { |
|---|
| .. | .. |
|---|
| 436 | 430 | { |
|---|
| 437 | 431 | void *ptr = NULL; |
|---|
| 438 | 432 | u8 best = 0xff; |
|---|
| 439 | | - int bestnode = -1, node, anynode = 0; |
|---|
| 433 | + int bestnode = NUMA_NO_NODE, node, anynode = 0; |
|---|
| 440 | 434 | |
|---|
| 441 | 435 | for_each_online_node(node) { |
|---|
| 442 | 436 | if (node_isset(node, memory_less_mask)) |
|---|
| .. | .. |
|---|
| 448 | 442 | anynode = node; |
|---|
| 449 | 443 | } |
|---|
| 450 | 444 | |
|---|
| 451 | | - if (bestnode == -1) |
|---|
| 445 | + if (bestnode == NUMA_NO_NODE) |
|---|
| 452 | 446 | bestnode = anynode; |
|---|
| 453 | 447 | |
|---|
| 454 | | - ptr = __alloc_bootmem_node(pgdat_list[bestnode], pernodesize, |
|---|
| 455 | | - PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); |
|---|
| 448 | + ptr = memblock_alloc_try_nid(pernodesize, PERCPU_PAGE_SIZE, |
|---|
| 449 | + __pa(MAX_DMA_ADDRESS), |
|---|
| 450 | + MEMBLOCK_ALLOC_ACCESSIBLE, |
|---|
| 451 | + bestnode); |
|---|
| 452 | + if (!ptr) |
|---|
| 453 | + panic("%s: Failed to allocate %lu bytes align=0x%lx nid=%d from=%lx\n", |
|---|
| 454 | + __func__, pernodesize, PERCPU_PAGE_SIZE, bestnode, |
|---|
| 455 | + __pa(MAX_DMA_ADDRESS)); |
|---|
| 456 | 456 | |
|---|
| 457 | 457 | return ptr; |
|---|
| 458 | 458 | } |
|---|
| .. | .. |
|---|
| 600 | 600 | |
|---|
| 601 | 601 | max_dma = virt_to_phys((void *) MAX_DMA_ADDRESS) >> PAGE_SHIFT; |
|---|
| 602 | 602 | |
|---|
| 603 | | - sparse_memory_present_with_active_regions(MAX_NUMNODES); |
|---|
| 604 | 603 | sparse_init(); |
|---|
| 605 | 604 | |
|---|
| 606 | 605 | #ifdef CONFIG_VIRTUAL_MEM_MAP |
|---|
| .. | .. |
|---|
| 626 | 625 | max_zone_pfns[ZONE_DMA32] = max_dma; |
|---|
| 627 | 626 | #endif |
|---|
| 628 | 627 | max_zone_pfns[ZONE_NORMAL] = max_pfn; |
|---|
| 629 | | - free_area_init_nodes(max_zone_pfns); |
|---|
| 628 | + free_area_init(max_zone_pfns); |
|---|
| 630 | 629 | |
|---|
| 631 | 630 | zero_page_memmap_ptr = virt_to_page(ia64_imva(empty_zero_page)); |
|---|
| 632 | 631 | } |
|---|
| .. | .. |
|---|
| 655 | 654 | int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, |
|---|
| 656 | 655 | struct vmem_altmap *altmap) |
|---|
| 657 | 656 | { |
|---|
| 658 | | - return vmemmap_populate_basepages(start, end, node); |
|---|
| 657 | + return vmemmap_populate_basepages(start, end, node, NULL); |
|---|
| 659 | 658 | } |
|---|
| 660 | 659 | |
|---|
| 661 | 660 | void vmemmap_free(unsigned long start, unsigned long end, |
|---|