// SPDX-License-Identifier: GPL-2.0 /* * NUMA support for s390 * * Implement NUMA core code. * * Copyright IBM Corp. 2015 */ #define KMSG_COMPONENT "numa" #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt #include #include #include #include #include #include #include #include #include "numa_mode.h" pg_data_t *node_data[MAX_NUMNODES]; EXPORT_SYMBOL(node_data); cpumask_t node_to_cpumask_map[MAX_NUMNODES]; EXPORT_SYMBOL(node_to_cpumask_map); static void plain_setup(void) { node_set(0, node_possible_map); } const struct numa_mode numa_mode_plain = { .name = "plain", .setup = plain_setup, }; static const struct numa_mode *mode = &numa_mode_plain; int numa_pfn_to_nid(unsigned long pfn) { return mode->__pfn_to_nid ? mode->__pfn_to_nid(pfn) : 0; } void numa_update_cpu_topology(void) { if (mode->update_cpu_topology) mode->update_cpu_topology(); } int __node_distance(int a, int b) { return mode->distance ? mode->distance(a, b) : 0; } int numa_debug_enabled; /* * alloc_node_data() - Allocate node data */ static __init pg_data_t *alloc_node_data(void) { pg_data_t *res; res = (pg_data_t *) memblock_alloc(sizeof(pg_data_t), 8); memset(res, 0, sizeof(pg_data_t)); return res; } /* * numa_setup_memory() - Assign bootmem to nodes * * The memory is first added to memblock without any respect to nodes. * This is fixed before remaining memblock memory is handed over to the * buddy allocator. * An important side effect is that large bootmem allocations might easily * cross node boundaries, which can be needed for large allocations with * smaller memory stripes in each node (i.e. when using NUMA emulation). * * Memory defines nodes: * Therefore this routine also sets the nodes online with memory. */ static void __init numa_setup_memory(void) { unsigned long cur_base, align, end_of_dram; int nid = 0; end_of_dram = memblock_end_of_DRAM(); align = mode->align ? mode->align() : ULONG_MAX; /* * Step through all available memory and assign it to the nodes * indicated by the mode implementation. * All nodes which are seen here will be set online. */ cur_base = 0; do { nid = numa_pfn_to_nid(PFN_DOWN(cur_base)); node_set_online(nid); memblock_set_node(cur_base, align, &memblock.memory, nid); cur_base += align; } while (cur_base < end_of_dram); /* Allocate and fill out node_data */ for (nid = 0; nid < MAX_NUMNODES; nid++) NODE_DATA(nid) = alloc_node_data(); for_each_online_node(nid) { unsigned long start_pfn, end_pfn; unsigned long t_start, t_end; int i; start_pfn = ULONG_MAX; end_pfn = 0; for_each_mem_pfn_range(i, nid, &t_start, &t_end, NULL) { if (t_start < start_pfn) start_pfn = t_start; if (t_end > end_pfn) end_pfn = t_end; } NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; NODE_DATA(nid)->node_id = nid; } } /* * numa_setup() - Earliest initialization * * Assign the mode and call the mode's setup routine. */ void __init numa_setup(void) { pr_info("NUMA mode: %s\n", mode->name); nodes_clear(node_possible_map); if (mode->setup) mode->setup(); numa_setup_memory(); memblock_dump_all(); } /* * numa_init_early() - Initialization initcall * * This runs when only one CPU is online and before the first * topology update is called for by the scheduler. */ static int __init numa_init_early(void) { /* Attach all possible CPUs to node 0 for now. */ cpumask_copy(&node_to_cpumask_map[0], cpu_possible_mask); return 0; } early_initcall(numa_init_early); /* * numa_init_late() - Initialization initcall * * Register NUMA nodes. */ static int __init numa_init_late(void) { int nid; for_each_online_node(nid) register_one_node(nid); return 0; } arch_initcall(numa_init_late); static int __init parse_debug(char *parm) { numa_debug_enabled = 1; return 0; } early_param("numa_debug", parse_debug); static int __init parse_numa(char *parm) { if (strcmp(parm, numa_mode_plain.name) == 0) mode = &numa_mode_plain; #ifdef CONFIG_NUMA_EMU if (strcmp(parm, numa_mode_emu.name) == 0) mode = &numa_mode_emu; #endif return 0; } early_param("numa", parse_numa);