mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2025-01-13 01:08:50 +00:00
x86: change NR_CPUS arrays in numa_64 fixup
Change the following static arrays sized by NR_CPUS to per_cpu data variables: char cpu_to_node_map[NR_CPUS]; fixup: - Split cpu_to_node function into "early" and "late" versions so that x86_cpu_to_node_map_early_ptr is not EXPORT'ed and the cpu_to_node inline function is more streamlined. - This also involves setting up the percpu maps as early as possible. - Fix X86_32 NUMA build errors that previous version of this patch caused. V2->V3: - add early_cpu_to_node function to keep cpu_to_node efficient - move and rename smp_set_apicids() to setup_percpu_maps() - call setup_percpu_maps() as early as possible V1->V2: - Removed extraneous casts - Fix !NUMA builds with '#ifdef CONFIG_NUMA" Signed-off-by: Mike Travis <travis@sgi.com> Reviewed-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
This commit is contained in:
parent
ac72e7888a
commit
834beda15e
@ -83,6 +83,40 @@ static int __init nonx32_setup(char *str)
|
|||||||
}
|
}
|
||||||
__setup("noexec32=", nonx32_setup);
|
__setup("noexec32=", nonx32_setup);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Copy data used in early init routines from the initial arrays to the
|
||||||
|
* per cpu data areas. These arrays then become expendable and the *_ptrs
|
||||||
|
* are zeroed indicating that the static arrays are gone.
|
||||||
|
*/
|
||||||
|
void __init setup_percpu_maps(void)
|
||||||
|
{
|
||||||
|
int cpu;
|
||||||
|
|
||||||
|
for_each_possible_cpu(cpu) {
|
||||||
|
#ifdef CONFIG_SMP
|
||||||
|
if (per_cpu_offset(cpu)) {
|
||||||
|
#endif
|
||||||
|
per_cpu(x86_cpu_to_apicid, cpu) =
|
||||||
|
x86_cpu_to_apicid_init[cpu];
|
||||||
|
#ifdef CONFIG_NUMA
|
||||||
|
per_cpu(x86_cpu_to_node_map, cpu) =
|
||||||
|
x86_cpu_to_node_map_init[cpu];
|
||||||
|
#endif
|
||||||
|
#ifdef CONFIG_SMP
|
||||||
|
}
|
||||||
|
else
|
||||||
|
printk(KERN_NOTICE "per_cpu_offset zero for cpu %d\n",
|
||||||
|
cpu);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
/* indicate the early static arrays are gone */
|
||||||
|
x86_cpu_to_apicid_early_ptr = NULL;
|
||||||
|
#ifdef CONFIG_NUMA
|
||||||
|
x86_cpu_to_node_map_early_ptr = NULL;
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Great future plan:
|
* Great future plan:
|
||||||
* Declare PDA itself and support (irqstack,tss,pgd) as per cpu data.
|
* Declare PDA itself and support (irqstack,tss,pgd) as per cpu data.
|
||||||
@ -104,18 +138,21 @@ void __init setup_per_cpu_areas(void)
|
|||||||
for_each_cpu_mask (i, cpu_possible_map) {
|
for_each_cpu_mask (i, cpu_possible_map) {
|
||||||
char *ptr;
|
char *ptr;
|
||||||
|
|
||||||
if (!NODE_DATA(cpu_to_node(i))) {
|
if (!NODE_DATA(early_cpu_to_node(i))) {
|
||||||
printk("cpu with no node %d, num_online_nodes %d\n",
|
printk("cpu with no node %d, num_online_nodes %d\n",
|
||||||
i, num_online_nodes());
|
i, num_online_nodes());
|
||||||
ptr = alloc_bootmem_pages(size);
|
ptr = alloc_bootmem_pages(size);
|
||||||
} else {
|
} else {
|
||||||
ptr = alloc_bootmem_pages_node(NODE_DATA(cpu_to_node(i)), size);
|
ptr = alloc_bootmem_pages_node(NODE_DATA(early_cpu_to_node(i)), size);
|
||||||
}
|
}
|
||||||
if (!ptr)
|
if (!ptr)
|
||||||
panic("Cannot allocate cpu data for CPU %d\n", i);
|
panic("Cannot allocate cpu data for CPU %d\n", i);
|
||||||
cpu_pda(i)->data_offset = ptr - __per_cpu_start;
|
cpu_pda(i)->data_offset = ptr - __per_cpu_start;
|
||||||
memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start);
|
memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* setup percpu data maps early */
|
||||||
|
setup_percpu_maps();
|
||||||
}
|
}
|
||||||
|
|
||||||
void pda_init(int cpu)
|
void pda_init(int cpu)
|
||||||
|
@ -465,7 +465,7 @@ cpumask_t node_to_cpumask_map[MAX_NUMNODES] __read_mostly =
|
|||||||
{ [0 ... MAX_NUMNODES-1] = CPU_MASK_NONE };
|
{ [0 ... MAX_NUMNODES-1] = CPU_MASK_NONE };
|
||||||
EXPORT_SYMBOL(node_to_cpumask_map);
|
EXPORT_SYMBOL(node_to_cpumask_map);
|
||||||
/* which node each logical CPU is on */
|
/* which node each logical CPU is on */
|
||||||
int cpu_to_node_map[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = 0 };
|
u8 cpu_to_node_map[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = 0 };
|
||||||
EXPORT_SYMBOL(cpu_to_node_map);
|
EXPORT_SYMBOL(cpu_to_node_map);
|
||||||
|
|
||||||
/* set up a mapping between cpu and node. */
|
/* set up a mapping between cpu and node. */
|
||||||
|
@ -851,39 +851,6 @@ static int __init smp_sanity_check(unsigned max_cpus)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Copy data used in early init routines from the initial arrays to the
|
|
||||||
* per cpu data areas. These arrays then become expendable and the
|
|
||||||
* *_ptrs are zeroed indicating that the static arrays are gone.
|
|
||||||
*/
|
|
||||||
void __init smp_set_apicids(void)
|
|
||||||
{
|
|
||||||
int cpu;
|
|
||||||
|
|
||||||
for_each_possible_cpu(cpu) {
|
|
||||||
if (per_cpu_offset(cpu)) {
|
|
||||||
per_cpu(x86_cpu_to_apicid, cpu) =
|
|
||||||
x86_cpu_to_apicid_init[cpu];
|
|
||||||
#ifdef CONFIG_NUMA
|
|
||||||
per_cpu(x86_cpu_to_node_map, cpu) =
|
|
||||||
x86_cpu_to_node_map_init[cpu];
|
|
||||||
#endif
|
|
||||||
per_cpu(x86_bios_cpu_apicid, cpu) =
|
|
||||||
x86_bios_cpu_apicid_init[cpu];
|
|
||||||
}
|
|
||||||
else
|
|
||||||
printk(KERN_NOTICE "per_cpu_offset zero for cpu %d\n",
|
|
||||||
cpu);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* indicate the early static arrays are gone */
|
|
||||||
x86_cpu_to_apicid_early_ptr = NULL;
|
|
||||||
#ifdef CONFIG_NUMA
|
|
||||||
x86_cpu_to_node_map_early_ptr = NULL;
|
|
||||||
#endif
|
|
||||||
x86_bios_cpu_apicid_early_ptr = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void __init smp_cpu_index_default(void)
|
static void __init smp_cpu_index_default(void)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
@ -906,7 +873,6 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
|
|||||||
smp_cpu_index_default();
|
smp_cpu_index_default();
|
||||||
current_cpu_data = boot_cpu_data;
|
current_cpu_data = boot_cpu_data;
|
||||||
current_thread_info()->cpu = 0; /* needed? */
|
current_thread_info()->cpu = 0; /* needed? */
|
||||||
smp_set_apicids();
|
|
||||||
set_cpu_sibling_map(0);
|
set_cpu_sibling_map(0);
|
||||||
|
|
||||||
if (smp_sanity_check(max_cpus) < 0) {
|
if (smp_sanity_check(max_cpus) < 0) {
|
||||||
|
@ -382,9 +382,10 @@ int __init acpi_scan_nodes(unsigned long start, unsigned long end)
|
|||||||
setup_node_bootmem(i, nodes[i].start, nodes[i].end);
|
setup_node_bootmem(i, nodes[i].start, nodes[i].end);
|
||||||
|
|
||||||
for (i = 0; i < NR_CPUS; i++) {
|
for (i = 0; i < NR_CPUS; i++) {
|
||||||
if (cpu_to_node(i) == NUMA_NO_NODE)
|
int node = cpu_to_node(i);
|
||||||
|
if (node == NUMA_NO_NODE)
|
||||||
continue;
|
continue;
|
||||||
if (!node_isset(cpu_to_node(i), node_possible_map))
|
if (!node_isset(node, node_possible_map))
|
||||||
numa_set_node(i, NUMA_NO_NODE);
|
numa_set_node(i, NUMA_NO_NODE);
|
||||||
}
|
}
|
||||||
numa_init_array();
|
numa_init_array();
|
||||||
|
@ -30,15 +30,29 @@
|
|||||||
#include <asm/mpspec.h>
|
#include <asm/mpspec.h>
|
||||||
|
|
||||||
/* Mappings between logical cpu number and node number */
|
/* Mappings between logical cpu number and node number */
|
||||||
|
#ifdef CONFIG_X86_32
|
||||||
|
extern u8 cpu_to_node_map[];
|
||||||
|
|
||||||
|
#else
|
||||||
DECLARE_PER_CPU(u16, x86_cpu_to_node_map);
|
DECLARE_PER_CPU(u16, x86_cpu_to_node_map);
|
||||||
extern u16 x86_cpu_to_node_map_init[];
|
extern u16 x86_cpu_to_node_map_init[];
|
||||||
extern void *x86_cpu_to_node_map_early_ptr;
|
extern void *x86_cpu_to_node_map_early_ptr;
|
||||||
|
#endif
|
||||||
|
|
||||||
extern cpumask_t node_to_cpumask_map[];
|
extern cpumask_t node_to_cpumask_map[];
|
||||||
|
|
||||||
#define NUMA_NO_NODE ((u16)(~0))
|
#define NUMA_NO_NODE ((u16)(~0))
|
||||||
|
|
||||||
/* Returns the number of the node containing CPU 'cpu' */
|
/* Returns the number of the node containing CPU 'cpu' */
|
||||||
|
#ifdef CONFIG_X86_32
|
||||||
|
#define early_cpu_to_node(cpu) cpu_to_node(cpu)
|
||||||
static inline int cpu_to_node(int cpu)
|
static inline int cpu_to_node(int cpu)
|
||||||
|
{
|
||||||
|
return cpu_to_node_map[cpu];
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* CONFIG_X86_64 */
|
||||||
|
static inline int early_cpu_to_node(int cpu)
|
||||||
{
|
{
|
||||||
u16 *cpu_to_node_map = x86_cpu_to_node_map_early_ptr;
|
u16 *cpu_to_node_map = x86_cpu_to_node_map_early_ptr;
|
||||||
|
|
||||||
@ -50,6 +64,15 @@ static inline int cpu_to_node(int cpu)
|
|||||||
return NUMA_NO_NODE;
|
return NUMA_NO_NODE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline int cpu_to_node(int cpu)
|
||||||
|
{
|
||||||
|
if(per_cpu_offset(cpu))
|
||||||
|
return per_cpu(x86_cpu_to_node_map, cpu);
|
||||||
|
else
|
||||||
|
return NUMA_NO_NODE;
|
||||||
|
}
|
||||||
|
#endif /* CONFIG_X86_64 */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Returns the number of the node containing Node 'node'. This
|
* Returns the number of the node containing Node 'node'. This
|
||||||
* architecture is flat, so it is a pretty simple function!
|
* architecture is flat, so it is a pretty simple function!
|
||||||
|
Loading…
x
Reference in New Issue
Block a user