mirror of https://github.com/xqemu/xqemu.git
pc: get numa node mapping from possible_cpus instead of numa_get_node_for_cpu()
Signed-off-by: Igor Mammedov <imammedo@redhat.com> Reviewed-by: Andrew Jones <drjones@redhat.com> Message-Id: <1494415802-227633-9-git-send-email-imammedo@redhat.com> Signed-off-by: Eduardo Habkost <ehabkost@redhat.com>
This commit is contained in:
parent
af9b20e8d2
commit
ea2650724c
|
@ -503,7 +503,6 @@ void build_cpus_aml(Aml *table, MachineState *machine, CPUHotplugFeatures opts,
|
||||||
|
|
||||||
/* build Processor object for each processor */
|
/* build Processor object for each processor */
|
||||||
for (i = 0; i < arch_ids->len; i++) {
|
for (i = 0; i < arch_ids->len; i++) {
|
||||||
int j;
|
|
||||||
Aml *dev;
|
Aml *dev;
|
||||||
Aml *uid = aml_int(i);
|
Aml *uid = aml_int(i);
|
||||||
GArray *madt_buf = g_array_new(0, 1, 1);
|
GArray *madt_buf = g_array_new(0, 1, 1);
|
||||||
|
@ -557,9 +556,9 @@ void build_cpus_aml(Aml *table, MachineState *machine, CPUHotplugFeatures opts,
|
||||||
* as a result _PXM is required for all CPUs which might
|
* as a result _PXM is required for all CPUs which might
|
||||||
* be hot-plugged. For simplicity, add it for all CPUs.
|
* be hot-plugged. For simplicity, add it for all CPUs.
|
||||||
*/
|
*/
|
||||||
j = numa_get_node_for_cpu(i);
|
if (arch_ids->cpus[i].props.has_node_id) {
|
||||||
if (j < nb_numa_nodes) {
|
aml_append(dev, aml_name_decl("_PXM",
|
||||||
aml_append(dev, aml_name_decl("_PXM", aml_int(j)));
|
aml_int(arch_ids->cpus[i].props.node_id)));
|
||||||
}
|
}
|
||||||
|
|
||||||
aml_append(cpus_dev, dev);
|
aml_append(cpus_dev, dev);
|
||||||
|
|
|
@ -2335,7 +2335,8 @@ build_srat(GArray *table_data, BIOSLinker *linker, MachineState *machine)
|
||||||
srat->reserved1 = cpu_to_le32(1);
|
srat->reserved1 = cpu_to_le32(1);
|
||||||
|
|
||||||
for (i = 0; i < apic_ids->len; i++) {
|
for (i = 0; i < apic_ids->len; i++) {
|
||||||
int j = numa_get_node_for_cpu(i);
|
int node_id = apic_ids->cpus[i].props.has_node_id ?
|
||||||
|
apic_ids->cpus[i].props.node_id : 0;
|
||||||
uint32_t apic_id = apic_ids->cpus[i].arch_id;
|
uint32_t apic_id = apic_ids->cpus[i].arch_id;
|
||||||
|
|
||||||
if (apic_id < 255) {
|
if (apic_id < 255) {
|
||||||
|
@ -2345,9 +2346,7 @@ build_srat(GArray *table_data, BIOSLinker *linker, MachineState *machine)
|
||||||
core->type = ACPI_SRAT_PROCESSOR_APIC;
|
core->type = ACPI_SRAT_PROCESSOR_APIC;
|
||||||
core->length = sizeof(*core);
|
core->length = sizeof(*core);
|
||||||
core->local_apic_id = apic_id;
|
core->local_apic_id = apic_id;
|
||||||
if (j < nb_numa_nodes) {
|
core->proximity_lo = node_id;
|
||||||
core->proximity_lo = j;
|
|
||||||
}
|
|
||||||
memset(core->proximity_hi, 0, 3);
|
memset(core->proximity_hi, 0, 3);
|
||||||
core->local_sapic_eid = 0;
|
core->local_sapic_eid = 0;
|
||||||
core->flags = cpu_to_le32(1);
|
core->flags = cpu_to_le32(1);
|
||||||
|
@ -2358,9 +2357,7 @@ build_srat(GArray *table_data, BIOSLinker *linker, MachineState *machine)
|
||||||
core->type = ACPI_SRAT_PROCESSOR_x2APIC;
|
core->type = ACPI_SRAT_PROCESSOR_x2APIC;
|
||||||
core->length = sizeof(*core);
|
core->length = sizeof(*core);
|
||||||
core->x2apic_id = cpu_to_le32(apic_id);
|
core->x2apic_id = cpu_to_le32(apic_id);
|
||||||
if (j < nb_numa_nodes) {
|
core->proximity_domain = cpu_to_le32(node_id);
|
||||||
core->proximity_domain = cpu_to_le32(j);
|
|
||||||
}
|
|
||||||
core->flags = cpu_to_le32(1);
|
core->flags = cpu_to_le32(1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
18
hw/i386/pc.c
18
hw/i386/pc.c
|
@ -747,7 +747,9 @@ static FWCfgState *bochs_bios_init(AddressSpace *as, PCMachineState *pcms)
|
||||||
{
|
{
|
||||||
FWCfgState *fw_cfg;
|
FWCfgState *fw_cfg;
|
||||||
uint64_t *numa_fw_cfg;
|
uint64_t *numa_fw_cfg;
|
||||||
int i, j;
|
int i;
|
||||||
|
const CPUArchIdList *cpus;
|
||||||
|
MachineClass *mc = MACHINE_GET_CLASS(pcms);
|
||||||
|
|
||||||
fw_cfg = fw_cfg_init_io_dma(FW_CFG_IO_BASE, FW_CFG_IO_BASE + 4, as);
|
fw_cfg = fw_cfg_init_io_dma(FW_CFG_IO_BASE, FW_CFG_IO_BASE + 4, as);
|
||||||
fw_cfg_add_i16(fw_cfg, FW_CFG_NB_CPUS, pcms->boot_cpus);
|
fw_cfg_add_i16(fw_cfg, FW_CFG_NB_CPUS, pcms->boot_cpus);
|
||||||
|
@ -782,12 +784,12 @@ static FWCfgState *bochs_bios_init(AddressSpace *as, PCMachineState *pcms)
|
||||||
*/
|
*/
|
||||||
numa_fw_cfg = g_new0(uint64_t, 1 + pcms->apic_id_limit + nb_numa_nodes);
|
numa_fw_cfg = g_new0(uint64_t, 1 + pcms->apic_id_limit + nb_numa_nodes);
|
||||||
numa_fw_cfg[0] = cpu_to_le64(nb_numa_nodes);
|
numa_fw_cfg[0] = cpu_to_le64(nb_numa_nodes);
|
||||||
for (i = 0; i < max_cpus; i++) {
|
cpus = mc->possible_cpu_arch_ids(MACHINE(pcms));
|
||||||
unsigned int apic_id = x86_cpu_apic_id_from_index(i);
|
for (i = 0; i < cpus->len; i++) {
|
||||||
|
unsigned int apic_id = cpus->cpus[i].arch_id;
|
||||||
assert(apic_id < pcms->apic_id_limit);
|
assert(apic_id < pcms->apic_id_limit);
|
||||||
j = numa_get_node_for_cpu(i);
|
if (cpus->cpus[i].props.has_node_id) {
|
||||||
if (j < nb_numa_nodes) {
|
numa_fw_cfg[apic_id + 1] = cpu_to_le64(cpus->cpus[i].props.node_id);
|
||||||
numa_fw_cfg[apic_id + 1] = cpu_to_le64(j);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for (i = 0; i < nb_numa_nodes; i++) {
|
for (i = 0; i < nb_numa_nodes; i++) {
|
||||||
|
@ -1984,8 +1986,8 @@ static void pc_cpu_pre_plug(HotplugHandler *hotplug_dev,
|
||||||
cs = CPU(cpu);
|
cs = CPU(cpu);
|
||||||
cs->cpu_index = idx;
|
cs->cpu_index = idx;
|
||||||
|
|
||||||
node_id = numa_get_node_for_cpu(cs->cpu_index);
|
node_id = cpu_slot->props.node_id;
|
||||||
if (node_id == nb_numa_nodes) {
|
if (!cpu_slot->props.has_node_id) {
|
||||||
/* by default CPUState::numa_node was 0 if it's not set via CLI
|
/* by default CPUState::numa_node was 0 if it's not set via CLI
|
||||||
* keep it this way for now but in future we probably should
|
* keep it this way for now but in future we probably should
|
||||||
* refuse to start up with incomplete numa mapping */
|
* refuse to start up with incomplete numa mapping */
|
||||||
|
|
Loading…
Reference in New Issue