Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows
code reuse.
Signed-off-by: Salil Mehta <salil.mehta@huawei.com>
---
hw/arm/virt.c | 261 ++++++++++++++++++++++++++++--------------
include/hw/arm/virt.h | 4 +
2 files changed, 182 insertions(+), 83 deletions(-)
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
index 3e1c4d2d2f..2e0ec7d869 100644
--- a/hw/arm/virt.c
+++ b/hw/arm/virt.c
@@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx)
return arm_build_mp_affinity(idx, clustersz);
}
+static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid)
+{
+ VirtMachineState *vms = VIRT_MACHINE(ms);
+ CPUArchId *found_cpu;
+ uint64_t mp_affinity;
+
+ assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len);
+
+ mp_affinity = virt_cpu_mp_affinity(vms, vcpuid);
+ found_cpu = &ms->possible_cpus->cpus[vcpuid];
+
+ assert(found_cpu->arch_id == mp_affinity);
+
+ /*
+ * RFC: Question:
+ * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity)
+ * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id.
+ * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is
+ * more related to machine? Current code assumes slot-id and vcpu-id are
+ * same i.e. meaning of slot is bit vague.
+ *
+ * Q1: Is there any requirement to clearly represent slot and dissociate it
+ * from vcpu-id?
+ * Q2: Should we make MPIDR within host KVM user configurable?
+ *
+ * +----+----+----+----+----+----+----+----+
+ * MPIDR ||| Res | Aff2 | Aff1 | Aff0 |
+ * +----+----+----+----+----+----+----+----+
+ * \ \ \ | |
+ * \ 8bit \ 8bit \ |4bit|
+ * \<------->\<------->\ |<-->|
+ * \ \ \| |
+ * +----+----+----+----+----+----+----+----+
+ * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 |
+ * +----+----+----+----+----+----+----+----+
+ */
+
+ return found_cpu;
+}
+
static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms,
int index)
{
@@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem)
}
}
+static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot,
+ Error **errp)
+{
+ MachineState *ms = MACHINE(qdev_get_machine());
+ VirtMachineState *vms = VIRT_MACHINE(ms);
+ Error *local_err = NULL;
+ VirtMachineClass *vmc;
+
+ vmc = VIRT_MACHINE_GET_CLASS(ms);
+
+ /* now, set the cpu object property values */
+ numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL);
+
+ if (!vms->secure) {
+ object_property_set_bool(cpuobj, "has_el3", false, NULL);
+ }
+
+ if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
+ object_property_set_bool(cpuobj, "has_el2", false, NULL);
+ }
+
+ if (vmc->kvm_no_adjvtime &&
+ object_property_find(cpuobj, "kvm-no-adjvtime")) {
+ object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
+ }
+
+ if (vmc->no_kvm_steal_time &&
+ object_property_find(cpuobj, "kvm-steal-time")) {
+ object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
+ }
+
+ if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
+ object_property_set_bool(cpuobj, "pmu", false, NULL);
+ }
+
+ if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
+ object_property_set_bool(cpuobj, "lpa2", false, NULL);
+ }
+
+ if (object_property_find(cpuobj, "reset-cbar")) {
+ object_property_set_int(cpuobj, "reset-cbar",
+ vms->memmap[VIRT_CPUPERIPHS].base,
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ /* link already initialized {secure,tag}-memory regions to this cpu */
+ object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ if (vms->secure) {
+ object_property_set_link(cpuobj, "secure-memory",
+ OBJECT(vms->secure_sysmem), &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ if (vms->mte) {
+ if (!object_property_find(cpuobj, "tag-memory")) {
+ error_setg(&local_err, "MTE requested, but not supported "
+ "by the guest CPU");
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem),
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ if (vms->secure) {
+ object_property_set_link(cpuobj, "secure-tag-memory",
+ OBJECT(vms->secure_tag_sysmem),
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+ }
+
+ /*
+ * RFC: Question: this must only be called for the hotplugged cpus. For the
+ * cold booted secondary cpus this is being taken care in arm_load_kernel()
+ * in boot.c. Perhaps we should remove that code now?
+ */
+ if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) {
+ object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit,
+ NULL);
+
+ /* Secondary CPUs start in PSCI powered-down state */
+ if (CPU(cpuobj)->cpu_index > 0) {
+ object_property_set_bool(cpuobj, "start-powered-off", true, NULL);
+ }
+ }
+
+out:
+ if (local_err) {
+ error_propagate(errp, local_err);
+ }
+}
+
static void machvirt_init(MachineState *machine)
{
VirtMachineState *vms = VIRT_MACHINE(machine);
VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine);
MachineClass *mc = MACHINE_GET_CLASS(machine);
const CPUArchIdList *possible_cpus;
- MemoryRegion *sysmem = get_system_memory();
+ MemoryRegion *secure_tag_sysmem = NULL;
MemoryRegion *secure_sysmem = NULL;
MemoryRegion *tag_sysmem = NULL;
- MemoryRegion *secure_tag_sysmem = NULL;
+ MemoryRegion *sysmem;
int n, virt_max_cpus;
bool firmware_loaded;
bool aarch64 = true;
@@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine)
/* uses smp.max_cpus to initialize all possible vCPUs */
possible_cpus = mc->possible_cpu_arch_ids(machine);
+ sysmem = vms->sysmem = get_system_memory();
+
if (vms->secure) {
/*
* The Secure view of the world is the same as the NonSecure,
@@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine)
* containing the system memory at low priority; any secure-only
* devices go in at higher priority and take precedence.
*/
- secure_sysmem = g_new(MemoryRegion, 1);
+ secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1);
memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory",
UINT64_MAX);
memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1);
@@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine)
exit(1);
}
+ if (vms->mte) {
+ /* Create the memory region only once, but link to all cpus later */
+ tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1);
+ memory_region_init(tag_sysmem, OBJECT(machine),
+ "tag-memory", UINT64_MAX / 32);
+
+ if (vms->secure) {
+ secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1);
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
+ "secure-tag-memory", UINT64_MAX / 32);
+
+ /* As with ram, secure-tag takes precedence over tag. */
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
+ tag_sysmem, -1);
+ }
+ }
+
create_fdt(vms);
assert(possible_cpus->len == max_cpus);
for (n = 0; n < possible_cpus->len; n++) {
+ CPUArchId *cpu_slot;
Object *cpuobj;
CPUState *cs;
@@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine)
}
cpuobj = object_new(possible_cpus->cpus[n].type);
- object_property_set_int(cpuobj, "mp-affinity",
- possible_cpus->cpus[n].arch_id, NULL);
cs = CPU(cpuobj);
cs->cpu_index = n;
- numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj),
- &error_fatal);
-
aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL);
object_property_set_int(cpuobj, "socket-id",
virt_get_socket_id(machine, n), NULL);
@@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine)
object_property_set_int(cpuobj, "thread-id",
virt_get_thread_id(machine, n), NULL);
- if (!vms->secure) {
- object_property_set_bool(cpuobj, "has_el3", false, NULL);
- }
-
- if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
- object_property_set_bool(cpuobj, "has_el2", false, NULL);
- }
-
- if (vmc->kvm_no_adjvtime &&
- object_property_find(cpuobj, "kvm-no-adjvtime")) {
- object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
- }
-
- if (vmc->no_kvm_steal_time &&
- object_property_find(cpuobj, "kvm-steal-time")) {
- object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
- }
-
- if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
- object_property_set_bool(cpuobj, "pmu", false, NULL);
- }
-
- if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
- object_property_set_bool(cpuobj, "lpa2", false, NULL);
- }
-
- if (object_property_find(cpuobj, "reset-cbar")) {
- object_property_set_int(cpuobj, "reset-cbar",
- vms->memmap[VIRT_CPUPERIPHS].base,
- &error_abort);
- }
-
- object_property_set_link(cpuobj, "memory", OBJECT(sysmem),
- &error_abort);
- if (vms->secure) {
- object_property_set_link(cpuobj, "secure-memory",
- OBJECT(secure_sysmem), &error_abort);
- }
-
- if (vms->mte) {
- /* Create the memory region only once, but link to all cpus. */
- if (!tag_sysmem) {
- /*
- * The property exists only if MemTag is supported.
- * If it is, we must allocate the ram to back that up.
- */
- if (!object_property_find(cpuobj, "tag-memory")) {
- error_report("MTE requested, but not supported "
- "by the guest CPU");
- exit(1);
- }
-
- tag_sysmem = g_new(MemoryRegion, 1);
- memory_region_init(tag_sysmem, OBJECT(machine),
- "tag-memory", UINT64_MAX / 32);
-
- if (vms->secure) {
- secure_tag_sysmem = g_new(MemoryRegion, 1);
- memory_region_init(secure_tag_sysmem, OBJECT(machine),
- "secure-tag-memory", UINT64_MAX / 32);
-
- /* As with ram, secure-tag takes precedence over tag. */
- memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
- tag_sysmem, -1);
- }
- }
-
- object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem),
- &error_abort);
- if (vms->secure) {
- object_property_set_link(cpuobj, "secure-tag-memory",
- OBJECT(secure_tag_sysmem),
- &error_abort);
- }
- }
+ cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index);
+ virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal);
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
object_unref(cpuobj);
diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h
index 6f9a7bb60b..780bd53ceb 100644
--- a/include/hw/arm/virt.h
+++ b/include/hw/arm/virt.h
@@ -139,6 +139,10 @@ struct VirtMachineState {
DeviceState *platform_bus_dev;
FWCfgState *fw_cfg;
PFlashCFI01 *flash[2];
+ MemoryRegion *sysmem;
+ MemoryRegion *secure_sysmem;
+ MemoryRegion *tag_sysmem;
+ MemoryRegion *secure_tag_sysmem;
bool secure;
bool highmem;
bool highmem_compact;
--
2.34.1
On 6/14/24 9:36 AM, Salil Mehta wrote: > Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows > code reuse. > > Signed-off-by: Salil Mehta <salil.mehta@huawei.com> > --- > hw/arm/virt.c | 261 ++++++++++++++++++++++++++++-------------- > include/hw/arm/virt.h | 4 + > 2 files changed, 182 insertions(+), 83 deletions(-) > If this series are going to be split for easier review, this is an candidate for a prepatory patch. > diff --git a/hw/arm/virt.c b/hw/arm/virt.c > index 3e1c4d2d2f..2e0ec7d869 100644 > --- a/hw/arm/virt.c > +++ b/hw/arm/virt.c > @@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx) > return arm_build_mp_affinity(idx, clustersz); > } > > +static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid) > +{ > + VirtMachineState *vms = VIRT_MACHINE(ms); > + CPUArchId *found_cpu; > + uint64_t mp_affinity; > + > + assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len); > + > + mp_affinity = virt_cpu_mp_affinity(vms, vcpuid); > + found_cpu = &ms->possible_cpus->cpus[vcpuid]; > + > + assert(found_cpu->arch_id == mp_affinity); > + > + /* > + * RFC: Question: > + * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity) > + * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id. > + * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is > + * more related to machine? Current code assumes slot-id and vcpu-id are > + * same i.e. meaning of slot is bit vague. > + * > + * Q1: Is there any requirement to clearly represent slot and dissociate it > + * from vcpu-id? > + * Q2: Should we make MPIDR within host KVM user configurable? > + * > + * +----+----+----+----+----+----+----+----+ > + * MPIDR ||| Res | Aff2 | Aff1 | Aff0 | > + * +----+----+----+----+----+----+----+----+ > + * \ \ \ | | > + * \ 8bit \ 8bit \ |4bit| > + * \<------->\<------->\ |<-->| > + * \ \ \| | > + * +----+----+----+----+----+----+----+----+ > + * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 | > + * +----+----+----+----+----+----+----+----+ > + */ > + > + return found_cpu; > +} > + I don't see why virt_find_cpu_slot() is needed. Apart from the sanity check, it basically returns ms->possible_cpus->cpus[]. The caller can dereference ms->possible_cpus->cpus[] directly. About the sanity check, the mp_affinity has been properly populated in virt_possible_cpu_arch_ids(). I don't see why we it needs to be checked again. > static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms, > int index) > { > @@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem) > } > } > > +static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot, > + Error **errp) > +{ > + MachineState *ms = MACHINE(qdev_get_machine()); > + VirtMachineState *vms = VIRT_MACHINE(ms); > + Error *local_err = NULL; > + VirtMachineClass *vmc; > + > + vmc = VIRT_MACHINE_GET_CLASS(ms); > + > + /* now, set the cpu object property values */ > + numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err); > + if (local_err) { > + goto out; > + } > + > + object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL); > + > + if (!vms->secure) { > + object_property_set_bool(cpuobj, "has_el3", false, NULL); > + } > + > + if (!vms->virt && object_property_find(cpuobj, "has_el2")) { > + object_property_set_bool(cpuobj, "has_el2", false, NULL); > + } > + > + if (vmc->kvm_no_adjvtime && > + object_property_find(cpuobj, "kvm-no-adjvtime")) { > + object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); > + } > + > + if (vmc->no_kvm_steal_time && > + object_property_find(cpuobj, "kvm-steal-time")) { > + object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); > + } > + > + if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { > + object_property_set_bool(cpuobj, "pmu", false, NULL); > + } > + > + if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { > + object_property_set_bool(cpuobj, "lpa2", false, NULL); > + } > + > + if (object_property_find(cpuobj, "reset-cbar")) { > + object_property_set_int(cpuobj, "reset-cbar", > + vms->memmap[VIRT_CPUPERIPHS].base, > + &local_err); > + if (local_err) { > + goto out; > + } > + } > + > + /* link already initialized {secure,tag}-memory regions to this cpu */ > + object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err); > + if (local_err) { > + goto out; > + } > + > + if (vms->secure) { > + object_property_set_link(cpuobj, "secure-memory", > + OBJECT(vms->secure_sysmem), &local_err); > + if (local_err) { > + goto out; > + } > + } > + > + if (vms->mte) { > + if (!object_property_find(cpuobj, "tag-memory")) { > + error_setg(&local_err, "MTE requested, but not supported " > + "by the guest CPU"); > + if (local_err) { > + goto out; > + } > + } > + > + object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem), > + &local_err); > + if (local_err) { > + goto out; > + } > + > + if (vms->secure) { > + object_property_set_link(cpuobj, "secure-tag-memory", > + OBJECT(vms->secure_tag_sysmem), > + &local_err); > + if (local_err) { > + goto out; > + } > + } > + } > + > + /* > + * RFC: Question: this must only be called for the hotplugged cpus. For the > + * cold booted secondary cpus this is being taken care in arm_load_kernel() > + * in boot.c. Perhaps we should remove that code now? > + */ > + if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) { > + object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit, > + NULL); > + > + /* Secondary CPUs start in PSCI powered-down state */ > + if (CPU(cpuobj)->cpu_index > 0) { > + object_property_set_bool(cpuobj, "start-powered-off", true, NULL); > + } > + } > + > +out: > + if (local_err) { > + error_propagate(errp, local_err); > + } > +} > + > static void machvirt_init(MachineState *machine) > { > VirtMachineState *vms = VIRT_MACHINE(machine); > VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine); > MachineClass *mc = MACHINE_GET_CLASS(machine); > const CPUArchIdList *possible_cpus; > - MemoryRegion *sysmem = get_system_memory(); > + MemoryRegion *secure_tag_sysmem = NULL; > MemoryRegion *secure_sysmem = NULL; > MemoryRegion *tag_sysmem = NULL; > - MemoryRegion *secure_tag_sysmem = NULL; > + MemoryRegion *sysmem; > int n, virt_max_cpus; > bool firmware_loaded; > bool aarch64 = true; > @@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine) > /* uses smp.max_cpus to initialize all possible vCPUs */ > possible_cpus = mc->possible_cpu_arch_ids(machine); > > + sysmem = vms->sysmem = get_system_memory(); > + > if (vms->secure) { > /* > * The Secure view of the world is the same as the NonSecure, > @@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine) > * containing the system memory at low priority; any secure-only > * devices go in at higher priority and take precedence. > */ > - secure_sysmem = g_new(MemoryRegion, 1); > + secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1); > memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory", > UINT64_MAX); > memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1); > @@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine) > exit(1); > } > > + if (vms->mte) { > + /* Create the memory region only once, but link to all cpus later */ > + tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1); > + memory_region_init(tag_sysmem, OBJECT(machine), > + "tag-memory", UINT64_MAX / 32); > + > + if (vms->secure) { > + secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1); > + memory_region_init(secure_tag_sysmem, OBJECT(machine), > + "secure-tag-memory", UINT64_MAX / 32); > + > + /* As with ram, secure-tag takes precedence over tag. */ > + memory_region_add_subregion_overlap(secure_tag_sysmem, 0, > + tag_sysmem, -1); > + } > + } > + > create_fdt(vms); > > assert(possible_cpus->len == max_cpus); > for (n = 0; n < possible_cpus->len; n++) { > + CPUArchId *cpu_slot; > Object *cpuobj; > CPUState *cs; > > @@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine) > } > > cpuobj = object_new(possible_cpus->cpus[n].type); > - object_property_set_int(cpuobj, "mp-affinity", > - possible_cpus->cpus[n].arch_id, NULL); > > cs = CPU(cpuobj); > cs->cpu_index = n; > > - numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj), > - &error_fatal); > - > aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL); > object_property_set_int(cpuobj, "socket-id", > virt_get_socket_id(machine, n), NULL); > @@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine) > object_property_set_int(cpuobj, "thread-id", > virt_get_thread_id(machine, n), NULL); > > - if (!vms->secure) { > - object_property_set_bool(cpuobj, "has_el3", false, NULL); > - } > - > - if (!vms->virt && object_property_find(cpuobj, "has_el2")) { > - object_property_set_bool(cpuobj, "has_el2", false, NULL); > - } > - > - if (vmc->kvm_no_adjvtime && > - object_property_find(cpuobj, "kvm-no-adjvtime")) { > - object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); > - } > - > - if (vmc->no_kvm_steal_time && > - object_property_find(cpuobj, "kvm-steal-time")) { > - object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); > - } > - > - if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { > - object_property_set_bool(cpuobj, "pmu", false, NULL); > - } > - > - if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { > - object_property_set_bool(cpuobj, "lpa2", false, NULL); > - } > - > - if (object_property_find(cpuobj, "reset-cbar")) { > - object_property_set_int(cpuobj, "reset-cbar", > - vms->memmap[VIRT_CPUPERIPHS].base, > - &error_abort); > - } > - > - object_property_set_link(cpuobj, "memory", OBJECT(sysmem), > - &error_abort); > - if (vms->secure) { > - object_property_set_link(cpuobj, "secure-memory", > - OBJECT(secure_sysmem), &error_abort); > - } > - > - if (vms->mte) { > - /* Create the memory region only once, but link to all cpus. */ > - if (!tag_sysmem) { > - /* > - * The property exists only if MemTag is supported. > - * If it is, we must allocate the ram to back that up. > - */ > - if (!object_property_find(cpuobj, "tag-memory")) { > - error_report("MTE requested, but not supported " > - "by the guest CPU"); > - exit(1); > - } > - > - tag_sysmem = g_new(MemoryRegion, 1); > - memory_region_init(tag_sysmem, OBJECT(machine), > - "tag-memory", UINT64_MAX / 32); > - > - if (vms->secure) { > - secure_tag_sysmem = g_new(MemoryRegion, 1); > - memory_region_init(secure_tag_sysmem, OBJECT(machine), > - "secure-tag-memory", UINT64_MAX / 32); > - > - /* As with ram, secure-tag takes precedence over tag. */ > - memory_region_add_subregion_overlap(secure_tag_sysmem, 0, > - tag_sysmem, -1); > - } > - } > - > - object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem), > - &error_abort); > - if (vms->secure) { > - object_property_set_link(cpuobj, "secure-tag-memory", > - OBJECT(secure_tag_sysmem), > - &error_abort); > - } > - } > + cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index); > + virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal); > > qdev_realize(DEVICE(cpuobj), NULL, &error_fatal); > object_unref(cpuobj); > diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h > index 6f9a7bb60b..780bd53ceb 100644 > --- a/include/hw/arm/virt.h > +++ b/include/hw/arm/virt.h > @@ -139,6 +139,10 @@ struct VirtMachineState { > DeviceState *platform_bus_dev; > FWCfgState *fw_cfg; > PFlashCFI01 *flash[2]; > + MemoryRegion *sysmem; > + MemoryRegion *secure_sysmem; > + MemoryRegion *tag_sysmem; > + MemoryRegion *secure_tag_sysmem; > bool secure; > bool highmem; > bool highmem_compact; Thanks, Gavin
© 2016 - 2024 Red Hat, Inc.