Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows
code reuse.
Signed-off-by: Salil Mehta <salil.mehta@huawei.com>
---
hw/arm/virt.c | 261 ++++++++++++++++++++++++++++--------------
include/hw/arm/virt.h | 4 +
2 files changed, 182 insertions(+), 83 deletions(-)
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
index 3e1c4d2d2f..2e0ec7d869 100644
--- a/hw/arm/virt.c
+++ b/hw/arm/virt.c
@@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx)
return arm_build_mp_affinity(idx, clustersz);
}
+static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid)
+{
+ VirtMachineState *vms = VIRT_MACHINE(ms);
+ CPUArchId *found_cpu;
+ uint64_t mp_affinity;
+
+ assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len);
+
+ mp_affinity = virt_cpu_mp_affinity(vms, vcpuid);
+ found_cpu = &ms->possible_cpus->cpus[vcpuid];
+
+ assert(found_cpu->arch_id == mp_affinity);
+
+ /*
+ * RFC: Question:
+ * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity)
+ * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id.
+ * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is
+ * more related to machine? Current code assumes slot-id and vcpu-id are
+ * same i.e. meaning of slot is bit vague.
+ *
+ * Q1: Is there any requirement to clearly represent slot and dissociate it
+ * from vcpu-id?
+ * Q2: Should we make MPIDR within host KVM user configurable?
+ *
+ * +----+----+----+----+----+----+----+----+
+ * MPIDR ||| Res | Aff2 | Aff1 | Aff0 |
+ * +----+----+----+----+----+----+----+----+
+ * \ \ \ | |
+ * \ 8bit \ 8bit \ |4bit|
+ * \<------->\<------->\ |<-->|
+ * \ \ \| |
+ * +----+----+----+----+----+----+----+----+
+ * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 |
+ * +----+----+----+----+----+----+----+----+
+ */
+
+ return found_cpu;
+}
+
static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms,
int index)
{
@@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem)
}
}
+static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot,
+ Error **errp)
+{
+ MachineState *ms = MACHINE(qdev_get_machine());
+ VirtMachineState *vms = VIRT_MACHINE(ms);
+ Error *local_err = NULL;
+ VirtMachineClass *vmc;
+
+ vmc = VIRT_MACHINE_GET_CLASS(ms);
+
+ /* now, set the cpu object property values */
+ numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL);
+
+ if (!vms->secure) {
+ object_property_set_bool(cpuobj, "has_el3", false, NULL);
+ }
+
+ if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
+ object_property_set_bool(cpuobj, "has_el2", false, NULL);
+ }
+
+ if (vmc->kvm_no_adjvtime &&
+ object_property_find(cpuobj, "kvm-no-adjvtime")) {
+ object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
+ }
+
+ if (vmc->no_kvm_steal_time &&
+ object_property_find(cpuobj, "kvm-steal-time")) {
+ object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
+ }
+
+ if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
+ object_property_set_bool(cpuobj, "pmu", false, NULL);
+ }
+
+ if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
+ object_property_set_bool(cpuobj, "lpa2", false, NULL);
+ }
+
+ if (object_property_find(cpuobj, "reset-cbar")) {
+ object_property_set_int(cpuobj, "reset-cbar",
+ vms->memmap[VIRT_CPUPERIPHS].base,
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ /* link already initialized {secure,tag}-memory regions to this cpu */
+ object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ if (vms->secure) {
+ object_property_set_link(cpuobj, "secure-memory",
+ OBJECT(vms->secure_sysmem), &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ if (vms->mte) {
+ if (!object_property_find(cpuobj, "tag-memory")) {
+ error_setg(&local_err, "MTE requested, but not supported "
+ "by the guest CPU");
+ if (local_err) {
+ goto out;
+ }
+ }
+
+ object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem),
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+
+ if (vms->secure) {
+ object_property_set_link(cpuobj, "secure-tag-memory",
+ OBJECT(vms->secure_tag_sysmem),
+ &local_err);
+ if (local_err) {
+ goto out;
+ }
+ }
+ }
+
+ /*
+ * RFC: Question: this must only be called for the hotplugged cpus. For the
+ * cold booted secondary cpus this is being taken care in arm_load_kernel()
+ * in boot.c. Perhaps we should remove that code now?
+ */
+ if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) {
+ object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit,
+ NULL);
+
+ /* Secondary CPUs start in PSCI powered-down state */
+ if (CPU(cpuobj)->cpu_index > 0) {
+ object_property_set_bool(cpuobj, "start-powered-off", true, NULL);
+ }
+ }
+
+out:
+ if (local_err) {
+ error_propagate(errp, local_err);
+ }
+}
+
static void machvirt_init(MachineState *machine)
{
VirtMachineState *vms = VIRT_MACHINE(machine);
VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine);
MachineClass *mc = MACHINE_GET_CLASS(machine);
const CPUArchIdList *possible_cpus;
- MemoryRegion *sysmem = get_system_memory();
+ MemoryRegion *secure_tag_sysmem = NULL;
MemoryRegion *secure_sysmem = NULL;
MemoryRegion *tag_sysmem = NULL;
- MemoryRegion *secure_tag_sysmem = NULL;
+ MemoryRegion *sysmem;
int n, virt_max_cpus;
bool firmware_loaded;
bool aarch64 = true;
@@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine)
/* uses smp.max_cpus to initialize all possible vCPUs */
possible_cpus = mc->possible_cpu_arch_ids(machine);
+ sysmem = vms->sysmem = get_system_memory();
+
if (vms->secure) {
/*
* The Secure view of the world is the same as the NonSecure,
@@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine)
* containing the system memory at low priority; any secure-only
* devices go in at higher priority and take precedence.
*/
- secure_sysmem = g_new(MemoryRegion, 1);
+ secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1);
memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory",
UINT64_MAX);
memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1);
@@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine)
exit(1);
}
+ if (vms->mte) {
+ /* Create the memory region only once, but link to all cpus later */
+ tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1);
+ memory_region_init(tag_sysmem, OBJECT(machine),
+ "tag-memory", UINT64_MAX / 32);
+
+ if (vms->secure) {
+ secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1);
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
+ "secure-tag-memory", UINT64_MAX / 32);
+
+ /* As with ram, secure-tag takes precedence over tag. */
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
+ tag_sysmem, -1);
+ }
+ }
+
create_fdt(vms);
assert(possible_cpus->len == max_cpus);
for (n = 0; n < possible_cpus->len; n++) {
+ CPUArchId *cpu_slot;
Object *cpuobj;
CPUState *cs;
@@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine)
}
cpuobj = object_new(possible_cpus->cpus[n].type);
- object_property_set_int(cpuobj, "mp-affinity",
- possible_cpus->cpus[n].arch_id, NULL);
cs = CPU(cpuobj);
cs->cpu_index = n;
- numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj),
- &error_fatal);
-
aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL);
object_property_set_int(cpuobj, "socket-id",
virt_get_socket_id(machine, n), NULL);
@@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine)
object_property_set_int(cpuobj, "thread-id",
virt_get_thread_id(machine, n), NULL);
- if (!vms->secure) {
- object_property_set_bool(cpuobj, "has_el3", false, NULL);
- }
-
- if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
- object_property_set_bool(cpuobj, "has_el2", false, NULL);
- }
-
- if (vmc->kvm_no_adjvtime &&
- object_property_find(cpuobj, "kvm-no-adjvtime")) {
- object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
- }
-
- if (vmc->no_kvm_steal_time &&
- object_property_find(cpuobj, "kvm-steal-time")) {
- object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
- }
-
- if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
- object_property_set_bool(cpuobj, "pmu", false, NULL);
- }
-
- if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
- object_property_set_bool(cpuobj, "lpa2", false, NULL);
- }
-
- if (object_property_find(cpuobj, "reset-cbar")) {
- object_property_set_int(cpuobj, "reset-cbar",
- vms->memmap[VIRT_CPUPERIPHS].base,
- &error_abort);
- }
-
- object_property_set_link(cpuobj, "memory", OBJECT(sysmem),
- &error_abort);
- if (vms->secure) {
- object_property_set_link(cpuobj, "secure-memory",
- OBJECT(secure_sysmem), &error_abort);
- }
-
- if (vms->mte) {
- /* Create the memory region only once, but link to all cpus. */
- if (!tag_sysmem) {
- /*
- * The property exists only if MemTag is supported.
- * If it is, we must allocate the ram to back that up.
- */
- if (!object_property_find(cpuobj, "tag-memory")) {
- error_report("MTE requested, but not supported "
- "by the guest CPU");
- exit(1);
- }
-
- tag_sysmem = g_new(MemoryRegion, 1);
- memory_region_init(tag_sysmem, OBJECT(machine),
- "tag-memory", UINT64_MAX / 32);
-
- if (vms->secure) {
- secure_tag_sysmem = g_new(MemoryRegion, 1);
- memory_region_init(secure_tag_sysmem, OBJECT(machine),
- "secure-tag-memory", UINT64_MAX / 32);
-
- /* As with ram, secure-tag takes precedence over tag. */
- memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
- tag_sysmem, -1);
- }
- }
-
- object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem),
- &error_abort);
- if (vms->secure) {
- object_property_set_link(cpuobj, "secure-tag-memory",
- OBJECT(secure_tag_sysmem),
- &error_abort);
- }
- }
+ cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index);
+ virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal);
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
object_unref(cpuobj);
diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h
index 6f9a7bb60b..780bd53ceb 100644
--- a/include/hw/arm/virt.h
+++ b/include/hw/arm/virt.h
@@ -139,6 +139,10 @@ struct VirtMachineState {
DeviceState *platform_bus_dev;
FWCfgState *fw_cfg;
PFlashCFI01 *flash[2];
+ MemoryRegion *sysmem;
+ MemoryRegion *secure_sysmem;
+ MemoryRegion *tag_sysmem;
+ MemoryRegion *secure_tag_sysmem;
bool secure;
bool highmem;
bool highmem_compact;
--
2.34.1
On 6/14/24 9:36 AM, Salil Mehta wrote:
> Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows
> code reuse.
>
> Signed-off-by: Salil Mehta <salil.mehta@huawei.com>
> ---
> hw/arm/virt.c | 261 ++++++++++++++++++++++++++++--------------
> include/hw/arm/virt.h | 4 +
> 2 files changed, 182 insertions(+), 83 deletions(-)
>
If this series are going to be split for easier review, this is an candidate
for a prepatory patch.
> diff --git a/hw/arm/virt.c b/hw/arm/virt.c
> index 3e1c4d2d2f..2e0ec7d869 100644
> --- a/hw/arm/virt.c
> +++ b/hw/arm/virt.c
> @@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx)
> return arm_build_mp_affinity(idx, clustersz);
> }
>
> +static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid)
> +{
> + VirtMachineState *vms = VIRT_MACHINE(ms);
> + CPUArchId *found_cpu;
> + uint64_t mp_affinity;
> +
> + assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len);
> +
> + mp_affinity = virt_cpu_mp_affinity(vms, vcpuid);
> + found_cpu = &ms->possible_cpus->cpus[vcpuid];
> +
> + assert(found_cpu->arch_id == mp_affinity);
> +
> + /*
> + * RFC: Question:
> + * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity)
> + * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id.
> + * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is
> + * more related to machine? Current code assumes slot-id and vcpu-id are
> + * same i.e. meaning of slot is bit vague.
> + *
> + * Q1: Is there any requirement to clearly represent slot and dissociate it
> + * from vcpu-id?
> + * Q2: Should we make MPIDR within host KVM user configurable?
> + *
> + * +----+----+----+----+----+----+----+----+
> + * MPIDR ||| Res | Aff2 | Aff1 | Aff0 |
> + * +----+----+----+----+----+----+----+----+
> + * \ \ \ | |
> + * \ 8bit \ 8bit \ |4bit|
> + * \<------->\<------->\ |<-->|
> + * \ \ \| |
> + * +----+----+----+----+----+----+----+----+
> + * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 |
> + * +----+----+----+----+----+----+----+----+
> + */
> +
> + return found_cpu;
> +}
> +
I don't see why virt_find_cpu_slot() is needed. Apart from the sanity check, it
basically returns ms->possible_cpus->cpus[]. The caller can dereference
ms->possible_cpus->cpus[] directly. About the sanity check, the mp_affinity
has been properly populated in virt_possible_cpu_arch_ids(). I don't see why
we it needs to be checked again.
> static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms,
> int index)
> {
> @@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem)
> }
> }
>
> +static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot,
> + Error **errp)
> +{
> + MachineState *ms = MACHINE(qdev_get_machine());
> + VirtMachineState *vms = VIRT_MACHINE(ms);
> + Error *local_err = NULL;
> + VirtMachineClass *vmc;
> +
> + vmc = VIRT_MACHINE_GET_CLASS(ms);
> +
> + /* now, set the cpu object property values */
> + numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err);
> + if (local_err) {
> + goto out;
> + }
> +
> + object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL);
> +
> + if (!vms->secure) {
> + object_property_set_bool(cpuobj, "has_el3", false, NULL);
> + }
> +
> + if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
> + object_property_set_bool(cpuobj, "has_el2", false, NULL);
> + }
> +
> + if (vmc->kvm_no_adjvtime &&
> + object_property_find(cpuobj, "kvm-no-adjvtime")) {
> + object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
> + }
> +
> + if (vmc->no_kvm_steal_time &&
> + object_property_find(cpuobj, "kvm-steal-time")) {
> + object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
> + }
> +
> + if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
> + object_property_set_bool(cpuobj, "pmu", false, NULL);
> + }
> +
> + if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
> + object_property_set_bool(cpuobj, "lpa2", false, NULL);
> + }
> +
> + if (object_property_find(cpuobj, "reset-cbar")) {
> + object_property_set_int(cpuobj, "reset-cbar",
> + vms->memmap[VIRT_CPUPERIPHS].base,
> + &local_err);
> + if (local_err) {
> + goto out;
> + }
> + }
> +
> + /* link already initialized {secure,tag}-memory regions to this cpu */
> + object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err);
> + if (local_err) {
> + goto out;
> + }
> +
> + if (vms->secure) {
> + object_property_set_link(cpuobj, "secure-memory",
> + OBJECT(vms->secure_sysmem), &local_err);
> + if (local_err) {
> + goto out;
> + }
> + }
> +
> + if (vms->mte) {
> + if (!object_property_find(cpuobj, "tag-memory")) {
> + error_setg(&local_err, "MTE requested, but not supported "
> + "by the guest CPU");
> + if (local_err) {
> + goto out;
> + }
> + }
> +
> + object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem),
> + &local_err);
> + if (local_err) {
> + goto out;
> + }
> +
> + if (vms->secure) {
> + object_property_set_link(cpuobj, "secure-tag-memory",
> + OBJECT(vms->secure_tag_sysmem),
> + &local_err);
> + if (local_err) {
> + goto out;
> + }
> + }
> + }
> +
> + /*
> + * RFC: Question: this must only be called for the hotplugged cpus. For the
> + * cold booted secondary cpus this is being taken care in arm_load_kernel()
> + * in boot.c. Perhaps we should remove that code now?
> + */
> + if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) {
> + object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit,
> + NULL);
> +
> + /* Secondary CPUs start in PSCI powered-down state */
> + if (CPU(cpuobj)->cpu_index > 0) {
> + object_property_set_bool(cpuobj, "start-powered-off", true, NULL);
> + }
> + }
> +
> +out:
> + if (local_err) {
> + error_propagate(errp, local_err);
> + }
> +}
> +
> static void machvirt_init(MachineState *machine)
> {
> VirtMachineState *vms = VIRT_MACHINE(machine);
> VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine);
> MachineClass *mc = MACHINE_GET_CLASS(machine);
> const CPUArchIdList *possible_cpus;
> - MemoryRegion *sysmem = get_system_memory();
> + MemoryRegion *secure_tag_sysmem = NULL;
> MemoryRegion *secure_sysmem = NULL;
> MemoryRegion *tag_sysmem = NULL;
> - MemoryRegion *secure_tag_sysmem = NULL;
> + MemoryRegion *sysmem;
> int n, virt_max_cpus;
> bool firmware_loaded;
> bool aarch64 = true;
> @@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine)
> /* uses smp.max_cpus to initialize all possible vCPUs */
> possible_cpus = mc->possible_cpu_arch_ids(machine);
>
> + sysmem = vms->sysmem = get_system_memory();
> +
> if (vms->secure) {
> /*
> * The Secure view of the world is the same as the NonSecure,
> @@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine)
> * containing the system memory at low priority; any secure-only
> * devices go in at higher priority and take precedence.
> */
> - secure_sysmem = g_new(MemoryRegion, 1);
> + secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1);
> memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory",
> UINT64_MAX);
> memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1);
> @@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine)
> exit(1);
> }
>
> + if (vms->mte) {
> + /* Create the memory region only once, but link to all cpus later */
> + tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1);
> + memory_region_init(tag_sysmem, OBJECT(machine),
> + "tag-memory", UINT64_MAX / 32);
> +
> + if (vms->secure) {
> + secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1);
> + memory_region_init(secure_tag_sysmem, OBJECT(machine),
> + "secure-tag-memory", UINT64_MAX / 32);
> +
> + /* As with ram, secure-tag takes precedence over tag. */
> + memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
> + tag_sysmem, -1);
> + }
> + }
> +
> create_fdt(vms);
>
> assert(possible_cpus->len == max_cpus);
> for (n = 0; n < possible_cpus->len; n++) {
> + CPUArchId *cpu_slot;
> Object *cpuobj;
> CPUState *cs;
>
> @@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine)
> }
>
> cpuobj = object_new(possible_cpus->cpus[n].type);
> - object_property_set_int(cpuobj, "mp-affinity",
> - possible_cpus->cpus[n].arch_id, NULL);
>
> cs = CPU(cpuobj);
> cs->cpu_index = n;
>
> - numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj),
> - &error_fatal);
> -
> aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL);
> object_property_set_int(cpuobj, "socket-id",
> virt_get_socket_id(machine, n), NULL);
> @@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine)
> object_property_set_int(cpuobj, "thread-id",
> virt_get_thread_id(machine, n), NULL);
>
> - if (!vms->secure) {
> - object_property_set_bool(cpuobj, "has_el3", false, NULL);
> - }
> -
> - if (!vms->virt && object_property_find(cpuobj, "has_el2")) {
> - object_property_set_bool(cpuobj, "has_el2", false, NULL);
> - }
> -
> - if (vmc->kvm_no_adjvtime &&
> - object_property_find(cpuobj, "kvm-no-adjvtime")) {
> - object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL);
> - }
> -
> - if (vmc->no_kvm_steal_time &&
> - object_property_find(cpuobj, "kvm-steal-time")) {
> - object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL);
> - }
> -
> - if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) {
> - object_property_set_bool(cpuobj, "pmu", false, NULL);
> - }
> -
> - if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) {
> - object_property_set_bool(cpuobj, "lpa2", false, NULL);
> - }
> -
> - if (object_property_find(cpuobj, "reset-cbar")) {
> - object_property_set_int(cpuobj, "reset-cbar",
> - vms->memmap[VIRT_CPUPERIPHS].base,
> - &error_abort);
> - }
> -
> - object_property_set_link(cpuobj, "memory", OBJECT(sysmem),
> - &error_abort);
> - if (vms->secure) {
> - object_property_set_link(cpuobj, "secure-memory",
> - OBJECT(secure_sysmem), &error_abort);
> - }
> -
> - if (vms->mte) {
> - /* Create the memory region only once, but link to all cpus. */
> - if (!tag_sysmem) {
> - /*
> - * The property exists only if MemTag is supported.
> - * If it is, we must allocate the ram to back that up.
> - */
> - if (!object_property_find(cpuobj, "tag-memory")) {
> - error_report("MTE requested, but not supported "
> - "by the guest CPU");
> - exit(1);
> - }
> -
> - tag_sysmem = g_new(MemoryRegion, 1);
> - memory_region_init(tag_sysmem, OBJECT(machine),
> - "tag-memory", UINT64_MAX / 32);
> -
> - if (vms->secure) {
> - secure_tag_sysmem = g_new(MemoryRegion, 1);
> - memory_region_init(secure_tag_sysmem, OBJECT(machine),
> - "secure-tag-memory", UINT64_MAX / 32);
> -
> - /* As with ram, secure-tag takes precedence over tag. */
> - memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
> - tag_sysmem, -1);
> - }
> - }
> -
> - object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem),
> - &error_abort);
> - if (vms->secure) {
> - object_property_set_link(cpuobj, "secure-tag-memory",
> - OBJECT(secure_tag_sysmem),
> - &error_abort);
> - }
> - }
> + cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index);
> + virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal);
>
> qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
> object_unref(cpuobj);
> diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h
> index 6f9a7bb60b..780bd53ceb 100644
> --- a/include/hw/arm/virt.h
> +++ b/include/hw/arm/virt.h
> @@ -139,6 +139,10 @@ struct VirtMachineState {
> DeviceState *platform_bus_dev;
> FWCfgState *fw_cfg;
> PFlashCFI01 *flash[2];
> + MemoryRegion *sysmem;
> + MemoryRegion *secure_sysmem;
> + MemoryRegion *tag_sysmem;
> + MemoryRegion *secure_tag_sysmem;
> bool secure;
> bool highmem;
> bool highmem_compact;
Thanks,
Gavin
© 2016 - 2026 Red Hat, Inc.