From: "Edgar E. Iglesias" <edgar.iglesias@amd.com>
Add support for optionally creating a PCIe/GPEX controller.
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@amd.com>
---
hw/xen/xen-pvh-common.c | 76 +++++++++++++++++++++++++++++++++
include/hw/xen/xen-pvh-common.h | 29 +++++++++++++
2 files changed, 105 insertions(+)
diff --git a/hw/xen/xen-pvh-common.c b/hw/xen/xen-pvh-common.c
index 295f920442..28d7168446 100644
--- a/hw/xen/xen-pvh-common.c
+++ b/hw/xen/xen-pvh-common.c
@@ -122,6 +122,64 @@ static void xen_enable_tpm(XenPVHMachineState *s)
}
#endif
+/*
+ * We use the GPEX PCIe controller with its internal INTX PCI interrupt
+ * swizzling. This swizzling is emulated in QEMU and routes all INTX
+ * interrupts from endpoints down to only 4 INTX interrupts.
+ * See include/hw/pci/pci.h : pci_swizzle()
+ */
+static inline void xenpvh_gpex_init(XenPVHMachineState *s,
+ XenPVHMachineClass *xpc,
+ MemoryRegion *sysmem)
+{
+ MemoryRegion *ecam_reg;
+ MemoryRegion *mmio_reg;
+ DeviceState *dev;
+ int i;
+
+ object_initialize_child(OBJECT(s), "gpex", &s->pci.gpex,
+ TYPE_GPEX_HOST);
+ dev = DEVICE(&s->pci.gpex);
+ sysbus_realize_and_unref(SYS_BUS_DEVICE(dev), &error_fatal);
+
+ ecam_reg = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
+ memory_region_add_subregion(sysmem, s->cfg.pci_ecam.base, ecam_reg);
+
+ mmio_reg = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 1);
+
+ if (s->cfg.pci_mmio.size) {
+ memory_region_init_alias(&s->pci.mmio_alias, OBJECT(dev), "pcie-mmio",
+ mmio_reg,
+ s->cfg.pci_mmio.base, s->cfg.pci_mmio.size);
+ memory_region_add_subregion(sysmem, s->cfg.pci_mmio.base,
+ &s->pci.mmio_alias);
+ }
+
+ if (s->cfg.pci_mmio_high.size) {
+ memory_region_init_alias(&s->pci.mmio_high_alias, OBJECT(dev),
+ "pcie-mmio-high",
+ mmio_reg, s->cfg.pci_mmio_high.base, s->cfg.pci_mmio_high.size);
+ memory_region_add_subregion(sysmem, s->cfg.pci_mmio_high.base,
+ &s->pci.mmio_high_alias);
+ }
+
+ /*
+ * PVH implementations with PCI enabled must provide set_pci_intx_irq()
+ * and optionally an implementation of set_pci_link_route().
+ */
+ assert(xpc->set_pci_intx_irq);
+
+ for (i = 0; i < GPEX_NUM_IRQS; i++) {
+ qemu_irq irq = qemu_allocate_irq(xpc->set_pci_intx_irq, s, i);
+
+ sysbus_connect_irq(SYS_BUS_DEVICE(dev), i, irq);
+ gpex_set_irq_num(GPEX_HOST(dev), i, s->cfg.pci_intx_irq_base + i);
+ if (xpc->set_pci_link_route) {
+ xpc->set_pci_link_route(i, s->cfg.pci_intx_irq_base + i);
+ }
+ }
+}
+
static void xen_pvh_init(MachineState *ms)
{
XenPVHMachineState *s = XEN_PVH_MACHINE(ms);
@@ -152,6 +210,15 @@ static void xen_pvh_init(MachineState *ms)
}
#endif
+ /* Non-zero pci-ecam-size enables PCI. */
+ if (s->cfg.pci_ecam.size) {
+ if (s->cfg.pci_ecam.size != 256 * MiB) {
+ error_report("pci-ecam-size only supports values 0 or 0x10000000");
+ exit(EXIT_FAILURE);
+ }
+ xenpvh_gpex_init(s, xpc, sysmem);
+ }
+
/* Call the implementation specific init. */
if (xpc->init) {
xpc->init(ms);
@@ -200,6 +267,9 @@ XEN_PVH_PROP_MEMMAP(ram_high)
/* TPM only has a base-addr option. */
XEN_PVH_PROP_MEMMAP_BASE(tpm)
XEN_PVH_PROP_MEMMAP(virtio_mmio)
+XEN_PVH_PROP_MEMMAP(pci_ecam)
+XEN_PVH_PROP_MEMMAP(pci_mmio)
+XEN_PVH_PROP_MEMMAP(pci_mmio_high)
void xen_pvh_class_setup_common_props(XenPVHMachineClass *xpc)
{
@@ -242,6 +312,12 @@ do { \
OC_MEMMAP_PROP(oc, "virtio-mmio", virtio_mmio);
}
+ if (xpc->has_pci) {
+ OC_MEMMAP_PROP(oc, "pci-ecam", pci_ecam);
+ OC_MEMMAP_PROP(oc, "pci-mmio", pci_mmio);
+ OC_MEMMAP_PROP(oc, "pci-mmio-high", pci_mmio_high);
+ }
+
#ifdef CONFIG_TPM
if (xpc->has_tpm) {
object_class_property_add(oc, "tpm-base-addr", "uint64_t",
diff --git a/include/hw/xen/xen-pvh-common.h b/include/hw/xen/xen-pvh-common.h
index 77fd98b9fe..bc09eea936 100644
--- a/include/hw/xen/xen-pvh-common.h
+++ b/include/hw/xen/xen-pvh-common.h
@@ -25,10 +25,29 @@ struct XenPVHMachineClass {
/* PVH implementation specific init. */
void (*init)(MachineState *state);
+ /*
+ * set_pci_intx_irq - Deliver INTX irqs to the guest.
+ *
+ * @opaque: pointer to XenPVHMachineState.
+ * @irq: IRQ after swizzling, between 0-3.
+ * @level: IRQ level.
+ */
+ void (*set_pci_intx_irq)(void *opaque, int irq, int level);
+
+ /*
+ * set_pci_link_route: - optional implementation call to setup
+ * routing between INTX IRQ (0 - 3) and GSI's.
+ *
+ * @line: line the INTx line (0 => A .. 3 => B)
+ * @irq: GSI
+ */
+ int (*set_pci_link_route)(uint8_t line, uint8_t irq);
+
/*
* Each implementation can optionally enable features that it
* supports and are known to work.
*/
+ bool has_pci;
bool has_tpm;
bool has_virtio_mmio;
};
@@ -44,6 +63,12 @@ struct XenPVHMachineState {
MemoryRegion high;
} ram;
+ struct {
+ GPEXHost gpex;
+ MemoryRegion mmio_alias;
+ MemoryRegion mmio_high_alias;
+ } pci;
+
struct {
MemMapEntry ram_low, ram_high;
MemMapEntry tpm;
@@ -52,6 +77,10 @@ struct XenPVHMachineState {
MemMapEntry virtio_mmio;
uint32_t virtio_mmio_num;
uint32_t virtio_mmio_irq_base;
+
+ /* PCI */
+ MemMapEntry pci_ecam, pci_mmio, pci_mmio_high;
+ uint32_t pci_intx_irq_base;
} cfg;
};
--
2.43.0
On Tue, 20 Aug 2024, Edgar E. Iglesias wrote:
> From: "Edgar E. Iglesias" <edgar.iglesias@amd.com>
>
> Add support for optionally creating a PCIe/GPEX controller.
>
> Signed-off-by: Edgar E. Iglesias <edgar.iglesias@amd.com>
Reviewed-by: Stefano Stabellini <sstabellini@kernel.org>
> ---
> hw/xen/xen-pvh-common.c | 76 +++++++++++++++++++++++++++++++++
> include/hw/xen/xen-pvh-common.h | 29 +++++++++++++
> 2 files changed, 105 insertions(+)
>
> diff --git a/hw/xen/xen-pvh-common.c b/hw/xen/xen-pvh-common.c
> index 295f920442..28d7168446 100644
> --- a/hw/xen/xen-pvh-common.c
> +++ b/hw/xen/xen-pvh-common.c
> @@ -122,6 +122,64 @@ static void xen_enable_tpm(XenPVHMachineState *s)
> }
> #endif
>
> +/*
> + * We use the GPEX PCIe controller with its internal INTX PCI interrupt
> + * swizzling. This swizzling is emulated in QEMU and routes all INTX
> + * interrupts from endpoints down to only 4 INTX interrupts.
> + * See include/hw/pci/pci.h : pci_swizzle()
> + */
> +static inline void xenpvh_gpex_init(XenPVHMachineState *s,
> + XenPVHMachineClass *xpc,
> + MemoryRegion *sysmem)
> +{
> + MemoryRegion *ecam_reg;
> + MemoryRegion *mmio_reg;
> + DeviceState *dev;
> + int i;
> +
> + object_initialize_child(OBJECT(s), "gpex", &s->pci.gpex,
> + TYPE_GPEX_HOST);
> + dev = DEVICE(&s->pci.gpex);
> + sysbus_realize_and_unref(SYS_BUS_DEVICE(dev), &error_fatal);
> +
> + ecam_reg = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
> + memory_region_add_subregion(sysmem, s->cfg.pci_ecam.base, ecam_reg);
> +
> + mmio_reg = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 1);
> +
> + if (s->cfg.pci_mmio.size) {
> + memory_region_init_alias(&s->pci.mmio_alias, OBJECT(dev), "pcie-mmio",
> + mmio_reg,
> + s->cfg.pci_mmio.base, s->cfg.pci_mmio.size);
> + memory_region_add_subregion(sysmem, s->cfg.pci_mmio.base,
> + &s->pci.mmio_alias);
> + }
> +
> + if (s->cfg.pci_mmio_high.size) {
> + memory_region_init_alias(&s->pci.mmio_high_alias, OBJECT(dev),
> + "pcie-mmio-high",
> + mmio_reg, s->cfg.pci_mmio_high.base, s->cfg.pci_mmio_high.size);
> + memory_region_add_subregion(sysmem, s->cfg.pci_mmio_high.base,
> + &s->pci.mmio_high_alias);
> + }
> +
> + /*
> + * PVH implementations with PCI enabled must provide set_pci_intx_irq()
> + * and optionally an implementation of set_pci_link_route().
> + */
> + assert(xpc->set_pci_intx_irq);
> +
> + for (i = 0; i < GPEX_NUM_IRQS; i++) {
> + qemu_irq irq = qemu_allocate_irq(xpc->set_pci_intx_irq, s, i);
> +
> + sysbus_connect_irq(SYS_BUS_DEVICE(dev), i, irq);
> + gpex_set_irq_num(GPEX_HOST(dev), i, s->cfg.pci_intx_irq_base + i);
> + if (xpc->set_pci_link_route) {
> + xpc->set_pci_link_route(i, s->cfg.pci_intx_irq_base + i);
> + }
> + }
> +}
> +
> static void xen_pvh_init(MachineState *ms)
> {
> XenPVHMachineState *s = XEN_PVH_MACHINE(ms);
> @@ -152,6 +210,15 @@ static void xen_pvh_init(MachineState *ms)
> }
> #endif
>
> + /* Non-zero pci-ecam-size enables PCI. */
> + if (s->cfg.pci_ecam.size) {
> + if (s->cfg.pci_ecam.size != 256 * MiB) {
> + error_report("pci-ecam-size only supports values 0 or 0x10000000");
> + exit(EXIT_FAILURE);
> + }
> + xenpvh_gpex_init(s, xpc, sysmem);
> + }
> +
> /* Call the implementation specific init. */
> if (xpc->init) {
> xpc->init(ms);
> @@ -200,6 +267,9 @@ XEN_PVH_PROP_MEMMAP(ram_high)
> /* TPM only has a base-addr option. */
> XEN_PVH_PROP_MEMMAP_BASE(tpm)
> XEN_PVH_PROP_MEMMAP(virtio_mmio)
> +XEN_PVH_PROP_MEMMAP(pci_ecam)
> +XEN_PVH_PROP_MEMMAP(pci_mmio)
> +XEN_PVH_PROP_MEMMAP(pci_mmio_high)
>
> void xen_pvh_class_setup_common_props(XenPVHMachineClass *xpc)
> {
> @@ -242,6 +312,12 @@ do { \
> OC_MEMMAP_PROP(oc, "virtio-mmio", virtio_mmio);
> }
>
> + if (xpc->has_pci) {
> + OC_MEMMAP_PROP(oc, "pci-ecam", pci_ecam);
> + OC_MEMMAP_PROP(oc, "pci-mmio", pci_mmio);
> + OC_MEMMAP_PROP(oc, "pci-mmio-high", pci_mmio_high);
> + }
> +
> #ifdef CONFIG_TPM
> if (xpc->has_tpm) {
> object_class_property_add(oc, "tpm-base-addr", "uint64_t",
> diff --git a/include/hw/xen/xen-pvh-common.h b/include/hw/xen/xen-pvh-common.h
> index 77fd98b9fe..bc09eea936 100644
> --- a/include/hw/xen/xen-pvh-common.h
> +++ b/include/hw/xen/xen-pvh-common.h
> @@ -25,10 +25,29 @@ struct XenPVHMachineClass {
> /* PVH implementation specific init. */
> void (*init)(MachineState *state);
>
> + /*
> + * set_pci_intx_irq - Deliver INTX irqs to the guest.
> + *
> + * @opaque: pointer to XenPVHMachineState.
> + * @irq: IRQ after swizzling, between 0-3.
> + * @level: IRQ level.
> + */
> + void (*set_pci_intx_irq)(void *opaque, int irq, int level);
> +
> + /*
> + * set_pci_link_route: - optional implementation call to setup
> + * routing between INTX IRQ (0 - 3) and GSI's.
> + *
> + * @line: line the INTx line (0 => A .. 3 => B)
> + * @irq: GSI
> + */
> + int (*set_pci_link_route)(uint8_t line, uint8_t irq);
> +
> /*
> * Each implementation can optionally enable features that it
> * supports and are known to work.
> */
> + bool has_pci;
> bool has_tpm;
> bool has_virtio_mmio;
> };
> @@ -44,6 +63,12 @@ struct XenPVHMachineState {
> MemoryRegion high;
> } ram;
>
> + struct {
> + GPEXHost gpex;
> + MemoryRegion mmio_alias;
> + MemoryRegion mmio_high_alias;
> + } pci;
> +
> struct {
> MemMapEntry ram_low, ram_high;
> MemMapEntry tpm;
> @@ -52,6 +77,10 @@ struct XenPVHMachineState {
> MemMapEntry virtio_mmio;
> uint32_t virtio_mmio_num;
> uint32_t virtio_mmio_irq_base;
> +
> + /* PCI */
> + MemMapEntry pci_ecam, pci_mmio, pci_mmio_high;
> + uint32_t pci_intx_irq_base;
> } cfg;
> };
>
> --
> 2.43.0
>
© 2016 - 2026 Red Hat, Inc.