Compute the host reserved regions in virtio_iommu_set_iommu_device().
The usable IOVA regions are retrieved from the HOSTIOMMUDevice.
The virtio_iommu_set_host_iova_ranges() helper turns usable regions
into complementary reserved regions while testing the inclusion
into existing ones. virtio_iommu_set_host_iova_ranges() reuse the
implementation of virtio_iommu_set_iova_ranges() which will be
removed in subsequent patches. rebuild_resv_regions() is just moved.
Signed-off-by: Eric Auger <eric.auger@redhat.com>
---
hw/virtio/virtio-iommu.c | 151 ++++++++++++++++++++++++++++++---------
1 file changed, 117 insertions(+), 34 deletions(-)
diff --git a/hw/virtio/virtio-iommu.c b/hw/virtio/virtio-iommu.c
index 0680a357f0..33e9682b83 100644
--- a/hw/virtio/virtio-iommu.c
+++ b/hw/virtio/virtio-iommu.c
@@ -494,12 +494,114 @@ get_host_iommu_device(VirtIOIOMMU *viommu, PCIBus *bus, int devfn) {
return g_hash_table_lookup(viommu->host_iommu_devices, &key);
}
+/**
+ * rebuild_resv_regions: rebuild resv regions with both the
+ * info of host resv ranges and property set resv ranges
+ */
+static int rebuild_resv_regions(IOMMUDevice *sdev)
+{
+ GList *l;
+ int i = 0;
+
+ /* free the existing list and rebuild it from scratch */
+ g_list_free_full(sdev->resv_regions, g_free);
+ sdev->resv_regions = NULL;
+
+ /* First add host reserved regions if any, all tagged as RESERVED */
+ for (l = sdev->host_resv_ranges; l; l = l->next) {
+ ReservedRegion *reg = g_new0(ReservedRegion, 1);
+ Range *r = (Range *)l->data;
+
+ reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
+ range_set_bounds(®->range, range_lob(r), range_upb(r));
+ sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
+ trace_virtio_iommu_host_resv_regions(sdev->iommu_mr.parent_obj.name, i,
+ range_lob(®->range),
+ range_upb(®->range));
+ i++;
+ }
+ /*
+ * then add higher priority reserved regions set by the machine
+ * through properties
+ */
+ add_prop_resv_regions(sdev);
+ return 0;
+}
+
+static int virtio_iommu_set_host_iova_ranges(VirtIOIOMMU *s, PCIBus *bus,
+ int devfn, GList *iova_ranges,
+ Error **errp)
+{
+ IOMMUPciBus *sbus = g_hash_table_lookup(s->as_by_busptr, bus);
+ IOMMUDevice *sdev;
+ GList *current_ranges;
+ GList *l, *tmp, *new_ranges = NULL;
+ int ret = -EINVAL;
+
+ if (!sbus) {
+ error_report("%s no sbus", __func__);
+ }
+
+ sdev = sbus->pbdev[devfn];
+
+ current_ranges = sdev->host_resv_ranges;
+
+ if (sdev->probe_done) {
+ error_setg(errp,
+ "%s: Notified about new host reserved regions after probe",
+ __func__);
+ goto out;
+ }
+
+ /* check that each new resv region is included in an existing one */
+ if (sdev->host_resv_ranges) {
+ range_inverse_array(iova_ranges,
+ &new_ranges,
+ 0, UINT64_MAX);
+
+ for (tmp = new_ranges; tmp; tmp = tmp->next) {
+ Range *newr = (Range *)tmp->data;
+ bool included = false;
+
+ for (l = current_ranges; l; l = l->next) {
+ Range * r = (Range *)l->data;
+
+ if (range_contains_range(r, newr)) {
+ included = true;
+ break;
+ }
+ }
+ if (!included) {
+ goto error;
+ }
+ }
+ /* all new reserved ranges are included in existing ones */
+ ret = 0;
+ goto out;
+ }
+
+ range_inverse_array(iova_ranges,
+ &sdev->host_resv_ranges,
+ 0, UINT64_MAX);
+ rebuild_resv_regions(sdev);
+
+ return 0;
+error:
+ error_setg(errp, "%s Conflicting host reserved ranges set!",
+ __func__);
+out:
+ g_list_free_full(new_ranges, g_free);
+ return ret;
+}
+
static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
HostIOMMUDevice *hiod, Error **errp)
{
VirtIOIOMMU *viommu = opaque;
VirtioHostIOMMUDevice *vhiod;
+ HostIOMMUDeviceClass *hiodc = HOST_IOMMU_DEVICE_GET_CLASS(hiod);
struct hiod_key *new_key;
+ GList *host_iova_ranges = NULL;
assert(hiod);
@@ -509,6 +611,20 @@ static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
return false;
}
+ if (hiodc->get_iova_ranges) {
+ int ret;
+ host_iova_ranges = hiodc->get_iova_ranges(hiod, errp);
+ if (!host_iova_ranges) {
+ return true; /* some old kernels may not support that capability */
+ }
+ ret = virtio_iommu_set_host_iova_ranges(viommu, bus, devfn,
+ host_iova_ranges, errp);
+ if (ret) {
+ g_list_free_full(host_iova_ranges, g_free);
+ return false;
+ }
+ }
+
vhiod = g_malloc0(sizeof(VirtioHostIOMMUDevice));
vhiod->bus = bus;
vhiod->devfn = (uint8_t)devfn;
@@ -521,6 +637,7 @@ static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
object_ref(hiod);
g_hash_table_insert(viommu->host_iommu_devices, new_key, vhiod);
+ g_list_free_full(host_iova_ranges, g_free);
return true;
}
@@ -1243,40 +1360,6 @@ static int virtio_iommu_set_page_size_mask(IOMMUMemoryRegion *mr,
return 0;
}
-/**
- * rebuild_resv_regions: rebuild resv regions with both the
- * info of host resv ranges and property set resv ranges
- */
-static int rebuild_resv_regions(IOMMUDevice *sdev)
-{
- GList *l;
- int i = 0;
-
- /* free the existing list and rebuild it from scratch */
- g_list_free_full(sdev->resv_regions, g_free);
- sdev->resv_regions = NULL;
-
- /* First add host reserved regions if any, all tagged as RESERVED */
- for (l = sdev->host_resv_ranges; l; l = l->next) {
- ReservedRegion *reg = g_new0(ReservedRegion, 1);
- Range *r = (Range *)l->data;
-
- reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
- range_set_bounds(®->range, range_lob(r), range_upb(r));
- sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
- trace_virtio_iommu_host_resv_regions(sdev->iommu_mr.parent_obj.name, i,
- range_lob(®->range),
- range_upb(®->range));
- i++;
- }
- /*
- * then add higher priority reserved regions set by the machine
- * through properties
- */
- add_prop_resv_regions(sdev);
- return 0;
-}
-
/**
* virtio_iommu_set_iova_ranges: Conveys the usable IOVA ranges
*
--
2.41.0
>-----Original Message-----
>From: Eric Auger <eric.auger@redhat.com>
>Subject: [RFC v2 4/7] virtio-iommu: Compute host reserved regions
>
>Compute the host reserved regions in virtio_iommu_set_iommu_device().
>The usable IOVA regions are retrieved from the HOSTIOMMUDevice.
>The virtio_iommu_set_host_iova_ranges() helper turns usable regions
>into complementary reserved regions while testing the inclusion
>into existing ones. virtio_iommu_set_host_iova_ranges() reuse the
>implementation of virtio_iommu_set_iova_ranges() which will be
>removed in subsequent patches. rebuild_resv_regions() is just moved.
>
>Signed-off-by: Eric Auger <eric.auger@redhat.com>
>---
> hw/virtio/virtio-iommu.c | 151 ++++++++++++++++++++++++++++++--------
>-
> 1 file changed, 117 insertions(+), 34 deletions(-)
>
>diff --git a/hw/virtio/virtio-iommu.c b/hw/virtio/virtio-iommu.c
>index 0680a357f0..33e9682b83 100644
>--- a/hw/virtio/virtio-iommu.c
>+++ b/hw/virtio/virtio-iommu.c
>@@ -494,12 +494,114 @@ get_host_iommu_device(VirtIOIOMMU
>*viommu, PCIBus *bus, int devfn) {
> return g_hash_table_lookup(viommu->host_iommu_devices, &key);
> }
>
>+/**
>+ * rebuild_resv_regions: rebuild resv regions with both the
>+ * info of host resv ranges and property set resv ranges
>+ */
>+static int rebuild_resv_regions(IOMMUDevice *sdev)
>+{
>+ GList *l;
>+ int i = 0;
>+
>+ /* free the existing list and rebuild it from scratch */
>+ g_list_free_full(sdev->resv_regions, g_free);
>+ sdev->resv_regions = NULL;
>+
>+ /* First add host reserved regions if any, all tagged as RESERVED */
>+ for (l = sdev->host_resv_ranges; l; l = l->next) {
>+ ReservedRegion *reg = g_new0(ReservedRegion, 1);
>+ Range *r = (Range *)l->data;
>+
>+ reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
>+ range_set_bounds(®->range, range_lob(r), range_upb(r));
>+ sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
>+ trace_virtio_iommu_host_resv_regions(sdev-
>>iommu_mr.parent_obj.name, i,
>+ range_lob(®->range),
>+ range_upb(®->range));
>+ i++;
>+ }
>+ /*
>+ * then add higher priority reserved regions set by the machine
>+ * through properties
>+ */
>+ add_prop_resv_regions(sdev);
>+ return 0;
>+}
>+
>+static int virtio_iommu_set_host_iova_ranges(VirtIOIOMMU *s, PCIBus
>*bus,
>+ int devfn, GList *iova_ranges,
>+ Error **errp)
>+{
>+ IOMMUPciBus *sbus = g_hash_table_lookup(s->as_by_busptr, bus);
>+ IOMMUDevice *sdev;
>+ GList *current_ranges;
>+ GList *l, *tmp, *new_ranges = NULL;
>+ int ret = -EINVAL;
>+
>+ if (!sbus) {
>+ error_report("%s no sbus", __func__);
>+ }
>+
>+ sdev = sbus->pbdev[devfn];
>+
>+ current_ranges = sdev->host_resv_ranges;
>+
>+ if (sdev->probe_done) {
Will this still happen with new interface?
>+ error_setg(errp,
>+ "%s: Notified about new host reserved regions after probe",
>+ __func__);
>+ goto out;
>+ }
>+
>+ /* check that each new resv region is included in an existing one */
>+ if (sdev->host_resv_ranges) {
Same here.
>+ range_inverse_array(iova_ranges,
>+ &new_ranges,
>+ 0, UINT64_MAX);
>+
>+ for (tmp = new_ranges; tmp; tmp = tmp->next) {
>+ Range *newr = (Range *)tmp->data;
>+ bool included = false;
>+
>+ for (l = current_ranges; l; l = l->next) {
>+ Range * r = (Range *)l->data;
>+
>+ if (range_contains_range(r, newr)) {
>+ included = true;
>+ break;
>+ }
>+ }
>+ if (!included) {
>+ goto error;
>+ }
>+ }
>+ /* all new reserved ranges are included in existing ones */
>+ ret = 0;
>+ goto out;
>+ }
>+
>+ range_inverse_array(iova_ranges,
>+ &sdev->host_resv_ranges,
>+ 0, UINT64_MAX);
>+ rebuild_resv_regions(sdev);
>+
>+ return 0;
>+error:
>+ error_setg(errp, "%s Conflicting host reserved ranges set!",
>+ __func__);
>+out:
>+ g_list_free_full(new_ranges, g_free);
>+ return ret;
>+}
>+
> static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque,
>int devfn,
> HostIOMMUDevice *hiod, Error **errp)
> {
> VirtIOIOMMU *viommu = opaque;
> VirtioHostIOMMUDevice *vhiod;
>+ HostIOMMUDeviceClass *hiodc =
>HOST_IOMMU_DEVICE_GET_CLASS(hiod);
> struct hiod_key *new_key;
>+ GList *host_iova_ranges = NULL;
g_autoptr(GList)?
Thanks
Zhenzhong
>
> assert(hiod);
>
>@@ -509,6 +611,20 @@ static bool
>virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
> return false;
> }
>
>+ if (hiodc->get_iova_ranges) {
>+ int ret;
>+ host_iova_ranges = hiodc->get_iova_ranges(hiod, errp);
>+ if (!host_iova_ranges) {
>+ return true; /* some old kernels may not support that capability */
>+ }
>+ ret = virtio_iommu_set_host_iova_ranges(viommu, bus, devfn,
>+ host_iova_ranges, errp);
>+ if (ret) {
>+ g_list_free_full(host_iova_ranges, g_free);
>+ return false;
>+ }
>+ }
>+
> vhiod = g_malloc0(sizeof(VirtioHostIOMMUDevice));
> vhiod->bus = bus;
> vhiod->devfn = (uint8_t)devfn;
>@@ -521,6 +637,7 @@ static bool
>virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
>
> object_ref(hiod);
> g_hash_table_insert(viommu->host_iommu_devices, new_key, vhiod);
>+ g_list_free_full(host_iova_ranges, g_free);
>
> return true;
> }
>@@ -1243,40 +1360,6 @@ static int
>virtio_iommu_set_page_size_mask(IOMMUMemoryRegion *mr,
> return 0;
> }
>
>-/**
>- * rebuild_resv_regions: rebuild resv regions with both the
>- * info of host resv ranges and property set resv ranges
>- */
>-static int rebuild_resv_regions(IOMMUDevice *sdev)
>-{
>- GList *l;
>- int i = 0;
>-
>- /* free the existing list and rebuild it from scratch */
>- g_list_free_full(sdev->resv_regions, g_free);
>- sdev->resv_regions = NULL;
>-
>- /* First add host reserved regions if any, all tagged as RESERVED */
>- for (l = sdev->host_resv_ranges; l; l = l->next) {
>- ReservedRegion *reg = g_new0(ReservedRegion, 1);
>- Range *r = (Range *)l->data;
>-
>- reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
>- range_set_bounds(®->range, range_lob(r), range_upb(r));
>- sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
>- trace_virtio_iommu_host_resv_regions(sdev-
>>iommu_mr.parent_obj.name, i,
>- range_lob(®->range),
>- range_upb(®->range));
>- i++;
>- }
>- /*
>- * then add higher priority reserved regions set by the machine
>- * through properties
>- */
>- add_prop_resv_regions(sdev);
>- return 0;
>-}
>-
> /**
> * virtio_iommu_set_iova_ranges: Conveys the usable IOVA ranges
> *
>--
>2.41.0
Hi Zhenzhong,
On 6/11/24 05:25, Duan, Zhenzhong wrote:
>
>> -----Original Message-----
>> From: Eric Auger <eric.auger@redhat.com>
>> Subject: [RFC v2 4/7] virtio-iommu: Compute host reserved regions
>>
>> Compute the host reserved regions in virtio_iommu_set_iommu_device().
>> The usable IOVA regions are retrieved from the HOSTIOMMUDevice.
>> The virtio_iommu_set_host_iova_ranges() helper turns usable regions
>> into complementary reserved regions while testing the inclusion
>> into existing ones. virtio_iommu_set_host_iova_ranges() reuse the
>> implementation of virtio_iommu_set_iova_ranges() which will be
>> removed in subsequent patches. rebuild_resv_regions() is just moved.
>>
>> Signed-off-by: Eric Auger <eric.auger@redhat.com>
>> ---
>> hw/virtio/virtio-iommu.c | 151 ++++++++++++++++++++++++++++++--------
>> -
>> 1 file changed, 117 insertions(+), 34 deletions(-)
>>
>> diff --git a/hw/virtio/virtio-iommu.c b/hw/virtio/virtio-iommu.c
>> index 0680a357f0..33e9682b83 100644
>> --- a/hw/virtio/virtio-iommu.c
>> +++ b/hw/virtio/virtio-iommu.c
>> @@ -494,12 +494,114 @@ get_host_iommu_device(VirtIOIOMMU
>> *viommu, PCIBus *bus, int devfn) {
>> return g_hash_table_lookup(viommu->host_iommu_devices, &key);
>> }
>>
>> +/**
>> + * rebuild_resv_regions: rebuild resv regions with both the
>> + * info of host resv ranges and property set resv ranges
>> + */
>> +static int rebuild_resv_regions(IOMMUDevice *sdev)
>> +{
>> + GList *l;
>> + int i = 0;
>> +
>> + /* free the existing list and rebuild it from scratch */
>> + g_list_free_full(sdev->resv_regions, g_free);
>> + sdev->resv_regions = NULL;
>> +
>> + /* First add host reserved regions if any, all tagged as RESERVED */
>> + for (l = sdev->host_resv_ranges; l; l = l->next) {
>> + ReservedRegion *reg = g_new0(ReservedRegion, 1);
>> + Range *r = (Range *)l->data;
>> +
>> + reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
>> + range_set_bounds(®->range, range_lob(r), range_upb(r));
>> + sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
>> + trace_virtio_iommu_host_resv_regions(sdev-
>>> iommu_mr.parent_obj.name, i,
>> + range_lob(®->range),
>> + range_upb(®->range));
>> + i++;
>> + }
>> + /*
>> + * then add higher priority reserved regions set by the machine
>> + * through properties
>> + */
>> + add_prop_resv_regions(sdev);
>> + return 0;
>> +}
>> +
>> +static int virtio_iommu_set_host_iova_ranges(VirtIOIOMMU *s, PCIBus
>> *bus,
>> + int devfn, GList *iova_ranges,
>> + Error **errp)
>> +{
>> + IOMMUPciBus *sbus = g_hash_table_lookup(s->as_by_busptr, bus);
>> + IOMMUDevice *sdev;
>> + GList *current_ranges;
>> + GList *l, *tmp, *new_ranges = NULL;
>> + int ret = -EINVAL;
>> +
>> + if (!sbus) {
>> + error_report("%s no sbus", __func__);
>> + }
>> +
>> + sdev = sbus->pbdev[devfn];
>> +
>> + current_ranges = sdev->host_resv_ranges;
>> +
>> + if (sdev->probe_done) {
> Will this still happen with new interface?
no this shouldn't. Replaced by a g_assert(!sdev->probe_done) to make
sure the i/f is used properly.
>
>> + error_setg(errp,
>> + "%s: Notified about new host reserved regions after probe",
>> + __func__);
>> + goto out;
>> + }
>> +
>> + /* check that each new resv region is included in an existing one */
>> + if (sdev->host_resv_ranges) {
> Same here.
To me this one can still happen in case several devices belong to the
same group.
>
>> + range_inverse_array(iova_ranges,
>> + &new_ranges,
>> + 0, UINT64_MAX);
>> +
>> + for (tmp = new_ranges; tmp; tmp = tmp->next) {
>> + Range *newr = (Range *)tmp->data;
>> + bool included = false;
>> +
>> + for (l = current_ranges; l; l = l->next) {
>> + Range * r = (Range *)l->data;
>> +
>> + if (range_contains_range(r, newr)) {
>> + included = true;
>> + break;
>> + }
>> + }
>> + if (!included) {
>> + goto error;
>> + }
>> + }
>> + /* all new reserved ranges are included in existing ones */
>> + ret = 0;
>> + goto out;
>> + }
>> +
>> + range_inverse_array(iova_ranges,
>> + &sdev->host_resv_ranges,
>> + 0, UINT64_MAX);
>> + rebuild_resv_regions(sdev);
>> +
>> + return 0;
>> +error:
>> + error_setg(errp, "%s Conflicting host reserved ranges set!",
>> + __func__);
>> +out:
>> + g_list_free_full(new_ranges, g_free);
>> + return ret;
>> +}
>> +
>> static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque,
>> int devfn,
>> HostIOMMUDevice *hiod, Error **errp)
>> {
>> VirtIOIOMMU *viommu = opaque;
>> VirtioHostIOMMUDevice *vhiod;
>> + HostIOMMUDeviceClass *hiodc =
>> HOST_IOMMU_DEVICE_GET_CLASS(hiod);
>> struct hiod_key *new_key;
>> + GList *host_iova_ranges = NULL;
> g_autoptr(GList)?
are you sure this frees all the elements of the list? As of now I would
be tempted to leave the code as is.
Thanks
Eric
>
> Thanks
> Zhenzhong
>
>> assert(hiod);
>>
>> @@ -509,6 +611,20 @@ static bool
>> virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
>> return false;
>> }
>>
>> + if (hiodc->get_iova_ranges) {
>> + int ret;
>> + host_iova_ranges = hiodc->get_iova_ranges(hiod, errp);
>> + if (!host_iova_ranges) {
>> + return true; /* some old kernels may not support that capability */
>> + }
>> + ret = virtio_iommu_set_host_iova_ranges(viommu, bus, devfn,
>> + host_iova_ranges, errp);
>> + if (ret) {
>> + g_list_free_full(host_iova_ranges, g_free);
>> + return false;
>> + }
>> + }
>> +
>> vhiod = g_malloc0(sizeof(VirtioHostIOMMUDevice));
>> vhiod->bus = bus;
>> vhiod->devfn = (uint8_t)devfn;
>> @@ -521,6 +637,7 @@ static bool
>> virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque, int devfn,
>>
>> object_ref(hiod);
>> g_hash_table_insert(viommu->host_iommu_devices, new_key, vhiod);
>> + g_list_free_full(host_iova_ranges, g_free);
>>
>> return true;
>> }
>> @@ -1243,40 +1360,6 @@ static int
>> virtio_iommu_set_page_size_mask(IOMMUMemoryRegion *mr,
>> return 0;
>> }
>>
>> -/**
>> - * rebuild_resv_regions: rebuild resv regions with both the
>> - * info of host resv ranges and property set resv ranges
>> - */
>> -static int rebuild_resv_regions(IOMMUDevice *sdev)
>> -{
>> - GList *l;
>> - int i = 0;
>> -
>> - /* free the existing list and rebuild it from scratch */
>> - g_list_free_full(sdev->resv_regions, g_free);
>> - sdev->resv_regions = NULL;
>> -
>> - /* First add host reserved regions if any, all tagged as RESERVED */
>> - for (l = sdev->host_resv_ranges; l; l = l->next) {
>> - ReservedRegion *reg = g_new0(ReservedRegion, 1);
>> - Range *r = (Range *)l->data;
>> -
>> - reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
>> - range_set_bounds(®->range, range_lob(r), range_upb(r));
>> - sdev->resv_regions = resv_region_list_insert(sdev->resv_regions, reg);
>> - trace_virtio_iommu_host_resv_regions(sdev-
>>> iommu_mr.parent_obj.name, i,
>> - range_lob(®->range),
>> - range_upb(®->range));
>> - i++;
>> - }
>> - /*
>> - * then add higher priority reserved regions set by the machine
>> - * through properties
>> - */
>> - add_prop_resv_regions(sdev);
>> - return 0;
>> -}
>> -
>> /**
>> * virtio_iommu_set_iova_ranges: Conveys the usable IOVA ranges
>> *
>> --
>> 2.41.0
>-----Original Message-----
>From: Eric Auger <eric.auger@redhat.com>
>Subject: Re: [RFC v2 4/7] virtio-iommu: Compute host reserved regions
>
>Hi Zhenzhong,
>
>On 6/11/24 05:25, Duan, Zhenzhong wrote:
>>
>>> -----Original Message-----
>>> From: Eric Auger <eric.auger@redhat.com>
>>> Subject: [RFC v2 4/7] virtio-iommu: Compute host reserved regions
>>>
>>> Compute the host reserved regions in virtio_iommu_set_iommu_device().
>>> The usable IOVA regions are retrieved from the HOSTIOMMUDevice.
>>> The virtio_iommu_set_host_iova_ranges() helper turns usable regions
>>> into complementary reserved regions while testing the inclusion
>>> into existing ones. virtio_iommu_set_host_iova_ranges() reuse the
>>> implementation of virtio_iommu_set_iova_ranges() which will be
>>> removed in subsequent patches. rebuild_resv_regions() is just moved.
>>>
>>> Signed-off-by: Eric Auger <eric.auger@redhat.com>
>>> ---
>>> hw/virtio/virtio-iommu.c | 151 ++++++++++++++++++++++++++++++----
>----
>>> -
>>> 1 file changed, 117 insertions(+), 34 deletions(-)
>>>
>>> diff --git a/hw/virtio/virtio-iommu.c b/hw/virtio/virtio-iommu.c
>>> index 0680a357f0..33e9682b83 100644
>>> --- a/hw/virtio/virtio-iommu.c
>>> +++ b/hw/virtio/virtio-iommu.c
>>> @@ -494,12 +494,114 @@ get_host_iommu_device(VirtIOIOMMU
>>> *viommu, PCIBus *bus, int devfn) {
>>> return g_hash_table_lookup(viommu->host_iommu_devices, &key);
>>> }
>>>
>>> +/**
>>> + * rebuild_resv_regions: rebuild resv regions with both the
>>> + * info of host resv ranges and property set resv ranges
>>> + */
>>> +static int rebuild_resv_regions(IOMMUDevice *sdev)
>>> +{
>>> + GList *l;
>>> + int i = 0;
>>> +
>>> + /* free the existing list and rebuild it from scratch */
>>> + g_list_free_full(sdev->resv_regions, g_free);
>>> + sdev->resv_regions = NULL;
>>> +
>>> + /* First add host reserved regions if any, all tagged as RESERVED */
>>> + for (l = sdev->host_resv_ranges; l; l = l->next) {
>>> + ReservedRegion *reg = g_new0(ReservedRegion, 1);
>>> + Range *r = (Range *)l->data;
>>> +
>>> + reg->type = VIRTIO_IOMMU_RESV_MEM_T_RESERVED;
>>> + range_set_bounds(®->range, range_lob(r), range_upb(r));
>>> + sdev->resv_regions = resv_region_list_insert(sdev->resv_regions,
>reg);
>>> + trace_virtio_iommu_host_resv_regions(sdev-
>>>> iommu_mr.parent_obj.name, i,
>>> + range_lob(®->range),
>>> + range_upb(®->range));
>>> + i++;
>>> + }
>>> + /*
>>> + * then add higher priority reserved regions set by the machine
>>> + * through properties
>>> + */
>>> + add_prop_resv_regions(sdev);
>>> + return 0;
>>> +}
>>> +
>>> +static int virtio_iommu_set_host_iova_ranges(VirtIOIOMMU *s, PCIBus
>>> *bus,
>>> + int devfn, GList *iova_ranges,
>>> + Error **errp)
>>> +{
>>> + IOMMUPciBus *sbus = g_hash_table_lookup(s->as_by_busptr, bus);
>>> + IOMMUDevice *sdev;
>>> + GList *current_ranges;
>>> + GList *l, *tmp, *new_ranges = NULL;
>>> + int ret = -EINVAL;
>>> +
>>> + if (!sbus) {
>>> + error_report("%s no sbus", __func__);
>>> + }
>>> +
>>> + sdev = sbus->pbdev[devfn];
>>> +
>>> + current_ranges = sdev->host_resv_ranges;
>>> +
>>> + if (sdev->probe_done) {
>> Will this still happen with new interface?
>no this shouldn't. Replaced by a g_assert(!sdev->probe_done) to make
>sure the i/f is used properly.
>>
>>> + error_setg(errp,
>>> + "%s: Notified about new host reserved regions after probe",
>>> + __func__);
>>> + goto out;
>>> + }
>>> +
>>> + /* check that each new resv region is included in an existing one */
>>> + if (sdev->host_resv_ranges) {
>> Same here.
>To me this one can still happen in case several devices belong to the
>same group.
If same slot is used to plug/unplug vfio devices with different reserved
ranges, the second device plug may check failed.
It looks sdev->host_resv_ranges is not freed after unplug.
>>
>>> + range_inverse_array(iova_ranges,
>>> + &new_ranges,
>>> + 0, UINT64_MAX);
>>> +
>>> + for (tmp = new_ranges; tmp; tmp = tmp->next) {
>>> + Range *newr = (Range *)tmp->data;
>>> + bool included = false;
>>> +
>>> + for (l = current_ranges; l; l = l->next) {
>>> + Range * r = (Range *)l->data;
>>> +
>>> + if (range_contains_range(r, newr)) {
>>> + included = true;
>>> + break;
>>> + }
>>> + }
>>> + if (!included) {
>>> + goto error;
>>> + }
>>> + }
>>> + /* all new reserved ranges are included in existing ones */
>>> + ret = 0;
>>> + goto out;
>>> + }
>>> +
>>> + range_inverse_array(iova_ranges,
>>> + &sdev->host_resv_ranges,
>>> + 0, UINT64_MAX);
>>> + rebuild_resv_regions(sdev);
>>> +
>>> + return 0;
>>> +error:
>>> + error_setg(errp, "%s Conflicting host reserved ranges set!",
>>> + __func__);
>>> +out:
>>> + g_list_free_full(new_ranges, g_free);
>>> + return ret;
>>> +}
>>> +
>>> static bool virtio_iommu_set_iommu_device(PCIBus *bus, void *opaque,
>>> int devfn,
>>> HostIOMMUDevice *hiod, Error **errp)
>>> {
>>> VirtIOIOMMU *viommu = opaque;
>>> VirtioHostIOMMUDevice *vhiod;
>>> + HostIOMMUDeviceClass *hiodc =
>>> HOST_IOMMU_DEVICE_GET_CLASS(hiod);
>>> struct hiod_key *new_key;
>>> + GList *host_iova_ranges = NULL;
>> g_autoptr(GList)?
>are you sure this frees all the elements of the list?
Not quite sure, just see some code in qemu use it that way.
> As of now I would
>be tempted to leave the code as is.
Sure, that's fine.
Thanks
Zhenzhong
© 2016 - 2026 Red Hat, Inc.