hw/virtio/vhost-vdpa.c | 13 +++++++++++++ 1 file changed, 13 insertions(+)
Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory
that used to be discarded will get re-populated and if we
discard+re-access memory after mapping+pinning, the pages mapped into the
vDPA IOMMU will go out of sync with the actual pages mapped into the user
space page tables.
Set discarding of RAM broken such that:
- virtio-mem and vhost-vdpa run mutually exclusive
- virtio-balloon is inhibited and no memory discards will get issued
In the future, we might be able to support coordinated discarding of RAM
as used by virtio-mem and as planned for VFIO.
Cc: Jason Wang <jasowang@redhat.com>
Cc: Michael S. Tsirkin <mst@redhat.com>
Cc: Cindy Lu <lulu@redhat.com>
Signed-off-by: David Hildenbrand <david@redhat.com>
---
Note: I was not actually able to reproduce/test as I fail to get the
vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, vhost_vdpa,
vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa devices
appear under /sys/bus/vdpa/devices/ or /dev/).
---
hw/virtio/vhost-vdpa.c | 13 +++++++++++++
1 file changed, 13 insertions(+)
diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c
index 01d2101d09..86058d4041 100644
--- a/hw/virtio/vhost-vdpa.c
+++ b/hw/virtio/vhost-vdpa.c
@@ -278,6 +278,17 @@ static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque)
uint64_t features;
assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
trace_vhost_vdpa_init(dev, opaque);
+ int ret;
+
+ /*
+ * Similar to VFIO, we end up pinning all guest memory and have to
+ * disable discarding of RAM.
+ */
+ ret = ram_block_discard_disable(true);
+ if (ret) {
+ error_report("Cannot set discarding of RAM broken");
+ return ret;
+ }
v = opaque;
v->dev = dev;
@@ -302,6 +313,8 @@ static int vhost_vdpa_cleanup(struct vhost_dev *dev)
memory_listener_unregister(&v->listener);
dev->opaque = NULL;
+ ram_block_discard_disable(false);
+
return 0;
}
--
2.29.2
On 2021/3/3 12:21 上午, David Hildenbrand wrote:
> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory
> that used to be discarded will get re-populated and if we
> discard+re-access memory after mapping+pinning, the pages mapped into the
> vDPA IOMMU will go out of sync with the actual pages mapped into the user
> space page tables.
>
> Set discarding of RAM broken such that:
> - virtio-mem and vhost-vdpa run mutually exclusive
> - virtio-balloon is inhibited and no memory discards will get issued
>
> In the future, we might be able to support coordinated discarding of RAM
> as used by virtio-mem and as planned for VFIO.
>
> Cc: Jason Wang <jasowang@redhat.com>
> Cc: Michael S. Tsirkin <mst@redhat.com>
> Cc: Cindy Lu <lulu@redhat.com>
> Signed-off-by: David Hildenbrand <david@redhat.com>
Acked-by: Jason Wang <jasowang@redhat.com>
> ---
>
> Note: I was not actually able to reproduce/test as I fail to get the
> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, vhost_vdpa,
> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa devices
> appear under /sys/bus/vdpa/devices/ or /dev/).
The device creation was switched to use vdpa tool that is integrated
with iproue2[1].
[1]
https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18
>
> ---
> hw/virtio/vhost-vdpa.c | 13 +++++++++++++
> 1 file changed, 13 insertions(+)
>
> diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c
> index 01d2101d09..86058d4041 100644
> --- a/hw/virtio/vhost-vdpa.c
> +++ b/hw/virtio/vhost-vdpa.c
> @@ -278,6 +278,17 @@ static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque)
> uint64_t features;
> assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
> trace_vhost_vdpa_init(dev, opaque);
> + int ret;
> +
> + /*
> + * Similar to VFIO, we end up pinning all guest memory and have to
> + * disable discarding of RAM.
> + */
> + ret = ram_block_discard_disable(true);
> + if (ret) {
> + error_report("Cannot set discarding of RAM broken");
> + return ret;
> + }
vDPA will support non pinning (shared VM) backend soon[2]. So I guess we
need a flag to be advertised to usersapce then we can conditionly enable
the discard here.
[2] https://www.spinics.net/lists/netdev/msg723944.html
Thanks
>
> v = opaque;
> v->dev = dev;
> @@ -302,6 +313,8 @@ static int vhost_vdpa_cleanup(struct vhost_dev *dev)
> memory_listener_unregister(&v->listener);
>
> dev->opaque = NULL;
> + ram_block_discard_disable(false);
> +
> return 0;
> }
>
On 03.03.21 03:53, Jason Wang wrote:
>
> On 2021/3/3 12:21 上午, David Hildenbrand wrote:
>> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory
>> that used to be discarded will get re-populated and if we
>> discard+re-access memory after mapping+pinning, the pages mapped into the
>> vDPA IOMMU will go out of sync with the actual pages mapped into the user
>> space page tables.
>>
>> Set discarding of RAM broken such that:
>> - virtio-mem and vhost-vdpa run mutually exclusive
>> - virtio-balloon is inhibited and no memory discards will get issued
>>
>> In the future, we might be able to support coordinated discarding of RAM
>> as used by virtio-mem and as planned for VFIO.
>>
>> Cc: Jason Wang <jasowang@redhat.com>
>> Cc: Michael S. Tsirkin <mst@redhat.com>
>> Cc: Cindy Lu <lulu@redhat.com>
>> Signed-off-by: David Hildenbrand <david@redhat.com>
>
>
> Acked-by: Jason Wang <jasowang@redhat.com>
>
>
>> ---
>>
>> Note: I was not actually able to reproduce/test as I fail to get the
>> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, vhost_vdpa,
>> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa devices
>> appear under /sys/bus/vdpa/devices/ or /dev/).
>
>
> The device creation was switched to use vdpa tool that is integrated
> with iproue2[1].
>
> [1]
> https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18
It would be great to document that somewhere if not already done. I only
found older RH documentations that were not aware of that. I'll give it
a try - thanks!
>
>
>>
>> ---
>> hw/virtio/vhost-vdpa.c | 13 +++++++++++++
>> 1 file changed, 13 insertions(+)
>>
>> diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c
>> index 01d2101d09..86058d4041 100644
>> --- a/hw/virtio/vhost-vdpa.c
>> +++ b/hw/virtio/vhost-vdpa.c
>> @@ -278,6 +278,17 @@ static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque)
>> uint64_t features;
>> assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
>> trace_vhost_vdpa_init(dev, opaque);
>> + int ret;
>> +
>> + /*
>> + * Similar to VFIO, we end up pinning all guest memory and have to
>> + * disable discarding of RAM.
>> + */
>> + ret = ram_block_discard_disable(true);
>> + if (ret) {
>> + error_report("Cannot set discarding of RAM broken");
>> + return ret;
>> + }
>
>
> vDPA will support non pinning (shared VM) backend soon[2]. So I guess we
> need a flag to be advertised to usersapce then we can conditionly enable
> the discard here.
I thought that was already the default (because I stumbled over
enforcing guest IOMMU) but was surprised when I had a look at the
implementation.
Having a flag sounds good.
BTW: I assume iommu support is not fully working yet, right? I don't see
special casing for iommu regions, including registering the listener and
updating the mapping.
--
Thanks,
David / dhildenb
On 03.03.21 11:26, David Hildenbrand wrote: > On 03.03.21 03:53, Jason Wang wrote: >> >> On 2021/3/3 12:21 上午, David Hildenbrand wrote: >>> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory >>> that used to be discarded will get re-populated and if we >>> discard+re-access memory after mapping+pinning, the pages mapped into the >>> vDPA IOMMU will go out of sync with the actual pages mapped into the user >>> space page tables. >>> >>> Set discarding of RAM broken such that: >>> - virtio-mem and vhost-vdpa run mutually exclusive >>> - virtio-balloon is inhibited and no memory discards will get issued >>> >>> In the future, we might be able to support coordinated discarding of RAM >>> as used by virtio-mem and as planned for VFIO. >>> >>> Cc: Jason Wang <jasowang@redhat.com> >>> Cc: Michael S. Tsirkin <mst@redhat.com> >>> Cc: Cindy Lu <lulu@redhat.com> >>> Signed-off-by: David Hildenbrand <david@redhat.com> >> >> >> Acked-by: Jason Wang <jasowang@redhat.com> >> >> >>> --- >>> >>> Note: I was not actually able to reproduce/test as I fail to get the >>> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, vhost_vdpa, >>> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa devices >>> appear under /sys/bus/vdpa/devices/ or /dev/). >> >> >> The device creation was switched to use vdpa tool that is integrated >> with iproue2[1]. >> >> [1] >> https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18 > > It would be great to document that somewhere if not already done. I only > found older RH documentations that were not aware of that. I'll give it > a try - thanks! Seems to work just fine: $ sudo ./build/qemu-system-x86_64 -m 2G,maxmem=4G --enable-kvm -object memory-backend-ram,id=mem0,size=2G -device virtio-mem-pci,id=vmem0,memdev=mem0,node=0,requested-size=0G -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-0,id=vhost-vdpa1 -device virtio-net-pci,netdev=vhost-vdpa1,mac=00:e8:ca:33:ba:05,disable-modern=off,page-per-vq=on -nographic qemu-system-x86_64: -device virtio-mem-pci,id=vmem0,memdev=mem0,node=0,requested-size=0G: Discarding RAM is disabled I think the -netdev is always processed/initialized before the "-device virtio-mem-pci", which is why we always fail from virtio-mem code right now and not from vhost-vdpa code. -- Thanks, David / dhildenb
On 2021/3/3 6:26 下午, David Hildenbrand wrote:
> On 03.03.21 03:53, Jason Wang wrote:
>>
>> On 2021/3/3 12:21 上午, David Hildenbrand wrote:
>>> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory
>>> that used to be discarded will get re-populated and if we
>>> discard+re-access memory after mapping+pinning, the pages mapped
>>> into the
>>> vDPA IOMMU will go out of sync with the actual pages mapped into the
>>> user
>>> space page tables.
>>>
>>> Set discarding of RAM broken such that:
>>> - virtio-mem and vhost-vdpa run mutually exclusive
>>> - virtio-balloon is inhibited and no memory discards will get issued
>>>
>>> In the future, we might be able to support coordinated discarding of
>>> RAM
>>> as used by virtio-mem and as planned for VFIO.
>>>
>>> Cc: Jason Wang <jasowang@redhat.com>
>>> Cc: Michael S. Tsirkin <mst@redhat.com>
>>> Cc: Cindy Lu <lulu@redhat.com>
>>> Signed-off-by: David Hildenbrand <david@redhat.com>
>>
>>
>> Acked-by: Jason Wang <jasowang@redhat.com>
>>
>>
>>> ---
>>>
>>> Note: I was not actually able to reproduce/test as I fail to get the
>>> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa,
>>> vhost_vdpa,
>>> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa
>>> devices
>>> appear under /sys/bus/vdpa/devices/ or /dev/).
>>
>>
>> The device creation was switched to use vdpa tool that is integrated
>> with iproue2[1].
>>
>> [1]
>> https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18
>>
>
> It would be great to document that somewhere if not already done. I
> only found older RH documentations that were not aware of that. I'll
> give it a try - thanks!
Will think about this. Which RH doc do you refer here? Is this the
redhat blog?
>
>>
>>
>>>
>>> ---
>>> hw/virtio/vhost-vdpa.c | 13 +++++++++++++
>>> 1 file changed, 13 insertions(+)
>>>
>>> diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c
>>> index 01d2101d09..86058d4041 100644
>>> --- a/hw/virtio/vhost-vdpa.c
>>> +++ b/hw/virtio/vhost-vdpa.c
>>> @@ -278,6 +278,17 @@ static int vhost_vdpa_init(struct vhost_dev
>>> *dev, void *opaque)
>>> uint64_t features;
>>> assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
>>> trace_vhost_vdpa_init(dev, opaque);
>>> + int ret;
>>> +
>>> + /*
>>> + * Similar to VFIO, we end up pinning all guest memory and have to
>>> + * disable discarding of RAM.
>>> + */
>>> + ret = ram_block_discard_disable(true);
>>> + if (ret) {
>>> + error_report("Cannot set discarding of RAM broken");
>>> + return ret;
>>> + }
>>
>>
>> vDPA will support non pinning (shared VM) backend soon[2]. So I guess we
>> need a flag to be advertised to usersapce then we can conditionly enable
>> the discard here.
>
> I thought that was already the default (because I stumbled over
> enforcing guest IOMMU) but was surprised when I had a look at the
> implementation.
>
> Having a flag sounds good.
>
> BTW: I assume iommu support is not fully working yet, right? I don't
> see special casing for iommu regions, including registering the
> listener and updating the mapping.
It's not yet implemented. Yes, it's something like what VFIO did right
now, e.g to use IOMMU notifiers.
Thanks
On 04.03.21 10:32, Jason Wang wrote: > > On 2021/3/3 6:26 下午, David Hildenbrand wrote: >> On 03.03.21 03:53, Jason Wang wrote: >>> >>> On 2021/3/3 12:21 上午, David Hildenbrand wrote: >>>> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory >>>> that used to be discarded will get re-populated and if we >>>> discard+re-access memory after mapping+pinning, the pages mapped >>>> into the >>>> vDPA IOMMU will go out of sync with the actual pages mapped into the >>>> user >>>> space page tables. >>>> >>>> Set discarding of RAM broken such that: >>>> - virtio-mem and vhost-vdpa run mutually exclusive >>>> - virtio-balloon is inhibited and no memory discards will get issued >>>> >>>> In the future, we might be able to support coordinated discarding of >>>> RAM >>>> as used by virtio-mem and as planned for VFIO. >>>> >>>> Cc: Jason Wang <jasowang@redhat.com> >>>> Cc: Michael S. Tsirkin <mst@redhat.com> >>>> Cc: Cindy Lu <lulu@redhat.com> >>>> Signed-off-by: David Hildenbrand <david@redhat.com> >>> >>> >>> Acked-by: Jason Wang <jasowang@redhat.com> >>> >>> >>>> --- >>>> >>>> Note: I was not actually able to reproduce/test as I fail to get the >>>> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, >>>> vhost_vdpa, >>>> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa >>>> devices >>>> appear under /sys/bus/vdpa/devices/ or /dev/). >>> >>> >>> The device creation was switched to use vdpa tool that is integrated >>> with iproue2[1]. >>> >>> [1] >>> https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18 >>> >> >> It would be great to document that somewhere if not already done. I >> only found older RH documentations that were not aware of that. I'll >> give it a try - thanks! > > > Will think about this. Which RH doc do you refer here? Is this the > redhat blog? https://www.redhat.com/en/blog/hands-vdpa-what-do-you-do-when-you-aint-got-hardware As it's supposed to be from October 14, 2020 I was surprised to not get it running (even with older kernels IIRC). -- Thanks, David / dhildenb
On 2021/3/4 5:34 下午, David Hildenbrand wrote: > On 04.03.21 10:32, Jason Wang wrote: >> >> On 2021/3/3 6:26 下午, David Hildenbrand wrote: >>> On 03.03.21 03:53, Jason Wang wrote: >>>> >>>> On 2021/3/3 12:21 上午, David Hildenbrand wrote: >>>>> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. >>>>> Memory >>>>> that used to be discarded will get re-populated and if we >>>>> discard+re-access memory after mapping+pinning, the pages mapped >>>>> into the >>>>> vDPA IOMMU will go out of sync with the actual pages mapped into the >>>>> user >>>>> space page tables. >>>>> >>>>> Set discarding of RAM broken such that: >>>>> - virtio-mem and vhost-vdpa run mutually exclusive >>>>> - virtio-balloon is inhibited and no memory discards will get issued >>>>> >>>>> In the future, we might be able to support coordinated discarding of >>>>> RAM >>>>> as used by virtio-mem and as planned for VFIO. >>>>> >>>>> Cc: Jason Wang <jasowang@redhat.com> >>>>> Cc: Michael S. Tsirkin <mst@redhat.com> >>>>> Cc: Cindy Lu <lulu@redhat.com> >>>>> Signed-off-by: David Hildenbrand <david@redhat.com> >>>> >>>> >>>> Acked-by: Jason Wang <jasowang@redhat.com> >>>> >>>> >>>>> --- >>>>> >>>>> Note: I was not actually able to reproduce/test as I fail to get the >>>>> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, >>>>> vhost_vdpa, >>>>> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa >>>>> devices >>>>> appear under /sys/bus/vdpa/devices/ or /dev/). >>>> >>>> >>>> The device creation was switched to use vdpa tool that is integrated >>>> with iproue2[1]. >>>> >>>> [1] >>>> https://git.kernel.org/pub/scm/network/iproute2/iproute2-next.git/commit/?id=143610383da51e1f868c6d5a2a5e2fb552293d18 >>>> >>>> >>> >>> It would be great to document that somewhere if not already done. I >>> only found older RH documentations that were not aware of that. I'll >>> give it a try - thanks! >> >> >> Will think about this. Which RH doc do you refer here? Is this the >> redhat blog? > > https://www.redhat.com/en/blog/hands-vdpa-what-do-you-do-when-you-aint-got-hardware > > > As it's supposed to be from October 14, 2020 I was surprised to not > get it running (even with older kernels IIRC). Right, the mgmt API is just merged. Will try to see if we can fix the blog. Thanks
On 02.03.21 17:21, David Hildenbrand wrote:
> Similar to VFIO, vDPA will go ahead an map+pin all guest memory. Memory
> that used to be discarded will get re-populated and if we
> discard+re-access memory after mapping+pinning, the pages mapped into the
> vDPA IOMMU will go out of sync with the actual pages mapped into the user
> space page tables.
>
> Set discarding of RAM broken such that:
> - virtio-mem and vhost-vdpa run mutually exclusive
> - virtio-balloon is inhibited and no memory discards will get issued
>
> In the future, we might be able to support coordinated discarding of RAM
> as used by virtio-mem and as planned for VFIO.
>
> Cc: Jason Wang <jasowang@redhat.com>
> Cc: Michael S. Tsirkin <mst@redhat.com>
> Cc: Cindy Lu <lulu@redhat.com>
> Signed-off-by: David Hildenbrand <david@redhat.com>
> ---
>
> Note: I was not actually able to reproduce/test as I fail to get the
> vdpa_sim/vdpa_sim_net running on upstream Linux (whetever vdpa, vhost_vdpa,
> vdpa_sim, vdpa_sim_net modules I probe, and in which order, no vdpa devices
> appear under /sys/bus/vdpa/devices/ or /dev/).
>
> ---
> hw/virtio/vhost-vdpa.c | 13 +++++++++++++
> 1 file changed, 13 insertions(+)
>
> diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c
> index 01d2101d09..86058d4041 100644
> --- a/hw/virtio/vhost-vdpa.c
> +++ b/hw/virtio/vhost-vdpa.c
> @@ -278,6 +278,17 @@ static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque)
> uint64_t features;
> assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
> trace_vhost_vdpa_init(dev, opaque);
> + int ret;
> +
> + /*
> + * Similar to VFIO, we end up pinning all guest memory and have to
> + * disable discarding of RAM.
> + */
> + ret = ram_block_discard_disable(true);
> + if (ret) {
> + error_report("Cannot set discarding of RAM broken");
> + return ret;
> + }
>
> v = opaque;
> v->dev = dev;
> @@ -302,6 +313,8 @@ static int vhost_vdpa_cleanup(struct vhost_dev *dev)
> memory_listener_unregister(&v->listener);
>
> dev->opaque = NULL;
> + ram_block_discard_disable(false);
> +
> return 0;
> }
>
>
@MST, do you have this on your radar? thanks
--
Thanks,
David / dhildenb
© 2016 - 2026 Red Hat, Inc.