On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote:
> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the
> 64-bit pci io window, failing during boot with errors like:
Well. Why people would use *that* ubuntu version is not clear to me.
It's *loooooong* out of support. Even the LTS version from that year
(16.04) is not supported any more. But it is at least available for
download still, so I gave it a spin.
Turns out it apparently can't deal with PCI bars mapped above 16TB (aka
44 phys-bits). Test patch below.
take care,
Gerd
diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c
index bb44dc296047..a43876a931c9 100644
--- a/src/fw/pciinit.c
+++ b/src/fw/pciinit.c
@@ -1189,11 +1189,16 @@ pci_setup(void)
if (CPUPhysBits) {
pci_mem64_top = 1LL << CPUPhysBits;
- if (CPUPhysBits > 46) {
- // Old linux kernels have trouble dealing with more than 46
- // phys-bits, so avoid that for now. Seems to be a bug in the
- // virtio-pci driver. Reported: centos-7, ubuntu-18.04
- pci_mem64_top = 1LL << 46;
+ if (CPUPhysBits > 44) {
+ // Old linux kernels have trouble dealing with more than 44/46
+ // phys-bits. Seems to be a bug in the virtio-pci driver.
+ // 46: centos-7, ubuntu-18.04
+ // 44: ubuntu-16.04
+ // Limit the used address space to mitigate the bug, except we are
+ // running in a guest with more than 1TB of memory installed.
+ if (RamSizeOver4G < (1LL << 40)) {
+ pci_mem64_top = 1LL << 44;
+ }
}
}
_______________________________________________
SeaBIOS mailing list -- seabios@seabios.org
To unsubscribe send an email to seabios-leave@seabios.org
Hi, Am 21.06.24 um 14:05 schrieb Gerd Hoffmann: > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: >> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the >> 64-bit pci io window, failing during boot with errors like: > > Well. Why people would use *that* ubuntu version is not clear to me. > It's *loooooong* out of support. Even the LTS version from that year > (16.04) is not supported any more. But it is at least available for > download still, so I gave it a spin. > > Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > 44 phys-bits). Test patch below. > > take care, > Gerd > > diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c > index bb44dc296047..a43876a931c9 100644 > --- a/src/fw/pciinit.c > +++ b/src/fw/pciinit.c > @@ -1189,11 +1189,16 @@ pci_setup(void) > > if (CPUPhysBits) { > pci_mem64_top = 1LL << CPUPhysBits; > - if (CPUPhysBits > 46) { > - // Old linux kernels have trouble dealing with more than 46 > - // phys-bits, so avoid that for now. Seems to be a bug in the > - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 > - pci_mem64_top = 1LL << 46; > + if (CPUPhysBits > 44) { > + // Old linux kernels have trouble dealing with more than 44/46 > + // phys-bits. Seems to be a bug in the virtio-pci driver. > + // 46: centos-7, ubuntu-18.04 > + // 44: ubuntu-16.04 > + // Limit the used address space to mitigate the bug, except we are > + // running in a guest with more than 1TB of memory installed. > + if (RamSizeOver4G < (1LL << 40)) { > + pci_mem64_top = 1LL << 44; > + } > } > } > we've also had two reports about issues with 32-bit guests now[0][1] (both had '-m 4096' in the QEMU commandline), which I was able to reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The QEMU commandline is below[2]. Unfortunately, it still fails to boot, even with the "limit address space used for pci devices, part two" patch applied on top of rel-1.16.3 (and using current QEMU master). It boots fine with '-m 3583', but not anymore with '-m 3584'. So is bumping the limit for the check necessary after all? Best Regards, Fiona [0]: https://forum.proxmox.com/threads/150217/ [1]: https://forum.proxmox.com/threads/149772/post-683562 [2]: > ./qemu-system-x86_64 \ > -accel 'kvm' \ > -cpu 'host' \ > -chardev 'socket,id=qmp,path=/var/run/qemu-server/121.qmp,server=on,wait=off' \ > -mon 'chardev=qmp,mode=control' \ > -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ > -mon 'chardev=qmp-event,mode=control' \ > -pidfile /var/run/qemu-server/121.pid \ > -smp '4,sockets=1,cores=4,maxcpus=4' \ > -nodefaults \ > -vnc 'unix:/var/run/qemu-server/121.vnc,password=on' \ > -m 4096 \ > -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ > -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ > -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ > -drive 'file=/dev/lvmthinbig/vm-121-disk-0,if=none,id=drive-scsi0,format=raw,cache=none,aio=io_uring,detect-zeroes=on' \ > -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ > -machine 'type=pc' _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Am 12.07.24 um 14:24 schrieb Fiona Ebner: > Hi, > > Am 21.06.24 um 14:05 schrieb Gerd Hoffmann: >> On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: >>> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the >>> 64-bit pci io window, failing during boot with errors like: >> >> Well. Why people would use *that* ubuntu version is not clear to me. >> It's *loooooong* out of support. Even the LTS version from that year >> (16.04) is not supported any more. But it is at least available for >> download still, so I gave it a spin. >> >> Turns out it apparently can't deal with PCI bars mapped above 16TB (aka >> 44 phys-bits). Test patch below. >> >> take care, >> Gerd >> >> diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c >> index bb44dc296047..a43876a931c9 100644 >> --- a/src/fw/pciinit.c >> +++ b/src/fw/pciinit.c >> @@ -1189,11 +1189,16 @@ pci_setup(void) >> >> if (CPUPhysBits) { >> pci_mem64_top = 1LL << CPUPhysBits; >> - if (CPUPhysBits > 46) { >> - // Old linux kernels have trouble dealing with more than 46 >> - // phys-bits, so avoid that for now. Seems to be a bug in the >> - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 >> - pci_mem64_top = 1LL << 46; >> + if (CPUPhysBits > 44) { >> + // Old linux kernels have trouble dealing with more than 44/46 >> + // phys-bits. Seems to be a bug in the virtio-pci driver. >> + // 46: centos-7, ubuntu-18.04 >> + // 44: ubuntu-16.04 >> + // Limit the used address space to mitigate the bug, except we are >> + // running in a guest with more than 1TB of memory installed. >> + if (RamSizeOver4G < (1LL << 40)) { >> + pci_mem64_top = 1LL << 44; >> + } >> } >> } >> > > we've also had two reports about issues with 32-bit guests now[0][1] > (both had '-m 4096' in the QEMU commandline), which I was able to > reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The > QEMU commandline is below[2]. > > Unfortunately, it still fails to boot, even with the "limit address > space used for pci devices, part two" patch applied on top of rel-1.16.3 > (and using current QEMU master). > > It boots fine with '-m 3583', but not anymore with '-m 3584'. So is > bumping the limit for the check necessary after all? > Sorry for bumping this again, but we got a new report about OpenBSD 7.6 being affected too (see [3] and the following posts in the forum thread). In particular, the user reports a regression with PCIe passthrough failing and seeing "mem address conflict" messages in dmesg. I could reproduce the latter using [4], i.e. seeing the messages and checked that it is caused by 96a8d130 ("be less conservative with the 64bit pci io window"). The user reported that using 2 GiB of RAM makes the passthrough work again and the messages disappear. The messages are still present when I test with current SeaBIOS master 2424e4c0 ("esp-scsi: indicate acceptance of MESSAGE IN phase data") and QEMU v9.1.1. Note that I also get the messages without the vfio-pci device, but since I wasn't able to reproduce the passthrough failure, it is what I have to work with. Best Regards, Fiona [3]: https://forum.proxmox.com/threads/149772/post-713842 [4]: https://ftp2.eu.openbsd.org/pub/OpenBSD/7.6/amd64/install76.iso QEMU commandline with which I observed the messages: > ./qemu-system-x86_64 \ > -accel kvm \ > -pidfile /var/run/qemu-server/122.pid \ > -chardev 'socket,id=qmp,path=/var/run/qemu-server/122.qmp,server=on,wait=off' \ > -mon 'chardev=qmp,mode=control' \ > -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ > -mon 'chardev=qmp-event,mode=control' \ > -smp '8,sockets=1,cores=8,maxcpus=8' \ > -nodefaults \ > -vnc 'unix:/var/run/qemu-server/122.vnc,password=on' \ > -cpu host \ > -m 4096 \ > -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ > -device 'vfio-pci,host=0000:00:1b.0,id=hostpci0,bus=pci.0,addr=0x10' \ > -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ > -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ > -drive 'file=/dev/zvol/zfs/vm-122-disk-0,if=none,id=drive-scsi0,format=raw' \ > -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ > -machine 'type=pc' > > [0]: https://forum.proxmox.com/threads/150217/ > [1]: https://forum.proxmox.com/threads/149772/post-683562 > [2]: > >> ./qemu-system-x86_64 \ >> -accel 'kvm' \ >> -cpu 'host' \ >> -chardev 'socket,id=qmp,path=/var/run/qemu-server/121.qmp,server=on,wait=off' \ >> -mon 'chardev=qmp,mode=control' \ >> -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ >> -mon 'chardev=qmp-event,mode=control' \ >> -pidfile /var/run/qemu-server/121.pid \ >> -smp '4,sockets=1,cores=4,maxcpus=4' \ >> -nodefaults \ >> -vnc 'unix:/var/run/qemu-server/121.vnc,password=on' \ >> -m 4096 \ >> -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ >> -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ >> -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ >> -drive 'file=/dev/lvmthinbig/vm-121-disk-0,if=none,id=drive-scsi0,format=raw,cache=none,aio=io_uring,detect-zeroes=on' \ >> -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ >> -machine 'type=pc' > > _______________________________________________ > SeaBIOS mailing list -- seabios@seabios.org > To unsubscribe send an email to seabios-leave@seabios.org > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Hi, > we've also had two reports about issues with 32-bit guests now[0][1] > (both had '-m 4096' in the QEMU commandline), which I was able to > reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The > QEMU commandline is below[2]. > > Unfortunately, it still fails to boot, even with the "limit address > space used for pci devices, part two" patch applied on top of rel-1.16.3 > (and using current QEMU master). Reproduces here. Seems there is something seriously wrong in that debian kernel. Even when going for 36 phys bits (-cpu host,host-phys-bits-limit=36), which is the physical address space of PAE mode when it was introduced by the pentium pro, the guest fails to boot. That *really* should not happen. I guess what you are seeing here is i386 support in the linux kernel starting to bitrot due to everybody moving to 64-bit ... take care, Gerd _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On 7/17/24 09:41, Gerd Hoffmann wrote: > Hi, > >> we've also had two reports about issues with 32-bit guests now[0][1] >> (both had '-m 4096' in the QEMU commandline), which I was able to >> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The >> QEMU commandline is below[2]. >> >> Unfortunately, it still fails to boot, even with the "limit address >> space used for pci devices, part two" patch applied on top of rel-1.16.3 >> (and using current QEMU master). > > Reproduces here. Seems there is something seriously wrong in that > debian kernel. We can complain about every kernel out there being "wrong", but ultimately what good is a BIOS if it is not able to boot reliably even older OSes? What do we lose by just reverting things to the state that actually worked? > > Even when going for 36 phys bits (-cpu host,host-phys-bits-limit=36), > which is the physical address space of PAE mode when it was introduced > by the pentium pro, the guest fails to boot. That *really* should not > happen. > > I guess what you are seeing here is i386 support in the linux kernel > starting to bitrot due to everybody moving to 64-bit ... > > take care, > Gerd > > _______________________________________________ > SeaBIOS mailing list -- seabios@seabios.org > To unsubscribe send an email to seabios-leave@seabios.org _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Thu, Oct 24, 2024 at 12:36:16PM +0200, Claudio Fontana wrote: > What do we lose by just reverting things to the state that actually worked? Well, the world is moving forward, devices getting more memory and larger pci bars. GPUs can have multi-GB pci bars these days. nvidia has a NPU with a 128 GB pci bar. Note this is larger than the whole (initial) PAE address space. I want seabios supporting that kind of devices well. The range of configurations becomes more and more wide, with huge devices as mentioned above on one end and old 32-bit guests on the other end. Have seabios automatically setup itself turned out to be a hard problem (as this thread clearly shows). We already have a runtime config switch to force 32-bit friendly setup (turn off long mode support in the vcpu). We already have a heuristic to select the 32-bit friendly setup, which right now is simply "no memory above 4G is present". The options I see to move this forward: (1) Tweak the heuristic. (1a) Raise the memory limit. (1b) Maybe also look at other things such as the machine type. The 'pc' machine type does not support pci express, so it is highly unlikely that devices with huge pci bars will be used with it and we could use that as another hint to better select the 32-bit friendly setup. (1c) Other ideas? (2) Add a compile-time option (CONFIG_...) to force 32-bit friendly setup unconditionally. Comments? take care, Gerd PS: sorry for the long delay, was offline in october and had quite a email backlog afterwards ... _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Tue, Nov 12, 2024 at 12:33:58PM +0100, Gerd Hoffmann wrote: > We already have a runtime config switch to force 32-bit friendly setup > (turn off long mode support in the vcpu). > > We already have a heuristic to select the 32-bit friendly setup, which > right now is simply "no memory above 4G is present". > > The options I see to move this forward: I think there is another option: - Revert the commits that added the pci sparse mapping heuristic (starting with commit 96a8d130) and replace with an explicit indicator from qemu when seabios should attempt to allocate sparse pci mappings. Alas, this isn't a great option as commit 96a8d130 has been around for so long that changing it now seems just as likely to create new regressions. > (1) Tweak the heuristic. > (1a) Raise the memory limit. I think this seems okay to me (or at least, least bad). If there are a large number of guests that require sparse pci mappings with low guest memory, then perhaps we could add a new "etc/pci_sparse_mapping" config option to allow users to force a different setting. > (1b) Maybe also look at other things such as the machine type. The > 'pc' machine type does not support pci express, so it is highly > unlikely that devices with huge pci bars will be used with it and > we could use that as another hint to better select the 32-bit > friendly setup. FWIW, this doesn't sound like a great option. I fear the more heuristics we add the more likely we'll get subtle breakage. > (1c) Other ideas? > > (2) Add a compile-time option (CONFIG_...) to force 32-bit friendly > setup unconditionally. FWIW, this doesn't sound like a great option. If admins don't want to force a guest to work by changing qemu options (eg, turn off long mode) then they also wont want to manually select a different SeaBIOS binary. Cheers, -Kevin _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Tue, 12 Nov 2024 12:33:58 +0100 Gerd Hoffmann <kraxel@redhat.com> wrote: > On Thu, Oct 24, 2024 at 12:36:16PM +0200, Claudio Fontana wrote: > > > What do we lose by just reverting things to the state that actually worked? > > Well, the world is moving forward, devices getting more memory and > larger pci bars. GPUs can have multi-GB pci bars these days. nvidia > has a NPU with a 128 GB pci bar. Note this is larger than the whole > (initial) PAE address space. I want seabios supporting that kind of > devices well. > > The range of configurations becomes more and more wide, with huge > devices as mentioned above on one end and old 32-bit guests on the > other end. Have seabios automatically setup itself turned out to be a > hard problem (as this thread clearly shows). > > We already have a runtime config switch to force 32-bit friendly setup > (turn off long mode support in the vcpu). > > We already have a heuristic to select the 32-bit friendly setup, which > right now is simply "no memory above 4G is present". > > The options I see to move this forward: > > (1) Tweak the heuristic. > (1a) Raise the memory limit. that would be improvement for legacy but worsen modern guest situation. Perhaps we at the point where heuristic can't help anymore. > (1b) Maybe also look at other things such as the machine type. The > 'pc' machine type does not support pci express, so it is highly > unlikely that devices with huge pci bars will be used with it and > we could use that as another hint to better select the 32-bit > friendly setup. If I'm not wrong it's possible to assign a device with large PCI bar to 'pc' machine. so just machine type won't be sufficient. > (1c) Other ideas? Perhaps a fwcfg file '32_bit_guest' passed to Seabios can serve as switch. But given Seabios and QEMU have no clue about OS nor should they, onus would be on management to configure qemu with compat option. Which leads to #2 > (2) Add a compile-time option (CONFIG_...) to force 32-bit friendly > setup unconditionally. Given that mgmt would have configure QEMU one way or another to handle legacy guests. This option might be preferable to runtime switch. That also could help us drop some heuristics and pick more suitable defaults for 32 and 64 bit variant instead of a poorly working compromise. > Comments? > > take care, > Gerd > > PS: sorry for the long delay, was offline in october and had quite a > email backlog afterwards ... > > _______________________________________________ > SeaBIOS mailing list -- seabios@seabios.org > To unsubscribe send an email to seabios-leave@seabios.org > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
> > We already have a heuristic to select the 32-bit friendly setup, which > > right now is simply "no memory above 4G is present". > > > > The options I see to move this forward: > > > > (1) Tweak the heuristic. > > (1a) Raise the memory limit. > > that would be improvement for legacy but worsen modern guest situation. Well, guests having devices with huge memory bars are unlikely to have only small amounts of main memory, so I think this shouldn't be too bad. Depends a bit on the new limit chosen (when going down that route) obviously. > Perhaps we at the point where heuristic can't help anymore. Could be. > > (1b) Maybe also look at other things such as the machine type. The > > 'pc' machine type does not support pci express, so it is highly > > unlikely that devices with huge pci bars will be used with it and > > we could use that as another hint to better select the 32-bit > > friendly setup. > > If I'm not wrong it's possible to assign a device with large PCI bar to 'pc' machine. > so just machine type won't be sufficient. It's possible, sure. But having devices with multi-GB bars is a relatively recent thing (when compared to the age of the pci express 1.0 spec). I don't think many (if any) legacy pci devices with multi-GB bars exist. Placing pci express devices into a legacy pci slot of the pc machine type is possible, but it comes with a number of quirks like not being able to access all pci config space. It's one of the reasons why we added the pcie-capable q35 chipset emulation to qemu. Which why I think going by machine type would be a sensible idea. > > (1c) Other ideas? > > Perhaps a fwcfg file '32_bit_guest' passed to Seabios can serve as switch. > But given Seabios and QEMU have no clue about OS nor should they, > onus would be on management to configure qemu with compat option. Well, we have that already. The config option is "turn off long mode support". Drawback is that this way seabios can't choose the default guest setup, qemu default for long mode is enabled and it is not going to change. The other way around would make sense, i.e. have seabios default to 32-bit guest setup and switch into 64-bit guest mode with a fwcfg file (see also the reply by Max). > > (2) Add a compile-time option (CONFIG_...) to force 32-bit friendly > > setup unconditionally. > > Given that mgmt would have configure QEMU one way or another to handle > legacy guests. This option might be preferable to runtime switch. Well, a runtime switch surely is better than a compile time switch. That's why seabios checks long mode support of the cpu already. But I don't see an agreement emerging on what the default should be ... The vast majority of guests should be 64-bit these days. I think when it comes to guests officially supported on RHEL hosts there are no 32-bit OSes left these days. So defaulting to 64-bit makes sense here. For some of the providers discussing here there are apparently enough 32-bit guests still being active that requiring manual config invention to keep them working is too much of a customer support burden, even if it is only a small minority. So they would probably prefer a 32-bit default. Maybe we should have a 'guest-hint' fwcfg file, carrying either '32' or '64' as value to choose the default at runtime, and a CONFIG_DEFAULT_GUEST_HINT compile time option which decides what to do in case the fwcfg file is not present. take care, Gerd _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On 11/12, Igor Mammedov wrote: > !-------------------------------------------------------------------| > This Message Is From an External Sender > This message came from outside your organization. > |-------------------------------------------------------------------! > > On Tue, 12 Nov 2024 12:33:58 +0100 > Gerd Hoffmann <kraxel@redhat.com> wrote: > > > On Thu, Oct 24, 2024 at 12:36:16PM +0200, Claudio Fontana wrote: > > > > > What do we lose by just reverting things to the state that actually worked? > > > > Well, the world is moving forward, devices getting more memory and > > larger pci bars. GPUs can have multi-GB pci bars these days. nvidia > > has a NPU with a 128 GB pci bar. Note this is larger than the whole > > (initial) PAE address space. I want seabios supporting that kind of > > devices well. > > > > The range of configurations becomes more and more wide, with huge > > devices as mentioned above on one end and old 32-bit guests on the > > other end. Have seabios automatically setup itself turned out to be a > > hard problem (as this thread clearly shows). > > > > We already have a runtime config switch to force 32-bit friendly setup > > (turn off long mode support in the vcpu). > > > > We already have a heuristic to select the 32-bit friendly setup, which > > right now is simply "no memory above 4G is present". > > > > The options I see to move this forward: > > > > (1) Tweak the heuristic. > > (1a) Raise the memory limit. It seems to me that the thinking here is backwards, the change to the heuristic behavior was a regression from the standpoint of QEMU, an upgraded bios bricked existing guests. Should it not be the case that '32bit-friendly' is enabled by default, and a runtime config switch is provided to expose the expanded PCI BAR space? If you want the new thing - supply the new flag? -- Max Tottenham | Senior Software Engineer /(* Akamai Technologies _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, 12 Jul 2024 14:24:40 +0200 Fiona Ebner <f.ebner@proxmox.com> wrote: > Hi, > > Am 21.06.24 um 14:05 schrieb Gerd Hoffmann: > > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > >> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > >> 64-bit pci io window, failing during boot with errors like: > > > > Well. Why people would use *that* ubuntu version is not clear to me. > > It's *loooooong* out of support. Even the LTS version from that year > > (16.04) is not supported any more. But it is at least available for > > download still, so I gave it a spin. > > > > Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > > 44 phys-bits). Test patch below. > > > > take care, > > Gerd > > > > diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c > > index bb44dc296047..a43876a931c9 100644 > > --- a/src/fw/pciinit.c > > +++ b/src/fw/pciinit.c > > @@ -1189,11 +1189,16 @@ pci_setup(void) > > > > if (CPUPhysBits) { > > pci_mem64_top = 1LL << CPUPhysBits; > > - if (CPUPhysBits > 46) { > > - // Old linux kernels have trouble dealing with more than 46 > > - // phys-bits, so avoid that for now. Seems to be a bug in the > > - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 > > - pci_mem64_top = 1LL << 46; > > + if (CPUPhysBits > 44) { > > + // Old linux kernels have trouble dealing with more than 44/46 > > + // phys-bits. Seems to be a bug in the virtio-pci driver. > > + // 46: centos-7, ubuntu-18.04 > > + // 44: ubuntu-16.04 > > + // Limit the used address space to mitigate the bug, except we are > > + // running in a guest with more than 1TB of memory installed. > > + if (RamSizeOver4G < (1LL << 40)) { > > + pci_mem64_top = 1LL << 44; > > + } > > } > > } > > > > we've also had two reports about issues with 32-bit guests now[0][1] > (both had '-m 4096' in the QEMU commandline), which I was able to > reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The > QEMU commandline is below[2]. is it also reproducible with upstream kernel? if yes, it would be better to fix that on guest kernel side, rather than in SeaBIOS which has no idea what guest OS is going to be running after it. > Unfortunately, it still fails to boot, even with the "limit address > space used for pci devices, part two" patch applied on top of rel-1.16.3 > (and using current QEMU master). > > It boots fine with '-m 3583', but not anymore with '-m 3584'. So is > bumping the limit for the check necessary after all? > > Best Regards, > Fiona > > [0]: https://forum.proxmox.com/threads/150217/ > [1]: https://forum.proxmox.com/threads/149772/post-683562 well, with current QEMU master branch (I assume it haven't even got topic patch yet) ./qemu-system-x86_64 --enable-kvm -cpu host -smp 4 -snapshot -m 4096 -M pc-i440fx-6.1 winxp_x86_build2600 following CLI boots just fine an 32-bit XP on Haswell host. Also using 'host' with ancient OS, is basically asking for trouble. If it works for user with old 'XP contemporary' CPU model, user should use that instead or workarounds (aka it's management task to configure CLI in compatible with OS manner). From suspicions config options in that post I see 'viommu' and 'vmgenid', are you sure XP even knows what to do with that, perhaps it triggers BSOD. > [2]: > > > ./qemu-system-x86_64 \ > > -accel 'kvm' \ > > -cpu 'host' \ > > -chardev 'socket,id=qmp,path=/var/run/qemu-server/121.qmp,server=on,wait=off' \ > > -mon 'chardev=qmp,mode=control' \ > > -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ > > -mon 'chardev=qmp-event,mode=control' \ > > -pidfile /var/run/qemu-server/121.pid \ > > -smp '4,sockets=1,cores=4,maxcpus=4' \ > > -nodefaults \ > > -vnc 'unix:/var/run/qemu-server/121.vnc,password=on' \ > > -m 4096 \ > > -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ > > -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ > > -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ > > -drive 'file=/dev/lvmthinbig/vm-121-disk-0,if=none,id=drive-scsi0,format=raw,cache=none,aio=io_uring,detect-zeroes=on' \ > > -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ > > -machine 'type=pc' can you try to boot with q35 + intel iommu enabled and/or force virtio into legacy mode. > > _______________________________________________ > SeaBIOS mailing list -- seabios@seabios.org > To unsubscribe send an email to seabios-leave@seabios.org > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Am 12.07.24 um 15:26 schrieb Igor Mammedov: > On Fri, 12 Jul 2024 14:24:40 +0200 > Fiona Ebner <f.ebner@proxmox.com> wrote: > >> Hi, >> >> Am 21.06.24 um 14:05 schrieb Gerd Hoffmann: >>> On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: >>>> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the >>>> 64-bit pci io window, failing during boot with errors like: >>> >>> Well. Why people would use *that* ubuntu version is not clear to me. >>> It's *loooooong* out of support. Even the LTS version from that year >>> (16.04) is not supported any more. But it is at least available for >>> download still, so I gave it a spin. >>> >>> Turns out it apparently can't deal with PCI bars mapped above 16TB (aka >>> 44 phys-bits). Test patch below. >>> >>> take care, >>> Gerd >>> >>> diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c >>> index bb44dc296047..a43876a931c9 100644 >>> --- a/src/fw/pciinit.c >>> +++ b/src/fw/pciinit.c >>> @@ -1189,11 +1189,16 @@ pci_setup(void) >>> >>> if (CPUPhysBits) { >>> pci_mem64_top = 1LL << CPUPhysBits; >>> - if (CPUPhysBits > 46) { >>> - // Old linux kernels have trouble dealing with more than 46 >>> - // phys-bits, so avoid that for now. Seems to be a bug in the >>> - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 >>> - pci_mem64_top = 1LL << 46; >>> + if (CPUPhysBits > 44) { >>> + // Old linux kernels have trouble dealing with more than 44/46 >>> + // phys-bits. Seems to be a bug in the virtio-pci driver. >>> + // 46: centos-7, ubuntu-18.04 >>> + // 44: ubuntu-16.04 >>> + // Limit the used address space to mitigate the bug, except we are >>> + // running in a guest with more than 1TB of memory installed. >>> + if (RamSizeOver4G < (1LL << 40)) { >>> + pci_mem64_top = 1LL << 44; >>> + } >>> } >>> } >>> >> >> we've also had two reports about issues with 32-bit guests now[0][1] >> (both had '-m 4096' in the QEMU commandline), which I was able to >> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The >> QEMU commandline is below[2]. > > is it also reproducible with upstream kernel? > if yes, it would be better to fix that on guest kernel side, > rather than in SeaBIOS which has no idea what guest OS is going to be > running after it. > Turns out it's only kernels with PAE. The 6.1 Debian kernel without PAE boots fine (but the PAE was the one installed by default). I built a 6.10 kernel and it also boots fine, a 6.10 build with PAE doesn't. >> Unfortunately, it still fails to boot, even with the "limit address >> space used for pci devices, part two" patch applied on top of rel-1.16.3 >> (and using current QEMU master). >> >> It boots fine with '-m 3583', but not anymore with '-m 3584'. So is >> bumping the limit for the check necessary after all? >> >> Best Regards, >> Fiona >> >> [0]: https://forum.proxmox.com/threads/150217/ > >> [1]: https://forum.proxmox.com/threads/149772/post-683562 > well, with current QEMU master branch (I assume it haven't even got topic patch yet) > ./qemu-system-x86_64 --enable-kvm -cpu host -smp 4 -snapshot -m 4096 -M pc-i440fx-6.1 winxp_x86_build2600 > following CLI boots just fine an 32-bit XP on Haswell host. > > Also using 'host' with ancient OS, is basically asking for trouble. > If it works for user with old 'XP contemporary' CPU model, user should use that > instead or workarounds (aka it's management task to configure CLI in > compatible with OS manner). > > From suspicions config options in that post I see 'viommu' and 'vmgenid', > are you sure XP even knows what to do with that, perhaps it triggers BSOD. > No idea why the user enabled viommu for XP. vmgenid is added by our management stack by default and I don't remember it having ever caused problems. The user said the VM booted fine after adding lm=off to the CPU options. I'm not super interested in the Windows case to be honest O:) >> [2]: >> >>> ./qemu-system-x86_64 \ >>> -accel 'kvm' \ >>> -cpu 'host' \ >>> -chardev 'socket,id=qmp,path=/var/run/qemu-server/121.qmp,server=on,wait=off' \ >>> -mon 'chardev=qmp,mode=control' \ >>> -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ >>> -mon 'chardev=qmp-event,mode=control' \ >>> -pidfile /var/run/qemu-server/121.pid \ >>> -smp '4,sockets=1,cores=4,maxcpus=4' \ >>> -nodefaults \ >>> -vnc 'unix:/var/run/qemu-server/121.vnc,password=on' \ >>> -m 4096 \ >>> -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ >>> -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ >>> -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ >>> -drive 'file=/dev/lvmthinbig/vm-121-disk-0,if=none,id=drive-scsi0,format=raw,cache=none,aio=io_uring,detect-zeroes=on' \ >>> -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ >>> -machine 'type=pc' > > can you try to boot with q35 + intel iommu enabled > and/or force virtio into legacy mode. > I tried > -device 'intel-iommu' \ > -machine 'type=q35' and intel_iommu=on in the guest's kernel cmdline, but it still failed with kernels with PAE. Appending 'disable-modern=on,disable-legacy=off' to the virtio-scsi-pci line made it work however (also with pc machine) :) Best Regards, Fiona _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Tue, 16 Jul 2024 13:56:08 +0200 Fiona Ebner <f.ebner@proxmox.com> wrote: > Am 12.07.24 um 15:26 schrieb Igor Mammedov: > > On Fri, 12 Jul 2024 14:24:40 +0200 > > Fiona Ebner <f.ebner@proxmox.com> wrote: > > > >> Hi, > >> > >> Am 21.06.24 um 14:05 schrieb Gerd Hoffmann: > >>> On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > >>>> Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > >>>> 64-bit pci io window, failing during boot with errors like: > >>> > >>> Well. Why people would use *that* ubuntu version is not clear to me. > >>> It's *loooooong* out of support. Even the LTS version from that year > >>> (16.04) is not supported any more. But it is at least available for > >>> download still, so I gave it a spin. > >>> > >>> Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > >>> 44 phys-bits). Test patch below. > >>> > >>> take care, > >>> Gerd > >>> > >>> diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c > >>> index bb44dc296047..a43876a931c9 100644 > >>> --- a/src/fw/pciinit.c > >>> +++ b/src/fw/pciinit.c > >>> @@ -1189,11 +1189,16 @@ pci_setup(void) > >>> > >>> if (CPUPhysBits) { > >>> pci_mem64_top = 1LL << CPUPhysBits; > >>> - if (CPUPhysBits > 46) { > >>> - // Old linux kernels have trouble dealing with more than 46 > >>> - // phys-bits, so avoid that for now. Seems to be a bug in the > >>> - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 > >>> - pci_mem64_top = 1LL << 46; > >>> + if (CPUPhysBits > 44) { > >>> + // Old linux kernels have trouble dealing with more than 44/46 > >>> + // phys-bits. Seems to be a bug in the virtio-pci driver. > >>> + // 46: centos-7, ubuntu-18.04 > >>> + // 44: ubuntu-16.04 > >>> + // Limit the used address space to mitigate the bug, except we are > >>> + // running in a guest with more than 1TB of memory installed. > >>> + if (RamSizeOver4G < (1LL << 40)) { > >>> + pci_mem64_top = 1LL << 44; > >>> + } > >>> } > >>> } > >>> > >> > >> we've also had two reports about issues with 32-bit guests now[0][1] > >> (both had '-m 4096' in the QEMU commandline), which I was able to > >> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The > >> QEMU commandline is below[2]. > > > > is it also reproducible with upstream kernel? > > if yes, it would be better to fix that on guest kernel side, > > rather than in SeaBIOS which has no idea what guest OS is going to be > > running after it. > > > > Turns out it's only kernels with PAE. The 6.1 Debian kernel without PAE > boots fine (but the PAE was the one installed by default). I built a > 6.10 kernel and it also boots fine, a 6.10 build with PAE doesn't. > > >> Unfortunately, it still fails to boot, even with the "limit address > >> space used for pci devices, part two" patch applied on top of rel-1.16.3 > >> (and using current QEMU master). > >> > >> It boots fine with '-m 3583', but not anymore with '-m 3584'. So is > >> bumping the limit for the check necessary after all? > >> > >> Best Regards, > >> Fiona > >> > >> [0]: https://forum.proxmox.com/threads/150217/ > > > >> [1]: https://forum.proxmox.com/threads/149772/post-683562 > > well, with current QEMU master branch (I assume it haven't even got topic patch yet) > > ./qemu-system-x86_64 --enable-kvm -cpu host -smp 4 -snapshot -m 4096 -M pc-i440fx-6.1 winxp_x86_build2600 > > following CLI boots just fine an 32-bit XP on Haswell host. > > > > Also using 'host' with ancient OS, is basically asking for trouble. > > If it works for user with old 'XP contemporary' CPU model, user should use that > > instead or workarounds (aka it's management task to configure CLI in > > compatible with OS manner). > > > > From suspicions config options in that post I see 'viommu' and 'vmgenid', > > are you sure XP even knows what to do with that, perhaps it triggers BSOD. > > > > No idea why the user enabled viommu for XP. vmgenid is added by our > management stack by default and I don't remember it having ever caused > problems. The user said the VM booted fine after adding lm=off to the > CPU options. I'm not super interested in the Windows case to be honest O:) > >> [2]: > >> > >>> ./qemu-system-x86_64 \ > >>> -accel 'kvm' \ > >>> -cpu 'host' \ > >>> -chardev 'socket,id=qmp,path=/var/run/qemu-server/121.qmp,server=on,wait=off' \ > >>> -mon 'chardev=qmp,mode=control' \ > >>> -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' \ > >>> -mon 'chardev=qmp-event,mode=control' \ > >>> -pidfile /var/run/qemu-server/121.pid \ > >>> -smp '4,sockets=1,cores=4,maxcpus=4' \ > >>> -nodefaults \ > >>> -vnc 'unix:/var/run/qemu-server/121.vnc,password=on' \ > >>> -m 4096 \ > >>> -device 'pci-bridge,id=pci.3,chassis_nr=3,bus=pci.0,addr=0x5' \ > >>> -device 'VGA,id=vga,bus=pci.0,addr=0x2' \ > >>> -device 'virtio-scsi-pci,id=virtioscsi0,bus=pci.3,addr=0x1' \ > >>> -drive 'file=/dev/lvmthinbig/vm-121-disk-0,if=none,id=drive-scsi0,format=raw,cache=none,aio=io_uring,detect-zeroes=on' \ > >>> -device 'scsi-hd,bus=virtioscsi0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' \ > >>> -machine 'type=pc' > > > > can you try to boot with q35 + intel iommu enabled > > and/or force virtio into legacy mode. > > > > I tried > > > -device 'intel-iommu' \ > > -machine 'type=q35' > > and intel_iommu=on in the guest's kernel cmdline, but it still failed > with kernels with PAE. > > Appending 'disable-modern=on,disable-legacy=off' to the virtio-scsi-pci > line made it work however (also with pc machine) :) does it also help in Windows case? Perhaps it's a better workaround (compared to lm=off or going back to older bios or dumb-ing down bios to suit PAE guests) for use on mgmt side, as it directly targets bug in guest's virtio-driver. Can we put this config tweak into libosinfo somehow, so that provisioning tools/mgmt could all reuse that to properly configure virtio for PAE kernels? > Best Regards, > Fiona > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Am 16.07.24 um 14:48 schrieb Igor Mammedov: > On Tue, 16 Jul 2024 13:56:08 +0200 > Fiona Ebner <f.ebner@proxmox.com> wrote: > >> Am 12.07.24 um 15:26 schrieb Igor Mammedov: >>> On Fri, 12 Jul 2024 14:24:40 +0200 >>> Fiona Ebner <f.ebner@proxmox.com> wrote: >>>> we've also had two reports about issues with 32-bit guests now[0][1] >>>> (both had '-m 4096' in the QEMU commandline), which I was able to >>>> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The >>>> QEMU commandline is below[2]. >>> >>> is it also reproducible with upstream kernel? >>> if yes, it would be better to fix that on guest kernel side, >>> rather than in SeaBIOS which has no idea what guest OS is going to be >>> running after it. >>> >> >> Turns out it's only kernels with PAE. The 6.1 Debian kernel without PAE >> boots fine (but the PAE was the one installed by default). I built a >> 6.10 kernel and it also boots fine, a 6.10 build with PAE doesn't. >> >> >> Appending 'disable-modern=on,disable-legacy=off' to the virtio-scsi-pci >> line made it work however (also with pc machine) :) > does it also help in Windows case? > Sorry, I haven't set it up right now. I only reproduced the issue with Debian. > Perhaps it's a better workaround (compared to lm=off or going back to > older bios or dumb-ing down bios to suit PAE guests) for use on mgmt side, > as it directly targets bug in guest's virtio-driver. > > Can we put this config tweak into libosinfo somehow, so that provisioning > tools/mgmt could all reuse that to properly configure virtio for PAE kernels? How would you detect whether the guest kernel is PAE or not before starting QEMU? Best Regards, Fiona _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Tue, 16 Jul 2024 15:58:35 +0200 Fiona Ebner <f.ebner@proxmox.com> wrote: > Am 16.07.24 um 14:48 schrieb Igor Mammedov: > > On Tue, 16 Jul 2024 13:56:08 +0200 > > Fiona Ebner <f.ebner@proxmox.com> wrote: > > > >> Am 12.07.24 um 15:26 schrieb Igor Mammedov: > >>> On Fri, 12 Jul 2024 14:24:40 +0200 > >>> Fiona Ebner <f.ebner@proxmox.com> wrote: > >>>> we've also had two reports about issues with 32-bit guests now[0][1] > >>>> (both had '-m 4096' in the QEMU commandline), which I was able to > >>>> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The > >>>> QEMU commandline is below[2]. > >>> > >>> is it also reproducible with upstream kernel? > >>> if yes, it would be better to fix that on guest kernel side, > >>> rather than in SeaBIOS which has no idea what guest OS is going to be > >>> running after it. > >>> > >> > >> Turns out it's only kernels with PAE. The 6.1 Debian kernel without PAE > >> boots fine (but the PAE was the one installed by default). I built a > >> 6.10 kernel and it also boots fine, a 6.10 build with PAE doesn't. > >> > >> > >> Appending 'disable-modern=on,disable-legacy=off' to the virtio-scsi-pci > >> line made it work however (also with pc machine) :) > > does it also help in Windows case? > > > > Sorry, I haven't set it up right now. I only reproduced the issue with > Debian. > > > Perhaps it's a better workaround (compared to lm=off or going back to > > older bios or dumb-ing down bios to suit PAE guests) for use on mgmt side, > > as it directly targets bug in guest's virtio-driver. > > > > Can we put this config tweak into libosinfo somehow, so that provisioning > > tools/mgmt could all reuse that to properly configure virtio for PAE kernels? > > How would you detect whether the guest kernel is PAE or not before > starting QEMU? that's what mgmt can do by poking in install media/disk image or asking user only mgmt can potentially know what target OS would be and properly configure QEMU (it's not something qemu or firmware can reasonably deal with on their own). (libosinfo can detect OS on install media, perhaps it also could be taught to probe what kernel would be used) As Gerd have said, all we can do at firmware level is heuristics, and in this case there is no good solution, we either hurt PAE guests with buggy driver by increasing size or hurt 64-bit guests by reducing it. In both cases we would get bug reports, and I'd expect a shift in numbers from PAE reports towards large device hotplug as time goes on. > Best Regards, > Fiona > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Am 16.07.24 um 17:15 schrieb Igor Mammedov: > On Tue, 16 Jul 2024 15:58:35 +0200 > Fiona Ebner <f.ebner@proxmox.com> wrote: > >> Am 16.07.24 um 14:48 schrieb Igor Mammedov: >>> On Tue, 16 Jul 2024 13:56:08 +0200 >>> Fiona Ebner <f.ebner@proxmox.com> wrote: >>> >>>> Am 12.07.24 um 15:26 schrieb Igor Mammedov: >>>>> On Fri, 12 Jul 2024 14:24:40 +0200 >>>>> Fiona Ebner <f.ebner@proxmox.com> wrote: >>>>>> we've also had two reports about issues with 32-bit guests now[0][1] >>>>>> (both had '-m 4096' in the QEMU commandline), which I was able to >>>>>> reproduce with a 32-bit Debian 12.6 install, so nothing ancient ;) The >>>>>> QEMU commandline is below[2]. >>>>> >>>>> is it also reproducible with upstream kernel? >>>>> if yes, it would be better to fix that on guest kernel side, >>>>> rather than in SeaBIOS which has no idea what guest OS is going to be >>>>> running after it. >>>>> >>>> >>>> Turns out it's only kernels with PAE. The 6.1 Debian kernel without PAE >>>> boots fine (but the PAE was the one installed by default). I built a >>>> 6.10 kernel and it also boots fine, a 6.10 build with PAE doesn't. >>>> >>>> >>>> Appending 'disable-modern=on,disable-legacy=off' to the virtio-scsi-pci >>>> line made it work however (also with pc machine) :) >>> does it also help in Windows case? >>> >> >> Sorry, I haven't set it up right now. I only reproduced the issue with >> Debian. >> >>> Perhaps it's a better workaround (compared to lm=off or going back to >>> older bios or dumb-ing down bios to suit PAE guests) for use on mgmt side, >>> as it directly targets bug in guest's virtio-driver. >>> >>> Can we put this config tweak into libosinfo somehow, so that provisioning >>> tools/mgmt could all reuse that to properly configure virtio for PAE kernels? >> >> How would you detect whether the guest kernel is PAE or not before >> starting QEMU? > > that's what mgmt can do by poking in install media/disk image or asking user > only mgmt can potentially know what target OS would be and > properly configure QEMU (it's not something qemu or firmware can reasonably > deal with on their own). (libosinfo can detect OS on install media, perhaps > it also could be taught to probe what kernel would be used) > > As Gerd have said, all we can do at firmware level is heuristics, > and in this case there is no good solution, we either hurt PAE guests > with buggy driver by increasing size or hurt 64-bit guests by reducing it. > In both cases we would get bug reports, and I'd expect a shift in numbers > from PAE reports towards large device hotplug as time goes on. > I see, thank you very much for the explanations! Best Regards, Fiona _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, 21 Jun 2024 14:05:17 +0200 Gerd Hoffmann <kraxel@redhat.com> wrote: > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > > Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > > 64-bit pci io window, failing during boot with errors like: > > Well. Why people would use *that* ubuntu version is not clear to me. > It's *loooooong* out of support. Even the LTS version from that year > (16.04) is not supported any more. But it is at least available for > download still, so I gave it a spin. > > Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > 44 phys-bits). Test patch below. > > take care, > Gerd > > diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c > index bb44dc296047..a43876a931c9 100644 > --- a/src/fw/pciinit.c > +++ b/src/fw/pciinit.c > @@ -1189,11 +1189,16 @@ pci_setup(void) > > if (CPUPhysBits) { > pci_mem64_top = 1LL << CPUPhysBits; > - if (CPUPhysBits > 46) { > - // Old linux kernels have trouble dealing with more than 46 > - // phys-bits, so avoid that for now. Seems to be a bug in the > - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 > - pci_mem64_top = 1LL << 46; > + if (CPUPhysBits > 44) { > + // Old linux kernels have trouble dealing with more than 44/46 > + // phys-bits. Seems to be a bug in the virtio-pci driver. > + // 46: centos-7, ubuntu-18.04 > + // 44: ubuntu-16.04 > + // Limit the used address space to mitigate the bug, except we are > + // running in a guest with more than 1TB of memory installed. Is it possible to fix those broken drivers (centos-7 for example) and ditch this heuristic altogether? The rest of downstream can pick it up from there if they care about their customers. > + if (RamSizeOver4G < (1LL << 40)) { > + pci_mem64_top = 1LL << 44; > + } > } > } > > > _______________________________________________ > SeaBIOS mailing list -- seabios@seabios.org > To unsubscribe send an email to seabios-leave@seabios.org > _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, Jun 21, 2024 at 03:20:19PM GMT, Igor Mammedov wrote: > > > diff --git a/src/fw/pciinit.c b/src/fw/pciinit.c > > index bb44dc296047..a43876a931c9 100644 > > --- a/src/fw/pciinit.c > > +++ b/src/fw/pciinit.c > > @@ -1189,11 +1189,16 @@ pci_setup(void) > > > > if (CPUPhysBits) { > > pci_mem64_top = 1LL << CPUPhysBits; > > - if (CPUPhysBits > 46) { > > - // Old linux kernels have trouble dealing with more than 46 > > - // phys-bits, so avoid that for now. Seems to be a bug in the > > - // virtio-pci driver. Reported: centos-7, ubuntu-18.04 > > - pci_mem64_top = 1LL << 46; > > + if (CPUPhysBits > 44) { > > + // Old linux kernels have trouble dealing with more than 44/46 > > + // phys-bits. Seems to be a bug in the virtio-pci driver. > > + // 46: centos-7, ubuntu-18.04 > > + // 44: ubuntu-16.04 > > + // Limit the used address space to mitigate the bug, except we are > > + // running in a guest with more than 1TB of memory installed. > > Is it possible to fix those broken drivers (centos-7 for example) > and ditch this heuristic altogether? > The rest of downstream can pick it up from there if they care about > their customers. Some further testing showed that this is not version-specific but arch specific. Old 32-bit kernels fail >44, old 64-bit kernels fail >46. Note that 44 = 32 + 12, i.e. this could be pfn (page frame number) hitting MAX_UINT32. Should that be the case the fix is probably not easy (didn't check the kernel source though). Also note that releasing a kernel fix is not enough, you also have to respin install media. Distros which are *that* old typically don't get regular install media updates any more ... In short: The idea to fix distros and drop the heuristic is IMHO not realistic. take care, Gerd _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
Hi, Dne 21. 06. 24 v 15:20 Igor Mammedov napsal(a): >> + // Old linux kernels have trouble dealing with more than 44/46 >> + // phys-bits. Seems to be a bug in the virtio-pci driver. >> + // 46: centos-7, ubuntu-18.04 >> + // 44: ubuntu-16.04 >> + // Limit the used address space to mitigate the bug, except we are >> + // running in a guest with more than 1TB of memory installed. > > Is it possible to fix those broken drivers (centos-7 for example) > and ditch this heuristic altogether? Does this code ever runs in some baremetal use cases as well? Or at least, meybe this could be applied only when virtio devices are present (maybe even just when transitional devices are present) ? Thanks, Rudolf _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, Jun 21, 2024 at 10:29:45PM +0200, Rudolf Marek wrote: > Hi, > > Dne 21. 06. 24 v 15:20 Igor Mammedov napsal(a): > > > + // Old linux kernels have trouble dealing with more than 44/46 > > > + // phys-bits. Seems to be a bug in the virtio-pci driver. > > > + // 46: centos-7, ubuntu-18.04 > > > + // 44: ubuntu-16.04 > > > + // Limit the used address space to mitigate the bug, except we are > > > + // running in a guest with more than 1TB of memory installed. > > > > Is it possible to fix those broken drivers (centos-7 for example) > > and ditch this heuristic altogether? > > Does this code ever runs in some baremetal use cases as well? > > Or at least, meybe this could be applied only when virtio devices are present (maybe even just when transitional devices are present) ? This code is only used on qemu (and derivatives) - specifically CONFIG_QEMU must be true. When running on coreboot, SeaBIOS expects coreboot to map all the PCI devices. Cheers, -Kevin _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, Jun 21, 2024 at 02:05:17PM +0200, Gerd Hoffmann wrote: > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > > Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > > 64-bit pci io window, failing during boot with errors like: > > Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > 44 phys-bits). Test patch below. Thanks for the patch, I can confirm this also works with Ubuntu 14.04 (oldest we had to hand) as well as a couple of 32-bit Windows VMs. This is a much better fix! > Even the LTS version from that year (16.04) is not supported any more. Even 14.04 is not yet end of life. If you're prepared to pay, they'll still support you. https://wiki.ubuntu.com/Releases > Well. Why people would use *that* ubuntu version is not clear to me. > It's *loooooong* out of support. You're in the IT dept of a large corporation. You have some critical application running on some old Dell server with Ubuntu 16.04. A move to virtualization has been mandated across the org, so you need to decommission that server. The application was built by some contractor - before your time - and the source code was long lost, due to a misadventure with a misconfigured array - again before your time. You've tried to use a newer version, but the application depends on lots of libraries that didn't take compatibility seriously (like, say, GNOME), so it simply can't run on newer versions. You've tried for some time to work around this by building and installing dependencies but you're not an expert on dynamic linkers, and could never get that last C++ symbol to resolve. There's no funding to build a new replacement for the app. You're aware that the OS is out of full support, so you do your best to lock down any network access and mitigate the relevant CVEs. Now you try to upgrade your virtualization cluster, and your VM doesn't boot any more. This kind of situation is very common. It's Long tail is long :( regards john _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, Jun 21, 2024 at 01:37:24PM GMT, John Levon wrote: > On Fri, Jun 21, 2024 at 02:05:17PM +0200, Gerd Hoffmann wrote: > > > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > > > Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > > > 64-bit pci io window, failing during boot with errors like: > > > Well. Why people would use *that* ubuntu version is not clear to me. > > It's *loooooong* out of support. > > You're in the IT dept of a large corporation. You have some critical > application running on some old Dell server with Ubuntu 16.04. I was specifically referring to the "16.10" listed above, i.e. running critical stuff on non-LTS distros in the first place. 16.04-LTS is a much better choice, and that there are plenty of things which can happen in real life which can delay or even prevent moving applications to a newer LTS version is pretty clear. take care, Gerd _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
On Fri, Jun 21, 2024 at 01:37:24PM +0100, John Levon wrote: > On Fri, Jun 21, 2024 at 02:05:17PM +0200, Gerd Hoffmann wrote: > > > On Wed, Jun 19, 2024 at 11:21:14AM GMT, John Levon wrote: > > > Older 32-bit Linux VMs (including Ubuntu 16.10) have issues with the > > > 64-bit pci io window, failing during boot with errors like: > > > > Turns out it apparently can't deal with PCI bars mapped above 16TB (aka > > 44 phys-bits). Test patch below. > > Thanks for the patch, I can confirm this also works with Ubuntu 14.04 (oldest we > had to hand) as well as a couple of 32-bit Windows VMs. This is a much better > fix! Thanks for tracking this down and testing. I agree this looks like an improved fix. -Kevin _______________________________________________ SeaBIOS mailing list -- seabios@seabios.org To unsubscribe send an email to seabios-leave@seabios.org
© 2016 - 2024 Red Hat, Inc.