On Tue, Jul 30, 2024 at 10:57 AM Akihiko Odaki <akihiko.odaki@daynix.com> wrote:
On 2024/07/30 11:04, Jason Wang wrote:
On Tue, Jul 30, 2024 at 12:43 AM Akihiko Odaki <akihiko.odaki@daynix.com> wrote:
On 2024/07/29 23:29, Peter Xu wrote:
On Mon, Jul 29, 2024 at 01:45:12PM +0900, Akihiko Odaki wrote:
On 2024/07/29 12:50, Jason Wang wrote:
On Sun, Jul 28, 2024 at 11:19 PM Akihiko Odaki <akihiko.odaki@daynix.com> wrote: > > On 2024/07/27 5:47, Peter Xu wrote: >> On Fri, Jul 26, 2024 at 04:17:12PM +0100, Daniel P. Berrangé wrote: >>> On Fri, Jul 26, 2024 at 10:43:42AM -0400, Peter Xu wrote: >>>> On Fri, Jul 26, 2024 at 09:48:02AM +0100, Daniel P. Berrangé wrote: >>>>> On Fri, Jul 26, 2024 at 09:03:24AM +0200, Thomas Huth wrote: >>>>>> On 26/07/2024 08.08, Michael S. Tsirkin wrote: >>>>>>> On Thu, Jul 25, 2024 at 06:18:20PM -0400, Peter Xu wrote: >>>>>>>> On Tue, Aug 01, 2023 at 01:31:48AM +0300, Yuri Benditovich wrote: >>>>>>>>> USO features of virtio-net device depend on kernel ability >>>>>>>>> to support them, for backward compatibility by default the >>>>>>>>> features are disabled on 8.0 and earlier. >>>>>>>>> >>>>>>>>> Signed-off-by: Yuri Benditovich <yuri.benditovich@daynix.com> >>>>>>>>> Signed-off-by: Andrew Melnychecnko <andrew@daynix.com> >>>>>>>> >>>>>>>> Looks like this patch broke migration when the VM starts on a host that has >>>>>>>> USO supported, to another host that doesn't.. >>>>>>> >>>>>>> This was always the case with all offloads. The answer at the moment is, >>>>>>> don't do this. >>>>>> >>>>>> May I ask for my understanding: >>>>>> "don't do this" = don't automatically enable/disable virtio features in QEMU >>>>>> depending on host kernel features, or "don't do this" = don't try to migrate >>>>>> between machines that have different host kernel features? >>>>>> >>>>>>> Long term, we need to start exposing management APIs >>>>>>> to discover this, and management has to disable unsupported features. >>>>>> >>>>>> Ack, this likely needs some treatments from the libvirt side, too. >>>>> >>>>> When QEMU automatically toggles machine type featuers based on host >>>>> kernel, relying on libvirt to then disable them again is impractical, >>>>> as we cannot assume that the libvirt people are using knows about >>>>> newly introduced features. Even if libvirt is updated to know about >>>>> it, people can easily be using a previous libvirt release. >>>>> >>>>> QEMU itself needs to make the machine types do that they are there >>>>> todo, which is to define a stable machine ABI. >>>>> >>>>> What QEMU is missing here is a "platform ABI" concept, to encode >>>>> sets of features which are tied to specific platform generations. >>>>> As long as we don't have that we'll keep having these broken >>>>> migration problems from machine types dynamically changing instead >>>>> of providing a stable guest ABI. >>>> >>>> Any more elaboration on this idea? Would it be easily feasible in >>>> implementation? >>> >>> In terms of launching QEMU I'd imagine: >>> >>> $QEMU -machine pc-q35-9.1 -platform linux-6.9 ...args... >>> >>> Any virtual machine HW features which are tied to host kernel features >>> would have their defaults set based on the requested -platform. The >>> -machine will be fully invariant wrt the host kernel. >>> >>> You would have -platform hlep to list available platforms, and >>> corresonding QMP "query-platforms" command to list what platforms >>> are supported on a given host OS. >>> >>> Downstream distros can provide their own platforms definitions >>> (eg "linux-rhel-9.5") if they have kernels whose feature set >>> diverges from upstream due to backports. >>> >>> Mgmt apps won't need to be taught about every single little QEMU >>> setting whose default is derived from the kernel. Individual >>> defaults are opaque and controlled by the requested platform. >>> >>> Live migration has clearly defined semantics, and mgmt app can >>> use query-platforms to validate two hosts are compatible. >>> >>> Omitting -platform should pick the very latest platform that is >>> cmpatible with the current host (not neccessarily the latest >>> platform built-in to QEMU). >> >> This seems to add one more layer to maintain, and so far I don't know >> whether it's a must. >> >> To put it simple, can we simply rely on qemu cmdline as "the guest ABI"? I >> thought it was mostly the case already, except some extremely rare >> outliers. >> >> When we have one host that boots up a VM using: >> >> $QEMU1 $cmdline >> >> Then another host boots up: >> >> $QEMU2 $cmdline -incoming XXX >> >> Then migration should succeed if $cmdline is exactly the same, and the VM >> can boot up all fine without errors on both sides. >> >> AFAICT this has nothing to do with what kernel is underneath, even not >> Linux? I think either QEMU1 / QEMU2 has the option to fail. But if it >> didn't, I thought the ABI should be guaranteed. >> >> That's why I think this is a migration violation, as 99.99% of other device >> properties should be following this rule. The issue here is, we have the >> same virtio-net-pci cmdline on both sides in this case, but the ABI got >> break. >> >> That's also why I was suggesting if the property contributes to the guest >> ABI, then AFAIU QEMU needs to: >> >> - Firstly, never quietly flipping any bit that affects the ABI... >> >> - Have a default value of off, then QEMU will always allow the VM to boot >> by default, while advanced users can opt-in on new features. We can't >> make this ON by default otherwise some VMs can already fail to boot, > > It may not be necessary the case that old features are supported by > every systems. In an extreme case, a user may migrate a VM from Linux to > Windows, which probably doesn't support any offloading at all. A more > convincing scenario is RSS offloading with eBPF; using eBPF requires a > privilege so we cannot assume it is always available even on the latest > version of Linux.
I don't get why eBPF matters here. It is something that is not noticed by the guest and we have a fallback anyhow.
It is noticeable for the guest, and the fallback is not effective with vhost.
It's a bug then. Qemu can fallback to tuntap if it sees issues in vhost.
We can certainly fallback to in-QEMU RSS by disabling vhost, but I would not say lack of such fallback is a bug.
Such fallback is by design since the introduction of vhost.
We don't provide in-QEMU fallback for other offloads.
Yes but what I want to say is that eBPF RSS is different from those segmentation offloads. And technically, Qemu can do fallback for offloads (as RSC did). Thanks
Regards, Akihiko Odaki