On Wed, Aug 30, 2017 at 11:19:08PM +0300, Michael S. Tsirkin wrote:
On Wed, Aug 30, 2017 at 04:25:10PM -0300, Eduardo Habkost wrote:
> On Wed, Aug 30, 2017 at 05:23:39PM +0300, Michael S. Tsirkin wrote:
> > On Wed, Aug 30, 2017 at 10:17:27AM -0300, Eduardo Habkost wrote:
> > > I'm CCing libvir-list and qemu-devel because I would like to get
> > > feedback from libvirt and QEMU developers too.
> > >
> > > On Tue, Aug 08, 2017 at 10:49:21PM +0300, Michael S. Tsirkin wrote:
> > > > On Tue, Jul 18, 2017 at 03:42:08PM +0200, Maxime Coquelin wrote:
> > > > > This is an revival from a thread I initiated earlier this year
[0], that
> > > > > I had to postpone due to other priorities.
> > > > >
> > > > > First, I'd like to thanks reviewers of my first proposal,
this new
> > > > > version tries to address the comments made:
> > > > > 1.This is Nova's role and not Libvirt's to query hosts
supported
> > > > > compatibility modes and to select one, since Nova adds the
vhost-user
> > > > > ports and has visibility on other hosts. Hence I remove libvirt
ML and
> > > > > add Openstack one in the recipient list.
> > > > > 2. By default, the compatibility version selected is the most
recent
> > > > > one, except if the admin selects on older compat version.
> > > > >
> > > > > The goal of this thread is to draft a solution based on the
outcomes
> > > > > of discussions with contributors of the different parties
(DPDK/OVS
> > > > > /Nova/...).
> > > > >
> > > > > I'm really interested on feedback from OVS & Nova
contributors,
> > > > > as my experience with these projects is rather limited.
> > > > >
> > > > > Problem statement:
> > > > > ==================
> > > > >
> > > > > When migrating a VM from one host to another, the interfaces
exposed by
> > > > > QEMU must stay unchanged in order to guarantee a successful
migration.
> > > > > In the case of vhost-user interface, parameters like supported
Virtio
> > > > > feature set, max number of queues, max vring sizes,... must
remain
> > > > > compatible. Indeed, the frontend not being re-initialized, no
> > > > > re-negotiation happens at migration time.
> > > > >
> > > > > For example, we have a VM that runs on host A, which has its
vhost-user
> > > > > backend advertising VIRTIO_F_RING_INDIRECT_DESC feature. Since
the Guest
> > > > > also support this feature, it is successfully negotiated, and
guest
> > > > > transmit packets using indirect descriptor tables, that the
backend
> > > > > knows to handle.
> > > > >
> > > > > At some point, the VM is being migrated to host B, which runs an
older
> > > > > version of the backend not supporting this
VIRTIO_F_RING_INDIRECT_DESC
> > > > > feature. The migration would break, because the Guest still have
the
> > > > > VIRTIO_F_RING_INDIRECT_DESC bit sets, and the virtqueue contains
some
> > > > > decriptors pointing to indirect tables, that backend B
doesn't know to
> > > > > handle.
> > > > > This is just an example about Virtio features compatibility, but
other
> > > > > backend implementation details could cause other failures.
(e.g.
> > > > > configurable queues sizes)
> > > > >
> > > > > What we need is to be able to query the destination host's
backend to
> > > > > ensure migration is possible before it is initiated.
> > > >
> > > > This remided me strongly of the issues around the virtual CPU
modeling
> > > > in KVM, see
> > > >
https://wiki.qemu.org/index.php/Features/CPUModels#Querying_host_capabili...
> > > >
> > > > QEMU recently gained query-cpu-model-expansion to allow capability
queries.
> > > >
> > > > Cc Eduardo accordingly. Eduardo, could you please take a look -
> > > > how is the problem solved on the KVM/VCPU side? Do the above
> > > > problem and solution for vhost look similar?
> > >
> > > (Sorry for taking so long to reply)
> > >
> > > CPU configuration in QEMU has the additional problem of features
> > > depending on host hardware and kernel capabilities (not just QEMU
> > > software capabilities). Do you have vhost-user features that
> > > depend on the host kernel or hardware too, or all of them just
> > > depend on the vhost-user backend software?
> >
> > vhost-net features depend on host kernel.
> >
> > > If it depends only on software, a solution similar to how
> > > machine-types work in QEMU sound enough. If features depend on
> > > host kernel or host hardware too, it is a bit more complex: it
> > > means you need an interface to find out if each configurable
> > > feature/version is really available on the host.
> > >
> > > (In the case of CPU models, we started with an interface that
> > > reported which CPU models were runnable on the host. But as
> > > libvirt allows enabling/disabling individual CPU features, the
> > > interface had to be extended to report which CPU features were
> > > available/unavailable on the host.)
> > >
> > > * * *
> > >
> > > Now, there's one thing that seems very different here: the
> > > guest-visible interface is not defined only by QEMU, but also by
> > > the vhost-user backend. Is that correct?
> >
> > Not exactly. As long as there are no bugs it's defined by QEMU but
> > depends on backend capabilities. Bugs in a backend could be guest
> > visible - same as kvm really.
>
> I'm a bit confused here.
>
> I will try to enumerate the steps involved in the process, for
> clarity:
> 1) Querying which features are available on a host;
> 2) Choosing a reasonable default based on what's available on the
> relevant host(s), before starting a VM;
> 3) Actually configuring what will be seen by the guest, based on
> (1), (2) (and optionally user input/configuration).
>
> Above you say that (1) on vhost-net depend on host kernel too.
> That's OK.
>
> I also understand that (2) can't be done by libvirt and QEMU
> alone, because they don't have information about the vhost-user
> backend before the VM is configured. That's OK too.
>
> However, I don't see the data flow of the configuration step (3)
> clearly. If the guest ABI is only defined by QEMU, does that
> mean configuring the guest-visible features would always be done
> through libvirt+QEMU?
>
> In other words, would the corresponding
> vhostuser_compat.virtio_features value (or other knobs that
> affect guest ABI) always flow this way:
> OVS -> libvirt -> QEMU -> vhost-user-backend -> guest
> and not directly this way:
> OVS -> vhost-user-backend -> guest
> ?
Barring bugs, after init it works like this
libvirt -----------------------+
|
v
OVS -> vhost-user-backend <-> QEMU -> guest
On device init qemu queries all features, enables
the subset configured by libvirt and also supplies
them (with possible modifications) to guest.
Good. I thought there was no interface to configure the
guest-visible virtio features through libvirt, and OVS would
bypass that by configuring vhost-user-backend directly.
Which variables affect the set of available features, exactly?
Does it depend on QEMU capabilities too on some cases, or just
vhost-backend + kernel capabilities?
Is there anything libvirt+QEMU need do to help implement (1)
(querying host capabilities) so the rest of the stack can
implement (2) (choosing what to enable based on host
capabilities)? Or this can be done without QEMU and libvirt
being involved at all?
>
> >
> > > This means QEMU won't fully control the resulting guest ABI
> > > anymore. I would really prefer if we could keep libvirt+QEMU in
> > > control of the guest ABI as usual, making QEMU configure all the
> > > guest-visible vhost-user features. But I understand this would
> > > require additional interfaces between QEMU and libvirt, and
> > > extending the libvirt APIs.
> > >
> > > So, if QEMU is really not going to control the resulting guest
> > > ABI completely, can we at least provide a mechanism which QEMU
> > > can use to ask vhost-user for guest ABI details on migration, and
> > > block migration if vhost-user was misconfigured on the
> > > destination host when migrating?
> > >
> > >
> > > >
> > > > > The below proposal has been drafted based on how Qemu manages
machine types:
> > > > >
> > > > > Proposal
> > > > > ========
> > > > >
> > > > > The idea is to have a table of supported version strings in
OVS,
> > > > > associated to key/value pairs. Nova or any other management tool
could
> > > > > query OVS for the list of supported versions strings for each
hosts.
> > > > > By default, the latest compatibility version will be selected,
but the
> > > > > admin can select manually an older compatibility mode in order
to ensure
> > > > > successful migration to an older destination host.
> > > > >
> > > > > Then, Nova would add OVS's vhost-user port with adding the
selected
> > > > > version (compatibility mode) as an extra parameter.
> > > > >
> > > > > Before starting the VM migration, Nova will ensure both source
and
> > > > > destination hosts' vhost-user interfaces run in the same
compatibility
> > > > > modes, and will prevent it if this is not the case.
> > > > >
> > > > > For example host A runs OVS-2.7, and host B OVS-2.6.
> > > > > Host A's OVS-2.7 has an OVS-2.6 compatibility mode (e.g.
with indirect
> > > > > descriptors disabled), which should be selected at vhost-user
port add
> > > > > time to ensure migration will succeed to host B.
> > > > >
> > > > > Advantage of doing so is that Nova does not need any update if
new keys
> > > > > are introduced (i.e. it does not need to know how the new keys
have to
> > > > > be handled), all these checks remain in OVS's vhost-user
implementation.
> > > > >
> > > > > Ideally, we would support per vhost-user interface compatibility
mode,
> > > > > which may have an impact also on DPDK API, as the Virtio feature
update
> > > > > API is global, and not per port.
> > > > >
> > > > > - Implementation:
> > > > > -----------------
> > > > >
> > > > > Goal here is just to illustrate this proposal, I'm sure you
will have
> > > > > good suggestion to improve it.
> > > > > In OVS vhost-user library, we would introduce a new structure,
for
> > > > > example (neither compiled nor tested):
> > > > >
> > > > > struct vhostuser_compat {
> > > > > char *version;
> > > > > uint64_t virtio_features;
> > > > > uint32_t max_rx_queue_sz;
> > > > > uint32_t max_nr_queues;
> > > > > };
> > > > >
> > > > > *version* field is the compatibility version string. It could
be
> > > > > something like: "upstream.ovs-dpdk.v2.6". In case for
example Fedora
> > > > > adds some more patches to its package that would break migration
to
> > > > > upstream version, it could have a dedicated compatibility
string:
> > > > > "fc26.ovs-dpdk.v2.6". In case OVS-v2.7 does not break
compatibility with
> > > > > previous OVS-v2.6 version, then no need to create a new entry,
just keep
> > > > > v2.6 one.
> > > > >
> > > > > *virtio_features* field is the Virtio features set for a given
> > > > > compatibility version. When an OVS tag is to be created, it
would be
> > > > > associated to a DPDK version. The Virtio features for these
version
> > > > > would be stored in this field. It would allow to upgrade the
DPDK
> > > > > package for example from v16.07 to v16.11 without breaking
migration.
> > > > > In case the distribution wants to benefit from latests Virtio
> > > > > features, it would have to create a new entry to ensure
migration
> > > > > won't be broken.
> > > > >
> > > > > *max_rx_queue_sz*
> > > > > *max_nr_queues* fields are just here for example, don't
think this is
> > > > > needed today. I just want to illustrate that we have to
anticipate
> > > > > other parameters than the Virtio feature set, even if not
necessary
> > > > > at the moment.
> > > > >
> > > > > We create a table with different compatibility versions in OVS
> > > > > vhost-user lib:
> > > > >
> > > > > static struct vhostuser_compat vu_compat[] = {
> > > > > {
> > > > > .version = "upstream.ovs-dpdk.v2.7",
> > > > > .virtio_features = 0x12045694,
> > > > > .max_rx_queue_sz = 512,
> > > > > },
> > > > > {
> > > > > .version = "upstream.ovs-dpdk.v2.6",
> > > > > .virtio_features = 0x10045694,
> > > > > .max_rx_queue_sz = 1024,
> > > > > },
> > > > > };
> > > > >
> > > > > At some time during installation, or system init, the table
would be
> > > > > parsed, and compatibility version strings would be stored into
the OVS
> > > > > database, or a new tool would be created to list these strings,
or a
> > > > > config file packaged with OVS stores the list of compatibiliy
versions.
> > > > >
> > > > > Before launching the VM, Nova will query the version strings for
the
> > > > > host so that the admin can select an older compatibility mode.
If none
> > > > > selected by the admin, then the most recent one will be used by
default,
> > > > > and passed to the OVS's add-port command as parameter. Note
that if no
> > > > > compatibility mode is passed to the add-port command, the most
recent
> > > > > one is selected by OVS as default.
> > > > >
> > > > > When the vhost-user connection is initiated, OVS would know in
which
> > > > > compatibility mode to init the interface, for example by
restricting the
> > > > > support Virtio features of the interface.
> > > > >
> > > > > Cheers,
> > > > > Maxime
> > > > >
> > > > > [0]:
> > > > >
https://mail.openvswitch.org/pipermail/ovs-dev/2017-February/328257.html
> > > > > <b2a5501c-7df7-ad2a-002f-d731c445a502 at redhat.com>
> > >
> > > --
> > > Eduardo
>
> --
> Eduardo
--
Eduardo