On Wed, Aug 30, 2017 at 04:25:10PM -0300, Eduardo Habkost wrote: > On Wed, Aug 30, 2017 at 05:23:39PM +0300, Michael S. Tsirkin wrote: > > On Wed, Aug 30, 2017 at 10:17:27AM -0300, Eduardo Habkost wrote: > > > I'm CCing libvir-list and qemu-devel because I would like to get > > > feedback from libvirt and QEMU developers too. > > > > > > On Tue, Aug 08, 2017 at 10:49:21PM +0300, Michael S. Tsirkin wrote: > > > > On Tue, Jul 18, 2017 at 03:42:08PM +0200, Maxime Coquelin wrote: > > > > > This is an revival from a thread I initiated earlier this year [0], that > > > > > I had to postpone due to other priorities. > > > > > > > > > > First, I'd like to thanks reviewers of my first proposal, this new > > > > > version tries to address the comments made: > > > > > 1.This is Nova's role and not Libvirt's to query hosts supported > > > > > compatibility modes and to select one, since Nova adds the vhost-user > > > > > ports and has visibility on other hosts. Hence I remove libvirt ML and > > > > > add Openstack one in the recipient list. > > > > > 2. By default, the compatibility version selected is the most recent > > > > > one, except if the admin selects on older compat version. > > > > > > > > > > The goal of this thread is to draft a solution based on the outcomes > > > > > of discussions with contributors of the different parties (DPDK/OVS > > > > > /Nova/...). > > > > > > > > > > I'm really interested on feedback from OVS & Nova contributors, > > > > > as my experience with these projects is rather limited. > > > > > > > > > > Problem statement: > > > > > ================== > > > > > > > > > > When migrating a VM from one host to another, the interfaces exposed by > > > > > QEMU must stay unchanged in order to guarantee a successful migration. > > > > > In the case of vhost-user interface, parameters like supported Virtio > > > > > feature set, max number of queues, max vring sizes,... must remain > > > > > compatible. Indeed, the frontend not being re-initialized, no > > > > > re-negotiation happens at migration time. > > > > > > > > > > For example, we have a VM that runs on host A, which has its vhost-user > > > > > backend advertising VIRTIO_F_RING_INDIRECT_DESC feature. Since the Guest > > > > > also support this feature, it is successfully negotiated, and guest > > > > > transmit packets using indirect descriptor tables, that the backend > > > > > knows to handle. > > > > > > > > > > At some point, the VM is being migrated to host B, which runs an older > > > > > version of the backend not supporting this VIRTIO_F_RING_INDIRECT_DESC > > > > > feature. The migration would break, because the Guest still have the > > > > > VIRTIO_F_RING_INDIRECT_DESC bit sets, and the virtqueue contains some > > > > > decriptors pointing to indirect tables, that backend B doesn't know to > > > > > handle. > > > > > This is just an example about Virtio features compatibility, but other > > > > > backend implementation details could cause other failures. (e.g. > > > > > configurable queues sizes) > > > > > > > > > > What we need is to be able to query the destination host's backend to > > > > > ensure migration is possible before it is initiated. > > > > > > > > This remided me strongly of the issues around the virtual CPU modeling > > > > in KVM, see > > > > https://wiki.qemu.org/index.php/Features/CPUModels#Querying_host_capabilities > > > > > > > > QEMU recently gained query-cpu-model-expansion to allow capability queries. > > > > > > > > Cc Eduardo accordingly. Eduardo, could you please take a look - > > > > how is the problem solved on the KVM/VCPU side? Do the above > > > > problem and solution for vhost look similar? > > > > > > (Sorry for taking so long to reply) > > > > > > CPU configuration in QEMU has the additional problem of features > > > depending on host hardware and kernel capabilities (not just QEMU > > > software capabilities). Do you have vhost-user features that > > > depend on the host kernel or hardware too, or all of them just > > > depend on the vhost-user backend software? > > > > vhost-net features depend on host kernel. > > > > > If it depends only on software, a solution similar to how > > > machine-types work in QEMU sound enough. If features depend on > > > host kernel or host hardware too, it is a bit more complex: it > > > means you need an interface to find out if each configurable > > > feature/version is really available on the host. > > > > > > (In the case of CPU models, we started with an interface that > > > reported which CPU models were runnable on the host. But as > > > libvirt allows enabling/disabling individual CPU features, the > > > interface had to be extended to report which CPU features were > > > available/unavailable on the host.) > > > > > > * * * > > > > > > Now, there's one thing that seems very different here: the > > > guest-visible interface is not defined only by QEMU, but also by > > > the vhost-user backend. Is that correct? > > > > Not exactly. As long as there are no bugs it's defined by QEMU but > > depends on backend capabilities. Bugs in a backend could be guest > > visible - same as kvm really. > > I'm a bit confused here. > > I will try to enumerate the steps involved in the process, for > clarity: > 1) Querying which features are available on a host; > 2) Choosing a reasonable default based on what's available on the > relevant host(s), before starting a VM; > 3) Actually configuring what will be seen by the guest, based on > (1), (2) (and optionally user input/configuration). > > Above you say that (1) on vhost-net depend on host kernel too. > That's OK. > > I also understand that (2) can't be done by libvirt and QEMU > alone, because they don't have information about the vhost-user > backend before the VM is configured. That's OK too. > > However, I don't see the data flow of the configuration step (3) > clearly. If the guest ABI is only defined by QEMU, does that > mean configuring the guest-visible features would always be done > through libvirt+QEMU? > > In other words, would the corresponding > vhostuser_compat.virtio_features value (or other knobs that > affect guest ABI) always flow this way: > OVS -> libvirt -> QEMU -> vhost-user-backend -> guest > and not directly this way: > OVS -> vhost-user-backend -> guest > ? Barring bugs, after init it works like this libvirt -----------------------+ | v OVS -> vhost-user-backend <-> QEMU -> guest On device init qemu queries all features, enables the subset configured by libvirt and also supplies them (with possible modifications) to guest. > > > > > > This means QEMU won't fully control the resulting guest ABI > > > anymore. I would really prefer if we could keep libvirt+QEMU in > > > control of the guest ABI as usual, making QEMU configure all the > > > guest-visible vhost-user features. But I understand this would > > > require additional interfaces between QEMU and libvirt, and > > > extending the libvirt APIs. > > > > > > So, if QEMU is really not going to control the resulting guest > > > ABI completely, can we at least provide a mechanism which QEMU > > > can use to ask vhost-user for guest ABI details on migration, and > > > block migration if vhost-user was misconfigured on the > > > destination host when migrating? > > > > > > > > > > > > > > > The below proposal has been drafted based on how Qemu manages machine types: > > > > > > > > > > Proposal > > > > > ======== > > > > > > > > > > The idea is to have a table of supported version strings in OVS, > > > > > associated to key/value pairs. Nova or any other management tool could > > > > > query OVS for the list of supported versions strings for each hosts. > > > > > By default, the latest compatibility version will be selected, but the > > > > > admin can select manually an older compatibility mode in order to ensure > > > > > successful migration to an older destination host. > > > > > > > > > > Then, Nova would add OVS's vhost-user port with adding the selected > > > > > version (compatibility mode) as an extra parameter. > > > > > > > > > > Before starting the VM migration, Nova will ensure both source and > > > > > destination hosts' vhost-user interfaces run in the same compatibility > > > > > modes, and will prevent it if this is not the case. > > > > > > > > > > For example host A runs OVS-2.7, and host B OVS-2.6. > > > > > Host A's OVS-2.7 has an OVS-2.6 compatibility mode (e.g. with indirect > > > > > descriptors disabled), which should be selected at vhost-user port add > > > > > time to ensure migration will succeed to host B. > > > > > > > > > > Advantage of doing so is that Nova does not need any update if new keys > > > > > are introduced (i.e. it does not need to know how the new keys have to > > > > > be handled), all these checks remain in OVS's vhost-user implementation. > > > > > > > > > > Ideally, we would support per vhost-user interface compatibility mode, > > > > > which may have an impact also on DPDK API, as the Virtio feature update > > > > > API is global, and not per port. > > > > > > > > > > - Implementation: > > > > > ----------------- > > > > > > > > > > Goal here is just to illustrate this proposal, I'm sure you will have > > > > > good suggestion to improve it. > > > > > In OVS vhost-user library, we would introduce a new structure, for > > > > > example (neither compiled nor tested): > > > > > > > > > > struct vhostuser_compat { > > > > > char *version; > > > > > uint64_t virtio_features; > > > > > uint32_t max_rx_queue_sz; > > > > > uint32_t max_nr_queues; > > > > > }; > > > > > > > > > > *version* field is the compatibility version string. It could be > > > > > something like: "upstream.ovs-dpdk.v2.6". In case for example Fedora > > > > > adds some more patches to its package that would break migration to > > > > > upstream version, it could have a dedicated compatibility string: > > > > > "fc26.ovs-dpdk.v2.6". In case OVS-v2.7 does not break compatibility with > > > > > previous OVS-v2.6 version, then no need to create a new entry, just keep > > > > > v2.6 one. > > > > > > > > > > *virtio_features* field is the Virtio features set for a given > > > > > compatibility version. When an OVS tag is to be created, it would be > > > > > associated to a DPDK version. The Virtio features for these version > > > > > would be stored in this field. It would allow to upgrade the DPDK > > > > > package for example from v16.07 to v16.11 without breaking migration. > > > > > In case the distribution wants to benefit from latests Virtio > > > > > features, it would have to create a new entry to ensure migration > > > > > won't be broken. > > > > > > > > > > *max_rx_queue_sz* > > > > > *max_nr_queues* fields are just here for example, don't think this is > > > > > needed today. I just want to illustrate that we have to anticipate > > > > > other parameters than the Virtio feature set, even if not necessary > > > > > at the moment. > > > > > > > > > > We create a table with different compatibility versions in OVS > > > > > vhost-user lib: > > > > > > > > > > static struct vhostuser_compat vu_compat[] = { > > > > > { > > > > > .version = "upstream.ovs-dpdk.v2.7", > > > > > .virtio_features = 0x12045694, > > > > > .max_rx_queue_sz = 512, > > > > > }, > > > > > { > > > > > .version = "upstream.ovs-dpdk.v2.6", > > > > > .virtio_features = 0x10045694, > > > > > .max_rx_queue_sz = 1024, > > > > > }, > > > > > }; > > > > > > > > > > At some time during installation, or system init, the table would be > > > > > parsed, and compatibility version strings would be stored into the OVS > > > > > database, or a new tool would be created to list these strings, or a > > > > > config file packaged with OVS stores the list of compatibiliy versions. > > > > > > > > > > Before launching the VM, Nova will query the version strings for the > > > > > host so that the admin can select an older compatibility mode. If none > > > > > selected by the admin, then the most recent one will be used by default, > > > > > and passed to the OVS's add-port command as parameter. Note that if no > > > > > compatibility mode is passed to the add-port command, the most recent > > > > > one is selected by OVS as default. > > > > > > > > > > When the vhost-user connection is initiated, OVS would know in which > > > > > compatibility mode to init the interface, for example by restricting the > > > > > support Virtio features of the interface. > > > > > > > > > > Cheers, > > > > > Maxime > > > > > > > > > > [0]: > > > > > https://mail.openvswitch.org/pipermail/ovs-dev/2017-February/328257.html > > > > > <b2a5501c-7df7-ad2a-002f-d731c445a502 at redhat.com> > > > > > > -- > > > Eduardo > > -- > Eduardo -- libvir-list mailing list libvir-list@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/libvir-list