The following series of (2) patches adds a QEMU driver to libvirt. The first patch provides a daemon for managing QEMU instances, the second provides a driver letting libvirt manage QEMU via the daemon. Basic architecture ------------------ The reason for the daemon architecture is two fold: - At this time, there is no (practical) way to enumerate QEMU instances, or reliably connect to the monitor console of an existing process. There is also no way to determine the guest configuration associated with a daemon. - It is desirable to be able to manage QEMU instances using either an unprivilegd local client, or a remote client. The daemon can provide connectivity via UNIX domain sockets, or IPv4 / IPv6 and layer in suitable authentication / encryption via TLS and/or SASL protocols. Anthony Ligouri is working on patches for QEMU with the goal of addressing the first point. For example, an extra command line argument will cause QEMU to save a PID file and create a UNIX socket for its monitor at a well-defined path. More functionality in the monitor console will allow the guest configuration to be reverse engineered from a running guest. Even with those patches, however, it will still be desirable to have a daemon to provide more flexible connectivity, and to facilitate implementation libvirt APIs which are host (rather than guest) related. Thus I expect that over time we can simply enhance the daemon to take advantage of newer capabilities in the QEMU monitor, but keep the same basic libvirt driver architecture. Considering some of the other hypervisor technologies out there, in particular User Mode Linux, and lhype, it may well become possible to let this QEMU daemon also provide the management of these guests - allowing re-use of the single driver backend in the libvirt client library itself. XML format ---------- As discussed in the previous mail thread, the XML format for describing guests with the QEMU backend is the same structure as that for Xen guests, with following enhancements: - The 'type' attribute on the top level <domain> tag can take one of the values 'qemu', 'kqemu' or 'kvm' instead of 'xen'. This selects between the different virtualization approaches QEMU can provide. - The '<type>' attribute within the <os> block of the XML (for now) is still expected to the 'hvm' (indicating full virtualization), although I'm trying to think of a better name, since its not technically hardware accelerated unless you're using KVM - The '<type>' attribute within the <os> block of the XML can have two optional 'arch' and 'machine' attributes. The former selects the CPU architecture to be emulated; the latter the specific machine to have QEMU emulate (determine those supported by QEMU using 'qemu -M ?'). - The <kernel>, <initrd>, <cmdline> elements can be used to specify an explicit kernel to boot off[1], otherwise it'll do a boot of the cdrom, harddisk / floppy (based on <boot> element). Well,the kernel bits are parsed at least. I've not got around to using them when building the QEMU argv yet. - The disk devices are configured in same way as Xen HVM guests. eg you have to use hda -> hdd, and/or fda -> fdb. Only hdc can be selected as a cdrom device. - The network configuration is work in progress. QEMU has many ways to setup networking. I use the 'type' attribute to select between the different approachs 'user', 'tap', 'server', 'client', 'mcast' mapping them directly onto QEMU command line arguments. You can specify a MAC address as usual too. I need to implement auto-generation of MAC addresses if omitted. Most of them have extra bits of metadata though which I've not figured out appropriate XML for yet. Thus when building the QEMU argv I currently just hardcode 'user' networking. - The QEMU binary is determined automatically based on the requested CPU architecture, defaulting to i686 if non specified. It is possible to override the default binary using the <emulator> element within the <devices> section. This is different to previously discussed, because recent work by Anthony merging VMI + KVM to give paravirt guests means that the <loader> element is best kept to refer to the VMI ROM (or other ROM like files :-) - this is also closer to Xen semantics anyway. Connectivity ------------ The namespace under which all connection URIs come is 'qemud'. Thereafter there are several options. First, two well-known local hypervisor connections - qemud:///session This is a per-user private hypervisor connection. The libvirt daemon and qemu guest processes just run as whatever UNIX user your client app is running. This lets unprivileged users use the qemu driver without needing any kind admin rights. Obviously you can't use KQEMU or KVM accelerators unless the /dev/ device node is chmod/chown'd to give you access. The communication goes over a UNIX domain socket which is mode 0600 created in the abstract namespace at $HOME/.qemud.d/sock. - qemud:///system This is a system-wide privileged hypervisor connection. There is only one of these on any given machine. The libvirt_qemud daemon would be started ahead of time (by an init script), possibly running as root, or maybe under a dedicated system user account (and the KQEMU/KVM devices chown'd to match). The admin would optionally also make it listen on IPv4/6 addrs to allow remote communication. (see next URI example) The local communication goes over one of two possible UNIX domain sockets Both in the abstract namespace under the directory /var/run. The first socket called 'qemud' is mode 0600, so only privileged apps (ie root) can access it, and gives full control capabilities. The other called 'qemud-ro' is mode 0666 and any clients connecting to it will be restricted to only read-only libvirt operations by the server. - qemud://hostname:port/ This lets you connect to a daemon over IPv4 or IPv6. If omitted the port is 8123 (will probably change it). This lets you connect to a system daemon on a remote host - assuming it was configured to listen on IPv4/6 interfaces. Currently there is zero auth or encryption, but I'm planning to make it mandortory to use the TLS protocol - using the GNU TLS library. This will give encryption, and mutual authentication using either x509 certificates or PGP keys & trustdbs or perhaps both :-) Will probably start off by implementing PGP since I understand it better. So if you wanted to remotely manage a server, you'd copy the server's certificate/public key to the client into a well known location. Similarly you'd generate a keypair for the client & copy its public key to the server. Perhaps I'll allow clients without a key to connect in read-only mode. Need to prototype it first and then write up some ideas. Server architecture ------------------- The server is a fairly simple beast. It is single-threaded using non-blocking I/O and poll() for all operations. It will listen on multiple sockets for incoming connections. The protocol used for client-server comms is a very simple binary message format close to the existing libvirt_proxy. Client sends a message, server receives it, performs appropriate operation & sends a reply to the client. The client (ie libvirt driver) blocks after sending its message until it gets a reply. The server does non-blocking reads from the client buffering until it has a single complete message, then processes it and populates the buffer with a reply and does non-blocking writes to send it back to the client. It won't try to read a further message from the client until its sent the entire reply back. ie, it is a totally synchronous message flow - no batching/pipelining of messages. During the time the server is processes a message it is not dealing with any other I/O, but thus far all the operations are very fast to implement, so this isn't a serious issue, and there ways to deal with it if there are operations which turn out to take a long time. I certainly want to avoid multi-threading in the server at all costs! As well as monitoring the client & client sockets, the poll() event loop in the server also captures stdout & stderr from the QEMU processes. Currently we just dump this to stdout of the daemon, but I expect we can log it somewhere. When we start accessing the QEMU monitor there will be another fd in the event loop - ie the pseduo-TTY (or UNIX socket) on which we talk to the monitor. Inactive guests --------------- Guests created using 'virsh create' (or equiv API) are treated as 'transient' domains - ie their config files are not saved to disk. This is consistent with the behaviour in the Xen backend. Guests created using 'virsh define', however, are saved out to disk in $HOME/.qemud.d for the per-user session daemon. The system-wide daemon should use /etc/qemud.d, but currently its still /root/.qemud.d The config files are simply saved as the libvirt XML blob ensuring no data conversion issues. In any case, QEMU doesn't currently have any config file format we can leverage. The list of inactive guests is loaded at startup of the daemon. New config files are expected to be created via the API - files manually created in the directory after initial startup are not seen. Might like to change this later. XML Examples ------------ This is a guest using plain qemu, with x86_64 architecture and a ISA-only (ie no PCI) machine emulation. I was actually running this on a 32-bit host :-) VNC is configured to run on port 5906. QEMU can't automatically choose a VNC port, so if one isn't specified we assign one based on the domain ID. This should be fixed in QEMU.... <domain type='qemu'> <name>demo1</name> <uuid>4dea23b3-1d52-d8f3-2516-782e98a23fa0</uuid> <memory>131072</memory> <vcpu>1</vcpu> <os> <type arch='x86_64' machine='isapc'>hvm</type> </os> <devices> <disk type='file' device='disk'> <source file='/home/berrange/fedora/diskboot.img'/> <target dev='hda'/> </disk> <interface type='user'> <mac address='24:42:53:21:52:45'/> </interface> <graphics type='vnc' port='5906'/> </devices> </domain> A second example, this time using KVM acceleration. Note how I specify a non-default path to QEMU to pick up the KVM build of QEMU. Normally KVM binary will default to /usr/bin/qemu-kvm - this may change depending on how distro packaging of KVM turns out - it may even be merged into regular QEMU binaries. <domain type='kvm'> <name>demo2</name> <uuid>4dea24b3-1d52-d8f3-2516-782e98a23fa0</uuid> <memory>131072</memory> <vcpu>1</vcpu> <os> <type>hvm</type> </os> <devices> <emulator>/home/berrange/usr/kvm-devel/bin/qemu-system-x86_64</emulator> <disk type='file' device='disk'> <source file='/home/berrange/fedora/diskboot.img'/> <target dev='hda'/> </disk> <interface type='user'> <mac address='24:42:53:21:52:45'/> </interface> <graphics type='vnc' port='-1'/> </devices> </domain> Outstanding work ---------------- - TLS support. Need to add TLS encryption & authentication to both the client and server side for IPv4/6 communications. This will obviously add a dependancy on libgnutls.so in libvirt & the daemon. I don't consider this a major problem since every non-trivial network app these days uses TLS. The other possible impl of OpenSSL has GPL-compatability issues, so is not considered. - Change the wire format to use fixed size data types (ie, int8, int16, int32, etc) instead of the size-dependant int/long types. At same time define some rules for the byte ordering. Client must match server ordering ? Server must accept client's desired ordering ? Everyone must use BE regardless of server/client format ? I'm inclined to say client must match server, since it distributes the byte-swapping overhead to all clients and lets the common case of x86->x86 be a no-op. - Add a protocol version message as first option to let use protocol at will later while maintaining compat with older libvirt client libraries. - Improve support for describing the various QEMU network configurations - Finish boot options - boot device order & explicit kernel - Open & use connection to QEMU monitor which will let us implement pause/resume, suspend/restore drivers, and device hotplug / media changes. - Return sensible data for virNodeInfo - will need to have operating system dependant code here - parsing /proc for Linux to determine available RAM & CPU speed. Who knows what for Solaris / BSD ?!? Anyone know of remotely standard ways for doing this. Accurate host memory reporting is the only really critical data item we need. - There is a fair bit of duplicate in various helper functions between the daemon, and various libvirt driver backends. We should probably pull this stuff out into a separate lib/ directoy, build it into a static library and then link that into both libvirt, virsh & the qemud daemon as needed. Dan. -- |=- Red Hat, Engineering, Emerging Technologies, Boston. +1 978 392 2496 -=| |=- Perl modules: http://search.cpan.org/~danberr/ -=| |=- Projects: http://freshmeat.net/~danielpb/ -=| |=- GnuPG: 7D3B9505 F3C9 553F A1DA 4AC2 5648 23C1 B3DF F742 7D3B 9505 -=|