Re: [PATCH v2] libxl: Implement basic video device selection

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 2014-04-04 at 14:51 +0200, Daniel P. Berrange wrote:
> On Fri, Apr 04, 2014 at 11:34:17AM +0100, Ian Campbell wrote:
> > On Fri, 2014-04-04 at 12:31 +0200, Stefan Bader wrote:
> > > On 04.04.2014 11:48, Ian Campbell wrote:
> > > > On Fri, 2014-04-04 at 11:36 +0200, Stefan Bader wrote:
> > > >> +    /*
> > > >> +     * Take the first defined video device (graphics card) to display
> > > >> +     * on the first graphics device (display).
> > > >> +     * Right now only type and vram info is used and anything beside
> > > >> +     * type xen and vga is mapped to cirrus.
> > > >> +     */
> > > >> +    if (def->nvideos) {
> > > >> +        unsigned int min_vram = 8 * 1024;
> > > >> +
> > > >> +        switch (def->videos[0]->type) {
> > > >> +            case VIR_DOMAIN_VIDEO_TYPE_VGA:
> > > >> +            case VIR_DOMAIN_VIDEO_TYPE_XEN:
> > > >> +                b_info->u.hvm.vga.kind = LIBXL_VGA_INTERFACE_TYPE_STD;
> > > >> +                /*
> > > >> +                 * Libxl enforces a minimal VRAM size of 8M when using
> > > >> +                 * LIBXL_DEVICE_MODEL_VERSION_QEMU_XEN_TRADITIONAL or
> > > >> +                 * 16M for LIBXL_DEVICE_MODEL_VERSION_QEMU_XEN.
> > > >> +                 * Avoid build failures and go with the minimum if less
> > > >> +                 * is specified.
> > > > 
> > > > Build failures? Do you mean "domain build" rather than "libvirt build"?
> > > > 
> > > > I'm not sure about this fixing up the GIGO from the user side, but
> > > > that's a libvirt policy decision I suppose? If it were me I would just
> > > > let libxl provide the error and expect people to fix their domain config
> > > > rather than silently giving them something other than what they asked
> > > > for. What if increasing the VRAM causes a cascading failure due e.g. to
> > > > lack of memory? That's going to be tricky to debug I think!
> > > 
> > > In the end its a start a domain with such a config. Which seems to be what I
> > > would end up with in my testing with an admittedly older version of virt-manager
> > > connecting to a libvirtd running an initial version of this patch without that part.
> > > The error seen on the front-end was something along the lines of "failed to get
> > > enough memory to start the guest" (the libxl log on the other side had the
> > > better error message). And the gui always reduces the memory below the minimum
> > > for both the options (VGA and "Xen").
> > > That is the reason I went for "meh, go for the minimum anyways".
> > 
> > Does the libvirt protocol require the client to provide all the sizes
> > etc with no provision for asking the server side to pick a sane default?
> 
> The XML does not have to include the VGA ram size. If it is omitted the
> we fill in a default value after initial parsing is done.

I guess the issue is that whatever client Stefan is using is including
the VGA ram size, with a value which it turns out is not allowed.

> That defualt can be hypervisor specific

Great!

Ian.
> > > And btw, it is already confusing enough as with cirrus, I get a 9M by default
> > > which is passed on to qemu on the command line and then ignored by it and one
> > > gets 32M in any way...
> 
> Yeah, we were silently ignoring ram size for cirrus since QEMU doesn't let
> it be configured. Only QXL and I think stdvga is configurable.
> 
> 
> Regards,
> Daniel


--
libvir-list mailing list
libvir-list@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/libvir-list




[Index of Archives]     [Virt Tools]     [Libvirt Users]     [Lib OS Info]     [Fedora Users]     [Fedora Desktop]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite News]     [KDE Users]     [Fedora Tools]