Re: kvm crashes with spice while loading qxl

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sat, Mar 05, 2011 at 01:35:58PM -0300, Marcelo Tosatti wrote:
> On Sat, Feb 26, 2011 at 01:29:01PM +0100, Jan Kiszka wrote:
> > >     at /var/tmp/portage/app-emulation/qemu-kvm-0.14.0/work/qemu-kvm-0.14.0/qemu-kvm.c:1466
> > > #12 0x00007ffff77bb944 in start_thread () from /lib/libpthread.so.0
> > > #13 0x00007ffff5e491dd in clone () from /lib/libc.so.6
> > > (gdb)
> > 
> > That's a spice bug. In fact, there are a lot of
> > qemu_mutex_lock/unlock_iothread in that subsystem. I bet at least a few
> > of them can cause even more subtle problems.
> > 
> > Two general issues with dropping the global mutex like this:
> >  - The caller of mutex_unlock is responsible for maintaining
> >    cpu_single_env across the unlocked phase (that's related to the
> >    abort above).
> >  - Dropping the lock in the middle of a callback is risky. That may
> >    enable re-entrances of code sections that weren't designed for this
> >    (I'm skeptic about the side effects of
> >    qemu_spice_vm_change_state_handler - why dropping the lock here?).
> > 
> > Spice requires a careful review regarding such issues. Or it should
> > pioneer with introducing its own lock so that we can handle at least
> > related I/O activities over the VCPUs without holding the global mutex
> > (but I bet it's not the simplest candidate for such a new scheme).
> > 
> > Jan
> > 
> 
> Agree with the concern regarding spice.
> 

What are the pros and cons of (re)introducing a spice specific lock?
 + simplicity. Only spice touches the spice lock.
 - ? what were the original reasons for Gerd dropping the spice lock?

I have no problem reintroducing this lock, I'm just concerned that it's
wasted effort because after I send that patch someone will jump and remind
me why it was removed in the first place.

> Regarding global mutex, TCG and KVM execution behaviour can become more
> similar wrt locking by dropping qemu_global_mutex during generation and
> execution of TBs.
> 
> Of course for memory or PIO accesses from vcpu context qemu_global_mutex
> must be acquired.
> 
> With that in place, it becomes easier to justify further improvements
> regarding parallelization, such as using a read-write lock for
> l1_phys_map / phys_page_find_alloc.
> 
> 
>  21.62%               sh            3d38920b3f  [.] 0x00003d38920b3f                  
>   6.38%               sh  qemu-system-x86_64    [.] phys_page_find_alloc              
>   4.90%               sh  qemu-system-x86_64    [.] tb_find_fast                      
>   4.34%               sh  qemu-system-x86_64    [.] tlb_flush  
> 
> --
> To unsubscribe from this list: send the line "unsubscribe kvm" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux