Re: gluster and kvm livemigration

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 17.01.2014 16:45:20, Samuli Heinonen wrote:
Hello Bernhard,

Can you test if setting option network.remote-dio to enable allows you to use cache=none?


Hi Samuli,

nope network.remote-dio enable didn't changed it.

got a too unclean system meanwhile,

will do a full reinstall of the hosts

and than come back to this topic

thnx 

Bernhard


-samuli

Bernhard Glomm <bernhard.glomm@xxxxxxxxxxx> kirjoitti 17.1.2014 kello 16.41:

Pranith,
I stopped the volume
started it again,
mounted it on both hosts
started the VM
did the livemigration
and collected the logs:
- etc-glusterfs-glusterd.vol.log
- glustershd.log
- srv-vm_infrastructure-vm-atom01.log
- cli.log
from the beginning of the gluster volume start.
You can found them here (part 1 to 3):

further more:
gluster --version: glusterfs 3.4.2 built on Jan 11 2014 03:21:47
ubuntu: raring
filesystem on the gluster bricks: zfs-0.6.2

gluster volume info fs_vm_atom01 
Volume Name: fs_vm_atom01
Type: Replicate
Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01
Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01
Options Reconfigured:
diagnostics.client-log-level: DEBUG
server.allow-insecure: on

disk part of VM configuration

<emulator>/usr/bin/kvm-spice</emulator>
    <disk type='file' device='disk'>
      <driver name='qemu' type='raw' cache='writethrough'/>
      <source file='/srv/vm_infrastructure/vm/atom01/atom01.img'/>
      <target dev='vda' bus='virtio'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/>
    </disk>

can't use <source protocol='gluster' ...
as josh suggested because couldn't get
my qemu recompiled with gluster enabled yet.

Are there other special tuning parameter for kvm/qemu/ to set on gluster?
as mentioned: all works except the livemigration (disk image file becomes read only)
and I have to use something different than cache=none...

TIA

Bernhard


On 17.01.2014 05:04:52, Pranith Kumar Karampuri wrote:
Bernhard,
Configuration seems ok. Could you please give the log files of the bricks and mount please. If you think it is not a big procedure to do this live migration, could you set client-log-level to DEBUG and provide the log files of that run.

Pranith

----- Original Message -----
From: "Bernhard Glomm" <bernhard.glomm@xxxxxxxxxxx>
To: pkarampu@xxxxxxxxxx
Cc: gluster-users@xxxxxxxxxxx
Sent: Thursday, January 16, 2014 5:58:17 PM
Subject: Re: gluster and kvm livemigration


hi Pranith

# gluster volume info fs_vm_atom01
 
Volume Name: fs_vm_atom01
Type: Replicate
Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01
Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01
Options Reconfigured:
diagnostics.client-log-level: ERROR


TIA
Bernhard


On 16.01.2014 13:05:12, Pranith Kumar Karampuri wrote:
hi Bernhard,
Could you give gluster volume info output?

Pranith

----- Original Message -----
From: "Bernhard Glomm" <> > bernhard.glomm@xxxxxxxxxxx> > >
To: > > gluster-users@xxxxxxxxxxx
Sent: Thursday, January 16, 2014 4:22:36 PM
Subject: gluster and kvm livemigration

I experienced a strange behavior of glusterfs during livemigration
of a qemu-kvm guest
using a 10GB file on a mirrored gluster 3.4.2 volume
(both on ubuntu 13.04)
I run
virsh migrate --verbose --live --unsafe --p2p --domain atom01 --desturi
qemu+ssh://<target_ip>/system
and the migration works,
the running machine is pingable and keeps sending pings.
nevertheless, when I let the machine touch a file during migration
it stops, complaining that it's filesystem is read only (from that moment
that
migration finished)
A reboot from inside the machine failes,
machine goes down and comes up with an error
unable to write to sector xxxxxx on hd0
(than falling into the initrd).
a
virsh destroy VM && virsh start VM
leads to a perfect running VM again,
no matter on which of the two hosts I start the machine
anybody better experience with livemigration?
any hint on a procedure how to debug that?
TIA
Bernhard

--


Bernhard Glomm
IT Administration

Phone: +49 (30) 86880 134
Fax: +49 (30) 86880 100
Skype: bernhard.glomm.ecologic

Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717
Berlin
| Germany
GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.:
DE811963464
Ecologic is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux