You need to look at the 3.1 release. It is supposed to get rid of this problem. On Fri, Oct 15, 2010 at 7:26 AM, Richard de Vries <rdevries1000 at gmail.com>wrote: > The mysql database only runs on one node at a time. > I still find it hard to understand why you need to restart the service > if a brick goes down and comes back again. > > this is the volume file I'm using > > > glusterfsd.vol server file: > > > volume posix1 > type storage/posix > option directory /export/database > end-volume > > volume locks1 > type features/locks > subvolumes posix1 > end-volume > > volume database > type performance/io-threads > option thread-count 8 > subvolumes locks1 > end-volume > > volume server > type protocol/server > option transport-type tcp/server > option auth.addr.database.allow * > option transport.socket.listen-port 6996 > option transport.socket.nodelay on > subvolumes database > end-volume > > > > > database.vol file: > > volume databasenode1 > type protocol/client > option transport-type tcp > option transport.socket.nodelay on > option remote-port 6996 > option ping-timeout 2 > option remote-host node1 > option remote-subvolume database > end-volume > > volume databasenode2 > type protocol/client > option transport-type tcp > option transport.socket.nodelay on > option remote-port 6996 > option ping-timeout 2 > option remote-host node2 > option remote-subvolume database > end-volume > > volume replicate > type cluster/replicate > subvolumes databasenode1 databasenode2 > end-volume > > volume stat-performance > type performance/stat-prefetch > subvolumes replicate > end-volume > > Maybe the stat-performance translator has influence on the this stat > output. > > I'll try to disable this and test again. > > Regards, > Richard > > On Fri, Oct 15, 2010 at 12:44 PM, Deadpan110 <deadpan110 at gmail.com> wrote: > > I am very new to this list, but here is my 2 cents... > > > > In the past I used DRBD between 2 nodes to provide a master/slave > > setup with mySQL data stored on the filesystem. > > > > Upon a failover situation, mySQL would start on the remaining server > > and pick up where things left off. > > > > DRBD (8.0 +) now supports master/master but it would be unwise to run > > mySQL on such a setup live on 2 servers. > > > > mySQL has also advanced and replication is not restricted to > master/slave. > > > > I use (and am loving) glusterfs in various guises on my 3 node cluster > > for my client filesystems. > > > > For mySQL I use master/master/master circular replication without > > depending on any type of clustered filesystem (only local on each > > node) - there have been people frowning on such a setup, but things > > have advanced with the latest stable mySQL versions and as such, I > > have been successfully using it in a clustered environment. > > > > Martin > > > > On 15 October 2010 20:33, Richard de Vries <rdevries1000 at gmail.com> > wrote: > >> Hello Beat, > >> > >> This is a pitty. Because a stop of the service only to resync te > >> standby node is not so nice... > >> > >> The stat of the database file in: /opt/test/database after a reboot of > >> the node 2 shows different output, > >> one time from the node 1 and another time from node 2. > >> > >> What is the role of self heal in this? It is noticed that the files > >> are not equal (via stat). > >> > >> Would you see the same behaviour for example with qemu-kvm that keeps > >> also files open? > >> > >> Regards, > >> Richard > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > -- Thanks, Richard Shade Integration Engineer RightScale - http://www.rightscale.com/ phone: 8055004164x1018