Hi Bala, thank you very much for your reply. Yes I tried both methods, one building a vol file and one retrieving the configuration from the server and both work fine However my problem do not regards the mount process but it regards the spof represented by node01. If it is unavailable I can't access my volume. Why this ? and why it happens only with the first node ? Regards M4dG -----Messaggio originale----- Da: Bala.JA [mailto:bala at gluster.com] Inviato: domenica 23 maggio 2010 20.58 A: roberto.lucignani at caleidos.it Cc: gluster-users at gluster.org Oggetto: Re: SPOF question Hi Roberto, Gluster Storage Platform provides client volume spec file through server for created volumes. You can mount using mount -t glusterfs <server>:<volume>-<transport> <your-mount-point> for example, mount -t glusterfs node01:gluster01-tcp /mnt/gluster01 Its not required to write your own spec file for mounting it. Thanks, Regards, Bala Roberto Lucignani wrote: > Hi all, > > I installed two Gluster Storage Platorm 3.0.4 on two servers node01 e > node02. > > I created a volume called gluster01 than I mounted it on a Debian box in > this way: > > > > mount -t glusterfs /etc/glusterfs/gluster01-tcp.vol /mnt/gluster01/ > > > > the gluster01-tcp.vol is the following: > > > > > > volume 192.168.0.200-1 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.200 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick1 > > end-volume > > > > volume 192.168.0.200-2 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.200 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick2 > > end-volume > > > > volume 192.168.0.201-1 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.201 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick1 > > end-volume > > > > volume 192.168.0.201-2 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.201 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick2 > > end-volume > > > > volume mirror-0 > > type cluster/replicate > > subvolumes 192.168.0.201-1 192.168.0.200-1 > > end-volume > > > > volume mirror-1 > > type cluster/replicate > > subvolumes 192.168.0.201-2 192.168.0.200-2 > > end-volume > > > > volume distribute > > type cluster/distribute > > subvolumes mirror-0 mirror-1 > > end-volume > > > > volume readahead > > type performance/read-ahead > > option page-count 4 > > subvolumes distribute > > end-volume > > > > volume iocache > > type performance/io-cache > > option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed > 's/[^0-9]//g') / 5120 ))`MB > > option cache-timeout 1 > > subvolumes readahead > > end-volume > > > > volume quickread > > type performance/quick-read > > option cache-timeout 1 > > option max-file-size 64kB > > subvolumes iocache > > end-volume > > > > volume writebehind > > type performance/write-behind > > option cache-size 4MB > > subvolumes quickread > > end-volume > > > > volume statprefetch > > type performance/stat-prefetch > > subvolumes writebehind > > end-volume > > > > > > > > all works fine and smooth, I can write and read on that volume without any > problem. > > The problem is when the node01 is unavailable, I can't access the volume via > mount on the Debian box. This doesn't happen if it is the node02 to be > unavailable. > > I expected the same behavior in the two cases, in this way the node01 > represents an SPOF, am I wrong ? am I missing something the configuration ? > > > > Tnx in advance > > Rpberto > > > > > ------------------------------------------------------------------------ > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users