Hi Roberto, Gluster Storage Platform provides client volume spec file through server for created volumes. You can mount using mount -t glusterfs <server>:<volume>-<transport> <your-mount-point> for example, mount -t glusterfs node01:gluster01-tcp /mnt/gluster01 Its not required to write your own spec file for mounting it. Thanks, Regards, Bala Roberto Lucignani wrote: > Hi all, > > I installed two Gluster Storage Platorm 3.0.4 on two servers node01 e > node02. > > I created a volume called gluster01 than I mounted it on a Debian box in > this way: > > > > mount -t glusterfs /etc/glusterfs/gluster01-tcp.vol /mnt/gluster01/ > > > > the gluster01-tcp.vol is the following: > > > > > > volume 192.168.0.200-1 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.200 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick1 > > end-volume > > > > volume 192.168.0.200-2 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.200 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick2 > > end-volume > > > > volume 192.168.0.201-1 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.201 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick1 > > end-volume > > > > volume 192.168.0.201-2 > > type protocol/client > > option transport-type tcp > > option remote-host 192.168.0.201 > > option transport.socket.nodelay on > > option transport.remote-port 10012 > > option remote-subvolume brick2 > > end-volume > > > > volume mirror-0 > > type cluster/replicate > > subvolumes 192.168.0.201-1 192.168.0.200-1 > > end-volume > > > > volume mirror-1 > > type cluster/replicate > > subvolumes 192.168.0.201-2 192.168.0.200-2 > > end-volume > > > > volume distribute > > type cluster/distribute > > subvolumes mirror-0 mirror-1 > > end-volume > > > > volume readahead > > type performance/read-ahead > > option page-count 4 > > subvolumes distribute > > end-volume > > > > volume iocache > > type performance/io-cache > > option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed > 's/[^0-9]//g') / 5120 ))`MB > > option cache-timeout 1 > > subvolumes readahead > > end-volume > > > > volume quickread > > type performance/quick-read > > option cache-timeout 1 > > option max-file-size 64kB > > subvolumes iocache > > end-volume > > > > volume writebehind > > type performance/write-behind > > option cache-size 4MB > > subvolumes quickread > > end-volume > > > > volume statprefetch > > type performance/stat-prefetch > > subvolumes writebehind > > end-volume > > > > > > > > all works fine and smooth, I can write and read on that volume without any > problem. > > The problem is when the node01 is unavailable, I can't access the volume via > mount on the Debian box. This doesn't happen if it is the node02 to be > unavailable. > > I expected the same behavior in the two cases, in this way the node01 > represents an SPOF, am I wrong ? am I missing something the configuration ? > > > > Tnx in advance > > Rpberto > > > > > ------------------------------------------------------------------------ > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users