---------- Forwarded message ---------- From: Stas Oskin <stas.oskin at gmail.com> Date: 2009/3/29 Subject: Re: DF reports incorrect sizes To: Vikas Gorur ?vikas at zresearch.com? Hi. After erasing all the data from my lab setup, and restarting all, it happened again in less then 5 hours. Here is what I see: Client: df -h: glusterfs 31G 29G 0 100% /mnt/media Server 1: df -h: /dev/hda4 31G 29G 0 100% /media Server 2: df -h: /dev/hda4 31G 20G 8.7G 70% /media This means that again the server lost each other. Perhaps it's related to the fact that the space go filled out. Any idea how to diagnose it? Regards. 2009/3/26 Stas Oskin <stas.oskin at gmail.com> > Hi. > > It occurs that 1 of the 2 AFR volumes is not synchronized. > > Meaning erasing or creating files on mounts performed only on 1 node - but > the free space reported from the both nodes. > > Any idea what's went wrong? > > Regards. > > > 2009/3/26 Stas Oskin <stas.oskin at gmail.com> > >> Hi. >> >> Same as advised on this list, see below. >> >> By the way, I restarted both the clients and servers, and the reported >> size is still the same. >> Whichever it is, it stuck quite persistently :). >> >> server.vol >> >> volume home1 >> type storage/posix # POSIX FS translator >> option directory /media/storage # Export this directory >> end-volume >> >> volume posix-locks-home1 >> type features/posix-locks >> option mandatory-locks on >> subvolumes home1 >> end-volume >> >> ### Add network serving capability to above home. >> volume server >> type protocol/server >> option transport-type tcp >> subvolumes posix-locks-home1 >> option auth.addr.posix-locks-home1.allow * # Allow access to "home1" >> volume >> end-volume >> >> >> client.vol >> >> ## Reference volume "home1" from remote server >> volume home1 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.253.41 # IP address of remote host >> option remote-subvolume posix-locks-home1 # use home1 on remote host >> option transport-timeout 10 # value in seconds; it should be >> set relatively low >> end-volume >> >> ## Reference volume "home2" from remote server >> volume home2 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.253.42 # IP address of remote host >> option remote-subvolume posix-locks-home1 # use home1 on remote host >> option transport-timeout 10 # value in seconds; it should be >> set relatively low >> end-volume >> >> volume home >> type cluster/afr >> option metadata-self-heal on >> subvolumes home1 home2 >> end-volume >> >> volume writebehind >> type performance/write-behind >> option aggregate-size 128KB >> option window-size 1MB >> subvolumes home >> end-volume >> >> volume cache >> type performance/io-cache >> option cache-size 512MB >> subvolumes writebehind >> end-volume >> >> >> Regards. >> >> 2009/3/26 Vikas Gorur <vikas at zresearch.com> >> >> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>: >>> > Hi. >>> > >>> > We erased all the data from our mount point, but the df still reports >>> > it's almost full: >>> > >>> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs >>> > >>> > Running du either in the mount point, or in the back-end directory, >>> > reports 914M. >>> > >>> > How do we get the space back? >>> >>> What is your client and server configuration? >>> >>> Vikas >>> -- >>> Engineer - Z Research >>> http://gluster.com/ >>> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090330/570576df/attachment.htm>