Fwd: DF reports incorrect sizes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The issue is that 1 of the 2 AFR volumes is not synchronized.

Meaning erasing or creating files on mounts performed only on 1 node - but
the free space reported from the both nodes.

Any idea what's went wrong?

Regards.


2009/3/26 Stas Oskin <stas.oskin at gmail.com>

> Hi.
>
> Same as advised on this list, see below.
>
> By the way, I restarted both the clients and servers, and the reported size
> is still the same.
> Whichever it is, it stuck quite persistently :).
>
> server.vol
>
> volume home1
>  type storage/posix                   # POSIX FS translator
>  option directory /media/storage        # Export this directory
> end-volume
>
> volume posix-locks-home1
>  type features/posix-locks
>  option mandatory-locks on
>  subvolumes home1
> end-volume
>
> ### Add network serving capability to above home.
> volume server
>  type protocol/server
>  option transport-type tcp
>  subvolumes posix-locks-home1
>  option auth.addr.posix-locks-home1.allow * # Allow access to "home1"
> volume
> end-volume
>
>
> client.vol
>
> ## Reference volume "home1" from remote server
> volume home1
>  type protocol/client
>  option transport-type tcp/client
>  option remote-host 192.168.253.41      # IP address of remote host
>  option remote-subvolume posix-locks-home1     # use home1 on remote host
>  option transport-timeout 10           # value in seconds; it should be set
> relatively low
> end-volume
>
> ## Reference volume "home2" from remote server
> volume home2
>  type protocol/client
>  option transport-type tcp/client
>  option remote-host 192.168.253.42      # IP address of remote host
>  option remote-subvolume posix-locks-home1     # use home1 on remote host
>  option transport-timeout 10           # value in seconds; it should be set
> relatively low
> end-volume
>
> volume home
>  type cluster/afr
>  option metadata-self-heal on
>  subvolumes home1 home2
> end-volume
>
> volume writebehind
>   type performance/write-behind
>   option aggregate-size 128KB
>   option window-size 1MB
>   subvolumes home
> end-volume
>
> volume cache
>   type performance/io-cache
>   option cache-size 512MB
>   subvolumes writebehind
> end-volume
>
>
> Regards.
>
> 2009/3/26 Vikas Gorur <vikas at zresearch.com>
>
> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>:
>> > Hi.
>> >
>> > We erased all the data from our mount point, but the df still reports
>> > it's almost full:
>> >
>> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs
>> >
>> > Running du either in the mount point, or in the back-end directory,
>> > reports 914M.
>> >
>> > How do we get the space back?
>>
>> What is your client and server configuration?
>>
>> Vikas
>> --
>> Engineer - Z Research
>> http://gluster.com/
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090330/304a27ef/attachment-0001.htm>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux