DF reports incorrect sizes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi.

Same as advised on this list, see below.

By the way, I restarted both the clients and servers, and the reported size
is still the same.
Whichever it is, it stuck quite persistently :).

server.vol

volume home1
 type storage/posix                   # POSIX FS translator
 option directory /media/storage        # Export this directory
end-volume

volume posix-locks-home1
 type features/posix-locks
 option mandatory-locks on
 subvolumes home1
end-volume

### Add network serving capability to above home.
volume server
 type protocol/server
 option transport-type tcp
 subvolumes posix-locks-home1
 option auth.addr.posix-locks-home1.allow * # Allow access to "home1" volume
end-volume


client.vol

## Reference volume "home1" from remote server
volume home1
 type protocol/client
 option transport-type tcp/client
 option remote-host 192.168.253.41      # IP address of remote host
 option remote-subvolume posix-locks-home1     # use home1 on remote host
 option transport-timeout 10           # value in seconds; it should be set
relatively low
end-volume

## Reference volume "home2" from remote server
volume home2
 type protocol/client
 option transport-type tcp/client
 option remote-host 192.168.253.42      # IP address of remote host
 option remote-subvolume posix-locks-home1     # use home1 on remote host
 option transport-timeout 10           # value in seconds; it should be set
relatively low
end-volume

volume home
 type cluster/afr
 option metadata-self-heal on
 subvolumes home1 home2
end-volume

volume writebehind
  type performance/write-behind
  option aggregate-size 128KB
  option window-size 1MB
  subvolumes home
end-volume

volume cache
  type performance/io-cache
  option cache-size 512MB
  subvolumes writebehind
end-volume


Regards.

2009/3/26 Vikas Gorur <vikas at zresearch.com>

> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>:
> > Hi.
> >
> > We erased all the data from our mount point, but the df still reports
> > it's almost full:
> >
> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs
> >
> > Running du either in the mount point, or in the back-end directory,
> > reports 914M.
> >
> > How do we get the space back?
>
> What is your client and server configuration?
>
> Vikas
> --
> Engineer - Z Research
> http://gluster.com/
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090326/b3f08ef3/attachment-0001.htm>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux