Re: How namespace works in mainline-2.5?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Amar for the confirmations. I just checked out up to patch 185 and haven't had an issue yet while rsyncing to the volume. Been running about 1/2 hour now. Love the new debug output.

As for the doc/example...I see the cluster-client.vol was fixed, but the bricks-ns isnt configured in any of the cluster-server#.vol files.

Huge Regards,
Dale

Amar S. Tumballi wrote:
Hi,
Sorry for less info regarding the changes made to unify in 2.5 branch. We haven't yet updated wiki, as release is not yet made. but wrong example in doc/* is my mistake :(

Yes, you have figured it right, namespace is NOT a subvolume of unify. Also about the schedulers complaining about the not enough space, all the schedulers now uses % based min-disk-size, (option alu.limits.min-disk-size 10 #don't consider the node for scheduling if the free disk space is less than 10%).

About the 'ls' failing, i may need little more info. First, which patch are you using. And, an output of 'bt' after doing gdb glusterfs -c /core.<pid>.

Regards,
Amar

On 6/21/07, *Dale Dude* <dale@xxxxxxxxxxxxxxx <mailto:dale@xxxxxxxxxxxxxxx>> wrote:

    How do I use the namespace option correctly? In the docs/examples I
    think your using it wrong.

    In cluster-client.vol you have volumes client1,client2,client3
    configured and this line in cluster/unify:
    option namespace client1 # this will not be storage child of unify.
    subvolumes client1 client2 client3

    I get the impression that namespace shouldnt be configured as a
    subvolume of unify. If I use a namespace of a volume that is a
    subvolume
    of unify it complains as such. If I create a volume on the server
    specifically for the namespace (i.e. 1 storage/posix called
    volumenamespace) and in the client config use that then it doesnt
    complain anymore. But I cant even ls on the mounted volume. I get this
    debug found below. A 'df -h' looks correct (glusterfs 5.9T  400M  5.9T
    1% /volumes).

    Btw, I have to use the ALU scheduler because any other scheduler keeps
    saying that there isnt enough space on any of the "bricks".



[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux