Thanks Amar for the confirmations. I just checked out up to patch 185
and haven't had an issue yet while rsyncing to the volume. Been running
about 1/2 hour now. Love the new debug output.
As for the doc/example...I see the cluster-client.vol was fixed, but the
bricks-ns isnt configured in any of the cluster-server#.vol files.
Huge Regards,
Dale
Amar S. Tumballi wrote:
Hi,
Sorry for less info regarding the changes made to unify in 2.5
branch. We haven't yet updated wiki, as release is not yet made. but
wrong example in doc/* is my mistake :(
Yes, you have figured it right, namespace is NOT a subvolume of
unify. Also about the schedulers complaining about the not enough
space, all the schedulers now uses % based min-disk-size, (option
alu.limits.min-disk-size 10 #don't consider the node for scheduling if
the free disk space is less than 10%).
About the 'ls' failing, i may need little more info. First, which
patch are you using. And, an output of 'bt' after doing gdb glusterfs
-c /core.<pid>.
Regards,
Amar
On 6/21/07, *Dale Dude* <dale@xxxxxxxxxxxxxxx
<mailto:dale@xxxxxxxxxxxxxxx>> wrote:
How do I use the namespace option correctly? In the docs/examples I
think your using it wrong.
In cluster-client.vol you have volumes client1,client2,client3
configured and this line in cluster/unify:
option namespace client1 # this will not be storage child of unify.
subvolumes client1 client2 client3
I get the impression that namespace shouldnt be configured as a
subvolume of unify. If I use a namespace of a volume that is a
subvolume
of unify it complains as such. If I create a volume on the server
specifically for the namespace (i.e. 1 storage/posix called
volumenamespace) and in the client config use that then it doesnt
complain anymore. But I cant even ls on the mounted volume. I get this
debug found below. A 'df -h' looks correct (glusterfs 5.9T 400M 5.9T
1% /volumes).
Btw, I have to use the ALU scheduler because any other scheduler keeps
saying that there isnt enough space on any of the "bricks".