Hi! > Assuming you are doing a straight distribute there(?), if the user in Yes's it's a distributed volume. > question is hashed onto the brick that is 100% full you'll get a space Is there a way around this other than moving files away from this one brick by hand? > error. Not sure I followed your migration details though, when you say > "user directories were moved into one of the above folders" do you mean > copied directly onto the individual storage bricks? Yes, eg. user_a had the following directories server5:/storage/5/user_a server6:/storage/6/user_a Then we performed a move: ssh server5 "mv /storage/5/user_a/ /storage/5/cluster/user_a" ssh server6 "mv /storage/6/user_a/ /storage/6/cluster/user_a" This was done as it would not cause any network traffic. Then the volume was created like that: Brick1: 192.168.101.249:/storage/4/cluster Brick2: 192.168.101.248:/storage/5/cluster Brick3: 192.168.101.250:/storage/6/cluster Brick4: 192.168.101.247:/storage/7/cluster Brick5: 192.168.101.246:/storage/8/cluster Regards, Daniel > > On 01/19/2011 05:01 AM, zander at ekp.uni-karlsruhe.de wrote: >> mv: cannot create regular file `/storage/cluster/<etc...>': No space >> left on device >> >> Doing df -h tells me, however: >> >> glusterfs#192.168.101.247:/lemmy >> 104T 69T 36T 66% /storage/cluster >> >> It may be of importance that one brick in the cluster is actually 100% >> used. Others are almost completely empty. I am using GlusterFS.3.1.1, >> the file servers are running debian lenny or ubuntu server 10.04, >> clients are SLC4, SLC5, CentOS and ubuntu server 10.04. >> > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users