>[root at 3d13 ~]# rm -rfv /flock/proj/tele2_holland/rnd/comp/010/v003 >rm: cannot remove directory `/flock/proj/tele2_holland/rnd/comp/010/v003': Directory not empty When I had this issue it was because I modified the files outside of glusterfs - so for example, when gluster was not running, I moved/modified files. I believe you have to run the scale-n-defrag.sh script that you'll find in the contrib directory of the gluster source. P On Mon, Sep 20, 2010 at 4:49 AM, Thomas Ericsson <thomas.ericsson at fido.se> wrote: > I can not delete quite a few directories on our glusterfs mounts. The error is "Directory not empty". A listing shows no files in the directory, however if i do a listing on the brick volumes some of them show files. > > Any idea as of how this can happen and how to remove the directories. Would it be safe to remove the "invisible" files straight from the brick volume? > > Best regards > Thomas > > > From a glusterfs client > [root at 3d13 ~]# ls -lai /flock/proj/tele2_holland/rnd/comp/010/v003/ > total 0 > 38939716797 drwxrwxr-x 2 poal FidoUsers 162 Sep 16 09:15 . > 60331700537 drwxrwxr-x 5 poal FidoUsers 536 Sep ?4 01:24 .. > [root at 3d13 ~]# rm -rfv /flock/proj/tele2_holland/rnd/comp/010/v003 > rm: cannot remove directory `/flock/proj/tele2_holland/rnd/comp/010/v003': Directory not empty > > From a glusterfsd brick > flock01 ~ # ls -lai /node04/storage/proj/tele2_holland/rnd/comp/010/v003/ > total 0 > ?305414438 drwxrwxr-x 2 1038 fido_user 57 Sep 16 09:15 . > 7541462567 drwxrwxr-x 5 1038 fido_user 61 Jul ?7 09:44 .. > ?305414403 ---------T 1 root root ? ? ? 0 Sep ?4 01:24 tele2_holland_010_comp_v003.0031.exr > > From another glusterfsd brick > flock04 ~ # ls -lai /node03/storage/proj/tele2_holland/rnd/comp/010/v003/ > total 0 > 4861583534 drwxrwxr-x 2 1038 ?500 57 Sep 16 09:15 . > ?280040615 drwxrwxr-x 5 1038 ?500 61 Jul ?7 09:44 .. > 4861671820 ---------T 1 root root ?0 Sep ?4 01:24 tele2_holland_010_comp_v003.0007.exr > > > -------------------------------------------------------------- > > Server and clients are vesion 2.0.8 with FUSE 2.7.4 > > Server config > flock04 ~ # cat /usr/local/etc/glusterfs/glusterfs.server > volume posix01 > ?type storage/posix > ?option directory /node01/storage > end-volume > > volume locks01 > ?type features/locks > ?subvolumes posix01 > end-volume > > volume brick01 > ?type performance/io-threads > ?option thread-count 2 > ?subvolumes locks01 > end-volume > > volume posix02 > ?type storage/posix > ?option directory /node02/storage > end-volume > > volume locks02 > ?type features/locks > ?subvolumes posix02 > end-volume > > volume brick02 > ?type performance/io-threads > ?option thread-count 2 > ?subvolumes locks02 > end-volume > > volume posix03 > ?type storage/posix > ?option directory /node03/storage > end-volume > > volume locks03 > ?type features/locks > ?subvolumes posix03 > end-volume > > volume brick03 > ?type performance/io-threads > ?option thread-count 32 > ?subvolumes locks03 > end-volume > > volume posix04 > ?type storage/posix > ?option directory /node04/storage > end-volume > > volume locks04 > ?type features/locks > ?subvolumes posix04 > end-volume > > volume brick04 > ?type performance/io-threads > ?option thread-count 32 > ?subvolumes locks04 > end-volume > > volume server > ?type protocol/server > ?option transport-type ib-verbs/server > ?option auth.addr.brick01.allow * > ?option auth.addr.brick02.allow * > ?option auth.addr.brick03.allow * > ?option auth.addr.brick04.allow * > ?subvolumes brick01 brick02 brick03 brick04 > end-volume > > volume tcp_server > ?type protocol/server > ?option transport-type tcp/server > ?option transport.socket.nodelay on > ?option auth.addr.brick01.allow * > ?option auth.addr.brick02.allow * > ?option auth.addr.brick03.allow * > ?option auth.addr.brick04.allow * > ?subvolumes brick01 brick02 brick03 brick04 > end-volume > > Client config > volume remote01 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock01 > ?option remote-subvolume brick03 > end-volume > > volume remote02 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock01 > ?option remote-subvolume brick04 > end-volume > > volume remote03 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock03 > ?option remote-subvolume brick03 > end-volume > > volume remote04 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock03 > ?option remote-subvolume brick04 > end-volume > > volume remote05 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock04 > ?option remote-subvolume brick03 > end-volume > > volume remote06 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock04 > ?option remote-subvolume brick04 > end-volume > > volume remote07 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock08 > ?option remote-subvolume brick03 > end-volume > > volume remote08 > ?type protocol/client > ?option transport-type ib-verbs/client > ?option remote-host flock08 > ?option remote-subvolume brick04 > end-volume > > volume distribute > ?type cluster/distribute > ?subvolumes remote01 remote02 remote03 remote04 remote05 remote06 remote07 remote08 > end-volume > > volume writebehind > ?type performance/write-behind > ?option cache-size 16MB > ?option flush-behind on > ?subvolumes distribute > end-volume > > volume cache01 > ?type performance/io-cache > ?option cache-size 512MB > ?option cache-timeout 30 > ?subvolumes writebehind > end-volume > > volume readahead > ?type performance/read-ahead > ?option page-count 16 ? ? ? ? ? # 2 is default option > ?option force-atime-update off # default is off > ?subvolumes cache01 > end-volume > > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > -- http://philcryer.com