gluster volume rebalance md1 status gives : Node Rebalanced-files size scanned failures skipped status run time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 3837 6.3GB 163881 0 0 completed 365.00 tsunami5 179 343.8MB 163882 0 0 completed 353.00 tsunami3 6786 4.7GB 163882 0 0 completed 416.00 tsunami6 0 0Bytes 163882 0 0 completed 353.00 tsunami4 0 0Bytes 163882 0 0 completed 353.00 tsunami2 0 0Bytes 163882 0 0 completed 353.00 volume rebalance: md1: success:
but no change on the bricks for the directory, still empty except on 2 bricks. Should I remove files in the .glusterfs directory on the 2 bricks associated to these "---T" files ?
Thanks,
A.
On Monday 23 February 2015 21:40:41 Ravishankar N wrote: On 02/23/2015 09:19 PM, Alessandro Ipe wrote:
On 4 of the 6 bricks, it is empty. However, on tsunami 3-4, ls -lsa gives total 16 d--------- 2 root root 61440 Feb 23 15:42 . drwxrwxrwx 3 gerb users 61 Feb 22 21:10 .. ---------T 2 gerb users 0 Apr 16 2014 akonadi-googledata-1.2.0-2.5.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 bluedevil-debugsource-1.2.2-1.8.3.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 bovo-4.7.4-3.12.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 digikam-debugsource-2.2.0-3.12.9.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 dolphin-debuginfo-4.7.4-4.22.6.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 freetds-doc-0.91-2.5.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kanagram-debuginfo-4.7.4-2.10.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kdebase4-runtime-4.7.4-3.17.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kdebindings-smokegen-debuginfo-4.7.4-2.9.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kdesdk4-strigi-debuginfo-4.7.4-3.12.5.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kradio-4.0.2-9.9.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kremotecontrol-4.7.4-2.12.9.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 kreversi-debuginfo-4.7.4-3.12.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 krfb-4.7.4-2.13.6.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 krusader-doc-2.0.0-23.9.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libalkimia-devel-4.3.1-2.5.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libdmtx0-0.7.4-2.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libdmtx0-debuginfo-0.7.4-2.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libkdegames4-debuginfo-4.7.4-3.12.7.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libksane0-4.7.4-2.10.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libkvkontakte-debugsource-1.0.0-2.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libmediawiki-debugsource-2.5.0-4.6.1.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 libsmokeqt-4.7.4-2.10.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 NetworkManager-vpnc-kde4-0.9.1git20111027-1.11.5.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 qtcurve-kde4-1.8.8-3.6.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 QtZeitgeist-devel-0.7.0-7.4.2.i586.rpm ---------T 2 gerb users 0 Apr 16 2014 umbrello-4.7.4-3.12.5.i586.rpm
so that might be the reason of error. How can I fix this ?
The '----T' files are DHT link-to files. The actual files must be present on the other distribute subolumes (tsunami 1-2 or tsunami 5-6) in the same path. But since that doesn't seem to be the case, the something went wrong with the re-balance process. You could run `gluster volume rebalance <volname> start+status` again and see if they disappear.
Thanks,
A.
On Monday 23 February 2015 21:06:58 Ravishankar N wrote: Just noticed that your `gluster volume status` shows that rebalance was triggered. Maybe DHT developers can help out. I see a similar bug has been fixed some time back. On 02/23/2015 08:15 PM, Alessandro Ipe wrote:
Hi,
Gluster version is 3.5.3-1. /var/log/gluster.log (client log) gives during the rm -rf the following logs: [2015-02-23 14:42:50.180091] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-2: remote operation failed: Directory not empty [2015-02-23 14:42:50.180134] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-3: remote operation failed: Directory not empty [2015-02-23 14:42:50.180740] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-5: remote operation failed: File exists. Path: /linux/suse/12.1/KDE4.7.4/i586 [2015-02-23 14:42:50.180772] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-4: remote operation failed: File exists. Path: /linux/suse/12.1/KDE4.7.4/i586 [2015-02-23 14:42:50.181129] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-3: remote operation failed: File exists. Path: /linux/suse/12.1/KDE4.7.4/i586 [2015-02-23 14:42:50.181160] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-2: remote operation failed: File exists. Path: /linux/suse/12.1/KDE4.7.4/i586 [2015-02-23 14:42:50.319213] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-3: remote operation failed: Directory not empty [2015-02-23 14:42:50.319762] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-2: remote operation failed: Directory not empty [2015-02-23 14:42:50.320501] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-0: remote operation failed: File exists. Path: /linux/suse/12.1/src-oss/suse/src [2015-02-23 14:42:50.320552] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-1: remote operation failed: File exists. Path: /linux/suse/12.1/src-oss/suse/src [2015-02-23 14:42:50.320842] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-2: remote operation failed: File exists. Path: /linux/suse/12.1/src-oss/suse/src [2015-02-23 14:42:50.320884] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-3: remote operation failed: File exists. Path: /linux/suse/12.1/src-oss/suse/src [2015-02-23 14:42:50.438982] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-3: remote operation failed: Directory not empty [2015-02-23 14:42:50.439347] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-2: remote operation failed: Directory not empty [2015-02-23 14:42:50.440235] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-0: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/noarch [2015-02-23 14:42:50.440344] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-1: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/noarch [2015-02-23 14:42:50.440603] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-2: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/noarch [2015-02-23 14:42:50.440665] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-3: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/noarch [2015-02-23 14:42:50.680827] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-2: remote operation failed: Directory not empty [2015-02-23 14:42:50.681721] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 0-md1-client-3: remote operation failed: Directory not empty [2015-02-23 14:42:50.682482] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-3: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/i586 [2015-02-23 14:42:50.682517] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 0-md1-client-2: remote operation failed: File exists. Path: /linux/suse/12.1/oss/suse/i586
Thanks,
A.
On Monday 23 February 2015 20:06:17 Ravishankar N wrote: On 02/23/2015 07:04 PM, Alessandro Ipe wrote:
Hi Ravi,
gluster volume status md1 returns Status of volume: md1 Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick tsunami1:/data/glusterfs/md1/brick1 49157 Y 2260 Brick tsunami2:/data/glusterfs/md1/brick1 49152 Y 2320 Brick tsunami3:/data/glusterfs/md1/brick1 49156 Y 20715 Brick tsunami4:/data/glusterfs/md1/brick1 49156 Y 10544 Brick tsunami5:/data/glusterfs/md1/brick1 49152 Y 12588 Brick tsunami6:/data/glusterfs/md1/brick1 49152 Y 12242 Self-heal Daemon on localhost N/A Y 2336 Self-heal Daemon on tsunami2 N/A Y 2359 Self-heal Daemon on tsunami5 N/A Y 27619 Self-heal Daemon on tsunami4 N/A Y 12318 Self-heal Daemon on tsunami3 N/A Y 19118 Self-heal Daemon on tsunami6 N/A Y 27650
Task Status of Volume md1 ------------------------------------------------------------------------------ Task : Rebalance ID : 9dfee1a2-49ac-4766-bdb6-00de5e5883f6 Status : completed so it seems that all brick server are up.
gluster volume heal md1 info returns Brick tsunami1.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Brick tsunami2.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Brick tsunami3.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Brick tsunami4.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Brick tsunami5.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Brick tsunami6.oma.be:/data/glusterfs/md1/brick1/ Number of entries: 0
Should I run "gluster volume heal md1 full" ?
Hi Alessandro,
Thanks,
A.
On Monday 23 February 2015 18:12:43 Ravishankar N wrote: On 02/23/2015 05:42 PM, Alessandro Ipe wrote:
Hi,
We have a "md1" volume under gluster 3.5.3 over 6 servers configured as distributed and replicated. When trying on a client, thourgh fuse mount (which turns out to be also a brick server) to delete (as root) recursively a directory with "rm -rf /home/.md1/linux/suse/12.1", I get the error messages
rm: cannot remove ‘/home/.md1/linux/suse/12.1/KDE4.7.4/i586’: Directory not empty rm: cannot remove ‘/home/.md1/linux/suse/12.1/src-oss/suse/src’: Directory not empty rm: cannot remove ‘/home/.md1/linux/suse/12.1/oss/suse/noarch’: Directory not empty rm: cannot remove ‘/home/.md1/linux/suse/12.1/oss/suse/i586’: Directory not empty (the same occurs as unprivileged user but with "Permission denied".)
while a "ls -Ral /home/.md1/linux/suse/12.1" gives me /home/.md1/linux/suse/12.1: total 0 drwxrwxrwx 5 gerb users 151 Feb 20 16:22 . drwxr-xr-x 6 gerb users 245 Feb 23 12:55 .. drwxrwxrwx 3 gerb users 95 Feb 23 13:03 KDE4.7.4 drwxrwxrwx 3 gerb users 311 Feb 20 16:57 oss drwxrwxrwx 3 gerb users 86 Feb 20 16:20 src-oss
/home/.md1/linux/suse/12.1/KDE4.7.4: total 28 drwxrwxrwx 3 gerb users 95 Feb 23 13:03 . drwxrwxrwx 5 gerb users 151 Feb 20 16:22 .. d--------- 2 root root 61452 Feb 23 13:03 i586
/home/.md1/linux/suse/12.1/KDE4.7.4/i586: total 28 d--------- 2 root root 61452 Feb 23 13:03 . drwxrwxrwx 3 gerb users 95 Feb 23 13:03 ..
/home/.md1/linux/suse/12.1/oss: total 0 drwxrwxrwx 3 gerb users 311 Feb 20 16:57 . drwxrwxrwx 5 gerb users 151 Feb 20 16:22 .. drwxrwxrwx 4 gerb users 90 Feb 23 13:03 suse
/home/.md1/linux/suse/12.1/oss/suse: total 536 drwxrwxrwx 4 gerb users 90 Feb 23 13:03 . drwxrwxrwx 3 gerb users 311 Feb 20 16:57 .. d--------- 2 root root 368652 Feb 23 13:03 i586 d--------- 2 root root 196620 Feb 23 13:03 noarch
/home/.md1/linux/suse/12.1/oss/suse/i586: total 360 d--------- 2 root root 368652 Feb 23 13:03 . drwxrwxrwx 4 gerb users 90 Feb 23 13:03 ..
/home/.md1/linux/suse/12.1/oss/suse/noarch: total 176 d--------- 2 root root 196620 Feb 23 13:03 . drwxrwxrwx 4 gerb users 90 Feb 23 13:03 ..
/home/.md1/linux/suse/12.1/src-oss: total 0 drwxrwxrwx 3 gerb users 86 Feb 20 16:20 . drwxrwxrwx 5 gerb users 151 Feb 20 16:22 .. drwxrwxrwx 3 gerb users 48 Feb 23 13:03 suse
/home/.md1/linux/suse/12.1/src-oss/suse: total 220 drwxrwxrwx 3 gerb users 48 Feb 23 13:03 . drwxrwxrwx 3 gerb users 86 Feb 20 16:20 .. d--------- 2 root root 225292 Feb 23 13:03 src
/home/.md1/linux/suse/12.1/src-oss/suse/src: total 220 d--------- 2 root root 225292 Feb 23 13:03 . drwxrwxrwx 3 gerb users 48 Feb 23 13:03 ..
Is there a cure such as manually forcing a healing on that directory ?
Many thanks,
Alessandro.
gluster volume info md1 outputs: Volume Name: md1 Type: Distributed-Replicate Volume ID: 6da4b915-1def-4df4-a41c-2f3300ebf16b Status: Started Number of Bricks: 3 x 2 = 6 Transport-type: tcp Bricks: Brick1: tsunami1:/data/glusterfs/md1/brick1 Brick2: tsunami2:/data/glusterfs/md1/brick1 Brick3: tsunami3:/data/glusterfs/md1/brick1 Brick4: tsunami4:/data/glusterfs/md1/brick1 Brick5: tsunami5:/data/glusterfs/md1/brick1 Brick6: tsunami6:/data/glusterfs/md1/brick1 Options Reconfigured: performance.write-behind: on performance.write-behind-window-size: 4MB performance.flush-behind: off performance.io-thread-count: 64 performance.cache-size: 512MB nfs.disable: on features.quota: off cluster.read-hash-mode: 2 server.allow-insecure: on cluster.lookup-unhashed: off _______________________________________________Gluster-users mailing listGluster-users@xxxxxxxxxxxhttp://www.gluster.org/mailman/listinfo/gluster-users
-- |
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users