3.4 volume rebalance status outputting nonsense

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ever since updating my 25x2 distributed-replicate volume from 3.3.1 to 
3.4, "gluster volume rebalance $myvolume status" has been outputting 
nonsensical information and hostnames:

[root at ml54 ~]# gluster volume rebalance bigdata status
                                     Node Rebalanced-files size       
scanned      failures         status run time in secs
                                ---------      ----------- -----------   
-----------   -----------   ------------ --------------
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                localhost                0 0Bytes       
1407934             5    in progress         60013.00
                                     ml59            22698 3.0GB       
1022093         23956    in progress         60013.00
volume rebalance: bigdata: success:

The first obvious weird thing here is that everything is listed as 
localhost, whereas one would expect to have one line per server. The 
other weird thing is that almost everything is at 0 even though this has 
been running for ~16 hours.

Same command from another server at almost the same time:
[root at ml01 ~]# gluster volume rebalance bigdata status
                                     Node Rebalanced-files size       
scanned      failures         status run time in secs
                                ---------      ----------- -----------   
-----------   -----------   ------------ --------------
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                localhost              670 9.4MB       
2917849         30813    in progress         60020.00
                                     ml26                0 0Bytes       
1406422             0    in progress         60019.00
volume rebalance: bigdata: success:

Running the command yet again from ml01 a few minutes later:

[root at ml01 ~]# gluster volume rebalance bigdata status
                                     Node Rebalanced-files size       
scanned      failures         status run time in secs
                                ---------      ----------- -----------   
-----------   -----------   ------------ --------------
                                localhost              670 9.4MB       
2918822         30813    in progress         60458.00
                                     ml57                0 0Bytes       
1418620             0    in progress         60457.00
                                     ml59            22908 3.0GB       
1025934         24101    in progress         60457.00
                                     ml47            19142 6.8GB       
1119561         44194    in progress         60457.00
                                     ml56             9789 1.4GB       
1276682         78928    in progress         60457.00
                                     ml55            23265 3.1GB       
1002220          8771    in progress         60457.00
                                     ml26                0 0Bytes       
1419357             0    in progress         60457.00
                                     ml30            23844 2.6GB        
957613         15464    in progress         60457.00
                                     ml29                0 0Bytes       
1398930             0    in progress         60457.00
                                     ml46                0 0Bytes       
1414131             0    in progress         60457.00
                                     ml44                0 0Bytes       
2948809             0    in progress         60457.00
                                     ml31                0 0Bytes       
1419496             0    in progress         60457.00
                                     ml25             3711 1.0GB       
2929044         48441    in progress         60457.00
                                     ml43            26180 6.6GB        
844032         11576    in progress         60457.00
                                     ml54                0 0Bytes       
1419523             5    in progress         60457.00
                                     ml45            26230 2.7GB        
732983         19163    in progress         60457.00
                                     ml40            20623 19.9GB       
1452570         38991    in progress         60457.00
                                     ml52                0 0Bytes       
2932022             0    in progress         60457.00
                                     ml48                0 0Bytes       
2918224             0    in progress         60457.00
                                     ml41                0 0Bytes       
2950754             0    in progress         60457.00
                                     ml51            27097 5.4GB        
564416          1206    in progress         60457.00

That output makes more sense, although the zeros for ml57, 29, 46, 44, 
31, 54, 52, 48 and 41 seem a bit worrisome. I couldn't find anything 
abnormal in the logs. As of this writing, ml54 is still outputting 
all-localhost nonsense.

Has anyone else encountered this issue?

Pierre-Francois



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux