"gluster volume replace-brick ... status" breaks when executed on multiple nodes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I've executed "gluster volume replace-brick ... status" on multiple peers at the same time, which resulted in quite an interesting breakage.

It's no longer possible to pause/abort/status/start the replace-brick operation.

Please advise. I'm running glusterfs 3.2.6.

root at ca2:~# gluster volume replace-brick sites ca1-int:/data/glusterfs ca2-int:/data/ca1 status
replace-brick status unknown
root at ca2:~# gluster volume replace-brick sites ca1-int:/data/glusterfs ca2-int:/data/ca1 pause
replace-brick pause failed
root at ca2:~# gluster volume replace-brick sites ca1-int:/data/glusterfs ca2-int:/data/ca1 abort
replace-brick abort failed
root at ca2:~# gluster volume replace-brick sites ca1-int:/data/glusterfs ca2-int:/data/ca1 start
replace-brick failed to start



-- 
Tomasz Chmielewski
http://www.ptraveler.com


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux