Re: How to remove mds from cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I try myself,  use these command and recover to Health Ok now.  But i do not know why were these command work, in my opinion , fail mds node first and rm failed mds node

root@ceph01-vm:~# ceph mds fail 0
failed mds.0
root@ceph01-vm:~# ceph -s
    cluster 075f1aae-48de-412e-b024-b0f014dbc8cf
     health HEALTH_ERR mds rank 0 has failed; mds cluster is degraded
     monmap e2: 3 mons at {ceph01-vm=192.168.123.251:6789/0,ceph02-vm=192.168.123.252:6789/0,ceph04-vm=192.168.123.250:6789/0}, election epoch 128, quorum 0,1,2 ceph04-vm,ceph01-vm,ceph02-vm
     mdsmap e68: 0/1/1 up, 1 failed
     osdmap e588: 8 osds: 8 up, 8 in
      pgmap v285967: 2392 pgs, 21 pools, 4990 MB data, 1391 objects
            15173 MB used, 2768 GB / 2790 GB avail
                2392 active+clean
root@ceph01-vm:~#  ceph mds rm 0 mds.ceph06-vm  
mds gid 0 dne
root@ceph01-vm:~# ceph -s
    cluster 075f1aae-48de-412e-b024-b0f014dbc8cf
     health HEALTH_ERR mds rank 0 has failed; mds cluster is degraded
     monmap e2: 3 mons at {ceph01-vm=192.168.123.251:6789/0,ceph02-vm=192.168.123.252:6789/0,ceph04-vm=192.168.123.250:6789/0}, election epoch 128, quorum 0,1,2 ceph04-vm,ceph01-vm,ceph02-vm
     mdsmap e69: 0/1/1 up, 1 failed
     osdmap e588: 8 osds: 8 up, 8 in
      pgmap v285970: 2392 pgs, 21 pools, 4990 MB data, 1391 objects
            15173 MB used, 2768 GB / 2790 GB avail
                2392 active+clean
root@ceph01-vm:~# ceph mds newfs 1 0 --yes-i-really-mean-it 
filesystem 'cephfs' already exists
root@ceph01-vm:~# ceph -s
    cluster 075f1aae-48de-412e-b024-b0f014dbc8cf
     health HEALTH_ERR mds rank 0 has failed; mds cluster is degraded
     monmap e2: 3 mons at {ceph01-vm=192.168.123.251:6789/0,ceph02-vm=192.168.123.252:6789/0,ceph04-vm=192.168.123.250:6789/0}, election epoch 128, quorum 0,1,2 ceph04-vm,ceph01-vm,ceph02-vm
     mdsmap e70: 0/1/1 up, 1 failed
     osdmap e588: 8 osds: 8 up, 8 in
      pgmap v285973: 2392 pgs, 21 pools, 4990 MB data, 1391 objects
            15173 MB used, 2768 GB / 2790 GB avail
                2392 active+clean
root@ceph01-vm:~# ceph mds rmfailed 0
root@ceph01-vm:~# ceph -s
    cluster 075f1aae-48de-412e-b024-b0f014dbc8cf
     health HEALTH_OK
     monmap e2: 3 mons at {ceph01-vm=192.168.123.251:6789/0,ceph02-vm=192.168.123.252:6789/0,ceph04-vm=192.168.123.250:6789/0}, election epoch 128, quorum 0,1,2 ceph04-vm,ceph01-vm,ceph02-vm
     mdsmap e71: 0/1/1 up
     osdmap e588: 8 osds: 8 up, 8 in
      pgmap v286028: 2392 pgs, 21 pools, 4990 MB data, 1391 objects
            15174 MB used, 2768 GB / 2790 GB avail
                2392 active+clean

2015-01-05 15:03 GMT+07:00 debian Only <onlydebian@xxxxxxxxx>:
i use 0.87,  in side ceph.conf, do not have mds.0  related config

i did 
root@ceph06-vm:~# ceph mds rm 0 mds.ceph06-vm    
mds gid 0 dne

2015-01-05 11:15 GMT+07:00 Lindsay Mathieson <lindsay.mathieson@xxxxxxxxx>:
Did you remove the mds.0 entry from ceph.conf?

On 5 January 2015 at 14:13, debian Only <onlydebian@xxxxxxxxx> wrote:
i have tried ' ceph mds newfs 1 0 --yes-i-really-mean-it'    but not fix the problem

2014-12-30 17:42 GMT+07:00 Lindsay Mathieson <lindsay.mathieson@xxxxxxxxx>:

On Tue, 30 Dec 2014 03:11:25 PM debian Only wrote:

> ceph 0.87 , Debian 7.5,   anyone can help ?

>

> 2014-12-29 20:03 GMT+07:00 debian Only <onlydebian@xxxxxxxxx>:

> i want to move mds from one host to another.

>

> how to do it ?

>

> what did i do as below, but ceph health not ok, mds was not removed :

>

> root@ceph06-vm:~# ceph mds rm 0 mds.ceph06-vm    

> mds gid 0 dne

>

> root@ceph06-vm:~# ceph health detail

> HEALTH_WARN mds ceph06-vm is laggy

> mds.ceph06-vm at 192.168.123.248:6800/4350 is laggy/unresponsive

 

I removed an mds using this guide:

 

http://www.sebastien-han.fr/blog/2012/07/04/remove-a-mds-server-from-a-ceph-cluster/

 

and ran into your problem, which is also mentioned there.

 

resolved it using the guide suggestion:

 

$ ceph mds newfs metadata data --yes-i-really-mean-it

 

--

Lindsay


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





--
Lindsay


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux