Re: Monitor rename / recreate issue -- probing state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 12/10/2015 04:00 AM, deeepdish wrote:
Hello,

I encountered a strange issue when rebuilding monitors reusing same
hostnames, however different IPs.

Steps to reproduce:

- Build monitor using ceph-deploy create mon <hostname1>
- Remove monitor
via http://docs.ceph.com/docs/master/rados/operations/add-or-rm-mons/
(remove monitor) ? I didn?t realize there was a ceph-deploy mon destroy
command at this point.
- Build a new monitor on same hardware using ceph-deploy create mon
<hostname1a>  # reason = to rename / change IP of monitor as per above link
- Monitor ends up in probing mode.   When connecting via the admin
socket, I see that there are no peers avail.   

The above behavior of only when reinstalling monitors.   I even tried
reinstalling the OS, however there?s a monmap embedded somewhere causing
the previous monitor hostnames / IPs to conflict with the new monitor?s
peering ability.  

[root@b02s08 ~]# 

On a reinstalled (not working) monitor:

sudo ceph --cluster=ceph --admin-daemon
/var/run/ceph/ceph-mon.smg01.asok mon_status
{
   "name": "smg01",
   "rank": 0,
   "state": "probing",
   "election_epoch": 0,
   "quorum": [],
   "outside_quorum": [
       "smg01"
   ],
   "extra_probe_peers": [
       "10.20.1.8:6789\/0",
       "10.20.10.14:6789\/0",
       "10.20.10.16:6789\/0",
       "10.20.10.18:6789\/0",
       "10.20.10.251:6789\/0",
       "10.20.10.252:6789\/0"
   ],
[snip]
}



This appears to be consistent with a wrongly populated 'mon_host' and
'mon_initial_members' in your ceph.conf.

 -Joao


Thanks Joao.   I had a look but my other 3 monitors are working just fine.   To be clear, I’ve confirmed the same behaviour on other monitor nodes that have been removed from the cluster and rebuild with a new IP (however same name).

[global]
fsid = (hidden)
mon_initial_members = smg01, smon01s, smon02s, b02s08
mon_host = 10.20.10.250, 10.20.10.251, 10.20.10.252, 10.20.1.8
public network = 10.20.10.0/24, 10.20.1.0/24
cluster network = 10.20.41.0/24

. . . 

[mon.smg01s]
host = smg01s
addr = 10.20.10.250:6789

[mon.smon01s]
host = smon01s
addr = 10.20.10.251:6789

[mon.smon02s]
host = smon02s
addr = 10.20.10.252:6789

[mon.b02s08]
host = b02s08
addr = 10.20.1.8:6789

# sudo ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.smg01.asok mon_status
{
    "name": "smg01",
    "rank": 0,
    "state": "probing",
    "election_epoch": 0,
    "quorum": [],
    "outside_quorum": [
        "smg01"
    ],
    "extra_probe_peers": [
        "10.20.1.8:6789\/0",
        "10.20.10.251:6789\/0",
        "10.20.10.252:6789\/0"
    ],
    "sync_provider": [],
    "monmap": {
        "epoch": 0,
        "fsid": “(hidden)",
        "modified": "0.000000",
        "created": "0.000000",
        "mons": [
            {
                "rank": 0,
                "name": "smg01",
                "addr": "10.20.10.250:6789\/0"
            },
            {
                "rank": 1,
                "name": "smon01s",
                "addr": "0.0.0.0:0\/1"
            },
            {
                "rank": 2,
                "name": "smon02s",
                "addr": "0.0.0.0:0\/2"
            },
            {
                "rank": 3,
                "name": "b02s08",
                "addr": "0.0.0.0:0\/3"
            }
        ]
    }
}

Processes running on the monitor node that’s in probing state:

# ps -ef | grep ceph
root      1140     1  0 Dec11 ?        00:05:07 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01
root      6406     1  0 Dec11 ?        00:05:10 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01
root      7712     1  0 Dec11 ?        00:05:09 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01
root      9105     1  0 Dec11 ?        00:05:11 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01
root     13098 30548  0 07:18 pts/1    00:00:00 grep --color=auto ceph
root     14243     1  0 Dec11 ?        00:05:09 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01
root     31222     1  0 05:39 ?        00:00:00 /bin/bash -c ulimit -n 32768; /usr/bin/ceph-mon -i smg01 --pid-file /var/run/ceph/mon.smg01.pid -c /etc/ceph/ceph.conf --cluster ceph -f
root     31226 31222  1 05:39 ?        00:01:39 /usr/bin/ceph-mon -i smg01 --pid-file /var/run/ceph/mon.smg01.pid -c /etc/ceph/ceph.conf --cluster ceph -f
root     31228     1  0 05:39 pts/1    00:00:15 python /usr/sbin/ceph-create-keys --cluster ceph -i smg01

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux