Re: UGRENT: add mon failed and ceph monitor refreshlog crazily

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It sounds a bit like the extra load on mon.e from the synchronization is 
preventing it from joining the quorum?  If you stop and restart mon.f it 
should pick a different mon to pull from, though.  Perhaps see if that 
makes a different mon drop out?  Then at least we'd understand what is 
going on...

sage

On Fri, 13 Feb 2015, minchen wrote:

> 
> ceph version is 0.80.4
> 
> when add mon.f to {b,c,d,e}, mon.e is out quorum, 
> mon.b, mon.c, mon.d are electing in cycle(restart a new election after
> leader win) .
> so,  i think current 4 monitors can exchange messages to each other
> successfully.
> 
> In addtion, mon.f is stuck at state synchronizing, and geting data from
> mon.e after probing. 
> 
> When I stop mon.f , mon.e goes back to quorum after a while, then ceph
> cluster becomes HEALTH_OK.
> But, all mon.b, mon.c, mon.d and mon.e logs are refreshing paxos acitive or
> updating messages many times per second, 
> and paxos commit seq is increasing fastly.  while the same situation not
> occurs in cluster of ?ceph-0.80.7 
> 
> If you are still confused, maybe I should reproduct this in our cluster, and
> get complete mon logs ...
> 
> ------------------ Original ------------------
> From:  "sweil";<sweil@xxxxxxxxxx>;
> Date:  Fri, Feb 13, 2015 10:28 PM
> To:  "minchen"<minchen@xxxxxxxxxxxxxxx>;
> Cc:  "ceph-users"<ceph-users@xxxxxxxxxxxxxx>; "joao"<joao@xxxxxxxxxx>;
> Subject:  Re:  UGRENT: add mon failed and ceph monitor
> refreshlog crazily
> 
> What version is this?
> 
> It's hard to tell from the logs below, but it looks like there might be a
> connectivity problem?  Is it able to exchange messages with the other
> monitors?
> 
> Perhaps more improtantly, though, if you simply stop the new mon.f, can
> mon.e join?  What is in its log?
> 
> sage
> 
> 
> On Fri, 13 Feb 2015, minchen wrote:
> 
> > Hi , 
> >   all developers and users
> > when i add a new mon to current mon cluter, failed with 2 mon out of
> quorum.
> >
> >
> > there are 5 mons in our ceph cluster:
> > epoch 7
> > fsid 0dfd2bd5-1896-4712-916b-ec02dcc7b049
> > last_changed 2015-02-13 09:11:45.758839
> > created 0.000000
> > 0: 10.117.16.17:6789/0 mon.b
> > 1: 10.118.32.7:6789/0 mon.cHEALTH_WARN 2 mons down, quorum 0,1,2 b,c,d
> > mon.e (rank 3) addr 10.122.0.9:6789/0 is down (out of quorum)
> > mon.f (rank 4) addr 10.122.48.11:6789/0 is down (out of quorum)
> > 2: 10.119.16.11:6789/0 mon.d
> > 3: 10.122.0.9:6789/0 mon.e
> > 4: 10.122.48.11:6789/0 mon.f
> >
> >
> > mon.f is newly added to montior cluster, but when starting mon.f,
> > it caused both  mon.e and mon.f out of quorum:
> > HEALTH_WARN 2 mons down, quorum 0,1,2 b,c,d
> > mon.e (rank 3) addr 10.122.0.9:6789/0 is down (out of quorum)
> > mon.f (rank 4) addr 10.122.48.11:6789/0 is down (out of quorum)
> >
> >
> > mon.b ,mon.c, mon.d, log refresh crazily as following:
> > Feb 13 09:37:34 root ceph-mon: 2015-02-13 09:37:34.063628 7f7b64e14700  1
> mon.b@0(leader).paxos(paxos active c 11818589..11819234) is_readable
> now=2015-02-13 09:37:34.063629 lease_expire=2015-02-13 09:37:38.205219 has
> v0 lc 11819234
> > Feb 13 09:37:34 root ceph-mon: 2015-02-13 09:37:34.090647 7f7b64e14700  1
> mon.b@0(leader).paxos(paxos active c 11818589..11819234) is_readable
> now=2015-02-13 09:37:34.090648 lease_expire=2015-02-13 09:37:38.205219 has
> v0 lc 11819234
> > Feb 13 09:37:34 root ceph-mon: 2015-02-13 09:37:34.090661 7f7b64e14700  1
> mon.b@0(leader).paxos(paxos active c 11818589..11819234) is_readable
> now=2015-02-13 09:37:34.090662 lease_expire=2015-02-13 09:37:38.205219 has
> v0 lc 11819234
> > ......
> >
> >
> > and mon.f log :
> >
> >
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.526676 7f3931dfd7c0  0
> ceph version 0.80.4 (7c241cfaa6c8c068bc9da8578ca00b9f4fc7567f), process
> ceph-mon, pid 30639
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.607412 7f3931dfd7c0  0
> mon.f does not exist in monmap, will attempt to join an existing cluster
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.609838 7f3931dfd7c0  0
> starting mon.f rank -1 at 10.122.48.11:6789/0 mon_data /osd/ceph/mon fsid
> 0dfd2bd5-1896-4712-916b-ec02dcc7b049
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.610076 7f3931dfd7c0  1
> mon.f@-1(probing) e0 preinit fsid 0dfd2bd5-1896-4712-916b-ec02dcc7b049
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.636499 7f392a504700  0
> -- 10.122.48.11:6789/0 >> 10.119.16.11:6789/0 pipe(0x7f3934ebfb80 sd=26
> :6789 s=0 pgs=0 cs=0 l=0 c=0x7f3934ea9ce0).accept connect_seq 0 vs existing
> 0 state wait
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.636797 7f392a201700  0
> -- 10.122.48.11:6789/0 >> 10.122.0.9:6789/0 pipe(0x7f3934ec0800 sd=29 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa940).accept connect_seq 0 vs existing 0
> state wait
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.636968 7f392a403700  0
> -- 10.122.48.11:6789/0 >> 10.118.32.7:6789/0 pipe(0x7f3934ec0080 sd=27 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934ea9e40).accept connect_seq 0 vs existing 0
> state wait
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.637037 7f392a302700  0
> -- 10.122.48.11:6789/0 >> 10.117.16.17:6789/0 pipe(0x7f3934ebfe00 sd=28
> :6789 s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa260).accept connect_seq 0 vs existing
> 0 state wait
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.638854 7f392c00a700  0
> mon.f@-1(probing) e7  my rank is now 4 (was -1)
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639365 7f392c00a700  1
> mon.f@4(synchronizing) e7 sync_obtain_latest_monmap
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639494 7f392b008700  0
> -- 10.122.48.11:6789/0 >> 10.122.0.9:6789/0 pipe(0x7f3934ec0580 sd=17 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa680).accept connect_seq 2 vs existing 0
> state connecting
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639513 7f392b008700  0
> -- 10.122.48.11:6789/0 >> 10.122.0.9:6789/0 pipe(0x7f3934ec0580 sd=17 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa680).accept we reset (peer sent cseq 2,
> 0x7f3934ebf400.cseq = 0), sending RESETSESSION
> > ......
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.643159 7f392af07700  0
> -- 10.122.48.11:6789/0 >> 10.119.16.11:6789/0 pipe(0x7f3934ec1700 sd=28
> :6789 s=0 pgs=0 cs=0 l=0 c=0x7f3934eab2e0).accept connect_seq 0 vs existing
> 0 state wait
> >
> >
> >
> >
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.637037 7f392a302700  0
> -- 10.122.48.11:6789/0 >> 10.117.16.17:6789/0 pipe(0x7f3934ebfe00 sd=28
> :6789 s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa260).accept connect_seq 0 vs existing
> 0 state wait
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.638854 7f392c00a700  0
> mon.f@-1(probing) e7  my rank is now 4 (was -1)
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639365 7f392c00a700  1
> mon.f@4(synchronizing) e7 sync_obtain_latest_monmap
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639494 7f392b008700  0
> -- 10.122.48.11:6789/0 >> 10.122.0.9:6789/0 pipe(0x7f3934ec0580 sd=17 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa680).accept connect_seq 2 vs existing 0
> state connecting
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.639513 7f392b008700  0
> -- 10.122.48.11:6789/0 >> 10.122.0.9:6789/0 pipe(0x7f3934ec0580 sd=17 :6789
> s=0 pgs=0 cs=0 l=0 c=0x7f3934eaa680).accept we reset (peer sent cseq 2,
> 0x7f3934ebf400.cseq = 0), sending RESETSESSION
> > ......
> > Feb 13 09:16:26 root ceph-mon: 2015-02-13 09:16:26.643273 7f392c00a700  1
> mon.f@4(synchronizing) e7 sync_obtain_latest_monmap obtained monmap e7
> > Feb 13 09:17:26 root ceph-mon: 2015-02-13 09:17:26.611550 7f392c80b700  0
> mon.f@4(synchronizing).data_health(0) update_stats avail 99% total 911815680
> used 33132 avail 911782548
> > Feb 13 09:17:26 root ceph-mon: 2015-02-13 09:17:26.708961 7f392c00a700  1
> mon.f@4(synchronizing) e7 sync_obtain_latest_monmap
> > Feb 13 09:17:26 root ceph-mon: 2015-02-13 09:17:26.709063 7f392c00a700  1
> mon.f@4(synchronizing) e7 sync_obtain_latest_monmap obtained monmap e7
> >
> >
> >
> >
> > someone can help? thank you!
> >
> >
> > minchen
> >
> >
> >
> >
> >
> >
> >
> 
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux