Re: Monitor Restart triggers half of our OSDs marked down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi ,

Even I have seen same logs in osds , but the steps I followed were:

Setup : 4 osds-nodes node1,node2,node3,node4
Each node contains 8 osds.

Node1 got rebooted.

But osd in node2 went down.

Logs from monitor:

2015-02-02 20:28:28.766087 7fbaabea4700  1 mon.rack1-ram-6@0(leader).osd e248 prepare_failure osd.2 10.242.42.114:6871/25606 from osd.13 10.242.42.110:6835/25881 is reporting failure:1
2015-02-02 20:28:28.766099 7fbaabea4700  0 log_channel(cluster) log [DBG] : osd.2 10.242.42.114:6871/25606 reported failed by osd.13 10.242.42.110:6835/25881
2015-02-02 20:28:32.977235 7fbaabea4700  1 mon.rack1-ram-6@0(leader).osd e248 prepare_failure osd.2 10.242.42.114:6871/25606 from osd.23 10.242.42.112:6810/23707 is reporting failure:1
2015-02-02 20:28:32.977244 7fbaabea4700  0 log_channel(cluster) log [DBG] : osd.2 10.242.42.114:6871/25606 reported failed by osd.23 10.242.42.112:6810/23707
2015-02-02 20:28:33.012054 7fbaabea4700  1 mon.rack1-ram-6@0(leader).osd e248 prepare_failure osd.2 10.242.42.114:6871/25606 from osd.22 10.242.42.112:6822/23764 is reporting failure:1
2015-02-02 20:28:33.012075 7fbaabea4700  0 log_channel(cluster) log [DBG] : osd.2 10.242.42.114:6871/25606 reported failed by osd.22 10.242.42.112:6822/23764
2015-02-02 20:28:33.012087 7fbaabea4700  1 mon.rack1-ram-6@0(leader).osd e248  we have enough reports/reporters to mark osd.2 down
2015-02-02 20:28:33.012098 7fbaabea4700  0 log_channel(cluster) log [INF] : osd.2 10.242.42.114:6871/25606 failed (3 reports from 3 peers after 21.000102 >= grace 20.000000)
2015-02-02 20:28:33.071370 7fbaae0ea700  1 mon.rack1-ram-6@0(leader).osd e249 e249: 32 osds: 23 up, 32 in

Thanks
Sahana


-----Original Message-----
From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Dan van der Ster
Sent: Thursday, February 05, 2015 2:41 PM
To: Sage Weil
Cc: ceph-users@xxxxxxxxxxxxxx
Subject: Re:  Monitor Restart triggers half of our OSDs marked down

On Thu, Feb 5, 2015 at 9:54 AM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> On Thu, 5 Feb 2015, Dan van der Ster wrote:
>> Hi,
>> We also have seen this once after upgrading to 0.80.8 (from dumpling).
>> Last week we had a network outage which marked out around 1/3rd of
>> our OSDs. The outage lasted less than a minute -- all the OSDs were
>> brought up once the network was restored.
>>
>> Then 30 minutes later I restarted one monitor to roll out a small
>> config change (changing leveldb log path). Surprisingly that resulted
>> in many OSDs (but seemingly fewer than before) being marked out again
>> then quickly marked in again.
>
> Did the 'wrongly marked down' messages appear in ceph.log?
>

Yes. Here's the start of the real outage:

2015-01-29 10:59:52.452367 mon.0 128.142.35.220:6789/0 417 : [INF] pgmap v35845354: 24608 pgs: 10 active+clean+scrubbing+deep, 24596
active+clean, 2 active+clean+scrubbing; 125 TB
 data, 377 TB used, 2021 TB / 2399 TB avail; 5137 kB/s rd, 31239 kB/s wr, 641 op/s
2015-01-29 10:59:52.618591 mon.0 128.142.35.220:6789/0 431 : [INF]
osd.1129 128.142.23.100:6912/90010 failed (3 reports from 3 peers after 20.000159 >= grace 20.000000)
2015-01-29 10:59:52.939018 mon.0 128.142.35.220:6789/0 479 : [INF]
osd.1080 128.142.23.111:6869/124575 failed (3 reports from 3 peers after 20.000181 >= grace 20.000000)
2015-01-29 10:59:53.147616 mon.0 128.142.35.220:6789/0 510 : [INF]
osd.1113 128.142.23.107:6810/25247 failed (3 reports from 3 peers after 20.525957 >= grace 20.000000)
2015-01-29 10:59:53.342428 mon.0 128.142.35.220:6789/0 538 : [INF]
osd.1136 128.142.23.100:6864/86061 failed (3 reports from 3 peers after 20.403032 >= grace 20.000000)
2015-01-29 10:59:53.342557 mon.0 128.142.35.220:6789/0 540 : [INF]
osd.1144 128.142.23.100:6883/91968 failed (3 reports from 3 peers after 20.403104 >= grace 20.000000)

Then many "wrongly marked me down" messages around a minute later when the network came back.

But then when I restarted the (peon) monitor:

2015-01-29 11:29:18.250750 mon.0 128.142.35.220:6789/0 10570 : [INF] pgmap v35847068: 24608 pgs: 1 active+clean+scrubbing+deep, 24602
active+clean, 5 active+clean+scrubbing; 125 T
B data, 377 TB used, 2021 TB / 2399 TB avail; 193 MB/s rd, 238 MB/s wr, 7410 op/s
2015-01-29 11:29:28.844678 mon.3 128.142.39.77:6789/0 1 : [INF] mon.2 calling new monitor election
2015-01-29 11:29:33.846946 mon.2 128.142.36.229:6789/0 9 : [INF] mon.4 calling new monitor election
2015-01-29 11:29:33.847022 mon.4 128.142.39.144:6789/0 7 : [INF] mon.3 calling new monitor election
2015-01-29 11:29:33.847085 mon.1 128.142.36.227:6789/0 24 : [INF]
mon.1 calling new monitor election
2015-01-29 11:29:33.853498 mon.3 128.142.39.77:6789/0 2 : [INF] mon.2 calling new monitor election
2015-01-29 11:29:33.895660 mon.0 128.142.35.220:6789/0 10860 : [INF]
mon.0 calling new monitor election
2015-01-29 11:29:33.901335 mon.0 128.142.35.220:6789/0 10861 : [INF]
mon.0@0 won leader election with quorum 0,1,2,3,4
2015-01-29 11:29:34.004028 mon.0 128.142.35.220:6789/0 10862 : [INF] monmap e5: 5 mons at {0=128.142.35.220:6789/0,1=128.142.36.227:6789/0,2=128.142.39.77:6789/0,3=128.142.39.144:6789/0,4=128.142.36.229:6789/0}
2015-01-29 11:29:34.005808 mon.0 128.142.35.220:6789/0 10863 : [INF] pgmap v35847069: 24608 pgs: 1 active+clean+scrubbing+deep, 24602
active+clean, 5 active+clean+scrubbing; 125 TB data, 377 TB used, 2021
TB / 2399 TB avail; 54507 kB/s rd, 85412 kB/s wr, 1967 op/s
2015-01-29 11:29:34.006111 mon.0 128.142.35.220:6789/0 10864 : [INF] mdsmap e157: 1/1/1 up {0=0=up:active}
2015-01-29 11:29:34.007165 mon.0 128.142.35.220:6789/0 10865 : [INF] osdmap e132055: 880 osds: 880 up, 880 in
2015-01-29 11:29:34.037367 mon.0 128.142.35.220:6789/0 11055 : [INF]
osd.1202 128.142.23.104:6801/98353 failed (4 reports from 3 peers after 29.673699 >= grace 28.948726)
2015-01-29 11:29:34.050478 mon.0 128.142.35.220:6789/0 11139 : [INF]
osd.1164 128.142.23.102:6850/22486 failed (3 reports from 2 peers after 30.685537 >= grace 28.946983)


and then just after:

2015-01-29 11:29:35.210184 osd.1202 128.142.23.104:6801/98353 59 :
[WRN] map e132056 wrongly marked me down
2015-01-29 11:29:35.441922 osd.1164 128.142.23.102:6850/22486 25 :
[WRN] map e132056 wrongly marked me down

plus many other OSDs like that.

Here is the full ceph.log from that day:
https://www.dropbox.com/s/kzb2e3h3uirhejm/ceph.log-20150130.gz?dl=0
And here is ceph-mon.1.log:
https://www.dropbox.com/s/w1mx8rcflffp5e8/ceph-mon.0.log-20150130.gz?dl=0
It has a bit more info (prepare_failure, 'ignoring stats from non-active osd')  but alas I had only debug mon = 1

Cheers, Dan

>> I only have the lowest level logs from this incident -- but I think
>> it's easily reproducable.
>
> Logs with debug ms = 1 and debug mon = 20 would be best if someone is
> able to reproduce this.  I think our QA wouldn't capture this case
> because when we are either thrashing monitors and not OSDs (so OSD is
> never marked
> down) or we are thrashing OSDs and not mons (and ignoring
> wrongly-marked-down events).
>
> sage
>
>>
>> Cheers, Dan
>>
>>
>> On Wed, Feb 4, 2015 at 12:06 PM, Christian Eichelmann
>> <christian.eichelmann@xxxxxxxx> wrote:
>> > Hi Greg,
>> >
>> > the behaviour is indeed strange. Today I was trying to reproduce
>> > the problem, but no matter which monitor I've restarted, no matter
>> > how many times, the bahviour was like expected: A new monitor
>> > election was called and everything contiuned normally.
>> >
>> > Then I continued my failover tests and simulated the failure of two
>> > racks with iptables (for us: 2 MON and & 6 OSD Server with in sum
>> > 360 OSDs)
>> >
>> > Afterwards I tried again to restart one monitor and again about 240
>> > OSDs got marked as down.
>> >
>> > There was no load on our monitor servers in that period. On one of
>> > the OSDs which got marked down I found lot's of those messages:
>> >
>> > 2015-02-04 11:55:22.788245 7fc48fa48700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.8:6806/3303244 pipe(0x7a1b600 sd=198 :59766 s=2 pgs=1353
>> > cs=1 l=0 c=0x4e562c0).fault with nothing to send, going to standby
>> > 2015-02-04 11:55:22.788371 7fc48be0c700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.8:6842/12012876 pipe(0x895e840 sd=188 :49283 s=2
>> >>> pgs=36873
>> > cs=1 l=0 c=0x13226f20).fault with nothing to send, going to standby
>> > 2015-02-04 11:55:22.788458 7fc494e9c700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.13:6870/13021609 pipe(0x13ace2c0 sd=117 :64130 s=2
>> >>> pgs=38011
>> > cs=1 l=0 c=0x52b4840).fault with nothing to send, going to standby
>> > 2015-02-04 11:55:22.797107 7fc46459d700  0 -- 10.76.70.4:0/94790 >>
>> > 10.76.70.11:6980/37144571 pipe(0xba0c580 sd=30 :0 s=1 pgs=0 cs=0
>> > l=1 c=0x4e51600).fault
>> > 2015-02-04 11:55:22.799350 7fc482d7d700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.10:6887/30410592 pipe(0x6a0cb00 sd=271 :53090 s=2
>> >>> pgs=15372
>> > cs=1 l=0 c=0xf3a6f20).fault with nothing to send, going to standby
>> > 2015-02-04 11:55:22.800018 7fc46429a700  0 -- 10.76.70.4:0/94790 >>
>> > 10.76.28.41:7076/37144571 pipe(0xba0c840 sd=59 :0 s=1 pgs=0 cs=0
>> > l=1 c=0xf339760).fault
>> > 2015-02-04 11:55:22.803086 7fc482272700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.5:6867/17011547 pipe(0x12f998c0 sd=294 :6997 s=2
>> >>> pgs=46095
>> > cs=1 l=0 c=0x8382000).fault with nothing to send, going to standby
>> > 2015-02-04 11:55:22.804736 7fc4892e1700  0 --
>> > 10.76.70.4:6997/17094790
>> >>> 10.76.70.13:6852/9142109 pipe(0x12fa5b80 sd=163 :57056 s=2
>> >>> pgs=45269
>> > cs=1 l=0 c=0x189d1600).fault with nothing to send, going to standby
>> >
>> > The IPs mentioned there are all OSD Server.
>> >
>> > For me it feels like the monitors still have some "memory" about
>> > the failed OSDs and something is happening when one of the goes
>> > down. If I can provide you any more information to clarify the
>> > issue, just tell me what you need.
>> >
>> > Regards,
>> > Christian
>> >
>> > Am 03.02.2015 18:10, schrieb Gregory Farnum:
>> >> On Tue, Feb 3, 2015 at 3:38 AM, Christian Eichelmann
>> >> <christian.eichelmann@xxxxxxxx> wrote:
>> >>> Hi all,
>> >>>
>> >>> during some failover tests and some configuration tests, we
>> >>> currently discover a strange phenomenon:
>> >>>
>> >>> Restarting one of our monitors (5 in sum) triggers about 300 of
>> >>> the following events:
>> >>>
>> >>> osd.669 10.76.28.58:6935/149172 failed (20 reports from 20 peers
>> >>> after
>> >>> 22.005858 >= grace 20.000000)
>> >>>
>> >>> The osds come back up shortly after the have been marked down.
>> >>> What I don't understand is: How can a restart of one monitor
>> >>> prevent the osds from talking to each other and marking them down?
>> >>>
>> >>> FYI:
>> >>> We are currently using the following settings:
>> >>> mon osd adjust hearbeat grace = false mon osd min down reporters
>> >>> = 20 mon osd adjust down out interval = false
>> >>
>> >> That's really strange. I think maybe you're seeing some kind of
>> >> secondary effect; what kind of CPU usage are you seeing on the
>> >> monitors during this time? Have you checked the log on any OSDs
>> >> which have been marked down?
>> >>
>> >> I have a suspicion that maybe the OSDs are detecting their failed
>> >> monitor connection and not being able to reconnect to another
>> >> monitor quickly enough, but I'm not certain what the overlaps are there.
>> >> -Greg
>> >>
>> >
>> >
>> > --
>> > Christian Eichelmann
>> > Systemadministrator
>> >
>> > 1&1 Internet AG - IT Operations Mail & Media Advertising &
>> > Targeting Brauerstraße 48 · DE-76135 Karlsruhe
>> > Telefon: +49 721 91374-8026
>> > christian.eichelmann@xxxxxxxx
>> >
>> > Amtsgericht Montabaur / HRB 6484
>> > Vorstände: Henning Ahlert, Ralph Dommermuth, Matthias Ehrlich,
>> > Robert Hoffmann, Markus Huhn, Hans-Henning Kettler, Dr. Oliver
>> > Mauss, Jan Oetjen
>> > Aufsichtsratsvorsitzender: Michael Scheeren
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

________________________________

PLEASE NOTE: The information contained in this electronic mail message is intended only for the use of the designated recipient(s) named above. If the reader of this message is not the intended recipient, you are hereby notified that you have received this message in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify the sender by telephone or e-mail (as shown above) immediately and destroy any and all copies of this message in your possession (whether hard copies or electronically stored copies).

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux