Re: Another cluster completely hang

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

it does not.

But in your case, you have 10 OSD, and 7 of them have incomplete PG's.

So since your proxmox vps's are not on single PG's but spread across
many PG's you have a good chance that at least some data of any vps is
on one of the defect PG's.

-- 
Mit freundlichen Gruessen / Best regards

Oliver Dzombic
IP-Interactive

mailto:info@xxxxxxxxxxxxxxxxx

Anschrift:

IP Interactive UG ( haftungsbeschraenkt )
Zum Sonnenberg 1-3
63571 Gelnhausen

HRB 93402 beim Amtsgericht Hanau
Geschäftsführung: Oliver Dzombic

Steuer Nr.: 35 236 3622 1
UST ID: DE274086107


Am 29.06.2016 um 13:09 schrieb Mario Giammarco:
> Just one question: why when ceph has some incomplete pgs it refuses to
> do I/o on good pgs?
> 
> Il giorno mer 29 giu 2016 alle ore 12:55 Oliver Dzombic
> <info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>> ha scritto:
> 
>     Hi,
> 
>     again:
> 
>     You >must< check all your logs ( as fucky as it is for sure ).
> 
>     Means on the ceph nodes in /var/log/ceph/*
> 
>     And go back to the time where things went down the hill.
> 
>     There must be something else going on, beyond normal osd crash.
> 
>     And your manual pg repair/pg remove/pg set complete is, most probably,
>     just getting your situation worst.
> 
>     So really, if you want to have a chance to find out whats going on, you
>     must check all the logs. Especially the OSD logs, especially the OSD log
>     of the OSD you removed, and then the OSD logs of those pg, which are
>     incomplete/stuck/what_ever_not_good.
> 
>     --
>     Mit freundlichen Gruessen / Best regards
> 
>     Oliver Dzombic
>     IP-Interactive
> 
>     mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
> 
>     Anschrift:
> 
>     IP Interactive UG ( haftungsbeschraenkt )
>     Zum Sonnenberg 1-3
>     63571 Gelnhausen
> 
>     HRB 93402 beim Amtsgericht Hanau
>     Geschäftsführung: Oliver Dzombic
> 
>     Steuer Nr.: 35 236 3622 1
>     UST ID: DE274086107
> 
> 
>     Am 29.06.2016 um 12:33 schrieb Mario Giammarco:
>     > Thanks,
>     > I can put in osds but the do not stay in, and I am pretty sure
>     that are
>     > not broken.
>     >
>     > Il giorno mer 29 giu 2016 alle ore 12:07 Oliver Dzombic
>     > <info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
>     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>> ha
>     scritto:
>     >
>     >     hi,
>     >
>     >     ceph osd set noscrub
>     >     ceph osd set nodeep-scrub
>     >
>     >     ceph osd in <id>
>     >
>     >
>     >     --
>     >     Mit freundlichen Gruessen / Best regards
>     >
>     >     Oliver Dzombic
>     >     IP-Interactive
>     >
>     >     mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
>     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >
>     >     Anschrift:
>     >
>     >     IP Interactive UG ( haftungsbeschraenkt )
>     >     Zum Sonnenberg 1-3
>     >     63571 Gelnhausen
>     >
>     >     HRB 93402 beim Amtsgericht Hanau
>     >     Geschäftsführung: Oliver Dzombic
>     >
>     >     Steuer Nr.: 35 236 3622 1
>     >     UST ID: DE274086107
>     >
>     >
>     >     Am 29.06.2016 um 12:00 schrieb Mario Giammarco:
>     >     > Now the problem is that ceph has put out two disks because
>     scrub  has
>     >     > failed (I think it is not a disk fault but due to mark-complete)
>     >     > How can I:
>     >     > - disable scrub
>     >     > - put in again the two disks
>     >     >
>     >     > I will wait anyway the end of recovery to be sure it really
>     works
>     >     again
>     >     >
>     >     > Il giorno mer 29 giu 2016 alle ore 11:16 Mario Giammarco
>     >     > <mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>
>     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>
>     >     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>
>     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>>> ha
>     scritto:
>     >     >
>     >     >     Infact I am worried because:
>     >     >
>     >     >     1) ceph is under proxmox, and proxmox may decide to reboot a
>     >     server
>     >     >     if it is not responding
>     >     >     2) probably a server was rebooted while ceph was
>     reconstructing
>     >     >     3) even using max=3 do not help
>     >     >
>     >     >     Anyway this is the "unofficial" procedure that I am
>     using, much
>     >     >     simpler than blog post:
>     >     >
>     >     >     1) find host where is pg
>     >     >     2) stop ceph in that host
>     >     >     3) ceph-objectstore-tool --pgid 1.98 --op mark-complete
>     >     --data-path
>     >     >     /var/lib/ceph/osd/ceph-9 --journal-path
>     >     >     /var/lib/ceph/osd/ceph-9/journal
>     >     >     4) start ceph
>     >     >     5) look finally it reconstructing
>     >     >
>     >     >     Il giorno mer 29 giu 2016 alle ore 11:11 Oliver Dzombic
>     >     >     <info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
>     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
>     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>>> ha
>     >     scritto:
>     >     >
>     >     >         Hi,
>     >     >
>     >     >         removing ONE disk while your replication is 2, is no
>     problem.
>     >     >
>     >     >         You dont need to wait a single second to replace of
>     remove
>     >     it. Its
>     >     >         anyway not used and out/down. So from ceph's point
>     of view its
>     >     >         not existent.
>     >     >
>     >     >         ----------------
>     >     >
>     >     >         But as christian told you already, what we see now
>     fits to a
>     >     >         szenario
>     >     >         where you lost the osd and eighter you did something, or
>     >     >         something else
>     >     >         happens, but the data were not recovered again.
>     >     >
>     >     >         Eighter because another OSD was broken, or because
>     you did
>     >     >         something.
>     >     >
>     >     >         Maybe, because of the "too many PGs per OSD (307 >
>     max 300)"
>     >     >         ceph never
>     >     >         recovered.
>     >     >
>     >     >         What i can see from http://pastebin.com/VZD7j2vN is that
>     >     >
>     >     >         OSD 5,13,9,0,6,2,3 and maybe others, are the OSD's
>     holding the
>     >     >         incomplete data.
>     >     >
>     >     >         This are 7 OSD's from 10. So something happend to that
>     >     OSD's or
>     >     >         the data
>     >     >         in them. And that had nothing to do with a single disk
>     >     failing.
>     >     >
>     >     >         Something else must have been happend.
>     >     >
>     >     >         And as christian already wrote: you will have to go
>     >     through your
>     >     >         logs
>     >     >         back until the point were things going down.
>     >     >
>     >     >         Because a fail of a single OSD, no matter what your
>     >     replication
>     >     >         size is,
>     >     >         can ( normally ) not harm the consistency of 7 other
>     OSD's,
>     >     >         means 70% of
>     >     >         your total cluster.
>     >     >
>     >     >         --
>     >     >         Mit freundlichen Gruessen / Best regards
>     >     >
>     >     >         Oliver Dzombic
>     >     >         IP-Interactive
>     >     >
>     >     >         mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>> <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>>
>     >     >
>     >     >         Anschrift:
>     >     >
>     >     >         IP Interactive UG ( haftungsbeschraenkt )
>     >     >         Zum Sonnenberg 1-3
>     >     >         63571 Gelnhausen
>     >     >
>     >     >         HRB 93402 beim Amtsgericht Hanau
>     >     >         Geschäftsführung: Oliver Dzombic
>     >     >
>     >     >         Steuer Nr.: 35 236 3622 1
>     >     >         UST ID: DE274086107
>     >     >
>     >     >
>     >     >         Am 29.06.2016 um 10:56 schrieb Mario Giammarco:
>     >     >         > Yes I have removed it from crush because it was
>     broken.
>     >     I have
>     >     >         waited 24
>     >     >         > hours to see if cephs would like to heals itself.
>     Then I
>     >     >         removed the
>     >     >         > disk completely (it was broken...) and I waited 24
>     hours
>     >     >         again. Then I
>     >     >         > start getting worried.
>     >     >         > Are you saying to me that I should not remove a broken
>     >     disk from
>     >     >         > cluster? 24 hours were not enough?
>     >     >         >
>     >     >         > Il giorno mer 29 giu 2016 alle ore 10:53 Zoltan
>     Arnold Nagy
>     >     >         > <zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>
>     >     <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>>
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx <mailto:zoltan@xxxxxxxxxxxxxxxxxx>
>     >     <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>>>
>     >     >         <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>
>     >     <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>
>     >     <mailto:zoltan@xxxxxxxxxxxxxxxxxx
>     <mailto:zoltan@xxxxxxxxxxxxxxxxxx>>>>> ha scritto:
>     >     >         >
>     >     >         >     Just loosing one disk doesn’t automagically delete
>     >     it from
>     >     >         CRUSH,
>     >     >         >     but in the output you had 10 disks listed, so
>     there
>     >     must be
>     >     >         >     something else going - did you delete the disk
>     from the
>     >     >         crush map as
>     >     >         >     well?
>     >     >         >
>     >     >         >     Ceph waits by default 300 secs AFAIK to mark
>     an OSD out
>     >     >         after it
>     >     >         >     will start to recover.
>     >     >         >
>     >     >         >
>     >     >         >>     On 29 Jun 2016, at 10:42, Mario Giammarco
>     >     >         <mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>
>     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>
>     >     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>
>     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>>
>     >     >         >>     <mailto:mgiammarco@xxxxxxxxx
>     <mailto:mgiammarco@xxxxxxxxx>
>     >     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>
>     >     >         <mailto:mgiammarco@xxxxxxxxx
>     <mailto:mgiammarco@xxxxxxxxx>
>     >     <mailto:mgiammarco@xxxxxxxxx <mailto:mgiammarco@xxxxxxxxx>>>>>
>     wrote:
>     >     >         >>
>     >     >         >>     I thank you for your reply so I can add my
>     experience:
>     >     >         >>
>     >     >         >>     1) the other time this thing happened to me I
>     had a
>     >     >         cluster with
>     >     >         >>     min_size=2 and size=3 and the problem was the
>     same.
>     >     That
>     >     >         time I
>     >     >         >>     put min_size=1 to recover the pool but it did
>     not help.
>     >     >         So I do
>     >     >         >>     not understand where is the advantage to put
>     three
>     >     copies
>     >     >         when
>     >     >         >>     ceph can decide to discard all three.
>     >     >         >>     2) I started with 11 hdds. The hard disk failed.
>     >     Ceph waited
>     >     >         >>     forever for hard disk coming back. But hard
>     disk is
>     >     really
>     >     >         >>     completelly broken so I have followed the
>     procedure
>     >     to really
>     >     >         >>     delete from cluster. Anyway ceph did not recover.
>     >     >         >>     3) I have 307 pgs more than 300 but it is due to
>     >     the fact
>     >     >         that I
>     >     >         >>     had 11 hdds now only 10. I will add more hdds
>     after I
>     >     >         repair the pool
>     >     >         >>     4) I have reduced the monitors to 3
>     >     >         >>
>     >     >         >>
>     >     >         >>
>     >     >         >>     Il giorno mer 29 giu 2016 alle ore 10:25
>     Christian
>     >     Balzer
>     >     >         >>     <chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>
>     >     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>>
>     >     >         <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>
>     >     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>>>> ha scritto:
>     >     >         >>
>     >     >         >>
>     >     >         >>         Hello,
>     >     >         >>
>     >     >         >>         On Wed, 29 Jun 2016 06:02:59 +0000 Mario
>     >     Giammarco wrote:
>     >     >         >>
>     >     >         >>         > pool 0 'rbd' replicated size 2 min_size 1
>     >     >         crush_ruleset 0
>     >     >         >>         object_hash
>     >     >         >>                                        ^
>     >     >         >>         And that's the root cause of all your woes.
>     >     >         >>         The default replication size is 3 for a
>     reason and
>     >     >         while I do
>     >     >         >>         run pools
>     >     >         >>         with replication of 2 they are either HDD
>     RAIDs or
>     >     >         extremely
>     >     >         >>         trustworthy
>     >     >         >>         and well monitored SSD.
>     >     >         >>
>     >     >         >>         That said, something more than a single
>     HDD failure
>     >     >         must have
>     >     >         >>         happened
>     >     >         >>         here, you should check the logs and backtrace
>     >     all the
>     >     >         step you
>     >     >         >>         did after
>     >     >         >>         that OSD failed.
>     >     >         >>
>     >     >         >>         You said there were 11 HDDs and your
>     first ceph -s
>     >     >         output showed:
>     >     >         >>         ---
>     >     >         >>              osdmap e10182: 10 osds: 10 up, 10 in
>     >     >         >>         ----
>     >     >         >>         And your crush map states the same.
>     >     >         >>
>     >     >         >>         So how and WHEN did you remove that OSD?
>     >     >         >>         My suspicion would be it was removed before
>     >     recovery
>     >     >         was complete.
>     >     >         >>
>     >     >         >>         Also, as I think was mentioned before, 7
>     mons are
>     >     >         overkill 3-5
>     >     >         >>         would be a
>     >     >         >>         saner number.
>     >     >         >>
>     >     >         >>         Christian
>     >     >         >>
>     >     >         >>         > rjenkins pg_num 512 pgp_num 512 last_change
>     >     9313 flags
>     >     >         >>         hashpspool
>     >     >         >>         > stripe_width 0
>     >     >         >>         >        removed_snaps [1~3]
>     >     >         >>         > pool 1 'rbd2' replicated size 2 min_size 1
>     >     >         crush_ruleset 0
>     >     >         >>         object_hash
>     >     >         >>         > rjenkins pg_num 512 pgp_num 512 last_change
>     >     9314 flags
>     >     >         >>         hashpspool
>     >     >         >>         > stripe_width 0
>     >     >         >>         >        removed_snaps [1~3]
>     >     >         >>         > pool 2 'rbd3' replicated size 2 min_size 1
>     >     >         crush_ruleset 0
>     >     >         >>         object_hash
>     >     >         >>         > rjenkins pg_num 512 pgp_num 512 last_change
>     >     10537 flags
>     >     >         >>         hashpspool
>     >     >         >>         > stripe_width 0
>     >     >         >>         >        removed_snaps [1~3]
>     >     >         >>         >
>     >     >         >>         >
>     >     >         >>         > ID WEIGHT  REWEIGHT SIZE   USE   AVAIL
>     %USE  VAR
>     >     >         >>         > 5 1.81000  1.00000  1857G  984G  872G
>     53.00 0.86
>     >     >         >>         > 6 1.81000  1.00000  1857G 1202G  655G
>     64.73 1.05
>     >     >         >>         > 2 1.81000  1.00000  1857G 1158G  698G
>     62.38 1.01
>     >     >         >>         > 3 1.35999  1.00000  1391G  906G  485G
>     65.12 1.06
>     >     >         >>         > 4 0.89999  1.00000   926G  702G  223G
>     75.88 1.23
>     >     >         >>         > 7 1.81000  1.00000  1857G 1063G  793G
>     57.27 0.93
>     >     >         >>         > 8 1.81000  1.00000  1857G 1011G  846G
>     54.44 0.88
>     >     >         >>         > 9 0.89999  1.00000   926G  573G  352G
>     61.91 1.01
>     >     >         >>         > 0 1.81000  1.00000  1857G 1227G  629G
>     66.10 1.07
>     >     >         >>         > 13 0.45000  1.00000   460G  307G  153G
>     66.74 1.08
>     >     >         >>         >              TOTAL 14846G 9136G 5710G 61.54
>     >     >         >>         > MIN/MAX VAR: 0.86/1.23  STDDEV: 6.47
>     >     >         >>         >
>     >     >         >>         >
>     >     >         >>         >
>     >     >         >>         > ceph version 0.94.7
>     >     >         (d56bdf93ced6b80b07397d57e3fa68fe68304432)
>     >     >         >>         >
>     >     >         >>         > http://pastebin.com/SvGfcSHb
>     >     >         >>         > http://pastebin.com/gYFatsNS
>     >     >         >>         > http://pastebin.com/VZD7j2vN
>     >     >         >>         >
>     >     >         >>         > I do not understand why I/O on ENTIRE
>     cluster is
>     >     >         blocked
>     >     >         >>         when only few
>     >     >         >>         > pgs are incomplete.
>     >     >         >>         >
>     >     >         >>         > Many thanks,
>     >     >         >>         > Mario
>     >     >         >>         >
>     >     >         >>         >
>     >     >         >>         > Il giorno mar 28 giu 2016 alle ore
>     19:34 Stefan
>     >     >         Priebe -
>     >     >         >>         Profihost AG <
>     >     >         >>         > s.priebe@xxxxxxxxxxxx
>     <mailto:s.priebe@xxxxxxxxxxxx>
>     >     <mailto:s.priebe@xxxxxxxxxxxx <mailto:s.priebe@xxxxxxxxxxxx>>
>     >     >         <mailto:s.priebe@xxxxxxxxxxxx
>     <mailto:s.priebe@xxxxxxxxxxxx>
>     >     <mailto:s.priebe@xxxxxxxxxxxx <mailto:s.priebe@xxxxxxxxxxxx>>>
>     <mailto:s.priebe@xxxxxxxxxxxx <mailto:s.priebe@xxxxxxxxxxxx>
>     >     <mailto:s.priebe@xxxxxxxxxxxx <mailto:s.priebe@xxxxxxxxxxxx>>
>     >     >         <mailto:s.priebe@xxxxxxxxxxxx
>     <mailto:s.priebe@xxxxxxxxxxxx>
>     >     <mailto:s.priebe@xxxxxxxxxxxx
>     <mailto:s.priebe@xxxxxxxxxxxx>>>>> ha
>     >     >         >>         scritto:
>     >     >         >>         >
>     >     >         >>         > > And ceph health detail
>     >     >         >>         > >
>     >     >         >>         > > Stefan
>     >     >         >>         > >
>     >     >         >>         > > Excuse my typo sent from my mobile phone.
>     >     >         >>         > >
>     >     >         >>         > > Am 28.06.2016 um 19:28 schrieb Oliver
>     Dzombic
>     >     >         >>         <info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>>> <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>>>>>:
>     >     >         >>         > >
>     >     >         >>         > > Hi Mario,
>     >     >         >>         > >
>     >     >         >>         > > please give some more details:
>     >     >         >>         > >
>     >     >         >>         > > Please the output of:
>     >     >         >>         > >
>     >     >         >>         > > ceph osd pool ls detail
>     >     >         >>         > > ceph osd df
>     >     >         >>         > > ceph --version
>     >     >         >>         > >
>     >     >         >>         > > ceph -w for 10 seconds ( use
>     >     http://pastebin.com/
>     >     >         please )
>     >     >         >>         > >
>     >     >         >>         > > ceph osd crush dump ( also pastebin pls )
>     >     >         >>         > >
>     >     >         >>         > > --
>     >     >         >>         > > Mit freundlichen Gruessen / Best regards
>     >     >         >>         > >
>     >     >         >>         > > Oliver Dzombic
>     >     >         >>         > > IP-Interactive
>     >     >         >>         > >
>     >     >         >>         > > mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>>
>     >     >         >>         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>>>> <info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>>
>     >     >         >>         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>
>     >     >         <mailto:info@xxxxxxxxxxxxxxxxx
>     <mailto:info@xxxxxxxxxxxxxxxxx>
>     >     <mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>>>>>
>     >     >         >>         > >
>     >     >         >>         > > Anschrift:
>     >     >         >>         > >
>     >     >         >>         > > IP Interactive UG ( haftungsbeschraenkt )
>     >     >         >>         > > Zum Sonnenberg 1-3
>     >     >         >>         > > 63571 Gelnhausen
>     >     >         >>         > >
>     >     >         >>         > > HRB 93402 beim Amtsgericht Hanau
>     >     >         >>         > > Geschäftsführung: Oliver Dzombic
>     >     >         >>         > >
>     >     >         >>         > > Steuer Nr.: 35 236 3622 1
>     >     >         >>         > > UST ID: DE274086107
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > > Am 28.06.2016 um 18:59 schrieb Mario
>     Giammarco:
>     >     >         >>         > >
>     >     >         >>         > > Hello,
>     >     >         >>         > >
>     >     >         >>         > > this is the second time that happens
>     to me, I
>     >     >         hope that
>     >     >         >>         someone can
>     >     >         >>         > >
>     >     >         >>         > > explain what I can do.
>     >     >         >>         > >
>     >     >         >>         > > Proxmox ceph cluster with 8 servers,
>     11 hdd.
>     >     >         Min_size=1,
>     >     >         >>         size=2.
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > > One hdd goes down due to bad sectors.
>     >     >         >>         > >
>     >     >         >>         > > Ceph recovers but it ends with:
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > > cluster
>     f2a8dd7d-949a-4a29-acab-11d4900249f4
>     >     >         >>         > >
>     >     >         >>         > >     health HEALTH_WARN
>     >     >         >>         > >
>     >     >         >>         > >            3 pgs down
>     >     >         >>         > >
>     >     >         >>         > >            19 pgs incomplete
>     >     >         >>         > >
>     >     >         >>         > >            19 pgs stuck inactive
>     >     >         >>         > >
>     >     >         >>         > >            19 pgs stuck unclean
>     >     >         >>         > >
>     >     >         >>         > >            7 requests are blocked >
>     32 sec
>     >     >         >>         > >
>     >     >         >>         > >     monmap e11: 7 mons at
>     >     >         >>         > >
>     >     >         >>         > >
>     >     {0=192.168.0.204:6789/0,1=192.168.0.201:6789/0
>     <http://192.168.0.204:6789/0,1=192.168.0.201:6789/0>
>     >     <http://192.168.0.204:6789/0,1=192.168.0.201:6789/0>
>     >     >         <http://192.168.0.204:6789/0,1=192.168.0.201:6789/0>
>     >     >         >>
>     >      <http://192.168.0.204:6789/0,1=192.168.0.201:6789/0>,
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>
>     >     >
>     >     2=192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202
>     <http://192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202>
>     >   
>      <http://192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202>
>     >     >
>     >     
>     <http://192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202>
>     >     >         >>
>     >     >
>     >   
>      <http://192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202>:
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         6789/0,5=192.168.0.206:6789/0,6=192.168.0.207:6789/0
>     <http://192.168.0.206:6789/0,6=192.168.0.207:6789/0>
>     >     <http://192.168.0.206:6789/0,6=192.168.0.207:6789/0>
>     >     >         <http://192.168.0.206:6789/0,6=192.168.0.207:6789/0>
>     >     >         >>
>     >      <http://192.168.0.206:6789/0,6=192.168.0.207:6789/0>}
>     >     >         >>         > >
>     >     >         >>         > >            election epoch 722, quorum
>     >     >         >>         > >
>     >     >         >>         > > 0,1,2,3,4,5,6 1,4,2,0,3,5,6
>     >     >         >>         > >
>     >     >         >>         > >     osdmap e10182: 10 osds: 10 up, 10 in
>     >     >         >>         > >
>     >     >         >>         > >      pgmap v3295880: 1024 pgs, 2
>     pools, 4563 GB
>     >     >         data, 1143
>     >     >         >>         kobjects
>     >     >         >>         > >
>     >     >         >>         > >            9136 GB used, 5710 GB /
>     14846 GB
>     >     avail
>     >     >         >>         > >
>     >     >         >>         > >                1005 active+clean
>     >     >         >>         > >
>     >     >         >>         > >                  16 incomplete
>     >     >         >>         > >
>     >     >         >>         > >                   3 down+incomplete
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > > Unfortunately "7 requests blocked"
>     means no
>     >     virtual
>     >     >         >>         machine can boot
>     >     >         >>         > >
>     >     >         >>         > > because ceph has stopped i/o.
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > > I can accept to lose some data, but
>     not ALL
>     >     data!
>     >     >         >>         > >
>     >     >         >>         > > Can you help me please?
>     >     >         >>         > >
>     >     >         >>         > > Thanks,
>     >     >         >>         > >
>     >     >         >>         > > Mario
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > >
>     _______________________________________________
>     >     >         >>         > >
>     >     >         >>         > > ceph-users mailing list
>     >     >         >>         > >
>     >     >         >>         > > ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>>
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >         >>         > >
>     >     >         >>         > >
>     >     >         >>         > >
>     _______________________________________________
>     >     >         >>         > > ceph-users mailing list
>     >     >         >>         > > ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>>
>     >     >         >>         > >
>     >     >         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >         >>         > >
>     >     >         >>         > >
>     _______________________________________________
>     >     >         >>         > > ceph-users mailing list
>     >     >         >>         > > ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>>
>     >     >         >>         > >
>     >     >         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >         >>         > >
>     >     >         >>
>     >     >         >>
>     >     >         >>         --
>     >     >         >>         Christian Balzer        Network/Systems
>     Engineer
>     >     >         >>         chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>
>     >     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>>
>     >     >         <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>
>     >     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>
>     <mailto:chibi@xxxxxxx <mailto:chibi@xxxxxxx>>>>           Global
>     >     >         OnLine
>     >     >         >>         Japan/Rakuten Communications
>     >     >         >>         http://www.gol.com/
>     >     >         >>
>     >     >         >>     _______________________________________________
>     >     >         >>     ceph-users mailing list
>     >     >         >>     ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     >         <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>>
>     >     >         >>   
>      http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >         >
>     >     >         >
>     >     >         >
>     >     >         > _______________________________________________
>     >     >         > ceph-users mailing list
>     >     >         > ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >         >
>     >     >         _______________________________________________
>     >     >         ceph-users mailing list
>     >     >         ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >     <mailto:ceph-users@xxxxxxxxxxxxxx
>     <mailto:ceph-users@xxxxxxxxxxxxxx>>>
>     >     >         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >     >
>     >     _______________________________________________
>     >     ceph-users mailing list
>     >     ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>
>     >     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >
>     _______________________________________________
>     ceph-users mailing list
>     ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux