Re: No Ceph Recovery : Is it a bug ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 19 May 2016, Gaurav Bafna wrote:
> On Thu, May 19, 2016 at 3:34 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> > On Thu, 19 May 2016, Gaurav Bafna wrote:
> >> Hi Sage & hzwulibin
> >>
> >> Many thanks for your reply :)
> >>
> >> Ceph osd tree : http://pastebin.com/3cC8brcF
> >>
> >> Crushmap : http://pastebin.com/K2BNSHys
> >
> > Can you paste the output from
> >
> >  ceph osd crush show-tunables
> 
> {
>     "choose_local_tries": 0,
>     "choose_local_fallback_tries": 0,
>     "choose_total_tries": 50,
>     "chooseleaf_descend_once": 1,
>     "chooseleaf_vary_r": 0,

This might be part of it.  If you don't have too much data in the cluster 
yet, 'ceph osd crush tunables hammer' may help.  Note that it will move a 
lot of data around.

sage

>     "straw_calc_version": 1,
>     "allowed_bucket_algs": 22,
>     "profile": "unknown",
>     "optimal_tunables": 0,
>     "legacy_tunables": 0,
>     "require_feature_tunables": 1,
>     "require_feature_tunables2": 1,
>     "require_feature_tunables3": 0,
>     "has_v2_rules": 0,
>     "has_v3_rules": 0,
>     "has_v4_buckets": 0
> }
> 
> 
> >
> >> Since I removed the 12 osds from the crush map, there are no more stuck PGs .
> >
> > Hmm...
> >
> >> Version : Hammer 0.94.5.    Do you think it can just be a reporting
> >> issue ? The osds that went down had around 300 PGs mapped to them out
> >> of which it was primary for 0 . We have two IDCs. All the primary are
> >> in one IDC only for now.
> >>
> >> Sage, I did not get yor first point. What is not matching ?  In our
> >> cluster there were 2758 running osds before the 12 osd crash . After
> >> that event , there were 2746 osds left.
> >>
> >> PG Stat for your reference :   v3733136: 75680 pgs: 75680
> >> active+clean; 409 GB data, 16745 GB used, 14299 TB / 14315 TB avail;
> >> 443 B/s wr, 0 op/s
> >>
> >> Also, I have seen this issue 3 times with our cluster. Even when 1 osd
> >> goes down in IDC1 or IDC2, some PG always remain undersized. I am sure
> >> this issue will come up once again . So can you tell me what things
> >> should I look when an OSD goes down next time ?
> >>
> >> When I manually stop an osd daemon in IDC1 or IDC2 by running,
> >> /etc/init.d/ceph  stop osd.x , the cluster recovers nicely. That is
> >> what making the issue so complex.
> >
> > Here you say restarting an OSD was enough to clear the problem, but
> > above you say you also removed the down OSDs from the CRUSH map.  Can you
> > be clear about when the PGs stopped being undersized?
> 
> I didn't restart the OSD as the disk was gone bad. When I removed from
> crush map, the PGs stopped being undersized.
> 
> 
> >
> > I can't tell from this information whether it is a reporting issue or
> > whether CRUSH to do a proper mapping on this cluster.
> >
> > If you can reproduce this, there are a few things to do:
> >
> > 1) Grab a copy of the OSD map:
> >
> >   get osd getmap -o osdmap
> >
> > 2) Get a list of undersized pgs:
> >
> >  ceph pg ls undersized > pgs.txt
> >
> > 3) query one of the undersized pgs:
> >
> >  tail pgs.txt
> >  ceph tell <one of the pgids> query > query.txt
> >
> > 4) Share the result with us
> >
> >  ceph-post-file osdmap pgs.txt query.txt
> >
> > (or attach it to an email).
> 
> Sure. I will do that the  next time it occurs.
> 
> Very grateful for your help,
> Gaurav
> >
> > Thanks!
> > sage
> >
> >
> >
> >>
> >> It is a big cluster and we have just started . We will need the
> >> community help to make it run smoothly and expand further :)
> >>
> >> Thanks
> >> Gaurav
> >>
> >> On Thu, May 19, 2016 at 1:17 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> >> > On Thu, 19 May 2016, Gaurav Bafna wrote:
> >> >> Hi Cephers ,
> >> >>
> >> >> In our production cluster at Reliance Jio, when as osd goes corrupt
> >> >> and crashes, Cluster remains unhealthy even after 4 hours.
> >> >>
> >> >>     cluster fac04d85-db48-4564-b821-deebda046261
> >> >>      health HEALTH_WARN
> >> >>             658 pgs degraded
> >> >>             658 pgs stuck degraded
> >> >>             688 pgs stuck unclean
> >> >>             658 pgs stuck undersized
> >> >>             658 pgs undersized
> >> >               ^^^ this...
> >> >
> >> >>             recovery 3064/1981308 objects degraded (0.155%)
> >> >>             recovery 124/1981308 objects misplaced (0.006%)
> >> >>      monmap e11: 11 mons at
> >> >> {dssmon2=10.140.208.224:6789/0,dssmon3=10.140.208.225:6789/0,dssmon31=10.135.38.141:6789/0,dssmon32=10.135.38.142:6789/0,dssmon33=10.135.38.143:6789/0,dssmon34=10.135.38.144:6789/0,dssmon35=10.135.38.145:6789/0,dssmon4=10.140.208.226:6789/0,dssmon5=10.140.208.227:6789/0,dssmon6=10.140.208.228:6789/0,dssmonleader1=10.140.208.223:6789/0}
> >> >>             election epoch 792, quorum 0,1,2,3,4,5,6,7,8,9,10
> >> >> dssmon31,dssmon32,dssmon33,dssmon34,dssmon35,dssmonleader1,dssmon2,dssmon3,dssmon4,dssmon5,dssmon6
> >> >>      osdmap e8778: 2774 osds: 2746 up, 2746 in; 30 remapped pgs
> >> >                                doesn't match this ^^
> >> >
> >> > which makes it look like a problem with OSDs reporting PG state to the
> >> > mon.  The fact that an OSD restarts supports that theory.
> >> >
> >> > What version is this?  A bunch of the osd -> mon pg reporting code was
> >> > recently rewritten (between infernalis and jewel), so the new code is
> >> > hopefully more robust.  (OTOH, it is also new, so we may have missed
> >> > something.)
> >> >
> >> > Nice big cluster!
> >> >
> >> > sage
> >>
> >>
> >>
> >> --
> >> Gaurav Bafna
> >> 9540631400
> >>
> >>
> 
> 
> 
> -- 
> Gaurav Bafna
> 9540631400
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux