Re: why are there "degraded" PGs when adding OSDs?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sam,
I think I may have the problem:  I noticed that the new host was created with 
straw2 instead of straw.  Would this account for 50% of PGs being degraded?

(I'm removing the OSDs on that host and will recreate with 'firefly' tunables.)

Thanks!
Chad.

On Monday, July 27, 2015 15:09:21 Chad William Seys wrote:
> Hi Sam,
> 	I'll need help getting the osdmap and pg dump prior to addition.
> 	I can remove the OSDs and add again if the osdmap (etc.) is not logged
> somewhere.
> 
> Chad.
> 
> > Hmm, that's odd.  Can you attach the osdmap and ceph pg dump prior to the
> > addition (with all pgs active+clean), then the osdmap and ceph pg dump
> > afterwards? -Sam
> > 
> > ----- Original Message -----
> > From: "Chad William Seys" <cwseys@xxxxxxxxxxxxxxxx>
> > To: "Samuel Just" <sjust@xxxxxxxxxx>, "ceph-users" <ceph-users@xxxxxxxx>
> > Sent: Monday, July 27, 2015 12:57:23 PM
> > Subject: Re:  why are there "degraded" PGs when adding OSDs?
> > 
> > Hi Sam,
> > 
> > > The pg might also be degraded right after a map change which changes the
> > > up/acting sets since the few objects updated right before the map change
> > > might be new on some replicas and old on the other replicas.  While in
> > > that
> > > state, those specific objects are degraded, and the pg would report
> > > degraded until they are recovered (which would happen asap, prior to
> > > backfilling the new replica). -Sam
> > 
> > That sounds like only a few PGs should be degraded.  I instead have about
> > 45% (and higher earlier).
> > 
> > # ceph -s
> > 
> >     cluster 7797e50e-f4b3-42f6-8454-2e2b19fa41d6
> >     
> >      health HEALTH_WARN
> >      
> >             2081 pgs backfill
> >             6745 pgs degraded
> >             17 pgs recovering
> >             6728 pgs recovery_wait
> >             6745 pgs stuck degraded
> >             8826 pgs stuck unclean
> >             recovery 2530124/5557452 objects degraded (45.527%)
> >             recovery 33594/5557452 objects misplaced (0.604%)
> >      
> >      monmap e5: 3 mons at
> > 
> > {mon01=128.104.164.197:6789/0,mon02=128.104.164.198:6789/0,mon03=10.128.19
> > 8. 51:6789/0} election epoch 16458, quorum 0,1,2 mon03,mon01,mon02
> > 
> >      mdsmap e3032: 1/1/1 up {0=mds01.hep.wisc.edu=up:active}
> >      osdmap e149761: 27 osds: 27 up, 27 in; 2083 remapped pgs
> >      
> >       pgmap v13464928: 18432 pgs, 9 pools, 5401 GB data, 1364 kobjects
> >       
> >             11122 GB used, 11786 GB / 22908 GB avail
> >             2530124/5557452 objects degraded (45.527%)
> >             33594/5557452 objects misplaced (0.604%)
> >             
> >                 9606 active+clean
> >                 6726 active+recovery_wait+degraded
> >                 2081 active+remapped+wait_backfill
> >                 
> >                   17 active+recovering+degraded
> >                   
> >                    2 active+recovery_wait+degraded+remapped
> > 
> > recovery io 24861 kB/s, 6 objects/s
> > 
> > Chad.
> > 
> > > ----- Original Message -----
> > > From: "Chad William Seys" <cwseys@xxxxxxxxxxxxxxxx>
> > > To: "ceph-users" <ceph-users@xxxxxxxx>
> > > Sent: Monday, July 27, 2015 12:27:26 PM
> > > Subject:  why are there "degraded" PGs when adding OSDs?
> > > 
> > > Hi All,
> > > 
> > > I recently added some OSDs to the Ceph cluster (0.94.2). I noticed that
> > > 'ceph -s' reported both misplaced AND degraded PGs.
> > > 
> > > Why should any PGs become degraded?  Seems as though Ceph should only be
> > > reporting misplaced PGs?
> > > 
> > > From the Giant release notes:
> > > Degraded vs misplaced: the Ceph health reports from ‘ceph -s’ and
> > > related
> > > commands now make a distinction between data that is degraded (there are
> > > fewer than the desired number of copies) and data that is misplaced
> > > (stored
> > > in the wrong location in the cluster). The distinction is important
> > > because
> > > the latter does not compromise data safety.
> > > 
> > > Does Ceph delete some replicas of the PGs (leading to degradation)
> > > before
> > > re- replicating on the new OSD?
> > > 
> > > This does not seem to be the safest algorithm.
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@xxxxxxxxxxxxxx
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux