Re: Kicking 'Remapped' PGs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This is pretty weird to me. Normally those PGs should be reported as
active, or stale, or something else in addition to remapped. Sam
suggests that they're probably stuck activating for some reason (which
is a state in new enough code, but not all versions), but I can't tell
or imagine why from these settings. You might have hit a bug I'm not
familiar with that will be jostled by just restarting the OSDs in
question... :/
-Greg


On Tue, May 5, 2015 at 7:46 AM, Paul Evans <paul@xxxxxxxxxxxx> wrote:
>  Gregory Farnum <greg@xxxxxxxxxxx> wrote:
>
> Oh. That's strange; they are all mapped to two OSDs but are placed on
> two different ones. I'm...not sure why that would happen. Are these
> PGs active? What's the full output of "ceph -s"?
>
>
> Those 4 PG’s went inactive at some point, and we had the luxury of  time to
> understand how we arrived at this state before we truly have to fix it (but
> that time is soon).
> So...We kicked a couple of OSD’s out yesterday to let the cluster re-shuffle
> things (osd.19 and osd.34…both of which were non-primary copies of the
> ‘acting’ PG map) and now the cluster status is even more interesting, IMHO:
>
> ceph@nc48-n1:/ceph-deploy/nautilus$ ceph -s
>     cluster 68bc69c1-1382-4c30-9bf8-480e32cc5b92
>      health HEALTH_WARN 2 pgs stuck inactive; 2 pgs stuck unclean;
> nodeep-scrub flag(s) set; crush map has legacy tunables
>      monmap e1: 3 mons at
> {nc48-n1=10.253.50.211:6789/0,nc48-n2=10.253.50.212:6789/0,nc48-n3=10.253.50.213:6789/0},
> election epoch 564, quorum 0,1,2 nc48-n1,nc48-n2,nc48-n3
>      osdmap e80862: 94 osds: 94 up, 92 in
>             flags nodeep-scrub
>       pgmap v1954234: 6144 pgs, 2 pools, 35251 GB data, 4419 kobjects
>             91727 GB used, 245 TB / 334 TB avail
>                 6140 active+clean
>                    2 remapped
>                    2 active+clean+scrubbing
> ceph@nc48-n1:/ceph-deploy/nautilus$ ceph pg dump_stuck
> ok
> pg_stat    objects    mip    degr    unf    bytes    log    disklog    state
> state_stamp    v    reported    up    up_primary    acting    acting_primary
> last_scrub    scrub_stamp    last_deep_scrub    deep_scrub_stamp
> 11.e2f    280    0    0    0    2339844181    3001    3001    remapped
> 2015-04-23 13:18:59.299589    68310'51082    80862:121916    [77,4]    77
> [77,34]    77    68310'51082    2015-04-23 11:40:11.565487    0'0
> 2014-10-20 13:41:46.122624
> 11.323    282    0    0    0    2357186647    3001    3001    remapped
> 2015-04-23 13:18:58.970396    70105'48961    80862:126346    [0,37]    0
> [0,19]    0    70105'48961    2015-04-23 11:47:02.980145    8145'44375
> 2015-03-30 16:09:36.975875
>
>
> --
> Paul
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux