stale+incomplete pgs on new cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Aug 19, 2014 at 3:36 PM, Gregory Farnum <greg at inktank.com> wrote:

> [Re-adding the list]
>
> On Tue, Aug 19, 2014 at 2:24 PM, Randy Smith <rbsmith at adams.edu> wrote:
> > Gregory,
> >
> > # ceph osd tree
> > # id    weight  type name       up/down reweight
> > -1      0.2     root default
> > -2      0.2             host cs00
> > 0       0.09999                 osd.0   up      1
> > 1       0.09999                 osd.1   up      1
> >
> > The min_size on the ruleset is 1.  Here's the relevant portion of my
> > crushmap.
>
> Sorry, by "min_size" I mean the member associated with the pool, not
> with the rule. You should be able to see it by looking at the pools
> section of "ceph osd dump", among others. I believe that by default it
> should be 2, and the full size is 3...which if you have 2 OSDs and
> aren't segregating across hosts should be good,


Ah, yes, of course. All of the pools were set to size 3 and min_size 2.
I've reset those to 2 and 1 since I only have two osds in this test.


> but I'm guessing your
> pools are using the default rules that will want to segregate across
> hosts.
>

I changed the crushmap to use osd rather than host in the 'rule_replicated'
rule which I though would take care of that.

I guess I missed something since I'm still seeing the incomplete pgs.


> -Greg
> Software Engineer #42 @ http://inktank.com | http://ceph.com
>
>
> >
> > # buckets
> > host cs00 {
> >         id -2           # do not change unnecessarily
> >         # weight 0.200
> >         alg straw
> >         hash 0  # rjenkins1
> >         item osd.0 weight 0.100
> >         item osd.1 weight 0.100
> > }
> > root default {
> >         id -1           # do not change unnecessarily
> >         # weight 0.200
> >         alg straw
> >         hash 0  # rjenkins1
> >         item cs00 weight 0.200
> > }
> >
> > # rules
> > rule replicated_ruleset {
> >         ruleset 0
> >         type replicated
> >         min_size 1
> >         max_size 10
> >         step take default
> >         step chooseleaf firstn 0 type osd
> >         step emit
> > }
> >
> > This is a single host "cluster" as it's just for testing.
> >
> >
> >
> > On Tue, Aug 19, 2014 at 2:44 PM, Gregory Farnum <greg at inktank.com>
> wrote:
> >>
> >> On Tue, Aug 19, 2014 at 1:37 PM, Randy Smith <rbsmith at adams.edu> wrote:
> >> > Greetings,
> >> >
> >> > I'm creating a new ceph cluster for testing and it's reporting "192
> >> > stale+incomplete" pgs.
> >> >
> >> > `ceph health detail` lists all of the pgs that are stuck. Here's a
> >> > representative line.
> >> >
> >> >   pg 2.2c is stuck stale for 3076.510998, current state
> >> > stale+incomplete,
> >> > last acting [0]
> >> >
> >> > But when I run `ceph pg 2.2c query`, I get
> >> >
> >> >   Error ENOENT: i don't have pgid 2.2c
> >> >
> >> > What can I do to clean up the cluster? I have no problems wiping the
> >> > system
> >> > because it's a new test cluster.
> >>
> >> Incomplete generally means the PGs aren't mapped to enough OSDs.
> >> What's the output of "ceph osd tree" and what is the min_size set to
> >> for your pools?
> >> -Greg
> >> Software Engineer #42 @ http://inktank.com | http://ceph.com
> >
> >
> >
> >
> > --
> > Randall Smith
> > Computing Services
> > Adams State University
> > http://www.adams.edu/
> > 719-587-7741
>



-- 
Randall Smith
Computing Services
Adams State University
http://www.adams.edu/
719-587-7741
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140819/078192a6/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux