Re: [ceph-users] PG down & incomplete

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Another thing... since your osd.10 is near full, your cluster may be
fairly close to capacity for the purposes of rebalancing.  Have a look
at:

http://ceph.com/docs/master/rados/configuration/mon-config-ref/#storage-capacity
http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/#no-free-drive-space

Maybe we can get some others to look at this.  It's not clear to me
why the other OSD crashes after you take osd.25 out. It could be
capacity, but that shouldn't make it crash. Have you tried adding more
OSDs to increase capacity?



On Fri, May 17, 2013 at 11:27 AM, John Wilkins <john.wilkins@xxxxxxxxxxx> wrote:
> It looks like you have the "noout" flag set:
>
> "noout flag(s) set; 1 mons down, quorum 0,1,2,3 a,b,c,e
>    monmap e7: 5 mons at
> {a=10.0.0.1:6789/0,b=10.0.0.2:6789/0,c=10.0.0.5:6789/0,d=10.0.0.6:6789/0,e=10.0.0.3:6789/0},
> election epoch 2584, quorum 0,1,2,3 a,b,c,e
>    osdmap e82502: 50 osds: 48 up, 48 in"
>
> http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/#stopping-w-out-rebalancing
>
> If you have down OSDs that don't get marked out, that would certainly
> cause problems. Have you tried restarting the failed OSDs?
>
> What do the logs look like for osd.15 and osd.25?
>
> On Fri, May 17, 2013 at 1:31 AM, Olivier Bonvalet <ceph.list@xxxxxxxxx> wrote:
>> Hi,
>>
>> thanks for your answer. In fact I have several different problems, which
>> I tried to solve separatly :
>>
>> 1) I loose 2 OSD, and some pools have only 2 replicas. So some data was
>> lost.
>> 2) One monitor refuse the Cuttlefish upgrade, so I only have 4 of 5
>> monitors running.
>> 3) I have 4 old inconsistent PG that I can't repair.
>>
>>
>> So the status :
>>
>>    health HEALTH_ERR 15 pgs incomplete; 4 pgs inconsistent; 15 pgs stuck
>> inactive; 15 pgs stuck unclean; 1 near full osd(s); 19 scrub errors;
>> noout flag(s) set; 1 mons down, quorum 0,1,2,3 a,b,c,e
>>    monmap e7: 5 mons at
>> {a=10.0.0.1:6789/0,b=10.0.0.2:6789/0,c=10.0.0.5:6789/0,d=10.0.0.6:6789/0,e=10.0.0.3:6789/0}, election epoch 2584, quorum 0,1,2,3 a,b,c,e
>>    osdmap e82502: 50 osds: 48 up, 48 in
>>     pgmap v12807617: 7824 pgs: 7803 active+clean, 1 active+clean
>> +scrubbing, 15 incomplete, 4 active+clean+inconsistent, 1 active+clean
>> +scrubbing+deep; 5676 GB data, 18948 GB used, 18315 GB / 37263 GB avail;
>> 137KB/s rd, 1852KB/s wr, 199op/s
>>    mdsmap e1: 0/0/1 up
>>
>>
>>
>> The tree :
>>
>> # id    weight  type name       up/down reweight
>> -8      14.26   root SSDroot
>> -27     8               datacenter SSDrbx2
>> -26     8                       room SSDs25
>> -25     8                               net SSD188-165-12
>> -24     8                                       rack SSD25B09
>> -23     8                                               host lyll
>> 46      2                                                       osd.46  up      1
>> 47      2                                                       osd.47  up      1
>> 48      2                                                       osd.48  up      1
>> 49      2                                                       osd.49  up      1
>> -10     4.26            datacenter SSDrbx3
>> -12     2                       room SSDs43
>> -13     2                               net SSD178-33-122
>> -16     2                                       rack SSD43S01
>> -17     2                                               host kaino
>> 42      1                                                       osd.42  up      1
>> 43      1                                                       osd.43  up      1
>> -22     2.26                    room SSDs45
>> -21     2.26                            net SSD5-135-138
>> -20     2.26                                    rack SSD45F01
>> -19     2.26                                            host taman
>> 44      1.13                                                    osd.44  up      1
>> 45      1.13                                                    osd.45  up      1
>> -9      2               datacenter SSDrbx4
>> -11     2                       room SSDs52
>> -14     2                               net SSD176-31-226
>> -15     2                                       rack SSD52B09
>> -18     2                                               host dragan
>> 40      1                                                       osd.40  up      1
>> 41      1                                                       osd.41  up      1
>> -1      33.43   root SASroot
>> -100    15.9            datacenter SASrbx1
>> -90     15.9                    room SASs15
>> -72     15.9                            net SAS188-165-15
>> -40     8                                       rack SAS15B01
>> -3      8                                               host brontes
>> 0       1                                                       osd.0   up      1
>> 1       1                                                       osd.1   up      1
>> 2       1                                                       osd.2   up      1
>> 3       1                                                       osd.3   up      1
>> 4       1                                                       osd.4   up      1
>> 5       1                                                       osd.5   up      1
>> 6       1                                                       osd.6   up      1
>> 7       1                                                       osd.7   up      1
>> -41     7.9                                     rack SAS15B02
>> -6      7.9                                             host alim
>> 24      1                                                       osd.24  up      1
>> 25      1                                                       osd.25  down    0
>> 26      1                                                       osd.26  up      1
>> 27      1                                                       osd.27  up      1
>> 28      1                                                       osd.28  up      1
>> 29      1                                                       osd.29  up      1
>> 30      1                                                       osd.30  up      1
>> 31      0.9                                                     osd.31  up      1
>> -101    17.53           datacenter SASrbx2
>> -91     17.53                   room SASs27
>> -70     1.6                             net SAS188-165-13
>> -44     0                                       rack SAS27B04
>> -7      0                                               host bul
>> -45     1.6                                     rack SAS27B06
>> -4      1.6                                             host okko
>> 32      0.2                                                     osd.32  up      1
>> 33      0.2                                                     osd.33  up      1
>> 34      0.2                                                     osd.34  up      1
>> 35      0.2                                                     osd.35  up      1
>> 36      0.2                                                     osd.36  up      1
>> 37      0.2                                                     osd.37  up      1
>> 38      0.2                                                     osd.38  up      1
>> 39      0.2                                                     osd.39  up      1
>> -71     15.93                           net SAS188-165-14
>> -42     8                                       rack SAS27A03
>> -5      8                                               host noburo
>> 8       1                                                       osd.8   up      1
>> 9       1                                                       osd.9   up      1
>> 18      1                                                       osd.18  up      1
>> 19      1                                                       osd.19  up      1
>> 20      1                                                       osd.20  up      1
>> 21      1                                                       osd.21  up      1
>> 22      1                                                       osd.22  up      1
>> 23      1                                                       osd.23  up      1
>> -43     7.93                                    rack SAS27A04
>> -2      7.93                                            host keron
>> 10      0.97                                                    osd.10  up      1
>> 11      1                                                       osd.11  up      1
>> 12      1                                                       osd.12  up      1
>> 13      1                                                       osd.13  up      1
>> 14      0.98                                                    osd.14  up      1
>> 15      1                                                       osd.15  down    0
>> 16      0.98                                                    osd.16  up      1
>> 17      1                                                       osd.17  up      1
>>
>>
>> Here I have 2 roots : SSDroot and SASroot. All my OSD/PG problems are on
>> the SAS branch, and my CRUSH rules use per "net" replication.
>>
>> The osd.15 have a failling disk since long time, its data was correctly
>> moved (= OSD was out until the cluster obtain HEALTH_OK).
>> The osd.25 is a buggy OSD that I can't remove or change : if I balance
>> it's PG on other OSD, then this others OSD crash. That problem occur
>> before I loose the osd.19 : OSD was unable to mark that PG as
>> inconsistent since it was crashing during scrub. For me, all
>> inconsistencies come from this OSD.
>> The osd.19 was a failling disk, that I changed.
>>
>>
>> And the health detail :
>>
>> HEALTH_ERR 15 pgs incomplete; 4 pgs inconsistent; 15 pgs stuck inactive;
>> 15 pgs stuck unclean; 1 near full osd(s); 19 scrub errors; noout flag(s)
>> set; 1 mons down, quorum 0,1,2,3 a,b,c,e
>> pg 4.5c is stuck inactive since forever, current state incomplete, last
>> acting [19,30]
>> pg 8.71d is stuck inactive since forever, current state incomplete, last
>> acting [24,19]
>> pg 8.3fa is stuck inactive since forever, current state incomplete, last
>> acting [19,31]
>> pg 8.3e0 is stuck inactive since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.56c is stuck inactive since forever, current state incomplete, last
>> acting [19,28]
>> pg 8.19f is stuck inactive since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.792 is stuck inactive since forever, current state incomplete, last
>> acting [19,28]
>> pg 4.0 is stuck inactive since forever, current state incomplete, last
>> acting [28,19]
>> pg 8.78a is stuck inactive since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.23e is stuck inactive since forever, current state incomplete, last
>> acting [32,13]
>> pg 8.2ff is stuck inactive since forever, current state incomplete, last
>> acting [6,19]
>> pg 8.5e2 is stuck inactive since forever, current state incomplete, last
>> acting [0,19]
>> pg 8.528 is stuck inactive since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.20f is stuck inactive since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.372 is stuck inactive since forever, current state incomplete, last
>> acting [19,24]
>> pg 4.5c is stuck unclean since forever, current state incomplete, last
>> acting [19,30]
>> pg 8.71d is stuck unclean since forever, current state incomplete, last
>> acting [24,19]
>> pg 8.3fa is stuck unclean since forever, current state incomplete, last
>> acting [19,31]
>> pg 8.3e0 is stuck unclean since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.56c is stuck unclean since forever, current state incomplete, last
>> acting [19,28]
>> pg 8.19f is stuck unclean since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.792 is stuck unclean since forever, current state incomplete, last
>> acting [19,28]
>> pg 4.0 is stuck unclean since forever, current state incomplete, last
>> acting [28,19]
>> pg 8.78a is stuck unclean since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.23e is stuck unclean since forever, current state incomplete, last
>> acting [32,13]
>> pg 8.2ff is stuck unclean since forever, current state incomplete, last
>> acting [6,19]
>> pg 8.5e2 is stuck unclean since forever, current state incomplete, last
>> acting [0,19]
>> pg 8.528 is stuck unclean since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.20f is stuck unclean since forever, current state incomplete, last
>> acting [31,19]
>> pg 8.372 is stuck unclean since forever, current state incomplete, last
>> acting [19,24]
>> pg 8.792 is incomplete, acting [19,28]
>> pg 8.78a is incomplete, acting [31,19]
>> pg 8.71d is incomplete, acting [24,19]
>> pg 8.5e2 is incomplete, acting [0,19]
>> pg 8.56c is incomplete, acting [19,28]
>> pg 8.528 is incomplete, acting [31,19]
>> pg 8.3fa is incomplete, acting [19,31]
>> pg 8.3e0 is incomplete, acting [31,19]
>> pg 8.372 is incomplete, acting [19,24]
>> pg 8.2ff is incomplete, acting [6,19]
>> pg 8.23e is incomplete, acting [32,13]
>> pg 8.20f is incomplete, acting [31,19]
>> pg 8.19f is incomplete, acting [31,19]
>> pg 3.7c is active+clean+inconsistent, acting [24,13,39]
>> pg 3.6b is active+clean+inconsistent, acting [28,23,5]
>> pg 4.5c is incomplete, acting [19,30]
>> pg 3.d is active+clean+inconsistent, acting [29,4,11]
>> pg 4.0 is incomplete, acting [28,19]
>> pg 3.1 is active+clean+inconsistent, acting [28,19,5]
>> osd.10 is near full at 85%
>> 19 scrub errors
>> noout flag(s) set
>> mon.d (rank 4) addr 10.0.0.6:6789/0 is down (out of quorum)
>>
>>
>> Pools 4 and 8 have only 2 replica, and pool 3 have 3 replica but
>> inconsistent data.
>>
>> Thanks in advance.
>>
>> Le vendredi 17 mai 2013 à 00:14 -0700, John Wilkins a écrit :
>>> If you can follow the documentation here:
>>> http://ceph.com/docs/master/rados/operations/monitoring-osd-pg/  and
>>> http://ceph.com/docs/master/rados/troubleshooting/  to provide some
>>> additional information, we may be better able to help you.
>>>
>>> For example, "ceph osd tree" would help us understand the status of
>>> your cluster a bit better.
>>>
>>> On Thu, May 16, 2013 at 10:32 PM, Olivier Bonvalet <ceph.list@xxxxxxxxx> wrote:
>>> > Le mercredi 15 mai 2013 à 00:15 +0200, Olivier Bonvalet a écrit :
>>> >> Hi,
>>> >>
>>> >> I have some PG in state down and/or incomplete on my cluster, because I
>>> >> loose 2 OSD and a pool was having only 2 replicas. So of course that
>>> >> data is lost.
>>> >>
>>> >> My problem now is that I can't retreive a "HEALTH_OK" status : if I try
>>> >> to remove, read or overwrite the corresponding RBD images, near all OSD
>>> >> hang (well... they don't do anything and requests stay in a growing
>>> >> queue, until the production will be done).
>>> >>
>>> >> So, what can I do to remove that corrupts images ?
>>> >>
>>> >> _______________________________________________
>>> >> ceph-users mailing list
>>> >> ceph-users@xxxxxxxxxxxxxx
>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >>
>>> >
>>> > Up. Nobody can help me on that problem ?
>>> >
>>> > Thanks,
>>> >
>>> > Olivier
>>> >
>>> > _______________________________________________
>>> > ceph-users mailing list
>>> > ceph-users@xxxxxxxxxxxxxx
>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>>
>>> --
>>> John Wilkins
>>> Senior Technical Writer
>>> Intank
>>> john.wilkins@xxxxxxxxxxx
>>> (415) 425-9599
>>> http://inktank.com
>>>
>>
>>
>
>
>
> --
> John Wilkins
> Senior Technical Writer
> Intank
> john.wilkins@xxxxxxxxxxx
> (415) 425-9599
> http://inktank.com



-- 
John Wilkins
Senior Technical Writer
Intank
john.wilkins@xxxxxxxxxxx
(415) 425-9599
http://inktank.com
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux