Re: Undersized pgs problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Bob,
Thanks for explanation, sounds resonable! But how it could happen that
host is down and its OSDs are still IN cluster?
I mean NOOUT flag is not set and my timeouts are fully default...

But if I remember correctly host was not completely down, it was
pingable but not other services were reachable like SSH or any others.
Is it possible that OSDs were still sending some information to
monitors making them look like IN?

2015-11-29 2:10 GMT+08:00 Bob R <bobr@xxxxxxxxxxxxxx>:
> Vasiliy,
>
> Your OSDs are marked as 'down' but 'in'.
>
> "Ceph OSDs have two known states that can be combined. Up and Down only
> tells you whether the OSD is actively involved in the cluster. OSD states
> also are expressed in terms of cluster replication: In and Out. Only when a
> Ceph OSD is tagged as Out does the self-healing process occur"
>
> Bob
>
> On Fri, Nov 27, 2015 at 6:15 AM, Mart van Santen <mart@xxxxxxxxxxxx> wrote:
>>
>>
>> Dear Vasilily,
>>
>>
>>
>> On 11/27/2015 02:00 PM, Irek Fasikhov wrote:
>>
>> You have time to synchronize?
>>
>> С уважением, Фасихов Ирек Нургаязович
>> Моб.: +79229045757
>>
>> 2015-11-27 15:57 GMT+03:00 Vasiliy Angapov <angapov@xxxxxxxxx>:
>>>
>>> > It seams that you played around with crushmap, and done something
>>> > wrong.
>>> > Compare the look of 'ceph osd tree' and crushmap. There are some 'osd'
>>> > devices renamed to 'device' think threre is you problem.
>>> Is this a mistake actually? What I did is removed a bunch of OSDs from
>>> my cluster that's why the numeration is sparse. But is it an issue to
>>> a have a sparse numeration of OSDs?
>>
>>
>> I think this is normal and should be no problem. I had this also
>> previously.
>>
>>>
>>> > Hi.
>>> > Vasiliy, Yes it is a problem with crusmap. Look at height:
>>> > -3 14.56000     host slpeah001
>>> > -2 14.56000     host slpeah002
>>> What exactly is wrong here?
>>
>>
>> I do not know how the weight of the hosts contribute to determine were to
>> store the 3-th copy of the PG. As you explained, you have enough space on
>> all hosts, but maybe if the weights of the hosts do not count up and the
>> crushmap maybe come to the conclusion it is not able to place the PGs. What
>> you can try, is to artificially raise the weights of these hosts, to see if
>> it starts mapping the thirth copies for the pg's onto the available host.
>>
>> I had a similiar problem in the past, this was solved by upgrading to the
>> latest crush tunables. But be aware, that can create massive datamovement
>> behavior.
>>
>>
>>>
>>> I also found out that my OSD logs are full of such records:
>>> 2015-11-26 08:31:19.273268 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:19.273276 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x41fd1000
>>> sd=79 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee1a520).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:24.273207 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:24.273225 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:24.273231 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x3f90b000
>>> sd=79 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee1a3c0).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:29.273199 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:29.273215 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:29.273222 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x41fd1000
>>> sd=79 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee1a260).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:34.273469 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:34.273482 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:34.273486 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x3f90b000
>>> sd=79 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee1a100).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:39.273310 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:39.273331 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:39.273342 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x41fcc000
>>> sd=98 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee19fa0).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:44.273753 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:44.273769 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:44.273776 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x41fcc000
>>> sd=98 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee189a0).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:49.273412 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>> 2015-11-26 08:31:49.273431 7fe4f49b1700  0 cephx: verify_authorizer
>>> could not get service secret for service osd secret_id=2924
>>> 2015-11-26 08:31:49.273455 7fe4f49b1700  0 --
>>> 192.168.254.18:6816/110740 >> 192.168.254.12:0/1011754 pipe(0x41fd1000
>>> sd=98 :6816 s=0 pgs=0 cs=0 l=1 c=0x3ee19080).accept: got bad
>>> authorizer
>>> 2015-11-26 08:31:54.273293 7fe4f49b1700  0 auth: could not find
>>> secret_id=2924
>>>
>>> What does it mean? Google sais it might be a time sync issue, but my
>>> clocks are perfectly synchronized...
>>
>>
>> Normally you get an error warning in "ceph status" if time is out of sync.
>> Nevertheless, you can try to restart the OSD's. I had issues with timing in
>> the past and discovered it sometime helps to restart the daemons *after*
>> syncing the times, before the accepted the new timings. But this was mostly
>> the case with monitors though.
>>
>>
>>
>> Regards,
>>
>>
>> Mart
>>
>>
>>
>>
>>>
>>> 2015-11-26 21:05 GMT+08:00 Irek Fasikhov <malmyzh@xxxxxxxxx>:
>>> > Hi.
>>> > Vasiliy, Yes it is a problem with crusmap. Look at height:
>>> > " -3 14.56000     host slpeah001
>>> >  -2 14.56000     host slpeah002
>>> >  "
>>> >
>>> > С уважением, Фасихов Ирек Нургаязович
>>> > Моб.: +79229045757
>>> >
>>> > 2015-11-26 13:16 GMT+03:00 ЦИТ РТ-Курамшин Камиль Фидаилевич
>>> > <Kamil.Kuramshin@xxxxxxxx>:
>>> >>
>>> >> It seams that you played around with crushmap, and done something
>>> >> wrong.
>>> >> Compare the look of 'ceph osd tree' and crushmap. There are some 'osd'
>>> >> devices renamed to 'device' think threre is you problem.
>>> >>
>>> >> Отправлено с мобильного устройства.
>>> >>
>>> >>
>>> >> -----Original Message-----
>>> >> From: Vasiliy Angapov <angapov@xxxxxxxxx>
>>> >> To: ceph-users <ceph-users@xxxxxxxxxxxxxx>
>>> >> Sent: чт, 26 нояб. 2015 7:53
>>> >> Subject:  Undersized pgs problem
>>> >>
>>> >> Hi, colleagues!
>>> >>
>>> >> I have small 4-node CEPH cluster (0.94.2), all pools have size 3,
>>> >> min_size
>>> >> 1.
>>> >> This night one host failed and cluster was unable to rebalance saying
>>> >> there are a lot of undersized pgs.
>>> >>
>>> >> root@slpeah002:[~]:# ceph -s
>>> >>     cluster 78eef61a-3e9c-447c-a3ec-ce84c617d728
>>> >>      health HEALTH_WARN
>>> >>             1486 pgs degraded
>>> >>             1486 pgs stuck degraded
>>> >>             2257 pgs stuck unclean
>>> >>             1486 pgs stuck undersized
>>> >>             1486 pgs undersized
>>> >>             recovery 80429/555185 objects degraded (14.487%)
>>> >>             recovery 40079/555185 objects misplaced (7.219%)
>>> >>             4/20 in osds are down
>>> >>             1 mons down, quorum 1,2 slpeah002,slpeah007
>>> >>      monmap e7: 3 mons at
>>> >>
>>> >>
>>> >> {slpeah001=192.168.254.11:6780/0,slpeah002=192.168.254.12:6780/0,slpeah007=172.31.252.46:6789/0}
>>> >>             election epoch 710, quorum 1,2 slpeah002,slpeah007
>>> >>      osdmap e14062: 20 osds: 16 up, 20 in; 771 remapped pgs
>>> >>       pgmap v7021316: 4160 pgs, 5 pools, 1045 GB data, 180 kobjects
>>> >>             3366 GB used, 93471 GB / 96838 GB avail
>>> >>             80429/555185 objects degraded (14.487%)
>>> >>             40079/555185 objects misplaced (7.219%)
>>> >>                 1903 active+clean
>>> >>                 1486 active+undersized+degraded
>>> >>                  771 active+remapped
>>> >>   client io 0 B/s rd, 246 kB/s wr, 67 op/s
>>> >>
>>> >>   root@slpeah002:[~]:# ceph osd tree
>>> >> ID  WEIGHT   TYPE NAME          UP/DOWN REWEIGHT PRIMARY-AFFINITY
>>> >>  -1 94.63998 root default
>>> >>  -9 32.75999     host slpeah007
>>> >>  72  5.45999         osd.72          up  1.00000          1.00000
>>> >>  73  5.45999         osd.73          up  1.00000          1.00000
>>> >>  74  5.45999         osd.74          up  1.00000          1.00000
>>> >>  75  5.45999         osd.75          up  1.00000          1.00000
>>> >>  76  5.45999         osd.76          up  1.00000          1.00000
>>> >>  77  5.45999         osd.77          up  1.00000          1.00000
>>> >> -10 32.75999     host slpeah008
>>> >>  78  5.45999         osd.78          up  1.00000          1.00000
>>> >>  79  5.45999         osd.79          up  1.00000          1.00000
>>> >>  80  5.45999         osd.80          up  1.00000          1.00000
>>> >>  81  5.45999         osd.81          up  1.00000          1.00000
>>> >>  82  5.45999         osd.82          up  1.00000          1.00000
>>> >>  83  5.45999         osd.83          up  1.00000          1.00000
>>> >>  -3 14.56000     host slpeah001
>>> >>   1  3.64000          osd.1         down  1.00000          1.00000
>>> >>  33  3.64000         osd.33        down  1.00000          1.00000
>>> >>  34  3.64000         osd.34        down  1.00000          1.00000
>>> >>  35  3.64000         osd.35        down  1.00000          1.00000
>>> >>  -2 14.56000     host slpeah002
>>> >>   0  3.64000         osd.0           up  1.00000          1.00000
>>> >>  36  3.64000         osd.36          up  1.00000          1.00000
>>> >>  37  3.64000         osd.37          up  1.00000          1.00000
>>> >>  38  3.64000         osd.38          up  1.00000          1.00000
>>> >>
>>> >> Crushmap:
>>> >>
>>> >>  # begin crush map
>>> >> tunable choose_local_tries 0
>>> >> tunable choose_local_fallback_tries 0
>>> >> tunable choose_total_tries 50
>>> >> tunable chooseleaf_descend_once 1
>>> >> tunable chooseleaf_vary_r 1
>>> >> tunable straw_calc_version 1
>>> >> tunable allowed_bucket_algs 54
>>> >>
>>> >> # devices
>>> >> device 0 osd.0
>>> >> device 1 osd.1
>>> >> device 2 device2
>>> >> device 3 device3
>>> >> device 4 device4
>>> >> device 5 device5
>>> >> device 6 device6
>>> >> device 7 device7
>>> >> device 8 device8
>>> >> device 9 device9
>>> >> device 10 device10
>>> >> device 11 device11
>>> >> device 12 device12
>>> >> device 13 device13
>>> >> device 14 device14
>>> >> device 15 device15
>>> >> device 16 device16
>>> >> device 17 device17
>>> >> device 18 device18
>>> >> device 19 device19
>>> >> device 20 device20
>>> >> device 21 device21
>>> >> device 22 device22
>>> >> device 23 device23
>>> >> device 24 device24
>>> >> device 25 device25
>>> >> device 26 device26
>>> >> device 27 device27
>>> >> device 28 device28
>>> >> device 29 device29
>>> >> device 30 device30
>>> >> device 31 device31
>>> >> device 32 device32
>>> >> device 33 osd.33
>>> >> device 34 osd.34
>>> >> device 35 osd.35
>>> >> device 36 osd.36
>>> >> device 37 osd.37
>>> >> device 38 osd.38
>>> >> device 39 device39
>>> >> device 40 device40
>>> >> device 41 device41
>>> >> device 42 device42
>>> >> device 43 device43
>>> >> device 44 device44
>>> >> device 45 device45
>>> >> device 46 device46
>>> >> device 47 device47
>>> >> device 48 device48
>>> >> device 49 device49
>>> >> device 50 device50
>>> >> device 51 device51
>>> >> device 52 device52
>>> >> device 53 device53
>>> >> device 54 device54
>>> >> device 55 device55
>>> >> device 56 device56
>>> >> device 57 device57
>>> >> device 58 device58
>>> >> device 59 device59
>>> >> device 60 device60
>>> >> device 61 device61
>>> >> device 62 device62
>>> >> device 63 device63
>>> >> device 64 device64
>>> >> device 65 device65
>>> >> device 66 device66
>>> >> device 67 device67
>>> >> device 68 device68
>>> >> device 69 device69
>>> >> device 70 device70
>>> >> device 71 device71
>>> >> device 72 osd.72
>>> >> device 73 osd.73
>>> >> device 74 osd.74
>>> >> device 75 osd.75
>>> >> device 76 osd.76
>>> >> device 77 osd.77
>>> >> device 78 osd.78
>>> >> device 79 osd.79
>>> >> device 80 osd.80
>>> >> device 81 osd.81
>>> >> device 82 osd.82
>>> >> device 83 osd.83
>>> >>
>>> >> # types
>>> >> type 0 osd
>>> >> type 1 host
>>> >> type 2 chassis
>>> >> type 3 rack
>>> >> type 4 row
>>> >> type 5 pdu
>>> >> type 6 pod
>>> >> type 7 room
>>> >> type 8 datacenter
>>> >> type 9 region
>>> >> type 10 root
>>> >>
>>> >> # buckets
>>> >> host slpeah007 {
>>> >>         id -9           # do not change unnecessarily
>>> >>         # weight 32.760
>>> >>         alg straw
>>> >>         hash 0  # rjenkins1
>>> >>         item osd.72 weight 5.460
>>> >>         item osd.73 weight 5.460
>>> >>         item osd.74 weight 5.460
>>> >>         item osd.75 weight 5.460
>>> >>         item osd.76 weight 5.460
>>> >>         item osd.77 weight 5.460
>>> >> }
>>> >> host slpeah008 {
>>> >>         id -10          # do not change unnecessarily
>>> >>         # weight 32.760
>>> >>         alg straw
>>> >>         hash 0  # rjenkins1
>>> >>         item osd.78 weight 5.460
>>> >>         item osd.79 weight 5.460
>>> >>         item osd.80 weight 5.460
>>> >>         item osd.81 weight 5.460
>>> >>         item osd.82 weight 5.460
>>> >>         item osd.83 weight 5.460
>>> >> }
>>> >> host slpeah001 {
>>> >>         id -3           # do not change unnecessarily
>>> >>         # weight 14.560
>>> >>         alg straw
>>> >>         hash 0  # rjenkins1
>>> >>         item osd.1 weight 3.640
>>> >>         item osd.33 weight 3.640
>>> >>         item osd.34 weight 3.640
>>> >>         item osd.35 weight 3.640
>>> >> }
>>> >> host slpeah002 {
>>> >>         id -2           # do not change unnecessarily
>>> >>         # weight 14.560
>>> >>         alg straw
>>> >>         hash 0  # rjenkins1
>>> >>         item osd.0 weight 3.640
>>> >>         item osd.36 weight 3.640
>>> >>         item osd.37 weight 3.640
>>> >>         item osd.38 weight 3.640
>>> >> }
>>> >> root default {
>>> >>         id -1           # do not change unnecessarily
>>> >>         # weight 94.640
>>> >>         alg straw
>>> >>         hash 0  # rjenkins1
>>> >>         item slpeah007 weight 32.760
>>> >>         item slpeah008 weight 32.760
>>> >>         item slpeah001 weight 14.560
>>> >>         item slpeah002 weight 14.560
>>> >> }
>>> >>
>>> >> # rules
>>> >> rule default {
>>> >>         ruleset 0
>>> >>         type replicated
>>> >>         min_size 1
>>> >>         max_size 10
>>> >>         step take default
>>> >>         step chooseleaf firstn 0 type host
>>> >>         step emit
>>> >> }
>>> >>
>>> >> # end crush map
>>> >>
>>> >>
>>> >>
>>> >> This is odd because pools have size 3 and I have 3 hosts alive, so why
>>> >> it is saying that undersized pgs are present? It makes me feel like
>>> >> CRUSH is not working properly.
>>> >> There is not much data currently in cluster, something about 3TB and
>>> >> as you can see from osd tree - each host have minimum of 14TB disk
>>> >> space on OSDs.
>>> >> So I'm a bit stuck now...
>>> >> How can I find the source of trouble?
>>> >>
>>> >> Thanks in advance!
>>> >> _______________________________________________
>>> >> ceph-users mailing list
>>> >> ceph-users@xxxxxxxxxxxxxx
>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >>
>>> >> _______________________________________________
>>> >> ceph-users mailing list
>>> >> ceph-users@xxxxxxxxxxxxxx
>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >>
>>> >
>>
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> --
>> Mart van Santen
>> Greenhost
>> E: mart@xxxxxxxxxxxx
>> T: +31 20 4890444
>> W: https://greenhost.nl
>>
>> A PGP signature can be attached to this e-mail,
>> you need PGP software to verify it.
>> My public key is available in keyserver(s)
>> see: http://tinyurl.com/openpgp-manual
>>
>> PGP Fingerprint: CA85 EB11 2B70 042D AF66  B29A 6437 01A1 10A3 D3A5
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux