HEALTH_WARN 4 pgs incomplete; 4 pgs stuck inactive; 4 pgs stuck unclean

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It sound likes you need to throttle recovery.  I have this in my ceph.conf:
[osd]
  osd max backfills = 1
  osd recovery max active = 1
  osd recovery op priority = 1


Those configs, plus SSD journals, really helped the stability of my cluster
during recovery.  Before I made those changes, I would see OSDs get voted
down by other OSDs for not responding to heartbeats quickly.  Messages in
ceph.log like:
osd.# IP:PORT 420 : [WRN] map e41738 wrongly marked me down

are an indication that OSDs are so overloaded that they're getting kicked
out.



I also ran into problems when OSDs were getting kicked repeatedly.  It
caused those really large sections in pg query's
[recovery_state][past_intervals]
that you also have. I would restart an OSD, it would peer, and then suicide
timeout 300 seconds after starting the peering process.  When I first saw
it, it was only affecting a few OSDs.  If you're seeing repeated suicide
timeouts in the OSD's logs, there's a manual process to catch them up.



On Thu, Aug 14, 2014 at 12:25 AM, Riederer, Michael <Michael.Riederer at br.de>
wrote:

>  Hi Craig,
>
> Yes we have stability problems. The cluster is definitely not suitable for
> a production environment. I will not describe the details here. I want to get
> to know ceph and this is possible with the Test-cluster. Some osds are
> very slow, less than 15 MB / sec writable. Also increases the load on the
> ceph nodes to over 30 when a osd is removed and a reorganistation of the
> data is necessary. If the load is very high (over 30) I have seen exactly
> what you describe. osds go down and out and come back up and in.
>
> OK. I'll try the slow osd to remove and then to scrub, deep-scrub the pgs.
>
> Many thanks for your help.
>
> Regards,
> Mike
>
>  ------------------------------
> *Von:* Craig Lewis [clewis at centraldesktop.com]
> *Gesendet:* Mittwoch, 13. August 2014 19:48
>
> *An:* Riederer, Michael
> *Cc:* Karan Singh; ceph-users at lists.ceph.com
> *Betreff:* Re: [ceph-users] HEALTH_WARN 4 pgs incomplete; 4 pgs stuck
> inactive; 4 pgs stuck unclean
>
>   Yes, ceph pg <PGID> query, not dump.  Sorry about that.
>
>  Are you having problems with OSD stability?  There's a lot of history in
> the [recovery_state][past_intervals]. That's normal when OSDs go down,
> and out, and come back up and in. You have a lot of history there. You
> might even be getting into the point that you have so much failover
> history, the OSDs can't process it all before they hit the suicide timeout.
>
>  [recovery_state][probing_osds] lists a lot of OSDs that have recently
> owned these PGs. If the OSDs are crashing frequently, you need to get that
> under control before proceeding.
>
>  Once the OSDs are stable, I think Ceph just needs to scrub and
> deep-scrub those PGs.
>
>
>  Until Ceph clears out the [recovery_state][probing_osds] section in the
> pg query, it's not going to do anything.  ceph osd lost hears you, but
> doesn't trust you.  Ceph won't do anything until it's actually checked
> those OSDs itself.  Scrubbing and Deep scrubbing should convince it.
>
>  Once that [recovery_state][probing_osds] section is gone, you should see
> the [recovery_state][past_intervals] section shrink or disappear. I don't
> have either section in my pg query. Once that happens, your ceph pg repair
> or ceph pg force_create_pg should finally have some effect.  You may or
> may not need to re-issue those commands.
>
>
>
>
> On Tue, Aug 12, 2014 at 9:32 PM, Riederer, Michael <Michael.Riederer at br.de
> > wrote:
>
>>  Hi Craig,
>>
>> # ceph pg 2.587 query
>> # ceph pg 2.c1 query
>> # ceph pg 2.92 query
>> # ceph pg 2.e3 query
>>
>> Please download the output form here:
>> http://server.riederer.org/ceph-user/
>>
>> #####################################
>>
>>
>> It is not possible to map a rbd:
>>
>> # rbd map testshareone --pool rbd --name client.admin
>> rbd: add failed: (5) Input/output error
>>
>> I found that:
>> http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/11405
>>  # ceph osd getcrushmap -o crushmap.bin
>>  got crush map from osdmap epoch 3741
>> # crushtool -i crushmap.bin --set-chooseleaf_vary_r 0 -o crushmap-new.bin
>> # ceph osd setcrushmap -i crushmap-new.bin
>> set crush map
>>
>> The Cluster had to do some. Now it looks a bit different.
>>
>> It is still not possible to map a rbd.
>>
>>  root at ceph-admin-storage:~# ceph -s
>>     cluster 6b481875-8be5-4508-b075-e1f660fd7b33
>>      health HEALTH_WARN 4 pgs incomplete; 4 pgs stuck inactive; 4 pgs
>> stuck unclean
>>      monmap e2: 3 mons at {ceph-1-storage=
>> 10.65.150.101:6789/0,ceph-2-storage=10.65.150.102:6789/0,ceph-3-storage=10.65.150.103:6789/0},
>> election epoch 5010, quorum 0,1,2
>> ceph-1-storage,ceph-2-storage,ceph-3-storage
>>       osdmap e34206: 55 osds: 55 up, 55 in
>>       pgmap v10838368: 6144 pgs, 3 pools, 11002 GB data, 2762 kobjects
>>             22078 GB used, 79932 GB / 102010 GB avail
>>                 6140 active+clean
>>                    4 incomplete
>>
>> root at ceph-admin-storage:~# ceph health detail
>> HEALTH_WARN 4 pgs incomplete; 4 pgs stuck inactive; 4 pgs stuck unclean
>> pg 2.92 is stuck inactive since forever, current state incomplete, last
>> acting [8,13]
>>  pg 2.c1 is stuck inactive since forever, current state incomplete, last
>> acting [13,8]
>> pg 2.e3 is stuck inactive since forever, current state incomplete, last
>> acting [20,8]
>> pg 2.587 is stuck inactive since forever, current state incomplete, last
>> acting [13,8]
>>
>> pg 2.92 is stuck unclean since forever, current state incomplete, last
>> acting [8,13]
>>  pg 2.c1 is stuck unclean since forever, current state incomplete, last
>> acting [13,8]
>> pg 2.e3 is stuck unclean since forever, current state incomplete, last
>> acting [20,8]
>> pg 2.587 is stuck unclean since forever, current state incomplete, last
>> acting [13,8]
>> pg 2.587 is incomplete, acting [13,8]
>> pg 2.e3 is incomplete, acting [20,8]
>> pg 2.c1 is incomplete, acting [13,8]
>>
>> pg 2.92 is incomplete, acting [8,13]
>>
>>  #######################################################################
>>
>> After updating to firefly, I did the following:
>>
>> # ceph health detail
>> HEALTH_WARN crush map has legacy tunables crush map has legacy tunables;
>> see http://ceph.com/docs/master/rados/operations/crush-map/#tunables
>>
>> # ceph osd crush tunables optimal
>> adjusted tunables profile to optimal
>>
>> Mike
>>  ------------------------------
>> *Von:* Craig Lewis [clewis at centraldesktop.com]
>> *Gesendet:* Dienstag, 12. August 2014 20:02
>> *An:* Riederer, Michael
>> *Cc:* Karan Singh; ceph-users at lists.ceph.com
>>
>> *Betreff:* Re: [ceph-users] HEALTH_WARN 4 pgs incomplete; 4 pgs stuck
>> inactive; 4 pgs stuck unclean
>>
>>    For the incomplete PGs, can you give me the output of
>> ceph pg <PGID> dump
>>
>>  I'm interested in the recovery_state key of that JSON data.
>>
>>
>>
>> On Tue, Aug 12, 2014 at 5:29 AM, Riederer, Michael <
>> Michael.Riederer at br.de> wrote:
>>
>>>  Sorry, but I think that does not help me. I forgot to mention something about
>>> the operating system:
>>>
>>> root at ceph-1-storage:~# dpkg -l | grep libleveldb1
>>> ii  libleveldb1                       1.12.0-1precise.ceph
>>> fast key-value storage library
>>> root at ceph-1-storage:~# lsb_release -a
>>> No LSB modules are available.
>>> Distributor ID: Ubuntu
>>> Description:    Ubuntu 12.04.5 LTS
>>> Release:        12.04
>>> Codename:       precise
>>> root at ceph-1-storage:~# uname -a
>>> Linux ceph-1-storage 3.5.0-52-generic #79~precise1-Ubuntu SMP Fri Jul 4
>>> 21:03:49 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux
>>>
>>> libleveldb1 is greater than the mentioned version 1.9.0-1 ~ bpo70 + 1.
>>>
>>> All ceph nodes are IBM x3650 with Intel Xeon CPUs 2.00 GHz and 8 GB RAM,
>>> ok all very old, about eight years,
>>> but are still running.
>>>
>>> Mike
>>>
>>>
>>>
>>>  ------------------------------
>>> *Von:* Karan Singh [karan.singh at csc.fi]
>>> *Gesendet:* Dienstag, 12. August 2014 13:00
>>>
>>> *An:* Riederer, Michael
>>> *Cc:* ceph-users at lists.ceph.com
>>> *Betreff:* Re: [ceph-users] HEALTH_WARN 4 pgs incomplete; 4 pgs stuck
>>> inactive; 4 pgs stuck unclean
>>>
>>>   I am not sure if this helps , but have a look
>>> https://www.mail-archive.com/ceph-users at lists.ceph.com/msg10078.html
>>>
>>> - Karan -
>>>
>>>  On 12 Aug 2014, at 12:04, Riederer, Michael <Michael.Riederer at br.de>
>>> wrote:
>>>
>>>  Hi Karan,
>>>
>>> root at ceph-admin-storage:~/ceph-cluster/crush-map-4-ceph-user-list# ceph
>>> osd getcrushmap -o crushmap.bin
>>> got crush map from osdmap epoch 30748
>>> root at ceph-admin-storage:~/ceph-cluster/crush-map-4-ceph-user-list#
>>> crushtool -d crushmap.bin -o crushmap.txt
>>> root at ceph-admin-storage:~/ceph-cluster/crush-map-4-ceph-user-list# cat
>>> crushmap.txt
>>> # begin crush map
>>> tunable choose_local_tries 0
>>> tunable choose_local_fallback_tries 0
>>> tunable choose_total_tries 50
>>> tunable chooseleaf_descend_once 1
>>> tunable chooseleaf_vary_r 1
>>>
>>> # devices
>>> device 0 osd.0
>>> device 1 osd.1
>>> device 2 osd.2
>>> device 3 osd.3
>>> device 4 osd.4
>>> device 5 osd.5
>>> device 6 osd.6
>>> device 7 osd.7
>>> device 8 osd.8
>>> device 9 osd.9
>>> device 10 osd.10
>>> device 11 osd.11
>>> device 12 osd.12
>>> device 13 osd.13
>>> device 14 osd.14
>>> device 15 osd.15
>>> device 16 osd.16
>>> device 17 osd.17
>>> device 18 osd.18
>>> device 19 osd.19
>>> device 20 osd.20
>>> device 21 device21
>>> device 22 osd.22
>>> device 23 osd.23
>>> device 24 osd.24
>>> device 25 osd.25
>>> device 26 osd.26
>>> device 27 device27
>>> device 28 osd.28
>>> device 29 osd.29
>>> device 30 osd.30
>>> device 31 osd.31
>>> device 32 osd.32
>>> device 33 osd.33
>>> device 34 osd.34
>>> device 35 osd.35
>>> device 36 osd.36
>>> device 37 osd.37
>>> device 38 osd.38
>>> device 39 osd.39
>>> device 40 device40
>>> device 41 device41
>>> device 42 osd.42
>>> device 43 osd.43
>>> device 44 osd.44
>>> device 45 osd.45
>>> device 46 osd.46
>>> device 47 osd.47
>>> device 48 osd.48
>>> device 49 osd.49
>>> device 50 osd.50
>>> device 51 osd.51
>>> device 52 osd.52
>>> device 53 osd.53
>>> device 54 osd.54
>>> device 55 osd.55
>>> device 56 osd.56
>>> device 57 osd.57
>>> device 58 osd.58
>>>
>>> # types
>>> type 0 osd
>>> type 1 host
>>> type 2 rack
>>> type 3 row
>>> type 4 room
>>> type 5 datacenter
>>> type 6 root
>>>
>>> # buckets
>>> host ceph-1-storage {
>>>     id -2        # do not change unnecessarily
>>>     # weight 19.330
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.0 weight 0.910
>>>     item osd.2 weight 0.910
>>>     item osd.3 weight 0.910
>>>     item osd.4 weight 1.820
>>>     item osd.9 weight 1.360
>>>     item osd.11 weight 0.680
>>>     item osd.6 weight 3.640
>>>     item osd.5 weight 1.820
>>>     item osd.7 weight 3.640
>>>     item osd.8 weight 3.640
>>> }
>>> host ceph-2-storage {
>>>     id -3        # do not change unnecessarily
>>>     # weight 20.000
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.14 weight 3.640
>>>     item osd.18 weight 1.360
>>>     item osd.19 weight 1.360
>>>     item osd.15 weight 3.640
>>>     item osd.1 weight 3.640
>>>     item osd.12 weight 3.640
>>>     item osd.22 weight 0.680
>>>     item osd.23 weight 0.680
>>>     item osd.26 weight 0.680
>>>     item osd.36 weight 0.680
>>> }
>>> host ceph-5-storage {
>>>     id -4        # do not change unnecessarily
>>>     # weight 11.730
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.32 weight 0.270
>>>     item osd.37 weight 0.270
>>>     item osd.42 weight 0.270
>>>     item osd.43 weight 1.820
>>>     item osd.44 weight 1.820
>>>     item osd.45 weight 1.820
>>>     item osd.46 weight 1.820
>>>     item osd.47 weight 1.820
>>>     item osd.48 weight 1.820
>>> }
>>> room room0 {
>>>     id -8        # do not change unnecessarily
>>>     # weight 51.060
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item ceph-1-storage weight 19.330
>>>     item ceph-2-storage weight 20.000
>>>     item ceph-5-storage weight 11.730
>>> }
>>> host ceph-3-storage {
>>>     id -5        # do not change unnecessarily
>>>     # weight 15.920
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.24 weight 1.820
>>>     item osd.25 weight 1.820
>>>     item osd.29 weight 1.360
>>>     item osd.10 weight 3.640
>>>     item osd.13 weight 3.640
>>>     item osd.20 weight 3.640
>>> }
>>> host ceph-4-storage {
>>>     id -6        # do not change unnecessarily
>>>     # weight 20.000
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.34 weight 3.640
>>>     item osd.38 weight 1.360
>>>     item osd.39 weight 1.360
>>>     item osd.16 weight 3.640
>>>     item osd.30 weight 0.680
>>>     item osd.35 weight 3.640
>>>     item osd.17 weight 3.640
>>>     item osd.28 weight 0.680
>>>     item osd.31 weight 0.680
>>>     item osd.33 weight 0.680
>>> }
>>> host ceph-6-storage {
>>>     id -7        # do not change unnecessarily
>>>     # weight 12.720
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item osd.49 weight 0.450
>>>     item osd.50 weight 0.450
>>>     item osd.51 weight 0.450
>>>     item osd.52 weight 0.450
>>>     item osd.53 weight 1.820
>>>     item osd.54 weight 1.820
>>>     item osd.55 weight 1.820
>>>     item osd.56 weight 1.820
>>>     item osd.57 weight 1.820
>>>     item osd.58 weight 1.820
>>> }
>>> room room1 {
>>>     id -9        # do not change unnecessarily
>>>     # weight 48.640
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item ceph-3-storage weight 15.920
>>>     item ceph-4-storage weight 20.000
>>>     item ceph-6-storage weight 12.720
>>> }
>>> root default {
>>>     id -1        # do not change unnecessarily
>>>     # weight 99.700
>>>     alg straw
>>>     hash 0    # rjenkins1
>>>     item room0 weight 51.060
>>>     item room1 weight 48.640
>>> }
>>>
>>> # rules
>>> rule data {
>>>     ruleset 0
>>>     type replicated
>>>     min_size 1
>>>     max_size 10
>>>     step take default
>>>     step chooseleaf firstn 0 type host
>>>     step emit
>>> }
>>> rule metadata {
>>>     ruleset 1
>>>     type replicated
>>>     min_size 1
>>>     max_size 10
>>>     step take default
>>>     step chooseleaf firstn 0 type host
>>>     step emit
>>> }
>>> rule rbd {
>>>     ruleset 2
>>>     type replicated
>>>     min_size 1
>>>     max_size 10
>>>     step take default
>>>     step chooseleaf firstn 0 type host
>>>     step emit
>>> }
>>>
>>> # end crush map
>>>
>>> root at ceph-admin-storage:~# ceph osd dump | grep -i pool
>>> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
>>> rjenkins pg_num 2048 pgp_num 2048 last_change 4623 crash_replay_interval 45
>>> stripe_width 0
>>> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1
>>> object_hash rjenkins pg_num 2048 pgp_num 2048 last_change 4627 stripe_width
>>> 0
>>> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 2 object_hash
>>> rjenkins pg_num 2048 pgp_num 2048 last_change 4632 stripe_width 0
>>>
>>>
>>> Mike
>>>  ------------------------------
>>> *Von:* Karan Singh [karan.singh at csc.fi]
>>> *Gesendet:* Dienstag, 12. August 2014 10:35
>>> *An:* Riederer, Michael
>>> *Cc:* ceph-users at lists.ceph.com
>>> *Betreff:* Re: [ceph-users] HEALTH_WARN 4 pgs incomplete; 4 pgs stuck
>>> inactive; 4 pgs stuck unclean
>>>
>>>   Can you provide your cluster?s ceph osd dump | grep -i pool    and
>>> crush map output.
>>>
>>>
>>> - Karan -
>>>
>>>  On 12 Aug 2014, at 10:40, Riederer, Michael <Michael.Riederer at br.de>
>>> wrote:
>>>
>>>  Hi all,
>>>
>>> How do I get my Ceph Cluster back to a healthy state?
>>>
>>> root at ceph-admin-storage:~# ceph -v
>>> ceph version 0.80.5 (38b73c67d375a2552d8ed67843c8a65c2c0feba6)
>>> root at ceph-admin-storage:~# ceph -s
>>>     cluster 6b481875-8be5-4508-b075-e1f660fd7b33
>>>      health HEALTH_WARN 4 pgs incomplete; 4 pgs stuck inactive; 4 pgs
>>> stuck unclean
>>>      monmap e2: 3 mons at {ceph-1-storage=
>>> 10.65.150.101:6789/0,ceph-2-storage=10.65.150.102:6789/0,ceph-3-storage=10.65.150.103:6789/0},
>>> election epoch 5010, quorum 0,1,2
>>> ceph-1-storage,ceph-2-storage,ceph-3-storage
>>>      osdmap e30748: 55 osds: 55 up, 55 in
>>>       pgmap v10800465: 6144 pgs, 3 pools, 11002 GB data, 2762 kobjects
>>>             22077 GB used, 79933 GB / 102010 GB avail
>>>                 6138 active+clean
>>>                    4 incomplete
>>>                    2 active+clean+replay
>>> root at ceph-admin-storage:~# ceph health detail
>>> HEALTH_WARN 4 pgs incomplete; 4 pgs stuck inactive; 4 pgs stuck unclean
>>> pg 2.92 is stuck inactive since forever, current state incomplete, last
>>> acting [8,13]
>>> pg 2.c1 is stuck inactive since forever, current state incomplete, last
>>> acting [13,7]
>>> pg 2.e3 is stuck inactive since forever, current state incomplete, last
>>> acting [20,7]
>>> pg 2.587 is stuck inactive since forever, current state incomplete, last
>>> acting [13,5]
>>> pg 2.92 is stuck unclean since forever, current state incomplete, last
>>> acting [8,13]
>>> pg 2.c1 is stuck unclean since forever, current state incomplete, last
>>> acting [13,7]
>>> pg 2.e3 is stuck unclean since forever, current state incomplete, last
>>> acting [20,7]
>>> pg 2.587 is stuck unclean since forever, current state incomplete, last
>>> acting [13,5]
>>> pg 2.587 is incomplete, acting [13,5]
>>> pg 2.e3 is incomplete, acting [20,7]
>>> pg 2.c1 is incomplete, acting [13,7]
>>> pg 2.92 is incomplete, acting [8,13]
>>> root at ceph-admin-storage:~# ceph pg dump_stuck inactive
>>> ok
>>> pg_stat    objects    mip    degr    unf    bytes    log    disklog
>>> state    state_stamp    v    reported    up    up_primary    acting
>>> acting_primary    last_scrub    scrub_stamp    last_deep_scrub
>>> deep_scrub_stamp
>>> 2.92    0    0    0    0    0    0    0    incomplete    2014-08-08
>>> 12:39:20.204592    0'0    30748:7729    [8,13]    8    [8,13]    8
>>> 13503'1390419    2014-06-26 01:57:48.727625    13503'1390419    2014-06-22
>>> 01:57:30.114186
>>> 2.c1    0    0    0    0    0    0    0    incomplete    2014-08-08
>>> 12:39:18.846542    0'0    30748:7117    [13,7]    13    [13,7]    13
>>> 13503'1687017    2014-06-26 20:52:51.249864    13503'1687017    2014-06-22
>>> 14:24:22.633554
>>> 2.e3    0    0    0    0    0    0    0    incomplete    2014-08-08
>>> 12:39:29.311552    0'0    30748:8027    [20,7]    20    [20,7]    20
>>> 13503'1398727    2014-06-26 07:03:25.899254    13503'1398727    2014-06-21
>>> 07:02:31.393053
>>> 2.587    0    0    0    0    0    0    0    incomplete    2014-08-08
>>> 12:39:19.715724    0'0    30748:7060    [13,5]    13    [13,5]    13
>>> 13646'1542934    2014-06-26 07:48:42.089935    13646'1542934    2014-06-22
>>> 07:46:20.363695
>>> root at ceph-admin-storage:~# ceph osd tree
>>> # id    weight    type name    up/down    reweight
>>> -1    99.7    root default
>>> -8    51.06        room room0
>>> -2    19.33            host ceph-1-storage
>>> 0    0.91                osd.0    up    1
>>> 2    0.91                osd.2    up    1
>>> 3    0.91                osd.3    up    1
>>> 4    1.82                osd.4    up    1
>>> 9    1.36                osd.9    up    1
>>> 11    0.68                osd.11    up    1
>>> 6    3.64                osd.6    up    1
>>> 5    1.82                osd.5    up    1
>>> 7    3.64                osd.7    up    1
>>> 8    3.64                osd.8    up    1
>>> -3    20            host ceph-2-storage
>>> 14    3.64                osd.14    up    1
>>> 18    1.36                osd.18    up    1
>>> 19    1.36                osd.19    up    1
>>> 15    3.64                osd.15    up    1
>>> 1    3.64                osd.1    up    1
>>> 12    3.64                osd.12    up    1
>>> 22    0.68                osd.22    up    1
>>> 23    0.68                osd.23    up    1
>>> 26    0.68                osd.26    up    1
>>> 36    0.68                osd.36    up    1
>>> -4    11.73            host ceph-5-storage
>>> 32    0.27                osd.32    up    1
>>> 37    0.27                osd.37    up    1
>>> 42    0.27                osd.42    up    1
>>> 43    1.82                osd.43    up    1
>>> 44    1.82                osd.44    up    1
>>> 45    1.82                osd.45    up    1
>>> 46    1.82                osd.46    up    1
>>> 47    1.82                osd.47    up    1
>>> 48    1.82                osd.48    up    1
>>> -9    48.64        room room1
>>> -5    15.92            host ceph-3-storage
>>> 24    1.82                osd.24    up    1
>>> 25    1.82                osd.25    up    1
>>> 29    1.36                osd.29    up    1
>>> 10    3.64                osd.10    up    1
>>> 13    3.64                osd.13    up    1
>>> 20    3.64                osd.20    up    1
>>> -6    20            host ceph-4-storage
>>> 34    3.64                osd.34    up    1
>>> 38    1.36                osd.38    up    1
>>> 39    1.36                osd.39    up    1
>>> 16    3.64                osd.16    up    1
>>> 30    0.68                osd.30    up    1
>>> 35    3.64                osd.35    up    1
>>> 17    3.64                osd.17    up    1
>>> 28    0.68                osd.28    up    1
>>> 31    0.68                osd.31    up    1
>>> 33    0.68                osd.33    up    1
>>> -7    12.72            host ceph-6-storage
>>> 49    0.45                osd.49    up    1
>>> 50    0.45                osd.50    up    1
>>> 51    0.45                osd.51    up    1
>>> 52    0.45                osd.52    up    1
>>> 53    1.82                osd.53    up    1
>>> 54    1.82                osd.54    up    1
>>> 55    1.82                osd.55    up    1
>>> 56    1.82                osd.56    up    1
>>> 57    1.82                osd.57    up    1
>>> 58    1.82                osd.58    up    1
>>>
>>> What I have tried so far:
>>> ceph pg repair 2.587 [2.e3 2.c1 2.92]
>>> ceph pg force_create_pg 2.587 [2.e3 2.c1 2.92]
>>> ceph osd lost 5 --yes-i-really-mean-it [7 8 13 20]
>>>
>>> The history in brief:
>>> I installed Cuttlefish and updated to Dumpling and to Emperor. The
>>> Cluster was healthy. Maybe I made ??a mistake during repair of 8 broken
>>> osds, but from then on I had incompletepgs. At last I have updated from
>>> Emperor to Firefly.
>>>
>>> Regards,
>>> Mike
>>>  --------------------------------------------------------------------------------------------------
>>>  Bayerischer Rundfunk; Rundfunkplatz 1; 80335 M?nchen  Telefon: +49 89
>>> 590001; E-Mail: info at BR.de; Website: http://www.BR.de
>>> <http://www.br.de/> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users at lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>>   --------------------------------------------------------------------------------------------------
>>>  Bayerischer Rundfunk; Rundfunkplatz 1; 80335 M?nchen  Telefon: +49 89
>>> 590001; E-Mail: info at BR.de; Website: http://www.BR.de
>>> <http://www.br.de/>
>>>
>>>
>>>     --------------------------------------------------------------------------------------------------
>>>  Bayerischer Rundfunk; Rundfunkplatz 1; 80335 M?nchen  Telefon: +49 89
>>> 590001; E-Mail: info at BR.de; Website: http://www.BR.de
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users at lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140814/a92bfcab/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux