Re: Fix active+remapped situation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



You most likely have pool size set to 3, but your crush rule requires
replicas to be separated across DCs, of which you have only 2.
-Sam

On Mon, Dec 14, 2015 at 11:12 AM, Reno Rainz <rainzreno@xxxxxxxxx> wrote:
> Thank you for your answer, but I don't really understand what do you mean.
>
> I use this map to distribute replicat into 2 differents dc, but I don't know
> where the mistake is.
>
> Le 14 déc. 2015 7:56 PM, "Samuel Just" <sjust@xxxxxxxxxx> a écrit :
>>
>> 2 datacenters.
>> -Sam
>>
>> On Mon, Dec 14, 2015 at 10:17 AM, Reno Rainz <rainzreno@xxxxxxxxx> wrote:
>> > Hi,
>> >
>> > I got a functionnal  and operationnal ceph cluster (in version 0.94.5),
>> > with
>> > 3 nodes (acting for MON and OSD), everything was fine.
>> >
>> > I added a 4th osd node (same configuration than 3 others) and now
>> > cluster
>> > status is health warn (active+remapped).
>> >
>> >
>> >     cluster e821c68f-995c-41a9-9c46-dbbd0a28b8c7
>> >      health HEALTH_WARN
>> >             256 pgs stuck unclean
>> >             recovery 279/1245 objects degraded (22.410%)
>> >             recovery 415/1245 objects misplaced (33.333%)
>> >             pool rbd pg_num 128 > pgp_num 64
>> >             pool data pg_num 128 > pgp_num 100
>> >      monmap e1: 3 mons at
>> >
>> > {ceph-osd-1=10.200.1.11:6789/0,ceph-osd-2=10.200.1.12:6789/0,ceph-osd-3=10.200.1.13:6789/0}
>> >             election epoch 4, quorum 0,1,2
>> > ceph-osd-1,ceph-osd-2,ceph-osd-3
>> >      osdmap e57: 8 osds: 8 up, 8 in; 256 remapped pgs
>> >       pgmap v948: 256 pgs, 2 pools, 1566 MB data, 415 objects
>> >             14929 MB used, 38237 MB / 55717 MB avail
>> >             279/1245 objects degraded (22.410%)
>> >             415/1245 objects misplaced (33.333%)
>> >                  256 active+remapped
>> >
>> >
>> > OSD Tree
>> > root@ceph-osd-1:~# ceph osd tree
>> > ID WEIGHT  TYPE NAME                     UP/DOWN REWEIGHT
>> > PRIMARY-AFFINITY
>> > -8 4.00000 root default
>> > -7 4.00000     region eu-west-1
>> > -5 1.00000         datacenter eu-west-1a
>> > -2 1.00000             host ceph-osd-1
>> >  0 1.00000                 osd.0              up  1.00000
>> > 1.00000
>> >  1 1.00000                 osd.1              up  1.00000
>> > 1.00000
>> > -4 1.00000             host ceph-osd-3
>> >  4 1.00000                 osd.4              up  1.00000
>> > 1.00000
>> >  5 1.00000                 osd.5              up  1.00000
>> > 1.00000
>> > -6 1.00000         datacenter eu-west-1b
>> > -3 1.00000             host ceph-osd-2
>> >  2 1.00000                 osd.2              up  1.00000
>> > 1.00000
>> >  3 1.00000                 osd.3              up  1.00000
>> > 1.00000
>> > -9 1.00000             host ceph-osd-4
>> >  6 1.00000                 osd.6              up  1.00000
>> > 1.00000
>> >  7 1.00000                 osd.7              up  1.00000
>> > 1.00000
>> > root@ceph-osd-1:~#
>> >
>> >  I'm using this crush map :
>> >
>> > {
>> >     "devices": [
>> >         {
>> >             "id": 0,
>> >             "name": "osd.0"
>> >         },
>> >         {
>> >             "id": 1,
>> >             "name": "osd.1"
>> >         },
>> >         {
>> >             "id": 2,
>> >             "name": "osd.2"
>> >         },
>> >         {
>> >             "id": 3,
>> >             "name": "osd.3"
>> >         },
>> >         {
>> >             "id": 4,
>> >             "name": "osd.4"
>> >         },
>> >         {
>> >             "id": 5,
>> >             "name": "osd.5"
>> >         },
>> >         {
>> >             "id": 6,
>> >             "name": "osd.6"
>> >         },
>> >         {
>> >             "id": 7,
>> >             "name": "osd.7"
>> >         }
>> >     ],
>> >     "types": [
>> >         {
>> >             "type_id": 0,
>> >             "name": "osd"
>> >         },
>> >         {
>> >             "type_id": 1,
>> >             "name": "host"
>> >         },
>> >         {
>> >             "type_id": 2,
>> >             "name": "chassis"
>> >         },
>> >         {
>> >             "type_id": 3,
>> >             "name": "rack"
>> >         },
>> >         {
>> >             "type_id": 4,
>> >             "name": "row"
>> >         },
>> >         {
>> >             "type_id": 5,
>> >             "name": "pdu"
>> >         },
>> >         {
>> >             "type_id": 6,
>> >             "name": "pod"
>> >         },
>> >         {
>> >             "type_id": 7,
>> >             "name": "room"
>> >         },
>> >         {
>> >             "type_id": 8,
>> >             "name": "datacenter"
>> >         },
>> >         {
>> >             "type_id": 9,
>> >             "name": "region"
>> >         },
>> >         {
>> >             "type_id": 10,
>> >             "name": "root"
>> >         }
>> >     ],
>> >     "buckets": [
>> >         {
>> >             "id": -2,
>> >             "name": "ceph-osd-1",
>> >             "type_id": 1,
>> >             "type_name": "host",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": 0,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": 1,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -3,
>> >             "name": "ceph-osd-2",
>> >             "type_id": 1,
>> >             "type_name": "host",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": 2,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": 3,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -4,
>> >             "name": "ceph-osd-3",
>> >             "type_id": 1,
>> >             "type_name": "host",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": 4,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": 5,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -5,
>> >             "name": "eu-west-1a",
>> >             "type_id": 8,
>> >             "type_name": "datacenter",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": -2,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": -4,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -6,
>> >             "name": "eu-west-1b",
>> >             "type_id": 8,
>> >             "type_name": "datacenter",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": -3,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": -9,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -7,
>> >             "name": "eu-west-1",
>> >             "type_id": 9,
>> >             "type_name": "region",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": -5,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": -6,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -8,
>> >             "name": "default",
>> >             "type_id": 10,
>> >             "type_name": "root",
>> >             "weight": 262144,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": -7,
>> >                     "weight": 262144,
>> >                     "pos": 0
>> >                 }
>> >             ]
>> >         },
>> >         {
>> >             "id": -9,
>> >             "name": "ceph-osd-4",
>> >             "type_id": 1,
>> >             "type_name": "host",
>> >             "weight": 131072,
>> >             "alg": "straw",
>> >             "hash": "rjenkins1",
>> >             "items": [
>> >                 {
>> >                     "id": 6,
>> >                     "weight": 65536,
>> >                     "pos": 0
>> >                 },
>> >                 {
>> >                     "id": 7,
>> >                     "weight": 65536,
>> >                     "pos": 1
>> >                 }
>> >             ]
>> >         }
>> >     ],
>> >     "rules": [
>> >         {
>> >             "rule_id": 0,
>> >             "rule_name": "replicated_ruleset",
>> >             "ruleset": 0,
>> >             "type": 1,
>> >             "min_size": 1,
>> >             "max_size": 10,
>> >             "steps": [
>> >                 {
>> >                     "op": "take",
>> >                     "item": -8,
>> >                     "item_name": "default"
>> >                 },
>> >                 {
>> >                     "op": "choose_firstn",
>> >                     "num": 0,
>> >                     "type": "datacenter"
>> >                 },
>> >                 {
>> >                     "op": "chooseleaf_firstn",
>> >                     "num": 1,
>> >                     "type": "host"
>> >                 },
>> >                 {
>> >                     "op": "emit"
>> >                 }
>> >             ]
>> >         }
>> >     ],
>> >     "tunables": {
>> >         "choose_local_tries": 0,
>> >         "choose_local_fallback_tries": 0,
>> >         "choose_total_tries": 50,
>> >         "chooseleaf_descend_once": 1,
>> >         "chooseleaf_vary_r": 1,
>> >         "straw_calc_version": 1,
>> >         "allowed_bucket_algs": 54,
>> >         "profile": "hammer",
>> >         "optimal_tunables": 0,
>> >         "legacy_tunables": 0,
>> >         "require_feature_tunables": 1,
>> >         "require_feature_tunables2": 1,
>> >         "require_feature_tunables3": 1,
>> >         "has_v2_rules": 0,
>> >         "has_v3_rules": 0,
>> >         "has_v4_buckets": 0
>> >     }
>> > }
>> >
>> >
>> > I read a thread
>> >
>> > (http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-November/006017.html)
>> > from this mailling list, I tried everything (tunnable to optimal, add
>> > more
>> > pg, use the same weight ), but I still got this issue.
>> >
>> > Do you have any ideas to fix this situation ?
>> >
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux