You most likely have pool size set to 3, but your crush rule requires replicas to be separated across DCs, of which you have only 2. -Sam On Mon, Dec 14, 2015 at 11:12 AM, Reno Rainz <rainzreno@xxxxxxxxx> wrote: > Thank you for your answer, but I don't really understand what do you mean. > > I use this map to distribute replicat into 2 differents dc, but I don't know > where the mistake is. > > Le 14 déc. 2015 7:56 PM, "Samuel Just" <sjust@xxxxxxxxxx> a écrit : >> >> 2 datacenters. >> -Sam >> >> On Mon, Dec 14, 2015 at 10:17 AM, Reno Rainz <rainzreno@xxxxxxxxx> wrote: >> > Hi, >> > >> > I got a functionnal and operationnal ceph cluster (in version 0.94.5), >> > with >> > 3 nodes (acting for MON and OSD), everything was fine. >> > >> > I added a 4th osd node (same configuration than 3 others) and now >> > cluster >> > status is health warn (active+remapped). >> > >> > >> > cluster e821c68f-995c-41a9-9c46-dbbd0a28b8c7 >> > health HEALTH_WARN >> > 256 pgs stuck unclean >> > recovery 279/1245 objects degraded (22.410%) >> > recovery 415/1245 objects misplaced (33.333%) >> > pool rbd pg_num 128 > pgp_num 64 >> > pool data pg_num 128 > pgp_num 100 >> > monmap e1: 3 mons at >> > >> > {ceph-osd-1=10.200.1.11:6789/0,ceph-osd-2=10.200.1.12:6789/0,ceph-osd-3=10.200.1.13:6789/0} >> > election epoch 4, quorum 0,1,2 >> > ceph-osd-1,ceph-osd-2,ceph-osd-3 >> > osdmap e57: 8 osds: 8 up, 8 in; 256 remapped pgs >> > pgmap v948: 256 pgs, 2 pools, 1566 MB data, 415 objects >> > 14929 MB used, 38237 MB / 55717 MB avail >> > 279/1245 objects degraded (22.410%) >> > 415/1245 objects misplaced (33.333%) >> > 256 active+remapped >> > >> > >> > OSD Tree >> > root@ceph-osd-1:~# ceph osd tree >> > ID WEIGHT TYPE NAME UP/DOWN REWEIGHT >> > PRIMARY-AFFINITY >> > -8 4.00000 root default >> > -7 4.00000 region eu-west-1 >> > -5 1.00000 datacenter eu-west-1a >> > -2 1.00000 host ceph-osd-1 >> > 0 1.00000 osd.0 up 1.00000 >> > 1.00000 >> > 1 1.00000 osd.1 up 1.00000 >> > 1.00000 >> > -4 1.00000 host ceph-osd-3 >> > 4 1.00000 osd.4 up 1.00000 >> > 1.00000 >> > 5 1.00000 osd.5 up 1.00000 >> > 1.00000 >> > -6 1.00000 datacenter eu-west-1b >> > -3 1.00000 host ceph-osd-2 >> > 2 1.00000 osd.2 up 1.00000 >> > 1.00000 >> > 3 1.00000 osd.3 up 1.00000 >> > 1.00000 >> > -9 1.00000 host ceph-osd-4 >> > 6 1.00000 osd.6 up 1.00000 >> > 1.00000 >> > 7 1.00000 osd.7 up 1.00000 >> > 1.00000 >> > root@ceph-osd-1:~# >> > >> > I'm using this crush map : >> > >> > { >> > "devices": [ >> > { >> > "id": 0, >> > "name": "osd.0" >> > }, >> > { >> > "id": 1, >> > "name": "osd.1" >> > }, >> > { >> > "id": 2, >> > "name": "osd.2" >> > }, >> > { >> > "id": 3, >> > "name": "osd.3" >> > }, >> > { >> > "id": 4, >> > "name": "osd.4" >> > }, >> > { >> > "id": 5, >> > "name": "osd.5" >> > }, >> > { >> > "id": 6, >> > "name": "osd.6" >> > }, >> > { >> > "id": 7, >> > "name": "osd.7" >> > } >> > ], >> > "types": [ >> > { >> > "type_id": 0, >> > "name": "osd" >> > }, >> > { >> > "type_id": 1, >> > "name": "host" >> > }, >> > { >> > "type_id": 2, >> > "name": "chassis" >> > }, >> > { >> > "type_id": 3, >> > "name": "rack" >> > }, >> > { >> > "type_id": 4, >> > "name": "row" >> > }, >> > { >> > "type_id": 5, >> > "name": "pdu" >> > }, >> > { >> > "type_id": 6, >> > "name": "pod" >> > }, >> > { >> > "type_id": 7, >> > "name": "room" >> > }, >> > { >> > "type_id": 8, >> > "name": "datacenter" >> > }, >> > { >> > "type_id": 9, >> > "name": "region" >> > }, >> > { >> > "type_id": 10, >> > "name": "root" >> > } >> > ], >> > "buckets": [ >> > { >> > "id": -2, >> > "name": "ceph-osd-1", >> > "type_id": 1, >> > "type_name": "host", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": 0, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": 1, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -3, >> > "name": "ceph-osd-2", >> > "type_id": 1, >> > "type_name": "host", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": 2, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": 3, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -4, >> > "name": "ceph-osd-3", >> > "type_id": 1, >> > "type_name": "host", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": 4, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": 5, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -5, >> > "name": "eu-west-1a", >> > "type_id": 8, >> > "type_name": "datacenter", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": -2, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": -4, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -6, >> > "name": "eu-west-1b", >> > "type_id": 8, >> > "type_name": "datacenter", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": -3, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": -9, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -7, >> > "name": "eu-west-1", >> > "type_id": 9, >> > "type_name": "region", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": -5, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": -6, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > }, >> > { >> > "id": -8, >> > "name": "default", >> > "type_id": 10, >> > "type_name": "root", >> > "weight": 262144, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": -7, >> > "weight": 262144, >> > "pos": 0 >> > } >> > ] >> > }, >> > { >> > "id": -9, >> > "name": "ceph-osd-4", >> > "type_id": 1, >> > "type_name": "host", >> > "weight": 131072, >> > "alg": "straw", >> > "hash": "rjenkins1", >> > "items": [ >> > { >> > "id": 6, >> > "weight": 65536, >> > "pos": 0 >> > }, >> > { >> > "id": 7, >> > "weight": 65536, >> > "pos": 1 >> > } >> > ] >> > } >> > ], >> > "rules": [ >> > { >> > "rule_id": 0, >> > "rule_name": "replicated_ruleset", >> > "ruleset": 0, >> > "type": 1, >> > "min_size": 1, >> > "max_size": 10, >> > "steps": [ >> > { >> > "op": "take", >> > "item": -8, >> > "item_name": "default" >> > }, >> > { >> > "op": "choose_firstn", >> > "num": 0, >> > "type": "datacenter" >> > }, >> > { >> > "op": "chooseleaf_firstn", >> > "num": 1, >> > "type": "host" >> > }, >> > { >> > "op": "emit" >> > } >> > ] >> > } >> > ], >> > "tunables": { >> > "choose_local_tries": 0, >> > "choose_local_fallback_tries": 0, >> > "choose_total_tries": 50, >> > "chooseleaf_descend_once": 1, >> > "chooseleaf_vary_r": 1, >> > "straw_calc_version": 1, >> > "allowed_bucket_algs": 54, >> > "profile": "hammer", >> > "optimal_tunables": 0, >> > "legacy_tunables": 0, >> > "require_feature_tunables": 1, >> > "require_feature_tunables2": 1, >> > "require_feature_tunables3": 1, >> > "has_v2_rules": 0, >> > "has_v3_rules": 0, >> > "has_v4_buckets": 0 >> > } >> > } >> > >> > >> > I read a thread >> > >> > (http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-November/006017.html) >> > from this mailling list, I tried everything (tunnable to optimal, add >> > more >> > pg, use the same weight ), but I still got this issue. >> > >> > Do you have any ideas to fix this situation ? >> > >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com