Thanks for the insight and links. > As I can see you are on Luminous. Since Luminous Balancer plugin is > available [1], you should use it instead reweight's in place, especially > in upmap mode [2] I'll try it out again - last I tried it complanied about older clients - it should be better now. > Also, may be I can catch another crush mistakes, can I see `ceph osd > crush show-tunables, `ceph osd crush rule dump`, `ceph osd pool ls > detail`? Here: $ sudo ceph osd crush show-tunables { "choose_local_tries": 0, "choose_local_fallback_tries": 0, "choose_total_tries": 50, "chooseleaf_descend_once": 1, "chooseleaf_vary_r": 1, "chooseleaf_stable": 0, "straw_calc_version": 1, "allowed_bucket_algs": 54, "profile": "hammer", "optimal_tunables": 0, "legacy_tunables": 0, "minimum_required_version": "hammer", "require_feature_tunables": 1, "require_feature_tunables2": 1, "has_v2_rules": 1, "require_feature_tunables3": 1, "has_v3_rules": 0, "has_v4_buckets": 1, "require_feature_tunables5": 0, "has_v5_rules": 0 } $ sudo ceph osd crush rule dump [ { "rule_id": 0, "rule_name": "replicated_ruleset_hdd", "ruleset": 0, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -1, "item_name": "default~hdd" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 1, "rule_name": "replicated_ruleset_hdd_fast", "ruleset": 1, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -28, "item_name": "default~hdd_fast" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 2, "rule_name": "replicated_ruleset_ssd", "ruleset": 2, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -21, "item_name": "default~ssd" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 3, "rule_name": "cephfs_data_ec42", "ruleset": 3, "type": 3, "min_size": 3, "max_size": 6, "steps": [ { "op": "set_chooseleaf_tries", "num": 5 }, { "op": "set_choose_tries", "num": 100 }, { "op": "take", "item": -1, "item_name": "default~hdd" }, { "op": "chooseleaf_indep", "num": 0, "type": "host" }, { "op": "emit" } ] } ] $ sudo ceph osd pool ls detail pool 6 'kube' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 41045 flags hashpspool stripe_width 0 application rbd removed_snaps [1~3] pool 15 'default.rgw.buckets.data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 17 'default.rgw.users.keys' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 41045 lfor 0/36590 flags hashpspool stripe_width 0 application rgw pool 18 'default.rgw.buckets.non-ec' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 41045 lfor 0/36595 flags hashpspool stripe_width 0 application rgw pool 19 'default.rgw.users.uid' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 41045 lfor 0/36608 flags hashpspool stripe_width 0 application rgw pool 20 'rbd' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 41045 flags hashpspool stripe_width 0 application rbd pool 26 'default.rgw.data.root' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 27 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 28 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 29 'default.rgw.gc' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 30 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 31 'default.rgw.buckets.index' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 last_change 41045 flags hashpspool stripe_width 0 application rgw pool 32 'cephfs_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 2048 pgp_num 2048 last_change 45215 lfor 0/45204 flags hashpspool stripe_width 0 application cephfs pool 33 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 2 object_hash rjenkins pg_num 256 pgp_num 256 last_change 34756 flags hashpspool stripe_width 0 application cephfs pool 34 'default.rgw.usage' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 41045 lfor 0/36615 flags hashpspool stripe_width 0 application rgw pool 44 'cephfs_data_ec42' erasure size 6 min_size 4 crush_rule 3 object_hash rjenkins pg_num 1024 pgp_num 1024 last_change 43464 lfor 0/43453 flags hashpspool,ec_overwrites stripe_width 16384 application cephfs _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com