here yiu are The osd that have been added is osd.40 ssd, and it's an nautilus cluster Thanks for helping ------------- Crush Map ID CLASS WEIGHT (compat) TYPE NAME -26 33.37372 root HDD10 -33 6.67474 6.67474 host HDD10-ceph01 1 hdd10 1.66869 1.66869 osd.1 5 hdd10 1.66869 1.66869 osd.5 10 hdd10 1.66869 1.66869 osd.10 14 hdd10 1.66869 1.66869 osd.14 -34 6.67474 6.67474 host HDD10-ceph02 2 hdd10 1.66869 1.66869 osd.2 7 hdd10 1.66869 1.66869 osd.7 11 hdd10 1.66869 1.66869 osd.11 15 hdd10 1.66869 1.66869 osd.15 -35 6.67474 6.67474 host HDD10-ceph03 0 hdd10 1.66869 1.66869 osd.0 6 hdd10 1.66869 1.66869 osd.6 12 hdd10 1.66869 1.66869 osd.12 16 hdd10 1.66869 1.66869 osd.16 -36 6.67474 6.67474 host HDD10-ceph04 3 hdd10 1.66869 1.66869 osd.3 8 hdd10 1.66869 1.66869 osd.8 13 hdd10 1.66869 1.66869 osd.13 17 hdd10 1.66869 1.66869 osd.17 -37 6.67474 6.67474 host HDD10-ceph05 4 hdd10 1.66869 1.66869 osd.4 34 hdd10 1.66869 1.66869 osd.34 35 hdd10 1.66869 1.66869 osd.35 36 hdd10 1.66869 1.66869 osd.36 -25 10.53297 root SSD -42 3.54659 3.54659 host SSD-ceph01 31 ssd 1.74660 1.74660 osd.31 40 ssd 1.79999 1.79999 osd.40 -41 1.74660 1.74660 host SSD-ceph02 30 ssd 1.74660 1.74660 osd.30 -40 1.74660 1.74660 host SSD-ceph03 33 ssd 1.74660 1.74660 osd.33 -39 1.74660 1.74660 host SSD-ceph04 32 ssd 1.74660 1.74660 osd.32 -38 1.74660 1.74660 host SSD-ceph05 39 ssd 1.74660 1.74660 osd.39 -1 166.64085 root default -3 33.32817 33.32817 host ceph01 18 hdd 11.10939 11.10939 osd.18 22 hdd 11.10939 11.10939 osd.22 26 hdd 11.10939 11.10939 osd.26 -5 33.32817 33.32817 host ceph02 19 hdd 11.10939 11.10939 osd.19 23 hdd 11.10939 11.10939 osd.23 27 hdd 11.10939 11.10939 osd.27 -9 33.32817 33.32817 host ceph03 21 hdd 11.10939 11.10939 osd.21 25 hdd 11.10939 11.10939 osd.25 29 hdd 11.10939 11.10939 osd.29 -7 33.32817 33.32817 host ceph04 20 hdd 11.10939 11.10939 osd.20 24 hdd 11.10939 11.10939 osd.24 28 hdd 11.10939 11.10939 osd.28 -11 33.32817 33.32817 host ceph05 9 hdd 11.10939 11.10939 osd.9 37 hdd 11.10939 11.10939 osd.37 38 hdd 11.10939 11.10939 osd.38 ------------- Crush rules [ { "rule_id": 0, "rule_name": "replicated_rule", "ruleset": 0, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -1, "item_name": "default" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 1, "rule_name": "replicated-ssd", "ruleset": 1, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -32, "item_name": "SSD~ssd" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 2, "rule_name": "replicated-hdd10", "ruleset": 2, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -27, "item_name": "HDD10~hdd10" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 3, "rule_name": "replicated-hdd72", "ruleset": 3, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -18, "item_name": "default~hdd" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 4, "rule_name": "ec-ssd", "ruleset": 4, "type": 3, "min_size": 3, "max_size": 5, "steps": [ { "op": "set_chooseleaf_tries", "num": 5 }, { "op": "set_choose_tries", "num": 100 }, { "op": "take", "item": -32, "item_name": "SSD~ssd" }, { "op": "chooseleaf_indep", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 5, "rule_name": "ec-hdd10", "ruleset": 5, "type": 3, "min_size": 3, "max_size": 5, "steps": [ { "op": "set_chooseleaf_tries", "num": 5 }, { "op": "set_choose_tries", "num": 100 }, { "op": "take", "item": -27, "item_name": "HDD10~hdd10" }, { "op": "chooseleaf_indep", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 6, "rule_name": "ec-hdd72", "ruleset": 6, "type": 3, "min_size": 3, "max_size": 5, "steps": [ { "op": "set_chooseleaf_tries", "num": 5 }, { "op": "set_choose_tries", "num": 100 }, { "op": "take", "item": -18, "item_name": "default~hdd" }, { "op": "chooseleaf_indep", "num": 0, "type": "host" }, { "op": "emit" } ] } { "rule_id": 6, "rule_name": "ec-hdd72", "ruleset": 6, "type": 3, "min_size": 3, "max_size": 5, "steps": [ { "op": "set_chooseleaf_tries", "num": 5 }, { "op": "set_choose_tries", "num": 100 }, { "op": "take", "item": -18, "item_name": "default~hdd" }, { "op": "chooseleaf_indep", "num": 0, "type": "host" }, { "op": "emit" } ] } ] ------------- Crush Map cluster: id: 8700e000-d2ac-4393-8380-1cf4779166b5 health: HEALTH_WARN nodeep-scrub flag(s) set 71 pgs not deep-scrubbed in time services: mon: 5 daemons, quorum ceph01,ceph02,ceph03,ceph05,ceph04 (age 26h) mgr: ceph04(active, since 2d), standbys: ceph03, ceph05, ceph02, ceph01 mds: cephfs:1 {0=ceph02=up:active} 4 up:standby osd: 41 osds: 41 up (since 2h), 41 in (since 2h) flags nodeep-scrub rgw: 5 daemons active (ceph01.rgw0, ceph02.rgw0, ceph03.rgw0, ceph04.rgw0, ceph05.rgw0) data: pools: 23 pools, 1576 pgs objects: 11.55M objects, 44 TiB usage: 84 TiB used, 126 TiB / 210 TiB avail pgs: 1576 active+clean io: client: 154 MiB/s rd, 51 MiB/s wr, 2.02k op/s rd, 1.88k op/s wr cache: 23 MiB/s flush, 144 MiB/s evict, 44 op/s promote, 1 PGs flushing _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx