Hello List, any ideas why my OSDs are that unbalanced ? root@ceph01:~# ceph -s cluster: id: 5436dd5d-83d4-4dc8-a93b-60ab5db145df health: HEALTH_WARN 1 nearfull osd(s) 4 pool(s) nearfull services: mon: 3 daemons, quorum ceph03,ceph01,ceph02 (age 2w) mgr: ceph03(active, since 4M), standbys: ceph02.jwvivm mds: backup:1 {0=backup.ceph06.hdjehi=up:active} 3 up:standby osd: 56 osds: 56 up (since 29h), 56 in (since 3d) task status: scrub status: mds.backup.ceph06.hdjehi: idle data: pools: 4 pools, 1185 pgs objects: 24.29M objects, 44 TiB usage: 151 TiB used, 55 TiB / 206 TiB avail pgs: 675 active+clean 476 active+clean+snaptrim_wait 30 active+clean+snaptrim 4 active+clean+scrubbing+deep root@ceph01:~# ceph osd df tree ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS TYPE NAME -1 206.79979 - 206 TiB 151 TiB 151 TiB 36 GiB 503 GiB 55 TiB 73.23 1.00 - root default -2 28.89995 - 29 TiB 20 TiB 20 TiB 5.5 GiB 74 GiB 8.9 TiB 69.19 0.94 - host ceph01 0 hdd 2.70000 1.00000 2.7 TiB 1.8 TiB 1.8 TiB 590 MiB 6.9 GiB 908 GiB 66.81 0.91 44 up osd.0 1 hdd 2.70000 1.00000 2.7 TiB 1.6 TiB 1.6 TiB 411 MiB 6.5 GiB 1.1 TiB 60.43 0.83 39 up osd.1 4 hdd 2.70000 1.00000 2.7 TiB 1.8 TiB 1.8 TiB 501 MiB 6.8 GiB 898 GiB 67.15 0.92 43 up osd.4 8 hdd 2.70000 1.00000 2.7 TiB 2.0 TiB 2.0 TiB 453 MiB 7.0 GiB 700 GiB 74.39 1.02 47 up osd.8 11 hdd 1.70000 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 356 MiB 5.6 GiB 433 GiB 75.39 1.03 31 up osd.11 12 hdd 2.70000 1.00000 2.7 TiB 2.1 TiB 2.1 TiB 471 MiB 7.0 GiB 591 GiB 78.40 1.07 48 up osd.12 14 hdd 2.70000 1.00000 2.7 TiB 1.6 TiB 1.6 TiB 448 MiB 6.0 GiB 1.1 TiB 59.68 0.82 38 up osd.14 18 hdd 2.70000 1.00000 2.7 TiB 1.7 TiB 1.7 TiB 515 MiB 6.2 GiB 980 GiB 64.15 0.88 41 up osd.18 22 hdd 1.70000 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 360 MiB 4.2 GiB 491 GiB 72.06 0.98 29 up osd.22 30 hdd 1.70000 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 366 MiB 4.7 GiB 558 GiB 68.26 0.93 28 up osd.30 33 hdd 1.59999 1.00000 1.6 TiB 1.2 TiB 1.2 TiB 406 MiB 4.9 GiB 427 GiB 74.28 1.01 29 up osd.33 64 hdd 3.29999 1.00000 3.3 TiB 2.4 TiB 2.4 TiB 736 MiB 8.6 GiB 915 GiB 73.22 1.00 60 up osd.64 -3 29.69995 - 30 TiB 22 TiB 22 TiB 5.4 GiB 81 GiB 7.9 TiB 73.20 1.00 - host ceph02 2 hdd 1.70000 1.00000 1.7 TiB 1.3 TiB 1.2 TiB 402 MiB 5.2 GiB 476 GiB 72.93 1.00 30 up osd.2 3 hdd 2.70000 1.00000 2.7 TiB 2.0 TiB 2.0 TiB 653 MiB 7.8 GiB 652 GiB 76.15 1.04 49 up osd.3 7 hdd 2.70000 1.00000 2.7 TiB 2.5 TiB 2.5 TiB 456 MiB 7.7 GiB 209 GiB 92.36 1.26 56 up osd.7 9 hdd 2.70000 1.00000 2.7 TiB 1.9 TiB 1.9 TiB 434 MiB 7.2 GiB 781 GiB 71.46 0.98 46 up osd.9 13 hdd 2.39999 1.00000 2.4 TiB 1.6 TiB 1.6 TiB 451 MiB 6.1 GiB 823 GiB 66.28 0.91 38 up osd.13 16 hdd 2.70000 1.00000 2.7 TiB 1.6 TiB 1.6 TiB 375 MiB 6.4 GiB 1.1 TiB 59.84 0.82 39 up osd.16 19 hdd 1.70000 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 323 MiB 4.7 GiB 601 GiB 65.80 0.90 27 up osd.19 23 hdd 2.70000 1.00000 2.7 TiB 2.2 TiB 2.2 TiB 471 MiB 7.7 GiB 520 GiB 80.99 1.11 50 up osd.23 24 hdd 1.70000 1.00000 1.7 TiB 1.4 TiB 1.4 TiB 371 MiB 5.5 GiB 273 GiB 84.44 1.15 32 up osd.24 28 hdd 2.70000 1.00000 2.7 TiB 1.9 TiB 1.9 TiB 428 MiB 7.4 GiB 818 GiB 70.07 0.96 44 up osd.28 31 hdd 2.70000 1.00000 2.7 TiB 2.0 TiB 2.0 TiB 516 MiB 7.4 GiB 660 GiB 75.85 1.04 48 up osd.31 32 hdd 3.29999 1.00000 3.3 TiB 2.2 TiB 2.2 TiB 661 MiB 7.9 GiB 1.2 TiB 64.86 0.89 52 up osd.32 -4 26.29996 - 26 TiB 18 TiB 18 TiB 4.3 GiB 73 GiB 8.0 TiB 69.58 0.95 - host ceph03 5 hdd 1.70000 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 298 MiB 5.2 GiB 541 GiB 69.21 0.95 29 up osd.5 6 hdd 1.70000 1.00000 1.7 TiB 1.0 TiB 1.0 TiB 321 MiB 4.4 GiB 697 GiB 60.34 0.82 25 up osd.6 10 hdd 2.70000 1.00000 2.7 TiB 1.9 TiB 1.9 TiB 431 MiB 7.5 GiB 796 GiB 70.89 0.97 46 up osd.10 15 hdd 2.70000 1.00000 2.7 TiB 1.9 TiB 1.9 TiB 500 MiB 6.6 GiB 805 GiB 70.55 0.96 44 up osd.15 17 hdd 1.59999 1.00000 1.6 TiB 1.1 TiB 1.1 TiB 377 MiB 4.9 GiB 530 GiB 68.05 0.93 27 up osd.17 20 hdd 1.70000 1.00000 1.7 TiB 1.0 TiB 1.0 TiB 223 MiB 4.7 GiB 685 GiB 61.03 0.83 25 up osd.20 21 hdd 2.70000 1.00000 2.7 TiB 1.7 TiB 1.7 TiB 392 MiB 6.7 GiB 951 GiB 65.23 0.89 42 up osd.21 25 hdd 1.70000 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 157 MiB 5.1 GiB 601 GiB 65.83 0.90 27 up osd.25 26 hdd 2.70000 1.00000 2.7 TiB 2.1 TiB 2.1 TiB 512 MiB 7.6 GiB 573 GiB 79.06 1.08 50 up osd.26 27 hdd 2.70000 1.00000 2.7 TiB 1.9 TiB 1.9 TiB 473 MiB 7.6 GiB 805 GiB 70.55 0.96 46 up osd.27 29 hdd 2.70000 1.00000 2.7 TiB 2.1 TiB 2.1 TiB 478 MiB 7.3 GiB 539 GiB 80.29 1.10 50 up osd.29 63 hdd 1.70000 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 195 MiB 5.1 GiB 646 GiB 63.23 0.86 26 up osd.63 -11 24.79999 - 25 TiB 18 TiB 18 TiB 4.1 GiB 59 GiB 6.3 TiB 74.51 1.02 - host ceph04 34 hdd 5.20000 1.00000 5.2 TiB 3.9 TiB 3.8 TiB 954 MiB 13 GiB 1.4 TiB 73.48 1.00 94 up osd.34 42 hdd 5.20000 1.00000 5.2 TiB 3.9 TiB 3.8 TiB 841 MiB 13 GiB 1.4 TiB 73.43 1.00 94 up osd.42 44 hdd 7.20000 1.00000 7.2 TiB 5.5 TiB 5.5 TiB 1.2 GiB 17 GiB 1.6 TiB 77.54 1.06 133 up osd.44 45 hdd 7.20000 1.00000 7.2 TiB 5.2 TiB 5.2 TiB 1.2 GiB 16 GiB 1.9 TiB 73.03 1.00 125 up osd.45 -13 30.09998 - 30 TiB 22 TiB 22 TiB 5.1 GiB 72 GiB 8.0 TiB 73.48 1.00 - host ceph05 39 hdd 7.20000 1.00000 7.2 TiB 5.6 TiB 5.6 TiB 1.4 GiB 17 GiB 1.6 TiB 77.89 1.06 126 up osd.39 40 hdd 7.20000 1.00000 7.2 TiB 5.3 TiB 5.3 TiB 1.2 GiB 17 GiB 1.9 TiB 73.87 1.01 124 up osd.40 41 hdd 7.20000 1.00000 7.2 TiB 5.4 TiB 5.3 TiB 1.1 GiB 17 GiB 1.8 TiB 74.92 1.02 128 up osd.41 43 hdd 5.20000 1.00000 5.2 TiB 3.7 TiB 3.7 TiB 853 MiB 13 GiB 1.5 TiB 71.28 0.97 91 up osd.43 60 hdd 3.29999 1.00000 3.3 TiB 2.1 TiB 2.1 TiB 573 MiB 7.9 GiB 1.2 TiB 63.53 0.87 52 up osd.60 -9 17.59999 - 18 TiB 12 TiB 12 TiB 3.0 GiB 40 GiB 5.3 TiB 70.10 0.96 - host ceph06 35 hdd 7.20000 1.00000 7.2 TiB 5.2 TiB 5.2 TiB 1.3 GiB 16 GiB 1.9 TiB 72.80 0.99 125 up osd.35 36 hdd 5.20000 1.00000 5.2 TiB 3.5 TiB 3.5 TiB 804 MiB 12 GiB 1.8 TiB 66.50 0.91 85 up osd.36 38 hdd 5.20000 1.00000 5.2 TiB 3.7 TiB 3.7 TiB 978 MiB 12 GiB 1.6 TiB 70.02 0.96 88 up osd.38 -15 24.89998 - 25 TiB 18 TiB 18 TiB 4.3 GiB 58 GiB 6.5 TiB 73.75 1.01 - host ceph07 66 hdd 7.20000 1.00000 7.2 TiB 5.3 TiB 5.3 TiB 1.1 GiB 17 GiB 1.8 TiB 74.74 1.02 126 up osd.66 67 hdd 7.20000 1.00000 7.2 TiB 5.3 TiB 5.3 TiB 1.2 GiB 17 GiB 1.8 TiB 74.51 1.02 121 up osd.67 68 hdd 3.29999 1.00000 3.3 TiB 2.3 TiB 2.3 TiB 720 MiB 7.9 GiB 1.0 TiB 68.63 0.94 55 up osd.68 69 hdd 7.20000 1.00000 7.2 TiB 5.3 TiB 5.3 TiB 1.2 GiB 17 GiB 1.8 TiB 74.40 1.02 129 up osd.69 -17 24.50000 - 24 TiB 20 TiB 20 TiB 4.1 GiB 47 GiB 4.4 TiB 82.08 1.12 - host ceph08 37 hdd 9.50000 1.00000 9.5 TiB 7.8 TiB 7.7 TiB 1.5 GiB 18 GiB 1.8 TiB 81.39 1.11 166 up osd.37 46 hdd 5.00000 1.00000 5.0 TiB 4.0 TiB 3.9 TiB 889 MiB 9.3 GiB 1.0 TiB 79.67 1.09 87 up osd.46 47 hdd 5.00000 1.00000 5.0 TiB 4.2 TiB 4.2 TiB 863 MiB 9.7 GiB 817 GiB 83.90 1.15 90 up osd.47 48 hdd 5.00000 1.00000 5.0 TiB 4.2 TiB 4.2 TiB 969 MiB 10 GiB 813 GiB 83.99 1.15 91 up osd.48 TOTAL 206 TiB 151 TiB 151 TiB 36 GiB 503 GiB 55 TiB 73.23 MIN/MAX VAR: 0.82/1.26 STDDEV: 7.00 root@ceph01:~# ceph balancer status { "active": true, "last_optimize_duration": "0:00:00.016174", "last_optimize_started": "Fri Apr 16 12:54:47 2021", "mode": "upmap", "optimize_result": "Unable to find further optimization, or pool(s) pg_num is decreasing, or distribution is already perfect", "plans": [] } root@ceph01:~# ceph versions { "mon": { "ceph version 15.2.5 (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 3 }, "mgr": { "ceph version 15.2.5 (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 2 }, "osd": { "ceph version 15.2.5 (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 56 }, "mds": { "ceph version 15.2.5 (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 4 }, "overall": { "ceph version 15.2.5 (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 65 } } root@ceph01:~# ceph osd crush rule ls replicated_ruleset root@ceph01:~# ceph osd crush rule dump replicated_ruleset { "rule_id": 0, "rule_name": "replicated_ruleset", "ruleset": 0, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -1, "item_name": "default" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] } Cheers, Michael _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx