Hey,
I have a cluster of 30 OSD's that is near perfect distribution minus two OSD's.
I am running ceph version 14.2.6 however has been the same for the previous versions, I have the balance module enabled in upmap and it says no improvements, I have also tried in crush mode.
ceph balancer status
{
"last_optimize_duration": "0:00:01.123659",
"plans": [],
"mode": "upmap",
"active": true,
"optimize_result": "Unable to find further optimization, or pool(s)' pg_num is decreasing, or distribution is already perfect",
"last_optimize_started": "Fri Jan 10 06:11:08 2020"
}
I have read a few email threads on the ML recently about similar cases but not sure if I am hitting the same "bug" as its only two that are off the rest are almost perfect.
ceph osd df
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS
23 hdd 0.00999 1.00000 10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21 33 up
24 hdd 0.00999 1.00000 10 GiB 1.4 GiB 441 MiB 48 KiB 1024 MiB 8.6 GiB 14.31 0.21 34 up
25 hdd 0.00999 1.00000 10 GiB 1.4 GiB 435 MiB 24 KiB 1024 MiB 8.6 GiB 14.26 0.21 34 up
26 hdd 0.00999 1.00000 10 GiB 1.4 GiB 436 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.27 0.21 34 up
27 hdd 0.00999 1.00000 10 GiB 1.4 GiB 437 MiB 16 KiB 1024 MiB 8.6 GiB 14.27 0.21 33 up
28 hdd 0.00999 1.00000 10 GiB 1.4 GiB 436 MiB 36 KiB 1024 MiB 8.6 GiB 14.26 0.21 34 up
3 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 76 KiB 19 GiB 3.0 TiB 67.26 1.00 170 up
4 hdd 9.09599 1.00000 9.1 TiB 6.2 TiB 6.1 TiB 44 KiB 19 GiB 2.9 TiB 67.77 1.01 172 up
5 hdd 9.09599 1.00000 9.1 TiB 6.3 TiB 6.3 TiB 112 KiB 20 GiB 2.8 TiB 69.50 1.03 176 up
6 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 17 KiB 19 GiB 2.9 TiB 67.58 1.01 171 up
7 hdd 9.09599 1.00000 9.1 TiB 6.7 TiB 6.7 TiB 88 KiB 21 GiB 2.4 TiB 73.98 1.10 187 up
8 hdd 9.09599 1.00000 9.1 TiB 6.5 TiB 6.5 TiB 76 KiB 20 GiB 2.6 TiB 71.84 1.07 182 up
9 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 120 KiB 19 GiB 3.0 TiB 67.24 1.00 170 up
10 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 72 KiB 19 GiB 3.0 TiB 67.19 1.00 170 up
11 hdd 9.09599 1.00000 9.1 TiB 6.2 TiB 6.2 TiB 40 KiB 19 GiB 2.9 TiB 68.06 1.01 172 up
12 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 28 KiB 19 GiB 3.0 TiB 67.48 1.00 170 up
13 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 36 KiB 19 GiB 3.0 TiB 67.04 1.00 170 up
14 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 108 KiB 19 GiB 3.0 TiB 67.30 1.00 170 up
15 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 68 KiB 19 GiB 3.0 TiB 67.41 1.00 170 up
16 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 152 KiB 19 GiB 2.9 TiB 67.61 1.01 171 up
17 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 36 KiB 19 GiB 3.0 TiB 67.16 1.00 170 up
18 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 41 KiB 19 GiB 3.0 TiB 67.19 1.00 170 up
19 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 64 KiB 19 GiB 3.0 TiB 67.49 1.00 171 up
20 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 12 KiB 19 GiB 3.0 TiB 67.55 1.01 171 up
21 hdd 9.09599 1.00000 9.1 TiB 6.2 TiB 6.1 TiB 76 KiB 19 GiB 2.9 TiB 67.76 1.01 171 up
22 hdd 9.09599 1.00000 9.1 TiB 6.2 TiB 6.2 TiB 12 KiB 19 GiB 2.9 TiB 68.05 1.01 172 up
29 hdd 9.09599 1.00000 9.1 TiB 5.8 TiB 5.8 TiB 108 KiB 17 GiB 3.3 TiB 63.59 0.95 163 up
30 hdd 9.09599 1.00000 9.1 TiB 5.9 TiB 5.9 TiB 24 KiB 18 GiB 3.2 TiB 65.18 0.97 167 up
31 hdd 9.09599 1.00000 9.1 TiB 6.1 TiB 6.1 TiB 44 KiB 18 GiB 3.0 TiB 66.74 0.99 171 up
32 hdd 9.09599 1.00000 9.1 TiB 6.0 TiB 6.0 TiB 220 KiB 18 GiB 3.1 TiB 66.31 0.99 170 up
33 hdd 9.09599 1.00000 9.1 TiB 6.0 TiB 5.9 TiB 36 KiB 18 GiB 3.1 TiB 65.54 0.98 168 up
34 hdd 9.09599 1.00000 9.1 TiB 6.0 TiB 6.0 TiB 44 KiB 18 GiB 3.1 TiB 66.33 0.99 170 up
35 hdd 9.09599 1.00000 9.1 TiB 5.9 TiB 5.9 TiB 68 KiB 18 GiB 3.2 TiB 64.77 0.96 166 up
36 hdd 9.09599 1.00000 9.1 TiB 5.8 TiB 5.8 TiB 168 KiB 17 GiB 3.3 TiB 63.60 0.95 163 up
37 hdd 9.09599 1.00000 9.1 TiB 6.0 TiB 6.0 TiB 60 KiB 18 GiB 3.1 TiB 65.91 0.98 169 up
38 hdd 9.09599 1.00000 9.1 TiB 5.9 TiB 5.9 TiB 68 KiB 18 GiB 3.2 TiB 65.15 0.97 167 up
0 hdd 0.00999 1.00000 10 GiB 1.4 GiB 437 MiB 28 KiB 1024 MiB 8.6 GiB 14.27 0.21 34 up
1 hdd 0.00999 1.00000 10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21 34 up
2 hdd 0.00999 1.00000 10 GiB 1.4 GiB 439 MiB 36 KiB 1024 MiB 8.6 GiB 14.29 0.21 33 up
23 hdd 0.00999 1.00000 10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21 33 up
24 hdd 0.00999 1.00000 10 GiB 1.4 GiB 441 MiB 48 KiB 1024 MiB 8.6 GiB 14.31 0.21 34 up
25 hdd 0.00999 1.00000 10 GiB 1.4 GiB 435 MiB 24 KiB 1024 MiB 8.6 GiB 14.26 0.21 34 up
26 hdd 0.00999 1.00000 10 GiB 1.4 GiB 436 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.27 0.21 34 up
27 hdd 0.00999 1.00000 10 GiB 1.4 GiB 437 MiB 16 KiB 1024 MiB 8.6 GiB 14.27 0.21 33 up
28 hdd 0.00999 1.00000 10 GiB 1.4 GiB 436 MiB 36 KiB 1024 MiB 8.6 GiB 14.26 0.21 34 up
TOTAL 273 TiB 183 TiB 183 TiB 6.4 MiB 567 GiB 90 TiB 67.17
ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-12 0.05798 root default
-11 0.02899 host sn-m02
23 hdd 0.00999 osd.23 up 1.00000 1.00000
24 hdd 0.00999 osd.24 up 1.00000 1.00000
25 hdd 0.00999 osd.25 up 1.00000 1.00000
-15 0.02899 host sn-m03
26 hdd 0.00999 osd.26 up 1.00000 1.00000
27 hdd 0.00999 osd.27 up 1.00000 1.00000
28 hdd 0.00999 osd.28 up 1.00000 1.00000
-6 272.87100 root ec
-5 90.95700 host sn-s01
3 hdd 9.09599 osd.3 up 1.00000 1.00000
4 hdd 9.09599 osd.4 up 1.00000 1.00000
5 hdd 9.09599 osd.5 up 1.00000 1.00000
6 hdd 9.09599 osd.6 up 1.00000 1.00000
7 hdd 9.09599 osd.7 up 1.00000 1.00000
8 hdd 9.09599 osd.8 up 1.00000 1.00000
9 hdd 9.09599 osd.9 up 1.00000 1.00000
10 hdd 9.09599 osd.10 up 1.00000 1.00000
11 hdd 9.09599 osd.11 up 1.00000 1.00000
12 hdd 9.09599 osd.12 up 1.00000 1.00000
-9 90.95700 host sn-s02
13 hdd 9.09599 osd.13 up 1.00000 1.00000
14 hdd 9.09599 osd.14 up 1.00000 1.00000
15 hdd 9.09599 osd.15 up 1.00000 1.00000
16 hdd 9.09599 osd.16 up 1.00000 1.00000
17 hdd 9.09599 osd.17 up 1.00000 1.00000
18 hdd 9.09599 osd.18 up 1.00000 1.00000
19 hdd 9.09599 osd.19 up 1.00000 1.00000
20 hdd 9.09599 osd.20 up 1.00000 1.00000
21 hdd 9.09599 osd.21 up 1.00000 1.00000
22 hdd 9.09599 osd.22 up 1.00000 1.00000
-17 90.95700 host sn-s03
29 hdd 9.09599 osd.29 up 1.00000 1.00000
30 hdd 9.09599 osd.30 up 1.00000 1.00000
31 hdd 9.09599 osd.31 up 1.00000 1.00000
32 hdd 9.09599 osd.32 up 1.00000 1.00000
33 hdd 9.09599 osd.33 up 1.00000 1.00000
34 hdd 9.09599 osd.34 up 1.00000 1.00000
35 hdd 9.09599 osd.35 up 1.00000 1.00000
36 hdd 9.09599 osd.36 up 1.00000 1.00000
37 hdd 9.09599 osd.37 up 1.00000 1.00000
38 hdd 9.09599 osd.38 up 1.00000 1.00000
-1 0.08698 root meta
-3 0.02899 host sn-m01
0 hdd 0.00999 osd.0 up 1.00000 1.00000
1 hdd 0.00999 osd.1 up 1.00000 1.00000
2 hdd 0.00999 osd.2 up 1.00000 1.00000
-11 0.02899 host sn-m02
23 hdd 0.00999 osd.23 up 1.00000 1.00000
24 hdd 0.00999 osd.24 up 1.00000 1.00000
25 hdd 0.00999 osd.25 up 1.00000 1.00000
-15 0.02899 host sn-m03
26 hdd 0.00999 osd.26 up 1.00000 1.00000
27 hdd 0.00999 osd.27 up 1.00000 1.00000
28 hdd 0.00999 osd.28 up 1.00000 1.00000
OSD 7,8 are the issue OSD's sitting at 182,187 PG's where the others are all sitting at 170,171.
Am I hitting the same issue? Or is there something I can do to re balance these extra PG's across the rest of the OSD better?
Thanks
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com