Near Perfect PG distrubtion apart from two OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey,

I have a cluster of 30 OSD's that is near perfect distribution minus two OSD's.

I am running ceph version 14.2.6 however has been the same for the previous versions, I have the balance module enabled in upmap and it says no improvements, I have also tried in crush mode.

ceph balancer status
{
    "last_optimize_duration": "0:00:01.123659",
    "plans": [],
    "mode": "upmap",
    "active": true,
    "optimize_result": "Unable to find further optimization, or pool(s)' pg_num is decreasing, or distribution is already perfect",
    "last_optimize_started": "Fri Jan 10 06:11:08 2020"
}

I have read a few email threads on the ML recently about similar cases but not sure if I am hitting the same "bug" as its only two that are off the rest are almost perfect.

ceph osd df
ID CLASS WEIGHT  REWEIGHT SIZE    RAW USE DATA    OMAP    META     AVAIL   %USE  VAR  PGS STATUS
23   hdd 0.00999  1.00000  10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21  33     up
24   hdd 0.00999  1.00000  10 GiB 1.4 GiB 441 MiB  48 KiB 1024 MiB 8.6 GiB 14.31 0.21  34     up
25   hdd 0.00999  1.00000  10 GiB 1.4 GiB 435 MiB  24 KiB 1024 MiB 8.6 GiB 14.26 0.21  34     up
26   hdd 0.00999  1.00000  10 GiB 1.4 GiB 436 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.27 0.21  34     up
27   hdd 0.00999  1.00000  10 GiB 1.4 GiB 437 MiB  16 KiB 1024 MiB 8.6 GiB 14.27 0.21  33     up
28   hdd 0.00999  1.00000  10 GiB 1.4 GiB 436 MiB  36 KiB 1024 MiB 8.6 GiB 14.26 0.21  34     up
 3   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  76 KiB   19 GiB 3.0 TiB 67.26 1.00 170     up
 4   hdd 9.09599  1.00000 9.1 TiB 6.2 TiB 6.1 TiB  44 KiB   19 GiB 2.9 TiB 67.77 1.01 172     up
 5   hdd 9.09599  1.00000 9.1 TiB 6.3 TiB 6.3 TiB 112 KiB   20 GiB 2.8 TiB 69.50 1.03 176     up
 6   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  17 KiB   19 GiB 2.9 TiB 67.58 1.01 171     up
 7   hdd 9.09599  1.00000 9.1 TiB 6.7 TiB 6.7 TiB  88 KiB   21 GiB 2.4 TiB 73.98 1.10 187     up
 8   hdd 9.09599  1.00000 9.1 TiB 6.5 TiB 6.5 TiB  76 KiB   20 GiB 2.6 TiB 71.84 1.07 182     up
 9   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB 120 KiB   19 GiB 3.0 TiB 67.24 1.00 170     up
10   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  72 KiB   19 GiB 3.0 TiB 67.19 1.00 170     up
11   hdd 9.09599  1.00000 9.1 TiB 6.2 TiB 6.2 TiB  40 KiB   19 GiB 2.9 TiB 68.06 1.01 172     up
12   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  28 KiB   19 GiB 3.0 TiB 67.48 1.00 170     up
13   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  36 KiB   19 GiB 3.0 TiB 67.04 1.00 170     up
14   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB 108 KiB   19 GiB 3.0 TiB 67.30 1.00 170     up
15   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  68 KiB   19 GiB 3.0 TiB 67.41 1.00 170     up
16   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB 152 KiB   19 GiB 2.9 TiB 67.61 1.01 171     up
17   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  36 KiB   19 GiB 3.0 TiB 67.16 1.00 170     up
18   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  41 KiB   19 GiB 3.0 TiB 67.19 1.00 170     up
19   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  64 KiB   19 GiB 3.0 TiB 67.49 1.00 171     up
20   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  12 KiB   19 GiB 3.0 TiB 67.55 1.01 171     up
21   hdd 9.09599  1.00000 9.1 TiB 6.2 TiB 6.1 TiB  76 KiB   19 GiB 2.9 TiB 67.76 1.01 171     up
22   hdd 9.09599  1.00000 9.1 TiB 6.2 TiB 6.2 TiB  12 KiB   19 GiB 2.9 TiB 68.05 1.01 172     up
29   hdd 9.09599  1.00000 9.1 TiB 5.8 TiB 5.8 TiB 108 KiB   17 GiB 3.3 TiB 63.59 0.95 163     up
30   hdd 9.09599  1.00000 9.1 TiB 5.9 TiB 5.9 TiB  24 KiB   18 GiB 3.2 TiB 65.18 0.97 167     up
31   hdd 9.09599  1.00000 9.1 TiB 6.1 TiB 6.1 TiB  44 KiB   18 GiB 3.0 TiB 66.74 0.99 171     up
32   hdd 9.09599  1.00000 9.1 TiB 6.0 TiB 6.0 TiB 220 KiB   18 GiB 3.1 TiB 66.31 0.99 170     up
33   hdd 9.09599  1.00000 9.1 TiB 6.0 TiB 5.9 TiB  36 KiB   18 GiB 3.1 TiB 65.54 0.98 168     up
34   hdd 9.09599  1.00000 9.1 TiB 6.0 TiB 6.0 TiB  44 KiB   18 GiB 3.1 TiB 66.33 0.99 170     up
35   hdd 9.09599  1.00000 9.1 TiB 5.9 TiB 5.9 TiB  68 KiB   18 GiB 3.2 TiB 64.77 0.96 166     up
36   hdd 9.09599  1.00000 9.1 TiB 5.8 TiB 5.8 TiB 168 KiB   17 GiB 3.3 TiB 63.60 0.95 163     up
37   hdd 9.09599  1.00000 9.1 TiB 6.0 TiB 6.0 TiB  60 KiB   18 GiB 3.1 TiB 65.91 0.98 169     up
38   hdd 9.09599  1.00000 9.1 TiB 5.9 TiB 5.9 TiB  68 KiB   18 GiB 3.2 TiB 65.15 0.97 167     up
 0   hdd 0.00999  1.00000  10 GiB 1.4 GiB 437 MiB  28 KiB 1024 MiB 8.6 GiB 14.27 0.21  34     up
 1   hdd 0.00999  1.00000  10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21  34     up
 2   hdd 0.00999  1.00000  10 GiB 1.4 GiB 439 MiB  36 KiB 1024 MiB 8.6 GiB 14.29 0.21  33     up
23   hdd 0.00999  1.00000  10 GiB 1.4 GiB 434 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.24 0.21  33     up
24   hdd 0.00999  1.00000  10 GiB 1.4 GiB 441 MiB  48 KiB 1024 MiB 8.6 GiB 14.31 0.21  34     up
25   hdd 0.00999  1.00000  10 GiB 1.4 GiB 435 MiB  24 KiB 1024 MiB 8.6 GiB 14.26 0.21  34     up
26   hdd 0.00999  1.00000  10 GiB 1.4 GiB 436 MiB 1.4 MiB 1023 MiB 8.6 GiB 14.27 0.21  34     up
27   hdd 0.00999  1.00000  10 GiB 1.4 GiB 437 MiB  16 KiB 1024 MiB 8.6 GiB 14.27 0.21  33     up
28   hdd 0.00999  1.00000  10 GiB 1.4 GiB 436 MiB  36 KiB 1024 MiB 8.6 GiB 14.26 0.21  34     up
                    TOTAL 273 TiB 183 TiB 183 TiB 6.4 MiB  567 GiB  90 TiB 67.17

ceph osd tree
ID  CLASS WEIGHT    TYPE NAME       STATUS REWEIGHT PRI-AFF
-12         0.05798 root default
-11         0.02899     host sn-m02
 23   hdd   0.00999         osd.23      up  1.00000 1.00000
 24   hdd   0.00999         osd.24      up  1.00000 1.00000
 25   hdd   0.00999         osd.25      up  1.00000 1.00000
-15         0.02899     host sn-m03
 26   hdd   0.00999         osd.26      up  1.00000 1.00000
 27   hdd   0.00999         osd.27      up  1.00000 1.00000
 28   hdd   0.00999         osd.28      up  1.00000 1.00000
 -6       272.87100 root ec
 -5        90.95700     host sn-s01
  3   hdd   9.09599         osd.3       up  1.00000 1.00000
  4   hdd   9.09599         osd.4       up  1.00000 1.00000
  5   hdd   9.09599         osd.5       up  1.00000 1.00000
  6   hdd   9.09599         osd.6       up  1.00000 1.00000
  7   hdd   9.09599         osd.7       up  1.00000 1.00000
  8   hdd   9.09599         osd.8       up  1.00000 1.00000
  9   hdd   9.09599         osd.9       up  1.00000 1.00000
 10   hdd   9.09599         osd.10      up  1.00000 1.00000
 11   hdd   9.09599         osd.11      up  1.00000 1.00000
 12   hdd   9.09599         osd.12      up  1.00000 1.00000
 -9        90.95700     host sn-s02
 13   hdd   9.09599         osd.13      up  1.00000 1.00000
 14   hdd   9.09599         osd.14      up  1.00000 1.00000
 15   hdd   9.09599         osd.15      up  1.00000 1.00000
 16   hdd   9.09599         osd.16      up  1.00000 1.00000
 17   hdd   9.09599         osd.17      up  1.00000 1.00000
 18   hdd   9.09599         osd.18      up  1.00000 1.00000
 19   hdd   9.09599         osd.19      up  1.00000 1.00000
 20   hdd   9.09599         osd.20      up  1.00000 1.00000
 21   hdd   9.09599         osd.21      up  1.00000 1.00000
 22   hdd   9.09599         osd.22      up  1.00000 1.00000
-17        90.95700     host sn-s03
 29   hdd   9.09599         osd.29      up  1.00000 1.00000
 30   hdd   9.09599         osd.30      up  1.00000 1.00000
 31   hdd   9.09599         osd.31      up  1.00000 1.00000
 32   hdd   9.09599         osd.32      up  1.00000 1.00000
 33   hdd   9.09599         osd.33      up  1.00000 1.00000
 34   hdd   9.09599         osd.34      up  1.00000 1.00000
 35   hdd   9.09599         osd.35      up  1.00000 1.00000
 36   hdd   9.09599         osd.36      up  1.00000 1.00000
 37   hdd   9.09599         osd.37      up  1.00000 1.00000
 38   hdd   9.09599         osd.38      up  1.00000 1.00000
 -1         0.08698 root meta
 -3         0.02899     host sn-m01
  0   hdd   0.00999         osd.0       up  1.00000 1.00000
  1   hdd   0.00999         osd.1       up  1.00000 1.00000
  2   hdd   0.00999         osd.2       up  1.00000 1.00000
-11         0.02899     host sn-m02
 23   hdd   0.00999         osd.23      up  1.00000 1.00000
 24   hdd   0.00999         osd.24      up  1.00000 1.00000
 25   hdd   0.00999         osd.25      up  1.00000 1.00000
-15         0.02899     host sn-m03
 26   hdd   0.00999         osd.26      up  1.00000 1.00000
 27   hdd   0.00999         osd.27      up  1.00000 1.00000
 28   hdd   0.00999         osd.28      up  1.00000 1.00000

OSD 7,8 are the issue OSD's sitting at 182,187 PG's where the others are all sitting at 170,171.

Am I hitting the same issue? Or is there something I can do to re balance these extra PG's across the rest of the OSD better?

Thanks




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux