Re: Usage of devices in SSD pool vary very much

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 1/5/19 4:17 PM, Kevin Olbrich wrote:
root@adminnode:~# ceph osd tree
ID  CLASS WEIGHT   TYPE NAME                     STATUS REWEIGHT PRI-AFF
  -1       30.82903 root default
-16       30.82903     datacenter dc01
-19       30.82903         pod dc01-agg01
-10       17.43365             rack dc01-rack02
  -4        7.20665                 host node1001
   0   hdd  0.90999                     osd.0         up  1.00000 1.00000
   1   hdd  0.90999                     osd.1         up  1.00000 1.00000
   5   hdd  0.90999                     osd.5         up  1.00000 1.00000
  29   hdd  0.90970                     osd.29        up  1.00000 1.00000
  32   hdd  0.90970                     osd.32      down        0 1.00000
  33   hdd  0.90970                     osd.33        up  1.00000 1.00000
   2   ssd  0.43700                     osd.2         up  1.00000 1.00000
   3   ssd  0.43700                     osd.3         up  1.00000 1.00000
   4   ssd  0.43700                     osd.4         up  1.00000 1.00000
  30   ssd  0.43660                     osd.30        up  1.00000 1.00000
  -7        6.29724                 host node1002
   9   hdd  0.90999                     osd.9         up  1.00000 1.00000
  10   hdd  0.90999                     osd.10        up  1.00000 1.00000
  11   hdd  0.90999                     osd.11        up  1.00000 1.00000
  12   hdd  0.90999                     osd.12        up  1.00000 1.00000
  35   hdd  0.90970                     osd.35        up  1.00000 1.00000
   6   ssd  0.43700                     osd.6         up  1.00000 1.00000
   7   ssd  0.43700                     osd.7         up  1.00000 1.00000
   8   ssd  0.43700                     osd.8         up  1.00000 1.00000
  31   ssd  0.43660                     osd.31        up  1.00000 1.00000
-28        2.18318                 host node1005
  34   ssd  0.43660                     osd.34        up  1.00000 1.00000
  36   ssd  0.87329                     osd.36        up  1.00000 1.00000
  37   ssd  0.87329                     osd.37        up  1.00000 1.00000
-29        1.74658                 host node1006
  42   ssd  0.87329                     osd.42        up  1.00000 1.00000
  43   ssd  0.87329                     osd.43        up  1.00000 1.00000
-11       13.39537             rack dc01-rack03
-22        5.38794                 host node1003
  17   hdd  0.90999                     osd.17        up  1.00000 1.00000
  18   hdd  0.90999                     osd.18        up  1.00000 1.00000
  24   hdd  0.90999                     osd.24        up  1.00000 1.00000
  26   hdd  0.90999                     osd.26        up  1.00000 1.00000
  13   ssd  0.43700                     osd.13        up  1.00000 1.00000
  14   ssd  0.43700                     osd.14        up  1.00000 1.00000
  15   ssd  0.43700                     osd.15        up  1.00000 1.00000
  16   ssd  0.43700                     osd.16        up  1.00000 1.00000
-25        5.38765                 host node1004
  23   hdd  0.90999                     osd.23        up  1.00000 1.00000
  25   hdd  0.90999                     osd.25        up  1.00000 1.00000
  27   hdd  0.90999                     osd.27        up  1.00000 1.00000
  28   hdd  0.90970                     osd.28        up  1.00000 1.00000
  19   ssd  0.43700                     osd.19        up  1.00000 1.00000
  20   ssd  0.43700                     osd.20        up  1.00000 1.00000
  21   ssd  0.43700                     osd.21        up  1.00000 1.00000
  22   ssd  0.43700                     osd.22        up  1.00000 1.00000
-30        2.61978                 host node1007
  38   ssd  0.43660                     osd.38        up  1.00000 1.00000
  39   ssd  0.43660                     osd.39        up  1.00000 1.00000
  40   ssd  0.87329                     osd.40        up  1.00000 1.00000
  41   ssd  0.87329                     osd.41        up  1.00000 1.00000

========================================================
root@adminnode:~# ceph osd df tree
ID  CLASS WEIGHT   REWEIGHT SIZE    USE     AVAIL   %USE  VAR  PGS
TYPE NAME
  -1       30.82903        - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00   -
root default
-16       30.82903        - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00   -
datacenter dc01
-19       30.82903        - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00   -
     pod dc01-agg01
-10       17.43365        - 16.5TiB 7.31TiB 9.21TiB 44.26 0.95   -
         rack dc01-rack02
  -4        7.20665        - 6.29TiB 2.76TiB 3.54TiB 43.83 0.94   -
             host node1001
   0   hdd  0.90999  1.00000  932GiB  356GiB  575GiB 38.22 0.82  95
                 osd.0
   1   hdd  0.90999  1.00000  932GiB  397GiB  534GiB 42.66 0.91 106
                 osd.1
   5   hdd  0.90999  1.00000  932GiB  284GiB  647GiB 30.50 0.65  76
                 osd.5
  29   hdd  0.90970  1.00000  932GiB  366GiB  566GiB 39.29 0.84  98
                 osd.29
  32   hdd  0.90970        0      0B      0B      0B     0    0   0
                 osd.32
  33   hdd  0.90970  1.00000  932GiB  369GiB  563GiB 39.57 0.85  99
                 osd.33
   2   ssd  0.43700  1.00000  447GiB  271GiB  176GiB 60.67 1.30  50
                 osd.2
   3   ssd  0.43700  1.00000  447GiB  249GiB  198GiB 55.62 1.19  58
                 osd.3
   4   ssd  0.43700  1.00000  447GiB  297GiB  150GiB 66.39 1.42  56
                 osd.4
  30   ssd  0.43660  1.00000  447GiB  236GiB  211GiB 52.85 1.13  48
                 osd.30
  -7        6.29724        - 6.29TiB 2.74TiB 3.55TiB 43.53 0.93   -
             host node1002
   9   hdd  0.90999  1.00000  932GiB  354GiB  578GiB 37.96 0.81  95
                 osd.9
  10   hdd  0.90999  1.00000  932GiB  357GiB  575GiB 38.28 0.82  96
                 osd.10
  11   hdd  0.90999  1.00000  932GiB  318GiB  613GiB 34.18 0.73  86
                 osd.11
  12   hdd  0.90999  1.00000  932GiB  373GiB  558GiB 40.09 0.86 100
                 osd.12
  35   hdd  0.90970  1.00000  932GiB  343GiB  588GiB 36.83 0.79  92
                 osd.35
   6   ssd  0.43700  1.00000  447GiB  269GiB  178GiB 60.20 1.29  60
                 osd.6
   7   ssd  0.43700  1.00000  447GiB  249GiB  198GiB 55.69 1.19  56
                 osd.7
   8   ssd  0.43700  1.00000  447GiB  286GiB  161GiB 63.95 1.37  56
                 osd.8
  31   ssd  0.43660  1.00000  447GiB  257GiB  190GiB 57.47 1.23  55
                 osd.31
-28        2.18318        - 2.18TiB  968GiB 1.24TiB 43.29 0.93   -
             host node1005
  34   ssd  0.43660  1.00000  447GiB  202GiB  245GiB 45.14 0.97  47
                 osd.34
  36   ssd  0.87329  1.00000  894GiB  405GiB  489GiB 45.28 0.97  91
                 osd.36
  37   ssd  0.87329  1.00000  894GiB  361GiB  533GiB 40.38 0.87  79
                 osd.37
-29        1.74658        - 1.75TiB  888GiB  900GiB 49.65 1.06   -
             host node1006
  42   ssd  0.87329  1.00000  894GiB  417GiB  477GiB 46.68 1.00  92
                 osd.42
  43   ssd  0.87329  1.00000  894GiB  471GiB  424GiB 52.63 1.13  90
                 osd.43
-11       13.39537        - 13.4TiB 6.64TiB 6.75TiB 49.60 1.06   -
         rack dc01-rack03
-22        5.38794        - 5.39TiB 2.70TiB 2.69TiB 50.14 1.07   -
             host node1003
  17   hdd  0.90999  1.00000  932GiB  371GiB  560GiB 39.83 0.85 100
                 osd.17
  18   hdd  0.90999  1.00000  932GiB  390GiB  542GiB 41.82 0.90 105
                 osd.18
  24   hdd  0.90999  1.00000  932GiB  352GiB  580GiB 37.77 0.81  94
                 osd.24
  26   hdd  0.90999  1.00000  932GiB  387GiB  545GiB 41.54 0.89 104
                 osd.26
  13   ssd  0.43700  1.00000  447GiB  319GiB  128GiB 71.32 1.53  77
                 osd.13
  14   ssd  0.43700  1.00000  447GiB  303GiB  144GiB 67.76 1.45  70
                 osd.14
  15   ssd  0.43700  1.00000  447GiB  361GiB 86.4GiB 80.67 1.73  77
                 osd.15
  16   ssd  0.43700  1.00000  447GiB  283GiB  164GiB 63.29 1.36  71
                 osd.16
-25        5.38765        - 5.39TiB 2.83TiB 2.56TiB 52.55 1.13   -
             host node1004
  23   hdd  0.90999  1.00000  932GiB  382GiB  549GiB 41.05 0.88 102
                 osd.23
  25   hdd  0.90999  1.00000  932GiB  412GiB  520GiB 44.20 0.95 111
                 osd.25
  27   hdd  0.90999  1.00000  932GiB  385GiB  546GiB 41.36 0.89 103
                 osd.27
  28   hdd  0.90970  1.00000  932GiB  462GiB  469GiB 49.64 1.06 124
                 osd.28
  19   ssd  0.43700  1.00000  447GiB  314GiB  133GiB 70.22 1.51  75
                 osd.19
  20   ssd  0.43700  1.00000  447GiB  327GiB  120GiB 73.06 1.57  76
                 osd.20
  21   ssd  0.43700  1.00000  447GiB  324GiB  123GiB 72.45 1.55  77
                 osd.21
  22   ssd  0.43700  1.00000  447GiB  292GiB  156GiB 65.21 1.40  68
                 osd.22
-30        2.61978        - 2.62TiB 1.11TiB 1.51TiB 42.43 0.91   -
             host node1007
  38   ssd  0.43660  1.00000  447GiB  165GiB  283GiB 36.82 0.79  46
                 osd.38
  39   ssd  0.43660  1.00000  447GiB  156GiB  292GiB 34.79 0.75  42
                 osd.39
  40   ssd  0.87329  1.00000  894GiB  429GiB  466GiB 47.94 1.03  98
                 osd.40
  41   ssd  0.87329  1.00000  894GiB  389GiB  505GiB 43.55 0.93 103
                 osd.41
                       TOTAL 29.9TiB 14.0TiB 16.0TiB 46.65
MIN/MAX VAR: 0.65/1.73  STDDEV: 13.30

=============================================================
root@adminnode:~# ceph df && ceph -v
GLOBAL:
     SIZE        AVAIL       RAW USED     %RAW USED
     29.9TiB     16.0TiB      14.0TiB         46.65
POOLS:
     NAME                  ID     USED        %USED     MAX AVAIL     OBJECTS
     rbd_vms_ssd           2       986GiB     49.83        993GiB      262606
     rbd_vms_hdd           3      3.76TiB     48.94       3.92TiB      992255
     rbd_vms_ssd_01        4       372KiB         0        662GiB         148
     rbd_vms_ssd_01_ec     6      2.85TiB     68.81       1.29TiB      770506

ceph version 12.2.8 (ae699615bac534ea496ee965ac6192cb7e0e07c0) luminous (stable)

Looks good. You should always delete your down osd's from crush map before replace it. After delete this osd try balancer again.



k

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux