Unbalanced OSDs when pg_autoscale enabled

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

I have a 9 node cluster running *Pacific 16.2.10*. OSDs live on 9 of the
nodes with each one having 4 x 1.8T ssd and 8 x 10.9T hdd for a total of
108 OSDs. We create three crush roots as belows.

1. The hdds (8x9=72) of all nodes form a large crush root, which is used as
a data pool, and object storage and cephfs share this crush root.
2. Take 3 ssds from the 4 ssds of each node as rbd block storage.
3. An ssd on each remaining node is used as an index pool for cephfs and
object storage.

[root@node01 smd]# ceph osd treeID    CLASS  WEIGHT     TYPE NAME
                                       STATUS  REWEIGHT  PRI-AFF
                                -92          15.71910  root
root-1c31624a-ad18-445e-8e42-86b71c1fd76f
                                         -112           1.74657
host node01-fa2cdf3e-7212-4b5f-b62a-3ab1e803547f
                                            13    ssd    1.74657
   osd.13                                            up   1.00000
1.00000                                      -103           1.74657
  host node02-4e232f27-fe4b-4d0e-bd2a-67d5006a0cdd
                                              34    ssd    1.74657
     osd.34                                            up   1.00000
1.00000                                      -109           1.74657
  host node03-3ae63d7a-9f65-4bea-b2ba-ff3fe342753d
                                              28    ssd    1.74657
     osd.28                                            up   1.00000
1.00000                                      -118           1.74657
  host node04-37a3f92a-f6d8-41f9-a774-3069fc2f50b8
                                              54    ssd    1.74657
     osd.54                                            up   1.00000
1.00000                                      -106           1.74657
  host node05-f667fa27-cc13-4b93-ad56-5dc4c31ffd77
                                              53    ssd    1.74657
     osd.53                                            up   1.00000
1.00000                                       -91           1.74657
  host node06-3808c8f6-8e10-47c7-8456-62c1e0e800ed
                                              61    ssd    1.74657
     osd.61                                            up   1.00000
1.00000                                       -97           1.74657
  host node07-78216b0d-0999-44e8-905d-8737a5f6f51f
                                              50    ssd    1.74657
     osd.50                                            up   1.00000
1.00000                                      -115           1.74657
  host node08-947bd556-fb06-497d-8f2c-c4a679d2b06f
                                              86    ssd    1.74657
     osd.86                                            up   1.00000
1.00000                                      -100           1.74657
  host node09-d9ae9046-0716-454f-ba0c-b03cf9986ba8
        85    ssd    1.74657          osd.85
                 up   1.00000  1.00000

-38         785.80701  root root-6041a4dc-7c9a-44ed-999c-a847cca81012
                                 -85          87.31189      host
node01-e5646053-2cf8-4ba5-90d5-bb1a63b1234c
  1    hdd   10.91399          osd.1
          up   1.00000  1.00000  22    hdd   10.91399          osd.22
                                          up   0.90002  1.00000  31
hdd   10.91399          osd.31
   up   1.00000  1.00000  51    hdd   10.91399          osd.51
                                   up   1.00000  1.00000  60    hdd
10.91399          osd.60                                            up
  1.00000  1.00000  70    hdd   10.91399          osd.70
                             up   1.00000  1.00000  78    hdd
10.91399          osd.78                                            up
  1.00000  1.00000                                        96    hdd
10.91399          osd.96                                            up
  1.00000  1.00000                                       -37
87.31189      host node02-be9925fd-60de-4147-81eb-720d7145715f
                                                           9    hdd
10.91399          osd.9                                             up
  1.00000  1.00000                                        19    hdd
10.91399          osd.19                                            up
  1.00000  1.00000                                        29    hdd
10.91399          osd.29                                            up
  1.00000  1.00000                                        47    hdd
10.91399          osd.47                                            up
  1.00000  1.00000                                        56    hdd
10.91399          osd.56                                            up
  1.00000  1.00000                                        65    hdd
10.91399          osd.65                                            up
  1.00000  1.00000                                        88    hdd
10.91399          osd.88                                            up
  1.00000  1.00000                                        98    hdd
10.91399          osd.98                                            up
  1.00000  1.00000                                       -52
87.31189      host node03-7828653d-6033-4e88-92b0-d8709b0ab218
                                                           2    hdd
10.91399          osd.2                                             up
  1.00000  1.00000                                        30    hdd
10.91399          osd.30                                            up
  1.00000  1.00000                                        40    hdd
10.91399          osd.40                                            up
  1.00000  1.00000                                        48    hdd
10.91399          osd.48                                            up
  1.00000  1.00000                                        58    hdd
10.91399          osd.58                                            up
  1.00000  1.00000                                        74    hdd
10.91399          osd.74                                            up
  1.00000  1.00000  83    hdd   10.91399          osd.83
                             up   1.00000  1.00000  92    hdd
10.91399          osd.92                                            up
  1.00000  1.00000 -46          87.31189      host
node04-e986c3fc-a21b-44ff-9b02-b60b82ee63d7
  12    hdd   10.91399          osd.12
           up   1.00000  1.00000   23    hdd   10.91399
osd.23                                            up   1.00000
1.00000  32    hdd   10.91399          osd.32
                  up   1.00000  1.00000  43    hdd   10.91399
osd.43                                            up   1.00000
1.00000  52    hdd   10.91399          osd.52
                  up   1.00000  1.00000  71    hdd   10.91399
osd.71                                            up   1.00000
1.00000                                        95    hdd   10.91399
      osd.95                                            up   1.00000
1.00000  104    hdd   10.91399          osd.104
                   up   1.00000  1.00000 -88          87.31189
host node05-fe31d85f-b3b9-4393-b24b-030dbcdfacea
       3    hdd   10.91399          osd.3
               up   1.00000  1.00000  24    hdd   10.91399
osd.24                                            up   1.00000
1.00000  33    hdd   10.91399          osd.33
                  up   1.00000  1.00000  45    hdd   10.91399
osd.45                                            up   1.00000
1.00000  69    hdd   10.91399          osd.69
                  up   1.00000  1.00000  79    hdd   10.91399
osd.79                                            up   1.00000
1.00000                                        89    hdd   10.91399
      osd.89                                            up   1.00000
1.00000                                        99    hdd   10.91399
      osd.99                                            up   1.00000
1.00000                                       -55          87.31189
  host node06-6f16ba4b-0082-472a-b243-b1a058070918
                                               5    hdd   10.91399
     osd.5                                             up   1.00000
1.00000                                        15    hdd   10.91399
      osd.15                                            up   1.00000
1.00000                                        25    hdd   10.91399
      osd.25                                            up   1.00000
1.00000                                        44    hdd   10.91399
      osd.44                                            up   1.00000
1.00000                                        63    hdd   10.91399
      osd.63                                            up   1.00000
1.00000                                        72    hdd   10.91399
      osd.72                                            up   1.00000
1.00000                                        81    hdd   10.91399
      osd.81                                            up   1.00000
1.00000                                        91    hdd   10.91399
      osd.91                                            up   1.00000
1.00000                                       -43          87.31189
  host node07-5dee846a-2814-4e04-bcfd-ff689d49795c
                                               7    hdd   10.91399
     osd.7                                             up   1.00000
1.00000                                        37    hdd   10.91399
      osd.37                                            up   1.00000
1.00000                                        46    hdd   10.91399
      osd.46                                            up   1.00000
1.00000                                        55    hdd   10.91399
      osd.55                                            up   1.00000
1.00000                                        64    hdd   10.91399
      osd.64                                            up   1.00000
1.00000  84    hdd   10.91399          osd.84
                  up   1.00000  1.00000  94    hdd   10.91399
osd.94                                            up   1.00000
1.00000 102    hdd   10.91399          osd.102
                  up   1.00000  1.00000  -58          87.31189
host node08-2d6b7ab3-2067-4e94-b77c-24d6e626e396
        6    hdd   10.91399          osd.6
                up   1.00000  1.00000  18    hdd   10.91399
osd.18                                            up   1.00000
1.00000  41    hdd   10.91399          osd.41
                  up   1.00000  1.00000  49    hdd   10.91399
osd.49                                            up   1.00000
1.00000  59    hdd   10.91399          osd.59
                  up   1.00000  1.00000
        68    hdd   10.91399          osd.68
                 up   1.00000  1.00000   97    hdd   10.91399
osd.97                                            up   1.00000
1.00000 105    hdd   10.91399          osd.105
                  up   1.00000  1.00000 -49          87.31189
host node09-e6bf0642-e3b4-48f0-9d0e-3d87ceacced8
       8    hdd   10.91399          osd.8
               up   1.00000  1.00000  17    hdd   10.91399
osd.17                                            up   1.00000
1.00000  27    hdd   10.91399          osd.27
                  up   1.00000  1.00000  38    hdd   10.91399
osd.38                                            up   1.00000
1.00000  57    hdd   10.91399          osd.57
                  up   1.00000  1.00000  67    hdd   10.91399
osd.67                                            up   1.00000
1.00000  93    hdd   10.91399          osd.93
                  up   1.00000  1.00000 103    hdd   10.91399
osd.103                                           up   1.00000
1.00000  -8          47.16115  root
root-ea7a1878-722e-49d5-8a91-c618a6aefe29
    -13           5.24013      host
node01-4c465825-6bd1-42a5-b087-51a82cb2865c
  0    ssd    1.74657          osd.0
          up   0.95001  1.00000  11    ssd    1.74699          osd.11
                                          up   1.00000  1.00000  16
ssd    1.74657          osd.16
   up   1.00000  1.00000 -22           5.24013      host
node02-0e3e418d-9129-44ee-8453-53680171270e
  4    ssd    1.74657          osd.4
          up   1.00000  1.00000  14    ssd    1.74657          osd.14
                                          up   0.95001  1.00000  82
ssd    1.74699          osd.82
   up   1.00000  1.00000 -28           5.24013      host
node03-274397b4-8559-4a17-a9eb-1c16920ad432
 10    ssd    1.74657          osd.10
          up   1.00000  1.00000  76    ssd    1.74657          osd.76
                                          up   1.00000  1.00000 100
ssd    1.74699          osd.100
   up   0.95001  1.00000 -31           5.24013      host
node04-df05c3cc-08ef-426e-85b8-088cb8c1b4e2
 20    ssd    1.74657          osd.20
          up   1.00000  1.00000  39    ssd    1.74657          osd.39
                                          up   0.95001  1.00000  90
ssd    1.74699          osd.90
   up   0.95001  1.00000 -25           5.24013      host
node05-d07e7e92-8290-49ba-b09c-ad34bddb1eae
 26    ssd    1.74657          osd.26
          up   0.95001  1.00000   62    ssd    1.74699          osd.62
                                           up   1.00000  1.00000  73
 ssd    1.74657          osd.73
    up   1.00000  1.00000 -16           5.24013      host
node06-a7a40371-a128-4ab1-90a2-62c99e040036
 42    ssd    1.74657          osd.42
          up   0.95001  1.00000  75    ssd    1.74657          osd.75
                                          up   0.95001  1.00000 107
ssd    1.74699          osd.107
   up   1.00000  1.00000  -7           5.24013      host
node07-971dee6e-dec4-4c0a-86d1-54d0b23832bd
 21    ssd    1.74699          osd.21
          up   1.00000  1.00000  66    ssd    1.74657          osd.66
                                          up   0.95001  1.00000  77
ssd    1.74657          osd.77
   up   1.00000  1.00000 -19           5.24013      host
node08-9a47282b-9530-4cc7-9e29-7f6c0b2f5184
 35    ssd    1.74699          osd.35
          up   1.00000  1.00000  80    ssd    1.74657          osd.80
                                          up   1.00000  1.00000 101
ssd    1.74657          osd.101
   up   1.00000  1.00000 -34           5.24013      host
node09-f09a8013-2426-4ced-b9c7-02c06ca9d6fc
  36    ssd    1.74699          osd.36
           up   0.95001  1.00000  87    ssd    1.74657          osd.87
                                           up   1.00000  1.00000 106
 ssd    1.74657          osd.106
    up   1.00000  1.00000

After ceph has been running normally for a period of time, I found that the
distribution of osd data is seriously uneven, and autoscale PGs were turned
on in my envs. My hdds are identical, but utilization ranges from 6.14% to
74.35%

root@node09:/# ceph osd df | grep hdd ID   CLASS  WEIGHT    REWEIGHT
SIZE     RAW USE  DATA      OMAP     META     AVAIL    %USE   VAR
PGS  STATUS  1    hdd  10.91399   1.00000   11 TiB  5.3 TiB   5.2 TiB
 15 KiB   53 GiB  5.7 TiB  48.15  1.38  174      up 22    hdd
10.91399   0.90002   11 TiB  8.1 TiB   8.0 TiB   12 KiB   71 GiB  2.8
TiB  74.35  2.13  160      up 31    hdd  10.91399   1.00000   11 TiB
3.0 TiB   2.9 TiB   13 KiB   27 GiB  8.0 TiB  27.14  0.78  179      up
51    hdd  10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB    9 KiB
36 GiB  7.2 TiB  34.12  0.98  173      up 60    hdd  10.91399
1.00000   11 TiB  973 GiB   962 GiB   14 KiB   11 GiB   10 TiB   8.71
0.25  167      up 70    hdd  10.91399   1.00000   11 TiB  2.0 TiB
2.0 TiB   11 KiB  3.9 GiB  8.9 TiB  18.42  0.53  173      up 78    hdd
 10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   12 KiB   28 GiB  7.9
TiB  27.18  0.78  173      up 96    hdd  10.91399   1.00000   11 TiB
3.7 TiB   3.7 TiB   10 KiB   36 GiB  7.2 TiB  34.15  0.98  176      up
 9    hdd  10.91399   1.00000   11 TiB  2.8 TiB   2.8 TiB   16 KiB
13 GiB  8.1 TiB  25.38  0.73  178      up 19    hdd  10.91399
1.00000   11 TiB  3.8 TiB   3.8 TiB   13 KiB   43 GiB  7.1 TiB  35.04
1.00  177      up 29    hdd  10.91399   1.00000   11 TiB  5.1 TiB
5.0 TiB   14 KiB   38 GiB  5.9 TiB  46.38  1.33  182      up 47    hdd
 10.91399   1.00000   11 TiB  686 GiB   684 GiB    9 KiB  1.4 GiB   10
TiB   6.14  0.18  172      up 56    hdd  10.91399   1.00000   11 TiB
3.7 TiB   3.7 TiB   12 KiB   36 GiB  7.2 TiB  34.16  0.98  174      up
65    hdd  10.91399   1.00000   11 TiB  5.9 TiB   5.9 TiB    7 KiB
54 GiB  5.0 TiB  54.27  1.55  183      up 88    hdd  10.91399
1.00000   11 TiB  4.5 TiB   4.4 TiB   12 KiB   44 GiB  6.4 TiB  41.17
1.18  173      up 98    hdd  10.91399   1.00000   11 TiB  7.3 TiB
7.2 TiB   13 KiB   56 GiB  3.7 TiB  66.49  1.90  179      up  2    hdd
 10.91399   1.00000   11 TiB  1.5 TiB   1.5 TiB   12 KiB   18 GiB  9.4
TiB  14.02  0.40  177      up 30    hdd  10.91399   1.00000   11 TiB
4.4 TiB   4.4 TiB   10 KiB   37 GiB  6.5 TiB  40.29  1.15  180      up
40    hdd  10.91399   1.00000   11 TiB  1.4 TiB   1.4 TiB   10 KiB
11 GiB  9.5 TiB  13.14  0.38  172      up 48    hdd  10.91399
1.00000   11 TiB  1.4 TiB   1.4 TiB   12 KiB   11 GiB  9.5 TiB  13.17
0.38  168      up 58    hdd  10.91399   1.00000   11 TiB  7.5 TiB
7.4 TiB   14 KiB   70 GiB  3.5 TiB  68.31  1.96  182      up 74    hdd
 10.91399   1.00000   11 TiB  6.0 TiB   6.0 TiB   10 KiB   61 GiB  4.9
TiB  55.15  1.58  181      up 83    hdd  10.91399   1.00000   11 TiB
2.2 TiB   2.2 TiB   22 KiB   19 GiB  8.7 TiB  20.13  0.58  174      up
92    hdd  10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   11 KiB
27 GiB  8.0 TiB  27.15  0.78  171      up 12    hdd  10.91399
1.00000   11 TiB  5.1 TiB   5.0 TiB    9 KiB   38 GiB  5.9 TiB  46.37
1.33  176      up 23    hdd  10.91399   1.00000   11 TiB  4.4 TiB
4.4 TiB   22 KiB   37 GiB  6.5 TiB  40.24  1.15  173      up 32    hdd
 10.91399   1.00000   11 TiB  2.2 TiB   2.2 TiB   10 KiB   19 GiB  8.7
TiB  20.13  0.58  176      up 43    hdd  10.91399   1.00000   11 TiB
2.0 TiB   2.0 TiB   10 KiB  4.4 GiB  8.9 TiB  18.41  0.53  176      up
52    hdd  10.91399   1.00000   11 TiB  3.6 TiB   3.6 TiB   12 KiB
29 GiB  7.3 TiB  33.27  0.95  171      up 71    hdd  10.91399
1.00000   11 TiB  3.0 TiB   2.9 TiB   15 KiB   27 GiB  8.0 TiB  27.13
0.78  170      up 95    hdd  10.91399   1.00000   11 TiB  5.4 TiB
5.3 TiB   10 KiB   60 GiB  5.6 TiB  49.04  1.40  173      up104    hdd
 10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB    8 KiB   36 GiB  7.2
TiB  34.13  0.98  171      up  3    hdd  10.91399   1.00000   11 TiB
2.2 TiB   2.2 TiB    9 KiB   19 GiB  8.7 TiB  20.14  0.58  173      up
24    hdd  10.91399   1.00000   11 TiB  2.2 TiB   2.2 TiB    9 KiB
19 GiB  8.7 TiB  20.12  0.58  170      up 33    hdd  10.91399
1.00000   11 TiB  6.8 TiB   6.7 TiB   14 KiB   69 GiB  4.1 TiB  62.15
1.78  181      up 45    hdd  10.91399   1.00000   11 TiB  2.9 TiB
2.8 TiB   10 KiB   20 GiB  8.0 TiB  26.27  0.75  178      up 69    hdd
 10.91399   1.00000   11 TiB  2.9 TiB   2.8 TiB   15 KiB   20 GiB  8.0
TiB  26.25  0.75  173      up 79    hdd  10.91399   1.00000   11 TiB
5.9 TiB   5.9 TiB   28 KiB   54 GiB  5.0 TiB  54.25  1.55  178      up
89    hdd  10.91399   1.00000   11 TiB  2.1 TiB   2.1 TiB   16 KiB
11 GiB  8.8 TiB  19.25  0.55  173      up 99    hdd  10.91399
1.00000   11 TiB  4.5 TiB   4.4 TiB   13 KiB   44 GiB  6.4 TiB  41.14
1.18  174      up  5    hdd  10.91399   1.00000   11 TiB  2.2 TiB
2.2 TiB   14 KiB   19 GiB  8.7 TiB  20.12  0.58  172      up 15    hdd
 10.91399   1.00000   11 TiB  7.3 TiB   7.2 TiB    7 KiB   56 GiB  3.7
TiB  66.48  1.90  177      up 25    hdd  10.91399   1.00000   11 TiB
1.4 TiB   1.4 TiB   11 KiB   10 GiB  9.5 TiB  13.13  0.38  176      up
44    hdd  10.91399   1.00000   11 TiB  3.1 TiB   3.0 TiB    9 KiB
34 GiB  7.9 TiB  28.03  0.80  179      up 63    hdd  10.91399
1.00000   11 TiB  2.9 TiB   2.8 TiB   10 KiB   20 GiB  8.0 TiB  26.26
0.75  178      up 72    hdd  10.91399   1.00000   11 TiB  3.7 TiB
3.7 TiB   12 KiB   36 GiB  7.2 TiB  34.13  0.98  180      up 81    hdd
 10.91399   1.00000   11 TiB  6.7 TiB   6.6 TiB   11 KiB   62 GiB  4.2
TiB  61.30  1.76  181      up 91    hdd  10.91399   1.00000   11 TiB
2.2 TiB   2.2 TiB   16 KiB   19 GiB  8.7 TiB  20.11  0.58  176      up
 7    hdd  10.91399   1.00000   11 TiB  4.6 TiB   4.5 TiB   13 KiB
52 GiB  6.3 TiB  42.03  1.20  180      up 37    hdd  10.91399
1.00000   11 TiB  5.3 TiB   5.2 TiB   14 KiB   52 GiB  5.7 TiB  48.16
1.38  182      up 46    hdd  10.91399   1.00000   11 TiB  784 GiB
775 GiB   14 KiB  8.8 GiB   10 TiB   7.01  0.20  173      up 55    hdd
 10.91399   1.00000   11 TiB  5.9 TiB   5.9 TiB   20 KiB   54 GiB  5.0
TiB  54.25  1.55  177      up 64    hdd  10.91399   1.00000   11 TiB
3.6 TiB   3.6 TiB   18 KiB   28 GiB  7.3 TiB  33.23  0.95  176      up
84    hdd  10.91399   1.00000   11 TiB  784 GiB   775 GiB   10 KiB
9.0 GiB   10 TiB   7.01  0.20  176      up 94    hdd  10.91399
1.00000   11 TiB  2.1 TiB   2.1 TiB    8 KiB   12 GiB  8.8 TiB  19.28
0.55  169      up102    hdd  10.91399   1.00000   11 TiB  3.0 TiB
2.9 TiB   10 KiB   28 GiB  8.0 TiB  27.13  0.78  170      up  6    hdd
 10.91399   1.00000   11 TiB  4.5 TiB   4.4 TiB   23 KiB   44 GiB  6.4
TiB  41.14  1.18  181      up 18    hdd  10.91399   1.00000   11 TiB
6.0 TiB   6.0 TiB    9 KiB   61 GiB  4.9 TiB  55.15  1.58  183      up
41    hdd  10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   15 KiB
27 GiB  8.0 TiB  27.12  0.78  179      up 49    hdd  10.91399
1.00000   11 TiB  785 GiB   775 GiB   19 KiB  9.5 GiB   10 TiB   7.02
0.20  176      up 59    hdd  10.91399   1.00000   11 TiB  2.2 TiB
2.2 TiB    8 KiB   19 GiB  8.7 TiB  20.12  0.58  178      up 68    hdd
 10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB   13 KiB   36 GiB  7.2
TiB  34.15  0.98  172      up 97    hdd  10.91399   1.00000   11 TiB
3.0 TiB   2.9 TiB   15 KiB   27 GiB  8.0 TiB  27.13  0.78  178
up105    hdd  10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB   13 KiB
 36 GiB  7.2 TiB  34.14  0.98  179      up  8    hdd  10.91399
1.00000   11 TiB  6.6 TiB   6.5 TiB   13 KiB   55 GiB  4.3 TiB  60.40
1.73  181      up 17    hdd  10.91399   1.00000   11 TiB  4.6 TiB
4.5 TiB   14 KiB   52 GiB  6.3 TiB  42.03  1.20  177      up 27    hdd
 10.91399   1.00000   11 TiB  783 GiB   774 GiB    8 KiB  9.4 GiB   10
TiB   7.01  0.20  176      up 38    hdd  10.91399   1.00000   11 TiB
2.2 TiB   2.2 TiB   15 KiB   19 GiB  8.7 TiB  20.11  0.58  174      up
57    hdd  10.91399   1.00000   11 TiB  2.3 TiB   2.3 TiB   12 KiB
26 GiB  8.6 TiB  21.01  0.60  177      up 67    hdd  10.91399
1.00000   11 TiB  4.5 TiB   4.4 TiB    9 KiB   44 GiB  6.4 TiB  41.14
1.18  176      up 93    hdd  10.91399   1.00000   11 TiB  3.1 TiB
3.0 TiB    7 KiB   35 GiB  7.9 TiB  28.01  0.80  171      up103    hdd
 10.91399   1.00000   11 TiB  4.2 TiB   4.2 TiB   10 KiB   22 GiB  6.7
TiB  38.52  1.10  176      up

three pools (deeproute-replica-hdd-pool,
os-dsglczutvqsgowpz.rgw.buckets.data, and cephfs-replicated-pool ) are set
to use the same device class hdd. I noticed that the *effective ratio* of
the three pools is very different, not sure if it will affect the data
rebalancing.

[root@node01 smd]# ceph osd pool autoscale-statusPOOL
                    SIZE  TARGET SIZE                RATE  RAW
CAPACITY   RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW
PG_NUM  AUTOSCALE  BULK   device_health_metrics
278.5M                              3.0        48289G  0.0000
                        1.0       1              on         False
deeproute-replica-hdd-pool              12231M
     3.0        785.8T  0.9901       50.0000           0.9901   1.0
4096              on         False  deeproute-replica-ssd-pool
     12333G                              3.0        48289G  0.9901
  50.0000           0.9901   1.0    1024              on         False
 .rgw.root                                5831
      3.0        48289G  0.0099        0.5000           0.0099   1.0
    8              on         False  default.rgw.log
        182                               3.0        48289G  0.0000
                              1.0      32              on
False  default.rgw.control                         0
            3.0        48289G  0.0000
1.0      32              on         False  default.rgw.meta
                0                               3.0        48289G
0.0000                                  4.0       8              on
     False  os-dsglczutvqsgowpz.rgw.control             0
                 3.0        16096G  0.1667        0.5000
0.1667   1.0      64              on         False
os-dsglczutvqsgowpz.rgw.meta            99.81k
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  64              on         False
os-dsglczutvqsgowpz.rgw.buckets.index   59437M
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  32              on         False
os-dsglczutvqsgowpz.rgw.buckets.non-ec  498.7M
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  32              on         False  os-dsglczutvqsgowpz.rgw.log
      1331M                              3.0        16096G  0.1667
   0.5000           0.1667   1.0      32              on         False
 os-dsglczutvqsgowpz.rgw.buckets.data    143.1T
1.3333333730697632        785.8T  0.2429        0.5000
0.0099   1.0      32              on         False  cephfs-metadata
                      3161M                              3.0
16096G  0.0006                                  4.0      32
  on         False  cephfs-replicated-pool                  21862G
                         3.0        785.8T  0.0815
             1.0      32              on         False  .nfs
                         84599                               3.0
 48289G  0.0000                                  1.0      32
   on         False  os-dsglczutvqsgowpz.rgw.otp                 0
                          3.0        16096G  0.1667        0.5000

[root@node01 smd]# ceph df--- RAW STORAGE ---CLASS     SIZE    AVAIL
  USED  RAW USED  %RAW USEDhdd    786 TiB  545 TiB  240 TiB   240 TiB
    30.60ssd     63 TiB   25 TiB   38 TiB    38 TiB      59.95TOTAL
849 TiB  571 TiB  278 TiB   278 TiB      32.77 --- POOLS ---POOL
                             ID   PGS   STORED  OBJECTS     USED
%USED  MAX AVAILdevice_health_metrics                    1     1  280
MiB      135  841 MiB   0.02    1.8 TiBdeeproute-replica-hdd-pool
        11  4096   12 GiB    9.44k   36 GiB   0.02     69
TiBdeeproute-replica-ssd-pool              12  1024   12 TiB    3.43M
 37 TiB  87.62    1.8 TiB.rgw.root                               25
 8  5.7 KiB       20  228 KiB      0    1.8 TiBdefault.rgw.log
                26    32    182 B        2   24 KiB      0    1.8
TiBdefault.rgw.control                     27    32      0 B        8
    0 B      0    1.8 TiBdefault.rgw.meta                        28
 8      0 B        0      0 B      0    1.8
TiBos-dsglczutvqsgowpz.rgw.control         29    64      0 B        8
    0 B      0    4.9 TiBos-dsglczutvqsgowpz.rgw.meta            30
64  100 KiB      456  4.3 MiB      0    4.9
TiBos-dsglczutvqsgowpz.rgw.buckets.index   31    32   58 GiB   10.40k
174 GiB   1.15    4.9 TiBos-dsglczutvqsgowpz.rgw.buckets.non-ec  32
32  500 MiB  247.78k  4.3 GiB   0.03    4.9
TiBos-dsglczutvqsgowpz.rgw.log             33    32  689 MiB      432
2.0 GiB   0.01    4.9 TiBos-dsglczutvqsgowpz.rgw.buckets.data    34
32  130 TiB  134.96M  174 TiB  45.69    155 TiBcephfs-metadata
                35    32  3.1 GiB    1.67M  9.3 GiB   0.06    4.9
TiBcephfs-replicated-pool                  36    32   21 TiB    8.25M
 64 TiB  23.79     69 TiB.nfs                                    37
32   89 KiB        9  341 KiB      0    1.8
TiBos-dsglczutvqsgowpz.rgw.otp             38     8      0 B        0
    0 B      0    4.9 TiB

[root@node01 smd]# ceph -s
  cluster:
    id:     f7f1c8ba-f793-436b-bb73-0964108a30c1
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum a,b,c (age 7w)
    mgr: a(active, since 2w), standbys: b
    mds: 1/1 daemons up, 1 hot standby
    osd: 108 osds: 108 up (since 5h), 108 in (since 4M); 1 remapped pgs
    rgw: 9 daemons active (9 hosts, 1 zones)

  data:
    volumes: 1/1 healthy
    pools:   17 pools, 5561 pgs
    objects: 148.37M objects, 162 TiB
    usage:   276 TiB used, 572 TiB / 849 TiB avail
    pgs:     2786556/1119434807 objects misplaced (0.249%)
             5559 active+clean
             1    active+remapped+backfilling
             1    active+clean+scrubbing+deep

  io:
    client:   212 MiB/s rd, 27 MiB/s wr, 4.04k op/s rd, 2.01k op/s wr
    recovery: 4.0 MiB/s, 4 objects/s

  progress:
    Global Recovery Event (2d)
      [===========================.] (remaining: 39m)


So my question is, how can I adjust to make the osd data more evenly
distributed? Thanks!

Best Regards,
Liang Zheng
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux