Re: Octopus - unbalanced OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Anyone an idea? :)

On Fri, Apr 16, 2021 at 3:09 PM Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
>
> Hello List,
>
> any ideas why my OSDs are that unbalanced ?
>
> root@ceph01:~# ceph -s
>   cluster:
>     id:     5436dd5d-83d4-4dc8-a93b-60ab5db145df
>     health: HEALTH_WARN
>             1 nearfull osd(s)
>             4 pool(s) nearfull
>
>   services:
>     mon: 3 daemons, quorum ceph03,ceph01,ceph02 (age 2w)
>     mgr: ceph03(active, since 4M), standbys: ceph02.jwvivm
>     mds: backup:1 {0=backup.ceph06.hdjehi=up:active} 3 up:standby
>     osd: 56 osds: 56 up (since 29h), 56 in (since 3d)
>
>   task status:
>     scrub status:
>         mds.backup.ceph06.hdjehi: idle
>
>   data:
>     pools:   4 pools, 1185 pgs
>     objects: 24.29M objects, 44 TiB
>     usage:   151 TiB used, 55 TiB / 206 TiB avail
>     pgs:     675 active+clean
>              476 active+clean+snaptrim_wait
>              30  active+clean+snaptrim
>              4   active+clean+scrubbing+deep
>
> root@ceph01:~# ceph osd df tree
> ID   CLASS  WEIGHT     REWEIGHT  SIZE     RAW USE  DATA     OMAP
> META     AVAIL    %USE   VAR   PGS  STATUS  TYPE NAME
>  -1         206.79979         -  206 TiB  151 TiB  151 TiB   36 GiB
> 503 GiB   55 TiB  73.23  1.00    -          root default
>  -2          28.89995         -   29 TiB   20 TiB   20 TiB  5.5 GiB
> 74 GiB  8.9 TiB  69.19  0.94    -              host ceph01
>   0    hdd    2.70000   1.00000  2.7 TiB  1.8 TiB  1.8 TiB  590 MiB
> 6.9 GiB  908 GiB  66.81  0.91   44      up          osd.0
>   1    hdd    2.70000   1.00000  2.7 TiB  1.6 TiB  1.6 TiB  411 MiB
> 6.5 GiB  1.1 TiB  60.43  0.83   39      up          osd.1
>   4    hdd    2.70000   1.00000  2.7 TiB  1.8 TiB  1.8 TiB  501 MiB
> 6.8 GiB  898 GiB  67.15  0.92   43      up          osd.4
>   8    hdd    2.70000   1.00000  2.7 TiB  2.0 TiB  2.0 TiB  453 MiB
> 7.0 GiB  700 GiB  74.39  1.02   47      up          osd.8
>  11    hdd    1.70000   1.00000  1.7 TiB  1.3 TiB  1.3 TiB  356 MiB
> 5.6 GiB  433 GiB  75.39  1.03   31      up          osd.11
>  12    hdd    2.70000   1.00000  2.7 TiB  2.1 TiB  2.1 TiB  471 MiB
> 7.0 GiB  591 GiB  78.40  1.07   48      up          osd.12
>  14    hdd    2.70000   1.00000  2.7 TiB  1.6 TiB  1.6 TiB  448 MiB
> 6.0 GiB  1.1 TiB  59.68  0.82   38      up          osd.14
>  18    hdd    2.70000   1.00000  2.7 TiB  1.7 TiB  1.7 TiB  515 MiB
> 6.2 GiB  980 GiB  64.15  0.88   41      up          osd.18
>  22    hdd    1.70000   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  360 MiB
> 4.2 GiB  491 GiB  72.06  0.98   29      up          osd.22
>  30    hdd    1.70000   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  366 MiB
> 4.7 GiB  558 GiB  68.26  0.93   28      up          osd.30
>  33    hdd    1.59999   1.00000  1.6 TiB  1.2 TiB  1.2 TiB  406 MiB
> 4.9 GiB  427 GiB  74.28  1.01   29      up          osd.33
>  64    hdd    3.29999   1.00000  3.3 TiB  2.4 TiB  2.4 TiB  736 MiB
> 8.6 GiB  915 GiB  73.22  1.00   60      up          osd.64
>  -3          29.69995         -   30 TiB   22 TiB   22 TiB  5.4 GiB
> 81 GiB  7.9 TiB  73.20  1.00    -              host ceph02
>   2    hdd    1.70000   1.00000  1.7 TiB  1.3 TiB  1.2 TiB  402 MiB
> 5.2 GiB  476 GiB  72.93  1.00   30      up          osd.2
>   3    hdd    2.70000   1.00000  2.7 TiB  2.0 TiB  2.0 TiB  653 MiB
> 7.8 GiB  652 GiB  76.15  1.04   49      up          osd.3
>   7    hdd    2.70000   1.00000  2.7 TiB  2.5 TiB  2.5 TiB  456 MiB
> 7.7 GiB  209 GiB  92.36  1.26   56      up          osd.7
>   9    hdd    2.70000   1.00000  2.7 TiB  1.9 TiB  1.9 TiB  434 MiB
> 7.2 GiB  781 GiB  71.46  0.98   46      up          osd.9
>  13    hdd    2.39999   1.00000  2.4 TiB  1.6 TiB  1.6 TiB  451 MiB
> 6.1 GiB  823 GiB  66.28  0.91   38      up          osd.13
>  16    hdd    2.70000   1.00000  2.7 TiB  1.6 TiB  1.6 TiB  375 MiB
> 6.4 GiB  1.1 TiB  59.84  0.82   39      up          osd.16
>  19    hdd    1.70000   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  323 MiB
> 4.7 GiB  601 GiB  65.80  0.90   27      up          osd.19
>  23    hdd    2.70000   1.00000  2.7 TiB  2.2 TiB  2.2 TiB  471 MiB
> 7.7 GiB  520 GiB  80.99  1.11   50      up          osd.23
>  24    hdd    1.70000   1.00000  1.7 TiB  1.4 TiB  1.4 TiB  371 MiB
> 5.5 GiB  273 GiB  84.44  1.15   32      up          osd.24
>  28    hdd    2.70000   1.00000  2.7 TiB  1.9 TiB  1.9 TiB  428 MiB
> 7.4 GiB  818 GiB  70.07  0.96   44      up          osd.28
>  31    hdd    2.70000   1.00000  2.7 TiB  2.0 TiB  2.0 TiB  516 MiB
> 7.4 GiB  660 GiB  75.85  1.04   48      up          osd.31
>  32    hdd    3.29999   1.00000  3.3 TiB  2.2 TiB  2.2 TiB  661 MiB
> 7.9 GiB  1.2 TiB  64.86  0.89   52      up          osd.32
>  -4          26.29996         -   26 TiB   18 TiB   18 TiB  4.3 GiB
> 73 GiB  8.0 TiB  69.58  0.95    -              host ceph03
>   5    hdd    1.70000   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  298 MiB
> 5.2 GiB  541 GiB  69.21  0.95   29      up          osd.5
>   6    hdd    1.70000   1.00000  1.7 TiB  1.0 TiB  1.0 TiB  321 MiB
> 4.4 GiB  697 GiB  60.34  0.82   25      up          osd.6
>  10    hdd    2.70000   1.00000  2.7 TiB  1.9 TiB  1.9 TiB  431 MiB
> 7.5 GiB  796 GiB  70.89  0.97   46      up          osd.10
>  15    hdd    2.70000   1.00000  2.7 TiB  1.9 TiB  1.9 TiB  500 MiB
> 6.6 GiB  805 GiB  70.55  0.96   44      up          osd.15
>  17    hdd    1.59999   1.00000  1.6 TiB  1.1 TiB  1.1 TiB  377 MiB
> 4.9 GiB  530 GiB  68.05  0.93   27      up          osd.17
>  20    hdd    1.70000   1.00000  1.7 TiB  1.0 TiB  1.0 TiB  223 MiB
> 4.7 GiB  685 GiB  61.03  0.83   25      up          osd.20
>  21    hdd    2.70000   1.00000  2.7 TiB  1.7 TiB  1.7 TiB  392 MiB
> 6.7 GiB  951 GiB  65.23  0.89   42      up          osd.21
>  25    hdd    1.70000   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  157 MiB
> 5.1 GiB  601 GiB  65.83  0.90   27      up          osd.25
>  26    hdd    2.70000   1.00000  2.7 TiB  2.1 TiB  2.1 TiB  512 MiB
> 7.6 GiB  573 GiB  79.06  1.08   50      up          osd.26
>  27    hdd    2.70000   1.00000  2.7 TiB  1.9 TiB  1.9 TiB  473 MiB
> 7.6 GiB  805 GiB  70.55  0.96   46      up          osd.27
>  29    hdd    2.70000   1.00000  2.7 TiB  2.1 TiB  2.1 TiB  478 MiB
> 7.3 GiB  539 GiB  80.29  1.10   50      up          osd.29
>  63    hdd    1.70000   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  195 MiB
> 5.1 GiB  646 GiB  63.23  0.86   26      up          osd.63
> -11          24.79999         -   25 TiB   18 TiB   18 TiB  4.1 GiB
> 59 GiB  6.3 TiB  74.51  1.02    -              host ceph04
>  34    hdd    5.20000   1.00000  5.2 TiB  3.9 TiB  3.8 TiB  954 MiB
> 13 GiB  1.4 TiB  73.48  1.00   94      up          osd.34
>  42    hdd    5.20000   1.00000  5.2 TiB  3.9 TiB  3.8 TiB  841 MiB
> 13 GiB  1.4 TiB  73.43  1.00   94      up          osd.42
>  44    hdd    7.20000   1.00000  7.2 TiB  5.5 TiB  5.5 TiB  1.2 GiB
> 17 GiB  1.6 TiB  77.54  1.06  133      up          osd.44
>  45    hdd    7.20000   1.00000  7.2 TiB  5.2 TiB  5.2 TiB  1.2 GiB
> 16 GiB  1.9 TiB  73.03  1.00  125      up          osd.45
> -13          30.09998         -   30 TiB   22 TiB   22 TiB  5.1 GiB
> 72 GiB  8.0 TiB  73.48  1.00    -              host ceph05
>  39    hdd    7.20000   1.00000  7.2 TiB  5.6 TiB  5.6 TiB  1.4 GiB
> 17 GiB  1.6 TiB  77.89  1.06  126      up          osd.39
>  40    hdd    7.20000   1.00000  7.2 TiB  5.3 TiB  5.3 TiB  1.2 GiB
> 17 GiB  1.9 TiB  73.87  1.01  124      up          osd.40
>  41    hdd    7.20000   1.00000  7.2 TiB  5.4 TiB  5.3 TiB  1.1 GiB
> 17 GiB  1.8 TiB  74.92  1.02  128      up          osd.41
>  43    hdd    5.20000   1.00000  5.2 TiB  3.7 TiB  3.7 TiB  853 MiB
> 13 GiB  1.5 TiB  71.28  0.97   91      up          osd.43
>  60    hdd    3.29999   1.00000  3.3 TiB  2.1 TiB  2.1 TiB  573 MiB
> 7.9 GiB  1.2 TiB  63.53  0.87   52      up          osd.60
>  -9          17.59999         -   18 TiB   12 TiB   12 TiB  3.0 GiB
> 40 GiB  5.3 TiB  70.10  0.96    -              host ceph06
>  35    hdd    7.20000   1.00000  7.2 TiB  5.2 TiB  5.2 TiB  1.3 GiB
> 16 GiB  1.9 TiB  72.80  0.99  125      up          osd.35
>  36    hdd    5.20000   1.00000  5.2 TiB  3.5 TiB  3.5 TiB  804 MiB
> 12 GiB  1.8 TiB  66.50  0.91   85      up          osd.36
>  38    hdd    5.20000   1.00000  5.2 TiB  3.7 TiB  3.7 TiB  978 MiB
> 12 GiB  1.6 TiB  70.02  0.96   88      up          osd.38
> -15          24.89998         -   25 TiB   18 TiB   18 TiB  4.3 GiB
> 58 GiB  6.5 TiB  73.75  1.01    -              host ceph07
>  66    hdd    7.20000   1.00000  7.2 TiB  5.3 TiB  5.3 TiB  1.1 GiB
> 17 GiB  1.8 TiB  74.74  1.02  126      up          osd.66
>  67    hdd    7.20000   1.00000  7.2 TiB  5.3 TiB  5.3 TiB  1.2 GiB
> 17 GiB  1.8 TiB  74.51  1.02  121      up          osd.67
>  68    hdd    3.29999   1.00000  3.3 TiB  2.3 TiB  2.3 TiB  720 MiB
> 7.9 GiB  1.0 TiB  68.63  0.94   55      up          osd.68
>  69    hdd    7.20000   1.00000  7.2 TiB  5.3 TiB  5.3 TiB  1.2 GiB
> 17 GiB  1.8 TiB  74.40  1.02  129      up          osd.69
> -17          24.50000         -   24 TiB   20 TiB   20 TiB  4.1 GiB
> 47 GiB  4.4 TiB  82.08  1.12    -              host ceph08
>  37    hdd    9.50000   1.00000  9.5 TiB  7.8 TiB  7.7 TiB  1.5 GiB
> 18 GiB  1.8 TiB  81.39  1.11  166      up          osd.37
>  46    hdd    5.00000   1.00000  5.0 TiB  4.0 TiB  3.9 TiB  889 MiB
> 9.3 GiB  1.0 TiB  79.67  1.09   87      up          osd.46
>  47    hdd    5.00000   1.00000  5.0 TiB  4.2 TiB  4.2 TiB  863 MiB
> 9.7 GiB  817 GiB  83.90  1.15   90      up          osd.47
>  48    hdd    5.00000   1.00000  5.0 TiB  4.2 TiB  4.2 TiB  969 MiB
> 10 GiB  813 GiB  83.99  1.15   91      up          osd.48
>                           TOTAL  206 TiB  151 TiB  151 TiB   36 GiB
> 503 GiB   55 TiB  73.23
> MIN/MAX VAR: 0.82/1.26  STDDEV: 7.00
>
>
> root@ceph01:~# ceph balancer status
> {
>     "active": true,
>     "last_optimize_duration": "0:00:00.016174",
>     "last_optimize_started": "Fri Apr 16 12:54:47 2021",
>     "mode": "upmap",
>     "optimize_result": "Unable to find further optimization, or
> pool(s) pg_num is decreasing, or distribution is already perfect",
>     "plans": []
> }
>
>
> root@ceph01:~# ceph versions
> {
>     "mon": {
>         "ceph version 15.2.5
> (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 3
>     },
>     "mgr": {
>         "ceph version 15.2.5
> (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 2
>     },
>     "osd": {
>         "ceph version 15.2.5
> (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 56
>     },
>     "mds": {
>         "ceph version 15.2.5
> (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 4
>     },
>     "overall": {
>         "ceph version 15.2.5
> (2c93eff00150f0cc5f106a559557a58d3d7b6f1f) octopus (stable)": 65
>     }
> }
>
> root@ceph01:~# ceph osd crush rule ls
> replicated_ruleset
>
> root@ceph01:~# ceph osd crush rule dump replicated_ruleset
> {
>     "rule_id": 0,
>     "rule_name": "replicated_ruleset",
>     "ruleset": 0,
>     "type": 1,
>     "min_size": 1,
>     "max_size": 10,
>     "steps": [
>         {
>             "op": "take",
>             "item": -1,
>             "item_name": "default"
>         },
>         {
>             "op": "chooseleaf_firstn",
>             "num": 0,
>             "type": "host"
>         },
>         {
>             "op": "emit"
>         }
>     ]
> }
>
>
>
> Cheers,
> Michael
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux