Re: OSD weight on Luminous

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



unless uou have enabled some balancing - then this is very normal (actually pretty good normal)

Jesper


Sent from myMail for iOS


Thursday, 14 May 2020, 09.35 +0200 from Florent B.  <florent@xxxxxxxxxxx>:
>Hi,
>
>I have something strange on a Ceph Luminous cluster.
>
>All OSDs have the same size, the same weight, and one of them is used at
>88% by Ceph (osd.3) while others are around 40 to 50% usage :
>
>#  ceph osd df
>ID CLASS WEIGHT  REWEIGHT SIZE    USE     DATA    OMAP    META   
>AVAIL   %USE  VAR  PGS
> 2   hdd 0.49179  1.00000  504GiB  264GiB  263GiB 63.7MiB  960MiB 
>240GiB 52.34 1.14  81
>13   hdd 0.49179  1.00000  504GiB  267GiB  266GiB 55.7MiB 1.37GiB 
>236GiB 53.09 1.16  94
>20   hdd 0.49179  1.00000  504GiB  235GiB  234GiB 62.5MiB  962MiB 
>268GiB 46.70 1.02  99
>21   hdd 0.49179  1.00000  504GiB  306GiB  305GiB 65.2MiB  991MiB 
>198GiB 60.75 1.32  87
>22   hdd 0.49179  1.00000  504GiB  185GiB  184GiB 51.9MiB  972MiB 
>318GiB 36.83 0.80  73
>23   hdd 0.49179  1.00000  504GiB  167GiB  166GiB 60.9MiB  963MiB 
>337GiB 33.07 0.72  80
>24   hdd 0.49179  1.00000  504GiB  235GiB  234GiB 67.5MiB  956MiB 
>268GiB 46.74 1.02  90
>25   hdd 0.49179  1.00000  504GiB  183GiB  182GiB 68.8MiB  955MiB 
>321GiB 36.32 0.79 100
> 3   hdd 0.49179  1.00000  504GiB  442GiB  440GiB 77.5MiB 1.15GiB
>61.9GiB  87.70 1.91 103
>26   hdd 0.49179  1.00000  504GiB  220GiB  219GiB 61.2MiB  963MiB 
>283GiB 43.78 0.95  80
>29   hdd 0.49179  1.00000  504GiB  298GiB  296GiB 77.4MiB 1013MiB 
>206GiB 59.09 1.29 106
>30   hdd 0.49179  1.00000  504GiB  183GiB  182GiB 60.2MiB  964MiB 
>321GiB 36.32 0.79  88
>10   hdd 0.49179  1.00000  504GiB  176GiB  175GiB 56.5MiB  968MiB 
>327GiB 35.02 0.76  85
>11   hdd 0.49179  1.00000  504GiB  209GiB  208GiB 62.5MiB  961MiB 
>295GiB 41.42 0.90  89
> 0   hdd 0.49179  1.00000  504GiB  253GiB  252GiB 55.7MiB  968MiB 
>251GiB 50.18 1.09  76
> 1   hdd 0.49179  1.00000  504GiB  199GiB  198GiB 60.4MiB  964MiB 
>305GiB 39.51 0.86  92
>16   hdd 0.49179  1.00000  504GiB  219GiB  218GiB 58.2MiB  966MiB 
>284GiB 43.51 0.95  85
>17   hdd 0.49179  1.00000  504GiB  231GiB  230GiB 69.0MiB  955MiB 
>272GiB 45.97 1.00  97
>14   hdd 0.49179  1.00000  504GiB  210GiB  209GiB 61.0MiB  963MiB 
>293GiB 41.72 0.91  74
>15   hdd 0.49179  1.00000  504GiB  182GiB  181GiB 50.7MiB  973MiB 
>322GiB 36.10 0.79  72
>18   hdd 0.49179  1.00000  504GiB  297GiB  296GiB 53.7MiB  978MiB 
>206GiB 59.03 1.29  87
>19   hdd 0.49179  1.00000  504GiB  125GiB  124GiB 61.9MiB  962MiB 
>379GiB 24.81 0.54  82
>                    TOTAL 10.8TiB 4.97TiB 4.94TiB 1.33GiB 21.4GiB
>5.85TiB 45.91         
>MIN/MAX VAR: 0.54/1.91  STDDEV: 12.80
>
>
>Is it a normal situation ? Is there any way to let Ceph handle this
>alone or am I forced to reweight the OSD manually ?
>
>Thank you.
>
>Florent
>_______________________________________________
>ceph-users mailing list --  ceph-users@xxxxxxx
>To unsubscribe send an email to  ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux