rebalance near full osd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi

I've just had a warning ( from ceph -s) that one of the osds is near full. Having investigated the warning, i've located that osd.6 is 86% full. The data distribution is nowhere near to being equal on my osds as you can see from the df command output below:

/dev/sdj1       2.8T  2.4T  413G  86% /var/lib/ceph/osd/ceph-6
/dev/sdb1       2.8T  2.1T  625G  78% /var/lib/ceph/osd/ceph-0
/dev/sdc1       2.8T  2.0T  824G  71% /var/lib/ceph/osd/ceph-1
/dev/sdd1       2.8T  1.5T  1.3T  55% /var/lib/ceph/osd/ceph-2
/dev/sde1       2.8T  1.7T  1.1T  63% /var/lib/ceph/osd/ceph-3
/dev/sdh1       2.8T  1.7T  1.1T  62% /var/lib/ceph/osd/ceph-4
/dev/sdf1       2.8T  1.9T  932G  67% /var/lib/ceph/osd/ceph-8
/dev/sdi1       2.8T  1.9T  880G  69% /var/lib/ceph/osd/ceph-5
/dev/sdg1       2.8T  2.0T  798G  72% /var/lib/ceph/osd/ceph-7

I seem to have the spread of over 30% disk utilisation between the osds, despite all my osds having the identical weight (ceph osd tree output):


 -2 24.56999                     host arh-ibstorage1-ib                                       
  1  2.73000                         osd.1                       up  1.00000          1.00000
  3  2.73000                         osd.3                       up  1.00000          1.00000
  5  2.73000                         osd.5                       up  1.00000          1.00000
  6  2.73000                         osd.6                       up  1.00000          1.00000
  7  2.73000                         osd.7                       up  1.00000          1.00000
  8  2.73000                         osd.8                       up  1.00000          1.00000
  4  2.73000                         osd.4                       up  1.00000          1.00000
  0  2.73000                         osd.0                       up  1.00000          1.00000
  2  2.73000                         osd.2                       up  1.00000          1.00000


What would be the best way to correct the issue without having significant impact on the cluster IO?

Many thanks

Andrei
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux