Hi This been bugging me for some time now, the distribution of data on the OSD is not balanced so some OSD are near full, i did ceph osd reweight-by-utilization but it not helping much.
[root@controller-node ~]# ceph osd tree
# id weight type name up/down reweight
-1 98.28 root default
-2 32.76 host ceph-node-1
0 3.64 osd.0 up 1
1 3.64 osd.1 up 1
2 3.64 osd.2 up 1
3 3.64 osd.3 up 1
4 3.64 osd.4 up 1
5 3.64 osd.5 up 1
6 3.64 osd.6 up 1
7 3.64 osd.7 up 1
8 3.64 osd.8 up 1
-3 32.76 host ceph-node-2
9 3.64 osd.9 up 1
10 3.64 osd.10 up 1
11 3.64 osd.11 up 1
12 3.64 osd.12 up 1
13 3.64 osd.13 up 1
14 3.64 osd.14 up 1
15 3.64 osd.15 up 1
16 3.64 osd.16 up 1
17 3.64 osd.17 up 1
-4 32.76 host ceph-node-3
18 3.64 osd.18 up 1
19 3.64 osd.19 up 1
20 3.64 osd.20 up 1
21 3.64 osd.21 up 1
22 3.64 osd.22 up 1
23 3.64 osd.23 up 1
24 3.64 osd.24 up 1
25 3.64 osd.25 up 1
26 3.64 osd.26 up 1
[root@controller-node ~]#
[root@controller-node ~]# /opt/df-osd.sh
ceph-node-1
=======================================================================
/dev/sdb1 3.7T 2.0T 1.7T 54% /var/lib/ceph/osd/ceph-0
/dev/sdc1 3.7T 2.7T 1.1T 72% /var/lib/ceph/osd/ceph-1
/dev/sdd1 3.7T 3.3T 431G 89% /var/lib/ceph/osd/ceph-2
/dev/sde1 3.7T 2.8T 879G 77% /var/lib/ceph/osd/ceph-3
/dev/sdf1 3.7T 3.3T 379G 90% /var/lib/ceph/osd/ceph-4
/dev/sdg1 3.7T 2.9T 762G 80% /var/lib/ceph/osd/ceph-5
/dev/sdh1 3.7T 3.0T 733G 81% /var/lib/ceph/osd/ceph-6
/dev/sdi1 3.7T 3.4T 284G 93% /var/lib/ceph/osd/ceph-7
/dev/sdj1 3.7T 3.4T 342G 91% /var/lib/ceph/osd/ceph-8
=======================================================================
ceph-node-2
=======================================================================
/dev/sdb1 3.7T 3.1T 622G 84% /var/lib/ceph/osd/ceph-9
/dev/sdc1 3.7T 2.7T 1.1T 72% /var/lib/ceph/osd/ceph-10
/dev/sdd1 3.7T 3.1T 557G 86% /var/lib/ceph/osd/ceph-11
/dev/sde1 3.7T 3.3T 392G 90% /var/lib/ceph/osd/ceph-12
/dev/sdf1 3.7T 2.6T 1.1T 72% /var/lib/ceph/osd/ceph-13
/dev/sdg1 3.7T 2.8T 879G 77% /var/lib/ceph/osd/ceph-14
/dev/sdh1 3.7T 2.7T 984G 74% /var/lib/ceph/osd/ceph-15
/dev/sdi1 3.7T 3.2T 463G 88% /var/lib/ceph/osd/ceph-16
/dev/sdj1 3.7T 3.1T 594G 85% /var/lib/ceph/osd/ceph-17
=======================================================================
ceph-node-3
=======================================================================
/dev/sdb1 3.7T 2.8T 910G 76% /var/lib/ceph/osd/ceph-18
/dev/sdc1 3.7T 2.7T 1012G 73% /var/lib/ceph/osd/ceph-19
/dev/sdd1 3.7T 3.2T 537G 86% /var/lib/ceph/osd/ceph-20
/dev/sde1 3.7T 3.2T 465G 88% /var/lib/ceph/osd/ceph-21
/dev/sdf1 3.7T 3.0T 663G 83% /var/lib/ceph/osd/ceph-22
/dev/sdg1 3.7T 3.4T 248G 94% /var/lib/ceph/osd/ceph-23
/dev/sdh1 3.7T 2.8T 928G 76% /var/lib/ceph/osd/ceph-24
/dev/sdi1 3.7T 2.9T 802G 79% /var/lib/ceph/osd/ceph-25
/dev/sdj1 3.7T 2.7T 1.1T 73% /var/lib/ceph/osd/ceph-26
=======================================================================
[root@controller-node ~]#
[root@controller-node ~]# ceph health detail
HEALTH_ERR 2 pgs inconsistent; 10 near full osd(s); 2 scrub errors
pg 5.7f is active+clean+inconsistent, acting [2,12,18]
pg 5.38 is active+clean+inconsistent, acting [7,9,24]
osd.2 is near full at 88%
osd.4 is near full at 89%
osd.7 is near full at 92%
osd.8 is near full at 90%
osd.11 is near full at 85%
osd.12 is near full at 89%
osd.16 is near full at 87%
osd.20 is near full at 85%
osd.21 is near full at 87%
osd.23 is near full at 93%
2 scrub errors
[root@controller-node ~]#
[root@controller-node ~]# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
100553G 19796G 80757G 80.31
POOLS:
NAME ID USED %USED OBJECTS
images 4 8680G 8.63 1111395
volumes 5 18192G 18.09 4675359
[root@controller-node ~]#
[root@controller-node ~]# ceph osd pool get images pg_num
pg_num: 1024
[root@controller-node ~]#
[root@controller-node ~]# ceph osd pool get volumes pg_num
pg_num: 300
[root@controller-node ~]#
Thanks.
/Vlad
ᐧ
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com