[root@controller-node ~]# /opt/df-osd.sh
ceph-node-1
=======================================================================
/dev/sdb1 3.7T 2.0T 1.7T 54% /var/lib/ceph/osd/ceph-0
/dev/sdc1 3.7T 2.7T 1.1T 72% /var/lib/ceph/osd/ceph-1
/dev/sdd1 3.7T 3.3T 431G 89% /var/lib/ceph/osd/ceph-2
/dev/sde1 3.7T 2.8T 879G 77% /var/lib/ceph/osd/ceph-3
/dev/sdf1 3.7T 3.3T 379G 90% /var/lib/ceph/osd/ceph-4
/dev/sdg1 3.7T 2.9T 762G 80% /var/lib/ceph/osd/ceph-5
/dev/sdh1 3.7T 3.0T 733G 81% /var/lib/ceph/osd/ceph-6
/dev/sdi1 3.7T 3.4T 284G 93% /var/lib/ceph/osd/ceph-7
/dev/sdj1 3.7T 3.4T 342G 91% /var/lib/ceph/osd/ceph-8
=======================================================================
ceph-node-2
=======================================================================
/dev/sdb1 3.7T 3.1T 622G 84% /var/lib/ceph/osd/ceph-9
/dev/sdc1 3.7T 2.7T 1.1T 72% /var/lib/ceph/osd/ceph-10
/dev/sdd1 3.7T 3.1T 557G 86% /var/lib/ceph/osd/ceph-11
/dev/sde1 3.7T 3.3T 392G 90% /var/lib/ceph/osd/ceph-12
/dev/sdf1 3.7T 2.6T 1.1T 72% /var/lib/ceph/osd/ceph-13
/dev/sdg1 3.7T 2.8T 879G 77% /var/lib/ceph/osd/ceph-14
/dev/sdh1 3.7T 2.7T 984G 74% /var/lib/ceph/osd/ceph-15
/dev/sdi1 3.7T 3.2T 463G 88% /var/lib/ceph/osd/ceph-16
/dev/sdj1 3.7T 3.1T 594G 85% /var/lib/ceph/osd/ceph-17
=======================================================================
ceph-node-3
=======================================================================
/dev/sdb1 3.7T 2.8T 910G 76% /var/lib/ceph/osd/ceph-18
/dev/sdc1 3.7T 2.7T 1012G 73% /var/lib/ceph/osd/ceph-19
/dev/sdd1 3.7T 3.2T 537G 86% /var/lib/ceph/osd/ceph-20
/dev/sde1 3.7T 3.2T 465G 88% /var/lib/ceph/osd/ceph-21
/dev/sdf1 3.7T 3.0T 663G 83% /var/lib/ceph/osd/ceph-22
/dev/sdg1 3.7T 3.4T 248G 94% /var/lib/ceph/osd/ceph-23
/dev/sdh1 3.7T 2.8T 928G 76% /var/lib/ceph/osd/ceph-24
/dev/sdi1 3.7T 2.9T 802G 79% /var/lib/ceph/osd/ceph-25
/dev/sdj1 3.7T 2.7T 1.1T 73% /var/lib/ceph/osd/ceph-26
=======================================================================
[root@controller-node ~]#
[root@controller-node ~]# ceph health detail
HEALTH_ERR 2 pgs inconsistent; 10 near full osd(s); 2 scrub errors
pg 5.7f is active+clean+inconsistent, acting [2,12,18]
pg 5.38 is active+clean+inconsistent, acting [7,9,24]
osd.2 is near full at 88%
osd.4 is near full at 89%
osd.7 is near full at 92%
osd.8 is near full at 90%
osd.11 is near full at 85%
osd.12 is near full at 89%
osd.16 is near full at 87%
osd.20 is near full at 85%
osd.21 is near full at 87%
osd.23 is near full at 93%
2 scrub errors
[root@controller-node ~]#
[root@controller-node ~]# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
100553G 19796G 80757G 80.31
POOLS:
NAME ID USED %USED OBJECTS
images 4 8680G 8.63 1111395
volumes 5 18192G 18.09 4675359
[root@controller-node ~]#
[root@controller-node ~]# ceph osd pool get images pg_num
pg_num: 1024
[root@controller-node ~]#
[root@controller-node ~]# ceph osd pool get volumes pg_num
pg_num: 300
[root@controller-node ~]#
Thanks.
/Vlad