Hi All,
We are new with cephfs and we have 5 OSD and each size has 3.3TB. As of now data has been stored around 12 TB size, unfortunately osd5 went down and while remapped+backfill below error is showing even though we have around 2TB free spaces. Kindly provide the solution to solve this issue.
cluster b466e09c-f7ae-4e89-99a7-99d30eba0a13
health HEALTH_WARN
51 pgs backfill_toofull
57 pgs degraded
57 pgs stuck unclean
57 pgs undersized
recovery 1139974/17932626 objects degraded (6.357%)
recovery 1139974/17932626 objects misplaced (6.357%)
3 near full osd(s)
monmap e2: 3 mons at {intcfs-mon1=192.168.113.113:6789/0,intcfs-mon2=192.168.113.114:6789/0,intcfs-mon3=192.168.113.72:6789/0}
election epoch 10, quorum 0,1,2 intcfs-mon3,intcfs-mon1,intcfs-mon2
fsmap e26: 1/1/1 up {0=intcfs-osd1=up:active}, 1 up:standby
osdmap e2349: 5 osds: 4 up, 4 in; 57 remapped pgs
flags sortbitwise
pgmap v681178: 564 pgs, 3 pools, 5811 GB data, 8756 kobjects
11001 GB used, 2393 GB / 13394 GB avail
1139974/17932626 objects degraded (6.357%)
1139974/17932626 objects misplaced (6.357%)
506 active+clean
51 active+undersized+degraded+remapped+backfill_toofull
6 active+undersized+degraded+remapped
1 active+clean+scrubbing
192.168.113.113,192.168.113.114,192.168.113.72:6789:/ ceph 14T 11T 2.4T 83% /home/build/cephfsdownloads
Regards
Prabu GJ
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com