Hello Everyone, We seem to be having a problem on one of our ceph clusters post the OS patch and reboot of one of the nodes. The three other nodes are showing OSD fill rates of 77%-81%, but the 60 OSDs contained in the host that was just rebooted are varying between 64% and 90% since the reboot occurred. The three other nodes have not yet been patched or rebooted. The result is: health: HEALTH_WARN 15 nearfull osd(s) 7 pool(s) nearfull Low space hindering backfill (add storage if this doesn't resolve itself): 15 pgs backfill_toofull Degraded data redundancy: 170940/1437684990 objects degraded (0.012%), 4 pgs degraded, 4 pgs undersized services: mon: 3 daemons, quorum prdceph01,prdceph02,prdceph03 (age 6h) mgr: prdceph01(active, since 5w), standbys: prdceph02, prdceph03, prdceph04 mds: ArchiveRepository:1 {0=prdceph01=up:active} 3 up:standby osd: 240 osds: 240 up (since 6h), 240 in (since 27h); 16 remapped pgs task status: scrub status: mds.prdceph01: idle data: pools: 7 pools, 8384 pgs objects: 479.23M objects, 557 TiB usage: 1.7 PiB used, 454 TiB / 2.1 PiB avail pgs: 170940/1437684990 objects degraded (0.012%) 4155186/1437684990 objects misplaced (0.289%) 8332 active+clean 36 active+clean+scrubbing+deep 11 active+remapped+backfill_toofull 2 active+undersized+degraded+remapped+backfill_toofull 2 active+forced_recovery+undersized+degraded+remapped+forced_backfill+backfill_toofull 1 active+remapped+backfilling io: client: 9.6 MiB/s rd, 820 KiB/s wr, 1.02k op/s rd, 189 op/s wr recovery: 0 B/s, 25 keys/s, 10 objects/s Any suggestions would be greatly appreciated, as currently it is not able to complete the repair, nor will it backfill, even when attempting to force. Many thanks in advance. Marco _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx