Hey guys, I'm trying to understand what is happening in my cluster, I see the number of degraded objects increasing, while all OSD's are still up and running. Can someone explain what's happening? I would expect the number of misplaced objects to increase when ceph's balancing algorithm decides blocks should be on different OSDs, but I would only expect the degraded objects to increase when OSD's die?? 6/15/23 11:31:25 AM[WRN]Health check update: Degraded data redundancy: 809035/219783315 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:31:20 AM[WRN]Health check update: Degraded data redundancy: 809044/219788544 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:31:15 AM[WRN]Health check update: Degraded data redundancy: 809044/219788616 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:31:10 AM[WRN]Health check update: Degraded data redundancy: 808944/219777540 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:31:05 AM[WRN]Health check update: Degraded data redundancy: 808944/219776271 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:31:00 AM[WRN]Health check update: Degraded data redundancy: 808821/219683475 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:30:55 AM[WRN]Health check update: Degraded data redundancy: 808740/219672240 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) 6/15/23 11:30:50 AM[WRN]Health check update: Degraded data redundancy: 808667/219645417 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized (PG_DEGRADED) health: health: HEALTH_WARN Degraded data redundancy: 810779/220602543 objects degraded (0.368%), 6 pgs degraded, 6 pgs undersized services: mon: 3 daemons, quorum ceph-mon01,ceph-mon02,ceph-mon03 (age 12h) mgr: ceph-mon01.vzbglj(active, since 12h), standbys: ceph-mon02.qtuntk mds: 1/1 daemons up, 2 standby osd: 118 osds: 118 up (since 12h), 118 in (since 23h); 510 remapped pgs data: volumes: 1/1 healthy pools: 5 pools, 1249 pgs objects: 44.55M objects, 60 TiB usage: 118 TiB used, 2.0 PiB / 2.1 PiB avail pgs: 810779/220602543 objects degraded (0.368%) 39586388/220602543 objects misplaced (17.945%) 739 active+clean 406 active+remapped+backfill_wait 98 active+remapped+backfilling 6 active+undersized+degraded+remapped+backfilling io: client: 585 KiB/s rd, 22 MiB/s wr, 357 op/s rd, 1.41k op/s wr recovery: 150 MiB/s, 82 objects/s Kind regards, Angelo Hongens _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx