Hi all, I'm running a quincy 17.2.5 ceph cluster, with 3 nodes that each have 3 disks with replica size 3 min_size 2, my cluster was running fine before, and suddenly it can't read and write because ceph slowops on all osd. I tried restarting osd, then degraded and misplaced data appeared, and slow ops reappeared. Any suggestions guys to solve this issues ? # ceph -s cluster: id: bb6590a9-155c-4830-aea7-ef4b5a098466 health: HEALTH_ERR 2 failed cephadm daemon(s) 1/282019 objects unfound (0.000%) noout flag(s) set Reduced data availability: 79 pgs inactive, 47 pgs peering Possible data damage: 1 pg recovery_unfound Degraded data redundancy: 26340/838103 objects degraded (3.143%), 16 pgs degraded, 11 pgs undersized 302 slow ops, oldest one blocked for 226773 sec, daemons [osd.2,osd.3,osd.4,osd.6,osd.7] have slow ops. services: mon: 3 daemons, quorum r2c2,r2c3,r2c1 (age 2h) mgr: r2c3(active, since 15m), standbys: r2c2, r2c1.ivwibd osd: 9 osds: 9 up (since 9m), 9 in (since 3d); 30 remapped pgs flags noout data: pools: 12 pools, 329 pgs objects: 282.02k objects, 1.3 TiB usage: 4.9 TiB used, 94 TiB / 99 TiB avail pgs: 24.924% pgs not active 26340/838103 objects degraded (3.143%) 5411/838103 objects misplaced (0.646%) 1/282019 objects unfound (0.000%) 240 active+clean 31 peering 23 activating 16 remapped+peering 7 activating+undersized+degraded+remapped 2 activating+remapped 2 active+undersized+degraded+remapped+backfilling 2 active+recovering+degraded 2 activating+degraded 1 active+recovery_unfound+undersized+degraded+remapped 1 activating+degraded+remapped 1 active+clean+scrubbing+deep 1 active+recovery_wait+undersized+degraded+remapped io: client: 12 KiB/s rd, 1.3 KiB/s wr, 7 op/s rd, 1 op/s wr Thanks, Pahrial _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx