Hi,
On 01/08/2018 05:40 PM, Alessandro De Salvo wrote:
Thanks Lincoln,
indeed, as I said the cluster is recovering, so there are pending ops:
pgs: 21.034% pgs not active
1692310/24980804 objects degraded (6.774%)
5612149/24980804 objects misplaced (22.466%)
458 active+clean
329 active+remapped+backfill_wait
159 activating+remapped
100 active+undersized+degraded+remapped+backfill_wait
58 activating+undersized+degraded+remapped
27 activating
22 active+undersized+degraded+remapped+backfilling
6 active+remapped+backfilling
1 active+recovery_wait+degraded
If it's just a matter to wait for the system to complete the recovery
it's fine, I'll deal with that, but I was wondendering if there is a
more suble problem here.
OK, I'll wait for the recovery to complete and see what happens, thanks.
The blocked MDS might be caused by the 'activating' PGs. Do you have a
warning about too much PGs per OSD? If that is the case,
activating/creating/peering/whatever on the affected OSDs is blocked,
which leads to blocked requests etc.
You can resolve this be increasing the number of allowed PGs per OSD
('mon_max_pg_per_osd'). AFAIK it needs to be set for mon, mgr and osd
instances. There was also been some discussion about this setting on the
mailing list in the last weeks.
Regards,
Burkhard
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com