Hi all,
During a crash disaster we had destroyed and reinstalled with a
different number a few osds.
As an example osd 3 was destroyed and recreated with id 101 by command
ceph osd purge 3 --yes-i-really-mean-it + ceph osd create (to block id
3) + ceph-deploy osd create --data /dev/sdxx <server> and finally ceph
osd rm 3).
Some of our pgs are now incomplet (which can be understood) but blocked
by some of the removed osd :
ex: here is an part of the ceph pg 30.3 query
{
"state": "incomplete",
"snap_trimq": "[]",
"snap_trimq_len": 0,
"epoch": 384075,
"up": [
103,
43,
29,
2,
66
],
"acting": [
103,
43,
29,
2,
66
],
........
"peer_info": [
{
"peer": "2(3)",
"pgid": "30.3s3",
"last_update": "373570'105925965",
"last_complete": "373570'105925965",
.......
},
"up": [
103,
43,
29,
2,
66
],
"acting": [
103,
43,
29,
2,
66
],
"avail_no_missing": [],
"object_location_counts": [],
*"blocked_by": [**
** 3,**
** 49**
** ],*
............
"down_osds_we_would_probe": [
*3*
],
"peering_blocked_by": [],
"peering_blocked_by_detail": [
{
* "detail": "peering_blocked_by_history_les_bound"*
}
]
I don't understand why the removed osd are still considered and present
in the pg infos.
Is there a way to get rid of that ?
Moreover, we have tons of slow ops (more than 15 000) but I guess that
the two problems are linked.
Thanks for your help.
F.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx