Re: pgs stuck unclean after removing OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



David Turner wrote:
: A couple things.  You didn't `ceph osd crush remove osd.21` after doing the
: other bits.  Also you will want to remove the bucket (re: host) from the
: crush map as it will now be empty.  Right now you have a host in the crush
: map with a weight, but no osds to put that data on.  It has a weight
: because of the 2 OSDs that are still in it that were removed from the
: cluster but not from the crush map.  It's confusing to your cluster.

	OK, this helped. I have removed osd.20 and osd.21 from the crush
map, as well as the bucket for the faulty host. PGs got unstuck, and after
some time, my system now reports HEALTH_OK.

	Thanks for the hint!

-Yenya

-- 
| Jan "Yenya" Kasprzak <kas at {fi.muni.cz - work | yenya.net - private}> |
| http://www.fi.muni.cz/~kas/                         GPG: 4096R/A45477D5 |
> That's why this kind of vulnerability is a concern: deploying stuff is  <
> often about collecting an obscene number of .jar files and pushing them <
> up to the application server.                          --pboddie at LWN <
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux