Hi all, I use puppet to deploy and manage my clusters. Recently, as I have been doing a removal of old hardware and adding of new I've noticed that sometimes the "ceph osd create" is returning repeated IDs. Usually it's on the same server, but yesterday I saw it in different servers. I was expecting the OSD ID's to be unique, and when they come on the same server puppet starts spewing errors - which is desirable - but when it's in different servers it broke those OSDs in Ceph. As they hadn't backfill any full PGs I just wiped, removed and started anew. As for the process itself: The OSDs are marked out and removed from crush, when empty they are auth del and osd rm. After building the server puppet will osd create, and use the generated ID for crush move and mkfs. Unfortunately I haven't been able to reproduce in isolation, and being a production cluster logging is tuned way down. This has happened in several different clusters, but they are all running 10.2.7. Any ideas? thanks, _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com