Re: CRUSH odd bucket affinity / persistence

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

> I’m having a (strange) issue with OSD bucket persistence / affinity on my test cluster..   
> 
> The cluster is PoC / test, by no means production.   Consists of a single OSD / MON host + another MON running on a KVM VM.  
> 
> Out of 12 OSDs I’m trying to get osd.10 and osd.11 to be part of the ssd bucket in my CRUSH map.   This works fine when either editing the CRUSH map by hand (exporting, decompile, edit, compile, import), or via the ceph osd crush set command:
> 
> "ceph osd crush set osd.11 0.140 root=ssd”
> 
> I’m able to verify that the OSD / MON host and another MON I have running see the same CRUSH map.     
> 
> After rebooting OSD / MON host, both osd.10 and osd.11 become part of the default bucket.   How can I ensure that ODSs persist in their configured buckets?

I guess you have set "osd crush update on start = true" (http://ceph.com/docs/master/rados/operations/crush-map/ ) and only the default „root“-entry.

Either fix the „root“-Entry in the ceph.conf or set osd crush update on start = false.

greetings

Johannes
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux