Diffrent OSD capacity & what is the weight of item

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi, 
        I have four storage machines to build a ceph storage cluster as storage nodes. Each of them is attached a 120 GB HDD and a 1 TB HDD. Is it OK to think that those storage devices are same when write a ceph.conf?
        For example, when setting osd pool default pg num , I thought: osd pool default pg num = (100 * 8 ) / 3 =  266, where osd pool default size = 3 and the number of OSDs is 8 (one Daemon per device).
   
        And, when Add the OSD to the CRUSH map so that it can begin receiving data. You may also decompile the CRUSH map, add the OSD to the device list, add the host as a bucket (if it’s not already in the CRUSH map), add the device as an item in the host, assign it a weight, recompile it and set it.
        ceph [--cluster {cluster-name}] osd crush add {id-or-name} {weight} [{bucket-type}={bucket-name} ...]
        What is the meaning of weight? How should I set it to satisfy my hardware condition ?



--
Zhen Wang


 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux