On 2014.05.07 20:28, *sm1Ly wrote: > I got deploy my cluster with this commans. > > mkdir "clustername" > > cd "clustername" > > ceph-deploy install mon1 mon2 mon3 mds1 mds2 mds3 osd200 > > ceph-deploy new mon1 mon2 mon3 > > ceph-deploy mon create mon1 mon2 mon3 > > ceph-deploy gatherkeys mon1 mon2 mon3 > > ceph-deploy osd prepare --fs-type ext4 osd200:/osd/osd1 > osd200:/osd/osd2 osd200:/osd/osd3 osd200:/osd/osd4 osd200:/osd/osd5 > osd200:/osd/osd6 osd200:/osd/osd7 osd200:/osd/osd8 osd200:/osd/osd9 > osd200:/osd/osd10 osd200:/osd/osd11 osd200:/osd/osd12 > > ceph-deploy osd activate osd200:/osd/osd1 osd200:/osd/osd2 > osd200:/osd/osd3 osd200:/osd/osd4 osd200:/osd/osd5 osd200:/osd/osd6 > osd200:/osd/osd7 osd200:/osd/osd8 osd200:/osd/osd9 osd200:/osd/osd10 > osd200:/osd/osd11 osd200:/osd/osd12 > > > ceph-deploy admin mon1 mon2 mon3 mds1 mds2 mds3 osd200 salt1 > > ceph-deploy mds create mds1 mds2 mds3 > > but in the end...: > > sudo ceph -s > > [sm1ly at salt1 ceph]$ sudo ceph -s > cluster 0b2c9c20-985a-4a39-af8e-ef2325234744 > health HEALTH_WARN 19 pgs degraded; 192 pgs stuck unclean; > recovery 21/42 objects degraded (50.000%); too few pgs per osd (16 < > min 20) > monmap e1: 3 mons at > {mon1=10.60.0.110:6789/0,mon2=10.60.0.111:6789/0,mon3=10.60.0.112:6789/0 > <http://10.60.0.110:6789/0,mon2=10.60.0.111:6789/0,mon3=10.60.0.112:6789/0>}, > election epoch 6, quorum 0,1,2 mon1,mon2,mon3 > mdsmap e6: 1/1/1 up {0=mds1=up:active}, 2 up:standby > osdmap e61: 12 osds: 12 up, 12 in > pgmap v103: 192 pgs, 3 pools, 9470 bytes data, 21 objects > 63751 MB used, 3069 GB / 3299 GB avail > 21/42 objects degraded (50.000%) > 159 active > 14 active+remapped > 19 active+degraded > > > mon[123] and mds[123] are vms. osd200 - hardware server, cause on vms > it shows bad perfomance? > > some searching talks me that the problem that I have only one osd > node. can I ignore it for tests? "19 pgs degraded; 192 pgs stuck unclean; recovery 21/42 objects degraded (50.000%)" you can ignore it, or edit crush map, so failure domain would be osd, not host > another search talks me about pg groups, but I cant find how to get pgid. "too few pgs per osd (16 < min 20)" - increase pg_num and pgp_num > > > -- > yours respectfully, Alexander Vasin. > > 8 926 1437200 > icq: 9906064 > > > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140507/63b20e93/attachment.htm>