Hi Alfredo, See the steps I executed below and the weird error I am getting when trying to activate OSDs- the last series of error messages are in an infinite loop- still printing > 2 days . FYI, /etc/ceph existed on all nodes after ceph-deploy
install. I checked after doing ceph-deploy install. Do you need the ceph.log file?
[ceph@ceph-admin-node-centos-6-4 ~]# mkdir my-cluster [ceph@ceph-admin-node-centos-6-4 ~]# cd my-cluster/ [ceph@ceph-admin-node-centos-6-4 ~]# ceph-deploy new ceph-node1-mon-centos-6-4 The above command creates a ceph.conf file with the cluster information in it. A log file by the name of ceph.log will also be created
[ceph@ceph-admin-node-centos-6-4 my-cluster]# ceph-deploy install ceph-node1-mon-centos-6-4 ceph-node2-osd0-centos-6-4 ceph-node3-osd1-centos-6-4 This will install ceph on all the nodes I added a ceph monitor node: [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy mon create ceph-node1-mon-centos-6-4 Gather keys: [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy gatherkeys ceph-node1-mon-centos-6-4 After gathering keys, made sure the directory should have - monitoring, admin, osd, mds keyrings:
[ceph@ceph-admin-node-centos-6-4 mycluster]$ ls ceph.bootstrap-mds.keyring ceph.bootstrap-osd.keyring ceph.client.admin.keyring ceph.conf ceph.log ceph.mon.keyring Addede two OSDs: Ssh;ed to both OSD nodes i.e. ceph-node2-osd0-centos-6-4 and ceph-node3-osd1-centos-6-4 and created two directories to be used as Ceph OSD daemons:
[ceph@ceph-admin-node-centos-6-4 mycluster]$ ssh ceph-node2-osd0-centos-6-4 Last login: Wed Oct 30 10:51:11 2013 from ceph-admin-node-centos-6-4 [ceph@ceph-node2-osd0-centos-6-4 ~]$ sudo mkdir -p /ceph/osd0 [ceph@ceph-node2-osd0-centos-6-4 ~]$ exit logout Connection to ceph-node2-osd0-centos-6-4 closed. [ceph@ceph-admin-node-centos-6-4 mycluster]$ ssh ceph-node3-osd1-centos-6-4 Last login: Wed Oct 30 10:51:20 2013 from ceph-admin-node-centos-6-4 [ceph@ceph-node3-osd1-centos-6-4 ~]$ sudo mkdir -p /ceph/osd1 [ceph@ceph-node3-osd1-centos-6-4 ~]$ exit logout Connection to ceph-node3-osd1-centos-6-4 closed. [ceph@ceph-admin-node-centos-6-4 mycluster]$ Used ceph-deploy to prepare the OSDs: [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy osd prepare ceph-node2-osd0-centos-6-4:/ceph/osd0 ceph-node3-osd1-centos-6-4:/ceph/osd1 [ceph_deploy.cli][INFO ] Invoked (1.2.7): /usr/bin/ceph-deploy osd prepare ceph-node2-osd0-centos-6-4:/ceph/osd0 ceph-node3-osd1-centos-6-4:/ceph/osd1 . . . [ceph-node3-osd1-centos-6-4][INFO ] create mon keyring file [ceph-node3-osd1-centos-6-4][INFO ] Running command: udevadm trigger --subsystem-match=block --action=""> [ceph_deploy.osd][DEBUG ] Preparing host ceph-node3-osd1-centos-6-4 disk /ceph/osd1 journal None activate False [ceph-node3-osd1-centos-6-4][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /ceph/osd1 [ceph_deploy.osd][DEBUG ] Host ceph-node3-osd1-centos-6-4 is now ready for osd use. [ceph@ceph-admin-node-centos-6-4 mycluster]$ Finally I activated the osds: [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy osd activate ceph-node2-osd0-centos-6-4:/ceph/osd0 ceph-node3-osd1-centos-6-4:/ceph/osd1 2013-11-06 14:26:53,373 [ceph_deploy.cli][INFO ] Invoked (1.3): /usr/bin/ceph-deploy osd activate ceph-node2-osd0-centos-6-4:/ceph/osd0 ceph-node3-osd1-centos-6-4:/ceph/osd1 2013-11-06 14:26:53,373 [ceph_deploy.osd][DEBUG ] Activating cluster ceph disks ceph-node2-osd0-centos-6-4:/ceph/osd0: ceph-node3-osd1-centos-6-4:/ceph/osd1: 2013-11-06 14:26:53,646 [ceph-node2-osd0-centos-6-4][DEBUG ] connected to host: ceph-node2-osd0-centos-6-4
2013-11-06 14:26:53,646 [ceph-node2-osd0-centos-6-4][DEBUG ] detect platform information from remote host 2013-11-06 14:26:53,662 [ceph-node2-osd0-centos-6-4][DEBUG ] detect machine type 2013-11-06 14:26:53,670 [ceph_deploy.osd][INFO ] Distro info: CentOS 6.4 Final 2013-11-06 14:26:53,670 [ceph_deploy.osd][DEBUG ] activating host ceph-node2-osd0-centos-6-4 disk /ceph/osd0 2013-11-06 14:26:53,670 [ceph_deploy.osd][DEBUG ] will use init type: sysvinit 2013-11-06 14:26:53,670 [ceph-node2-osd0-centos-6-4][INFO ] Running command: sudo ceph-disk-activate --mark-init sysvinit --mount /ceph/osd0 2013-11-06 14:26:53,891 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:26:54.835529 7f589c9b7700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe(0x7f5898024480 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f58980246e0).fault 2013-11-06 14:26:56,914 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:26:57.830775 7f589c8b6700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe(0x7f588c000c00 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f588c000e60).fault 2013-11-06 14:26:59,886 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:27:00.831031 7f589c9b7700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe(0x7f588c003010 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f588c003270).fault 2013-11-06 14:27:03,914 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:27:04.831257 7f589c8b6700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe(0x7f588c003a70 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f588c003cd0).fault 2013-11-06 14:27:06,886 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:27:07.831298 7f589c9b7700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe(0x7f588c005550 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f588c0057b0).fault 2013-11-06 14:27:09,909 [ceph-node2-osd0-centos-6-4][ERROR ] 2013-11-06 14:27:10.831239 7f589c8b6700 0 -- :/1019489 >> 10.12.0.70:6789/0 pipe -----Original Message----- On Thu, Nov 7, 2013 at 3:25 PM, Trivedi, Narendra <Narendra.Trivedi@xxxxxxxxxx> wrote: > I can't install Ubuntu... I am not sure why would it do on a new install of CentOS. I wanted to try this to if I can take it as RBD/Radosgw backend for OpenStack production but I can't believe it has taken forever to get it running
and I am not there yet! > This message contains information which may be confidential and/or privileged. Unless you are the intended recipient (or authorized to receive for the intended recipient), you may not read, use, copy or disclose to anyone the message or any information contained in the message. If you have received the message in error, please advise the sender by reply e-mail and delete the message and any attachment(s) thereto without retaining any copies. |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com