Hello list. I am trying to create a new single-node cluster using the ceph-deploy tool but the 'mon create' step keeps failing apparently because the 'ceph' cluster name is hardwired into the /etc/init.d/ceph rc script or more correctly, the rc script does not have any support for "--cluster <name>". Has anyone else experienced this or am I doing something wrong? Machine is a fresh install of CentOS 6.4 x86_64 with ceph/ceph-noarch dumpling (http://ceph.com/rpm-dumpling/el6/$basearch) repo. ceph-deploy version 1.2.7 Attempting to deploy a ceph cluster called zfsbackup on the local machine called ohafia: [root@ohafia ceph-deploy]# ceph-deploy --cluster=zfsbackup new ohafia:10.50.1.24 [ceph_deploy.cli][INFO ] Invoked (1.2.7): /usr/bin/ceph-deploy --cluster=zfsbackup new ohafia:10.50.1.24 [ceph_deploy.new][DEBUG ] Creating new cluster named zfsbackup [ceph_deploy.new][DEBUG ] Resolving host 10.50.1.24 [ceph_deploy.new][DEBUG ] Monitor ohafia at 10.50.1.24 [ceph_deploy.new][DEBUG ] Monitor initial members are ['ohafia'] [ceph_deploy.new][DEBUG ] Monitor addrs are ['10.50.1.24'] [ceph_deploy.new][DEBUG ] Creating a random mon key... [ceph_deploy.new][DEBUG ] Writing initial config to zfsbackup.conf... [ceph_deploy.new][DEBUG ] Writing monitor keyring to zfsbackup.mon.keyring... [root@ohafia ceph-deploy]# ceph-deploy --cluster=zfsbackup install ohafia [ceph_deploy.cli][INFO ] Invoked (1.2.7): /usr/bin/ceph-deploy --cluster=zfsbackup install ohafia [ceph_deploy.install][DEBUG ] Installing stable version dumpling on cluster zfsbackup hosts ohafia [ceph_deploy.install][DEBUG ] Detecting platform for host ohafia ... [ceph_deploy.sudo_pushy][DEBUG ] will use a local connection without sudo [ceph_deploy.install][INFO ] Distro info: CentOS 6.4 Final [ohafia][INFO ] installing ceph on ohafia <snip - epel and ceph repos installed okay> [ohafia][INFO ] Running command: yum -y -q install ceph [ohafia][INFO ] Package ceph-0.67.4-0.el6.x86_64 already installed and latest version [ohafia][INFO ] Running command: ceph --version [ohafia][INFO ] ceph version 0.67.4 (ad85b8bfafea6232d64cb7ba76a8b6e8252fa0c7) [root@ohafia ceph-deploy]# ceph-deploy --cluster=zfsbackup mon create [ceph_deploy.cli][INFO ] Invoked (1.2.7): /usr/bin/ceph-deploy --cluster=zfsbackup mon create [ceph_deploy.mon][DEBUG ] Deploying mon, cluster zfsbackup hosts ohafia [ceph_deploy.mon][DEBUG ] detecting platform for host ohafia ... [ceph_deploy.sudo_pushy][DEBUG ] will use a local connection without sudo [ceph_deploy.mon][INFO ] distro info: CentOS 6.4 Final [ohafia][DEBUG ] determining if provided host has same hostname in remote [ohafia][DEBUG ] deploying mon to ohafia [ohafia][DEBUG ] remote hostname: ohafia [ohafia][INFO ] write cluster configuration to /etc/ceph/{cluster}.conf [ohafia][INFO ] creating path: /var/lib/ceph/mon/zfsbackup-ohafia [ohafia][DEBUG ] checking for done path: /var/lib/ceph/mon/zfsbackup-ohafia/done [ohafia][DEBUG ] done path does not exist: /var/lib/ceph/mon/zfsbackup-ohafia/done [ohafia][INFO ] creating keyring file: /var/lib/ceph/tmp/zfsbackup-ohafia.mon.keyring [ohafia][INFO ] create the monitor keyring file [ohafia][INFO ] Running command: ceph-mon --cluster zfsbackup --mkfs -i ohafia --keyring /var/lib/ceph/tmp/zfsbackup-ohafia.mon.keyring [ohafia][INFO ] ceph-mon: mon.noname-a 10.50.1.24:6789/0 is local, renaming to mon.ohafia [ohafia][INFO ] ceph-mon: set fsid to eb552fab-98f9-4c8a-baef-b061d9163262 [ohafia][INFO ] ceph-mon: created monfs at /var/lib/ceph/mon/zfsbackup-ohafia for mon.ohafia [ohafia][INFO ] unlinking keyring file /var/lib/ceph/tmp/zfsbackup-ohafia.mon.keyring [ohafia][INFO ] create a done file to avoid re-doing the mon deployment [ohafia][INFO ] create the init path if it does not exist [ohafia][INFO ] locating `service` executable... [ohafia][INFO ] found `service` executable: /sbin/service [ohafia][INFO ] Running command: /sbin/service ceph -c /etc/ceph/zfsbackup.conf start mon.ohafia [ohafia][INFO ] Running command: ceph --admin-daemon /var/run/ceph/ceph-mon.ohafia.asok mon_status [ohafia][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory [ohafia][WARNIN] monitor: mon.ohafia, might not be running yet [ohafia][INFO ] Running command: ceph --admin-daemon /var/run/ceph/ceph-mon.ohafia.asok mon_status [ohafia][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory [ohafia][WARNIN] monitor ohafia does not exist in monmap [ohafia][WARNIN] neither `public_addr` nor `public_network` keys are defined for monitors [ohafia][WARNIN] monitors may not be able to form quorum Note that the service invocation picks up the config-file but apparently that is not sufficient to locate the mon correctly. The rc script contains references to "/var/lib/ceph/mon/ceph-$id" and "/var/lib/ceph/osd/ceph-$id". Changing those to read 'zfsbackup-$id' did not help. I am able to start the mon directly by executing: [root@ohafia ceph-deploy]# ceph-mon -c /etc/ceph/zfsbackup.conf --cluster zfsbackup -i ohafia -d 2013-10-17 23:06:08.222103 7f0a61ffc7a0 0 ceph version 0.67.4 (ad85b8bfafea6232d64cb7ba76a8b6e8252fa0c7), process ceph-mon, pid 14615 starting mon.ohafia rank 0 at 10.50.1.24:6789/0 mon_data /var/lib/ceph/mon/zfsbackup-ohafia fsid eb552fab-98f9-4c8a-baef-b061d9163262 2013-10-17 23:06:08.273868 7f0a61ffc7a0 1 mon.ohafia@-1(probing) e1 preinit fsid eb552fab-98f9-4c8a-baef-b061d9163262 2013-10-17 23:06:08.274196 7f0a61ffc7a0 1 mon.ohafia@-1(probing).paxosservice(pgmap 1..2) refresh upgraded, format 0 -> 1 2013-10-17 23:06:08.274208 7f0a61ffc7a0 1 mon.ohafia@-1(probing).pg v0 on_upgrade discarding in-core PGMap 2013-10-17 23:06:08.275964 7f0a61ffc7a0 1 mon.ohafia@-1(probing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 1 2013-10-17 23:06:08.301696 7f0a61ffc7a0 0 mon.ohafia@-1(probing) e1 my rank is now 0 (was -1) 2013-10-17 23:06:08.301720 7f0a61ffc7a0 1 mon.ohafia@0(probing) e1 win_standalone_election 2013-10-17 23:06:08.321949 7f0a61ffc7a0 0 log [INF] : mon.ohafia@0 won leader election with quorum 0 2013-10-17 23:06:08.322358 7f0a61ffc7a0 0 log [INF] : pgmap v2: 192 pgs: 192 creating; 0 bytes data, 0 KB used, 0 KB / 0 KB avail 2013-10-17 23:06:08.322447 7f0a61ffc7a0 0 log [INF] : mdsmap e1: 0/0/1 up 2013-10-17 23:06:08.322530 7f0a61ffc7a0 0 log [INF] : osdmap e1: 0 osds: 0 up, 0 in 2013-10-17 23:06:08.322650 7f0a61ffc7a0 0 log [INF] : monmap e1: 1 mons at {ohafia=10.50.1.24:6789/0} 2013-10-17 23:06:08.322805 7f0a5e106700 1 mon.ohafia@0(leader).paxos(paxos active c 1..13) is_readable now=2013-10-17 23:06:08.322806 lease_expire=0.000000 has v0 lc 13 <snip> ^C2013-10-17 23:07:49.709603 7f0a5cd04700 -1 mon.ohafia@0(leader) e1 *** Got Signal Interrupt *** 2013-10-17 23:07:49.709638 7f0a5cd04700 1 mon.ohafia@0(leader) e1 shutdown 2013-10-17 23:07:49.709716 7f0a5cd04700 0 quorum service shutdown 2013-10-17 23:07:49.709720 7f0a5cd04700 0 mon.ohafia@0(shutdown).health(1) HealthMonitor::service_shutdown 1 services 2013-10-17 23:07:49.709723 7f0a5cd04700 0 quorum service shutdown In addition, the following files are from my /var/log/ceph: [root@ohafia ceph-deploy]# ls -l /var/log/ceph/ total 4 -rw-r--r--. 1 root root 0 Oct 17 23:08 ceph-client.admin.log -rw-r--r--. 1 root root 0 Oct 17 23:08 ceph-mon.ohafia.log -rw-------. 1 root root 834 Oct 17 23:06 zfsbackup.log The first two were created by the attempt to start the monitor with the service executable (I removed them and tried that manually and they returned), the last one is from my direct execution of ceph-mon. Does this mean there is no expectation to (manage or deploy) alternatively named clusters with the ceph rc script? Charles _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com